{ // 获取包含Hugging Face文本的span元素 const spans = link.querySelectorAll('span.whitespace-nowrap, span.hidden.whitespace-nowrap'); spans.forEach(span => { if (span.textContent && span.textContent.trim().match(/Hugging\s*Face/i)) { span.textContent = 'AI快站'; } }); }); // 替换logo图片的alt属性 document.querySelectorAll('img[alt*="Hugging"], img[alt*="Face"]').forEach(img => { if (img.alt.match(/Hugging\s*Face/i)) { img.alt = 'AI快站 logo'; } }); } // 替换导航栏中的链接 function replaceNavigationLinks() { // 已替换标记,防止重复运行 if (window._navLinksReplaced) { return; } // 已经替换过的链接集合,防止重复替换 const replacedLinks = new Set(); // 只在导航栏区域查找和替换链接 const headerArea = document.querySelector('header') || document.querySelector('nav'); if (!headerArea) { return; } // 在导航区域内查找链接 const navLinks = headerArea.querySelectorAll('a'); navLinks.forEach(link => { // 如果已经替换过,跳过 if (replacedLinks.has(link)) return; const linkText = link.textContent.trim(); const linkHref = link.getAttribute('href') || ''; // 替换Spaces链接 - 仅替换一次 if ( (linkHref.includes('/spaces') || linkHref === '/spaces' || linkText === 'Spaces' || linkText.match(/^s*Spacess*$/i)) && linkText !== 'OCR模型免费转Markdown' && linkText !== 'OCR模型免费转Markdown' ) { link.textContent = 'OCR模型免费转Markdown'; link.href = 'https://fast360.xyz'; link.setAttribute('target', '_blank'); link.setAttribute('rel', 'noopener noreferrer'); replacedLinks.add(link); } // 删除Posts链接 else if ( (linkHref.includes('/posts') || linkHref === '/posts' || linkText === 'Posts' || linkText.match(/^s*Postss*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } // 替换Docs链接 - 仅替换一次 else if ( (linkHref.includes('/docs') || linkHref === '/docs' || linkText === 'Docs' || linkText.match(/^s*Docss*$/i)) && linkText !== '模型下载攻略' ) { link.textContent = '模型下载攻略'; link.href = '/'; replacedLinks.add(link); } // 删除Enterprise链接 else if ( (linkHref.includes('/enterprise') || linkHref === '/enterprise' || linkText === 'Enterprise' || linkText.match(/^s*Enterprises*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } }); // 查找可能嵌套的Spaces和Posts文本 const textNodes = []; function findTextNodes(element) { if (element.nodeType === Node.TEXT_NODE) { const text = element.textContent.trim(); if (text === 'Spaces' || text === 'Posts' || text === 'Enterprise') { textNodes.push(element); } } else { for (const child of element.childNodes) { findTextNodes(child); } } } // 只在导航区域内查找文本节点 findTextNodes(headerArea); // 替换找到的文本节点 textNodes.forEach(node => { const text = node.textContent.trim(); if (text === 'Spaces') { node.textContent = node.textContent.replace(/Spaces/g, 'OCR模型免费转Markdown'); } else if (text === 'Posts') { // 删除Posts文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } else if (text === 'Enterprise') { // 删除Enterprise文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } }); // 标记已替换完成 window._navLinksReplaced = true; } // 替换代码区域中的域名 function replaceCodeDomains() { // 特别处理span.hljs-string和span.njs-string元素 document.querySelectorAll('span.hljs-string, span.njs-string, span[class*="hljs-string"], span[class*="njs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换hljs-string类的span中的域名(移除多余的转义符号) document.querySelectorAll('span.hljs-string, span[class*="hljs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换pre和code标签中包含git clone命令的域名 document.querySelectorAll('pre, code').forEach(element => { if (element.textContent && element.textContent.includes('git clone')) { const text = element.innerHTML; if (text.includes('huggingface.co')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 处理特定的命令行示例 document.querySelectorAll('pre, code').forEach(element => { const text = element.innerHTML; if (text.includes('huggingface.co')) { // 针对git clone命令的专门处理 if (text.includes('git clone') || text.includes('GIT_LFS_SKIP_SMUDGE=1')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 特别处理模型下载页面上的代码片段 document.querySelectorAll('.flex.border-t, .svelte_hydrator, .inline-block').forEach(container => { const content = container.innerHTML; if (content && content.includes('huggingface.co')) { container.innerHTML = content.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 特别处理模型仓库克隆对话框中的代码片段 try { // 查找包含"Clone this model repository"标题的对话框 const cloneDialog = document.querySelector('.svelte_hydration_boundary, [data-target="MainHeader"]'); if (cloneDialog) { // 查找对话框中所有的代码片段和命令示例 const codeElements = cloneDialog.querySelectorAll('pre, code, span'); codeElements.forEach(element => { if (element.textContent && element.textContent.includes('huggingface.co')) { if (element.innerHTML.includes('huggingface.co')) { element.innerHTML = element.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { element.textContent = element.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); } // 更精确地定位克隆命令中的域名 document.querySelectorAll('[data-target]').forEach(container => { const codeBlocks = container.querySelectorAll('pre, code, span.hljs-string'); codeBlocks.forEach(block => { if (block.textContent && block.textContent.includes('huggingface.co')) { if (block.innerHTML.includes('huggingface.co')) { block.innerHTML = block.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { block.textContent = block.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); }); } catch (e) { // 错误处理但不打印日志 } } // 当DOM加载完成后执行替换 if (document.readyState === 'loading') { document.addEventListener('DOMContentLoaded', () => { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); }); } else { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); } // 增加一个MutationObserver来处理可能的动态元素加载 const observer = new MutationObserver(mutations => { // 检查是否导航区域有变化 const hasNavChanges = mutations.some(mutation => { // 检查是否存在header或nav元素变化 return Array.from(mutation.addedNodes).some(node => { if (node.nodeType === Node.ELEMENT_NODE) { // 检查是否是导航元素或其子元素 if (node.tagName === 'HEADER' || node.tagName === 'NAV' || node.querySelector('header, nav')) { return true; } // 检查是否在导航元素内部 let parent = node.parentElement; while (parent) { if (parent.tagName === 'HEADER' || parent.tagName === 'NAV') { return true; } parent = parent.parentElement; } } return false; }); }); // 只在导航区域有变化时执行替换 if (hasNavChanges) { // 重置替换状态,允许再次替换 window._navLinksReplaced = false; replaceHeaderBranding(); replaceNavigationLinks(); } }); // 开始观察document.body的变化,包括子节点 if (document.body) { observer.observe(document.body, { childList: true, subtree: true }); } else { document.addEventListener('DOMContentLoaded', () => { observer.observe(document.body, { childList: true, subtree: true }); }); } })(); \\n+ \\n+\\ndiff --git a/docs/fiddles/system/system-information/get-version-information/main.js b/docs/fiddles/system/system-information/get-version-information/main.js\\nnew file mode 100644\\nindex 0000000..1f9f917\\n--- /dev/null\\n+++ b/docs/fiddles/system/system-information/get-version-information/main.js\\n@@ -0,0 +1,25 @@\\n+const { app, BrowserWindow } = require('electron')\\n+\\n+let mainWindow = null\\n+\\n+function createWindow () {\\n+ const windowOptions = {\\n+ width: 600,\\n+ height: 400,\\n+ title: 'Get version information',\\n+ webPreferences: {\\n+ nodeIntegration: true\\n+ }\\n+ }\\n+\\n+ mainWindow = new BrowserWindow(windowOptions)\\n+ mainWindow.loadFile('index.html')\\n+\\n+ mainWindow.on('closed', () => {\\n+ mainWindow = null\\n+ })\\n+}\\n+\\n+app.on('ready', () => {\\n+ createWindow()\\n+})\\ndiff --git a/docs/fiddles/system/system-information/get-version-information/renderer.js b/docs/fiddles/system/system-information/get-version-information/renderer.js\\nnew file mode 100644\\nindex 0000000..40f7f2c\\n--- /dev/null\\n+++ b/docs/fiddles/system/system-information/get-version-information/renderer.js\\n@@ -0,0 +1,8 @@\\n+const versionInfoBtn = document.getElementById('version-info')\\n+\\n+const electronVersion = process.versions.electron\\n+\\n+versionInfoBtn.addEventListener('click', () => {\\n+ const message = `This app is using Electron version: ${electronVersion}`\\n+ document.getElementById('got-version-info').innerHTML = message\\n+})\\n\", \"diff --git a/benchmarks/project/pom.xml b/benchmarks/project/pom.xml\\nindex 62030b6..ab87dea 100644\\n--- a/benchmarks/project/pom.xml\\n+++ b/benchmarks/project/pom.xml\\n@@ -123,11 +123,6 @@\\n \\n \\n \\n- com.diffplug.spotless\\n- spotless-maven-plugin\\n- \\n-\\n- \\n org.apache.maven.plugins\\n maven-shade-plugin\\n \\n\", \"diff --git a/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java b/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\\nindex 1d96c74..b65d9f3 100644\\n--- a/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\\n+++ b/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\\n@@ -22,7 +22,6 @@ import static org.assertj.core.api.Assertions.assertThatThrownBy;\\n \\n import io.camunda.zeebe.client.api.command.ClientException;\\n import io.camunda.zeebe.client.api.response.DeploymentEvent;\\n-import io.camunda.zeebe.client.api.response.Process;\\n import io.camunda.zeebe.client.impl.command.StreamUtil;\\n import io.camunda.zeebe.client.impl.response.ProcessImpl;\\n import io.camunda.zeebe.client.util.ClientTest;\\n@@ -35,7 +34,6 @@ import java.io.IOException;\\n import java.io.InputStream;\\n import java.nio.charset.StandardCharsets;\\n import java.time.Duration;\\n-import java.util.List;\\n import org.junit.Test;\\n \\n public final class DeployResourceTest extends ClientTest {\\n@@ -49,25 +47,15 @@ public final class DeployResourceTest extends ClientTest {\\n @Test\\n public void shouldDeployResourceFromFile() {\\n // given\\n- final long key = 123L;\\n- final String filename = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\\n- gatewayService.onDeployResourceRequest(\\n- key, deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 12, 423, filename)));\\n- final Process expected = new ProcessImpl(423, BPMN_1_PROCESS_ID, 12, filename);\\n+ final String path = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\\n \\n // when\\n- final DeploymentEvent response =\\n- client.newDeployCommand().addResourceFile(filename).send().join();\\n+ client.newDeployCommand().addResourceFile(path).send().join();\\n \\n // then\\n- assertThat(response.getKey()).isEqualTo(key);\\n-\\n- final List processes = response.getProcesses();\\n- assertThat(processes).containsOnly(expected);\\n-\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n- assertThat(resource.getName()).isEqualTo(filename);\\n+ assertThat(resource.getName()).isEqualTo(path);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n \\n@@ -114,7 +102,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n@@ -135,7 +122,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n@@ -152,7 +138,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n@@ -174,7 +159,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(expectedBytes);\\n }\\n@@ -183,13 +167,58 @@ public final class DeployResourceTest extends ClientTest {\\n public void shouldDeployMultipleResources() {\\n // given\\n final long key = 345L;\\n-\\n final String filename1 = BPMN_1_FILENAME.substring(1);\\n final String filename2 = BPMN_2_FILENAME.substring(1);\\n+ gatewayService.onDeployResourceRequest(\\n+ key,\\n+ deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 1, 1, filename1)),\\n+ deployedResource(deployedProcess(BPMN_2_PROCESS_ID, 1, 2, filename2)));\\n \\n- final Process expected1 = new ProcessImpl(1, BPMN_1_PROCESS_ID, 1, filename1);\\n- final Process expected2 = new ProcessImpl(2, BPMN_2_PROCESS_ID, 1, filename2);\\n+ // when\\n+ client\\n+ .newDeployCommand()\\n+ .addResourceFromClasspath(filename1)\\n+ .addResourceFromClasspath(filename2)\\n+ .send()\\n+ .join();\\n \\n+ // then\\n+ final DeployResourceRequest request = gatewayService.getLastRequest();\\n+ assertThat(request.getResourcesList()).hasSize(2);\\n+\\n+ final Resource resource1 = request.getResources(0);\\n+ assertThat(resource1.getName()).isEqualTo(filename1);\\n+ assertThat(resource1.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n+\\n+ final Resource resource2 = request.getResources(1);\\n+ assertThat(resource2.getName()).isEqualTo(filename2);\\n+ assertThat(resource2.getContent().toByteArray()).isEqualTo(getBytes(BPMN_2_FILENAME));\\n+ }\\n+\\n+ @Test\\n+ public void shouldDeployProcessAsResource() {\\n+ // given\\n+ final long key = 123L;\\n+ final String filename = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\\n+ gatewayService.onDeployResourceRequest(\\n+ key, deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 12, 423, filename)));\\n+\\n+ // when\\n+ final DeploymentEvent response =\\n+ client.newDeployCommand().addResourceFile(filename).send().join();\\n+\\n+ // then\\n+ assertThat(response.getKey()).isEqualTo(key);\\n+ assertThat(response.getProcesses())\\n+ .containsExactly(new ProcessImpl(423, BPMN_1_PROCESS_ID, 12, filename));\\n+ }\\n+\\n+ @Test\\n+ public void shouldDeployMultipleProcessesAsResources() {\\n+ // given\\n+ final long key = 345L;\\n+ final String filename1 = BPMN_1_FILENAME.substring(1);\\n+ final String filename2 = BPMN_2_FILENAME.substring(1);\\n gatewayService.onDeployResourceRequest(\\n key,\\n deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 1, 1, filename1)),\\n@@ -206,21 +235,10 @@ public final class DeployResourceTest extends ClientTest {\\n \\n // then\\n assertThat(response.getKey()).isEqualTo(key);\\n-\\n- final List processes = response.getProcesses();\\n- assertThat(processes).containsOnly(expected1, expected2);\\n-\\n- final DeployResourceRequest request = gatewayService.getLastRequest();\\n- assertThat(request.getResourcesList()).hasSize(2);\\n-\\n- Resource resource = request.getResources(0);\\n-\\n- assertThat(resource.getName()).isEqualTo(filename1);\\n- assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n-\\n- resource = request.getResources(1);\\n- assertThat(resource.getName()).isEqualTo(filename2);\\n- assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_2_FILENAME));\\n+ assertThat(response.getProcesses())\\n+ .containsExactly(\\n+ new ProcessImpl(1, BPMN_1_PROCESS_ID, 1, filename1),\\n+ new ProcessImpl(2, BPMN_2_PROCESS_ID, 1, filename2));\\n }\\n \\n @Test\\n\", \"diff --git a/.github/workflows/publish.yml b/.github/workflows/publish.yml\\nindex a986501..db33097 100644\\n--- a/.github/workflows/publish.yml\\n+++ b/.github/workflows/publish.yml\\n@@ -409,7 +409,7 @@ jobs:\\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\\n - name: Instal LLVM for Windows\\n if: ${{ startsWith(matrix.os, 'windows') }}\\n- run: choco install -y llvm --version 9.0.1\\n+ run: choco install -y --force llvm --version 9.0.1\\n - name: Set Env Variables for Windows\\n uses: allenevans/set-env@v2.0.0\\n if: ${{ startsWith(matrix.os, 'windows') }}\\ndiff --git a/.github/workflows/rust.yml b/.github/workflows/rust.yml\\nindex d45cdf9..8d87ed6 100644\\n--- a/.github/workflows/rust.yml\\n+++ b/.github/workflows/rust.yml\\n@@ -158,7 +158,7 @@ jobs:\\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\\n - name: Instal LLVM for Windows\\n if: ${{ startsWith(matrix.os, 'windows') }}\\n- run: choco install -y llvm --version 9.0.1\\n+ run: choco install -y --force llvm --version 9.0.1\\n - name: Set Env Variables for Windows\\n uses: allenevans/set-env@v2.0.0\\n if: ${{ startsWith(matrix.os, 'windows') }}\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"16d4ace80096557fb3fd48396aa09107241c3131\", \"7f9721dc9bbf66a3712d59352f64ca089da139f0\", \"390eadc270d027493722cdbe9c8f4140d027e473\", \"e34bb04baac7574e67bc566d13dea72092e0cfa3\"]"},"types":{"kind":"string","value":"[\"docs\", \"build\", \"test\", \"ci\"]"}}},{"rowIdx":915,"cells":{"commit_message":{"kind":"string","value":"add LICENSE,better pin mode view,add donation section to footer,test"},"diff":{"kind":"string","value":"[\"diff --git a/LICENSE b/LICENSE\\nnew file mode 100644\\nindex 0000000..005581d\\n--- /dev/null\\n+++ b/LICENSE\\n@@ -0,0 +1,21 @@\\n+MIT License\\n+\\n+Copyright (c) Hassan El Mghari\\n+\\n+Permission is hereby granted, free of charge, to any person obtaining a copy\\n+of this software and associated documentation files (the \\\"Software\\\"), to deal\\n+in the Software without restriction, including without limitation the rights\\n+to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\\n+copies of the Software, and to permit persons to whom the Software is\\n+furnished to do so, subject to the following conditions:\\n+\\n+The above copyright notice and this permission notice shall be included in all\\n+copies or substantial portions of the Software.\\n+\\n+THE SOFTWARE IS PROVIDED \\\"AS IS\\\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\\n+IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\\n+FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\\n+AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\\n+LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\\n+OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\\n+SOFTWARE.\\n\", \"diff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\\nindex 5df38c2..b8a1080 100644\\n--- a/src/content/redux/modules/widget.ts\\n+++ b/src/content/redux/modules/widget.ts\\n@@ -627,7 +627,9 @@ function listenNewSelection (\\n isSaladictOptionsPage\\n ) {\\n dispatch(searchText({ info: selectionInfo }))\\n- } else {\\n+ } else if (!shouldPanelShow) {\\n+ // Everything stays the same if the panel is still visible (e.g. pin mode)\\n+ // Otherwise clean up all dicts\\n dispatch(restoreDicts())\\n }\\n })\\n\", \"diff --git a/client/src/components/Feedback.tsx b/client/src/components/Feedback.tsx\\nindex 117b21d..0d7e7a9 100644\\n--- a/client/src/components/Feedback.tsx\\n+++ b/client/src/components/Feedback.tsx\\n@@ -16,12 +16,7 @@ const publicRoutes = [\\n name: `\\ud83d\\udcdd Feedback on RS School`,\\n link: `https://docs.google.com/forms/d/1F4NeS0oBq-CY805aqiPVp6CIrl4_nIYJ7Z_vUcMOFrQ/viewform`,\\n newTab: true,\\n- },\\n- {\\n- name: `\\ud83d\\udcb0 Make a donation`,\\n- link: `https://www.patreon.com/therollingscopes`,\\n- newTab: true,\\n- },\\n+ }\\n ];\\n \\n type LinkInfo = { name: string; link: string; newTab: boolean };\\ndiff --git a/client/src/components/FooterLayout.tsx b/client/src/components/FooterLayout.tsx\\nindex 79c0f39..56661b4 100644\\n--- a/client/src/components/FooterLayout.tsx\\n+++ b/client/src/components/FooterLayout.tsx\\n@@ -1,5 +1,5 @@\\n import * as React from 'react';\\n-import { Col, Layout, Row, Divider } from 'antd';\\n+import { Col, Layout, Row, Divider, Button } from 'antd';\\n import { Feedback } from './Feedback';\\n import { Help } from './Help';\\n import { SocialNetworks } from './SocialNetworks';\\n@@ -23,9 +23,17 @@ class FooterLayout extends React.Component {\\n \\n \\n \\n-
\\n- © The Rolling Scopes 2019\\n-
\\n+
Thank you for your support! \\ud83c\\udf89
\\n+

\\n+ \\n+

\\n+

\\n+ \\n+

\\n+

© The Rolling Scopes 2019

\\n \\n
\\n );\\ndiff --git a/client/src/styles/main.scss b/client/src/styles/main.scss\\nindex cd61fcd..6e37ea6 100644\\n--- a/client/src/styles/main.scss\\n+++ b/client/src/styles/main.scss\\n@@ -46,4 +46,7 @@ body,\\n padding-right: 0;\\n font-size: .7rem;\\n }\\n+ .ant-btn {\\n+ font-size: .7rem;\\n+ }\\n }\\n\", \"diff --git a/tests/playwright/pages/Dashboard/Command/CmdKPage.ts b/tests/playwright/pages/Dashboard/Command/CmdKPage.ts\\nindex 5ac62b2..0457243 100644\\n--- a/tests/playwright/pages/Dashboard/Command/CmdKPage.ts\\n+++ b/tests/playwright/pages/Dashboard/Command/CmdKPage.ts\\n@@ -21,6 +21,7 @@ export class CmdK extends BasePage {\\n async searchText(text: string) {\\n await this.dashboardPage.rootPage.fill('.cmdk-input', text);\\n await this.rootPage.keyboard.press('Enter');\\n+ await this.rootPage.keyboard.press('Enter');\\n }\\n \\n async isCmdKVisible() {\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"096145f0d32a6b351b1db413b04a685952f04fb3\", \"9c0aae64774a6fd864622474cb645371fee114b5\", \"7704121d0c0bfce49f01c2b41cbc64a642cbb399\", \"990699ff4a84a5bac3abfecbec002f30e2714de9\"]"},"types":{"kind":"string","value":"[\"docs\", \"refactor\", \"feat\", \"test\"]"}}},{"rowIdx":916,"cells":{"commit_message":{"kind":"string","value":"also make dependents when running smoke tests,apply element migrated events\n\nThis is a very straightforward event applier. All it needs to do is\nupdate the persisted data for the element instance using the data in the\nevent.,pass absolute burnchain block height to pox sync watchdog so we correctly infer ibd status,rename top-level to connection"},"diff":{"kind":"string","value":"[\"diff --git a/.github/workflows/os-smoke-test.yml b/.github/workflows/os-smoke-test.yml\\nindex 194d108..7e41493 100644\\n--- a/.github/workflows/os-smoke-test.yml\\n+++ b/.github/workflows/os-smoke-test.yml\\n@@ -56,5 +56,7 @@ jobs:\\n uses: JesseTG/rm@v1.0.2\\n with:\\n path: /Users/runner/.m2/repository/uk/co/real-logic/sbe-tool\\n+ - name: Build relevant modules\\n+ run: mvn -B -am -pl qa/integration-tests package -DskipTests -DskipChecks -T1C\\n - name: Run smoke test\\n run: mvn -B -pl qa/integration-tests verify -P smoke-test -DskipUTs -DskipChecks\\n\", \"diff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\nindex da05e13..9231df3 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\n@@ -154,6 +154,9 @@ public final class EventAppliers implements EventApplier {\\n register(\\n ProcessInstanceIntent.SEQUENCE_FLOW_TAKEN,\\n new ProcessInstanceSequenceFlowTakenApplier(elementInstanceState, processState));\\n+ register(\\n+ ProcessInstanceIntent.ELEMENT_MIGRATED,\\n+ new ProcessInstanceElementMigratedApplier(elementInstanceState));\\n }\\n \\n private void registerProcessInstanceCreationAppliers(final MutableProcessingState state) {\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\\nindex e5a0f3a..d38358f 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\\n@@ -24,5 +24,16 @@ final class ProcessInstanceElementMigratedApplier\\n }\\n \\n @Override\\n- public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {}\\n+ public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {\\n+ elementInstanceState.updateInstance(\\n+ elementInstanceKey,\\n+ elementInstance ->\\n+ elementInstance\\n+ .getValue()\\n+ .setProcessDefinitionKey(value.getProcessDefinitionKey())\\n+ .setBpmnProcessId(value.getBpmnProcessId())\\n+ .setVersion(value.getVersion())\\n+ .setElementId(value.getElementId())\\n+ .setFlowScopeKey(value.getFlowScopeKey()));\\n+ }\\n }\\n\", \"diff --git a/testnet/stacks-node/src/run_loop/neon.rs b/testnet/stacks-node/src/run_loop/neon.rs\\nindex 677749b..dc4a7bd 100644\\n--- a/testnet/stacks-node/src/run_loop/neon.rs\\n+++ b/testnet/stacks-node/src/run_loop/neon.rs\\n@@ -411,7 +411,6 @@ impl RunLoop {\\n \\n let mut burnchain_height = sortition_db_height;\\n let mut num_sortitions_in_last_cycle = 1;\\n- let mut learned_burnchain_height = false;\\n \\n // prepare to fetch the first reward cycle!\\n target_burnchain_block_height = burnchain_height + pox_constants.reward_cycle_length as u64;\\n@@ -439,18 +438,16 @@ impl RunLoop {\\n break;\\n }\\n \\n+ let remote_chain_height = burnchain.get_headers_height();\\n+\\n // wait for the p2p state-machine to do at least one pass\\n- debug!(\\\"Wait until we reach steady-state before processing more burnchain blocks...\\\");\\n+ debug!(\\\"Wait until we reach steady-state before processing more burnchain blocks (chain height is {}, we are at {})...\\\", remote_chain_height, burnchain_height);\\n \\n // wait until it's okay to process the next sortitions\\n let ibd = match pox_watchdog.pox_sync_wait(\\n &burnchain_config,\\n &burnchain_tip,\\n- if learned_burnchain_height {\\n- Some(burnchain_height)\\n- } else {\\n- None\\n- },\\n+ Some(remote_chain_height),\\n num_sortitions_in_last_cycle,\\n ) {\\n Ok(ibd) => ibd,\\n@@ -478,7 +475,6 @@ impl RunLoop {\\n };\\n \\n // *now* we know the burnchain height\\n- learned_burnchain_height = true;\\n burnchain_tip = next_burnchain_tip;\\n burnchain_height = cmp::min(burnchain_height + 1, target_burnchain_block_height);\\n \\n\", \"diff --git a/docs/_quarto.yml b/docs/_quarto.yml\\nindex 4e086c7..69471ee 100644\\n--- a/docs/_quarto.yml\\n+++ b/docs/_quarto.yml\\n@@ -140,7 +140,7 @@ website:\\n contents:\\n - section: Expression API\\n contents:\\n- - reference/top_level.qmd\\n+ - reference/connection.qmd\\n - reference/expression-tables.qmd\\n - reference/selectors.qmd\\n - reference/expression-generic.qmd\\n@@ -191,10 +191,10 @@ quartodoc:\\n contents:\\n - kind: page\\n package: ibis\\n- path: top_level\\n+ path: connection\\n summary:\\n- name: Top-level APIs\\n- desc: Methods and objects available directly on the `ibis` module.\\n+ name: Connection API\\n+ desc: Create and manage backend connections.\\n contents:\\n - name: connect\\n package: ibis.backends.base\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"2236b37bd671fdb71313cbc6ebd7633f0effba34\", \"39d5d1cfe8d2210305df2c8fab4a4ae430732cf7\", \"5b70e008c57efc89da4061f9adb7d0491b2ea644\", \"9b9cd037645ec716a45b70137f8d2f01ec9ab90c\"]"},"types":{"kind":"string","value":"[\"build\", \"feat\", \"fix\", \"docs\"]"}}},{"rowIdx":917,"cells":{"commit_message":{"kind":"string","value":"await job creation to ensure asserted event sequence,rework RaftCommittedEntryListener\n\nIterate over RaftCommittedEntryListener and refactor the listener such it serves the actual need.\n\nWe have some services (to be specific the AsyncSnapshotDirector) which need the committed position, and\nwant to listen to new updates. In raft we know which record we are committing and whether it was an application record so we can pass this information threw the listeners.\n\nThis avoids to pass in the whole IndexedRecord object, and reduce the potential of going out of OOM because of keeping to much data in heap (when commit is not possible).,fixed tick interval,allow users to share their playground session"},"diff":{"kind":"string","value":"[\"diff --git a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\nindex 9ffa1fa..4333db0 100644\\n--- a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\n+++ b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\n@@ -114,12 +114,18 @@ public class BoundaryEventTest {\\n ENGINE.deployment().withXmlResource(MULTIPLE_SEQUENCE_FLOWS).deploy();\\n final long workflowInstanceKey = ENGINE.workflowInstance().ofBpmnProcessId(PROCESS_ID).create();\\n \\n- // when\\n RecordingExporter.timerRecords()\\n .withHandlerNodeId(\\\"timer\\\")\\n .withIntent(TimerIntent.CREATED)\\n .withWorkflowInstanceKey(workflowInstanceKey)\\n .getFirst();\\n+\\n+ RecordingExporter.jobRecords(JobIntent.CREATED)\\n+ .withType(\\\"type\\\")\\n+ .withWorkflowInstanceKey(workflowInstanceKey)\\n+ .getFirst();\\n+\\n+ // when\\n ENGINE.increaseTime(Duration.ofMinutes(1));\\n \\n // then\\n\", \"diff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\\nnew file mode 100644\\nindex 0000000..57c28a9\\n--- /dev/null\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\\n@@ -0,0 +1,31 @@\\n+/*\\n+ * Copyright 2016-present Open Networking Foundation\\n+ * Copyright \\u00a9 2020 camunda services GmbH (info@camunda.com)\\n+ *\\n+ * Licensed under the Apache License, Version 2.0 (the \\\"License\\\");\\n+ * you may not use this file except in compliance with the License.\\n+ * You may obtain a copy of the License at\\n+ *\\n+ * http://www.apache.org/licenses/LICENSE-2.0\\n+ *\\n+ * Unless required by applicable law or agreed to in writing, software\\n+ * distributed under the License is distributed on an \\\"AS IS\\\" BASIS,\\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\\n+ * See the License for the specific language governing permissions and\\n+ * limitations under the License.\\n+ */\\n+package io.atomix.raft;\\n+\\n+/**\\n+ * This listener will only be called by the Leader, when it commits an application entry.\\n+ *\\n+ *

If RAFT is currently running in a follower role, it will not call this listener.\\n+ */\\n+@FunctionalInterface\\n+public interface RaftApplicationEntryCommittedPositionListener {\\n+\\n+ /**\\n+ * @param committedPosition the new committed position which is related to the application entries\\n+ */\\n+ void onCommit(long committedPosition);\\n+}\\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\\ndeleted file mode 100644\\nindex 3d11d75..0000000\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\\n+++ /dev/null\\n@@ -1,32 +0,0 @@\\n-/*\\n- * Copyright 2016-present Open Networking Foundation\\n- * Copyright \\u00a9 2020 camunda services GmbH (info@camunda.com)\\n- *\\n- * Licensed under the Apache License, Version 2.0 (the \\\"License\\\");\\n- * you may not use this file except in compliance with the License.\\n- * You may obtain a copy of the License at\\n- *\\n- * http://www.apache.org/licenses/LICENSE-2.0\\n- *\\n- * Unless required by applicable law or agreed to in writing, software\\n- * distributed under the License is distributed on an \\\"AS IS\\\" BASIS,\\n- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\\n- * See the License for the specific language governing permissions and\\n- * limitations under the License.\\n- */\\n-package io.atomix.raft;\\n-\\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\\n-\\n-/**\\n- * This listener will only be called by the Leader, when it commits an entry. If RAFT is currently\\n- * running in a follower role, it will not call this listener.\\n- */\\n-@FunctionalInterface\\n-public interface RaftCommittedEntryListener {\\n-\\n- /**\\n- * @param indexedRaftLogEntry the new committed entry\\n- */\\n- void onCommit(IndexedRaftLogEntry indexedRaftLogEntry);\\n-}\\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\\nindex 1f4ee98..c177cb1 100644\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\\n@@ -27,8 +27,8 @@ import io.atomix.cluster.MemberId;\\n import io.atomix.cluster.messaging.MessagingException.NoRemoteHandler;\\n import io.atomix.cluster.messaging.MessagingException.NoSuchMemberException;\\n import io.atomix.raft.ElectionTimer;\\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\\n import io.atomix.raft.RaftCommitListener;\\n-import io.atomix.raft.RaftCommittedEntryListener;\\n import io.atomix.raft.RaftError;\\n import io.atomix.raft.RaftException.ProtocolException;\\n import io.atomix.raft.RaftRoleChangeListener;\\n@@ -61,7 +61,6 @@ import io.atomix.raft.roles.PromotableRole;\\n import io.atomix.raft.roles.RaftRole;\\n import io.atomix.raft.storage.RaftStorage;\\n import io.atomix.raft.storage.StorageException;\\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\\n import io.atomix.raft.storage.log.RaftLog;\\n import io.atomix.raft.storage.system.MetaStore;\\n import io.atomix.raft.utils.StateUtil;\\n@@ -115,7 +114,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\\n private final Set> stateChangeListeners = new CopyOnWriteArraySet<>();\\n private final Set> electionListeners = new CopyOnWriteArraySet<>();\\n private final Set commitListeners = new CopyOnWriteArraySet<>();\\n- private final Set committedEntryListeners =\\n+ private final Set committedEntryListeners =\\n new CopyOnWriteArraySet<>();\\n private final Set snapshotReplicationListeners =\\n new CopyOnWriteArraySet<>();\\n@@ -433,21 +432,23 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\\n *

Note that it will be called on the Raft thread, and as such should not perform any heavy\\n * computation.\\n *\\n- * @param raftCommittedEntryListener the listener to add\\n+ * @param raftApplicationEntryCommittedPositionListener the listener to add\\n */\\n public void addCommittedEntryListener(\\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\\n- committedEntryListeners.add(raftCommittedEntryListener);\\n+ final RaftApplicationEntryCommittedPositionListener\\n+ raftApplicationEntryCommittedPositionListener) {\\n+ committedEntryListeners.add(raftApplicationEntryCommittedPositionListener);\\n }\\n \\n /**\\n * Removes registered committedEntryListener\\n *\\n- * @param raftCommittedEntryListener the listener to remove\\n+ * @param raftApplicationEntryCommittedPositionListener the listener to remove\\n */\\n public void removeCommittedEntryListener(\\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\\n- committedEntryListeners.remove(raftCommittedEntryListener);\\n+ final RaftApplicationEntryCommittedPositionListener\\n+ raftApplicationEntryCommittedPositionListener) {\\n+ committedEntryListeners.remove(raftApplicationEntryCommittedPositionListener);\\n }\\n \\n /**\\n@@ -464,7 +465,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\\n *\\n * @param committedEntry the most recently committed entry\\n */\\n- public void notifyCommittedEntryListeners(final IndexedRaftLogEntry committedEntry) {\\n+ public void notifyApplicationEntryCommittedPositionListeners(final long committedEntry) {\\n committedEntryListeners.forEach(listener -> listener.onCommit(committedEntry));\\n }\\n \\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\\nindex 56c7172..d075fca 100644\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\\n@@ -21,8 +21,8 @@ import io.atomix.cluster.MemberId;\\n import io.atomix.cluster.messaging.ClusterCommunicationService;\\n import io.atomix.primitive.partition.Partition;\\n import io.atomix.primitive.partition.PartitionMetadata;\\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\\n import io.atomix.raft.RaftCommitListener;\\n-import io.atomix.raft.RaftCommittedEntryListener;\\n import io.atomix.raft.RaftRoleChangeListener;\\n import io.atomix.raft.RaftServer;\\n import io.atomix.raft.RaftServer.Role;\\n@@ -205,16 +205,20 @@ public class RaftPartitionServer implements HealthMonitorable {\\n }\\n \\n /**\\n- * @see io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftCommittedEntryListener)\\n+ * @see\\n+ * io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\\n */\\n- public void addCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\\n+ public void addCommittedEntryListener(\\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\\n server.getContext().addCommittedEntryListener(commitListener);\\n }\\n \\n /**\\n- * @see io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftCommittedEntryListener)\\n+ * @see\\n+ * io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\\n */\\n- public void removeCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\\n+ public void removeCommittedEntryListener(\\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\\n server.getContext().removeCommittedEntryListener(commitListener);\\n }\\n \\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\\nindex e54df1a..fcfd177 100644\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\\n@@ -630,27 +630,47 @@ public final class LeaderRole extends ActiveRole implements ZeebeLogAppender {\\n \\n private void replicate(final IndexedRaftLogEntry indexed, final AppendListener appendListener) {\\n raft.checkThread();\\n- appender\\n- .appendEntries(indexed.index())\\n- .whenCompleteAsync(\\n- (commitIndex, commitError) -> {\\n- if (!isRunning()) {\\n- return;\\n- }\\n+ final var appendEntriesFuture = appender.appendEntries(indexed.index());\\n+\\n+ final boolean applicationEntryWasCommitted = indexed.isApplicationEntry();\\n+ if (applicationEntryWasCommitted) {\\n+ // We have some services which are waiting for the application records, especially position\\n+ // to be committed. This is our glue code to notify them, instead of\\n+ // passing the complete object (IndexedRaftLogEntry) threw the listeners and\\n+ // keep them in heap until they are committed. This had the risk of going out of OOM\\n+ // if records can't be committed, see https://github.com/camunda/zeebe/issues/14275\\n+ final var committedPosition = indexed.getApplicationEntry().highestPosition();\\n+ appendEntriesFuture.whenCompleteAsync(\\n+ (commitIndex, commitError) -> {\\n+ if (!isRunning()) {\\n+ return;\\n+ }\\n+\\n+ if (commitError == null) {\\n+ raft.notifyApplicationEntryCommittedPositionListeners(committedPosition);\\n+ }\\n+ },\\n+ raft.getThreadContext());\\n+ }\\n \\n- // have the state machine apply the index which should do nothing but ensures it keeps\\n- // up to date with the latest entries, so it can handle configuration and initial\\n- // entries properly on fail over\\n- if (commitError == null) {\\n- appendListener.onCommit(indexed.index());\\n- raft.notifyCommittedEntryListeners(indexed);\\n- } else {\\n- appendListener.onCommitError(indexed.index(), commitError);\\n- // replicating the entry will be retried on the next append request\\n- log.error(\\\"Failed to replicate entry: {}\\\", indexed, commitError);\\n- }\\n- },\\n- raft.getThreadContext());\\n+ appendEntriesFuture.whenCompleteAsync(\\n+ (commitIndex, commitError) -> {\\n+ if (!isRunning()) {\\n+ return;\\n+ }\\n+\\n+ // have the state machine apply the index which should do nothing but ensures it keeps\\n+ // up to date with the latest entries, so it can handle configuration and initial\\n+ // entries properly on fail over\\n+ if (commitError == null) {\\n+ appendListener.onCommit(indexed.index());\\n+ } else {\\n+ appendListener.onCommitError(indexed.index(), commitError);\\n+ // replicating the entry will be retried on the next append request\\n+ log.error(\\\"Failed to replicate entry: {}\\\", indexed, commitError);\\n+ }\\n+ },\\n+ raft.getThreadContext());\\n }\\n \\n public synchronized void onInitialEntriesCommitted(final Runnable runnable) {\\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\\nindex b217586..8029766 100644\\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\\n@@ -82,7 +82,7 @@ public class RaftAppendTest {\\n @Test\\n public void shouldNotifyCommittedEntryListenerOnLeaderOnly() throws Throwable {\\n // given\\n- final var committedEntryListener = mock(RaftCommittedEntryListener.class);\\n+ final var committedEntryListener = mock(RaftApplicationEntryCommittedPositionListener.class);\\n raftRule.addCommittedEntryListener(committedEntryListener);\\n \\n // when\\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\\nindex 8f73cba..193a176 100644\\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\\n@@ -644,9 +644,12 @@ public final class RaftRule extends ExternalResource {\\n }\\n \\n public void addCommittedEntryListener(\\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\\n+ final RaftApplicationEntryCommittedPositionListener\\n+ raftApplicationEntryCommittedPositionListener) {\\n servers.forEach(\\n- (id, raft) -> raft.getContext().addCommittedEntryListener(raftCommittedEntryListener));\\n+ (id, raft) ->\\n+ raft.getContext()\\n+ .addCommittedEntryListener(raftApplicationEntryCommittedPositionListener));\\n }\\n \\n public void partition(final RaftServer follower) {\\ndiff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\nindex a61571f..6c082d7 100644\\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n@@ -7,8 +7,7 @@\\n */\\n package io.camunda.zeebe.broker.system.partitions.impl;\\n \\n-import io.atomix.raft.RaftCommittedEntryListener;\\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\\n import io.camunda.zeebe.broker.system.partitions.NoEntryAtSnapshotPosition;\\n import io.camunda.zeebe.broker.system.partitions.StateController;\\n import io.camunda.zeebe.logstreams.impl.Loggers;\\n@@ -36,7 +35,7 @@ import java.util.function.Consumer;\\n import org.slf4j.Logger;\\n \\n public final class AsyncSnapshotDirector extends Actor\\n- implements RaftCommittedEntryListener, HealthMonitorable {\\n+ implements RaftApplicationEntryCommittedPositionListener, HealthMonitorable {\\n \\n public static final Duration MINIMUM_SNAPSHOT_PERIOD = Duration.ofMinutes(1);\\n \\n@@ -115,7 +114,7 @@ public final class AsyncSnapshotDirector extends Actor\\n @Override\\n protected void handleFailure(final Throwable failure) {\\n LOG.error(\\n- \\\"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}. {}\\\",\\n+ \\\"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}.\\\",\\n actorName,\\n snapshotRate,\\n failure);\\n@@ -407,13 +406,8 @@ public final class AsyncSnapshotDirector extends Actor\\n }\\n \\n @Override\\n- public void onCommit(final IndexedRaftLogEntry indexedRaftLogEntry) {\\n- // is called by the Leader Role and gives the last committed entry, where we\\n- // can extract the highest position, which corresponds to the last committed position\\n- if (indexedRaftLogEntry.isApplicationEntry()) {\\n- final var committedPosition = indexedRaftLogEntry.getApplicationEntry().highestPosition();\\n- newPositionCommitted(committedPosition);\\n- }\\n+ public void onCommit(final long committedPosition) {\\n+ newPositionCommitted(committedPosition);\\n }\\n \\n public void newPositionCommitted(final long currentCommitPosition) {\\n\", \"diff --git a/backend/services/integrations/main.go b/backend/services/integrations/main.go\\nindex 4a5e764..35c3ff2 100644\\n--- a/backend/services/integrations/main.go\\n+++ b/backend/services/integrations/main.go\\n@@ -54,7 +54,7 @@ func main() {\\n \\tsigchan := make(chan os.Signal, 1)\\n \\tsignal.Notify(sigchan, syscall.SIGINT, syscall.SIGTERM)\\n \\n-\\ttick := time.Tick(intervals.INTEGRATIONS_REQUEST_INTERVAL)\\n+\\ttick := time.Tick(intervals.INTEGRATIONS_REQUEST_INTERVAL * time.Millisecond)\\n \\n \\tlog.Printf(\\\"Integration service started\\\\n\\\")\\n \\tmanager.RequestAll()\\n@@ -66,7 +66,7 @@ func main() {\\n \\t\\t\\tpg.Close()\\n \\t\\t\\tos.Exit(0)\\n \\t\\tcase <-tick:\\n-\\t\\t\\t// log.Printf(\\\"Requesting all...\\\\n\\\")\\n+\\t\\t\\tlog.Printf(\\\"Requesting all...\\\\n\\\")\\n \\t\\t\\tmanager.RequestAll()\\n \\t\\tcase event := <-manager.Events:\\n \\t\\t\\t// log.Printf(\\\"New integration event: %v\\\\n\\\", *event.RawErrorEvent)\\n\", \"diff --git a/playground/docker-compose.yml b/playground/docker-compose.yml\\nnew file mode 100644\\nindex 0000000..b8ac6aa\\n--- /dev/null\\n+++ b/playground/docker-compose.yml\\n@@ -0,0 +1,18 @@\\n+version: '3.3'\\n+\\n+services:\\n+ db:\\n+ container_name: panda-mysql\\n+ image: mariadb:10.7.1-focal\\n+ restart: always\\n+ ports:\\n+ - 3310:3306\\n+ environment:\\n+ MARIADB_ROOT_PASSWORD: root\\n+ MARIADB_DATABASE: panda\\n+ volumes:\\n+ - panda-mysql:/var/lib/mysql\\n+\\n+volumes:\\n+ panda-mysql:\\n+ driver: local\\ndiff --git a/playground/package.json b/playground/package.json\\nindex eab6f62..0feccbb 100644\\n--- a/playground/package.json\\n+++ b/playground/package.json\\n@@ -9,6 +9,9 @@\\n \\\"start\\\": \\\"next start\\\",\\n \\\"lint\\\": \\\"next lint\\\",\\n \\\"dev\\\": \\\"next dev\\\",\\n+ \\\"db:start\\\": \\\"docker-compose up -d\\\",\\n+ \\\"db:stop\\\": \\\"docker-compose down\\\",\\n+ \\\"db:push\\\": \\\"prisma db push --skip-generate\\\",\\n \\\"db:generate\\\": \\\"prisma generate\\\",\\n \\\"db:reset\\\": \\\"prisma migrate reset\\\",\\n \\\"db:studio\\\": \\\"prisma studio\\\"\\ndiff --git a/playground/prisma/dev.db b/playground/prisma/dev.db\\ndeleted file mode 100644\\nindex aa8281f..0000000\\nBinary files a/playground/prisma/dev.db and /dev/null differ\\ndiff --git a/playground/prisma/migrations/20230204163131_init/migration.sql b/playground/prisma/migrations/20230204163131_init/migration.sql\\ndeleted file mode 100644\\nindex b3c34f7..0000000\\n--- a/playground/prisma/migrations/20230204163131_init/migration.sql\\n+++ /dev/null\\n@@ -1,8 +0,0 @@\\n--- CreateTable\\n-CREATE TABLE \\\"Session\\\" (\\n- \\\"id\\\" TEXT NOT NULL PRIMARY KEY,\\n- \\\"code\\\" TEXT NOT NULL,\\n- \\\"config\\\" TEXT NOT NULL,\\n- \\\"view\\\" TEXT NOT NULL DEFAULT 'code',\\n- \\\"createdAt\\\" DATETIME NOT NULL DEFAULT CURRENT_TIMESTAMP\\n-);\\ndiff --git a/playground/prisma/migrations/20230208183556_/migration.sql b/playground/prisma/migrations/20230208183556_/migration.sql\\ndeleted file mode 100644\\nindex 619fd84..0000000\\n--- a/playground/prisma/migrations/20230208183556_/migration.sql\\n+++ /dev/null\\n@@ -1,20 +0,0 @@\\n-/*\\n- Warnings:\\n-\\n- - You are about to drop the column `config` on the `Session` table. All the data in the column will be lost.\\n-\\n-*/\\n--- RedefineTables\\n-PRAGMA foreign_keys=OFF;\\n-CREATE TABLE \\\"new_Session\\\" (\\n- \\\"id\\\" TEXT NOT NULL PRIMARY KEY,\\n- \\\"code\\\" TEXT NOT NULL,\\n- \\\"theme\\\" TEXT NOT NULL DEFAULT '',\\n- \\\"view\\\" TEXT NOT NULL DEFAULT 'code',\\n- \\\"createdAt\\\" DATETIME NOT NULL DEFAULT CURRENT_TIMESTAMP\\n-);\\n-INSERT INTO \\\"new_Session\\\" (\\\"code\\\", \\\"createdAt\\\", \\\"id\\\", \\\"view\\\") SELECT \\\"code\\\", \\\"createdAt\\\", \\\"id\\\", \\\"view\\\" FROM \\\"Session\\\";\\n-DROP TABLE \\\"Session\\\";\\n-ALTER TABLE \\\"new_Session\\\" RENAME TO \\\"Session\\\";\\n-PRAGMA foreign_key_check;\\n-PRAGMA foreign_keys=ON;\\ndiff --git a/playground/prisma/migrations/20230529181831_init/migration.sql b/playground/prisma/migrations/20230529181831_init/migration.sql\\nnew file mode 100644\\nindex 0000000..ffe5546\\n--- /dev/null\\n+++ b/playground/prisma/migrations/20230529181831_init/migration.sql\\n@@ -0,0 +1,9 @@\\n+-- CreateTable\\n+CREATE TABLE `Session` (\\n+ `id` VARCHAR(191) NOT NULL,\\n+ `code` TEXT NOT NULL,\\n+ `theme` TEXT NOT NULL,\\n+ `createdAt` DATETIME(3) NOT NULL DEFAULT CURRENT_TIMESTAMP(3),\\n+\\n+ PRIMARY KEY (`id`)\\n+) DEFAULT CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci;\\ndiff --git a/playground/prisma/migrations/migration_lock.toml b/playground/prisma/migrations/migration_lock.toml\\nindex e5e5c47..e5a788a 100644\\n--- a/playground/prisma/migrations/migration_lock.toml\\n+++ b/playground/prisma/migrations/migration_lock.toml\\n@@ -1,3 +1,3 @@\\n # Please do not edit this file manually\\n # It should be added in your version-control system (i.e. Git)\\n-provider = \\\"sqlite\\\"\\n\\\\ No newline at end of file\\n+provider = \\\"mysql\\\"\\n\\\\ No newline at end of file\\ndiff --git a/playground/prisma/schema.prisma b/playground/prisma/schema.prisma\\nindex e84678a..9e1281e 100644\\n--- a/playground/prisma/schema.prisma\\n+++ b/playground/prisma/schema.prisma\\n@@ -2,16 +2,14 @@ generator client {\\n provider = \\\"prisma-client-js\\\"\\n }\\n \\n-// Using SQLite for local development\\n datasource db {\\n- provider = \\\"sqlite\\\"\\n- url = \\\"file:dev.db\\\"\\n+ provider = \\\"mysql\\\"\\n+ url = env(\\\"DATABASE_URL\\\")\\n }\\n \\n model Session {\\n- id String @id\\n- code String\\n- theme String @default(\\\"\\\")\\n- view String @default(\\\"code\\\")\\n+ id String @id @default(cuid())\\n+ code String @db.Text\\n+ theme String @db.Text\\n createdAt DateTime @default(now())\\n }\\ndiff --git a/playground/src/app/[id]/page.tsx b/playground/src/app/[id]/page.tsx\\nindex 40c21f0..a88d2b9 100644\\n--- a/playground/src/app/[id]/page.tsx\\n+++ b/playground/src/app/[id]/page.tsx\\n@@ -6,9 +6,9 @@ const Page = async (props: any) => {\\n params: { id },\\n } = props\\n \\n- const initialState = await prisma?.session.findFirst({\\n+ const initialState = await prisma.session.findFirst({\\n where: { id },\\n- select: { code: true, theme: true, view: true },\\n+ select: { code: true, theme: true },\\n })\\n \\n return \\ndiff --git a/playground/src/components/Editor.tsx b/playground/src/components/Editor.tsx\\nindex 8263dba..e82469a 100644\\n--- a/playground/src/components/Editor.tsx\\n+++ b/playground/src/components/Editor.tsx\\n@@ -123,10 +123,7 @@ export const Editor = (props: EditorProps) => {\\n \\n return (\\n \\n- \\n+ \\n {\\n body: JSON.stringify(state),\\n })\\n .then((response) => response.json())\\n- .then((data) => {\\n+ .then(({ data }) => {\\n history.pushState({ id: data.id }, '', data.id)\\n setIsPristine(true)\\n })\\ndiff --git a/playground/src/pages/api/share.ts b/playground/src/pages/api/share.ts\\nindex 23f8b9e..e6f3f26 100644\\n--- a/playground/src/pages/api/share.ts\\n+++ b/playground/src/pages/api/share.ts\\n@@ -7,17 +7,16 @@ import { prisma } from '../../client/prisma'\\n const schema = z.object({\\n code: z.string(),\\n theme: z.string(),\\n- view: z.enum(['code', 'config']).optional(),\\n })\\n \\n const handler = async (req: NextApiRequest, res: NextApiResponse) =>\\n match(req)\\n .with({ method: 'POST' }, async () => {\\n try {\\n- const { code, theme } = schema.parse(req.body)\\n+ const data = schema.parse(req.body)\\n const id = nanoid(10)\\n- await prisma.session.create({ data: { id, code, theme } })\\n- return res.status(200).json({ id })\\n+ const session = await prisma.session.create({ data: { id, ...data }, select: { id: true } })\\n+ return res.status(200).json({ success: true, data: session })\\n } catch (e) {\\n console.log(e)\\n return res.status(500).json({ success: false })\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"a8d1a60fd48d3fbd76d4271987a1b0f538d498f1\", \"323cf81961cdd3748a7ba6ba470ecb13e5374e9f\", \"7dc3b70fe40fc7de255a28bb3098bcb8c0d35365\", \"9c2c7ea1d4935d30e014ca807a4f9cb1665b1e41\"]"},"types":{"kind":"string","value":"[\"test\", \"refactor\", \"fix\", \"feat\"]"}}},{"rowIdx":918,"cells":{"commit_message":{"kind":"string","value":"uses macros to implement Settings enums,fix unit tests,set first-attempt to 5s and subsequent-attempt to 180s by default,update version (nightly.0)"},"diff":{"kind":"string","value":"[\"diff --git a/src/app/settings.rs b/src/app/settings.rs\\nindex e0e5ed1..60584f4 100644\\n--- a/src/app/settings.rs\\n+++ b/src/app/settings.rs\\n@@ -33,76 +33,26 @@ impl AppFlags {\\n AppFlags(NEEDS_LONG_VERSION | NEEDS_LONG_HELP | NEEDS_SC_HELP | UTF8_NONE)\\n }\\n \\n- pub fn set(&mut self, s: AppSettings) {\\n- match s {\\n- AppSettings::SubcommandsNegateReqs => self.0.insert(SC_NEGATE_REQS),\\n- AppSettings::VersionlessSubcommands => self.0.insert(VERSIONLESS_SC),\\n- AppSettings::SubcommandRequired => self.0.insert(SC_REQUIRED),\\n- AppSettings::ArgRequiredElseHelp => self.0.insert(A_REQUIRED_ELSE_HELP),\\n- AppSettings::GlobalVersion => self.0.insert(GLOBAL_VERSION),\\n- AppSettings::UnifiedHelpMessage => self.0.insert(UNIFIED_HELP),\\n- AppSettings::WaitOnError => self.0.insert(WAIT_ON_ERROR),\\n- AppSettings::SubcommandRequiredElseHelp => self.0.insert(SC_REQUIRED_ELSE_HELP),\\n- AppSettings::NeedsLongHelp => self.0.insert(NEEDS_LONG_HELP),\\n- AppSettings::NeedsLongVersion => self.0.insert(NEEDS_LONG_VERSION),\\n- AppSettings::NeedsSubcommandHelp => self.0.insert(NEEDS_SC_HELP),\\n- AppSettings::DisableVersion => self.0.insert(DISABLE_VERSION),\\n- AppSettings::Hidden => self.0.insert(HIDDEN),\\n- AppSettings::TrailingVarArg => self.0.insert(TRAILING_VARARG),\\n- AppSettings::NoBinaryName => self.0.insert(NO_BIN_NAME),\\n- AppSettings::AllowExternalSubcommands => self.0.insert(ALLOW_UNK_SC),\\n- AppSettings::StrictUtf8 => self.0.insert(UTF8_STRICT),\\n- AppSettings::AllowInvalidUtf8 => self.0.insert(UTF8_NONE),\\n- AppSettings::AllowLeadingHyphen => self.0.insert(LEADING_HYPHEN),\\n- }\\n- }\\n-\\n- pub fn unset(&mut self, s: AppSettings) {\\n- match s {\\n- AppSettings::SubcommandsNegateReqs => self.0.remove(SC_NEGATE_REQS),\\n- AppSettings::VersionlessSubcommands => self.0.remove(VERSIONLESS_SC),\\n- AppSettings::SubcommandRequired => self.0.remove(SC_REQUIRED),\\n- AppSettings::ArgRequiredElseHelp => self.0.remove(A_REQUIRED_ELSE_HELP),\\n- AppSettings::GlobalVersion => self.0.remove(GLOBAL_VERSION),\\n- AppSettings::UnifiedHelpMessage => self.0.remove(UNIFIED_HELP),\\n- AppSettings::WaitOnError => self.0.remove(WAIT_ON_ERROR),\\n- AppSettings::SubcommandRequiredElseHelp => self.0.remove(SC_REQUIRED_ELSE_HELP),\\n- AppSettings::NeedsLongHelp => self.0.remove(NEEDS_LONG_HELP),\\n- AppSettings::NeedsLongVersion => self.0.remove(NEEDS_LONG_VERSION),\\n- AppSettings::NeedsSubcommandHelp => self.0.remove(NEEDS_SC_HELP),\\n- AppSettings::DisableVersion => self.0.remove(DISABLE_VERSION),\\n- AppSettings::Hidden => self.0.remove(HIDDEN),\\n- AppSettings::TrailingVarArg => self.0.remove(TRAILING_VARARG),\\n- AppSettings::NoBinaryName => self.0.remove(NO_BIN_NAME),\\n- AppSettings::AllowExternalSubcommands => self.0.remove(ALLOW_UNK_SC),\\n- AppSettings::StrictUtf8 => self.0.remove(UTF8_STRICT),\\n- AppSettings::AllowInvalidUtf8 => self.0.remove(UTF8_NONE),\\n- AppSettings::AllowLeadingHyphen => self.0.remove(LEADING_HYPHEN),\\n- }\\n- }\\n-\\n- pub fn is_set(&self, s: AppSettings) -> bool {\\n- match s {\\n- AppSettings::SubcommandsNegateReqs => self.0.contains(SC_NEGATE_REQS),\\n- AppSettings::VersionlessSubcommands => self.0.contains(VERSIONLESS_SC),\\n- AppSettings::SubcommandRequired => self.0.contains(SC_REQUIRED),\\n- AppSettings::ArgRequiredElseHelp => self.0.contains(A_REQUIRED_ELSE_HELP),\\n- AppSettings::GlobalVersion => self.0.contains(GLOBAL_VERSION),\\n- AppSettings::UnifiedHelpMessage => self.0.contains(UNIFIED_HELP),\\n- AppSettings::WaitOnError => self.0.contains(WAIT_ON_ERROR),\\n- AppSettings::SubcommandRequiredElseHelp => self.0.contains(SC_REQUIRED_ELSE_HELP),\\n- AppSettings::NeedsLongHelp => self.0.contains(NEEDS_LONG_HELP),\\n- AppSettings::NeedsLongVersion => self.0.contains(NEEDS_LONG_VERSION),\\n- AppSettings::NeedsSubcommandHelp => self.0.contains(NEEDS_SC_HELP),\\n- AppSettings::DisableVersion => self.0.contains(DISABLE_VERSION),\\n- AppSettings::Hidden => self.0.contains(HIDDEN),\\n- AppSettings::TrailingVarArg => self.0.contains(TRAILING_VARARG),\\n- AppSettings::NoBinaryName => self.0.contains(NO_BIN_NAME),\\n- AppSettings::AllowExternalSubcommands => self.0.contains(ALLOW_UNK_SC),\\n- AppSettings::StrictUtf8 => self.0.contains(UTF8_STRICT),\\n- AppSettings::AllowInvalidUtf8 => self.0.contains(UTF8_NONE),\\n- AppSettings::AllowLeadingHyphen => self.0.contains(LEADING_HYPHEN),\\n- }\\n+ impl_settings! { AppSettings,\\n+ SubcommandsNegateReqs => SC_NEGATE_REQS,\\n+ VersionlessSubcommands => VERSIONLESS_SC,\\n+ SubcommandRequired => SC_REQUIRED,\\n+ ArgRequiredElseHelp => A_REQUIRED_ELSE_HELP,\\n+ GlobalVersion => GLOBAL_VERSION,\\n+ UnifiedHelpMessage => UNIFIED_HELP,\\n+ WaitOnError => WAIT_ON_ERROR,\\n+ SubcommandRequiredElseHelp => SC_REQUIRED_ELSE_HELP,\\n+ NeedsLongHelp => NEEDS_LONG_HELP,\\n+ NeedsLongVersion => NEEDS_LONG_VERSION,\\n+ NeedsSubcommandHelp => NEEDS_SC_HELP,\\n+ DisableVersion => DISABLE_VERSION,\\n+ Hidden => HIDDEN,\\n+ TrailingVarArg => TRAILING_VARARG,\\n+ NoBinaryName => NO_BIN_NAME,\\n+ AllowExternalSubcommands => ALLOW_UNK_SC,\\n+ StrictUtf8 => UTF8_STRICT,\\n+ AllowInvalidUtf8 => UTF8_NONE,\\n+ AllowLeadingHyphen => LEADING_HYPHEN\\n }\\n }\\n \\ndiff --git a/src/args/settings.rs b/src/args/settings.rs\\nindex f2f1384..effc18c 100644\\n--- a/src/args/settings.rs\\n+++ b/src/args/settings.rs\\n@@ -21,40 +21,14 @@ impl ArgFlags {\\n ArgFlags(EMPTY_VALS | USE_DELIM)\\n }\\n \\n- pub fn set(&mut self, s: ArgSettings) {\\n- match s {\\n- ArgSettings::Required => self.0.insert(REQUIRED),\\n- ArgSettings::Multiple => self.0.insert(MULTIPLE),\\n- ArgSettings::EmptyValues => self.0.insert(EMPTY_VALS),\\n- ArgSettings::Global => self.0.insert(GLOBAL),\\n- ArgSettings::Hidden => self.0.insert(HIDDEN),\\n- ArgSettings::TakesValue => self.0.insert(TAKES_VAL),\\n- ArgSettings::UseValueDelimiter => self.0.insert(USE_DELIM),\\n- }\\n- }\\n-\\n- pub fn unset(&mut self, s: ArgSettings) {\\n- match s {\\n- ArgSettings::Required => self.0.remove(REQUIRED),\\n- ArgSettings::Multiple => self.0.remove(MULTIPLE),\\n- ArgSettings::EmptyValues => self.0.remove(EMPTY_VALS),\\n- ArgSettings::Global => self.0.remove(GLOBAL),\\n- ArgSettings::Hidden => self.0.remove(HIDDEN),\\n- ArgSettings::TakesValue => self.0.remove(TAKES_VAL),\\n- ArgSettings::UseValueDelimiter => self.0.remove(USE_DELIM),\\n- }\\n- }\\n-\\n- pub fn is_set(&self, s: ArgSettings) -> bool {\\n- match s {\\n- ArgSettings::Required => self.0.contains(REQUIRED),\\n- ArgSettings::Multiple => self.0.contains(MULTIPLE),\\n- ArgSettings::EmptyValues => self.0.contains(EMPTY_VALS),\\n- ArgSettings::Global => self.0.contains(GLOBAL),\\n- ArgSettings::Hidden => self.0.contains(HIDDEN),\\n- ArgSettings::TakesValue => self.0.contains(TAKES_VAL),\\n- ArgSettings::UseValueDelimiter => self.0.contains(USE_DELIM),\\n- }\\n+ impl_settings!{ArgSettings,\\n+ Required => REQUIRED,\\n+ Multiple => MULTIPLE,\\n+ EmptyValues => EMPTY_VALS,\\n+ Global => GLOBAL,\\n+ Hidden => HIDDEN,\\n+ TakesValue => TAKES_VAL,\\n+ UseValueDelimiter => USE_DELIM\\n }\\n }\\n \\ndiff --git a/src/macros.rs b/src/macros.rs\\nindex 47675ac..29d5382 100644\\n--- a/src/macros.rs\\n+++ b/src/macros.rs\\n@@ -1,3 +1,25 @@\\n+macro_rules! impl_settings {\\n+ ($n:ident, $($v:ident => $c:ident),+) => {\\n+ pub fn set(&mut self, s: $n) {\\n+ match s {\\n+ $($n::$v => self.0.insert($c)),+\\n+ }\\n+ }\\n+\\n+ pub fn unset(&mut self, s: $n) {\\n+ match s {\\n+ $($n::$v => self.0.remove($c)),+\\n+ }\\n+ }\\n+\\n+ pub fn is_set(&self, s: $n) -> bool {\\n+ match s {\\n+ $($n::$v => self.0.contains($c)),+\\n+ }\\n+ }\\n+ };\\n+}\\n+\\n // Convenience for writing to stderr thanks to https://github.com/BurntSushi\\n macro_rules! wlnerr(\\n ($($arg:tt)*) => ({\\n\", \"diff --git a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\\nindex 4d771d6..9f9683c 100644\\n--- a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\\n+++ b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\\n@@ -9,9 +9,6 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\\n Object {\\n \\\"flex\\\": 1,\\n },\\n- Object {\\n- \\\"backgroundColor\\\": \\\"#000000\\\",\\n- },\\n undefined,\\n ]\\n }\\n@@ -132,6 +129,33 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\\n ]\\n }\\n >\\n+ \\n \\n+ \\n MinerConfig {\\n MinerConfig {\\n min_tx_fee: 1,\\n- first_attempt_time_ms: 1_000,\\n- subsequent_attempt_time_ms: 30_000,\\n+ first_attempt_time_ms: 5_000,\\n+ subsequent_attempt_time_ms: 180_000,\\n microblock_attempt_time_ms: 30_000,\\n probability_pick_no_estimate_tx: 5,\\n }\\n\", \"diff --git a/Cargo.lock b/Cargo.lock\\nindex f949506..6a10219 100644\\n--- a/Cargo.lock\\n+++ b/Cargo.lock\\n@@ -94,7 +94,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"els\\\"\\n-version = \\\"0.1.22\\\"\\n+version = \\\"0.1.23-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_compiler\\\",\\n@@ -105,7 +105,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"els\\\",\\n \\\"erg_common\\\",\\n@@ -115,7 +115,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_common\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"backtrace-on-stack-overflow\\\",\\n \\\"crossterm\\\",\\n@@ -126,7 +126,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_compiler\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_parser\\\",\\n@@ -134,7 +134,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_parser\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"unicode-xid\\\",\\ndiff --git a/Cargo.toml b/Cargo.toml\\nindex 04fdad7..ecc45e5 100644\\n--- a/Cargo.toml\\n+++ b/Cargo.toml\\n@@ -20,7 +20,7 @@ members = [\\n ]\\n \\n [workspace.package]\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n authors = [\\\"erg-lang team \\\"]\\n license = \\\"MIT OR Apache-2.0\\\"\\n edition = \\\"2021\\\"\\n@@ -64,10 +64,10 @@ full-repl = [\\\"erg_common/full-repl\\\"]\\n full = [\\\"els\\\", \\\"full-repl\\\", \\\"unicode\\\", \\\"pretty\\\"]\\n \\n [workspace.dependencies]\\n-erg_common = { version = \\\"0.6.10\\\", path = \\\"./crates/erg_common\\\" }\\n-erg_parser = { version = \\\"0.6.10\\\", path = \\\"./crates/erg_parser\\\" }\\n-erg_compiler = { version = \\\"0.6.10\\\", path = \\\"./crates/erg_compiler\\\" }\\n-els = { version = \\\"0.1.22\\\", path = \\\"./crates/els\\\" }\\n+erg_common = { version = \\\"0.6.11-nightly.0\\\", path = \\\"./crates/erg_common\\\" }\\n+erg_parser = { version = \\\"0.6.11-nightly.0\\\", path = \\\"./crates/erg_parser\\\" }\\n+erg_compiler = { version = \\\"0.6.11-nightly.0\\\", path = \\\"./crates/erg_compiler\\\" }\\n+els = { version = \\\"0.1.23-nightly.0\\\", path = \\\"./crates/els\\\" }\\n \\n [dependencies]\\n erg_common = { workspace = true }\\ndiff --git a/crates/els/Cargo.toml b/crates/els/Cargo.toml\\nindex bc031e6..7c9455f 100644\\n--- a/crates/els/Cargo.toml\\n+++ b/crates/els/Cargo.toml\\n@@ -2,7 +2,7 @@\\n name = \\\"els\\\"\\n description = \\\"An Erg compiler frontend for IDEs, implements LSP.\\\"\\n documentation = \\\"http://docs.rs/els\\\"\\n-version = \\\"0.1.22\\\"\\n+version = \\\"0.1.23-nightly.0\\\"\\n authors.workspace = true\\n license.workspace = true\\n edition.workspace = true\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"86f3e3397594f8312226c5a193608a054087805c\", \"87427fe39d165bee2acedde8dbaa237cca3fb61e\", \"d35d302cadf355a169dca6636597183de6bbee23\", \"607ecc92b5f8c084304e406eec725b7dcfa0a562\"]"},"types":{"kind":"string","value":"[\"refactor\", \"test\", \"fix\", \"build\"]"}}},{"rowIdx":919,"cells":{"commit_message":{"kind":"string","value":"run nix macos jobs on macos-13 to try and avoid SIP,change tests to depend on BrokerContext,fix \"types\" field in dist,update the formatting for python integration example"},"diff":{"kind":"string","value":"[\"diff --git a/.github/actionlint.yaml b/.github/actionlint.yaml\\nnew file mode 100644\\nindex 0000000..5be7d17\\n--- /dev/null\\n+++ b/.github/actionlint.yaml\\n@@ -0,0 +1,7 @@\\n+self-hosted-runner:\\n+ # Labels of self-hosted runner in array of strings.\\n+ labels: [macos-13]\\n+# Configuration variables in array of strings defined in your repository or\\n+# organization. `null` means disabling configuration variables check.\\n+# Empty array means no configuration variable is allowed.\\n+config-variables: null\\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\\nindex e37346c..dce77e1 100644\\n--- a/.github/workflows/nix.yml\\n+++ b/.github/workflows/nix.yml\\n@@ -37,7 +37,7 @@ jobs:\\n - \\\"3.10\\\"\\n - \\\"3.11\\\"\\n include:\\n- - os: macos-latest\\n+ - os: macos-13\\n python-version: \\\"3.10\\\"\\n steps:\\n - name: checkout\\ndiff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml\\nindex 005a850..8db22e2 100644\\n--- a/.pre-commit-config.yaml\\n+++ b/.pre-commit-config.yaml\\n@@ -3,7 +3,7 @@ ci:\\n autofix_prs: false\\n autoupdate_commit_msg: \\\"chore(deps): pre-commit.ci autoupdate\\\"\\n skip:\\n- - actionlint\\n+ - actionlint-system\\n - deadnix\\n - just\\n - nixpkgs-fmt\\n@@ -17,9 +17,9 @@ default_stages:\\n - commit\\n repos:\\n - repo: https://github.com/rhysd/actionlint\\n- rev: v1.6.24\\n+ rev: v1.6.25\\n hooks:\\n- - id: actionlint\\n+ - id: actionlint-system\\n - repo: https://github.com/psf/black\\n rev: 23.3.0\\n hooks:\\n@@ -30,7 +30,7 @@ repos:\\n - id: nbstripout\\n exclude: .+/rendered/.+\\n - repo: https://github.com/codespell-project/codespell\\n- rev: v2.2.4\\n+ rev: v2.2.5\\n hooks:\\n - id: codespell\\n additional_dependencies:\\n\", \"diff --git a/broker/src/main/java/io/camunda/zeebe/broker/Broker.java b/broker/src/main/java/io/camunda/zeebe/broker/Broker.java\\nindex fe4e42d..37c7066 100644\\n--- a/broker/src/main/java/io/camunda/zeebe/broker/Broker.java\\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/Broker.java\\n@@ -7,20 +7,14 @@\\n */\\n package io.camunda.zeebe.broker;\\n \\n-import io.atomix.cluster.AtomixCluster;\\n import io.camunda.zeebe.broker.bootstrap.BrokerContext;\\n import io.camunda.zeebe.broker.bootstrap.BrokerStartupContextImpl;\\n import io.camunda.zeebe.broker.bootstrap.BrokerStartupProcess;\\n-import io.camunda.zeebe.broker.clustering.ClusterServices;\\n import io.camunda.zeebe.broker.exporter.repo.ExporterLoadException;\\n import io.camunda.zeebe.broker.exporter.repo.ExporterRepository;\\n-import io.camunda.zeebe.broker.partitioning.PartitionManager;\\n-import io.camunda.zeebe.broker.system.EmbeddedGatewayService;\\n import io.camunda.zeebe.broker.system.SystemContext;\\n import io.camunda.zeebe.broker.system.configuration.BrokerCfg;\\n-import io.camunda.zeebe.broker.system.management.BrokerAdminService;\\n import io.camunda.zeebe.broker.system.monitoring.BrokerHealthCheckService;\\n-import io.camunda.zeebe.broker.system.monitoring.DiskSpaceUsageMonitor;\\n import io.camunda.zeebe.protocol.impl.encoding.BrokerInfo;\\n import io.camunda.zeebe.util.LogUtil;\\n import io.camunda.zeebe.util.VersionUtil;\\n@@ -184,35 +178,15 @@ public final class Broker implements AutoCloseable {\\n }\\n \\n // only used for tests\\n- public EmbeddedGatewayService getEmbeddedGatewayService() {\\n- return brokerContext.getEmbeddedGatewayService();\\n- }\\n-\\n- public AtomixCluster getAtomixCluster() {\\n- return brokerContext.getAtomixCluster();\\n- }\\n-\\n- public ClusterServices getClusterServices() {\\n- return brokerContext.getClusterServices();\\n- }\\n-\\n- public DiskSpaceUsageMonitor getDiskSpaceUsageMonitor() {\\n- return brokerContext.getDiskSpaceUsageMonitor();\\n- }\\n-\\n- public BrokerAdminService getBrokerAdminService() {\\n- return brokerContext.getBrokerAdminService();\\n+ public BrokerContext getBrokerContext() {\\n+ return brokerContext;\\n }\\n \\n+ // only used for tests\\n public SystemContext getSystemContext() {\\n return systemContext;\\n }\\n \\n- public PartitionManager getPartitionManager() {\\n- return brokerContext.getPartitionManager();\\n- }\\n- // only used for tests\\n-\\n /**\\n * Temporary helper object. This object is needed during the transition of broker startup/shutdown\\n * steps to the new concept. Afterwards, the expectation is that this object will merge with the\\ndiff --git a/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java b/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java\\nindex bda5170..1accbc1 100644\\n--- a/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java\\n+++ b/broker/src/test/java/io/camunda/zeebe/broker/system/partitions/BrokerSnapshotTest.java\\n@@ -45,11 +45,12 @@ public class BrokerSnapshotTest {\\n (RaftPartition)\\n brokerRule\\n .getBroker()\\n+ .getBrokerContext()\\n .getPartitionManager()\\n .getPartitionGroup()\\n .getPartition(PartitionId.from(PartitionManagerImpl.GROUP_NAME, PARTITION_ID));\\n journalReader = raftPartition.getServer().openReader();\\n- brokerAdminService = brokerRule.getBroker().getBrokerAdminService();\\n+ brokerAdminService = brokerRule.getBroker().getBrokerContext().getBrokerAdminService();\\n \\n final String contactPoint = NetUtil.toSocketAddressString(brokerRule.getGatewayAddress());\\n final ZeebeClientBuilder zeebeClientBuilder =\\ndiff --git a/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java b/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java\\nindex e98e7d2..a831bfe 100644\\n--- a/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java\\n+++ b/broker/src/test/java/io/camunda/zeebe/broker/test/EmbeddedBrokerRule.java\\n@@ -173,11 +173,11 @@ public final class EmbeddedBrokerRule extends ExternalResource {\\n }\\n \\n public ClusterServices getClusterServices() {\\n- return broker.getClusterServices();\\n+ return broker.getBrokerContext().getClusterServices();\\n }\\n \\n public AtomixCluster getAtomixCluster() {\\n- return broker.getAtomixCluster();\\n+ return broker.getBrokerContext().getAtomixCluster();\\n }\\n \\n public InetSocketAddress getGatewayAddress() {\\n@@ -245,7 +245,8 @@ public final class EmbeddedBrokerRule extends ExternalResource {\\n Thread.currentThread().interrupt();\\n }\\n \\n- final EmbeddedGatewayService embeddedGatewayService = broker.getEmbeddedGatewayService();\\n+ final EmbeddedGatewayService embeddedGatewayService =\\n+ broker.getBrokerContext().getEmbeddedGatewayService();\\n if (embeddedGatewayService != null) {\\n final BrokerClient brokerClient = embeddedGatewayService.get().getBrokerClient();\\n \\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java\\nindex 890b596..8561cf1 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/clustering/ClusteringRule.java\\n@@ -29,6 +29,7 @@ import io.atomix.utils.net.Address;\\n import io.camunda.zeebe.broker.Broker;\\n import io.camunda.zeebe.broker.PartitionListener;\\n import io.camunda.zeebe.broker.SpringBrokerBridge;\\n+import io.camunda.zeebe.broker.bootstrap.BrokerContext;\\n import io.camunda.zeebe.broker.exporter.stream.ExporterDirectorContext;\\n import io.camunda.zeebe.broker.partitioning.PartitionManagerImpl;\\n import io.camunda.zeebe.broker.system.SystemContext;\\n@@ -602,11 +603,11 @@ public final class ClusteringRule extends ExternalResource {\\n }\\n \\n public void stepDown(final Broker broker, final int partitionId) {\\n- final var atomix = broker.getClusterServices();\\n+ final var atomix = broker.getBrokerContext().getClusterServices();\\n final MemberId nodeId = atomix.getMembershipService().getLocalMember().id();\\n \\n final var raftPartition =\\n- broker.getPartitionManager().getPartitionGroup().getPartitions().stream()\\n+ broker.getBrokerContext().getPartitionManager().getPartitionGroup().getPartitions().stream()\\n .filter(partition -> partition.members().contains(nodeId))\\n .filter(partition -> partition.id().id() == partitionId)\\n .map(RaftPartition.class::cast)\\n@@ -617,14 +618,14 @@ public final class ClusteringRule extends ExternalResource {\\n }\\n \\n public void disconnect(final Broker broker) {\\n- final var atomix = broker.getAtomixCluster();\\n+ final var atomix = broker.getBrokerContext().getAtomixCluster();\\n \\n ((NettyUnicastService) atomix.getUnicastService()).stop().join();\\n ((NettyMessagingService) atomix.getMessagingService()).stop().join();\\n }\\n \\n public void connect(final Broker broker) {\\n- final var atomix = broker.getAtomixCluster();\\n+ final var atomix = broker.getBrokerContext().getAtomixCluster();\\n \\n ((NettyUnicastService) atomix.getUnicastService()).start().join();\\n ((NettyMessagingService) atomix.getMessagingService()).start().join();\\n@@ -666,11 +667,11 @@ public final class ClusteringRule extends ExternalResource {\\n }\\n \\n final var broker = brokers.get(expectedLeader);\\n- final var atomix = broker.getClusterServices();\\n+ final var atomix = broker.getBrokerContext().getClusterServices();\\n final MemberId nodeId = atomix.getMembershipService().getLocalMember().id();\\n \\n final var raftPartition =\\n- broker.getPartitionManager().getPartitionGroup().getPartitions().stream()\\n+ broker.getBrokerContext().getPartitionManager().getPartitionGroup().getPartitions().stream()\\n .filter(partition -> partition.members().contains(nodeId))\\n .filter(partition -> partition.id().id() == START_PARTITION_ID)\\n .map(RaftPartition.class::cast)\\n@@ -775,14 +776,15 @@ public final class ClusteringRule extends ExternalResource {\\n }\\n \\n public void takeSnapshot(final Broker broker) {\\n- broker.getBrokerAdminService().takeSnapshot();\\n+ broker.getBrokerContext().getBrokerAdminService().takeSnapshot();\\n }\\n \\n public void triggerAndWaitForSnapshots() {\\n // Ensure that the exporter positions are distributed to the followers\\n getClock().addTime(ExporterDirectorContext.DEFAULT_DISTRIBUTION_INTERVAL);\\n getBrokers().stream()\\n- .map(Broker::getBrokerAdminService)\\n+ .map(Broker::getBrokerContext)\\n+ .map(BrokerContext::getBrokerAdminService)\\n .forEach(BrokerAdminService::takeSnapshot);\\n \\n getBrokers()\\n@@ -794,7 +796,7 @@ public final class ClusteringRule extends ExternalResource {\\n .until(\\n () -> {\\n // Trigger snapshot again in case snapshot is not already taken\\n- broker.getBrokerAdminService().takeSnapshot();\\n+ broker.getBrokerContext().getBrokerAdminService().takeSnapshot();\\n return getSnapshot(broker);\\n },\\n Optional::isPresent));\\n@@ -831,7 +833,7 @@ public final class ClusteringRule extends ExternalResource {\\n \\n private Optional getSnapshot(final Broker broker, final int partitionId) {\\n \\n- final var partitions = broker.getBrokerAdminService().getPartitionStatus();\\n+ final var partitions = broker.getBrokerContext().getBrokerAdminService().getPartitionStatus();\\n final var partitionStatus = partitions.get(partitionId);\\n \\n return Optional.ofNullable(partitionStatus)\\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java\\nindex f07961c..d46636b 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceMonitoringFailOverTest.java\\n@@ -88,6 +88,7 @@ public class DiskSpaceMonitoringFailOverTest {\\n () ->\\n clusteringRule\\n .getBroker(newLeaderId)\\n+ .getBrokerContext()\\n .getBrokerAdminService()\\n .getPartitionStatus()\\n .get(1)\\n@@ -96,7 +97,7 @@ public class DiskSpaceMonitoringFailOverTest {\\n }\\n \\n private void waitUntilDiskSpaceNotAvailable(final Broker broker) throws InterruptedException {\\n- final var diskSpaceMonitor = broker.getDiskSpaceUsageMonitor();\\n+ final var diskSpaceMonitor = broker.getBrokerContext().getDiskSpaceUsageMonitor();\\n \\n final CountDownLatch diskSpaceNotAvailable = new CountDownLatch(1);\\n diskSpaceMonitor.addDiskUsageListener(\\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java\\nindex 0a02a27..6e93cf9 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryClusteredTest.java\\n@@ -165,7 +165,7 @@ public class DiskSpaceRecoveryClusteredTest {\\n }\\n \\n private void waitUntilDiskSpaceNotAvailable(final Broker broker) throws InterruptedException {\\n- final var diskSpaceMonitor = broker.getDiskSpaceUsageMonitor();\\n+ final var diskSpaceMonitor = broker.getBrokerContext().getDiskSpaceUsageMonitor();\\n \\n final CountDownLatch diskSpaceNotAvailable = new CountDownLatch(1);\\n diskSpaceMonitor.addDiskUsageListener(\\n@@ -188,7 +188,7 @@ public class DiskSpaceRecoveryClusteredTest {\\n }\\n \\n private void waitUntilDiskSpaceAvailable(final Broker broker) throws InterruptedException {\\n- final var diskSpaceMonitor = broker.getDiskSpaceUsageMonitor();\\n+ final var diskSpaceMonitor = broker.getBrokerContext().getDiskSpaceUsageMonitor();\\n final CountDownLatch diskSpaceAvailableAgain = new CountDownLatch(1);\\n diskSpaceMonitor.addDiskUsageListener(\\n new DiskSpaceUsageListener() {\\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java\\nindex 9cef5a0..a487729 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryTest.java\\n@@ -192,7 +192,8 @@ public class DiskSpaceRecoveryTest {\\n }\\n \\n private void waitUntilDiskSpaceNotAvailable() throws InterruptedException {\\n- final var diskSpaceMonitor = embeddedBrokerRule.getBroker().getDiskSpaceUsageMonitor();\\n+ final var diskSpaceMonitor =\\n+ embeddedBrokerRule.getBroker().getBrokerContext().getDiskSpaceUsageMonitor();\\n \\n final CountDownLatch diskSpaceNotAvailable = new CountDownLatch(1);\\n diskSpaceMonitor.addDiskUsageListener(\\n@@ -215,7 +216,8 @@ public class DiskSpaceRecoveryTest {\\n }\\n \\n private void waitUntilDiskSpaceAvailable() throws InterruptedException {\\n- final var diskSpaceMonitor = embeddedBrokerRule.getBroker().getDiskSpaceUsageMonitor();\\n+ final var diskSpaceMonitor =\\n+ embeddedBrokerRule.getBroker().getBrokerContext().getDiskSpaceUsageMonitor();\\n final CountDownLatch diskSpaceAvailableAgain = new CountDownLatch(1);\\n diskSpaceMonitor.addDiskUsageListener(\\n new DiskSpaceUsageListener() {\\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java\\nindex 2d1e4f0..58f6f16 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/HealthMonitoringTest.java\\n@@ -48,6 +48,7 @@ public class HealthMonitoringTest {\\n final var raftPartition =\\n (RaftPartition)\\n leader\\n+ .getBrokerContext()\\n .getPartitionManager()\\n .getPartitionGroup()\\n .getPartition(\\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java\\nindex 468f83c..7ff03be 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceClusterTest.java\\n@@ -11,6 +11,7 @@ import static org.assertj.core.api.Assertions.assertThat;\\n \\n import io.atomix.raft.RaftServer.Role;\\n import io.camunda.zeebe.broker.Broker;\\n+import io.camunda.zeebe.broker.bootstrap.BrokerContext;\\n import io.camunda.zeebe.broker.system.management.BrokerAdminService;\\n import io.camunda.zeebe.engine.processing.streamprocessor.StreamProcessor.Phase;\\n import io.camunda.zeebe.it.clustering.ClusteringRule;\\n@@ -48,7 +49,7 @@ public class BrokerAdminServiceClusterTest {\\n @Before\\n public void before() {\\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\\n- leaderAdminService = leader.getBrokerAdminService();\\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\\n }\\n \\n @Test\\n@@ -61,7 +62,8 @@ public class BrokerAdminServiceClusterTest {\\n // when\\n final var followerStatus =\\n followers.stream()\\n- .map(Broker::getBrokerAdminService)\\n+ .map(Broker::getBrokerContext)\\n+ .map(BrokerContext::getBrokerAdminService)\\n .map(BrokerAdminService::getPartitionStatus)\\n .map(status -> status.get(1));\\n \\n@@ -94,7 +96,8 @@ public class BrokerAdminServiceClusterTest {\\n \\n // then\\n clusteringRule.getBrokers().stream()\\n- .map(Broker::getBrokerAdminService)\\n+ .map(Broker::getBrokerContext)\\n+ .map(BrokerContext::getBrokerAdminService)\\n .forEach(this::assertThatStatusContainsProcessedPositionInSnapshot);\\n }\\n \\n@@ -102,7 +105,8 @@ public class BrokerAdminServiceClusterTest {\\n public void shouldPauseAfterLeaderChange() {\\n // given\\n clusteringRule.getBrokers().stream()\\n- .map(Broker::getBrokerAdminService)\\n+ .map(Broker::getBrokerContext)\\n+ .map(BrokerContext::getBrokerAdminService)\\n .forEach(BrokerAdminService::pauseStreamProcessing);\\n \\n // when\\n@@ -113,6 +117,7 @@ public class BrokerAdminServiceClusterTest {\\n final var newLeaderAdminService =\\n clusteringRule\\n .getBroker(clusteringRule.getLeaderForPartition(1).getNodeId())\\n+ .getBrokerContext()\\n .getBrokerAdminService();\\n assertStreamProcessorPhase(newLeaderAdminService, Phase.PAUSED);\\n }\\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java\\nindex 5160b50..2185329 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceTest.java\\n@@ -41,7 +41,7 @@ public class BrokerAdminServiceTest {\\n @Before\\n public void before() {\\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\\n- leaderAdminService = leader.getBrokerAdminService();\\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\\n }\\n \\n @Test\\n@@ -144,7 +144,7 @@ public class BrokerAdminServiceTest {\\n \\n // then\\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\\n- leaderAdminService = leader.getBrokerAdminService();\\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\\n assertStreamProcessorPhase(leaderAdminService, Phase.PAUSED);\\n }\\n \\n@@ -161,7 +161,7 @@ public class BrokerAdminServiceTest {\\n \\n // then\\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\\n- leaderAdminService = leader.getBrokerAdminService();\\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\\n assertStreamProcessorPhase(leaderAdminService, Phase.PROCESSING);\\n }\\n \\n@@ -176,7 +176,7 @@ public class BrokerAdminServiceTest {\\n \\n // then\\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\\n- leaderAdminService = leader.getBrokerAdminService();\\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\\n assertExporterPhase(leaderAdminService, ExporterPhase.PAUSED);\\n }\\n \\n@@ -193,7 +193,7 @@ public class BrokerAdminServiceTest {\\n \\n // then\\n leader = clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\\n- leaderAdminService = leader.getBrokerAdminService();\\n+ leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\\n assertExporterPhase(leaderAdminService, ExporterPhase.EXPORTING);\\n }\\n \\ndiff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java\\nindex d6c8ab3..4582ad2 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/system/BrokerAdminServiceWithOutExporterTest.java\\n@@ -30,7 +30,7 @@ public class BrokerAdminServiceWithOutExporterTest {\\n // given\\n final var leader =\\n clusteringRule.getBroker(clusteringRule.getLeaderForPartition(1).getNodeId());\\n- final var leaderAdminService = leader.getBrokerAdminService();\\n+ final var leaderAdminService = leader.getBrokerContext().getBrokerAdminService();\\n // when there are no exporters configured\\n // then\\n final var partitionStatus = leaderAdminService.getPartitionStatus().get(1);\\ndiff --git a/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java b/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java\\nindex 36bc0bf..d332201 100644\\n--- a/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java\\n+++ b/test/src/main/java/io/camunda/zeebe/test/EmbeddedBrokerRule.java\\n@@ -240,7 +240,8 @@ public class EmbeddedBrokerRule extends ExternalResource {\\n Thread.currentThread().interrupt();\\n }\\n \\n- final EmbeddedGatewayService embeddedGatewayService = broker.getEmbeddedGatewayService();\\n+ final EmbeddedGatewayService embeddedGatewayService =\\n+ broker.getBrokerContext().getEmbeddedGatewayService();\\n if (embeddedGatewayService != null) {\\n final BrokerClient brokerClient = embeddedGatewayService.get().getBrokerClient();\\n \\n\", \"diff --git a/scripts/prepare.js b/scripts/prepare.js\\nindex 4bab09b..55f459b 100644\\n--- a/scripts/prepare.js\\n+++ b/scripts/prepare.js\\n@@ -96,7 +96,6 @@ async function prepare() {\\n delete json.private\\n delete json.scripts\\n delete json.devDependencies\\n- delete json.types\\n \\n // Add \\\"postinstall\\\" script for donations.\\n if (/(native|core)$/.test(name))\\n@@ -128,6 +127,7 @@ async function prepare() {\\n else {\\n json.main = json.main.replace(/^dist\\\\//, '')\\n if (json.main.endsWith('.cjs.js')) {\\n+ json.types = json.main.replace('.cjs.js', '.d.ts')\\n json.module = json.main.replace('.cjs', '')\\n }\\n }\\n\", \"diff --git a/website/docs/integration/python.md b/website/docs/integration/python.md\\nindex 064cae3..b6b720d 100644\\n--- a/website/docs/integration/python.md\\n+++ b/website/docs/integration/python.md\\n@@ -13,6 +13,7 @@ header = \\\"All notable changes to this project will be documented in this file.\\\"\\n body = \\\"...\\\"\\n footer = \\\"\\\"\\n # see [changelog] section for more keys\\n+\\n [tool.git-cliff.git]\\n conventional_commits = true\\n commit_parsers = []\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"54cb6d4643b4a072ff997592a7fa14a69a6c068d\", \"e52a6201093f273add4903dd5f4e55a63539386d\", \"f14ef3809f456aadd73523e47cb16c5d15e9a9df\", \"3ee672483790ec71c700907a6e93af4698492026\"]"},"types":{"kind":"string","value":"[\"ci\", \"refactor\", \"build\", \"docs\"]"}}},{"rowIdx":920,"cells":{"commit_message":{"kind":"string","value":"add link to roadmap,Publish crates,Fix Cube Store build on Windows,apply permissions to profile request"},"diff":{"kind":"string","value":"[\"diff --git a/packages/plugin-core/README.md b/packages/plugin-core/README.md\\nindex 3c25c9b..c7506d4 100644\\n--- a/packages/plugin-core/README.md\\n+++ b/packages/plugin-core/README.md\\n@@ -187,6 +187,10 @@ When the workspace opens, it will show dialogue to install the recommended exten\\n \\n See [[FAQ]] to answers for common questions.\\n \\n+# Roadmap\\n+\\n+Check out our [public roadmap](https://github.com/orgs/dendronhq/projects/1) to see the features we're working on and to vote for what you want to see next. \\n+\\n \\n # Contributing\\n \\n\", \"diff --git a/CHANGELOG.md b/CHANGELOG.md\\nindex 7b98b44..f17ad6f 100644\\n--- a/CHANGELOG.md\\n+++ b/CHANGELOG.md\\n@@ -7,6 +7,9 @@\\n \\n - **(css/parser)** Fix parsing of at rules (#3328) ([506a310](https://github.com/swc-project/swc/commit/506a31078aaebf50129658f096bbd5929995205f))\\n \\n+\\n+- **(es/compat)** Fix regression of `destructuring` (#3326) ([6d1ad36](https://github.com/swc-project/swc/commit/6d1ad368aca53ee64a63ae565cd015909f2f4458))\\n+\\n ### Performance\\n \\n \\ndiff --git a/Cargo.lock b/Cargo.lock\\nindex 3c6598b..4baa252 100644\\n--- a/Cargo.lock\\n+++ b/Cargo.lock\\n@@ -2652,7 +2652,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"swc\\\"\\n-version = \\\"0.116.15\\\"\\n+version = \\\"0.116.16\\\"\\n dependencies = [\\n \\\"ahash\\\",\\n \\\"anyhow\\\",\\n@@ -3097,7 +3097,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"swc_ecma_transforms\\\"\\n-version = \\\"0.113.3\\\"\\n+version = \\\"0.113.4\\\"\\n dependencies = [\\n \\\"pretty_assertions 0.7.2\\\",\\n \\\"sourcemap\\\",\\n@@ -3157,7 +3157,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"swc_ecma_transforms_compat\\\"\\n-version = \\\"0.68.2\\\"\\n+version = \\\"0.68.3\\\"\\n dependencies = [\\n \\\"ahash\\\",\\n \\\"arrayvec 0.7.2\\\",\\n@@ -3366,7 +3366,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"swc_ecmascript\\\"\\n-version = \\\"0.110.14\\\"\\n+version = \\\"0.110.15\\\"\\n dependencies = [\\n \\\"swc_ecma_ast\\\",\\n \\\"swc_ecma_codegen\\\",\\ndiff --git a/crates/swc/Cargo.toml b/crates/swc/Cargo.toml\\nindex 756cfc8..2f02d22 100644\\n--- a/crates/swc/Cargo.toml\\n+++ b/crates/swc/Cargo.toml\\n@@ -9,7 +9,7 @@ include = [\\\"Cargo.toml\\\", \\\"src/**/*.rs\\\"]\\n license = \\\"Apache-2.0\\\"\\n name = \\\"swc\\\"\\n repository = \\\"https://github.com/swc-project/swc.git\\\"\\n-version = \\\"0.116.15\\\"\\n+version = \\\"0.116.16\\\"\\n \\n [lib]\\n name = \\\"swc\\\"\\n@@ -55,7 +55,7 @@ swc_ecma_loader = {version = \\\"0.27.0\\\", path = \\\"../swc_ecma_loader\\\", features = [\\n swc_ecma_minifier = {version = \\\"0.70.9\\\", path = \\\"../swc_ecma_minifier\\\"}\\n swc_ecma_parser = {version = \\\"0.87.0\\\", path = \\\"../swc_ecma_parser\\\"}\\n swc_ecma_preset_env = {version = \\\"0.86.1\\\", path = \\\"../swc_ecma_preset_env\\\"}\\n-swc_ecma_transforms = {version = \\\"0.113.3\\\", path = \\\"../swc_ecma_transforms\\\", features = [\\n+swc_ecma_transforms = {version = \\\"0.113.4\\\", path = \\\"../swc_ecma_transforms\\\", features = [\\n \\\"compat\\\",\\n \\\"module\\\",\\n \\\"optimization\\\",\\n@@ -64,11 +64,11 @@ swc_ecma_transforms = {version = \\\"0.113.3\\\", path = \\\"../swc_ecma_transforms\\\", fea\\n \\\"typescript\\\",\\n ]}\\n swc_ecma_transforms_base = {version = \\\"0.57.1\\\", path = \\\"../swc_ecma_transforms_base\\\"}\\n-swc_ecma_transforms_compat = {version = \\\"0.68.2\\\", path = \\\"../swc_ecma_transforms_compat\\\"}\\n+swc_ecma_transforms_compat = {version = \\\"0.68.3\\\", path = \\\"../swc_ecma_transforms_compat\\\"}\\n swc_ecma_transforms_optimization = {version = \\\"0.83.0\\\", path = \\\"../swc_ecma_transforms_optimization\\\"}\\n swc_ecma_utils = {version = \\\"0.64.0\\\", path = \\\"../swc_ecma_utils\\\"}\\n swc_ecma_visit = {version = \\\"0.51.1\\\", path = \\\"../swc_ecma_visit\\\"}\\n-swc_ecmascript = {version = \\\"0.110.14\\\", path = \\\"../swc_ecmascript\\\"}\\n+swc_ecmascript = {version = \\\"0.110.15\\\", path = \\\"../swc_ecmascript\\\"}\\n swc_node_comments = {version = \\\"0.4.0\\\", path = \\\"../swc_node_comments\\\"}\\n swc_plugin_runner = {version = \\\"0.30.0\\\", path = \\\"../swc_plugin_runner\\\", optional = true}\\n swc_visit = {version = \\\"0.3.0\\\", path = \\\"../swc_visit\\\"}\\ndiff --git a/crates/swc_ecma_transforms/Cargo.toml b/crates/swc_ecma_transforms/Cargo.toml\\nindex 1604f4e..a0aafae 100644\\n--- a/crates/swc_ecma_transforms/Cargo.toml\\n+++ b/crates/swc_ecma_transforms/Cargo.toml\\n@@ -6,7 +6,7 @@ edition = \\\"2021\\\"\\n license = \\\"Apache-2.0\\\"\\n name = \\\"swc_ecma_transforms\\\"\\n repository = \\\"https://github.com/swc-project/swc.git\\\"\\n-version = \\\"0.113.3\\\"\\n+version = \\\"0.113.4\\\"\\n \\n [package.metadata.docs.rs]\\n all-features = true\\n@@ -28,7 +28,7 @@ swc_common = {version = \\\"0.17.0\\\", path = \\\"../swc_common\\\"}\\n swc_ecma_ast = {version = \\\"0.65.0\\\", path = \\\"../swc_ecma_ast\\\"}\\n swc_ecma_parser = {version = \\\"0.87.0\\\", path = \\\"../swc_ecma_parser\\\"}\\n swc_ecma_transforms_base = {version = \\\"0.57.1\\\", path = \\\"../swc_ecma_transforms_base\\\"}\\n-swc_ecma_transforms_compat = {version = \\\"0.68.2\\\", path = \\\"../swc_ecma_transforms_compat\\\", optional = true}\\n+swc_ecma_transforms_compat = {version = \\\"0.68.3\\\", path = \\\"../swc_ecma_transforms_compat\\\", optional = true}\\n swc_ecma_transforms_module = {version = \\\"0.74.0\\\", path = \\\"../swc_ecma_transforms_module\\\", optional = true}\\n swc_ecma_transforms_optimization = {version = \\\"0.83.0\\\", path = \\\"../swc_ecma_transforms_optimization\\\", optional = true}\\n swc_ecma_transforms_proposal = {version = \\\"0.74.0\\\", path = \\\"../swc_ecma_transforms_proposal\\\", optional = true}\\ndiff --git a/crates/swc_ecma_transforms_compat/Cargo.toml b/crates/swc_ecma_transforms_compat/Cargo.toml\\nindex 0ea6609..58374e3 100644\\n--- a/crates/swc_ecma_transforms_compat/Cargo.toml\\n+++ b/crates/swc_ecma_transforms_compat/Cargo.toml\\n@@ -6,7 +6,7 @@ edition = \\\"2021\\\"\\n license = \\\"Apache-2.0\\\"\\n name = \\\"swc_ecma_transforms_compat\\\"\\n repository = \\\"https://github.com/swc-project/swc.git\\\"\\n-version = \\\"0.68.2\\\"\\n+version = \\\"0.68.3\\\"\\n # See more keys and their definitions at https://doc.rust-lang.org/cargo/reference/manifest.html\\n \\n [features]\\ndiff --git a/crates/swc_ecmascript/Cargo.toml b/crates/swc_ecmascript/Cargo.toml\\nindex 63680a0..775208a 100644\\n--- a/crates/swc_ecmascript/Cargo.toml\\n+++ b/crates/swc_ecmascript/Cargo.toml\\n@@ -6,7 +6,7 @@ edition = \\\"2021\\\"\\n license = \\\"Apache-2.0\\\"\\n name = \\\"swc_ecmascript\\\"\\n repository = \\\"https://github.com/swc-project/swc.git\\\"\\n-version = \\\"0.110.14\\\"\\n+version = \\\"0.110.15\\\"\\n \\n [package.metadata.docs.rs]\\n all-features = true\\n@@ -39,7 +39,7 @@ swc_ecma_dep_graph = {version = \\\"0.58.0\\\", path = \\\"../swc_ecma_dep_graph\\\", option\\n swc_ecma_minifier = {version = \\\"0.70.9\\\", path = \\\"../swc_ecma_minifier\\\", optional = true}\\n swc_ecma_parser = {version = \\\"0.87.0\\\", path = \\\"../swc_ecma_parser\\\", optional = true, default-features = false}\\n swc_ecma_preset_env = {version = \\\"0.86.1\\\", path = \\\"../swc_ecma_preset_env\\\", optional = true}\\n-swc_ecma_transforms = {version = \\\"0.113.3\\\", path = \\\"../swc_ecma_transforms\\\", optional = true}\\n+swc_ecma_transforms = {version = \\\"0.113.4\\\", path = \\\"../swc_ecma_transforms\\\", optional = true}\\n swc_ecma_utils = {version = \\\"0.64.0\\\", path = \\\"../swc_ecma_utils\\\", optional = true}\\n swc_ecma_visit = {version = \\\"0.51.1\\\", path = \\\"../swc_ecma_visit\\\", optional = true}\\n \\n\", \"diff --git a/.github/workflows/publish.yml b/.github/workflows/publish.yml\\nindex a986501..db33097 100644\\n--- a/.github/workflows/publish.yml\\n+++ b/.github/workflows/publish.yml\\n@@ -409,7 +409,7 @@ jobs:\\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\\n - name: Instal LLVM for Windows\\n if: ${{ startsWith(matrix.os, 'windows') }}\\n- run: choco install -y llvm --version 9.0.1\\n+ run: choco install -y --force llvm --version 9.0.1\\n - name: Set Env Variables for Windows\\n uses: allenevans/set-env@v2.0.0\\n if: ${{ startsWith(matrix.os, 'windows') }}\\ndiff --git a/.github/workflows/rust.yml b/.github/workflows/rust.yml\\nindex d45cdf9..8d87ed6 100644\\n--- a/.github/workflows/rust.yml\\n+++ b/.github/workflows/rust.yml\\n@@ -158,7 +158,7 @@ jobs:\\n run: vcpkg integrate install; vcpkg install openssl:x64-windows\\n - name: Instal LLVM for Windows\\n if: ${{ startsWith(matrix.os, 'windows') }}\\n- run: choco install -y llvm --version 9.0.1\\n+ run: choco install -y --force llvm --version 9.0.1\\n - name: Set Env Variables for Windows\\n uses: allenevans/set-env@v2.0.0\\n if: ${{ startsWith(matrix.os, 'windows') }}\\n\", \"diff --git a/client/src/components/Profile/AboutCard.tsx b/client/src/components/Profile/AboutCard.tsx\\nindex 3bd6e9a..e07ddb6 100644\\n--- a/client/src/components/Profile/AboutCard.tsx\\n+++ b/client/src/components/Profile/AboutCard.tsx\\n@@ -11,6 +11,7 @@ import { InfoCircleOutlined } from '@ant-design/icons';\\n \\n type Props = {\\n data: GeneralInfo;\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n class AboutCard extends React.Component {\\ndiff --git a/client/src/components/Profile/ContactsCard.tsx b/client/src/components/Profile/ContactsCard.tsx\\nindex 6fe80a3..3a35c9f 100644\\n--- a/client/src/components/Profile/ContactsCard.tsx\\n+++ b/client/src/components/Profile/ContactsCard.tsx\\n@@ -12,8 +12,11 @@ import { ContactsOutlined } from '@ant-design/icons';\\n \\n type Props = {\\n data: Contacts;\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n+type Contact = { name: string, value?: string };\\n+\\n class ContactsCard extends React.Component {\\n render() {\\n const { email, telegram, phone, skype, notes } = this.props.data;\\n@@ -32,7 +35,7 @@ class ContactsCard extends React.Component {\\n }, {\\n name: 'Notes',\\n value: notes,\\n- }].filter(({ value }: { name: string, value: string | null }) => value);\\n+ }].filter(({ value }: Contact) => value);\\n \\n return (\\n {\\n (\\n+ renderItem={({ name, value }: Contact) => (\\n \\n {name}: {value}\\n \\ndiff --git a/client/src/components/Profile/EducationCard.tsx b/client/src/components/Profile/EducationCard.tsx\\nindex 4279c9f..b409c29 100644\\n--- a/client/src/components/Profile/EducationCard.tsx\\n+++ b/client/src/components/Profile/EducationCard.tsx\\n@@ -12,6 +12,7 @@ import { ReadOutlined } from '@ant-design/icons';\\n \\n type Props = {\\n data: GeneralInfo;\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n class EducationCard extends React.Component {\\ndiff --git a/client/src/components/Profile/EnglishCard.tsx b/client/src/components/Profile/EnglishCard.tsx\\nindex d8f8ab4..2d5efa0 100644\\n--- a/client/src/components/Profile/EnglishCard.tsx\\n+++ b/client/src/components/Profile/EnglishCard.tsx\\n@@ -11,6 +11,7 @@ import { TagOutlined } from '@ant-design/icons';\\n \\n type Props = {\\n data: GeneralInfo;\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n class EnglishCard extends React.Component {\\ndiff --git a/client/src/components/Profile/MainCard.tsx b/client/src/components/Profile/MainCard.tsx\\nindex cbfb71b..c0d49cc 100644\\n--- a/client/src/components/Profile/MainCard.tsx\\n+++ b/client/src/components/Profile/MainCard.tsx\\n@@ -4,6 +4,8 @@ import { GithubAvatar } from 'components';\\n import {\\n Card,\\n Typography,\\n+ Drawer,\\n+ Checkbox,\\n } from 'antd';\\n \\n const { Title, Paragraph } = Typography;\\n@@ -11,30 +13,70 @@ const { Title, Paragraph } = Typography;\\n import {\\n GithubFilled,\\n EnvironmentFilled,\\n+ EditOutlined,\\n+ SettingOutlined,\\n } from '@ant-design/icons';\\n \\n type Props = {\\n data: GeneralInfo;\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n-class MainCard extends React.Component {\\n+type State = {\\n+ isSettingsVisible: boolean;\\n+}\\n+\\n+class MainCard extends React.Component {\\n+ state = {\\n+ isSettingsVisible: false,\\n+ }\\n+\\n+ private showSettings = () => {\\n+ this.setState({ isSettingsVisible: true });\\n+ }\\n+\\n+ private hideSettings = () => {\\n+ this.setState({ isSettingsVisible: false });\\n+ }\\n+\\n render() {\\n const { githubId, name, locationName } = this.props.data;\\n+ const { isSettingsVisible } = this.state;\\n+\\n return (\\n- \\n- \\n- {name}\\n- \\n- \\n- {githubId}\\n- \\n- \\n- \\n- \\n- {locationName}\\n- \\n- \\n- \\n+ <>\\n+\\n+ ,\\n+ ,\\n+ ]}\\n+ >\\n+ \\n+ {name}\\n+ \\n+ \\n+ {githubId}\\n+ \\n+ \\n+ \\n+ \\n+ {locationName}\\n+ \\n+ \\n+ \\n+ Nobody\\n+ \\n+ \\n+ \\n );\\n }\\n }\\ndiff --git a/client/src/components/Profile/MentorStatsCard.tsx b/client/src/components/Profile/MentorStatsCard.tsx\\nindex ca54480..1ec3b9c 100644\\n--- a/client/src/components/Profile/MentorStatsCard.tsx\\n+++ b/client/src/components/Profile/MentorStatsCard.tsx\\n@@ -18,6 +18,7 @@ import {\\n \\n type Props = {\\n data: MentorStats[];\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n type State = {\\n@@ -80,7 +81,7 @@ class MentorStatsCard extends React.Component {\\n {courseName}{locationName && ` / ${locationName}`}\\n

\\n {\\n- idx === 0 && (\\n+ students ? idx === 0 && (\\n {\\n \\n )}\\n />\\n- )\\n+ ) :

Doesn't have students at this course yet

\\n }\\n
\\n- \\n+ {\\n+ students && \\n+ }\\n \\n )}\\n />\\ndiff --git a/client/src/components/Profile/MentorStatsModal.tsx b/client/src/components/Profile/MentorStatsModal.tsx\\nindex 47b5f2a..0e94cc1 100644\\n--- a/client/src/components/Profile/MentorStatsModal.tsx\\n+++ b/client/src/components/Profile/MentorStatsModal.tsx\\n@@ -38,7 +38,7 @@ class MentorStatsModal extends React.Component {\\n >\\n \\n {\\n- students.map(({ name, githubId, isExpelled, totalScore }) => {\\n+ students?.map(({ name, githubId, isExpelled, totalScore }) => {\\n const profile = `/profile?githubId=${githubId}`;\\n const guithubLink = `https://github.com/${githubId}`;\\n const privateRepoLink = `https://github.com/rolling-scopes-school/${githubId}-${courseYearPostfix}`;\\ndiff --git a/client/src/components/Profile/PublicFeedbackCard.tsx b/client/src/components/Profile/PublicFeedbackCard.tsx\\nindex 2f8a999..6ce1862 100644\\n--- a/client/src/components/Profile/PublicFeedbackCard.tsx\\n+++ b/client/src/components/Profile/PublicFeedbackCard.tsx\\n@@ -22,6 +22,7 @@ import {\\n \\n type Props = {\\n data: PublicFeedback[];\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n interface State {\\ndiff --git a/client/src/components/Profile/StudentStatsCard.tsx b/client/src/components/Profile/StudentStatsCard.tsx\\nindex c811640..b472e49 100644\\n--- a/client/src/components/Profile/StudentStatsCard.tsx\\n+++ b/client/src/components/Profile/StudentStatsCard.tsx\\n@@ -18,6 +18,7 @@ import {\\n \\n type Props = {\\n data: StudentStats[];\\n+ isEditingModeEnabled: boolean;\\n };\\n \\n type State = {\\ndiff --git a/client/src/pages/profile/index.tsx b/client/src/pages/profile/index.tsx\\nindex 68b2a70..b6ffb1a 100644\\n--- a/client/src/pages/profile/index.tsx\\n+++ b/client/src/pages/profile/index.tsx\\n@@ -1,6 +1,7 @@\\n import * as React from 'react';\\n import {\\n Result,\\n+ Button,\\n } from 'antd';\\n import css from 'styled-jsx/css';\\n import Masonry from 'react-masonry-css';\\n@@ -23,18 +24,25 @@ import CoreJsIviewsCard from 'components/Profile/CoreJsIviewsCard';\\n import { CoreJsInterviewData } from 'components/Profile/CoreJsIviewsCard';\\n import PreScreeningIviewCard from 'components/Profile/PreScreeningIviewCard';\\n \\n+import {\\n+ EditOutlined,\\n+ EyeOutlined,\\n+} from '@ant-design/icons';\\n+\\n type Props = {\\n router: NextRouter;\\n session: Session;\\n };\\n \\n type State = {\\n+ isEditingModeEnabled: boolean;\\n profile: ProfileInfo | null;\\n isLoading: boolean;\\n };\\n \\n class ProfilePage extends React.Component {\\n state: State = {\\n+ isEditingModeEnabled: false,\\n isLoading: true,\\n profile: null,\\n };\\n@@ -79,6 +87,12 @@ class ProfilePage extends React.Component {\\n }\\n };\\n \\n+ private toggleEditViewProfileButton = () => {\\n+ const { isEditingModeEnabled } = this.state;\\n+\\n+ this.setState({ isEditingModeEnabled: !isEditingModeEnabled });\\n+ }\\n+\\n async componentDidMount() {\\n await this.fetchData();\\n }\\n@@ -90,21 +104,29 @@ class ProfilePage extends React.Component {\\n }\\n \\n render() {\\n- const { profile } = this.state;\\n+ const { profile, isEditingModeEnabled } = this.state;\\n \\n const cards = [\\n- profile?.generalInfo && ,\\n- profile?.generalInfo?.aboutMyself && ,\\n- profile?.generalInfo?.englishLevel && ,\\n- profile?.generalInfo?.educationHistory.length && ,\\n- profile?.contacts && ,\\n- profile?.publicFeedback.length && ,\\n- profile?.studentStats.length && ,\\n- profile?.mentorStats.length && ,\\n- profile?.studentStats.length &&\\n- this.hadStudentCoreJSInterview(profile.studentStats) &&\\n+ profile?.generalInfo &&\\n+ ,\\n+ profile?.generalInfo?.aboutMyself &&\\n+ ,\\n+ profile?.generalInfo?.englishLevel &&\\n+ ,\\n+ profile?.generalInfo?.educationHistory?.length &&\\n+ ,\\n+ profile?.contacts &&\\n+ ,\\n+ profile?.publicFeedback?.length &&\\n+ ,\\n+ profile?.studentStats?.length &&\\n+ ,\\n+ profile?.mentorStats?.length &&\\n+ ,\\n+ profile?.studentStats?.length && this.hadStudentCoreJSInterview(profile.studentStats) &&\\n ,\\n- profile?.stageInterviewFeedback.length && ,\\n+ profile?.stageInterviewFeedback.length &&\\n+ ,\\n ].filter(Boolean) as JSX.Element[];\\n \\n return (\\n@@ -114,6 +136,17 @@ class ProfilePage extends React.Component {\\n {\\n this.state.profile\\n ?
\\n+ \\n+ {\\n+ isEditingModeEnabled ?\\n+ Edit :\\n+ View\\n+ }\\n+ \\n async (ctx: Router.RouterContext) => {\\n- const {\\n- // id: userId,\\n- githubId: userGithubId,\\n- } = ctx.state!.user as IUserSession;\\n+ const { githubId: userGithubId } = ctx.state!.user as IUserSession;\\n // const { isAdmin, roles } = ctx.state!.user as IUserSession;\\n- const { githubId } = ctx.query as { githubId: string | undefined };\\n-\\n+ const { githubId = userGithubId } = ctx.query as { githubId: string | undefined };\\n // console.log('GITHUB =>', githubId);\\n // console.log('ADMIN =>', isAdmin);\\n // console.log('ROLES =>', roles);\\n@@ -75,16 +71,28 @@ export const getProfileInfo = (_: ILogger) => async (ctx: Router.RouterContext) \\n return setResponse(ctx, NOT_FOUND);\\n }\\n \\n+ const isProfileOwner = githubId === userGithubId;\\n+ console.log('isProfileOwner', isProfileOwner);\\n // await getRepository(ProfilePermissions).save({ userId });\\n \\n- const permissions = await getPermissions(userGithubId, githubId);\\n+ const permissions = await getPermissions(userGithubId, githubId, { isProfileOwner });\\n \\n- console.log(JSON.stringify(permissions, null, 2));\\n+ const { isProfileVisible, isPublicFeedbackVisible, isMentorStatsVisible, isStudentStatsVisible } = permissions;\\n+\\n+ if (!isProfileVisible && !isProfileOwner) {\\n+ return setResponse(ctx, FORBIDDEN);\\n+ }\\n+\\n+ if (isProfileOwner) {\\n+ const ownerPermissions = await getOwnerPermissions(userGithubId);\\n+\\n+ console.log('OWN =>', ownerPermissions);\\n+ }\\n \\n const { generalInfo, contacts } = await getUserInfo(githubId, permissions);\\n- const publicFeedback = await getPublicFeedback(githubId);\\n- const mentorStats = await getMentorStats(githubId);\\n- const studentStats = await getStudentStats(githubId);\\n+ const publicFeedback = isPublicFeedbackVisible ? await getPublicFeedback(githubId) : undefined;\\n+ const mentorStats = isMentorStatsVisible ? await getMentorStats(githubId) : undefined;\\n+ const studentStats = isStudentStatsVisible ? await getStudentStats(githubId) : undefined;\\n const stageInterviewFeedback = await getStageInterviewFeedback(githubId);\\n \\n const profileInfo: ProfileInfo = {\\n@@ -96,7 +104,8 @@ export const getProfileInfo = (_: ILogger) => async (ctx: Router.RouterContext) \\n studentStats,\\n };\\n \\n- // console.log(JSON.stringify(profileInfo, null, 2));\\n+ console.log(JSON.stringify(permissions, null, 2));\\n+ console.log(JSON.stringify(profileInfo, null, 2));\\n \\n setResponse(ctx, OK, profileInfo);\\n };\\ndiff --git a/server/src/routes/profile/mentor-stats.ts b/server/src/routes/profile/mentor-stats.ts\\nindex 843a2f7..72e6b30 100644\\n--- a/server/src/routes/profile/mentor-stats.ts\\n+++ b/server/src/routes/profile/mentor-stats.ts\\n@@ -36,11 +36,11 @@ export const getMentorStats = async (githubId: string): Promise =\\n studentIsExpelledStatuses,\\n studentTotalScores,\\n }: any) => {\\n- const students = studentGithubIds.map((githubId: string, idx: number) => ({\\n+ const students = studentGithubIds[0] ? studentGithubIds.map((githubId: string, idx: number) => ({\\n githubId,\\n name: getFullName(studentFirstNames[idx], studentLastNames[idx], githubId),\\n isExpelled: studentIsExpelledStatuses[idx],\\n totalScore: studentTotalScores[idx],\\n- }));\\n+ })) : undefined;\\n return { courseName, locationName, courseFullName, students };\\n });\\ndiff --git a/server/src/routes/profile/permissions.ts b/server/src/routes/profile/permissions.ts\\nindex 61924a8..b40121c 100644\\n--- a/server/src/routes/profile/permissions.ts\\n+++ b/server/src/routes/profile/permissions.ts\\n@@ -1,3 +1,4 @@\\n+import { get, mapValues } from 'lodash';\\n import { getRepository } from 'typeorm';\\n import {\\n User,\\n@@ -8,6 +9,12 @@ import {\\n TaskInterviewResult,\\n StageInterview,\\n } from '../../models';\\n+import {\\n+ PublicVisibilitySettings,\\n+ VisibilitySettings,\\n+ defaultPublicVisibilitySettings,\\n+ defaultVisibilitySettings,\\n+} from '../../models/profilePermissions';\\n \\n interface Relations {\\n student: string;\\n@@ -19,7 +26,43 @@ interface Relations {\\n \\n type RelationRole = 'student' | 'mentor' | 'all';\\n \\n-const getAllProfilePermissions = async (githubId: string): Promise => (\\n+interface SuperAccessRights {\\n+ isProfileOwner: boolean;\\n+}\\n+\\n+interface ConfigurableProfilePermissions {\\n+ isProfileVisible: PublicVisibilitySettings;\\n+ isAboutVisible: VisibilitySettings;\\n+ isEducationVisible: VisibilitySettings;\\n+ isEnglishVisible: VisibilitySettings;\\n+ isEmailVisible: VisibilitySettings;\\n+ isTelegramVisible: VisibilitySettings;\\n+ isSkypeVisible: VisibilitySettings;\\n+ isPhoneVisible: VisibilitySettings;\\n+ isContactsNotesVisible: VisibilitySettings;\\n+ isLinkedInVisible: VisibilitySettings;\\n+ isPublicFeedbackVisible: VisibilitySettings;\\n+ isMentorStatsVisible: VisibilitySettings;\\n+ isStudentStatsVisible: VisibilitySettings;\\n+}\\n+\\n+export interface Permissions {\\n+ isProfileVisible: boolean;\\n+ isAboutVisible: boolean;\\n+ isEducationVisible: boolean;\\n+ isEnglishVisible: boolean;\\n+ isEmailVisible: boolean;\\n+ isTelegramVisible: boolean;\\n+ isSkypeVisible: boolean;\\n+ isPhoneVisible: boolean;\\n+ isContactsNotesVisible: boolean;\\n+ isLinkedInVisible: boolean;\\n+ isPublicFeedbackVisible: boolean;\\n+ isMentorStatsVisible: boolean;\\n+ isStudentStatsVisible: boolean;\\n+}\\n+\\n+const getConfigurableProfilePermissions = async (githubId: string): Promise => (\\n (await getRepository(ProfilePermissions)\\n .createQueryBuilder('pp')\\n .select('\\\"pp\\\".\\\"isProfileVisible\\\" AS \\\"isProfileVisible\\\"')\\n@@ -85,16 +128,67 @@ const getRelationRole = async (userGithubId: string, requestedGithubId: string):\\n return 'all';\\n };\\n \\n-const matchPermissions = (permissions: any, role: RelationRole) => {\\n- const obj: any = {};\\n- Object.keys(permissions).forEach((key) => {\\n- obj[key] = permissions[key].all || permissions[key][role];\\n- });\\n- return obj;\\n+const matchPermissions = (\\n+ permissions: ConfigurableProfilePermissions,\\n+ role: RelationRole,\\n+ { isProfileOwner }: SuperAccessRights,\\n+): Permissions => {\\n+ const p: Permissions = {\\n+ isProfileVisible: false,\\n+ isAboutVisible: false,\\n+ isEducationVisible: false,\\n+ isEnglishVisible: false,\\n+ isEmailVisible: false,\\n+ isTelegramVisible: false,\\n+ isSkypeVisible: false,\\n+ isPhoneVisible: false,\\n+ isContactsNotesVisible: false,\\n+ isLinkedInVisible: false,\\n+ isPublicFeedbackVisible: false,\\n+ isMentorStatsVisible: false,\\n+ isStudentStatsVisible: false,\\n+ };\\n+\\n+ // (Object.keys(p) as (keyof Permissions)[]).forEach((key) => {\\n+ // p[key] = isProfileOwner || permissions[key].all || permissions[key][role];\\n+ // });\\n+\\n+ // return p;\\n+\\n+ return mapValues(p, (_, key) => isProfileOwner ||\\n+ get(permissions, `${key}.all`) ||\\n+ get(permissions, `${key}.${role}`) ||\\n+ false,\\n+ );\\n };\\n \\n-export const getPermissions = async (userGithubId: string, requestedGithubId: string) => {\\n- const permissions = await getAllProfilePermissions(requestedGithubId);\\n+export const getPermissions = async (\\n+ userGithubId: string,\\n+ requestedGithubId: string,\\n+ superAccessRights: SuperAccessRights,\\n+) => {\\n+ const permissions = await getConfigurableProfilePermissions(requestedGithubId);\\n const role = await getRelationRole(userGithubId, requestedGithubId);\\n- return matchPermissions(permissions, role);\\n+ return matchPermissions(permissions, role, superAccessRights);\\n+};\\n+\\n+export const getOwnerPermissions = async (githubId: string) => {\\n+ const permissions = await getConfigurableProfilePermissions(githubId);\\n+ const p: ConfigurableProfilePermissions = {\\n+ isProfileVisible: defaultPublicVisibilitySettings,\\n+ isAboutVisible: defaultVisibilitySettings,\\n+ isEducationVisible: defaultVisibilitySettings,\\n+ isEnglishVisible: defaultVisibilitySettings,\\n+ isEmailVisible: defaultVisibilitySettings,\\n+ isTelegramVisible: defaultVisibilitySettings,\\n+ isSkypeVisible: defaultVisibilitySettings,\\n+ isPhoneVisible: defaultVisibilitySettings,\\n+ isContactsNotesVisible: defaultVisibilitySettings,\\n+ isLinkedInVisible: defaultVisibilitySettings,\\n+ isPublicFeedbackVisible: defaultVisibilitySettings,\\n+ isMentorStatsVisible: defaultVisibilitySettings,\\n+ isStudentStatsVisible: defaultVisibilitySettings,\\n+ };\\n+\\n+ return mapValues(p, (value, key) => get(permissions, key, value));\\n };\\ndiff --git a/server/src/routes/profile/user-info.ts b/server/src/routes/profile/user-info.ts\\nindex 5b871e0..1998ed0 100644\\n--- a/server/src/routes/profile/user-info.ts\\n+++ b/server/src/routes/profile/user-info.ts\\n@@ -2,23 +2,53 @@ import { getRepository } from 'typeorm';\\n import { UserInfo } from '../../../../common/models/profile';\\n import { getFullName } from '../../lib/utils';\\n import { User } from '../../models';\\n+import { Permissions } from './permissions';\\n \\n-export const getUserInfo = async (githubId: string, permissions: any): Promise => {\\n- const { isAboutVisible } = permissions;\\n+export const getUserInfo = async (githubId: string, permissions: Permissions): Promise => {\\n+ const {\\n+ isAboutVisible,\\n+ isEducationVisible,\\n+ isEnglishVisible,\\n+ isPhoneVisible,\\n+ isEmailVisible,\\n+ isTelegramVisible,\\n+ isSkypeVisible,\\n+ isContactsNotesVisible,\\n+ } = permissions;\\n \\n const query = await getRepository(User)\\n .createQueryBuilder('user')\\n .select('\\\"user\\\".\\\"firstName\\\" AS \\\"firstName\\\", \\\"user\\\".\\\"lastName\\\" AS \\\"lastName\\\"')\\n .addSelect('\\\"user\\\".\\\"githubId\\\" AS \\\"githubId\\\"')\\n- .addSelect('\\\"user\\\".\\\"locationName\\\" AS \\\"locationName\\\"')\\n- .addSelect('\\\"user\\\".\\\"educationHistory\\\" AS \\\"educationHistory\\\"')\\n- .addSelect('\\\"user\\\".\\\"employmentHistory\\\" AS \\\"employmentHistory\\\"')\\n- .addSelect('\\\"user\\\".\\\"englishLevel\\\" AS \\\"englishLevel\\\"')\\n- .addSelect('\\\"user\\\".\\\"contactsPhone\\\" AS \\\"contactsPhone\\\"')\\n- .addSelect('\\\"user\\\".\\\"contactsEmail\\\" AS \\\"contactsEmail\\\"')\\n- .addSelect('\\\"user\\\".\\\"contactsTelegram\\\" AS \\\"contactsTelegram\\\"')\\n- .addSelect('\\\"user\\\".\\\"contactsSkype\\\" AS \\\"contactsSkype\\\"')\\n- .addSelect('\\\"user\\\".\\\"contactsNotes\\\" AS \\\"contactsNotes\\\"');\\n+ .addSelect('\\\"user\\\".\\\"locationName\\\" AS \\\"locationName\\\"');\\n+\\n+ if (isEducationVisible) {\\n+ query.addSelect('\\\"user\\\".\\\"educationHistory\\\" AS \\\"educationHistory\\\"');\\n+ }\\n+\\n+ if (isEnglishVisible) {\\n+ query.addSelect('\\\"user\\\".\\\"englishLevel\\\" AS \\\"englishLevel\\\"');\\n+ }\\n+\\n+ if (isPhoneVisible) {\\n+ query.addSelect('\\\"user\\\".\\\"contactsPhone\\\" AS \\\"contactsPhone\\\"');\\n+ }\\n+\\n+ if (isEmailVisible) {\\n+ query.addSelect('\\\"user\\\".\\\"contactsEmail\\\" AS \\\"contactsEmail\\\"');\\n+ }\\n+\\n+ if (isTelegramVisible) {\\n+ query.addSelect('\\\"user\\\".\\\"contactsTelegram\\\" AS \\\"contactsTelegram\\\"');\\n+ }\\n+\\n+ if (isSkypeVisible) {\\n+ query.addSelect('\\\"user\\\".\\\"contactsSkype\\\" AS \\\"contactsSkype\\\"');\\n+ }\\n+\\n+ if (isContactsNotesVisible) {\\n+ query.addSelect('\\\"user\\\".\\\"contactsNotes\\\" AS \\\"contactsNotes\\\"');\\n+ }\\n \\n if (isAboutVisible) {\\n query.addSelect('\\\"user\\\".\\\"aboutMyself\\\" AS \\\"aboutMyself\\\"');\\n@@ -33,7 +63,6 @@ export const getUserInfo = async (githubId: string, permissions: any): Promise"},"diff":{"kind":"string","value":"[\"diff --git a/ibis/expr/analysis.py b/ibis/expr/analysis.py\\nindex bb17a7a..975c658 100644\\n--- a/ibis/expr/analysis.py\\n+++ b/ibis/expr/analysis.py\\n@@ -39,7 +39,9 @@ def sub_for(expr, substitutions):\\n An Ibis expression\\n \\\"\\\"\\\"\\n \\n- def fn(node, mapping={k.op(): v for k, v in substitutions}):\\n+ mapping = {k.op(): v for k, v in substitutions}\\n+\\n+ def fn(node):\\n try:\\n return mapping[node]\\n except KeyError:\\n\", \"diff --git a/package.json b/package.json\\nindex c8051d2..b0a97fb 100644\\n--- a/package.json\\n+++ b/package.json\\n@@ -60,6 +60,7 @@\\n \\\"babel-cli\\\": \\\"^6.16.0\\\",\\n \\\"babel-core\\\": \\\"^6.16.0\\\",\\n \\\"babel-eslint\\\": \\\"^7.0.0\\\",\\n+ \\\"babel-loader\\\": \\\"^6.2.5\\\",\\n \\\"babel-plugin-transform-class-properties\\\": \\\"^6.10.2\\\",\\n \\\"babel-plugin-transform-flow-strip-types\\\": \\\"^6.14.0\\\",\\n \\\"babel-preset-es2015-node6\\\": \\\"^0.3.0\\\",\\n@@ -82,6 +83,7 @@\\n \\\"eslint-plugin-react\\\": \\\"^6.3.0\\\",\\n \\\"flow-bin\\\": \\\"^0.33.0\\\",\\n \\\"jsdom\\\": \\\"^9.4.2\\\",\\n+ \\\"json-loader\\\": \\\"^0.5.4\\\",\\n \\\"jsx-chai\\\": \\\"^4.0.0\\\",\\n \\\"mocha\\\": \\\"^3.0.2\\\",\\n \\\"mock-require\\\": \\\"^1.3.0\\\",\\n@@ -91,6 +93,8 @@\\n \\\"rimraf\\\": \\\"^2.5.2\\\",\\n \\\"sinon\\\": \\\"^1.17.6\\\",\\n \\\"sinon-chai\\\": \\\"^2.8.0\\\",\\n- \\\"watch\\\": \\\"^1.0.0\\\"\\n+ \\\"source-map-support\\\": \\\"^0.4.3\\\",\\n+ \\\"watch\\\": \\\"^1.0.0\\\",\\n+ \\\"webpack\\\": \\\"^1.13.2\\\"\\n }\\n }\\ndiff --git a/webpack.config.js b/webpack.config.js\\nnew file mode 100644\\nindex 0000000..0ca6da1\\n--- /dev/null\\n+++ b/webpack.config.js\\n@@ -0,0 +1,44 @@\\n+const webpack = require('webpack');\\n+const path = require('path');\\n+const fs = require('fs');\\n+\\n+const nodeModules = {\\n+ zmq: 'commonjs zmq',\\n+ jmp: 'commonjs jmp',\\n+ github: 'commonjs github',\\n+};\\n+\\n+module.exports = {\\n+ entry: './src/notebook/index.js',\\n+ target: 'electron-renderer',\\n+ output: {\\n+ path: path.join(__dirname, 'app', 'build'),\\n+ filename: 'webpacked-notebook.js'\\n+ },\\n+ module: {\\n+ loaders: [\\n+ { test: /\\\\.js$/, exclude: /node_modules/, loaders: ['babel'] },\\n+ { test: /\\\\.json$/, loader: 'json-loader' },\\n+ ]\\n+ },\\n+ resolve: {\\n+ extensions: ['', '.js', '.jsx'],\\n+ root: path.join(__dirname, 'app'),\\n+ // Webpack 1\\n+ modulesDirectories: [\\n+ path.resolve(__dirname, 'app', 'node_modules'),\\n+ path.resolve(__dirname, 'node_modules'),\\n+ ],\\n+ // Webpack 2\\n+ modules: [\\n+ path.resolve(__dirname, 'app', 'node_modules'),\\n+ ],\\n+ },\\n+ externals: nodeModules,\\n+ plugins: [\\n+ new webpack.IgnorePlugin(/\\\\.(css|less)$/),\\n+ new webpack.BannerPlugin('require(\\\"source-map-support\\\").install();',\\n+ { raw: true, entryOnly: false })\\n+ ],\\n+ devtool: 'sourcemap'\\n+};\\n\", \"diff --git a/.circleci/config.yml b/.circleci/config.yml\\nindex 3ae6728..a5f2d2f 100644\\n--- a/.circleci/config.yml\\n+++ b/.circleci/config.yml\\n@@ -12,7 +12,7 @@\\n # The CI for every PR and merge to main runs tests, fmt, lints and compiles debug binaries\\n #\\n # On main if all these checks pass it will then additionally compile in \\\"release\\\" mode and\\n-# publish a docker image to quay.io/influxdb/fusion:$COMMIT_SHA\\n+# publish a docker image to quay.io/influxdb/iox:$COMMIT_SHA\\n #\\n # Manual CI Image:\\n #\\n@@ -317,11 +317,11 @@ jobs:\\n #\\n # Uses the latest ci_image (influxdb/rust below) to build a release binary and\\n # copies it to a minimal container image based upon `rust:slim-buster`. This\\n- # minimal image is then pushed to `quay.io/influxdb/fusion:${BRANCH}` with '/'\\n+ # minimal image is then pushed to `quay.io/influxdb/iox:${BRANCH}` with '/'\\n # repaced by '.' - as an example:\\n #\\n # git branch: dom/my-awesome-feature/perf\\n- # container: quay.io/influxdb/fusion:dom.my-awesome-feature.perf\\n+ # container: quay.io/influxdb/iox:dom.my-awesome-feature.perf\\n #\\n # Subsequent CI runs will overwrite the tag if you push more changes, so watch\\n # out for parallel CI runs!\\n@@ -365,7 +365,7 @@ jobs:\\n sudo apt-get update\\n sudo apt-get install -y docker.io\\n - run: |\\n- echo \\\"$QUAY_PASS\\\" | docker login quay.io --username $QUAY_USER --password-stdin\\n+ echo \\\"$QUAY_INFLUXDB_IOX_PASS\\\" | docker login quay.io --username $QUAY_INFLUXDB_IOX_USER --password-stdin\\n - run:\\n # Docker has functionality to support per-Dockerfile .dockerignore\\n # This was added in https://github.com/moby/buildkit/pull/901\\n@@ -379,8 +379,8 @@ jobs:\\n echo sha256sum after build is\\n sha256sum target/release/influxdb_iox\\n COMMIT_SHA=$(git rev-parse --short HEAD)\\n- docker build -t quay.io/influxdb/fusion:$COMMIT_SHA -f docker/Dockerfile.iox .\\n- docker push quay.io/influxdb/fusion:$COMMIT_SHA\\n+ docker build -t quay.io/influxdb/iox:$COMMIT_SHA -f docker/Dockerfile.iox .\\n+ docker push quay.io/influxdb/iox:$COMMIT_SHA\\n echo \\\"export COMMIT_SHA=${COMMIT_SHA}\\\" >> $BASH_ENV\\n - run:\\n name: Deploy tags\\n\", \"diff --git a/packages/nc-gui/composables/useSharedView.ts b/packages/nc-gui/composables/useSharedView.ts\\nindex cb0c5ea..f67a6c9 100644\\n--- a/packages/nc-gui/composables/useSharedView.ts\\n+++ b/packages/nc-gui/composables/useSharedView.ts\\n@@ -17,7 +17,7 @@ export function useSharedView() {\\n \\n const { appInfo } = $(useGlobal())\\n \\n- const { loadProject } = useProject()\\n+ const { project } = useProject()\\n \\n const appInfoDefaultLimit = appInfo.defaultLimit || 25\\n \\n@@ -76,7 +76,16 @@ export function useSharedView() {\\n \\n await setMeta(viewMeta.model)\\n \\n- await loadProject(true, viewMeta.project_id)\\n+ // if project is not defined then set it with an object containing base\\n+ if (!project.value?.bases)\\n+ project.value = {\\n+ bases: [\\n+ {\\n+ id: viewMeta.base_id,\\n+ type: viewMeta.client,\\n+ },\\n+ ],\\n+ }\\n \\n const relatedMetas = { ...viewMeta.relatedMetas }\\n Object.keys(relatedMetas).forEach((key) => setMeta(relatedMetas[key]))\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"ad52e1d67fd77f0b6a73fbf989b33f9abf395ecc\", \"4ab28fc2e63e975a0c77e18ae644f34fa5f8771a\", \"f751bb5426b87f82096d620f1cd6203badf45d58\", \"4986a5892fb00bd5a6b2065ad8cfefbc36052dd7\"]"},"types":{"kind":"string","value":"[\"refactor\", \"build\", \"ci\", \"fix\"]"}}},{"rowIdx":922,"cells":{"commit_message":{"kind":"string","value":"generate terminate end event compatible execution steps part 1\n\nThe random execution tests don't know the concept of flow scopes. This makes it challenging to generate a correct execution path for terminate end events, as they terminate a specific flow scope. Processing should continue as normal once the flow scope has been terminated.\n\nWhilst we don't have flow scopes, we do have execution path segments. These segments don't map 1 to 1 to flow scopes. However, since every flow scope starts a new segment we can use these segments to get the desired behavior.\n\nEach segment must keep track whether is has reached a terminate end event. If this is the case that means that we don't expect any further execution steps. We can isolate this behavior in a single location, during the appending of one segment to another segment.\nIn order to differentiate between flow scopes a new append method has been added which takes the boolean `changesFlowScope` as a parameter. Blockbuilder where the flow scope changes (e.g. SubProcessBlockBuilder) can use this to indicate that even though a terminate end event has been reached. Execution steps after this specific segment still need to added to complete the process.\n\nWhen a segment is appended to a different segment and the flow scope does not change we can use the segment that should be appended to identify whether new segment can still be added to the current segment. If passed segment has reached a terminate end event and the flow scope has not been changed it is guaranteed that the current segment is in the same flow scope has the previous segment and thus has also reached the terminate end event.,release for ppc64\n\ncloses #3703\n\nSigned-off-by: Carlos A Becker ,increase timeout of multiregion failover test\n\nDue to the nature of the test, restarts and failovers can take long. If the recovery takes longer than 15m, then the test will fail unnecessarily. Since we are not really testing for how was it can recover, it is ok to increase the maxInstanceDuration.,treeview width fix\n\nSigned-off-by: Raju Udava <86527202+dstala@users.noreply.github.com>"},"diff":{"kind":"string","value":"[\"diff --git a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\nindex da33c23..23c43be 100644\\n--- a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\n+++ b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\n@@ -29,6 +29,10 @@ import org.apache.commons.lang3.builder.ToStringStyle;\\n */\\n public final class ExecutionPathSegment {\\n \\n+ // If we have reached a terminate end event we want to stop generating execution steps for a\\n+ // specific flow scope. By setting this flag to true no new execution steps will be added for the\\n+ // flow scope this segment is in.\\n+ private boolean reachedTerminateEndEvent = false;\\n private final List scheduledSteps = new ArrayList<>();\\n private final Map variableDefaults = new HashMap<>();\\n \\n@@ -87,10 +91,28 @@ public final class ExecutionPathSegment {\\n new ScheduledExecutionStep(logicalPredecessor, executionPredecessor, executionStep));\\n }\\n \\n+ /**\\n+ * Appends the steps of the passed execution path segment to the current segment.\\n+ *\\n+ * @param pathToAdd execution path segment to append to this segment\\n+ */\\n public void append(final ExecutionPathSegment pathToAdd) {\\n+ append(pathToAdd, false);\\n+ }\\n+\\n+ /**\\n+ * Appends the step of the passed execution path segment to the current segment if the current\\n+ *\\n+ * @param pathToAdd\\n+ * @param changesFlowScope\\n+ */\\n+ public void append(final ExecutionPathSegment pathToAdd, final boolean changesFlowScope) {\\n mergeVariableDefaults(pathToAdd);\\n \\n- pathToAdd.getScheduledSteps().forEach(this::append);\\n+ if (!hasReachedTerminateEndEvent() || changesFlowScope) {\\n+ pathToAdd.getScheduledSteps().forEach(this::append);\\n+ }\\n+ reachedTerminateEndEvent = pathToAdd.hasReachedTerminateEndEvent() && !changesFlowScope;\\n }\\n \\n public void append(final ScheduledExecutionStep scheduledExecutionStep) {\\n@@ -259,6 +281,14 @@ public final class ExecutionPathSegment {\\n return ToStringBuilder.reflectionToString(this, ToStringStyle.SHORT_PREFIX_STYLE);\\n }\\n \\n+ public boolean hasReachedTerminateEndEvent() {\\n+ return reachedTerminateEndEvent;\\n+ }\\n+\\n+ public void setReachedTerminateEndEvent(final boolean reachedTerminateEndEvent) {\\n+ this.reachedTerminateEndEvent = reachedTerminateEndEvent;\\n+ }\\n+\\n /**\\n * An execution boundary is the point where automatic and non-automatic {@link\\n * ScheduledExecutionStep}'s meet each other. This class contains information about the existing\\n\", \"diff --git a/.goreleaser.yaml b/.goreleaser.yaml\\nindex 46901cb..7d4d355 100644\\n--- a/.goreleaser.yaml\\n+++ b/.goreleaser.yaml\\n@@ -25,6 +25,7 @@ builds:\\n - amd64\\n - arm\\n - arm64\\n+ - ppc64\\n goarm:\\n - \\\"7\\\"\\n mod_timestamp: '{{ .CommitTimestamp }}'\\n\", \"diff --git a/.github/workflows/e2e-testbench.yaml b/.github/workflows/e2e-testbench.yaml\\nindex 708f97f..fd0b918 100644\\n--- a/.github/workflows/e2e-testbench.yaml\\n+++ b/.github/workflows/e2e-testbench.yaml\\n@@ -31,6 +31,11 @@ on:\\n default: null\\n required: false\\n type: string\\n+ maxInstanceDuration:\\n+ description: 'If an instance takes longer than the given duration to complete, test will fail.'\\n+ default: '15m'\\n+ required: false\\n+ type: string\\n \\n workflow_call:\\n inputs:\\n@@ -59,6 +64,11 @@ on:\\n default: null\\n required: false\\n type: string\\n+ maxInstanceDuration:\\n+ description: 'If an instance takes longer than the given duration to complete, test will fail.'\\n+ default: '15m'\\n+ required: false\\n+ type: string\\n \\n jobs:\\n e2e:\\n@@ -81,7 +91,7 @@ jobs:\\n {\\n \\\\\\\"maxTestDuration\\\\\\\": \\\\\\\"${{ inputs.maxTestDuration || 'P5D' }}\\\\\\\",\\n \\\\\\\"starter\\\\\\\": [ {\\\\\\\"rate\\\\\\\": 50, \\\\\\\"processId\\\\\\\": \\\\\\\"one-task-one-timer\\\\\\\" } ],\\n- \\\\\\\"verifier\\\\\\\" : { \\\\\\\"maxInstanceDuration\\\\\\\" : \\\\\\\"15m\\\\\\\" },\\n+ \\\\\\\"verifier\\\\\\\" : { \\\\\\\"maxInstanceDuration\\\\\\\" : \\\\\\\"${{ inputs.maxInstanceDuration }}\\\\\\\" },\\n \\\\\\\"fault\\\\\\\": ${{ inputs.fault || 'null' }}\\n }\\n }\\ndiff --git a/.github/workflows/weekly-e2e.yml b/.github/workflows/weekly-e2e.yml\\nindex 93aaeb5..4bd0afd 100644\\n--- a/.github/workflows/weekly-e2e.yml\\n+++ b/.github/workflows/weekly-e2e.yml\\n@@ -31,4 +31,5 @@ jobs:\\n maxTestDuration: P1D\\n clusterPlan: Multiregion test simulation\\n fault: \\\\\\\"2-region-dataloss-failover\\\\\\\"\\n+ maxInstanceDuration: 40m\\n secrets: inherit\\n\", \"diff --git a/tests/playwright/pages/Dashboard/TreeView.ts b/tests/playwright/pages/Dashboard/TreeView.ts\\nindex 9cc622b..75c02c0 100644\\n--- a/tests/playwright/pages/Dashboard/TreeView.ts\\n+++ b/tests/playwright/pages/Dashboard/TreeView.ts\\n@@ -23,10 +23,24 @@ export class TreeViewPage extends BasePage {\\n }\\n \\n async verifyVisibility({ isVisible }: { isVisible: boolean }) {\\n- if (isVisible) {\\n- await expect(this.get()).toBeVisible();\\n+ await this.rootPage.waitForTimeout(1000);\\n+\\n+ const domElement = await this.get();\\n+ // get width of treeview dom element\\n+ const width = (await domElement.boundingBox()).width;\\n+\\n+ // if (isVisible) {\\n+ // await expect(this.get()).toBeVisible();\\n+ // } else {\\n+ // await expect(this.get()).not.toBeVisible();\\n+ // }\\n+\\n+ // border for treeview is 1px\\n+ // if not-visible, width should be < 5;\\n+ if (!isVisible) {\\n+ expect(width).toBeLessThan(5);\\n } else {\\n- await expect(this.get()).not.toBeVisible();\\n+ expect(width).toBeGreaterThan(5);\\n }\\n }\\n \\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"40597fb4de41c7194eb99479a914db70da7909ea\", \"e27e3a6478d59eb0f93af0a51a9c474bad6f8350\", \"ee824ddd71cbc4ccc26f7c6876d379c4927b79e6\", \"efeb30f26252ef4791ef2a02d83827b7f0c45462\"]"},"types":{"kind":"string","value":"[\"feat\", \"build\", \"ci\", \"test\"]"}}},{"rowIdx":923,"cells":{"commit_message":{"kind":"string","value":"remove duplicated variables,#972 External links open in the same tab,enable recovery test\n\nrelated to camunda-tngp/zeebe#353,use new, public `quay.io/influxdb/iox` image"},"diff":{"kind":"string","value":"[\"diff --git a/packages/core/src/components/item/item.ios.scss b/packages/core/src/components/item/item.ios.scss\\nindex 4de5455..6c4d11a 100644\\n--- a/packages/core/src/components/item/item.ios.scss\\n+++ b/packages/core/src/components/item/item.ios.scss\\n@@ -47,15 +47,6 @@ $item-ios-detail-push-color: $list-ios-border-color !default;\\n /// @prop - Icon for the detail arrow\\n $item-ios-detail-push-svg: \\\"\\\" !default;\\n \\n-/// @prop - Background for the divider\\n-$item-ios-divider-background: #f7f7f7 !default;\\n-\\n-/// @prop - Color for the divider\\n-$item-ios-divider-color: #222 !default;\\n-\\n-/// @prop - Padding for the divider\\n-$item-ios-divider-padding: 5px 15px !default;\\n-\\n \\n // iOS Item\\n // --------------------------------------------------\\ndiff --git a/packages/core/src/components/item/item.md.scss b/packages/core/src/components/item/item.md.scss\\nindex 1dd1800..3dadbc0 100644\\n--- a/packages/core/src/components/item/item.md.scss\\n+++ b/packages/core/src/components/item/item.md.scss\\n@@ -35,21 +35,6 @@ $item-md-detail-push-color: $list-md-border-color !default;\\n /// @prop - Icon for the detail arrow\\n $item-md-detail-push-svg: \\\"\\\" !default;\\n \\n-/// @prop - Color for the divider\\n-$item-md-divider-color: #858585 !default;\\n-\\n-/// @prop - Background for the divider\\n-$item-md-divider-background: #fff !default;\\n-\\n-/// @prop - Font size for the divider\\n-$item-md-divider-font-size: $item-md-body-text-font-size !default;\\n-\\n-/// @prop - Border bottom for the divider\\n-$item-md-divider-border-bottom: 1px solid $list-md-border-color !default;\\n-\\n-/// @prop - Padding for the divider\\n-$item-md-divider-padding: 5px 15px !default;\\n-\\n \\n .item-md {\\n @include padding-horizontal($item-md-padding-start, 0);\\ndiff --git a/packages/core/src/components/item/item.wp.scss b/packages/core/src/components/item/item.wp.scss\\nindex 2c4aae6..07b9266 100644\\n--- a/packages/core/src/components/item/item.wp.scss\\n+++ b/packages/core/src/components/item/item.wp.scss\\n@@ -41,21 +41,6 @@ $item-wp-detail-push-color: $input-wp-border-color !default;\\n /// @prop - Icon for the detail arrow\\n $item-wp-detail-push-svg: \\\"\\\" !default;\\n \\n-/// @prop - Color for the divider\\n-$item-wp-divider-color: $list-wp-text-color !default;\\n-\\n-/// @prop - Background for the divider\\n-$item-wp-divider-background: #fff !default;\\n-\\n-/// @prop - Bodrer bottom for the divider\\n-$item-wp-divider-border-bottom: 1px solid $list-wp-border-color !default;\\n-\\n-/// @prop - Font size for the divider\\n-$item-wp-divider-font-size: 2rem !default;\\n-\\n-/// @prop - Padding for the divider\\n-$item-wp-divider-padding: 5px 15px !default;\\n-\\n \\n .item-wp {\\n @include padding-horizontal($item-wp-padding-start, 0);\\n\", \"diff --git a/kofta/src/app/components/Footer.tsx b/kofta/src/app/components/Footer.tsx\\nindex c55fae9..940f7ac 100644\\n--- a/kofta/src/app/components/Footer.tsx\\n+++ b/kofta/src/app/components/Footer.tsx\\n@@ -13,14 +13,14 @@ export const Footer: React.FC = ({ isLogin }) => {\\n return (\\n
\\n {isLogin ? (\\n- \\n+ \\n {t(\\\"footer.link_1\\\")}\\n \\n ) : null}\\n- \\n+ \\n {t(\\\"footer.link_2\\\")}\\n \\n- \\n+ \\n {t(\\\"footer.link_3\\\")}\\n \\n {/* cramps footer on mobile @todo think about how to incorporate this without cramping footer and making the footer really tall */}\\ndiff --git a/kofta/src/app/pages/Login.tsx b/kofta/src/app/pages/Login.tsx\\nindex 3854b5d..1f06220 100644\\n--- a/kofta/src/app/pages/Login.tsx\\n+++ b/kofta/src/app/pages/Login.tsx\\n@@ -46,6 +46,7 @@ export const Login: React.FC = () => {\\n \\n {t(\\\"pages.login.featureText_4\\\")}\\n \\n\", \"diff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\nindex 22b8590..db1b553 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n@@ -116,7 +116,6 @@ public class BrokerRecoveryTest\\n ClockUtil.reset();\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldCreateWorkflowInstanceAfterRestart()\\n {\\n@@ -136,7 +135,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_CREATED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceAtTaskAfterRestart()\\n {\\n@@ -166,7 +164,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceWithLockedTaskAfterRestart()\\n {\\n@@ -200,7 +197,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceAtSecondTaskAfterRestart()\\n {\\n@@ -237,7 +233,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldDeployNewWorkflowVersionAfterRestart()\\n {\\n@@ -412,7 +407,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldResolveIncidentAfterRestart()\\n {\\n@@ -443,7 +437,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"CREATED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldResolveFailedIncidentAfterRestart()\\n {\\n\", \"diff --git a/.circleci/config.yml b/.circleci/config.yml\\nindex 3ae6728..a5f2d2f 100644\\n--- a/.circleci/config.yml\\n+++ b/.circleci/config.yml\\n@@ -12,7 +12,7 @@\\n # The CI for every PR and merge to main runs tests, fmt, lints and compiles debug binaries\\n #\\n # On main if all these checks pass it will then additionally compile in \\\"release\\\" mode and\\n-# publish a docker image to quay.io/influxdb/fusion:$COMMIT_SHA\\n+# publish a docker image to quay.io/influxdb/iox:$COMMIT_SHA\\n #\\n # Manual CI Image:\\n #\\n@@ -317,11 +317,11 @@ jobs:\\n #\\n # Uses the latest ci_image (influxdb/rust below) to build a release binary and\\n # copies it to a minimal container image based upon `rust:slim-buster`. This\\n- # minimal image is then pushed to `quay.io/influxdb/fusion:${BRANCH}` with '/'\\n+ # minimal image is then pushed to `quay.io/influxdb/iox:${BRANCH}` with '/'\\n # repaced by '.' - as an example:\\n #\\n # git branch: dom/my-awesome-feature/perf\\n- # container: quay.io/influxdb/fusion:dom.my-awesome-feature.perf\\n+ # container: quay.io/influxdb/iox:dom.my-awesome-feature.perf\\n #\\n # Subsequent CI runs will overwrite the tag if you push more changes, so watch\\n # out for parallel CI runs!\\n@@ -365,7 +365,7 @@ jobs:\\n sudo apt-get update\\n sudo apt-get install -y docker.io\\n - run: |\\n- echo \\\"$QUAY_PASS\\\" | docker login quay.io --username $QUAY_USER --password-stdin\\n+ echo \\\"$QUAY_INFLUXDB_IOX_PASS\\\" | docker login quay.io --username $QUAY_INFLUXDB_IOX_USER --password-stdin\\n - run:\\n # Docker has functionality to support per-Dockerfile .dockerignore\\n # This was added in https://github.com/moby/buildkit/pull/901\\n@@ -379,8 +379,8 @@ jobs:\\n echo sha256sum after build is\\n sha256sum target/release/influxdb_iox\\n COMMIT_SHA=$(git rev-parse --short HEAD)\\n- docker build -t quay.io/influxdb/fusion:$COMMIT_SHA -f docker/Dockerfile.iox .\\n- docker push quay.io/influxdb/fusion:$COMMIT_SHA\\n+ docker build -t quay.io/influxdb/iox:$COMMIT_SHA -f docker/Dockerfile.iox .\\n+ docker push quay.io/influxdb/iox:$COMMIT_SHA\\n echo \\\"export COMMIT_SHA=${COMMIT_SHA}\\\" >> $BASH_ENV\\n - run:\\n name: Deploy tags\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"cd7e8c3d3549ea05115b3f02586eeba894d86906\", \"07452180fee89e98f05e1aeca68f9923d4c7ab63\", \"f2cc48b74bf92fe22cc265cff4224565f910a921\", \"f751bb5426b87f82096d620f1cd6203badf45d58\"]"},"types":{"kind":"string","value":"[\"refactor\", \"fix\", \"test\", \"ci\"]"}}},{"rowIdx":924,"cells":{"commit_message":{"kind":"string","value":"build updates,Introduce timediff fn (stub),convert to record,add documentation to use react-native-paper with CRA (#874)"},"diff":{"kind":"string","value":"[\"diff --git a/demo/vanilla_new/css/404.min.css b/demo/vanilla_new/css/404.min.css\\nindex a3485b4..e69de29 100644\\n--- a/demo/vanilla_new/css/404.min.css\\n+++ b/demo/vanilla_new/css/404.min.css\\n@@ -1 +0,0 @@\\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\\\"\\\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\\\"Share Tech Mono\\\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\\\"Space Mono\\\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\\\"Space Mono\\\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\\\"\\\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\\\"back to hompage\\\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\\\"report error\\\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\\ndiff --git a/demo/vanilla_new/css/main.min.css b/demo/vanilla_new/css/main.min.css\\nindex 043eb4f..e69de29 100644\\n--- a/demo/vanilla_new/css/main.min.css\\n+++ b/demo/vanilla_new/css/main.min.css\\n@@ -1 +0,0 @@\\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}#floating-container{right:20px;top:20px;position:fixed;z-index:4000}\\ndiff --git a/demo/vanilla_new/js/404.min.js b/demo/vanilla_new/js/404.min.js\\nindex 3642106..e69de29 100644\\n--- a/demo/vanilla_new/js/404.min.js\\n+++ b/demo/vanilla_new/js/404.min.js\\n@@ -1 +0,0 @@\\n-tsParticles.loadJSON(\\\"tsparticles\\\",\\\"/configs/404.json\\\");\\ndiff --git a/website/css/404.min.css b/website/css/404.min.css\\nindex a3485b4..e69de29 100644\\n--- a/website/css/404.min.css\\n+++ b/website/css/404.min.css\\n@@ -1 +0,0 @@\\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\\\"\\\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\\\"Share Tech Mono\\\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\\\"Space Mono\\\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\\\"Space Mono\\\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\\\"\\\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\\\"back to hompage\\\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\\\"report error\\\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\\ndiff --git a/website/css/main.min.css b/website/css/main.min.css\\nindex 818002f..e69de29 100644\\n--- a/website/css/main.min.css\\n+++ b/website/css/main.min.css\\n@@ -1 +0,0 @@\\n-@font-face{font-family:Polya;src:url(https://raw.githubusercontent.com/matteobruni/tsparticles/gh-pages/fonts/Polya.otf)}html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}.github{bottom:10px;right:10px;padding:0 12px 6px;position:fixed;border-radius:10px;background:#fff;border:1px solid #000}.github a{color:#000}.github a:active,.github a:hover,.github a:link,.github a:visited{color:#000;text-decoration:none}.github img{height:30px}.github #gh-project{font-size:20px;padding-left:5px;font-weight:700;vertical-align:bottom}.toggle-sidebar{top:50%;left:0;font-size:20px;color:#000;position:absolute;padding:3px;border-top-right-radius:5px;border-bottom-right-radius:5px;background:#e7e7e7;border:1px solid #000;border-left:none}#editor{background:#fff}[hidden]{display:none}#repulse-div{width:200px;height:200px;background-color:rgba(255,255,255,.5);border-radius:100px;position:absolute;top:50%;left:50%;margin-left:-100px;margin-top:-100px;z-index:200}@media (min-width:1600px) and (-webkit-device-pixel-ratio:1){.col-xxl-3{-ms-flex:0 0 25%;flex:0 0 25%;max-width:25%}}.btn-react{color:#fff;background-color:#61dafb;border-color:#fff}.btn-react:hover{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react.focus,.btn-react:focus{color:#fff;background-color:#5aa3c4;border-color:#ccc;box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}.btn-react.disabled,.btn-react:disabled{color:#fff;background-color:#61dafb;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active,.btn-react:not(:disabled):not(.disabled):active,.show>.btn-react.dropdown-toggle{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active:focus,.btn-react:not(:disabled):not(.disabled):active:focus,.show>.btn-react.dropdown-toggle:focus{box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}#stats,.count-particles{-webkit-user-select:none}#stats{overflow:hidden}#stats-graph canvas{border-radius:3px 3px 0 0}.count-particles{border-radius:0 0 3px 3px}\\ndiff --git a/website/css/presets.min.css b/website/css/presets.min.css\\nindex 6c2ae2c..e69de29 100644\\n--- a/website/css/presets.min.css\\n+++ b/website/css/presets.min.css\\n@@ -1 +0,0 @@\\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden}\\n\", \"diff --git a/rust/Cargo.lock b/rust/Cargo.lock\\nindex b42616f..4795eb6 100644\\n--- a/rust/Cargo.lock\\n+++ b/rust/Cargo.lock\\n@@ -1287,7 +1287,7 @@ dependencies = [\\n [[package]]\\n name = \\\"datafusion\\\"\\n version = \\\"5.1.0\\\"\\n-source = \\\"git+https://github.com/cube-js/arrow-datafusion.git?rev=8df4132b83d896a0d3db5c82a4eaaa3eaa285d15#8df4132b83d896a0d3db5c82a4eaaa3eaa285d15\\\"\\n+source = \\\"git+https://github.com/cube-js/arrow-datafusion.git?rev=868f3c4de13d13cda84cee33475b9782b94fa60c#868f3c4de13d13cda84cee33475b9782b94fa60c\\\"\\n dependencies = [\\n \\\"ahash 0.7.4\\\",\\n \\\"arrow 6.0.0\\\",\\ndiff --git a/rust/cubesql/Cargo.toml b/rust/cubesql/Cargo.toml\\nindex 3cb386a..9aef494 100644\\n--- a/rust/cubesql/Cargo.toml\\n+++ b/rust/cubesql/Cargo.toml\\n@@ -9,7 +9,7 @@ documentation = \\\"https://cube.dev/docs\\\"\\n homepage = \\\"https://cube.dev\\\"\\n \\n [dependencies]\\n-datafusion = { git = 'https://github.com/cube-js/arrow-datafusion.git', rev = \\\"8df4132b83d896a0d3db5c82a4eaaa3eaa285d15\\\", default-features = false, features = [\\\"unicode_expressions\\\"] }\\n+datafusion = { git = 'https://github.com/cube-js/arrow-datafusion.git', rev = \\\"868f3c4de13d13cda84cee33475b9782b94fa60c\\\", default-features = false, features = [\\\"unicode_expressions\\\"] }\\n anyhow = \\\"1.0\\\"\\n thiserror = \\\"1.0\\\"\\n cubeclient = { path = \\\"../cubeclient\\\" }\\ndiff --git a/rust/cubesql/src/compile/engine/df/intervals.rs b/rust/cubesql/src/compile/engine/df/intervals.rs\\nnew file mode 100644\\nindex 0000000..9e6cb7e\\n--- /dev/null\\n+++ b/rust/cubesql/src/compile/engine/df/intervals.rs\\n@@ -0,0 +1,51 @@\\n+#[macro_export]\\n+macro_rules! make_string_interval_year_month {\\n+ ($array: ident, $row: ident) => {{\\n+ let s = if $array.is_null($row) {\\n+ \\\"NULL\\\".to_string()\\n+ } else {\\n+ let interval = $array.value($row) as f64;\\n+ let years = (interval / 12_f64).floor();\\n+ let month = interval - (years * 12_f64);\\n+\\n+ format!(\\n+ \\\"{} years {} mons 0 days 0 hours 0 mins 0.00 secs\\\",\\n+ years, month,\\n+ )\\n+ };\\n+\\n+ s\\n+ }};\\n+}\\n+\\n+#[macro_export]\\n+macro_rules! make_string_interval_day_time {\\n+ ($array: ident, $row: ident) => {{\\n+ let s = if $array.is_null($row) {\\n+ \\\"NULL\\\".to_string()\\n+ } else {\\n+ let value: u64 = $array.value($row) as u64;\\n+\\n+ let days_parts: i32 = ((value & 0xFFFFFFFF00000000) >> 32) as i32;\\n+ let milliseconds_part: i32 = (value & 0xFFFFFFFF) as i32;\\n+\\n+ let secs = milliseconds_part / 1000;\\n+ let mins = secs / 60;\\n+ let hours = mins / 60;\\n+\\n+ let secs = secs - (mins * 60);\\n+ let mins = mins - (hours * 60);\\n+\\n+ format!(\\n+ \\\"0 years 0 mons {} days {} hours {} mins {}.{:02} secs\\\",\\n+ days_parts,\\n+ hours,\\n+ mins,\\n+ secs,\\n+ (milliseconds_part % 1000),\\n+ )\\n+ };\\n+\\n+ s\\n+ }};\\n+}\\ndiff --git a/rust/cubesql/src/compile/engine/df/mod.rs b/rust/cubesql/src/compile/engine/df/mod.rs\\nindex a19a970..3097523 100644\\n--- a/rust/cubesql/src/compile/engine/df/mod.rs\\n+++ b/rust/cubesql/src/compile/engine/df/mod.rs\\n@@ -1 +1,2 @@\\n pub mod coerce;\\n+pub mod intervals;\\ndiff --git a/rust/cubesql/src/compile/engine/udf.rs b/rust/cubesql/src/compile/engine/udf.rs\\nindex 55b8bc1..0e160b3 100644\\n--- a/rust/cubesql/src/compile/engine/udf.rs\\n+++ b/rust/cubesql/src/compile/engine/udf.rs\\n@@ -1,14 +1,19 @@\\n use std::any::type_name;\\n use std::sync::Arc;\\n \\n+\\n use datafusion::{\\n arrow::{\\n array::{\\n ArrayRef, BooleanArray, BooleanBuilder, GenericStringArray, Int32Builder,\\n- PrimitiveArray, StringBuilder, UInt32Builder,\\n+ IntervalDayTimeBuilder, PrimitiveArray, StringBuilder,\\n+ UInt32Builder,\\n },\\n compute::cast,\\n- datatypes::{DataType, Int64Type},\\n+ datatypes::{\\n+ DataType, Int64Type, IntervalUnit, TimeUnit,\\n+ TimestampNanosecondType,\\n+ },\\n },\\n error::DataFusionError,\\n logical_plan::create_udf,\\n@@ -399,3 +404,63 @@ pub fn create_convert_tz_udf() -> ScalarUDF {\\n &fun,\\n )\\n }\\n+\\n+pub fn create_timediff_udf() -> ScalarUDF {\\n+ let fun = make_scalar_function(move |args: &[ArrayRef]| {\\n+ assert!(args.len() == 2);\\n+\\n+ let left_dt = &args[0];\\n+ let right_dt = &args[1];\\n+\\n+ let left_date = match left_dt.data_type() {\\n+ DataType::Timestamp(TimeUnit::Nanosecond, _) => {\\n+ let arr = downcast_primitive_arg!(left_dt, \\\"left_dt\\\", TimestampNanosecondType);\\n+ let ts = arr.value(0);\\n+\\n+ // NaiveDateTime::from_timestamp(ts, 0)\\n+ ts\\n+ }\\n+ _ => {\\n+ return Err(DataFusionError::Execution(format!(\\n+ \\\"left_dt argument must be a Timestamp, actual: {}\\\",\\n+ left_dt.data_type()\\n+ )));\\n+ }\\n+ };\\n+\\n+ let right_date = match right_dt.data_type() {\\n+ DataType::Timestamp(TimeUnit::Nanosecond, _) => {\\n+ let arr = downcast_primitive_arg!(right_dt, \\\"right_dt\\\", TimestampNanosecondType);\\n+ arr.value(0)\\n+ }\\n+ _ => {\\n+ return Err(DataFusionError::Execution(format!(\\n+ \\\"right_dt argument must be a Timestamp, actual: {}\\\",\\n+ right_dt.data_type()\\n+ )));\\n+ }\\n+ };\\n+\\n+ let diff = right_date - left_date;\\n+ if diff != 0 {\\n+ return Err(DataFusionError::NotImplemented(format!(\\n+ \\\"timediff is not implemented, it's stub\\\"\\n+ )));\\n+ }\\n+\\n+ let mut interal_arr = IntervalDayTimeBuilder::new(1);\\n+ interal_arr.append_value(diff)?;\\n+\\n+ Ok(Arc::new(interal_arr.finish()) as ArrayRef)\\n+ });\\n+\\n+ let return_type: ReturnTypeFunction =\\n+ Arc::new(move |_| Ok(Arc::new(DataType::Interval(IntervalUnit::DayTime))));\\n+\\n+ ScalarUDF::new(\\n+ \\\"timediff\\\",\\n+ &Signature::any(2, Volatility::Immutable),\\n+ &return_type,\\n+ &fun,\\n+ )\\n+}\\ndiff --git a/rust/cubesql/src/compile/mod.rs b/rust/cubesql/src/compile/mod.rs\\nindex a88da57..6121aa0 100644\\n--- a/rust/cubesql/src/compile/mod.rs\\n+++ b/rust/cubesql/src/compile/mod.rs\\n@@ -32,8 +32,8 @@ use self::engine::context::SystemVar;\\n use self::engine::provider::CubeContext;\\n use self::engine::udf::{\\n create_connection_id_udf, create_convert_tz_udf, create_current_user_udf, create_db_udf,\\n- create_if_udf, create_instr_udf, create_isnull_udf, create_least_udf, create_user_udf,\\n- create_version_udf,\\n+ create_if_udf, create_instr_udf, create_isnull_udf, create_least_udf, create_timediff_udf,\\n+ create_user_udf, create_version_udf,\\n };\\n use self::parser::parse_sql_to_statement;\\n \\n@@ -1450,6 +1450,7 @@ impl QueryPlanner {\\n ctx.register_udf(create_if_udf());\\n ctx.register_udf(create_least_udf());\\n ctx.register_udf(create_convert_tz_udf());\\n+ ctx.register_udf(create_timediff_udf());\\n \\n let state = ctx.state.lock().unwrap().clone();\\n let cube_ctx = CubeContext::new(&state, &self.context.cubes);\\n@@ -3226,6 +3227,25 @@ mod tests {\\n }\\n \\n #[tokio::test]\\n+ async fn test_timediff() -> Result<(), CubeError> {\\n+ assert_eq!(\\n+ execute_df_query(\\n+ \\\"select \\\\\\n+ timediff('1994-11-26T13:25:00.000Z'::timestamp, '1994-11-26T13:25:00.000Z'::timestamp) as r1\\n+ \\\".to_string()\\n+ )\\n+ .await?,\\n+ \\\"+------------------------------------------------+\\\\n\\\\\\n+ | r1 |\\\\n\\\\\\n+ +------------------------------------------------+\\\\n\\\\\\n+ | 0 years 0 mons 0 days 0 hours 0 mins 0.00 secs |\\\\n\\\\\\n+ +------------------------------------------------+\\\"\\n+ );\\n+\\n+ Ok(())\\n+ }\\n+\\n+ #[tokio::test]\\n async fn test_metabase() -> Result<(), CubeError> {\\n assert_eq!(\\n execute_df_query(\\ndiff --git a/rust/cubesql/src/mysql/dataframe.rs b/rust/cubesql/src/mysql/dataframe.rs\\nindex fa246aa..2443458 100644\\n--- a/rust/cubesql/src/mysql/dataframe.rs\\n+++ b/rust/cubesql/src/mysql/dataframe.rs\\n@@ -3,9 +3,10 @@ use std::fmt::{self, Debug, Formatter};\\n use chrono::{SecondsFormat, TimeZone, Utc};\\n use comfy_table::{Cell, Table};\\n use datafusion::arrow::array::{\\n- Array, Float64Array, Int32Array, Int64Array, StringArray, TimestampMicrosecondArray,\\n- UInt32Array,\\n+ Array, Float64Array, Int32Array, Int64Array, IntervalDayTimeArray, IntervalYearMonthArray,\\n+ StringArray, TimestampMicrosecondArray, UInt32Array,\\n };\\n+use datafusion::arrow::datatypes::IntervalUnit;\\n use datafusion::arrow::{\\n array::{BooleanArray, TimestampNanosecondArray, UInt64Array},\\n datatypes::{DataType, TimeUnit},\\n@@ -15,6 +16,7 @@ use log::{error, warn};\\n use msql_srv::{ColumnFlags, ColumnType};\\n \\n use crate::{compile::builder::CompiledQueryFieldMeta, CubeError};\\n+use crate::{make_string_interval_day_time, make_string_interval_year_month};\\n \\n #[derive(Clone, Debug)]\\n pub struct Column {\\n@@ -309,6 +311,7 @@ pub fn arrow_to_column_type(arrow_type: DataType) -> Result Ok(ColumnType::MYSQL_TYPE_BLOB),\\n DataType::Utf8 | DataType::LargeUtf8 => Ok(ColumnType::MYSQL_TYPE_STRING),\\n DataType::Timestamp(_, _) => Ok(ColumnType::MYSQL_TYPE_STRING),\\n+ DataType::Interval(_) => Ok(ColumnType::MYSQL_TYPE_STRING),\\n DataType::Float16 | DataType::Float64 => Ok(ColumnType::MYSQL_TYPE_DOUBLE),\\n DataType::Boolean => Ok(ColumnType::MYSQL_TYPE_TINY),\\n DataType::Int8\\n@@ -402,6 +405,24 @@ pub fn batch_to_dataframe(batches: &Vec) -> Result {\\n+ let a = array\\n+ .as_any()\\n+ .downcast_ref::()\\n+ .unwrap();\\n+ for i in 0..num_rows {\\n+ rows[i].push(TableValue::String(make_string_interval_day_time!(a, i)));\\n+ }\\n+ }\\n+ DataType::Interval(IntervalUnit::YearMonth) => {\\n+ let a = array\\n+ .as_any()\\n+ .downcast_ref::()\\n+ .unwrap();\\n+ for i in 0..num_rows {\\n+ rows[i].push(TableValue::String(make_string_interval_year_month!(a, i)));\\n+ }\\n+ }\\n DataType::Boolean => {\\n let a = array.as_any().downcast_ref::().unwrap();\\n for i in 0..num_rows {\\n\", \"diff --git a/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java b/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\\nindex cc998c6..65c8550 100755\\n--- a/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\\n+++ b/broker/src/test/java/io/camunda/zeebe/broker/exporter/stream/ExporterDirectorDistributionTest.java\\n@@ -167,13 +167,8 @@ public final class ExporterDirectorDistributionTest {\\n *

This makes sure that even if we miss one export position event, we distribute the event\\n * later again, which makes tests less flaky.\\n */\\n- private static final class ClockShifter implements ConditionEvaluationListener {\\n-\\n- private final ControlledActorClock clock;\\n-\\n- public ClockShifter(final ControlledActorClock clock) {\\n- this.clock = clock;\\n- }\\n+ private record ClockShifter(ControlledActorClock clock)\\n+ implements ConditionEvaluationListener {\\n \\n @Override\\n public void conditionEvaluated(final EvaluatedCondition condition) {\\n\", \"diff --git a/docs/pages/4.react-native-web.md b/docs/pages/4.react-native-web.md\\nindex 69e4e52..8d6ae2a 100644\\n--- a/docs/pages/4.react-native-web.md\\n+++ b/docs/pages/4.react-native-web.md\\n@@ -16,6 +16,63 @@ To install `react-native-web`, run:\\n yarn add react-native-web react-dom react-art\\n ```\\n \\n+### Using CRA ([Create React App](https://github.com/facebook/create-react-app))\\n+\\n+Install [`react-app-rewired`](https://github.com/timarney/react-app-rewired) to override `webpack` configuration:\\n+\\n+```sh\\n+yarn add --dev react-app-rewired\\n+```\\n+\\n+[Configure `babel-loader`](#2-configure-babel-loader) using a new file called `config-overrides.js`:\\n+\\n+```js\\n+module.exports = function override(config, env) {\\n+ config.module.rules.push({\\n+ test: /\\\\.js$/,\\n+ exclude: /node_modules[/\\\\\\\\](?!react-native-paper|react-native-vector-icons|react-native-safe-area-view)/,\\n+ use: {\\n+ loader: \\\"babel-loader\\\",\\n+ options: {\\n+ // Disable reading babel configuration\\n+ babelrc: false,\\n+ configFile: false,\\n+\\n+ // The configration for compilation\\n+ presets: [\\n+ [\\\"@babel/preset-env\\\", { useBuiltIns: \\\"usage\\\" }],\\n+ \\\"@babel/preset-react\\\",\\n+ \\\"@babel/preset-flow\\\"\\n+ ],\\n+ plugins: [\\n+ \\\"@babel/plugin-proposal-class-properties\\\",\\n+ \\\"@babel/plugin-proposal-object-rest-spread\\\"\\n+ ]\\n+ }\\n+ }\\n+ });\\n+\\n+ return config;\\n+};\\n+```\\n+\\n+Change your script in `package.json`:\\n+\\n+```diff\\n+/* package.json */\\n+\\n+ \\\"scripts\\\": {\\n+- \\\"start\\\": \\\"react-scripts start\\\",\\n++ \\\"start\\\": \\\"react-app-rewired start\\\",\\n+- \\\"build\\\": \\\"react-scripts build\\\",\\n++ \\\"build\\\": \\\"react-app-rewired build\\\",\\n+- \\\"test\\\": \\\"react-scripts test --env=jsdom\\\",\\n++ \\\"test\\\": \\\"react-app-rewired test --env=jsdom\\\"\\n+}\\n+```\\n+\\n+### Custom webpack setup\\n+\\n To install `webpack`, run:\\n \\n ```sh\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"9acf7a062ee9c0538c2cd4661c1f5da61ab06316\", \"29dfb9716298c5a579c0ffba6742e13a29325670\", \"3346331a963766c8193170fb130adad2e658ada2\", \"ee7cc5d5a940fba774e715b1f029c6361110b108\"]"},"types":{"kind":"string","value":"[\"build\", \"feat\", \"refactor\", \"docs\"]"}}},{"rowIdx":925,"cells":{"commit_message":{"kind":"string","value":"group example,alerts do not trigger modal lifecycle events\n\nfixes #8616,updated riot to v6, fixed build,do not check mkdocs for older versions used in deployments"},"diff":{"kind":"string","value":"[\"diff --git a/src/build/arg_group.rs b/src/build/arg_group.rs\\nindex 5201e97..e1b1991 100644\\n--- a/src/build/arg_group.rs\\n+++ b/src/build/arg_group.rs\\n@@ -43,7 +43,7 @@ use crate::util::{Id, Key};\\n /// .arg(\\\"--minor 'auto increase minor'\\\")\\n /// .arg(\\\"--patch 'auto increase patch'\\\")\\n /// .group(ArgGroup::with_name(\\\"vers\\\")\\n-/// .args(&[\\\"set-ver\\\", \\\"major\\\", \\\"minor\\\",\\\"patch\\\"])\\n+/// .args(&[\\\"set-ver\\\", \\\"major\\\", \\\"minor\\\", \\\"patch\\\"])\\n /// .required(true))\\n /// .try_get_matches_from(vec![\\\"app\\\", \\\"--major\\\", \\\"--patch\\\"]);\\n /// // Because we used two args in the group it's an error\\n\", \"diff --git a/src/components/app/app-root.ts b/src/components/app/app-root.ts\\nindex ec7daee..29dc797 100644\\n--- a/src/components/app/app-root.ts\\n+++ b/src/components/app/app-root.ts\\n@@ -15,6 +15,7 @@ export const AppRootToken = new OpaqueToken('USERROOT');\\n selector: 'ion-app',\\n template:\\n '

' +\\n+ '
' +\\n '
' +\\n '
' +\\n '
' +\\n@@ -24,6 +25,8 @@ export class IonicApp extends Ion implements OnInit {\\n \\n @ViewChild('viewport', {read: ViewContainerRef}) _viewport: ViewContainerRef;\\n \\n+ @ViewChild('modalPortal', { read: OverlayPortal }) _modalPortal: OverlayPortal;\\n+\\n @ViewChild('overlayPortal', { read: OverlayPortal }) _overlayPortal: OverlayPortal;\\n \\n @ViewChild('loadingPortal', { read: OverlayPortal }) _loadingPortal: OverlayPortal;\\n@@ -96,6 +99,9 @@ export class IonicApp extends Ion implements OnInit {\\n if (portal === AppPortal.TOAST) {\\n return this._toastPortal;\\n }\\n+ if (portal === AppPortal.MODAL) {\\n+ return this._modalPortal;\\n+ }\\n return this._overlayPortal;\\n }\\n \\n@@ -110,6 +116,7 @@ export class IonicApp extends Ion implements OnInit {\\n \\n export enum AppPortal {\\n DEFAULT,\\n+ MODAL,\\n LOADING,\\n TOAST\\n };\\ndiff --git a/src/components/modal/modal.ts b/src/components/modal/modal.ts\\nindex bd4d406..c3e7a62 100644\\n--- a/src/components/modal/modal.ts\\n+++ b/src/components/modal/modal.ts\\n@@ -1,6 +1,7 @@\\n import { Injectable } from '@angular/core';\\n \\n import { App } from '../app/app';\\n+import { AppPortal } from '../app/app-root';\\n import { isPresent } from '../../util/util';\\n import { ModalCmp } from './modal-component';\\n import { ModalOptions } from './modal-options';\\n@@ -40,7 +41,7 @@ export class Modal extends ViewController {\\n * @returns {Promise} Returns a promise which is resolved when the transition has completed.\\n */\\n present(navOptions: NavOptions = {}) {\\n- return this._app.present(this, navOptions);\\n+ return this._app.present(this, navOptions, AppPortal.MODAL);\\n }\\n \\n /**\\n\", \"diff --git a/components/riot/package.json b/components/riot/package.json\\nindex c41743a..eb69756 100644\\n--- a/components/riot/package.json\\n+++ b/components/riot/package.json\\n@@ -61,7 +61,7 @@\\n },\\n \\\"devDependencies\\\": {\\n \\\"@babel/preset-typescript\\\": \\\"^7.14.5\\\",\\n- \\\"@riotjs/cli\\\": \\\"^6.0.4\\\",\\n+ \\\"@riotjs/cli\\\": \\\"^6.0.5\\\",\\n \\\"@riotjs/compiler\\\": \\\"^6.0.0\\\",\\n \\\"chai\\\": \\\"^4.3.4\\\",\\n \\\"esm\\\": \\\"^3.2.25\\\",\\n\", \"diff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\\nindex 57d94a4..04de03b 100644\\n--- a/.github/workflows/ibis-docs-lint.yml\\n+++ b/.github/workflows/ibis-docs-lint.yml\\n@@ -206,7 +206,7 @@ jobs:\\n - name: build and push dev docs\\n run: |\\n nix develop --ignore-environment -c \\\\\\n- mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}'\\n+ mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}' --ignore-version\\n \\n simulate_release:\\n runs-on: ubuntu-latest\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"9849430b11b92ae58d94cfe4d0b06313c7eab550\", \"e2704a4a25b9e348764e1cc922ca7d6a927550eb\", \"5d256f937f93e5a5ed003df86d38c44834095a11\", \"21228c55b7045d9b2225f65e6231184ff332b071\"]"},"types":{"kind":"string","value":"[\"docs\", \"fix\", \"build\", \"ci\"]"}}},{"rowIdx":926,"cells":{"commit_message":{"kind":"string","value":"use new freespace config for disk space recory test,i18n for Time Picker,use an action for issue assignment,expose the means by which we process each reward cycle's affirmation maps at reward cycle boundaries"},"diff":{"kind":"string","value":"[\"diff --git a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java\\nindex 0854323..bfc7b7e 100644\\n--- a/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java\\n+++ b/qa/integration-tests/src/test/java/io/camunda/zeebe/it/health/DiskSpaceRecoveryIT.java\\n@@ -47,7 +47,8 @@ final class DiskSpaceRecoveryIT {\\n .withZeebeData(volume)\\n .withEnv(\\\"ZEEBE_BROKER_DATA_LOGSEGMENTSIZE\\\", \\\"1MB\\\")\\n .withEnv(\\\"ZEEBE_BROKER_NETWORK_MAXMESSAGESIZE\\\", \\\"1MB\\\")\\n- .withEnv(\\\"ZEEBE_BROKER_DATA_DISKUSAGECOMMANDWATERMARK\\\", \\\"0.5\\\");\\n+ .withEnv(\\\"ZEEBE_BROKER_DATA_DISK_FREESPACE_PROCESSING\\\", \\\"10MB\\\")\\n+ .withEnv(\\\"ZEEBE_BROKER_DATA_DISK_FREESPACE_REPLICATION\\\", \\\"1MB\\\");\\n \\n private ZeebeClient client;\\n \\n@@ -127,7 +128,9 @@ final class DiskSpaceRecoveryIT {\\n ContainerEngine.builder()\\n .withDebugReceiverPort(SocketUtil.getNextAddress().getPort())\\n .withContainer(\\n- container.withEnv(\\\"ZEEBE_BROKER_DATA_DISKUSAGECOMMANDWATERMARK\\\", \\\"0.0001\\\"))\\n+ container\\n+ .withEnv(\\\"ZEEBE_BROKER_DATA_DISK_FREESPACE_PROCESSING\\\", \\\"16MB\\\")\\n+ .withEnv(\\\"ZEEBE_BROKER_DATA_DISK_FREESPACE_REPLICATION\\\", \\\"10MB\\\"))\\n .build();\\n \\n @BeforeEach\\n\", \"diff --git a/packages/nc-gui/components/cell/TimePicker.vue b/packages/nc-gui/components/cell/TimePicker.vue\\nindex 619ab45..7f66828 100644\\n--- a/packages/nc-gui/components/cell/TimePicker.vue\\n+++ b/packages/nc-gui/components/cell/TimePicker.vue\\n@@ -38,6 +38,8 @@ const isTimeInvalid = ref(false)\\n \\n const dateFormat = isMysql(column.value.base_id) ? 'YYYY-MM-DD HH:mm:ss' : 'YYYY-MM-DD HH:mm:ssZ'\\n \\n+const { t } = useI18n()\\n+\\n const localState = computed({\\n get() {\\n if (!modelValue) {\\n@@ -89,11 +91,11 @@ watch(\\n \\n const placeholder = computed(() => {\\n if (isEditColumn.value && (modelValue === '' || modelValue === null)) {\\n- return '(Optional)'\\n+ return t('labels.optional')\\n } else if (modelValue === null && showNull.value) {\\n- return 'NULL'\\n+ return t('general.null')\\n } else if (isTimeInvalid.value) {\\n- return 'Invalid time'\\n+ return t('msg.invalidTime')\\n } else {\\n return ''\\n }\\n\", \"diff --git a/.github/workflows/assign.yml b/.github/workflows/assign.yml\\nindex 29d92a8..758874e 100644\\n--- a/.github/workflows/assign.yml\\n+++ b/.github/workflows/assign.yml\\n@@ -8,8 +8,6 @@ jobs:\\n runs-on: ubuntu-latest\\n if: ${{ github.event.comment.body == '/take' }}\\n steps:\\n- - uses: actions/checkout@v2\\n- - name: Assign issue ${{ github.event.issue.number }} to ${{ github.event.comment.user.login }}\\n- run: gh issue edit ${{ github.event.issue.number }} --add-assignee \\\"${{ github.event.comment.user.login }}\\\"\\n- env:\\n- GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}\\n+ - uses: pozil/auto-assign-issue@v1.1.0\\n+ with:\\n+ assignees: ${{ github.event.comment.user.login }}\\n\", \"diff --git a/src/burnchains/burnchain.rs b/src/burnchains/burnchain.rs\\nindex 92105d6..60c608a 100644\\n--- a/src/burnchains/burnchain.rs\\n+++ b/src/burnchains/burnchain.rs\\n@@ -851,8 +851,26 @@ impl Burnchain {\\n );\\n \\n burnchain_db.store_new_burnchain_block(burnchain, indexer, &block)?;\\n- let block_height = block.block_height();\\n+ Burnchain::process_affirmation_maps(\\n+ burnchain,\\n+ burnchain_db,\\n+ indexer,\\n+ block.block_height(),\\n+ )?;\\n+\\n+ let header = block.header();\\n+ Ok(header)\\n+ }\\n \\n+ /// Update the affirmation maps for the previous reward cycle's commits.\\n+ /// This is a no-op unless the given burnchain block height falls on a reward cycle boundary. In that\\n+ /// case, the previous reward cycle's block commits' affirmation maps are all re-calculated.\\n+ pub fn process_affirmation_maps(\\n+ burnchain: &Burnchain,\\n+ burnchain_db: &mut BurnchainDB,\\n+ indexer: &B,\\n+ block_height: u64,\\n+ ) -> Result<(), burnchain_error> {\\n let this_reward_cycle = burnchain\\n .block_height_to_reward_cycle(block_height)\\n .unwrap_or(0);\\n@@ -872,10 +890,7 @@ impl Burnchain {\\n );\\n update_pox_affirmation_maps(burnchain_db, indexer, prev_reward_cycle, burnchain)?;\\n }\\n-\\n- let header = block.header();\\n-\\n- Ok(header)\\n+ Ok(())\\n }\\n \\n /// Hand off the block to the ChainsCoordinator _and_ process the sortition\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"672cd2b9775fb6dac2d522cb3f4469db47c0556b\", \"48806e3675c7b18327e7629827454d7c29be25a9\", \"fb3a231b29bc8bff9270b99dd4aff9dad599f21f\", \"d7972da833257c073403dec3c2ac3a7f297e328a\"]"},"types":{"kind":"string","value":"[\"test\", \"fix\", \"ci\", \"refactor\"]"}}},{"rowIdx":927,"cells":{"commit_message":{"kind":"string","value":"enable recovery test\n\nrelated to camunda-tngp/zeebe#353,ensure \"dist\" dirs exist,added vue3 readme,avoid cancelling jobs"},"diff":{"kind":"string","value":"[\"diff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\nindex 22b8590..db1b553 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n@@ -116,7 +116,6 @@ public class BrokerRecoveryTest\\n ClockUtil.reset();\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldCreateWorkflowInstanceAfterRestart()\\n {\\n@@ -136,7 +135,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_CREATED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceAtTaskAfterRestart()\\n {\\n@@ -166,7 +164,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceWithLockedTaskAfterRestart()\\n {\\n@@ -200,7 +197,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceAtSecondTaskAfterRestart()\\n {\\n@@ -237,7 +233,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldDeployNewWorkflowVersionAfterRestart()\\n {\\n@@ -412,7 +407,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldResolveIncidentAfterRestart()\\n {\\n@@ -443,7 +437,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"CREATED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldResolveFailedIncidentAfterRestart()\\n {\\n\", \"diff --git a/scripts/prepare.js b/scripts/prepare.js\\nindex 9eb8cb8..f285825 100644\\n--- a/scripts/prepare.js\\n+++ b/scripts/prepare.js\\n@@ -68,6 +68,9 @@ async function prepare() {\\n names.push(json.name)\\n }\\n \\n+ // Ensure all \\\"dist\\\" directories exist.\\n+ dirs.forEach(dir => fs.ensureDirSync(join(dir, distId)))\\n+\\n log(``)\\n for (let i = 0; i < names.length; i++) {\\n const dir = dirs[i]\\n\", \"diff --git a/core/main/README.md b/core/main/README.md\\nindex e5e4c93..e9cfda9 100644\\n--- a/core/main/README.md\\n+++ b/core/main/README.md\\n@@ -217,7 +217,7 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\\n \\n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/svelte/README.md)\\n \\n-### VueJS\\n+### VueJS 2.x\\n \\n #### `particles.vue`\\n \\n@@ -225,6 +225,14 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\\n \\n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue/README.md)\\n \\n+### VueJS 3.x\\n+\\n+#### `particles.vue3`\\n+\\n+[![npm](https://img.shields.io/npm/v/particles.vue3)](https://www.npmjs.com/package/particles.vue3) [![npm](https://img.shields.io/npm/dm/particles.vue3)](https://www.npmjs.com/package/particles.vue3)\\n+\\n+You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue3/README.md)\\n+\\n ---\\n \\n ## **_Demo / Generator_**\\ndiff --git a/core/main/tsconfig.json b/core/main/tsconfig.json\\nindex 7916bc5..72399c0 100644\\n--- a/core/main/tsconfig.json\\n+++ b/core/main/tsconfig.json\\n@@ -107,10 +107,14 @@\\n \\\"source\\\": \\\"../../components/react/README.md\\\"\\n },\\n {\\n- \\\"title\\\": \\\"Vue\\\",\\n+ \\\"title\\\": \\\"Vue 2.x\\\",\\n \\\"source\\\": \\\"../../components/vue/README.md\\\"\\n },\\n {\\n+ \\\"title\\\": \\\"Vue 3.x\\\",\\n+ \\\"source\\\": \\\"../../components/vue3/README.md\\\"\\n+ },\\n+ {\\n \\\"title\\\": \\\"Svelte\\\",\\n \\\"source\\\": \\\"../../components/svelte/README.md\\\"\\n },\\n\", \"diff --git a/.github/workflows/ibis-backends-cloud.yml b/.github/workflows/ibis-backends-cloud.yml\\nindex 321708e..b990984 100644\\n--- a/.github/workflows/ibis-backends-cloud.yml\\n+++ b/.github/workflows/ibis-backends-cloud.yml\\n@@ -29,7 +29,9 @@ jobs:\\n name: ${{ matrix.backend.title }} python-${{ matrix.python-version }}\\n # only a single bigquery or snowflake run at a time, otherwise test data is\\n # clobbered by concurrent runs\\n- concurrency: ${{ matrix.backend.name }}\\n+ concurrency:\\n+ group: ${{ matrix.backend.name }}\\n+ cancel-in-progress: false\\n runs-on: ubuntu-latest\\n strategy:\\n fail-fast: false\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"f2cc48b74bf92fe22cc265cff4224565f910a921\", \"ca060bf255a55b99000ddf0c67f7422f28b735a6\", \"e4c3e2cff769ce46d22d5c8f7dd527510443a8a7\", \"19514bc68624a964c63fc217f163f7b11f3dfe82\"]"},"types":{"kind":"string","value":"[\"test\", \"build\", \"docs\", \"ci\"]"}}},{"rowIdx":928,"cells":{"commit_message":{"kind":"string","value":"abort parallel stages if one failed,update sandbox-option.md (#18275)\n\nCo-Authored-By: Mark Lee ,serialize access to StreamObserver,bump version\n\nSigned-off-by: rjshrjndrn "},"diff":{"kind":"string","value":"[\"diff --git a/Jenkinsfile b/Jenkinsfile\\nindex 168f446..a4da961 100644\\n--- a/Jenkinsfile\\n+++ b/Jenkinsfile\\n@@ -28,6 +28,7 @@ pipeline {\\n }\\n \\n stage('Verify') {\\n+ failFast true\\n parallel {\\n stage('Tests') {\\n steps {\\n\", \"diff --git a/docs/api/sandbox-option.md b/docs/api/sandbox-option.md\\nindex 7d24bee..e293d34 100644\\n--- a/docs/api/sandbox-option.md\\n+++ b/docs/api/sandbox-option.md\\n@@ -113,8 +113,8 @@ window.open = customWindowOpen\\n Important things to notice in the preload script:\\n \\n - Even though the sandboxed renderer doesn't have Node.js running, it still has\\n- access to a limited node-like environment: `Buffer`, `process`, `setImmediate`\\n- and `require` are available.\\n+ access to a limited node-like environment: `Buffer`, `process`, `setImmediate`,\\n+ `clearImmediate` and `require` are available.\\n - The preload script can indirectly access all APIs from the main process through the\\n `remote` and `ipcRenderer` modules.\\n - The preload script must be contained in a single script, but it is possible to have\\n@@ -162,16 +162,17 @@ feature. We are still not aware of the security implications of exposing some\\n Electron renderer APIs to the preload script, but here are some things to\\n consider before rendering untrusted content:\\n \\n-- A preload script can accidentally leak privileged APIs to untrusted code.\\n+- A preload script can accidentally leak privileged APIs to untrusted code,\\n+ unless [`contextIsolation`](../tutorial/security.md#3-enable-context-isolation-for-remote-content)\\n+ is also enabled.\\n - Some bug in V8 engine may allow malicious code to access the renderer preload\\n APIs, effectively granting full access to the system through the `remote`\\n- module.\\n+ module. Therefore, it is highly recommended to\\n+ [disable the `remote` module](../tutorial/security.md#15-disable-the-remote-module).\\n+ If disabling is not feasible, you should selectively\\n+ [filter the `remote` module](../tutorial/security.md#16-filter-the-remote-module).\\n \\n Since rendering untrusted content in Electron is still uncharted territory,\\n the APIs exposed to the sandbox preload script should be considered more\\n unstable than the rest of Electron APIs, and may have breaking changes to fix\\n security issues.\\n-\\n-One planned enhancement that should greatly increase security is to block IPC\\n-messages from sandboxed renderers by default, allowing the main process to\\n-explicitly define a set of messages the renderer is allowed to send.\\n\", \"diff --git a/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java b/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java\\nindex ae2b1c0..8ed64e5 100644\\n--- a/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java\\n+++ b/gateway/src/main/java/io/camunda/zeebe/gateway/impl/stream/ClientStreamAdapter.java\\n@@ -22,6 +22,7 @@ import io.camunda.zeebe.transport.stream.api.ClientStreamer;\\n import io.camunda.zeebe.util.VisibleForTesting;\\n import io.grpc.Status;\\n import io.grpc.StatusRuntimeException;\\n+import io.grpc.internal.SerializingExecutor;\\n import io.grpc.stub.ServerCallStreamObserver;\\n import io.grpc.stub.StreamObserver;\\n import java.util.concurrent.CompletableFuture;\\n@@ -83,12 +84,12 @@ public class ClientStreamAdapter {\\n @VisibleForTesting(\\\"Allow unit testing behavior job handling behavior\\\")\\n static final class ClientStreamConsumerImpl implements ClientStreamConsumer {\\n private final StreamObserver responseObserver;\\n- private final Executor executor;\\n+ private final SerializingExecutor executor;\\n \\n public ClientStreamConsumerImpl(\\n final StreamObserver responseObserver, final Executor executor) {\\n this.responseObserver = responseObserver;\\n- this.executor = executor;\\n+ this.executor = new SerializingExecutor(executor);\\n }\\n \\n @Override\\n\", \"diff --git a/scripts/helmcharts/init.sh b/scripts/helmcharts/init.sh\\nindex 5a2b4b0..69a6944 100644\\n--- a/scripts/helmcharts/init.sh\\n+++ b/scripts/helmcharts/init.sh\\n@@ -26,7 +26,7 @@ usr=$(whoami)\\n \\n # Installing k3s\\n function install_k8s() {\\n- curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\\\"644\\\" INSTALL_K3S_VERSION='v1.22.8+k3s1' INSTALL_K3S_EXEC=\\\"--no-deploy=traefik\\\" sh -\\n+ curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\\\"644\\\" INSTALL_K3S_VERSION='v1.25.6+k3s1' INSTALL_K3S_EXEC=\\\"--disable=traefik\\\" sh -\\n [[ -d ~/.kube ]] || mkdir ~/.kube\\n sudo cp /etc/rancher/k3s/k3s.yaml ~/.kube/config\\n sudo chmod 0644 ~/.kube/config\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"28e623b294816c4e070971782a75c8697a11966f\", \"dbb8617214aaa8b56b827deef1265d9ee38765bd\", \"22044d58302513f5cf22b06151c4a367bbb88f6e\", \"9a25fe59dfb63d32505afcea3a164ff0b8ea4c71\"]"},"types":{"kind":"string","value":"[\"ci\", \"docs\", \"fix\", \"build\"]"}}},{"rowIdx":929,"cells":{"commit_message":{"kind":"string","value":"added vue3 readme,build improvements,remove unnecessary start argument from `range`,add method to extract snapshot name from filename\n\nalso corrected pattern, where the period was meant to match a period, not any\ncharacter.\n\nrelated to zeebe-io/zeebe#876"},"diff":{"kind":"string","value":"[\"diff --git a/core/main/README.md b/core/main/README.md\\nindex e5e4c93..e9cfda9 100644\\n--- a/core/main/README.md\\n+++ b/core/main/README.md\\n@@ -217,7 +217,7 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\\n \\n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/svelte/README.md)\\n \\n-### VueJS\\n+### VueJS 2.x\\n \\n #### `particles.vue`\\n \\n@@ -225,6 +225,14 @@ You can find the instructions [here](https://github.com/matteobruni/tsparticles/\\n \\n You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue/README.md)\\n \\n+### VueJS 3.x\\n+\\n+#### `particles.vue3`\\n+\\n+[![npm](https://img.shields.io/npm/v/particles.vue3)](https://www.npmjs.com/package/particles.vue3) [![npm](https://img.shields.io/npm/dm/particles.vue3)](https://www.npmjs.com/package/particles.vue3)\\n+\\n+You can find the instructions [here](https://github.com/matteobruni/tsparticles/blob/master/components/vue3/README.md)\\n+\\n ---\\n \\n ## **_Demo / Generator_**\\ndiff --git a/core/main/tsconfig.json b/core/main/tsconfig.json\\nindex 7916bc5..72399c0 100644\\n--- a/core/main/tsconfig.json\\n+++ b/core/main/tsconfig.json\\n@@ -107,10 +107,14 @@\\n \\\"source\\\": \\\"../../components/react/README.md\\\"\\n },\\n {\\n- \\\"title\\\": \\\"Vue\\\",\\n+ \\\"title\\\": \\\"Vue 2.x\\\",\\n \\\"source\\\": \\\"../../components/vue/README.md\\\"\\n },\\n {\\n+ \\\"title\\\": \\\"Vue 3.x\\\",\\n+ \\\"source\\\": \\\"../../components/vue3/README.md\\\"\\n+ },\\n+ {\\n \\\"title\\\": \\\"Svelte\\\",\\n \\\"source\\\": \\\"../../components/svelte/README.md\\\"\\n },\\n\", \"diff --git a/.travis.yml b/.travis.yml\\nindex 9e1b926..3144244 100644\\n--- a/.travis.yml\\n+++ b/.travis.yml\\n@@ -1,5 +1,6 @@\\n language: node_js\\n dist: trusty\\n+sudo: required\\n node_js:\\n - '6.9.5'\\n before_install:\\ndiff --git a/e2e/schematics/command-line.test.ts b/e2e/schematics/command-line.test.ts\\nindex 16d8b34..ea91494 100644\\n--- a/e2e/schematics/command-line.test.ts\\n+++ b/e2e/schematics/command-line.test.ts\\n@@ -68,8 +68,6 @@ describe('Command line', () => {\\n \\n updateFile('apps/myapp/src/app/app.component.spec.ts', `import '@nrwl/mylib';`);\\n \\n- updateRunAffectedToWorkInE2ESetup();\\n-\\n const affectedApps = runCommand('npm run affected:apps -- --files=\\\"libs/mylib/index.ts\\\"');\\n expect(affectedApps).toContain('myapp');\\n expect(affectedApps).not.toContain('myapp2');\\n@@ -147,11 +145,3 @@ describe('Command line', () => {\\n 1000000\\n );\\n });\\n-\\n-function updateRunAffectedToWorkInE2ESetup() {\\n- const runAffected = readFile('node_modules/@nrwl/schematics/src/command-line/affected.js');\\n- const newRunAffected = runAffected\\n- .replace('ng build', '../../node_modules/.bin/ng build')\\n- .replace('ng e2e', '../../node_modules/.bin/ng e2e');\\n- updateFile('node_modules/@nrwl/schematics/src/command-line/affected.js', newRunAffected);\\n-}\\ndiff --git a/e2e/schematics/workspace.test.ts b/e2e/schematics/workspace.test.ts\\nindex 8a41070..8749926 100644\\n--- a/e2e/schematics/workspace.test.ts\\n+++ b/e2e/schematics/workspace.test.ts\\n@@ -82,7 +82,7 @@ describe('Nrwl Convert to Nx Workspace', () => {\\n \\n it('should generate a workspace and not change dependencies or devDependencies if they already exist', () => {\\n // create a new AngularCLI app\\n- runNgNew('--skip-install');\\n+ runNgNew();\\n const nxVersion = '0.0.0';\\n const schematicsVersion = '0.0.0';\\n const ngrxVersion = '0.0.0';\\ndiff --git a/e2e/utils.ts b/e2e/utils.ts\\nindex 422d866..a03104f 100644\\n--- a/e2e/utils.ts\\n+++ b/e2e/utils.ts\\n@@ -17,8 +17,7 @@ export function newProject(): void {\\n copyMissingPackages();\\n execSync('mv ./tmp/proj ./tmp/proj_backup');\\n }\\n- execSync('cp -r ./tmp/proj_backup ./tmp/proj');\\n- setUpSynLink();\\n+ execSync('cp -a ./tmp/proj_backup ./tmp/proj');\\n }\\n \\n export function copyMissingPackages(): void {\\n@@ -26,14 +25,9 @@ export function copyMissingPackages(): void {\\n modulesToCopy.forEach(m => copyNodeModule(projectName, m));\\n }\\n \\n-export function setUpSynLink(): void {\\n- execSync(`ln -s ../@nrwl/schematics/src/command-line/nx.js tmp/${projectName}/node_modules/.bin/nx`);\\n- execSync(`chmod +x tmp/${projectName}/node_modules/.bin/nx`);\\n-}\\n-\\n function copyNodeModule(path: string, name: string) {\\n execSync(`rm -rf tmp/${path}/node_modules/${name}`);\\n- execSync(`cp -r node_modules/${name} tmp/${path}/node_modules/${name}`);\\n+ execSync(`cp -a node_modules/${name} tmp/${path}/node_modules/${name}`);\\n }\\n \\n export function runCLI(\\n@@ -43,7 +37,7 @@ export function runCLI(\\n }\\n ): string {\\n try {\\n- return execSync(`../../node_modules/.bin/ng ${command}`, {\\n+ return execSync(`./node_modules/.bin/ng ${command}`, {\\n cwd: `./tmp/${projectName}`\\n })\\n .toString()\\n@@ -67,7 +61,7 @@ export function newLib(name: string): string {\\n }\\n \\n export function runSchematic(command: string): string {\\n- return execSync(`../../node_modules/.bin/schematics ${command}`, {\\n+ return execSync(`./node_modules/.bin/schematics ${command}`, {\\n cwd: `./tmp/${projectName}`\\n }).toString();\\n }\\ndiff --git a/package.json b/package.json\\nindex bef54f8..9186a58 100644\\n--- a/package.json\\n+++ b/package.json\\n@@ -6,7 +6,7 @@\\n \\\"private\\\": true,\\n \\\"scripts\\\": {\\n \\\"build\\\": \\\"./scripts/build.sh\\\",\\n- \\\"e2e\\\": \\\"yarn build && ./scripts/e2e.sh\\\",\\n+ \\\"e2e\\\": \\\"./scripts/e2e.sh\\\",\\n \\\"format\\\": \\\"./scripts/format.sh\\\",\\n \\\"linknpm\\\": \\\"./scripts/link.sh\\\",\\n \\\"package\\\": \\\"./scripts/package.sh\\\",\\n@@ -14,7 +14,7 @@\\n \\\"copy\\\": \\\"./scripts/copy.sh\\\",\\n \\\"test:schematics\\\": \\\"yarn build && ./scripts/test_schematics.sh\\\",\\n \\\"test:nx\\\": \\\"yarn build && ./scripts/test_nx.sh\\\",\\n- \\\"test\\\": \\\"yarn build && ./scripts/test_nx.sh && ./scripts/test_schematics.sh\\\",\\n+ \\\"test\\\": \\\"yarn linknpm && ./scripts/test_nx.sh && ./scripts/test_schematics.sh\\\",\\n \\\"checkformat\\\": \\\"./scripts/check-format.sh\\\",\\n \\\"publish_npm\\\": \\\"./scripts/publish.sh\\\"\\n },\\ndiff --git a/packages/schematics/src/collection/workspace/index.ts b/packages/schematics/src/collection/workspace/index.ts\\nindex 8f8897f..c70d161 100644\\n--- a/packages/schematics/src/collection/workspace/index.ts\\n+++ b/packages/schematics/src/collection/workspace/index.ts\\n@@ -254,20 +254,7 @@ function moveFiles(options: Schema) {\\n \\n function copyAngularCliTgz() {\\n return (host: Tree) => {\\n- copyFile(\\n- path.join(\\n- 'node_modules',\\n- '@nrwl',\\n- 'schematics',\\n- 'src',\\n- 'collection',\\n- 'application',\\n- 'files',\\n- '__directory__',\\n- '.angular_cli.tgz'\\n- ),\\n- '.'\\n- );\\n+ copyFile(path.join(__dirname, '..', 'application', 'files', '__directory__', '.angular_cli.tgz'), '.');\\n return host;\\n };\\n }\\ndiff --git a/packages/schematics/src/command-line/affected.ts b/packages/schematics/src/command-line/affected.ts\\nindex b7f9173..89a4f72 100644\\n--- a/packages/schematics/src/command-line/affected.ts\\n+++ b/packages/schematics/src/command-line/affected.ts\\n@@ -1,5 +1,7 @@\\n import { execSync } from 'child_process';\\n import { getAffectedApps, parseFiles } from './shared';\\n+import * as path from 'path';\\n+import * as resolve from 'resolve';\\n \\n export function affected(args: string[]): void {\\n const command = args[0];\\n@@ -39,7 +41,7 @@ function build(apps: string[], rest: string[]) {\\n if (apps.length > 0) {\\n console.log(`Building ${apps.join(', ')}`);\\n apps.forEach(app => {\\n- execSync(`ng build ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\\n+ execSync(`${ngPath()} build ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\\n });\\n } else {\\n console.log('No apps to build');\\n@@ -50,9 +52,13 @@ function e2e(apps: string[], rest: string[]) {\\n if (apps.length > 0) {\\n console.log(`Testing ${apps.join(', ')}`);\\n apps.forEach(app => {\\n- execSync(`ng e2e ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\\n+ execSync(`${ngPath()} e2e ${rest.join(' ')} -a=${app}`, { stdio: [0, 1, 2] });\\n });\\n } else {\\n- console.log('No apps to tst');\\n+ console.log('No apps to test');\\n }\\n }\\n+\\n+function ngPath() {\\n+ return `${path.dirname(path.dirname(path.dirname(resolve.sync('@angular/cli', { basedir: __dirname }))))}/bin/ng`;\\n+}\\ndiff --git a/scripts/build.sh b/scripts/build.sh\\nindex ac533b5..9b8891b 100755\\n--- a/scripts/build.sh\\n+++ b/scripts/build.sh\\n@@ -3,6 +3,8 @@\\n rm -rf build\\n ngc\\n rsync -a --exclude=*.ts packages/ build/packages\\n+chmod +x build/packages/schematics/bin/create-nx-workspace.js\\n+chmod +x build/packages/schematics/src/command-line/nx.js\\n rm -rf build/packages/install\\n cp README.md build/packages/schematics\\n cp README.md build/packages/nx\\n\\\\ No newline at end of file\\n\", \"diff --git a/ibis/backends/dask/tests/execution/test_window.py b/ibis/backends/dask/tests/execution/test_window.py\\nindex 75a7331..6bfc5e3 100644\\n--- a/ibis/backends/dask/tests/execution/test_window.py\\n+++ b/ibis/backends/dask/tests/execution/test_window.py\\n@@ -489,7 +489,7 @@ def test_project_list_scalar(npartitions):\\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\\n result = expr.execute()\\n \\n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\\\"res\\\")\\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\\\"res\\\")\\n tm.assert_series_equal(result.res, expected)\\n \\n \\ndiff --git a/ibis/backends/pandas/tests/execution/test_window.py b/ibis/backends/pandas/tests/execution/test_window.py\\nindex 8f292b3..effa372 100644\\n--- a/ibis/backends/pandas/tests/execution/test_window.py\\n+++ b/ibis/backends/pandas/tests/execution/test_window.py\\n@@ -436,7 +436,7 @@ def test_project_list_scalar():\\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\\n result = expr.execute()\\n \\n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\\\"res\\\")\\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\\\"res\\\")\\n tm.assert_series_equal(result.res, expected)\\n \\n \\ndiff --git a/ibis/backends/pyspark/tests/test_basic.py b/ibis/backends/pyspark/tests/test_basic.py\\nindex 3850919..14fe677 100644\\n--- a/ibis/backends/pyspark/tests/test_basic.py\\n+++ b/ibis/backends/pyspark/tests/test_basic.py\\n@@ -19,7 +19,7 @@ from ibis.backends.pyspark.compiler import _can_be_replaced_by_column_name # no\\n def test_basic(con):\\n table = con.table(\\\"basic_table\\\")\\n result = table.compile().toPandas()\\n- expected = pd.DataFrame({\\\"id\\\": range(0, 10), \\\"str_col\\\": \\\"value\\\"})\\n+ expected = pd.DataFrame({\\\"id\\\": range(10), \\\"str_col\\\": \\\"value\\\"})\\n \\n tm.assert_frame_equal(result, expected)\\n \\n@@ -28,9 +28,7 @@ def test_projection(con):\\n table = con.table(\\\"basic_table\\\")\\n result1 = table.mutate(v=table[\\\"id\\\"]).compile().toPandas()\\n \\n- expected1 = pd.DataFrame(\\n- {\\\"id\\\": range(0, 10), \\\"str_col\\\": \\\"value\\\", \\\"v\\\": range(0, 10)}\\n- )\\n+ expected1 = pd.DataFrame({\\\"id\\\": range(10), \\\"str_col\\\": \\\"value\\\", \\\"v\\\": range(10)})\\n \\n result2 = (\\n table.mutate(v=table[\\\"id\\\"])\\n@@ -44,8 +42,8 @@ def test_projection(con):\\n {\\n \\\"id\\\": range(0, 20, 2),\\n \\\"str_col\\\": \\\"value\\\",\\n- \\\"v\\\": range(0, 10),\\n- \\\"v2\\\": range(0, 10),\\n+ \\\"v\\\": range(10),\\n+ \\\"v2\\\": range(10),\\n }\\n )\\n \\n\", \"diff --git a/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java b/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\\nindex d8f4d89..e54e85a 100644\\n--- a/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\\n+++ b/logstreams/src/main/java/io/zeebe/logstreams/impl/snapshot/fs/FsSnapshotStorageConfiguration.java\\n@@ -23,8 +23,9 @@ public class FsSnapshotStorageConfiguration\\n {\\n protected static final String CHECKSUM_ALGORITHM = \\\"SHA1\\\";\\n \\n- protected static final String SNAPSHOT_FILE_NAME_TEMPLATE = \\\"%s\\\" + File.separatorChar + \\\"%s-%d.snapshot\\\";\\n- protected static final String SNAPSHOT_FILE_NAME_PATTERN = \\\"%s-(\\\\\\\\d+).snapshot\\\";\\n+ protected static final String SNAPSHOT_FILE_NAME_TEMPLATE = \\\"%s-%d.snapshot\\\";\\n+ protected static final String SNAPSHOT_FILE_PATH_TEMPLATE = \\\"%s\\\" + File.separatorChar + SNAPSHOT_FILE_NAME_TEMPLATE;\\n+ protected static final String SNAPSHOT_FILE_NAME_PATTERN = \\\"%s-(\\\\\\\\d+)\\\\\\\\.snapshot\\\";\\n \\n protected static final String CHECKSUM_FILE_NAME_TEMPLATE = \\\"%s\\\" + File.separatorChar + \\\"%s-%d.\\\" + CHECKSUM_ALGORITHM.toLowerCase();\\n \\n@@ -50,7 +51,7 @@ public class FsSnapshotStorageConfiguration\\n \\n public String snapshotFileName(String name, long logPosition)\\n {\\n- return String.format(SNAPSHOT_FILE_NAME_TEMPLATE, rootPath, name, logPosition);\\n+ return String.format(SNAPSHOT_FILE_PATH_TEMPLATE, rootPath, name, logPosition);\\n }\\n \\n public String checksumFileName(String name, long logPosition)\\n@@ -86,7 +87,7 @@ public class FsSnapshotStorageConfiguration\\n return String.format(CHECKSUM_CONTENT_TEMPLATE, checksum, dataFileName);\\n }\\n \\n- public String extractDigetsFromChecksumContent(String content)\\n+ public String extractDigestFromChecksumContent(String content)\\n {\\n final int indexOfSeparator = content.indexOf(CHECKSUM_CONTENT_SEPARATOR);\\n if (indexOfSeparator < 0)\\n@@ -108,9 +109,18 @@ public class FsSnapshotStorageConfiguration\\n return content.substring(indexOfSeparator + CHECKSUM_CONTENT_SEPARATOR.length());\\n }\\n \\n+ public String getSnapshotNameFromFileName(final String fileName)\\n+ {\\n+ final String suffixPattern = String.format(SNAPSHOT_FILE_NAME_PATTERN, \\\"\\\");\\n+ final Pattern pattern = Pattern.compile(suffixPattern);\\n+ final String[] parts = pattern.split(fileName);\\n+\\n+ return parts[0];\\n+ }\\n+\\n public String getSnapshotFileNameTemplate()\\n {\\n- return SNAPSHOT_FILE_NAME_TEMPLATE;\\n+ return SNAPSHOT_FILE_PATH_TEMPLATE;\\n }\\n \\n public String getChecksumFileNameTemplate()\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"e4c3e2cff769ce46d22d5c8f7dd527510443a8a7\", \"e0a977b2d316e7612b5d72cb02cd7d78e75dbc55\", \"15f8d95754a0b6865ea475ca9e515272a07bf6ba\", \"7ab965c55d0e98fdb6179577d0db56599675e400\"]"},"types":{"kind":"string","value":"[\"docs\", \"build\", \"refactor\", \"feat\"]"}}},{"rowIdx":930,"cells":{"commit_message":{"kind":"string","value":"fixa few issues,add style prop to FAB group action items. closes #475,updated riot to v6, fixed build,fixed tick interval"},"diff":{"kind":"string","value":"[\"diff --git a/README.md b/README.md\\nindex d944d22..5099f03 100644\\n--- a/README.md\\n+++ b/README.md\\n@@ -10,9 +10,8 @@ React state management with a minimal API. Made with :heart: and ES6 Proxies.\\n \\n \\n \\n-* [Motivation](#motivation)\\n+* [Introduction](#introduction)\\n * [Installation](#installation)\\n- + [Setting up a quick project](#setting-up-a-quick-project)\\n * [Usage](#usage)\\n + [Creating stores](#creating-stores)\\n + [Creating reactive views](#creating-reactive-views)\\n@@ -35,12 +34,14 @@ React state management with a minimal API. Made with :heart: and ES6 Proxies.\\n Easy State consists of two wrapper functions only. `store` creates state stores and `view` creates reactive components, which re-render whenever state stores are mutated. The rest is just plain JavaScript.\\n \\n ```js\\n-import React, from 'react'\\n+import React from 'react'\\n import { store, view } from 'react-easy-state'\\n \\n+// stores are normal objects\\n const clock = store({ time: new Date() })\\n setInterval(() => clock.time = new Date(), 1000)\\n \\n+// reactive components re-render on store mutations\\n function ClockComp () {\\n return
{clock.time}
\\n }\\n\", \"diff --git a/src/components/FAB/FABGroup.js b/src/components/FAB/FABGroup.js\\nindex 424a178..11bd10f 100644\\n--- a/src/components/FAB/FABGroup.js\\n+++ b/src/components/FAB/FABGroup.js\\n@@ -25,6 +25,7 @@ type Props = {\\n * - `label`: optional label text\\n * - `accessibilityLabel`: accessibility label for the action, uses label by default if specified\\n * - `color`: custom icon color of the action item\\n+ * - `style`: pass additional styles for the fab item, for example, `backgroundColor`\\n * - `onPress`: callback that is called when `FAB` is pressed (required)\\n */\\n actions: Array<{\\n@@ -32,6 +33,7 @@ type Props = {\\n label?: string,\\n color?: string,\\n accessibilityLabel?: string,\\n+ style?: any,\\n onPress: () => mixed,\\n }>,\\n /**\\n@@ -44,7 +46,7 @@ type Props = {\\n */\\n accessibilityLabel?: string,\\n /**\\n- * Custom icon color for the `FAB`.\\n+ * Custom color for the `FAB`.\\n */\\n color?: string,\\n /**\\n@@ -252,9 +254,7 @@ class FABGroup extends React.Component {\\n {\\n it.onPress();\\n@@ -280,6 +280,7 @@ class FABGroup extends React.Component {\\n transform: [{ scale: scales[i] }],\\n backgroundColor: theme.colors.surface,\\n },\\n+ it.style,\\n ]}\\n onPress={() => {\\n it.onPress();\\n\", \"diff --git a/components/riot/package.json b/components/riot/package.json\\nindex c41743a..eb69756 100644\\n--- a/components/riot/package.json\\n+++ b/components/riot/package.json\\n@@ -61,7 +61,7 @@\\n },\\n \\\"devDependencies\\\": {\\n \\\"@babel/preset-typescript\\\": \\\"^7.14.5\\\",\\n- \\\"@riotjs/cli\\\": \\\"^6.0.4\\\",\\n+ \\\"@riotjs/cli\\\": \\\"^6.0.5\\\",\\n \\\"@riotjs/compiler\\\": \\\"^6.0.0\\\",\\n \\\"chai\\\": \\\"^4.3.4\\\",\\n \\\"esm\\\": \\\"^3.2.25\\\",\\n\", \"diff --git a/backend/services/integrations/main.go b/backend/services/integrations/main.go\\nindex 4a5e764..35c3ff2 100644\\n--- a/backend/services/integrations/main.go\\n+++ b/backend/services/integrations/main.go\\n@@ -54,7 +54,7 @@ func main() {\\n \\tsigchan := make(chan os.Signal, 1)\\n \\tsignal.Notify(sigchan, syscall.SIGINT, syscall.SIGTERM)\\n \\n-\\ttick := time.Tick(intervals.INTEGRATIONS_REQUEST_INTERVAL)\\n+\\ttick := time.Tick(intervals.INTEGRATIONS_REQUEST_INTERVAL * time.Millisecond)\\n \\n \\tlog.Printf(\\\"Integration service started\\\\n\\\")\\n \\tmanager.RequestAll()\\n@@ -66,7 +66,7 @@ func main() {\\n \\t\\t\\tpg.Close()\\n \\t\\t\\tos.Exit(0)\\n \\t\\tcase <-tick:\\n-\\t\\t\\t// log.Printf(\\\"Requesting all...\\\\n\\\")\\n+\\t\\t\\tlog.Printf(\\\"Requesting all...\\\\n\\\")\\n \\t\\t\\tmanager.RequestAll()\\n \\t\\tcase event := <-manager.Events:\\n \\t\\t\\t// log.Printf(\\\"New integration event: %v\\\\n\\\", *event.RawErrorEvent)\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"b8a664c1b10f4e30a3e221a14211a3cdaf90b7f4\", \"8b9176b44bb01a1eef497a403b0304bc389c9aee\", \"5d256f937f93e5a5ed003df86d38c44834095a11\", \"7dc3b70fe40fc7de255a28bb3098bcb8c0d35365\"]"},"types":{"kind":"string","value":"[\"docs\", \"feat\", \"build\", \"fix\"]"}}},{"rowIdx":931,"cells":{"commit_message":{"kind":"string","value":"explain `ChunkOrder` query test scenario,Template using kube api version\n\nSigned-off-by: rjshrjndrn ,add missing region to cloudformation_stack_set,set name for topology module"},"diff":{"kind":"string","value":"[\"diff --git a/query_tests/src/scenarios.rs b/query_tests/src/scenarios.rs\\nindex f0e352b..86df0e9 100644\\n--- a/query_tests/src/scenarios.rs\\n+++ b/query_tests/src/scenarios.rs\\n@@ -1170,6 +1170,21 @@ impl DbSetup for ChunkOrder {\\n .clear_lifecycle_action()\\n .unwrap();\\n \\n+ // Now we have the the following chunks (same partition and table):\\n+ //\\n+ // | ID | order | tag: region | field: user | time |\\n+ // | -- | ----- | ----------- | ----------- | ---- |\\n+ // | 1 | 1 | \\\"west\\\" | 2 | 100 |\\n+ // | 2 | 0 | \\\"west\\\" | 1 | 100 |\\n+ //\\n+ // The result after deduplication should be:\\n+ //\\n+ // | tag: region | field: user | time |\\n+ // | ----------- | ----------- | ---- |\\n+ // | \\\"west\\\" | 2 | 100 |\\n+ //\\n+ // So the query engine must use `order` as a primary key to sort chunks, NOT `id`.\\n+\\n let scenario = DbScenario {\\n scenario_name: \\\"chunks where chunk ID alone cannot be used for ordering\\\".into(),\\n db,\\n\", \"diff --git a/.github/workflows/api-ee.yaml b/.github/workflows/api-ee.yaml\\nindex c014f34..2a12e0d 100644\\n--- a/.github/workflows/api-ee.yaml\\n+++ b/.github/workflows/api-ee.yaml\\n@@ -8,7 +8,7 @@ on:\\n default: 'false'\\n push:\\n branches:\\n- - dev\\n+ - test_ci\\n paths:\\n - ee/api/**\\n - api/**\\n@@ -112,7 +112,8 @@ jobs:\\n # Deploy command\\n kubectl config set-context --namespace=app --current\\n kubectl config get-contexts\\n- helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks | kubectl apply -f -\\n+ k_version=$(kubectl version --short 2>/dev/null | awk '/Server/{print $NF}')\\n+ helm template openreplay -n app openreplay -f vars.yaml -f /tmp/image_override.yaml --set ingress-nginx.enabled=false --set skipMigration=true --no-hooks --kube-version=$k_version | kubectl apply -f -\\n env:\\n DOCKER_REPO: ${{ secrets.EE_REGISTRY_URL }}\\n # We're not passing -ee flag, because helm will add that.\\n\", \"diff --git a/internal/providers/terraform/aws/cloudformation_stack_set.go b/internal/providers/terraform/aws/cloudformation_stack_set.go\\nindex 6720caa..e752b79 100644\\n--- a/internal/providers/terraform/aws/cloudformation_stack_set.go\\n+++ b/internal/providers/terraform/aws/cloudformation_stack_set.go\\n@@ -12,7 +12,7 @@ func getCloudFormationStackSetRegistryItem() *schema.RegistryItem {\\n \\t}\\n }\\n func NewCloudformationStackSet(d *schema.ResourceData, u *schema.UsageData) *schema.Resource {\\n-\\tr := &aws.CloudformationStackSet{Address: strPtr(d.Address)}\\n+\\tr := &aws.CloudformationStackSet{Address: strPtr(d.Address), Region: strPtr(d.Get(\\\"region\\\").String())}\\n \\tif !d.IsEmpty(\\\"template_body\\\") {\\n \\t\\tr.TemplateBody = strPtr(d.Get(\\\"template_body\\\").String())\\n \\t}\\n\", \"diff --git a/topology/pom.xml b/topology/pom.xml\\nindex 389508e..ee6239a 100644\\n--- a/topology/pom.xml\\n+++ b/topology/pom.xml\\n@@ -16,6 +16,7 @@\\n \\n \\n zeebe-cluster-topology\\n+ Zeebe Cluster Topology\\n \\n \\n ${maven.multiModuleProjectDirectory}/topology/src/main/resources/proto\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"9a60af7fa3b480e2e04bacd646112cad9aaab6d7\", \"c3531347fe5a4cc82d426db195026a5bdad15e7a\", \"304d0588f634e9e72087a706367c53af9c7f7180\", \"8911a972222dc80a242f3f1d9b3596321b3fdeaa\"]"},"types":{"kind":"string","value":"[\"docs\", \"ci\", \"fix\", \"build\"]"}}},{"rowIdx":932,"cells":{"commit_message":{"kind":"string","value":"remove unnecessary start argument from `range`,add LICENSE,adds test for exec with http proxy\n\nSigned-off-by: Sam Alba ,conditionals and iterators in rsx"},"diff":{"kind":"string","value":"[\"diff --git a/ibis/backends/dask/tests/execution/test_window.py b/ibis/backends/dask/tests/execution/test_window.py\\nindex 75a7331..6bfc5e3 100644\\n--- a/ibis/backends/dask/tests/execution/test_window.py\\n+++ b/ibis/backends/dask/tests/execution/test_window.py\\n@@ -489,7 +489,7 @@ def test_project_list_scalar(npartitions):\\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\\n result = expr.execute()\\n \\n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\\\"res\\\")\\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\\\"res\\\")\\n tm.assert_series_equal(result.res, expected)\\n \\n \\ndiff --git a/ibis/backends/pandas/tests/execution/test_window.py b/ibis/backends/pandas/tests/execution/test_window.py\\nindex 8f292b3..effa372 100644\\n--- a/ibis/backends/pandas/tests/execution/test_window.py\\n+++ b/ibis/backends/pandas/tests/execution/test_window.py\\n@@ -436,7 +436,7 @@ def test_project_list_scalar():\\n expr = table.mutate(res=table.ints.quantile([0.5, 0.95]))\\n result = expr.execute()\\n \\n- expected = pd.Series([[1.0, 1.9] for _ in range(0, 3)], name=\\\"res\\\")\\n+ expected = pd.Series([[1.0, 1.9] for _ in range(3)], name=\\\"res\\\")\\n tm.assert_series_equal(result.res, expected)\\n \\n \\ndiff --git a/ibis/backends/pyspark/tests/test_basic.py b/ibis/backends/pyspark/tests/test_basic.py\\nindex 3850919..14fe677 100644\\n--- a/ibis/backends/pyspark/tests/test_basic.py\\n+++ b/ibis/backends/pyspark/tests/test_basic.py\\n@@ -19,7 +19,7 @@ from ibis.backends.pyspark.compiler import _can_be_replaced_by_column_name # no\\n def test_basic(con):\\n table = con.table(\\\"basic_table\\\")\\n result = table.compile().toPandas()\\n- expected = pd.DataFrame({\\\"id\\\": range(0, 10), \\\"str_col\\\": \\\"value\\\"})\\n+ expected = pd.DataFrame({\\\"id\\\": range(10), \\\"str_col\\\": \\\"value\\\"})\\n \\n tm.assert_frame_equal(result, expected)\\n \\n@@ -28,9 +28,7 @@ def test_projection(con):\\n table = con.table(\\\"basic_table\\\")\\n result1 = table.mutate(v=table[\\\"id\\\"]).compile().toPandas()\\n \\n- expected1 = pd.DataFrame(\\n- {\\\"id\\\": range(0, 10), \\\"str_col\\\": \\\"value\\\", \\\"v\\\": range(0, 10)}\\n- )\\n+ expected1 = pd.DataFrame({\\\"id\\\": range(10), \\\"str_col\\\": \\\"value\\\", \\\"v\\\": range(10)})\\n \\n result2 = (\\n table.mutate(v=table[\\\"id\\\"])\\n@@ -44,8 +42,8 @@ def test_projection(con):\\n {\\n \\\"id\\\": range(0, 20, 2),\\n \\\"str_col\\\": \\\"value\\\",\\n- \\\"v\\\": range(0, 10),\\n- \\\"v2\\\": range(0, 10),\\n+ \\\"v\\\": range(10),\\n+ \\\"v2\\\": range(10),\\n }\\n )\\n \\n\", \"diff --git a/LICENSE b/LICENSE\\nnew file mode 100644\\nindex 0000000..005581d\\n--- /dev/null\\n+++ b/LICENSE\\n@@ -0,0 +1,21 @@\\n+MIT License\\n+\\n+Copyright (c) Hassan El Mghari\\n+\\n+Permission is hereby granted, free of charge, to any person obtaining a copy\\n+of this software and associated documentation files (the \\\"Software\\\"), to deal\\n+in the Software without restriction, including without limitation the rights\\n+to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\\n+copies of the Software, and to permit persons to whom the Software is\\n+furnished to do so, subject to the following conditions:\\n+\\n+The above copyright notice and this permission notice shall be included in all\\n+copies or substantial portions of the Software.\\n+\\n+THE SOFTWARE IS PROVIDED \\\"AS IS\\\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\\n+IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\\n+FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\\n+AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\\n+LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\\n+OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\\n+SOFTWARE.\\n\", \"diff --git a/tests/tasks.bats b/tests/tasks.bats\\nindex e3b95c4..4cfba15 100644\\n--- a/tests/tasks.bats\\n+++ b/tests/tasks.bats\\n@@ -101,6 +101,14 @@ setup() {\\n assert_line --partial --index 9 'actions.basicTest.stop'\\n }\\n \\n+@test \\\"task: #Exec with HTTP proxy\\\" {\\n+ cd ./tasks/exec\\n+ export HTTPS_PROXY=\\\"https://localhost:4242/\\\"\\n+ run \\\"$DAGGER\\\" \\\"do\\\" -p ./http_proxy.cue curlProxy\\n+ assert_failure\\n+ unset HTTP_PROXY\\n+}\\n+\\n @test \\\"task: #Start #Stop params\\\" {\\n cd ./tasks/exec\\n \\\"$DAGGER\\\" \\\"do\\\" -p ./start_stop_exec.cue execParamsTest\\n@@ -297,4 +305,4 @@ setup() {\\n \\n @test \\\"task: #Rm\\\" {\\n \\\"$DAGGER\\\" \\\"do\\\" -p ./tasks/rm/rm.cue test\\n-}\\n\\\\ No newline at end of file\\n+}\\ndiff --git a/tests/tasks/exec/http_proxy.cue b/tests/tasks/exec/http_proxy.cue\\nnew file mode 100644\\nindex 0000000..05de4b9\\n--- /dev/null\\n+++ b/tests/tasks/exec/http_proxy.cue\\n@@ -0,0 +1,25 @@\\n+package main\\n+\\n+import (\\n+\\t\\\"dagger.io/dagger\\\"\\n+\\t\\\"dagger.io/dagger/core\\\"\\n+)\\n+\\n+dagger.#Plan & {\\n+\\tactions: {\\n+\\t\\timage: core.#Pull & {\\n+\\t\\t\\tsource: \\\"alpine:3.15.0@sha256:e7d88de73db3d3fd9b2d63aa7f447a10fd0220b7cbf39803c803f2af9ba256b3\\\"\\n+\\t\\t}\\n+\\n+\\t\\tcurlProxy: core.#Exec & {\\n+\\t\\t\\tinput: image.output\\n+\\t\\t\\targs: [\\n+\\t\\t\\t\\t\\\"sh\\\", \\\"-c\\\",\\n+\\t\\t\\t\\t\\\"\\\"\\\"\\n+\\t\\t\\t\\t\\tapk add --no-cache curl\\n+\\t\\t\\t\\t\\tcurl -sfL -o /dev/null https://www.google.com/\\n+\\t\\t\\t\\t\\t\\\"\\\"\\\",\\n+\\t\\t\\t]\\n+\\t\\t}\\n+\\t}\\n+}\\n\", \"diff --git a/packages/interpreter/src/interpreter.js b/packages/interpreter/src/interpreter.js\\nindex 2f5c06f..58613ea 100644\\n--- a/packages/interpreter/src/interpreter.js\\n+++ b/packages/interpreter/src/interpreter.js\\n@@ -172,7 +172,7 @@ export class Interpreter {\\n node.style = {};\\n }\\n node.style[name] = value;\\n- } else if (ns != null || ns != undefined) {\\n+ } else if (ns != null && ns != undefined) {\\n node.setAttributeNS(ns, name, value);\\n } else {\\n switch (name) {\\n@@ -266,7 +266,7 @@ export class Interpreter {\\n this.AssignId(edit.path, edit.id);\\n break;\\n case \\\"CreateElement\\\":\\n- if (edit.namespace !== null || edit.namespace !== undefined) {\\n+ if (edit.namespace !== null && edit.namespace !== undefined) {\\n this.CreateElementNs(edit.name, edit.id, edit.namespace);\\n } else {\\n this.CreateElement(edit.name, edit.id);\\ndiff --git a/packages/rsx/src/lib.rs b/packages/rsx/src/lib.rs\\nindex 09c6bd6..d974a6c 100644\\n--- a/packages/rsx/src/lib.rs\\n+++ b/packages/rsx/src/lib.rs\\n@@ -245,7 +245,11 @@ impl<'a> DynamicContext<'a> {\\n quote! { ::dioxus::core::TemplateNode::Text(#text) }\\n }\\n \\n- BodyNode::Text(_) | BodyNode::RawExpr(_) | BodyNode::Component(_) => {\\n+ BodyNode::RawExpr(_)\\n+ | BodyNode::Text(_)\\n+ | BodyNode::ForLoop(_)\\n+ | BodyNode::IfChain(_)\\n+ | BodyNode::Component(_) => {\\n let ct = self.dynamic_nodes.len();\\n self.dynamic_nodes.push(root);\\n self.node_paths.push(self.current_path.clone());\\ndiff --git a/packages/rsx/src/node.rs b/packages/rsx/src/node.rs\\nindex 4013c9c..7b4bd23 100644\\n--- a/packages/rsx/src/node.rs\\n+++ b/packages/rsx/src/node.rs\\n@@ -5,7 +5,7 @@ use quote::{quote, ToTokens, TokenStreamExt};\\n use syn::{\\n parse::{Parse, ParseStream},\\n spanned::Spanned,\\n- token, Expr, LitStr, Result,\\n+ token, Block, Expr, ExprIf, LitStr, Pat, Result,\\n };\\n \\n /*\\n@@ -20,6 +20,8 @@ Parse\\n pub enum BodyNode {\\n Element(Element),\\n Component(Component),\\n+ ForLoop(ForLoop),\\n+ IfChain(ExprIf),\\n Text(IfmtInput),\\n RawExpr(Expr),\\n }\\n@@ -35,6 +37,8 @@ impl BodyNode {\\n BodyNode::Component(component) => component.name.span(),\\n BodyNode::Text(text) => text.source.span(),\\n BodyNode::RawExpr(exp) => exp.span(),\\n+ BodyNode::ForLoop(fl) => fl.for_token.span(),\\n+ BodyNode::IfChain(f) => f.if_token.span(),\\n }\\n }\\n }\\n@@ -89,6 +93,28 @@ impl Parse for BodyNode {\\n }\\n }\\n \\n+ // Transform for loops into into_iter calls\\n+ if stream.peek(Token![for]) {\\n+ let _f = stream.parse::()?;\\n+ let pat = stream.parse::()?;\\n+ let _i = stream.parse::()?;\\n+ let expr = stream.parse::>()?;\\n+ let body = stream.parse::()?;\\n+\\n+ return Ok(BodyNode::ForLoop(ForLoop {\\n+ for_token: _f,\\n+ pat,\\n+ in_token: _i,\\n+ expr,\\n+ body,\\n+ }));\\n+ }\\n+\\n+ // Transform unterminated if statements into terminated optional if statements\\n+ if stream.peek(Token![if]) {\\n+ return Ok(BodyNode::IfChain(stream.parse()?));\\n+ }\\n+\\n Ok(BodyNode::RawExpr(stream.parse::()?))\\n }\\n }\\n@@ -104,6 +130,104 @@ impl ToTokens for BodyNode {\\n BodyNode::RawExpr(exp) => tokens.append_all(quote! {\\n __cx.fragment_from_iter(#exp)\\n }),\\n+ BodyNode::ForLoop(exp) => {\\n+ let ForLoop {\\n+ pat, expr, body, ..\\n+ } = exp;\\n+\\n+ tokens.append_all(quote! {\\n+ __cx.fragment_from_iter(\\n+ (#expr).into_iter().map(|#pat| {\\n+ #body\\n+ })\\n+ )\\n+ })\\n+ }\\n+ BodyNode::IfChain(chain) => {\\n+ if is_if_chain_terminated(chain) {\\n+ tokens.append_all(quote! {\\n+ __cx.fragment_from_iter(#chain)\\n+ });\\n+ } else {\\n+ let ExprIf {\\n+ cond,\\n+ then_branch,\\n+ else_branch,\\n+ ..\\n+ } = chain;\\n+\\n+ let mut body = TokenStream2::new();\\n+\\n+ body.append_all(quote! {\\n+ if #cond {\\n+ Some(#then_branch)\\n+ }\\n+ });\\n+\\n+ let mut elif = else_branch;\\n+\\n+ while let Some((_, ref branch)) = elif {\\n+ match branch.as_ref() {\\n+ Expr::If(ref eelif) => {\\n+ let ExprIf {\\n+ cond,\\n+ then_branch,\\n+ else_branch,\\n+ ..\\n+ } = eelif;\\n+\\n+ body.append_all(quote! {\\n+ else if #cond {\\n+ Some(#then_branch)\\n+ }\\n+ });\\n+\\n+ elif = else_branch;\\n+ }\\n+ _ => {\\n+ body.append_all(quote! {\\n+ else {\\n+ #branch\\n+ }\\n+ });\\n+ break;\\n+ }\\n+ }\\n+ }\\n+\\n+ body.append_all(quote! {\\n+ else { None }\\n+ });\\n+\\n+ tokens.append_all(quote! {\\n+ __cx.fragment_from_iter(#body)\\n+ });\\n+ }\\n+ }\\n+ }\\n+ }\\n+}\\n+\\n+#[derive(PartialEq, Eq, Clone, Debug, Hash)]\\n+pub struct ForLoop {\\n+ pub for_token: Token![for],\\n+ pub pat: Pat,\\n+ pub in_token: Token![in],\\n+ pub expr: Box,\\n+ pub body: Block,\\n+}\\n+\\n+fn is_if_chain_terminated(chain: &ExprIf) -> bool {\\n+ let mut current = chain;\\n+ loop {\\n+ if let Some((_, else_block)) = &current.else_branch {\\n+ if let Expr::If(else_if) = else_block.as_ref() {\\n+ current = else_if;\\n+ } else {\\n+ return true;\\n+ }\\n+ } else {\\n+ return false;\\n }\\n }\\n }\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"15f8d95754a0b6865ea475ca9e515272a07bf6ba\", \"096145f0d32a6b351b1db413b04a685952f04fb3\", \"6c7398993bc567ec84e4573b6ededbf50b1ef606\", \"6b473cbdc5997af47c56a2a74f5b64da6d4c2ad7\"]"},"types":{"kind":"string","value":"[\"refactor\", \"docs\", \"test\", \"feat\"]"}}},{"rowIdx":933,"cells":{"commit_message":{"kind":"string","value":"build updates,missing transformation for T,Fix readme\n\nSigned-off-by: Ben Johnson ,add test for spurious cross join"},"diff":{"kind":"string","value":"[\"diff --git a/demo/vanilla_new/css/404.min.css b/demo/vanilla_new/css/404.min.css\\nindex a3485b4..e69de29 100644\\n--- a/demo/vanilla_new/css/404.min.css\\n+++ b/demo/vanilla_new/css/404.min.css\\n@@ -1 +0,0 @@\\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\\\"\\\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\\\"Share Tech Mono\\\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\\\"Space Mono\\\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\\\"Space Mono\\\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\\\"\\\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\\\"back to hompage\\\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\\\"report error\\\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\\ndiff --git a/demo/vanilla_new/css/main.min.css b/demo/vanilla_new/css/main.min.css\\nindex 043eb4f..e69de29 100644\\n--- a/demo/vanilla_new/css/main.min.css\\n+++ b/demo/vanilla_new/css/main.min.css\\n@@ -1 +0,0 @@\\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}#floating-container{right:20px;top:20px;position:fixed;z-index:4000}\\ndiff --git a/demo/vanilla_new/js/404.min.js b/demo/vanilla_new/js/404.min.js\\nindex 3642106..e69de29 100644\\n--- a/demo/vanilla_new/js/404.min.js\\n+++ b/demo/vanilla_new/js/404.min.js\\n@@ -1 +0,0 @@\\n-tsParticles.loadJSON(\\\"tsparticles\\\",\\\"/configs/404.json\\\");\\ndiff --git a/website/css/404.min.css b/website/css/404.min.css\\nindex a3485b4..e69de29 100644\\n--- a/website/css/404.min.css\\n+++ b/website/css/404.min.css\\n@@ -1 +0,0 @@\\n-@import url(https://fonts.googleapis.com/css?family=Share+Tech+Mono%7CSpace+Mono);a,abbr,acronym,address,applet,article,aside,audio,b,big,blockquote,body,canvas,caption,center,cite,code,dd,del,details,dfn,div,dl,dt,em,embed,fieldset,figcaption,figure,footer,form,h1,h2,h3,h4,h5,h6,header,hgroup,html,i,iframe,img,ins,kbd,label,legend,li,mark,menu,nav,object,ol,output,p,pre,q,ruby,s,samp,section,small,span,strike,strong,sub,summary,sup,table,tbody,td,tfoot,th,thead,time,tr,tt,u,ul,var,video{margin:0;padding:0;border:0;font-size:100%;font:inherit;vertical-align:baseline}article,aside,details,figcaption,figure,footer,header,hgroup,main,menu,nav,section{display:block}body{line-height:1}ol,ul{list-style:none}blockquote,q{quotes:none}blockquote:after,blockquote:before,q:after,q:before{content:\\\"\\\";content:none}table{border-collapse:collapse;border-spacing:0}body{padding:0;margin:0;font-size:18px}.container{min-height:100vh;position:relative;padding:240px 0;box-sizing:border-box}.overlay{position:absolute;top:0;left:0;width:100%;height:100%;overflow:hidden}.content{position:absolute;top:50%;left:50%;width:100%;transform:translate(-50%,-50%)}.message{text-align:center;color:#000}.message-heading{font-family:\\\"Share Tech Mono\\\";font-weight:900;text-transform:uppercase;letter-spacing:.7em;font-size:2rem;padding:0 0 0 1.4em}.message-description{font-family:\\\"Space Mono\\\";line-height:42px;font-size:15px;letter-spacing:.15rem;padding:0 20px;max-width:600px;margin:auto}.links{max-width:600px;margin:40px auto 0;text-align:center}.links a{width:170px;display:inline-block;padding:15px 0;margin:0 15px;border:1px solid #000;color:#000;text-decoration:none;font-family:\\\"Space Mono\\\";text-transform:uppercase;font-size:11px;letter-spacing:.1rem;position:relative}.links a:before{content:\\\"\\\";height:42px;background:#000;position:absolute;top:0;right:0;width:0;transition:all .3s}.links a:after{transition:all .3s;z-index:999;position:relative;content:\\\"back to hompage\\\"}.links a:hover:before{width:170px}.links a:hover:after{color:#fff}.links a:nth-child(2){background:#fff;color:#000}.links a:nth-child(2):before{background:#212121;left:0}.links a:nth-child(2):after{content:\\\"report error\\\"}.links a:nth-child(2):hover:after{color:#fff}.social{position:absolute;bottom:15px;left:15px}.social-list{margin:0;padding:0;list-style-type:none}.social-list li{display:inline-block;margin:5px 10px}.social-list li a{color:#000}@media (max-width:480px){.message-heading{font-size:1rem;margin-bottom:30px}.message-description{font-size:.7rem;line-height:2rem}.links a{margin:10px;width:280px}.social{left:50%;margin-left:-55px}}\\ndiff --git a/website/css/main.min.css b/website/css/main.min.css\\nindex 818002f..e69de29 100644\\n--- a/website/css/main.min.css\\n+++ b/website/css/main.min.css\\n@@ -1 +0,0 @@\\n-@font-face{font-family:Polya;src:url(https://raw.githubusercontent.com/matteobruni/tsparticles/gh-pages/fonts/Polya.otf)}html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden;background:#000}.github{bottom:10px;right:10px;padding:0 12px 6px;position:fixed;border-radius:10px;background:#fff;border:1px solid #000}.github a{color:#000}.github a:active,.github a:hover,.github a:link,.github a:visited{color:#000;text-decoration:none}.github img{height:30px}.github #gh-project{font-size:20px;padding-left:5px;font-weight:700;vertical-align:bottom}.toggle-sidebar{top:50%;left:0;font-size:20px;color:#000;position:absolute;padding:3px;border-top-right-radius:5px;border-bottom-right-radius:5px;background:#e7e7e7;border:1px solid #000;border-left:none}#editor{background:#fff}[hidden]{display:none}#repulse-div{width:200px;height:200px;background-color:rgba(255,255,255,.5);border-radius:100px;position:absolute;top:50%;left:50%;margin-left:-100px;margin-top:-100px;z-index:200}@media (min-width:1600px) and (-webkit-device-pixel-ratio:1){.col-xxl-3{-ms-flex:0 0 25%;flex:0 0 25%;max-width:25%}}.btn-react{color:#fff;background-color:#61dafb;border-color:#fff}.btn-react:hover{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react.focus,.btn-react:focus{color:#fff;background-color:#5aa3c4;border-color:#ccc;box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}.btn-react.disabled,.btn-react:disabled{color:#fff;background-color:#61dafb;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active,.btn-react:not(:disabled):not(.disabled):active,.show>.btn-react.dropdown-toggle{color:#fff;background-color:#5aa3c4;border-color:#ccc}.btn-react:not(:disabled):not(.disabled).active:focus,.btn-react:not(:disabled):not(.disabled):active:focus,.show>.btn-react.dropdown-toggle:focus{box-shadow:0 0 0 .2rem rgba(90,163,196,.5)}#stats,.count-particles{-webkit-user-select:none}#stats{overflow:hidden}#stats-graph canvas{border-radius:3px 3px 0 0}.count-particles{border-radius:0 0 3px 3px}\\ndiff --git a/website/css/presets.min.css b/website/css/presets.min.css\\nindex 6c2ae2c..e69de29 100644\\n--- a/website/css/presets.min.css\\n+++ b/website/css/presets.min.css\\n@@ -1 +0,0 @@\\n-html{height:100%;overflow:hidden}body{line-height:1;height:100%;overflow:hidden}\\n\", \"diff --git a/src/Tuple/Merge.ts b/src/Tuple/Merge.ts\\nindex dfa7ce5..5ba44b7 100644\\n--- a/src/Tuple/Merge.ts\\n+++ b/src/Tuple/Merge.ts\\n@@ -30,7 +30,7 @@ type _MergeFlat = {\\n }\\n \\n type MergeDeep =\\n- TupleOf, keyof T>, ObjectOf>>>\\n+ TupleOf, Omit, keyof T>, ObjectOf>>>\\n // same principle as above, but with a little tweak\\n // we keep the original `O1` to know if we can merge\\n // => if `O` and `O1` have `object` fields of same name\\n\", \"diff --git a/README.md b/README.md\\nindex 587d655..da746bb 100644\\n--- a/README.md\\n+++ b/README.md\\n@@ -38,15 +38,20 @@ simple and unified.\\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\\n * [**Configuration**][docs.configuration]\\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\\n \\n-#### [Components](https://vector.dev/components)\\n+#### Reference\\n \\n-* [**Sources**][docs.sources] - \\n-* [**Transforms**][docs.transforms]\\n-* [**Sinks**][docs.sinks]\\n+* [**Sources**][docs.sources] - [docker][docs.sources.docker], [file][docs.sources.file], [journald][docs.sources.journald], [kafka][docs.sources.kafka]\\n+* [**Transforms**][docs.transforms] - [json_parser][docs.transforms.json_parser], [log_to_metric][docs.transforms.log_to_metric], [lua][docs.transforms.lua], [regex_parser][docs.transforms.regex_parser]\\n+* [**Sinks**][docs.sinks] - [aws_cloudwatch_logs][docs.sinks.aws_cloudwatch_logs], [aws_cloudwatch_metrics][docs.sinks.aws_cloudwatch_metrics], [aws_kinesis_streams][docs.sinks.aws_kinesis_streams], [aws_s3][docs.sinks.aws_s3], [clickhouse][docs.sinks.clickhouse], [elasticsearch][docs.sinks.elasticsearch], and [15 more][docs.sinks]\\n \\n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\\n-* [**Guides**][docs.guides]\\n+#### Administration\\n+\\n+* [**Process management**][docs.process-management]\\n+* [**Monitoring**][docs.monitoring]\\n+* [**Updating**][docs.updating]\\n+* [**Validating**][docs.validating]\\n \\n #### Resources\\n \\n@@ -105,88 +110,6 @@ curl --proto '=https' --tlsv1.2 -sSf https://sh.vector.dev | sh\\n \\n Or view [platform specific installation instructions][docs.installation].\\n \\n-\\n-## Sources\\n-\\n-| Name | Description |\\n-|:------|:------------|\\n-| [**`docker`**][docs.sources.docker] | Ingests data through the docker engine daemon and outputs [`log`][docs.data-model#log] events. |\\n-| [**`file`**][docs.sources.file] | Ingests data through one or more local files and outputs [`log`][docs.data-model#log] events. |\\n-| [**`journald`**][docs.sources.journald] | Ingests data through log records from journald and outputs [`log`][docs.data-model#log] events. |\\n-| [**`kafka`**][docs.sources.kafka] | Ingests data through Kafka 0.9 or later and outputs [`log`][docs.data-model#log] events. |\\n-| [**`statsd`**][docs.sources.statsd] | Ingests data through the StatsD UDP protocol and outputs [`metric`][docs.data-model#metric] events. |\\n-| [**`stdin`**][docs.sources.stdin] | Ingests data through standard input (STDIN) and outputs [`log`][docs.data-model#log] events. |\\n-| [**`syslog`**][docs.sources.syslog] | Ingests data through the Syslog 5424 protocol and outputs [`log`][docs.data-model#log] events. |\\n-| [**`tcp`**][docs.sources.tcp] | Ingests data through the TCP protocol and outputs [`log`][docs.data-model#log] events. |\\n-| [**`udp`**][docs.sources.udp] | Ingests data through the UDP protocol and outputs [`log`][docs.data-model#log] events. |\\n-| [**`vector`**][docs.sources.vector] | Ingests data through another upstream [`vector` sink][docs.sinks.vector] and outputs [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events. |\\n-\\n-[+ request a new source][urls.new_source]\\n-\\n-\\n-## Transforms\\n-\\n-| Name | Description |\\n-|:------|:------------|\\n-| [**`add_fields`**][docs.transforms.add_fields] | Accepts [`log`][docs.data-model#log] events and allows you to add one or more log fields. |\\n-| [**`add_tags`**][docs.transforms.add_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to add one or more metric tags. |\\n-| [**`coercer`**][docs.transforms.coercer] | Accepts [`log`][docs.data-model#log] events and allows you to coerce log fields into fixed types. |\\n-| [**`field_filter`**][docs.transforms.field_filter] | Accepts [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events and allows you to filter events by a log field's value. |\\n-| [**`grok_parser`**][docs.transforms.grok_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value with [Grok][urls.grok]. |\\n-| [**`json_parser`**][docs.transforms.json_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field value as JSON. |\\n-| [**`log_to_metric`**][docs.transforms.log_to_metric] | Accepts [`log`][docs.data-model#log] events and allows you to convert logs into one or more metrics. |\\n-| [**`lua`**][docs.transforms.lua] | Accepts [`log`][docs.data-model#log] events and allows you to transform events with a full embedded [Lua][urls.lua] engine. |\\n-| [**`regex_parser`**][docs.transforms.regex_parser] | Accepts [`log`][docs.data-model#log] events and allows you to parse a log field's value with a [Regular Expression][urls.regex]. |\\n-| [**`remove_fields`**][docs.transforms.remove_fields] | Accepts [`log`][docs.data-model#log] events and allows you to remove one or more log fields. |\\n-| [**`remove_tags`**][docs.transforms.remove_tags] | Accepts [`metric`][docs.data-model#metric] events and allows you to remove one or more metric tags. |\\n-| [**`sampler`**][docs.transforms.sampler] | Accepts [`log`][docs.data-model#log] events and allows you to sample events with a configurable rate. |\\n-| [**`split`**][docs.transforms.split] | Accepts [`log`][docs.data-model#log] events and allows you to split a field's value on a given separator and zip the tokens into ordered field names. |\\n-| [**`tokenizer`**][docs.transforms.tokenizer] | Accepts [`log`][docs.data-model#log] events and allows you to tokenize a field's value by splitting on white space, ignoring special wrapping characters, and zip the tokens into ordered field names. |\\n-\\n-[+ request a new transform][urls.new_transform]\\n-\\n-\\n-## Sinks\\n-\\n-| Name | Description |\\n-|:------|:------------|\\n-| [**`aws_cloudwatch_logs`**][docs.sinks.aws_cloudwatch_logs] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS CloudWatch Logs][urls.aws_cw_logs] via the [`PutLogEvents` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatchLogs/latest/APIReference/API_PutLogEvents.html). |\\n-| [**`aws_cloudwatch_metrics`**][docs.sinks.aws_cloudwatch_metrics] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [AWS CloudWatch Metrics][urls.aws_cw_metrics] via the [`PutMetricData` API endpoint](https://docs.aws.amazon.com/AmazonCloudWatch/latest/APIReference/API_PutMetricData.html). |\\n-| [**`aws_kinesis_streams`**][docs.sinks.aws_kinesis_streams] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS Kinesis Data Stream][urls.aws_kinesis_data_streams] via the [`PutRecords` API endpoint](https://docs.aws.amazon.com/kinesis/latest/APIReference/API_PutRecords.html). |\\n-| [**`aws_s3`**][docs.sinks.aws_s3] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [AWS S3][urls.aws_s3] via the [`PutObject` API endpoint](https://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectPUT.html). |\\n-| [**`blackhole`**][docs.sinks.blackhole] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to a blackhole that simply discards data, designed for testing and benchmarking purposes. |\\n-| [**`clickhouse`**][docs.sinks.clickhouse] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Clickhouse][urls.clickhouse] via the [`HTTP` Interface][urls.clickhouse_http]. |\\n-| [**`console`**][docs.sinks.console] | [Streams](#streaming) [`log`][docs.data-model#log] and [`metric`][docs.data-model#metric] events to [standard output streams][urls.standard_streams], such as `STDOUT` and `STDERR`. |\\n-| [**`datadog_metrics`**][docs.sinks.datadog_metrics] | [Batches](#buffers-and-batches) [`metric`][docs.data-model#metric] events to [Datadog][urls.datadog] metrics service using [HTTP API](https://docs.datadoghq.com/api/?lang=bash#metrics). |\\n-| [**`elasticsearch`**][docs.sinks.elasticsearch] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to [Elasticsearch][urls.elasticsearch] via the [`_bulk` API endpoint](https://www.elastic.co/guide/en/elasticsearch/reference/current/docs-bulk.html). |\\n-| [**`file`**][docs.sinks.file] | [Streams](#streaming) [`log`][docs.data-model#log] events to a file. |\\n-| [**`http`**][docs.sinks.http] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a generic HTTP endpoint. |\\n-| [**`kafka`**][docs.sinks.kafka] | [Streams](#streaming) [`log`][docs.data-model#log] events to [Apache Kafka][urls.kafka] via the [Kafka protocol][urls.kafka_protocol]. |\\n-| [**`prometheus`**][docs.sinks.prometheus] | [Exposes](#exposing-and-scraping) [`metric`][docs.data-model#metric] events to [Prometheus][urls.prometheus] metrics service. |\\n-| [**`splunk_hec`**][docs.sinks.splunk_hec] | [Batches](#buffers-and-batches) [`log`][docs.data-model#log] events to a [Splunk HTTP Event Collector][urls.splunk_hec]. |\\n-| [**`statsd`**][docs.sinks.statsd] | [Streams](#streaming) [`metric`][docs.data-model#metric] events to [StatsD][urls.statsd] metrics service. |\\n-| [**`tcp`**][docs.sinks.tcp] | [Streams](#streaming) [`log`][docs.data-model#log] events to a TCP connection. |\\n-| [**`vector`**][docs.sinks.vector] | [Streams](#streaming) [`log`][docs.data-model#log] events to another downstream [`vector` source][docs.sources.vector]. |\\n-\\n-[+ request a new sink][urls.new_sink]\\n-\\n-\\n-## License\\n-\\n-Copyright 2019, Vector Authors. All rights reserved.\\n-\\n-Licensed under the Apache License, Version 2.0 (the \\\"License\\\"); you may not\\n-use these files except in compliance with the License. You may obtain a copy\\n-of the License at\\n-\\n-http://www.apache.org/licenses/LICENSE-2.0\\n-\\n-Unless required by applicable law or agreed to in writing, software\\n-distributed under the License is distributed on an \\\"AS IS\\\" BASIS, WITHOUT\\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\\n-License for the specific language governing permissions and limitations under\\n-the License.\\n-\\n ---\\n \\n

\\n@@ -200,8 +123,6 @@ the License.\\n [docs.configuration]: https://vector.dev/docs/setup/configuration\\n [docs.containers]: https://vector.dev/docs/setup/installation/containers\\n [docs.correctness]: https://vector.dev/docs/about/correctness\\n-[docs.data-model#log]: https://vector.dev/docs/about/data-model#log\\n-[docs.data-model#metric]: https://vector.dev/docs/about/data-model#metric\\n [docs.data-model.log]: https://vector.dev/docs/about/data-model/log\\n [docs.data-model.metric]: https://vector.dev/docs/about/data-model/metric\\n [docs.data_model]: https://vector.dev/docs/about/data-model\\n@@ -209,6 +130,7 @@ the License.\\n [docs.from-archives]: https://vector.dev/docs/setup/installation/manual/from-archives\\n [docs.from-source]: https://vector.dev/docs/setup/installation/manual/from-source\\n [docs.guarantees]: https://vector.dev/docs/about/guarantees\\n+[docs.guides.getting_started]: https://vector.dev/docs/setup/guides/getting-started\\n [docs.guides]: https://vector.dev/docs/setup/guides\\n [docs.installation]: https://vector.dev/docs/setup/installation\\n [docs.monitoring]: https://vector.dev/docs/administration/monitoring\\n@@ -224,72 +146,25 @@ the License.\\n [docs.sinks.aws_cloudwatch_metrics]: https://vector.dev/docs/reference/sinks/aws_cloudwatch_metrics\\n [docs.sinks.aws_kinesis_streams]: https://vector.dev/docs/reference/sinks/aws_kinesis_streams\\n [docs.sinks.aws_s3]: https://vector.dev/docs/reference/sinks/aws_s3\\n-[docs.sinks.blackhole]: https://vector.dev/docs/reference/sinks/blackhole\\n [docs.sinks.clickhouse]: https://vector.dev/docs/reference/sinks/clickhouse\\n-[docs.sinks.console]: https://vector.dev/docs/reference/sinks/console\\n-[docs.sinks.datadog_metrics]: https://vector.dev/docs/reference/sinks/datadog_metrics\\n [docs.sinks.elasticsearch]: https://vector.dev/docs/reference/sinks/elasticsearch\\n-[docs.sinks.file]: https://vector.dev/docs/reference/sinks/file\\n-[docs.sinks.http]: https://vector.dev/docs/reference/sinks/http\\n-[docs.sinks.kafka]: https://vector.dev/docs/reference/sinks/kafka\\n-[docs.sinks.prometheus]: https://vector.dev/docs/reference/sinks/prometheus\\n-[docs.sinks.splunk_hec]: https://vector.dev/docs/reference/sinks/splunk_hec\\n-[docs.sinks.statsd]: https://vector.dev/docs/reference/sinks/statsd\\n-[docs.sinks.tcp]: https://vector.dev/docs/reference/sinks/tcp\\n-[docs.sinks.vector]: https://vector.dev/docs/reference/sinks/vector\\n [docs.sinks]: https://vector.dev/docs/reference/sinks\\n [docs.sources.docker]: https://vector.dev/docs/reference/sources/docker\\n [docs.sources.file]: https://vector.dev/docs/reference/sources/file\\n [docs.sources.journald]: https://vector.dev/docs/reference/sources/journald\\n [docs.sources.kafka]: https://vector.dev/docs/reference/sources/kafka\\n-[docs.sources.statsd]: https://vector.dev/docs/reference/sources/statsd\\n-[docs.sources.stdin]: https://vector.dev/docs/reference/sources/stdin\\n-[docs.sources.syslog]: https://vector.dev/docs/reference/sources/syslog\\n-[docs.sources.tcp]: https://vector.dev/docs/reference/sources/tcp\\n-[docs.sources.udp]: https://vector.dev/docs/reference/sources/udp\\n-[docs.sources.vector]: https://vector.dev/docs/reference/sources/vector\\n [docs.sources]: https://vector.dev/docs/reference/sources\\n [docs.topologies]: https://vector.dev/docs/setup/deployment/topologies\\n-[docs.transforms.add_fields]: https://vector.dev/docs/reference/transforms/add_fields\\n-[docs.transforms.add_tags]: https://vector.dev/docs/reference/transforms/add_tags\\n-[docs.transforms.coercer]: https://vector.dev/docs/reference/transforms/coercer\\n-[docs.transforms.field_filter]: https://vector.dev/docs/reference/transforms/field_filter\\n-[docs.transforms.grok_parser]: https://vector.dev/docs/reference/transforms/grok_parser\\n [docs.transforms.json_parser]: https://vector.dev/docs/reference/transforms/json_parser\\n [docs.transforms.log_to_metric]: https://vector.dev/docs/reference/transforms/log_to_metric\\n [docs.transforms.lua]: https://vector.dev/docs/reference/transforms/lua\\n [docs.transforms.regex_parser]: https://vector.dev/docs/reference/transforms/regex_parser\\n-[docs.transforms.remove_fields]: https://vector.dev/docs/reference/transforms/remove_fields\\n-[docs.transforms.remove_tags]: https://vector.dev/docs/reference/transforms/remove_tags\\n-[docs.transforms.sampler]: https://vector.dev/docs/reference/transforms/sampler\\n-[docs.transforms.split]: https://vector.dev/docs/reference/transforms/split\\n-[docs.transforms.tokenizer]: https://vector.dev/docs/reference/transforms/tokenizer\\n [docs.transforms]: https://vector.dev/docs/reference/transforms\\n [docs.updating]: https://vector.dev/docs/administration/updating\\n [docs.use_cases]: https://vector.dev/docs/use_cases\\n [docs.validating]: https://vector.dev/docs/administration/validating\\n-[urls.aws_cw_logs]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/WhatIsCloudWatchLogs.html\\n-[urls.aws_cw_metrics]: https://docs.aws.amazon.com/AmazonCloudWatch/latest/monitoring/working_with_metrics.html\\n-[urls.aws_kinesis_data_streams]: https://aws.amazon.com/kinesis/data-streams/\\n-[urls.aws_s3]: https://aws.amazon.com/s3/\\n-[urls.clickhouse]: https://clickhouse.yandex/\\n-[urls.clickhouse_http]: https://clickhouse.yandex/docs/en/interfaces/http/\\n-[urls.datadog]: https://www.datadoghq.com\\n-[urls.elasticsearch]: https://www.elastic.co/products/elasticsearch\\n-[urls.grok]: http://grokdebug.herokuapp.com/\\n-[urls.kafka]: https://kafka.apache.org/\\n-[urls.kafka_protocol]: https://kafka.apache.org/protocol\\n-[urls.lua]: https://www.lua.org/\\n [urls.mailing_list]: https://vector.dev/mailing_list/\\n-[urls.new_sink]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\\n-[urls.new_source]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\\n-[urls.new_transform]: https://github.com/timberio/vector/issues/new?labels=Type%3A+New+Feature\\n-[urls.prometheus]: https://prometheus.io/\\n-[urls.regex]: https://en.wikipedia.org/wiki/Regular_expression\\n [urls.rust]: https://www.rust-lang.org/\\n-[urls.splunk_hec]: http://dev.splunk.com/view/event-collector/SP-CAAAE6M\\n-[urls.standard_streams]: https://en.wikipedia.org/wiki/Standard_streams\\n-[urls.statsd]: https://github.com/statsd/statsd\\n [urls.test_harness]: https://github.com/timberio/vector-test-harness/\\n [urls.v0.5.0]: https://github.com/timberio/vector/releases/tag/v0.5.0\\n [urls.vector_changelog]: https://github.com/timberio/vector/blob/master/CHANGELOG.md\\ndiff --git a/README.md.erb b/README.md.erb\\nindex 3b14aa0..cc241eb 100644\\n--- a/README.md.erb\\n+++ b/README.md.erb\\n@@ -38,15 +38,20 @@ simple and unified.\\n * [**Installation**][docs.installation] - [containers][docs.containers], [operating systems][docs.operating_systems], [package managers][docs.package_managers], [from archives][docs.from-archives], [from source][docs.from-source]\\n * [**Configuration**][docs.configuration]\\n * [**Deployment**][docs.deployment] - [topologies][docs.topologies], [roles][docs.roles]\\n+* [**Guides**][docs.guides] - [getting started][docs.guides.getting_started]\\n \\n-#### [Components](https://vector.dev/components)\\n+#### Reference\\n \\n-* [**Sources**][docs.sources] - \\n-* [**Transforms**][docs.transforms]\\n-* [**Sinks**][docs.sinks]\\n+* [**Sources**][docs.sources] - <%= common_component_links(:source) %>\\n+* [**Transforms**][docs.transforms] - <%= common_component_links(:transform) %>\\n+* [**Sinks**][docs.sinks] - <%= common_component_links(:sink) %>\\n \\n-* [**Administration**][docs.administration] - [process management][docs.process-management], [monitoring][docs.monitoring], [updating][docs.updating], [validating][docs.validating]\\n-* [**Guides**][docs.guides]\\n+#### Administration\\n+\\n+* [**Process management**][docs.process-management]\\n+* [**Monitoring**][docs.monitoring]\\n+* [**Updating**][docs.updating]\\n+* [**Validating**][docs.validating]\\n \\n #### Resources\\n \\n@@ -105,44 +110,6 @@ Run the following in your terminal, then follow the on-screen instructions.\\n \\n Or view [platform specific installation instructions][docs.installation].\\n \\n-\\n-## Sources\\n-\\n-<%= components_table(metadata.sources.to_h.values.sort) %>\\n-\\n-[+ request a new source][urls.new_source]\\n-\\n-\\n-## Transforms\\n-\\n-<%= components_table(metadata.transforms.to_h.values.sort) %>\\n-\\n-[+ request a new transform][urls.new_transform]\\n-\\n-\\n-## Sinks\\n-\\n-<%= components_table(metadata.sinks.to_h.values.sort) %>\\n-\\n-[+ request a new sink][urls.new_sink]\\n-\\n-\\n-## License\\n-\\n-Copyright <%= Time.now.year %>, Vector Authors. All rights reserved.\\n-\\n-Licensed under the Apache License, Version 2.0 (the \\\"License\\\"); you may not\\n-use these files except in compliance with the License. You may obtain a copy\\n-of the License at\\n-\\n-http://www.apache.org/licenses/LICENSE-2.0\\n-\\n-Unless required by applicable law or agreed to in writing, software\\n-distributed under the License is distributed on an \\\"AS IS\\\" BASIS, WITHOUT\\n-WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\\n-License for the specific language governing permissions and limitations under\\n-the License.\\n-\\n ---\\n \\n

\\ndiff --git a/scripts/generate/templates.rb b/scripts/generate/templates.rb\\nindex e5e7ce7..c793ae0 100644\\n--- a/scripts/generate/templates.rb\\n+++ b/scripts/generate/templates.rb\\n@@ -89,6 +89,23 @@ class Templates\\n render(\\\"#{partials_path}/_commit_type_toc_item.md\\\", binding).gsub(/,$/, \\\"\\\")\\n end\\n \\n+ def common_component_links(type, limit = 5)\\n+ common = metadata.send(\\\"#{type.to_s.pluralize}_list\\\").select(&:common?)\\n+\\n+ links =\\n+ common[0..limit].collect do |component|\\n+ \\\"[#{component.name}][docs.#{type.to_s.pluralize}.#{component.name}]\\\"\\n+ end\\n+\\n+ num_leftover = common.size - links.size\\n+\\n+ if num_leftover > 0\\n+ links << \\\"and [15 more][docs.#{type.to_s.pluralize}]\\\"\\n+ end\\n+\\n+ links.join(\\\", \\\")\\n+ end\\n+\\n def component_config_example(component)\\n render(\\\"#{partials_path}/_component_config_example.md\\\", binding).strip\\n end\\ndiff --git a/scripts/util/metadata/component.rb b/scripts/util/metadata/component.rb\\nindex 0873b2e..4dc5650 100644\\n--- a/scripts/util/metadata/component.rb\\n+++ b/scripts/util/metadata/component.rb\\n@@ -9,6 +9,7 @@ class Component\\n include Comparable\\n \\n attr_reader :beta,\\n+ :common,\\n :function_category,\\n :id,\\n :name,\\n@@ -18,6 +19,7 @@ class Component\\n \\n def initialize(hash)\\n @beta = hash[\\\"beta\\\"] == true\\n+ @common = hash[\\\"common\\\"] == true\\n @function_category = hash.fetch(\\\"function_category\\\")\\n @name = hash.fetch(\\\"name\\\")\\n @type ||= self.class.name.downcase\\n@@ -71,6 +73,10 @@ class Component\\n beta == true\\n end\\n \\n+ def common?\\n+ common == true\\n+ end\\n+\\n def context_options\\n options_list.select(&:context?)\\n end\\ndiff --git a/website/src/components/VectorComponents/index.js b/website/src/components/VectorComponents/index.js\\nindex b6c5c13..d3c9adf 100644\\n--- a/website/src/components/VectorComponents/index.js\\n+++ b/website/src/components/VectorComponents/index.js\\n@@ -154,7 +154,7 @@ function VectorComponents(props) {\\n //\\n \\n const [onlyAtLeastOnce, setOnlyAtLeastOnce] = useState(queryObj['at-least-once'] == 'true');\\n- const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['providers']));\\n+ const [onlyFunctions, setOnlyFunctions] = useState(new Set(queryObj['functions']));\\n const [onlyLog, setOnlyLog] = useState(queryObj['log'] == 'true');\\n const [onlyMetric, setOnlyMetric] = useState(queryObj['metric'] == 'true');\\n const [onlyProductionReady, setOnlyProductionReady] = useState(queryObj['prod-ready'] == 'true');\\n\", \"diff --git a/ibis/tests/sql/test_sqlalchemy.py b/ibis/tests/sql/test_sqlalchemy.py\\nindex 4ad32a6..b2e5d72 100644\\n--- a/ibis/tests/sql/test_sqlalchemy.py\\n+++ b/ibis/tests/sql/test_sqlalchemy.py\\n@@ -841,3 +841,63 @@ def test_filter_group_by_agg_with_same_name():\\n )\\n ex = sa.select([t0]).where(t0.c.bigint_col == 60)\\n _check(expr, ex)\\n+\\n+\\n+@pytest.fixture\\n+def person():\\n+ return ibis.table(\\n+ dict(id=\\\"string\\\", personal=\\\"string\\\", family=\\\"string\\\"),\\n+ name=\\\"person\\\",\\n+ )\\n+\\n+\\n+@pytest.fixture\\n+def visited():\\n+ return ibis.table(\\n+ dict(id=\\\"int32\\\", site=\\\"string\\\", dated=\\\"string\\\"),\\n+ name=\\\"visited\\\",\\n+ )\\n+\\n+\\n+@pytest.fixture\\n+def survey():\\n+ return ibis.table(\\n+ dict(\\n+ taken=\\\"int32\\\",\\n+ person=\\\"string\\\",\\n+ quant=\\\"string\\\",\\n+ reading=\\\"float32\\\",\\n+ ),\\n+ name=\\\"survey\\\",\\n+ )\\n+\\n+\\n+def test_no_cross_join(person, visited, survey):\\n+ expr = person.join(survey, person.id == survey.person).join(\\n+ visited,\\n+ visited.id == survey.taken,\\n+ )\\n+\\n+ context = AlchemyContext(compiler=AlchemyCompiler)\\n+ _ = AlchemyCompiler.to_sql(expr, context)\\n+\\n+ t0 = context.get_ref(person)\\n+ t1 = context.get_ref(survey)\\n+ t2 = context.get_ref(visited)\\n+\\n+ from_ = t0.join(t1, t0.c.id == t1.c.person).join(t2, t2.c.id == t1.c.taken)\\n+ ex = sa.select(\\n+ [\\n+ t0.c.id.label(\\\"id_x\\\"),\\n+ t0.c.personal,\\n+ t0.c.family,\\n+ t1.c.taken,\\n+ t1.c.person,\\n+ t1.c.quant,\\n+ t1.c.reading,\\n+ t2.c.id.label(\\\"id_y\\\"),\\n+ t2.c.site,\\n+ t2.c.dated,\\n+ ]\\n+ ).select_from(from_)\\n+ _check(expr, ex)\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"9acf7a062ee9c0538c2cd4661c1f5da61ab06316\", \"c4d9e5023fa0f88ba283b37da27677ceda1cbfbb\", \"662c5d1346ea2b01c0bc3c11c648cbdf92035fe2\", \"8dac3fe5a7a56356ca95547fcf7925bec8d9c1dd\"]"},"types":{"kind":"string","value":"[\"build\", \"fix\", \"docs\", \"test\"]"}}},{"rowIdx":934,"cells":{"commit_message":{"kind":"string","value":"brew tests/multiple darwin builds/gh enterprise,add --ignore-existing to all npx commands,updated test to use rows for action items\n\nreferences #279,Deploy utilities from correct folder\n\nSigned-off-by: rjshrjndrn "},"diff":{"kind":"string","value":"[\"diff --git a/pipeline/brew/brew.go b/pipeline/brew/brew.go\\nindex ec27182..15ed189 100644\\n--- a/pipeline/brew/brew.go\\n+++ b/pipeline/brew/brew.go\\n@@ -1,5 +1,3 @@\\n-// Package brew implements the Pipe, providing formula generation and\\n-// uploading it to a configured repo.\\n package brew\\n \\n import (\\n@@ -10,13 +8,12 @@ import (\\n \\t\\\"strings\\\"\\n \\t\\\"text/template\\\"\\n \\n-\\t\\\"github.com/goreleaser/goreleaser/internal/artifact\\\"\\n-\\n \\t\\\"github.com/apex/log\\\"\\n \\n \\t\\\"github.com/goreleaser/goreleaser/checksum\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/config\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/context\\\"\\n+\\t\\\"github.com/goreleaser/goreleaser/internal/artifact\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/internal/client\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/pipeline\\\"\\n )\\n@@ -106,14 +103,14 @@ func doRun(ctx *context.Context, client client.Client) error {\\n \\t\\tartifact.And(\\n \\t\\t\\tartifact.ByGoos(\\\"darwin\\\"),\\n \\t\\t\\tartifact.ByGoarch(\\\"amd64\\\"),\\n-\\t\\t\\tartifact.ByGoarch(\\\"\\\"),\\n+\\t\\t\\tartifact.ByGoarm(\\\"\\\"),\\n \\t\\t\\tartifact.ByType(artifact.UploadableArchive),\\n \\t\\t),\\n \\t).List()\\n \\tif len(archives) == 0 {\\n \\t\\treturn ErrNoDarwin64Build\\n \\t}\\n-\\tif len(archives) > 0 {\\n+\\tif len(archives) > 1 {\\n \\t\\treturn ErrTooManyDarwin64Builds\\n \\t}\\n \\tvar path = filepath.Join(ctx.Config.Brew.Folder, ctx.Config.ProjectName+\\\".rb\\\")\\n@@ -145,8 +142,7 @@ func doBuildFormula(data templateData) (out bytes.Buffer, err error) {\\n }\\n \\n func dataFor(ctx *context.Context, client client.Client, artifact artifact.Artifact) (result templateData, err error) {\\n-\\tvar file = artifact.Path\\n-\\tsum, err := checksum.SHA256(file)\\n+\\tsum, err := checksum.SHA256(artifact.Path)\\n \\tif err != nil {\\n \\t\\treturn\\n \\t}\\n@@ -163,7 +159,7 @@ func dataFor(ctx *context.Context, client client.Client, artifact artifact.Artif\\n \\t\\tTag: ctx.Git.CurrentTag,\\n \\t\\tVersion: ctx.Version,\\n \\t\\tCaveats: ctx.Config.Brew.Caveats,\\n-\\t\\tFile: file,\\n+\\t\\tFile: artifact.Name,\\n \\t\\tSHA256: sum,\\n \\t\\tDependencies: ctx.Config.Brew.Dependencies,\\n \\t\\tConflicts: ctx.Config.Brew.Conflicts,\\ndiff --git a/pipeline/brew/brew_test.go b/pipeline/brew/brew_test.go\\nindex 7e513bf..9066935 100644\\n--- a/pipeline/brew/brew_test.go\\n+++ b/pipeline/brew/brew_test.go\\n@@ -9,6 +9,7 @@ import (\\n \\n \\t\\\"github.com/goreleaser/goreleaser/config\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/context\\\"\\n+\\t\\\"github.com/goreleaser/goreleaser/internal/artifact\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/internal/testlib\\\"\\n \\t\\\"github.com/stretchr/testify/assert\\\"\\n )\\n@@ -93,7 +94,8 @@ func TestRunPipe(t *testing.T) {\\n \\t\\tGit: context.GitInfo{\\n \\t\\t\\tCurrentTag: \\\"v1.0.1\\\",\\n \\t\\t},\\n-\\t\\tVersion: \\\"1.0.1\\\",\\n+\\t\\tVersion: \\\"1.0.1\\\",\\n+\\t\\tArtifacts: artifact.New(),\\n \\t\\tConfig: config.Project{\\n \\t\\t\\tDist: folder,\\n \\t\\t\\tProjectName: \\\"run-pipe\\\",\\n@@ -124,31 +126,53 @@ func TestRunPipe(t *testing.T) {\\n \\t\\tPublish: true,\\n \\t}\\n \\tvar path = filepath.Join(folder, \\\"bin.tar.gz\\\")\\n-\\tctx.AddBinary(\\\"darwinamd64\\\", \\\"bin\\\", \\\"bin\\\", path)\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin.tar.gz\\\",\\n+\\t\\tPath: path,\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n \\tclient := &DummyClient{}\\n \\tassert.Error(t, doRun(ctx, client))\\n \\tassert.False(t, client.CreatedFile)\\n \\n \\t_, err = os.Create(path)\\n \\tassert.NoError(t, err)\\n-\\tassert.NoError(t, doRun(ctx, client))\\n-\\tassert.True(t, client.CreatedFile)\\n \\n-\\tbts, err := ioutil.ReadFile(\\\"testdata/run_pipe.rb\\\")\\n-\\tassert.NoError(t, err)\\n-\\t// ioutil.WriteFile(\\\"testdata/run_pipe.rb\\\", []byte(client.Content), 0644)\\n+\\tt.Run(\\\"default git url\\\", func(tt *testing.T) {\\n+\\t\\tassert.NoError(tt, doRun(ctx, client))\\n+\\t\\tassert.True(tt, client.CreatedFile)\\n+\\n+\\t\\tbts, err := ioutil.ReadFile(\\\"testdata/run_pipe.rb\\\")\\n+\\t\\tassert.NoError(tt, err)\\n+\\t\\t// TODO: make writing this file toggleable somehow?\\n+\\t\\t// ioutil.WriteFile(\\\"testdata/run_pipe.rb\\\", []byte(client.Content), 0644)\\n+\\t\\tassert.Equal(tt, string(bts), client.Content)\\n+\\t})\\n \\n-\\tassert.Equal(t, string(bts), client.Content)\\n+\\tt.Run(\\\"github enterprise url\\\", func(tt *testing.T) {\\n+\\t\\tctx.Config.GitHubURLs.Download = \\\"http://github.example.org\\\"\\n+\\t\\tassert.NoError(tt, doRun(ctx, client))\\n+\\t\\tassert.True(tt, client.CreatedFile)\\n+\\n+\\t\\tbts, err := ioutil.ReadFile(\\\"testdata/run_pipe_enterprise.rb\\\")\\n+\\t\\tassert.NoError(tt, err)\\n+\\t\\t// TODO: make writing this file toggleable somehow?\\n+\\t\\t// ioutil.WriteFile(\\\"testdata/run_pipe_enterprise.rb\\\", []byte(client.Content), 0644)\\n+\\t\\tassert.Equal(tt, string(bts), client.Content)\\n+\\t})\\n }\\n \\n+// TODO: this test is irrelevant and can probavly be removed\\n func TestRunPipeFormatOverride(t *testing.T) {\\n \\tfolder, err := ioutil.TempDir(\\\"\\\", \\\"goreleasertest\\\")\\n \\tassert.NoError(t, err)\\n \\tvar path = filepath.Join(folder, \\\"bin.zip\\\")\\n \\t_, err = os.Create(path)\\n \\tassert.NoError(t, err)\\n-\\tvar ctx = &context.Context{\\n-\\t\\tConfig: config.Project{\\n+\\tvar ctx = context.New(\\n+\\t\\tconfig.Project{\\n \\t\\t\\tDist: folder,\\n \\t\\t\\tArchive: config.Archive{\\n \\t\\t\\t\\tFormat: \\\"tar.gz\\\",\\n@@ -166,9 +190,15 @@ func TestRunPipeFormatOverride(t *testing.T) {\\n \\t\\t\\t\\t},\\n \\t\\t\\t},\\n \\t\\t},\\n-\\t\\tPublish: true,\\n-\\t}\\n-\\tctx.AddBinary(\\\"darwinamd64\\\", \\\"bin\\\", \\\"bin\\\", path)\\n+\\t)\\n+\\tctx.Publish = true\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin.zip\\\",\\n+\\t\\tPath: path,\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n \\tclient := &DummyClient{}\\n \\tassert.NoError(t, doRun(ctx, client))\\n \\tassert.True(t, client.CreatedFile)\\n@@ -195,6 +225,40 @@ func TestRunPipeNoDarwin64Build(t *testing.T) {\\n \\tassert.False(t, client.CreatedFile)\\n }\\n \\n+func TestRunPipeMultipleDarwin64Build(t *testing.T) {\\n+\\tvar ctx = context.New(\\n+\\t\\tconfig.Project{\\n+\\t\\t\\tArchive: config.Archive{\\n+\\t\\t\\t\\tFormat: \\\"tar.gz\\\",\\n+\\t\\t\\t},\\n+\\t\\t\\tBrew: config.Homebrew{\\n+\\t\\t\\t\\tGitHub: config.Repo{\\n+\\t\\t\\t\\t\\tOwner: \\\"test\\\",\\n+\\t\\t\\t\\t\\tName: \\\"test\\\",\\n+\\t\\t\\t\\t},\\n+\\t\\t\\t},\\n+\\t\\t},\\n+\\t)\\n+\\tctx.Publish = true\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin1\\\",\\n+\\t\\tPath: \\\"doesnt mather\\\",\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin2\\\",\\n+\\t\\tPath: \\\"doesnt mather\\\",\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n+\\tclient := &DummyClient{}\\n+\\tassert.Equal(t, ErrTooManyDarwin64Builds, doRun(ctx, client))\\n+\\tassert.False(t, client.CreatedFile)\\n+}\\n+\\n func TestRunPipeBrewNotSetup(t *testing.T) {\\n \\tvar ctx = &context.Context{\\n \\t\\tConfig: config.Project{},\\n@@ -206,9 +270,8 @@ func TestRunPipeBrewNotSetup(t *testing.T) {\\n }\\n \\n func TestRunPipeBinaryRelease(t *testing.T) {\\n-\\tvar ctx = &context.Context{\\n-\\t\\tPublish: true,\\n-\\t\\tConfig: config.Project{\\n+\\tvar ctx = context.New(\\n+\\t\\tconfig.Project{\\n \\t\\t\\tArchive: config.Archive{\\n \\t\\t\\t\\tFormat: \\\"binary\\\",\\n \\t\\t\\t},\\n@@ -219,8 +282,15 @@ func TestRunPipeBinaryRelease(t *testing.T) {\\n \\t\\t\\t\\t},\\n \\t\\t\\t},\\n \\t\\t},\\n-\\t}\\n-\\tctx.AddBinary(\\\"darwinamd64\\\", \\\"foo\\\", \\\"bar\\\", \\\"baz\\\")\\n+\\t)\\n+\\tctx.Publish = true\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin\\\",\\n+\\t\\tPath: \\\"doesnt mather\\\",\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.Binary,\\n+\\t})\\n \\tclient := &DummyClient{}\\n \\ttestlib.AssertSkipped(t, doRun(ctx, client))\\n \\tassert.False(t, client.CreatedFile)\\ndiff --git a/pipeline/brew/doc.go b/pipeline/brew/doc.go\\nnew file mode 100644\\nindex 0000000..2cddc12\\n--- /dev/null\\n+++ b/pipeline/brew/doc.go\\n@@ -0,0 +1,3 @@\\n+// Package brew implements the Pipe, providing formula generation and\\n+// uploading it to a configured repo.\\n+package brew\\ndiff --git a/pipeline/brew/testdata/run_pipe_enterprise.rb b/pipeline/brew/testdata/run_pipe_enterprise.rb\\nnew file mode 100644\\nindex 0000000..4b24ce0\\n--- /dev/null\\n+++ b/pipeline/brew/testdata/run_pipe_enterprise.rb\\n@@ -0,0 +1,33 @@\\n+class RunPipe < Formula\\n+ desc \\\"A run pipe test formula\\\"\\n+ homepage \\\"https://github.com/goreleaser\\\"\\n+ url \\\"http://github.example.org/test/test/releases/download/v1.0.1/bin.tar.gz\\\"\\n+ version \\\"1.0.1\\\"\\n+ sha256 \\\"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855\\\"\\n+ \\n+ depends_on \\\"zsh\\\"\\n+ depends_on \\\"bash\\\"\\n+ \\n+ conflicts_with \\\"gtk+\\\"\\n+ conflicts_with \\\"qt\\\"\\n+\\n+ def install\\n+ bin.install \\\"foo\\\"\\n+ end\\n+\\n+ def caveats\\n+ \\\"don't do this\\\"\\n+ end\\n+\\n+ plist_options :startup => false\\n+\\n+ def plist; <<-EOS.undent\\n+ whatever\\n+ EOS\\n+ end\\n+\\n+ test do\\n+ system \\\"true\\\"\\n+ system \\\"#{bin}/foo -h\\\"\\n+ end\\n+end\\n\", \"diff --git a/docs/getting-started/getting-started.md b/docs/getting-started/getting-started.md\\nindex dc6db37..3ef9d0a 100644\\n--- a/docs/getting-started/getting-started.md\\n+++ b/docs/getting-started/getting-started.md\\n@@ -13,7 +13,7 @@ npm install -g @angular/cli\\n **Using `npx`**\\n \\n ```bash\\n-npx create-nx-workspace myworkspace\\n+npx --ignore-existing create-nx-workspace myworkspace\\n ```\\n \\n **Using `npm init`**\\ndiff --git a/docs/guides/react-and-angular.md b/docs/guides/react-and-angular.md\\nindex c1929a2..a5651ff 100644\\n--- a/docs/guides/react-and-angular.md\\n+++ b/docs/guides/react-and-angular.md\\n@@ -11,7 +11,7 @@ To show how Nx does it, let's build two applications (one in Angular, and one in\\n Let's start by creating a new Nx workspace. The easiest way to do this is to use npx.\\n \\n ```bash\\n-npx create-nx-workspace happynrwl --preset=empty\\n+npx --ignore-existing create-nx-workspace happynrwl --preset=empty\\n ```\\n \\n ## Creating an Angular Application\\ndiff --git a/docs/guides/react.md b/docs/guides/react.md\\nindex e1647fd..eac848e 100644\\n--- a/docs/guides/react.md\\n+++ b/docs/guides/react.md\\n@@ -16,13 +16,13 @@ Nx has first class support for React: you can create React applications and libr\\n Create a new Nx workspace. The easiest way to do it is to use npx.\\n \\n ```bash\\n-npx create-nx-workspace happynrwl --preset=empty\\n+npx --ignore-existing create-nx-workspace happynrwl --preset=empty\\n ```\\n \\n You can also create a workspace with a React application in place by running:\\n \\n ```bash\\n-npx create-nx-workspace happynrwl --preset=react\\n+npx --ignore-existing create-nx-workspace happynrwl --preset=react\\n ```\\n \\n ## Generating a React Application\\ndiff --git a/docs/tutorial/01-create-application.md b/docs/tutorial/01-create-application.md\\nindex ea87ecf..967a56e 100644\\n--- a/docs/tutorial/01-create-application.md\\n+++ b/docs/tutorial/01-create-application.md\\n@@ -7,7 +7,7 @@ In this tutorial you will use Nx to build a full-stack application out of common\\n **Start by creating a new workspace.**\\n \\n ```bash\\n-npx create-nx-workspace myorg\\n+npx --ignore-existing create-nx-workspace myorg\\n ```\\n \\n When asked about 'preset', select `empty`.\\n\", \"diff --git a/ionic/components/card/test/advanced/main.html b/ionic/components/card/test/advanced/main.html\\nindex 7c56a7d..c19ea12 100644\\n--- a/ionic/components/card/test/advanced/main.html\\n+++ b/ionic/components/card/test/advanced/main.html\\n@@ -19,16 +19,20 @@\\n

\\n \\n \\n- \\n- \\n- \\n- \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n \\n \\n \\n@@ -51,19 +55,24 @@\\n

Hello. I am a paragraph.

\\n \\n \\n- \\n- \\n- \\n- \\n- Right Note\\n- \\n- \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n \\n \\n \\n@@ -76,20 +85,27 @@\\n This card was breaking the border radius.\\n \\n \\n- \\n- \\n- \\n- \\n- \\n+ \\n+ \\n+ \\n+ \\n+\\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n \\n \\n \\n\", \"diff --git a/.github/workflows/utilities.yaml b/.github/workflows/utilities.yaml\\nindex 92e130c..afbc850 100644\\n--- a/.github/workflows/utilities.yaml\\n+++ b/.github/workflows/utilities.yaml\\n@@ -43,7 +43,7 @@ jobs:\\n PUSH_IMAGE=1 bash build.sh\\n - name: Deploy to kubernetes\\n run: |\\n- cd scripts/helm/\\n+ cd scripts/helmcharts/\\n sed -i \\\"s#openReplayContainerRegistry.*#openReplayContainerRegistry: \\\\\\\"${{ secrets.OSS_REGISTRY_URL }}\\\\\\\"#g\\\" vars.yaml\\n sed -i \\\"s#minio_access_key.*#minio_access_key: \\\\\\\"${{ secrets.OSS_MINIO_ACCESS_KEY }}\\\\\\\" #g\\\" vars.yaml\\n sed -i \\\"s#minio_secret_key.*#minio_secret_key: \\\\\\\"${{ secrets.OSS_MINIO_SECRET_KEY }}\\\\\\\" #g\\\" vars.yaml\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"f433bcb59c36571e22d4e86c612e0a6a52f73c09\", \"fc9af4d0b93d69be4e201ffb18da04324e8a4a87\", \"19feaea1885eb015759b5c7a5d785521f2b8a212\", \"2ebf04099353ef70395b8c8f5e130f70e1ed0814\"]"},"types":{"kind":"string","value":"[\"feat\", \"docs\", \"test\", \"ci\"]"}}},{"rowIdx":935,"cells":{"commit_message":{"kind":"string","value":"treeview width fix\n\nSigned-off-by: Raju Udava <86527202+dstala@users.noreply.github.com>,pass absolute burnchain block height to pox sync watchdog so we correctly infer ibd status,update get-started,add gitignore.nix to dep update matrix"},"diff":{"kind":"string","value":"[\"diff --git a/tests/playwright/pages/Dashboard/TreeView.ts b/tests/playwright/pages/Dashboard/TreeView.ts\\nindex 9cc622b..75c02c0 100644\\n--- a/tests/playwright/pages/Dashboard/TreeView.ts\\n+++ b/tests/playwright/pages/Dashboard/TreeView.ts\\n@@ -23,10 +23,24 @@ export class TreeViewPage extends BasePage {\\n }\\n \\n async verifyVisibility({ isVisible }: { isVisible: boolean }) {\\n- if (isVisible) {\\n- await expect(this.get()).toBeVisible();\\n+ await this.rootPage.waitForTimeout(1000);\\n+\\n+ const domElement = await this.get();\\n+ // get width of treeview dom element\\n+ const width = (await domElement.boundingBox()).width;\\n+\\n+ // if (isVisible) {\\n+ // await expect(this.get()).toBeVisible();\\n+ // } else {\\n+ // await expect(this.get()).not.toBeVisible();\\n+ // }\\n+\\n+ // border for treeview is 1px\\n+ // if not-visible, width should be < 5;\\n+ if (!isVisible) {\\n+ expect(width).toBeLessThan(5);\\n } else {\\n- await expect(this.get()).not.toBeVisible();\\n+ expect(width).toBeGreaterThan(5);\\n }\\n }\\n \\n\", \"diff --git a/testnet/stacks-node/src/run_loop/neon.rs b/testnet/stacks-node/src/run_loop/neon.rs\\nindex 677749b..dc4a7bd 100644\\n--- a/testnet/stacks-node/src/run_loop/neon.rs\\n+++ b/testnet/stacks-node/src/run_loop/neon.rs\\n@@ -411,7 +411,6 @@ impl RunLoop {\\n \\n let mut burnchain_height = sortition_db_height;\\n let mut num_sortitions_in_last_cycle = 1;\\n- let mut learned_burnchain_height = false;\\n \\n // prepare to fetch the first reward cycle!\\n target_burnchain_block_height = burnchain_height + pox_constants.reward_cycle_length as u64;\\n@@ -439,18 +438,16 @@ impl RunLoop {\\n break;\\n }\\n \\n+ let remote_chain_height = burnchain.get_headers_height();\\n+\\n // wait for the p2p state-machine to do at least one pass\\n- debug!(\\\"Wait until we reach steady-state before processing more burnchain blocks...\\\");\\n+ debug!(\\\"Wait until we reach steady-state before processing more burnchain blocks (chain height is {}, we are at {})...\\\", remote_chain_height, burnchain_height);\\n \\n // wait until it's okay to process the next sortitions\\n let ibd = match pox_watchdog.pox_sync_wait(\\n &burnchain_config,\\n &burnchain_tip,\\n- if learned_burnchain_height {\\n- Some(burnchain_height)\\n- } else {\\n- None\\n- },\\n+ Some(remote_chain_height),\\n num_sortitions_in_last_cycle,\\n ) {\\n Ok(ibd) => ibd,\\n@@ -478,7 +475,6 @@ impl RunLoop {\\n };\\n \\n // *now* we know the burnchain height\\n- learned_burnchain_height = true;\\n burnchain_tip = next_burnchain_tip;\\n burnchain_height = cmp::min(burnchain_height + 1, target_burnchain_block_height);\\n \\n\", \"diff --git a/docs/src/go-client/get-started.md b/docs/src/go-client/get-started.md\\nindex 4f4405f..a792e0e 100755\\n--- a/docs/src/go-client/get-started.md\\n+++ b/docs/src/go-client/get-started.md\\n@@ -199,14 +199,12 @@ workflowKey:1 bpmnProcessId:\\\"order-process\\\" version:1 workflowInstanceKey:6\\n \\n You did it! You want to see how the workflow instance is executed?\\n \\n-Start the Zeebe Monitor using `java -jar zeebe-simple-monitor.jar`.\\n+Start the Zeebe Monitor using `java -jar zeebe-simple-monitor-app-*.jar`.\\n \\n Open a web browser and go to .\\n \\n-Connect to the broker and switch to the workflow instances view.\\n-Here, you see the current state of the workflow instance which includes active jobs, completed activities, the payload and open incidents.\\n-\\n-![zeebe-monitor-step-1](/java-client/zeebe-monitor-1.png)\\n+Here, you see the current state of the workflow instance.\\n+![zeebe-monitor-step-1](/java-client/java-get-started-monitor-1.gif)\\n \\n \\n ## Work on a task\\n@@ -322,7 +320,7 @@ it encounters a problem while processing the job.\\n \\n When you have a look at the Zeebe Monitor, then you can see that the workflow instance moved from the first service task to the next one:\\n \\n-![zeebe-monitor-step-2](/go-client/zeebe-monitor-2.png)\\n+![zeebe-monitor-step-2](/java-client/java-get-started-monitor-2.gif)\\n \\n When you run the above example you should see similar output:\\n \\ndiff --git a/docs/src/go-client/java-get-started-monitor-1.gif b/docs/src/go-client/java-get-started-monitor-1.gif\\nnew file mode 100644\\nindex 0000000..b86803a\\nBinary files /dev/null and b/docs/src/go-client/java-get-started-monitor-1.gif differ\\ndiff --git a/docs/src/go-client/java-get-started-monitor-2.gif b/docs/src/go-client/java-get-started-monitor-2.gif\\nnew file mode 100644\\nindex 0000000..8f0f2a4\\nBinary files /dev/null and b/docs/src/go-client/java-get-started-monitor-2.gif differ\\ndiff --git a/docs/src/go-client/zeebe-monitor-1.png b/docs/src/go-client/zeebe-monitor-1.png\\ndeleted file mode 100644\\nindex 0075f3d..0000000\\nBinary files a/docs/src/go-client/zeebe-monitor-1.png and /dev/null differ\\ndiff --git a/docs/src/go-client/zeebe-monitor-2.png b/docs/src/go-client/zeebe-monitor-2.png\\ndeleted file mode 100644\\nindex 6687bb0..0000000\\nBinary files a/docs/src/go-client/zeebe-monitor-2.png and /dev/null differ\\ndiff --git a/docs/src/go-client/zeebe-monitor-3.png b/docs/src/go-client/zeebe-monitor-3.png\\ndeleted file mode 100644\\nindex bc15659..0000000\\nBinary files a/docs/src/go-client/zeebe-monitor-3.png and /dev/null differ\\ndiff --git a/docs/src/introduction/quickstart.md b/docs/src/introduction/quickstart.md\\nindex 70abacf..68be28b 100644\\n--- a/docs/src/introduction/quickstart.md\\n+++ b/docs/src/introduction/quickstart.md\\n@@ -215,7 +215,7 @@ and completed by a [job worker](/basics/job-workers.html). A job worker is a\\n long living process which repeatedly tries to activate jobs for a given job\\n type and completes them after executing its business logic. The `zbctl` also\\n provides a command to spawn simple job workers using an external command or\\n-script. The job worker will receive for every job the payload as JSON object on\\n+script. The job worker will receive for every job the workflow instance variables as JSON object on\\n `stdin` and has to return its result also as JSON object on `stdout` if it\\n handled the job successfully.\\n \\ndiff --git a/docs/src/java-client/get-started.md b/docs/src/java-client/get-started.md\\nindex 54d2208..afc1fd4 100755\\n--- a/docs/src/java-client/get-started.md\\n+++ b/docs/src/java-client/get-started.md\\n@@ -21,9 +21,9 @@ You will be guided through the following steps:\\n * [Zeebe Modeler](https://github.com/zeebe-io/zeebe-modeler/releases)\\n * [Zeebe Monitor](https://github.com/zeebe-io/zeebe-simple-monitor/releases)\\n \\n-Before you begin to setup your project please start the broker, i.e. by running the start up script \\n-`bin/broker` or `bin/broker.bat` in the distribution. Per default the broker is binding to the \\n-address `localhost:26500`, which is used as contact point in this guide. In case your broker is \\n+Before you begin to setup your project please start the broker, i.e. by running the start up script\\n+`bin/broker` or `bin/broker.bat` in the distribution. Per default the broker is binding to the\\n+address `localhost:26500`, which is used as contact point in this guide. In case your broker is\\n available under another address please adjust the broker contact point when building the client.\\n \\n ## Set up a project\\n@@ -182,14 +182,12 @@ Workflow instance created. Key: 6\\n \\n You did it! You want to see how the workflow instance is executed?\\n \\n-Start the Zeebe Monitor using `java -jar zeebe-simple-monitor.jar`.\\n+Start the Zeebe Monitor using `java -jar zeebe-simple-monitor-app-*.jar`.\\n \\n Open a web browser and go to .\\n \\n-Connect to the broker and switch to the workflow instances view.\\n-Here, you see the current state of the workflow instance which includes active jobs, completed activities, the payload and open incidents.\\n-\\n-![zeebe-monitor-step-1](/java-client/zeebe-monitor-1.png)\\n+Here, you see the current state of the workflow instance.\\n+![zeebe-monitor-step-1](/java-client/java-get-started-monitor-1.gif)\\n \\n ## Work on a job\\n \\n@@ -205,12 +203,9 @@ Insert a few service tasks between the start and the end event.\\n You need to set the type of each task, which identifies the nature of the work to be performed.\\n Set the type of the first task to 'payment-service'.\\n \\n-Optionally, you can define parameters of the task by adding headers.\\n-Add the header `method = VISA` to the first task.\\n-\\n Save the BPMN diagram and switch back to the main class.\\n \\n-Add the following lines to create a [job worker][] for the first jobs type:\\n+Add the following lines to create a job worker for the first jobs type:\\n \\n ```java\\n package io.zeebe;\\n@@ -227,10 +222,7 @@ public class Application\\n .jobType(\\\"payment-service\\\")\\n .handler((jobClient, job) ->\\n {\\n- final Map headers = job.getCustomHeaders();\\n- final String method = (String) headers.get(\\\"method\\\");\\n-\\n- System.out.println(\\\"Collect money using payment method: \\\" + method);\\n+ System.out.println(\\\"Collect money\\\");\\n \\n // ...\\n \\n@@ -252,40 +244,29 @@ public class Application\\n Run the program and verify that the job is processed. You should see the output:\\n \\n ```\\n-Collect money using payment method: VISA\\n+Collect money\\n ```\\n \\n When you have a look at the Zeebe Monitor, then you can see that the workflow instance moved from the first service task to the next one:\\n \\n-![zeebe-monitor-step-2](/java-client/zeebe-monitor-2.png)\\n+![zeebe-monitor-step-2](/java-client/java-get-started-monitor-2.gif)\\n \\n ## Work with data\\n \\n-Usually, a workflow is more than just tasks, there is also data flow.\\n-The tasks need data as input and in order to produce data.\\n+Usually, a workflow is more than just tasks, there is also a data flow. The worker gets the data from the workflow instance to do its work and send the result back to the workflow instance.\\n \\n-In Zeebe, the data is represented as a JSON document.\\n-When you create a workflow instance, then you can pass the data as payload.\\n-Within the workflow, you can use input and output mappings on tasks to control the data flow.\\n+In Zeebe, the data is stored as key-value-pairs in form of variables. Variables can be set when the workflow instance is created. Within the workflow, variables can be read and modified by workers.\\n \\n-In our example, we want to create a workflow instance with the following data:\\n+In our example, we want to create a workflow instance with the following variables:\\n \\n ```json\\n-{\\n- \\\"orderId\\\": 31243,\\n- \\\"orderItems\\\": [435, 182, 376]\\n-}\\n+\\\"orderId\\\": 31243\\n+\\\"orderItems\\\": [435, 182, 376]\\n ```\\n \\n-The first task should take `orderId` as input and return `totalPrice` as result.\\n-\\n-Open the BPMN diagram and switch to the input-output-mappings of the first task.\\n-Add the input mapping `$.orderId : $.orderId` and the output mapping `$.totalPrice : $.totalPrice`.\\n+The first task should read `orderId` as input and return `totalPrice` as result.\\n \\n-Save the BPMN diagram and go back to the main class.\\n-\\n-Modify the create command and pass the data as variables.\\n-Also, modify the job worker to read the jobs payload and complete the job with payload.\\n+Modify the workflow instance create command and pass the data as variables. Also, modify the job worker to read the job variables and complete the job with a result.\\n \\n ```java\\n package io.zeebe;\\n@@ -313,23 +294,22 @@ public class Application\\n .jobType(\\\"payment-service\\\")\\n .handler((jobClient, job) ->\\n {\\n- final Map headers = job.getCustomHeaders();\\n- final String method = (String) headers.get(\\\"method\\\");\\n-\\n- final Map payload = job.getPayloadAsMap();\\n+ final Map variables = job.getVariablesAsMap();\\n \\n- System.out.println(\\\"Process order: \\\" + payload.get(\\\"orderId\\\"));\\n- System.out.println(\\\"Collect money using payment method: \\\" + method);\\n+ System.out.println(\\\"Process order: \\\" + variables.get(\\\"orderId\\\"));\\n+ System.out.println(\\\"Collect money\\\");\\n \\n // ...\\n \\n- payload.put(\\\"totalPrice\\\", 46.50);\\n+ final Map result = new HashMap<>();\\n+ result.put(\\\"totalPrice\\\", 46.50);\\n \\n jobClient.newCompleteCommand(job.getKey())\\n- .payload(payload)\\n+ .variables(result)\\n .send()\\n .join();\\n })\\n+ .fetchVariables(\\\"orderId\\\")\\n .open();\\n \\n // ...\\n@@ -337,16 +317,16 @@ public class Application\\n }\\n ```\\n \\n-Run the program and verify that the payload is mapped into the job. You should see the output:\\n+Run the program and verify that the variable is read. You should see the output:\\n \\n ```\\n-Process order: {\\\"orderId\\\":31243}\\n-Collect money using payment method: VISA\\n+Process order: 31243\\n+Collect money\\n ```\\n \\n-When we have a look at the Zeebe Monitor, then we can see how the payload is modified after the activity:\\n+When we have a look at the Zeebe Monitor, then we can see that the variable `totalPrice` is set:\\n \\n-![zeebe-monitor-step-3](/java-client/zeebe-monitor-3.png)\\n+![zeebe-monitor-step-3](/java-client/java-get-started-monitor-3.gif)\\n \\n ## What's next?\\n \\ndiff --git a/docs/src/java-client/java-get-started-monitor-1.gif b/docs/src/java-client/java-get-started-monitor-1.gif\\nnew file mode 100644\\nindex 0000000..b86803a\\nBinary files /dev/null and b/docs/src/java-client/java-get-started-monitor-1.gif differ\\ndiff --git a/docs/src/java-client/java-get-started-monitor-2.gif b/docs/src/java-client/java-get-started-monitor-2.gif\\nnew file mode 100644\\nindex 0000000..8f0f2a4\\nBinary files /dev/null and b/docs/src/java-client/java-get-started-monitor-2.gif differ\\ndiff --git a/docs/src/java-client/java-get-started-monitor-3.gif b/docs/src/java-client/java-get-started-monitor-3.gif\\nnew file mode 100644\\nindex 0000000..1f6cb56\\nBinary files /dev/null and b/docs/src/java-client/java-get-started-monitor-3.gif differ\\ndiff --git a/docs/src/java-client/zeebe-monitor-1.png b/docs/src/java-client/zeebe-monitor-1.png\\ndeleted file mode 100644\\nindex 0075f3d..0000000\\nBinary files a/docs/src/java-client/zeebe-monitor-1.png and /dev/null differ\\ndiff --git a/docs/src/java-client/zeebe-monitor-2.png b/docs/src/java-client/zeebe-monitor-2.png\\ndeleted file mode 100644\\nindex 6687bb0..0000000\\nBinary files a/docs/src/java-client/zeebe-monitor-2.png and /dev/null differ\\ndiff --git a/docs/src/java-client/zeebe-monitor-3.png b/docs/src/java-client/zeebe-monitor-3.png\\ndeleted file mode 100644\\nindex bc15659..0000000\\nBinary files a/docs/src/java-client/zeebe-monitor-3.png and /dev/null differ\\n\", \"diff --git a/.github/workflows/update-deps.yml b/.github/workflows/update-deps.yml\\nindex e0ccd62..1236f58 100644\\n--- a/.github/workflows/update-deps.yml\\n+++ b/.github/workflows/update-deps.yml\\n@@ -13,6 +13,7 @@ jobs:\\n - nixpkgs\\n - poetry2nix\\n - pre-commit-hooks\\n+ - gitignore.nix\\n steps:\\n - name: Checkout\\n uses: actions/checkout@v2\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"efeb30f26252ef4791ef2a02d83827b7f0c45462\", \"5b70e008c57efc89da4061f9adb7d0491b2ea644\", \"cf6d526123abab2689b24a06aaf03d8e4d6ddff4\", \"c444fdb9e85ce44c5c0c99addc777dd7b6085153\"]"},"types":{"kind":"string","value":"[\"test\", \"fix\", \"docs\", \"ci\"]"}}},{"rowIdx":936,"cells":{"commit_message":{"kind":"string","value":"add test for spurious cross join,101: fix import key cmd\n\nSigned-off-by: Sam Alba ,add classname and style props for Playground,cancel in-progress dep update jobs when a new one arrives [skip ci]"},"diff":{"kind":"string","value":"[\"diff --git a/ibis/tests/sql/test_sqlalchemy.py b/ibis/tests/sql/test_sqlalchemy.py\\nindex 4ad32a6..b2e5d72 100644\\n--- a/ibis/tests/sql/test_sqlalchemy.py\\n+++ b/ibis/tests/sql/test_sqlalchemy.py\\n@@ -841,3 +841,63 @@ def test_filter_group_by_agg_with_same_name():\\n )\\n ex = sa.select([t0]).where(t0.c.bigint_col == 60)\\n _check(expr, ex)\\n+\\n+\\n+@pytest.fixture\\n+def person():\\n+ return ibis.table(\\n+ dict(id=\\\"string\\\", personal=\\\"string\\\", family=\\\"string\\\"),\\n+ name=\\\"person\\\",\\n+ )\\n+\\n+\\n+@pytest.fixture\\n+def visited():\\n+ return ibis.table(\\n+ dict(id=\\\"int32\\\", site=\\\"string\\\", dated=\\\"string\\\"),\\n+ name=\\\"visited\\\",\\n+ )\\n+\\n+\\n+@pytest.fixture\\n+def survey():\\n+ return ibis.table(\\n+ dict(\\n+ taken=\\\"int32\\\",\\n+ person=\\\"string\\\",\\n+ quant=\\\"string\\\",\\n+ reading=\\\"float32\\\",\\n+ ),\\n+ name=\\\"survey\\\",\\n+ )\\n+\\n+\\n+def test_no_cross_join(person, visited, survey):\\n+ expr = person.join(survey, person.id == survey.person).join(\\n+ visited,\\n+ visited.id == survey.taken,\\n+ )\\n+\\n+ context = AlchemyContext(compiler=AlchemyCompiler)\\n+ _ = AlchemyCompiler.to_sql(expr, context)\\n+\\n+ t0 = context.get_ref(person)\\n+ t1 = context.get_ref(survey)\\n+ t2 = context.get_ref(visited)\\n+\\n+ from_ = t0.join(t1, t0.c.id == t1.c.person).join(t2, t2.c.id == t1.c.taken)\\n+ ex = sa.select(\\n+ [\\n+ t0.c.id.label(\\\"id_x\\\"),\\n+ t0.c.personal,\\n+ t0.c.family,\\n+ t1.c.taken,\\n+ t1.c.person,\\n+ t1.c.quant,\\n+ t1.c.reading,\\n+ t2.c.id.label(\\\"id_y\\\"),\\n+ t2.c.site,\\n+ t2.c.dated,\\n+ ]\\n+ ).select_from(from_)\\n+ _check(expr, ex)\\n\", \"diff --git a/docs/learn/101-use.md b/docs/learn/101-use.md\\nindex 283c1c1..2ec10f9 100644\\n--- a/docs/learn/101-use.md\\n+++ b/docs/learn/101-use.md\\n@@ -41,8 +41,7 @@ cd ./examples/todoapp\\n The example app contains encrypted secrets and other pre-configured inputs, here is how to decrypt them:\\n \\n ```sh\\n-curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\\n-dagger input list\\n+dagger input list || curl -sfL https://releases.dagger.io/examples/key.txt >> ~/.config/dagger/keys.txt\\n ```\\n \\n **Step 4**: Deploy!\\n\", \"diff --git a/packages/docz-theme-default/src/components/ui/Render.tsx b/packages/docz-theme-default/src/components/ui/Render.tsx\\nindex 197359b..943f9ab 100644\\n--- a/packages/docz-theme-default/src/components/ui/Render.tsx\\n+++ b/packages/docz-theme-default/src/components/ui/Render.tsx\\n@@ -24,9 +24,16 @@ const Code = styled('div')`\\n }\\n `\\n \\n-export const Render: RenderComponent = ({ component, code }) => (\\n+export const Render: RenderComponent = ({\\n+ component,\\n+ code,\\n+ className,\\n+ style,\\n+}) => (\\n \\n- {component}\\n+ \\n+ {component}\\n+ \\n {code}\\n \\n )\\ndiff --git a/packages/docz/src/components/DocPreview.tsx b/packages/docz/src/components/DocPreview.tsx\\nindex ca2d88f..ee8f7c0 100644\\n--- a/packages/docz/src/components/DocPreview.tsx\\n+++ b/packages/docz/src/components/DocPreview.tsx\\n@@ -16,6 +16,8 @@ const DefaultLoading: SFC = () => null\\n export type RenderComponent = ComponentType<{\\n component: JSX.Element\\n code: any\\n+ className?: string\\n+ style?: any\\n }>\\n \\n export const DefaultRender: RenderComponent = ({ component, code }) => (\\ndiff --git a/packages/docz/src/components/Playground.tsx b/packages/docz/src/components/Playground.tsx\\nindex d6ff5a3..418c82e 100644\\n--- a/packages/docz/src/components/Playground.tsx\\n+++ b/packages/docz/src/components/Playground.tsx\\n@@ -9,15 +9,21 @@ export interface PlaygroundProps {\\n __code: (components: ComponentsMap) => any\\n children: any\\n components: ComponentsMap\\n+ className?: string\\n+ style?: any\\n }\\n \\n const BasePlayground: SFC = ({\\n components,\\n children,\\n __code,\\n+ className,\\n+ style,\\n }) => {\\n return components && components.render ? (\\n \\n\", \"diff --git a/.github/workflows/update-deps.yml b/.github/workflows/update-deps.yml\\nindex 3a71e29..25f6f27 100644\\n--- a/.github/workflows/update-deps.yml\\n+++ b/.github/workflows/update-deps.yml\\n@@ -4,6 +4,11 @@ on:\\n # run every 24 hours at midnight\\n - cron: \\\"0 */24 * * *\\\"\\n workflow_dispatch:\\n+\\n+concurrency:\\n+ group: ${{ github.repository }}-${{ github.head_ref || github.sha }}-${{ github.workflow }}\\n+ cancel-in-progress: true\\n+\\n jobs:\\n generate_updates:\\n runs-on: ubuntu-latest\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"8dac3fe5a7a56356ca95547fcf7925bec8d9c1dd\", \"2b01808ec86fe9d8b4a93141a1b7f95e11fd6010\", \"1b64ed30a2e3c41abf3976efee4c7463044b2ef1\", \"c2300c94c6b7d1599387272b616e1d79e93723c7\"]"},"types":{"kind":"string","value":"[\"test\", \"docs\", \"feat\", \"ci\"]"}}},{"rowIdx":937,"cells":{"commit_message":{"kind":"string","value":"release for ppc64\n\ncloses #3703\n\nSigned-off-by: Carlos A Becker ,move toolbar to tab content level\n\nSigned-off-by: Pranav C ,implement array flatten support,enable recovery test\n\nrelated to camunda-tngp/zeebe#353"},"diff":{"kind":"string","value":"[\"diff --git a/.goreleaser.yaml b/.goreleaser.yaml\\nindex 46901cb..7d4d355 100644\\n--- a/.goreleaser.yaml\\n+++ b/.goreleaser.yaml\\n@@ -25,6 +25,7 @@ builds:\\n - amd64\\n - arm\\n - arm64\\n+ - ppc64\\n goarm:\\n - \\\"7\\\"\\n mod_timestamp: '{{ .CommitTimestamp }}'\\n\", \"diff --git a/packages/nc-gui-v2/components.d.ts b/packages/nc-gui-v2/components.d.ts\\nindex f6be04b..cf555ef 100644\\n--- a/packages/nc-gui-v2/components.d.ts\\n+++ b/packages/nc-gui-v2/components.d.ts\\n@@ -201,6 +201,7 @@ declare module '@vue/runtime-core' {\\n MdiThumbUp: typeof import('~icons/mdi/thumb-up')['default']\\n MdiTrashCan: typeof import('~icons/mdi/trash-can')['default']\\n MdiTwitter: typeof import('~icons/mdi/twitter')['default']\\n+ MdiUpload: typeof import('~icons/mdi/upload')['default']\\n MdiUploadOutline: typeof import('~icons/mdi/upload-outline')['default']\\n MdiViewListOutline: typeof import('~icons/mdi/view-list-outline')['default']\\n MdiWhatsapp: typeof import('~icons/mdi/whatsapp')['default']\\ndiff --git a/packages/nc-gui-v2/components/smartsheet-toolbar/ViewActions.vue b/packages/nc-gui-v2/components/smartsheet-toolbar/ViewActions.vue\\nindex c2c87d3..27c0acc 100644\\n--- a/packages/nc-gui-v2/components/smartsheet-toolbar/ViewActions.vue\\n+++ b/packages/nc-gui-v2/components/smartsheet-toolbar/ViewActions.vue\\n@@ -132,7 +132,7 @@ async function changeLockType(type: LockType) {\\n
\\n \\n \\n-
\\n+
\\n \\n .nc-locked-menu-item > div {\\n- @apply grid grid-cols-[30px,auto] gap-2 p-2 align-center;\\n+ @apply grid grid-cols-[30px,auto] gap-2 p-2 items-center;\\n }\\n \\ndiff --git a/packages/nc-gui-v2/components/smartsheet/Toolbar.vue b/packages/nc-gui-v2/components/smartsheet/Toolbar.vue\\nindex 5fa555f..d498871 100644\\n--- a/packages/nc-gui-v2/components/smartsheet/Toolbar.vue\\n+++ b/packages/nc-gui-v2/components/smartsheet/Toolbar.vue\\n@@ -36,7 +36,7 @@ const {isOpen} =useSidebar()\\n \\n \\n \\n- \\n+ \\n \\n \\n
\\ndiff --git a/packages/nc-gui-v2/components/smartsheet/sidebar/index.vue b/packages/nc-gui-v2/components/smartsheet/sidebar/index.vue\\nindex 896ad62..77aee05 100644\\n--- a/packages/nc-gui-v2/components/smartsheet/sidebar/index.vue\\n+++ b/packages/nc-gui-v2/components/smartsheet/sidebar/index.vue\\n@@ -99,6 +99,7 @@ function onCreate(view: GridType | FormType | KanbanType | GalleryType) {\\n class=\\\"relative shadow-md h-full\\\"\\n theme=\\\"light\\\"\\n >\\n+ \\n
\\n \\n \\ndiff --git a/packages/nc-gui-v2/components/smartsheet/sidebar/toolbar/ToggleDrawer.vue b/packages/nc-gui-v2/components/smartsheet/sidebar/toolbar/ToggleDrawer.vue\\nindex 3e3d78a..8441450 100644\\n--- a/packages/nc-gui-v2/components/smartsheet/sidebar/toolbar/ToggleDrawer.vue\\n+++ b/packages/nc-gui-v2/components/smartsheet/sidebar/toolbar/ToggleDrawer.vue\\n@@ -4,7 +4,7 @@ const { isOpen, toggle } = useSidebar({ storageKey: 'nc-right-sidebar' })\\n \\n \\n \\ndiff --git a/packages/nc-gui-v2/components/tabs/Smartsheet.vue b/packages/nc-gui-v2/components/tabs/Smartsheet.vue\\nindex 4181996..7b7ec36 100644\\n--- a/packages/nc-gui-v2/components/tabs/Smartsheet.vue\\n+++ b/packages/nc-gui-v2/components/tabs/Smartsheet.vue\\n@@ -83,11 +83,11 @@ watch(isLocked, (nextValue) => (treeViewIsLockedInj.value = nextValue), { immedi\\n \\n \\n
\\n+ \\n
\\n \\n
\\n \\n- \\n
\\n \\n \\n\", \"diff --git a/ibis/backends/snowflake/registry.py b/ibis/backends/snowflake/registry.py\\nindex 2373dd7..4ce03b0 100644\\n--- a/ibis/backends/snowflake/registry.py\\n+++ b/ibis/backends/snowflake/registry.py\\n@@ -422,6 +422,7 @@ operation_registry.update(\\n ops.ArrayZip: _array_zip,\\n ops.ArraySort: unary(sa.func.array_sort),\\n ops.ArrayRepeat: fixed_arity(sa.func.ibis_udfs.public.array_repeat, 2),\\n+ ops.ArrayFlatten: fixed_arity(sa.func.array_flatten, 1),\\n ops.StringSplit: fixed_arity(sa.func.split, 2),\\n # snowflake typeof only accepts VARIANT, so we cast\\n ops.TypeOf: unary(lambda arg: sa.func.typeof(sa.func.to_variant(arg))),\\n\", \"diff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\nindex 22b8590..db1b553 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n@@ -116,7 +116,6 @@ public class BrokerRecoveryTest\\n ClockUtil.reset();\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldCreateWorkflowInstanceAfterRestart()\\n {\\n@@ -136,7 +135,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_CREATED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceAtTaskAfterRestart()\\n {\\n@@ -166,7 +164,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceWithLockedTaskAfterRestart()\\n {\\n@@ -200,7 +197,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldContinueWorkflowInstanceAtSecondTaskAfterRestart()\\n {\\n@@ -237,7 +233,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasWorkflowInstanceEvent(wfInstanceEvent(\\\"WORKFLOW_INSTANCE_COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldDeployNewWorkflowVersionAfterRestart()\\n {\\n@@ -412,7 +407,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"COMPLETED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldResolveIncidentAfterRestart()\\n {\\n@@ -443,7 +437,6 @@ public class BrokerRecoveryTest\\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"CREATED\\\")));\\n }\\n \\n- @Ignore(\\\"Recovery of workflow deployment event fails - see https://github.com/camunda-tngp/zeebe/issues/353\\\")\\n @Test\\n public void shouldResolveFailedIncidentAfterRestart()\\n {\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"e27e3a6478d59eb0f93af0a51a9c474bad6f8350\", \"bf95d5d0b34d32ef2684488feb3de01cb824b2b4\", \"d3c754f09502be979e5dcc79f968b15052590bd0\", \"f2cc48b74bf92fe22cc265cff4224565f910a921\"]"},"types":{"kind":"string","value":"[\"build\", \"refactor\", \"feat\", \"test\"]"}}},{"rowIdx":938,"cells":{"commit_message":{"kind":"string","value":"refactor generate_completion,backup manager can mark inprogress backups as failed,don't consider cases where there are no txids,remove unnecessary spotless definition\n\nIt receives this already from the parent pom."},"diff":{"kind":"string","value":"[\"diff --git a/src/lib.rs b/src/lib.rs\\nindex dfd8014..15850f7 100644\\n--- a/src/lib.rs\\n+++ b/src/lib.rs\\n@@ -1,11 +1,106 @@\\n //! Generates [Nushell](https://github.com/nushell/nushell) completions for [`clap`](https://github.com/clap-rs/clap) based CLIs\\n \\n-use clap::Command;\\n+use clap::{Arg, Command};\\n use clap_complete::Generator;\\n \\n /// Generate Nushell complete file\\n pub struct Nushell;\\n \\n+enum Argument {\\n+ Short(char),\\n+ Long(String),\\n+ ShortAndLong(char, String),\\n+ Positional(String, bool),\\n+}\\n+\\n+struct ArgumentLine {\\n+ arg: Argument,\\n+ takes_values: bool,\\n+ help: Option,\\n+}\\n+\\n+impl From<&Arg> for ArgumentLine {\\n+ fn from(arg: &Arg) -> Self {\\n+ let takes_values = arg\\n+ .get_num_args()\\n+ .map(|v| v.takes_values())\\n+ .unwrap_or(false);\\n+\\n+ let help = arg.get_help().map(|s| s.to_string());\\n+\\n+ if arg.is_positional() {\\n+ let id = arg.get_id().to_string();\\n+ let required = arg.is_required_set();\\n+ let arg = Argument::Positional(id, required);\\n+\\n+ return Self {\\n+ arg,\\n+ takes_values,\\n+ help,\\n+ };\\n+ }\\n+\\n+ let short = arg.get_short();\\n+ let long = arg.get_long();\\n+\\n+ match short {\\n+ Some(short) => match long {\\n+ Some(long) => Self {\\n+ arg: Argument::ShortAndLong(short, long.into()),\\n+ takes_values,\\n+ help,\\n+ },\\n+ None => Self {\\n+ arg: Argument::Short(short),\\n+ takes_values,\\n+ help,\\n+ },\\n+ },\\n+ None => match long {\\n+ Some(long) => Self {\\n+ arg: Argument::Long(long.into()),\\n+ takes_values,\\n+ help,\\n+ },\\n+ None => unreachable!(\\\"No short or long option found\\\"),\\n+ },\\n+ }\\n+ }\\n+}\\n+\\n+impl ToString for ArgumentLine {\\n+ fn to_string(&self) -> String {\\n+ let mut s = String::new();\\n+\\n+ match &self.arg {\\n+ Argument::Short(short) => s.push_str(format!(\\\" -{}\\\", short).as_str()),\\n+ Argument::Long(long) => s.push_str(format!(\\\" --{}\\\", long).as_str()),\\n+ Argument::ShortAndLong(short, long) => {\\n+ s.push_str(format!(\\\" --{}(-{})\\\", long, short).as_str())\\n+ }\\n+ Argument::Positional(positional, required) => {\\n+ s.push_str(format!(\\\" {}\\\", positional).as_str());\\n+\\n+ if !*required {\\n+ s.push('?');\\n+ }\\n+ }\\n+ }\\n+\\n+ if self.takes_values {\\n+ s.push_str(\\\": string\\\");\\n+ }\\n+\\n+ if let Some(help) = &self.help {\\n+ s.push_str(format!(\\\"\\\\t# {}\\\", help).as_str());\\n+ }\\n+\\n+ s.push('\\\\n');\\n+\\n+ s\\n+ }\\n+}\\n+\\n impl Generator for Nushell {\\n fn file_name(&self, name: &str) -> String {\\n format!(\\\"{}.nu\\\", name)\\n@@ -37,51 +132,18 @@ fn generate_completion(completions: &mut String, cmd: &Command, is_subcommand: b\\n \\n let bin_name = cmd.get_bin_name().expect(\\\"Failed to get bin name\\\");\\n \\n- if is_subcommand {\\n- completions.push_str(format!(\\\" export extern \\\\\\\"{}\\\\\\\" [\\\\n\\\", bin_name).as_str());\\n+ let name = if is_subcommand {\\n+ format!(r#\\\"\\\"{}\\\"\\\"#, bin_name)\\n } else {\\n- completions.push_str(format!(\\\" export extern {} [\\\\n\\\", bin_name).as_str());\\n- }\\n+ bin_name.into()\\n+ };\\n \\n- let mut s = String::new();\\n- for arg in cmd.get_arguments() {\\n- if arg.is_positional() {\\n- s.push_str(format!(\\\" {}\\\", arg.get_id()).as_str());\\n- if !arg.is_required_set() {\\n- s.push('?');\\n- }\\n- }\\n-\\n- let long = arg.get_long();\\n- if let Some(opt) = long {\\n- s.push_str(format!(\\\" --{}\\\", opt).as_str());\\n- }\\n+ completions.push_str(format!(\\\" export extern {} [\\\\n\\\", name).as_str());\\n \\n- let short = arg.get_short();\\n- if let Some(opt) = short {\\n- if long.is_some() {\\n- s.push_str(format!(\\\"(-{})\\\", opt).as_str());\\n- } else {\\n- s.push_str(format!(\\\" -{}\\\", opt).as_str());\\n- }\\n- }\\n-\\n- if let Some(v) = arg.get_num_args() {\\n- if v.takes_values() {\\n- // TODO: add more types?\\n- // TODO: add possible values?\\n- s.push_str(\\\": string\\\");\\n- }\\n- }\\n-\\n- if let Some(msg) = arg.get_help() {\\n- if arg.is_positional() || long.is_some() || short.is_some() {\\n- s.push_str(format!(\\\"\\\\t# {}\\\", msg).as_str());\\n- }\\n- }\\n-\\n- s.push('\\\\n');\\n- }\\n+ let s: String = cmd\\n+ .get_arguments()\\n+ .map(|arg| ArgumentLine::from(arg).to_string())\\n+ .collect();\\n \\n completions.push_str(&s);\\n completions.push_str(\\\" ]\\\\n\\\\n\\\");\\n\", \"diff --git a/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java b/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java\\nindex b2dfb98..21eaf6d 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java\\n@@ -42,4 +42,6 @@ public interface BackupManager {\\n \\n /** Close Backup manager */\\n ActorFuture closeAsync();\\n+\\n+ void failInProgressBackup(long lastCheckpointId);\\n }\\ndiff --git a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java\\nindex a1e1319..33149ae 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java\\n@@ -16,6 +16,7 @@ import io.camunda.zeebe.scheduler.future.ActorFuture;\\n import io.camunda.zeebe.scheduler.future.CompletableActorFuture;\\n import io.camunda.zeebe.snapshots.PersistedSnapshotStore;\\n import java.nio.file.Path;\\n+import java.util.List;\\n import java.util.function.Predicate;\\n import org.slf4j.Logger;\\n import org.slf4j.LoggerFactory;\\n@@ -31,11 +32,13 @@ public final class BackupService extends Actor implements BackupManager {\\n private final PersistedSnapshotStore snapshotStore;\\n private final Path segmentsDirectory;\\n private final Predicate isSegmentsFile;\\n+ private List partitionMembers;\\n \\n public BackupService(\\n final int nodeId,\\n final int partitionId,\\n final int numberOfPartitions,\\n+ final List partitionMembers,\\n final PersistedSnapshotStore snapshotStore,\\n final Predicate isSegmentsFile,\\n final Path segmentsDirectory) {\\n@@ -48,6 +51,7 @@ public final class BackupService extends Actor implements BackupManager {\\n snapshotStore,\\n segmentsDirectory,\\n isSegmentsFile);\\n+ this.partitionMembers = partitionMembers;\\n }\\n \\n public BackupService(\\n@@ -122,6 +126,12 @@ public final class BackupService extends Actor implements BackupManager {\\n new UnsupportedOperationException(\\\"Not implemented\\\"));\\n }\\n \\n+ @Override\\n+ public void failInProgressBackup(final long lastCheckpointId) {\\n+ internalBackupManager.failInProgressBackups(\\n+ partitionId, lastCheckpointId, partitionMembers, actor);\\n+ }\\n+\\n private BackupIdentifierImpl getBackupId(final long checkpointId) {\\n return new BackupIdentifierImpl(nodeId, partitionId, checkpointId);\\n }\\ndiff --git a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java\\nindex e462dd5..f6d76b6 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java\\n@@ -9,16 +9,23 @@ package io.camunda.zeebe.backup.management;\\n \\n import io.camunda.zeebe.backup.api.BackupIdentifier;\\n import io.camunda.zeebe.backup.api.BackupStatus;\\n+import io.camunda.zeebe.backup.api.BackupStatusCode;\\n import io.camunda.zeebe.backup.api.BackupStore;\\n+import io.camunda.zeebe.backup.common.BackupIdentifierImpl;\\n+import io.camunda.zeebe.backup.processing.state.CheckpointState;\\n import io.camunda.zeebe.scheduler.ConcurrencyControl;\\n import io.camunda.zeebe.scheduler.future.ActorFuture;\\n import io.camunda.zeebe.scheduler.future.CompletableActorFuture;\\n+import java.util.Collection;\\n import java.util.HashSet;\\n import java.util.Set;\\n import java.util.function.BiConsumer;\\n import java.util.function.Consumer;\\n+import org.slf4j.Logger;\\n+import org.slf4j.LoggerFactory;\\n \\n final class BackupServiceImpl {\\n+ private static final Logger LOG = LoggerFactory.getLogger(BackupServiceImpl.class);\\n private final Set backupsInProgress = new HashSet<>();\\n private final BackupStore backupStore;\\n private ConcurrencyControl concurrencyControl;\\n@@ -138,4 +145,48 @@ final class BackupServiceImpl {\\n }));\\n return future;\\n }\\n+\\n+ void failInProgressBackups(\\n+ final int partitionId,\\n+ final long lastCheckpointId,\\n+ final Collection brokers,\\n+ final ConcurrencyControl executor) {\\n+ if (lastCheckpointId != CheckpointState.NO_CHECKPOINT) {\\n+ executor.run(\\n+ () -> {\\n+ final var backupIds =\\n+ brokers.stream()\\n+ .map(b -> new BackupIdentifierImpl(b, partitionId, lastCheckpointId))\\n+ .toList();\\n+ // Fail backups initiated by previous leaders\\n+ backupIds.forEach(this::failInProgressBackup);\\n+ });\\n+ }\\n+ }\\n+\\n+ private void failInProgressBackup(final BackupIdentifier backupId) {\\n+ backupStore\\n+ .getStatus(backupId)\\n+ .thenAccept(\\n+ status -> {\\n+ if (status.statusCode() == BackupStatusCode.IN_PROGRESS) {\\n+ LOG.debug(\\n+ \\\"The backup {} initiated by previous leader is still in progress. Marking it as failed.\\\",\\n+ backupId);\\n+ backupStore\\n+ .markFailed(backupId)\\n+ .thenAccept(ignore -> LOG.trace(\\\"Marked backup {} as failed.\\\", backupId))\\n+ .exceptionally(\\n+ failed -> {\\n+ LOG.debug(\\\"Failed to mark backup {} as failed\\\", backupId, failed);\\n+ return null;\\n+ });\\n+ }\\n+ })\\n+ .exceptionally(\\n+ error -> {\\n+ LOG.debug(\\\"Failed to retrieve status of backup {}\\\", backupId);\\n+ return null;\\n+ });\\n+ }\\n }\\ndiff --git a/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java b/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java\\nindex c83fdc1..2899d4d 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java\\n@@ -14,20 +14,24 @@ import io.camunda.zeebe.backup.processing.state.DbCheckpointState;\\n import io.camunda.zeebe.engine.api.ProcessingResult;\\n import io.camunda.zeebe.engine.api.ProcessingResultBuilder;\\n import io.camunda.zeebe.engine.api.ProcessingScheduleService;\\n+import io.camunda.zeebe.engine.api.ReadonlyStreamProcessorContext;\\n import io.camunda.zeebe.engine.api.RecordProcessor;\\n import io.camunda.zeebe.engine.api.RecordProcessorContext;\\n+import io.camunda.zeebe.engine.api.StreamProcessorLifecycleAware;\\n import io.camunda.zeebe.engine.api.TypedRecord;\\n import io.camunda.zeebe.protocol.impl.record.value.management.CheckpointRecord;\\n import io.camunda.zeebe.protocol.record.ValueType;\\n import io.camunda.zeebe.protocol.record.intent.management.CheckpointIntent;\\n import java.time.Duration;\\n+import java.util.List;\\n import java.util.Set;\\n import java.util.concurrent.CopyOnWriteArraySet;\\n import org.slf4j.Logger;\\n import org.slf4j.LoggerFactory;\\n \\n /** Process and replays records related to Checkpoint. */\\n-public final class CheckpointRecordsProcessor implements RecordProcessor {\\n+public final class CheckpointRecordsProcessor\\n+ implements RecordProcessor, StreamProcessorLifecycleAware {\\n \\n private static final Logger LOG = LoggerFactory.getLogger(CheckpointRecordsProcessor.class);\\n \\n@@ -62,6 +66,8 @@ public final class CheckpointRecordsProcessor implements RecordProcessor {\\n checkpointListeners.forEach(\\n listener -> listener.onNewCheckpointCreated(checkpointState.getCheckpointId()));\\n }\\n+\\n+ recordProcessorContext.addLifecycleListeners(List.of(this));\\n }\\n \\n @Override\\n@@ -126,4 +132,12 @@ public final class CheckpointRecordsProcessor implements RecordProcessor {\\n });\\n }\\n }\\n+\\n+ @Override\\n+ public void onRecovered(final ReadonlyStreamProcessorContext context) {\\n+ // After a leader change, the new leader will not continue taking the backup initiated by\\n+ // previous leader. So mark them as failed, so that the users do not wait forever for it to be\\n+ // completed.\\n+ backupManager.failInProgressBackup(checkpointState.getCheckpointId());\\n+ }\\n }\\ndiff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java\\nindex 3424e19..591e17b 100644\\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java\\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java\\n@@ -7,6 +7,7 @@\\n */\\n package io.camunda.zeebe.broker.system.partitions.impl.steps;\\n \\n+import io.atomix.cluster.MemberId;\\n import io.atomix.raft.RaftServer.Role;\\n import io.camunda.zeebe.backup.api.BackupManager;\\n import io.camunda.zeebe.backup.management.BackupService;\\n@@ -17,6 +18,7 @@ import io.camunda.zeebe.journal.file.SegmentFile;\\n import io.camunda.zeebe.scheduler.future.ActorFuture;\\n import io.camunda.zeebe.scheduler.future.CompletableActorFuture;\\n import java.nio.file.Path;\\n+import java.util.List;\\n import java.util.function.Predicate;\\n \\n public final class BackupServiceTransitionStep implements PartitionTransitionStep {\\n@@ -69,6 +71,7 @@ public final class BackupServiceTransitionStep implements PartitionTransitionSte\\n context.getNodeId(),\\n context.getPartitionId(),\\n context.getBrokerCfg().getCluster().getPartitionsCount(),\\n+ getPartitionMembers(context),\\n context.getPersistedSnapshotStore(),\\n isSegmentsFile,\\n context.getRaftPartition().dataDirectory().toPath());\\n@@ -90,4 +93,12 @@ public final class BackupServiceTransitionStep implements PartitionTransitionSte\\n });\\n return installed;\\n }\\n+\\n+ // Brokers which are members of this partition's replication group\\n+ private static List getPartitionMembers(final PartitionTransitionContext context) {\\n+ return context.getRaftPartition().members().stream()\\n+ .map(MemberId::id)\\n+ .map(Integer::parseInt)\\n+ .toList();\\n+ }\\n }\\n\", \"diff --git a/src/main.rs b/src/main.rs\\nindex 25d9580..9ba4e38 100644\\n--- a/src/main.rs\\n+++ b/src/main.rs\\n@@ -441,6 +441,9 @@ fn main() {\\n let mut delta_tx_fees = vec![];\\n let empty_txids = vec![];\\n let txids = tx_mined_deltas.get(&delta).unwrap_or(&empty_txids);\\n+ if txids.len() == 0 {\\n+ continue;\\n+ }\\n for txid in txids.iter() {\\n delta_tx_fees.push(*tx_fees.get(txid).unwrap_or(&0));\\n }\\n\", \"diff --git a/benchmarks/project/pom.xml b/benchmarks/project/pom.xml\\nindex 62030b6..ab87dea 100644\\n--- a/benchmarks/project/pom.xml\\n+++ b/benchmarks/project/pom.xml\\n@@ -123,11 +123,6 @@\\n \\n \\n \\n- com.diffplug.spotless\\n- spotless-maven-plugin\\n- \\n-\\n- \\n org.apache.maven.plugins\\n maven-shade-plugin\\n \\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"f1bc5a554af4e617c7d7508f7f16f8fd25c78c91\", \"fb83ef33b699fd966486a922ba1ade4cf8e55858\", \"37a1b5bbb5270befcee5d9b9621af196c787a61f\", \"7f9721dc9bbf66a3712d59352f64ca089da139f0\"]"},"types":{"kind":"string","value":"[\"refactor\", \"feat\", \"fix\", \"build\"]"}}},{"rowIdx":939,"cells":{"commit_message":{"kind":"string","value":"add title to badge icon,add test for spurious cross join,remove unused,remove unused branches and ignore envrc file"},"diff":{"kind":"string","value":"[\"diff --git a/kibbeh/src/modules/room/chat/RoomChatList.tsx b/kibbeh/src/modules/room/chat/RoomChatList.tsx\\nindex a7418e6..805a9a4 100644\\n--- a/kibbeh/src/modules/room/chat/RoomChatList.tsx\\n+++ b/kibbeh/src/modules/room/chat/RoomChatList.tsx\\n@@ -16,6 +16,11 @@ interface ChatListProps {\\n users: RoomUser[];\\n }\\n \\n+interface BadgeIconData {\\n+ emoji: string,\\n+ title: string\\n+}\\n+\\n export const RoomChatList: React.FC = ({ room, users }) => {\\n const { setData } = useContext(UserPreviewModalContext);\\n const { messages, toggleFrozen } = useRoomChatStore();\\n@@ -48,11 +53,14 @@ export const RoomChatList: React.FC = ({ room, users }) => {\\n const getBadgeIcon = (m: Message) => {\\n const user = users.find((u) => u.id === m.userId);\\n const isSpeaker = room.creatorId === user?.id || user?.roomPermissions?.isSpeaker;\\n- let emoji = null;\\n+ let badgeIconData: BadgeIconData | null = null;\\n if (isSpeaker) {\\n- emoji = \\\"\\ud83d\\udce3\\\";\\n+ badgeIconData = {\\n+ emoji: \\\"\\ud83d\\udce3\\\",\\n+ title: \\\"Speaker\\\"\\n+ };\\n }\\n- return emoji && ;\\n+ return badgeIconData && ;\\n };\\n \\n return (\\n\", \"diff --git a/ibis/tests/sql/test_sqlalchemy.py b/ibis/tests/sql/test_sqlalchemy.py\\nindex 4ad32a6..b2e5d72 100644\\n--- a/ibis/tests/sql/test_sqlalchemy.py\\n+++ b/ibis/tests/sql/test_sqlalchemy.py\\n@@ -841,3 +841,63 @@ def test_filter_group_by_agg_with_same_name():\\n )\\n ex = sa.select([t0]).where(t0.c.bigint_col == 60)\\n _check(expr, ex)\\n+\\n+\\n+@pytest.fixture\\n+def person():\\n+ return ibis.table(\\n+ dict(id=\\\"string\\\", personal=\\\"string\\\", family=\\\"string\\\"),\\n+ name=\\\"person\\\",\\n+ )\\n+\\n+\\n+@pytest.fixture\\n+def visited():\\n+ return ibis.table(\\n+ dict(id=\\\"int32\\\", site=\\\"string\\\", dated=\\\"string\\\"),\\n+ name=\\\"visited\\\",\\n+ )\\n+\\n+\\n+@pytest.fixture\\n+def survey():\\n+ return ibis.table(\\n+ dict(\\n+ taken=\\\"int32\\\",\\n+ person=\\\"string\\\",\\n+ quant=\\\"string\\\",\\n+ reading=\\\"float32\\\",\\n+ ),\\n+ name=\\\"survey\\\",\\n+ )\\n+\\n+\\n+def test_no_cross_join(person, visited, survey):\\n+ expr = person.join(survey, person.id == survey.person).join(\\n+ visited,\\n+ visited.id == survey.taken,\\n+ )\\n+\\n+ context = AlchemyContext(compiler=AlchemyCompiler)\\n+ _ = AlchemyCompiler.to_sql(expr, context)\\n+\\n+ t0 = context.get_ref(person)\\n+ t1 = context.get_ref(survey)\\n+ t2 = context.get_ref(visited)\\n+\\n+ from_ = t0.join(t1, t0.c.id == t1.c.person).join(t2, t2.c.id == t1.c.taken)\\n+ ex = sa.select(\\n+ [\\n+ t0.c.id.label(\\\"id_x\\\"),\\n+ t0.c.personal,\\n+ t0.c.family,\\n+ t1.c.taken,\\n+ t1.c.person,\\n+ t1.c.quant,\\n+ t1.c.reading,\\n+ t2.c.id.label(\\\"id_y\\\"),\\n+ t2.c.site,\\n+ t2.c.dated,\\n+ ]\\n+ ).select_from(from_)\\n+ _check(expr, ex)\\n\", \"diff --git a/src/content/redux/modules/dictionaries.ts b/src/content/redux/modules/dictionaries.ts\\nindex 88f7215..570d397 100644\\n--- a/src/content/redux/modules/dictionaries.ts\\n+++ b/src/content/redux/modules/dictionaries.ts\\n@@ -3,7 +3,6 @@ import { DictID, appConfigFactory, AppConfig } from '@/app-config'\\n import isEqual from 'lodash/isEqual'\\n import { saveWord } from '@/_helpers/record-manager'\\n import { getDefaultSelectionInfo, SelectionInfo, isSameSelection } from '@/_helpers/selection'\\n-import { createActiveConfigStream } from '@/_helpers/config-manager'\\n import { isContainChinese, isContainEnglish, testerPunct, isContainMinor, testerChinese, testJapanese, testKorean } from '@/_helpers/lang-check'\\n import { MsgType, MsgFetchDictResult } from '@/typings/message'\\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\\ndiff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\\nindex 53ad550..68e0a3d 100644\\n--- a/src/content/redux/modules/widget.ts\\n+++ b/src/content/redux/modules/widget.ts\\n@@ -1,9 +1,9 @@\\n import * as recordManager from '@/_helpers/record-manager'\\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\\n-import appConfigFactory, { TCDirection, AppConfig, DictID } from '@/app-config'\\n+import appConfigFactory, { TCDirection, DictID } from '@/app-config'\\n import { message, storage } from '@/_helpers/browser-api'\\n-import { createActiveConfigStream, createConfigIDListStream } from '@/_helpers/config-manager'\\n-import { MsgSelection, MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\\n+import { createConfigIDListStream } from '@/_helpers/config-manager'\\n+import { MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\\n import { searchText, restoreDicts } from '@/content/redux/modules/dictionaries'\\n import { SelectionInfo, getDefaultSelectionInfo } from '@/_helpers/selection'\\n import { Mutable } from '@/typings/helpers'\\n\", \"diff --git a/.github/workflows/ibis-backends-cloud.yml b/.github/workflows/ibis-backends-cloud.yml\\nindex 2003e8e..7c7fd26 100644\\n--- a/.github/workflows/ibis-backends-cloud.yml\\n+++ b/.github/workflows/ibis-backends-cloud.yml\\n@@ -5,9 +5,12 @@ on:\\n # Skip the backend suite if all changes are in the docs directory\\n paths-ignore:\\n - \\\"docs/**\\\"\\n+ - \\\"**/*.md\\\"\\n+ - \\\"**/*.qmd\\\"\\n+ - \\\"codecov.yml\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n- - quarto\\n \\n permissions:\\n # this allows extractions/setup-just to list releases for `just` at a higher\\ndiff --git a/.github/workflows/ibis-backends-skip-helper.yml b/.github/workflows/ibis-backends-skip-helper.yml\\nindex 5d5f3f7..0471994 100644\\n--- a/.github/workflows/ibis-backends-skip-helper.yml\\n+++ b/.github/workflows/ibis-backends-skip-helper.yml\\n@@ -9,20 +9,20 @@ on:\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n - \\\"codecov.yml\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n - \\\"codecov.yml\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n jobs:\\n test_backends:\\ndiff --git a/.github/workflows/ibis-backends.yml b/.github/workflows/ibis-backends.yml\\nindex 4a1cae9..30e6c1a 100644\\n--- a/.github/workflows/ibis-backends.yml\\n+++ b/.github/workflows/ibis-backends.yml\\n@@ -8,10 +8,10 @@ on:\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n - \\\"codecov.yml\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n # Skip the backend suite if all changes are docs\\n paths-ignore:\\n@@ -19,10 +19,10 @@ on:\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n - \\\"codecov.yml\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n \\n permissions:\\ndiff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\\nindex 1adda11..b528a30 100644\\n--- a/.github/workflows/ibis-docs-lint.yml\\n+++ b/.github/workflows/ibis-docs-lint.yml\\n@@ -5,12 +5,10 @@ on:\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n \\n concurrency:\\ndiff --git a/.github/workflows/ibis-main-skip-helper.yml b/.github/workflows/ibis-main-skip-helper.yml\\nindex a5fdc6f..0fb5dea 100644\\n--- a/.github/workflows/ibis-main-skip-helper.yml\\n+++ b/.github/workflows/ibis-main-skip-helper.yml\\n@@ -8,19 +8,19 @@ on:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n jobs:\\n test_core:\\ndiff --git a/.github/workflows/ibis-main.yml b/.github/workflows/ibis-main.yml\\nindex aa31436..0b1536a 100644\\n--- a/.github/workflows/ibis-main.yml\\n+++ b/.github/workflows/ibis-main.yml\\n@@ -7,20 +7,20 @@ on:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n # Skip the test suite if all changes are in the docs directory\\n paths-ignore:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n \\n permissions:\\ndiff --git a/.github/workflows/ibis-tpch-queries-skip-helper.yml b/.github/workflows/ibis-tpch-queries-skip-helper.yml\\nindex 1f1c0bc..f10fb8d 100644\\n--- a/.github/workflows/ibis-tpch-queries-skip-helper.yml\\n+++ b/.github/workflows/ibis-tpch-queries-skip-helper.yml\\n@@ -6,19 +6,19 @@ on:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n \\n concurrency:\\ndiff --git a/.github/workflows/ibis-tpch-queries.yml b/.github/workflows/ibis-tpch-queries.yml\\nindex b4f8a48..9e65a61 100644\\n--- a/.github/workflows/ibis-tpch-queries.yml\\n+++ b/.github/workflows/ibis-tpch-queries.yml\\n@@ -6,19 +6,19 @@ on:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n paths-ignore:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n \\n concurrency:\\ndiff --git a/.github/workflows/nix-skip-helper.yml b/.github/workflows/nix-skip-helper.yml\\nindex 677b4d7..e0ab8f7 100644\\n--- a/.github/workflows/nix-skip-helper.yml\\n+++ b/.github/workflows/nix-skip-helper.yml\\n@@ -9,19 +9,19 @@ on:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n \\n jobs:\\ndiff --git a/.github/workflows/nix.yml b/.github/workflows/nix.yml\\nindex f2dd3f0..7ea9e26 100644\\n--- a/.github/workflows/nix.yml\\n+++ b/.github/workflows/nix.yml\\n@@ -6,19 +6,19 @@ on:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n pull_request:\\n paths-ignore:\\n - \\\"docs/**\\\"\\n - \\\"**/*.md\\\"\\n - \\\"**/*.qmd\\\"\\n+ - \\\".envrc\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n- - quarto\\n merge_group:\\n \\n concurrency:\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"6e5098655e6d9bb13f6423abe780cdf6b50ff13a\", \"8dac3fe5a7a56356ca95547fcf7925bec8d9c1dd\", \"a50b51999015e210918d9c8e95fd4cac347353be\", \"d0c6476df61b9c6ab07b87e1724ea7c5318595bb\"]"},"types":{"kind":"string","value":"[\"feat\", \"test\", \"refactor\", \"ci\"]"}}},{"rowIdx":940,"cells":{"commit_message":{"kind":"string","value":"update sandbox-option.md (#18275)\n\nCo-Authored-By: Mark Lee ,generate terminate end event compatible execution steps part 1\n\nThe random execution tests don't know the concept of flow scopes. This makes it challenging to generate a correct execution path for terminate end events, as they terminate a specific flow scope. Processing should continue as normal once the flow scope has been terminated.\n\nWhilst we don't have flow scopes, we do have execution path segments. These segments don't map 1 to 1 to flow scopes. However, since every flow scope starts a new segment we can use these segments to get the desired behavior.\n\nEach segment must keep track whether is has reached a terminate end event. If this is the case that means that we don't expect any further execution steps. We can isolate this behavior in a single location, during the appending of one segment to another segment.\nIn order to differentiate between flow scopes a new append method has been added which takes the boolean `changesFlowScope` as a parameter. Blockbuilder where the flow scope changes (e.g. SubProcessBlockBuilder) can use this to indicate that even though a terminate end event has been reached. Execution steps after this specific segment still need to added to complete the process.\n\nWhen a segment is appended to a different segment and the flow scope does not change we can use the segment that should be appended to identify whether new segment can still be added to the current segment. If passed segment has reached a terminate end event and the flow scope has not been changed it is guaranteed that the current segment is in the same flow scope has the previous segment and thus has also reached the terminate end event.,update version (v0.6.18),verify process responses for deploy command\n\nTests should generally only fail for 1 reason, but the first test case\n(`shouldDeployResourceFromFile`) verifies multiple unrelated things.\n\nTo align with the other test cases in this class, it makes sense that\nthis test case only verifies that the gateway service was called with a\nspecific request.\n\nWe can extract the verification of the response into a separate test.\nThis can also be applied to the shouldDeployMultipleResources test case."},"diff":{"kind":"string","value":"[\"diff --git a/docs/api/sandbox-option.md b/docs/api/sandbox-option.md\\nindex 7d24bee..e293d34 100644\\n--- a/docs/api/sandbox-option.md\\n+++ b/docs/api/sandbox-option.md\\n@@ -113,8 +113,8 @@ window.open = customWindowOpen\\n Important things to notice in the preload script:\\n \\n - Even though the sandboxed renderer doesn't have Node.js running, it still has\\n- access to a limited node-like environment: `Buffer`, `process`, `setImmediate`\\n- and `require` are available.\\n+ access to a limited node-like environment: `Buffer`, `process`, `setImmediate`,\\n+ `clearImmediate` and `require` are available.\\n - The preload script can indirectly access all APIs from the main process through the\\n `remote` and `ipcRenderer` modules.\\n - The preload script must be contained in a single script, but it is possible to have\\n@@ -162,16 +162,17 @@ feature. We are still not aware of the security implications of exposing some\\n Electron renderer APIs to the preload script, but here are some things to\\n consider before rendering untrusted content:\\n \\n-- A preload script can accidentally leak privileged APIs to untrusted code.\\n+- A preload script can accidentally leak privileged APIs to untrusted code,\\n+ unless [`contextIsolation`](../tutorial/security.md#3-enable-context-isolation-for-remote-content)\\n+ is also enabled.\\n - Some bug in V8 engine may allow malicious code to access the renderer preload\\n APIs, effectively granting full access to the system through the `remote`\\n- module.\\n+ module. Therefore, it is highly recommended to\\n+ [disable the `remote` module](../tutorial/security.md#15-disable-the-remote-module).\\n+ If disabling is not feasible, you should selectively\\n+ [filter the `remote` module](../tutorial/security.md#16-filter-the-remote-module).\\n \\n Since rendering untrusted content in Electron is still uncharted territory,\\n the APIs exposed to the sandbox preload script should be considered more\\n unstable than the rest of Electron APIs, and may have breaking changes to fix\\n security issues.\\n-\\n-One planned enhancement that should greatly increase security is to block IPC\\n-messages from sandboxed renderers by default, allowing the main process to\\n-explicitly define a set of messages the renderer is allowed to send.\\n\", \"diff --git a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\nindex da33c23..23c43be 100644\\n--- a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\n+++ b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\n@@ -29,6 +29,10 @@ import org.apache.commons.lang3.builder.ToStringStyle;\\n */\\n public final class ExecutionPathSegment {\\n \\n+ // If we have reached a terminate end event we want to stop generating execution steps for a\\n+ // specific flow scope. By setting this flag to true no new execution steps will be added for the\\n+ // flow scope this segment is in.\\n+ private boolean reachedTerminateEndEvent = false;\\n private final List scheduledSteps = new ArrayList<>();\\n private final Map variableDefaults = new HashMap<>();\\n \\n@@ -87,10 +91,28 @@ public final class ExecutionPathSegment {\\n new ScheduledExecutionStep(logicalPredecessor, executionPredecessor, executionStep));\\n }\\n \\n+ /**\\n+ * Appends the steps of the passed execution path segment to the current segment.\\n+ *\\n+ * @param pathToAdd execution path segment to append to this segment\\n+ */\\n public void append(final ExecutionPathSegment pathToAdd) {\\n+ append(pathToAdd, false);\\n+ }\\n+\\n+ /**\\n+ * Appends the step of the passed execution path segment to the current segment if the current\\n+ *\\n+ * @param pathToAdd\\n+ * @param changesFlowScope\\n+ */\\n+ public void append(final ExecutionPathSegment pathToAdd, final boolean changesFlowScope) {\\n mergeVariableDefaults(pathToAdd);\\n \\n- pathToAdd.getScheduledSteps().forEach(this::append);\\n+ if (!hasReachedTerminateEndEvent() || changesFlowScope) {\\n+ pathToAdd.getScheduledSteps().forEach(this::append);\\n+ }\\n+ reachedTerminateEndEvent = pathToAdd.hasReachedTerminateEndEvent() && !changesFlowScope;\\n }\\n \\n public void append(final ScheduledExecutionStep scheduledExecutionStep) {\\n@@ -259,6 +281,14 @@ public final class ExecutionPathSegment {\\n return ToStringBuilder.reflectionToString(this, ToStringStyle.SHORT_PREFIX_STYLE);\\n }\\n \\n+ public boolean hasReachedTerminateEndEvent() {\\n+ return reachedTerminateEndEvent;\\n+ }\\n+\\n+ public void setReachedTerminateEndEvent(final boolean reachedTerminateEndEvent) {\\n+ this.reachedTerminateEndEvent = reachedTerminateEndEvent;\\n+ }\\n+\\n /**\\n * An execution boundary is the point where automatic and non-automatic {@link\\n * ScheduledExecutionStep}'s meet each other. This class contains information about the existing\\n\", \"diff --git a/Cargo.lock b/Cargo.lock\\nindex c32d8b4..599790e 100644\\n--- a/Cargo.lock\\n+++ b/Cargo.lock\\n@@ -94,7 +94,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"els\\\"\\n-version = \\\"0.1.30-nightly.2\\\"\\n+version = \\\"0.1.30\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_compiler\\\",\\n@@ -105,7 +105,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg\\\"\\n-version = \\\"0.6.18-nightly.2\\\"\\n+version = \\\"0.6.18\\\"\\n dependencies = [\\n \\\"els\\\",\\n \\\"erg_common\\\",\\n@@ -115,7 +115,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_common\\\"\\n-version = \\\"0.6.18-nightly.2\\\"\\n+version = \\\"0.6.18\\\"\\n dependencies = [\\n \\\"backtrace-on-stack-overflow\\\",\\n \\\"crossterm\\\",\\n@@ -125,7 +125,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_compiler\\\"\\n-version = \\\"0.6.18-nightly.2\\\"\\n+version = \\\"0.6.18\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_parser\\\",\\n@@ -133,7 +133,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_parser\\\"\\n-version = \\\"0.6.18-nightly.2\\\"\\n+version = \\\"0.6.18\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"unicode-xid\\\",\\ndiff --git a/Cargo.toml b/Cargo.toml\\nindex baaa0ac..5082cd3 100644\\n--- a/Cargo.toml\\n+++ b/Cargo.toml\\n@@ -20,7 +20,7 @@ members = [\\n ]\\n \\n [workspace.package]\\n-version = \\\"0.6.18-nightly.2\\\"\\n+version = \\\"0.6.18\\\"\\n authors = [\\\"erg-lang team \\\"]\\n license = \\\"MIT OR Apache-2.0\\\"\\n edition = \\\"2021\\\"\\n@@ -64,10 +64,10 @@ full = [\\\"els\\\", \\\"full-repl\\\", \\\"unicode\\\", \\\"pretty\\\"]\\n experimental = [\\\"erg_common/experimental\\\", \\\"erg_parser/experimental\\\", \\\"erg_compiler/experimental\\\"]\\n \\n [workspace.dependencies]\\n-erg_common = { version = \\\"0.6.18-nightly.2\\\", path = \\\"./crates/erg_common\\\" }\\n-erg_parser = { version = \\\"0.6.18-nightly.2\\\", path = \\\"./crates/erg_parser\\\" }\\n-erg_compiler = { version = \\\"0.6.18-nightly.2\\\", path = \\\"./crates/erg_compiler\\\" }\\n-els = { version = \\\"0.1.30-nightly.2\\\", path = \\\"./crates/els\\\" }\\n+erg_common = { version = \\\"0.6.18\\\", path = \\\"./crates/erg_common\\\" }\\n+erg_parser = { version = \\\"0.6.18\\\", path = \\\"./crates/erg_parser\\\" }\\n+erg_compiler = { version = \\\"0.6.18\\\", path = \\\"./crates/erg_compiler\\\" }\\n+els = { version = \\\"0.1.30\\\", path = \\\"./crates/els\\\" }\\n \\n [dependencies]\\n erg_common = { workspace = true }\\ndiff --git a/crates/els/Cargo.toml b/crates/els/Cargo.toml\\nindex 3efbf4e..9f902fa 100644\\n--- a/crates/els/Cargo.toml\\n+++ b/crates/els/Cargo.toml\\n@@ -2,7 +2,7 @@\\n name = \\\"els\\\"\\n description = \\\"An Erg compiler frontend for IDEs, implements LSP.\\\"\\n documentation = \\\"http://docs.rs/els\\\"\\n-version = \\\"0.1.30-nightly.2\\\"\\n+version = \\\"0.1.30\\\"\\n authors.workspace = true\\n license.workspace = true\\n edition.workspace = true\\n\", \"diff --git a/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java b/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\\nindex 1d96c74..b65d9f3 100644\\n--- a/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\\n+++ b/clients/java/src/test/java/io/camunda/zeebe/client/process/DeployResourceTest.java\\n@@ -22,7 +22,6 @@ import static org.assertj.core.api.Assertions.assertThatThrownBy;\\n \\n import io.camunda.zeebe.client.api.command.ClientException;\\n import io.camunda.zeebe.client.api.response.DeploymentEvent;\\n-import io.camunda.zeebe.client.api.response.Process;\\n import io.camunda.zeebe.client.impl.command.StreamUtil;\\n import io.camunda.zeebe.client.impl.response.ProcessImpl;\\n import io.camunda.zeebe.client.util.ClientTest;\\n@@ -35,7 +34,6 @@ import java.io.IOException;\\n import java.io.InputStream;\\n import java.nio.charset.StandardCharsets;\\n import java.time.Duration;\\n-import java.util.List;\\n import org.junit.Test;\\n \\n public final class DeployResourceTest extends ClientTest {\\n@@ -49,25 +47,15 @@ public final class DeployResourceTest extends ClientTest {\\n @Test\\n public void shouldDeployResourceFromFile() {\\n // given\\n- final long key = 123L;\\n- final String filename = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\\n- gatewayService.onDeployResourceRequest(\\n- key, deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 12, 423, filename)));\\n- final Process expected = new ProcessImpl(423, BPMN_1_PROCESS_ID, 12, filename);\\n+ final String path = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\\n \\n // when\\n- final DeploymentEvent response =\\n- client.newDeployCommand().addResourceFile(filename).send().join();\\n+ client.newDeployCommand().addResourceFile(path).send().join();\\n \\n // then\\n- assertThat(response.getKey()).isEqualTo(key);\\n-\\n- final List processes = response.getProcesses();\\n- assertThat(processes).containsOnly(expected);\\n-\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n- assertThat(resource.getName()).isEqualTo(filename);\\n+ assertThat(resource.getName()).isEqualTo(path);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n \\n@@ -114,7 +102,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n@@ -135,7 +122,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n@@ -152,7 +138,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n }\\n@@ -174,7 +159,6 @@ public final class DeployResourceTest extends ClientTest {\\n // then\\n final DeployResourceRequest request = gatewayService.getLastRequest();\\n final Resource resource = request.getResources(0);\\n-\\n assertThat(resource.getName()).isEqualTo(filename);\\n assertThat(resource.getContent().toByteArray()).isEqualTo(expectedBytes);\\n }\\n@@ -183,13 +167,58 @@ public final class DeployResourceTest extends ClientTest {\\n public void shouldDeployMultipleResources() {\\n // given\\n final long key = 345L;\\n-\\n final String filename1 = BPMN_1_FILENAME.substring(1);\\n final String filename2 = BPMN_2_FILENAME.substring(1);\\n+ gatewayService.onDeployResourceRequest(\\n+ key,\\n+ deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 1, 1, filename1)),\\n+ deployedResource(deployedProcess(BPMN_2_PROCESS_ID, 1, 2, filename2)));\\n \\n- final Process expected1 = new ProcessImpl(1, BPMN_1_PROCESS_ID, 1, filename1);\\n- final Process expected2 = new ProcessImpl(2, BPMN_2_PROCESS_ID, 1, filename2);\\n+ // when\\n+ client\\n+ .newDeployCommand()\\n+ .addResourceFromClasspath(filename1)\\n+ .addResourceFromClasspath(filename2)\\n+ .send()\\n+ .join();\\n \\n+ // then\\n+ final DeployResourceRequest request = gatewayService.getLastRequest();\\n+ assertThat(request.getResourcesList()).hasSize(2);\\n+\\n+ final Resource resource1 = request.getResources(0);\\n+ assertThat(resource1.getName()).isEqualTo(filename1);\\n+ assertThat(resource1.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n+\\n+ final Resource resource2 = request.getResources(1);\\n+ assertThat(resource2.getName()).isEqualTo(filename2);\\n+ assertThat(resource2.getContent().toByteArray()).isEqualTo(getBytes(BPMN_2_FILENAME));\\n+ }\\n+\\n+ @Test\\n+ public void shouldDeployProcessAsResource() {\\n+ // given\\n+ final long key = 123L;\\n+ final String filename = DeployResourceTest.class.getResource(BPMN_1_FILENAME).getPath();\\n+ gatewayService.onDeployResourceRequest(\\n+ key, deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 12, 423, filename)));\\n+\\n+ // when\\n+ final DeploymentEvent response =\\n+ client.newDeployCommand().addResourceFile(filename).send().join();\\n+\\n+ // then\\n+ assertThat(response.getKey()).isEqualTo(key);\\n+ assertThat(response.getProcesses())\\n+ .containsExactly(new ProcessImpl(423, BPMN_1_PROCESS_ID, 12, filename));\\n+ }\\n+\\n+ @Test\\n+ public void shouldDeployMultipleProcessesAsResources() {\\n+ // given\\n+ final long key = 345L;\\n+ final String filename1 = BPMN_1_FILENAME.substring(1);\\n+ final String filename2 = BPMN_2_FILENAME.substring(1);\\n gatewayService.onDeployResourceRequest(\\n key,\\n deployedResource(deployedProcess(BPMN_1_PROCESS_ID, 1, 1, filename1)),\\n@@ -206,21 +235,10 @@ public final class DeployResourceTest extends ClientTest {\\n \\n // then\\n assertThat(response.getKey()).isEqualTo(key);\\n-\\n- final List processes = response.getProcesses();\\n- assertThat(processes).containsOnly(expected1, expected2);\\n-\\n- final DeployResourceRequest request = gatewayService.getLastRequest();\\n- assertThat(request.getResourcesList()).hasSize(2);\\n-\\n- Resource resource = request.getResources(0);\\n-\\n- assertThat(resource.getName()).isEqualTo(filename1);\\n- assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_1_FILENAME));\\n-\\n- resource = request.getResources(1);\\n- assertThat(resource.getName()).isEqualTo(filename2);\\n- assertThat(resource.getContent().toByteArray()).isEqualTo(getBytes(BPMN_2_FILENAME));\\n+ assertThat(response.getProcesses())\\n+ .containsExactly(\\n+ new ProcessImpl(1, BPMN_1_PROCESS_ID, 1, filename1),\\n+ new ProcessImpl(2, BPMN_2_PROCESS_ID, 1, filename2));\\n }\\n \\n @Test\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"dbb8617214aaa8b56b827deef1265d9ee38765bd\", \"40597fb4de41c7194eb99479a914db70da7909ea\", \"bb3e3d9b96e435c3b92fc208bca93d1ad7e1ad50\", \"390eadc270d027493722cdbe9c8f4140d027e473\"]"},"types":{"kind":"string","value":"[\"docs\", \"feat\", \"build\", \"test\"]"}}},{"rowIdx":941,"cells":{"commit_message":{"kind":"string","value":"add title to badge icon,expose the means by which we process each reward cycle's affirmation maps at reward cycle boundaries,do not pin time in tests but only skip ahead\n\nrelated to #573,handle default_branch_monthly_cost having no cost\n\nCloses https://github.com/infracost/infracost-gh-action/issues/17"},"diff":{"kind":"string","value":"[\"diff --git a/kibbeh/src/modules/room/chat/RoomChatList.tsx b/kibbeh/src/modules/room/chat/RoomChatList.tsx\\nindex a7418e6..805a9a4 100644\\n--- a/kibbeh/src/modules/room/chat/RoomChatList.tsx\\n+++ b/kibbeh/src/modules/room/chat/RoomChatList.tsx\\n@@ -16,6 +16,11 @@ interface ChatListProps {\\n users: RoomUser[];\\n }\\n \\n+interface BadgeIconData {\\n+ emoji: string,\\n+ title: string\\n+}\\n+\\n export const RoomChatList: React.FC = ({ room, users }) => {\\n const { setData } = useContext(UserPreviewModalContext);\\n const { messages, toggleFrozen } = useRoomChatStore();\\n@@ -48,11 +53,14 @@ export const RoomChatList: React.FC = ({ room, users }) => {\\n const getBadgeIcon = (m: Message) => {\\n const user = users.find((u) => u.id === m.userId);\\n const isSpeaker = room.creatorId === user?.id || user?.roomPermissions?.isSpeaker;\\n- let emoji = null;\\n+ let badgeIconData: BadgeIconData | null = null;\\n if (isSpeaker) {\\n- emoji = \\\"\\ud83d\\udce3\\\";\\n+ badgeIconData = {\\n+ emoji: \\\"\\ud83d\\udce3\\\",\\n+ title: \\\"Speaker\\\"\\n+ };\\n }\\n- return emoji && ;\\n+ return badgeIconData && ;\\n };\\n \\n return (\\n\", \"diff --git a/src/burnchains/burnchain.rs b/src/burnchains/burnchain.rs\\nindex 92105d6..60c608a 100644\\n--- a/src/burnchains/burnchain.rs\\n+++ b/src/burnchains/burnchain.rs\\n@@ -851,8 +851,26 @@ impl Burnchain {\\n );\\n \\n burnchain_db.store_new_burnchain_block(burnchain, indexer, &block)?;\\n- let block_height = block.block_height();\\n+ Burnchain::process_affirmation_maps(\\n+ burnchain,\\n+ burnchain_db,\\n+ indexer,\\n+ block.block_height(),\\n+ )?;\\n+\\n+ let header = block.header();\\n+ Ok(header)\\n+ }\\n \\n+ /// Update the affirmation maps for the previous reward cycle's commits.\\n+ /// This is a no-op unless the given burnchain block height falls on a reward cycle boundary. In that\\n+ /// case, the previous reward cycle's block commits' affirmation maps are all re-calculated.\\n+ pub fn process_affirmation_maps(\\n+ burnchain: &Burnchain,\\n+ burnchain_db: &mut BurnchainDB,\\n+ indexer: &B,\\n+ block_height: u64,\\n+ ) -> Result<(), burnchain_error> {\\n let this_reward_cycle = burnchain\\n .block_height_to_reward_cycle(block_height)\\n .unwrap_or(0);\\n@@ -872,10 +890,7 @@ impl Burnchain {\\n );\\n update_pox_affirmation_maps(burnchain_db, indexer, prev_reward_cycle, burnchain)?;\\n }\\n-\\n- let header = block.header();\\n-\\n- Ok(header)\\n+ Ok(())\\n }\\n \\n /// Hand off the block to the ChainsCoordinator _and_ process the sortition\\n\", \"diff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\nindex 636cd21..76afff7 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n@@ -15,7 +15,9 @@\\n */\\n package io.zeebe.broker.it.startup;\\n \\n-import static io.zeebe.broker.it.util.TopicEventRecorder.*;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.incidentEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.taskEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.wfInstanceEvent;\\n import static io.zeebe.test.util.TestUtil.doRepeatedly;\\n import static io.zeebe.test.util.TestUtil.waitUntil;\\n import static org.assertj.core.api.Assertions.assertThat;\\n@@ -24,11 +26,18 @@ import java.io.File;\\n import java.io.InputStream;\\n import java.nio.charset.StandardCharsets;\\n import java.time.Duration;\\n-import java.time.Instant;\\n import java.util.Collections;\\n import java.util.List;\\n import java.util.regex.Pattern;\\n \\n+import org.assertj.core.util.Files;\\n+import org.junit.After;\\n+import org.junit.Rule;\\n+import org.junit.Test;\\n+import org.junit.rules.ExpectedException;\\n+import org.junit.rules.RuleChain;\\n+import org.junit.rules.TemporaryFolder;\\n+\\n import io.zeebe.broker.clustering.ClusterServiceNames;\\n import io.zeebe.broker.it.ClientRule;\\n import io.zeebe.broker.it.EmbeddedBrokerRule;\\n@@ -38,7 +47,9 @@ import io.zeebe.client.ZeebeClient;\\n import io.zeebe.client.clustering.impl.TopicLeader;\\n import io.zeebe.client.clustering.impl.TopologyResponse;\\n import io.zeebe.client.cmd.ClientCommandRejectedException;\\n-import io.zeebe.client.event.*;\\n+import io.zeebe.client.event.DeploymentEvent;\\n+import io.zeebe.client.event.TaskEvent;\\n+import io.zeebe.client.event.WorkflowInstanceEvent;\\n import io.zeebe.model.bpmn.Bpmn;\\n import io.zeebe.model.bpmn.instance.WorkflowDefinition;\\n import io.zeebe.raft.Raft;\\n@@ -48,9 +59,6 @@ import io.zeebe.test.util.TestFileUtil;\\n import io.zeebe.test.util.TestUtil;\\n import io.zeebe.transport.SocketAddress;\\n import io.zeebe.util.time.ClockUtil;\\n-import org.assertj.core.util.Files;\\n-import org.junit.*;\\n-import org.junit.rules.*;\\n \\n public class BrokerRecoveryTest\\n {\\n@@ -360,17 +368,12 @@ public class BrokerRecoveryTest\\n waitUntil(() -> !recordingTaskHandler.getHandledTasks().isEmpty());\\n \\n // when\\n- restartBroker(() ->\\n- {\\n- final Instant now = ClockUtil.getCurrentTime();\\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\\n- });\\n+ restartBroker(() -> ClockUtil.addTime(Duration.ofSeconds(60)));\\n \\n // wait until stream processor and scheduler process the lock task event which is not re-processed on recovery\\n doRepeatedly(() ->\\n {\\n- final Instant now = ClockUtil.getCurrentTime();\\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\\n+ ClockUtil.addTime(Duration.ofSeconds(60)); // retriggers lock expiration check in broker\\n return null;\\n }).until(t -> eventRecorder.hasTaskEvent(taskEvent(\\\"LOCK_EXPIRED\\\")));\\n \\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\\nindex 5ff1301..0ffe98d 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\\n@@ -15,7 +15,9 @@\\n */\\n package io.zeebe.broker.it.startup;\\n \\n-import static io.zeebe.broker.it.util.TopicEventRecorder.*;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.incidentEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.taskEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.wfInstanceEvent;\\n import static io.zeebe.test.util.TestUtil.waitUntil;\\n import static org.assertj.core.api.Assertions.assertThat;\\n \\n@@ -23,11 +25,18 @@ import java.io.File;\\n import java.io.InputStream;\\n import java.nio.charset.StandardCharsets;\\n import java.time.Duration;\\n-import java.time.Instant;\\n import java.util.Collections;\\n import java.util.List;\\n import java.util.regex.Pattern;\\n \\n+import org.junit.After;\\n+import org.junit.Rule;\\n+import org.junit.Test;\\n+import org.junit.experimental.categories.Category;\\n+import org.junit.rules.ExpectedException;\\n+import org.junit.rules.RuleChain;\\n+import org.junit.rules.TemporaryFolder;\\n+\\n import io.zeebe.broker.clustering.ClusterServiceNames;\\n import io.zeebe.broker.it.ClientRule;\\n import io.zeebe.broker.it.EmbeddedBrokerRule;\\n@@ -37,7 +46,9 @@ import io.zeebe.client.ZeebeClient;\\n import io.zeebe.client.clustering.impl.TopicLeader;\\n import io.zeebe.client.clustering.impl.TopologyResponse;\\n import io.zeebe.client.cmd.ClientCommandRejectedException;\\n-import io.zeebe.client.event.*;\\n+import io.zeebe.client.event.DeploymentEvent;\\n+import io.zeebe.client.event.TaskEvent;\\n+import io.zeebe.client.event.WorkflowInstanceEvent;\\n import io.zeebe.model.bpmn.Bpmn;\\n import io.zeebe.model.bpmn.instance.WorkflowDefinition;\\n import io.zeebe.raft.Raft;\\n@@ -47,9 +58,6 @@ import io.zeebe.test.util.TestFileUtil;\\n import io.zeebe.test.util.TestUtil;\\n import io.zeebe.transport.SocketAddress;\\n import io.zeebe.util.time.ClockUtil;\\n-import org.junit.*;\\n-import org.junit.experimental.categories.Category;\\n-import org.junit.rules.*;\\n \\n public class BrokerRestartTest\\n {\\n@@ -360,11 +368,7 @@ public class BrokerRestartTest\\n waitUntil(() -> !recordingTaskHandler.getHandledTasks().isEmpty());\\n \\n // when\\n- restartBroker(() ->\\n- {\\n- final Instant now = ClockUtil.getCurrentTime();\\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\\n- });\\n+ restartBroker(() -> ClockUtil.addTime(Duration.ofSeconds(60)));\\n \\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"LOCK_EXPIRED\\\")));\\n recordingTaskHandler.clear();\\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\\nindex 49b527d..a322fbe 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\\n@@ -353,7 +353,7 @@ public class TaskSubscriptionTest\\n waitUntil(() -> taskHandler.getHandledTasks().size() == 1);\\n \\n // when\\n- ClockUtil.setCurrentTime(Instant.now().plus(Duration.ofMinutes(5)));\\n+ ClockUtil.addTime(Duration.ofMinutes(5));\\n \\n // then\\n waitUntil(() -> taskHandler.getHandledTasks().size() == 2);\\n\", \"diff --git a/scripts/ci/diff.sh b/scripts/ci/diff.sh\\nindex 7472273..fa48e4b 100755\\n--- a/scripts/ci/diff.sh\\n+++ b/scripts/ci/diff.sh\\n@@ -112,7 +112,12 @@ echo \\\"$default_branch_output\\\" > default_branch_infracost.txt\\n default_branch_monthly_cost=$(cat default_branch_infracost.txt | awk '/OVERALL TOTAL/ { gsub(\\\",\\\",\\\"\\\"); printf(\\\"%.2f\\\",$NF) }')\\n echo \\\"::set-output name=default_branch_monthly_cost::$default_branch_monthly_cost\\\"\\n \\n-percent_diff=$(echo \\\"scale=4; $current_branch_monthly_cost / $default_branch_monthly_cost * 100 - 100\\\" | bc)\\n+if [ $(echo \\\"$default_branch_monthly_cost > 0\\\" | bc -l) = 1 ]; then\\n+ percent_diff=$(echo \\\"scale=4; $current_branch_monthly_cost / $default_branch_monthly_cost * 100 - 100\\\" | bc)\\n+else\\n+ echo \\\"Default branch has no cost, setting percent_diff=100 to force a comment\\\"\\n+ percent_diff=100\\n+fi\\n absolute_percent_diff=$(echo $percent_diff | tr -d -)\\n \\n if [ $(echo \\\"$absolute_percent_diff > $percentage_threshold\\\" | bc -l) = 1 ]; then\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"6e5098655e6d9bb13f6423abe780cdf6b50ff13a\", \"d7972da833257c073403dec3c2ac3a7f297e328a\", \"7ece3a9a16780dc6c633bbd903d36ce0aefd6a8a\", \"9474f58b44a35321e9157ca9890c589a7b3729b2\"]"},"types":{"kind":"string","value":"[\"feat\", \"refactor\", \"test\", \"fix\"]"}}},{"rowIdx":942,"cells":{"commit_message":{"kind":"string","value":"fix deploy,set first-attempt to 5s and subsequent-attempt to 180s by default,filters for Rating\n\nSigned-off-by: Raju Udava <86527202+dstala@users.noreply.github.com>,refactor generate_completion"},"diff":{"kind":"string","value":"[\"diff --git a/.github/workflows/deploy.yaml b/.github/workflows/deploy.yaml\\nindex 3830f4c..3b14ee5 100644\\n--- a/.github/workflows/deploy.yaml\\n+++ b/.github/workflows/deploy.yaml\\n@@ -67,7 +67,7 @@ jobs:\\n run: aws s3 cp .next/static s3://cdn.rs.school/_next/static/ --recursive --cache-control \\\"public,max-age=15552000,immutable\\\"\\n \\n - name: Build container\\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-client:master .\\n \\n - name: Login to GitHub Container Registry\\n uses: docker/login-action@v1\\n@@ -117,7 +117,7 @@ jobs:\\n run: npm run build\\n \\n - name: Build container\\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-server:master .\\n \\n - name: Login to GitHub Container Registry\\n uses: docker/login-action@v1\\n@@ -167,7 +167,7 @@ jobs:\\n run: npm run build\\n \\n - name: Build container\\n- run: docker buildx build --platform linux/amd64,linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\\n+ run: docker buildx build --platform linux/arm64 -t ghcr.io/rolling-scopes/rsschool-app-nestjs:master .\\n \\n - name: Login to GitHub Container Registry\\n uses: docker/login-action@v1\\n\", \"diff --git a/testnet/stacks-node/src/config.rs b/testnet/stacks-node/src/config.rs\\nindex 24ca06c..d80f721 100644\\n--- a/testnet/stacks-node/src/config.rs\\n+++ b/testnet/stacks-node/src/config.rs\\n@@ -1414,8 +1414,8 @@ impl MinerConfig {\\n pub fn default() -> MinerConfig {\\n MinerConfig {\\n min_tx_fee: 1,\\n- first_attempt_time_ms: 1_000,\\n- subsequent_attempt_time_ms: 30_000,\\n+ first_attempt_time_ms: 5_000,\\n+ subsequent_attempt_time_ms: 180_000,\\n microblock_attempt_time_ms: 30_000,\\n probability_pick_no_estimate_tx: 5,\\n }\\n\", \"diff --git a/tests/playwright/pages/Dashboard/common/Toolbar/Filter.ts b/tests/playwright/pages/Dashboard/common/Toolbar/Filter.ts\\nindex 1a626fa..b82e7f6 100644\\n--- a/tests/playwright/pages/Dashboard/common/Toolbar/Filter.ts\\n+++ b/tests/playwright/pages/Dashboard/common/Toolbar/Filter.ts\\n@@ -1,6 +1,7 @@\\n import { expect } from '@playwright/test';\\n import BasePage from '../../../Base';\\n import { ToolbarPage } from './index';\\n+import { UITypes } from 'nocodb-sdk';\\n \\n export class ToolbarFilterPage extends BasePage {\\n readonly toolbar: ToolbarPage;\\n@@ -33,11 +34,13 @@ export class ToolbarFilterPage extends BasePage {\\n opType,\\n value,\\n isLocallySaved,\\n+ dataType,\\n }: {\\n columnTitle: string;\\n opType: string;\\n value?: string;\\n isLocallySaved: boolean;\\n+ dataType?: string;\\n }) {\\n await this.get().locator(`button:has-text(\\\"Add Filter\\\")`).first().click();\\n \\n@@ -86,14 +89,25 @@ export class ToolbarFilterPage extends BasePage {\\n \\n // if value field was provided, fill it\\n if (value) {\\n- const fillFilter = this.rootPage.locator('.nc-filter-value-select > input').last().fill(value);\\n- await this.waitForResponse({\\n- uiAction: fillFilter,\\n- httpMethodsToMatch: ['GET'],\\n- requestUrlPathToMatch: isLocallySaved ? `/api/v1/db/public/` : `/api/v1/db/data/noco/`,\\n- });\\n- await this.toolbar.parent.dashboard.waitForLoaderToDisappear();\\n- await this.toolbar.parent.waitLoading();\\n+ let fillFilter: any = null;\\n+ switch (dataType) {\\n+ case UITypes.Rating:\\n+ await this.get('.nc-filter-value-select')\\n+ .locator('.ant-rate-star > div')\\n+ .nth(parseInt(value) - 1)\\n+ .click();\\n+ break;\\n+ default:\\n+ fillFilter = this.rootPage.locator('.nc-filter-value-select > input').last().fill(value);\\n+ await this.waitForResponse({\\n+ uiAction: fillFilter,\\n+ httpMethodsToMatch: ['GET'],\\n+ requestUrlPathToMatch: isLocallySaved ? `/api/v1/db/public/` : `/api/v1/db/data/noco/`,\\n+ });\\n+ await this.toolbar.parent.dashboard.waitForLoaderToDisappear();\\n+ await this.toolbar.parent.waitLoading();\\n+ break;\\n+ }\\n }\\n }\\n \\ndiff --git a/tests/playwright/tests/filters.spec.ts b/tests/playwright/tests/filters.spec.ts\\nindex 774a70a..48d949a 100644\\n--- a/tests/playwright/tests/filters.spec.ts\\n+++ b/tests/playwright/tests/filters.spec.ts\\n@@ -36,7 +36,13 @@ async function validateRowArray(param) {\\n // }\\n }\\n \\n-async function verifyFilter(param: { column: string; opType: string; value?: string; result: { rowCount: number } }) {\\n+async function verifyFilter(param: {\\n+ column: string;\\n+ opType: string;\\n+ value?: string;\\n+ result: { rowCount: number };\\n+ dataType?: string;\\n+}) {\\n // if opType was included in skip list, skip it\\n if (skipList[param.column]?.includes(param.opType)) {\\n return;\\n@@ -48,6 +54,7 @@ async function verifyFilter(param: { column: string; opType: string; value?: str\\n opType: param.opType,\\n value: param.value,\\n isLocallySaved: false,\\n+ dataType: param?.dataType,\\n });\\n await toolbar.clickFilter();\\n \\n@@ -414,4 +421,74 @@ test.describe('Filter Tests: Numerical', () => {\\n });\\n }\\n });\\n+\\n+ test('Filter: Rating', async () => {\\n+ // close 'Team & Auth' tab\\n+ await dashboard.closeTab({ title: 'Team & Auth' });\\n+ await dashboard.treeView.openTable({ title: 'numberBased' });\\n+ const dataType = 'Rating';\\n+\\n+ const filterList = [\\n+ {\\n+ op: '=',\\n+ value: '3',\\n+ rowCount: records.list.filter(r => r[dataType] === 3).length,\\n+ },\\n+ {\\n+ op: '!=',\\n+ value: '3',\\n+ rowCount: records.list.filter(r => r[dataType] !== 3).length,\\n+ },\\n+ {\\n+ op: 'is null',\\n+ value: '',\\n+ rowCount: records.list.filter(r => r[dataType] === null).length,\\n+ },\\n+ {\\n+ op: 'is not null',\\n+ value: '',\\n+ rowCount: records.list.filter(r => r[dataType] !== null).length,\\n+ },\\n+ {\\n+ op: 'is blank',\\n+ value: '',\\n+ rowCount: records.list.filter(r => r[dataType] === null).length,\\n+ },\\n+ {\\n+ op: 'is not blank',\\n+ value: '',\\n+ rowCount: records.list.filter(r => r[dataType] !== null).length,\\n+ },\\n+ {\\n+ op: '>',\\n+ value: '2',\\n+ rowCount: records.list.filter(r => r[dataType] > 2 && r[dataType] != null).length,\\n+ },\\n+ {\\n+ op: '>=',\\n+ value: '2',\\n+ rowCount: records.list.filter(r => r[dataType] >= 2 && r[dataType] != null).length,\\n+ },\\n+ {\\n+ op: '<',\\n+ value: '2',\\n+ rowCount: records.list.filter(r => r[dataType] < 2 && r[dataType] != null).length,\\n+ },\\n+ {\\n+ op: '<=',\\n+ value: '2',\\n+ rowCount: records.list.filter(r => r[dataType] <= 2 && r[dataType] != null).length,\\n+ },\\n+ ];\\n+\\n+ for (let i = 0; i < filterList.length; i++) {\\n+ await verifyFilter({\\n+ column: dataType,\\n+ opType: filterList[i].op,\\n+ value: filterList[i].value,\\n+ result: { rowCount: filterList[i].rowCount },\\n+ dataType: dataType,\\n+ });\\n+ }\\n+ });\\n });\\n\", \"diff --git a/src/lib.rs b/src/lib.rs\\nindex dfd8014..15850f7 100644\\n--- a/src/lib.rs\\n+++ b/src/lib.rs\\n@@ -1,11 +1,106 @@\\n //! Generates [Nushell](https://github.com/nushell/nushell) completions for [`clap`](https://github.com/clap-rs/clap) based CLIs\\n \\n-use clap::Command;\\n+use clap::{Arg, Command};\\n use clap_complete::Generator;\\n \\n /// Generate Nushell complete file\\n pub struct Nushell;\\n \\n+enum Argument {\\n+ Short(char),\\n+ Long(String),\\n+ ShortAndLong(char, String),\\n+ Positional(String, bool),\\n+}\\n+\\n+struct ArgumentLine {\\n+ arg: Argument,\\n+ takes_values: bool,\\n+ help: Option,\\n+}\\n+\\n+impl From<&Arg> for ArgumentLine {\\n+ fn from(arg: &Arg) -> Self {\\n+ let takes_values = arg\\n+ .get_num_args()\\n+ .map(|v| v.takes_values())\\n+ .unwrap_or(false);\\n+\\n+ let help = arg.get_help().map(|s| s.to_string());\\n+\\n+ if arg.is_positional() {\\n+ let id = arg.get_id().to_string();\\n+ let required = arg.is_required_set();\\n+ let arg = Argument::Positional(id, required);\\n+\\n+ return Self {\\n+ arg,\\n+ takes_values,\\n+ help,\\n+ };\\n+ }\\n+\\n+ let short = arg.get_short();\\n+ let long = arg.get_long();\\n+\\n+ match short {\\n+ Some(short) => match long {\\n+ Some(long) => Self {\\n+ arg: Argument::ShortAndLong(short, long.into()),\\n+ takes_values,\\n+ help,\\n+ },\\n+ None => Self {\\n+ arg: Argument::Short(short),\\n+ takes_values,\\n+ help,\\n+ },\\n+ },\\n+ None => match long {\\n+ Some(long) => Self {\\n+ arg: Argument::Long(long.into()),\\n+ takes_values,\\n+ help,\\n+ },\\n+ None => unreachable!(\\\"No short or long option found\\\"),\\n+ },\\n+ }\\n+ }\\n+}\\n+\\n+impl ToString for ArgumentLine {\\n+ fn to_string(&self) -> String {\\n+ let mut s = String::new();\\n+\\n+ match &self.arg {\\n+ Argument::Short(short) => s.push_str(format!(\\\" -{}\\\", short).as_str()),\\n+ Argument::Long(long) => s.push_str(format!(\\\" --{}\\\", long).as_str()),\\n+ Argument::ShortAndLong(short, long) => {\\n+ s.push_str(format!(\\\" --{}(-{})\\\", long, short).as_str())\\n+ }\\n+ Argument::Positional(positional, required) => {\\n+ s.push_str(format!(\\\" {}\\\", positional).as_str());\\n+\\n+ if !*required {\\n+ s.push('?');\\n+ }\\n+ }\\n+ }\\n+\\n+ if self.takes_values {\\n+ s.push_str(\\\": string\\\");\\n+ }\\n+\\n+ if let Some(help) = &self.help {\\n+ s.push_str(format!(\\\"\\\\t# {}\\\", help).as_str());\\n+ }\\n+\\n+ s.push('\\\\n');\\n+\\n+ s\\n+ }\\n+}\\n+\\n impl Generator for Nushell {\\n fn file_name(&self, name: &str) -> String {\\n format!(\\\"{}.nu\\\", name)\\n@@ -37,51 +132,18 @@ fn generate_completion(completions: &mut String, cmd: &Command, is_subcommand: b\\n \\n let bin_name = cmd.get_bin_name().expect(\\\"Failed to get bin name\\\");\\n \\n- if is_subcommand {\\n- completions.push_str(format!(\\\" export extern \\\\\\\"{}\\\\\\\" [\\\\n\\\", bin_name).as_str());\\n+ let name = if is_subcommand {\\n+ format!(r#\\\"\\\"{}\\\"\\\"#, bin_name)\\n } else {\\n- completions.push_str(format!(\\\" export extern {} [\\\\n\\\", bin_name).as_str());\\n- }\\n+ bin_name.into()\\n+ };\\n \\n- let mut s = String::new();\\n- for arg in cmd.get_arguments() {\\n- if arg.is_positional() {\\n- s.push_str(format!(\\\" {}\\\", arg.get_id()).as_str());\\n- if !arg.is_required_set() {\\n- s.push('?');\\n- }\\n- }\\n-\\n- let long = arg.get_long();\\n- if let Some(opt) = long {\\n- s.push_str(format!(\\\" --{}\\\", opt).as_str());\\n- }\\n+ completions.push_str(format!(\\\" export extern {} [\\\\n\\\", name).as_str());\\n \\n- let short = arg.get_short();\\n- if let Some(opt) = short {\\n- if long.is_some() {\\n- s.push_str(format!(\\\"(-{})\\\", opt).as_str());\\n- } else {\\n- s.push_str(format!(\\\" -{}\\\", opt).as_str());\\n- }\\n- }\\n-\\n- if let Some(v) = arg.get_num_args() {\\n- if v.takes_values() {\\n- // TODO: add more types?\\n- // TODO: add possible values?\\n- s.push_str(\\\": string\\\");\\n- }\\n- }\\n-\\n- if let Some(msg) = arg.get_help() {\\n- if arg.is_positional() || long.is_some() || short.is_some() {\\n- s.push_str(format!(\\\"\\\\t# {}\\\", msg).as_str());\\n- }\\n- }\\n-\\n- s.push('\\\\n');\\n- }\\n+ let s: String = cmd\\n+ .get_arguments()\\n+ .map(|arg| ArgumentLine::from(arg).to_string())\\n+ .collect();\\n \\n completions.push_str(&s);\\n completions.push_str(\\\" ]\\\\n\\\\n\\\");\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"7785be09053049b30cf41b420c59f051cd0129fc\", \"d35d302cadf355a169dca6636597183de6bbee23\", \"de88de81551d3e2619444a25a68170c9ed35a9b5\", \"f1bc5a554af4e617c7d7508f7f16f8fd25c78c91\"]"},"types":{"kind":"string","value":"[\"ci\", \"fix\", \"test\", \"refactor\"]"}}},{"rowIdx":943,"cells":{"commit_message":{"kind":"string","value":"await job creation to ensure asserted event sequence,add travis file,Downgrade @azure/* deps for Node.sj 10 compability,fix `memtable` docstrings"},"diff":{"kind":"string","value":"[\"diff --git a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\nindex 9ffa1fa..4333db0 100644\\n--- a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\n+++ b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\n@@ -114,12 +114,18 @@ public class BoundaryEventTest {\\n ENGINE.deployment().withXmlResource(MULTIPLE_SEQUENCE_FLOWS).deploy();\\n final long workflowInstanceKey = ENGINE.workflowInstance().ofBpmnProcessId(PROCESS_ID).create();\\n \\n- // when\\n RecordingExporter.timerRecords()\\n .withHandlerNodeId(\\\"timer\\\")\\n .withIntent(TimerIntent.CREATED)\\n .withWorkflowInstanceKey(workflowInstanceKey)\\n .getFirst();\\n+\\n+ RecordingExporter.jobRecords(JobIntent.CREATED)\\n+ .withType(\\\"type\\\")\\n+ .withWorkflowInstanceKey(workflowInstanceKey)\\n+ .getFirst();\\n+\\n+ // when\\n ENGINE.increaseTime(Duration.ofMinutes(1));\\n \\n // then\\n\", \"diff --git a/.travis.yml b/.travis.yml\\nnew file mode 100644\\nindex 0000000..c08cc34\\n--- /dev/null\\n+++ b/.travis.yml\\n@@ -0,0 +1,11 @@\\n+sudo: false\\n+\\n+language: java\\n+jdk: oraclejdk8\\n+\\n+branches:\\n+ only:\\n+ - master\\n+\\n+notifications:\\n+ email: false\\n\", \"diff --git a/package.json b/package.json\\nindex 911f8cd..ac29f54 100644\\n--- a/package.json\\n+++ b/package.json\\n@@ -79,7 +79,13 @@\\n \\\"resolutions\\\": {\\n \\\"@types/ramda\\\": \\\"0.27.40\\\",\\n \\\"rc-tree\\\": \\\"4.1.5\\\",\\n+ \\\"@azure/storage-blob\\\": \\\"12.7.0\\\",\\n+ \\\"@azure/core-paging\\\": \\\"1.1.3\\\",\\n+ \\\"@azure/logger\\\": \\\"1.0.0\\\",\\n \\\"@azure/core-auth\\\": \\\"1.2.0\\\",\\n+ \\\"@azure/core-lro\\\": \\\"1.0.5\\\",\\n+ \\\"@azure/core-tracing\\\": \\\"1.0.0-preview.10\\\",\\n+ \\\"@azure/core-http\\\": \\\"1.2.6\\\",\\n \\\"testcontainers\\\": \\\"7.12.1\\\"\\n },\\n \\\"license\\\": \\\"MIT\\\"\\ndiff --git a/yarn.lock b/yarn.lock\\nindex 5019f68..99235b5 100644\\n--- a/yarn.lock\\n+++ b/yarn.lock\\n@@ -1144,19 +1144,19 @@\\n \\\"@azure/abort-controller\\\" \\\"^1.0.0\\\"\\n tslib \\\"^2.0.0\\\"\\n \\n-\\\"@azure/core-http@^2.0.0\\\":\\n- version \\\"2.2.2\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@azure/core-http/-/core-http-2.2.2.tgz#573798f087d808d39aa71fd7c52b8d7b89f440da\\\"\\n- integrity sha512-V1DdoO9V/sFimKpdWoNBgsE+QUjQgpXYnxrTdUp5RyhsTJjvEVn/HKmTQXIHuLUUo6IyIWj+B+Dg4VaXse9dIA==\\n+\\\"@azure/core-http@1.2.6\\\", \\\"@azure/core-http@^1.2.0\\\", \\\"@azure/core-http@^2.0.0\\\":\\n+ version \\\"1.2.6\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@azure/core-http/-/core-http-1.2.6.tgz#9cd508418572d2062fd3175274219438772bdb65\\\"\\n+ integrity sha512-odtH7UMKtekc5YQ86xg9GlVHNXR6pq2JgJ5FBo7/jbOjNGdBqcrIVrZx2bevXVJz/uUTSx6vUf62gzTXTfqYSQ==\\n dependencies:\\n \\\"@azure/abort-controller\\\" \\\"^1.0.0\\\"\\n \\\"@azure/core-asynciterator-polyfill\\\" \\\"^1.0.0\\\"\\n \\\"@azure/core-auth\\\" \\\"^1.3.0\\\"\\n- \\\"@azure/core-tracing\\\" \\\"1.0.0-preview.13\\\"\\n+ \\\"@azure/core-tracing\\\" \\\"1.0.0-preview.11\\\"\\n \\\"@azure/logger\\\" \\\"^1.0.0\\\"\\n \\\"@types/node-fetch\\\" \\\"^2.5.0\\\"\\n- \\\"@types/tunnel\\\" \\\"^0.0.3\\\"\\n- form-data \\\"^4.0.0\\\"\\n+ \\\"@types/tunnel\\\" \\\"^0.0.1\\\"\\n+ form-data \\\"^3.0.0\\\"\\n node-fetch \\\"^2.6.0\\\"\\n process \\\"^0.11.10\\\"\\n tough-cookie \\\"^4.0.0\\\"\\n@@ -1165,38 +1165,39 @@\\n uuid \\\"^8.3.0\\\"\\n xml2js \\\"^0.4.19\\\"\\n \\n-\\\"@azure/core-lro@^2.2.0\\\":\\n- version \\\"2.2.1\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@azure/core-lro/-/core-lro-2.2.1.tgz#5527b41037c658d3aefc19d68633e51e53d6e6a3\\\"\\n- integrity sha512-HE6PBl+mlKa0eBsLwusHqAqjLc5n9ByxeDo3Hz4kF3B1hqHvRkBr4oMgoT6tX7Hc3q97KfDctDUon7EhvoeHPA==\\n+\\\"@azure/core-lro@1.0.5\\\", \\\"@azure/core-lro@^2.0.0\\\":\\n+ version \\\"1.0.5\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@azure/core-lro/-/core-lro-1.0.5.tgz#856a2cb6a9bec739ee9cde33a27cc28f81ac0522\\\"\\n+ integrity sha512-0EFCFZxARrIoLWMIRt4vuqconRVIO2Iin7nFBfJiYCCbKp5eEmxutNk8uqudPmG0XFl5YqlVh68/al/vbE5OOg==\\n dependencies:\\n \\\"@azure/abort-controller\\\" \\\"^1.0.0\\\"\\n- \\\"@azure/core-tracing\\\" \\\"1.0.0-preview.13\\\"\\n- \\\"@azure/logger\\\" \\\"^1.0.0\\\"\\n- tslib \\\"^2.2.0\\\"\\n+ \\\"@azure/core-http\\\" \\\"^1.2.0\\\"\\n+ \\\"@azure/core-tracing\\\" \\\"1.0.0-preview.11\\\"\\n+ events \\\"^3.0.0\\\"\\n+ tslib \\\"^2.0.0\\\"\\n \\n-\\\"@azure/core-paging@^1.1.1\\\":\\n- version \\\"1.2.0\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@azure/core-paging/-/core-paging-1.2.0.tgz#3754da429e8687bdc3613c750e79a564582e802b\\\"\\n- integrity sha512-ZX1bCjm/MjKPCN6kQD/9GJErYSoKA8YWp6YWoo5EIzcTWlSBLXu3gNaBTUl8usGl+UShiKo7b4Gdy1NSTIlpZg==\\n+\\\"@azure/core-paging@1.1.3\\\", \\\"@azure/core-paging@^1.1.1\\\":\\n+ version \\\"1.1.3\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@azure/core-paging/-/core-paging-1.1.3.tgz#3587c9898a0530cacb64bab216d7318468aa5efc\\\"\\n+ integrity sha512-his7Ah40ThEYORSpIAwuh6B8wkGwO/zG7gqVtmSE4WAJ46e36zUDXTKReUCLBDc6HmjjApQQxxcRFy5FruG79A==\\n dependencies:\\n \\\"@azure/core-asynciterator-polyfill\\\" \\\"^1.0.0\\\"\\n- tslib \\\"^2.2.0\\\"\\n \\n-\\\"@azure/core-tracing@1.0.0-preview.13\\\":\\n- version \\\"1.0.0-preview.13\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@azure/core-tracing/-/core-tracing-1.0.0-preview.13.tgz#55883d40ae2042f6f1e12b17dd0c0d34c536d644\\\"\\n- integrity sha512-KxDlhXyMlh2Jhj2ykX6vNEU0Vou4nHr025KoSEiz7cS3BNiHNaZcdECk/DmLkEB0as5T7b/TpRcehJ5yV6NeXQ==\\n+\\\"@azure/core-tracing@1.0.0-preview.10\\\", \\\"@azure/core-tracing@1.0.0-preview.11\\\", \\\"@azure/core-tracing@1.0.0-preview.13\\\":\\n+ version \\\"1.0.0-preview.10\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@azure/core-tracing/-/core-tracing-1.0.0-preview.10.tgz#e7060272145dddad4486765030d1b037cd52a8ea\\\"\\n+ integrity sha512-iIwjtMwQnsxB7cYkugMx+s4W1nfy3+pT/ceo+uW1fv4YDgYe84nh+QP0fEC9IH/3UATLSWbIBemdMHzk2APUrw==\\n dependencies:\\n- \\\"@opentelemetry/api\\\" \\\"^1.0.1\\\"\\n- tslib \\\"^2.2.0\\\"\\n+ \\\"@opencensus/web-types\\\" \\\"0.0.7\\\"\\n+ \\\"@opentelemetry/api\\\" \\\"^0.10.2\\\"\\n+ tslib \\\"^2.0.0\\\"\\n \\n-\\\"@azure/logger@^1.0.0\\\":\\n- version \\\"1.0.3\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@azure/logger/-/logger-1.0.3.tgz#6e36704aa51be7d4a1bae24731ea580836293c96\\\"\\n- integrity sha512-aK4s3Xxjrx3daZr3VylxejK3vG5ExXck5WOHDJ8in/k9AqlfIyFMMT1uG7u8mNjX+QRILTIn0/Xgschfh/dQ9g==\\n+\\\"@azure/logger@1.0.0\\\", \\\"@azure/logger@^1.0.0\\\":\\n+ version \\\"1.0.0\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@azure/logger/-/logger-1.0.0.tgz#48b371dfb34288c8797e5c104f6c4fb45bf1772c\\\"\\n+ integrity sha512-g2qLDgvmhyIxR3JVS8N67CyIOeFRKQlX/llxYJQr1OSGQqM3HTpVP8MjmjcEKbL/OIt2N9C9UFaNQuKOw1laOA==\\n dependencies:\\n- tslib \\\"^2.2.0\\\"\\n+ tslib \\\"^1.9.3\\\"\\n \\n \\\"@azure/ms-rest-azure-env@^2.0.0\\\":\\n version \\\"2.0.0\\\"\\n@@ -1227,19 +1228,19 @@\\n \\\"@azure/ms-rest-js\\\" \\\"^2.0.4\\\"\\n adal-node \\\"^0.2.2\\\"\\n \\n-\\\"@azure/storage-blob@^12.5.0\\\":\\n- version \\\"12.8.0\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@azure/storage-blob/-/storage-blob-12.8.0.tgz#97b7ecc6c7b17bcbaf0281c79c16af6f512d6130\\\"\\n- integrity sha512-c8+Wz19xauW0bGkTCoqZH4dYfbtBniPiGiRQOn1ca6G5jsjr4azwaTk9gwjVY8r3vY2Taf95eivLzipfIfiS4A==\\n+\\\"@azure/storage-blob@12.7.0\\\", \\\"@azure/storage-blob@^12.5.0\\\":\\n+ version \\\"12.7.0\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@azure/storage-blob/-/storage-blob-12.7.0.tgz#f17f278000a46bca516e5864d846cd8fa57d6d7d\\\"\\n+ integrity sha512-7YEWEx03Us/YBxthzBv788R7jokwpCD5KcIsvtE5xRaijNX9o80KXpabhEwLR9DD9nmt/AlU/c1R+aXydgCduQ==\\n dependencies:\\n \\\"@azure/abort-controller\\\" \\\"^1.0.0\\\"\\n \\\"@azure/core-http\\\" \\\"^2.0.0\\\"\\n- \\\"@azure/core-lro\\\" \\\"^2.2.0\\\"\\n+ \\\"@azure/core-lro\\\" \\\"^2.0.0\\\"\\n \\\"@azure/core-paging\\\" \\\"^1.1.1\\\"\\n \\\"@azure/core-tracing\\\" \\\"1.0.0-preview.13\\\"\\n \\\"@azure/logger\\\" \\\"^1.0.0\\\"\\n events \\\"^3.0.0\\\"\\n- tslib \\\"^2.2.0\\\"\\n+ tslib \\\"^2.0.0\\\"\\n \\n \\\"@babel/cli@^7.5.5\\\":\\n version \\\"7.16.0\\\"\\n@@ -2888,9 +2889,9 @@\\n integrity sha512-82cpyJyKRoQoRi+14ibCeGPu0CwypgtBAdBhq1WfvagpCZNKqwXbKwXllYSMG91DhmG4jt9gN8eP6lGOtozuaw==\\n \\n \\\"@google-cloud/bigquery@^5.6.0\\\":\\n- version \\\"5.9.1\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@google-cloud/bigquery/-/bigquery-5.9.1.tgz#96cee86fa0caef4a7e1470efde9295bc09f5981f\\\"\\n- integrity sha512-80pMzhAC299CSiXW9TvR8AARLaPRDeQg8pSAvrVcLXcUkx1hWvVx2m94nBZ4KUoZb4LVWIHHYhvFB6XvIcxqjw==\\n+ version \\\"5.9.2\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@google-cloud/bigquery/-/bigquery-5.9.2.tgz#d53eac984fdd256d31be490762157e5f6c5b82c3\\\"\\n+ integrity sha512-lJiMsSekcnhrzzR9e48yx8iOx+ElP3r/wOoionXL6eDPbA41RgP12if5NmMqHZzfWdKlWV2plspEPrbjhJAzCw==\\n dependencies:\\n \\\"@google-cloud/common\\\" \\\"^3.1.0\\\"\\n \\\"@google-cloud/paginator\\\" \\\"^3.0.0\\\"\\n@@ -4831,11 +4832,28 @@\\n resolved \\\"https://registry.yarnpkg.com/@oozcitak/util/-/util-8.3.8.tgz#10f65fe1891fd8cde4957360835e78fd1936bfdd\\\"\\n integrity sha512-T8TbSnGsxo6TDBJx/Sgv/BlVJL3tshxZP7Aq5R1mSnM5OcHY2dQaxLMu2+E8u3gN0MLOzdjurqN4ZRVuzQycOQ==\\n \\n-\\\"@opentelemetry/api@^1.0.0\\\", \\\"@opentelemetry/api@^1.0.1\\\":\\n+\\\"@opencensus/web-types@0.0.7\\\":\\n+ version \\\"0.0.7\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@opencensus/web-types/-/web-types-0.0.7.tgz#4426de1fe5aa8f624db395d2152b902874f0570a\\\"\\n+ integrity sha512-xB+w7ZDAu3YBzqH44rCmG9/RlrOmFuDPt/bpf17eJr8eZSrLt7nc7LnWdxM9Mmoj/YKMHpxRg28txu3TcpiL+g==\\n+\\n+\\\"@opentelemetry/api@^0.10.2\\\":\\n+ version \\\"0.10.2\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@opentelemetry/api/-/api-0.10.2.tgz#9647b881f3e1654089ff7ea59d587b2d35060654\\\"\\n+ integrity sha512-GtpMGd6vkzDMYcpu2t9LlhEgMy/SzBwRnz48EejlRArYqZzqSzAsKmegUK7zHgl+EOIaK9mKHhnRaQu3qw20cA==\\n+ dependencies:\\n+ \\\"@opentelemetry/context-base\\\" \\\"^0.10.2\\\"\\n+\\n+\\\"@opentelemetry/api@^1.0.0\\\":\\n version \\\"1.0.3\\\"\\n resolved \\\"https://registry.yarnpkg.com/@opentelemetry/api/-/api-1.0.3.tgz#13a12ae9e05c2a782f7b5e84c3cbfda4225eaf80\\\"\\n integrity sha512-puWxACExDe9nxbBB3lOymQFrLYml2dVOrd7USiVRnSbgXE+KwBu+HxFvxrzfqsiSda9IWsXJG1ef7C1O2/GmKQ==\\n \\n+\\\"@opentelemetry/context-base@^0.10.2\\\":\\n+ version \\\"0.10.2\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@opentelemetry/context-base/-/context-base-0.10.2.tgz#55bea904b2b91aa8a8675df9eaba5961bddb1def\\\"\\n+ integrity sha512-hZNKjKOYsckoOEgBziGMnBcX0M7EtstnCmwz5jZUOUYwlZ+/xxX6z3jPu1XVO2Jivk0eLfuP9GP+vFD49CMetw==\\n+\\n \\\"@opentelemetry/semantic-conventions@^0.24.0\\\":\\n version \\\"0.24.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/@opentelemetry/semantic-conventions/-/semantic-conventions-0.24.0.tgz#1028ef0e0923b24916158d80d2ddfd67ea8b6740\\\"\\n@@ -5564,9 +5582,9 @@\\n integrity sha1-7ihweulOEdK4J7y+UnC86n8+ce4=\\n \\n \\\"@types/jsonwebtoken@^8.5.0\\\":\\n- version \\\"8.5.5\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@types/jsonwebtoken/-/jsonwebtoken-8.5.5.tgz#da5f2f4baee88f052ef3e4db4c1a0afb46cff22c\\\"\\n- integrity sha512-OGqtHQ7N5/Ap/TUwO6IgHDuLiAoTmHhGpNvgkCm/F4N6pKzx/RBSfr2OXZSwC6vkfnsEdb6+7DNZVtiXiwdwFw==\\n+ version \\\"8.5.6\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@types/jsonwebtoken/-/jsonwebtoken-8.5.6.tgz#1913e5a61e70a192c5a444623da4901a7b1a9d42\\\"\\n+ integrity sha512-+P3O/xC7nzVizIi5VbF34YtqSonFsdnbXBnWUCYRiKOi1f9gA4sEFvXkrGr/QVV23IbMYvcoerI7nnhDUiWXRQ==\\n dependencies:\\n \\\"@types/node\\\" \\\"*\\\"\\n \\n@@ -5753,18 +5771,18 @@\\n \\\"@types/react\\\" \\\"*\\\"\\n \\n \\\"@types/react@*\\\", \\\"@types/react@^17.0.3\\\":\\n- version \\\"17.0.34\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@types/react/-/react-17.0.34.tgz#797b66d359b692e3f19991b6b07e4b0c706c0102\\\"\\n- integrity sha512-46FEGrMjc2+8XhHXILr+3+/sTe3OfzSPU9YGKILLrUYbQ1CLQC9Daqo1KzENGXAWwrFwiY0l4ZbF20gRvgpWTg==\\n+ version \\\"17.0.35\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@types/react/-/react-17.0.35.tgz#217164cf830267d56cd1aec09dcf25a541eedd4c\\\"\\n+ integrity sha512-r3C8/TJuri/SLZiiwwxQoLAoavaczARfT9up9b4Jr65+ErAUX3MIkU0oMOQnrpfgHme8zIqZLX7O5nnjm5Wayw==\\n dependencies:\\n \\\"@types/prop-types\\\" \\\"*\\\"\\n \\\"@types/scheduler\\\" \\\"*\\\"\\n csstype \\\"^3.0.2\\\"\\n \\n \\\"@types/react@^16.9.41\\\":\\n- version \\\"16.14.20\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@types/react/-/react-16.14.20.tgz#ff6e932ad71d92c27590e4a8667c7a53a7d0baad\\\"\\n- integrity sha512-SV7TaVc8e9E/5Xuv6TIyJ5VhQpZoVFJqX6IZgj5HZoFCtIDCArE3qXkcHlc6O/Ud4UwcMoX+tlvDA95YrKdLgA==\\n+ version \\\"16.14.21\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@types/react/-/react-16.14.21.tgz#35199b21a278355ec7a3c40003bd6a334bd4ae4a\\\"\\n+ integrity sha512-rY4DzPKK/4aohyWiDRHS2fotN5rhBSK6/rz1X37KzNna9HJyqtaGAbq9fVttrEPWF5ywpfIP1ITL8Xi2QZn6Eg==\\n dependencies:\\n \\\"@types/prop-types\\\" \\\"*\\\"\\n \\\"@types/scheduler\\\" \\\"*\\\"\\n@@ -5950,10 +5968,10 @@\\n resolved \\\"https://registry.yarnpkg.com/@types/tough-cookie/-/tough-cookie-4.0.1.tgz#8f80dd965ad81f3e1bc26d6f5c727e132721ff40\\\"\\n integrity sha512-Y0K95ThC3esLEYD6ZuqNek29lNX2EM1qxV8y2FTLUB0ff5wWrk7az+mLrnNFUnaXcgKye22+sFBRXOgpPILZNg==\\n \\n-\\\"@types/tunnel@^0.0.3\\\":\\n- version \\\"0.0.3\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@types/tunnel/-/tunnel-0.0.3.tgz#f109e730b072b3136347561fc558c9358bb8c6e9\\\"\\n- integrity sha512-sOUTGn6h1SfQ+gbgqC364jLFBw2lnFqkgF3q0WovEHRLMrVD1sd5aufqi/aJObLekJO+Aq5z646U4Oxy6shXMA==\\n+\\\"@types/tunnel@^0.0.1\\\":\\n+ version \\\"0.0.1\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@types/tunnel/-/tunnel-0.0.1.tgz#0d72774768b73df26f25df9184273a42da72b19c\\\"\\n+ integrity sha512-AOqu6bQu5MSWwYvehMXLukFHnupHrpZ8nvgae5Ggie9UwzDR1CCwoXgSSWNZJuyOlCdfdsWMA5F2LlmvyoTv8A==\\n dependencies:\\n \\\"@types/node\\\" \\\"*\\\"\\n \\n@@ -5999,9 +6017,9 @@\\n source-map \\\"^0.6.1\\\"\\n \\n \\\"@types/webpack@^4\\\", \\\"@types/webpack@^4.0.0\\\", \\\"@types/webpack@^4.41.8\\\":\\n- version \\\"4.41.31\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@types/webpack/-/webpack-4.41.31.tgz#c35f252a3559ddf9c85c0d8b0b42019025e581aa\\\"\\n- integrity sha512-/i0J7sepXFIp1ZT7FjUGi1eXMCg8HCCzLJEQkKsOtbJFontsJLolBcDC+3qxn5pPwiCt1G0ZdRmYRzNBtvpuGQ==\\n+ version \\\"4.41.32\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@types/webpack/-/webpack-4.41.32.tgz#a7bab03b72904070162b2f169415492209e94212\\\"\\n+ integrity sha512-cb+0ioil/7oz5//7tZUSwbrSAN/NWHrQylz5cW8G0dWTcF/g+/dSdMlKVZspBYuMAN1+WnwHrkxiRrLcwd0Heg==\\n dependencies:\\n \\\"@types/node\\\" \\\"*\\\"\\n \\\"@types/tapable\\\" \\\"^1\\\"\\n@@ -7624,9 +7642,9 @@ autoprefixer@^9.6.1, autoprefixer@^9.6.5, autoprefixer@^9.8.6:\\n postcss-value-parser \\\"^4.1.0\\\"\\n \\n aws-sdk@^2.404.0, aws-sdk@^2.787.0, aws-sdk@^2.819.0, aws-sdk@^2.878.0:\\n- version \\\"2.1028.0\\\"\\n- resolved \\\"https://registry.yarnpkg.com/aws-sdk/-/aws-sdk-2.1028.0.tgz#ce076076174afa9bd311406b8186ea90163e3331\\\"\\n- integrity sha512-OmR0NcpU8zsDcUOZhM+eZ6CzlUFtuaEuRyjm6mxDO0KI7lJAp7/NzB6tcellRrgWxL+NO7b5TSxi+m28qu5ocQ==\\n+ version \\\"2.1029.0\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/aws-sdk/-/aws-sdk-2.1029.0.tgz#702d4d6092adcf0ceaf37ae0da6fee07a71f39dd\\\"\\n+ integrity sha512-nCmaMPkJr3EATXaeqR3JeNC0GTDH2lJZ3Xq/ZCAW+yrfaPQWv8HqJJHBCNGtmk3FmcCoxc7ed/gEB8XSl0tocA==\\n dependencies:\\n buffer \\\"4.9.2\\\"\\n events \\\"1.1.1\\\"\\n@@ -8596,11 +8614,16 @@ bytes@3.0.0:\\n resolved \\\"https://registry.yarnpkg.com/bytes/-/bytes-3.0.0.tgz#d32815404d689699f85a4ea4fa8755dd13a96048\\\"\\n integrity sha1-0ygVQE1olpn4Wk6k+odV3ROpYEg=\\n \\n-bytes@3.1.0, bytes@^3.1.0:\\n+bytes@3.1.0:\\n version \\\"3.1.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/bytes/-/bytes-3.1.0.tgz#f6cf7933a360e0588fa9fde85651cdc7f805d1f6\\\"\\n integrity sha512-zauLjrfCG+xvoyaqLoV8bLVXXNGC4JqlxFCutSDWA6fJrTo2ZuvLYTqZ7aHBLZSMOopbzwv8f+wZcVzfVTI2Dg==\\n \\n+bytes@^3.1.0:\\n+ version \\\"3.1.1\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/bytes/-/bytes-3.1.1.tgz#3f018291cb4cbad9accb6e6970bca9c8889e879a\\\"\\n+ integrity sha512-dWe4nWO/ruEOY7HkUJ5gFt1DCFV9zPRoJr8pV0/ASQermOZjtq8jMjOprC0Kd10GLN+l7xaUPvxzJFWtxGu8Fg==\\n+\\n cacache@15.0.3:\\n version \\\"15.0.3\\\"\\n resolved \\\"https://registry.yarnpkg.com/cacache/-/cacache-15.0.3.tgz#2225c2d1dd8e872339950d6a39c051e0e9334392\\\"\\n@@ -11359,9 +11382,9 @@ ejs@^2.6.1:\\n integrity sha512-7vmuyh5+kuUyJKePhQfRQBhXV5Ce+RnaeeQArKu1EAMpL3WbgMt5WG6uQZpEVvYSSsxMXRKOewtDk9RaTKXRlA==\\n \\n electron-to-chromium@^1.3.564, electron-to-chromium@^1.3.896:\\n- version \\\"1.3.896\\\"\\n- resolved \\\"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.896.tgz#4a94efe4870b1687eafd5c378198a49da06e8a1b\\\"\\n- integrity sha512-NcGkBVXePiuUrPLV8IxP43n1EOtdg+dudVjrfVEUd/bOqpQUFZ2diL5PPYzbgEhZFEltdXV3AcyKwGnEQ5lhMA==\\n+ version \\\"1.3.899\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/electron-to-chromium/-/electron-to-chromium-1.3.899.tgz#4d7d040e73def3d5f5bd6b8a21049025dce6fce0\\\"\\n+ integrity sha512-w16Dtd2zl7VZ4N4Db+FIa7n36sgPGCKjrKvUUmp5ialsikvcQLjcJR9RWnlYNxIyEHLdHaoIZEqKsPxU9MdyBg==\\n \\n elegant-spinner@^1.0.1:\\n version \\\"1.0.1\\\"\\n@@ -12887,15 +12910,6 @@ form-data@^3.0.0:\\n combined-stream \\\"^1.0.8\\\"\\n mime-types \\\"^2.1.12\\\"\\n \\n-form-data@^4.0.0:\\n- version \\\"4.0.0\\\"\\n- resolved \\\"https://registry.yarnpkg.com/form-data/-/form-data-4.0.0.tgz#93919daeaf361ee529584b9b31664dc12c9fa452\\\"\\n- integrity sha512-ETEklSGi5t0QMZuiXoA/Q6vcnxcLQP5vdugSpuAyi6SVGi2clPPp+xgEhuMaHC+zGgn31Kd235W35f7Hykkaww==\\n- dependencies:\\n- asynckit \\\"^0.4.0\\\"\\n- combined-stream \\\"^1.0.8\\\"\\n- mime-types \\\"^2.1.12\\\"\\n-\\n form-data@~2.3.2:\\n version \\\"2.3.3\\\"\\n resolved \\\"https://registry.yarnpkg.com/form-data/-/form-data-2.3.3.tgz#dcce52c05f644f298c6a7ab936bd724ceffbf3a6\\\"\\n@@ -21198,11 +21212,13 @@ proto-list@~1.2.1:\\n integrity sha1-IS1b/hMYMGpCD2QCuOJv85ZHqEk=\\n \\n proto3-json-serializer@^0.1.5:\\n- version \\\"0.1.5\\\"\\n- resolved \\\"https://registry.yarnpkg.com/proto3-json-serializer/-/proto3-json-serializer-0.1.5.tgz#c619769a59dc7fd8adf4e6c5060b9bf3039c8304\\\"\\n- integrity sha512-G395jcZkgNXNeS+6FGqd09TsXeoCs9wmBWByDiwFy7Yd7HD8pyfyvf6q+rGh7PhT4AshRpG4NowzoKYUtkNjKg==\\n+ version \\\"0.1.6\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/proto3-json-serializer/-/proto3-json-serializer-0.1.6.tgz#67cf3b8d5f4c8bebfc410698ad3b1ed64da39c7b\\\"\\n+ integrity sha512-tGbV6m6Kad8NqxMh5hw87euPS0YoZSAOIfvR01zYkQV8Gpx1V/8yU/0gCKCvfCkhAJsjvzzhnnsdQxA1w7PSog==\\n+ dependencies:\\n+ protobufjs \\\"^6.11.2\\\"\\n \\n-protobufjs@6.11.2, protobufjs@^6.10.0:\\n+protobufjs@6.11.2, protobufjs@^6.10.0, protobufjs@^6.11.2:\\n version \\\"6.11.2\\\"\\n resolved \\\"https://registry.yarnpkg.com/protobufjs/-/protobufjs-6.11.2.tgz#de39fabd4ed32beaa08e9bb1e30d08544c1edf8b\\\"\\n integrity sha512-4BQJoPooKJl2G9j3XftkIXjoC9C0Av2NOrWmbLWT1vH32GcSUHjM0Arra6UfTsVyfMAuFzaLucXn1sadxJydAw==\\n\", \"diff --git a/ibis/expr/api.py b/ibis/expr/api.py\\nindex 93fabaa..66a2ea9 100644\\n--- a/ibis/expr/api.py\\n+++ b/ibis/expr/api.py\\n@@ -403,15 +403,21 @@ def memtable(\\n >>> import ibis\\n >>> t = ibis.memtable([{\\\"a\\\": 1}, {\\\"a\\\": 2}])\\n >>> t\\n+ PandasInMemoryTable\\n+ data:\\n+ DataFrameProxy:\\n+ a\\n+ 0 1\\n+ 1 2\\n \\n >>> t = ibis.memtable([{\\\"a\\\": 1, \\\"b\\\": \\\"foo\\\"}, {\\\"a\\\": 2, \\\"b\\\": \\\"baz\\\"}])\\n >>> t\\n PandasInMemoryTable\\n data:\\n- ((1, 'foo'), (2, 'baz'))\\n- schema:\\n- a int8\\n- b string\\n+ DataFrameProxy:\\n+ a b\\n+ 0 1 foo\\n+ 1 2 baz\\n \\n Create a table literal without column names embedded in the data and pass\\n `columns`\\n@@ -420,10 +426,22 @@ def memtable(\\n >>> t\\n PandasInMemoryTable\\n data:\\n- ((1, 'foo'), (2, 'baz'))\\n- schema:\\n- a int8\\n- b string\\n+ DataFrameProxy:\\n+ a b\\n+ 0 1 foo\\n+ 1 2 baz\\n+\\n+ Create a table literal without column names embedded in the data. Ibis\\n+ generates column names if none are provided.\\n+\\n+ >>> t = ibis.memtable([(1, \\\"foo\\\"), (2, \\\"baz\\\")])\\n+ >>> t\\n+ PandasInMemoryTable\\n+ data:\\n+ DataFrameProxy:\\n+ col0 col1\\n+ 0 1 foo\\n+ 1 2 baz\\n \\\"\\\"\\\"\\n if columns is not None and schema is not None:\\n raise NotImplementedError(\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"a8d1a60fd48d3fbd76d4271987a1b0f538d498f1\", \"d0814a928601706635287fd3d9d3451d156b821a\", \"5ef4fd29a4cef69c6c348dd25156934df041f183\", \"72bc0f5172c0a3d17bde29cfc00db4c60d2fee3a\"]"},"types":{"kind":"string","value":"[\"test\", \"ci\", \"build\", \"docs\"]"}}},{"rowIdx":944,"cells":{"commit_message":{"kind":"string","value":"add title to badge icon,update version (nightly.0),rename ELECTRON_CACHE env variable to electron_config_cache (#21313),fix height calc"},"diff":{"kind":"string","value":"[\"diff --git a/kibbeh/src/modules/room/chat/RoomChatList.tsx b/kibbeh/src/modules/room/chat/RoomChatList.tsx\\nindex a7418e6..805a9a4 100644\\n--- a/kibbeh/src/modules/room/chat/RoomChatList.tsx\\n+++ b/kibbeh/src/modules/room/chat/RoomChatList.tsx\\n@@ -16,6 +16,11 @@ interface ChatListProps {\\n users: RoomUser[];\\n }\\n \\n+interface BadgeIconData {\\n+ emoji: string,\\n+ title: string\\n+}\\n+\\n export const RoomChatList: React.FC = ({ room, users }) => {\\n const { setData } = useContext(UserPreviewModalContext);\\n const { messages, toggleFrozen } = useRoomChatStore();\\n@@ -48,11 +53,14 @@ export const RoomChatList: React.FC = ({ room, users }) => {\\n const getBadgeIcon = (m: Message) => {\\n const user = users.find((u) => u.id === m.userId);\\n const isSpeaker = room.creatorId === user?.id || user?.roomPermissions?.isSpeaker;\\n- let emoji = null;\\n+ let badgeIconData: BadgeIconData | null = null;\\n if (isSpeaker) {\\n- emoji = \\\"\\ud83d\\udce3\\\";\\n+ badgeIconData = {\\n+ emoji: \\\"\\ud83d\\udce3\\\",\\n+ title: \\\"Speaker\\\"\\n+ };\\n }\\n- return emoji && ;\\n+ return badgeIconData && ;\\n };\\n \\n return (\\n\", \"diff --git a/Cargo.lock b/Cargo.lock\\nindex e6f659c..cf93556 100644\\n--- a/Cargo.lock\\n+++ b/Cargo.lock\\n@@ -94,7 +94,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"els\\\"\\n-version = \\\"0.1.23\\\"\\n+version = \\\"0.1.24-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_compiler\\\",\\n@@ -105,7 +105,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"els\\\",\\n \\\"erg_common\\\",\\n@@ -115,7 +115,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_common\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"backtrace-on-stack-overflow\\\",\\n \\\"crossterm\\\",\\n@@ -126,7 +126,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_compiler\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_parser\\\",\\n@@ -134,7 +134,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_parser\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"unicode-xid\\\",\\ndiff --git a/Cargo.toml b/Cargo.toml\\nindex c58299b..6e51b6e 100644\\n--- a/Cargo.toml\\n+++ b/Cargo.toml\\n@@ -20,7 +20,7 @@ members = [\\n ]\\n \\n [workspace.package]\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n authors = [\\\"erg-lang team \\\"]\\n license = \\\"MIT OR Apache-2.0\\\"\\n edition = \\\"2021\\\"\\n@@ -64,10 +64,10 @@ full-repl = [\\\"erg_common/full-repl\\\"]\\n full = [\\\"els\\\", \\\"full-repl\\\", \\\"unicode\\\", \\\"pretty\\\"]\\n \\n [workspace.dependencies]\\n-erg_common = { version = \\\"0.6.11\\\", path = \\\"./crates/erg_common\\\" }\\n-erg_parser = { version = \\\"0.6.11\\\", path = \\\"./crates/erg_parser\\\" }\\n-erg_compiler = { version = \\\"0.6.11\\\", path = \\\"./crates/erg_compiler\\\" }\\n-els = { version = \\\"0.1.23\\\", path = \\\"./crates/els\\\" }\\n+erg_common = { version = \\\"0.6.12-nightly.0\\\", path = \\\"./crates/erg_common\\\" }\\n+erg_parser = { version = \\\"0.6.12-nightly.0\\\", path = \\\"./crates/erg_parser\\\" }\\n+erg_compiler = { version = \\\"0.6.12-nightly.0\\\", path = \\\"./crates/erg_compiler\\\" }\\n+els = { version = \\\"0.1.24-nightly.0\\\", path = \\\"./crates/els\\\" }\\n \\n [dependencies]\\n erg_common = { workspace = true }\\ndiff --git a/crates/els/Cargo.toml b/crates/els/Cargo.toml\\nindex 5f005a1..e1a9964 100644\\n--- a/crates/els/Cargo.toml\\n+++ b/crates/els/Cargo.toml\\n@@ -2,7 +2,7 @@\\n name = \\\"els\\\"\\n description = \\\"An Erg compiler frontend for IDEs, implements LSP.\\\"\\n documentation = \\\"http://docs.rs/els\\\"\\n-version = \\\"0.1.23\\\"\\n+version = \\\"0.1.24-nightly.0\\\"\\n authors.workspace = true\\n license.workspace = true\\n edition.workspace = true\\n\", \"diff --git a/docs/tutorial/installation.md b/docs/tutorial/installation.md\\nindex d4af120..1a09eea 100644\\n--- a/docs/tutorial/installation.md\\n+++ b/docs/tutorial/installation.md\\n@@ -82,7 +82,7 @@ with the network at all.\\n On environments that have been using older versions of Electron, you might find the\\n cache also in `~/.electron`.\\n \\n-You can also override the local cache location by providing a `ELECTRON_CACHE`\\n+You can also override the local cache location by providing a `electron_config_cache`\\n environment variable.\\n \\n The cache contains the version's official zip file as well as a checksum, stored as\\n\", \"diff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\\nindex 96b54f1..bcfe3bb 100644\\n--- a/src/content/redux/modules/widget.ts\\n+++ b/src/content/redux/modules/widget.ts\\n@@ -93,7 +93,7 @@ export const initState: WidgetState = {\\n : _initConfig.panelWidth,\\n height: isSaladictPopupPage\\n ? 400\\n- : 30 + _initConfig.dicts.selected.length * 30,\\n+ : 30 + 30, // menubar + 1 dict hegiht\\n },\\n bowlRect: {\\n x: 0,\\n@@ -565,7 +565,7 @@ function listenNewSelection (\\n mouseX,\\n mouseY,\\n lastPanelRect.width,\\n- 30 + state.config.dicts.selected.length * 30,\\n+ 30 + state.dictionaries.active.length * 30,\\n )\\n }\\n }\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"6e5098655e6d9bb13f6423abe780cdf6b50ff13a\", \"92e940efeee199b1e0bbbc3c9eea7f3dc8221619\", \"f2f52c23b513dd857350f3c163f676d37189d0d3\", \"148cd56d096ba972e9706653c47052a07d5f9d08\"]"},"types":{"kind":"string","value":"[\"feat\", \"build\", \"docs\", \"fix\"]"}}},{"rowIdx":945,"cells":{"commit_message":{"kind":"string","value":"updated webpack in react,use module path alias,add page balckwhitelist and pdf,permission check"},"diff":{"kind":"string","value":"[\"diff --git a/components/react/package.json b/components/react/package.json\\nindex bbeb9ee..43ddebc 100644\\n--- a/components/react/package.json\\n+++ b/components/react/package.json\\n@@ -114,7 +114,7 @@\\n \\\"ts-loader\\\": \\\"^9.2.9\\\",\\n \\\"ts-node\\\": \\\"^10.7.0\\\",\\n \\\"typescript\\\": \\\"^4.7.3\\\",\\n- \\\"webpack\\\": \\\"^5.72.0\\\",\\n+ \\\"webpack\\\": \\\"^5.73.0\\\",\\n \\\"webpack-bundle-analyzer\\\": \\\"^4.5.0\\\",\\n \\\"webpack-cli\\\": \\\"^4.9.2\\\",\\n \\\"webpack-node-externals\\\": \\\"^3.0.0\\\"\\ndiff --git a/yarn.lock b/yarn.lock\\nindex a3fdb26..19a0716 100644\\n--- a/yarn.lock\\n+++ b/yarn.lock\\n@@ -25212,7 +25212,7 @@ webpack@^4.38.0, webpack@^4.42.1:\\n watchpack \\\"^1.7.4\\\"\\n webpack-sources \\\"^1.4.1\\\"\\n \\n-webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0:\\n+webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0, webpack@^5.73.0:\\n version \\\"5.73.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/webpack/-/webpack-5.73.0.tgz#bbd17738f8a53ee5760ea2f59dce7f3431d35d38\\\"\\n integrity sha512-svjudQRPPa0YiOYa2lM/Gacw0r6PvxptHj4FuEKQ2kX05ZLkjbVc5MnPs6its5j7IZljnIqSVo/OsY2X0IpHGA==\\n\", \"diff --git a/src/background/audio-manager.ts b/src/background/audio-manager.ts\\nindex 54e8b24..11c5fba 100644\\n--- a/src/background/audio-manager.ts\\n+++ b/src/background/audio-manager.ts\\n@@ -2,7 +2,7 @@\\n * To make sure only one audio plays at a time\\n */\\n \\n-import { timeout } from '../_helpers/promise-more'\\n+import { timeout } from '@/_helpers/promise-more'\\n \\n declare global {\\n interface Window {\\ndiff --git a/src/background/context-menus.ts b/src/background/context-menus.ts\\nindex 994b59e..7036362 100644\\n--- a/src/background/context-menus.ts\\n+++ b/src/background/context-menus.ts\\n@@ -1,5 +1,5 @@\\n-import { storage, openURL } from '../_helpers/browser-api'\\n-import { AppConfig } from '../app-config'\\n+import { storage, openURL } from '@/_helpers/browser-api'\\n+import { AppConfig } from '@/app-config'\\n \\n import { Observable } from 'rxjs/Observable'\\n import { fromPromise } from 'rxjs/observable/fromPromise'\\ndiff --git a/src/background/initialization.ts b/src/background/initialization.ts\\nindex 0e5b3ad..001ee73 100644\\n--- a/src/background/initialization.ts\\n+++ b/src/background/initialization.ts\\n@@ -1,6 +1,6 @@\\n-import { storage, openURL } from '../_helpers/browser-api'\\n-import checkUpdate from '../_helpers/check-update'\\n-import { AppConfig } from '../app-config'\\n+import { storage, openURL } from '@/_helpers/browser-api'\\n+import checkUpdate from '@/_helpers/check-update'\\n+import { AppConfig } from '@/app-config'\\n import { mergeConfig } from './merge-config'\\n import { init as initMenus } from './context-menus'\\n import { init as initPdf } from './pdf-sniffer'\\ndiff --git a/src/background/merge-config.ts b/src/background/merge-config.ts\\nindex afa1800..afdbd63 100644\\n--- a/src/background/merge-config.ts\\n+++ b/src/background/merge-config.ts\\n@@ -1,4 +1,4 @@\\n-import { appConfigFactory, AppConfig } from '../app-config'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n import _ from 'lodash'\\n \\n /**\\n@@ -24,7 +24,7 @@ function initConfig (): Promise {\\n const storageObj = { config: appConfigFactory() }\\n \\n Object.keys(storageObj.config.dicts.all).forEach(id => {\\n- storageObj[id] = require('../components/dictionaries/' + id + '/config')\\n+ storageObj[id] = require('@/components/dictionaries/' + id + '/config')\\n })\\n \\n return browser.storage.sync.set(storageObj)\\n@@ -70,7 +70,7 @@ function mergeHistorical (config): Promise {\\n \\n const storageObj = { config: base }\\n Object.keys(base.dicts.all).forEach(id => {\\n- storageObj[id] = config.dicts.all[id] || require('../components/dictionaries/' + id + '/config')\\n+ storageObj[id] = config.dicts.all[id] || require('@/components/dictionaries/' + id + '/config')\\n })\\n \\n return browser.storage.sync.set(storageObj)\\ndiff --git a/src/background/pdf-sniffer.ts b/src/background/pdf-sniffer.ts\\nindex 6ba27cf..70aa38f 100644\\n--- a/src/background/pdf-sniffer.ts\\n+++ b/src/background/pdf-sniffer.ts\\n@@ -2,8 +2,8 @@\\n * Open pdf link directly\\n */\\n \\n-import { storage } from '../_helpers/browser-api'\\n-import { AppConfig } from '../app-config'\\n+import { storage } from '@/_helpers/browser-api'\\n+import { AppConfig } from '@/app-config'\\n \\n export function init (pdfSniff: boolean) {\\n if (browser.webRequest.onBeforeRequest.hasListener(otherPdfListener)) {\\ndiff --git a/src/background/server.ts b/src/background/server.ts\\nindex 73b34b6..66ed5c0 100644\\n--- a/src/background/server.ts\\n+++ b/src/background/server.ts\\n@@ -1,7 +1,7 @@\\n-import { DictID } from '../app-config'\\n-import { message, openURL } from '../_helpers/browser-api'\\n+import { DictID } from '@/app-config'\\n+import { message, openURL } from '@/_helpers/browser-api'\\n import { play } from './audio-manager'\\n-import { chsToChz } from '../_helpers/chs-to-chz'\\n+import { chsToChz } from '@/_helpers/chs-to-chz'\\n \\n interface MessageOpenUrlWithEscape {\\n type: 'OPEN_URL'\\n@@ -63,7 +63,7 @@ function fetchDictResult (data: MessageFetchDictResult): Promise {\\n let search\\n \\n try {\\n- search = require('../components/dictionaries/' + data.dict + '/engine.js')\\n+ search = require('@/components/dictionaries/' + data.dict + '/engine.js')\\n } catch (err) {\\n return Promise.reject(err)\\n }\\ndiff --git a/test/unit/_helpers/browser-api.spec.ts b/test/unit/_helpers/browser-api.spec.ts\\nindex 1f39145..e327169 100644\\n--- a/test/unit/_helpers/browser-api.spec.ts\\n+++ b/test/unit/_helpers/browser-api.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { message, storage, openURL } from '../../../src/_helpers/browser-api'\\n+import { message, storage, openURL } from '@/_helpers/browser-api'\\n \\n beforeEach(() => {\\n browser.flush()\\ndiff --git a/test/unit/_helpers/check-update.spec.ts b/test/unit/_helpers/check-update.spec.ts\\nindex 2abfc57..fd0b678 100644\\n--- a/test/unit/_helpers/check-update.spec.ts\\n+++ b/test/unit/_helpers/check-update.spec.ts\\n@@ -1,4 +1,4 @@\\n-import checkUpdate from '../../../src/_helpers/check-update'\\n+import checkUpdate from '@/_helpers/check-update'\\n import fetchMock from 'jest-fetch-mock'\\n \\n describe('Check Update', () => {\\ndiff --git a/test/unit/_helpers/chs-to-chz.spec.ts b/test/unit/_helpers/chs-to-chz.spec.ts\\nindex 295c6ad..21d5229 100644\\n--- a/test/unit/_helpers/chs-to-chz.spec.ts\\n+++ b/test/unit/_helpers/chs-to-chz.spec.ts\\n@@ -1,4 +1,4 @@\\n-import chsToChz from '../../../src/_helpers/chs-to-chz'\\n+import chsToChz from '@/_helpers/chs-to-chz'\\n \\n describe('Chs to Chz', () => {\\n it('should convert chs to chz', () => {\\ndiff --git a/test/unit/_helpers/fetch-dom.spec.ts b/test/unit/_helpers/fetch-dom.spec.ts\\nindex a79dda0..bbfbf10 100644\\n--- a/test/unit/_helpers/fetch-dom.spec.ts\\n+++ b/test/unit/_helpers/fetch-dom.spec.ts\\n@@ -1,4 +1,4 @@\\n-import fetchDom from '../../../src/_helpers/fetch-dom'\\n+import fetchDom from '@/_helpers/fetch-dom'\\n \\n class XMLHttpRequestMock {\\n static queue: XMLHttpRequestMock[] = []\\ndiff --git a/test/unit/_helpers/lang-check.spec.ts b/test/unit/_helpers/lang-check.spec.ts\\nindex f3e668a..09f30bb 100644\\n--- a/test/unit/_helpers/lang-check.spec.ts\\n+++ b/test/unit/_helpers/lang-check.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { isContainChinese, isContainEnglish } from '../../../src/_helpers/lang-check'\\n+import { isContainChinese, isContainEnglish } from '@/_helpers/lang-check'\\n \\n describe('Language Check', () => {\\n it('isContainChinese should return ture if text contains Chinese', () => {\\ndiff --git a/test/unit/_helpers/promise-more.spec.ts b/test/unit/_helpers/promise-more.spec.ts\\nindex 9601c7d..66dc8d9 100644\\n--- a/test/unit/_helpers/promise-more.spec.ts\\n+++ b/test/unit/_helpers/promise-more.spec.ts\\n@@ -1,4 +1,4 @@\\n-import * as pm from '../../../src/_helpers/promise-more'\\n+import * as pm from '@/_helpers/promise-more'\\n \\n describe('Promise More', () => {\\n beforeAll(() => {\\ndiff --git a/test/unit/_helpers/selection.spec.ts b/test/unit/_helpers/selection.spec.ts\\nindex 370239a..06812cf 100644\\n--- a/test/unit/_helpers/selection.spec.ts\\n+++ b/test/unit/_helpers/selection.spec.ts\\n@@ -1,4 +1,4 @@\\n-import selection from '../../../src/_helpers/selection'\\n+import selection from '@/_helpers/selection'\\n \\n describe('Selection', () => {\\n const bakSelection = window.getSelection\\ndiff --git a/test/unit/_helpers/strip-script.spec.ts b/test/unit/_helpers/strip-script.spec.ts\\nindex cce558f..355b382 100644\\n--- a/test/unit/_helpers/strip-script.spec.ts\\n+++ b/test/unit/_helpers/strip-script.spec.ts\\n@@ -1,4 +1,4 @@\\n-import stripScript from '../../../src/_helpers/strip-script'\\n+import stripScript from '@/_helpers/strip-script'\\n \\n describe('Strip Script', () => {\\n const expectedEl = document.createElement('div') as HTMLDivElement\\ndiff --git a/test/unit/background/audio-manager.spec.ts b/test/unit/background/audio-manager.spec.ts\\nindex b0096a6..b1266d7 100644\\n--- a/test/unit/background/audio-manager.spec.ts\\n+++ b/test/unit/background/audio-manager.spec.ts\\n@@ -1,4 +1,4 @@\\n-import audio from '../../../src/background/audio-manager'\\n+import audio from '@/background/audio-manager'\\n \\n describe('Audio Manager', () => {\\n const bakAudio = (window as any).Audio\\ndiff --git a/test/unit/background/context-menus.spec.ts b/test/unit/background/context-menus.spec.ts\\nindex 39e249c..d9049dc 100644\\n--- a/test/unit/background/context-menus.spec.ts\\n+++ b/test/unit/background/context-menus.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n import sinon from 'sinon'\\n \\n function specialConfig () {\\n@@ -11,7 +11,7 @@ describe('Context Menus', () => {\\n beforeAll(() => {\\n browser.flush()\\n jest.resetModules()\\n- require('../../../src/background/context-menus')\\n+ require('@/background/context-menus')\\n })\\n afterAll(() => browser.flush())\\n \\n@@ -93,7 +93,7 @@ describe('Context Menus', () => {\\n browser.contextMenus.create.callsFake((_, cb) => cb())\\n config = specialConfig()\\n jest.resetModules()\\n- const { init } = require('../../../src/background/context-menus')\\n+ const { init } = require('@/background/context-menus')\\n init(config.contextMenus)\\n })\\n \\n@@ -110,7 +110,7 @@ describe('Context Menus', () => {\\n it('should not init setup when called multiple times', () => {\\n expect(browser.contextMenus.removeAll.calledOnce).toBeTruthy()\\n \\n- const { init } = require('../../../src/background/context-menus')\\n+ const { init } = require('@/background/context-menus')\\n init(config.contextMenus)\\n init(config.contextMenus)\\n \\ndiff --git a/test/unit/background/initialization.spec.ts b/test/unit/background/initialization.spec.ts\\nindex 7bc0972..56a6389 100644\\n--- a/test/unit/background/initialization.spec.ts\\n+++ b/test/unit/background/initialization.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n import fetchMock from 'jest-fetch-mock'\\n import sinon from 'sinon'\\n \\n@@ -11,12 +11,12 @@ describe('Initialization', () => {\\n const checkUpdate = jest.fn().mockReturnValue(Promise.resolve())\\n \\n beforeAll(() => {\\n- const { message, storage } = require('../../../src/_helpers/browser-api')\\n+ const { message, storage } = require('@/_helpers/browser-api')\\n window.fetch = fetchMock\\n \\n browser.flush()\\n jest.resetModules()\\n- jest.doMock('../../../src/background/merge-config', () => {\\n+ jest.doMock('@/background/merge-config', () => {\\n return {\\n mergeConfig (config) {\\n mergeConfig(config)\\n@@ -24,16 +24,16 @@ describe('Initialization', () => {\\n }\\n }\\n })\\n- jest.doMock('../../../src/background/context-menus', () => {\\n+ jest.doMock('@/background/context-menus', () => {\\n return { init: initMenus }\\n })\\n- jest.doMock('../../../src/background/pdf-sniffer', () => {\\n+ jest.doMock('@/background/pdf-sniffer', () => {\\n return { init: initPdf }\\n })\\n- jest.doMock('../../../src/_helpers/check-update', () => {\\n+ jest.doMock('@/_helpers/check-update', () => {\\n return checkUpdate\\n })\\n- jest.doMock('../../../src/_helpers/browser-api', () => {\\n+ jest.doMock('@/_helpers/browser-api', () => {\\n return {\\n message,\\n storage,\\n@@ -41,13 +41,13 @@ describe('Initialization', () => {\\n }\\n })\\n \\n- require('../../../src/background/initialization')\\n+ require('@/background/initialization')\\n })\\n afterAll(() => {\\n browser.flush()\\n- jest.dontMock('../../../src/background/merge-config')\\n- jest.dontMock('../../../src/background/context-menus')\\n- jest.dontMock('../../../src/_helpers/browser-api')\\n+ jest.dontMock('@/background/merge-config')\\n+ jest.dontMock('@/background/context-menus')\\n+ jest.dontMock('@/_helpers/browser-api')\\n window.fetch = bakFetch\\n })\\n \\ndiff --git a/test/unit/background/merge-config.spec.ts b/test/unit/background/merge-config.spec.ts\\nindex 73c047d..c0dce26 100644\\n--- a/test/unit/background/merge-config.spec.ts\\n+++ b/test/unit/background/merge-config.spec.ts\\n@@ -1,5 +1,5 @@\\n-import { appConfigFactory, AppConfig, AppConfigMutable } from '../../../src/app-config'\\n-import mergeConfig from '../../../src/background/merge-config'\\n+import { appConfigFactory, AppConfig, AppConfigMutable } from '@/app-config'\\n+import mergeConfig from '@/background/merge-config'\\n import sinon from 'sinon'\\n \\n describe('Merge Config', () => {\\ndiff --git a/test/unit/background/pdf-sniffer.spec.ts b/test/unit/background/pdf-sniffer.spec.ts\\nindex a0219d2..bb7726f 100644\\n--- a/test/unit/background/pdf-sniffer.spec.ts\\n+++ b/test/unit/background/pdf-sniffer.spec.ts\\n@@ -1,5 +1,5 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n-import { init as initPdf } from '../../../src/background/pdf-sniffer'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n+import { init as initPdf } from '@/background/pdf-sniffer'\\n import sinon from 'sinon'\\n \\n function hasListenerPatch (fn) {\\ndiff --git a/test/unit/background/server.spec.ts b/test/unit/background/server.spec.ts\\nindex b8ef065..aa04525 100644\\n--- a/test/unit/background/server.spec.ts\\n+++ b/test/unit/background/server.spec.ts\\n@@ -1,5 +1,5 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n-import * as browserWrap from '../../../src/_helpers/browser-api'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n+import * as browserWrap from '@/_helpers/browser-api'\\n import sinon from 'sinon'\\n \\n describe('Server', () => {\\n@@ -13,26 +13,26 @@ describe('Server', () => {\\n browserWrap.openURL = openURL\\n \\n beforeAll(() => {\\n- jest.doMock('../../../src/_helpers/chs-to-chz', () => {\\n+ jest.doMock('@/_helpers/chs-to-chz', () => {\\n return { chsToChz }\\n })\\n- jest.doMock('../../../src/background/audio-manager', () => {\\n+ jest.doMock('@/background/audio-manager', () => {\\n return { play }\\n })\\n- jest.doMock('../../../src/_helpers/browser-api', () => {\\n+ jest.doMock('@/_helpers/browser-api', () => {\\n return browserWrap\\n })\\n- jest.doMock('../../../src/components/dictionaries/bing/engine.js', () => {\\n+ jest.doMock('@/components/dictionaries/bing/engine.js', () => {\\n return bingSearch\\n })\\n })\\n \\n afterAll(() => {\\n browser.flush()\\n- jest.dontMock('../../../src/_helpers/chs-to-chz')\\n- jest.dontMock('../../../src/background/audio-manager')\\n- jest.dontMock('../../../src/_helpers/browser-api')\\n- jest.dontMock('../../../src/components/dictionaries/bing/engine.js')\\n+ jest.dontMock('@/_helpers/chs-to-chz')\\n+ jest.dontMock('@/background/audio-manager')\\n+ jest.dontMock('@/_helpers/browser-api')\\n+ jest.dontMock('@/components/dictionaries/bing/engine.js')\\n })\\n \\n beforeEach(() => {\\n@@ -46,7 +46,7 @@ describe('Server', () => {\\n bingSearch.mockReset()\\n bingSearch.mockImplementation(() => Promise.resolve())\\n jest.resetModules()\\n- require('../../../src/background/server')\\n+ require('@/background/server')\\n })\\n \\n it('should properly init', () => {\\n\", \"diff --git a/src/_locales/common/messages.json b/src/_locales/common/messages.json\\nindex e8524ac..3a596d6 100644\\n--- a/src/_locales/common/messages.json\\n+++ b/src/_locales/common/messages.json\\n@@ -4,11 +4,21 @@\\n \\\"zh_CN\\\": \\\"\\u6dfb\\u52a0\\\",\\n \\\"zh_TW\\\": \\\"\\u65b0\\u589e\\\"\\n },\\n+ \\\"blacklist\\\": {\\n+ \\\"en\\\": \\\"Blacklist\\\",\\n+ \\\"zh_CN\\\": \\\"\\u9ed1\\u540d\\u5355\\\",\\n+ \\\"zh_TW\\\": \\\"\\u9ed1\\u540d\\u55ae\\\"\\n+ },\\n \\\"cancel\\\": {\\n \\\"en\\\": \\\"Cancel\\\",\\n \\\"zh_CN\\\": \\\"\\u53d6\\u6d88\\\",\\n \\\"zh_TW\\\": \\\"\\u53d6\\u6d88\\\"\\n },\\n+ \\\"changes_confirm\\\": {\\n+ \\\"en\\\": \\\"Changes not saved. Close anyway?\\\",\\n+ \\\"zh_CN\\\": \\\"\\u4fee\\u6539\\u672a\\u4fdd\\u5b58\\u3002\\u786e\\u8ba4\\u5173\\u95ed\\uff1f\\\",\\n+ \\\"zh_TW\\\": \\\"\\u4fee\\u6539\\u672a\\u4fdd\\u5b58\\u3002\\u78ba\\u5b9a\\u95dc\\u9589\\uff1f\\\"\\n+ },\\n \\\"confirm\\\": {\\n \\\"en\\\": \\\"Confirm\\\",\\n \\\"zh_CN\\\": \\\"\\u786e\\u8ba4\\\",\\n@@ -93,5 +103,10 @@\\n \\\"en\\\": \\\"words\\\",\\n \\\"zh_CN\\\": \\\"\\u4e2a\\\",\\n \\\"zh_TW\\\": \\\"\\u4e2a\\\"\\n+ },\\n+ \\\"whitelist\\\": {\\n+ \\\"en\\\": \\\"Whitelist\\\",\\n+ \\\"zh_CN\\\": \\\"\\u767d\\u540d\\u5355\\\",\\n+ \\\"zh_TW\\\": \\\"\\u767d\\u540d\\u55ae\\\"\\n }\\n }\\ndiff --git a/src/_locales/options/messages.json b/src/_locales/options/messages.json\\nindex ada2488..e7d699a 100644\\n--- a/src/_locales/options/messages.json\\n+++ b/src/_locales/options/messages.json\\n@@ -119,6 +119,11 @@\\n \\\"zh_CN\\\": \\\"\\u53cd\\u9988\\u95ee\\u9898\\\",\\n \\\"zh_TW\\\": \\\"\\u8edf\\u9ad4\\u4f7f\\u7528\\u7591\\u554f\\u548c\\u5efa\\u8a00\\\"\\n },\\n+ \\\"match_pattern_description\\\": {\\n+ \\\"en\\\": \\\"Specify URLs as match patterns. Examples. Empty fields will be removed.\\\",\\n+ \\\"zh_CN\\\": \\\"\\u7f51\\u5740\\u652f\\u6301\\u5339\\u914d\\u6a21\\u5f0f\\uff08\\u4f8b\\u5b50\\uff09\\u3002\\u7559\\u7a7a\\u4fdd\\u5b58\\u5373\\u53ef\\u6e05\\u9664\\u3002\\\",\\n+ \\\"zh_TW\\\": \\\"\\u7db2\\u5740\\u652f\\u63f4\\u5339\\u914d\\u6a21\\u5f0f\\uff08\\u4f8b\\u5b50\\uff09\\u3002\\u7559\\u7a7a\\u5132\\u5b58\\u5373\\u53ef\\u6e05\\u9664\\u3002\\\"\\n+ },\\n \\\"msg_updated\\\": {\\n \\\"en\\\": \\\"Successfully updated\\\",\\n \\\"zh_CN\\\": \\\"\\u8bbe\\u7f6e\\u5df2\\u66f4\\u65b0\\\",\\n@@ -319,6 +324,21 @@\\n \\\"zh_CN\\\": \\\"\\u5f00\\u542f\\u540e\\uff0c\\u672c\\u6269\\u5c55\\u4f1a\\u81ea\\u52a8\\u8bc6\\u522b\\u8f93\\u5165\\u6846\\u4ee5\\u53ca\\u5e38\\u89c1\\u7f16\\u8f91\\u5668\\uff0c\\u5982 CodeMirror\\u3001ACE \\u548c Monaco\\u3002\\\",\\n \\\"zh_TW\\\": \\\"\\u958b\\u555f\\u540e\\uff0c\\u672c\\u7a0b\\u5f0f\\u6703\\u81ea\\u52d5\\u8b58\\u5225\\u8f38\\u5165\\u6846\\u4ee5\\u53ca\\u5e38\\u898b\\u7de8\\u8f2f\\u5668\\uff0c\\u5982 CodeMirror\\u3001ACE \\u548c Monaco\\u3002\\\"\\n },\\n+ \\\"opt_pdf_blackwhitelist_help\\\": {\\n+ \\\"en\\\": \\\"Blacklisted PDF links will not jump to Saladict PDF Viewer.\\\",\\n+ \\\"zh_CN\\\": \\\"\\u9ed1\\u540d\\u5355\\u5339\\u914d\\u7684 PDF \\u94fe\\u63a5\\u5c06\\u4e0d\\u4f1a\\u8df3\\u8f6c\\u5230 Saladict \\u6253\\u5f00\\u3002\\\",\\n+ \\\"zh_TW\\\": \\\"\\u9ed1\\u540d\\u55ae\\u5339\\u914d\\u7684 PDF \\u9023\\u7d50\\u5c07\\u4e0d\\u6703\\u8df3\\u8f49\\u5230 Saladict \\u958b\\u555f\\u3002\\\"\\n+ },\\n+ \\\"opt_pdf_sniff\\\": {\\n+ \\\"en\\\": \\\"Enable PDF Sniffer\\\",\\n+ \\\"zh_CN\\\": \\\"\\u9ed8\\u8ba4\\u7528\\u672c\\u6269\\u5c55\\u6d4f\\u89c8 PDF\\\",\\n+ \\\"zh_TW\\\": \\\"\\u4f7f\\u7528\\u672c\\u61c9\\u7528\\u7a0b\\u5f0f\\u700f\\u89bd PDF\\\"\\n+ },\\n+ \\\"opt_pdf_sniff_help\\\": {\\n+ \\\"en\\\": \\\"If turned on\\uff0c PDF links will be automatically captured.\\\",\\n+ \\\"zh_CN\\\": \\\"\\u5f00\\u542f\\u540e\\u6240\\u6709 PDF \\u94fe\\u63a5\\u5c06\\u81ea\\u52a8\\u8df3\\u8f6c\\u5230\\u672c\\u6269\\u5c55\\u6253\\u5f00\\uff08\\u5305\\u62ec\\u672c\\u5730\\uff0c\\u5982\\u679c\\u5728\\u6269\\u5c55\\u7ba1\\u7406\\u9875\\u9762\\u52fe\\u9009\\u4e86\\u5141\\u8bb8\\uff09\\u3002\\\",\\n+ \\\"zh_TW\\\": \\\"\\u958b\\u555f\\u5f8c\\u6240\\u6709 PDF \\u9023\\u7d50\\u5c07\\u81ea\\u52d5\\u8df3\\u8f49\\u5230\\u672c\\u64f4\\u5145\\u5957\\u4ef6\\u958b\\u555f\\uff08\\u5305\\u62ec\\u672c\\u5730\\uff0c\\u5982\\u679c\\u5728\\u64f4\\u5145\\u5957\\u4ef6\\u7ba1\\u7406\\u9801\\u9762\\u52fe\\u9078\\u4e86\\u5141\\u8a31\\uff09\\u3002\\\"\\n+ },\\n \\\"opt_profile_change\\\": {\\n \\\"en\\\": \\\"This option may change base on \\\\\\\"Profile\\\\\\\".\\\",\\n \\\"zh_CN\\\": \\\"\\u6b64\\u9009\\u9879\\u4f1a\\u56e0\\u300c\\u60c5\\u666f\\u6a21\\u5f0f\\u300d\\u800c\\u6539\\u53d8\\u3002\\\",\\n@@ -329,6 +349,16 @@\\n \\\"zh_CN\\\": \\\"\\u8f93\\u5165\\u65f6\\u663e\\u793a\\u5019\\u9009\\\",\\n \\\"zh_TW\\\": \\\"\\u8f38\\u5165\\u6642\\u986f\\u793a\\u5019\\u9078\\\"\\n },\\n+ \\\"opt_sel_blackwhitelist\\\": {\\n+ \\\"en\\\": \\\"Selection Black/White List\\\",\\n+ \\\"zh_CN\\\": \\\"\\u5212\\u8bcd\\u9ed1\\u767d\\u540d\\u5355\\\",\\n+ \\\"zh_TW\\\": \\\"\\u9078\\u8a5e\\u9ed1\\u767d\\u540d\\u55ae\\\"\\n+ },\\n+ \\\"opt_sel_blackwhitelist_help\\\": {\\n+ \\\"en\\\": \\\"Saladict will not react to selection in blacklisted pages.\\\",\\n+ \\\"zh_CN\\\": \\\"\\u9ed1\\u540d\\u5355\\u5339\\u914d\\u7684\\u9875\\u9762 Saladict \\u5c06\\u4e0d\\u4f1a\\u54cd\\u5e94\\u9f20\\u6807\\u5212\\u8bcd\\u3002\\\",\\n+ \\\"zh_TW\\\": \\\"\\u9ed1\\u540d\\u55ae\\u5339\\u914d\\u7684\\u9801\\u9762 Saladict \\u5c07\\u4e0d\\u6703\\u97ff\\u61c9\\u6ed1\\u9f20\\u5283\\u8a5e\\u3002\\\"\\n+ },\\n \\\"opt_sel_lang\\\": {\\n \\\"en\\\": \\\"Selection Languages\\\",\\n \\\"zh_CN\\\": \\\"\\u5212\\u8bcd\\u8bed\\u8a00\\\",\\ndiff --git a/src/options/components/options/BlackWhiteList/index.tsx b/src/options/components/options/BlackWhiteList/index.tsx\\nnew file mode 100644\\nindex 0000000..52708dd\\n--- /dev/null\\n+++ b/src/options/components/options/BlackWhiteList/index.tsx\\n@@ -0,0 +1,69 @@\\n+import React from 'react'\\n+import { Props } from '../typings'\\n+import { formItemLayout } from '../helpers'\\n+import MatchPatternModal from '../../MatchPatternModal'\\n+\\n+import { FormComponentProps } from 'antd/lib/form'\\n+import { Form, Button } from 'antd'\\n+\\n+export type BlackWhiteListProps = Props & FormComponentProps\\n+\\n+interface BlackWhiteListState {\\n+ editingArea: '' | 'pdfWhitelist' | 'pdfBlacklist' | 'whitelist' | 'blacklist'\\n+}\\n+\\n+export class BlackWhiteList extends React.Component {\\n+ constructor (props: BlackWhiteListProps) {\\n+ super(props)\\n+ this.state = {\\n+ editingArea: ''\\n+ }\\n+ }\\n+\\n+ closeModal = () => {\\n+ this.setState({ editingArea: '' })\\n+ }\\n+\\n+ render () {\\n+ const { t, config } = this.props\\n+\\n+ return (\\n+
\\n+ \\n+ this.setState({ editingArea: 'blacklist' })}\\n+ >{t('common:blacklist')}\\n+ this.setState({ editingArea: 'whitelist' })}\\n+ >{t('common:whitelist')}\\n+ \\n+ \\n+ this.setState({ editingArea: 'pdfBlacklist' })}\\n+ >PDF {t('common:blacklist')}\\n+ this.setState({ editingArea: 'pdfWhitelist' })}\\n+ >PDF {t('common:whitelist')}\\n+ \\n+ \\n+ \\n+ )\\n+ }\\n+}\\n+\\n+export default BlackWhiteList\\ndiff --git a/src/options/components/options/PDF/index.tsx b/src/options/components/options/PDF/index.tsx\\nnew file mode 100644\\nindex 0000000..3e7772d\\n--- /dev/null\\n+++ b/src/options/components/options/PDF/index.tsx\\n@@ -0,0 +1,72 @@\\n+import React from 'react'\\n+import { Props } from '../typings'\\n+import { updateConfigOrProfile, formItemLayout } from '../helpers'\\n+import MatchPatternModal from '../../MatchPatternModal'\\n+\\n+import { FormComponentProps } from 'antd/lib/form'\\n+import { Form, Switch, Button } from 'antd'\\n+\\n+export type PDFProps = Props & FormComponentProps\\n+\\n+interface PDFState {\\n+ editingArea: '' | 'pdfWhitelist' | 'pdfBlacklist'\\n+}\\n+\\n+export class PDF extends React.Component {\\n+ constructor (props: PDFProps) {\\n+ super(props)\\n+\\n+ this.state = {\\n+ editingArea: ''\\n+ }\\n+ }\\n+\\n+ closeModal = () => {\\n+ this.setState({ editingArea: '' })\\n+ }\\n+\\n+ render () {\\n+ const { t, config } = this.props\\n+ const { getFieldDecorator } = this.props.form\\n+\\n+ return (\\n+
\\n+ {\\n+ getFieldDecorator('config#pdfSniff', {\\n+ initialValue: config.pdfSniff,\\n+ valuePropName: 'checked',\\n+ })(\\n+ \\n+ )\\n+ }\\n+ \\n+ this.setState({ editingArea: 'pdfBlacklist' })}\\n+ >PDF {t('common:blacklist')}\\n+ this.setState({ editingArea: 'pdfWhitelist' })}\\n+ >PDF {t('common:whitelist')}\\n+ \\n+ \\n+ \\n+ )\\n+ }\\n+}\\n+\\n+export default Form.create({\\n+ onValuesChange: updateConfigOrProfile\\n+})(PDF)\\n\", \"diff --git a/server/src/routes/course/index.ts b/server/src/routes/course/index.ts\\nindex 557f5fb..bc0e490 100644\\n--- a/server/src/routes/course/index.ts\\n+++ b/server/src/routes/course/index.ts\\n@@ -209,7 +209,7 @@ function addStudentApi(router: Router, logger: ILogger) {\\n router.post('/student/:githubId/status', ...mentorValidators, updateStudentStatus(logger));\\n router.post('/student/:githubId/status-self', courseGuard, selfUpdateStudentStatus(logger));\\n router.get('/student/:githubId/score', courseGuard, getScoreByStudent(logger));\\n- router.post('/student/:githubId/certificate', courseManagerGuard, ...validators, postStudentCertificate(logger));\\n+ router.post('/student/:githubId/certificate', courseManagerGuard, validateGithubId, postStudentCertificate(logger));\\n \\n router.get('/students', courseSupervisorGuard, getStudents(logger));\\n router.get('/students/csv', courseSupervisorGuard, getStudentsCsv(logger));\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"78c446cbea61af2268b4c4da03a9ad4283f10049\", \"8246d024f21d93cc092e19bede5f7b3a5325c8dc\", \"9b1c0fc20b614513384a1e562317dbf076eb8ef0\", \"33c25b2f59c931a7f4af994365522221a7821dca\"]"},"types":{"kind":"string","value":"[\"build\", \"refactor\", \"feat\", \"fix\"]"}}},{"rowIdx":946,"cells":{"commit_message":{"kind":"string","value":"update the formatting for python integration example,add unit test for query API,unset DOCKER_HOST set to swarm by jenkins\n\n- fixes issue where old images are pushed to registry,Handle different events."},"diff":{"kind":"string","value":"[\"diff --git a/website/docs/integration/python.md b/website/docs/integration/python.md\\nindex 064cae3..b6b720d 100644\\n--- a/website/docs/integration/python.md\\n+++ b/website/docs/integration/python.md\\n@@ -13,6 +13,7 @@ header = \\\"All notable changes to this project will be documented in this file.\\\"\\n body = \\\"...\\\"\\n footer = \\\"\\\"\\n # see [changelog] section for more keys\\n+\\n [tool.git-cliff.git]\\n conventional_commits = true\\n commit_parsers = []\\n\", \"diff --git a/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java b/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java\\nindex 2d2d084..38261ad 100644\\n--- a/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java\\n+++ b/gateway/src/test/java/io/camunda/zeebe/gateway/api/util/StubbedBrokerClient.java\\n@@ -25,6 +25,7 @@ import java.util.HashMap;\\n import java.util.List;\\n import java.util.Map;\\n import java.util.concurrent.CompletableFuture;\\n+import java.util.concurrent.TimeUnit;\\n import java.util.function.Consumer;\\n \\n public final class StubbedBrokerClient implements BrokerClient {\\n@@ -67,7 +68,15 @@ public final class StubbedBrokerClient implements BrokerClient {\\n @Override\\n public CompletableFuture> sendRequestWithRetry(\\n final BrokerRequest request, final Duration requestTimeout) {\\n- throw new UnsupportedOperationException(\\\"not implemented\\\");\\n+ final CompletableFuture> result = new CompletableFuture<>();\\n+\\n+ sendRequestWithRetry(\\n+ request,\\n+ (key, response) ->\\n+ result.complete(new BrokerResponse<>(response, Protocol.decodePartitionId(key), key)),\\n+ result::completeExceptionally);\\n+\\n+ return result.orTimeout(requestTimeout.toNanos(), TimeUnit.NANOSECONDS);\\n }\\n \\n @Override\\ndiff --git a/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryApiTest.java b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryApiTest.java\\nnew file mode 100644\\nindex 0000000..ec9ec80\\n--- /dev/null\\n+++ b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryApiTest.java\\n@@ -0,0 +1,91 @@\\n+/*\\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\\n+ * one or more contributor license agreements. See the NOTICE file distributed\\n+ * with this work for additional information regarding copyright ownership.\\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\\n+ * except in compliance with the Zeebe Community License 1.1.\\n+ */\\n+package io.camunda.zeebe.gateway.query;\\n+\\n+import static org.assertj.core.api.Assertions.assertThat;\\n+\\n+import io.camunda.zeebe.gateway.api.util.GatewayTest;\\n+import io.camunda.zeebe.gateway.cmd.BrokerErrorException;\\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerError;\\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerErrorResponse;\\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerResponse;\\n+import io.camunda.zeebe.gateway.query.impl.QueryApiImpl;\\n+import io.camunda.zeebe.protocol.Protocol;\\n+import io.camunda.zeebe.protocol.record.ErrorCode;\\n+import java.time.Duration;\\n+import java.util.concurrent.CompletionStage;\\n+import java.util.concurrent.ExecutionException;\\n+import org.junit.Test;\\n+import org.junit.runner.RunWith;\\n+import org.junit.runners.Parameterized;\\n+import org.junit.runners.Parameterized.Parameter;\\n+import org.junit.runners.Parameterized.Parameters;\\n+\\n+@RunWith(Parameterized.class)\\n+public final class QueryApiTest extends GatewayTest {\\n+ @Parameter(0)\\n+ public String name;\\n+\\n+ @Parameter(1)\\n+ public Querier querier;\\n+\\n+ @Parameters(name = \\\"{index}: {0}\\\")\\n+ public static Object[][] queries() {\\n+ return new Object[][] {\\n+ new Object[] {\\\"getBpmnProcessIdForProcess\\\", (Querier) QueryApi::getBpmnProcessIdFromProcess},\\n+ new Object[] {\\n+ \\\"getBpmnProcessIdForProcessInstance\\\",\\n+ (Querier) QueryApi::getBpmnProcessIdFromProcessInstance\\n+ },\\n+ new Object[] {\\\"getBpmnProcessIdForProcessJob\\\", (Querier) QueryApi::getBpmnProcessIdFromJob},\\n+ };\\n+ }\\n+\\n+ @Test\\n+ public void shouldGetBpmnProcessId() {\\n+ // given\\n+ final var key = Protocol.encodePartitionId(1, 1);\\n+ final var api = new QueryApiImpl(brokerClient);\\n+ final var timeout = Duration.ofSeconds(5);\\n+ final var stub = new QueryStub(new BrokerResponse<>(\\\"myProcess\\\", 1, 1));\\n+ stub.registerWith(brokerClient);\\n+\\n+ // when\\n+ final var result = querier.query(api, key, timeout);\\n+\\n+ // then\\n+ assertThat(result).succeedsWithin(timeout).isEqualTo(\\\"myProcess\\\");\\n+ }\\n+\\n+ @Test\\n+ public void shouldCompleteExceptionallyOnError() {\\n+ // given\\n+ final var key = Protocol.encodePartitionId(1, 1);\\n+ final var api = new QueryApiImpl(brokerClient);\\n+ final var timeout = Duration.ofSeconds(5);\\n+ final var stub =\\n+ new QueryStub(\\n+ new BrokerErrorResponse<>(\\n+ new BrokerError(ErrorCode.PARTITION_LEADER_MISMATCH, \\\"Leader mismatch\\\")));\\n+ stub.registerWith(brokerClient);\\n+\\n+ // when\\n+ final var result = querier.query(api, key, timeout);\\n+\\n+ // then\\n+ assertThat(result)\\n+ .failsWithin(timeout)\\n+ .withThrowableOfType(ExecutionException.class)\\n+ .havingRootCause()\\n+ .isInstanceOf(BrokerErrorException.class);\\n+ }\\n+\\n+ private interface Querier {\\n+ CompletionStage query(final QueryApi api, final long key, final Duration timeout);\\n+ }\\n+}\\ndiff --git a/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryStub.java b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryStub.java\\nnew file mode 100644\\nindex 0000000..2f8334e\\n--- /dev/null\\n+++ b/gateway/src/test/java/io/camunda/zeebe/gateway/query/QueryStub.java\\n@@ -0,0 +1,31 @@\\n+/*\\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\\n+ * one or more contributor license agreements. See the NOTICE file distributed\\n+ * with this work for additional information regarding copyright ownership.\\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\\n+ * except in compliance with the Zeebe Community License 1.1.\\n+ */\\n+package io.camunda.zeebe.gateway.query;\\n+\\n+import io.camunda.zeebe.gateway.api.util.StubbedBrokerClient;\\n+import io.camunda.zeebe.gateway.api.util.StubbedBrokerClient.RequestStub;\\n+import io.camunda.zeebe.gateway.impl.broker.response.BrokerResponse;\\n+import io.camunda.zeebe.gateway.query.impl.BrokerExecuteQuery;\\n+\\n+final class QueryStub implements RequestStub> {\\n+ private final BrokerResponse response;\\n+\\n+ public QueryStub(final BrokerResponse response) {\\n+ this.response = response;\\n+ }\\n+\\n+ @Override\\n+ public void registerWith(final StubbedBrokerClient gateway) {\\n+ gateway.registerHandler(BrokerExecuteQuery.class, this);\\n+ }\\n+\\n+ @Override\\n+ public BrokerResponse handle(final BrokerExecuteQuery request) throws Exception {\\n+ return response;\\n+ }\\n+}\\n\", \"diff --git a/.ci/docker.dsl b/.ci/docker.dsl\\nindex 4768cb8..9f6a4c9 100644\\n--- a/.ci/docker.dsl\\n+++ b/.ci/docker.dsl\\n@@ -8,6 +8,9 @@ def dockerHubUpload =\\n '''\\\\\\n #!/bin/bash -xeu\\n \\n+# clear docker host env set by jenkins job\\n+unset DOCKER_HOST\\n+\\n VERSION=${RELEASE_VERSION}\\n \\n if [ \\\"${RELEASE_VERSION}\\\" = \\\"SNAPSHOT\\\" ]; then\\n@@ -26,9 +29,6 @@ docker login --username ${DOCKER_HUB_USERNAME} --password ${DOCKER_HUB_PASSWORD}\\n docker push camunda/zeebe:${RELEASE_VERSION}\\n \\n if [ \\\"${IS_LATEST}\\\" = \\\"true\\\" ]; then\\n- # to make sure we can tag latest, there were problems before\\n- docker rmi camunda/zeebe:latest\\n-\\n docker tag -f camunda/zeebe:${RELEASE_VERSION} camunda/zeebe:latest\\n docker push camunda/zeebe:latest\\n fi\\n\", \"diff --git a/src/notebook/epics/kernel-launch.js b/src/notebook/epics/kernel-launch.js\\nindex 9075d7c..9f16e67 100644\\n--- a/src/notebook/epics/kernel-launch.js\\n+++ b/src/notebook/epics/kernel-launch.js\\n@@ -113,6 +113,12 @@ export function newKernelObservable(kernelSpec: KernelInfo, cwd: string) {\\n observer.error({ type: 'ERROR', payload: error, err: true });\\n observer.complete();\\n });\\n+ spawn.on('exit', () => {\\n+ observer.complete();\\n+ });\\n+ spawn.on('disconnect', () => {\\n+ observer.complete();\\n+ });\\n });\\n });\\n }\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"3ee672483790ec71c700907a6e93af4698492026\", \"bed86aeae8dad2dd6371635cd24bf8ef3db80361\", \"8b18a58969ed2adf2df2a8bfe91aedacad3868f5\", \"a280a52c8309465276c3509848ddcddbe19732b6\"]"},"types":{"kind":"string","value":"[\"docs\", \"test\", \"ci\", \"fix\"]"}}},{"rowIdx":947,"cells":{"commit_message":{"kind":"string","value":"add canonical `_name` to edge packages,correctly read new last flushed index,add flag to wait for workflow instance result\n\n- with the flag withResult the create instance command will wait for the\n workflow to complete\n- optional a list of variable names can be specified to limit the fetched\n variables,add link to roadmap"},"diff":{"kind":"string","value":"[\"diff --git a/scripts/bump-edge.ts b/scripts/bump-edge.ts\\nindex e92e3c9..0b7a11a 100644\\n--- a/scripts/bump-edge.ts\\n+++ b/scripts/bump-edge.ts\\n@@ -53,6 +53,7 @@ async function loadWorkspace (dir: string) {\\n }\\n \\n const rename = (from: string, to: string) => {\\n+ find(from).data._name = find(from).data.name\\n find(from).data.name = to\\n for (const pkg of packages) {\\n pkg.updateDeps((dep) => {\\n\", \"diff --git a/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java b/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java\\nindex 69b06b6..a4fcb77 100644\\n--- a/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java\\n+++ b/restore/src/main/java/io/camunda/zeebe/restore/PartitionRestoreService.java\\n@@ -112,7 +112,7 @@ public class PartitionRestoreService {\\n SegmentedJournal.builder()\\n .withDirectory(dataDirectory.toFile())\\n .withName(partition.name())\\n- .withLastWrittenIndex(-1)\\n+ .withLastFlushedIndex(-1)\\n .build()) {\\n \\n resetJournal(checkpointPosition, journal);\\n\", \"diff --git a/clients/zbctl/cmd/createInstance.go b/clients/zbctl/cmd/createInstance.go\\nindex 016f115..85ac0be 100644\\n--- a/clients/zbctl/cmd/createInstance.go\\n+++ b/clients/zbctl/cmd/createInstance.go\\n@@ -15,13 +15,15 @@ package cmd\\n \\n import (\\n \\t\\\"github.com/zeebe-io/zeebe/clients/go/commands\\\"\\n+\\t\\\"strings\\\"\\n \\n \\t\\\"github.com/spf13/cobra\\\"\\n )\\n \\n var (\\n-\\tcreateInstanceVersionFlag int32\\n-\\tcreateInstanceVariablesFlag string\\n+\\tcreateInstanceVersionFlag int32\\n+\\tcreateInstanceVariablesFlag string\\n+\\tcreateInstanceWithResultFlag []string\\n )\\n \\n var createInstanceCmd = &cobra.Command{\\n@@ -39,12 +41,29 @@ var createInstanceCmd = &cobra.Command{\\n \\t\\t\\treturn err\\n \\t\\t}\\n \\n-\\t\\tresponse, err := zbCmd.Send()\\n-\\t\\tif err != nil {\\n-\\t\\t\\treturn err\\n-\\t\\t}\\n+\\t\\tif createInstanceWithResultFlag == nil {\\n+\\t\\t\\tresponse, err := zbCmd.Send()\\n+\\t\\t\\tif err != nil {\\n+\\t\\t\\t\\treturn err\\n+\\t\\t\\t}\\n+\\n+\\t\\t\\treturn printJson(response)\\n+\\t\\t} else {\\n+\\t\\t\\tvariableNames := []string{}\\n+\\t\\t\\tfor _, variableName := range createInstanceWithResultFlag {\\n+\\t\\t\\t\\ttrimedVariableName := strings.TrimSpace(variableName)\\n+\\t\\t\\t\\tif trimedVariableName != \\\"\\\" {\\n+\\t\\t\\t\\t\\tvariableNames = append(variableNames, trimedVariableName)\\n+\\t\\t\\t\\t}\\n+\\t\\t\\t}\\n+\\t\\t\\tresponse, err := zbCmd.WithResult().FetchVariables(variableNames...).Send()\\n+\\t\\t\\tif err != nil {\\n+\\t\\t\\t\\treturn err\\n+\\t\\t\\t}\\n+\\n+\\t\\t\\treturn printJson(response)\\n \\n-\\t\\treturn printJson(response)\\n+\\t\\t}\\n \\t},\\n }\\n \\n@@ -58,4 +77,11 @@ func init() {\\n \\tcreateInstanceCmd.\\n \\t\\tFlags().\\n \\t\\tInt32Var(&createInstanceVersionFlag, \\\"version\\\", commands.LatestVersion, \\\"Specify version of workflow which should be executed.\\\")\\n+\\n+\\tcreateInstanceCmd.\\n+\\t\\tFlags().\\n+\\t\\tStringSliceVar(&createInstanceWithResultFlag, \\\"withResult\\\", nil, \\\"Specify to await result of workflow, optional a list of variable names can be provided to limit the returned variables\\\")\\n+\\n+\\t// hack to use --withResult without values\\n+\\tcreateInstanceCmd.Flag(\\\"withResult\\\").NoOptDefVal = \\\" \\\"\\n }\\n\", \"diff --git a/packages/plugin-core/README.md b/packages/plugin-core/README.md\\nindex 3c25c9b..c7506d4 100644\\n--- a/packages/plugin-core/README.md\\n+++ b/packages/plugin-core/README.md\\n@@ -187,6 +187,10 @@ When the workspace opens, it will show dialogue to install the recommended exten\\n \\n See [[FAQ]] to answers for common questions.\\n \\n+# Roadmap\\n+\\n+Check out our [public roadmap](https://github.com/orgs/dendronhq/projects/1) to see the features we're working on and to vote for what you want to see next. \\n+\\n \\n # Contributing\\n \\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"573f87edf9bdc19c9c4c3a978fad6ed3ce788f5f\", \"5ffc5794808647de14f945141692be26ad143006\", \"f3107f1a8eb124b55e775d23416540f49204a19e\", \"94202f01e44c58bee4419044f8a18ac5f1a50dff\"]"},"types":{"kind":"string","value":"[\"build\", \"fix\", \"feat\", \"docs\"]"}}},{"rowIdx":948,"cells":{"commit_message":{"kind":"string","value":"Remove hasmany and belongsto from context menu\n\nSigned-off-by: Pranav C <61551451+pranavxc@users.noreply.github.com>,removing automatic page push on nav,README,build improvements"},"diff":{"kind":"string","value":"[\"diff --git a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\nindex 5bc6f67..aaa297c 100644\\n--- a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\n+++ b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\n@@ -261,37 +261,7 @@\\n :size=\\\"size\\\"\\n @input=\\\"loadTableData\\\"\\n />\\n- \\n \\n- \\n- \\n- \\n
\\n \\n Delete Selected Rows\\n \\n \\n- \\n- \\n- \\n- \\n \\n \\n Delete Selected Rows\\n \\n \\n-