{ // 获取包含Hugging Face文本的span元素 const spans = link.querySelectorAll('span.whitespace-nowrap, span.hidden.whitespace-nowrap'); spans.forEach(span => { if (span.textContent && span.textContent.trim().match(/Hugging\s*Face/i)) { span.textContent = 'AI快站'; } }); }); // 替换logo图片的alt属性 document.querySelectorAll('img[alt*="Hugging"], img[alt*="Face"]').forEach(img => { if (img.alt.match(/Hugging\s*Face/i)) { img.alt = 'AI快站 logo'; } }); } // 替换导航栏中的链接 function replaceNavigationLinks() { // 已替换标记,防止重复运行 if (window._navLinksReplaced) { return; } // 已经替换过的链接集合,防止重复替换 const replacedLinks = new Set(); // 只在导航栏区域查找和替换链接 const headerArea = document.querySelector('header') || document.querySelector('nav'); if (!headerArea) { return; } // 在导航区域内查找链接 const navLinks = headerArea.querySelectorAll('a'); navLinks.forEach(link => { // 如果已经替换过,跳过 if (replacedLinks.has(link)) return; const linkText = link.textContent.trim(); const linkHref = link.getAttribute('href') || ''; // 替换Spaces链接 - 仅替换一次 if ( (linkHref.includes('/spaces') || linkHref === '/spaces' || linkText === 'Spaces' || linkText.match(/^s*Spacess*$/i)) && linkText !== 'OCR模型免费转Markdown' && linkText !== 'OCR模型免费转Markdown' ) { link.textContent = 'OCR模型免费转Markdown'; link.href = 'https://fast360.xyz'; link.setAttribute('target', '_blank'); link.setAttribute('rel', 'noopener noreferrer'); replacedLinks.add(link); } // 删除Posts链接 else if ( (linkHref.includes('/posts') || linkHref === '/posts' || linkText === 'Posts' || linkText.match(/^s*Postss*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } // 替换Docs链接 - 仅替换一次 else if ( (linkHref.includes('/docs') || linkHref === '/docs' || linkText === 'Docs' || linkText.match(/^s*Docss*$/i)) && linkText !== '模型下载攻略' ) { link.textContent = '模型下载攻略'; link.href = '/'; replacedLinks.add(link); } // 删除Enterprise链接 else if ( (linkHref.includes('/enterprise') || linkHref === '/enterprise' || linkText === 'Enterprise' || linkText.match(/^s*Enterprises*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } }); // 查找可能嵌套的Spaces和Posts文本 const textNodes = []; function findTextNodes(element) { if (element.nodeType === Node.TEXT_NODE) { const text = element.textContent.trim(); if (text === 'Spaces' || text === 'Posts' || text === 'Enterprise') { textNodes.push(element); } } else { for (const child of element.childNodes) { findTextNodes(child); } } } // 只在导航区域内查找文本节点 findTextNodes(headerArea); // 替换找到的文本节点 textNodes.forEach(node => { const text = node.textContent.trim(); if (text === 'Spaces') { node.textContent = node.textContent.replace(/Spaces/g, 'OCR模型免费转Markdown'); } else if (text === 'Posts') { // 删除Posts文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } else if (text === 'Enterprise') { // 删除Enterprise文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } }); // 标记已替换完成 window._navLinksReplaced = true; } // 替换代码区域中的域名 function replaceCodeDomains() { // 特别处理span.hljs-string和span.njs-string元素 document.querySelectorAll('span.hljs-string, span.njs-string, span[class*="hljs-string"], span[class*="njs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换hljs-string类的span中的域名(移除多余的转义符号) document.querySelectorAll('span.hljs-string, span[class*="hljs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换pre和code标签中包含git clone命令的域名 document.querySelectorAll('pre, code').forEach(element => { if (element.textContent && element.textContent.includes('git clone')) { const text = element.innerHTML; if (text.includes('huggingface.co')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 处理特定的命令行示例 document.querySelectorAll('pre, code').forEach(element => { const text = element.innerHTML; if (text.includes('huggingface.co')) { // 针对git clone命令的专门处理 if (text.includes('git clone') || text.includes('GIT_LFS_SKIP_SMUDGE=1')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 特别处理模型下载页面上的代码片段 document.querySelectorAll('.flex.border-t, .svelte_hydrator, .inline-block').forEach(container => { const content = container.innerHTML; if (content && content.includes('huggingface.co')) { container.innerHTML = content.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 特别处理模型仓库克隆对话框中的代码片段 try { // 查找包含"Clone this model repository"标题的对话框 const cloneDialog = document.querySelector('.svelte_hydration_boundary, [data-target="MainHeader"]'); if (cloneDialog) { // 查找对话框中所有的代码片段和命令示例 const codeElements = cloneDialog.querySelectorAll('pre, code, span'); codeElements.forEach(element => { if (element.textContent && element.textContent.includes('huggingface.co')) { if (element.innerHTML.includes('huggingface.co')) { element.innerHTML = element.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { element.textContent = element.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); } // 更精确地定位克隆命令中的域名 document.querySelectorAll('[data-target]').forEach(container => { const codeBlocks = container.querySelectorAll('pre, code, span.hljs-string'); codeBlocks.forEach(block => { if (block.textContent && block.textContent.includes('huggingface.co')) { if (block.innerHTML.includes('huggingface.co')) { block.innerHTML = block.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { block.textContent = block.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); }); } catch (e) { // 错误处理但不打印日志 } } // 当DOM加载完成后执行替换 if (document.readyState === 'loading') { document.addEventListener('DOMContentLoaded', () => { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); }); } else { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); } // 增加一个MutationObserver来处理可能的动态元素加载 const observer = new MutationObserver(mutations => { // 检查是否导航区域有变化 const hasNavChanges = mutations.some(mutation => { // 检查是否存在header或nav元素变化 return Array.from(mutation.addedNodes).some(node => { if (node.nodeType === Node.ELEMENT_NODE) { // 检查是否是导航元素或其子元素 if (node.tagName === 'HEADER' || node.tagName === 'NAV' || node.querySelector('header, nav')) { return true; } // 检查是否在导航元素内部 let parent = node.parentElement; while (parent) { if (parent.tagName === 'HEADER' || parent.tagName === 'NAV') { return true; } parent = parent.parentElement; } } return false; }); }); // 只在导航区域有变化时执行替换 if (hasNavChanges) { // 重置替换状态,允许再次替换 window._navLinksReplaced = false; replaceHeaderBranding(); replaceNavigationLinks(); } }); // 开始观察document.body的变化,包括子节点 if (document.body) { observer.observe(document.body, { childList: true, subtree: true }); } else { document.addEventListener('DOMContentLoaded', () => { observer.observe(document.body, { childList: true, subtree: true }); }); } })(); \\n \\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"3346331a963766c8193170fb130adad2e658ada2\", \"7f611e4208d0cc36c635eb0641a4b3258999ae8d\", \"4ab28fc2e63e975a0c77e18ae644f34fa5f8771a\", \"c3b5dc77ff3d89d389f6f3a868b17d0a8ca63074\"]"},"types":{"kind":"string","value":"[\"refactor\", \"docs\", \"build\", \"test\"]"}}},{"rowIdx":1022,"cells":{"commit_message":{"kind":"string","value":"cancel in-progress dep update jobs when a new one arrives [skip ci],remove writers from interface,add postgres-driver typings,better tested publishing flow"},"diff":{"kind":"string","value":"[\"diff --git a/.github/workflows/update-deps.yml b/.github/workflows/update-deps.yml\\nindex 3a71e29..25f6f27 100644\\n--- a/.github/workflows/update-deps.yml\\n+++ b/.github/workflows/update-deps.yml\\n@@ -4,6 +4,11 @@ on:\\n # run every 24 hours at midnight\\n - cron: \\\"0 */24 * * *\\\"\\n workflow_dispatch:\\n+\\n+concurrency:\\n+ group: ${{ github.repository }}-${{ github.head_ref || github.sha }}-${{ github.workflow }}\\n+ cancel-in-progress: true\\n+\\n jobs:\\n generate_updates:\\n runs-on: ubuntu-latest\\n\", \"diff --git a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\\nindex 91f1b41..eb4b9a8 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\\n@@ -81,8 +81,6 @@ public class Engine implements RecordProcessor {\\n \\n engineContext.setLifecycleListeners(typedRecordProcessors.getLifecycleListeners());\\n recordProcessorMap = typedRecordProcessors.getRecordProcessorMap();\\n-\\n- engineContext.setWriters(writers);\\n }\\n \\n @Override\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\\nindex a8e5538..a27b6e6 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\\n@@ -15,7 +15,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.StreamProcessorListene\\n import io.camunda.zeebe.engine.processing.streamprocessor.TypedRecordProcessorFactory;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriter;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\\n import io.camunda.zeebe.engine.state.EventApplier;\\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\\n import java.util.Collections;\\n@@ -34,7 +33,6 @@ public final class EngineContext {\\n private final TypedRecordProcessorFactory typedRecordProcessorFactory;\\n private List lifecycleListeners = Collections.EMPTY_LIST;\\n private StreamProcessorListener streamProcessorListener;\\n- private Writers writers;\\n \\n public EngineContext(\\n final int partitionId,\\n@@ -102,12 +100,4 @@ public final class EngineContext {\\n public void setStreamProcessorListener(final StreamProcessorListener streamProcessorListener) {\\n this.streamProcessorListener = streamProcessorListener;\\n }\\n-\\n- public Writers getWriters() {\\n- return writers;\\n- }\\n-\\n- public void setWriters(final Writers writers) {\\n- this.writers = writers;\\n- }\\n }\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\\nindex f30c7cc..834b421 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\\n@@ -8,7 +8,6 @@\\n package io.camunda.zeebe.engine.api;\\n \\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\\n import io.camunda.zeebe.logstreams.log.LogStream;\\n \\n@@ -27,11 +26,6 @@ public interface ReadonlyStreamProcessorContext {\\n LegacyTypedStreamWriter getLogStreamWriter();\\n \\n /**\\n- * @return the specific writers, like command, response, etc\\n- */\\n- Writers getWriters();\\n-\\n- /**\\n * @return the state, where the data is stored during processing\\n */\\n MutableZeebeState getZeebeState();\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\\nindex 844e487..49fd8e2 100755\\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\\n@@ -346,7 +346,6 @@ public class StreamProcessor extends Actor implements HealthMonitorable, LogReco\\n if (listener != null) {\\n streamProcessorContext.listener(engineContext.getStreamProcessorListener());\\n }\\n- streamProcessorContext.writers(engineContext.getWriters());\\n }\\n \\n private long recoverFromSnapshot() {\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\\nindex d02b273..b527d3c 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\\n@@ -17,7 +17,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.writers.CommandRespons\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriterImpl;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.NoopLegacyTypedStreamWriter;\\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\\n import io.camunda.zeebe.engine.state.EventApplier;\\n import io.camunda.zeebe.engine.state.KeyGeneratorControls;\\n import io.camunda.zeebe.engine.state.ZeebeDbState;\\n@@ -55,7 +54,7 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\\n private StreamProcessorMode streamProcessorMode = StreamProcessorMode.PROCESSING;\\n private ProcessingScheduleService processingScheduleService;\\n private MutableLastProcessedPositionState lastProcessedPositionState;\\n- private Writers writers;\\n+\\n private LogStreamBatchWriter logStreamBatchWriter;\\n private CommandResponseWriter commandResponseWriter;\\n \\n@@ -85,11 +84,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\\n }\\n \\n @Override\\n- public Writers getWriters() {\\n- return writers;\\n- }\\n-\\n- @Override\\n public MutableZeebeState getZeebeState() {\\n return zeebeState;\\n }\\n@@ -216,10 +210,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\\n return streamProcessorMode;\\n }\\n \\n- public void writers(final Writers writers) {\\n- this.writers = writers;\\n- }\\n-\\n public void logStreamBatchWriter(final LogStreamBatchWriter batchWriter) {\\n logStreamBatchWriter = batchWriter;\\n }\\n\", \"diff --git a/packages/cubejs-postgres-driver/driver/index.d.ts b/packages/cubejs-postgres-driver/driver/index.d.ts\\nnew file mode 100644\\nindex 0000000..47dcada\\n--- /dev/null\\n+++ b/packages/cubejs-postgres-driver/driver/index.d.ts\\n@@ -0,0 +1,8 @@\\n+import { PoolConfig } from \\\"pg\\\";\\n+\\n+declare module \\\"@cubejs-backend/postgres-driver\\\" {\\n+ class PostgresDriver {\\n+ constructor(options?: PoolConfig);\\n+ }\\n+ export = PostgresDriver;\\n+}\\ndiff --git a/packages/cubejs-postgres-driver/package.json b/packages/cubejs-postgres-driver/package.json\\nindex 9db5a20..1e9a236 100644\\n--- a/packages/cubejs-postgres-driver/package.json\\n+++ b/packages/cubejs-postgres-driver/package.json\\n@@ -12,6 +12,7 @@\\n \\\"node\\\": \\\">=8.11.1\\\"\\n },\\n \\\"main\\\": \\\"driver/PostgresDriver.js\\\",\\n+ \\\"typings\\\": \\\"driver/index.d.ts\\\",\\n \\\"scripts\\\": {\\n \\\"lint\\\": \\\"eslint **/*.js\\\"\\n },\\n\", \"diff --git a/Makefile.toml b/Makefile.toml\\nindex e7d2b20..490d6e2 100644\\n--- a/Makefile.toml\\n+++ b/Makefile.toml\\n@@ -82,7 +82,7 @@ end\\n '''\\n \\n [tasks.build-plugins-release]\\n-env = { \\\"CARGO_MAKE_WORKSPACE_SKIP_MEMBERS\\\" = [\\\".\\\"] }\\n+env = { \\\"CARGO_MAKE_WORKSPACE_INCLUDE_MEMBERS\\\" = [\\\"default-plugins/status-bar\\\", \\\"default-plugins/strider\\\", \\\"default-plugins/tab-bar\\\"] }\\n run_task = { name = \\\"build-release\\\", fork = true }\\n \\n [tasks.wasm-opt-plugins]\\n@@ -129,15 +129,16 @@ args = [\\\"install\\\", \\\"cross\\\"]\\n [tasks.publish]\\n clear = true\\n workspace = false\\n-dependencies = [\\\"build-plugins-release\\\", \\\"wasm-opt-plugins\\\", \\\"release-commit\\\", \\\"build-release\\\", \\\"publish-zellij-tile\\\", \\\"publish-zellij-tile-utils\\\", \\\"publish-zellij-utils\\\", \\\"publish-zellij-client\\\", \\\"publish-zellij-server\\\"]\\n+dependencies = [\\\"build-plugins-release\\\", \\\"wasm-opt-plugins\\\", \\\"release-commit\\\"]\\n run_task = \\\"publish-zellij\\\"\\n \\n [tasks.release-commit]\\n dependencies = [\\\"commit-all\\\", \\\"tag-release\\\"]\\n command = \\\"git\\\"\\n-args = [\\\"push\\\", \\\"--atomic\\\", \\\"upstream\\\", \\\"main\\\", \\\"v${CARGO_MAKE_CRATE_VERSION}\\\"]\\n+args = [\\\"push\\\", \\\"--atomic\\\", \\\"origin\\\", \\\"main\\\", \\\"v${CARGO_MAKE_CRATE_VERSION}\\\"]\\n \\n [tasks.commit-all]\\n+ignore_errors = true\\n command = \\\"git\\\"\\n args = [\\\"commit\\\", \\\"-aem\\\", \\\"chore(release): v${CARGO_MAKE_CRATE_VERSION}\\\"]\\n \\n@@ -148,31 +149,32 @@ args = [\\\"tag\\\", \\\"v${CARGO_MAKE_CRATE_VERSION}\\\"]\\n [tasks.publish-zellij-tile]\\n ignore_errors = true\\n cwd = \\\"zellij-tile\\\"\\n-command = \\\"cargo publish && sleep 15\\\"\\n+script = \\\"cargo publish && sleep 15\\\"\\n \\n [tasks.publish-zellij-client]\\n+ignore_errors = true\\n dependencies = [\\\"publish-zellij-utils\\\"]\\n cwd = \\\"zellij-client\\\"\\n-command = \\\"cargo publish && sleep 15\\\"\\n+script = \\\"cargo publish && sleep 15\\\"\\n \\n [tasks.publish-zellij-server]\\n+ignore_errors = true\\n dependencies = [\\\"publish-zellij-utils\\\"]\\n cwd = \\\"zellij-server\\\"\\n-command = \\\"cargo publish && sleep 15\\\"\\n+script = \\\"cargo publish && sleep 15\\\"\\n \\n [tasks.publish-zellij-utils]\\n+ignore_errors = true\\n dependencies = [\\\"publish-zellij-tile\\\"]\\n cwd = \\\"zellij-utils\\\"\\n-command = \\\"cargo publish && sleep 15\\\"\\n+script = \\\"cargo publish && sleep 15\\\"\\n \\n [tasks.publish-zellij-tile-utils]\\n ignore_errors = true\\n cwd = \\\"zellij-tile-utils\\\"\\n-command = \\\"cargo publish && sleep 15\\\"\\n+script = \\\"cargo publish && sleep 15\\\"\\n \\n [tasks.publish-zellij]\\n dependencies = [\\\"publish-zellij-client\\\", \\\"publish-zellij-server\\\", \\\"publish-zellij-utils\\\"]\\n command = \\\"cargo\\\"\\n args = [\\\"publish\\\"]\\n-\\n-\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"c2300c94c6b7d1599387272b616e1d79e93723c7\", \"d2709cab63295109dcd1a49f57da9418110e9044\", \"364d9bf18b2ce73c04d5ec3a70aefa3e6b83cc12\", \"65574eea5da54bf4722ecb551b42f8ff6088f33b\"]"},"types":{"kind":"string","value":"[\"ci\", \"refactor\", \"feat\", \"build\"]"}}},{"rowIdx":1023,"cells":{"commit_message":{"kind":"string","value":"fix `get-deploy-tags.sh`,fix the contact icon in the e2e test\n\nreferences #6364,parallelize pybind11 build,fix pagination spacing"},"diff":{"kind":"string","value":"[\"diff --git a/.circleci/get-deploy-tags.sh b/.circleci/get-deploy-tags.sh\\nindex f80c8cb..7ddfa62 100755\\n--- a/.circleci/get-deploy-tags.sh\\n+++ b/.circleci/get-deploy-tags.sh\\n@@ -20,7 +20,7 @@\\n set -euo pipefail\\n \\n DOCKER_IMAGE_TAG=${1}\\n-DOCKER_IMAGE=\\\"quay.io/influxdb/fusion\\\"\\n+DOCKER_IMAGE=\\\"quay.io/influxdb/iox\\\"\\n APP_NAME=\\\"IOx\\\"\\n \\n DOCKER_IMAGE_DIGEST=\\\"$(docker image inspect \\\"${DOCKER_IMAGE}:${DOCKER_IMAGE_TAG}\\\" --format '{{ if eq (len .RepoDigests) 1 }}{{index .RepoDigests 0}}{{ end }}')\\\"\\n\", \"diff --git a/ionic/components/toolbar/test/colors/main.html b/ionic/components/toolbar/test/colors/main.html\\nindex 24e48ca..73fe306 100644\\n--- a/ionic/components/toolbar/test/colors/main.html\\n+++ b/ionic/components/toolbar/test/colors/main.html\\n@@ -9,8 +9,8 @@\\n \\n \\n \\n- \\n \\n \\n \\n+ \\n+
\\n+

The process module is built into Node.js (therefore you can use this in both the main and renderer processes) and in Electron apps this object has a few more useful properties on it.

\\n+

The example below gets the version of Electron in use by the app.

\\n+

See the process documentation (opens in new window) for more.

\\n+
\\n+ \\n+ \\n+ \\n+ \\n+\\ndiff --git a/docs/fiddles/system/system-information/get-version-information/main.js b/docs/fiddles/system/system-information/get-version-information/main.js\\nnew file mode 100644\\nindex 0000000..1f9f917\\n--- /dev/null\\n+++ b/docs/fiddles/system/system-information/get-version-information/main.js\\n@@ -0,0 +1,25 @@\\n+const { app, BrowserWindow } = require('electron')\\n+\\n+let mainWindow = null\\n+\\n+function createWindow () {\\n+ const windowOptions = {\\n+ width: 600,\\n+ height: 400,\\n+ title: 'Get version information',\\n+ webPreferences: {\\n+ nodeIntegration: true\\n+ }\\n+ }\\n+\\n+ mainWindow = new BrowserWindow(windowOptions)\\n+ mainWindow.loadFile('index.html')\\n+\\n+ mainWindow.on('closed', () => {\\n+ mainWindow = null\\n+ })\\n+}\\n+\\n+app.on('ready', () => {\\n+ createWindow()\\n+})\\ndiff --git a/docs/fiddles/system/system-information/get-version-information/renderer.js b/docs/fiddles/system/system-information/get-version-information/renderer.js\\nnew file mode 100644\\nindex 0000000..40f7f2c\\n--- /dev/null\\n+++ b/docs/fiddles/system/system-information/get-version-information/renderer.js\\n@@ -0,0 +1,8 @@\\n+const versionInfoBtn = document.getElementById('version-info')\\n+\\n+const electronVersion = process.versions.electron\\n+\\n+versionInfoBtn.addEventListener('click', () => {\\n+ const message = `This app is using Electron version: ${electronVersion}`\\n+ document.getElementById('got-version-info').innerHTML = message\\n+})\\n\", \"diff --git a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\\nindex 4d771d6..9f9683c 100644\\n--- a/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\\n+++ b/src/components/__tests__/__snapshots__/BottomNavigation.test.js.snap\\n@@ -9,9 +9,6 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\\n Object {\\n \\\"flex\\\": 1,\\n },\\n- Object {\\n- \\\"backgroundColor\\\": \\\"#000000\\\",\\n- },\\n undefined,\\n ]\\n }\\n@@ -132,6 +129,33 @@ exports[`renders custom icon and label in non-shifting bottom navigation 1`] = `\\n ]\\n }\\n >\\n+ \\n \\n+ \\n 0 {\\n+\\tif len(archives) > 1 {\\n \\t\\treturn ErrTooManyDarwin64Builds\\n \\t}\\n \\tvar path = filepath.Join(ctx.Config.Brew.Folder, ctx.Config.ProjectName+\\\".rb\\\")\\n@@ -145,8 +142,7 @@ func doBuildFormula(data templateData) (out bytes.Buffer, err error) {\\n }\\n \\n func dataFor(ctx *context.Context, client client.Client, artifact artifact.Artifact) (result templateData, err error) {\\n-\\tvar file = artifact.Path\\n-\\tsum, err := checksum.SHA256(file)\\n+\\tsum, err := checksum.SHA256(artifact.Path)\\n \\tif err != nil {\\n \\t\\treturn\\n \\t}\\n@@ -163,7 +159,7 @@ func dataFor(ctx *context.Context, client client.Client, artifact artifact.Artif\\n \\t\\tTag: ctx.Git.CurrentTag,\\n \\t\\tVersion: ctx.Version,\\n \\t\\tCaveats: ctx.Config.Brew.Caveats,\\n-\\t\\tFile: file,\\n+\\t\\tFile: artifact.Name,\\n \\t\\tSHA256: sum,\\n \\t\\tDependencies: ctx.Config.Brew.Dependencies,\\n \\t\\tConflicts: ctx.Config.Brew.Conflicts,\\ndiff --git a/pipeline/brew/brew_test.go b/pipeline/brew/brew_test.go\\nindex 7e513bf..9066935 100644\\n--- a/pipeline/brew/brew_test.go\\n+++ b/pipeline/brew/brew_test.go\\n@@ -9,6 +9,7 @@ import (\\n \\n \\t\\\"github.com/goreleaser/goreleaser/config\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/context\\\"\\n+\\t\\\"github.com/goreleaser/goreleaser/internal/artifact\\\"\\n \\t\\\"github.com/goreleaser/goreleaser/internal/testlib\\\"\\n \\t\\\"github.com/stretchr/testify/assert\\\"\\n )\\n@@ -93,7 +94,8 @@ func TestRunPipe(t *testing.T) {\\n \\t\\tGit: context.GitInfo{\\n \\t\\t\\tCurrentTag: \\\"v1.0.1\\\",\\n \\t\\t},\\n-\\t\\tVersion: \\\"1.0.1\\\",\\n+\\t\\tVersion: \\\"1.0.1\\\",\\n+\\t\\tArtifacts: artifact.New(),\\n \\t\\tConfig: config.Project{\\n \\t\\t\\tDist: folder,\\n \\t\\t\\tProjectName: \\\"run-pipe\\\",\\n@@ -124,31 +126,53 @@ func TestRunPipe(t *testing.T) {\\n \\t\\tPublish: true,\\n \\t}\\n \\tvar path = filepath.Join(folder, \\\"bin.tar.gz\\\")\\n-\\tctx.AddBinary(\\\"darwinamd64\\\", \\\"bin\\\", \\\"bin\\\", path)\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin.tar.gz\\\",\\n+\\t\\tPath: path,\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n \\tclient := &DummyClient{}\\n \\tassert.Error(t, doRun(ctx, client))\\n \\tassert.False(t, client.CreatedFile)\\n \\n \\t_, err = os.Create(path)\\n \\tassert.NoError(t, err)\\n-\\tassert.NoError(t, doRun(ctx, client))\\n-\\tassert.True(t, client.CreatedFile)\\n \\n-\\tbts, err := ioutil.ReadFile(\\\"testdata/run_pipe.rb\\\")\\n-\\tassert.NoError(t, err)\\n-\\t// ioutil.WriteFile(\\\"testdata/run_pipe.rb\\\", []byte(client.Content), 0644)\\n+\\tt.Run(\\\"default git url\\\", func(tt *testing.T) {\\n+\\t\\tassert.NoError(tt, doRun(ctx, client))\\n+\\t\\tassert.True(tt, client.CreatedFile)\\n+\\n+\\t\\tbts, err := ioutil.ReadFile(\\\"testdata/run_pipe.rb\\\")\\n+\\t\\tassert.NoError(tt, err)\\n+\\t\\t// TODO: make writing this file toggleable somehow?\\n+\\t\\t// ioutil.WriteFile(\\\"testdata/run_pipe.rb\\\", []byte(client.Content), 0644)\\n+\\t\\tassert.Equal(tt, string(bts), client.Content)\\n+\\t})\\n \\n-\\tassert.Equal(t, string(bts), client.Content)\\n+\\tt.Run(\\\"github enterprise url\\\", func(tt *testing.T) {\\n+\\t\\tctx.Config.GitHubURLs.Download = \\\"http://github.example.org\\\"\\n+\\t\\tassert.NoError(tt, doRun(ctx, client))\\n+\\t\\tassert.True(tt, client.CreatedFile)\\n+\\n+\\t\\tbts, err := ioutil.ReadFile(\\\"testdata/run_pipe_enterprise.rb\\\")\\n+\\t\\tassert.NoError(tt, err)\\n+\\t\\t// TODO: make writing this file toggleable somehow?\\n+\\t\\t// ioutil.WriteFile(\\\"testdata/run_pipe_enterprise.rb\\\", []byte(client.Content), 0644)\\n+\\t\\tassert.Equal(tt, string(bts), client.Content)\\n+\\t})\\n }\\n \\n+// TODO: this test is irrelevant and can probavly be removed\\n func TestRunPipeFormatOverride(t *testing.T) {\\n \\tfolder, err := ioutil.TempDir(\\\"\\\", \\\"goreleasertest\\\")\\n \\tassert.NoError(t, err)\\n \\tvar path = filepath.Join(folder, \\\"bin.zip\\\")\\n \\t_, err = os.Create(path)\\n \\tassert.NoError(t, err)\\n-\\tvar ctx = &context.Context{\\n-\\t\\tConfig: config.Project{\\n+\\tvar ctx = context.New(\\n+\\t\\tconfig.Project{\\n \\t\\t\\tDist: folder,\\n \\t\\t\\tArchive: config.Archive{\\n \\t\\t\\t\\tFormat: \\\"tar.gz\\\",\\n@@ -166,9 +190,15 @@ func TestRunPipeFormatOverride(t *testing.T) {\\n \\t\\t\\t\\t},\\n \\t\\t\\t},\\n \\t\\t},\\n-\\t\\tPublish: true,\\n-\\t}\\n-\\tctx.AddBinary(\\\"darwinamd64\\\", \\\"bin\\\", \\\"bin\\\", path)\\n+\\t)\\n+\\tctx.Publish = true\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin.zip\\\",\\n+\\t\\tPath: path,\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n \\tclient := &DummyClient{}\\n \\tassert.NoError(t, doRun(ctx, client))\\n \\tassert.True(t, client.CreatedFile)\\n@@ -195,6 +225,40 @@ func TestRunPipeNoDarwin64Build(t *testing.T) {\\n \\tassert.False(t, client.CreatedFile)\\n }\\n \\n+func TestRunPipeMultipleDarwin64Build(t *testing.T) {\\n+\\tvar ctx = context.New(\\n+\\t\\tconfig.Project{\\n+\\t\\t\\tArchive: config.Archive{\\n+\\t\\t\\t\\tFormat: \\\"tar.gz\\\",\\n+\\t\\t\\t},\\n+\\t\\t\\tBrew: config.Homebrew{\\n+\\t\\t\\t\\tGitHub: config.Repo{\\n+\\t\\t\\t\\t\\tOwner: \\\"test\\\",\\n+\\t\\t\\t\\t\\tName: \\\"test\\\",\\n+\\t\\t\\t\\t},\\n+\\t\\t\\t},\\n+\\t\\t},\\n+\\t)\\n+\\tctx.Publish = true\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin1\\\",\\n+\\t\\tPath: \\\"doesnt mather\\\",\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin2\\\",\\n+\\t\\tPath: \\\"doesnt mather\\\",\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.UploadableArchive,\\n+\\t})\\n+\\tclient := &DummyClient{}\\n+\\tassert.Equal(t, ErrTooManyDarwin64Builds, doRun(ctx, client))\\n+\\tassert.False(t, client.CreatedFile)\\n+}\\n+\\n func TestRunPipeBrewNotSetup(t *testing.T) {\\n \\tvar ctx = &context.Context{\\n \\t\\tConfig: config.Project{},\\n@@ -206,9 +270,8 @@ func TestRunPipeBrewNotSetup(t *testing.T) {\\n }\\n \\n func TestRunPipeBinaryRelease(t *testing.T) {\\n-\\tvar ctx = &context.Context{\\n-\\t\\tPublish: true,\\n-\\t\\tConfig: config.Project{\\n+\\tvar ctx = context.New(\\n+\\t\\tconfig.Project{\\n \\t\\t\\tArchive: config.Archive{\\n \\t\\t\\t\\tFormat: \\\"binary\\\",\\n \\t\\t\\t},\\n@@ -219,8 +282,15 @@ func TestRunPipeBinaryRelease(t *testing.T) {\\n \\t\\t\\t\\t},\\n \\t\\t\\t},\\n \\t\\t},\\n-\\t}\\n-\\tctx.AddBinary(\\\"darwinamd64\\\", \\\"foo\\\", \\\"bar\\\", \\\"baz\\\")\\n+\\t)\\n+\\tctx.Publish = true\\n+\\tctx.Artifacts.Add(artifact.Artifact{\\n+\\t\\tName: \\\"bin\\\",\\n+\\t\\tPath: \\\"doesnt mather\\\",\\n+\\t\\tGoos: \\\"darwin\\\",\\n+\\t\\tGoarch: \\\"amd64\\\",\\n+\\t\\tType: artifact.Binary,\\n+\\t})\\n \\tclient := &DummyClient{}\\n \\ttestlib.AssertSkipped(t, doRun(ctx, client))\\n \\tassert.False(t, client.CreatedFile)\\ndiff --git a/pipeline/brew/doc.go b/pipeline/brew/doc.go\\nnew file mode 100644\\nindex 0000000..2cddc12\\n--- /dev/null\\n+++ b/pipeline/brew/doc.go\\n@@ -0,0 +1,3 @@\\n+// Package brew implements the Pipe, providing formula generation and\\n+// uploading it to a configured repo.\\n+package brew\\ndiff --git a/pipeline/brew/testdata/run_pipe_enterprise.rb b/pipeline/brew/testdata/run_pipe_enterprise.rb\\nnew file mode 100644\\nindex 0000000..4b24ce0\\n--- /dev/null\\n+++ b/pipeline/brew/testdata/run_pipe_enterprise.rb\\n@@ -0,0 +1,33 @@\\n+class RunPipe < Formula\\n+ desc \\\"A run pipe test formula\\\"\\n+ homepage \\\"https://github.com/goreleaser\\\"\\n+ url \\\"http://github.example.org/test/test/releases/download/v1.0.1/bin.tar.gz\\\"\\n+ version \\\"1.0.1\\\"\\n+ sha256 \\\"e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855\\\"\\n+ \\n+ depends_on \\\"zsh\\\"\\n+ depends_on \\\"bash\\\"\\n+ \\n+ conflicts_with \\\"gtk+\\\"\\n+ conflicts_with \\\"qt\\\"\\n+\\n+ def install\\n+ bin.install \\\"foo\\\"\\n+ end\\n+\\n+ def caveats\\n+ \\\"don't do this\\\"\\n+ end\\n+\\n+ plist_options :startup => false\\n+\\n+ def plist; <<-EOS.undent\\n+ whatever\\n+ EOS\\n+ end\\n+\\n+ test do\\n+ system \\\"true\\\"\\n+ system \\\"#{bin}/foo -h\\\"\\n+ end\\n+end\\n\", \"diff --git a/server/src/services/courseService.ts b/server/src/services/courseService.ts\\nindex 89633f4..10bfc55 100644\\n--- a/server/src/services/courseService.ts\\n+++ b/server/src/services/courseService.ts\\n@@ -580,8 +580,7 @@ export async function getTaskSolutionCheckers(courseTaskId: number, minCheckedCo\\n .createQueryBuilder('tsr')\\n .select('tsr.\\\"studentId\\\", ROUND(AVG(tsr.score)) as \\\"score\\\"')\\n .where(qb => {\\n- // query students with 3 checked tasks\\n-\\n+ // query students who checked enough tasks\\n const query = qb\\n .subQuery()\\n .select('r.\\\"checkerId\\\"')\\n@@ -600,7 +599,7 @@ export async function getTaskSolutionCheckers(courseTaskId: number, minCheckedCo\\n })\\n .andWhere('tsr.\\\"courseTaskId\\\" = :courseTaskId', { courseTaskId })\\n .groupBy('tsr.\\\"studentId\\\"')\\n- .having(`COUNT(tsr.id) >= :count`, { count: minCheckedCount })\\n+ .having(`COUNT(tsr.id) >= :count`, { count: minCheckedCount - 1 })\\n .getRawMany();\\n \\n return records.map(record => ({ studentId: record.studentId, score: Number(record.score) }));\\n\", \"diff --git a/packages/core/src/components/action-sheet/action-sheet.tsx b/packages/core/src/components/action-sheet/action-sheet.tsx\\nindex 7166508..dad7daf 100644\\n--- a/packages/core/src/components/action-sheet/action-sheet.tsx\\n+++ b/packages/core/src/components/action-sheet/action-sheet.tsx\\n@@ -1,9 +1,9 @@\\n import { Component, CssClassMap, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n \\n-import { domControllerAsync, isDef, playAnimationAsync } from '../../utils/helpers';\\n+import { domControllerAsync } from '../../utils/helpers';\\n import { createThemedClasses, getClassMap } from '../../utils/theme';\\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\\n \\n import iosEnterAnimation from './animations/ios.enter';\\n import iosLeaveAnimation from './animations/ios.leave';\\n@@ -23,15 +23,15 @@ import mdLeaveAnimation from './animations/md.leave';\\n })\\n export class ActionSheet implements OverlayInterface {\\n \\n+ private presented = false;\\n+\\n mode: string;\\n color: string;\\n-\\n- private presented = false;\\n- private animation: Animation | null = null;\\n+ animation: Animation;\\n \\n @Element() private el: HTMLElement;\\n \\n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\\n @Prop({ context: 'config' }) config: Config;\\n @Prop({ context: 'dom' }) dom: DomController;\\n @Prop() overlayId: number;\\n@@ -178,25 +178,8 @@ export class ActionSheet implements OverlayInterface {\\n });\\n }\\n \\n- private playAnimation(animationBuilder: AnimationBuilder) {\\n- if (this.animation) {\\n- this.animation.destroy();\\n- this.animation = null;\\n- }\\n-\\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\\n- this.animation = animation;\\n- // Check if prop animate is false or if the config for animate is defined/false\\n- if (!this.willAnimate || (isDef(this.config.get('willAnimate')) && this.config.get('willAnimate') === false)) {\\n- // if the duration is 0, it won't actually animate I don't think\\n- // TODO - validate this\\n- this.animation = animation.duration(0);\\n- }\\n- return playAnimationAsync(animation);\\n- }).then((animation) => {\\n- animation.destroy();\\n- this.animation = null;\\n- });\\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise {\\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\\n }\\n \\n protected buttonClick(button: ActionSheetButton) {\\ndiff --git a/packages/core/src/components/alert/alert.tsx b/packages/core/src/components/alert/alert.tsx\\nindex 800b77b..bdf4fc5 100644\\n--- a/packages/core/src/components/alert/alert.tsx\\n+++ b/packages/core/src/components/alert/alert.tsx\\n@@ -1,8 +1,8 @@\\n import { Component, CssClassMap, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n-import { domControllerAsync, playAnimationAsync, autoFocus } from '../../utils/helpers';\\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n+import { domControllerAsync, autoFocus } from '../../utils/helpers';\\n import { createThemedClasses, getClassMap } from '../../utils/theme';\\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\\n \\n import iosEnterAnimation from './animations/ios.enter';\\n import iosLeaveAnimation from './animations/ios.leave';\\n@@ -21,18 +21,19 @@ import mdLeaveAnimation from './animations/md.leave';\\n }\\n })\\n export class Alert implements OverlayInterface {\\n- mode: string;\\n- color: string;\\n \\n private presented = false;\\n- private animation: Animation | null = null;\\n private activeId: string;\\n private inputType: string | null = null;\\n private hdrId: string;\\n \\n+ animation: Animation;\\n+ mode: string;\\n+ color: string;\\n+\\n @Element() private el: HTMLElement;\\n \\n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\\n @Prop({ context: 'config' }) config: Config;\\n @Prop({ context: 'dom' }) dom: DomController;\\n @Prop() overlayId: number;\\n@@ -264,25 +265,10 @@ export class Alert implements OverlayInterface {\\n return values;\\n }\\n \\n- private playAnimation(animationBuilder: AnimationBuilder) {\\n- if (this.animation) {\\n- this.animation.destroy();\\n- this.animation = null;\\n- }\\n-\\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\\n- this.animation = animation;\\n- if (!this.willAnimate) {\\n- animation.duration(0);\\n- }\\n- return playAnimationAsync(animation);\\n- }).then(animation => {\\n- animation.destroy();\\n- this.animation = null;\\n- });\\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise {\\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\\n }\\n \\n-\\n private renderCheckbox(inputs: AlertInput[]) {\\n if (inputs.length === 0) return null;\\n \\ndiff --git a/packages/core/src/components/loading/loading.tsx b/packages/core/src/components/loading/loading.tsx\\nindex f45eaf1..cc4f511 100644\\n--- a/packages/core/src/components/loading/loading.tsx\\n+++ b/packages/core/src/components/loading/loading.tsx\\n@@ -1,13 +1,13 @@\\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n+import { domControllerAsync } from '../../utils/helpers';\\n import { createThemedClasses, getClassMap } from '../../utils/theme';\\n \\n import iosEnterAnimation from './animations/ios.enter';\\n import iosLeaveAnimation from './animations/ios.leave';\\n import mdEnterAnimation from './animations/md.enter';\\n import mdLeaveAnimation from './animations/md.leave';\\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\\n \\n @Component({\\n tag: 'ion-loading',\\n@@ -21,16 +21,17 @@ import { OverlayInterface, BACKDROP } from '../../utils/overlays';\\n })\\n \\n export class Loading implements OverlayInterface {\\n- color: string;\\n- mode: string;\\n \\n private presented = false;\\n- private animation: Animation;\\n private durationTimeout: any;\\n \\n+ animation: Animation;\\n+ color: string;\\n+ mode: string;\\n+\\n @Element() private el: HTMLElement;\\n \\n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\\n @Prop({ context: 'config' }) config: Config;\\n @Prop({ context: 'dom' }) dom: DomController;\\n @Prop() overlayId: number;\\n@@ -199,24 +200,8 @@ export class Loading implements OverlayInterface {\\n });\\n }\\n \\n- private playAnimation(animationBuilder: AnimationBuilder) {\\n- if (this.animation) {\\n- this.animation.destroy();\\n- this.animation = null;\\n- }\\n-\\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\\n- this.animation = animation;\\n- if (!this.willAnimate) {\\n- // if the duration is 0, it won't actually animate I don't think\\n- // TODO - validate this\\n- animation.duration(0);\\n- }\\n- return playAnimationAsync(animation);\\n- }).then(animation => {\\n- animation.destroy();\\n- this.animation = null;\\n- });\\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise {\\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\\n }\\n \\n hostData() {\\ndiff --git a/packages/core/src/components/modal/modal.tsx b/packages/core/src/components/modal/modal.tsx\\nindex af50d63..2b7510c 100644\\n--- a/packages/core/src/components/modal/modal.tsx\\n+++ b/packages/core/src/components/modal/modal.tsx\\n@@ -1,10 +1,10 @@\\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n+import { Animation, AnimationBuilder, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n \\n import { DomFrameworkDelegate } from '../../utils/dom-framework-delegate';\\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\\n+import { domControllerAsync } from '../../utils/helpers';\\n import { createThemedClasses } from '../../utils/theme';\\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\\n \\n import iosEnterAnimation from './animations/ios.enter';\\n import iosLeaveAnimation from './animations/ios.leave';\\n@@ -25,14 +25,16 @@ import mdLeaveAnimation from './animations/md.leave';\\n export class Modal implements OverlayInterface {\\n \\n private presented = false;\\n- private animation: Animation;\\n private usersComponentElement: HTMLElement;\\n \\n+ animation: Animation;\\n+\\n @Element() private el: HTMLElement;\\n \\n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\\n @Prop({ context: 'config' }) config: Config;\\n @Prop({ context: 'dom' }) dom: DomController;\\n+\\n @Prop() overlayId: number;\\n @Prop({ mutable: true }) delegate: FrameworkDelegate;\\n \\n@@ -208,22 +210,8 @@ export class Modal implements OverlayInterface {\\n });\\n }\\n \\n- private playAnimation(animationBuilder: AnimationBuilder) {\\n- if (this.animation) {\\n- this.animation.destroy();\\n- this.animation = null;\\n- }\\n-\\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\\n- this.animation = animation;\\n- if (!this.willAnimate) {\\n- animation.duration(0);\\n- }\\n- return playAnimationAsync(animation);\\n- }).then((animation) => {\\n- animation.destroy();\\n- this.animation = null;\\n- });\\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise {\\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\\n }\\n \\n @Method()\\ndiff --git a/packages/core/src/components/picker/picker.tsx b/packages/core/src/components/picker/picker.tsx\\nindex 13faa3e..d70381e 100644\\n--- a/packages/core/src/components/picker/picker.tsx\\n+++ b/packages/core/src/components/picker/picker.tsx\\n@@ -1,9 +1,9 @@\\n import { Component, CssClassMap, Element, Event, EventEmitter, Listen, Method, Prop, State } from '@stencil/core';\\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n+import { Animation, AnimationBuilder, Config, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n \\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\\n+import { domControllerAsync } from '../../utils/helpers';\\n import { getClassMap } from '../../utils/theme';\\n-import { OverlayInterface } from '../../utils/overlays';\\n+import { OverlayInterface, overlayAnimation } from '../../utils/overlays';\\n \\n import iosEnterAnimation from './animations/ios.enter';\\n import iosLeaveAnimation from './animations/ios.leave';\\n@@ -21,16 +21,17 @@ import iosLeaveAnimation from './animations/ios.leave';\\n export class Picker implements OverlayInterface {\\n \\n private presented = false;\\n- private animation: Animation;\\n private durationTimeout: any;\\n private mode: string;\\n \\n+ animation: Animation;\\n+\\n @Element() private el: HTMLElement;\\n \\n @State() private showSpinner: boolean = null;\\n @State() private spinner: string;\\n \\n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\\n @Prop({ context: 'config' }) config: Config;\\n @Prop({ context: 'dom' }) dom: DomController;\\n @Prop() overlayId: number;\\n@@ -231,22 +232,8 @@ export class Picker implements OverlayInterface {\\n return this.columns;\\n }\\n \\n- private playAnimation(animationBuilder: AnimationBuilder) {\\n- if (this.animation) {\\n- this.animation.destroy();\\n- this.animation = null;\\n- }\\n-\\n- return this.animationCtrl.create(animationBuilder, this.el).then(animation => {\\n- this.animation = animation;\\n- if (!this.willAnimate) {\\n- animation.duration(0);\\n- }\\n- return playAnimationAsync(animation);\\n- }).then(animation => {\\n- animation.destroy();\\n- this.animation = null;\\n- })\\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise {\\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, undefined);\\n }\\n \\n private buttonClick(button: PickerButton) {\\ndiff --git a/packages/core/src/components/popover/popover.tsx b/packages/core/src/components/popover/popover.tsx\\nindex 65031ff..6a47bf6 100644\\n--- a/packages/core/src/components/popover/popover.tsx\\n+++ b/packages/core/src/components/popover/popover.tsx\\n@@ -1,10 +1,10 @@\\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\\n-import { Animation, AnimationBuilder, AnimationController, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n+import { Animation, AnimationBuilder, Config, DomController, FrameworkDelegate, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n \\n import { DomFrameworkDelegate } from '../../utils/dom-framework-delegate';\\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\\n+import { domControllerAsync } from '../../utils/helpers';\\n import { createThemedClasses } from '../../utils/theme';\\n-import { OverlayInterface, BACKDROP } from '../../utils/overlays';\\n+import { OverlayInterface, BACKDROP, overlayAnimation } from '../../utils/overlays';\\n \\n import iosEnterAnimation from './animations/ios.enter';\\n import iosLeaveAnimation from './animations/ios.leave';\\n@@ -24,12 +24,13 @@ import mdLeaveAnimation from './animations/md.leave';\\n export class Popover implements OverlayInterface {\\n \\n private presented = false;\\n- private animation: Animation;\\n private usersComponentElement: HTMLElement;\\n \\n+ animation: Animation;\\n+\\n @Element() private el: HTMLElement;\\n \\n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\\n @Prop({ context: 'config' }) config: Config;\\n @Prop({ context: 'dom' }) dom: DomController;\\n @Prop({ mutable: true }) delegate: FrameworkDelegate;\\n@@ -224,22 +225,8 @@ export class Popover implements OverlayInterface {\\n });\\n }\\n \\n- private playAnimation(animationBuilder: AnimationBuilder) {\\n- if (this.animation) {\\n- this.animation.destroy();\\n- this.animation = null;\\n- }\\n-\\n- return this.animationCtrl.create(animationBuilder, this.el, this.ev).then((animation) => {\\n- this.animation = animation;\\n- if (!this.willAnimate) {\\n- animation.duration(0);\\n- }\\n- return playAnimationAsync(animation);\\n- }).then(animation => {\\n- animation.destroy();\\n- this.animation = null;\\n- })\\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise {\\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, this.ev);\\n }\\n \\n hostData() {\\ndiff --git a/packages/core/src/components/toast/toast.tsx b/packages/core/src/components/toast/toast.tsx\\nindex 1afa318..372070a 100644\\n--- a/packages/core/src/components/toast/toast.tsx\\n+++ b/packages/core/src/components/toast/toast.tsx\\n@@ -1,9 +1,9 @@\\n import { Component, Element, Event, EventEmitter, Listen, Method, Prop } from '@stencil/core';\\n-import { Animation, AnimationBuilder, AnimationController, Config, CssClassMap, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n+import { Animation, AnimationBuilder, Config, CssClassMap, DomController, OverlayDismissEvent, OverlayDismissEventDetail } from '../../index';\\n \\n-import { domControllerAsync, playAnimationAsync } from '../../utils/helpers';\\n+import { domControllerAsync } from '../../utils/helpers';\\n import { createThemedClasses, getClassMap } from '../../utils/theme';\\n-import { OverlayInterface } from '../../utils/overlays';\\n+import { OverlayInterface, overlayAnimation } from '../../utils/overlays';\\n \\n import iosEnterAnimation from './animations/ios.enter';\\n import iosLeaveAnimation from './animations/ios.leave';\\n@@ -24,14 +24,14 @@ import mdLeaveAnimation from './animations/md.leave';\\n export class Toast implements OverlayInterface {\\n \\n private presented = false;\\n- private animation: Animation | null;\\n \\n @Element() private el: HTMLElement;\\n \\n mode: string;\\n color: string;\\n+ animation: Animation | null;\\n \\n- @Prop({ connect: 'ion-animation-controller' }) animationCtrl: AnimationController;\\n+ @Prop({ connect: 'ion-animation-controller' }) animationCtrl: HTMLIonAnimationControllerElement;\\n @Prop({ context: 'config' }) config: Config;\\n @Prop({ context: 'dom' }) dom: DomController;\\n @Prop() overlayId: number;\\n@@ -123,6 +123,22 @@ export class Toast implements OverlayInterface {\\n */\\n @Event() ionToastDidUnload: EventEmitter;\\n \\n+ componentDidLoad() {\\n+ this.ionToastDidLoad.emit();\\n+ }\\n+\\n+ componentDidUnload() {\\n+ this.ionToastDidUnload.emit();\\n+ }\\n+\\n+ @Listen('ionDismiss')\\n+ protected onDismiss(ev: UIEvent) {\\n+ ev.stopPropagation();\\n+ ev.preventDefault();\\n+\\n+ this.dismiss();\\n+ }\\n+\\n /**\\n * Present the toast overlay after it has been created.\\n */\\n@@ -169,38 +185,8 @@ export class Toast implements OverlayInterface {\\n });\\n }\\n \\n- playAnimation(animationBuilder: AnimationBuilder) {\\n- if (this.animation) {\\n- this.animation.destroy();\\n- this.animation = null;\\n- }\\n-\\n- return this.animationCtrl.create(animationBuilder, this.el, this.position).then(animation => {\\n- this.animation = animation;\\n- if (!this.willAnimate) {\\n- animation.duration(0);\\n- }\\n- return playAnimationAsync(animation);\\n- }).then((animation) => {\\n- animation.destroy();\\n- this.animation = null;\\n- });\\n- }\\n-\\n- componentDidLoad() {\\n- this.ionToastDidLoad.emit();\\n- }\\n-\\n- componentDidUnload() {\\n- this.ionToastDidUnload.emit();\\n- }\\n-\\n- @Listen('ionDismiss')\\n- protected onDismiss(ev: UIEvent) {\\n- ev.stopPropagation();\\n- ev.preventDefault();\\n-\\n- this.dismiss();\\n+ private playAnimation(animationBuilder: AnimationBuilder): Promise {\\n+ return overlayAnimation(this, animationBuilder, this.willAnimate, this.el, this.position);\\n }\\n \\n private wrapperClass(): CssClassMap {\\ndiff --git a/packages/core/src/utils/overlays.ts b/packages/core/src/utils/overlays.ts\\nindex 8926544..634df43 100644\\n--- a/packages/core/src/utils/overlays.ts\\n+++ b/packages/core/src/utils/overlays.ts\\n@@ -1,3 +1,5 @@\\n+import { AnimationBuilder, Animation } from \\\"..\\\";\\n+import { playAnimationAsync } from \\\"./helpers\\\";\\n \\n let lastId = 1;\\n \\n@@ -56,8 +58,33 @@ export function removeLastOverlay(overlays: OverlayMap) {\\n return toRemove ? toRemove.dismiss() : Promise.resolve();\\n }\\n \\n+export function overlayAnimation(\\n+ overlay: OverlayInterface,\\n+ animationBuilder: AnimationBuilder,\\n+ animate: boolean,\\n+ baseEl: HTMLElement,\\n+ opts: any\\n+): Promise {\\n+ if (overlay.animation) {\\n+ overlay.animation.destroy();\\n+ overlay.animation = null;\\n+ }\\n+ return overlay.animationCtrl.create(animationBuilder, baseEl, opts).then(animation => {\\n+ overlay.animation = animation;\\n+ if (!animate) {\\n+ animation.duration(0);\\n+ }\\n+ return playAnimationAsync(animation);\\n+ }).then((animation) => {\\n+ animation.destroy();\\n+ overlay.animation = null;\\n+ });\\n+}\\n+\\n export interface OverlayInterface {\\n overlayId: number;\\n+ animation: Animation;\\n+ animationCtrl: HTMLIonAnimationControllerElement;\\n \\n present(): Promise;\\n dismiss(data?: any, role?: string): Promise;\\n\", \"diff --git a/core/main/src/Core/Particle.ts b/core/main/src/Core/Particle.ts\\nindex 1aa6fba..6ea6ffc 100644\\n--- a/core/main/src/Core/Particle.ts\\n+++ b/core/main/src/Core/Particle.ts\\n@@ -271,7 +271,7 @@ export class Particle implements IParticle {\\n }\\n }\\n \\n- const sizeAnimation = this.options.size.animation;\\n+ const sizeAnimation = sizeOptions.animation;\\n \\n if (sizeAnimation.enable) {\\n this.size.status = AnimationStatus.increasing;\\n@@ -279,7 +279,8 @@ export class Particle implements IParticle {\\n if (!randomSize) {\\n switch (sizeAnimation.startValue) {\\n case StartValueType.min:\\n- this.size.value = sizeAnimation.minimumValue * pxRatio;\\n+ this.size.value = NumberUtils.getRangeMin(sizeOptions.value) * pxRatio;\\n+ this.size.status = AnimationStatus.increasing;\\n \\n break;\\n \\n@@ -287,11 +288,14 @@ export class Particle implements IParticle {\\n this.size.value = NumberUtils.randomInRange(\\n NumberUtils.setRangeValue(sizeAnimation.minimumValue * pxRatio, this.size.value)\\n );\\n+ this.size.status =\\n+ Math.random() >= 0.5 ? AnimationStatus.increasing : AnimationStatus.decreasing;\\n \\n break;\\n \\n case StartValueType.max:\\n default:\\n+ this.size.value = NumberUtils.getRangeMax(sizeOptions.value) * pxRatio;\\n this.size.status = AnimationStatus.decreasing;\\n \\n break;\\n@@ -393,7 +397,8 @@ export class Particle implements IParticle {\\n if (!randomOpacity) {\\n switch (opacityAnimation.startValue) {\\n case StartValueType.min:\\n- this.opacity.value = opacityAnimation.minimumValue;\\n+ this.opacity.value = NumberUtils.getRangeMin(this.opacity.value);\\n+ this.opacity.status = AnimationStatus.increasing;\\n \\n break;\\n \\n@@ -401,11 +406,14 @@ export class Particle implements IParticle {\\n this.opacity.value = NumberUtils.randomInRange(\\n NumberUtils.setRangeValue(opacityAnimation.minimumValue, this.opacity.value)\\n );\\n+ this.opacity.status =\\n+ Math.random() >= 0.5 ? AnimationStatus.increasing : AnimationStatus.decreasing;\\n \\n break;\\n \\n case StartValueType.max:\\n default:\\n+ this.opacity.value = NumberUtils.getRangeMax(this.opacity.value);\\n this.opacity.status = AnimationStatus.decreasing;\\n \\n break;\\ndiff --git a/presets/confetti/src/options.ts b/presets/confetti/src/options.ts\\nindex 7fc6225..a713425 100644\\n--- a/presets/confetti/src/options.ts\\n+++ b/presets/confetti/src/options.ts\\n@@ -28,7 +28,7 @@ export const loadOptions = (confettiOptions: RecursivePartial)\\n animation: {\\n enable: true,\\n minimumValue: 0,\\n- speed: 2,\\n+ speed: 0.5,\\n startValue: \\\"max\\\",\\n destroy: \\\"min\\\",\\n },\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"f433bcb59c36571e22d4e86c612e0a6a52f73c09\", \"fd849bd08363df60dbc8b9b6d55bac4f5ace88f4\", \"9e3f295bbfd4098ffda1ae6656699f60b86c1f92\", \"06960183db42cba1b1f1a8077660ba8c801c9e18\"]"},"types":{"kind":"string","value":"[\"feat\", \"docs\", \"refactor\", \"fix\"]"}}},{"rowIdx":1027,"cells":{"commit_message":{"kind":"string","value":"bump version\n\nSigned-off-by: rjshrjndrn ,process CommandDistribution ACKNOWLEDGED event\n\nAdds an EventApplier for the CommandDistribution ACKNOWLEDGED event. This applier will be responsible to remove a pending distribution from the state. This will be used to mark the distribution to a specific partition as completed.,ignore all markdown files for backend and main test suites,uses macros to implement Settings enums"},"diff":{"kind":"string","value":"[\"diff --git a/scripts/helmcharts/init.sh b/scripts/helmcharts/init.sh\\nindex 5a2b4b0..69a6944 100644\\n--- a/scripts/helmcharts/init.sh\\n+++ b/scripts/helmcharts/init.sh\\n@@ -26,7 +26,7 @@ usr=$(whoami)\\n \\n # Installing k3s\\n function install_k8s() {\\n- curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\\\"644\\\" INSTALL_K3S_VERSION='v1.22.8+k3s1' INSTALL_K3S_EXEC=\\\"--no-deploy=traefik\\\" sh -\\n+ curl -sL https://get.k3s.io | sudo K3S_KUBECONFIG_MODE=\\\"644\\\" INSTALL_K3S_VERSION='v1.25.6+k3s1' INSTALL_K3S_EXEC=\\\"--disable=traefik\\\" sh -\\n [[ -d ~/.kube ]] || mkdir ~/.kube\\n sudo cp /etc/rancher/k3s/k3s.yaml ~/.kube/config\\n sudo chmod 0644 ~/.kube/config\\n\", \"diff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java\\nnew file mode 100644\\nindex 0000000..4abf2e3\\n--- /dev/null\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/CommandDistributionAcknowledgedApplier.java\\n@@ -0,0 +1,28 @@\\n+/*\\n+ * Copyright Camunda Services GmbH and/or licensed to Camunda Services GmbH under\\n+ * one or more contributor license agreements. See the NOTICE file distributed\\n+ * with this work for additional information regarding copyright ownership.\\n+ * Licensed under the Zeebe Community License 1.1. You may not use this file\\n+ * except in compliance with the Zeebe Community License 1.1.\\n+ */\\n+package io.camunda.zeebe.engine.state.appliers;\\n+\\n+import io.camunda.zeebe.engine.state.TypedEventApplier;\\n+import io.camunda.zeebe.engine.state.mutable.MutableDistributionState;\\n+import io.camunda.zeebe.protocol.impl.record.value.distribution.CommandDistributionRecord;\\n+import io.camunda.zeebe.protocol.record.intent.CommandDistributionIntent;\\n+\\n+public final class CommandDistributionAcknowledgedApplier\\n+ implements TypedEventApplier {\\n+\\n+ private final MutableDistributionState distributionState;\\n+\\n+ public CommandDistributionAcknowledgedApplier(final MutableDistributionState distributionState) {\\n+ this.distributionState = distributionState;\\n+ }\\n+\\n+ @Override\\n+ public void applyState(final long key, final CommandDistributionRecord value) {\\n+ distributionState.removePendingDistribution(key, value.getPartitionId());\\n+ }\\n+}\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\nindex a72309b..4793315 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\n@@ -284,6 +284,9 @@ public final class EventAppliers implements EventApplier {\\n CommandDistributionIntent.DISTRIBUTING,\\n new CommandDistributionDistributingApplier(distributionState));\\n register(\\n+ CommandDistributionIntent.ACKNOWLEDGED,\\n+ new CommandDistributionAcknowledgedApplier(distributionState));\\n+ register(\\n CommandDistributionIntent.FINISHED,\\n new CommandDistributionFinishedApplier(distributionState));\\n }\\n\", \"diff --git a/.github/workflows/ibis-backends-skip-helper.yml b/.github/workflows/ibis-backends-skip-helper.yml\\nindex efd0953..058f8b6 100644\\n--- a/.github/workflows/ibis-backends-skip-helper.yml\\n+++ b/.github/workflows/ibis-backends-skip-helper.yml\\n@@ -7,6 +7,7 @@ on:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n@@ -14,6 +15,7 @@ on:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\ndiff --git a/.github/workflows/ibis-backends.yml b/.github/workflows/ibis-backends.yml\\nindex d18e62d..144562c 100644\\n--- a/.github/workflows/ibis-backends.yml\\n+++ b/.github/workflows/ibis-backends.yml\\n@@ -3,18 +3,20 @@ name: Backends\\n \\n on:\\n push:\\n- # Skip the backend suite if all changes are in the docs directory\\n+ # Skip the backend suite if all changes are docs\\n paths-ignore:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n pull_request:\\n- # Skip the backend suite if all changes are in the docs directory\\n+ # Skip the backend suite if all changes are docs\\n paths-ignore:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\ndiff --git a/.github/workflows/ibis-main-skip-helper.yml b/.github/workflows/ibis-main-skip-helper.yml\\nindex f6086e1..7d79af7 100644\\n--- a/.github/workflows/ibis-main-skip-helper.yml\\n+++ b/.github/workflows/ibis-main-skip-helper.yml\\n@@ -7,6 +7,7 @@ on:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n@@ -14,6 +15,7 @@ on:\\n paths:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\ndiff --git a/.github/workflows/ibis-main.yml b/.github/workflows/ibis-main.yml\\nindex d5b0735..3d22bff 100644\\n--- a/.github/workflows/ibis-main.yml\\n+++ b/.github/workflows/ibis-main.yml\\n@@ -7,6 +7,7 @@ on:\\n paths-ignore:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n@@ -15,6 +16,7 @@ on:\\n paths-ignore:\\n - \\\"docs/**\\\"\\n - \\\"mkdocs.yml\\\"\\n+ - \\\"**/*.md\\\"\\n branches:\\n - master\\n - \\\"*.x.x\\\"\\n\", \"diff --git a/src/app/settings.rs b/src/app/settings.rs\\nindex e0e5ed1..60584f4 100644\\n--- a/src/app/settings.rs\\n+++ b/src/app/settings.rs\\n@@ -33,76 +33,26 @@ impl AppFlags {\\n AppFlags(NEEDS_LONG_VERSION | NEEDS_LONG_HELP | NEEDS_SC_HELP | UTF8_NONE)\\n }\\n \\n- pub fn set(&mut self, s: AppSettings) {\\n- match s {\\n- AppSettings::SubcommandsNegateReqs => self.0.insert(SC_NEGATE_REQS),\\n- AppSettings::VersionlessSubcommands => self.0.insert(VERSIONLESS_SC),\\n- AppSettings::SubcommandRequired => self.0.insert(SC_REQUIRED),\\n- AppSettings::ArgRequiredElseHelp => self.0.insert(A_REQUIRED_ELSE_HELP),\\n- AppSettings::GlobalVersion => self.0.insert(GLOBAL_VERSION),\\n- AppSettings::UnifiedHelpMessage => self.0.insert(UNIFIED_HELP),\\n- AppSettings::WaitOnError => self.0.insert(WAIT_ON_ERROR),\\n- AppSettings::SubcommandRequiredElseHelp => self.0.insert(SC_REQUIRED_ELSE_HELP),\\n- AppSettings::NeedsLongHelp => self.0.insert(NEEDS_LONG_HELP),\\n- AppSettings::NeedsLongVersion => self.0.insert(NEEDS_LONG_VERSION),\\n- AppSettings::NeedsSubcommandHelp => self.0.insert(NEEDS_SC_HELP),\\n- AppSettings::DisableVersion => self.0.insert(DISABLE_VERSION),\\n- AppSettings::Hidden => self.0.insert(HIDDEN),\\n- AppSettings::TrailingVarArg => self.0.insert(TRAILING_VARARG),\\n- AppSettings::NoBinaryName => self.0.insert(NO_BIN_NAME),\\n- AppSettings::AllowExternalSubcommands => self.0.insert(ALLOW_UNK_SC),\\n- AppSettings::StrictUtf8 => self.0.insert(UTF8_STRICT),\\n- AppSettings::AllowInvalidUtf8 => self.0.insert(UTF8_NONE),\\n- AppSettings::AllowLeadingHyphen => self.0.insert(LEADING_HYPHEN),\\n- }\\n- }\\n-\\n- pub fn unset(&mut self, s: AppSettings) {\\n- match s {\\n- AppSettings::SubcommandsNegateReqs => self.0.remove(SC_NEGATE_REQS),\\n- AppSettings::VersionlessSubcommands => self.0.remove(VERSIONLESS_SC),\\n- AppSettings::SubcommandRequired => self.0.remove(SC_REQUIRED),\\n- AppSettings::ArgRequiredElseHelp => self.0.remove(A_REQUIRED_ELSE_HELP),\\n- AppSettings::GlobalVersion => self.0.remove(GLOBAL_VERSION),\\n- AppSettings::UnifiedHelpMessage => self.0.remove(UNIFIED_HELP),\\n- AppSettings::WaitOnError => self.0.remove(WAIT_ON_ERROR),\\n- AppSettings::SubcommandRequiredElseHelp => self.0.remove(SC_REQUIRED_ELSE_HELP),\\n- AppSettings::NeedsLongHelp => self.0.remove(NEEDS_LONG_HELP),\\n- AppSettings::NeedsLongVersion => self.0.remove(NEEDS_LONG_VERSION),\\n- AppSettings::NeedsSubcommandHelp => self.0.remove(NEEDS_SC_HELP),\\n- AppSettings::DisableVersion => self.0.remove(DISABLE_VERSION),\\n- AppSettings::Hidden => self.0.remove(HIDDEN),\\n- AppSettings::TrailingVarArg => self.0.remove(TRAILING_VARARG),\\n- AppSettings::NoBinaryName => self.0.remove(NO_BIN_NAME),\\n- AppSettings::AllowExternalSubcommands => self.0.remove(ALLOW_UNK_SC),\\n- AppSettings::StrictUtf8 => self.0.remove(UTF8_STRICT),\\n- AppSettings::AllowInvalidUtf8 => self.0.remove(UTF8_NONE),\\n- AppSettings::AllowLeadingHyphen => self.0.remove(LEADING_HYPHEN),\\n- }\\n- }\\n-\\n- pub fn is_set(&self, s: AppSettings) -> bool {\\n- match s {\\n- AppSettings::SubcommandsNegateReqs => self.0.contains(SC_NEGATE_REQS),\\n- AppSettings::VersionlessSubcommands => self.0.contains(VERSIONLESS_SC),\\n- AppSettings::SubcommandRequired => self.0.contains(SC_REQUIRED),\\n- AppSettings::ArgRequiredElseHelp => self.0.contains(A_REQUIRED_ELSE_HELP),\\n- AppSettings::GlobalVersion => self.0.contains(GLOBAL_VERSION),\\n- AppSettings::UnifiedHelpMessage => self.0.contains(UNIFIED_HELP),\\n- AppSettings::WaitOnError => self.0.contains(WAIT_ON_ERROR),\\n- AppSettings::SubcommandRequiredElseHelp => self.0.contains(SC_REQUIRED_ELSE_HELP),\\n- AppSettings::NeedsLongHelp => self.0.contains(NEEDS_LONG_HELP),\\n- AppSettings::NeedsLongVersion => self.0.contains(NEEDS_LONG_VERSION),\\n- AppSettings::NeedsSubcommandHelp => self.0.contains(NEEDS_SC_HELP),\\n- AppSettings::DisableVersion => self.0.contains(DISABLE_VERSION),\\n- AppSettings::Hidden => self.0.contains(HIDDEN),\\n- AppSettings::TrailingVarArg => self.0.contains(TRAILING_VARARG),\\n- AppSettings::NoBinaryName => self.0.contains(NO_BIN_NAME),\\n- AppSettings::AllowExternalSubcommands => self.0.contains(ALLOW_UNK_SC),\\n- AppSettings::StrictUtf8 => self.0.contains(UTF8_STRICT),\\n- AppSettings::AllowInvalidUtf8 => self.0.contains(UTF8_NONE),\\n- AppSettings::AllowLeadingHyphen => self.0.contains(LEADING_HYPHEN),\\n- }\\n+ impl_settings! { AppSettings,\\n+ SubcommandsNegateReqs => SC_NEGATE_REQS,\\n+ VersionlessSubcommands => VERSIONLESS_SC,\\n+ SubcommandRequired => SC_REQUIRED,\\n+ ArgRequiredElseHelp => A_REQUIRED_ELSE_HELP,\\n+ GlobalVersion => GLOBAL_VERSION,\\n+ UnifiedHelpMessage => UNIFIED_HELP,\\n+ WaitOnError => WAIT_ON_ERROR,\\n+ SubcommandRequiredElseHelp => SC_REQUIRED_ELSE_HELP,\\n+ NeedsLongHelp => NEEDS_LONG_HELP,\\n+ NeedsLongVersion => NEEDS_LONG_VERSION,\\n+ NeedsSubcommandHelp => NEEDS_SC_HELP,\\n+ DisableVersion => DISABLE_VERSION,\\n+ Hidden => HIDDEN,\\n+ TrailingVarArg => TRAILING_VARARG,\\n+ NoBinaryName => NO_BIN_NAME,\\n+ AllowExternalSubcommands => ALLOW_UNK_SC,\\n+ StrictUtf8 => UTF8_STRICT,\\n+ AllowInvalidUtf8 => UTF8_NONE,\\n+ AllowLeadingHyphen => LEADING_HYPHEN\\n }\\n }\\n \\ndiff --git a/src/args/settings.rs b/src/args/settings.rs\\nindex f2f1384..effc18c 100644\\n--- a/src/args/settings.rs\\n+++ b/src/args/settings.rs\\n@@ -21,40 +21,14 @@ impl ArgFlags {\\n ArgFlags(EMPTY_VALS | USE_DELIM)\\n }\\n \\n- pub fn set(&mut self, s: ArgSettings) {\\n- match s {\\n- ArgSettings::Required => self.0.insert(REQUIRED),\\n- ArgSettings::Multiple => self.0.insert(MULTIPLE),\\n- ArgSettings::EmptyValues => self.0.insert(EMPTY_VALS),\\n- ArgSettings::Global => self.0.insert(GLOBAL),\\n- ArgSettings::Hidden => self.0.insert(HIDDEN),\\n- ArgSettings::TakesValue => self.0.insert(TAKES_VAL),\\n- ArgSettings::UseValueDelimiter => self.0.insert(USE_DELIM),\\n- }\\n- }\\n-\\n- pub fn unset(&mut self, s: ArgSettings) {\\n- match s {\\n- ArgSettings::Required => self.0.remove(REQUIRED),\\n- ArgSettings::Multiple => self.0.remove(MULTIPLE),\\n- ArgSettings::EmptyValues => self.0.remove(EMPTY_VALS),\\n- ArgSettings::Global => self.0.remove(GLOBAL),\\n- ArgSettings::Hidden => self.0.remove(HIDDEN),\\n- ArgSettings::TakesValue => self.0.remove(TAKES_VAL),\\n- ArgSettings::UseValueDelimiter => self.0.remove(USE_DELIM),\\n- }\\n- }\\n-\\n- pub fn is_set(&self, s: ArgSettings) -> bool {\\n- match s {\\n- ArgSettings::Required => self.0.contains(REQUIRED),\\n- ArgSettings::Multiple => self.0.contains(MULTIPLE),\\n- ArgSettings::EmptyValues => self.0.contains(EMPTY_VALS),\\n- ArgSettings::Global => self.0.contains(GLOBAL),\\n- ArgSettings::Hidden => self.0.contains(HIDDEN),\\n- ArgSettings::TakesValue => self.0.contains(TAKES_VAL),\\n- ArgSettings::UseValueDelimiter => self.0.contains(USE_DELIM),\\n- }\\n+ impl_settings!{ArgSettings,\\n+ Required => REQUIRED,\\n+ Multiple => MULTIPLE,\\n+ EmptyValues => EMPTY_VALS,\\n+ Global => GLOBAL,\\n+ Hidden => HIDDEN,\\n+ TakesValue => TAKES_VAL,\\n+ UseValueDelimiter => USE_DELIM\\n }\\n }\\n \\ndiff --git a/src/macros.rs b/src/macros.rs\\nindex 47675ac..29d5382 100644\\n--- a/src/macros.rs\\n+++ b/src/macros.rs\\n@@ -1,3 +1,25 @@\\n+macro_rules! impl_settings {\\n+ ($n:ident, $($v:ident => $c:ident),+) => {\\n+ pub fn set(&mut self, s: $n) {\\n+ match s {\\n+ $($n::$v => self.0.insert($c)),+\\n+ }\\n+ }\\n+\\n+ pub fn unset(&mut self, s: $n) {\\n+ match s {\\n+ $($n::$v => self.0.remove($c)),+\\n+ }\\n+ }\\n+\\n+ pub fn is_set(&self, s: $n) -> bool {\\n+ match s {\\n+ $($n::$v => self.0.contains($c)),+\\n+ }\\n+ }\\n+ };\\n+}\\n+\\n // Convenience for writing to stderr thanks to https://github.com/BurntSushi\\n macro_rules! wlnerr(\\n ($($arg:tt)*) => ({\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"9a25fe59dfb63d32505afcea3a164ff0b8ea4c71\", \"6f4c06076abff94f8bb5c634beaba55483a78b72\", \"370830b8c9f971fa537f42308ab5e3ff356919f8\", \"86f3e3397594f8312226c5a193608a054087805c\"]"},"types":{"kind":"string","value":"[\"build\", \"feat\", \"ci\", \"refactor\"]"}}},{"rowIdx":1028,"cells":{"commit_message":{"kind":"string","value":"get ip from forwarded header,do not pin time in tests but only skip ahead\n\nrelated to #573,add automation for rebasing `*.x.x` branches,fix \"types\" field in dist"},"diff":{"kind":"string","value":"[\"diff --git a/kousa/lib/broth/socket_handler.ex b/kousa/lib/broth/socket_handler.ex\\nindex d142135..5828f30 100644\\n--- a/kousa/lib/broth/socket_handler.ex\\n+++ b/kousa/lib/broth/socket_handler.ex\\n@@ -22,7 +22,7 @@ defmodule Broth.SocketHandler do\\n ## initialization boilerplate\\n \\n @impl true\\n- def init(request = %{peer: {ip, _reverse_port}}, _state) do\\n+ def init(request, _state) do\\n props = :cowboy_req.parse_qs(request)\\n \\n compression =\\n@@ -37,10 +37,16 @@ defmodule Broth.SocketHandler do\\n _ -> :json\\n end\\n \\n+ ip =\\n+ case request.headers do\\n+ %{\\\"x-forwarded-for\\\" => v} -> v\\n+ _ -> nil\\n+ end\\n+\\n state = %__MODULE__{\\n awaiting_init: true,\\n user_id: nil,\\n- ip: IP.to_string(ip),\\n+ ip: ip,\\n encoding: encoding,\\n compression: compression,\\n callers: get_callers(request)\\ndiff --git a/kousa/test/_support/ws_client.ex b/kousa/test/_support/ws_client.ex\\nindex aeca704..125da17 100644\\n--- a/kousa/test/_support/ws_client.ex\\n+++ b/kousa/test/_support/ws_client.ex\\n@@ -19,7 +19,9 @@ defmodule BrothTest.WsClient do\\n \\n @api_url\\n |> Path.join(\\\"socket\\\")\\n- |> WebSockex.start_link(__MODULE__, nil, extra_headers: [{\\\"user-agent\\\", ancestors}])\\n+ |> WebSockex.start_link(__MODULE__, nil,\\n+ extra_headers: [{\\\"user-agent\\\", ancestors}, {\\\"x-forwarded-for\\\", \\\"127.0.0.1\\\"}]\\n+ )\\n end\\n \\n ###########################################################################\\n\", \"diff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\nindex 636cd21..76afff7 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRecoveryTest.java\\n@@ -15,7 +15,9 @@\\n */\\n package io.zeebe.broker.it.startup;\\n \\n-import static io.zeebe.broker.it.util.TopicEventRecorder.*;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.incidentEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.taskEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.wfInstanceEvent;\\n import static io.zeebe.test.util.TestUtil.doRepeatedly;\\n import static io.zeebe.test.util.TestUtil.waitUntil;\\n import static org.assertj.core.api.Assertions.assertThat;\\n@@ -24,11 +26,18 @@ import java.io.File;\\n import java.io.InputStream;\\n import java.nio.charset.StandardCharsets;\\n import java.time.Duration;\\n-import java.time.Instant;\\n import java.util.Collections;\\n import java.util.List;\\n import java.util.regex.Pattern;\\n \\n+import org.assertj.core.util.Files;\\n+import org.junit.After;\\n+import org.junit.Rule;\\n+import org.junit.Test;\\n+import org.junit.rules.ExpectedException;\\n+import org.junit.rules.RuleChain;\\n+import org.junit.rules.TemporaryFolder;\\n+\\n import io.zeebe.broker.clustering.ClusterServiceNames;\\n import io.zeebe.broker.it.ClientRule;\\n import io.zeebe.broker.it.EmbeddedBrokerRule;\\n@@ -38,7 +47,9 @@ import io.zeebe.client.ZeebeClient;\\n import io.zeebe.client.clustering.impl.TopicLeader;\\n import io.zeebe.client.clustering.impl.TopologyResponse;\\n import io.zeebe.client.cmd.ClientCommandRejectedException;\\n-import io.zeebe.client.event.*;\\n+import io.zeebe.client.event.DeploymentEvent;\\n+import io.zeebe.client.event.TaskEvent;\\n+import io.zeebe.client.event.WorkflowInstanceEvent;\\n import io.zeebe.model.bpmn.Bpmn;\\n import io.zeebe.model.bpmn.instance.WorkflowDefinition;\\n import io.zeebe.raft.Raft;\\n@@ -48,9 +59,6 @@ import io.zeebe.test.util.TestFileUtil;\\n import io.zeebe.test.util.TestUtil;\\n import io.zeebe.transport.SocketAddress;\\n import io.zeebe.util.time.ClockUtil;\\n-import org.assertj.core.util.Files;\\n-import org.junit.*;\\n-import org.junit.rules.*;\\n \\n public class BrokerRecoveryTest\\n {\\n@@ -360,17 +368,12 @@ public class BrokerRecoveryTest\\n waitUntil(() -> !recordingTaskHandler.getHandledTasks().isEmpty());\\n \\n // when\\n- restartBroker(() ->\\n- {\\n- final Instant now = ClockUtil.getCurrentTime();\\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\\n- });\\n+ restartBroker(() -> ClockUtil.addTime(Duration.ofSeconds(60)));\\n \\n // wait until stream processor and scheduler process the lock task event which is not re-processed on recovery\\n doRepeatedly(() ->\\n {\\n- final Instant now = ClockUtil.getCurrentTime();\\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\\n+ ClockUtil.addTime(Duration.ofSeconds(60)); // retriggers lock expiration check in broker\\n return null;\\n }).until(t -> eventRecorder.hasTaskEvent(taskEvent(\\\"LOCK_EXPIRED\\\")));\\n \\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\\nindex 5ff1301..0ffe98d 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/startup/BrokerRestartTest.java\\n@@ -15,7 +15,9 @@\\n */\\n package io.zeebe.broker.it.startup;\\n \\n-import static io.zeebe.broker.it.util.TopicEventRecorder.*;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.incidentEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.taskEvent;\\n+import static io.zeebe.broker.it.util.TopicEventRecorder.wfInstanceEvent;\\n import static io.zeebe.test.util.TestUtil.waitUntil;\\n import static org.assertj.core.api.Assertions.assertThat;\\n \\n@@ -23,11 +25,18 @@ import java.io.File;\\n import java.io.InputStream;\\n import java.nio.charset.StandardCharsets;\\n import java.time.Duration;\\n-import java.time.Instant;\\n import java.util.Collections;\\n import java.util.List;\\n import java.util.regex.Pattern;\\n \\n+import org.junit.After;\\n+import org.junit.Rule;\\n+import org.junit.Test;\\n+import org.junit.experimental.categories.Category;\\n+import org.junit.rules.ExpectedException;\\n+import org.junit.rules.RuleChain;\\n+import org.junit.rules.TemporaryFolder;\\n+\\n import io.zeebe.broker.clustering.ClusterServiceNames;\\n import io.zeebe.broker.it.ClientRule;\\n import io.zeebe.broker.it.EmbeddedBrokerRule;\\n@@ -37,7 +46,9 @@ import io.zeebe.client.ZeebeClient;\\n import io.zeebe.client.clustering.impl.TopicLeader;\\n import io.zeebe.client.clustering.impl.TopologyResponse;\\n import io.zeebe.client.cmd.ClientCommandRejectedException;\\n-import io.zeebe.client.event.*;\\n+import io.zeebe.client.event.DeploymentEvent;\\n+import io.zeebe.client.event.TaskEvent;\\n+import io.zeebe.client.event.WorkflowInstanceEvent;\\n import io.zeebe.model.bpmn.Bpmn;\\n import io.zeebe.model.bpmn.instance.WorkflowDefinition;\\n import io.zeebe.raft.Raft;\\n@@ -47,9 +58,6 @@ import io.zeebe.test.util.TestFileUtil;\\n import io.zeebe.test.util.TestUtil;\\n import io.zeebe.transport.SocketAddress;\\n import io.zeebe.util.time.ClockUtil;\\n-import org.junit.*;\\n-import org.junit.experimental.categories.Category;\\n-import org.junit.rules.*;\\n \\n public class BrokerRestartTest\\n {\\n@@ -360,11 +368,7 @@ public class BrokerRestartTest\\n waitUntil(() -> !recordingTaskHandler.getHandledTasks().isEmpty());\\n \\n // when\\n- restartBroker(() ->\\n- {\\n- final Instant now = ClockUtil.getCurrentTime();\\n- ClockUtil.setCurrentTime(now.plusSeconds(60));\\n- });\\n+ restartBroker(() -> ClockUtil.addTime(Duration.ofSeconds(60)));\\n \\n waitUntil(() -> eventRecorder.hasTaskEvent(taskEvent(\\\"LOCK_EXPIRED\\\")));\\n recordingTaskHandler.clear();\\ndiff --git a/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java b/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\\nindex 49b527d..a322fbe 100644\\n--- a/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\\n+++ b/qa/integration-tests/src/test/java/io/zeebe/broker/it/task/TaskSubscriptionTest.java\\n@@ -353,7 +353,7 @@ public class TaskSubscriptionTest\\n waitUntil(() -> taskHandler.getHandledTasks().size() == 1);\\n \\n // when\\n- ClockUtil.setCurrentTime(Instant.now().plus(Duration.ofMinutes(5)));\\n+ ClockUtil.addTime(Duration.ofMinutes(5));\\n \\n // then\\n waitUntil(() -> taskHandler.getHandledTasks().size() == 2);\\n\", \"diff --git a/.github/workflows/ibis-rebase-nightly.yml b/.github/workflows/ibis-rebase-nightly.yml\\nnew file mode 100644\\nindex 0000000..3d2f064\\n--- /dev/null\\n+++ b/.github/workflows/ibis-rebase-nightly.yml\\n@@ -0,0 +1,82 @@\\n+name: Update Dependencies\\n+on:\\n+ schedule:\\n+ # run every 24 hours at 1 AM\\n+ - cron: \\\"1 */24 * * *\\\"\\n+ workflow_dispatch:\\n+\\n+jobs:\\n+ generate_branches:\\n+ runs-on: ubuntu-latest\\n+ outputs:\\n+ matrix: ${{ steps.set-matrix.outputs.matrix }}\\n+ steps:\\n+ - name: output branches list\\n+ id: set-matrix\\n+ run: |\\n+ set -euo pipefail\\n+\\n+ branches=$(git ls-remote --heads https://github.com/ibis-project/ibis.git \\\\\\n+ | sed -e 's#\\\\t# #g' -e 's#refs/heads/##g' \\\\\\n+ | cut -d ' ' -f2 \\\\\\n+ | grep -P '\\\\d+\\\\.x\\\\.x' \\\\\\n+ | xargs -I {} printf '\\\"%s\\\"' \\\\\\n+ | jq -s '{branch: .}')\\n+\\n+ echo \\\"::set-output name=matrix::$branches\\\"\\n+\\n+ niv_update:\\n+ runs-on: ubuntu-latest\\n+ needs:\\n+ - generate_branches\\n+ strategy:\\n+ matrix: ${{ fromJSON(needs.generate_branches.outputs.matrix) }}\\n+ steps:\\n+ - uses: actions/checkout@v3\\n+\\n+ - uses: tibdex/github-app-token@v1\\n+ id: generate_pr_token\\n+ with:\\n+ app_id: ${{ secrets.SQUAWK_BOT_APP_ID }}\\n+ private_key: ${{ secrets.SQUAWK_BOT_APP_PRIVATE_KEY }}\\n+\\n+ - uses: tibdex/github-app-token@v1\\n+ id: generate_pr_approval_token\\n+ with:\\n+ app_id: ${{ secrets.PR_APPROVAL_BOT_APP_ID }}\\n+ private_key: ${{ secrets.PR_APPROVAL_BOT_APP_PRIVATE_KEY }}\\n+\\n+ - uses: cpcloud/compare-commits-action@v5.0.19\\n+ id: compare_commits\\n+ with:\\n+ token: ${{ steps.generate_pr_token.outputs.token }}\\n+ owner: ibis-project\\n+ repo: ibis\\n+ basehead: ${{ github.sha }}...${{ steps.get_current_commit.outputs.rev }}\\n+ include-merge-commits: false\\n+\\n+ - uses: peter-evans/create-pull-request@v4\\n+ id: create_pr\\n+ with:\\n+ token: ${{ steps.generate_pr_token.outputs.token }}\\n+ commit-message: \\\"chore(${{ matrix.branch }}): rebase onto upstream\\\"\\n+ branch: \\\"create-pull-request/rebase-${{ matrix.branch }}\\\"\\n+ base: ${{ matrix.branch }}\\n+ delete-branch: true\\n+ author: \\\"ibis-squawk-bot[bot] \\\"\\n+ title: \\\"chore(${{ matrix.branch }}): rebase onto upstream\\\"\\n+ body: ${{ steps.compare_commits.outputs.differences }}\\n+ labels: dependencies\\n+\\n+ - uses: juliangruber/approve-pull-request-action@v1.1.1\\n+ if: ${{ fromJSON(steps.create_pr.outputs.pull-request-number) != null }}\\n+ with:\\n+ github-token: ${{ steps.generate_pr_approval_token.outputs.token }}\\n+ number: ${{ steps.create_pr.outputs.pull-request-number }}\\n+\\n+ - uses: peter-evans/enable-pull-request-automerge@v2\\n+ if: ${{ fromJSON(steps.create_pr.outputs.pull-request-number) != null }}\\n+ with:\\n+ token: ${{ steps.generate_pr_token.outputs.token }}\\n+ pull-request-number: ${{ steps.create_pr.outputs.pull-request-number }}\\n+ merge-method: rebase\\n\", \"diff --git a/scripts/prepare.js b/scripts/prepare.js\\nindex 4bab09b..55f459b 100644\\n--- a/scripts/prepare.js\\n+++ b/scripts/prepare.js\\n@@ -96,7 +96,6 @@ async function prepare() {\\n delete json.private\\n delete json.scripts\\n delete json.devDependencies\\n- delete json.types\\n \\n // Add \\\"postinstall\\\" script for donations.\\n if (/(native|core)$/.test(name))\\n@@ -128,6 +127,7 @@ async function prepare() {\\n else {\\n json.main = json.main.replace(/^dist\\\\//, '')\\n if (json.main.endsWith('.cjs.js')) {\\n+ json.types = json.main.replace('.cjs.js', '.d.ts')\\n json.module = json.main.replace('.cjs', '')\\n }\\n }\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"2f5718743a830d40ddf272ad46f253dbb6d08cff\", \"7ece3a9a16780dc6c633bbd903d36ce0aefd6a8a\", \"e82254c4ca73fe55834f005f08bc2a060496f815\", \"f14ef3809f456aadd73523e47cb16c5d15e9a9df\"]"},"types":{"kind":"string","value":"[\"fix\", \"test\", \"ci\", \"build\"]"}}},{"rowIdx":1029,"cells":{"commit_message":{"kind":"string","value":"remove writers from interface,use new, public `quay.io/influxdb/iox` image,create DashboardDetails,change min checked results for score calculation"},"diff":{"kind":"string","value":"[\"diff --git a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\\nindex 91f1b41..eb4b9a8 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/Engine.java\\n@@ -81,8 +81,6 @@ public class Engine implements RecordProcessor {\\n \\n engineContext.setLifecycleListeners(typedRecordProcessors.getLifecycleListeners());\\n recordProcessorMap = typedRecordProcessors.getRecordProcessorMap();\\n-\\n- engineContext.setWriters(writers);\\n }\\n \\n @Override\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\\nindex a8e5538..a27b6e6 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/EngineContext.java\\n@@ -15,7 +15,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.StreamProcessorListene\\n import io.camunda.zeebe.engine.processing.streamprocessor.TypedRecordProcessorFactory;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriter;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\\n import io.camunda.zeebe.engine.state.EventApplier;\\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\\n import java.util.Collections;\\n@@ -34,7 +33,6 @@ public final class EngineContext {\\n private final TypedRecordProcessorFactory typedRecordProcessorFactory;\\n private List lifecycleListeners = Collections.EMPTY_LIST;\\n private StreamProcessorListener streamProcessorListener;\\n- private Writers writers;\\n \\n public EngineContext(\\n final int partitionId,\\n@@ -102,12 +100,4 @@ public final class EngineContext {\\n public void setStreamProcessorListener(final StreamProcessorListener streamProcessorListener) {\\n this.streamProcessorListener = streamProcessorListener;\\n }\\n-\\n- public Writers getWriters() {\\n- return writers;\\n- }\\n-\\n- public void setWriters(final Writers writers) {\\n- this.writers = writers;\\n- }\\n }\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\\nindex f30c7cc..834b421 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/api/ReadonlyStreamProcessorContext.java\\n@@ -8,7 +8,6 @@\\n package io.camunda.zeebe.engine.api;\\n \\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\\n import io.camunda.zeebe.engine.state.mutable.MutableZeebeState;\\n import io.camunda.zeebe.logstreams.log.LogStream;\\n \\n@@ -27,11 +26,6 @@ public interface ReadonlyStreamProcessorContext {\\n LegacyTypedStreamWriter getLogStreamWriter();\\n \\n /**\\n- * @return the specific writers, like command, response, etc\\n- */\\n- Writers getWriters();\\n-\\n- /**\\n * @return the state, where the data is stored during processing\\n */\\n MutableZeebeState getZeebeState();\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\\nindex 844e487..49fd8e2 100755\\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessor.java\\n@@ -346,7 +346,6 @@ public class StreamProcessor extends Actor implements HealthMonitorable, LogReco\\n if (listener != null) {\\n streamProcessorContext.listener(engineContext.getStreamProcessorListener());\\n }\\n- streamProcessorContext.writers(engineContext.getWriters());\\n }\\n \\n private long recoverFromSnapshot() {\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\\nindex d02b273..b527d3c 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/streamprocessor/StreamProcessorContext.java\\n@@ -17,7 +17,6 @@ import io.camunda.zeebe.engine.processing.streamprocessor.writers.CommandRespons\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedResponseWriterImpl;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.LegacyTypedStreamWriter;\\n import io.camunda.zeebe.engine.processing.streamprocessor.writers.NoopLegacyTypedStreamWriter;\\n-import io.camunda.zeebe.engine.processing.streamprocessor.writers.Writers;\\n import io.camunda.zeebe.engine.state.EventApplier;\\n import io.camunda.zeebe.engine.state.KeyGeneratorControls;\\n import io.camunda.zeebe.engine.state.ZeebeDbState;\\n@@ -55,7 +54,7 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\\n private StreamProcessorMode streamProcessorMode = StreamProcessorMode.PROCESSING;\\n private ProcessingScheduleService processingScheduleService;\\n private MutableLastProcessedPositionState lastProcessedPositionState;\\n- private Writers writers;\\n+\\n private LogStreamBatchWriter logStreamBatchWriter;\\n private CommandResponseWriter commandResponseWriter;\\n \\n@@ -85,11 +84,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\\n }\\n \\n @Override\\n- public Writers getWriters() {\\n- return writers;\\n- }\\n-\\n- @Override\\n public MutableZeebeState getZeebeState() {\\n return zeebeState;\\n }\\n@@ -216,10 +210,6 @@ public final class StreamProcessorContext implements ReadonlyStreamProcessorCont\\n return streamProcessorMode;\\n }\\n \\n- public void writers(final Writers writers) {\\n- this.writers = writers;\\n- }\\n-\\n public void logStreamBatchWriter(final LogStreamBatchWriter batchWriter) {\\n logStreamBatchWriter = batchWriter;\\n }\\n\", \"diff --git a/.circleci/config.yml b/.circleci/config.yml\\nindex 3ae6728..a5f2d2f 100644\\n--- a/.circleci/config.yml\\n+++ b/.circleci/config.yml\\n@@ -12,7 +12,7 @@\\n # The CI for every PR and merge to main runs tests, fmt, lints and compiles debug binaries\\n #\\n # On main if all these checks pass it will then additionally compile in \\\"release\\\" mode and\\n-# publish a docker image to quay.io/influxdb/fusion:$COMMIT_SHA\\n+# publish a docker image to quay.io/influxdb/iox:$COMMIT_SHA\\n #\\n # Manual CI Image:\\n #\\n@@ -317,11 +317,11 @@ jobs:\\n #\\n # Uses the latest ci_image (influxdb/rust below) to build a release binary and\\n # copies it to a minimal container image based upon `rust:slim-buster`. This\\n- # minimal image is then pushed to `quay.io/influxdb/fusion:${BRANCH}` with '/'\\n+ # minimal image is then pushed to `quay.io/influxdb/iox:${BRANCH}` with '/'\\n # repaced by '.' - as an example:\\n #\\n # git branch: dom/my-awesome-feature/perf\\n- # container: quay.io/influxdb/fusion:dom.my-awesome-feature.perf\\n+ # container: quay.io/influxdb/iox:dom.my-awesome-feature.perf\\n #\\n # Subsequent CI runs will overwrite the tag if you push more changes, so watch\\n # out for parallel CI runs!\\n@@ -365,7 +365,7 @@ jobs:\\n sudo apt-get update\\n sudo apt-get install -y docker.io\\n - run: |\\n- echo \\\"$QUAY_PASS\\\" | docker login quay.io --username $QUAY_USER --password-stdin\\n+ echo \\\"$QUAY_INFLUXDB_IOX_PASS\\\" | docker login quay.io --username $QUAY_INFLUXDB_IOX_USER --password-stdin\\n - run:\\n # Docker has functionality to support per-Dockerfile .dockerignore\\n # This was added in https://github.com/moby/buildkit/pull/901\\n@@ -379,8 +379,8 @@ jobs:\\n echo sha256sum after build is\\n sha256sum target/release/influxdb_iox\\n COMMIT_SHA=$(git rev-parse --short HEAD)\\n- docker build -t quay.io/influxdb/fusion:$COMMIT_SHA -f docker/Dockerfile.iox .\\n- docker push quay.io/influxdb/fusion:$COMMIT_SHA\\n+ docker build -t quay.io/influxdb/iox:$COMMIT_SHA -f docker/Dockerfile.iox .\\n+ docker push quay.io/influxdb/iox:$COMMIT_SHA\\n echo \\\"export COMMIT_SHA=${COMMIT_SHA}\\\" >> $BASH_ENV\\n - run:\\n name: Deploy tags\\n\", \"diff --git a/client/src/components/MentorSearch.tsx b/client/src/components/MentorSearch.tsx\\nindex 622560a..06f0114 100644\\n--- a/client/src/components/MentorSearch.tsx\\n+++ b/client/src/components/MentorSearch.tsx\\n@@ -7,8 +7,9 @@ type Props = UserProps & {\\n };\\n \\n export function MentorSearch(props: Props) {\\n- const courseService = useMemo(() => new CourseService(props.courseId), [props.courseId]);\\n+ const { courseId, ...otherProps } = props;\\n+ const courseService = useMemo(() => new CourseService(courseId), [courseId]);\\n const handleSearch = useCallback(async (value: string) => courseService.searchMentors(value), [courseService]);\\n \\n- return ;\\n+ return ;\\n }\\ndiff --git a/client/src/components/Student/DashboardDetails.tsx b/client/src/components/Student/DashboardDetails.tsx\\nnew file mode 100644\\nindex 0000000..30506ef\\n--- /dev/null\\n+++ b/client/src/components/Student/DashboardDetails.tsx\\n@@ -0,0 +1,89 @@\\n+import { BranchesOutlined, CloseCircleTwoTone, SolutionOutlined, UndoOutlined } from '@ant-design/icons';\\n+import { Button, Descriptions, Drawer } from 'antd';\\n+import { CommentModal, MentorSearch } from 'components';\\n+import { useState } from 'react';\\n+import { StudentDetails } from 'services/course';\\n+import { MentorBasic } from '../../../../common/models';\\n+import css from 'styled-jsx/css';\\n+\\n+type Props = {\\n+ details: StudentDetails | null;\\n+ courseId: number;\\n+ onClose: () => void;\\n+ onCreateRepository: () => void;\\n+ onRestoreStudent: () => void;\\n+ onExpelStudent: (comment: string) => void;\\n+ onIssueCertificate: () => void;\\n+ onUpdateMentor: (githubId: string) => void;\\n+};\\n+\\n+export function DashboardDetails(props: Props) {\\n+ const [expelMode, setExpelMode] = useState(false);\\n+ const { details } = props;\\n+ if (details == null) {\\n+ return null;\\n+ }\\n+ return (\\n+ <>\\n+ \\n+
\\n+ }\\n+ onClick={props.onCreateRepository}\\n+ >\\n+ Create Repository\\n+ \\n+ \\n+
\\n+ \\n+ setExpelMode(false)}\\n+ onOk={(text: string) => {\\n+ props.onExpelStudent(text);\\n+ setExpelMode(false);\\n+ }}\\n+ />\\n+ \\n+ \\n+ );\\n+}\\n+\\n+const styles = css`\\n+ .student-details-actions :global(.ant-btn) {\\n+ margin: 0 8px 8px 0;\\n+ }\\n+`;\\ndiff --git a/client/src/components/Student/index.ts b/client/src/components/Student/index.ts\\nindex 71e28de..076f0e2 100644\\n--- a/client/src/components/Student/index.ts\\n+++ b/client/src/components/Student/index.ts\\n@@ -1 +1,2 @@\\n export { default as AssignStudentModal } from './AssignStudentModal';\\n+export { DashboardDetails } from './DashboardDetails';\\ndiff --git a/client/src/components/StudentSearch.tsx b/client/src/components/StudentSearch.tsx\\nindex 5952aed..7c14263 100644\\n--- a/client/src/components/StudentSearch.tsx\\n+++ b/client/src/components/StudentSearch.tsx\\n@@ -7,8 +7,9 @@ type Props = UserProps & {\\n };\\n \\n export function StudentSearch(props: Props) {\\n- const courseService = useMemo(() => new CourseService(props.courseId), [props.courseId]);\\n+ const { courseId, ...otherProps } = props;\\n+ const courseService = useMemo(() => new CourseService(courseId), [courseId]);\\n const handleSearch = useCallback(async (value: string) => courseService.searchStudents(value), [courseService]);\\n \\n- return ;\\n+ return ;\\n }\\ndiff --git a/client/src/components/UserSearch.tsx b/client/src/components/UserSearch.tsx\\nindex ff95941..4075827 100644\\n--- a/client/src/components/UserSearch.tsx\\n+++ b/client/src/components/UserSearch.tsx\\n@@ -14,7 +14,7 @@ export type UserProps = SelectProps & {\\n \\n export function UserSearch(props: UserProps) {\\n const [data, setData] = useState([]);\\n- const { searchFn = defaultSearch, defaultValues } = props;\\n+ const { searchFn = defaultSearch, defaultValues, keyField, ...otherProps } = props;\\n \\n useEffect(() => {\\n setData(defaultValues ?? []);\\n@@ -29,7 +29,6 @@ export function UserSearch(props: UserProps) {\\n }\\n };\\n \\n- const { keyField, ...otherProps } = props;\\n return (\\n new CourseService(courseId), [courseId]);\\n const [students, setStudents] = useState([] as StudentDetails[]);\\n@@ -77,7 +72,6 @@ function Page(props: Props) {\\n await courseService.expelStudent(githubId, text);\\n message.info('Student has been expelled');\\n }\\n- setExpelMode(false);\\n });\\n \\n const restoreStudent = withLoading(async () => {\\n@@ -114,59 +108,20 @@ function Page(props: Props) {\\n
{renderToolbar()}
\\n \\n \\n- {\\n setDetails(null);\\n loadStudents();\\n }}\\n- visible={!!details}\\n- >\\n-
\\n- }\\n- onClick={createRepository}\\n- >\\n- Create Repository\\n- \\n- \\n-
\\n- \\n- setExpelMode(false)}\\n- onOk={expelStudent}\\n+ details={details}\\n+ courseId={props.course.id}\\n />\\n- \\n \\n );\\n }\\n@@ -306,14 +261,4 @@ function calculateStats(students: StudentDetails[]) {\\n };\\n }\\n \\n-const styles = css`\\n- :global(.rs-table-row-disabled) {\\n- opacity: 0.25;\\n- }\\n-\\n- .student-details-actions :global(.ant-btn) {\\n- margin: 0 8px 8px 0;\\n- }\\n-`;\\n-\\n export default withCourseData(withSession(Page));\\ndiff --git a/client/src/styles/main.css b/client/src/styles/main.css\\nindex 2ccac3c..df3cc8c 100644\\n--- a/client/src/styles/main.css\\n+++ b/client/src/styles/main.css\\n@@ -21,6 +21,10 @@ body,\\n display: none;\\n }\\n \\n+.ant-drawer-content-wrapper {\\n+ max-width: 85%;\\n+}\\n+\\n .footer-dark.ant-layout-footer {\\n background: #000;\\n color: #fff;\\n\", \"diff --git a/server/src/services/courseService.ts b/server/src/services/courseService.ts\\nindex 89633f4..10bfc55 100644\\n--- a/server/src/services/courseService.ts\\n+++ b/server/src/services/courseService.ts\\n@@ -580,8 +580,7 @@ export async function getTaskSolutionCheckers(courseTaskId: number, minCheckedCo\\n .createQueryBuilder('tsr')\\n .select('tsr.\\\"studentId\\\", ROUND(AVG(tsr.score)) as \\\"score\\\"')\\n .where(qb => {\\n- // query students with 3 checked tasks\\n-\\n+ // query students who checked enough tasks\\n const query = qb\\n .subQuery()\\n .select('r.\\\"checkerId\\\"')\\n@@ -600,7 +599,7 @@ export async function getTaskSolutionCheckers(courseTaskId: number, minCheckedCo\\n })\\n .andWhere('tsr.\\\"courseTaskId\\\" = :courseTaskId', { courseTaskId })\\n .groupBy('tsr.\\\"studentId\\\"')\\n- .having(`COUNT(tsr.id) >= :count`, { count: minCheckedCount })\\n+ .having(`COUNT(tsr.id) >= :count`, { count: minCheckedCount - 1 })\\n .getRawMany();\\n \\n return records.map(record => ({ studentId: record.studentId, score: Number(record.score) }));\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"d2709cab63295109dcd1a49f57da9418110e9044\", \"f751bb5426b87f82096d620f1cd6203badf45d58\", \"fd5f211916c989fddc2ee5afeeb7d46e6a2f51cb\", \"fd849bd08363df60dbc8b9b6d55bac4f5ace88f4\"]"},"types":{"kind":"string","value":"[\"refactor\", \"ci\", \"feat\", \"docs\"]"}}},{"rowIdx":1030,"cells":{"commit_message":{"kind":"string","value":"Port shard precreation service from InfluxDB 1.x\n\nProvides new configuration parameters:\n\n```\n--storage-shard-precreator-advance-period\n--storage-shard-precreator-check-interval\n```\n\nCloses #19520,simplify loadFiles code,add activatedElementInstanceKeys to modification record,import flux-lsp v0.5.21"},"diff":{"kind":"string","value":"[\"diff --git a/cmd/influxd/launcher/launcher.go b/cmd/influxd/launcher/launcher.go\\nindex e3548ef..5559e94 100644\\n--- a/cmd/influxd/launcher/launcher.go\\n+++ b/cmd/influxd/launcher/launcher.go\\n@@ -440,6 +440,16 @@ func launcherOpts(l *Launcher) []cli.Opt {\\n \\t\\t\\tFlag: \\\"storage-retention-check-interval\\\",\\n \\t\\t\\tDesc: \\\"The interval of time when retention policy enforcement checks run.\\\",\\n \\t\\t},\\n+\\t\\t{\\n+\\t\\t\\tDestP: &l.StorageConfig.PrecreatorConfig.CheckInterval,\\n+\\t\\t\\tFlag: \\\"storage-shard-precreator-check-interval\\\",\\n+\\t\\t\\tDesc: \\\"The interval of time when the check to pre-create new shards runs.\\\",\\n+\\t\\t},\\n+\\t\\t{\\n+\\t\\t\\tDestP: &l.StorageConfig.PrecreatorConfig.AdvancePeriod,\\n+\\t\\t\\tFlag: \\\"storage-shard-precreator-advance-period\\\",\\n+\\t\\t\\tDesc: \\\"The default period ahead of the endtime of a shard group that its successor group is created.\\\",\\n+\\t\\t},\\n \\n \\t\\t// InfluxQL Coordinator Config\\n \\t\\t{\\ndiff --git a/storage/config.go b/storage/config.go\\nindex ef953a2..d8e24db 100644\\n--- a/storage/config.go\\n+++ b/storage/config.go\\n@@ -2,6 +2,7 @@ package storage\\n \\n import (\\n \\t\\\"github.com/influxdata/influxdb/v2/tsdb\\\"\\n+\\t\\\"github.com/influxdata/influxdb/v2/v1/services/precreator\\\"\\n \\t\\\"github.com/influxdata/influxdb/v2/v1/services/retention\\\"\\n )\\n \\n@@ -10,6 +11,7 @@ type Config struct {\\n \\tData tsdb.Config\\n \\n \\tRetentionService retention.Config\\n+\\tPrecreatorConfig precreator.Config\\n }\\n \\n // NewConfig initialises a new config for an Engine.\\n@@ -17,5 +19,6 @@ func NewConfig() Config {\\n \\treturn Config{\\n \\t\\tData: tsdb.NewConfig(),\\n \\t\\tRetentionService: retention.NewConfig(),\\n+\\t\\tPrecreatorConfig: precreator.NewConfig(),\\n \\t}\\n }\\ndiff --git a/storage/engine.go b/storage/engine.go\\nindex 8518f48..ae37fdd 100644\\n--- a/storage/engine.go\\n+++ b/storage/engine.go\\n@@ -19,6 +19,7 @@ import (\\n \\t_ \\\"github.com/influxdata/influxdb/v2/tsdb/index/tsi1\\\"\\n \\t\\\"github.com/influxdata/influxdb/v2/v1/coordinator\\\"\\n \\t\\\"github.com/influxdata/influxdb/v2/v1/services/meta\\\"\\n+\\t\\\"github.com/influxdata/influxdb/v2/v1/services/precreator\\\"\\n \\t\\\"github.com/influxdata/influxdb/v2/v1/services/retention\\\"\\n \\t\\\"github.com/influxdata/influxql\\\"\\n \\t\\\"github.com/pkg/errors\\\"\\n@@ -42,7 +43,8 @@ type Engine struct {\\n \\t\\tWritePoints(database, retentionPolicy string, consistencyLevel models.ConsistencyLevel, user meta.User, points []models.Point) error\\n \\t}\\n \\n-\\tretentionService *retention.Service\\n+\\tretentionService *retention.Service\\n+\\tprecreatorService *precreator.Service\\n \\n \\tdefaultMetricLabels prometheus.Labels\\n \\n@@ -66,6 +68,7 @@ type MetaClient interface {\\n \\tDatabase(name string) (di *meta.DatabaseInfo)\\n \\tDatabases() []meta.DatabaseInfo\\n \\tDeleteShardGroup(database, policy string, id uint64) error\\n+\\tPrecreateShardGroups(now, cutoff time.Time) error\\n \\tPruneShardGroups() error\\n \\tRetentionPolicy(database, policy string) (*meta.RetentionPolicyInfo, error)\\n \\tShardGroupsByTimeRange(database, policy string, min, max time.Time) (a []meta.ShardGroupInfo, err error)\\n@@ -115,6 +118,9 @@ func NewEngine(path string, c Config, options ...Option) *Engine {\\n \\te.retentionService.TSDBStore = e.tsdbStore\\n \\te.retentionService.MetaClient = e.metaClient\\n \\n+\\te.precreatorService = precreator.NewService(c.PrecreatorConfig)\\n+\\te.precreatorService.MetaClient = e.metaClient\\n+\\n \\treturn e\\n }\\n \\n@@ -132,6 +138,10 @@ func (e *Engine) WithLogger(log *zap.Logger) {\\n \\tif e.retentionService != nil {\\n \\t\\te.retentionService.WithLogger(log)\\n \\t}\\n+\\n+\\tif e.precreatorService != nil {\\n+\\t\\te.precreatorService.WithLogger(log)\\n+\\t}\\n }\\n \\n // PrometheusCollectors returns all the prometheus collectors associated with\\n@@ -161,6 +171,10 @@ func (e *Engine) Open(ctx context.Context) (err error) {\\n \\t\\treturn err\\n \\t}\\n \\n+\\tif err := e.precreatorService.Open(ctx); err != nil {\\n+\\t\\treturn err\\n+\\t}\\n+\\n \\te.closing = make(chan struct{})\\n \\n \\treturn nil\\n@@ -194,6 +208,10 @@ func (e *Engine) Close() error {\\n \\n \\tvar retErr *multierror.Error\\n \\n+\\tif err := e.precreatorService.Close(); err != nil {\\n+\\t\\tretErr = multierror.Append(retErr, fmt.Errorf(\\\"error closing shard precreator service: %w\\\", err))\\n+\\t}\\n+\\n \\tif err := e.retentionService.Close(); err != nil {\\n \\t\\tretErr = multierror.Append(retErr, fmt.Errorf(\\\"error closing retention service: %w\\\", err))\\n \\t}\\ndiff --git a/v1/services/precreator/README.md b/v1/services/precreator/README.md\\nnew file mode 100644\\nindex 0000000..8830b73\\n--- /dev/null\\n+++ b/v1/services/precreator/README.md\\n@@ -0,0 +1,13 @@\\n+Shard Precreation\\n+============\\n+\\n+During normal operation when InfluxDB receives time-series data, it writes the data to files known as _shards_. Each shard only contains data for a specific range of time. Therefore, before data can be accepted by the system, the shards must exist and InfluxDB always checks that the required shards exist for every incoming data point. If the required shards do not exist, InfluxDB will create those shards. Because this requires a cluster to reach consensus, the process is not instantaneous and can temporarily impact write-throughput.\\n+\\n+Since almost all time-series data is written sequentially in time, the system has an excellent idea of the timestamps of future data. Shard precreation takes advantage of this fact by creating required shards ahead of time, thereby ensuring the required shards exist by the time new time-series data actually arrives. Write-throughput is therefore not affected when data is first received for a range of time that would normally trigger shard creation.\\n+\\n+Note that the shard-existence check must remain in place in the code, even with shard precreation. This is because while most data is written sequentially in time, this is not always the case. Data may be written with timestamps in the past, or farther in the future than shard precreation handles.\\n+\\n+## Configuration\\n+Shard precreation can be disabled if necessary, though this is not recommended. If it is disabled, then shards will be only be created when explicitly needed.\\n+\\n+The interval between runs of the shard precreation service, as well as the time-in-advance the shards are created, are also configurable. The defaults should work for most deployments.\\ndiff --git a/v1/services/precreator/config.go b/v1/services/precreator/config.go\\nnew file mode 100644\\nindex 0000000..5e994e6\\n--- /dev/null\\n+++ b/v1/services/precreator/config.go\\n@@ -0,0 +1,65 @@\\n+package precreator\\n+\\n+import (\\n+\\t\\\"errors\\\"\\n+\\t\\\"time\\\"\\n+\\n+\\t\\\"github.com/influxdata/influxdb/v2/toml\\\"\\n+\\t\\\"github.com/influxdata/influxdb/v2/v1/monitor/diagnostics\\\"\\n+)\\n+\\n+const (\\n+\\t// DefaultCheckInterval is the shard precreation check time if none is specified.\\n+\\tDefaultCheckInterval = 10 * time.Minute\\n+\\n+\\t// DefaultAdvancePeriod is the default period ahead of the endtime of a shard group\\n+\\t// that its successor group is created.\\n+\\tDefaultAdvancePeriod = 30 * time.Minute\\n+)\\n+\\n+// Config represents the configuration for shard precreation.\\n+type Config struct {\\n+\\tEnabled bool `toml:\\\"enabled\\\"`\\n+\\tCheckInterval toml.Duration `toml:\\\"check-interval\\\"`\\n+\\tAdvancePeriod toml.Duration `toml:\\\"advance-period\\\"`\\n+}\\n+\\n+// NewConfig returns a new Config with defaults.\\n+func NewConfig() Config {\\n+\\treturn Config{\\n+\\t\\tEnabled: true,\\n+\\t\\tCheckInterval: toml.Duration(DefaultCheckInterval),\\n+\\t\\tAdvancePeriod: toml.Duration(DefaultAdvancePeriod),\\n+\\t}\\n+}\\n+\\n+// Validate returns an error if the Config is invalid.\\n+func (c Config) Validate() error {\\n+\\tif !c.Enabled {\\n+\\t\\treturn nil\\n+\\t}\\n+\\n+\\tif c.CheckInterval <= 0 {\\n+\\t\\treturn errors.New(\\\"check-interval must be positive\\\")\\n+\\t}\\n+\\tif c.AdvancePeriod <= 0 {\\n+\\t\\treturn errors.New(\\\"advance-period must be positive\\\")\\n+\\t}\\n+\\n+\\treturn nil\\n+}\\n+\\n+// Diagnostics returns a diagnostics representation of a subset of the Config.\\n+func (c Config) Diagnostics() (*diagnostics.Diagnostics, error) {\\n+\\tif !c.Enabled {\\n+\\t\\treturn diagnostics.RowFromMap(map[string]interface{}{\\n+\\t\\t\\t\\\"enabled\\\": false,\\n+\\t\\t}), nil\\n+\\t}\\n+\\n+\\treturn diagnostics.RowFromMap(map[string]interface{}{\\n+\\t\\t\\\"enabled\\\": true,\\n+\\t\\t\\\"check-interval\\\": c.CheckInterval,\\n+\\t\\t\\\"advance-period\\\": c.AdvancePeriod,\\n+\\t}), nil\\n+}\\ndiff --git a/v1/services/precreator/config_test.go b/v1/services/precreator/config_test.go\\nnew file mode 100644\\nindex 0000000..2686001\\n--- /dev/null\\n+++ b/v1/services/precreator/config_test.go\\n@@ -0,0 +1,67 @@\\n+package precreator_test\\n+\\n+import (\\n+\\t\\\"testing\\\"\\n+\\t\\\"time\\\"\\n+\\n+\\t\\\"github.com/BurntSushi/toml\\\"\\n+\\t\\\"github.com/influxdata/influxdb/v2/v1/services/precreator\\\"\\n+)\\n+\\n+func TestConfig_Parse(t *testing.T) {\\n+\\t// Parse configuration.\\n+\\tvar c precreator.Config\\n+\\tif _, err := toml.Decode(`\\n+enabled = true\\n+check-interval = \\\"2m\\\"\\n+advance-period = \\\"10m\\\"\\n+`, &c); err != nil {\\n+\\n+\\t\\tt.Fatal(err)\\n+\\t}\\n+\\n+\\t// Validate configuration.\\n+\\tif !c.Enabled {\\n+\\t\\tt.Fatalf(\\\"unexpected enabled state: %v\\\", c.Enabled)\\n+\\t} else if time.Duration(c.CheckInterval) != 2*time.Minute {\\n+\\t\\tt.Fatalf(\\\"unexpected check interval: %s\\\", c.CheckInterval)\\n+\\t} else if time.Duration(c.AdvancePeriod) != 10*time.Minute {\\n+\\t\\tt.Fatalf(\\\"unexpected advance period: %s\\\", c.AdvancePeriod)\\n+\\t}\\n+}\\n+\\n+func TestConfig_Validate(t *testing.T) {\\n+\\tc := precreator.NewConfig()\\n+\\tif err := c.Validate(); err != nil {\\n+\\t\\tt.Fatalf(\\\"unexpected validation fail from NewConfig: %s\\\", err)\\n+\\t}\\n+\\n+\\tc = precreator.NewConfig()\\n+\\tc.CheckInterval = 0\\n+\\tif err := c.Validate(); err == nil {\\n+\\t\\tt.Fatal(\\\"expected error for check-interval = 0, got nil\\\")\\n+\\t}\\n+\\n+\\tc = precreator.NewConfig()\\n+\\tc.CheckInterval *= -1\\n+\\tif err := c.Validate(); err == nil {\\n+\\t\\tt.Fatal(\\\"expected error for negative check-interval, got nil\\\")\\n+\\t}\\n+\\n+\\tc = precreator.NewConfig()\\n+\\tc.AdvancePeriod = 0\\n+\\tif err := c.Validate(); err == nil {\\n+\\t\\tt.Fatal(\\\"expected error for advance-period = 0, got nil\\\")\\n+\\t}\\n+\\n+\\tc = precreator.NewConfig()\\n+\\tc.AdvancePeriod *= -1\\n+\\tif err := c.Validate(); err == nil {\\n+\\t\\tt.Fatal(\\\"expected error for negative advance-period, got nil\\\")\\n+\\t}\\n+\\n+\\tc.Enabled = false\\n+\\tif err := c.Validate(); err != nil {\\n+\\t\\tt.Fatalf(\\\"unexpected validation fail from disabled config: %s\\\", err)\\n+\\t}\\n+}\\ndiff --git a/v1/services/precreator/service.go b/v1/services/precreator/service.go\\nnew file mode 100644\\nindex 0000000..28e8f16\\n--- /dev/null\\n+++ b/v1/services/precreator/service.go\\n@@ -0,0 +1,93 @@\\n+// Package precreator provides the shard precreation service.\\n+package precreator // import \\\"github.com/influxdata/influxdb/v2/v1/services/precreator\\\"\\n+\\n+import (\\n+\\t\\\"context\\\"\\n+\\t\\\"sync\\\"\\n+\\t\\\"time\\\"\\n+\\n+\\t\\\"github.com/influxdata/influxdb/v2/logger\\\"\\n+\\t\\\"go.uber.org/zap\\\"\\n+)\\n+\\n+// Service manages the shard precreation service.\\n+type Service struct {\\n+\\tcheckInterval time.Duration\\n+\\tadvancePeriod time.Duration\\n+\\n+\\tLogger *zap.Logger\\n+\\n+\\tcancel context.CancelFunc\\n+\\twg sync.WaitGroup\\n+\\n+\\tMetaClient interface {\\n+\\t\\tPrecreateShardGroups(now, cutoff time.Time) error\\n+\\t}\\n+}\\n+\\n+// NewService returns an instance of the precreation service.\\n+func NewService(c Config) *Service {\\n+\\treturn &Service{\\n+\\t\\tcheckInterval: time.Duration(c.CheckInterval),\\n+\\t\\tadvancePeriod: time.Duration(c.AdvancePeriod),\\n+\\t\\tLogger: zap.NewNop(),\\n+\\t}\\n+}\\n+\\n+// WithLogger sets the logger for the service.\\n+func (s *Service) WithLogger(log *zap.Logger) {\\n+\\ts.Logger = log.With(zap.String(\\\"service\\\", \\\"shard-precreation\\\"))\\n+}\\n+\\n+// Open starts the precreation service.\\n+func (s *Service) Open(ctx context.Context) error {\\n+\\tif s.cancel != nil {\\n+\\t\\treturn nil\\n+\\t}\\n+\\n+\\ts.Logger.Info(\\\"Starting precreation service\\\",\\n+\\t\\tlogger.DurationLiteral(\\\"check_interval\\\", s.checkInterval),\\n+\\t\\tlogger.DurationLiteral(\\\"advance_period\\\", s.advancePeriod))\\n+\\n+\\tctx, s.cancel = context.WithCancel(ctx)\\n+\\n+\\ts.wg.Add(1)\\n+\\tgo s.runPrecreation(ctx)\\n+\\treturn nil\\n+}\\n+\\n+// Close stops the precreation service.\\n+func (s *Service) Close() error {\\n+\\tif s.cancel == nil {\\n+\\t\\treturn nil\\n+\\t}\\n+\\n+\\ts.cancel()\\n+\\ts.wg.Wait()\\n+\\ts.cancel = nil\\n+\\n+\\treturn nil\\n+}\\n+\\n+// runPrecreation continually checks if resources need precreation.\\n+func (s *Service) runPrecreation(ctx context.Context) {\\n+\\tdefer s.wg.Done()\\n+\\n+\\tfor {\\n+\\t\\tselect {\\n+\\t\\tcase <-time.After(s.checkInterval):\\n+\\t\\t\\tif err := s.precreate(time.Now().UTC()); err != nil {\\n+\\t\\t\\t\\ts.Logger.Info(\\\"Failed to precreate shards\\\", zap.Error(err))\\n+\\t\\t\\t}\\n+\\t\\tcase <-ctx.Done():\\n+\\t\\t\\ts.Logger.Info(\\\"Terminating precreation service\\\")\\n+\\t\\t\\treturn\\n+\\t\\t}\\n+\\t}\\n+}\\n+\\n+// precreate performs actual resource precreation.\\n+func (s *Service) precreate(now time.Time) error {\\n+\\tcutoff := now.Add(s.advancePeriod).UTC()\\n+\\treturn s.MetaClient.PrecreateShardGroups(now, cutoff)\\n+}\\ndiff --git a/v1/services/precreator/service_test.go b/v1/services/precreator/service_test.go\\nnew file mode 100644\\nindex 0000000..20289b7\\n--- /dev/null\\n+++ b/v1/services/precreator/service_test.go\\n@@ -0,0 +1,56 @@\\n+package precreator_test\\n+\\n+import (\\n+\\t\\\"context\\\"\\n+\\t\\\"os\\\"\\n+\\t\\\"testing\\\"\\n+\\t\\\"time\\\"\\n+\\n+\\t\\\"github.com/influxdata/influxdb/v2/logger\\\"\\n+\\t\\\"github.com/influxdata/influxdb/v2/toml\\\"\\n+\\t\\\"github.com/influxdata/influxdb/v2/v1/internal\\\"\\n+\\t\\\"github.com/influxdata/influxdb/v2/v1/services/precreator\\\"\\n+)\\n+\\n+func TestShardPrecreation(t *testing.T) {\\n+\\tdone := make(chan struct{})\\n+\\tprecreate := false\\n+\\n+\\tvar mc internal.MetaClientMock\\n+\\tmc.PrecreateShardGroupsFn = func(now, cutoff time.Time) error {\\n+\\t\\tif !precreate {\\n+\\t\\t\\tclose(done)\\n+\\t\\t\\tprecreate = true\\n+\\t\\t}\\n+\\t\\treturn nil\\n+\\t}\\n+\\n+\\ts := NewTestService()\\n+\\ts.MetaClient = &mc\\n+\\n+\\tif err := s.Open(context.Background()); err != nil {\\n+\\t\\tt.Fatalf(\\\"unexpected open error: %s\\\", err)\\n+\\t}\\n+\\tdefer s.Close() // double close should not cause a panic\\n+\\n+\\ttimer := time.NewTimer(100 * time.Millisecond)\\n+\\tselect {\\n+\\tcase <-done:\\n+\\t\\ttimer.Stop()\\n+\\tcase <-timer.C:\\n+\\t\\tt.Errorf(\\\"timeout exceeded while waiting for precreate\\\")\\n+\\t}\\n+\\n+\\tif err := s.Close(); err != nil {\\n+\\t\\tt.Fatalf(\\\"unexpected close error: %s\\\", err)\\n+\\t}\\n+}\\n+\\n+func NewTestService() *precreator.Service {\\n+\\tconfig := precreator.NewConfig()\\n+\\tconfig.CheckInterval = toml.Duration(10 * time.Millisecond)\\n+\\n+\\ts := precreator.NewService(config)\\n+\\ts.WithLogger(logger.New(os.Stderr))\\n+\\treturn s\\n+}\\n\", \"diff --git a/frontend/app/player/web/network/loadFiles.ts b/frontend/app/player/web/network/loadFiles.ts\\nindex ec174fc..d164333 100644\\n--- a/frontend/app/player/web/network/loadFiles.ts\\n+++ b/frontend/app/player/web/network/loadFiles.ts\\n@@ -1,43 +1,33 @@\\n import APIClient from 'App/api_client';\\n \\n-const NO_NTH_FILE = \\\"nnf\\\"\\n-const NO_UNPROCESSED_FILES = \\\"nuf\\\"\\n+const NO_FILE_OK = \\\"No-file-but-this-is-ok\\\"\\n+const NO_BACKUP_FILE = \\\"No-efs-file\\\"\\n \\n export const loadFiles = (\\n urls: string[],\\n onData: (data: Uint8Array) => void,\\n ): Promise => {\\n- const firstFileURL = urls[0]\\n- urls = urls.slice(1)\\n- if (!firstFileURL) {\\n+ if (!urls.length) {\\n return Promise.reject(\\\"No urls provided\\\")\\n }\\n- return window.fetch(firstFileURL)\\n- .then(r => {\\n- return processAPIStreamResponse(r, true)\\n- })\\n- .then(onData)\\n- .then(() =>\\n- urls.reduce((p, url) =>\\n- p.then(() =>\\n- window.fetch(url)\\n- .then(r => {\\n- return processAPIStreamResponse(r, false)\\n- })\\n- .then(onData)\\n- ),\\n- Promise.resolve(),\\n- )\\n+ return urls.reduce((p, url, index) =>\\n+ p.then(() =>\\n+ window.fetch(url)\\n+ .then(r => {\\n+ return processAPIStreamResponse(r, index===0)\\n+ })\\n+ .then(onData)\\n+ ),\\n+ Promise.resolve(),\\n )\\n .catch(e => {\\n- if (e === NO_NTH_FILE) {\\n+ if (e === NO_FILE_OK) {\\n return\\n }\\n throw e\\n })\\n }\\n \\n-\\n export async function requestEFSDom(sessionId: string) {\\n return await requestEFSMobFile(sessionId + \\\"/dom.mob\\\")\\n }\\n@@ -50,21 +40,18 @@ async function requestEFSMobFile(filename: string) {\\n const api = new APIClient()\\n const res = await api.fetch('/unprocessed/' + filename)\\n if (res.status >= 400) {\\n- throw NO_UNPROCESSED_FILES\\n+ throw NO_BACKUP_FILE\\n }\\n return await processAPIStreamResponse(res, false)\\n }\\n \\n-const processAPIStreamResponse = (response: Response, isFirstFile: boolean) => {\\n+const processAPIStreamResponse = (response: Response, canBeMissed: boolean) => {\\n return new Promise((res, rej) => {\\n- if (response.status === 404 && !isFirstFile) {\\n- return rej(NO_NTH_FILE)\\n+ if (response.status === 404 && canBeMissed) {\\n+ return rej(NO_FILE_OK)\\n }\\n if (response.status >= 400) {\\n- return rej(\\n- isFirstFile ? `no start file. status code ${ response.status }`\\n- : `Bad endfile status code ${response.status}`\\n- )\\n+ return rej(`Bad file status code ${response.status}. Url: ${response.url}`)\\n }\\n res(response.arrayBuffer())\\n }).then(buffer => new Uint8Array(buffer))\\n\", \"diff --git a/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java b/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java\\nindex 33410da..edd0588 100644\\n--- a/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java\\n+++ b/protocol-impl/src/test/java/io/camunda/zeebe/protocol/impl/JsonSerializableToJsonTest.java\\n@@ -787,7 +787,8 @@ final class JsonSerializableToJsonTest {\\n }\\n }],\\n \\\"elementId\\\": \\\"activity\\\"\\n- }]\\n+ }],\\n+ \\\"activatedElementInstanceKeys\\\": []\\n }\\n \\\"\\\"\\\"\\n },\\n@@ -803,7 +804,8 @@ final class JsonSerializableToJsonTest {\\n {\\n \\\"processInstanceKey\\\": 1,\\n \\\"terminateInstructions\\\": [],\\n- \\\"activateInstructions\\\": []\\n+ \\\"activateInstructions\\\": [],\\n+ \\\"activatedElementInstanceKeys\\\": []\\n }\\n \\\"\\\"\\\"\\n },\\n\", \"diff --git a/ui/package.json b/ui/package.json\\nindex 7a44aad..a36fc3d 100644\\n--- a/ui/package.json\\n+++ b/ui/package.json\\n@@ -134,7 +134,7 @@\\n \\\"dependencies\\\": {\\n \\\"@influxdata/clockface\\\": \\\"2.3.4\\\",\\n \\\"@influxdata/flux\\\": \\\"^0.5.1\\\",\\n- \\\"@influxdata/flux-lsp-browser\\\": \\\"0.5.20\\\",\\n+ \\\"@influxdata/flux-lsp-browser\\\": \\\"0.5.21\\\",\\n \\\"@influxdata/giraffe\\\": \\\"0.29.0\\\",\\n \\\"@influxdata/influx\\\": \\\"0.5.5\\\",\\n \\\"@influxdata/influxdb-templates\\\": \\\"0.9.0\\\",\\ndiff --git a/ui/yarn.lock b/ui/yarn.lock\\nindex 99ae766..e6e2a47 100644\\n--- a/ui/yarn.lock\\n+++ b/ui/yarn.lock\\n@@ -752,10 +752,10 @@\\n resolved \\\"https://registry.yarnpkg.com/@influxdata/clockface/-/clockface-2.3.4.tgz#9c496601253e1d49cbeae29a7b9cfb54862785f6\\\"\\n integrity sha512-mmz3YElK8Ho+1onEafuas6sVhIT638JA4NbDTO3bVJgK1TG7AnU4rQP+c6fj7vZSfvrIwtOwGaMONJTaww5o6w==\\n \\n-\\\"@influxdata/flux-lsp-browser@0.5.20\\\":\\n- version \\\"0.5.20\\\"\\n- resolved \\\"https://registry.yarnpkg.com/@influxdata/flux-lsp-browser/-/flux-lsp-browser-0.5.20.tgz#150d261bab869e130f6d00ee73ea4e859e8969e4\\\"\\n- integrity sha512-gUy19t/QndkJPmyv7Lb56zXxaW5v7R9TslTHt0hB0GJjo7lmYkRfkD7DELdFHrD2e/CLtcNQBnczIMIGkII8Bw==\\n+\\\"@influxdata/flux-lsp-browser@0.5.21\\\":\\n+ version \\\"0.5.21\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/@influxdata/flux-lsp-browser/-/flux-lsp-browser-0.5.21.tgz#d5632f45e925c09bae9501a00fbef2ed55567f9e\\\"\\n+ integrity sha512-lcUwKX1yj0QqGiusQFOVi7UPsvp6+qNX7Cwf9qqS5/dRwoh7c++nFVRdGNrSWlsbyRrPaAWBoZWEnghSnIf6DQ==\\n \\n \\\"@influxdata/flux@^0.5.1\\\":\\n version \\\"0.5.1\\\"\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"6f0cf049caa1a7982669ee685e86621452686551\", \"983fef55ef08ca2ca25349bb2d5bdff10ecf89f4\", \"f7cc7b263afeb27eef393b7497db8dad8ebb0518\", \"bfe32bf10e9b6d699f694fbd095af0b3f2e6275f\"]"},"types":{"kind":"string","value":"[\"feat\", \"refactor\", \"test\", \"build\"]"}}},{"rowIdx":1031,"cells":{"commit_message":{"kind":"string","value":"Deploy utilities from correct folder\n\nSigned-off-by: rjshrjndrn ,wire up fixed null encoding,implement array flatten support,init environ cache"},"diff":{"kind":"string","value":"[\"diff --git a/.github/workflows/utilities.yaml b/.github/workflows/utilities.yaml\\nindex 92e130c..afbc850 100644\\n--- a/.github/workflows/utilities.yaml\\n+++ b/.github/workflows/utilities.yaml\\n@@ -43,7 +43,7 @@ jobs:\\n PUSH_IMAGE=1 bash build.sh\\n - name: Deploy to kubernetes\\n run: |\\n- cd scripts/helm/\\n+ cd scripts/helmcharts/\\n sed -i \\\"s#openReplayContainerRegistry.*#openReplayContainerRegistry: \\\\\\\"${{ secrets.OSS_REGISTRY_URL }}\\\\\\\"#g\\\" vars.yaml\\n sed -i \\\"s#minio_access_key.*#minio_access_key: \\\\\\\"${{ secrets.OSS_MINIO_ACCESS_KEY }}\\\\\\\" #g\\\" vars.yaml\\n sed -i \\\"s#minio_secret_key.*#minio_secret_key: \\\\\\\"${{ secrets.OSS_MINIO_SECRET_KEY }}\\\\\\\" #g\\\" vars.yaml\\n\", \"diff --git a/read_buffer/src/row_group.rs b/read_buffer/src/row_group.rs\\nindex 91c9fb5..ca77f3c 100644\\n--- a/read_buffer/src/row_group.rs\\n+++ b/read_buffer/src/row_group.rs\\n@@ -958,24 +958,15 @@ impl From for RowGroup {\\n }\\n Some(InfluxColumnType::Field(_)) => {\\n let column_data = match arrow_column.data_type() {\\n- arrow::datatypes::DataType::Int64 => Column::from(\\n- arrow_column\\n- .as_any()\\n- .downcast_ref::()\\n- .unwrap(),\\n- ),\\n- arrow::datatypes::DataType::Float64 => Column::from(\\n- arrow_column\\n- .as_any()\\n- .downcast_ref::()\\n- .unwrap(),\\n- ),\\n- arrow::datatypes::DataType::UInt64 => Column::from(\\n- arrow_column\\n- .as_any()\\n- .downcast_ref::()\\n- .unwrap(),\\n- ),\\n+ arrow::datatypes::DataType::Int64 => {\\n+ Column::from(arrow::array::Int64Array::from(arrow_column.data()))\\n+ }\\n+ arrow::datatypes::DataType::Float64 => {\\n+ Column::from(arrow::array::Float64Array::from(arrow_column.data()))\\n+ }\\n+ arrow::datatypes::DataType::UInt64 => {\\n+ Column::from(arrow::array::UInt64Array::from(arrow_column.data()))\\n+ }\\n dt => unimplemented!(\\n \\\"data type {:?} currently not supported for field columns\\\",\\n dt\\n\", \"diff --git a/ibis/backends/snowflake/registry.py b/ibis/backends/snowflake/registry.py\\nindex 2373dd7..4ce03b0 100644\\n--- a/ibis/backends/snowflake/registry.py\\n+++ b/ibis/backends/snowflake/registry.py\\n@@ -422,6 +422,7 @@ operation_registry.update(\\n ops.ArrayZip: _array_zip,\\n ops.ArraySort: unary(sa.func.array_sort),\\n ops.ArrayRepeat: fixed_arity(sa.func.ibis_udfs.public.array_repeat, 2),\\n+ ops.ArrayFlatten: fixed_arity(sa.func.array_flatten, 1),\\n ops.StringSplit: fixed_arity(sa.func.split, 2),\\n # snowflake typeof only accepts VARIANT, so we cast\\n ops.TypeOf: unary(lambda arg: sa.func.typeof(sa.func.to_variant(arg))),\\n\", \"diff --git a/src/environment.go b/src/environment.go\\nindex ae5e26a..0c961c5 100644\\n--- a/src/environment.go\\n+++ b/src/environment.go\\n@@ -229,6 +229,7 @@ func (env *environment) environ() map[string]string {\\n \\tif env.environCache != nil {\\n \\t\\treturn env.environCache\\n \\t}\\n+\\tenv.environCache = make(map[string]string)\\n \\tconst separator = \\\"=\\\"\\n \\tvalues := os.Environ()\\n \\tfor value := range values {\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"2ebf04099353ef70395b8c8f5e130f70e1ed0814\", \"28b596b8834d1b51be3ac6a2ac30df28f37702d8\", \"d3c754f09502be979e5dcc79f968b15052590bd0\", \"dc50bd35462a49058c91a939fc8830ae7a9eb692\"]"},"types":{"kind":"string","value":"[\"ci\", \"refactor\", \"feat\", \"fix\"]"}}},{"rowIdx":1032,"cells":{"commit_message":{"kind":"string","value":"Added tooltip for Data sources table buttons only on small screen,fix height calc,removing automatic page push on nav,fix readme"},"diff":{"kind":"string","value":"[\"diff --git a/packages/nc-gui/components/dashboard/settings/DataSources.vue b/packages/nc-gui/components/dashboard/settings/DataSources.vue\\nindex 78caa98..0ed5df9 100644\\n--- a/packages/nc-gui/components/dashboard/settings/DataSources.vue\\n+++ b/packages/nc-gui/components/dashboard/settings/DataSources.vue\\n@@ -351,59 +351,78 @@ const isEditBaseModalOpen = computed({\\n \\n
\\n
\\n- \\n-
\\n- \\n-
\\n- {{ $t('tooltip.metaSync') }}\\n+ \\n+ \\n+ \\n+
\\n+ \\n+
\\n+ {{ $t('tooltip.metaSync') }}\\n+
\\n
\\n-
\\n- \\n- \\n-
\\n- \\n-
\\n- {{ $t('title.relations') }}\\n+ \\n+ \\n+ \\n+ \\n+ \\n+
\\n+ \\n+
\\n+ {{ $t('title.relations') }}\\n+
\\n
\\n-
\\n- \\n- \\n-
\\n- \\n-
\\n- {{ $t('labels.uiAcl') }}\\n+ \\n+ \\n+ \\n+ \\n+ \\n+
\\n+ \\n+
\\n+ {{ $t('labels.uiAcl') }}\\n+
\\n
\\n-
\\n- \\n- \\n-
\\n- \\n-
\\n- {{ $t('title.audit') }}\\n+ \\n+ \\n+ \\n+ \\n+ \\n+
\\n+ \\n+
\\n+ {{ $t('title.audit') }}\\n+
\\n
\\n-
\\n- \\n+ \\n+ \\n
\\n
\\n
\\n@@ -450,67 +469,92 @@ const isEditBaseModalOpen = computed({\\n \\n
\\n
\\n- \\n-
\\n- \\n-
\\n- {{ $t('title.relations') }}\\n+ \\n+ \\n+ \\n+
\\n+ \\n+
\\n+ {{ $t('title.relations') }}\\n+
\\n
\\n-
\\n- \\n+ \\n+ \\n+ \\n+ \\n+ \\n+
\\n+ \\n+
\\n+ {{ $t('labels.uiAcl') }}\\n+
\\n+
\\n+ \\n+
\\n+ \\n+ \\n+ \\n+
\\n+ \\n+
\\n+ {{ $t('tooltip.metaSync') }}\\n+
\\n+
\\n+ \\n+
\\n+
\\n+
\\n+
\\n+ \\n+ \\n \\n-
\\n- \\n-
\\n- {{ $t('labels.uiAcl') }}\\n-
\\n-
\\n+ \\n \\n+
\\n+ \\n+ \\n \\n-
\\n- \\n-
\\n- {{ $t('tooltip.metaSync') }}\\n-
\\n-
\\n+ \\n \\n-
\\n-
\\n-
\\n- \\n- \\n- \\n- \\n- \\n- \\n+ \\n
\\n
\\n \\ndiff --git a/packages/nc-gui/components/nc/Tooltip.vue b/packages/nc-gui/components/nc/Tooltip.vue\\nindex 0810b8b..97b159e 100644\\n--- a/packages/nc-gui/components/nc/Tooltip.vue\\n+++ b/packages/nc-gui/components/nc/Tooltip.vue\\n@@ -12,6 +12,7 @@ interface Props {\\n disabled?: boolean\\n placement?: TooltipPlacement | undefined\\n hideOnClick?: boolean\\n+ overlayClassName?: string\\n }\\n \\n const props = defineProps()\\n@@ -36,6 +37,8 @@ const attrs = useAttrs()\\n \\n const isKeyPressed = ref(false)\\n \\n+const overlayClassName = computed(() => props.overlayClassName)\\n+\\n onKeyStroke(\\n (e) => e.key === modifierKey.value,\\n (e) => {\\n@@ -100,7 +103,7 @@ const onClick = () => {\\n \\n- \\n- \\n- \\n
\\n \\n Delete Selected Rows\\n \\n \\n- \\ndiff --git a/packages/nc-gui/components/nc/Tooltip.vue b/packages/nc-gui/components/nc/Tooltip.vue\\nindex 0810b8b..97b159e 100644\\n--- a/packages/nc-gui/components/nc/Tooltip.vue\\n+++ b/packages/nc-gui/components/nc/Tooltip.vue\\n@@ -12,6 +12,7 @@ interface Props {\\n disabled?: boolean\\n placement?: TooltipPlacement | undefined\\n hideOnClick?: boolean\\n+ overlayClassName?: string\\n }\\n \\n const props = defineProps()\\n@@ -36,6 +37,8 @@ const attrs = useAttrs()\\n \\n const isKeyPressed = ref(false)\\n \\n+const overlayClassName = computed(() => props.overlayClassName)\\n+\\n onKeyStroke(\\n (e) => e.key === modifierKey.value,\\n (e) => {\\n@@ -100,7 +103,7 @@ const onClick = () => {\\n \\ndiff --git a/packages/nc-gui/components/nc/Tooltip.vue b/packages/nc-gui/components/nc/Tooltip.vue\\nindex 0810b8b..97b159e 100644\\n--- a/packages/nc-gui/components/nc/Tooltip.vue\\n+++ b/packages/nc-gui/components/nc/Tooltip.vue\\n@@ -12,6 +12,7 @@ interface Props {\\n disabled?: boolean\\n placement?: TooltipPlacement | undefined\\n hideOnClick?: boolean\\n+ overlayClassName?: string\\n }\\n \\n const props = defineProps()\\n@@ -36,6 +37,8 @@ const attrs = useAttrs()\\n \\n const isKeyPressed = ref(false)\\n \\n+const overlayClassName = computed(() => props.overlayClassName)\\n+\\n onKeyStroke(\\n (e) => e.key === modifierKey.value,\\n (e) => {\\n@@ -100,7 +103,7 @@ const onClick = () => {\\n \\n
\\n \\n- \\n
\\n \\n \\n\", \"diff --git a/ci/cue/lint.cue b/ci/cue/lint.cue\\nindex cdda698..6aac265 100644\\n--- a/ci/cue/lint.cue\\n+++ b/ci/cue/lint.cue\\n@@ -39,7 +39,7 @@ import (\\n \\t\\t\\t// CACHE: copy only *.cue files\\n \\t\\t\\tdocker.#Copy & {\\n \\t\\t\\t\\tcontents: source\\n-\\t\\t\\t\\tinclude: [\\\"*.cue\\\"]\\n+\\t\\t\\t\\tinclude: [\\\"*.cue\\\", \\\"**/*.cue\\\"]\\n \\t\\t\\t\\tdest: \\\"/cue\\\"\\n \\t\\t\\t},\\n \\n\", \"diff --git a/docs/api/sandbox-option.md b/docs/api/sandbox-option.md\\nindex 7d24bee..e293d34 100644\\n--- a/docs/api/sandbox-option.md\\n+++ b/docs/api/sandbox-option.md\\n@@ -113,8 +113,8 @@ window.open = customWindowOpen\\n Important things to notice in the preload script:\\n \\n - Even though the sandboxed renderer doesn't have Node.js running, it still has\\n- access to a limited node-like environment: `Buffer`, `process`, `setImmediate`\\n- and `require` are available.\\n+ access to a limited node-like environment: `Buffer`, `process`, `setImmediate`,\\n+ `clearImmediate` and `require` are available.\\n - The preload script can indirectly access all APIs from the main process through the\\n `remote` and `ipcRenderer` modules.\\n - The preload script must be contained in a single script, but it is possible to have\\n@@ -162,16 +162,17 @@ feature. We are still not aware of the security implications of exposing some\\n Electron renderer APIs to the preload script, but here are some things to\\n consider before rendering untrusted content:\\n \\n-- A preload script can accidentally leak privileged APIs to untrusted code.\\n+- A preload script can accidentally leak privileged APIs to untrusted code,\\n+ unless [`contextIsolation`](../tutorial/security.md#3-enable-context-isolation-for-remote-content)\\n+ is also enabled.\\n - Some bug in V8 engine may allow malicious code to access the renderer preload\\n APIs, effectively granting full access to the system through the `remote`\\n- module.\\n+ module. Therefore, it is highly recommended to\\n+ [disable the `remote` module](../tutorial/security.md#15-disable-the-remote-module).\\n+ If disabling is not feasible, you should selectively\\n+ [filter the `remote` module](../tutorial/security.md#16-filter-the-remote-module).\\n \\n Since rendering untrusted content in Electron is still uncharted territory,\\n the APIs exposed to the sandbox preload script should be considered more\\n unstable than the rest of Electron APIs, and may have breaking changes to fix\\n security issues.\\n-\\n-One planned enhancement that should greatly increase security is to block IPC\\n-messages from sandboxed renderers by default, allowing the main process to\\n-explicitly define a set of messages the renderer is allowed to send.\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"90d4c1d5bcc9f2dce6e1da0cb953f04f46fb1380\", \"bf95d5d0b34d32ef2684488feb3de01cb824b2b4\", \"4c44543a3d9eea37e90a2316717feb01c0e0d83a\", \"dbb8617214aaa8b56b827deef1265d9ee38765bd\"]"},"types":{"kind":"string","value":"[\"test\", \"refactor\", \"ci\", \"docs\"]"}}},{"rowIdx":1067,"cells":{"commit_message":{"kind":"string","value":"do not check mkdocs for older versions used in deployments,add canonical `_name` to edge packages,rename ELECTRON_CACHE env variable to electron_config_cache (#21313),entries updates"},"diff":{"kind":"string","value":"[\"diff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\\nindex 57d94a4..04de03b 100644\\n--- a/.github/workflows/ibis-docs-lint.yml\\n+++ b/.github/workflows/ibis-docs-lint.yml\\n@@ -206,7 +206,7 @@ jobs:\\n - name: build and push dev docs\\n run: |\\n nix develop --ignore-environment -c \\\\\\n- mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}'\\n+ mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}' --ignore-version\\n \\n simulate_release:\\n runs-on: ubuntu-latest\\n\", \"diff --git a/scripts/bump-edge.ts b/scripts/bump-edge.ts\\nindex e92e3c9..0b7a11a 100644\\n--- a/scripts/bump-edge.ts\\n+++ b/scripts/bump-edge.ts\\n@@ -53,6 +53,7 @@ async function loadWorkspace (dir: string) {\\n }\\n \\n const rename = (from: string, to: string) => {\\n+ find(from).data._name = find(from).data.name\\n find(from).data.name = to\\n for (const pkg of packages) {\\n pkg.updateDeps((dep) => {\\n\", \"diff --git a/docs/tutorial/installation.md b/docs/tutorial/installation.md\\nindex d4af120..1a09eea 100644\\n--- a/docs/tutorial/installation.md\\n+++ b/docs/tutorial/installation.md\\n@@ -82,7 +82,7 @@ with the network at all.\\n On environments that have been using older versions of Electron, you might find the\\n cache also in `~/.electron`.\\n \\n-You can also override the local cache location by providing a `ELECTRON_CACHE`\\n+You can also override the local cache location by providing a `electron_config_cache`\\n environment variable.\\n \\n The cache contains the version's official zip file as well as a checksum, stored as\\n\", \"diff --git a/packages/docz-core/src/DataServer.ts b/packages/docz-core/src/DataServer.ts\\nindex 0dad341..d1d95fb 100644\\n--- a/packages/docz-core/src/DataServer.ts\\n+++ b/packages/docz-core/src/DataServer.ts\\n@@ -34,13 +34,13 @@ export class DataServer {\\n public async processEntries(): Promise {\\n const config = this.config\\n const entries = new Entries(config)\\n- const map = await entries.getMap()\\n const watcher = chokidar.watch(this.config.files, {\\n ignored: /(^|[\\\\/\\\\\\\\])\\\\../,\\n })\\n \\n- const handleConnection = (socket: WS) => {\\n- const update = this.updateEntries(socket)\\n+ const handleConnection = async (socket: WS) => {\\n+ const update = this.updateEntries(entries, socket)\\n+ const map = await entries.getMap()\\n \\n watcher.on('change', async () => update(this.config))\\n watcher.on('unlink', async () => update(this.config))\\n@@ -51,12 +51,14 @@ export class DataServer {\\n })\\n \\n socket.send(this.entriesData(map))\\n+ await Entries.writeImports(map)\\n }\\n \\n this.server.on('connection', handleConnection)\\n this.server.on('close', () => watcher.close())\\n \\n- await Entries.write(config, map)\\n+ await Entries.writeGenerated(config)\\n+ await Entries.writeImports(await entries.getMap())\\n }\\n \\n public async processThemeConfig(): Promise {\\n@@ -88,14 +90,16 @@ export class DataServer {\\n return this.dataObj('docz.config', config.themeConfig)\\n }\\n \\n- private updateEntries(socket: WS): (config: Config) => Promise {\\n+ private updateEntries(\\n+ entries: Entries,\\n+ socket: WS\\n+ ): (config: Config) => Promise {\\n return async config => {\\n if (isSocketOpened(socket)) {\\n- const newEntries = new Entries(config)\\n- const newMap = await newEntries.getMap()\\n+ const map = await entries.getMap()\\n \\n- await Entries.rewrite(newMap)\\n- socket.send(this.entriesData(newMap))\\n+ await Entries.writeImports(map)\\n+ socket.send(this.entriesData(map))\\n }\\n }\\n }\\ndiff --git a/packages/docz-core/src/Entries.ts b/packages/docz-core/src/Entries.ts\\nindex 76178eb..6e1a370 100644\\n--- a/packages/docz-core/src/Entries.ts\\n+++ b/packages/docz-core/src/Entries.ts\\n@@ -77,14 +77,13 @@ const writeImports = async (entries: EntryMap): Promise => {\\n export type EntryMap = Record\\n \\n export class Entries {\\n- public static async write(config: Config, entries: EntryMap): Promise {\\n+ public static async writeGenerated(config: Config): Promise {\\n mkd(paths.docz)\\n await writeGeneratedFiles(config)\\n- await writeImports(entries)\\n }\\n \\n- public static async rewrite(map: EntryMap): Promise {\\n- await writeImports(map)\\n+ public static async writeImports(entries: EntryMap): Promise {\\n+ await writeImports(entries)\\n }\\n \\n public all: EntryMap\\ndiff --git a/packages/docz-core/templates/app.tpl.js b/packages/docz-core/templates/app.tpl.js\\nindex 22ad59b..bbb9081 100644\\n--- a/packages/docz-core/templates/app.tpl.js\\n+++ b/packages/docz-core/templates/app.tpl.js\\n@@ -18,15 +18,6 @@ class App extends React.Component {\\n state = {\\n config: {},\\n entries: {},\\n- imports: {},\\n- }\\n-\\n- static getDerivedStateFromProps(nextProps, prevState) {\\n- return {\\n- config: prevState.config,\\n- entries: prevState.entries,\\n- imports: nextProps.imports\\n- }\\n }\\n \\n async componentDidMount() {\\n@@ -44,7 +35,8 @@ class App extends React.Component {\\n }\\n \\n render() {\\n- return \\n+ const { imports } = this.props\\n+ return \\n }\\n }\\n \\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"21228c55b7045d9b2225f65e6231184ff332b071\", \"573f87edf9bdc19c9c4c3a978fad6ed3ce788f5f\", \"f2f52c23b513dd857350f3c163f676d37189d0d3\", \"7147ac1f43a3ca454c79a6709dda2c35162ec88c\"]"},"types":{"kind":"string","value":"[\"ci\", \"build\", \"docs\", \"fix\"]"}}},{"rowIdx":1068,"cells":{"commit_message":{"kind":"string","value":"switch to callback ref,apply element migrated events\n\nThis is a very straightforward event applier. All it needs to do is\nupdate the persisted data for the element instance using the data in the\nevent.,update basic test with colors,updated webpack in react"},"diff":{"kind":"string","value":"[\"diff --git a/src/notebook/components/transforms/html.js b/src/notebook/components/transforms/html.js\\nindex 83fc1fb..021cc65 100644\\n--- a/src/notebook/components/transforms/html.js\\n+++ b/src/notebook/components/transforms/html.js\\n@@ -8,16 +8,16 @@ type Props = {\\n \\n export default class HTMLDisplay extends React.Component {\\n props: Props;\\n+ el: HTMLElement;\\n \\n componentDidMount(): void {\\n- if (this.refs.here) {\\n- if (document.createRange && Range && Range.prototype.createContextualFragment) {\\n- const range = document.createRange();\\n- const fragment = range.createContextualFragment(this.props.data);\\n- ReactDOM.findDOMNode(this.refs.here).appendChild(fragment);\\n- } else {\\n- ReactDOM.findDOMNode(this.refs.here).innerHTML = this.props.data;\\n- }\\n+ // Create a range to ensure that scripts are invoked from within the HTML\\n+ if (document.createRange && Range && Range.prototype.createContextualFragment) {\\n+ const range = document.createRange();\\n+ const fragment = range.createContextualFragment(this.props.data);\\n+ this.el.appendChild(fragment);\\n+ } else {\\n+ this.el.innerHTML = this.props.data;\\n }\\n }\\n \\n@@ -27,7 +27,7 @@ export default class HTMLDisplay extends React.Component {\\n \\n render(): ?React.Element {\\n return (\\n-
\\n+
{ this.el = el; }} />\\n );\\n }\\n }\\n\", \"diff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\nindex da05e13..9231df3 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/EventAppliers.java\\n@@ -154,6 +154,9 @@ public final class EventAppliers implements EventApplier {\\n register(\\n ProcessInstanceIntent.SEQUENCE_FLOW_TAKEN,\\n new ProcessInstanceSequenceFlowTakenApplier(elementInstanceState, processState));\\n+ register(\\n+ ProcessInstanceIntent.ELEMENT_MIGRATED,\\n+ new ProcessInstanceElementMigratedApplier(elementInstanceState));\\n }\\n \\n private void registerProcessInstanceCreationAppliers(final MutableProcessingState state) {\\ndiff --git a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\\nindex e5a0f3a..d38358f 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/state/appliers/ProcessInstanceElementMigratedApplier.java\\n@@ -24,5 +24,16 @@ final class ProcessInstanceElementMigratedApplier\\n }\\n \\n @Override\\n- public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {}\\n+ public void applyState(final long elementInstanceKey, final ProcessInstanceRecord value) {\\n+ elementInstanceState.updateInstance(\\n+ elementInstanceKey,\\n+ elementInstance ->\\n+ elementInstance\\n+ .getValue()\\n+ .setProcessDefinitionKey(value.getProcessDefinitionKey())\\n+ .setBpmnProcessId(value.getBpmnProcessId())\\n+ .setVersion(value.getVersion())\\n+ .setElementId(value.getElementId())\\n+ .setFlowScopeKey(value.getFlowScopeKey()));\\n+ }\\n }\\n\", \"diff --git a/core/src/components/label/test/basic/index.html b/core/src/components/label/test/basic/index.html\\nindex d0b566c..377e58c 100644\\n--- a/core/src/components/label/test/basic/index.html\\n+++ b/core/src/components/label/test/basic/index.html\\n@@ -19,12 +19,32 @@\\n \\n \\n \\n+
\\n+ Default\\n+\\n+ Secondary\\n+\\n+ Tertiary\\n+\\n+ Danger\\n+\\n+ Custom\\n+
\\n+\\n \\n \\n Default\\n \\n \\n \\n+ Tertiary\\n+ \\n+ \\n+ \\n+ Custom\\n+ \\n+ \\n+ \\n Wrap label this label just goes on and on and on\\n \\n \\n@@ -42,6 +62,12 @@\\n \\n \\n
\\n+\\n+ \\n \\n \\n \\n\", \"diff --git a/components/react/package.json b/components/react/package.json\\nindex bbeb9ee..43ddebc 100644\\n--- a/components/react/package.json\\n+++ b/components/react/package.json\\n@@ -114,7 +114,7 @@\\n \\\"ts-loader\\\": \\\"^9.2.9\\\",\\n \\\"ts-node\\\": \\\"^10.7.0\\\",\\n \\\"typescript\\\": \\\"^4.7.3\\\",\\n- \\\"webpack\\\": \\\"^5.72.0\\\",\\n+ \\\"webpack\\\": \\\"^5.73.0\\\",\\n \\\"webpack-bundle-analyzer\\\": \\\"^4.5.0\\\",\\n \\\"webpack-cli\\\": \\\"^4.9.2\\\",\\n \\\"webpack-node-externals\\\": \\\"^3.0.0\\\"\\ndiff --git a/yarn.lock b/yarn.lock\\nindex a3fdb26..19a0716 100644\\n--- a/yarn.lock\\n+++ b/yarn.lock\\n@@ -25212,7 +25212,7 @@ webpack@^4.38.0, webpack@^4.42.1:\\n watchpack \\\"^1.7.4\\\"\\n webpack-sources \\\"^1.4.1\\\"\\n \\n-webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0:\\n+webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0, webpack@^5.73.0:\\n version \\\"5.73.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/webpack/-/webpack-5.73.0.tgz#bbd17738f8a53ee5760ea2f59dce7f3431d35d38\\\"\\n integrity sha512-svjudQRPPa0YiOYa2lM/Gacw0r6PvxptHj4FuEKQ2kX05ZLkjbVc5MnPs6its5j7IZljnIqSVo/OsY2X0IpHGA==\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"ee4bf61fb8836e249fb4ef3507dc938e70696b3f\", \"39d5d1cfe8d2210305df2c8fab4a4ae430732cf7\", \"c3b5dc77ff3d89d389f6f3a868b17d0a8ca63074\", \"78c446cbea61af2268b4c4da03a9ad4283f10049\"]"},"types":{"kind":"string","value":"[\"refactor\", \"feat\", \"test\", \"build\"]"}}},{"rowIdx":1069,"cells":{"commit_message":{"kind":"string","value":"use a closure,offset tests for min and max read cursors,render-svg,fix monorepo.dir prop\n\nSigned-off-by: Carlos Alexandro Becker "},"diff":{"kind":"string","value":"[\"diff --git a/ibis/expr/analysis.py b/ibis/expr/analysis.py\\nindex bb17a7a..975c658 100644\\n--- a/ibis/expr/analysis.py\\n+++ b/ibis/expr/analysis.py\\n@@ -39,7 +39,9 @@ def sub_for(expr, substitutions):\\n An Ibis expression\\n \\\"\\\"\\\"\\n \\n- def fn(node, mapping={k.op(): v for k, v in substitutions}):\\n+ mapping = {k.op(): v for k, v in substitutions}\\n+\\n+ def fn(node):\\n try:\\n return mapping[node]\\n except KeyError:\\n\", \"diff --git a/storage/reads/array_cursor_test.go b/storage/reads/array_cursor_test.go\\nindex 7c7ad0c..c1e6ff9 100644\\n--- a/storage/reads/array_cursor_test.go\\n+++ b/storage/reads/array_cursor_test.go\\n@@ -1541,6 +1541,34 @@ func TestWindowMinArrayCursor(t *testing.T) {\\n \\t\\t\\t},\\n \\t\\t},\\n \\t\\t{\\n+\\t\\t\\tname: \\\"window offset\\\",\\n+\\t\\t\\tevery: time.Hour,\\n+\\t\\t\\toffset: 30 * time.Minute,\\n+\\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n+\\t\\t\\t\\tmakeIntegerArray(\\n+\\t\\t\\t\\t\\t16,\\n+\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:00:00Z\\\"), 15*time.Minute,\\n+\\t\\t\\t\\t\\tfunc(i int64) int64 {\\n+\\t\\t\\t\\t\\t\\tbase := (i / 4) * 100\\n+\\t\\t\\t\\t\\t\\tm := (i % 4) * 15\\n+\\t\\t\\t\\t\\t\\treturn base + m\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t),\\n+\\t\\t\\t},\\n+\\t\\t\\twant: []*cursors.IntegerArray{\\n+\\t\\t\\t\\t{\\n+\\t\\t\\t\\t\\tTimestamps: []int64{\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:00:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:30:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T01:30:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T02:30:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T03:30:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t\\tValues: []int64{0, 30, 130, 230, 330},\\n+\\t\\t\\t\\t},\\n+\\t\\t\\t},\\n+\\t\\t},\\n+\\t\\t{\\n \\t\\t\\tname: \\\"window desc values\\\",\\n \\t\\t\\tevery: time.Hour,\\n \\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n@@ -1560,6 +1588,34 @@ func TestWindowMinArrayCursor(t *testing.T) {\\n \\t\\t\\t},\\n \\t\\t},\\n \\t\\t{\\n+\\t\\t\\tname: \\\"window offset desc values\\\",\\n+\\t\\t\\tevery: time.Hour,\\n+\\t\\t\\toffset: 30 * time.Minute,\\n+\\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n+\\t\\t\\t\\tmakeIntegerArray(\\n+\\t\\t\\t\\t\\t16,\\n+\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:00:00Z\\\"), 15*time.Minute,\\n+\\t\\t\\t\\t\\tfunc(i int64) int64 {\\n+\\t\\t\\t\\t\\t\\tbase := (i / 4) * 100\\n+\\t\\t\\t\\t\\t\\tm := 60 - (i%4)*15\\n+\\t\\t\\t\\t\\t\\treturn base + m\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t),\\n+\\t\\t\\t},\\n+\\t\\t\\twant: []*cursors.IntegerArray{\\n+\\t\\t\\t\\t{\\n+\\t\\t\\t\\t\\tTimestamps: []int64{\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:15:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:45:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T01:45:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T02:45:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T03:45:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t\\tValues: []int64{45, 15, 115, 215, 315},\\n+\\t\\t\\t\\t},\\n+\\t\\t\\t},\\n+\\t\\t},\\n+\\t\\t{\\n \\t\\t\\tname: \\\"window min int\\\",\\n \\t\\t\\tevery: time.Hour,\\n \\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n@@ -1693,6 +1749,34 @@ func TestWindowMaxArrayCursor(t *testing.T) {\\n \\t\\t\\t},\\n \\t\\t},\\n \\t\\t{\\n+\\t\\t\\tname: \\\"window offset\\\",\\n+\\t\\t\\tevery: time.Hour,\\n+\\t\\t\\toffset: 30 * time.Minute,\\n+\\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n+\\t\\t\\t\\tmakeIntegerArray(\\n+\\t\\t\\t\\t\\t16,\\n+\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:00:00Z\\\"), 15*time.Minute,\\n+\\t\\t\\t\\t\\tfunc(i int64) int64 {\\n+\\t\\t\\t\\t\\t\\tbase := (i / 4) * 100\\n+\\t\\t\\t\\t\\t\\tm := (i % 4) * 15\\n+\\t\\t\\t\\t\\t\\treturn base + m\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t),\\n+\\t\\t\\t},\\n+\\t\\t\\twant: []*cursors.IntegerArray{\\n+\\t\\t\\t\\t{\\n+\\t\\t\\t\\t\\tTimestamps: []int64{\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:15:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T01:15:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T02:15:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T03:15:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T03:45:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t\\tValues: []int64{15, 115, 215, 315, 345},\\n+\\t\\t\\t\\t},\\n+\\t\\t\\t},\\n+\\t\\t},\\n+\\t\\t{\\n \\t\\t\\tname: \\\"window desc values\\\",\\n \\t\\t\\tevery: time.Hour,\\n \\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n@@ -1712,6 +1796,34 @@ func TestWindowMaxArrayCursor(t *testing.T) {\\n \\t\\t\\t},\\n \\t\\t},\\n \\t\\t{\\n+\\t\\t\\tname: \\\"window offset desc values\\\",\\n+\\t\\t\\tevery: time.Hour,\\n+\\t\\t\\toffset: 30 * time.Minute,\\n+\\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n+\\t\\t\\t\\tmakeIntegerArray(\\n+\\t\\t\\t\\t\\t16,\\n+\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:00:00Z\\\"), 15*time.Minute,\\n+\\t\\t\\t\\t\\tfunc(i int64) int64 {\\n+\\t\\t\\t\\t\\t\\tbase := (i / 4) * 100\\n+\\t\\t\\t\\t\\t\\tm := 60 - (i%4)*15\\n+\\t\\t\\t\\t\\t\\treturn base + m\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t),\\n+\\t\\t\\t},\\n+\\t\\t\\twant: []*cursors.IntegerArray{\\n+\\t\\t\\t\\t{\\n+\\t\\t\\t\\t\\tTimestamps: []int64{\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T00:00:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T01:00:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T02:00:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T03:00:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t\\tmustParseTime(\\\"2010-01-01T03:30:00Z\\\").UnixNano(),\\n+\\t\\t\\t\\t\\t},\\n+\\t\\t\\t\\t\\tValues: []int64{60, 160, 260, 360, 330},\\n+\\t\\t\\t\\t},\\n+\\t\\t\\t},\\n+\\t\\t},\\n+\\t\\t{\\n \\t\\t\\tname: \\\"window min int\\\",\\n \\t\\t\\tevery: time.Hour,\\n \\t\\t\\tinputArrays: []*cursors.IntegerArray{\\n\", \"diff --git a/package.json b/package.json\\nindex 3f8e5fa..cc4e398 100644\\n--- a/package.json\\n+++ b/package.json\\n@@ -42,6 +42,7 @@\\n \\\"rollup\\\": \\\"^2.34.2\\\",\\n \\\"rollup-plugin-copy\\\": \\\"^3.3.0\\\",\\n \\\"rollup-plugin-dts\\\": \\\"^2.0.0\\\",\\n+ \\\"rollup-plugin-terser\\\": \\\"^7.0.2\\\",\\n \\\"rollup-plugin-typescript2\\\": \\\"^0.29.0\\\",\\n \\\"ts-jest\\\": \\\"^26.4.4\\\",\\n \\\"tsup\\\": \\\"^3.10.1\\\",\\ndiff --git a/packages/renderer-svg/package.json b/packages/renderer-svg/package.json\\nindex fa9c049..6a0654c 100644\\n--- a/packages/renderer-svg/package.json\\n+++ b/packages/renderer-svg/package.json\\n@@ -1,16 +1,27 @@\\n {\\n- \\\"name\\\": \\\"shiki-renderer-svg\\\",\\n+ \\\"name\\\": \\\"@antfu/shiki-renderer-svg\\\",\\n \\\"version\\\": \\\"0.2.0\\\",\\n \\\"description\\\": \\\"SVG renderer for shiki\\\",\\n \\\"author\\\": \\\"Pine Wu \\\",\\n \\\"homepage\\\": \\\"https://github.com/octref/shiki/tree/master/packages/renderer-svg\\\",\\n \\\"license\\\": \\\"MIT\\\",\\n- \\\"main\\\": \\\"dist/index.js\\\",\\n- \\\"types\\\": \\\"dist/index.d.ts\\\",\\n \\\"repository\\\": {\\n \\\"type\\\": \\\"git\\\",\\n \\\"url\\\": \\\"git+https://github.com/shikijs/shiki.git\\\"\\n },\\n+ \\\"main\\\": \\\"dist/index.js\\\",\\n+ \\\"module\\\": \\\"dist/index.mjs\\\",\\n+ \\\"types\\\": \\\"dist/index.d.ts\\\",\\n+ \\\"unpkg\\\": \\\"dist/index.iife.min.js\\\",\\n+ \\\"jsdelivr\\\": \\\"dist/index.iife.min.js\\\",\\n+ \\\"files\\\": [\\n+ \\\"dist\\\"\\n+ ],\\n+ \\\"scripts\\\": {\\n+ \\\"prepublishOnly\\\": \\\"npm run build\\\",\\n+ \\\"build\\\": \\\"rollup -c\\\",\\n+ \\\"watch\\\": \\\"rollup -c -w\\\"\\n+ },\\n \\\"dependencies\\\": {\\n \\\"puppeteer\\\": \\\"^5.2.1\\\"\\n },\\ndiff --git a/packages/renderer-svg/rollup.config.js b/packages/renderer-svg/rollup.config.js\\nnew file mode 100644\\nindex 0000000..d4e45ce\\n--- /dev/null\\n+++ b/packages/renderer-svg/rollup.config.js\\n@@ -0,0 +1,67 @@\\n+import { nodeResolve } from '@rollup/plugin-node-resolve'\\n+import commonjs from '@rollup/plugin-commonjs'\\n+import dts from 'rollup-plugin-dts'\\n+import typescript from 'rollup-plugin-typescript2'\\n+import replace from '@rollup/plugin-replace'\\n+import { terser } from 'rollup-plugin-terser'\\n+\\n+const external = ['shiki', 'puppeteer']\\n+\\n+export default [\\n+ {\\n+ input: 'src/index.ts',\\n+ external,\\n+ output: [\\n+ {\\n+ file: 'dist/index.js',\\n+ format: 'cjs'\\n+ },\\n+ {\\n+ file: 'dist/index.mjs',\\n+ format: 'esm'\\n+ }\\n+ ],\\n+ plugins: [\\n+ replace({\\n+ __BROWSER__: JSON.stringify(false)\\n+ }),\\n+ typescript(),\\n+ nodeResolve(),\\n+ commonjs()\\n+ ]\\n+ },\\n+ {\\n+ input: 'src/index.ts',\\n+ output: [\\n+ {\\n+ file: 'dist/index.iife.js',\\n+ format: 'iife',\\n+ name: 'ShikiRenderSVG'\\n+ },\\n+ {\\n+ file: 'dist/index.iife.min.js',\\n+ format: 'iife',\\n+ name: 'ShikiRenderSVG',\\n+ plugins: [terser()]\\n+ }\\n+ ],\\n+ plugins: [\\n+ replace({\\n+ __BROWSER__: JSON.stringify(true)\\n+ }),\\n+ typescript(),\\n+ nodeResolve(),\\n+ commonjs()\\n+ ]\\n+ },\\n+ {\\n+ input: 'src/index.ts',\\n+ output: [\\n+ {\\n+ file: 'dist/index.d.ts',\\n+ format: 'es'\\n+ }\\n+ ],\\n+ plugins: [dts()]\\n+ }\\n+]\\ndiff --git a/packages/renderer-svg/src/global.d.ts b/packages/renderer-svg/src/global.d.ts\\nnew file mode 100644\\nindex 0000000..08c128f\\n--- /dev/null\\n+++ b/packages/renderer-svg/src/global.d.ts\\n@@ -0,0 +1 @@\\n+declare var __BROWSER__: boolean\\ndiff --git a/packages/renderer-svg/src/index.ts b/packages/renderer-svg/src/index.ts\\nindex ae77136..8f92312 100644\\n--- a/packages/renderer-svg/src/index.ts\\n+++ b/packages/renderer-svg/src/index.ts\\n@@ -1,4 +1,4 @@\\n-import { IThemedToken } from 'shiki'\\n+import type { IThemedToken } from 'shiki'\\n import { measureMonospaceTypeface } from './measureMonospaceTypeface'\\n \\n interface SVGRendererOptions {\\ndiff --git a/packages/renderer-svg/src/measureMonospaceTypeface.ts b/packages/renderer-svg/src/measureMonospaceTypeface.ts\\nindex e28a1ff..6ab834d 100644\\n--- a/packages/renderer-svg/src/measureMonospaceTypeface.ts\\n+++ b/packages/renderer-svg/src/measureMonospaceTypeface.ts\\n@@ -1,58 +1,61 @@\\n-import puppeteer from 'puppeteer'\\n+function measureFont(fontName: string, fontSize: number) {\\n+ /**\\n+ * Measure `M` for width\\n+ */\\n+ var c = document.createElement('canvas')\\n+ var ctx = c.getContext('2d')!\\n+ ctx.font = `${fontSize}px \\\"${fontName}\\\"`\\n \\n-export async function measureMonospaceTypeface(\\n- fontName: string,\\n- fontSize: number\\n-): Promise<{ width: number; height: number }> {\\n- const browser = await puppeteer.launch({ headless: true })\\n- const page = await browser.newPage()\\n- const measurement = await page.evaluate(measureFont, fontName, fontSize)\\n- await browser.close()\\n- return measurement\\n+ const capMMeasurement = ctx.measureText('M')\\n \\n- function measureFont(fontName: string, fontSize: number) {\\n- /**\\n- * Measure `M` for width\\n- */\\n- var c = document.createElement('canvas')\\n- var ctx = c.getContext('2d')!\\n- ctx.font = `${fontSize}px \\\"${fontName}\\\"`\\n-\\n- const capMMeasurement = ctx.measureText('M')\\n+ /**\\n+ * Measure A-Z, a-z for height\\n+ * A - 65\\n+ * Z - 90\\n+ * a - 97\\n+ * z - 122\\n+ */\\n+ const characters = []\\n+ for (let i = 65; i <= 90; i++) {\\n+ characters.push(String.fromCharCode(i))\\n+ }\\n+ for (let i = 97; i <= 122; i++) {\\n+ characters.push(String.fromCharCode(i))\\n+ }\\n \\n- /**\\n- * Measure A-Z, a-z for height\\n- * A - 65\\n- * Z - 90\\n- * a - 97\\n- * z - 122\\n- */\\n- const characters = []\\n- for (let i = 65; i <= 90; i++) {\\n- characters.push(String.fromCharCode(i))\\n+ let highC, lowC\\n+ let highestAscent = 0\\n+ let lowestDescent = 0\\n+ characters.forEach(c => {\\n+ const m = ctx.measureText(c)\\n+ if (m.actualBoundingBoxAscent > highestAscent) {\\n+ highestAscent = m.actualBoundingBoxAscent\\n+ highC = c\\n }\\n- for (let i = 97; i <= 122; i++) {\\n- characters.push(String.fromCharCode(i))\\n+ if (m.actualBoundingBoxDescent > lowestDescent) {\\n+ lowestDescent = m.actualBoundingBoxDescent\\n+ lowC = c\\n }\\n+ })\\n \\n- let highC, lowC\\n- let highestAscent = 0\\n- let lowestDescent = 0\\n- characters.forEach(c => {\\n- const m = ctx.measureText(c)\\n- if (m.actualBoundingBoxAscent > highestAscent) {\\n- highestAscent = m.actualBoundingBoxAscent\\n- highC = c\\n- }\\n- if (m.actualBoundingBoxDescent > lowestDescent) {\\n- lowestDescent = m.actualBoundingBoxDescent\\n- lowC = c\\n- }\\n- })\\n+ return {\\n+ width: capMMeasurement.width,\\n+ height: highestAscent + lowestDescent\\n+ }\\n+}\\n \\n- return {\\n- width: capMMeasurement.width,\\n- height: highestAscent + lowestDescent\\n- }\\n+export async function measureMonospaceTypeface(\\n+ fontName: string,\\n+ fontSize: number\\n+): Promise<{ width: number; height: number }> {\\n+ if (__BROWSER__) {\\n+ return measureFont(fontName, fontSize)\\n+ } else {\\n+ const puppeteer = await import('puppeteer')\\n+ const browser = await puppeteer.launch({ headless: true })\\n+ const page = await browser.newPage()\\n+ const measurement = await page.evaluate(measureFont, fontName, fontSize)\\n+ await browser.close()\\n+ return measurement\\n }\\n }\\ndiff --git a/packages/renderer-svg/tsconfig.json b/packages/renderer-svg/tsconfig.json\\nindex 3613212..bc50ce3 100644\\n--- a/packages/renderer-svg/tsconfig.json\\n+++ b/packages/renderer-svg/tsconfig.json\\n@@ -1,9 +1,10 @@\\n {\\n- \\\"extends\\\": \\\"../../tsconfig.json\\\",\\n \\\"compilerOptions\\\": {\\n- \\\"composite\\\": true,\\n- \\\"rootDir\\\": \\\"src\\\",\\n- \\\"outDir\\\": \\\"dist\\\",\\n- \\\"lib\\\": [\\\"dom\\\"]\\n+ \\\"module\\\": \\\"esnext\\\",\\n+ \\\"target\\\": \\\"es2017\\\",\\n+ \\\"esModuleInterop\\\": true,\\n+ \\\"moduleResolution\\\": \\\"node\\\",\\n+ \\\"lib\\\": [\\\"esnext\\\", \\\"DOM\\\"],\\n+ \\\"sourceMap\\\": true\\n }\\n }\\ndiff --git a/packages/shiki/rollup.config.js b/packages/shiki/rollup.config.js\\nindex b8ba9e3..9078ea2 100644\\n--- a/packages/shiki/rollup.config.js\\n+++ b/packages/shiki/rollup.config.js\\n@@ -4,6 +4,7 @@ import dts from 'rollup-plugin-dts'\\n import typescript from 'rollup-plugin-typescript2'\\n import copy from 'rollup-plugin-copy'\\n import replace from '@rollup/plugin-replace'\\n+import { terser } from 'rollup-plugin-terser'\\n import { version } from './package.json'\\n \\n const external = ['onigasm', 'vscode-textmate']\\n@@ -22,7 +23,14 @@ export default [\\n format: 'esm'\\n }\\n ],\\n- plugins: [typescript(), nodeResolve(), commonjs()]\\n+ plugins: [\\n+ replace({\\n+ __BROWSER__: JSON.stringify(false)\\n+ }),\\n+ typescript(),\\n+ nodeResolve(),\\n+ commonjs()\\n+ ]\\n },\\n {\\n input: 'src/index.ts',\\n@@ -58,7 +66,15 @@ export default [\\n ]\\n }\\n ],\\n- plugins: [typescript(), nodeResolve(), commonjs()]\\n+ plugins: [\\n+ replace({\\n+ __BROWSER__: JSON.stringify(true)\\n+ }),\\n+ typescript(),\\n+ nodeResolve(),\\n+ commonjs(),\\n+ terser()\\n+ ]\\n },\\n {\\n input: 'src/index.ts',\\ndiff --git a/packages/shiki/src/global.d.ts b/packages/shiki/src/global.d.ts\\nnew file mode 100644\\nindex 0000000..08c128f\\n--- /dev/null\\n+++ b/packages/shiki/src/global.d.ts\\n@@ -0,0 +1 @@\\n+declare var __BROWSER__: boolean\\ndiff --git a/packages/shiki/src/loader.ts b/packages/shiki/src/loader.ts\\nindex 934cfbd..d9c3128 100644\\n--- a/packages/shiki/src/loader.ts\\n+++ b/packages/shiki/src/loader.ts\\n@@ -5,11 +5,16 @@ import type { ILanguageRegistration, IShikiTheme } from './types'\\n export const isBrowser = typeof window !== 'undefined' && typeof window.document !== 'undefined'\\n \\n let CDN_ROOT = '__CDN_ROOT__'\\n+let ONIGASM_WASM = ''\\n \\n export function setCDN(root: string) {\\n CDN_ROOT = root\\n }\\n \\n+export function setOnigasmWASM(path: string) {\\n+ ONIGASM_WASM = path\\n+}\\n+\\n let _onigasmPromise: Promise = null\\n \\n export async function getOnigasm(): Promise {\\n@@ -17,7 +22,7 @@ export async function getOnigasm(): Promise {\\n let loader: Promise\\n \\n if (isBrowser) {\\n- loader = Onigasm.loadWASM(_resolvePath('onigasm.wasm', 'dist/'))\\n+ loader = Onigasm.loadWASM(ONIGASM_WASM || _resolvePath('onigasm.wasm', 'dist/'))\\n } else {\\n const path = require('path')\\n const onigasmPath = path.join(require.resolve('onigasm'), '../onigasm.wasm')\\ndiff --git a/yarn.lock b/yarn.lock\\nindex c143969..dfd7540 100644\\n--- a/yarn.lock\\n+++ b/yarn.lock\\n@@ -2487,6 +2487,11 @@ combined-stream@^1.0.6, combined-stream@~1.0.6:\\n dependencies:\\n delayed-stream \\\"~1.0.0\\\"\\n \\n+commander@^2.20.0, commander@~2.20.3:\\n+ version \\\"2.20.3\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/commander/-/commander-2.20.3.tgz#fd485e84c03eb4881c20722ba48035e8531aeb33\\\"\\n+ integrity sha512-GpVkmM8vF2vQUkj2LvZmD35JxeJOLCwJ9cUkugyk2nuhbv3+mJvpLYYt+0+USMxE+oj+ey/lJEnhZw75x/OMcQ==\\n+\\n commander@^4.0.0:\\n version \\\"4.1.1\\\"\\n resolved \\\"https://registry.yarnpkg.com/commander/-/commander-4.1.1.tgz#9fd602bd936294e9e9ef46a3f4d6964044b18068\\\"\\n@@ -2497,11 +2502,6 @@ commander@^6.2.0:\\n resolved \\\"https://registry.yarnpkg.com/commander/-/commander-6.2.0.tgz#b990bfb8ac030aedc6d11bc04d1488ffef56db75\\\"\\n integrity sha512-zP4jEKbe8SHzKJYQmq8Y9gYjtO/POJLgIdKgV7B9qNmABVFVc+ctqSX6iXh4mCpJfRBOabiZ2YKPg8ciDw6C+Q==\\n \\n-commander@~2.20.3:\\n- version \\\"2.20.3\\\"\\n- resolved \\\"https://registry.yarnpkg.com/commander/-/commander-2.20.3.tgz#fd485e84c03eb4881c20722ba48035e8531aeb33\\\"\\n- integrity sha512-GpVkmM8vF2vQUkj2LvZmD35JxeJOLCwJ9cUkugyk2nuhbv3+mJvpLYYt+0+USMxE+oj+ey/lJEnhZw75x/OMcQ==\\n-\\n commondir@^1.0.1:\\n version \\\"1.0.1\\\"\\n resolved \\\"https://registry.yarnpkg.com/commondir/-/commondir-1.0.1.tgz#ddd800da0c66127393cca5950ea968a3aaf1253b\\\"\\n@@ -4799,7 +4799,7 @@ jest-watcher@^26.6.2:\\n jest-util \\\"^26.6.2\\\"\\n string-length \\\"^4.0.1\\\"\\n \\n-jest-worker@^26.6.2:\\n+jest-worker@^26.2.1, jest-worker@^26.6.2:\\n version \\\"26.6.2\\\"\\n resolved \\\"https://registry.yarnpkg.com/jest-worker/-/jest-worker-26.6.2.tgz#7f72cbc4d643c365e27b9fd775f9d0eaa9c7a8ed\\\"\\n integrity sha512-KWYVV1c4i+jbMpaBC+U++4Va0cp8OisU185o73T1vo99hqi7w8tSJfUXYswwqqrjzwxa6KpRK54WhPvwf5w6PQ==\\n@@ -6444,6 +6444,13 @@ quick-lru@^4.0.1:\\n resolved \\\"https://registry.yarnpkg.com/quick-lru/-/quick-lru-4.0.1.tgz#5b8878f113a58217848c6482026c73e1ba57727f\\\"\\n integrity sha512-ARhCpm70fzdcvNQfPoy49IaanKkTlRWF2JMzqhcJbhSFRZv7nPTvZJdcY7301IPmvW+/p0RgIWnQDLJxifsQ7g==\\n \\n+randombytes@^2.1.0:\\n+ version \\\"2.1.0\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/randombytes/-/randombytes-2.1.0.tgz#df6f84372f0270dc65cdf6291349ab7a473d4f2a\\\"\\n+ integrity sha512-vYl3iOX+4CKUWuxGi9Ukhie6fsqXqS9FE2Zaic4tNFD2N2QQaXOMFbuKK4QmDHC0JO6B1Zp41J0LpT0oR68amQ==\\n+ dependencies:\\n+ safe-buffer \\\"^5.1.0\\\"\\n+\\n react-is@^17.0.1:\\n version \\\"17.0.1\\\"\\n resolved \\\"https://registry.yarnpkg.com/react-is/-/react-is-17.0.1.tgz#5b3531bd76a645a4c9fb6e693ed36419e3301339\\\"\\n@@ -6812,6 +6819,16 @@ rollup-plugin-dts@^2.0.0:\\n optionalDependencies:\\n \\\"@babel/code-frame\\\" \\\"^7.10.4\\\"\\n \\n+rollup-plugin-terser@^7.0.2:\\n+ version \\\"7.0.2\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/rollup-plugin-terser/-/rollup-plugin-terser-7.0.2.tgz#e8fbba4869981b2dc35ae7e8a502d5c6c04d324d\\\"\\n+ integrity sha512-w3iIaU4OxcF52UUXiZNsNeuXIMDvFrr+ZXK6bFZ0Q60qyVfq4uLptoS4bbq3paG3x216eQllFZX7zt6TIImguQ==\\n+ dependencies:\\n+ \\\"@babel/code-frame\\\" \\\"^7.10.4\\\"\\n+ jest-worker \\\"^26.2.1\\\"\\n+ serialize-javascript \\\"^4.0.0\\\"\\n+ terser \\\"^5.0.0\\\"\\n+\\n rollup-plugin-typescript2@^0.29.0:\\n version \\\"0.29.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/rollup-plugin-typescript2/-/rollup-plugin-typescript2-0.29.0.tgz#b7ad83f5241dbc5bdf1e98d9c3fca005ffe39e1a\\\"\\n@@ -6873,7 +6890,7 @@ safe-buffer@^5.0.1, safe-buffer@^5.1.1, safe-buffer@^5.1.2, safe-buffer@~5.1.0, \\n resolved \\\"https://registry.yarnpkg.com/safe-buffer/-/safe-buffer-5.1.2.tgz#991ec69d296e0313747d59bdfd2b745c35f8828d\\\"\\n integrity sha512-Gd2UZBJDkXlY7GbJxfsE8/nvKkUEU1G38c1siN6QP6a9PT9MmHB8GnpscSmMJSoF8LOIrt8ud/wPtojys4G6+g==\\n \\n-safe-buffer@^5.2.0, safe-buffer@~5.2.0:\\n+safe-buffer@^5.1.0, safe-buffer@^5.2.0, safe-buffer@~5.2.0:\\n version \\\"5.2.1\\\"\\n resolved \\\"https://registry.yarnpkg.com/safe-buffer/-/safe-buffer-5.2.1.tgz#1eaf9fa9bdb1fdd4ec75f58f9cdb4e6b7827eec6\\\"\\n integrity sha512-rp3So07KcdmmKbGvgaNxQSJr7bGVSVk5S9Eq1F+ppbRo70+YeaDxkw5Dd8NPN+GD6bjnYm2VuPuCXmpuYvmCXQ==\\n@@ -6937,6 +6954,13 @@ semver@^6.0.0, semver@^6.2.0, semver@^6.3.0:\\n resolved \\\"https://registry.yarnpkg.com/semver/-/semver-6.3.0.tgz#ee0a64c8af5e8ceea67687b133761e1becbd1d3d\\\"\\n integrity sha512-b39TBaTSfV6yBrapU89p5fKekE2m/NwnDocOVruQFS1/veMgdzuPcnOM34M6CwxW8jH/lxEa5rBoDeUwu5HHTw==\\n \\n+serialize-javascript@^4.0.0:\\n+ version \\\"4.0.0\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/serialize-javascript/-/serialize-javascript-4.0.0.tgz#b525e1238489a5ecfc42afacc3fe99e666f4b1aa\\\"\\n+ integrity sha512-GaNA54380uFefWghODBWEGisLZFj00nS5ACs6yHa9nLqlLpVLO8ChDGeKRjZnV4Nh4n0Qi7nhYZD/9fCPzEqkw==\\n+ dependencies:\\n+ randombytes \\\"^2.1.0\\\"\\n+\\n set-blocking@^2.0.0, set-blocking@~2.0.0:\\n version \\\"2.0.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/set-blocking/-/set-blocking-2.0.0.tgz#045f9782d011ae9a6803ddd382b24392b3d890f7\\\"\\n@@ -7140,7 +7164,7 @@ source-map-resolve@^0.5.0:\\n source-map-url \\\"^0.4.0\\\"\\n urix \\\"^0.1.0\\\"\\n \\n-source-map-support@^0.5.6:\\n+source-map-support@^0.5.6, source-map-support@~0.5.19:\\n version \\\"0.5.19\\\"\\n resolved \\\"https://registry.yarnpkg.com/source-map-support/-/source-map-support-0.5.19.tgz#a98b62f86dcaf4f67399648c085291ab9e8fed61\\\"\\n integrity sha512-Wonm7zOCIJzBGQdB+thsPar0kYuCIzYvxZwlBa87yi/Mdjv7Tip2cyVbLj5o0cFPN4EVkuTwb3GDDyUx2DGnGw==\\n@@ -7163,7 +7187,7 @@ source-map@^0.6.0, source-map@^0.6.1, source-map@~0.6.1:\\n resolved \\\"https://registry.yarnpkg.com/source-map/-/source-map-0.6.1.tgz#74722af32e9614e9c287a8d0bbde48b5e2f1a263\\\"\\n integrity sha512-UjgapumWlbMhkBgzT7Ykc5YXUT46F0iKu8SGXq0bcwP5dz/h0Plj6enJqjz1Zbq2l5WaqYnrVbwWOWMyF3F47g==\\n \\n-source-map@^0.7.3:\\n+source-map@^0.7.3, source-map@~0.7.2:\\n version \\\"0.7.3\\\"\\n resolved \\\"https://registry.yarnpkg.com/source-map/-/source-map-0.7.3.tgz#5302f8169031735226544092e64981f751750383\\\"\\n integrity sha512-CkCj6giN3S+n9qrYiBTX5gystlENnRW5jZeNLHpe6aue+SrHcG5VYwujhW9s4dY31mEGsxBDrHR6oI69fTXsaQ==\\n@@ -7552,6 +7576,15 @@ terminal-link@^2.0.0:\\n ansi-escapes \\\"^4.2.1\\\"\\n supports-hyperlinks \\\"^2.0.0\\\"\\n \\n+terser@^5.0.0:\\n+ version \\\"5.5.1\\\"\\n+ resolved \\\"https://registry.yarnpkg.com/terser/-/terser-5.5.1.tgz#540caa25139d6f496fdea056e414284886fb2289\\\"\\n+ integrity sha512-6VGWZNVP2KTUcltUQJ25TtNjx/XgdDsBDKGt8nN0MpydU36LmbPPcMBd2kmtZNNGVVDLg44k7GKeHHj+4zPIBQ==\\n+ dependencies:\\n+ commander \\\"^2.20.0\\\"\\n+ source-map \\\"~0.7.2\\\"\\n+ source-map-support \\\"~0.5.19\\\"\\n+\\n test-exclude@^6.0.0:\\n version \\\"6.0.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/test-exclude/-/test-exclude-6.0.0.tgz#04a8698661d805ea6fa293b6cb9e63ac044ef15e\\\"\\n\", \"diff --git a/www/docs/customization/monorepo.md b/www/docs/customization/monorepo.md\\nindex 6d0e857..e45490f 100644\\n--- a/www/docs/customization/monorepo.md\\n+++ b/www/docs/customization/monorepo.md\\n@@ -18,7 +18,7 @@ project_name: subproj1\\n \\n monorepo:\\n tag_prefix: subproject1/\\n- folder: subproj1\\n+ dir: subproj1\\n ```\\n \\n Then, you can release with (from the project's root directory):\\n@@ -30,11 +30,11 @@ goreleaser release --rm-dist -f ./subproj1/.goreleaser.yml\\n Then, the following is different from a \\\"regular\\\" run:\\n \\n - GoReleaser will then look if current commit has a tag prefixed with `subproject1`, and also the previous tag with the same prefix;\\n-- Changelog will include only commits that contain changes to files within the `subproj1` folder;\\n+- Changelog will include only commits that contain changes to files within the `subproj1` directory;\\n - Release name gets prefixed with `{{ .ProjectName }} ` if empty;\\n-- All build's `dir` setting get set to `monorepo.folder` if empty;\\n+- All build's `dir` setting get set to `monorepo.dir` if empty;\\n - if yours is not, you might want to change that manually;\\n-- Extra files on the release, archives, Docker builds, etc are prefixed with `monorepo.folder`;\\n+- Extra files on the release, archives, Docker builds, etc are prefixed with `monorepo.dir`;\\n - On templates, `{{.PrefixedTag}}` will be `monorepo.prefix/tag` (aka the actual tag name), and `{{.Tag}}` has the prefix stripped;\\n \\n The rest of the release process should work as usual.\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"ad52e1d67fd77f0b6a73fbf989b33f9abf395ecc\", \"b7e2330fa3a8d7b8a9bff01b707c44e64b845c7b\", \"ace6b981c8098a68092d4a10e75daae7b8bfee9b\", \"9ed3c0c4a72af977fc9150512fb6538f20a94b22\"]"},"types":{"kind":"string","value":"[\"refactor\", \"test\", \"feat\", \"docs\"]"}}},{"rowIdx":1070,"cells":{"commit_message":{"kind":"string","value":"add travis file,update version (nightly.0),skip if related view/hook/column of a filter is not found\n\nSigned-off-by: Pranav C ,dedup redundant imports"},"diff":{"kind":"string","value":"[\"diff --git a/.travis.yml b/.travis.yml\\nnew file mode 100644\\nindex 0000000..c08cc34\\n--- /dev/null\\n+++ b/.travis.yml\\n@@ -0,0 +1,11 @@\\n+sudo: false\\n+\\n+language: java\\n+jdk: oraclejdk8\\n+\\n+branches:\\n+ only:\\n+ - master\\n+\\n+notifications:\\n+ email: false\\n\", \"diff --git a/Cargo.lock b/Cargo.lock\\nindex e6f659c..cf93556 100644\\n--- a/Cargo.lock\\n+++ b/Cargo.lock\\n@@ -94,7 +94,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"els\\\"\\n-version = \\\"0.1.23\\\"\\n+version = \\\"0.1.24-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_compiler\\\",\\n@@ -105,7 +105,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"els\\\",\\n \\\"erg_common\\\",\\n@@ -115,7 +115,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_common\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"backtrace-on-stack-overflow\\\",\\n \\\"crossterm\\\",\\n@@ -126,7 +126,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_compiler\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_parser\\\",\\n@@ -134,7 +134,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_parser\\\"\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"unicode-xid\\\",\\ndiff --git a/Cargo.toml b/Cargo.toml\\nindex c58299b..6e51b6e 100644\\n--- a/Cargo.toml\\n+++ b/Cargo.toml\\n@@ -20,7 +20,7 @@ members = [\\n ]\\n \\n [workspace.package]\\n-version = \\\"0.6.11\\\"\\n+version = \\\"0.6.12-nightly.0\\\"\\n authors = [\\\"erg-lang team \\\"]\\n license = \\\"MIT OR Apache-2.0\\\"\\n edition = \\\"2021\\\"\\n@@ -64,10 +64,10 @@ full-repl = [\\\"erg_common/full-repl\\\"]\\n full = [\\\"els\\\", \\\"full-repl\\\", \\\"unicode\\\", \\\"pretty\\\"]\\n \\n [workspace.dependencies]\\n-erg_common = { version = \\\"0.6.11\\\", path = \\\"./crates/erg_common\\\" }\\n-erg_parser = { version = \\\"0.6.11\\\", path = \\\"./crates/erg_parser\\\" }\\n-erg_compiler = { version = \\\"0.6.11\\\", path = \\\"./crates/erg_compiler\\\" }\\n-els = { version = \\\"0.1.23\\\", path = \\\"./crates/els\\\" }\\n+erg_common = { version = \\\"0.6.12-nightly.0\\\", path = \\\"./crates/erg_common\\\" }\\n+erg_parser = { version = \\\"0.6.12-nightly.0\\\", path = \\\"./crates/erg_parser\\\" }\\n+erg_compiler = { version = \\\"0.6.12-nightly.0\\\", path = \\\"./crates/erg_compiler\\\" }\\n+els = { version = \\\"0.1.24-nightly.0\\\", path = \\\"./crates/els\\\" }\\n \\n [dependencies]\\n erg_common = { workspace = true }\\ndiff --git a/crates/els/Cargo.toml b/crates/els/Cargo.toml\\nindex 5f005a1..e1a9964 100644\\n--- a/crates/els/Cargo.toml\\n+++ b/crates/els/Cargo.toml\\n@@ -2,7 +2,7 @@\\n name = \\\"els\\\"\\n description = \\\"An Erg compiler frontend for IDEs, implements LSP.\\\"\\n documentation = \\\"http://docs.rs/els\\\"\\n-version = \\\"0.1.23\\\"\\n+version = \\\"0.1.24-nightly.0\\\"\\n authors.workspace = true\\n license.workspace = true\\n edition.workspace = true\\n\", \"diff --git a/packages/nocodb/src/lib/version-upgrader/ncFilterUpgrader.ts b/packages/nocodb/src/lib/version-upgrader/ncFilterUpgrader.ts\\nindex 1515f88..6c250bd 100644\\n--- a/packages/nocodb/src/lib/version-upgrader/ncFilterUpgrader.ts\\n+++ b/packages/nocodb/src/lib/version-upgrader/ncFilterUpgrader.ts\\n@@ -21,7 +21,13 @@ export default async function ({ ncMeta }: NcUpgraderCtx) {\\n } else {\\n continue;\\n }\\n- if (filter.project_id != model.project_id) {\\n+\\n+ // skip if related model is not found\\n+ if (!model) {\\n+ continue;\\n+ }\\n+\\n+ if (filter.project_id !== model.project_id) {\\n await ncMeta.metaUpdate(\\n null,\\n null,\\n\", \"diff --git a/ibis/backends/base/__init__.py b/ibis/backends/base/__init__.py\\nindex effd44c..a59c0ec 100644\\n--- a/ibis/backends/base/__init__.py\\n+++ b/ibis/backends/base/__init__.py\\n@@ -31,7 +31,7 @@ import ibis.common.exceptions as exc\\n import ibis.config\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n \\n __all__ = ('BaseBackend', 'Database', 'connect')\\n \\ndiff --git a/ibis/backends/base/sql/__init__.py b/ibis/backends/base/sql/__init__.py\\nindex e4f2129..7bbdaf9 100644\\n--- a/ibis/backends/base/sql/__init__.py\\n+++ b/ibis/backends/base/sql/__init__.py\\n@@ -12,7 +12,7 @@ import ibis.expr.analysis as an\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base import BaseBackend\\n from ibis.backends.base.sql.compiler import Compiler\\n \\ndiff --git a/ibis/backends/base/sql/alchemy/__init__.py b/ibis/backends/base/sql/alchemy/__init__.py\\nindex 71cc0e8..ab89d7d 100644\\n--- a/ibis/backends/base/sql/alchemy/__init__.py\\n+++ b/ibis/backends/base/sql/alchemy/__init__.py\\n@@ -11,7 +11,7 @@ import ibis\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql import BaseSQLBackend\\n from ibis.backends.base.sql.alchemy.database import AlchemyDatabase, AlchemyTable\\n from ibis.backends.base.sql.alchemy.datatypes import (\\ndiff --git a/ibis/backends/base/sql/alchemy/query_builder.py b/ibis/backends/base/sql/alchemy/query_builder.py\\nindex 54c74ba..0ec432f 100644\\n--- a/ibis/backends/base/sql/alchemy/query_builder.py\\n+++ b/ibis/backends/base/sql/alchemy/query_builder.py\\n@@ -3,7 +3,7 @@ from __future__ import annotations\\n import functools\\n \\n import sqlalchemy as sa\\n-import sqlalchemy.sql as sql\\n+from sqlalchemy import sql\\n \\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\ndiff --git a/ibis/backends/base/sql/compiler/base.py b/ibis/backends/base/sql/compiler/base.py\\nindex 84102aa..fb44667 100644\\n--- a/ibis/backends/base/sql/compiler/base.py\\n+++ b/ibis/backends/base/sql/compiler/base.py\\n@@ -7,7 +7,7 @@ import toolz\\n \\n import ibis.expr.analysis as an\\n import ibis.expr.operations as ops\\n-import ibis.util as util\\n+from ibis import util\\n \\n \\n class DML(abc.ABC):\\ndiff --git a/ibis/backends/base/sql/compiler/query_builder.py b/ibis/backends/base/sql/compiler/query_builder.py\\nindex a2d5214..95f5e8d 100644\\n--- a/ibis/backends/base/sql/compiler/query_builder.py\\n+++ b/ibis/backends/base/sql/compiler/query_builder.py\\n@@ -8,7 +8,7 @@ import toolz\\n import ibis.common.exceptions as com\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.compiler.base import DML, QueryAST, SetOp\\n from ibis.backends.base.sql.compiler.select_builder import SelectBuilder, _LimitSpec\\n from ibis.backends.base.sql.compiler.translator import ExprTranslator, QueryContext\\ndiff --git a/ibis/backends/base/sql/registry/main.py b/ibis/backends/base/sql/registry/main.py\\nindex 77f70a5..586ace5 100644\\n--- a/ibis/backends/base/sql/registry/main.py\\n+++ b/ibis/backends/base/sql/registry/main.py\\n@@ -4,7 +4,7 @@ import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.registry import (\\n aggregate,\\n binary_infix,\\ndiff --git a/ibis/backends/base/sql/registry/timestamp.py b/ibis/backends/base/sql/registry/timestamp.py\\nindex 412eab1..3c8571f 100644\\n--- a/ibis/backends/base/sql/registry/timestamp.py\\n+++ b/ibis/backends/base/sql/registry/timestamp.py\\n@@ -3,7 +3,7 @@ from __future__ import annotations\\n import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n-import ibis.util as util\\n+from ibis import util\\n \\n \\n def extract_field(sql_attr):\\ndiff --git a/ibis/backends/clickhouse/tests/test_client.py b/ibis/backends/clickhouse/tests/test_client.py\\nindex 8db6672..bb1b9ba 100644\\n--- a/ibis/backends/clickhouse/tests/test_client.py\\n+++ b/ibis/backends/clickhouse/tests/test_client.py\\n@@ -3,9 +3,9 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.config as config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n+from ibis import config\\n from ibis.backends.clickhouse.tests.conftest import (\\n CLICKHOUSE_HOST,\\n CLICKHOUSE_PASS,\\ndiff --git a/ibis/backends/conftest.py b/ibis/backends/conftest.py\\nindex 3a974da..ba7ad75 100644\\n--- a/ibis/backends/conftest.py\\n+++ b/ibis/backends/conftest.py\\n@@ -20,7 +20,7 @@ if TYPE_CHECKING:\\n import pytest\\n \\n import ibis\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base import _get_backend_names\\n \\n TEST_TABLES = {\\ndiff --git a/ibis/backends/dask/execution/util.py b/ibis/backends/dask/execution/util.py\\nindex 61bff7e..7ed0c10 100644\\n--- a/ibis/backends/dask/execution/util.py\\n+++ b/ibis/backends/dask/execution/util.py\\n@@ -9,13 +9,13 @@ import pandas as pd\\n from dask.dataframe.groupby import SeriesGroupBy\\n \\n import ibis.backends.pandas.execution.util as pd_util\\n-import ibis.common.graph as graph\\n import ibis.expr.analysis as an\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n import ibis.util\\n from ibis.backends.dask.core import execute\\n from ibis.backends.pandas.trace import TraceTwoLevelDispatcher\\n+from ibis.common import graph\\n from ibis.expr.scope import Scope\\n \\n if TYPE_CHECKING:\\ndiff --git a/ibis/backends/duckdb/datatypes.py b/ibis/backends/duckdb/datatypes.py\\nindex fd6b8f5..52c0719 100644\\n--- a/ibis/backends/duckdb/datatypes.py\\n+++ b/ibis/backends/duckdb/datatypes.py\\n@@ -3,7 +3,7 @@ from __future__ import annotations\\n import parsy as p\\n import toolz\\n \\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.parsing import (\\n COMMA,\\n FIELD,\\ndiff --git a/ibis/backends/impala/__init__.py b/ibis/backends/impala/__init__.py\\nindex 4ad2057..8299a28 100644\\n--- a/ibis/backends/impala/__init__.py\\n+++ b/ibis/backends/impala/__init__.py\\n@@ -20,7 +20,7 @@ import ibis.config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.rules as rlz\\n import ibis.expr.schema as sch\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql import BaseSQLBackend\\n from ibis.backends.base.sql.ddl import (\\n CTAS,\\ndiff --git a/ibis/backends/impala/client.py b/ibis/backends/impala/client.py\\nindex 6655ce7..78d526f 100644\\n--- a/ibis/backends/impala/client.py\\n+++ b/ibis/backends/impala/client.py\\n@@ -10,7 +10,7 @@ import sqlalchemy as sa\\n import ibis.common.exceptions as com\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base import Database\\n from ibis.backends.base.sql.compiler import DDL, DML\\n from ibis.backends.base.sql.ddl import (\\ndiff --git a/ibis/backends/impala/pandas_interop.py b/ibis/backends/impala/pandas_interop.py\\nindex f410a8b..e687884 100644\\n--- a/ibis/backends/impala/pandas_interop.py\\n+++ b/ibis/backends/impala/pandas_interop.py\\n@@ -22,7 +22,7 @@ from posixpath import join as pjoin\\n import ibis.backends.pandas.client # noqa: F401\\n import ibis.common.exceptions as com\\n import ibis.expr.schema as sch\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.config import options\\n \\n \\ndiff --git a/ibis/backends/impala/tests/conftest.py b/ibis/backends/impala/tests/conftest.py\\nindex 1075ebe..a815be5 100644\\n--- a/ibis/backends/impala/tests/conftest.py\\n+++ b/ibis/backends/impala/tests/conftest.py\\n@@ -13,8 +13,7 @@ import pytest\\n \\n import ibis\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n-from ibis import options\\n+from ibis import options, util\\n from ibis.backends.base import BaseBackend\\n from ibis.backends.conftest import TEST_TABLES, _random_identifier\\n from ibis.backends.impala.compiler import ImpalaCompiler, ImpalaExprTranslator\\ndiff --git a/ibis/backends/impala/tests/test_client.py b/ibis/backends/impala/tests/test_client.py\\nindex 0b56054..3fcca3a 100644\\n--- a/ibis/backends/impala/tests/test_client.py\\n+++ b/ibis/backends/impala/tests/test_client.py\\n@@ -7,9 +7,9 @@ import pytz\\n \\n import ibis\\n import ibis.common.exceptions as com\\n-import ibis.config as config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n+from ibis import config\\n from ibis.tests.util import assert_equal\\n \\n pytest.importorskip(\\\"impala\\\")\\ndiff --git a/ibis/backends/impala/tests/test_ddl.py b/ibis/backends/impala/tests/test_ddl.py\\nindex 870c4dc..2346a3d 100644\\n--- a/ibis/backends/impala/tests/test_ddl.py\\n+++ b/ibis/backends/impala/tests/test_ddl.py\\n@@ -6,7 +6,7 @@ import ibis\\n import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.ddl import fully_qualified_re\\n from ibis.tests.util import assert_equal\\n \\ndiff --git a/ibis/backends/impala/tests/test_exprs.py b/ibis/backends/impala/tests/test_exprs.py\\nindex cfc8552..1d6f44f 100644\\n--- a/ibis/backends/impala/tests/test_exprs.py\\n+++ b/ibis/backends/impala/tests/test_exprs.py\\n@@ -5,10 +5,10 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.types as ir\\n from ibis import literal as L\\n from ibis.backends.impala.compiler import ImpalaCompiler\\n+from ibis.expr import api\\n from ibis.expr.datatypes import Category\\n \\n \\ndiff --git a/ibis/backends/impala/tests/test_partition.py b/ibis/backends/impala/tests/test_partition.py\\nindex 1f96e7d..44217a4 100644\\n--- a/ibis/backends/impala/tests/test_partition.py\\n+++ b/ibis/backends/impala/tests/test_partition.py\\n@@ -6,7 +6,7 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.tests.util import assert_equal\\n \\n pytest.importorskip(\\\"impala\\\")\\ndiff --git a/ibis/backends/impala/tests/test_udf.py b/ibis/backends/impala/tests/test_udf.py\\nindex 895918b..fd950d5 100644\\n--- a/ibis/backends/impala/tests/test_udf.py\\n+++ b/ibis/backends/impala/tests/test_udf.py\\n@@ -9,11 +9,11 @@ import ibis\\n import ibis.backends.impala as api\\n import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n-import ibis.expr.rules as rules\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.impala import ddl\\n from ibis.common.exceptions import IbisTypeError\\n+from ibis.expr import rules\\n \\n pytest.importorskip(\\\"impala\\\")\\n \\ndiff --git a/ibis/backends/impala/udf.py b/ibis/backends/impala/udf.py\\nindex c6f2ef6..8b8b552 100644\\n--- a/ibis/backends/impala/udf.py\\n+++ b/ibis/backends/impala/udf.py\\n@@ -21,7 +21,7 @@ import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.udf.validate as v\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.registry import fixed_arity, sql_type_names\\n from ibis.backends.impala.compiler import ImpalaExprTranslator\\n \\ndiff --git a/ibis/backends/mysql/__init__.py b/ibis/backends/mysql/__init__.py\\nindex c0ddacb..50b331a 100644\\n--- a/ibis/backends/mysql/__init__.py\\n+++ b/ibis/backends/mysql/__init__.py\\n@@ -8,7 +8,7 @@ import warnings\\n from typing import Literal\\n \\n import sqlalchemy as sa\\n-import sqlalchemy.dialects.mysql as mysql\\n+from sqlalchemy.dialects import mysql\\n \\n import ibis.expr.datatypes as dt\\n import ibis.expr.schema as sch\\ndiff --git a/ibis/backends/mysql/compiler.py b/ibis/backends/mysql/compiler.py\\nindex 13819cb..7456f71 100644\\n--- a/ibis/backends/mysql/compiler.py\\n+++ b/ibis/backends/mysql/compiler.py\\n@@ -1,7 +1,7 @@\\n from __future__ import annotations\\n \\n import sqlalchemy as sa\\n-import sqlalchemy.dialects.mysql as mysql\\n+from sqlalchemy.dialects import mysql\\n \\n import ibis.expr.datatypes as dt\\n from ibis.backends.base.sql.alchemy import AlchemyCompiler, AlchemyExprTranslator\\ndiff --git a/ibis/backends/postgres/tests/test_functions.py b/ibis/backends/postgres/tests/test_functions.py\\nindex 33c6d2e..0f377e3 100644\\n--- a/ibis/backends/postgres/tests/test_functions.py\\n+++ b/ibis/backends/postgres/tests/test_functions.py\\n@@ -11,9 +11,9 @@ import pytest\\n from pytest import param\\n \\n import ibis\\n-import ibis.config as config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n+from ibis import config\\n from ibis import literal as L\\n from ibis.expr.window import rows_with_max_lookback\\n \\ndiff --git a/ibis/backends/pyspark/__init__.py b/ibis/backends/pyspark/__init__.py\\nindex 1b42080..b994911 100644\\n--- a/ibis/backends/pyspark/__init__.py\\n+++ b/ibis/backends/pyspark/__init__.py\\n@@ -14,8 +14,7 @@ import ibis.config\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.expr.types as types\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql import BaseSQLBackend\\n from ibis.backends.base.sql.compiler import Compiler, TableSetFormatter\\n from ibis.backends.base.sql.ddl import (\\n@@ -217,16 +216,16 @@ class Backend(BaseSQLBackend):\\n **kwargs: Any,\\n ) -> Any:\\n \\\"\\\"\\\"Execute an expression.\\\"\\\"\\\"\\n- if isinstance(expr, types.Table):\\n+ if isinstance(expr, ir.Table):\\n return self.compile(expr, timecontext, params, **kwargs).toPandas()\\n- elif isinstance(expr, types.Column):\\n+ elif isinstance(expr, ir.Column):\\n # expression must be named for the projection\\n if not expr.has_name():\\n expr = expr.name(\\\"tmp\\\")\\n return self.compile(\\n expr.to_projection(), timecontext, params, **kwargs\\n ).toPandas()[expr.get_name()]\\n- elif isinstance(expr, types.Scalar):\\n+ elif isinstance(expr, ir.Scalar):\\n compiled = self.compile(expr, timecontext, params, **kwargs)\\n if isinstance(compiled, Column):\\n # attach result column to a fake DataFrame and\\ndiff --git a/ibis/backends/pyspark/tests/test_ddl.py b/ibis/backends/pyspark/tests/test_ddl.py\\nindex 0288062..ccc8a97 100644\\n--- a/ibis/backends/pyspark/tests/test_ddl.py\\n+++ b/ibis/backends/pyspark/tests/test_ddl.py\\n@@ -5,7 +5,7 @@ import pytest\\n \\n import ibis\\n import ibis.common.exceptions as com\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.tests.util import assert_equal\\n \\n pyspark = pytest.importorskip(\\\"pyspark\\\")\\ndiff --git a/ibis/backends/sqlite/tests/test_client.py b/ibis/backends/sqlite/tests/test_client.py\\nindex 95aa24d..ad64700 100644\\n--- a/ibis/backends/sqlite/tests/test_client.py\\n+++ b/ibis/backends/sqlite/tests/test_client.py\\n@@ -5,8 +5,8 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.config as config\\n import ibis.expr.types as ir\\n+from ibis import config\\n \\n pytest.importorskip(\\\"sqlalchemy\\\")\\n \\ndiff --git a/ibis/expr/format.py b/ibis/expr/format.py\\nindex e3d48cd..85fab3f 100644\\n--- a/ibis/expr/format.py\\n+++ b/ibis/expr/format.py\\n@@ -9,13 +9,13 @@ from typing import Any, Callable, Deque, Iterable, Mapping, Tuple\\n import rich.pretty\\n \\n import ibis\\n-import ibis.common.graph as graph\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n import ibis.expr.window as win\\n-import ibis.util as util\\n+from ibis import util\\n+from ibis.common import graph\\n \\n Aliases = Mapping[ops.TableNode, int]\\n Deps = Deque[Tuple[int, ops.TableNode]]\\ndiff --git a/ibis/expr/operations/relations.py b/ibis/expr/operations/relations.py\\nindex 080ddcd..de44a15 100644\\n--- a/ibis/expr/operations/relations.py\\n+++ b/ibis/expr/operations/relations.py\\n@@ -11,7 +11,7 @@ import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.annotations import attribute\\n from ibis.expr.deferred import Deferred\\n from ibis.expr.operations.core import Named, Node, Value\\ndiff --git a/ibis/expr/rules.py b/ibis/expr/rules.py\\nindex 9b1a3b7..d40700e 100644\\n--- a/ibis/expr/rules.py\\n+++ b/ibis/expr/rules.py\\n@@ -11,7 +11,7 @@ import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.annotations import attribute, optional\\n from ibis.common.validators import (\\n bool_,\\ndiff --git a/ibis/expr/timecontext.py b/ibis/expr/timecontext.py\\nindex 7ecd8e7..9620d6c 100644\\n--- a/ibis/expr/timecontext.py\\n+++ b/ibis/expr/timecontext.py\\n@@ -38,8 +38,8 @@ from typing import TYPE_CHECKING, Any\\n import numpy as np\\n \\n import ibis.common.exceptions as com\\n-import ibis.config as config\\n import ibis.expr.operations as ops\\n+from ibis import config\\n \\n if TYPE_CHECKING:\\n import pandas as pd\\ndiff --git a/ibis/expr/types/groupby.py b/ibis/expr/types/groupby.py\\nindex 138f92e..97aaaa2 100644\\n--- a/ibis/expr/types/groupby.py\\n+++ b/ibis/expr/types/groupby.py\\n@@ -22,7 +22,7 @@ from typing import Iterable, Sequence\\n import ibis.expr.analysis as an\\n import ibis.expr.types as ir\\n import ibis.expr.window as _window\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.expr.deferred import Deferred\\n \\n _function_types = tuple(\\ndiff --git a/ibis/expr/window.py b/ibis/expr/window.py\\nindex 5ef3bb1..3e0efdc 100644\\n--- a/ibis/expr/window.py\\n+++ b/ibis/expr/window.py\\n@@ -11,7 +11,7 @@ import toolz\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.exceptions import IbisInputError\\n from ibis.common.grounds import Comparable\\n \\ndiff --git a/ibis/tests/expr/test_decimal.py b/ibis/tests/expr/test_decimal.py\\nindex 85d8eb2..12b809b 100644\\n--- a/ibis/tests/expr/test_decimal.py\\n+++ b/ibis/tests/expr/test_decimal.py\\n@@ -3,10 +3,10 @@ import operator\\n import pytest\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n+from ibis.expr import api\\n \\n \\n def test_type_metadata(lineitem):\\ndiff --git a/ibis/tests/expr/test_interactive.py b/ibis/tests/expr/test_interactive.py\\nindex cea1945..0c5613b 100644\\n--- a/ibis/tests/expr/test_interactive.py\\n+++ b/ibis/tests/expr/test_interactive.py\\n@@ -14,7 +14,7 @@\\n \\n import pytest\\n \\n-import ibis.config as config\\n+from ibis import config\\n from ibis.tests.expr.mocks import MockBackend\\n \\n \\ndiff --git a/ibis/tests/expr/test_table.py b/ibis/tests/expr/test_table.py\\nindex 04f4a7d..3f77985 100644\\n--- a/ibis/tests/expr/test_table.py\\n+++ b/ibis/tests/expr/test_table.py\\n@@ -10,13 +10,13 @@ from pytest import param\\n import ibis\\n import ibis.common.exceptions as com\\n import ibis.expr.analysis as an\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n from ibis import _\\n from ibis import literal as L\\n from ibis.common.exceptions import RelationError\\n+from ibis.expr import api\\n from ibis.expr.types import Column, Table\\n from ibis.tests.expr.mocks import MockAlchemyBackend, MockBackend\\n from ibis.tests.util import assert_equal, assert_pickle_roundtrip\\ndiff --git a/ibis/tests/expr/test_temporal.py b/ibis/tests/expr/test_temporal.py\\nindex e76e71c..9a0f43f 100644\\n--- a/ibis/tests/expr/test_temporal.py\\n+++ b/ibis/tests/expr/test_temporal.py\\n@@ -5,10 +5,10 @@ import pytest\\n from pytest import param\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n+from ibis.expr import api\\n \\n \\n def test_temporal_literals():\\ndiff --git a/ibis/tests/expr/test_timestamp.py b/ibis/tests/expr/test_timestamp.py\\nindex 6601c8b..7782787 100644\\n--- a/ibis/tests/expr/test_timestamp.py\\n+++ b/ibis/tests/expr/test_timestamp.py\\n@@ -5,11 +5,11 @@ import pandas as pd\\n import pytest\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.types as ir\\n+from ibis.expr import api\\n \\n \\n def test_field_select(alltypes):\\ndiff --git a/ibis/tests/expr/test_value_exprs.py b/ibis/tests/expr/test_value_exprs.py\\nindex 4c3d475..9eb247c 100644\\n--- a/ibis/tests/expr/test_value_exprs.py\\n+++ b/ibis/tests/expr/test_value_exprs.py\\n@@ -15,13 +15,13 @@ from pytest import param\\n import ibis\\n import ibis.common.exceptions as com\\n import ibis.expr.analysis as L\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.types as ir\\n from ibis import _, literal\\n from ibis.common.exceptions import IbisTypeError\\n+from ibis.expr import api\\n from ibis.tests.util import assert_equal\\n \\n \\ndiff --git a/ibis/tests/expr/test_visualize.py b/ibis/tests/expr/test_visualize.py\\nindex 5525944..253564f 100644\\n--- a/ibis/tests/expr/test_visualize.py\\n+++ b/ibis/tests/expr/test_visualize.py\\n@@ -9,8 +9,8 @@ import ibis.expr.types as ir\\n \\n pytest.importorskip('graphviz')\\n \\n-import ibis.expr.api as api # noqa: E402\\n import ibis.expr.visualize as viz # noqa: E402\\n+from ibis.expr import api # noqa: E402\\n \\n pytestmark = pytest.mark.skipif(\\n int(os.environ.get('CONDA_BUILD', 0)) == 1, reason='CONDA_BUILD defined'\\ndiff --git a/ibis/tests/sql/test_sqlalchemy.py b/ibis/tests/sql/test_sqlalchemy.py\\nindex 2ad5453..3aa8c3d 100644\\n--- a/ibis/tests/sql/test_sqlalchemy.py\\n+++ b/ibis/tests/sql/test_sqlalchemy.py\\n@@ -15,8 +15,8 @@\\n import operator\\n \\n import pytest\\n-import sqlalchemy.sql as sql\\n from sqlalchemy import func as F\\n+from sqlalchemy import sql\\n from sqlalchemy import types as sat\\n \\n import ibis\\ndiff --git a/ibis/tests/util.py b/ibis/tests/util.py\\nindex f79d09a..025bfc7 100644\\n--- a/ibis/tests/util.py\\n+++ b/ibis/tests/util.py\\n@@ -5,7 +5,7 @@ from __future__ import annotations\\n import pickle\\n \\n import ibis\\n-import ibis.util as util\\n+from ibis import util\\n \\n \\n def assert_equal(left, right):\\ndiff --git a/pyproject.toml b/pyproject.toml\\nindex f2146d4..492ad9e 100644\\n--- a/pyproject.toml\\n+++ b/pyproject.toml\\n@@ -310,6 +310,7 @@ select = [\\n \\\"PGH\\\", # pygrep-hooks\\n \\\"PLC\\\", # pylint\\n \\\"PLE\\\", # pylint\\n+ \\\"PLR\\\", # pylint import style\\n \\\"PLW\\\", # pylint\\n \\\"RET\\\", # flake8-return\\n \\\"RUF\\\", # ruff-specific rules\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"d0814a928601706635287fd3d9d3451d156b821a\", \"92e940efeee199b1e0bbbc3c9eea7f3dc8221619\", \"ab1e60a97c6d5c688dacbd23bca40cb8f20c4ac3\", \"8d53d724275ebe4b2a0bb0bd7e2c2dfc399e049b\"]"},"types":{"kind":"string","value":"[\"ci\", \"build\", \"fix\", \"refactor\"]"}}},{"rowIdx":1071,"cells":{"commit_message":{"kind":"string","value":"remove unnecessary lines from verify-wal test,use new, public `quay.io/influxdb/iox` image,fix unstable MessageCorrelationTest,set cursor position in setHorizontalRule correctly, fix #2429"},"diff":{"kind":"string","value":"[\"diff --git a/storage/wal/verifier_test.go b/storage/wal/verifier_test.go\\nindex 61e1536..a44755f 100644\\n--- a/storage/wal/verifier_test.go\\n+++ b/storage/wal/verifier_test.go\\n@@ -138,22 +138,13 @@ func writeCorruptEntries(file *os.File, t *testing.T, n int) {\\n \\t\\t}\\n \\t}\\n \\n-\\n \\t// Write some random bytes to the file to simulate corruption.\\n \\tif _, err := file.Write(corruption); err != nil {\\n \\t\\tfatal(t, \\\"corrupt WAL segment\\\", err)\\n \\t}\\n-\\tcorrupt := []byte{1, 255, 0, 3, 45, 26, 110}\\n-\\n-\\twrote, err := file.Write(corrupt)\\n-\\tif err != nil {\\n-\\t\\tt.Fatal(err)\\n-\\t} else if wrote != len(corrupt) {\\n-\\t\\tt.Fatal(\\\"Error writing corrupt data to file\\\")\\n-\\t}\\n \\n \\tif err := file.Close(); err != nil {\\n-\\t\\tt.Fatalf(\\\"Error: filed to close file: %v\\\\n\\\", err)\\n+\\t\\tt.Fatalf(\\\"Error: failed to close file: %v\\\\n\\\", err)\\n \\t}\\n }\\n \\n\", \"diff --git a/.circleci/config.yml b/.circleci/config.yml\\nindex 3ae6728..a5f2d2f 100644\\n--- a/.circleci/config.yml\\n+++ b/.circleci/config.yml\\n@@ -12,7 +12,7 @@\\n # The CI for every PR and merge to main runs tests, fmt, lints and compiles debug binaries\\n #\\n # On main if all these checks pass it will then additionally compile in \\\"release\\\" mode and\\n-# publish a docker image to quay.io/influxdb/fusion:$COMMIT_SHA\\n+# publish a docker image to quay.io/influxdb/iox:$COMMIT_SHA\\n #\\n # Manual CI Image:\\n #\\n@@ -317,11 +317,11 @@ jobs:\\n #\\n # Uses the latest ci_image (influxdb/rust below) to build a release binary and\\n # copies it to a minimal container image based upon `rust:slim-buster`. This\\n- # minimal image is then pushed to `quay.io/influxdb/fusion:${BRANCH}` with '/'\\n+ # minimal image is then pushed to `quay.io/influxdb/iox:${BRANCH}` with '/'\\n # repaced by '.' - as an example:\\n #\\n # git branch: dom/my-awesome-feature/perf\\n- # container: quay.io/influxdb/fusion:dom.my-awesome-feature.perf\\n+ # container: quay.io/influxdb/iox:dom.my-awesome-feature.perf\\n #\\n # Subsequent CI runs will overwrite the tag if you push more changes, so watch\\n # out for parallel CI runs!\\n@@ -365,7 +365,7 @@ jobs:\\n sudo apt-get update\\n sudo apt-get install -y docker.io\\n - run: |\\n- echo \\\"$QUAY_PASS\\\" | docker login quay.io --username $QUAY_USER --password-stdin\\n+ echo \\\"$QUAY_INFLUXDB_IOX_PASS\\\" | docker login quay.io --username $QUAY_INFLUXDB_IOX_USER --password-stdin\\n - run:\\n # Docker has functionality to support per-Dockerfile .dockerignore\\n # This was added in https://github.com/moby/buildkit/pull/901\\n@@ -379,8 +379,8 @@ jobs:\\n echo sha256sum after build is\\n sha256sum target/release/influxdb_iox\\n COMMIT_SHA=$(git rev-parse --short HEAD)\\n- docker build -t quay.io/influxdb/fusion:$COMMIT_SHA -f docker/Dockerfile.iox .\\n- docker push quay.io/influxdb/fusion:$COMMIT_SHA\\n+ docker build -t quay.io/influxdb/iox:$COMMIT_SHA -f docker/Dockerfile.iox .\\n+ docker push quay.io/influxdb/iox:$COMMIT_SHA\\n echo \\\"export COMMIT_SHA=${COMMIT_SHA}\\\" >> $BASH_ENV\\n - run:\\n name: Deploy tags\\n\", \"diff --git a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\\nindex 0f5fed9..796393c 100644\\n--- a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\\n+++ b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\\n@@ -27,7 +27,6 @@ import static io.zeebe.test.util.MsgPackUtil.asMsgPack;\\n import static org.assertj.core.api.Assertions.assertThat;\\n import static org.assertj.core.api.Assertions.entry;\\n \\n-import io.zeebe.UnstableTest;\\n import io.zeebe.broker.test.EmbeddedBrokerRule;\\n import io.zeebe.model.bpmn.Bpmn;\\n import io.zeebe.model.bpmn.BpmnModelInstance;\\n@@ -50,7 +49,6 @@ import org.agrona.DirectBuffer;\\n import org.junit.Before;\\n import org.junit.Rule;\\n import org.junit.Test;\\n-import org.junit.experimental.categories.Category;\\n import org.junit.rules.RuleChain;\\n import org.junit.runner.RunWith;\\n import org.junit.runners.Parameterized;\\n@@ -165,7 +163,7 @@ public class MessageCorrelationTest {\\n \\\"receive-message\\\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\\n \\n final SubscribedRecord messageSubscription =\\n- findMessageSubscription(testClient, MessageSubscriptionIntent.OPENED);\\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\\n assertThat(messageSubscription.valueType()).isEqualTo(ValueType.MESSAGE_SUBSCRIPTION);\\n assertThat(messageSubscription.recordType()).isEqualTo(RecordType.EVENT);\\n assertThat(messageSubscription.value())\\n@@ -244,7 +242,7 @@ public class MessageCorrelationTest {\\n final long workflowInstanceKey =\\n testClient.createWorkflowInstance(\\\"wf\\\", asMsgPack(\\\"orderId\\\", \\\"order-123\\\"));\\n \\n- testClient.receiveFirstWorkflowInstanceEvent(WorkflowInstanceIntent.ELEMENT_ACTIVATED);\\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\\n \\n // when\\n testClient.publishMessage(\\\"order canceled\\\", \\\"order-123\\\", asMsgPack(\\\"foo\\\", \\\"bar\\\"));\\n@@ -308,13 +306,12 @@ public class MessageCorrelationTest {\\n }\\n \\n @Test\\n- @Category(UnstableTest.class) // => https://github.com/zeebe-io/zeebe/issues/1234\\n public void shouldCorrelateMessageWithZeroTTL() throws Exception {\\n // given\\n final long workflowInstanceKey =\\n testClient.createWorkflowInstance(\\\"wf\\\", asMsgPack(\\\"orderId\\\", \\\"order-123\\\"));\\n \\n- testClient.receiveElementInState(\\\"receive-message\\\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\\n \\n // when\\n testClient.publishMessage(\\\"order canceled\\\", \\\"order-123\\\", asMsgPack(\\\"foo\\\", \\\"bar\\\"), 0);\\n@@ -499,10 +496,9 @@ public class MessageCorrelationTest {\\n .containsEntry(\\\"activityInstanceKey\\\", catchEventEntered.key());\\n }\\n \\n- private SubscribedRecord findMessageSubscription(\\n- final TestPartitionClient client, final MessageSubscriptionIntent intent)\\n+ private SubscribedRecord findMessageSubscription(final MessageSubscriptionIntent intent)\\n throws AssertionError {\\n- return client\\n+ return testClient\\n .receiveEvents()\\n .filter(intent(intent))\\n .findFirst()\\n\", \"diff --git a/packages/extension-horizontal-rule/src/horizontal-rule.ts b/packages/extension-horizontal-rule/src/horizontal-rule.ts\\nindex 6f583e1..c905b63 100644\\n--- a/packages/extension-horizontal-rule/src/horizontal-rule.ts\\n+++ b/packages/extension-horizontal-rule/src/horizontal-rule.ts\\n@@ -49,15 +49,14 @@ export const HorizontalRule = Node.create({\\n // set cursor after horizontal rule\\n .command(({ tr, dispatch }) => {\\n if (dispatch) {\\n- const { parent, pos } = tr.selection.$from\\n- const posAfter = pos + 1\\n- const nodeAfter = tr.doc.nodeAt(posAfter)\\n+ const { $to } = tr.selection\\n+ const posAfter = $to.end()\\n \\n- if (nodeAfter) {\\n- tr.setSelection(TextSelection.create(tr.doc, posAfter))\\n+ if ($to.nodeAfter) {\\n+ tr.setSelection(TextSelection.create(tr.doc, $to.pos))\\n } else {\\n // add node after horizontal rule if it\\u2019s the end of the document\\n- const node = parent.type.contentMatch.defaultType?.create()\\n+ const node = $to.parent.type.contentMatch.defaultType?.create()\\n \\n if (node) {\\n tr.insert(posAfter, node)\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"fba4326c72fc22d81aba6976a9fef1e4b6154fd9\", \"f751bb5426b87f82096d620f1cd6203badf45d58\", \"98bed2a8137930149559bc1cae9bd34a1a75e556\", \"34d80114704679118e9bb6058e0d6c7aa03fd4b5\"]"},"types":{"kind":"string","value":"[\"refactor\", \"ci\", \"test\", \"fix\"]"}}},{"rowIdx":1072,"cells":{"commit_message":{"kind":"string","value":"update get-started,fix golden tests for aws_vpn_connection,extract lambdas,set name for topology module"},"diff":{"kind":"string","value":"[\"diff --git a/docs/src/go-client/get-started.md b/docs/src/go-client/get-started.md\\nindex 4f4405f..a792e0e 100755\\n--- a/docs/src/go-client/get-started.md\\n+++ b/docs/src/go-client/get-started.md\\n@@ -199,14 +199,12 @@ workflowKey:1 bpmnProcessId:\\\"order-process\\\" version:1 workflowInstanceKey:6\\n \\n You did it! You want to see how the workflow instance is executed?\\n \\n-Start the Zeebe Monitor using `java -jar zeebe-simple-monitor.jar`.\\n+Start the Zeebe Monitor using `java -jar zeebe-simple-monitor-app-*.jar`.\\n \\n Open a web browser and go to .\\n \\n-Connect to the broker and switch to the workflow instances view.\\n-Here, you see the current state of the workflow instance which includes active jobs, completed activities, the payload and open incidents.\\n-\\n-![zeebe-monitor-step-1](/java-client/zeebe-monitor-1.png)\\n+Here, you see the current state of the workflow instance.\\n+![zeebe-monitor-step-1](/java-client/java-get-started-monitor-1.gif)\\n \\n \\n ## Work on a task\\n@@ -322,7 +320,7 @@ it encounters a problem while processing the job.\\n \\n When you have a look at the Zeebe Monitor, then you can see that the workflow instance moved from the first service task to the next one:\\n \\n-![zeebe-monitor-step-2](/go-client/zeebe-monitor-2.png)\\n+![zeebe-monitor-step-2](/java-client/java-get-started-monitor-2.gif)\\n \\n When you run the above example you should see similar output:\\n \\ndiff --git a/docs/src/go-client/java-get-started-monitor-1.gif b/docs/src/go-client/java-get-started-monitor-1.gif\\nnew file mode 100644\\nindex 0000000..b86803a\\nBinary files /dev/null and b/docs/src/go-client/java-get-started-monitor-1.gif differ\\ndiff --git a/docs/src/go-client/java-get-started-monitor-2.gif b/docs/src/go-client/java-get-started-monitor-2.gif\\nnew file mode 100644\\nindex 0000000..8f0f2a4\\nBinary files /dev/null and b/docs/src/go-client/java-get-started-monitor-2.gif differ\\ndiff --git a/docs/src/go-client/zeebe-monitor-1.png b/docs/src/go-client/zeebe-monitor-1.png\\ndeleted file mode 100644\\nindex 0075f3d..0000000\\nBinary files a/docs/src/go-client/zeebe-monitor-1.png and /dev/null differ\\ndiff --git a/docs/src/go-client/zeebe-monitor-2.png b/docs/src/go-client/zeebe-monitor-2.png\\ndeleted file mode 100644\\nindex 6687bb0..0000000\\nBinary files a/docs/src/go-client/zeebe-monitor-2.png and /dev/null differ\\ndiff --git a/docs/src/go-client/zeebe-monitor-3.png b/docs/src/go-client/zeebe-monitor-3.png\\ndeleted file mode 100644\\nindex bc15659..0000000\\nBinary files a/docs/src/go-client/zeebe-monitor-3.png and /dev/null differ\\ndiff --git a/docs/src/introduction/quickstart.md b/docs/src/introduction/quickstart.md\\nindex 70abacf..68be28b 100644\\n--- a/docs/src/introduction/quickstart.md\\n+++ b/docs/src/introduction/quickstart.md\\n@@ -215,7 +215,7 @@ and completed by a [job worker](/basics/job-workers.html). A job worker is a\\n long living process which repeatedly tries to activate jobs for a given job\\n type and completes them after executing its business logic. The `zbctl` also\\n provides a command to spawn simple job workers using an external command or\\n-script. The job worker will receive for every job the payload as JSON object on\\n+script. The job worker will receive for every job the workflow instance variables as JSON object on\\n `stdin` and has to return its result also as JSON object on `stdout` if it\\n handled the job successfully.\\n \\ndiff --git a/docs/src/java-client/get-started.md b/docs/src/java-client/get-started.md\\nindex 54d2208..afc1fd4 100755\\n--- a/docs/src/java-client/get-started.md\\n+++ b/docs/src/java-client/get-started.md\\n@@ -21,9 +21,9 @@ You will be guided through the following steps:\\n * [Zeebe Modeler](https://github.com/zeebe-io/zeebe-modeler/releases)\\n * [Zeebe Monitor](https://github.com/zeebe-io/zeebe-simple-monitor/releases)\\n \\n-Before you begin to setup your project please start the broker, i.e. by running the start up script \\n-`bin/broker` or `bin/broker.bat` in the distribution. Per default the broker is binding to the \\n-address `localhost:26500`, which is used as contact point in this guide. In case your broker is \\n+Before you begin to setup your project please start the broker, i.e. by running the start up script\\n+`bin/broker` or `bin/broker.bat` in the distribution. Per default the broker is binding to the\\n+address `localhost:26500`, which is used as contact point in this guide. In case your broker is\\n available under another address please adjust the broker contact point when building the client.\\n \\n ## Set up a project\\n@@ -182,14 +182,12 @@ Workflow instance created. Key: 6\\n \\n You did it! You want to see how the workflow instance is executed?\\n \\n-Start the Zeebe Monitor using `java -jar zeebe-simple-monitor.jar`.\\n+Start the Zeebe Monitor using `java -jar zeebe-simple-monitor-app-*.jar`.\\n \\n Open a web browser and go to .\\n \\n-Connect to the broker and switch to the workflow instances view.\\n-Here, you see the current state of the workflow instance which includes active jobs, completed activities, the payload and open incidents.\\n-\\n-![zeebe-monitor-step-1](/java-client/zeebe-monitor-1.png)\\n+Here, you see the current state of the workflow instance.\\n+![zeebe-monitor-step-1](/java-client/java-get-started-monitor-1.gif)\\n \\n ## Work on a job\\n \\n@@ -205,12 +203,9 @@ Insert a few service tasks between the start and the end event.\\n You need to set the type of each task, which identifies the nature of the work to be performed.\\n Set the type of the first task to 'payment-service'.\\n \\n-Optionally, you can define parameters of the task by adding headers.\\n-Add the header `method = VISA` to the first task.\\n-\\n Save the BPMN diagram and switch back to the main class.\\n \\n-Add the following lines to create a [job worker][] for the first jobs type:\\n+Add the following lines to create a job worker for the first jobs type:\\n \\n ```java\\n package io.zeebe;\\n@@ -227,10 +222,7 @@ public class Application\\n .jobType(\\\"payment-service\\\")\\n .handler((jobClient, job) ->\\n {\\n- final Map headers = job.getCustomHeaders();\\n- final String method = (String) headers.get(\\\"method\\\");\\n-\\n- System.out.println(\\\"Collect money using payment method: \\\" + method);\\n+ System.out.println(\\\"Collect money\\\");\\n \\n // ...\\n \\n@@ -252,40 +244,29 @@ public class Application\\n Run the program and verify that the job is processed. You should see the output:\\n \\n ```\\n-Collect money using payment method: VISA\\n+Collect money\\n ```\\n \\n When you have a look at the Zeebe Monitor, then you can see that the workflow instance moved from the first service task to the next one:\\n \\n-![zeebe-monitor-step-2](/java-client/zeebe-monitor-2.png)\\n+![zeebe-monitor-step-2](/java-client/java-get-started-monitor-2.gif)\\n \\n ## Work with data\\n \\n-Usually, a workflow is more than just tasks, there is also data flow.\\n-The tasks need data as input and in order to produce data.\\n+Usually, a workflow is more than just tasks, there is also a data flow. The worker gets the data from the workflow instance to do its work and send the result back to the workflow instance.\\n \\n-In Zeebe, the data is represented as a JSON document.\\n-When you create a workflow instance, then you can pass the data as payload.\\n-Within the workflow, you can use input and output mappings on tasks to control the data flow.\\n+In Zeebe, the data is stored as key-value-pairs in form of variables. Variables can be set when the workflow instance is created. Within the workflow, variables can be read and modified by workers.\\n \\n-In our example, we want to create a workflow instance with the following data:\\n+In our example, we want to create a workflow instance with the following variables:\\n \\n ```json\\n-{\\n- \\\"orderId\\\": 31243,\\n- \\\"orderItems\\\": [435, 182, 376]\\n-}\\n+\\\"orderId\\\": 31243\\n+\\\"orderItems\\\": [435, 182, 376]\\n ```\\n \\n-The first task should take `orderId` as input and return `totalPrice` as result.\\n-\\n-Open the BPMN diagram and switch to the input-output-mappings of the first task.\\n-Add the input mapping `$.orderId : $.orderId` and the output mapping `$.totalPrice : $.totalPrice`.\\n+The first task should read `orderId` as input and return `totalPrice` as result.\\n \\n-Save the BPMN diagram and go back to the main class.\\n-\\n-Modify the create command and pass the data as variables.\\n-Also, modify the job worker to read the jobs payload and complete the job with payload.\\n+Modify the workflow instance create command and pass the data as variables. Also, modify the job worker to read the job variables and complete the job with a result.\\n \\n ```java\\n package io.zeebe;\\n@@ -313,23 +294,22 @@ public class Application\\n .jobType(\\\"payment-service\\\")\\n .handler((jobClient, job) ->\\n {\\n- final Map headers = job.getCustomHeaders();\\n- final String method = (String) headers.get(\\\"method\\\");\\n-\\n- final Map payload = job.getPayloadAsMap();\\n+ final Map variables = job.getVariablesAsMap();\\n \\n- System.out.println(\\\"Process order: \\\" + payload.get(\\\"orderId\\\"));\\n- System.out.println(\\\"Collect money using payment method: \\\" + method);\\n+ System.out.println(\\\"Process order: \\\" + variables.get(\\\"orderId\\\"));\\n+ System.out.println(\\\"Collect money\\\");\\n \\n // ...\\n \\n- payload.put(\\\"totalPrice\\\", 46.50);\\n+ final Map result = new HashMap<>();\\n+ result.put(\\\"totalPrice\\\", 46.50);\\n \\n jobClient.newCompleteCommand(job.getKey())\\n- .payload(payload)\\n+ .variables(result)\\n .send()\\n .join();\\n })\\n+ .fetchVariables(\\\"orderId\\\")\\n .open();\\n \\n // ...\\n@@ -337,16 +317,16 @@ public class Application\\n }\\n ```\\n \\n-Run the program and verify that the payload is mapped into the job. You should see the output:\\n+Run the program and verify that the variable is read. You should see the output:\\n \\n ```\\n-Process order: {\\\"orderId\\\":31243}\\n-Collect money using payment method: VISA\\n+Process order: 31243\\n+Collect money\\n ```\\n \\n-When we have a look at the Zeebe Monitor, then we can see how the payload is modified after the activity:\\n+When we have a look at the Zeebe Monitor, then we can see that the variable `totalPrice` is set:\\n \\n-![zeebe-monitor-step-3](/java-client/zeebe-monitor-3.png)\\n+![zeebe-monitor-step-3](/java-client/java-get-started-monitor-3.gif)\\n \\n ## What's next?\\n \\ndiff --git a/docs/src/java-client/java-get-started-monitor-1.gif b/docs/src/java-client/java-get-started-monitor-1.gif\\nnew file mode 100644\\nindex 0000000..b86803a\\nBinary files /dev/null and b/docs/src/java-client/java-get-started-monitor-1.gif differ\\ndiff --git a/docs/src/java-client/java-get-started-monitor-2.gif b/docs/src/java-client/java-get-started-monitor-2.gif\\nnew file mode 100644\\nindex 0000000..8f0f2a4\\nBinary files /dev/null and b/docs/src/java-client/java-get-started-monitor-2.gif differ\\ndiff --git a/docs/src/java-client/java-get-started-monitor-3.gif b/docs/src/java-client/java-get-started-monitor-3.gif\\nnew file mode 100644\\nindex 0000000..1f6cb56\\nBinary files /dev/null and b/docs/src/java-client/java-get-started-monitor-3.gif differ\\ndiff --git a/docs/src/java-client/zeebe-monitor-1.png b/docs/src/java-client/zeebe-monitor-1.png\\ndeleted file mode 100644\\nindex 0075f3d..0000000\\nBinary files a/docs/src/java-client/zeebe-monitor-1.png and /dev/null differ\\ndiff --git a/docs/src/java-client/zeebe-monitor-2.png b/docs/src/java-client/zeebe-monitor-2.png\\ndeleted file mode 100644\\nindex 6687bb0..0000000\\nBinary files a/docs/src/java-client/zeebe-monitor-2.png and /dev/null differ\\ndiff --git a/docs/src/java-client/zeebe-monitor-3.png b/docs/src/java-client/zeebe-monitor-3.png\\ndeleted file mode 100644\\nindex bc15659..0000000\\nBinary files a/docs/src/java-client/zeebe-monitor-3.png and /dev/null differ\\n\", \"diff --git a/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf b/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\\nindex d895677..cf10e3f 100644\\n--- a/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\\n+++ b/internal/providers/terraform/aws/testdata/vpn_connection_test/vpn_connection_test.tf\\n@@ -12,6 +12,7 @@ provider \\\"aws\\\" {\\n resource \\\"aws_vpn_connection\\\" \\\"vpn_connection\\\" {\\n customer_gateway_id = \\\"dummy-customer-gateway-id\\\"\\n type = \\\"ipsec.1\\\"\\n+ vpn_gateway_id = \\\"vpn-gateway-id\\\"\\n }\\n \\n resource \\\"aws_vpn_connection\\\" \\\"transit\\\" {\\n@@ -23,10 +24,11 @@ resource \\\"aws_vpn_connection\\\" \\\"transit\\\" {\\n resource \\\"aws_vpn_connection\\\" \\\"vpn_connection_withUsage\\\" {\\n customer_gateway_id = \\\"dummy-customer-gateway-id2\\\"\\n type = \\\"ipsec.1\\\"\\n+ vpn_gateway_id = \\\"vpn-gateway-id\\\"\\n }\\n \\n resource \\\"aws_vpn_connection\\\" \\\"transit_withUsage\\\" {\\n customer_gateway_id = \\\"dummy-customer-gateway-id2\\\"\\n type = \\\"ipsec.1\\\"\\n transit_gateway_id = \\\"dummy-transit-gateway-id2\\\"\\n-}\\n\\\\ No newline at end of file\\n+}\\n\", \"diff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\nindex 6ee5797..bcfcc72 100644\\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n@@ -224,7 +224,6 @@ public final class AsyncSnapshotDirector extends Actor\\n private void takeSnapshot() {\\n final var transientSnapshotFuture =\\n stateController.takeTransientSnapshot(lowerBoundSnapshotPosition);\\n-\\n transientSnapshotFuture.onComplete(\\n (optionalTransientSnapshot, snapshotTakenError) -> {\\n if (snapshotTakenError != null) {\\n@@ -237,27 +236,31 @@ public final class AsyncSnapshotDirector extends Actor\\n takingSnapshot = false;\\n return;\\n }\\n- pendingSnapshot = optionalTransientSnapshot.get();\\n- onRecovered();\\n-\\n- final ActorFuture lastWrittenPosition =\\n- streamProcessor.getLastWrittenPositionAsync();\\n- actor.runOnCompletion(\\n- lastWrittenPosition,\\n- (endPosition, error) -> {\\n- if (error == null) {\\n- LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\\n- lastWrittenEventPosition = endPosition;\\n- persistingSnapshot = false;\\n- persistSnapshotIfLastWrittenPositionCommitted();\\n- } else {\\n- resetStateOnFailure();\\n- LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\\n- }\\n- });\\n+ onTransientSnapshotTaken(optionalTransientSnapshot.get());\\n });\\n }\\n \\n+ private void onTransientSnapshotTaken(final TransientSnapshot transientSnapshot) {\\n+\\n+ pendingSnapshot = transientSnapshot;\\n+ onRecovered();\\n+\\n+ final ActorFuture lastWrittenPosition = streamProcessor.getLastWrittenPositionAsync();\\n+ actor.runOnCompletion(lastWrittenPosition, this::onLastWrittenPositionReceived);\\n+ }\\n+\\n+ private void onLastWrittenPositionReceived(final Long endPosition, final Throwable error) {\\n+ if (error == null) {\\n+ LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\\n+ lastWrittenEventPosition = endPosition;\\n+ persistingSnapshot = false;\\n+ persistSnapshotIfLastWrittenPositionCommitted();\\n+ } else {\\n+ resetStateOnFailure();\\n+ LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\\n+ }\\n+ }\\n+\\n private void onRecovered() {\\n if (healthStatus != HealthStatus.HEALTHY) {\\n healthStatus = HealthStatus.HEALTHY;\\n\", \"diff --git a/topology/pom.xml b/topology/pom.xml\\nindex 389508e..ee6239a 100644\\n--- a/topology/pom.xml\\n+++ b/topology/pom.xml\\n@@ -16,6 +16,7 @@\\n \\n \\n zeebe-cluster-topology\\n+ Zeebe Cluster Topology\\n \\n \\n ${maven.multiModuleProjectDirectory}/topology/src/main/resources/proto\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"cf6d526123abab2689b24a06aaf03d8e4d6ddff4\", \"9b059dd8245e72f0bf8c40fc633f9ef6fccae405\", \"14abf5c31523a551134aebe9e8f3505ef26ed421\", \"8911a972222dc80a242f3f1d9b3596321b3fdeaa\"]"},"types":{"kind":"string","value":"[\"docs\", \"test\", \"refactor\", \"build\"]"}}},{"rowIdx":1073,"cells":{"commit_message":{"kind":"string","value":"Introduce timediff fn (stub),removing automatic page push on nav,update version (nightly.0),use module path alias"},"diff":{"kind":"string","value":"[\"diff --git a/rust/Cargo.lock b/rust/Cargo.lock\\nindex b42616f..4795eb6 100644\\n--- a/rust/Cargo.lock\\n+++ b/rust/Cargo.lock\\n@@ -1287,7 +1287,7 @@ dependencies = [\\n [[package]]\\n name = \\\"datafusion\\\"\\n version = \\\"5.1.0\\\"\\n-source = \\\"git+https://github.com/cube-js/arrow-datafusion.git?rev=8df4132b83d896a0d3db5c82a4eaaa3eaa285d15#8df4132b83d896a0d3db5c82a4eaaa3eaa285d15\\\"\\n+source = \\\"git+https://github.com/cube-js/arrow-datafusion.git?rev=868f3c4de13d13cda84cee33475b9782b94fa60c#868f3c4de13d13cda84cee33475b9782b94fa60c\\\"\\n dependencies = [\\n \\\"ahash 0.7.4\\\",\\n \\\"arrow 6.0.0\\\",\\ndiff --git a/rust/cubesql/Cargo.toml b/rust/cubesql/Cargo.toml\\nindex 3cb386a..9aef494 100644\\n--- a/rust/cubesql/Cargo.toml\\n+++ b/rust/cubesql/Cargo.toml\\n@@ -9,7 +9,7 @@ documentation = \\\"https://cube.dev/docs\\\"\\n homepage = \\\"https://cube.dev\\\"\\n \\n [dependencies]\\n-datafusion = { git = 'https://github.com/cube-js/arrow-datafusion.git', rev = \\\"8df4132b83d896a0d3db5c82a4eaaa3eaa285d15\\\", default-features = false, features = [\\\"unicode_expressions\\\"] }\\n+datafusion = { git = 'https://github.com/cube-js/arrow-datafusion.git', rev = \\\"868f3c4de13d13cda84cee33475b9782b94fa60c\\\", default-features = false, features = [\\\"unicode_expressions\\\"] }\\n anyhow = \\\"1.0\\\"\\n thiserror = \\\"1.0\\\"\\n cubeclient = { path = \\\"../cubeclient\\\" }\\ndiff --git a/rust/cubesql/src/compile/engine/df/intervals.rs b/rust/cubesql/src/compile/engine/df/intervals.rs\\nnew file mode 100644\\nindex 0000000..9e6cb7e\\n--- /dev/null\\n+++ b/rust/cubesql/src/compile/engine/df/intervals.rs\\n@@ -0,0 +1,51 @@\\n+#[macro_export]\\n+macro_rules! make_string_interval_year_month {\\n+ ($array: ident, $row: ident) => {{\\n+ let s = if $array.is_null($row) {\\n+ \\\"NULL\\\".to_string()\\n+ } else {\\n+ let interval = $array.value($row) as f64;\\n+ let years = (interval / 12_f64).floor();\\n+ let month = interval - (years * 12_f64);\\n+\\n+ format!(\\n+ \\\"{} years {} mons 0 days 0 hours 0 mins 0.00 secs\\\",\\n+ years, month,\\n+ )\\n+ };\\n+\\n+ s\\n+ }};\\n+}\\n+\\n+#[macro_export]\\n+macro_rules! make_string_interval_day_time {\\n+ ($array: ident, $row: ident) => {{\\n+ let s = if $array.is_null($row) {\\n+ \\\"NULL\\\".to_string()\\n+ } else {\\n+ let value: u64 = $array.value($row) as u64;\\n+\\n+ let days_parts: i32 = ((value & 0xFFFFFFFF00000000) >> 32) as i32;\\n+ let milliseconds_part: i32 = (value & 0xFFFFFFFF) as i32;\\n+\\n+ let secs = milliseconds_part / 1000;\\n+ let mins = secs / 60;\\n+ let hours = mins / 60;\\n+\\n+ let secs = secs - (mins * 60);\\n+ let mins = mins - (hours * 60);\\n+\\n+ format!(\\n+ \\\"0 years 0 mons {} days {} hours {} mins {}.{:02} secs\\\",\\n+ days_parts,\\n+ hours,\\n+ mins,\\n+ secs,\\n+ (milliseconds_part % 1000),\\n+ )\\n+ };\\n+\\n+ s\\n+ }};\\n+}\\ndiff --git a/rust/cubesql/src/compile/engine/df/mod.rs b/rust/cubesql/src/compile/engine/df/mod.rs\\nindex a19a970..3097523 100644\\n--- a/rust/cubesql/src/compile/engine/df/mod.rs\\n+++ b/rust/cubesql/src/compile/engine/df/mod.rs\\n@@ -1 +1,2 @@\\n pub mod coerce;\\n+pub mod intervals;\\ndiff --git a/rust/cubesql/src/compile/engine/udf.rs b/rust/cubesql/src/compile/engine/udf.rs\\nindex 55b8bc1..0e160b3 100644\\n--- a/rust/cubesql/src/compile/engine/udf.rs\\n+++ b/rust/cubesql/src/compile/engine/udf.rs\\n@@ -1,14 +1,19 @@\\n use std::any::type_name;\\n use std::sync::Arc;\\n \\n+\\n use datafusion::{\\n arrow::{\\n array::{\\n ArrayRef, BooleanArray, BooleanBuilder, GenericStringArray, Int32Builder,\\n- PrimitiveArray, StringBuilder, UInt32Builder,\\n+ IntervalDayTimeBuilder, PrimitiveArray, StringBuilder,\\n+ UInt32Builder,\\n },\\n compute::cast,\\n- datatypes::{DataType, Int64Type},\\n+ datatypes::{\\n+ DataType, Int64Type, IntervalUnit, TimeUnit,\\n+ TimestampNanosecondType,\\n+ },\\n },\\n error::DataFusionError,\\n logical_plan::create_udf,\\n@@ -399,3 +404,63 @@ pub fn create_convert_tz_udf() -> ScalarUDF {\\n &fun,\\n )\\n }\\n+\\n+pub fn create_timediff_udf() -> ScalarUDF {\\n+ let fun = make_scalar_function(move |args: &[ArrayRef]| {\\n+ assert!(args.len() == 2);\\n+\\n+ let left_dt = &args[0];\\n+ let right_dt = &args[1];\\n+\\n+ let left_date = match left_dt.data_type() {\\n+ DataType::Timestamp(TimeUnit::Nanosecond, _) => {\\n+ let arr = downcast_primitive_arg!(left_dt, \\\"left_dt\\\", TimestampNanosecondType);\\n+ let ts = arr.value(0);\\n+\\n+ // NaiveDateTime::from_timestamp(ts, 0)\\n+ ts\\n+ }\\n+ _ => {\\n+ return Err(DataFusionError::Execution(format!(\\n+ \\\"left_dt argument must be a Timestamp, actual: {}\\\",\\n+ left_dt.data_type()\\n+ )));\\n+ }\\n+ };\\n+\\n+ let right_date = match right_dt.data_type() {\\n+ DataType::Timestamp(TimeUnit::Nanosecond, _) => {\\n+ let arr = downcast_primitive_arg!(right_dt, \\\"right_dt\\\", TimestampNanosecondType);\\n+ arr.value(0)\\n+ }\\n+ _ => {\\n+ return Err(DataFusionError::Execution(format!(\\n+ \\\"right_dt argument must be a Timestamp, actual: {}\\\",\\n+ right_dt.data_type()\\n+ )));\\n+ }\\n+ };\\n+\\n+ let diff = right_date - left_date;\\n+ if diff != 0 {\\n+ return Err(DataFusionError::NotImplemented(format!(\\n+ \\\"timediff is not implemented, it's stub\\\"\\n+ )));\\n+ }\\n+\\n+ let mut interal_arr = IntervalDayTimeBuilder::new(1);\\n+ interal_arr.append_value(diff)?;\\n+\\n+ Ok(Arc::new(interal_arr.finish()) as ArrayRef)\\n+ });\\n+\\n+ let return_type: ReturnTypeFunction =\\n+ Arc::new(move |_| Ok(Arc::new(DataType::Interval(IntervalUnit::DayTime))));\\n+\\n+ ScalarUDF::new(\\n+ \\\"timediff\\\",\\n+ &Signature::any(2, Volatility::Immutable),\\n+ &return_type,\\n+ &fun,\\n+ )\\n+}\\ndiff --git a/rust/cubesql/src/compile/mod.rs b/rust/cubesql/src/compile/mod.rs\\nindex a88da57..6121aa0 100644\\n--- a/rust/cubesql/src/compile/mod.rs\\n+++ b/rust/cubesql/src/compile/mod.rs\\n@@ -32,8 +32,8 @@ use self::engine::context::SystemVar;\\n use self::engine::provider::CubeContext;\\n use self::engine::udf::{\\n create_connection_id_udf, create_convert_tz_udf, create_current_user_udf, create_db_udf,\\n- create_if_udf, create_instr_udf, create_isnull_udf, create_least_udf, create_user_udf,\\n- create_version_udf,\\n+ create_if_udf, create_instr_udf, create_isnull_udf, create_least_udf, create_timediff_udf,\\n+ create_user_udf, create_version_udf,\\n };\\n use self::parser::parse_sql_to_statement;\\n \\n@@ -1450,6 +1450,7 @@ impl QueryPlanner {\\n ctx.register_udf(create_if_udf());\\n ctx.register_udf(create_least_udf());\\n ctx.register_udf(create_convert_tz_udf());\\n+ ctx.register_udf(create_timediff_udf());\\n \\n let state = ctx.state.lock().unwrap().clone();\\n let cube_ctx = CubeContext::new(&state, &self.context.cubes);\\n@@ -3226,6 +3227,25 @@ mod tests {\\n }\\n \\n #[tokio::test]\\n+ async fn test_timediff() -> Result<(), CubeError> {\\n+ assert_eq!(\\n+ execute_df_query(\\n+ \\\"select \\\\\\n+ timediff('1994-11-26T13:25:00.000Z'::timestamp, '1994-11-26T13:25:00.000Z'::timestamp) as r1\\n+ \\\".to_string()\\n+ )\\n+ .await?,\\n+ \\\"+------------------------------------------------+\\\\n\\\\\\n+ | r1 |\\\\n\\\\\\n+ +------------------------------------------------+\\\\n\\\\\\n+ | 0 years 0 mons 0 days 0 hours 0 mins 0.00 secs |\\\\n\\\\\\n+ +------------------------------------------------+\\\"\\n+ );\\n+\\n+ Ok(())\\n+ }\\n+\\n+ #[tokio::test]\\n async fn test_metabase() -> Result<(), CubeError> {\\n assert_eq!(\\n execute_df_query(\\ndiff --git a/rust/cubesql/src/mysql/dataframe.rs b/rust/cubesql/src/mysql/dataframe.rs\\nindex fa246aa..2443458 100644\\n--- a/rust/cubesql/src/mysql/dataframe.rs\\n+++ b/rust/cubesql/src/mysql/dataframe.rs\\n@@ -3,9 +3,10 @@ use std::fmt::{self, Debug, Formatter};\\n use chrono::{SecondsFormat, TimeZone, Utc};\\n use comfy_table::{Cell, Table};\\n use datafusion::arrow::array::{\\n- Array, Float64Array, Int32Array, Int64Array, StringArray, TimestampMicrosecondArray,\\n- UInt32Array,\\n+ Array, Float64Array, Int32Array, Int64Array, IntervalDayTimeArray, IntervalYearMonthArray,\\n+ StringArray, TimestampMicrosecondArray, UInt32Array,\\n };\\n+use datafusion::arrow::datatypes::IntervalUnit;\\n use datafusion::arrow::{\\n array::{BooleanArray, TimestampNanosecondArray, UInt64Array},\\n datatypes::{DataType, TimeUnit},\\n@@ -15,6 +16,7 @@ use log::{error, warn};\\n use msql_srv::{ColumnFlags, ColumnType};\\n \\n use crate::{compile::builder::CompiledQueryFieldMeta, CubeError};\\n+use crate::{make_string_interval_day_time, make_string_interval_year_month};\\n \\n #[derive(Clone, Debug)]\\n pub struct Column {\\n@@ -309,6 +311,7 @@ pub fn arrow_to_column_type(arrow_type: DataType) -> Result Ok(ColumnType::MYSQL_TYPE_BLOB),\\n DataType::Utf8 | DataType::LargeUtf8 => Ok(ColumnType::MYSQL_TYPE_STRING),\\n DataType::Timestamp(_, _) => Ok(ColumnType::MYSQL_TYPE_STRING),\\n+ DataType::Interval(_) => Ok(ColumnType::MYSQL_TYPE_STRING),\\n DataType::Float16 | DataType::Float64 => Ok(ColumnType::MYSQL_TYPE_DOUBLE),\\n DataType::Boolean => Ok(ColumnType::MYSQL_TYPE_TINY),\\n DataType::Int8\\n@@ -402,6 +405,24 @@ pub fn batch_to_dataframe(batches: &Vec) -> Result {\\n+ let a = array\\n+ .as_any()\\n+ .downcast_ref::()\\n+ .unwrap();\\n+ for i in 0..num_rows {\\n+ rows[i].push(TableValue::String(make_string_interval_day_time!(a, i)));\\n+ }\\n+ }\\n+ DataType::Interval(IntervalUnit::YearMonth) => {\\n+ let a = array\\n+ .as_any()\\n+ .downcast_ref::()\\n+ .unwrap();\\n+ for i in 0..num_rows {\\n+ rows[i].push(TableValue::String(make_string_interval_year_month!(a, i)));\\n+ }\\n+ }\\n DataType::Boolean => {\\n let a = array.as_any().downcast_ref::().unwrap();\\n for i in 0..num_rows {\\n\", \"diff --git a/ionic/components/nav/test/basic/index.ts b/ionic/components/nav/test/basic/index.ts\\nindex 4b1a8ea..2834f68 100644\\n--- a/ionic/components/nav/test/basic/index.ts\\n+++ b/ionic/components/nav/test/basic/index.ts\\n@@ -63,12 +63,6 @@ class FirstPage {\\n }\\n }\\n \\n- onPageDidEnter() {\\n- setTimeout(() => {\\n- this.nav.push(PrimaryHeaderPage);\\n- }, 1000);\\n- }\\n-\\n setPages() {\\n let items = [\\n PrimaryHeaderPage\\n\", \"diff --git a/Cargo.lock b/Cargo.lock\\nindex f949506..6a10219 100644\\n--- a/Cargo.lock\\n+++ b/Cargo.lock\\n@@ -94,7 +94,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"els\\\"\\n-version = \\\"0.1.22\\\"\\n+version = \\\"0.1.23-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_compiler\\\",\\n@@ -105,7 +105,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"els\\\",\\n \\\"erg_common\\\",\\n@@ -115,7 +115,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_common\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"backtrace-on-stack-overflow\\\",\\n \\\"crossterm\\\",\\n@@ -126,7 +126,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_compiler\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"erg_parser\\\",\\n@@ -134,7 +134,7 @@ dependencies = [\\n \\n [[package]]\\n name = \\\"erg_parser\\\"\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n dependencies = [\\n \\\"erg_common\\\",\\n \\\"unicode-xid\\\",\\ndiff --git a/Cargo.toml b/Cargo.toml\\nindex 04fdad7..ecc45e5 100644\\n--- a/Cargo.toml\\n+++ b/Cargo.toml\\n@@ -20,7 +20,7 @@ members = [\\n ]\\n \\n [workspace.package]\\n-version = \\\"0.6.10\\\"\\n+version = \\\"0.6.11-nightly.0\\\"\\n authors = [\\\"erg-lang team \\\"]\\n license = \\\"MIT OR Apache-2.0\\\"\\n edition = \\\"2021\\\"\\n@@ -64,10 +64,10 @@ full-repl = [\\\"erg_common/full-repl\\\"]\\n full = [\\\"els\\\", \\\"full-repl\\\", \\\"unicode\\\", \\\"pretty\\\"]\\n \\n [workspace.dependencies]\\n-erg_common = { version = \\\"0.6.10\\\", path = \\\"./crates/erg_common\\\" }\\n-erg_parser = { version = \\\"0.6.10\\\", path = \\\"./crates/erg_parser\\\" }\\n-erg_compiler = { version = \\\"0.6.10\\\", path = \\\"./crates/erg_compiler\\\" }\\n-els = { version = \\\"0.1.22\\\", path = \\\"./crates/els\\\" }\\n+erg_common = { version = \\\"0.6.11-nightly.0\\\", path = \\\"./crates/erg_common\\\" }\\n+erg_parser = { version = \\\"0.6.11-nightly.0\\\", path = \\\"./crates/erg_parser\\\" }\\n+erg_compiler = { version = \\\"0.6.11-nightly.0\\\", path = \\\"./crates/erg_compiler\\\" }\\n+els = { version = \\\"0.1.23-nightly.0\\\", path = \\\"./crates/els\\\" }\\n \\n [dependencies]\\n erg_common = { workspace = true }\\ndiff --git a/crates/els/Cargo.toml b/crates/els/Cargo.toml\\nindex bc031e6..7c9455f 100644\\n--- a/crates/els/Cargo.toml\\n+++ b/crates/els/Cargo.toml\\n@@ -2,7 +2,7 @@\\n name = \\\"els\\\"\\n description = \\\"An Erg compiler frontend for IDEs, implements LSP.\\\"\\n documentation = \\\"http://docs.rs/els\\\"\\n-version = \\\"0.1.22\\\"\\n+version = \\\"0.1.23-nightly.0\\\"\\n authors.workspace = true\\n license.workspace = true\\n edition.workspace = true\\n\", \"diff --git a/src/background/audio-manager.ts b/src/background/audio-manager.ts\\nindex 54e8b24..11c5fba 100644\\n--- a/src/background/audio-manager.ts\\n+++ b/src/background/audio-manager.ts\\n@@ -2,7 +2,7 @@\\n * To make sure only one audio plays at a time\\n */\\n \\n-import { timeout } from '../_helpers/promise-more'\\n+import { timeout } from '@/_helpers/promise-more'\\n \\n declare global {\\n interface Window {\\ndiff --git a/src/background/context-menus.ts b/src/background/context-menus.ts\\nindex 994b59e..7036362 100644\\n--- a/src/background/context-menus.ts\\n+++ b/src/background/context-menus.ts\\n@@ -1,5 +1,5 @@\\n-import { storage, openURL } from '../_helpers/browser-api'\\n-import { AppConfig } from '../app-config'\\n+import { storage, openURL } from '@/_helpers/browser-api'\\n+import { AppConfig } from '@/app-config'\\n \\n import { Observable } from 'rxjs/Observable'\\n import { fromPromise } from 'rxjs/observable/fromPromise'\\ndiff --git a/src/background/initialization.ts b/src/background/initialization.ts\\nindex 0e5b3ad..001ee73 100644\\n--- a/src/background/initialization.ts\\n+++ b/src/background/initialization.ts\\n@@ -1,6 +1,6 @@\\n-import { storage, openURL } from '../_helpers/browser-api'\\n-import checkUpdate from '../_helpers/check-update'\\n-import { AppConfig } from '../app-config'\\n+import { storage, openURL } from '@/_helpers/browser-api'\\n+import checkUpdate from '@/_helpers/check-update'\\n+import { AppConfig } from '@/app-config'\\n import { mergeConfig } from './merge-config'\\n import { init as initMenus } from './context-menus'\\n import { init as initPdf } from './pdf-sniffer'\\ndiff --git a/src/background/merge-config.ts b/src/background/merge-config.ts\\nindex afa1800..afdbd63 100644\\n--- a/src/background/merge-config.ts\\n+++ b/src/background/merge-config.ts\\n@@ -1,4 +1,4 @@\\n-import { appConfigFactory, AppConfig } from '../app-config'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n import _ from 'lodash'\\n \\n /**\\n@@ -24,7 +24,7 @@ function initConfig (): Promise {\\n const storageObj = { config: appConfigFactory() }\\n \\n Object.keys(storageObj.config.dicts.all).forEach(id => {\\n- storageObj[id] = require('../components/dictionaries/' + id + '/config')\\n+ storageObj[id] = require('@/components/dictionaries/' + id + '/config')\\n })\\n \\n return browser.storage.sync.set(storageObj)\\n@@ -70,7 +70,7 @@ function mergeHistorical (config): Promise {\\n \\n const storageObj = { config: base }\\n Object.keys(base.dicts.all).forEach(id => {\\n- storageObj[id] = config.dicts.all[id] || require('../components/dictionaries/' + id + '/config')\\n+ storageObj[id] = config.dicts.all[id] || require('@/components/dictionaries/' + id + '/config')\\n })\\n \\n return browser.storage.sync.set(storageObj)\\ndiff --git a/src/background/pdf-sniffer.ts b/src/background/pdf-sniffer.ts\\nindex 6ba27cf..70aa38f 100644\\n--- a/src/background/pdf-sniffer.ts\\n+++ b/src/background/pdf-sniffer.ts\\n@@ -2,8 +2,8 @@\\n * Open pdf link directly\\n */\\n \\n-import { storage } from '../_helpers/browser-api'\\n-import { AppConfig } from '../app-config'\\n+import { storage } from '@/_helpers/browser-api'\\n+import { AppConfig } from '@/app-config'\\n \\n export function init (pdfSniff: boolean) {\\n if (browser.webRequest.onBeforeRequest.hasListener(otherPdfListener)) {\\ndiff --git a/src/background/server.ts b/src/background/server.ts\\nindex 73b34b6..66ed5c0 100644\\n--- a/src/background/server.ts\\n+++ b/src/background/server.ts\\n@@ -1,7 +1,7 @@\\n-import { DictID } from '../app-config'\\n-import { message, openURL } from '../_helpers/browser-api'\\n+import { DictID } from '@/app-config'\\n+import { message, openURL } from '@/_helpers/browser-api'\\n import { play } from './audio-manager'\\n-import { chsToChz } from '../_helpers/chs-to-chz'\\n+import { chsToChz } from '@/_helpers/chs-to-chz'\\n \\n interface MessageOpenUrlWithEscape {\\n type: 'OPEN_URL'\\n@@ -63,7 +63,7 @@ function fetchDictResult (data: MessageFetchDictResult): Promise {\\n let search\\n \\n try {\\n- search = require('../components/dictionaries/' + data.dict + '/engine.js')\\n+ search = require('@/components/dictionaries/' + data.dict + '/engine.js')\\n } catch (err) {\\n return Promise.reject(err)\\n }\\ndiff --git a/test/unit/_helpers/browser-api.spec.ts b/test/unit/_helpers/browser-api.spec.ts\\nindex 1f39145..e327169 100644\\n--- a/test/unit/_helpers/browser-api.spec.ts\\n+++ b/test/unit/_helpers/browser-api.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { message, storage, openURL } from '../../../src/_helpers/browser-api'\\n+import { message, storage, openURL } from '@/_helpers/browser-api'\\n \\n beforeEach(() => {\\n browser.flush()\\ndiff --git a/test/unit/_helpers/check-update.spec.ts b/test/unit/_helpers/check-update.spec.ts\\nindex 2abfc57..fd0b678 100644\\n--- a/test/unit/_helpers/check-update.spec.ts\\n+++ b/test/unit/_helpers/check-update.spec.ts\\n@@ -1,4 +1,4 @@\\n-import checkUpdate from '../../../src/_helpers/check-update'\\n+import checkUpdate from '@/_helpers/check-update'\\n import fetchMock from 'jest-fetch-mock'\\n \\n describe('Check Update', () => {\\ndiff --git a/test/unit/_helpers/chs-to-chz.spec.ts b/test/unit/_helpers/chs-to-chz.spec.ts\\nindex 295c6ad..21d5229 100644\\n--- a/test/unit/_helpers/chs-to-chz.spec.ts\\n+++ b/test/unit/_helpers/chs-to-chz.spec.ts\\n@@ -1,4 +1,4 @@\\n-import chsToChz from '../../../src/_helpers/chs-to-chz'\\n+import chsToChz from '@/_helpers/chs-to-chz'\\n \\n describe('Chs to Chz', () => {\\n it('should convert chs to chz', () => {\\ndiff --git a/test/unit/_helpers/fetch-dom.spec.ts b/test/unit/_helpers/fetch-dom.spec.ts\\nindex a79dda0..bbfbf10 100644\\n--- a/test/unit/_helpers/fetch-dom.spec.ts\\n+++ b/test/unit/_helpers/fetch-dom.spec.ts\\n@@ -1,4 +1,4 @@\\n-import fetchDom from '../../../src/_helpers/fetch-dom'\\n+import fetchDom from '@/_helpers/fetch-dom'\\n \\n class XMLHttpRequestMock {\\n static queue: XMLHttpRequestMock[] = []\\ndiff --git a/test/unit/_helpers/lang-check.spec.ts b/test/unit/_helpers/lang-check.spec.ts\\nindex f3e668a..09f30bb 100644\\n--- a/test/unit/_helpers/lang-check.spec.ts\\n+++ b/test/unit/_helpers/lang-check.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { isContainChinese, isContainEnglish } from '../../../src/_helpers/lang-check'\\n+import { isContainChinese, isContainEnglish } from '@/_helpers/lang-check'\\n \\n describe('Language Check', () => {\\n it('isContainChinese should return ture if text contains Chinese', () => {\\ndiff --git a/test/unit/_helpers/promise-more.spec.ts b/test/unit/_helpers/promise-more.spec.ts\\nindex 9601c7d..66dc8d9 100644\\n--- a/test/unit/_helpers/promise-more.spec.ts\\n+++ b/test/unit/_helpers/promise-more.spec.ts\\n@@ -1,4 +1,4 @@\\n-import * as pm from '../../../src/_helpers/promise-more'\\n+import * as pm from '@/_helpers/promise-more'\\n \\n describe('Promise More', () => {\\n beforeAll(() => {\\ndiff --git a/test/unit/_helpers/selection.spec.ts b/test/unit/_helpers/selection.spec.ts\\nindex 370239a..06812cf 100644\\n--- a/test/unit/_helpers/selection.spec.ts\\n+++ b/test/unit/_helpers/selection.spec.ts\\n@@ -1,4 +1,4 @@\\n-import selection from '../../../src/_helpers/selection'\\n+import selection from '@/_helpers/selection'\\n \\n describe('Selection', () => {\\n const bakSelection = window.getSelection\\ndiff --git a/test/unit/_helpers/strip-script.spec.ts b/test/unit/_helpers/strip-script.spec.ts\\nindex cce558f..355b382 100644\\n--- a/test/unit/_helpers/strip-script.spec.ts\\n+++ b/test/unit/_helpers/strip-script.spec.ts\\n@@ -1,4 +1,4 @@\\n-import stripScript from '../../../src/_helpers/strip-script'\\n+import stripScript from '@/_helpers/strip-script'\\n \\n describe('Strip Script', () => {\\n const expectedEl = document.createElement('div') as HTMLDivElement\\ndiff --git a/test/unit/background/audio-manager.spec.ts b/test/unit/background/audio-manager.spec.ts\\nindex b0096a6..b1266d7 100644\\n--- a/test/unit/background/audio-manager.spec.ts\\n+++ b/test/unit/background/audio-manager.spec.ts\\n@@ -1,4 +1,4 @@\\n-import audio from '../../../src/background/audio-manager'\\n+import audio from '@/background/audio-manager'\\n \\n describe('Audio Manager', () => {\\n const bakAudio = (window as any).Audio\\ndiff --git a/test/unit/background/context-menus.spec.ts b/test/unit/background/context-menus.spec.ts\\nindex 39e249c..d9049dc 100644\\n--- a/test/unit/background/context-menus.spec.ts\\n+++ b/test/unit/background/context-menus.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n import sinon from 'sinon'\\n \\n function specialConfig () {\\n@@ -11,7 +11,7 @@ describe('Context Menus', () => {\\n beforeAll(() => {\\n browser.flush()\\n jest.resetModules()\\n- require('../../../src/background/context-menus')\\n+ require('@/background/context-menus')\\n })\\n afterAll(() => browser.flush())\\n \\n@@ -93,7 +93,7 @@ describe('Context Menus', () => {\\n browser.contextMenus.create.callsFake((_, cb) => cb())\\n config = specialConfig()\\n jest.resetModules()\\n- const { init } = require('../../../src/background/context-menus')\\n+ const { init } = require('@/background/context-menus')\\n init(config.contextMenus)\\n })\\n \\n@@ -110,7 +110,7 @@ describe('Context Menus', () => {\\n it('should not init setup when called multiple times', () => {\\n expect(browser.contextMenus.removeAll.calledOnce).toBeTruthy()\\n \\n- const { init } = require('../../../src/background/context-menus')\\n+ const { init } = require('@/background/context-menus')\\n init(config.contextMenus)\\n init(config.contextMenus)\\n \\ndiff --git a/test/unit/background/initialization.spec.ts b/test/unit/background/initialization.spec.ts\\nindex 7bc0972..56a6389 100644\\n--- a/test/unit/background/initialization.spec.ts\\n+++ b/test/unit/background/initialization.spec.ts\\n@@ -1,4 +1,4 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n import fetchMock from 'jest-fetch-mock'\\n import sinon from 'sinon'\\n \\n@@ -11,12 +11,12 @@ describe('Initialization', () => {\\n const checkUpdate = jest.fn().mockReturnValue(Promise.resolve())\\n \\n beforeAll(() => {\\n- const { message, storage } = require('../../../src/_helpers/browser-api')\\n+ const { message, storage } = require('@/_helpers/browser-api')\\n window.fetch = fetchMock\\n \\n browser.flush()\\n jest.resetModules()\\n- jest.doMock('../../../src/background/merge-config', () => {\\n+ jest.doMock('@/background/merge-config', () => {\\n return {\\n mergeConfig (config) {\\n mergeConfig(config)\\n@@ -24,16 +24,16 @@ describe('Initialization', () => {\\n }\\n }\\n })\\n- jest.doMock('../../../src/background/context-menus', () => {\\n+ jest.doMock('@/background/context-menus', () => {\\n return { init: initMenus }\\n })\\n- jest.doMock('../../../src/background/pdf-sniffer', () => {\\n+ jest.doMock('@/background/pdf-sniffer', () => {\\n return { init: initPdf }\\n })\\n- jest.doMock('../../../src/_helpers/check-update', () => {\\n+ jest.doMock('@/_helpers/check-update', () => {\\n return checkUpdate\\n })\\n- jest.doMock('../../../src/_helpers/browser-api', () => {\\n+ jest.doMock('@/_helpers/browser-api', () => {\\n return {\\n message,\\n storage,\\n@@ -41,13 +41,13 @@ describe('Initialization', () => {\\n }\\n })\\n \\n- require('../../../src/background/initialization')\\n+ require('@/background/initialization')\\n })\\n afterAll(() => {\\n browser.flush()\\n- jest.dontMock('../../../src/background/merge-config')\\n- jest.dontMock('../../../src/background/context-menus')\\n- jest.dontMock('../../../src/_helpers/browser-api')\\n+ jest.dontMock('@/background/merge-config')\\n+ jest.dontMock('@/background/context-menus')\\n+ jest.dontMock('@/_helpers/browser-api')\\n window.fetch = bakFetch\\n })\\n \\ndiff --git a/test/unit/background/merge-config.spec.ts b/test/unit/background/merge-config.spec.ts\\nindex 73c047d..c0dce26 100644\\n--- a/test/unit/background/merge-config.spec.ts\\n+++ b/test/unit/background/merge-config.spec.ts\\n@@ -1,5 +1,5 @@\\n-import { appConfigFactory, AppConfig, AppConfigMutable } from '../../../src/app-config'\\n-import mergeConfig from '../../../src/background/merge-config'\\n+import { appConfigFactory, AppConfig, AppConfigMutable } from '@/app-config'\\n+import mergeConfig from '@/background/merge-config'\\n import sinon from 'sinon'\\n \\n describe('Merge Config', () => {\\ndiff --git a/test/unit/background/pdf-sniffer.spec.ts b/test/unit/background/pdf-sniffer.spec.ts\\nindex a0219d2..bb7726f 100644\\n--- a/test/unit/background/pdf-sniffer.spec.ts\\n+++ b/test/unit/background/pdf-sniffer.spec.ts\\n@@ -1,5 +1,5 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n-import { init as initPdf } from '../../../src/background/pdf-sniffer'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n+import { init as initPdf } from '@/background/pdf-sniffer'\\n import sinon from 'sinon'\\n \\n function hasListenerPatch (fn) {\\ndiff --git a/test/unit/background/server.spec.ts b/test/unit/background/server.spec.ts\\nindex b8ef065..aa04525 100644\\n--- a/test/unit/background/server.spec.ts\\n+++ b/test/unit/background/server.spec.ts\\n@@ -1,5 +1,5 @@\\n-import { appConfigFactory, AppConfig } from '../../../src/app-config'\\n-import * as browserWrap from '../../../src/_helpers/browser-api'\\n+import { appConfigFactory, AppConfig } from '@/app-config'\\n+import * as browserWrap from '@/_helpers/browser-api'\\n import sinon from 'sinon'\\n \\n describe('Server', () => {\\n@@ -13,26 +13,26 @@ describe('Server', () => {\\n browserWrap.openURL = openURL\\n \\n beforeAll(() => {\\n- jest.doMock('../../../src/_helpers/chs-to-chz', () => {\\n+ jest.doMock('@/_helpers/chs-to-chz', () => {\\n return { chsToChz }\\n })\\n- jest.doMock('../../../src/background/audio-manager', () => {\\n+ jest.doMock('@/background/audio-manager', () => {\\n return { play }\\n })\\n- jest.doMock('../../../src/_helpers/browser-api', () => {\\n+ jest.doMock('@/_helpers/browser-api', () => {\\n return browserWrap\\n })\\n- jest.doMock('../../../src/components/dictionaries/bing/engine.js', () => {\\n+ jest.doMock('@/components/dictionaries/bing/engine.js', () => {\\n return bingSearch\\n })\\n })\\n \\n afterAll(() => {\\n browser.flush()\\n- jest.dontMock('../../../src/_helpers/chs-to-chz')\\n- jest.dontMock('../../../src/background/audio-manager')\\n- jest.dontMock('../../../src/_helpers/browser-api')\\n- jest.dontMock('../../../src/components/dictionaries/bing/engine.js')\\n+ jest.dontMock('@/_helpers/chs-to-chz')\\n+ jest.dontMock('@/background/audio-manager')\\n+ jest.dontMock('@/_helpers/browser-api')\\n+ jest.dontMock('@/components/dictionaries/bing/engine.js')\\n })\\n \\n beforeEach(() => {\\n@@ -46,7 +46,7 @@ describe('Server', () => {\\n bingSearch.mockReset()\\n bingSearch.mockImplementation(() => Promise.resolve())\\n jest.resetModules()\\n- require('../../../src/background/server')\\n+ require('@/background/server')\\n })\\n \\n it('should properly init', () => {\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"29dfb9716298c5a579c0ffba6742e13a29325670\", \"cd9e6a2ab17c5961b0f977bb8a06f8545da49a97\", \"607ecc92b5f8c084304e406eec725b7dcfa0a562\", \"8246d024f21d93cc092e19bede5f7b3a5325c8dc\"]"},"types":{"kind":"string","value":"[\"feat\", \"test\", \"build\", \"refactor\"]"}}},{"rowIdx":1074,"cells":{"commit_message":{"kind":"string","value":"coordinator accepts a request transformer instead of a list of operations\n\nThe request transformer can generate the operations from the current topology. This helps to\n- ensure that the operations are generated based on the latest topology. When concurrent changes\n happens, coordinator can detect it. Previously it was unclear because by the time handle apply\n operations, the cluster topology might have changed.\n- return the simulated final topology as part of the result,remove unnecessary import,exclude github.io from link checking to avoid rate limiting,change min checked results for score calculation"},"diff":{"kind":"string","value":"[\"diff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\\nindex 8bb5c3d..f8f5e24 100644\\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinator.java\\n@@ -10,6 +10,7 @@ package io.camunda.zeebe.topology.changes;\\n import io.camunda.zeebe.scheduler.future.ActorFuture;\\n import io.camunda.zeebe.topology.state.ClusterTopology;\\n import io.camunda.zeebe.topology.state.TopologyChangeOperation;\\n+import io.camunda.zeebe.util.Either;\\n import java.util.List;\\n \\n public interface TopologyChangeCoordinator {\\n@@ -39,4 +40,16 @@ public interface TopologyChangeCoordinator {\\n ActorFuture hasCompletedChanges(final long version);\\n \\n ActorFuture getCurrentTopology();\\n+\\n+ ActorFuture applyOperations(TopologyChangeRequest request);\\n+\\n+ record TopologyChangeResult(\\n+ ClusterTopology currentTopology,\\n+ ClusterTopology finalTopology,\\n+ List operations) {}\\n+\\n+ interface TopologyChangeRequest {\\n+ Either> operations(\\n+ final ClusterTopology currentTopology);\\n+ }\\n }\\ndiff --git a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\\nindex 13ec754..877fc3c 100644\\n--- a/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\\n+++ b/topology/src/main/java/io/camunda/zeebe/topology/changes/TopologyChangeCoordinatorImpl.java\\n@@ -103,6 +103,62 @@ public class TopologyChangeCoordinatorImpl implements TopologyChangeCoordinator \\n return clusterTopologyManager.getClusterTopology();\\n }\\n \\n+ @Override\\n+ public ActorFuture applyOperations(final TopologyChangeRequest request) {\\n+ final ActorFuture future = executor.createFuture();\\n+ clusterTopologyManager\\n+ .getClusterTopology()\\n+ .onComplete(\\n+ (currentClusterTopology, errorOnGettingTopology) -> {\\n+ if (errorOnGettingTopology != null) {\\n+ future.completeExceptionally(errorOnGettingTopology);\\n+ return;\\n+ }\\n+\\n+ final var operationsEither = request.operations(currentClusterTopology);\\n+ if (operationsEither.isLeft()) {\\n+ future.completeExceptionally(operationsEither.getLeft());\\n+ return;\\n+ }\\n+ final var operations = operationsEither.get();\\n+ if (operations.isEmpty()) {\\n+ // No operations to apply\\n+ future.complete(\\n+ new TopologyChangeResult(\\n+ currentClusterTopology, currentClusterTopology, operations));\\n+ return;\\n+ }\\n+\\n+ final ActorFuture validation =\\n+ validateTopologyChangeRequest(currentClusterTopology, operations);\\n+\\n+ validation.onComplete(\\n+ (simulatedFinalTopology, validationError) -> {\\n+ if (validationError != null) {\\n+ future.completeExceptionally(validationError);\\n+ return;\\n+ }\\n+\\n+ // if the validation was successful, apply the changes\\n+ final ActorFuture applyFuture = executor.createFuture();\\n+ applyTopologyChange(\\n+ operations, currentClusterTopology, simulatedFinalTopology, applyFuture);\\n+\\n+ applyFuture.onComplete(\\n+ (ignore, error) -> {\\n+ if (error == null) {\\n+ future.complete(\\n+ new TopologyChangeResult(\\n+ currentClusterTopology, simulatedFinalTopology, operations));\\n+ } else {\\n+ future.completeExceptionally(error);\\n+ }\\n+ });\\n+ });\\n+ });\\n+ return future;\\n+ }\\n+\\n private ActorFuture validateTopologyChangeRequest(\\n final ClusterTopology currentClusterTopology,\\n final List operations) {\\n\", \"diff --git a/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java b/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\\nindex 14c6f30..ebaef60 100644\\n--- a/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\\n+++ b/transport/src/main/java/io/camunda/zeebe/transport/stream/impl/LogicalId.java\\n@@ -8,7 +8,6 @@\\n package io.camunda.zeebe.transport.stream.impl;\\n \\n import io.camunda.zeebe.util.buffer.BufferUtil;\\n-import org.agrona.BitUtil;\\n import org.agrona.concurrent.UnsafeBuffer;\\n \\n /**\\n\", \"diff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\\nindex 90c5a27..db6457b 100644\\n--- a/.github/workflows/ibis-docs-lint.yml\\n+++ b/.github/workflows/ibis-docs-lint.yml\\n@@ -101,6 +101,7 @@ jobs:\\n --exclude-mail \\\\\\n --exclude fonts.gstatic.com \\\\\\n --exclude github.com \\\\\\n+ --exclude github.io \\\\\\n --no-progress \\\\\\n --github-token ${{ steps.generate_token.outputs.token }}\\n \\n\", \"diff --git a/server/src/services/courseService.ts b/server/src/services/courseService.ts\\nindex 89633f4..10bfc55 100644\\n--- a/server/src/services/courseService.ts\\n+++ b/server/src/services/courseService.ts\\n@@ -580,8 +580,7 @@ export async function getTaskSolutionCheckers(courseTaskId: number, minCheckedCo\\n .createQueryBuilder('tsr')\\n .select('tsr.\\\"studentId\\\", ROUND(AVG(tsr.score)) as \\\"score\\\"')\\n .where(qb => {\\n- // query students with 3 checked tasks\\n-\\n+ // query students who checked enough tasks\\n const query = qb\\n .subQuery()\\n .select('r.\\\"checkerId\\\"')\\n@@ -600,7 +599,7 @@ export async function getTaskSolutionCheckers(courseTaskId: number, minCheckedCo\\n })\\n .andWhere('tsr.\\\"courseTaskId\\\" = :courseTaskId', { courseTaskId })\\n .groupBy('tsr.\\\"studentId\\\"')\\n- .having(`COUNT(tsr.id) >= :count`, { count: minCheckedCount })\\n+ .having(`COUNT(tsr.id) >= :count`, { count: minCheckedCount - 1 })\\n .getRawMany();\\n \\n return records.map(record => ({ studentId: record.studentId, score: Number(record.score) }));\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"dec860436916ef216998f80f8b2f9c39d00c064d\", \"84529bcb10c6fe02e2c0079d069ab6c6ac7683d6\", \"ce0539a32b927a3559feebf8f5307e3863e992a1\", \"fd849bd08363df60dbc8b9b6d55bac4f5ace88f4\"]"},"types":{"kind":"string","value":"[\"feat\", \"refactor\", \"ci\", \"docs\"]"}}},{"rowIdx":1075,"cells":{"commit_message":{"kind":"string","value":"disable edit/delete if primary key missing\n\nSigned-off-by: Pranav C ,change tests to depend on BrokerContext,fix `get-deploy-tags.sh`,verify property exist in row object\n\nSigned-off-by: Pranav C "},"diff":{"kind":"string","value":"[\"diff --git a/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue b/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\\nindex 5f9841f..c414c8c 100644\\n--- a/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\\n+++ b/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\\n@@ -413,6 +413,9 @@ export default {\\n \\n await this.reload()\\n } else if (Object.keys(updatedObj).length) {\\n+ if (!id) {\\n+ return this.$toast.info('Update not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n await this.api.update(id, updatedObj, this.oldRow)\\n } else {\\n return this.$toast.info('No columns to update').goAway(3000)\\ndiff --git a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\nindex c2b4b81..1b9d6a0 100644\\n--- a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\n+++ b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\n@@ -62,7 +62,15 @@\\n \\n \\n \\n-\\n+ \\n+ \\n+ Update & Delete not allowed since the table doesn't have any primary key\\n+ \\n+ \\n \\n \\n \\n@@ -208,6 +216,7 @@\\n :meta=\\\"meta\\\"\\n :is-virtual=\\\"selectedView.type === 'vtable'\\\"\\n :api=\\\"api\\\"\\n+ :is-pk-avail=\\\"isPkAvail\\\"\\n @onNewColCreation=\\\"onNewColCreation\\\"\\n @onCellValueChange=\\\"onCellValueChange\\\"\\n @insertNewRow=\\\"insertNewRow\\\"\\n@@ -631,8 +640,8 @@ export default {\\n if (\\n !this.meta || (\\n (this.meta.hasMany && this.meta.hasMany.length) ||\\n- (this.meta.manyToMany && this.meta.manyToMany.length) ||\\n- (this.meta.belongsTo && this.meta.belongsTo.length))\\n+ (this.meta.manyToMany && this.meta.manyToMany.length) ||\\n+ (this.meta.belongsTo && this.meta.belongsTo.length))\\n ) {\\n return this.$toast.info('Please delete relations before deleting table.').goAway(3000)\\n }\\n@@ -817,6 +826,10 @@ export default {\\n \\n const id = this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\\n \\n+ if (!id) {\\n+ return this.$toast.info('Update not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n+\\n const newData = await this.api.update(id, {\\n [column._cn]: rowObj[column._cn]\\n }, { [column._cn]: oldRow[column._cn] })\\n@@ -841,6 +854,11 @@ export default {\\n const rowObj = this.rowContextMenu.row\\n if (!this.rowContextMenu.rowMeta.new) {\\n const id = this.meta && this.meta.columns && this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\\n+\\n+ if (!id) {\\n+ return this.$toast.info('Delete not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n+\\n await this.api.delete(id)\\n }\\n this.data.splice(this.rowContextMenu.index, 1)\\n@@ -859,6 +877,11 @@ export default {\\n }\\n if (!rowMeta.new) {\\n const id = this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\\n+\\n+ if (!id) {\\n+ return this.$toast.info('Delete not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n+\\n await this.api.delete(id)\\n }\\n this.data.splice(row, 1)\\n@@ -991,6 +1014,9 @@ export default {\\n }\\n },\\n computed: {\\n+ isPkAvail() {\\n+ return this.meta && this.meta.columns.some(c => c.pk)\\n+ },\\n isGallery() {\\n return this.selectedView && this.selectedView.show_as === 'gallery'\\n },\\ndiff --git a/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue b/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\\nindex 5497d05..c198784 100644\\n--- a/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\\n+++ b/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\\n@@ -27,7 +27,7 @@\\n @xcresized=\\\"resizingCol = null\\\"\\n >\\n \\n+-->\\n \\n \\n@@ -162,7 +162,8 @@\\n \\n \\n \\n \\n-
\\n+ \\n \\n \\n- \\n+ \\n
\\n \\n \\n \\n@@ -411,6 +417,7 @@ export default {\\n parentId: this.parentId,\\n is_group: true,\\n status: 'update',\\n+ logical_op: 'and',\\n });\\n this.filters = this.filters.slice();\\n const index = this.filters.length - 1;\\n@@ -478,4 +485,8 @@ export default {\\n column-gap: 6px;\\n row-gap: 6px;\\n }\\n+\\n+.nc-filter-value-select {\\n+ min-width: 100px;\\n+}\\n \\n\", \"diff --git a/bootstrap/scripts/publish-patch.sh b/bootstrap/scripts/publish-patch.sh\\nindex a1b6f12..0d849a5 100755\\n--- a/bootstrap/scripts/publish-patch.sh\\n+++ b/bootstrap/scripts/publish-patch.sh\\n@@ -5,4 +5,4 @@ lerna version patch\\n lerna publish from-package -y\\n git push\\n \\n-./pack_and_install.sh\\n\\\\ No newline at end of file\\n+./bootstrap/scripts/pack_and_install.sh\\n\\\\ No newline at end of file\\n\", \"diff --git a/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java b/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java\\nindex b2dfb98..21eaf6d 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/api/BackupManager.java\\n@@ -42,4 +42,6 @@ public interface BackupManager {\\n \\n /** Close Backup manager */\\n ActorFuture closeAsync();\\n+\\n+ void failInProgressBackup(long lastCheckpointId);\\n }\\ndiff --git a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java\\nindex a1e1319..33149ae 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupService.java\\n@@ -16,6 +16,7 @@ import io.camunda.zeebe.scheduler.future.ActorFuture;\\n import io.camunda.zeebe.scheduler.future.CompletableActorFuture;\\n import io.camunda.zeebe.snapshots.PersistedSnapshotStore;\\n import java.nio.file.Path;\\n+import java.util.List;\\n import java.util.function.Predicate;\\n import org.slf4j.Logger;\\n import org.slf4j.LoggerFactory;\\n@@ -31,11 +32,13 @@ public final class BackupService extends Actor implements BackupManager {\\n private final PersistedSnapshotStore snapshotStore;\\n private final Path segmentsDirectory;\\n private final Predicate isSegmentsFile;\\n+ private List partitionMembers;\\n \\n public BackupService(\\n final int nodeId,\\n final int partitionId,\\n final int numberOfPartitions,\\n+ final List partitionMembers,\\n final PersistedSnapshotStore snapshotStore,\\n final Predicate isSegmentsFile,\\n final Path segmentsDirectory) {\\n@@ -48,6 +51,7 @@ public final class BackupService extends Actor implements BackupManager {\\n snapshotStore,\\n segmentsDirectory,\\n isSegmentsFile);\\n+ this.partitionMembers = partitionMembers;\\n }\\n \\n public BackupService(\\n@@ -122,6 +126,12 @@ public final class BackupService extends Actor implements BackupManager {\\n new UnsupportedOperationException(\\\"Not implemented\\\"));\\n }\\n \\n+ @Override\\n+ public void failInProgressBackup(final long lastCheckpointId) {\\n+ internalBackupManager.failInProgressBackups(\\n+ partitionId, lastCheckpointId, partitionMembers, actor);\\n+ }\\n+\\n private BackupIdentifierImpl getBackupId(final long checkpointId) {\\n return new BackupIdentifierImpl(nodeId, partitionId, checkpointId);\\n }\\ndiff --git a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java\\nindex e462dd5..f6d76b6 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/management/BackupServiceImpl.java\\n@@ -9,16 +9,23 @@ package io.camunda.zeebe.backup.management;\\n \\n import io.camunda.zeebe.backup.api.BackupIdentifier;\\n import io.camunda.zeebe.backup.api.BackupStatus;\\n+import io.camunda.zeebe.backup.api.BackupStatusCode;\\n import io.camunda.zeebe.backup.api.BackupStore;\\n+import io.camunda.zeebe.backup.common.BackupIdentifierImpl;\\n+import io.camunda.zeebe.backup.processing.state.CheckpointState;\\n import io.camunda.zeebe.scheduler.ConcurrencyControl;\\n import io.camunda.zeebe.scheduler.future.ActorFuture;\\n import io.camunda.zeebe.scheduler.future.CompletableActorFuture;\\n+import java.util.Collection;\\n import java.util.HashSet;\\n import java.util.Set;\\n import java.util.function.BiConsumer;\\n import java.util.function.Consumer;\\n+import org.slf4j.Logger;\\n+import org.slf4j.LoggerFactory;\\n \\n final class BackupServiceImpl {\\n+ private static final Logger LOG = LoggerFactory.getLogger(BackupServiceImpl.class);\\n private final Set backupsInProgress = new HashSet<>();\\n private final BackupStore backupStore;\\n private ConcurrencyControl concurrencyControl;\\n@@ -138,4 +145,48 @@ final class BackupServiceImpl {\\n }));\\n return future;\\n }\\n+\\n+ void failInProgressBackups(\\n+ final int partitionId,\\n+ final long lastCheckpointId,\\n+ final Collection brokers,\\n+ final ConcurrencyControl executor) {\\n+ if (lastCheckpointId != CheckpointState.NO_CHECKPOINT) {\\n+ executor.run(\\n+ () -> {\\n+ final var backupIds =\\n+ brokers.stream()\\n+ .map(b -> new BackupIdentifierImpl(b, partitionId, lastCheckpointId))\\n+ .toList();\\n+ // Fail backups initiated by previous leaders\\n+ backupIds.forEach(this::failInProgressBackup);\\n+ });\\n+ }\\n+ }\\n+\\n+ private void failInProgressBackup(final BackupIdentifier backupId) {\\n+ backupStore\\n+ .getStatus(backupId)\\n+ .thenAccept(\\n+ status -> {\\n+ if (status.statusCode() == BackupStatusCode.IN_PROGRESS) {\\n+ LOG.debug(\\n+ \\\"The backup {} initiated by previous leader is still in progress. Marking it as failed.\\\",\\n+ backupId);\\n+ backupStore\\n+ .markFailed(backupId)\\n+ .thenAccept(ignore -> LOG.trace(\\\"Marked backup {} as failed.\\\", backupId))\\n+ .exceptionally(\\n+ failed -> {\\n+ LOG.debug(\\\"Failed to mark backup {} as failed\\\", backupId, failed);\\n+ return null;\\n+ });\\n+ }\\n+ })\\n+ .exceptionally(\\n+ error -> {\\n+ LOG.debug(\\\"Failed to retrieve status of backup {}\\\", backupId);\\n+ return null;\\n+ });\\n+ }\\n }\\ndiff --git a/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java b/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java\\nindex c83fdc1..2899d4d 100644\\n--- a/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java\\n+++ b/backup/src/main/java/io/camunda/zeebe/backup/processing/CheckpointRecordsProcessor.java\\n@@ -14,20 +14,24 @@ import io.camunda.zeebe.backup.processing.state.DbCheckpointState;\\n import io.camunda.zeebe.engine.api.ProcessingResult;\\n import io.camunda.zeebe.engine.api.ProcessingResultBuilder;\\n import io.camunda.zeebe.engine.api.ProcessingScheduleService;\\n+import io.camunda.zeebe.engine.api.ReadonlyStreamProcessorContext;\\n import io.camunda.zeebe.engine.api.RecordProcessor;\\n import io.camunda.zeebe.engine.api.RecordProcessorContext;\\n+import io.camunda.zeebe.engine.api.StreamProcessorLifecycleAware;\\n import io.camunda.zeebe.engine.api.TypedRecord;\\n import io.camunda.zeebe.protocol.impl.record.value.management.CheckpointRecord;\\n import io.camunda.zeebe.protocol.record.ValueType;\\n import io.camunda.zeebe.protocol.record.intent.management.CheckpointIntent;\\n import java.time.Duration;\\n+import java.util.List;\\n import java.util.Set;\\n import java.util.concurrent.CopyOnWriteArraySet;\\n import org.slf4j.Logger;\\n import org.slf4j.LoggerFactory;\\n \\n /** Process and replays records related to Checkpoint. */\\n-public final class CheckpointRecordsProcessor implements RecordProcessor {\\n+public final class CheckpointRecordsProcessor\\n+ implements RecordProcessor, StreamProcessorLifecycleAware {\\n \\n private static final Logger LOG = LoggerFactory.getLogger(CheckpointRecordsProcessor.class);\\n \\n@@ -62,6 +66,8 @@ public final class CheckpointRecordsProcessor implements RecordProcessor {\\n checkpointListeners.forEach(\\n listener -> listener.onNewCheckpointCreated(checkpointState.getCheckpointId()));\\n }\\n+\\n+ recordProcessorContext.addLifecycleListeners(List.of(this));\\n }\\n \\n @Override\\n@@ -126,4 +132,12 @@ public final class CheckpointRecordsProcessor implements RecordProcessor {\\n });\\n }\\n }\\n+\\n+ @Override\\n+ public void onRecovered(final ReadonlyStreamProcessorContext context) {\\n+ // After a leader change, the new leader will not continue taking the backup initiated by\\n+ // previous leader. So mark them as failed, so that the users do not wait forever for it to be\\n+ // completed.\\n+ backupManager.failInProgressBackup(checkpointState.getCheckpointId());\\n+ }\\n }\\ndiff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java\\nindex 3424e19..591e17b 100644\\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java\\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/steps/BackupServiceTransitionStep.java\\n@@ -7,6 +7,7 @@\\n */\\n package io.camunda.zeebe.broker.system.partitions.impl.steps;\\n \\n+import io.atomix.cluster.MemberId;\\n import io.atomix.raft.RaftServer.Role;\\n import io.camunda.zeebe.backup.api.BackupManager;\\n import io.camunda.zeebe.backup.management.BackupService;\\n@@ -17,6 +18,7 @@ import io.camunda.zeebe.journal.file.SegmentFile;\\n import io.camunda.zeebe.scheduler.future.ActorFuture;\\n import io.camunda.zeebe.scheduler.future.CompletableActorFuture;\\n import java.nio.file.Path;\\n+import java.util.List;\\n import java.util.function.Predicate;\\n \\n public final class BackupServiceTransitionStep implements PartitionTransitionStep {\\n@@ -69,6 +71,7 @@ public final class BackupServiceTransitionStep implements PartitionTransitionSte\\n context.getNodeId(),\\n context.getPartitionId(),\\n context.getBrokerCfg().getCluster().getPartitionsCount(),\\n+ getPartitionMembers(context),\\n context.getPersistedSnapshotStore(),\\n isSegmentsFile,\\n context.getRaftPartition().dataDirectory().toPath());\\n@@ -90,4 +93,12 @@ public final class BackupServiceTransitionStep implements PartitionTransitionSte\\n });\\n return installed;\\n }\\n+\\n+ // Brokers which are members of this partition's replication group\\n+ private static List getPartitionMembers(final PartitionTransitionContext context) {\\n+ return context.getRaftPartition().members().stream()\\n+ .map(MemberId::id)\\n+ .map(Integer::parseInt)\\n+ .toList();\\n+ }\\n }\\n\", \"diff --git a/docs/pages/4.react-native-web.md b/docs/pages/4.react-native-web.md\\nindex 69e4e52..8d6ae2a 100644\\n--- a/docs/pages/4.react-native-web.md\\n+++ b/docs/pages/4.react-native-web.md\\n@@ -16,6 +16,63 @@ To install `react-native-web`, run:\\n yarn add react-native-web react-dom react-art\\n ```\\n \\n+### Using CRA ([Create React App](https://github.com/facebook/create-react-app))\\n+\\n+Install [`react-app-rewired`](https://github.com/timarney/react-app-rewired) to override `webpack` configuration:\\n+\\n+```sh\\n+yarn add --dev react-app-rewired\\n+```\\n+\\n+[Configure `babel-loader`](#2-configure-babel-loader) using a new file called `config-overrides.js`:\\n+\\n+```js\\n+module.exports = function override(config, env) {\\n+ config.module.rules.push({\\n+ test: /\\\\.js$/,\\n+ exclude: /node_modules[/\\\\\\\\](?!react-native-paper|react-native-vector-icons|react-native-safe-area-view)/,\\n+ use: {\\n+ loader: \\\"babel-loader\\\",\\n+ options: {\\n+ // Disable reading babel configuration\\n+ babelrc: false,\\n+ configFile: false,\\n+\\n+ // The configration for compilation\\n+ presets: [\\n+ [\\\"@babel/preset-env\\\", { useBuiltIns: \\\"usage\\\" }],\\n+ \\\"@babel/preset-react\\\",\\n+ \\\"@babel/preset-flow\\\"\\n+ ],\\n+ plugins: [\\n+ \\\"@babel/plugin-proposal-class-properties\\\",\\n+ \\\"@babel/plugin-proposal-object-rest-spread\\\"\\n+ ]\\n+ }\\n+ }\\n+ });\\n+\\n+ return config;\\n+};\\n+```\\n+\\n+Change your script in `package.json`:\\n+\\n+```diff\\n+/* package.json */\\n+\\n+ \\\"scripts\\\": {\\n+- \\\"start\\\": \\\"react-scripts start\\\",\\n++ \\\"start\\\": \\\"react-app-rewired start\\\",\\n+- \\\"build\\\": \\\"react-scripts build\\\",\\n++ \\\"build\\\": \\\"react-app-rewired build\\\",\\n+- \\\"test\\\": \\\"react-scripts test --env=jsdom\\\",\\n++ \\\"test\\\": \\\"react-app-rewired test --env=jsdom\\\"\\n+}\\n+```\\n+\\n+### Custom webpack setup\\n+\\n To install `webpack`, run:\\n \\n ```sh\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"4f86f2570b274c45605cc59d9adb38f7ed30cd17\", \"3fcfb20b0feb371b357edc42fcb7c87085c9b82a\", \"fb83ef33b699fd966486a922ba1ade4cf8e55858\", \"ee7cc5d5a940fba774e715b1f029c6361110b108\"]"},"types":{"kind":"string","value":"[\"refactor\", \"build\", \"feat\", \"docs\"]"}}},{"rowIdx":1084,"cells":{"commit_message":{"kind":"string","value":"abort parallel stages if one failed,licensing,fix pagination spacing,extract lambdas"},"diff":{"kind":"string","value":"[\"diff --git a/Jenkinsfile b/Jenkinsfile\\nindex 168f446..a4da961 100644\\n--- a/Jenkinsfile\\n+++ b/Jenkinsfile\\n@@ -28,6 +28,7 @@ pipeline {\\n }\\n \\n stage('Verify') {\\n+ failFast true\\n parallel {\\n stage('Tests') {\\n steps {\\n\", \"diff --git a/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java b/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\\nindex a4aee6b..bb523fa 100644\\n--- a/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\\n+++ b/atomix/cluster/src/test/java/io/atomix/cluster/messaging/impl/NettyMessagingServiceTlsTest.java\\n@@ -1,3 +1,18 @@\\n+/*\\n+ * Copyright \\u00a9 2020 camunda services GmbH (info@camunda.com)\\n+ *\\n+ * Licensed under the Apache License, Version 2.0 (the \\\"License\\\");\\n+ * you may not use this file except in compliance with the License.\\n+ * You may obtain a copy of the License at\\n+ *\\n+ * http://www.apache.org/licenses/LICENSE-2.0\\n+ *\\n+ * Unless required by applicable law or agreed to in writing, software\\n+ * distributed under the License is distributed on an \\\"AS IS\\\" BASIS,\\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\\n+ * See the License for the specific language governing permissions and\\n+ * limitations under the License.\\n+ */\\n package io.atomix.cluster.messaging.impl;\\n \\n import static org.assertj.core.api.Assertions.assertThat;\\n\", \"diff --git a/website/layouts/Base.tsx b/website/layouts/Base.tsx\\nindex 22d36a2..40f7130 100644\\n--- a/website/layouts/Base.tsx\\n+++ b/website/layouts/Base.tsx\\n@@ -399,7 +399,7 @@ export function Base({ children, headings }: BaseProps) {\\n >\\n \\n \\n-
\\n+
\\n \\n Previous\\n \\n@@ -418,7 +418,7 @@ export function Base({ children, headings }: BaseProps) {\\n aria-label={`Go to ${next.resource?.label}`}\\n >\\n \\n-
\\n+
\\n \\n Next\\n \\n\", \"diff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\nindex 6ee5797..bcfcc72 100644\\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n@@ -224,7 +224,6 @@ public final class AsyncSnapshotDirector extends Actor\\n private void takeSnapshot() {\\n final var transientSnapshotFuture =\\n stateController.takeTransientSnapshot(lowerBoundSnapshotPosition);\\n-\\n transientSnapshotFuture.onComplete(\\n (optionalTransientSnapshot, snapshotTakenError) -> {\\n if (snapshotTakenError != null) {\\n@@ -237,27 +236,31 @@ public final class AsyncSnapshotDirector extends Actor\\n takingSnapshot = false;\\n return;\\n }\\n- pendingSnapshot = optionalTransientSnapshot.get();\\n- onRecovered();\\n-\\n- final ActorFuture lastWrittenPosition =\\n- streamProcessor.getLastWrittenPositionAsync();\\n- actor.runOnCompletion(\\n- lastWrittenPosition,\\n- (endPosition, error) -> {\\n- if (error == null) {\\n- LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\\n- lastWrittenEventPosition = endPosition;\\n- persistingSnapshot = false;\\n- persistSnapshotIfLastWrittenPositionCommitted();\\n- } else {\\n- resetStateOnFailure();\\n- LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\\n- }\\n- });\\n+ onTransientSnapshotTaken(optionalTransientSnapshot.get());\\n });\\n }\\n \\n+ private void onTransientSnapshotTaken(final TransientSnapshot transientSnapshot) {\\n+\\n+ pendingSnapshot = transientSnapshot;\\n+ onRecovered();\\n+\\n+ final ActorFuture lastWrittenPosition = streamProcessor.getLastWrittenPositionAsync();\\n+ actor.runOnCompletion(lastWrittenPosition, this::onLastWrittenPositionReceived);\\n+ }\\n+\\n+ private void onLastWrittenPositionReceived(final Long endPosition, final Throwable error) {\\n+ if (error == null) {\\n+ LOG.info(LOG_MSG_WAIT_UNTIL_COMMITTED, endPosition, commitPosition);\\n+ lastWrittenEventPosition = endPosition;\\n+ persistingSnapshot = false;\\n+ persistSnapshotIfLastWrittenPositionCommitted();\\n+ } else {\\n+ resetStateOnFailure();\\n+ LOG.error(ERROR_MSG_ON_RESOLVE_WRITTEN_POS, error);\\n+ }\\n+ }\\n+\\n private void onRecovered() {\\n if (healthStatus != HealthStatus.HEALTHY) {\\n healthStatus = HealthStatus.HEALTHY;\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"28e623b294816c4e070971782a75c8697a11966f\", \"cbe62140ce219da84772e21e7cfb4b5c2a25c1b8\", \"1e05a24486f15889ddf6bf1c711ea2bbffc1a88e\", \"14abf5c31523a551134aebe9e8f3505ef26ed421\"]"},"types":{"kind":"string","value":"[\"ci\", \"docs\", \"fix\", \"refactor\"]"}}},{"rowIdx":1085,"cells":{"commit_message":{"kind":"string","value":"remove unused,only restart if pages directory itself is changed\n\nresolves #429,await job creation to ensure asserted event sequence,new ShowDebug parameter\n\ncalculate each segment timing\nnew parameter to show/hide segment debug information\nset-poshprompt updated with the new showDebug parameter\n\nForce disabled segment to be visible for debug purpose"},"diff":{"kind":"string","value":"[\"diff --git a/src/content/redux/modules/dictionaries.ts b/src/content/redux/modules/dictionaries.ts\\nindex 88f7215..570d397 100644\\n--- a/src/content/redux/modules/dictionaries.ts\\n+++ b/src/content/redux/modules/dictionaries.ts\\n@@ -3,7 +3,6 @@ import { DictID, appConfigFactory, AppConfig } from '@/app-config'\\n import isEqual from 'lodash/isEqual'\\n import { saveWord } from '@/_helpers/record-manager'\\n import { getDefaultSelectionInfo, SelectionInfo, isSameSelection } from '@/_helpers/selection'\\n-import { createActiveConfigStream } from '@/_helpers/config-manager'\\n import { isContainChinese, isContainEnglish, testerPunct, isContainMinor, testerChinese, testJapanese, testKorean } from '@/_helpers/lang-check'\\n import { MsgType, MsgFetchDictResult } from '@/typings/message'\\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\\ndiff --git a/src/content/redux/modules/widget.ts b/src/content/redux/modules/widget.ts\\nindex 53ad550..68e0a3d 100644\\n--- a/src/content/redux/modules/widget.ts\\n+++ b/src/content/redux/modules/widget.ts\\n@@ -1,9 +1,9 @@\\n import * as recordManager from '@/_helpers/record-manager'\\n import { StoreState, DispatcherThunk, Dispatcher } from './index'\\n-import appConfigFactory, { TCDirection, AppConfig, DictID } from '@/app-config'\\n+import appConfigFactory, { TCDirection, DictID } from '@/app-config'\\n import { message, storage } from '@/_helpers/browser-api'\\n-import { createActiveConfigStream, createConfigIDListStream } from '@/_helpers/config-manager'\\n-import { MsgSelection, MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\\n+import { createConfigIDListStream } from '@/_helpers/config-manager'\\n+import { MsgType, MsgTempDisabledState, MsgEditWord, MsgOpenUrl, MsgFetchDictResult } from '@/typings/message'\\n import { searchText, restoreDicts } from '@/content/redux/modules/dictionaries'\\n import { SelectionInfo, getDefaultSelectionInfo } from '@/_helpers/selection'\\n import { Mutable } from '@/typings/helpers'\\n\", \"diff --git a/packages/cli/src/commands/dev.ts b/packages/cli/src/commands/dev.ts\\nindex 35d859e..d6d91ed 100644\\n--- a/packages/cli/src/commands/dev.ts\\n+++ b/packages/cli/src/commands/dev.ts\\n@@ -1,4 +1,4 @@\\n-import { resolve } from 'upath'\\n+import { resolve, relative } from 'upath'\\n import chokidar from 'chokidar'\\n import debounce from 'debounce-promise'\\n import type { Nuxt } from '@nuxt/kit'\\n@@ -27,9 +27,9 @@ export default defineNuxtCommand({\\n const { loadNuxt, buildNuxt } = requireModule('@nuxt/kit', rootDir) as typeof import('@nuxt/kit')\\n \\n let currentNuxt: Nuxt\\n- const load = async (isRestart: boolean) => {\\n+ const load = async (isRestart: boolean, reason?: string) => {\\n try {\\n- const message = `${isRestart ? 'Restarting' : 'Starting'} nuxt...`\\n+ const message = `${reason ? reason + '. ' : ''}${isRestart ? 'Restarting' : 'Starting'} nuxt...`\\n server.setApp(createLoadingHandler(message))\\n if (isRestart) {\\n console.log(message)\\n@@ -59,12 +59,8 @@ export default defineNuxtCommand({\\n const dLoad = debounce(load, 250)\\n const watcher = chokidar.watch([rootDir], { ignoreInitial: true, depth: 1 })\\n watcher.on('all', (_event, file) => {\\n- // Ignore any changes to files within the Nuxt build directory\\n- if (file.includes(currentNuxt.options.buildDir)) {\\n- return\\n- }\\n- if (file.includes('nuxt.config') || file.includes('modules') || file.includes('pages')) {\\n- dLoad(true)\\n+ if (file.match(/nuxt\\\\.config\\\\.(js|ts|mjs|cjs)$|pages$/)) {\\n+ dLoad(true, `${relative(rootDir, file)} updated`)\\n }\\n })\\n \\n\", \"diff --git a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\nindex 9ffa1fa..4333db0 100644\\n--- a/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\n+++ b/engine/src/test/java/io/zeebe/engine/processor/workflow/boundary/BoundaryEventTest.java\\n@@ -114,12 +114,18 @@ public class BoundaryEventTest {\\n ENGINE.deployment().withXmlResource(MULTIPLE_SEQUENCE_FLOWS).deploy();\\n final long workflowInstanceKey = ENGINE.workflowInstance().ofBpmnProcessId(PROCESS_ID).create();\\n \\n- // when\\n RecordingExporter.timerRecords()\\n .withHandlerNodeId(\\\"timer\\\")\\n .withIntent(TimerIntent.CREATED)\\n .withWorkflowInstanceKey(workflowInstanceKey)\\n .getFirst();\\n+\\n+ RecordingExporter.jobRecords(JobIntent.CREATED)\\n+ .withType(\\\"type\\\")\\n+ .withWorkflowInstanceKey(workflowInstanceKey)\\n+ .getFirst();\\n+\\n+ // when\\n ENGINE.increaseTime(Duration.ofMinutes(1));\\n \\n // then\\n\", \"diff --git a/engine.go b/engine.go\\nindex 6cc1ff3..4617ceb 100644\\n--- a/engine.go\\n+++ b/engine.go\\n@@ -67,6 +67,9 @@ func (e *engine) renderText(text string) {\\n \\tprefix := e.activeSegment.getValue(Prefix, \\\" \\\")\\n \\tpostfix := e.activeSegment.getValue(Postfix, \\\" \\\")\\n \\te.renderer.write(e.activeSegment.Background, e.activeSegment.Foreground, fmt.Sprintf(\\\"%s%s%s\\\", prefix, text, postfix))\\n+\\tif *e.env.getArgs().Debug {\\n+\\t\\te.renderer.write(e.activeSegment.Background, e.activeSegment.Foreground, fmt.Sprintf(\\\"(%s:%s)\\\", e.activeSegment.Type, e.activeSegment.timing))\\n+\\t}\\n }\\n \\n func (e *engine) renderSegmentText(text string) {\\n@@ -107,13 +110,11 @@ func (e *engine) setStringValues(segments []*Segment) {\\n \\twg.Add(len(segments))\\n \\tdefer wg.Wait()\\n \\tcwd := e.env.getcwd()\\n+\\tdebug := *e.env.getArgs().Debug\\n \\tfor _, segment := range segments {\\n \\t\\tgo func(s *Segment) {\\n \\t\\t\\tdefer wg.Done()\\n-\\t\\t\\terr := s.mapSegmentWithWriter(e.env)\\n-\\t\\t\\tif err == nil && !s.hasValue(IgnoreFolders, cwd) && s.enabled() {\\n-\\t\\t\\t\\ts.stringValue = s.string()\\n-\\t\\t\\t}\\n+\\t\\t\\ts.setStringValue(e.env, cwd, debug)\\n \\t\\t}(segment)\\n \\t}\\n }\\ndiff --git a/main.go b/main.go\\nindex 56ae8a5..d67a640 100644\\n--- a/main.go\\n+++ b/main.go\\n@@ -14,6 +14,7 @@ type args struct {\\n \\tConfig *string\\n \\tShell *string\\n \\tPWD *string\\n+\\tDebug *bool\\n }\\n \\n func main() {\\n@@ -42,6 +43,10 @@ func main() {\\n \\t\\t\\t\\\"pwd\\\",\\n \\t\\t\\t\\\"\\\",\\n \\t\\t\\t\\\"the path you are working in\\\"),\\n+\\t\\tDebug: flag.Bool(\\n+\\t\\t\\t\\\"debug\\\",\\n+\\t\\t\\tfalse,\\n+\\t\\t\\t\\\"Print debug information\\\"),\\n \\t}\\n \\tflag.Parse()\\n \\tenv := &environment{\\ndiff --git a/packages/powershell/oh-my-posh/oh-my-posh.psm1 b/packages/powershell/oh-my-posh/oh-my-posh.psm1\\nindex 9234fc6..1450eb3 100644\\n--- a/packages/powershell/oh-my-posh/oh-my-posh.psm1\\n+++ b/packages/powershell/oh-my-posh/oh-my-posh.psm1\\n@@ -5,6 +5,7 @@\\n \\n $global:PoshSettings = New-Object -TypeName PSObject -Property @{\\n Theme = \\\"$PSScriptRoot\\\\themes\\\\jandedobbeleer.json\\\";\\n+ ShowDebug = $false\\n }\\n \\n function Get-PoshCommand {\\n@@ -36,9 +37,14 @@ function Set-PoshPrompt {\\n param(\\n [Parameter(Mandatory = $false)]\\n [string]\\n- $Theme\\n+ $Theme,\\n+ [Parameter(Mandatory = $false)]\\n+ [bool]\\n+ $ShowDebug = $false\\n )\\n \\n+ $global:PoshSettings.ShowDebug = $ShowDebug\\n+\\n if (Test-Path \\\"$PSScriptRoot/themes/$Theme.json\\\") {\\n $global:PoshSettings.Theme = \\\"$PSScriptRoot/themes/$Theme.json\\\"\\n }\\n@@ -68,8 +74,9 @@ function Set-PoshPrompt {\\n $startInfo = New-Object System.Diagnostics.ProcessStartInfo\\n $startInfo.FileName = Get-PoshCommand\\n $config = $global:PoshSettings.Theme\\n+ $showDebug = $global:PoshSettings.ShowDebug\\n $cleanPWD = $PWD.ProviderPath.TrimEnd(\\\"\\\\\\\")\\n- $startInfo.Arguments = \\\"-config=\\\"\\\"$config\\\"\\\" -error=$errorCode -pwd=\\\"\\\"$cleanPWD\\\"\\\"\\\"\\n+ $startInfo.Arguments = \\\"-debug=\\\"\\\"$showDebug\\\"\\\" -config=\\\"\\\"$config\\\"\\\" -error=$errorCode -pwd=\\\"\\\"$cleanPWD\\\"\\\"\\\"\\n $startInfo.Environment[\\\"TERM\\\"] = \\\"xterm-256color\\\"\\n $startInfo.CreateNoWindow = $true\\n $startInfo.StandardOutputEncoding = [System.Text.Encoding]::UTF8\\ndiff --git a/segment.go b/segment.go\\nindex 27dd416..4015dac 100644\\n--- a/segment.go\\n+++ b/segment.go\\n@@ -1,6 +1,9 @@\\n package main\\n \\n-import \\\"errors\\\"\\n+import (\\n+\\t\\\"errors\\\"\\n+\\t\\\"time\\\"\\n+)\\n \\n // Segment represent a single segment and it's configuration\\n type Segment struct {\\n@@ -17,6 +20,7 @@ type Segment struct {\\n \\twriter SegmentWriter\\n \\tstringValue string\\n \\tactive bool\\n+\\ttiming time.Duration\\n }\\n \\n // SegmentWriter is the interface used to define what and if to write to the prompt\\n@@ -149,3 +153,26 @@ func (segment *Segment) mapSegmentWithWriter(env environmentInfo) error {\\n \\t}\\n \\treturn errors.New(\\\"unable to map writer\\\")\\n }\\n+\\n+func (segment *Segment) setStringValue(env environmentInfo, cwd string, debug bool) {\\n+\\terr := segment.mapSegmentWithWriter(env)\\n+\\tif err != nil || segment.hasValue(IgnoreFolders, cwd) {\\n+\\t\\treturn\\n+\\t}\\n+\\t// add timing only in debug\\n+\\tif debug {\\n+\\t\\tstart := time.Now()\\n+\\t\\tdefer (func() {\\n+\\t\\t\\t// force segment rendering to display the time it took\\n+\\t\\t\\t// to check if the segment is enabled or not\\n+\\t\\t\\t// depending on the segement, calling enabled()\\n+\\t\\t\\t// can be time consuming\\n+\\t\\t\\tsegment.active = true\\n+\\t\\t\\telapsed := time.Since(start)\\n+\\t\\t\\tsegment.timing = elapsed\\n+\\t\\t})()\\n+\\t}\\n+\\tif segment.enabled() {\\n+\\t\\tsegment.stringValue = segment.string()\\n+\\t}\\n+}\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"a50b51999015e210918d9c8e95fd4cac347353be\", \"cbce777addb3dd118232a9f28db9d425d4c937b2\", \"a8d1a60fd48d3fbd76d4271987a1b0f538d498f1\", \"bea32587586ca08f390c901a95e9b9c25263f4df\"]"},"types":{"kind":"string","value":"[\"refactor\", \"fix\", \"test\", \"feat\"]"}}},{"rowIdx":1086,"cells":{"commit_message":{"kind":"string","value":"refactor generate_completion,add react ecosystem,add important to override paragraphs in items,only run JMH on develop and master\n\n- reduces test duration of feature branches\n- reduces bors merge duration\n- show performance impact of PRs and Releases on the respective branches\n- number parallel stages to enforce ordering in Blue Ocean view\n- **note**: skipping a parallel stage in Blue Ocean triggers a bug where\n log will not be show until the stage finished https://issues.jenkins-ci.org/browse/JENKINS-48879"},"diff":{"kind":"string","value":"[\"diff --git a/src/lib.rs b/src/lib.rs\\nindex dfd8014..15850f7 100644\\n--- a/src/lib.rs\\n+++ b/src/lib.rs\\n@@ -1,11 +1,106 @@\\n //! Generates [Nushell](https://github.com/nushell/nushell) completions for [`clap`](https://github.com/clap-rs/clap) based CLIs\\n \\n-use clap::Command;\\n+use clap::{Arg, Command};\\n use clap_complete::Generator;\\n \\n /// Generate Nushell complete file\\n pub struct Nushell;\\n \\n+enum Argument {\\n+ Short(char),\\n+ Long(String),\\n+ ShortAndLong(char, String),\\n+ Positional(String, bool),\\n+}\\n+\\n+struct ArgumentLine {\\n+ arg: Argument,\\n+ takes_values: bool,\\n+ help: Option,\\n+}\\n+\\n+impl From<&Arg> for ArgumentLine {\\n+ fn from(arg: &Arg) -> Self {\\n+ let takes_values = arg\\n+ .get_num_args()\\n+ .map(|v| v.takes_values())\\n+ .unwrap_or(false);\\n+\\n+ let help = arg.get_help().map(|s| s.to_string());\\n+\\n+ if arg.is_positional() {\\n+ let id = arg.get_id().to_string();\\n+ let required = arg.is_required_set();\\n+ let arg = Argument::Positional(id, required);\\n+\\n+ return Self {\\n+ arg,\\n+ takes_values,\\n+ help,\\n+ };\\n+ }\\n+\\n+ let short = arg.get_short();\\n+ let long = arg.get_long();\\n+\\n+ match short {\\n+ Some(short) => match long {\\n+ Some(long) => Self {\\n+ arg: Argument::ShortAndLong(short, long.into()),\\n+ takes_values,\\n+ help,\\n+ },\\n+ None => Self {\\n+ arg: Argument::Short(short),\\n+ takes_values,\\n+ help,\\n+ },\\n+ },\\n+ None => match long {\\n+ Some(long) => Self {\\n+ arg: Argument::Long(long.into()),\\n+ takes_values,\\n+ help,\\n+ },\\n+ None => unreachable!(\\\"No short or long option found\\\"),\\n+ },\\n+ }\\n+ }\\n+}\\n+\\n+impl ToString for ArgumentLine {\\n+ fn to_string(&self) -> String {\\n+ let mut s = String::new();\\n+\\n+ match &self.arg {\\n+ Argument::Short(short) => s.push_str(format!(\\\" -{}\\\", short).as_str()),\\n+ Argument::Long(long) => s.push_str(format!(\\\" --{}\\\", long).as_str()),\\n+ Argument::ShortAndLong(short, long) => {\\n+ s.push_str(format!(\\\" --{}(-{})\\\", long, short).as_str())\\n+ }\\n+ Argument::Positional(positional, required) => {\\n+ s.push_str(format!(\\\" {}\\\", positional).as_str());\\n+\\n+ if !*required {\\n+ s.push('?');\\n+ }\\n+ }\\n+ }\\n+\\n+ if self.takes_values {\\n+ s.push_str(\\\": string\\\");\\n+ }\\n+\\n+ if let Some(help) = &self.help {\\n+ s.push_str(format!(\\\"\\\\t# {}\\\", help).as_str());\\n+ }\\n+\\n+ s.push('\\\\n');\\n+\\n+ s\\n+ }\\n+}\\n+\\n impl Generator for Nushell {\\n fn file_name(&self, name: &str) -> String {\\n format!(\\\"{}.nu\\\", name)\\n@@ -37,51 +132,18 @@ fn generate_completion(completions: &mut String, cmd: &Command, is_subcommand: b\\n \\n let bin_name = cmd.get_bin_name().expect(\\\"Failed to get bin name\\\");\\n \\n- if is_subcommand {\\n- completions.push_str(format!(\\\" export extern \\\\\\\"{}\\\\\\\" [\\\\n\\\", bin_name).as_str());\\n+ let name = if is_subcommand {\\n+ format!(r#\\\"\\\"{}\\\"\\\"#, bin_name)\\n } else {\\n- completions.push_str(format!(\\\" export extern {} [\\\\n\\\", bin_name).as_str());\\n- }\\n+ bin_name.into()\\n+ };\\n \\n- let mut s = String::new();\\n- for arg in cmd.get_arguments() {\\n- if arg.is_positional() {\\n- s.push_str(format!(\\\" {}\\\", arg.get_id()).as_str());\\n- if !arg.is_required_set() {\\n- s.push('?');\\n- }\\n- }\\n-\\n- let long = arg.get_long();\\n- if let Some(opt) = long {\\n- s.push_str(format!(\\\" --{}\\\", opt).as_str());\\n- }\\n+ completions.push_str(format!(\\\" export extern {} [\\\\n\\\", name).as_str());\\n \\n- let short = arg.get_short();\\n- if let Some(opt) = short {\\n- if long.is_some() {\\n- s.push_str(format!(\\\"(-{})\\\", opt).as_str());\\n- } else {\\n- s.push_str(format!(\\\" -{}\\\", opt).as_str());\\n- }\\n- }\\n-\\n- if let Some(v) = arg.get_num_args() {\\n- if v.takes_values() {\\n- // TODO: add more types?\\n- // TODO: add possible values?\\n- s.push_str(\\\": string\\\");\\n- }\\n- }\\n-\\n- if let Some(msg) = arg.get_help() {\\n- if arg.is_positional() || long.is_some() || short.is_some() {\\n- s.push_str(format!(\\\"\\\\t# {}\\\", msg).as_str());\\n- }\\n- }\\n-\\n- s.push('\\\\n');\\n- }\\n+ let s: String = cmd\\n+ .get_arguments()\\n+ .map(|arg| ArgumentLine::from(arg).to_string())\\n+ .collect();\\n \\n completions.push_str(&s);\\n completions.push_str(\\\" ]\\\\n\\\\n\\\");\\n\", \"diff --git a/package.json b/package.json\\nindex 1ba8c4f..d1de9a0 100644\\n--- a/package.json\\n+++ b/package.json\\n@@ -36,14 +36,19 @@\\n \\\"@types/node\\\": \\\"^9.3.0\\\",\\n \\\"@types/react\\\": \\\"^16.0.34\\\",\\n \\\"@types/react-dom\\\": \\\"^16.0.3\\\",\\n+ \\\"@types/react-motion\\\": \\\"^0.0.25\\\",\\n \\\"bootstrap-sass\\\": \\\"^3.3.7\\\",\\n \\\"highcharts\\\": \\\"^6.0.4\\\",\\n \\\"html2canvas\\\": \\\"^1.0.0-alpha.9\\\",\\n+ \\\"immer\\\": \\\"^1.2.1\\\",\\n \\\"lodash\\\": \\\"^4.17.4\\\",\\n \\\"moment\\\": \\\"^2.20.1\\\",\\n \\\"normalize.css\\\": \\\"^8.0.0\\\",\\n- \\\"react\\\": \\\"^16.2.0\\\",\\n- \\\"react-dom\\\": \\\"^16.2.0\\\",\\n+ \\\"react\\\": \\\"^16.3.1\\\",\\n+ \\\"react-dom\\\": \\\"^16.3.1\\\",\\n+ \\\"react-motion\\\": \\\"^0.5.2\\\",\\n+ \\\"react-redux\\\": \\\"^5.0.7\\\",\\n+ \\\"redux\\\": \\\"^3.7.2\\\",\\n \\\"rxjs\\\": \\\"^5.5.6\\\",\\n \\\"vue\\\": \\\"^2.5.13\\\",\\n \\\"vue-plugin-webextension-i18n\\\": \\\"^0.1.0\\\",\\ndiff --git a/yarn.lock b/yarn.lock\\nindex c8898d8..5d0fc9f 100644\\n--- a/yarn.lock\\n+++ b/yarn.lock\\n@@ -187,6 +187,12 @@\\n \\\"@types/node\\\" \\\"*\\\"\\n \\\"@types/react\\\" \\\"*\\\"\\n \\n+\\\"@types/react-motion@^0.0.25\\\":\\n+ version \\\"0.0.25\\\"\\n+ resolved \\\"https://registry.npmjs.org/@types/react-motion/-/react-motion-0.0.25.tgz#2445745ee8e8e6149faa47a36ff6b0d4c21dbf94\\\"\\n+ dependencies:\\n+ \\\"@types/react\\\" \\\"*\\\"\\n+\\n \\\"@types/react@*\\\", \\\"@types/react@^16.0.34\\\":\\n version \\\"16.0.40\\\"\\n resolved \\\"https://registry.npmjs.org/@types/react/-/react-16.0.40.tgz#caabc2296886f40b67f6fc80f0f3464476461df9\\\"\\n@@ -3837,6 +3843,10 @@ hoek@4.x.x:\\n version \\\"4.2.1\\\"\\n resolved \\\"https://registry.npmjs.org/hoek/-/hoek-4.2.1.tgz#9634502aa12c445dd5a7c5734b572bb8738aacbb\\\"\\n \\n+hoist-non-react-statics@^2.5.0:\\n+ version \\\"2.5.0\\\"\\n+ resolved \\\"https://registry.npmjs.org/hoist-non-react-statics/-/hoist-non-react-statics-2.5.0.tgz#d2ca2dfc19c5a91c5a6615ce8e564ef0347e2a40\\\"\\n+\\n home-or-tmp@^2.0.0:\\n version \\\"2.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/home-or-tmp/-/home-or-tmp-2.0.0.tgz#e36c3f2d2cae7d746a857e38d18d5f32a7882db8\\\"\\n@@ -4004,6 +4014,10 @@ ignore@^3.3.5:\\n version \\\"3.3.7\\\"\\n resolved \\\"https://registry.npmjs.org/ignore/-/ignore-3.3.7.tgz#612289bfb3c220e186a58118618d5be8c1bab021\\\"\\n \\n+immer@^1.2.1:\\n+ version \\\"1.2.1\\\"\\n+ resolved \\\"https://registry.npmjs.org/immer/-/immer-1.2.1.tgz#96e2ae29cdfc428f28120b832701931b92fa597c\\\"\\n+\\n import-local@^1.0.0:\\n version \\\"1.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/import-local/-/import-local-1.0.0.tgz#5e4ffdc03f4fe6c009c6729beb29631c2f8227bc\\\"\\n@@ -4104,7 +4118,7 @@ interpret@^1.0.0:\\n version \\\"1.1.0\\\"\\n resolved \\\"https://registry.npmjs.org/interpret/-/interpret-1.1.0.tgz#7ed1b1410c6a0e0f78cf95d3b8440c63f78b8614\\\"\\n \\n-invariant@^2.2.2:\\n+invariant@^2.0.0, invariant@^2.2.2:\\n version \\\"2.2.4\\\"\\n resolved \\\"https://registry.npmjs.org/invariant/-/invariant-2.2.4.tgz#610f3c92c9359ce1db616e538008d23ff35158e6\\\"\\n dependencies:\\n@@ -5040,6 +5054,10 @@ locate-path@^2.0.0:\\n p-locate \\\"^2.0.0\\\"\\n path-exists \\\"^3.0.0\\\"\\n \\n+lodash-es@^4.17.5, lodash-es@^4.2.1:\\n+ version \\\"4.17.8\\\"\\n+ resolved \\\"https://registry.npmjs.org/lodash-es/-/lodash-es-4.17.8.tgz#6fa8c8c5d337481df0bdf1c0d899d42473121e45\\\"\\n+\\n lodash._reinterpolate@~3.0.0:\\n version \\\"3.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/lodash._reinterpolate/-/lodash._reinterpolate-3.0.0.tgz#0ccf2d89166af03b3663c796538b75ac6e114d9d\\\"\\n@@ -5149,7 +5167,7 @@ lodash@4.17.2:\\n version \\\"4.17.2\\\"\\n resolved \\\"https://registry.npmjs.org/lodash/-/lodash-4.17.2.tgz#34a3055babe04ce42467b607d700072c7ff6bf42\\\"\\n \\n-lodash@4.x, lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\\n+lodash@4.x, lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.17.5, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\\n version \\\"4.17.5\\\"\\n resolved \\\"https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz#99a92d65c0272debe8c96b6057bc8fbfa3bed511\\\"\\n \\n@@ -6467,7 +6485,7 @@ promise@^7.1.1:\\n dependencies:\\n asap \\\"~2.0.3\\\"\\n \\n-prop-types@^15.6.0:\\n+prop-types@^15.5.8, prop-types@^15.6.0:\\n version \\\"15.6.1\\\"\\n resolved \\\"https://registry.npmjs.org/prop-types/-/prop-types-15.6.1.tgz#36644453564255ddda391191fb3a125cbdf654ca\\\"\\n dependencies:\\n@@ -6574,7 +6592,7 @@ quick-lru@^1.0.0:\\n version \\\"1.1.0\\\"\\n resolved \\\"https://registry.npmjs.org/quick-lru/-/quick-lru-1.1.0.tgz#4360b17c61136ad38078397ff11416e186dcfbb8\\\"\\n \\n-raf@3.4.0:\\n+raf@3.4.0, raf@^3.1.0:\\n version \\\"3.4.0\\\"\\n resolved \\\"https://registry.npmjs.org/raf/-/raf-3.4.0.tgz#a28876881b4bc2ca9117d4138163ddb80f781575\\\"\\n dependencies:\\n@@ -6645,9 +6663,9 @@ react-dev-utils@^5.0.0:\\n strip-ansi \\\"3.0.1\\\"\\n text-table \\\"0.2.0\\\"\\n \\n-react-dom@^16.2.0:\\n- version \\\"16.2.0\\\"\\n- resolved \\\"https://registry.npmjs.org/react-dom/-/react-dom-16.2.0.tgz#69003178601c0ca19b709b33a83369fe6124c044\\\"\\n+react-dom@^16.3.1:\\n+ version \\\"16.3.1\\\"\\n+ resolved \\\"https://registry.npmjs.org/react-dom/-/react-dom-16.3.1.tgz#6a3c90a4fb62f915bdbcf6204422d93a7d4ca573\\\"\\n dependencies:\\n fbjs \\\"^0.8.16\\\"\\n loose-envify \\\"^1.1.0\\\"\\n@@ -6658,9 +6676,28 @@ react-error-overlay@^4.0.0:\\n version \\\"4.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/react-error-overlay/-/react-error-overlay-4.0.0.tgz#d198408a85b4070937a98667f500c832f86bd5d4\\\"\\n \\n-react@^16.2.0:\\n- version \\\"16.2.0\\\"\\n- resolved \\\"https://registry.npmjs.org/react/-/react-16.2.0.tgz#a31bd2dab89bff65d42134fa187f24d054c273ba\\\"\\n+react-motion@^0.5.2:\\n+ version \\\"0.5.2\\\"\\n+ resolved \\\"https://registry.npmjs.org/react-motion/-/react-motion-0.5.2.tgz#0dd3a69e411316567927917c6626551ba0607316\\\"\\n+ dependencies:\\n+ performance-now \\\"^0.2.0\\\"\\n+ prop-types \\\"^15.5.8\\\"\\n+ raf \\\"^3.1.0\\\"\\n+\\n+react-redux@^5.0.7:\\n+ version \\\"5.0.7\\\"\\n+ resolved \\\"https://registry.npmjs.org/react-redux/-/react-redux-5.0.7.tgz#0dc1076d9afb4670f993ffaef44b8f8c1155a4c8\\\"\\n+ dependencies:\\n+ hoist-non-react-statics \\\"^2.5.0\\\"\\n+ invariant \\\"^2.0.0\\\"\\n+ lodash \\\"^4.17.5\\\"\\n+ lodash-es \\\"^4.17.5\\\"\\n+ loose-envify \\\"^1.1.0\\\"\\n+ prop-types \\\"^15.6.0\\\"\\n+\\n+react@^16.3.1:\\n+ version \\\"16.3.1\\\"\\n+ resolved \\\"https://registry.npmjs.org/react/-/react-16.3.1.tgz#4a2da433d471251c69b6033ada30e2ed1202cfd8\\\"\\n dependencies:\\n fbjs \\\"^0.8.16\\\"\\n loose-envify \\\"^1.1.0\\\"\\n@@ -6788,6 +6825,15 @@ reduce-function-call@^1.0.1:\\n dependencies:\\n balanced-match \\\"^0.4.2\\\"\\n \\n+redux@^3.7.2:\\n+ version \\\"3.7.2\\\"\\n+ resolved \\\"https://registry.npmjs.org/redux/-/redux-3.7.2.tgz#06b73123215901d25d065be342eb026bc1c8537b\\\"\\n+ dependencies:\\n+ lodash \\\"^4.2.1\\\"\\n+ lodash-es \\\"^4.2.1\\\"\\n+ loose-envify \\\"^1.1.0\\\"\\n+ symbol-observable \\\"^1.0.3\\\"\\n+\\n regenerate@^1.2.1:\\n version \\\"1.3.3\\\"\\n resolved \\\"https://registry.npmjs.org/regenerate/-/regenerate-1.3.3.tgz#0c336d3980553d755c39b586ae3b20aa49c82b7f\\\"\\n@@ -7811,6 +7857,10 @@ symbol-observable@1.0.1:\\n version \\\"1.0.1\\\"\\n resolved \\\"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.0.1.tgz#8340fc4702c3122df5d22288f88283f513d3fdd4\\\"\\n \\n+symbol-observable@^1.0.3:\\n+ version \\\"1.2.0\\\"\\n+ resolved \\\"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.2.0.tgz#c22688aed4eab3cdc2dfeacbb561660560a00804\\\"\\n+\\n symbol-tree@^3.2.2:\\n version \\\"3.2.2\\\"\\n resolved \\\"https://registry.npmjs.org/symbol-tree/-/symbol-tree-3.2.2.tgz#ae27db38f660a7ae2e1c3b7d1bc290819b8519e6\\\"\\n\", \"diff --git a/packages/core/src/components/text/text.ios.scss b/packages/core/src/components/text/text.ios.scss\\nindex a3c58e2..2a020ab 100644\\n--- a/packages/core/src/components/text/text.ios.scss\\n+++ b/packages/core/src/components/text/text.ios.scss\\n@@ -9,8 +9,9 @@\\n @each $color-name, $color-base, $color-contrast in get-colors($colors-ios) {\\n \\n .text-ios-#{$color-name},\\n- .text-ios-#{$color-name} a {\\n- color: $color-base;\\n+ .text-ios-#{$color-name} a,\\n+ .text-ios-#{$color-name} p {\\n+ color: $color-base !important\\n }\\n \\n }\\ndiff --git a/packages/core/src/components/text/text.md.scss b/packages/core/src/components/text/text.md.scss\\nindex b397acb..050af1a 100644\\n--- a/packages/core/src/components/text/text.md.scss\\n+++ b/packages/core/src/components/text/text.md.scss\\n@@ -9,8 +9,9 @@\\n @each $color-name, $color-base, $color-contrast in get-colors($colors-md) {\\n \\n .text-md-#{$color-name},\\n- .text-md-#{$color-name} a {\\n- color: $color-base;\\n+ .text-md-#{$color-name} a,\\n+ .text-md-#{$color-name} p {\\n+ color: $color-base !important;\\n }\\n \\n }\\n\", \"diff --git a/Jenkinsfile b/Jenkinsfile\\nindex 9fbd3a1..3e4f052 100644\\n--- a/Jenkinsfile\\n+++ b/Jenkinsfile\\n@@ -29,7 +29,7 @@ pipeline {\\n \\n stage('Verify') {\\n parallel {\\n- stage('Tests') {\\n+ stage('1 - Java Tests') {\\n steps {\\n withMaven(jdk: jdkVersion, maven: mavenVersion, mavenSettingsConfig: mavenSettingsConfig) {\\n sh 'mvn -B verify -P skip-unstable-ci'\\n@@ -42,7 +42,9 @@ pipeline {\\n }\\n }\\n \\n- stage('JMH') {\\n+ stage('2 - JMH') {\\n+ // delete this line to also run JMH on feature branch\\n+ when { anyOf { branch 'master'; branch 'develop' } }\\n agent { node { label 'ubuntu-large' } }\\n \\n steps {\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"f1bc5a554af4e617c7d7508f7f16f8fd25c78c91\", \"7e04a5e829d7416e312ac342a00a11787745753b\", \"7ab363f7ba2807b3eb9895e47f4fcd058f43ae5e\", \"83934807f4061980e7f5bf46d36eca70e238835d\"]"},"types":{"kind":"string","value":"[\"refactor\", \"build\", \"test\", \"ci\"]"}}},{"rowIdx":1087,"cells":{"commit_message":{"kind":"string","value":"fix readme,updated test to use rows for action items\n\nreferences #279,do not check mkdocs for older versions used in deployments,update build"},"diff":{"kind":"string","value":"[\"diff --git a/crates/dagger-sdk/README.md b/crates/dagger-sdk/README.md\\nindex ed96be1..974fb7f 100644\\n--- a/crates/dagger-sdk/README.md\\n+++ b/crates/dagger-sdk/README.md\\n@@ -29,9 +29,9 @@ fn main() -> eyre::Result<()> {\\n let client = dagger_sdk::connect()?;\\n \\n let version = client\\n- .container(None)\\n- .from(\\\"golang:1.19\\\".into())\\n- .with_exec(vec![\\\"go\\\".into(), \\\"version\\\".into()], None)\\n+ .container()\\n+ .from(\\\"golang:1.19\\\")\\n+ .with_exec(vec![\\\"go\\\", \\\"version\\\"])\\n .stdout()?;\\n \\n println!(\\\"Hello from Dagger and {}\\\", version.trim());\\n\", \"diff --git a/ionic/components/card/test/advanced/main.html b/ionic/components/card/test/advanced/main.html\\nindex 7c56a7d..c19ea12 100644\\n--- a/ionic/components/card/test/advanced/main.html\\n+++ b/ionic/components/card/test/advanced/main.html\\n@@ -19,16 +19,20 @@\\n

\\n \\n \\n- \\n- \\n- \\n- \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n \\n \\n \\n@@ -51,19 +55,24 @@\\n

Hello. I am a paragraph.

\\n \\n \\n- \\n- \\n- \\n- \\n- Right Note\\n- \\n- \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n \\n \\n \\n@@ -76,20 +85,27 @@\\n This card was breaking the border radius.\\n \\n \\n- \\n- \\n- \\n- \\n- \\n+ \\n+ \\n+ \\n+ \\n+\\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n+ \\n \\n \\n \\n\", \"diff --git a/.github/workflows/ibis-docs-lint.yml b/.github/workflows/ibis-docs-lint.yml\\nindex 57d94a4..04de03b 100644\\n--- a/.github/workflows/ibis-docs-lint.yml\\n+++ b/.github/workflows/ibis-docs-lint.yml\\n@@ -206,7 +206,7 @@ jobs:\\n - name: build and push dev docs\\n run: |\\n nix develop --ignore-environment -c \\\\\\n- mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}'\\n+ mkdocs gh-deploy --message 'docs: ibis@${{ github.sha }}' --ignore-version\\n \\n simulate_release:\\n runs-on: ubuntu-latest\\n\", \"diff --git a/bootstrap/scripts/publish-patch.sh b/bootstrap/scripts/publish-patch.sh\\nindex a1b6f12..0d849a5 100755\\n--- a/bootstrap/scripts/publish-patch.sh\\n+++ b/bootstrap/scripts/publish-patch.sh\\n@@ -5,4 +5,4 @@ lerna version patch\\n lerna publish from-package -y\\n git push\\n \\n-./pack_and_install.sh\\n\\\\ No newline at end of file\\n+./bootstrap/scripts/pack_and_install.sh\\n\\\\ No newline at end of file\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"04e70ce964b343e28b3dbd0c46d10ccda958ab8c\", \"19feaea1885eb015759b5c7a5d785521f2b8a212\", \"21228c55b7045d9b2225f65e6231184ff332b071\", \"3fcfb20b0feb371b357edc42fcb7c87085c9b82a\"]"},"types":{"kind":"string","value":"[\"docs\", \"test\", \"ci\", \"build\"]"}}},{"rowIdx":1088,"cells":{"commit_message":{"kind":"string","value":"add --ignore-existing to all npx commands,add workingDirectory option to shell.openExternal() (#15065)\n\nAllows passing `workingDirectory` to the underlying `ShellExecuteW` API on Windows._x000D_\n_x000D_\nthe motivation is that by default `ShellExecute` would use the current working directory, which would get locked on Windows and can prevent autoUpdater from working correctly. We need to be able specify a different `workingDirectory` to prevent this situation.,rework RaftCommittedEntryListener\n\nIterate over RaftCommittedEntryListener and refactor the listener such it serves the actual need.\n\nWe have some services (to be specific the AsyncSnapshotDirector) which need the committed position, and\nwant to listen to new updates. In raft we know which record we are committing and whether it was an application record so we can pass this information threw the listeners.\n\nThis avoids to pass in the whole IndexedRecord object, and reduce the potential of going out of OOM because of keeping to much data in heap (when commit is not possible).,enable performance test trigger\n\nThis reverts commit 146c7b58154a5b3de957f87e3b193447e0576547."},"diff":{"kind":"string","value":"[\"diff --git a/docs/getting-started/getting-started.md b/docs/getting-started/getting-started.md\\nindex dc6db37..3ef9d0a 100644\\n--- a/docs/getting-started/getting-started.md\\n+++ b/docs/getting-started/getting-started.md\\n@@ -13,7 +13,7 @@ npm install -g @angular/cli\\n **Using `npx`**\\n \\n ```bash\\n-npx create-nx-workspace myworkspace\\n+npx --ignore-existing create-nx-workspace myworkspace\\n ```\\n \\n **Using `npm init`**\\ndiff --git a/docs/guides/react-and-angular.md b/docs/guides/react-and-angular.md\\nindex c1929a2..a5651ff 100644\\n--- a/docs/guides/react-and-angular.md\\n+++ b/docs/guides/react-and-angular.md\\n@@ -11,7 +11,7 @@ To show how Nx does it, let's build two applications (one in Angular, and one in\\n Let's start by creating a new Nx workspace. The easiest way to do this is to use npx.\\n \\n ```bash\\n-npx create-nx-workspace happynrwl --preset=empty\\n+npx --ignore-existing create-nx-workspace happynrwl --preset=empty\\n ```\\n \\n ## Creating an Angular Application\\ndiff --git a/docs/guides/react.md b/docs/guides/react.md\\nindex e1647fd..eac848e 100644\\n--- a/docs/guides/react.md\\n+++ b/docs/guides/react.md\\n@@ -16,13 +16,13 @@ Nx has first class support for React: you can create React applications and libr\\n Create a new Nx workspace. The easiest way to do it is to use npx.\\n \\n ```bash\\n-npx create-nx-workspace happynrwl --preset=empty\\n+npx --ignore-existing create-nx-workspace happynrwl --preset=empty\\n ```\\n \\n You can also create a workspace with a React application in place by running:\\n \\n ```bash\\n-npx create-nx-workspace happynrwl --preset=react\\n+npx --ignore-existing create-nx-workspace happynrwl --preset=react\\n ```\\n \\n ## Generating a React Application\\ndiff --git a/docs/tutorial/01-create-application.md b/docs/tutorial/01-create-application.md\\nindex ea87ecf..967a56e 100644\\n--- a/docs/tutorial/01-create-application.md\\n+++ b/docs/tutorial/01-create-application.md\\n@@ -7,7 +7,7 @@ In this tutorial you will use Nx to build a full-stack application out of common\\n **Start by creating a new workspace.**\\n \\n ```bash\\n-npx create-nx-workspace myorg\\n+npx --ignore-existing create-nx-workspace myorg\\n ```\\n \\n When asked about 'preset', select `empty`.\\n\", \"diff --git a/atom/browser/atom_browser_client.cc b/atom/browser/atom_browser_client.cc\\nindex 97e5f26..df0774b 100644\\n--- a/atom/browser/atom_browser_client.cc\\n+++ b/atom/browser/atom_browser_client.cc\\n@@ -611,7 +611,7 @@ void OnOpenExternal(const GURL& escaped_url, bool allowed) {\\n #else\\n escaped_url,\\n #endif\\n- true);\\n+ platform_util::OpenExternalOptions());\\n }\\n \\n void HandleExternalProtocolInUI(\\ndiff --git a/atom/common/api/atom_api_shell.cc b/atom/common/api/atom_api_shell.cc\\nindex 1323cd6..7c67c7a 100644\\n--- a/atom/common/api/atom_api_shell.cc\\n+++ b/atom/common/api/atom_api_shell.cc\\n@@ -60,11 +60,12 @@ bool OpenExternal(\\n const GURL& url,\\n #endif\\n mate::Arguments* args) {\\n- bool activate = true;\\n+ platform_util::OpenExternalOptions options;\\n if (args->Length() >= 2) {\\n- mate::Dictionary options;\\n- if (args->GetNext(&options)) {\\n- options.Get(\\\"activate\\\", &activate);\\n+ mate::Dictionary obj;\\n+ if (args->GetNext(&obj)) {\\n+ obj.Get(\\\"activate\\\", &options.activate);\\n+ obj.Get(\\\"workingDirectory\\\", &options.working_dir);\\n }\\n }\\n \\n@@ -72,13 +73,13 @@ bool OpenExternal(\\n base::Callback)> callback;\\n if (args->GetNext(&callback)) {\\n platform_util::OpenExternal(\\n- url, activate,\\n+ url, options,\\n base::Bind(&OnOpenExternalFinished, args->isolate(), callback));\\n return true;\\n }\\n }\\n \\n- return platform_util::OpenExternal(url, activate);\\n+ return platform_util::OpenExternal(url, options);\\n }\\n \\n #if defined(OS_WIN)\\ndiff --git a/atom/common/platform_util.h b/atom/common/platform_util.h\\nindex 6fd8405..6686a4f 100644\\n--- a/atom/common/platform_util.h\\n+++ b/atom/common/platform_util.h\\n@@ -8,6 +8,7 @@\\n #include \\n \\n #include \\\"base/callback_forward.h\\\"\\n+#include \\\"base/files/file_path.h\\\"\\n #include \\\"build/build_config.h\\\"\\n \\n #if defined(OS_WIN)\\n@@ -16,10 +17,6 @@\\n \\n class GURL;\\n \\n-namespace base {\\n-class FilePath;\\n-}\\n-\\n namespace platform_util {\\n \\n typedef base::Callback OpenExternalCallback;\\n@@ -32,6 +29,11 @@ bool ShowItemInFolder(const base::FilePath& full_path);\\n // Must be called from the UI thread.\\n bool OpenItem(const base::FilePath& full_path);\\n \\n+struct OpenExternalOptions {\\n+ bool activate = true;\\n+ base::FilePath working_dir;\\n+};\\n+\\n // Open the given external protocol URL in the desktop's default manner.\\n // (For example, mailto: URLs in the default mail user agent.)\\n bool OpenExternal(\\n@@ -40,7 +42,7 @@ bool OpenExternal(\\n #else\\n const GURL& url,\\n #endif\\n- bool activate);\\n+ const OpenExternalOptions& options);\\n \\n // The asynchronous version of OpenExternal.\\n void OpenExternal(\\n@@ -49,7 +51,7 @@ void OpenExternal(\\n #else\\n const GURL& url,\\n #endif\\n- bool activate,\\n+ const OpenExternalOptions& options,\\n const OpenExternalCallback& callback);\\n \\n // Move a file to trash.\\ndiff --git a/atom/common/platform_util_linux.cc b/atom/common/platform_util_linux.cc\\nindex 63ee0bd..f17cbda 100644\\n--- a/atom/common/platform_util_linux.cc\\n+++ b/atom/common/platform_util_linux.cc\\n@@ -80,7 +80,7 @@ bool OpenItem(const base::FilePath& full_path) {\\n return XDGOpen(full_path.value(), false);\\n }\\n \\n-bool OpenExternal(const GURL& url, bool activate) {\\n+bool OpenExternal(const GURL& url, const OpenExternalOptions& options) {\\n // Don't wait for exit, since we don't want to wait for the browser/email\\n // client window to close before returning\\n if (url.SchemeIs(\\\"mailto\\\"))\\n@@ -90,10 +90,10 @@ bool OpenExternal(const GURL& url, bool activate) {\\n }\\n \\n void OpenExternal(const GURL& url,\\n- bool activate,\\n+ const OpenExternalOptions& options,\\n const OpenExternalCallback& callback) {\\n // TODO(gabriel): Implement async open if callback is specified\\n- callback.Run(OpenExternal(url, activate) ? \\\"\\\" : \\\"Failed to open\\\");\\n+ callback.Run(OpenExternal(url, options) ? \\\"\\\" : \\\"Failed to open\\\");\\n }\\n \\n bool MoveItemToTrash(const base::FilePath& full_path) {\\ndiff --git a/atom/common/platform_util_mac.mm b/atom/common/platform_util_mac.mm\\nindex b83b1e1..4cda8bf 100644\\n--- a/atom/common/platform_util_mac.mm\\n+++ b/atom/common/platform_util_mac.mm\\n@@ -139,16 +139,16 @@ bool OpenItem(const base::FilePath& full_path) {\\n launchIdentifiers:NULL];\\n }\\n \\n-bool OpenExternal(const GURL& url, bool activate) {\\n+bool OpenExternal(const GURL& url, const OpenExternalOptions& options) {\\n DCHECK([NSThread isMainThread]);\\n NSURL* ns_url = net::NSURLWithGURL(url);\\n if (ns_url)\\n- return OpenURL(ns_url, activate).empty();\\n+ return OpenURL(ns_url, options.activate).empty();\\n return false;\\n }\\n \\n void OpenExternal(const GURL& url,\\n- bool activate,\\n+ const OpenExternalOptions& options,\\n const OpenExternalCallback& callback) {\\n NSURL* ns_url = net::NSURLWithGURL(url);\\n if (!ns_url) {\\n@@ -157,13 +157,13 @@ void OpenExternal(const GURL& url,\\n }\\n \\n __block OpenExternalCallback c = callback;\\n- dispatch_async(dispatch_get_global_queue(DISPATCH_QUEUE_PRIORITY_DEFAULT, 0),\\n- ^{\\n- __block std::string error = OpenURL(ns_url, activate);\\n- dispatch_async(dispatch_get_main_queue(), ^{\\n- c.Run(error);\\n- });\\n- });\\n+ dispatch_async(\\n+ dispatch_get_global_queue(DISPATCH_QUEUE_PRIORITY_DEFAULT, 0), ^{\\n+ __block std::string error = OpenURL(ns_url, options.activate);\\n+ dispatch_async(dispatch_get_main_queue(), ^{\\n+ c.Run(error);\\n+ });\\n+ });\\n }\\n \\n bool MoveItemToTrash(const base::FilePath& full_path) {\\ndiff --git a/atom/common/platform_util_win.cc b/atom/common/platform_util_win.cc\\nindex 34576be..5712200 100644\\n--- a/atom/common/platform_util_win.cc\\n+++ b/atom/common/platform_util_win.cc\\n@@ -294,15 +294,18 @@ bool OpenItem(const base::FilePath& full_path) {\\n return ui::win::OpenFileViaShell(full_path);\\n }\\n \\n-bool OpenExternal(const base::string16& url, bool activate) {\\n+bool OpenExternal(const base::string16& url,\\n+ const OpenExternalOptions& options) {\\n // Quote the input scheme to be sure that the command does not have\\n // parameters unexpected by the external program. This url should already\\n // have been escaped.\\n base::string16 escaped_url = L\\\"\\\\\\\"\\\" + url + L\\\"\\\\\\\"\\\";\\n+ auto working_dir = options.working_dir.value();\\n \\n- if (reinterpret_cast(ShellExecuteW(\\n- NULL, L\\\"open\\\", escaped_url.c_str(), NULL, NULL, SW_SHOWNORMAL)) <=\\n- 32) {\\n+ if (reinterpret_cast(\\n+ ShellExecuteW(nullptr, L\\\"open\\\", escaped_url.c_str(), nullptr,\\n+ working_dir.empty() ? nullptr : working_dir.c_str(),\\n+ SW_SHOWNORMAL)) <= 32) {\\n // We fail to execute the call. We could display a message to the user.\\n // TODO(nsylvain): we should also add a dialog to warn on errors. See\\n // bug 1136923.\\n@@ -312,10 +315,10 @@ bool OpenExternal(const base::string16& url, bool activate) {\\n }\\n \\n void OpenExternal(const base::string16& url,\\n- bool activate,\\n+ const OpenExternalOptions& options,\\n const OpenExternalCallback& callback) {\\n // TODO(gabriel): Implement async open if callback is specified\\n- callback.Run(OpenExternal(url, activate) ? \\\"\\\" : \\\"Failed to open\\\");\\n+ callback.Run(OpenExternal(url, options) ? \\\"\\\" : \\\"Failed to open\\\");\\n }\\n \\n bool MoveItemToTrash(const base::FilePath& path) {\\ndiff --git a/docs/api/shell.md b/docs/api/shell.md\\nindex a469f94..b38348a 100644\\n--- a/docs/api/shell.md\\n+++ b/docs/api/shell.md\\n@@ -37,9 +37,10 @@ Open the given file in the desktop's default manner.\\n ### `shell.openExternal(url[, options, callback])`\\n \\n * `url` String - Max 2081 characters on windows, or the function returns false.\\n-* `options` Object (optional) _macOS_\\n- * `activate` Boolean - `true` to bring the opened application to the\\n- foreground. The default is `true`.\\n+* `options` Object (optional)\\n+ * `activate` Boolean (optional) - `true` to bring the opened application to the\\n+ foreground. The default is `true`. _macOS_\\n+ * `workingDirectory` String (optional) - The working directory. _Windows_\\n * `callback` Function (optional) _macOS_ - If specified will perform the open asynchronously.\\n * `error` Error\\n \\n\", \"diff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\\nnew file mode 100644\\nindex 0000000..57c28a9\\n--- /dev/null\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/RaftApplicationEntryCommittedPositionListener.java\\n@@ -0,0 +1,31 @@\\n+/*\\n+ * Copyright 2016-present Open Networking Foundation\\n+ * Copyright \\u00a9 2020 camunda services GmbH (info@camunda.com)\\n+ *\\n+ * Licensed under the Apache License, Version 2.0 (the \\\"License\\\");\\n+ * you may not use this file except in compliance with the License.\\n+ * You may obtain a copy of the License at\\n+ *\\n+ * http://www.apache.org/licenses/LICENSE-2.0\\n+ *\\n+ * Unless required by applicable law or agreed to in writing, software\\n+ * distributed under the License is distributed on an \\\"AS IS\\\" BASIS,\\n+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\\n+ * See the License for the specific language governing permissions and\\n+ * limitations under the License.\\n+ */\\n+package io.atomix.raft;\\n+\\n+/**\\n+ * This listener will only be called by the Leader, when it commits an application entry.\\n+ *\\n+ *

If RAFT is currently running in a follower role, it will not call this listener.\\n+ */\\n+@FunctionalInterface\\n+public interface RaftApplicationEntryCommittedPositionListener {\\n+\\n+ /**\\n+ * @param committedPosition the new committed position which is related to the application entries\\n+ */\\n+ void onCommit(long committedPosition);\\n+}\\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java b/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\\ndeleted file mode 100644\\nindex 3d11d75..0000000\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/RaftCommittedEntryListener.java\\n+++ /dev/null\\n@@ -1,32 +0,0 @@\\n-/*\\n- * Copyright 2016-present Open Networking Foundation\\n- * Copyright \\u00a9 2020 camunda services GmbH (info@camunda.com)\\n- *\\n- * Licensed under the Apache License, Version 2.0 (the \\\"License\\\");\\n- * you may not use this file except in compliance with the License.\\n- * You may obtain a copy of the License at\\n- *\\n- * http://www.apache.org/licenses/LICENSE-2.0\\n- *\\n- * Unless required by applicable law or agreed to in writing, software\\n- * distributed under the License is distributed on an \\\"AS IS\\\" BASIS,\\n- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\\n- * See the License for the specific language governing permissions and\\n- * limitations under the License.\\n- */\\n-package io.atomix.raft;\\n-\\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\\n-\\n-/**\\n- * This listener will only be called by the Leader, when it commits an entry. If RAFT is currently\\n- * running in a follower role, it will not call this listener.\\n- */\\n-@FunctionalInterface\\n-public interface RaftCommittedEntryListener {\\n-\\n- /**\\n- * @param indexedRaftLogEntry the new committed entry\\n- */\\n- void onCommit(IndexedRaftLogEntry indexedRaftLogEntry);\\n-}\\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\\nindex 1f4ee98..c177cb1 100644\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/impl/RaftContext.java\\n@@ -27,8 +27,8 @@ import io.atomix.cluster.MemberId;\\n import io.atomix.cluster.messaging.MessagingException.NoRemoteHandler;\\n import io.atomix.cluster.messaging.MessagingException.NoSuchMemberException;\\n import io.atomix.raft.ElectionTimer;\\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\\n import io.atomix.raft.RaftCommitListener;\\n-import io.atomix.raft.RaftCommittedEntryListener;\\n import io.atomix.raft.RaftError;\\n import io.atomix.raft.RaftException.ProtocolException;\\n import io.atomix.raft.RaftRoleChangeListener;\\n@@ -61,7 +61,6 @@ import io.atomix.raft.roles.PromotableRole;\\n import io.atomix.raft.roles.RaftRole;\\n import io.atomix.raft.storage.RaftStorage;\\n import io.atomix.raft.storage.StorageException;\\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\\n import io.atomix.raft.storage.log.RaftLog;\\n import io.atomix.raft.storage.system.MetaStore;\\n import io.atomix.raft.utils.StateUtil;\\n@@ -115,7 +114,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\\n private final Set> stateChangeListeners = new CopyOnWriteArraySet<>();\\n private final Set> electionListeners = new CopyOnWriteArraySet<>();\\n private final Set commitListeners = new CopyOnWriteArraySet<>();\\n- private final Set committedEntryListeners =\\n+ private final Set committedEntryListeners =\\n new CopyOnWriteArraySet<>();\\n private final Set snapshotReplicationListeners =\\n new CopyOnWriteArraySet<>();\\n@@ -433,21 +432,23 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\\n *

Note that it will be called on the Raft thread, and as such should not perform any heavy\\n * computation.\\n *\\n- * @param raftCommittedEntryListener the listener to add\\n+ * @param raftApplicationEntryCommittedPositionListener the listener to add\\n */\\n public void addCommittedEntryListener(\\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\\n- committedEntryListeners.add(raftCommittedEntryListener);\\n+ final RaftApplicationEntryCommittedPositionListener\\n+ raftApplicationEntryCommittedPositionListener) {\\n+ committedEntryListeners.add(raftApplicationEntryCommittedPositionListener);\\n }\\n \\n /**\\n * Removes registered committedEntryListener\\n *\\n- * @param raftCommittedEntryListener the listener to remove\\n+ * @param raftApplicationEntryCommittedPositionListener the listener to remove\\n */\\n public void removeCommittedEntryListener(\\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\\n- committedEntryListeners.remove(raftCommittedEntryListener);\\n+ final RaftApplicationEntryCommittedPositionListener\\n+ raftApplicationEntryCommittedPositionListener) {\\n+ committedEntryListeners.remove(raftApplicationEntryCommittedPositionListener);\\n }\\n \\n /**\\n@@ -464,7 +465,7 @@ public class RaftContext implements AutoCloseable, HealthMonitorable {\\n *\\n * @param committedEntry the most recently committed entry\\n */\\n- public void notifyCommittedEntryListeners(final IndexedRaftLogEntry committedEntry) {\\n+ public void notifyApplicationEntryCommittedPositionListeners(final long committedEntry) {\\n committedEntryListeners.forEach(listener -> listener.onCommit(committedEntry));\\n }\\n \\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\\nindex 56c7172..d075fca 100644\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/partition/impl/RaftPartitionServer.java\\n@@ -21,8 +21,8 @@ import io.atomix.cluster.MemberId;\\n import io.atomix.cluster.messaging.ClusterCommunicationService;\\n import io.atomix.primitive.partition.Partition;\\n import io.atomix.primitive.partition.PartitionMetadata;\\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\\n import io.atomix.raft.RaftCommitListener;\\n-import io.atomix.raft.RaftCommittedEntryListener;\\n import io.atomix.raft.RaftRoleChangeListener;\\n import io.atomix.raft.RaftServer;\\n import io.atomix.raft.RaftServer.Role;\\n@@ -205,16 +205,20 @@ public class RaftPartitionServer implements HealthMonitorable {\\n }\\n \\n /**\\n- * @see io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftCommittedEntryListener)\\n+ * @see\\n+ * io.atomix.raft.impl.RaftContext#addCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\\n */\\n- public void addCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\\n+ public void addCommittedEntryListener(\\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\\n server.getContext().addCommittedEntryListener(commitListener);\\n }\\n \\n /**\\n- * @see io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftCommittedEntryListener)\\n+ * @see\\n+ * io.atomix.raft.impl.RaftContext#removeCommittedEntryListener(RaftApplicationEntryCommittedPositionListener)\\n */\\n- public void removeCommittedEntryListener(final RaftCommittedEntryListener commitListener) {\\n+ public void removeCommittedEntryListener(\\n+ final RaftApplicationEntryCommittedPositionListener commitListener) {\\n server.getContext().removeCommittedEntryListener(commitListener);\\n }\\n \\ndiff --git a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\\nindex e54df1a..fcfd177 100644\\n--- a/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\\n+++ b/atomix/cluster/src/main/java/io/atomix/raft/roles/LeaderRole.java\\n@@ -630,27 +630,47 @@ public final class LeaderRole extends ActiveRole implements ZeebeLogAppender {\\n \\n private void replicate(final IndexedRaftLogEntry indexed, final AppendListener appendListener) {\\n raft.checkThread();\\n- appender\\n- .appendEntries(indexed.index())\\n- .whenCompleteAsync(\\n- (commitIndex, commitError) -> {\\n- if (!isRunning()) {\\n- return;\\n- }\\n+ final var appendEntriesFuture = appender.appendEntries(indexed.index());\\n+\\n+ final boolean applicationEntryWasCommitted = indexed.isApplicationEntry();\\n+ if (applicationEntryWasCommitted) {\\n+ // We have some services which are waiting for the application records, especially position\\n+ // to be committed. This is our glue code to notify them, instead of\\n+ // passing the complete object (IndexedRaftLogEntry) threw the listeners and\\n+ // keep them in heap until they are committed. This had the risk of going out of OOM\\n+ // if records can't be committed, see https://github.com/camunda/zeebe/issues/14275\\n+ final var committedPosition = indexed.getApplicationEntry().highestPosition();\\n+ appendEntriesFuture.whenCompleteAsync(\\n+ (commitIndex, commitError) -> {\\n+ if (!isRunning()) {\\n+ return;\\n+ }\\n+\\n+ if (commitError == null) {\\n+ raft.notifyApplicationEntryCommittedPositionListeners(committedPosition);\\n+ }\\n+ },\\n+ raft.getThreadContext());\\n+ }\\n \\n- // have the state machine apply the index which should do nothing but ensures it keeps\\n- // up to date with the latest entries, so it can handle configuration and initial\\n- // entries properly on fail over\\n- if (commitError == null) {\\n- appendListener.onCommit(indexed.index());\\n- raft.notifyCommittedEntryListeners(indexed);\\n- } else {\\n- appendListener.onCommitError(indexed.index(), commitError);\\n- // replicating the entry will be retried on the next append request\\n- log.error(\\\"Failed to replicate entry: {}\\\", indexed, commitError);\\n- }\\n- },\\n- raft.getThreadContext());\\n+ appendEntriesFuture.whenCompleteAsync(\\n+ (commitIndex, commitError) -> {\\n+ if (!isRunning()) {\\n+ return;\\n+ }\\n+\\n+ // have the state machine apply the index which should do nothing but ensures it keeps\\n+ // up to date with the latest entries, so it can handle configuration and initial\\n+ // entries properly on fail over\\n+ if (commitError == null) {\\n+ appendListener.onCommit(indexed.index());\\n+ } else {\\n+ appendListener.onCommitError(indexed.index(), commitError);\\n+ // replicating the entry will be retried on the next append request\\n+ log.error(\\\"Failed to replicate entry: {}\\\", indexed, commitError);\\n+ }\\n+ },\\n+ raft.getThreadContext());\\n }\\n \\n public synchronized void onInitialEntriesCommitted(final Runnable runnable) {\\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\\nindex b217586..8029766 100644\\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftAppendTest.java\\n@@ -82,7 +82,7 @@ public class RaftAppendTest {\\n @Test\\n public void shouldNotifyCommittedEntryListenerOnLeaderOnly() throws Throwable {\\n // given\\n- final var committedEntryListener = mock(RaftCommittedEntryListener.class);\\n+ final var committedEntryListener = mock(RaftApplicationEntryCommittedPositionListener.class);\\n raftRule.addCommittedEntryListener(committedEntryListener);\\n \\n // when\\ndiff --git a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\\nindex 8f73cba..193a176 100644\\n--- a/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\\n+++ b/atomix/cluster/src/test/java/io/atomix/raft/RaftRule.java\\n@@ -644,9 +644,12 @@ public final class RaftRule extends ExternalResource {\\n }\\n \\n public void addCommittedEntryListener(\\n- final RaftCommittedEntryListener raftCommittedEntryListener) {\\n+ final RaftApplicationEntryCommittedPositionListener\\n+ raftApplicationEntryCommittedPositionListener) {\\n servers.forEach(\\n- (id, raft) -> raft.getContext().addCommittedEntryListener(raftCommittedEntryListener));\\n+ (id, raft) ->\\n+ raft.getContext()\\n+ .addCommittedEntryListener(raftApplicationEntryCommittedPositionListener));\\n }\\n \\n public void partition(final RaftServer follower) {\\ndiff --git a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\nindex a61571f..6c082d7 100644\\n--- a/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n+++ b/broker/src/main/java/io/camunda/zeebe/broker/system/partitions/impl/AsyncSnapshotDirector.java\\n@@ -7,8 +7,7 @@\\n */\\n package io.camunda.zeebe.broker.system.partitions.impl;\\n \\n-import io.atomix.raft.RaftCommittedEntryListener;\\n-import io.atomix.raft.storage.log.IndexedRaftLogEntry;\\n+import io.atomix.raft.RaftApplicationEntryCommittedPositionListener;\\n import io.camunda.zeebe.broker.system.partitions.NoEntryAtSnapshotPosition;\\n import io.camunda.zeebe.broker.system.partitions.StateController;\\n import io.camunda.zeebe.logstreams.impl.Loggers;\\n@@ -36,7 +35,7 @@ import java.util.function.Consumer;\\n import org.slf4j.Logger;\\n \\n public final class AsyncSnapshotDirector extends Actor\\n- implements RaftCommittedEntryListener, HealthMonitorable {\\n+ implements RaftApplicationEntryCommittedPositionListener, HealthMonitorable {\\n \\n public static final Duration MINIMUM_SNAPSHOT_PERIOD = Duration.ofMinutes(1);\\n \\n@@ -115,7 +114,7 @@ public final class AsyncSnapshotDirector extends Actor\\n @Override\\n protected void handleFailure(final Throwable failure) {\\n LOG.error(\\n- \\\"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}. {}\\\",\\n+ \\\"No snapshot was taken due to failure in '{}'. Will try to take snapshot after snapshot period {}.\\\",\\n actorName,\\n snapshotRate,\\n failure);\\n@@ -407,13 +406,8 @@ public final class AsyncSnapshotDirector extends Actor\\n }\\n \\n @Override\\n- public void onCommit(final IndexedRaftLogEntry indexedRaftLogEntry) {\\n- // is called by the Leader Role and gives the last committed entry, where we\\n- // can extract the highest position, which corresponds to the last committed position\\n- if (indexedRaftLogEntry.isApplicationEntry()) {\\n- final var committedPosition = indexedRaftLogEntry.getApplicationEntry().highestPosition();\\n- newPositionCommitted(committedPosition);\\n- }\\n+ public void onCommit(final long committedPosition) {\\n+ newPositionCommitted(committedPosition);\\n }\\n \\n public void newPositionCommitted(final long currentCommitPosition) {\\n\", \"diff --git a/Jenkinsfile b/Jenkinsfile\\nindex 399f8b8..c3f8fde 100644\\n--- a/Jenkinsfile\\n+++ b/Jenkinsfile\\n@@ -120,6 +120,12 @@ pipeline {\\n }\\n }\\n \\n+ stage('Trigger Performance Tests') {\\n+ when { branch 'develop' }\\n+ steps {\\n+ build job: 'zeebe-cluster-performance-tests', wait: false\\n+ }\\n+ }\\n }\\n \\n post {\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"fc9af4d0b93d69be4e201ffb18da04324e8a4a87\", \"a9475f359061fcd6cd53557599fedf0df5e9ee00\", \"323cf81961cdd3748a7ba6ba470ecb13e5374e9f\", \"80944b7a513b442afcb2d0d6c7d71c0d79365dba\"]"},"types":{"kind":"string","value":"[\"docs\", \"feat\", \"refactor\", \"ci\"]"}}},{"rowIdx":1089,"cells":{"commit_message":{"kind":"string","value":"trigger build every hour for develop\n\nTo better track stability of the develop branch the build should be\ntriggered on commit and every hour. Other branches should not be\neffected.\n\n- add cron trigger to develop branch\n- extract variables to identify stable and develop branch,fix unstable MessageCorrelationTest,fix typos (#90),get ip from forwarded header"},"diff":{"kind":"string","value":"[\"diff --git a/Jenkinsfile b/Jenkinsfile\\nindex 2c58f61..9daa38f 100644\\n--- a/Jenkinsfile\\n+++ b/Jenkinsfile\\n@@ -4,9 +4,17 @@\\n \\n def buildName = \\\"${env.JOB_BASE_NAME.replaceAll(\\\"%2F\\\", \\\"-\\\").replaceAll(\\\"\\\\\\\\.\\\", \\\"-\\\").take(20)}-${env.BUILD_ID}\\\"\\n \\n+def masterBranchName = 'master'\\n+def isMasterBranch = env.BRANCH_NAME == masterBranchName\\n+def developBranchName = 'develop'\\n+def isDevelopBranch = env.BRANCH_NAME == developBranchName\\n+\\n //for develop branch keep builds for 7 days to be able to analyse build errors, for all other branches, keep the last 10 builds\\n-def daysToKeep = (env.BRANCH_NAME=='develop') ? '7' : '-1'\\n-def numToKeep = (env.BRANCH_NAME=='develop') ? '-1' : '10'\\n+def daysToKeep = isDevelopBranch ? '7' : '-1'\\n+def numToKeep = isDevelopBranch ? '-1' : '10'\\n+\\n+//the develop branch should be run hourly to detect flaky tests and instability, other branches only on commit\\n+def cronTrigger = isDevelopBranch ? '@hourly' : ''\\n \\n pipeline {\\n agent {\\n@@ -23,6 +31,10 @@ pipeline {\\n SONARCLOUD_TOKEN = credentials('zeebe-sonarcloud-token')\\n }\\n \\n+ triggers {\\n+ cron(cronTrigger)\\n+ }\\n+\\n options {\\n buildDiscarder(logRotator(daysToKeepStr: daysToKeep, numToKeepStr: numToKeep))\\n timestamps()\\n@@ -201,7 +213,7 @@ pipeline {\\n }\\n \\n stage('Upload') {\\n- when { branch 'develop' }\\n+ when { allOf { branch developBranchName ; not { triggeredBy 'TimerTrigger' } } }\\n steps {\\n retry(3) {\\n container('maven') {\\n@@ -214,9 +226,11 @@ pipeline {\\n }\\n \\n stage('Post') {\\n+ when { not { triggeredBy 'TimerTrigger' } }\\n+\\n parallel {\\n stage('Docker') {\\n- when { branch 'develop' }\\n+ when { branch developBranchName }\\n \\n environment {\\n VERSION = readMavenPom(file: 'parent/pom.xml').getVersion()\\n@@ -227,20 +241,20 @@ pipeline {\\n build job: 'zeebe-docker', parameters: [\\n string(name: 'BRANCH', value: env.BRANCH_NAME),\\n string(name: 'VERSION', value: env.VERSION),\\n- booleanParam(name: 'IS_LATEST', value: env.BRANCH_NAME == 'master'),\\n- booleanParam(name: 'PUSH', value: env.BRANCH_NAME == 'develop')\\n+ booleanParam(name: 'IS_LATEST', value: isMasterBranch),\\n+ booleanParam(name: 'PUSH', value: isDevelopBranch)\\n ]\\n }\\n }\\n }\\n \\n stage('Docs') {\\n- when { anyOf { branch 'master'; branch 'develop' } }\\n+ when { anyOf { branch masterBranchName; branch developBranchName } }\\n steps {\\n retry(3) {\\n build job: 'zeebe-docs', parameters: [\\n string(name: 'BRANCH', value: env.BRANCH_NAME),\\n- booleanParam(name: 'LIVE', value: env.BRANCH_NAME == 'master')\\n+ booleanParam(name: 'LIVE', value: isMasterBranch)\\n ]\\n }\\n }\\n\", \"diff --git a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\\nindex 0f5fed9..796393c 100644\\n--- a/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\\n+++ b/broker-core/src/test/java/io/zeebe/broker/workflow/MessageCorrelationTest.java\\n@@ -27,7 +27,6 @@ import static io.zeebe.test.util.MsgPackUtil.asMsgPack;\\n import static org.assertj.core.api.Assertions.assertThat;\\n import static org.assertj.core.api.Assertions.entry;\\n \\n-import io.zeebe.UnstableTest;\\n import io.zeebe.broker.test.EmbeddedBrokerRule;\\n import io.zeebe.model.bpmn.Bpmn;\\n import io.zeebe.model.bpmn.BpmnModelInstance;\\n@@ -50,7 +49,6 @@ import org.agrona.DirectBuffer;\\n import org.junit.Before;\\n import org.junit.Rule;\\n import org.junit.Test;\\n-import org.junit.experimental.categories.Category;\\n import org.junit.rules.RuleChain;\\n import org.junit.runner.RunWith;\\n import org.junit.runners.Parameterized;\\n@@ -165,7 +163,7 @@ public class MessageCorrelationTest {\\n \\\"receive-message\\\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\\n \\n final SubscribedRecord messageSubscription =\\n- findMessageSubscription(testClient, MessageSubscriptionIntent.OPENED);\\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\\n assertThat(messageSubscription.valueType()).isEqualTo(ValueType.MESSAGE_SUBSCRIPTION);\\n assertThat(messageSubscription.recordType()).isEqualTo(RecordType.EVENT);\\n assertThat(messageSubscription.value())\\n@@ -244,7 +242,7 @@ public class MessageCorrelationTest {\\n final long workflowInstanceKey =\\n testClient.createWorkflowInstance(\\\"wf\\\", asMsgPack(\\\"orderId\\\", \\\"order-123\\\"));\\n \\n- testClient.receiveFirstWorkflowInstanceEvent(WorkflowInstanceIntent.ELEMENT_ACTIVATED);\\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\\n \\n // when\\n testClient.publishMessage(\\\"order canceled\\\", \\\"order-123\\\", asMsgPack(\\\"foo\\\", \\\"bar\\\"));\\n@@ -308,13 +306,12 @@ public class MessageCorrelationTest {\\n }\\n \\n @Test\\n- @Category(UnstableTest.class) // => https://github.com/zeebe-io/zeebe/issues/1234\\n public void shouldCorrelateMessageWithZeroTTL() throws Exception {\\n // given\\n final long workflowInstanceKey =\\n testClient.createWorkflowInstance(\\\"wf\\\", asMsgPack(\\\"orderId\\\", \\\"order-123\\\"));\\n \\n- testClient.receiveElementInState(\\\"receive-message\\\", WorkflowInstanceIntent.ELEMENT_ACTIVATED);\\n+ findMessageSubscription(MessageSubscriptionIntent.OPENED);\\n \\n // when\\n testClient.publishMessage(\\\"order canceled\\\", \\\"order-123\\\", asMsgPack(\\\"foo\\\", \\\"bar\\\"), 0);\\n@@ -499,10 +496,9 @@ public class MessageCorrelationTest {\\n .containsEntry(\\\"activityInstanceKey\\\", catchEventEntered.key());\\n }\\n \\n- private SubscribedRecord findMessageSubscription(\\n- final TestPartitionClient client, final MessageSubscriptionIntent intent)\\n+ private SubscribedRecord findMessageSubscription(final MessageSubscriptionIntent intent)\\n throws AssertionError {\\n- return client\\n+ return testClient\\n .receiveEvents()\\n .filter(intent(intent))\\n .findFirst()\\n\", \"diff --git a/README.md b/README.md\\nindex de15ac5..5ad8b47 100755\\n--- a/README.md\\n+++ b/README.md\\n@@ -16,13 +16,13 @@ content that will be loaded, similar to Facebook cards loaders.\\n \\n ## Features\\n \\n-* :gear: **Complety customizable:** you can change the colors, speed and sizes;\\n+* :gear: **Completely customizable:** you can change the colors, speed and sizes;\\n * :pencil2: **Create your own loading:** use the\\n [create-react-content-loader](https://danilowoz.github.io/create-react-content-loader/) to create\\n- your customs loadings easily;\\n+ your custom loadings easily;\\n * :ok_hand: **You can use right now:** there are a lot of presets to use the loader, see the\\n [options](#options);\\n-* :rocket: **Perfomance:** react-content-loader uses pure SVG to work, so it's works without any extra scritpt,\\n+* :rocket: **Performance:** react-content-loader uses pure SVG to work, so it works without any extra scripts,\\n canvas, etc;\\n \\n ## Usage\\n\", \"diff --git a/kousa/lib/broth/socket_handler.ex b/kousa/lib/broth/socket_handler.ex\\nindex d142135..5828f30 100644\\n--- a/kousa/lib/broth/socket_handler.ex\\n+++ b/kousa/lib/broth/socket_handler.ex\\n@@ -22,7 +22,7 @@ defmodule Broth.SocketHandler do\\n ## initialization boilerplate\\n \\n @impl true\\n- def init(request = %{peer: {ip, _reverse_port}}, _state) do\\n+ def init(request, _state) do\\n props = :cowboy_req.parse_qs(request)\\n \\n compression =\\n@@ -37,10 +37,16 @@ defmodule Broth.SocketHandler do\\n _ -> :json\\n end\\n \\n+ ip =\\n+ case request.headers do\\n+ %{\\\"x-forwarded-for\\\" => v} -> v\\n+ _ -> nil\\n+ end\\n+\\n state = %__MODULE__{\\n awaiting_init: true,\\n user_id: nil,\\n- ip: IP.to_string(ip),\\n+ ip: ip,\\n encoding: encoding,\\n compression: compression,\\n callers: get_callers(request)\\ndiff --git a/kousa/test/_support/ws_client.ex b/kousa/test/_support/ws_client.ex\\nindex aeca704..125da17 100644\\n--- a/kousa/test/_support/ws_client.ex\\n+++ b/kousa/test/_support/ws_client.ex\\n@@ -19,7 +19,9 @@ defmodule BrothTest.WsClient do\\n \\n @api_url\\n |> Path.join(\\\"socket\\\")\\n- |> WebSockex.start_link(__MODULE__, nil, extra_headers: [{\\\"user-agent\\\", ancestors}])\\n+ |> WebSockex.start_link(__MODULE__, nil,\\n+ extra_headers: [{\\\"user-agent\\\", ancestors}, {\\\"x-forwarded-for\\\", \\\"127.0.0.1\\\"}]\\n+ )\\n end\\n \\n ###########################################################################\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"3bc1541d6c95ef8cb5ce5da741733f09c98e4b29\", \"98bed2a8137930149559bc1cae9bd34a1a75e556\", \"88257ee720ed8ba136d49087c0d31373e8397dd5\", \"2f5718743a830d40ddf272ad46f253dbb6d08cff\"]"},"types":{"kind":"string","value":"[\"ci\", \"test\", \"docs\", \"fix\"]"}}},{"rowIdx":1090,"cells":{"commit_message":{"kind":"string","value":"update pr condition,updated webpack in react,dedup redundant imports,generate terminate end event compatible execution steps part 1\n\nThe random execution tests don't know the concept of flow scopes. This makes it challenging to generate a correct execution path for terminate end events, as they terminate a specific flow scope. Processing should continue as normal once the flow scope has been terminated.\n\nWhilst we don't have flow scopes, we do have execution path segments. These segments don't map 1 to 1 to flow scopes. However, since every flow scope starts a new segment we can use these segments to get the desired behavior.\n\nEach segment must keep track whether is has reached a terminate end event. If this is the case that means that we don't expect any further execution steps. We can isolate this behavior in a single location, during the appending of one segment to another segment.\nIn order to differentiate between flow scopes a new append method has been added which takes the boolean `changesFlowScope` as a parameter. Blockbuilder where the flow scope changes (e.g. SubProcessBlockBuilder) can use this to indicate that even though a terminate end event has been reached. Execution steps after this specific segment still need to added to complete the process.\n\nWhen a segment is appended to a different segment and the flow scope does not change we can use the segment that should be appended to identify whether new segment can still be added to the current segment. If passed segment has reached a terminate end event and the flow scope has not been changed it is guaranteed that the current segment is in the same flow scope has the previous segment and thus has also reached the terminate end event."},"diff":{"kind":"string","value":"[\"diff --git a/.github/workflows/release-pr.yml b/.github/workflows/release-pr.yml\\nindex 697ca8e..23f4475 100644\\n--- a/.github/workflows/release-pr.yml\\n+++ b/.github/workflows/release-pr.yml\\n@@ -3,7 +3,6 @@ name: release\\n on:\\n issue_comment:\\n types: [created]\\n- contains: \\\"/trigger release\\\"\\n \\n env:\\n # 7 GiB by default on GitHub, setting to 6 GiB\\n@@ -11,6 +10,7 @@ env:\\n \\n jobs:\\n release-pr:\\n+ if: ${{ github.event.issue.pull_request && github.event.comment.body == '/trigger release' }}\\n permissions:\\n id-token: write\\n runs-on: ubuntu-latest\\n\", \"diff --git a/components/react/package.json b/components/react/package.json\\nindex bbeb9ee..43ddebc 100644\\n--- a/components/react/package.json\\n+++ b/components/react/package.json\\n@@ -114,7 +114,7 @@\\n \\\"ts-loader\\\": \\\"^9.2.9\\\",\\n \\\"ts-node\\\": \\\"^10.7.0\\\",\\n \\\"typescript\\\": \\\"^4.7.3\\\",\\n- \\\"webpack\\\": \\\"^5.72.0\\\",\\n+ \\\"webpack\\\": \\\"^5.73.0\\\",\\n \\\"webpack-bundle-analyzer\\\": \\\"^4.5.0\\\",\\n \\\"webpack-cli\\\": \\\"^4.9.2\\\",\\n \\\"webpack-node-externals\\\": \\\"^3.0.0\\\"\\ndiff --git a/yarn.lock b/yarn.lock\\nindex a3fdb26..19a0716 100644\\n--- a/yarn.lock\\n+++ b/yarn.lock\\n@@ -25212,7 +25212,7 @@ webpack@^4.38.0, webpack@^4.42.1:\\n watchpack \\\"^1.7.4\\\"\\n webpack-sources \\\"^1.4.1\\\"\\n \\n-webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0:\\n+webpack@^5.54.0, webpack@^5.71.0, webpack@^5.72.0, webpack@^5.73.0:\\n version \\\"5.73.0\\\"\\n resolved \\\"https://registry.yarnpkg.com/webpack/-/webpack-5.73.0.tgz#bbd17738f8a53ee5760ea2f59dce7f3431d35d38\\\"\\n integrity sha512-svjudQRPPa0YiOYa2lM/Gacw0r6PvxptHj4FuEKQ2kX05ZLkjbVc5MnPs6its5j7IZljnIqSVo/OsY2X0IpHGA==\\n\", \"diff --git a/ibis/backends/base/__init__.py b/ibis/backends/base/__init__.py\\nindex effd44c..a59c0ec 100644\\n--- a/ibis/backends/base/__init__.py\\n+++ b/ibis/backends/base/__init__.py\\n@@ -31,7 +31,7 @@ import ibis.common.exceptions as exc\\n import ibis.config\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n \\n __all__ = ('BaseBackend', 'Database', 'connect')\\n \\ndiff --git a/ibis/backends/base/sql/__init__.py b/ibis/backends/base/sql/__init__.py\\nindex e4f2129..7bbdaf9 100644\\n--- a/ibis/backends/base/sql/__init__.py\\n+++ b/ibis/backends/base/sql/__init__.py\\n@@ -12,7 +12,7 @@ import ibis.expr.analysis as an\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base import BaseBackend\\n from ibis.backends.base.sql.compiler import Compiler\\n \\ndiff --git a/ibis/backends/base/sql/alchemy/__init__.py b/ibis/backends/base/sql/alchemy/__init__.py\\nindex 71cc0e8..ab89d7d 100644\\n--- a/ibis/backends/base/sql/alchemy/__init__.py\\n+++ b/ibis/backends/base/sql/alchemy/__init__.py\\n@@ -11,7 +11,7 @@ import ibis\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql import BaseSQLBackend\\n from ibis.backends.base.sql.alchemy.database import AlchemyDatabase, AlchemyTable\\n from ibis.backends.base.sql.alchemy.datatypes import (\\ndiff --git a/ibis/backends/base/sql/alchemy/query_builder.py b/ibis/backends/base/sql/alchemy/query_builder.py\\nindex 54c74ba..0ec432f 100644\\n--- a/ibis/backends/base/sql/alchemy/query_builder.py\\n+++ b/ibis/backends/base/sql/alchemy/query_builder.py\\n@@ -3,7 +3,7 @@ from __future__ import annotations\\n import functools\\n \\n import sqlalchemy as sa\\n-import sqlalchemy.sql as sql\\n+from sqlalchemy import sql\\n \\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\ndiff --git a/ibis/backends/base/sql/compiler/base.py b/ibis/backends/base/sql/compiler/base.py\\nindex 84102aa..fb44667 100644\\n--- a/ibis/backends/base/sql/compiler/base.py\\n+++ b/ibis/backends/base/sql/compiler/base.py\\n@@ -7,7 +7,7 @@ import toolz\\n \\n import ibis.expr.analysis as an\\n import ibis.expr.operations as ops\\n-import ibis.util as util\\n+from ibis import util\\n \\n \\n class DML(abc.ABC):\\ndiff --git a/ibis/backends/base/sql/compiler/query_builder.py b/ibis/backends/base/sql/compiler/query_builder.py\\nindex a2d5214..95f5e8d 100644\\n--- a/ibis/backends/base/sql/compiler/query_builder.py\\n+++ b/ibis/backends/base/sql/compiler/query_builder.py\\n@@ -8,7 +8,7 @@ import toolz\\n import ibis.common.exceptions as com\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.compiler.base import DML, QueryAST, SetOp\\n from ibis.backends.base.sql.compiler.select_builder import SelectBuilder, _LimitSpec\\n from ibis.backends.base.sql.compiler.translator import ExprTranslator, QueryContext\\ndiff --git a/ibis/backends/base/sql/registry/main.py b/ibis/backends/base/sql/registry/main.py\\nindex 77f70a5..586ace5 100644\\n--- a/ibis/backends/base/sql/registry/main.py\\n+++ b/ibis/backends/base/sql/registry/main.py\\n@@ -4,7 +4,7 @@ import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.registry import (\\n aggregate,\\n binary_infix,\\ndiff --git a/ibis/backends/base/sql/registry/timestamp.py b/ibis/backends/base/sql/registry/timestamp.py\\nindex 412eab1..3c8571f 100644\\n--- a/ibis/backends/base/sql/registry/timestamp.py\\n+++ b/ibis/backends/base/sql/registry/timestamp.py\\n@@ -3,7 +3,7 @@ from __future__ import annotations\\n import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n-import ibis.util as util\\n+from ibis import util\\n \\n \\n def extract_field(sql_attr):\\ndiff --git a/ibis/backends/clickhouse/tests/test_client.py b/ibis/backends/clickhouse/tests/test_client.py\\nindex 8db6672..bb1b9ba 100644\\n--- a/ibis/backends/clickhouse/tests/test_client.py\\n+++ b/ibis/backends/clickhouse/tests/test_client.py\\n@@ -3,9 +3,9 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.config as config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n+from ibis import config\\n from ibis.backends.clickhouse.tests.conftest import (\\n CLICKHOUSE_HOST,\\n CLICKHOUSE_PASS,\\ndiff --git a/ibis/backends/conftest.py b/ibis/backends/conftest.py\\nindex 3a974da..ba7ad75 100644\\n--- a/ibis/backends/conftest.py\\n+++ b/ibis/backends/conftest.py\\n@@ -20,7 +20,7 @@ if TYPE_CHECKING:\\n import pytest\\n \\n import ibis\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base import _get_backend_names\\n \\n TEST_TABLES = {\\ndiff --git a/ibis/backends/dask/execution/util.py b/ibis/backends/dask/execution/util.py\\nindex 61bff7e..7ed0c10 100644\\n--- a/ibis/backends/dask/execution/util.py\\n+++ b/ibis/backends/dask/execution/util.py\\n@@ -9,13 +9,13 @@ import pandas as pd\\n from dask.dataframe.groupby import SeriesGroupBy\\n \\n import ibis.backends.pandas.execution.util as pd_util\\n-import ibis.common.graph as graph\\n import ibis.expr.analysis as an\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n import ibis.util\\n from ibis.backends.dask.core import execute\\n from ibis.backends.pandas.trace import TraceTwoLevelDispatcher\\n+from ibis.common import graph\\n from ibis.expr.scope import Scope\\n \\n if TYPE_CHECKING:\\ndiff --git a/ibis/backends/duckdb/datatypes.py b/ibis/backends/duckdb/datatypes.py\\nindex fd6b8f5..52c0719 100644\\n--- a/ibis/backends/duckdb/datatypes.py\\n+++ b/ibis/backends/duckdb/datatypes.py\\n@@ -3,7 +3,7 @@ from __future__ import annotations\\n import parsy as p\\n import toolz\\n \\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.parsing import (\\n COMMA,\\n FIELD,\\ndiff --git a/ibis/backends/impala/__init__.py b/ibis/backends/impala/__init__.py\\nindex 4ad2057..8299a28 100644\\n--- a/ibis/backends/impala/__init__.py\\n+++ b/ibis/backends/impala/__init__.py\\n@@ -20,7 +20,7 @@ import ibis.config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.rules as rlz\\n import ibis.expr.schema as sch\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql import BaseSQLBackend\\n from ibis.backends.base.sql.ddl import (\\n CTAS,\\ndiff --git a/ibis/backends/impala/client.py b/ibis/backends/impala/client.py\\nindex 6655ce7..78d526f 100644\\n--- a/ibis/backends/impala/client.py\\n+++ b/ibis/backends/impala/client.py\\n@@ -10,7 +10,7 @@ import sqlalchemy as sa\\n import ibis.common.exceptions as com\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base import Database\\n from ibis.backends.base.sql.compiler import DDL, DML\\n from ibis.backends.base.sql.ddl import (\\ndiff --git a/ibis/backends/impala/pandas_interop.py b/ibis/backends/impala/pandas_interop.py\\nindex f410a8b..e687884 100644\\n--- a/ibis/backends/impala/pandas_interop.py\\n+++ b/ibis/backends/impala/pandas_interop.py\\n@@ -22,7 +22,7 @@ from posixpath import join as pjoin\\n import ibis.backends.pandas.client # noqa: F401\\n import ibis.common.exceptions as com\\n import ibis.expr.schema as sch\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.config import options\\n \\n \\ndiff --git a/ibis/backends/impala/tests/conftest.py b/ibis/backends/impala/tests/conftest.py\\nindex 1075ebe..a815be5 100644\\n--- a/ibis/backends/impala/tests/conftest.py\\n+++ b/ibis/backends/impala/tests/conftest.py\\n@@ -13,8 +13,7 @@ import pytest\\n \\n import ibis\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n-from ibis import options\\n+from ibis import options, util\\n from ibis.backends.base import BaseBackend\\n from ibis.backends.conftest import TEST_TABLES, _random_identifier\\n from ibis.backends.impala.compiler import ImpalaCompiler, ImpalaExprTranslator\\ndiff --git a/ibis/backends/impala/tests/test_client.py b/ibis/backends/impala/tests/test_client.py\\nindex 0b56054..3fcca3a 100644\\n--- a/ibis/backends/impala/tests/test_client.py\\n+++ b/ibis/backends/impala/tests/test_client.py\\n@@ -7,9 +7,9 @@ import pytz\\n \\n import ibis\\n import ibis.common.exceptions as com\\n-import ibis.config as config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n+from ibis import config\\n from ibis.tests.util import assert_equal\\n \\n pytest.importorskip(\\\"impala\\\")\\ndiff --git a/ibis/backends/impala/tests/test_ddl.py b/ibis/backends/impala/tests/test_ddl.py\\nindex 870c4dc..2346a3d 100644\\n--- a/ibis/backends/impala/tests/test_ddl.py\\n+++ b/ibis/backends/impala/tests/test_ddl.py\\n@@ -6,7 +6,7 @@ import ibis\\n import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.ddl import fully_qualified_re\\n from ibis.tests.util import assert_equal\\n \\ndiff --git a/ibis/backends/impala/tests/test_exprs.py b/ibis/backends/impala/tests/test_exprs.py\\nindex cfc8552..1d6f44f 100644\\n--- a/ibis/backends/impala/tests/test_exprs.py\\n+++ b/ibis/backends/impala/tests/test_exprs.py\\n@@ -5,10 +5,10 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.types as ir\\n from ibis import literal as L\\n from ibis.backends.impala.compiler import ImpalaCompiler\\n+from ibis.expr import api\\n from ibis.expr.datatypes import Category\\n \\n \\ndiff --git a/ibis/backends/impala/tests/test_partition.py b/ibis/backends/impala/tests/test_partition.py\\nindex 1f96e7d..44217a4 100644\\n--- a/ibis/backends/impala/tests/test_partition.py\\n+++ b/ibis/backends/impala/tests/test_partition.py\\n@@ -6,7 +6,7 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.tests.util import assert_equal\\n \\n pytest.importorskip(\\\"impala\\\")\\ndiff --git a/ibis/backends/impala/tests/test_udf.py b/ibis/backends/impala/tests/test_udf.py\\nindex 895918b..fd950d5 100644\\n--- a/ibis/backends/impala/tests/test_udf.py\\n+++ b/ibis/backends/impala/tests/test_udf.py\\n@@ -9,11 +9,11 @@ import ibis\\n import ibis.backends.impala as api\\n import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n-import ibis.expr.rules as rules\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.impala import ddl\\n from ibis.common.exceptions import IbisTypeError\\n+from ibis.expr import rules\\n \\n pytest.importorskip(\\\"impala\\\")\\n \\ndiff --git a/ibis/backends/impala/udf.py b/ibis/backends/impala/udf.py\\nindex c6f2ef6..8b8b552 100644\\n--- a/ibis/backends/impala/udf.py\\n+++ b/ibis/backends/impala/udf.py\\n@@ -21,7 +21,7 @@ import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.udf.validate as v\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql.registry import fixed_arity, sql_type_names\\n from ibis.backends.impala.compiler import ImpalaExprTranslator\\n \\ndiff --git a/ibis/backends/mysql/__init__.py b/ibis/backends/mysql/__init__.py\\nindex c0ddacb..50b331a 100644\\n--- a/ibis/backends/mysql/__init__.py\\n+++ b/ibis/backends/mysql/__init__.py\\n@@ -8,7 +8,7 @@ import warnings\\n from typing import Literal\\n \\n import sqlalchemy as sa\\n-import sqlalchemy.dialects.mysql as mysql\\n+from sqlalchemy.dialects import mysql\\n \\n import ibis.expr.datatypes as dt\\n import ibis.expr.schema as sch\\ndiff --git a/ibis/backends/mysql/compiler.py b/ibis/backends/mysql/compiler.py\\nindex 13819cb..7456f71 100644\\n--- a/ibis/backends/mysql/compiler.py\\n+++ b/ibis/backends/mysql/compiler.py\\n@@ -1,7 +1,7 @@\\n from __future__ import annotations\\n \\n import sqlalchemy as sa\\n-import sqlalchemy.dialects.mysql as mysql\\n+from sqlalchemy.dialects import mysql\\n \\n import ibis.expr.datatypes as dt\\n from ibis.backends.base.sql.alchemy import AlchemyCompiler, AlchemyExprTranslator\\ndiff --git a/ibis/backends/postgres/tests/test_functions.py b/ibis/backends/postgres/tests/test_functions.py\\nindex 33c6d2e..0f377e3 100644\\n--- a/ibis/backends/postgres/tests/test_functions.py\\n+++ b/ibis/backends/postgres/tests/test_functions.py\\n@@ -11,9 +11,9 @@ import pytest\\n from pytest import param\\n \\n import ibis\\n-import ibis.config as config\\n import ibis.expr.datatypes as dt\\n import ibis.expr.types as ir\\n+from ibis import config\\n from ibis import literal as L\\n from ibis.expr.window import rows_with_max_lookback\\n \\ndiff --git a/ibis/backends/pyspark/__init__.py b/ibis/backends/pyspark/__init__.py\\nindex 1b42080..b994911 100644\\n--- a/ibis/backends/pyspark/__init__.py\\n+++ b/ibis/backends/pyspark/__init__.py\\n@@ -14,8 +14,7 @@ import ibis.config\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.expr.types as types\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.backends.base.sql import BaseSQLBackend\\n from ibis.backends.base.sql.compiler import Compiler, TableSetFormatter\\n from ibis.backends.base.sql.ddl import (\\n@@ -217,16 +216,16 @@ class Backend(BaseSQLBackend):\\n **kwargs: Any,\\n ) -> Any:\\n \\\"\\\"\\\"Execute an expression.\\\"\\\"\\\"\\n- if isinstance(expr, types.Table):\\n+ if isinstance(expr, ir.Table):\\n return self.compile(expr, timecontext, params, **kwargs).toPandas()\\n- elif isinstance(expr, types.Column):\\n+ elif isinstance(expr, ir.Column):\\n # expression must be named for the projection\\n if not expr.has_name():\\n expr = expr.name(\\\"tmp\\\")\\n return self.compile(\\n expr.to_projection(), timecontext, params, **kwargs\\n ).toPandas()[expr.get_name()]\\n- elif isinstance(expr, types.Scalar):\\n+ elif isinstance(expr, ir.Scalar):\\n compiled = self.compile(expr, timecontext, params, **kwargs)\\n if isinstance(compiled, Column):\\n # attach result column to a fake DataFrame and\\ndiff --git a/ibis/backends/pyspark/tests/test_ddl.py b/ibis/backends/pyspark/tests/test_ddl.py\\nindex 0288062..ccc8a97 100644\\n--- a/ibis/backends/pyspark/tests/test_ddl.py\\n+++ b/ibis/backends/pyspark/tests/test_ddl.py\\n@@ -5,7 +5,7 @@ import pytest\\n \\n import ibis\\n import ibis.common.exceptions as com\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.tests.util import assert_equal\\n \\n pyspark = pytest.importorskip(\\\"pyspark\\\")\\ndiff --git a/ibis/backends/sqlite/tests/test_client.py b/ibis/backends/sqlite/tests/test_client.py\\nindex 95aa24d..ad64700 100644\\n--- a/ibis/backends/sqlite/tests/test_client.py\\n+++ b/ibis/backends/sqlite/tests/test_client.py\\n@@ -5,8 +5,8 @@ import pandas.testing as tm\\n import pytest\\n \\n import ibis\\n-import ibis.config as config\\n import ibis.expr.types as ir\\n+from ibis import config\\n \\n pytest.importorskip(\\\"sqlalchemy\\\")\\n \\ndiff --git a/ibis/expr/format.py b/ibis/expr/format.py\\nindex e3d48cd..85fab3f 100644\\n--- a/ibis/expr/format.py\\n+++ b/ibis/expr/format.py\\n@@ -9,13 +9,13 @@ from typing import Any, Callable, Deque, Iterable, Mapping, Tuple\\n import rich.pretty\\n \\n import ibis\\n-import ibis.common.graph as graph\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n import ibis.expr.window as win\\n-import ibis.util as util\\n+from ibis import util\\n+from ibis.common import graph\\n \\n Aliases = Mapping[ops.TableNode, int]\\n Deps = Deque[Tuple[int, ops.TableNode]]\\ndiff --git a/ibis/expr/operations/relations.py b/ibis/expr/operations/relations.py\\nindex 080ddcd..de44a15 100644\\n--- a/ibis/expr/operations/relations.py\\n+++ b/ibis/expr/operations/relations.py\\n@@ -11,7 +11,7 @@ import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.annotations import attribute\\n from ibis.expr.deferred import Deferred\\n from ibis.expr.operations.core import Named, Node, Value\\ndiff --git a/ibis/expr/rules.py b/ibis/expr/rules.py\\nindex 9b1a3b7..d40700e 100644\\n--- a/ibis/expr/rules.py\\n+++ b/ibis/expr/rules.py\\n@@ -11,7 +11,7 @@ import ibis.common.exceptions as com\\n import ibis.expr.datatypes as dt\\n import ibis.expr.schema as sch\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.annotations import attribute, optional\\n from ibis.common.validators import (\\n bool_,\\ndiff --git a/ibis/expr/timecontext.py b/ibis/expr/timecontext.py\\nindex 7ecd8e7..9620d6c 100644\\n--- a/ibis/expr/timecontext.py\\n+++ b/ibis/expr/timecontext.py\\n@@ -38,8 +38,8 @@ from typing import TYPE_CHECKING, Any\\n import numpy as np\\n \\n import ibis.common.exceptions as com\\n-import ibis.config as config\\n import ibis.expr.operations as ops\\n+from ibis import config\\n \\n if TYPE_CHECKING:\\n import pandas as pd\\ndiff --git a/ibis/expr/types/groupby.py b/ibis/expr/types/groupby.py\\nindex 138f92e..97aaaa2 100644\\n--- a/ibis/expr/types/groupby.py\\n+++ b/ibis/expr/types/groupby.py\\n@@ -22,7 +22,7 @@ from typing import Iterable, Sequence\\n import ibis.expr.analysis as an\\n import ibis.expr.types as ir\\n import ibis.expr.window as _window\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.expr.deferred import Deferred\\n \\n _function_types = tuple(\\ndiff --git a/ibis/expr/window.py b/ibis/expr/window.py\\nindex 5ef3bb1..3e0efdc 100644\\n--- a/ibis/expr/window.py\\n+++ b/ibis/expr/window.py\\n@@ -11,7 +11,7 @@ import toolz\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.types as ir\\n-import ibis.util as util\\n+from ibis import util\\n from ibis.common.exceptions import IbisInputError\\n from ibis.common.grounds import Comparable\\n \\ndiff --git a/ibis/tests/expr/test_decimal.py b/ibis/tests/expr/test_decimal.py\\nindex 85d8eb2..12b809b 100644\\n--- a/ibis/tests/expr/test_decimal.py\\n+++ b/ibis/tests/expr/test_decimal.py\\n@@ -3,10 +3,10 @@ import operator\\n import pytest\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n+from ibis.expr import api\\n \\n \\n def test_type_metadata(lineitem):\\ndiff --git a/ibis/tests/expr/test_interactive.py b/ibis/tests/expr/test_interactive.py\\nindex cea1945..0c5613b 100644\\n--- a/ibis/tests/expr/test_interactive.py\\n+++ b/ibis/tests/expr/test_interactive.py\\n@@ -14,7 +14,7 @@\\n \\n import pytest\\n \\n-import ibis.config as config\\n+from ibis import config\\n from ibis.tests.expr.mocks import MockBackend\\n \\n \\ndiff --git a/ibis/tests/expr/test_table.py b/ibis/tests/expr/test_table.py\\nindex 04f4a7d..3f77985 100644\\n--- a/ibis/tests/expr/test_table.py\\n+++ b/ibis/tests/expr/test_table.py\\n@@ -10,13 +10,13 @@ from pytest import param\\n import ibis\\n import ibis.common.exceptions as com\\n import ibis.expr.analysis as an\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n from ibis import _\\n from ibis import literal as L\\n from ibis.common.exceptions import RelationError\\n+from ibis.expr import api\\n from ibis.expr.types import Column, Table\\n from ibis.tests.expr.mocks import MockAlchemyBackend, MockBackend\\n from ibis.tests.util import assert_equal, assert_pickle_roundtrip\\ndiff --git a/ibis/tests/expr/test_temporal.py b/ibis/tests/expr/test_temporal.py\\nindex e76e71c..9a0f43f 100644\\n--- a/ibis/tests/expr/test_temporal.py\\n+++ b/ibis/tests/expr/test_temporal.py\\n@@ -5,10 +5,10 @@ import pytest\\n from pytest import param\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.types as ir\\n+from ibis.expr import api\\n \\n \\n def test_temporal_literals():\\ndiff --git a/ibis/tests/expr/test_timestamp.py b/ibis/tests/expr/test_timestamp.py\\nindex 6601c8b..7782787 100644\\n--- a/ibis/tests/expr/test_timestamp.py\\n+++ b/ibis/tests/expr/test_timestamp.py\\n@@ -5,11 +5,11 @@ import pandas as pd\\n import pytest\\n \\n import ibis\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.types as ir\\n+from ibis.expr import api\\n \\n \\n def test_field_select(alltypes):\\ndiff --git a/ibis/tests/expr/test_value_exprs.py b/ibis/tests/expr/test_value_exprs.py\\nindex 4c3d475..9eb247c 100644\\n--- a/ibis/tests/expr/test_value_exprs.py\\n+++ b/ibis/tests/expr/test_value_exprs.py\\n@@ -15,13 +15,13 @@ from pytest import param\\n import ibis\\n import ibis.common.exceptions as com\\n import ibis.expr.analysis as L\\n-import ibis.expr.api as api\\n import ibis.expr.datatypes as dt\\n import ibis.expr.operations as ops\\n import ibis.expr.rules as rlz\\n import ibis.expr.types as ir\\n from ibis import _, literal\\n from ibis.common.exceptions import IbisTypeError\\n+from ibis.expr import api\\n from ibis.tests.util import assert_equal\\n \\n \\ndiff --git a/ibis/tests/expr/test_visualize.py b/ibis/tests/expr/test_visualize.py\\nindex 5525944..253564f 100644\\n--- a/ibis/tests/expr/test_visualize.py\\n+++ b/ibis/tests/expr/test_visualize.py\\n@@ -9,8 +9,8 @@ import ibis.expr.types as ir\\n \\n pytest.importorskip('graphviz')\\n \\n-import ibis.expr.api as api # noqa: E402\\n import ibis.expr.visualize as viz # noqa: E402\\n+from ibis.expr import api # noqa: E402\\n \\n pytestmark = pytest.mark.skipif(\\n int(os.environ.get('CONDA_BUILD', 0)) == 1, reason='CONDA_BUILD defined'\\ndiff --git a/ibis/tests/sql/test_sqlalchemy.py b/ibis/tests/sql/test_sqlalchemy.py\\nindex 2ad5453..3aa8c3d 100644\\n--- a/ibis/tests/sql/test_sqlalchemy.py\\n+++ b/ibis/tests/sql/test_sqlalchemy.py\\n@@ -15,8 +15,8 @@\\n import operator\\n \\n import pytest\\n-import sqlalchemy.sql as sql\\n from sqlalchemy import func as F\\n+from sqlalchemy import sql\\n from sqlalchemy import types as sat\\n \\n import ibis\\ndiff --git a/ibis/tests/util.py b/ibis/tests/util.py\\nindex f79d09a..025bfc7 100644\\n--- a/ibis/tests/util.py\\n+++ b/ibis/tests/util.py\\n@@ -5,7 +5,7 @@ from __future__ import annotations\\n import pickle\\n \\n import ibis\\n-import ibis.util as util\\n+from ibis import util\\n \\n \\n def assert_equal(left, right):\\ndiff --git a/pyproject.toml b/pyproject.toml\\nindex f2146d4..492ad9e 100644\\n--- a/pyproject.toml\\n+++ b/pyproject.toml\\n@@ -310,6 +310,7 @@ select = [\\n \\\"PGH\\\", # pygrep-hooks\\n \\\"PLC\\\", # pylint\\n \\\"PLE\\\", # pylint\\n+ \\\"PLR\\\", # pylint import style\\n \\\"PLW\\\", # pylint\\n \\\"RET\\\", # flake8-return\\n \\\"RUF\\\", # ruff-specific rules\\n\", \"diff --git a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\nindex da33c23..23c43be 100644\\n--- a/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\n+++ b/test-util/src/main/java/io/camunda/zeebe/test/util/bpmn/random/ExecutionPathSegment.java\\n@@ -29,6 +29,10 @@ import org.apache.commons.lang3.builder.ToStringStyle;\\n */\\n public final class ExecutionPathSegment {\\n \\n+ // If we have reached a terminate end event we want to stop generating execution steps for a\\n+ // specific flow scope. By setting this flag to true no new execution steps will be added for the\\n+ // flow scope this segment is in.\\n+ private boolean reachedTerminateEndEvent = false;\\n private final List scheduledSteps = new ArrayList<>();\\n private final Map variableDefaults = new HashMap<>();\\n \\n@@ -87,10 +91,28 @@ public final class ExecutionPathSegment {\\n new ScheduledExecutionStep(logicalPredecessor, executionPredecessor, executionStep));\\n }\\n \\n+ /**\\n+ * Appends the steps of the passed execution path segment to the current segment.\\n+ *\\n+ * @param pathToAdd execution path segment to append to this segment\\n+ */\\n public void append(final ExecutionPathSegment pathToAdd) {\\n+ append(pathToAdd, false);\\n+ }\\n+\\n+ /**\\n+ * Appends the step of the passed execution path segment to the current segment if the current\\n+ *\\n+ * @param pathToAdd\\n+ * @param changesFlowScope\\n+ */\\n+ public void append(final ExecutionPathSegment pathToAdd, final boolean changesFlowScope) {\\n mergeVariableDefaults(pathToAdd);\\n \\n- pathToAdd.getScheduledSteps().forEach(this::append);\\n+ if (!hasReachedTerminateEndEvent() || changesFlowScope) {\\n+ pathToAdd.getScheduledSteps().forEach(this::append);\\n+ }\\n+ reachedTerminateEndEvent = pathToAdd.hasReachedTerminateEndEvent() && !changesFlowScope;\\n }\\n \\n public void append(final ScheduledExecutionStep scheduledExecutionStep) {\\n@@ -259,6 +281,14 @@ public final class ExecutionPathSegment {\\n return ToStringBuilder.reflectionToString(this, ToStringStyle.SHORT_PREFIX_STYLE);\\n }\\n \\n+ public boolean hasReachedTerminateEndEvent() {\\n+ return reachedTerminateEndEvent;\\n+ }\\n+\\n+ public void setReachedTerminateEndEvent(final boolean reachedTerminateEndEvent) {\\n+ this.reachedTerminateEndEvent = reachedTerminateEndEvent;\\n+ }\\n+\\n /**\\n * An execution boundary is the point where automatic and non-automatic {@link\\n * ScheduledExecutionStep}'s meet each other. This class contains information about the existing\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"f8c7b34bdeedcf1a4628cd50b23920afeaf57cb6\", \"78c446cbea61af2268b4c4da03a9ad4283f10049\", \"8d53d724275ebe4b2a0bb0bd7e2c2dfc399e049b\", \"40597fb4de41c7194eb99479a914db70da7909ea\"]"},"types":{"kind":"string","value":"[\"ci\", \"build\", \"refactor\", \"feat\"]"}}},{"rowIdx":1091,"cells":{"commit_message":{"kind":"string","value":"add react ecosystem,stop playing audio on panel close\n\nCloses #824,replace tuple with record,Add the select function for logicflow"},"diff":{"kind":"string","value":"[\"diff --git a/package.json b/package.json\\nindex 1ba8c4f..d1de9a0 100644\\n--- a/package.json\\n+++ b/package.json\\n@@ -36,14 +36,19 @@\\n \\\"@types/node\\\": \\\"^9.3.0\\\",\\n \\\"@types/react\\\": \\\"^16.0.34\\\",\\n \\\"@types/react-dom\\\": \\\"^16.0.3\\\",\\n+ \\\"@types/react-motion\\\": \\\"^0.0.25\\\",\\n \\\"bootstrap-sass\\\": \\\"^3.3.7\\\",\\n \\\"highcharts\\\": \\\"^6.0.4\\\",\\n \\\"html2canvas\\\": \\\"^1.0.0-alpha.9\\\",\\n+ \\\"immer\\\": \\\"^1.2.1\\\",\\n \\\"lodash\\\": \\\"^4.17.4\\\",\\n \\\"moment\\\": \\\"^2.20.1\\\",\\n \\\"normalize.css\\\": \\\"^8.0.0\\\",\\n- \\\"react\\\": \\\"^16.2.0\\\",\\n- \\\"react-dom\\\": \\\"^16.2.0\\\",\\n+ \\\"react\\\": \\\"^16.3.1\\\",\\n+ \\\"react-dom\\\": \\\"^16.3.1\\\",\\n+ \\\"react-motion\\\": \\\"^0.5.2\\\",\\n+ \\\"react-redux\\\": \\\"^5.0.7\\\",\\n+ \\\"redux\\\": \\\"^3.7.2\\\",\\n \\\"rxjs\\\": \\\"^5.5.6\\\",\\n \\\"vue\\\": \\\"^2.5.13\\\",\\n \\\"vue-plugin-webextension-i18n\\\": \\\"^0.1.0\\\",\\ndiff --git a/yarn.lock b/yarn.lock\\nindex c8898d8..5d0fc9f 100644\\n--- a/yarn.lock\\n+++ b/yarn.lock\\n@@ -187,6 +187,12 @@\\n \\\"@types/node\\\" \\\"*\\\"\\n \\\"@types/react\\\" \\\"*\\\"\\n \\n+\\\"@types/react-motion@^0.0.25\\\":\\n+ version \\\"0.0.25\\\"\\n+ resolved \\\"https://registry.npmjs.org/@types/react-motion/-/react-motion-0.0.25.tgz#2445745ee8e8e6149faa47a36ff6b0d4c21dbf94\\\"\\n+ dependencies:\\n+ \\\"@types/react\\\" \\\"*\\\"\\n+\\n \\\"@types/react@*\\\", \\\"@types/react@^16.0.34\\\":\\n version \\\"16.0.40\\\"\\n resolved \\\"https://registry.npmjs.org/@types/react/-/react-16.0.40.tgz#caabc2296886f40b67f6fc80f0f3464476461df9\\\"\\n@@ -3837,6 +3843,10 @@ hoek@4.x.x:\\n version \\\"4.2.1\\\"\\n resolved \\\"https://registry.npmjs.org/hoek/-/hoek-4.2.1.tgz#9634502aa12c445dd5a7c5734b572bb8738aacbb\\\"\\n \\n+hoist-non-react-statics@^2.5.0:\\n+ version \\\"2.5.0\\\"\\n+ resolved \\\"https://registry.npmjs.org/hoist-non-react-statics/-/hoist-non-react-statics-2.5.0.tgz#d2ca2dfc19c5a91c5a6615ce8e564ef0347e2a40\\\"\\n+\\n home-or-tmp@^2.0.0:\\n version \\\"2.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/home-or-tmp/-/home-or-tmp-2.0.0.tgz#e36c3f2d2cae7d746a857e38d18d5f32a7882db8\\\"\\n@@ -4004,6 +4014,10 @@ ignore@^3.3.5:\\n version \\\"3.3.7\\\"\\n resolved \\\"https://registry.npmjs.org/ignore/-/ignore-3.3.7.tgz#612289bfb3c220e186a58118618d5be8c1bab021\\\"\\n \\n+immer@^1.2.1:\\n+ version \\\"1.2.1\\\"\\n+ resolved \\\"https://registry.npmjs.org/immer/-/immer-1.2.1.tgz#96e2ae29cdfc428f28120b832701931b92fa597c\\\"\\n+\\n import-local@^1.0.0:\\n version \\\"1.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/import-local/-/import-local-1.0.0.tgz#5e4ffdc03f4fe6c009c6729beb29631c2f8227bc\\\"\\n@@ -4104,7 +4118,7 @@ interpret@^1.0.0:\\n version \\\"1.1.0\\\"\\n resolved \\\"https://registry.npmjs.org/interpret/-/interpret-1.1.0.tgz#7ed1b1410c6a0e0f78cf95d3b8440c63f78b8614\\\"\\n \\n-invariant@^2.2.2:\\n+invariant@^2.0.0, invariant@^2.2.2:\\n version \\\"2.2.4\\\"\\n resolved \\\"https://registry.npmjs.org/invariant/-/invariant-2.2.4.tgz#610f3c92c9359ce1db616e538008d23ff35158e6\\\"\\n dependencies:\\n@@ -5040,6 +5054,10 @@ locate-path@^2.0.0:\\n p-locate \\\"^2.0.0\\\"\\n path-exists \\\"^3.0.0\\\"\\n \\n+lodash-es@^4.17.5, lodash-es@^4.2.1:\\n+ version \\\"4.17.8\\\"\\n+ resolved \\\"https://registry.npmjs.org/lodash-es/-/lodash-es-4.17.8.tgz#6fa8c8c5d337481df0bdf1c0d899d42473121e45\\\"\\n+\\n lodash._reinterpolate@~3.0.0:\\n version \\\"3.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/lodash._reinterpolate/-/lodash._reinterpolate-3.0.0.tgz#0ccf2d89166af03b3663c796538b75ac6e114d9d\\\"\\n@@ -5149,7 +5167,7 @@ lodash@4.17.2:\\n version \\\"4.17.2\\\"\\n resolved \\\"https://registry.npmjs.org/lodash/-/lodash-4.17.2.tgz#34a3055babe04ce42467b607d700072c7ff6bf42\\\"\\n \\n-lodash@4.x, lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\\n+lodash@4.x, lodash@^4.0.0, lodash@^4.13.1, lodash@^4.14.0, lodash@^4.16.3, lodash@^4.17.2, lodash@^4.17.3, lodash@^4.17.4, lodash@^4.17.5, lodash@^4.2.0, lodash@^4.2.1, lodash@^4.3.0, lodash@~4.17.4:\\n version \\\"4.17.5\\\"\\n resolved \\\"https://registry.npmjs.org/lodash/-/lodash-4.17.5.tgz#99a92d65c0272debe8c96b6057bc8fbfa3bed511\\\"\\n \\n@@ -6467,7 +6485,7 @@ promise@^7.1.1:\\n dependencies:\\n asap \\\"~2.0.3\\\"\\n \\n-prop-types@^15.6.0:\\n+prop-types@^15.5.8, prop-types@^15.6.0:\\n version \\\"15.6.1\\\"\\n resolved \\\"https://registry.npmjs.org/prop-types/-/prop-types-15.6.1.tgz#36644453564255ddda391191fb3a125cbdf654ca\\\"\\n dependencies:\\n@@ -6574,7 +6592,7 @@ quick-lru@^1.0.0:\\n version \\\"1.1.0\\\"\\n resolved \\\"https://registry.npmjs.org/quick-lru/-/quick-lru-1.1.0.tgz#4360b17c61136ad38078397ff11416e186dcfbb8\\\"\\n \\n-raf@3.4.0:\\n+raf@3.4.0, raf@^3.1.0:\\n version \\\"3.4.0\\\"\\n resolved \\\"https://registry.npmjs.org/raf/-/raf-3.4.0.tgz#a28876881b4bc2ca9117d4138163ddb80f781575\\\"\\n dependencies:\\n@@ -6645,9 +6663,9 @@ react-dev-utils@^5.0.0:\\n strip-ansi \\\"3.0.1\\\"\\n text-table \\\"0.2.0\\\"\\n \\n-react-dom@^16.2.0:\\n- version \\\"16.2.0\\\"\\n- resolved \\\"https://registry.npmjs.org/react-dom/-/react-dom-16.2.0.tgz#69003178601c0ca19b709b33a83369fe6124c044\\\"\\n+react-dom@^16.3.1:\\n+ version \\\"16.3.1\\\"\\n+ resolved \\\"https://registry.npmjs.org/react-dom/-/react-dom-16.3.1.tgz#6a3c90a4fb62f915bdbcf6204422d93a7d4ca573\\\"\\n dependencies:\\n fbjs \\\"^0.8.16\\\"\\n loose-envify \\\"^1.1.0\\\"\\n@@ -6658,9 +6676,28 @@ react-error-overlay@^4.0.0:\\n version \\\"4.0.0\\\"\\n resolved \\\"https://registry.npmjs.org/react-error-overlay/-/react-error-overlay-4.0.0.tgz#d198408a85b4070937a98667f500c832f86bd5d4\\\"\\n \\n-react@^16.2.0:\\n- version \\\"16.2.0\\\"\\n- resolved \\\"https://registry.npmjs.org/react/-/react-16.2.0.tgz#a31bd2dab89bff65d42134fa187f24d054c273ba\\\"\\n+react-motion@^0.5.2:\\n+ version \\\"0.5.2\\\"\\n+ resolved \\\"https://registry.npmjs.org/react-motion/-/react-motion-0.5.2.tgz#0dd3a69e411316567927917c6626551ba0607316\\\"\\n+ dependencies:\\n+ performance-now \\\"^0.2.0\\\"\\n+ prop-types \\\"^15.5.8\\\"\\n+ raf \\\"^3.1.0\\\"\\n+\\n+react-redux@^5.0.7:\\n+ version \\\"5.0.7\\\"\\n+ resolved \\\"https://registry.npmjs.org/react-redux/-/react-redux-5.0.7.tgz#0dc1076d9afb4670f993ffaef44b8f8c1155a4c8\\\"\\n+ dependencies:\\n+ hoist-non-react-statics \\\"^2.5.0\\\"\\n+ invariant \\\"^2.0.0\\\"\\n+ lodash \\\"^4.17.5\\\"\\n+ lodash-es \\\"^4.17.5\\\"\\n+ loose-envify \\\"^1.1.0\\\"\\n+ prop-types \\\"^15.6.0\\\"\\n+\\n+react@^16.3.1:\\n+ version \\\"16.3.1\\\"\\n+ resolved \\\"https://registry.npmjs.org/react/-/react-16.3.1.tgz#4a2da433d471251c69b6033ada30e2ed1202cfd8\\\"\\n dependencies:\\n fbjs \\\"^0.8.16\\\"\\n loose-envify \\\"^1.1.0\\\"\\n@@ -6788,6 +6825,15 @@ reduce-function-call@^1.0.1:\\n dependencies:\\n balanced-match \\\"^0.4.2\\\"\\n \\n+redux@^3.7.2:\\n+ version \\\"3.7.2\\\"\\n+ resolved \\\"https://registry.npmjs.org/redux/-/redux-3.7.2.tgz#06b73123215901d25d065be342eb026bc1c8537b\\\"\\n+ dependencies:\\n+ lodash \\\"^4.2.1\\\"\\n+ lodash-es \\\"^4.2.1\\\"\\n+ loose-envify \\\"^1.1.0\\\"\\n+ symbol-observable \\\"^1.0.3\\\"\\n+\\n regenerate@^1.2.1:\\n version \\\"1.3.3\\\"\\n resolved \\\"https://registry.npmjs.org/regenerate/-/regenerate-1.3.3.tgz#0c336d3980553d755c39b586ae3b20aa49c82b7f\\\"\\n@@ -7811,6 +7857,10 @@ symbol-observable@1.0.1:\\n version \\\"1.0.1\\\"\\n resolved \\\"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.0.1.tgz#8340fc4702c3122df5d22288f88283f513d3fdd4\\\"\\n \\n+symbol-observable@^1.0.3:\\n+ version \\\"1.2.0\\\"\\n+ resolved \\\"https://registry.npmjs.org/symbol-observable/-/symbol-observable-1.2.0.tgz#c22688aed4eab3cdc2dfeacbb561660560a00804\\\"\\n+\\n symbol-tree@^3.2.2:\\n version \\\"3.2.2\\\"\\n resolved \\\"https://registry.npmjs.org/symbol-tree/-/symbol-tree-3.2.2.tgz#ae27db38f660a7ae2e1c3b7d1bc290819b8519e6\\\"\\n\", \"diff --git a/src/background/audio-manager.ts b/src/background/audio-manager.ts\\nindex 84032f1..9e116fc 100644\\n--- a/src/background/audio-manager.ts\\n+++ b/src/background/audio-manager.ts\\n@@ -1,4 +1,4 @@\\n-import { timeout } from '@/_helpers/promise-more'\\n+import { timer } from '@/_helpers/promise-more'\\n \\n /**\\n * To make sure only one audio plays at a time\\n@@ -16,6 +16,8 @@ export class AudioManager {\\n \\n private audio?: HTMLAudioElement\\n \\n+ currentSrc?: string\\n+\\n reset() {\\n if (this.audio) {\\n this.audio.pause()\\n@@ -23,28 +25,33 @@ export class AudioManager {\\n this.audio.src = ''\\n this.audio.onended = null\\n }\\n+ this.currentSrc = ''\\n }\\n \\n load(src: string): HTMLAudioElement {\\n this.reset()\\n+ this.currentSrc = src\\n return (this.audio = new Audio(src))\\n }\\n \\n async play(src?: string): Promise {\\n- if (!src) {\\n+ if (!src || src === this.currentSrc) {\\n this.reset()\\n return\\n }\\n \\n const audio = this.load(src)\\n \\n- const onEnd = new Promise(resolve => {\\n- audio.onended = resolve\\n- })\\n+ const onEnd = Promise.race([\\n+ new Promise(resolve => {\\n+ audio.onended = resolve\\n+ }),\\n+ timer(20000)\\n+ ])\\n+\\n+ await audio.play()\\n+ await onEnd\\n \\n- await audio\\n- .play()\\n- .then(() => timeout(onEnd, 4000))\\n- .catch(() => {})\\n+ this.currentSrc = ''\\n }\\n }\\ndiff --git a/src/background/server.ts b/src/background/server.ts\\nindex 65f6f6c..4c70196 100644\\n--- a/src/background/server.ts\\n+++ b/src/background/server.ts\\n@@ -64,6 +64,9 @@ export class BackgroundServer {\\n return openURL(msg.payload.url, msg.payload.self)\\n case 'PLAY_AUDIO':\\n return AudioManager.getInstance().play(msg.payload)\\n+ case 'STOP_AUDIO':\\n+ AudioManager.getInstance().reset()\\n+ return\\n case 'FETCH_DICT_RESULT':\\n return this.fetchDictResult(msg.payload)\\n case 'DICT_ENGINE_METHOD':\\n@@ -79,6 +82,7 @@ export class BackgroundServer {\\n case 'OPEN_QS_PANEL':\\n return this.openQSPanel()\\n case 'CLOSE_QS_PANEL':\\n+ AudioManager.getInstance().reset()\\n return this.qsPanelManager.destroy()\\n case 'QS_SWITCH_SIDEBAR':\\n return this.qsPanelManager.toggleSidebar(msg.payload)\\n@@ -105,6 +109,16 @@ export class BackgroundServer {\\n return this.youdaoTranslateAjax(msg.payload)\\n }\\n })\\n+\\n+ browser.runtime.onConnect.addListener(port => {\\n+ if (port.name === 'popup') {\\n+ // This is a workaround for browser action page\\n+ // which does not fire beforeunload event\\n+ port.onDisconnect.addListener(() => {\\n+ AudioManager.getInstance().reset()\\n+ })\\n+ }\\n+ })\\n }\\n \\n async openQSPanel(): Promise {\\ndiff --git a/src/content/redux/epics/index.ts b/src/content/redux/epics/index.ts\\nindex b941c07..587b54d 100644\\n--- a/src/content/redux/epics/index.ts\\n+++ b/src/content/redux/epics/index.ts\\n@@ -1,6 +1,6 @@\\n import { combineEpics } from 'redux-observable'\\n import { from, of, EMPTY } from 'rxjs'\\n-import { map, mapTo, mergeMap, filter } from 'rxjs/operators'\\n+import { map, mapTo, mergeMap, filter, pairwise } from 'rxjs/operators'\\n \\n import { isPopupPage, isStandalonePage } from '@/_helpers/saladict'\\n import { saveWord } from '@/_helpers/record-manager'\\n@@ -11,6 +11,7 @@ import { ofType } from './utils'\\n import searchStartEpic from './searchStart.epic'\\n import newSelectionEpic from './newSelection.epic'\\n import { translateCtxs, genCtxText } from '@/_helpers/translateCtx'\\n+import { message } from '@/_helpers/browser-api'\\n \\n export const epics = combineEpics(\\n /** Start searching text. This will also send to Redux. */\\n@@ -28,6 +29,17 @@ export const epics = combineEpics(\\n )\\n ),\\n (action$, state$) =>\\n+ state$.pipe(\\n+ map(state => state.isShowDictPanel),\\n+ pairwise(),\\n+ mergeMap(([oldShow, newShow]) => {\\n+ if (oldShow && !newShow) {\\n+ message.send({ type: 'STOP_AUDIO' })\\n+ }\\n+ return EMPTY\\n+ })\\n+ ),\\n+ (action$, state$) =>\\n action$.pipe(\\n ofType('ADD_TO_NOTEBOOK'),\\n mergeMap(() => {\\ndiff --git a/src/popup/index.tsx b/src/popup/index.tsx\\nindex cbca1c0..a406bfd 100644\\n--- a/src/popup/index.tsx\\n+++ b/src/popup/index.tsx\\n@@ -21,6 +21,10 @@ import Popup from './Popup'\\n import Notebook from './Notebook'\\n import './_style.scss'\\n \\n+// This is a workaround for browser action page\\n+// which does not fire beforeunload event\\n+browser.runtime.connect({ name: 'popup' } as any) // wrong typing\\n+\\n const Title: FC = () => {\\n const { t } = useTranslate('popup')\\n return (\\ndiff --git a/src/typings/message.ts b/src/typings/message.ts\\nindex bdd6fad..63238cb 100644\\n--- a/src/typings/message.ts\\n+++ b/src/typings/message.ts\\n@@ -146,6 +146,8 @@ export type MessageConfig = MessageConfigType<{\\n payload: string\\n }\\n \\n+ STOP_AUDIO: {}\\n+\\n LAST_PLAY_AUDIO: {\\n response?: null | { src: string; timestamp: number }\\n }\\n\", \"diff --git a/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java b/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java\\nindex fa6f8d4..2185b1e 100644\\n--- a/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java\\n+++ b/engine/src/main/java/io/camunda/zeebe/engine/processing/processinstance/CreateProcessInstanceProcessor.java\\n@@ -37,7 +37,6 @@ import io.camunda.zeebe.protocol.record.intent.ProcessInstanceCreationIntent;\\n import io.camunda.zeebe.protocol.record.intent.ProcessInstanceIntent;\\n import io.camunda.zeebe.protocol.record.value.BpmnElementType;\\n import io.camunda.zeebe.util.Either;\\n-import io.camunda.zeebe.util.collection.Tuple;\\n import java.util.Arrays;\\n import java.util.HashMap;\\n import java.util.Map;\\n@@ -236,21 +235,22 @@ public final class CreateProcessInstanceProcessor\\n return startInstructions.stream()\\n .map(\\n instruction ->\\n- Tuple.of(\\n+ new ElementIdAndType(\\n instruction.getElementId(),\\n process.getElementById(instruction.getElementIdBuffer()).getElementType()))\\n- .filter(elementTuple -> UNSUPPORTED_ELEMENT_TYPES.contains(elementTuple.getRight()))\\n+ .filter(\\n+ elementIdAndType -> UNSUPPORTED_ELEMENT_TYPES.contains(elementIdAndType.elementType))\\n .findAny()\\n .map(\\n- elementTypeTuple ->\\n+ elementIdAndType ->\\n Either.left(\\n new Rejection(\\n RejectionType.INVALID_ARGUMENT,\\n (\\\"Expected to create instance of process with start instructions but the element with id '%s' targets unsupported element type '%s'. \\\"\\n + \\\"Supported element types are: %s\\\")\\n .formatted(\\n- elementTypeTuple.getLeft(),\\n- elementTypeTuple.getRight(),\\n+ elementIdAndType.elementId,\\n+ elementIdAndType.elementType,\\n Arrays.stream(BpmnElementType.values())\\n .filter(\\n elementType ->\\n@@ -493,4 +493,6 @@ public final class CreateProcessInstanceProcessor\\n }\\n \\n record Rejection(RejectionType type, String reason) {}\\n+\\n+ record ElementIdAndType(String elementId, BpmnElementType elementType) {}\\n }\\n\", \"diff --git a/packages/core/src/LogicFlow.tsx b/packages/core/src/LogicFlow.tsx\\nindex 0d913b7..dcc59b3 100644\\n--- a/packages/core/src/LogicFlow.tsx\\n+++ b/packages/core/src/LogicFlow.tsx\\n@@ -276,6 +276,12 @@ export default class LogicFlow {\\n this.translate(-TRANSLATE_X, -TRANSLATE_Y);\\n }\\n /**\\n+ * \\u5c06\\u56fe\\u5f62\\u9009\\u4e2d\\n+ */\\n+ select(id: string) {\\n+ this.graphModel.selectElementById(id);\\n+ }\\n+ /**\\n * \\u5c06\\u56fe\\u5f62\\u5b9a\\u4f4d\\u5230\\u753b\\u5e03\\u4e2d\\u5fc3\\n * @param focusOnArgs \\u652f\\u6301\\u7528\\u6237\\u4f20\\u5165\\u56fe\\u5f62\\u5f53\\u524d\\u7684\\u5750\\u6807\\u6216id\\uff0c\\u53ef\\u4ee5\\u901a\\u8fc7type\\u6765\\u533a\\u5206\\u662f\\u8282\\u70b9\\u8fd8\\u662f\\u8fde\\u7ebf\\u7684id\\uff0c\\u4e5f\\u53ef\\u4ee5\\u4e0d\\u4f20\\uff08\\u515c\\u5e95\\uff09\\n */\\ndiff --git a/packages/core/src/model/GraphModel.ts b/packages/core/src/model/GraphModel.ts\\nindex 94d0899..10280a9 100644\\n--- a/packages/core/src/model/GraphModel.ts\\n+++ b/packages/core/src/model/GraphModel.ts\\n@@ -481,6 +481,13 @@ class GraphModel {\\n this.selectElement?.setSelected(true);\\n }\\n \\n+ @action\\n+ selectElementById(id: string) {\\n+ this.selectElement?.setSelected(false);\\n+ this.selectElement = this.getElement(id) as BaseNodeModel | BaseEdgeModel;\\n+ this.selectElement?.setSelected(true);\\n+ }\\n+\\n /* \\u4fee\\u6539\\u8fde\\u7ebf\\u7c7b\\u578b */\\n @action\\n changeEdgeType(type: string): void {\\n\"]"},"concern_count":{"kind":"number","value":4,"string":"4"},"shas":{"kind":"string","value":"[\"7e04a5e829d7416e312ac342a00a11787745753b\", \"97cabf49e7aca7754edde247003fbcb4ea42dd59\", \"bb2ccc1a778452aebf233cf78b20f1f4bab4354b\", \"6ae067153cd2608018fd3da76bd6d00a08da4b3a\"]"},"types":{"kind":"string","value":"[\"build\", \"fix\", \"refactor\", \"feat\"]"}}},{"rowIdx":1092,"cells":{"commit_message":{"kind":"string","value":"Handle different events.,simplify loadFiles code,disable edit/delete if primary key missing\n\nSigned-off-by: Pranav C ,avoid cancelling jobs"},"diff":{"kind":"string","value":"[\"diff --git a/src/notebook/epics/kernel-launch.js b/src/notebook/epics/kernel-launch.js\\nindex 9075d7c..9f16e67 100644\\n--- a/src/notebook/epics/kernel-launch.js\\n+++ b/src/notebook/epics/kernel-launch.js\\n@@ -113,6 +113,12 @@ export function newKernelObservable(kernelSpec: KernelInfo, cwd: string) {\\n observer.error({ type: 'ERROR', payload: error, err: true });\\n observer.complete();\\n });\\n+ spawn.on('exit', () => {\\n+ observer.complete();\\n+ });\\n+ spawn.on('disconnect', () => {\\n+ observer.complete();\\n+ });\\n });\\n });\\n }\\n\", \"diff --git a/frontend/app/player/web/network/loadFiles.ts b/frontend/app/player/web/network/loadFiles.ts\\nindex ec174fc..d164333 100644\\n--- a/frontend/app/player/web/network/loadFiles.ts\\n+++ b/frontend/app/player/web/network/loadFiles.ts\\n@@ -1,43 +1,33 @@\\n import APIClient from 'App/api_client';\\n \\n-const NO_NTH_FILE = \\\"nnf\\\"\\n-const NO_UNPROCESSED_FILES = \\\"nuf\\\"\\n+const NO_FILE_OK = \\\"No-file-but-this-is-ok\\\"\\n+const NO_BACKUP_FILE = \\\"No-efs-file\\\"\\n \\n export const loadFiles = (\\n urls: string[],\\n onData: (data: Uint8Array) => void,\\n ): Promise => {\\n- const firstFileURL = urls[0]\\n- urls = urls.slice(1)\\n- if (!firstFileURL) {\\n+ if (!urls.length) {\\n return Promise.reject(\\\"No urls provided\\\")\\n }\\n- return window.fetch(firstFileURL)\\n- .then(r => {\\n- return processAPIStreamResponse(r, true)\\n- })\\n- .then(onData)\\n- .then(() =>\\n- urls.reduce((p, url) =>\\n- p.then(() =>\\n- window.fetch(url)\\n- .then(r => {\\n- return processAPIStreamResponse(r, false)\\n- })\\n- .then(onData)\\n- ),\\n- Promise.resolve(),\\n- )\\n+ return urls.reduce((p, url, index) =>\\n+ p.then(() =>\\n+ window.fetch(url)\\n+ .then(r => {\\n+ return processAPIStreamResponse(r, index===0)\\n+ })\\n+ .then(onData)\\n+ ),\\n+ Promise.resolve(),\\n )\\n .catch(e => {\\n- if (e === NO_NTH_FILE) {\\n+ if (e === NO_FILE_OK) {\\n return\\n }\\n throw e\\n })\\n }\\n \\n-\\n export async function requestEFSDom(sessionId: string) {\\n return await requestEFSMobFile(sessionId + \\\"/dom.mob\\\")\\n }\\n@@ -50,21 +40,18 @@ async function requestEFSMobFile(filename: string) {\\n const api = new APIClient()\\n const res = await api.fetch('/unprocessed/' + filename)\\n if (res.status >= 400) {\\n- throw NO_UNPROCESSED_FILES\\n+ throw NO_BACKUP_FILE\\n }\\n return await processAPIStreamResponse(res, false)\\n }\\n \\n-const processAPIStreamResponse = (response: Response, isFirstFile: boolean) => {\\n+const processAPIStreamResponse = (response: Response, canBeMissed: boolean) => {\\n return new Promise((res, rej) => {\\n- if (response.status === 404 && !isFirstFile) {\\n- return rej(NO_NTH_FILE)\\n+ if (response.status === 404 && canBeMissed) {\\n+ return rej(NO_FILE_OK)\\n }\\n if (response.status >= 400) {\\n- return rej(\\n- isFirstFile ? `no start file. status code ${ response.status }`\\n- : `Bad endfile status code ${response.status}`\\n- )\\n+ return rej(`Bad file status code ${response.status}. Url: ${response.url}`)\\n }\\n res(response.arrayBuffer())\\n }).then(buffer => new Uint8Array(buffer))\\n\", \"diff --git a/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue b/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\\nindex 5f9841f..c414c8c 100644\\n--- a/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\\n+++ b/packages/nc-gui/components/project/spreadsheet/components/expandedForm.vue\\n@@ -413,6 +413,9 @@ export default {\\n \\n await this.reload()\\n } else if (Object.keys(updatedObj).length) {\\n+ if (!id) {\\n+ return this.$toast.info('Update not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n await this.api.update(id, updatedObj, this.oldRow)\\n } else {\\n return this.$toast.info('No columns to update').goAway(3000)\\ndiff --git a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\nindex c2b4b81..1b9d6a0 100644\\n--- a/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\n+++ b/packages/nc-gui/components/project/spreadsheet/rowsXcDataTable.vue\\n@@ -62,7 +62,15 @@\\n \\n \\n \\n-\\n+ \\n+ \\n+ Update & Delete not allowed since the table doesn't have any primary key\\n+ \\n+ \\n \\n \\n \\n@@ -208,6 +216,7 @@\\n :meta=\\\"meta\\\"\\n :is-virtual=\\\"selectedView.type === 'vtable'\\\"\\n :api=\\\"api\\\"\\n+ :is-pk-avail=\\\"isPkAvail\\\"\\n @onNewColCreation=\\\"onNewColCreation\\\"\\n @onCellValueChange=\\\"onCellValueChange\\\"\\n @insertNewRow=\\\"insertNewRow\\\"\\n@@ -631,8 +640,8 @@ export default {\\n if (\\n !this.meta || (\\n (this.meta.hasMany && this.meta.hasMany.length) ||\\n- (this.meta.manyToMany && this.meta.manyToMany.length) ||\\n- (this.meta.belongsTo && this.meta.belongsTo.length))\\n+ (this.meta.manyToMany && this.meta.manyToMany.length) ||\\n+ (this.meta.belongsTo && this.meta.belongsTo.length))\\n ) {\\n return this.$toast.info('Please delete relations before deleting table.').goAway(3000)\\n }\\n@@ -817,6 +826,10 @@ export default {\\n \\n const id = this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\\n \\n+ if (!id) {\\n+ return this.$toast.info('Update not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n+\\n const newData = await this.api.update(id, {\\n [column._cn]: rowObj[column._cn]\\n }, { [column._cn]: oldRow[column._cn] })\\n@@ -841,6 +854,11 @@ export default {\\n const rowObj = this.rowContextMenu.row\\n if (!this.rowContextMenu.rowMeta.new) {\\n const id = this.meta && this.meta.columns && this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\\n+\\n+ if (!id) {\\n+ return this.$toast.info('Delete not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n+\\n await this.api.delete(id)\\n }\\n this.data.splice(this.rowContextMenu.index, 1)\\n@@ -859,6 +877,11 @@ export default {\\n }\\n if (!rowMeta.new) {\\n const id = this.meta.columns.filter(c => c.pk).map(c => rowObj[c._cn]).join('___')\\n+\\n+ if (!id) {\\n+ return this.$toast.info('Delete not allowed for table which doesn\\\\'t have primary Key').goAway(3000)\\n+ }\\n+\\n await this.api.delete(id)\\n }\\n this.data.splice(row, 1)\\n@@ -991,6 +1014,9 @@ export default {\\n }\\n },\\n computed: {\\n+ isPkAvail() {\\n+ return this.meta && this.meta.columns.some(c => c.pk)\\n+ },\\n isGallery() {\\n return this.selectedView && this.selectedView.show_as === 'gallery'\\n },\\ndiff --git a/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue b/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\\nindex 5497d05..c198784 100644\\n--- a/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\\n+++ b/packages/nc-gui/components/project/spreadsheet/views/xcGridView.vue\\n@@ -27,7 +27,7 @@\\n @xcresized=\\\"resizingCol = null\\\"\\n >\\n \\n+-->\\n \\n \\n@@ -162,7 +162,8 @@\\n \\n \\n

\\n \\n \\n- \\n- \\n- \\n \\n \\n Delete Selected Rows\\n \\n \\n-