{ // 获取包含Hugging Face文本的span元素 const spans = link.querySelectorAll('span.whitespace-nowrap, span.hidden.whitespace-nowrap'); spans.forEach(span => { if (span.textContent && span.textContent.trim().match(/Hugging\s*Face/i)) { span.textContent = 'AI快站'; } }); }); // 替换logo图片的alt属性 document.querySelectorAll('img[alt*="Hugging"], img[alt*="Face"]').forEach(img => { if (img.alt.match(/Hugging\s*Face/i)) { img.alt = 'AI快站 logo'; } }); } // 替换导航栏中的链接 function replaceNavigationLinks() { // 已替换标记,防止重复运行 if (window._navLinksReplaced) { return; } // 已经替换过的链接集合,防止重复替换 const replacedLinks = new Set(); // 只在导航栏区域查找和替换链接 const headerArea = document.querySelector('header') || document.querySelector('nav'); if (!headerArea) { return; } // 在导航区域内查找链接 const navLinks = headerArea.querySelectorAll('a'); navLinks.forEach(link => { // 如果已经替换过,跳过 if (replacedLinks.has(link)) return; const linkText = link.textContent.trim(); const linkHref = link.getAttribute('href') || ''; // 替换Spaces链接 - 仅替换一次 if ( (linkHref.includes('/spaces') || linkHref === '/spaces' || linkText === 'Spaces' || linkText.match(/^s*Spacess*$/i)) && linkText !== 'PDF TO Markdown' && linkText !== 'PDF TO Markdown' ) { link.textContent = 'PDF TO Markdown'; link.href = 'https://fast360.xyz'; link.setAttribute('target', '_blank'); link.setAttribute('rel', 'noopener noreferrer'); replacedLinks.add(link); } // 删除Posts链接 else if ( (linkHref.includes('/posts') || linkHref === '/posts' || linkText === 'Posts' || linkText.match(/^s*Postss*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } // 替换Docs链接 - 仅替换一次 else if ( (linkHref.includes('/docs') || linkHref === '/docs' || linkText === 'Docs' || linkText.match(/^s*Docss*$/i)) && linkText !== 'Voice Cloning' ) { link.textContent = 'Voice Cloning'; link.href = 'https://vibevoice.info/'; replacedLinks.add(link); } // 删除Enterprise链接 else if ( (linkHref.includes('/enterprise') || linkHref === '/enterprise' || linkText === 'Enterprise' || linkText.match(/^s*Enterprises*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } }); // 查找可能嵌套的Spaces和Posts文本 const textNodes = []; function findTextNodes(element) { if (element.nodeType === Node.TEXT_NODE) { const text = element.textContent.trim(); if (text === 'Spaces' || text === 'Posts' || text === 'Enterprise') { textNodes.push(element); } } else { for (const child of element.childNodes) { findTextNodes(child); } } } // 只在导航区域内查找文本节点 findTextNodes(headerArea); // 替换找到的文本节点 textNodes.forEach(node => { const text = node.textContent.trim(); if (text === 'Spaces') { node.textContent = node.textContent.replace(/Spaces/g, 'PDF TO Markdown'); } else if (text === 'Posts') { // 删除Posts文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } else if (text === 'Enterprise') { // 删除Enterprise文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } }); // 标记已替换完成 window._navLinksReplaced = true; } // 替换代码区域中的域名 function replaceCodeDomains() { // 特别处理span.hljs-string和span.njs-string元素 document.querySelectorAll('span.hljs-string, span.njs-string, span[class*="hljs-string"], span[class*="njs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换hljs-string类的span中的域名(移除多余的转义符号) document.querySelectorAll('span.hljs-string, span[class*="hljs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换pre和code标签中包含git clone命令的域名 document.querySelectorAll('pre, code').forEach(element => { if (element.textContent && element.textContent.includes('git clone')) { const text = element.innerHTML; if (text.includes('huggingface.co')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 处理特定的命令行示例 document.querySelectorAll('pre, code').forEach(element => { const text = element.innerHTML; if (text.includes('huggingface.co')) { // 针对git clone命令的专门处理 if (text.includes('git clone') || text.includes('GIT_LFS_SKIP_SMUDGE=1')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 特别处理模型下载页面上的代码片段 document.querySelectorAll('.flex.border-t, .svelte_hydrator, .inline-block').forEach(container => { const content = container.innerHTML; if (content && content.includes('huggingface.co')) { container.innerHTML = content.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 特别处理模型仓库克隆对话框中的代码片段 try { // 查找包含"Clone this model repository"标题的对话框 const cloneDialog = document.querySelector('.svelte_hydration_boundary, [data-target="MainHeader"]'); if (cloneDialog) { // 查找对话框中所有的代码片段和命令示例 const codeElements = cloneDialog.querySelectorAll('pre, code, span'); codeElements.forEach(element => { if (element.textContent && element.textContent.includes('huggingface.co')) { if (element.innerHTML.includes('huggingface.co')) { element.innerHTML = element.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { element.textContent = element.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); } // 更精确地定位克隆命令中的域名 document.querySelectorAll('[data-target]').forEach(container => { const codeBlocks = container.querySelectorAll('pre, code, span.hljs-string'); codeBlocks.forEach(block => { if (block.textContent && block.textContent.includes('huggingface.co')) { if (block.innerHTML.includes('huggingface.co')) { block.innerHTML = block.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { block.textContent = block.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); }); } catch (e) { // 错误处理但不打印日志 } } // 当DOM加载完成后执行替换 if (document.readyState === 'loading') { document.addEventListener('DOMContentLoaded', () => { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); }); } else { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); } // 增加一个MutationObserver来处理可能的动态元素加载 const observer = new MutationObserver(mutations => { // 检查是否导航区域有变化 const hasNavChanges = mutations.some(mutation => { // 检查是否存在header或nav元素变化 return Array.from(mutation.addedNodes).some(node => { if (node.nodeType === Node.ELEMENT_NODE) { // 检查是否是导航元素或其子元素 if (node.tagName === 'HEADER' || node.tagName === 'NAV' || node.querySelector('header, nav')) { return true; } // 检查是否在导航元素内部 let parent = node.parentElement; while (parent) { if (parent.tagName === 'HEADER' || parent.tagName === 'NAV') { return true; } parent = parent.parentElement; } } return false; }); }); // 只在导航区域有变化时执行替换 if (hasNavChanges) { // 重置替换状态,允许再次替换 window._navLinksReplaced = false; replaceHeaderBranding(); replaceNavigationLinks(); } }); // 开始观察document.body的变化,包括子节点 if (document.body) { observer.observe(document.body, { childList: true, subtree: true }); } else { document.addEventListener('DOMContentLoaded', () => { observer.observe(document.body, { childList: true, subtree: true }); }); } })(); `))\n\n\treturn\n}","func IndexHandler(ctx *Context, w http.ResponseWriter, r *http.Request) (int, error) {\n\tvar host string\n\tif r.Header.Get(\"X-Forwarded-Host\") != \"\" {\n\t\thost = r.Header.Get(\"X-Forwarded-Host\")\n\t} else {\n\t\thost = r.Host\n\t}\n\n\tvar hostConfig *gabs.Container\n\tvar ok bool\n\n\t// Check if host present in Config.\n\tchildren := ctx.Config.ChildrenMap()\n\tif hostConfig, ok = children[host]; !ok {\n\t\treturn 404, fmt.Errorf(\"Shortcut '%s' not found in Config.\", host)\n\t}\n\n\ttokens := tokenize(host + r.URL.Path)\n\n\t// Set up handles on token and Config. We might need to skip ahead if there's a custom schema set.\n\ttokensStart := tokens.Front()\n\tconf := ctx.Config\n\n\tvar path bytes.Buffer\n\tif s := hostConfig.Path(sslKey).Data(); s != nil && s.(bool) {\n\t\tpath.WriteString(httpPrefix)\n\t} else if s := hostConfig.Path(schemaKey).Data(); s != nil && s.(string) != \"\" {\n\t\tpath.WriteString(hostConfig.Path(schemaKey).Data().(string) + \":/\")\n\t\t// move one token ahead to parse expansions correctly.\n\t\tconf = conf.ChildrenMap()[tokensStart.Value.(string)]\n\t\ttokensStart = tokensStart.Next()\n\t} else {\n\t\t// Default to regular https prefix.\n\t\tpath.WriteString(httpsPrefix)\n\t}\n\n\tExpandPath(conf, tokensStart, &path)\n\n\t// send result\n\thttp.Redirect(w, r, path.String(), http.StatusFound)\n\n\treturn 302, nil\n}","func (srv *Server) ApplyHandlers() {\n\tsrv.router.Handle(\"/*\", http.FileServer(http.Dir(\"./web\")))\n\tsrv.router.Get(\"/socket\", srv.socketHandler)\n}","func homeHandler(env *Env, w http.ResponseWriter, r *http.Request) error {\n return renderTemplate(w, \"index\", \"base\", nil)\n}","func Mount(mux goahttp.Muxer, h *Server) {\n\tMountRegisterHandler(mux, h.Register)\n\tMountRetrieveHandler(mux, h.Retrieve)\n\tMountHmrcCallbackHandler(mux, h.HmrcCallback)\n}","func (mux *ServeMux) handler(host, path string) (h http.Handler, pattern string) {\n\tmux.mu.RLock()\n\tdefer mux.mu.RUnlock()\n\n\t// Host-specific pattern takes precedence over generic ones\n\tif mux.hosts {\n\t\th, pattern = mux.match(host + path)\n\t}\n\tif h == nil {\n\t\th, pattern = mux.match(path)\n\t}\n\tif h == nil {\n\t\th, pattern = http.NotFoundHandler(), \"\"\n\t}\n\treturn\n}","func (mux *ServeMux) handler(host, path string) (h Handler, pattern string) {\n\tmux.mu.RLock()\n\tdefer mux.mu.RUnlock()\n\n\t// Host-specific pattern takes precedence over generic ones\n\tif mux.hosts {\n\t\th, pattern = mux.match(host + path)\n\t}\n\tif h == nil {\n\t\th, pattern = mux.match(path)\n\t}\n\tif h == nil {\n\t\th, pattern = NotFoundHandler(), \"\"\n\t}\n\treturn\n}","func (r *Router) handle(c *Ctx) {\n\tvar handler HandlerFunc\n\treq := c.Request()\n\tw := c.Writer()\n\tpath := req.URL.Path\n\tmethod := req.Method\n\tres := r.trie.Match(path)\n\n\tif res.Node == nil {\n\t\t// FixedPathRedirect or TrailingSlashRedirect\n\t\tif res.TSR != \"\" || res.FPR != \"\" {\n\t\t\treq.URL.Path = res.TSR\n\t\t\tif res.FPR != \"\" {\n\t\t\t\treq.URL.Path = res.FPR\n\t\t\t}\n\t\t\tcode := 301\n\t\t\tif method != \"GET\" {\n\t\t\t\tcode = 307\n\t\t\t}\n\t\t\thttp.Redirect(w, req, req.URL.String(), code)\n\t\t\treturn\n\t\t}\n\t\tif r.noRoute == nil {\n\t\t\thttp.Error(w, fmt.Sprintf(`\"%s\" not implemented`, path), 501)\n\t\t\treturn\n\t\t}\n\t\thandler = r.noRoute\n\t} else {\n\t\t// ok := false\n\t\thd := res.Node.GetHandler(method)\n\t\thandler, _ = hd.(HandlerFunc)\n\t\t// handler = r.wrapHandler(hd)\n\t\t// if !ok {\n\t\t// \tpanic(\"handler error\")\n\t\t// }\n\t\tif handler == nil {\n\t\t\t// OPTIONS support\n\t\t\tif method == http.MethodOptions {\n\t\t\t\tw.Header().Set(\"Allow\", res.Node.GetAllow())\n\t\t\t\tw.WriteHeader(204)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif r.noMethod == nil {\n\t\t\t\t// If no route handler is returned, it's a 405 error\n\t\t\t\tw.Header().Set(\"Allow\", res.Node.GetAllow())\n\t\t\t\thttp.Error(w, fmt.Sprintf(`\"%s\" not allowed in \"%s\"`, method, path), 405)\n\t\t\t\treturn\n\t\t\t}\n\t\t\thandler = r.noMethod\n\t\t}\n\t}\n\n\tif len(res.Params) != 0 {\n\t\tc.params = res.Params\n\t}\n\tc.handlers = append(c.handlers, handler)\n\tc.Next()\n}","func MountApplication(mountpoints map[string]http.Handler, host *Host, app *Application) {\n listing := &DirList{Host: host}\n\n\t// Serve the static build files from the mountpoint path.\n\turl := app.PublishUrl()\n\tlog.Printf(\"Serving app %s from %s\", url, app.PublicDirectory())\n fileserver := http.FileServer(http.Dir(app.PublicDirectory()))\n mountpoints[url] = http.StripPrefix(url, PublicHandler{Listing: listing, App: app, FileServer: fileserver})\n}","func (s *Server) Start() *mux.Router {\n\trouter := mux.NewRouter()\n\n\trouter.HandleFunc(\"/\", s.Config.Index)\n\trouter.HandleFunc(\"/host/{host}\", s.Config.HostSummary)\n\trouter.HandleFunc(\"/group/{group}\", s.Config.GroupSummary)\n\trouter.PathPrefix(\"/\").Handler(http.FileServer(http.Dir(\"./assets/\")))\n\n\treturn router\n}","func rootHandler(w http.ResponseWriter, r *http.Request) {\n\t//fmt.Fprintf(w, \"

Hello All

\")\n\tt,_ := template.ParseFiles(\"root.html\")\n\tt.Execute(w, nil)\n\n}","func (s *server) routes() {\n s.router.HandleFunc(\"/\", s.homePageHandler)\n s.router.HandleFunc(\"/signup/\", s.signupHandler)\n s.router.HandleFunc(\"/signin/\", s.signinHandler)\n s.router.HandleFunc(\"/signout/\", s.makeHandler(s.signoutHandler))\n s.router.HandleFunc(\"/view/\", s.makeHandler(s.viewHandler))\n s.router.HandleFunc(\"/save/\", s.makeHandler(s.saveHandler))\n s.router.HandleFunc(\"/edit/\", s.makeHandler(s.editHandler))\n s.router.HandleFunc(\"/delete/\", s.makeHandler(s.deleteHandler))\n\n s.validPath = regexp.MustCompile(\n \"^/(new|view|save|edit|delete|signout)/([0-9]*)$\")\n}","func (mux *ServeMux) Handler(r *Request) (h Handler, pattern string) {\n\n\t// CONNECT requests are not canonicalized.\n\tif r.Method == \"CONNECT\" {\n\t\t// If r.URL.Path is /tree and its handler is not registered,\n\t\t// the /tree -> /tree/ redirect applies to CONNECT requests\n\t\t// but the path canonicalization does not.\n\t\tif u, ok := mux.redirectToPathSlash(r.URL.Host, r.URL.Path, r.URL); ok {\n\t\t\treturn RedirectHandler(u.String(), StatusMovedPermanently), u.Path\n\t\t}\n\n\t\treturn mux.handler(r.Host, r.URL.Path)\n\t}\n\n\t// All other requests have any port stripped and path cleaned\n\t// before passing to mux.handler.\n\thost := stripHostPort(r.Host)\n\tpath := cleanPath(r.URL.Path)\n\n\t// If the given path is /tree and its handler is not registered,\n\t// redirect for /tree/.\n\tif u, ok := mux.redirectToPathSlash(host, path, r.URL); ok {\n\t\treturn RedirectHandler(u.String(), StatusMovedPermanently), u.Path\n\t}\n\n\tif path != r.URL.Path {\n\t\t_, pattern = mux.handler(host, path)\n\t\tu := &url.URL{Path: path, RawQuery: r.URL.RawQuery}\n\t\treturn RedirectHandler(u.String(), StatusMovedPermanently), pattern\n\t}\n\n\treturn mux.handler(host, r.URL.Path)\n}","func (h *Host) initHostHandler() {\n\thttp.HandleFunc(h.url+\":\"+h.port+\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\th.reqLog.Access(r)\n\t\tw = h.addHeaders(w)\n\t\tif h.static != \"\" {\n\t\t\tif strings.HasPrefix(r.URL.Path, h.static) {\n\t\t\t\thttp.ServeFile(w, r, h.root+r.URL.Path)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\t//get next proxy\n\t\trewriteProxy := h.NextProxy()\n\t\tr.RequestURI = strings.Replace(r.RequestURI, \"/\", \"\", 1)\n\t\tr.URL.Path = strings.Replace(r.URL.Path, \"/\", \"\", 1)\n\t\trewriteProxy.ServeHTTP(w, r)\n\t})\n}","func InitRouter() {\n\tr := mux.NewRouter()\n\n\tr.HandleFunc(\"/login\", service.Session{}.Login)\n\tr.HandleFunc(\"/logmein\", service.Session{}.LogMeIn).Methods(\"GET\")\n\tr.HandleFunc(\"/logout\", service.Session{}.Logout).Methods(\"GET\")\n\n\tr.HandleFunc(\"/faq\", func(w http.ResponseWriter, r *http.Request) {\n\t\thc := &kinli.HttpContext{W: w, R: r}\n\t\tpage := kinli.NewPage(hc, \"Frequently Asked Questions\", \"\", \"\", nil)\n\t\tkinli.DisplayPage(w, \"faq\", page)\n\t}).Methods(\"GET\")\n\n\tr.HandleFunc(\"/example\", func(w http.ResponseWriter, r *http.Request) {\n\t\thc := &kinli.HttpContext{W: w, R: r}\n\t\tpage := kinli.NewPage(hc, \"Example Form\", \"\", \"\", nil)\n\t\tkinli.DisplayPage(w, \"example\", page)\n\t}).Methods(\"GET\")\n\n\tr.HandleFunc(\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\thc := &kinli.HttpContext{W: w, R: r}\n\t\tpage := kinli.NewPage(hc, \"hello page\", \"\", \"\", nil)\n\t\tkinli.DisplayPage(w, \"home\", page)\n\t}).Methods(\"GET\")\n\n\tr.HandleFunc(\"/{uid}\", service.FormSubmissionRequest).Methods(\"POST\")\n\n\tr.NotFoundHandler = http.HandlerFunc(notFound)\n\n\tinitStatic(r)\n\n\tsrv := &http.Server{\n\t\tHandler: r,\n\t\tAddr: common.Config.LocalServer,\n\t\tWriteTimeout: 60 * time.Second,\n\t\tReadTimeout: 60 * time.Second,\n\t}\n\tlog.Println(\"Starting server on\", common.Config.LocalServer)\n\tlog.Fatal(srv.ListenAndServe())\n\n}","func (this *Router) Handle(title, path string, handlers ...HandlerFunc) *Router {\n\tif len(path) < 1 || path[0] != '/' || strings.Contains(path, \"//\") {\n\t\tpanic(\"add router faild, invalid path \" + path)\n\t}\n\tif sepIndex := strings.Index(path[1:], \"/\") + 1; sepIndex > 1 {\n\t\troot := path[:sepIndex]\n\t\tsubpath := path[sepIndex:]\n\t\tvar group *Router = nil\n\t\tfor _, router := range this.children {\n\t\t\tif router.path == root {\n\t\t\t\tgroup = router\n\t\t\t}\n\t\t}\n\t\tif group == nil {\n\t\t\tgroup = this.Group(root)\n\t\t}\n\t\treturn group.Handle(title, subpath, handlers...)\n\t}\n\thandlerChain := append([]HandlerFunc{}, this.handlerChain...)\n\thandlerChain = append(handlerChain, handlers...)\n\trouter := &Router{\n\t\ttitle: title,\n\t\tpath: path,\n\t\trealPath: this.realPath + path,\n\t\thandlerChain: handlerChain,\n\t\tchildren: []*Router{},\n\t}\n\tthis.children = append(this.children, router)\n\tlog.Log(\"DEBUG\", \"add router\", router.realPath)\n\treturn router\n}","func (_m *RESTHandler) Mount(group *echo.Group) {\n\t_m.Called(group)\n}","func (osh *SystemHandler) Mount(source string, target string, fsType string, flags uintptr, data string) error {\n\treturn syscall.Mount(source, target, fsType, flags, data)\n}","func (hr *httpRouter) Handler() http.Handler {\n\n\tc, _ := console.New(console.Options{Color: true})\n\t_ = logger.Register(\"console\", logger.Config{Writer: c})\n\tcLogger, _ := logger.Get(\"console\")\n\tl := log.New(cLogger)\n\n\tfmt.Print(\"Loading Routes...\")\n\t//add files in a directory\n\tro := newHttpRouterExtended(hr)\n\n\tmw := middleware.Chain{}\n\n\t//adding files\n\tfor path, file := range hr.file {\n\t\tro.HandlerFunc(\"GET\", path, mw.Add(l.MW).Handle(\n\t\t\tfunc(w http.ResponseWriter, req *http.Request) {\n\t\t\t\thttp.ServeFile(w, req, hr.file[req.Context().Value(router.PATTERN).(string)])\n\t\t\t}))\n\t\tfmt.Printf(\"\\n\\x1b[32m %#v [GET]%v \\x1b[49m\\x1b[39m \", path, file)\n\t}\n\n\t// adding directories\n\tfor k, path := range hr.dir {\n\t\tfileServer := http.FileServer(http.Dir(path))\n\t\tpattern := k + \"/*filepath\"\n\t\tro.HandlerFunc(\"GET\", pattern, mw.Add(l.MW).Handle(\n\t\t\tfunc(w http.ResponseWriter, req *http.Request) {\n\t\t\t\t//disable directory listing\n\t\t\t\tif strings.HasSuffix(req.URL.Path, \"/\") {\n\t\t\t\t\thttp.NotFound(w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif val, ok := req.Context().Value(router.PARAMS).(map[string][]string)[\"filepath\"]; ok {\n\t\t\t\t\treq.URL.Path = val[0]\n\t\t\t\t\tfileServer.ServeHTTP(w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\thttp.NotFound(w, req)\n\t\t\t\treturn\n\n\t\t\t}))\n\t\tfmt.Printf(\"\\n\\x1b[32m %#v [GET]%v \\x1b[49m\\x1b[39m \", pattern, http.Dir(path))\n\t}\n\n\t//register all controller routes\n\tfor _, r := range hr.routes {\n\t\tfmt.Printf(\"\\n\\x1b[32m %#v :name \\x1b[49m\\x1b[39m \", r.pattern)\n\t\tfor method, fn := range r.controller.MappingBy(r.pattern) {\n\t\t\tif r.mws != nil {\n\t\t\t\tro.HandlerFunc(strings.ToUpper(method), r.pattern, r.mws.Handle(r.controller.ServeHTTP)) //TODO ????? error no url pattern\n\t\t\t} else {\n\t\t\t\tro.HandlerFunc(strings.ToUpper(method), r.pattern, r.controller.ServeHTTP)\n\t\t\t}\n\t\t\tfmt.Printf(\"\\x1b[32m [%v]%v name \\x1b[49m\\x1b[39m \", method, fn)\n\t\t}\n\t}\n\n\t//Not Found Handler\n\tif hr.notFound != nil {\n\t\tro.NotFound = hr.notFound\n\t}\n\n\treturn ro\n}","func rootHandler(w http.ResponseWriter, r *http.Request, title string) {\n\tp, err := loadRoot(title)\n\n\tp.Body = template.HTML(blackfriday.MarkdownCommon([]byte(p.Body)))\n\tp.Body = template.HTML(convertWikiMarkup([]byte(p.Body)))\n\n\terr = templates.ExecuteTemplate(w, \"root.html\", p)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t}\n}","func (mux *Mux) Handler(r *Request) Handler {\n\tscheme := r.URL.Scheme\n\thost := r.URL.Hostname()\n\tpath := cleanPath(r.URL.Path)\n\n\t// If the given path is /tree and its handler is not registered,\n\t// redirect for /tree/.\n\tif u, ok := mux.redirectToPathSlash(muxKey{scheme, host, path}, r.URL); ok {\n\t\treturn StatusHandler(StatusPermanentRedirect, u.String())\n\t}\n\n\tif path != r.URL.Path {\n\t\tu := *r.URL\n\t\tu.Path = path\n\t\treturn StatusHandler(StatusPermanentRedirect, u.String())\n\t}\n\n\tmux.mu.RLock()\n\tdefer mux.mu.RUnlock()\n\n\th := mux.match(muxKey{scheme, host, path})\n\tif h == nil {\n\t\t// Try wildcard\n\t\tif wildcard, ok := getWildcard(host); ok {\n\t\t\th = mux.match(muxKey{scheme, wildcard, path})\n\t\t}\n\t}\n\tif h == nil {\n\t\th = NotFoundHandler()\n\t}\n\treturn h\n}","func MountCORSHandler(mux goahttp.Muxer, h http.Handler) {\n\th = handleStationOrigin(h)\n\tf, ok := h.(http.HandlerFunc)\n\tif !ok {\n\t\tf = func(w http.ResponseWriter, r *http.Request) {\n\t\t\th.ServeHTTP(w, r)\n\t\t}\n\t}\n\tmux.Handle(\"OPTIONS\", \"/stations\", f)\n\tmux.Handle(\"OPTIONS\", \"/stations/{id}\", f)\n\tmux.Handle(\"OPTIONS\", \"/stations/@/{id}\", f)\n\tmux.Handle(\"OPTIONS\", \"/projects/{id}/stations\", f)\n\tmux.Handle(\"OPTIONS\", \"/stations/{id}/photo\", f)\n}","func (e *Engine) setupFallback() {\n\te.srv.HTTPErrorHandler = func(err error, c echo.Context) {\n\t\tif err != echo.ErrNotFound {\n\t\t\treturn\n\t\t}\n\t\turi := genericPath(c.Request().RequestURI)\n\t\tfs, err := os.Stat(uri)\n\t\tif err != nil {\n\t\t\tc.Error(err)\n\t\t\treturn\n\t\t}\n\n\t\tif fs.IsDir() {\n\t\t\t// if working at pwd, capture all dirs\n\t\t\tif len(e.dirs) == 1 && e.dirs[0] == \".\" {\n\t\t\t\t// check if file excluded\n\t\t\t\tif isExclude(uri) {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif err := e.doDispatch(c, fs); err != nil {\n\t\t\t\t\tc.Error(err)\n\t\t\t\t}\n\t\t\t}\n\t\t\t// only capture missing dir when serving pwd\n\t\t\t// return for any other dirs\n\t\t\treturn\n\t\t}\n\n\t\t// capture any files here\n\t\terr = e.doServeFile(uri)(c)\n\t\tif err != nil {\n\t\t\tc.Error(err)\n\t\t}\n\t}\n}","func NewHandler(s *Server) *Handler {\n\th := &Handler{\n\t\tserver: s,\n\t\tmux: pat.New(),\n\t}\n\n\t// Series routes.\n\th.mux.Get(\"/db/:db/series\", http.HandlerFunc(h.serveQuery))\n\th.mux.Post(\"/db/:db/series\", http.HandlerFunc(h.serveWriteSeries))\n\th.mux.Del(\"/db/:db/series/:series\", http.HandlerFunc(h.serveDeleteSeries))\n\th.mux.Get(\"/db\", http.HandlerFunc(h.serveDatabases))\n\th.mux.Post(\"/db\", http.HandlerFunc(h.serveCreateDatabase))\n\th.mux.Del(\"/db/:name\", http.HandlerFunc(h.serveDeleteDatabase))\n\n\t// Cluster admins routes.\n\th.mux.Get(\"/cluster_admins/authenticate\", http.HandlerFunc(h.serveAuthenticateClusterAdmin))\n\th.mux.Get(\"/cluster_admins\", http.HandlerFunc(h.serveClusterAdmins))\n\th.mux.Post(\"/cluster_admins\", http.HandlerFunc(h.serveCreateClusterAdmin))\n\th.mux.Post(\"/cluster_admins/:user\", http.HandlerFunc(h.serveUpdateClusterAdmin))\n\th.mux.Del(\"/cluster_admins/:user\", http.HandlerFunc(h.serveDeleteClusterAdmin))\n\n\t// Database users routes.\n\th.mux.Get(\"/db/:db/authenticate\", http.HandlerFunc(h.serveAuthenticateDBUser))\n\th.mux.Get(\"/db/:db/users\", http.HandlerFunc(h.serveDBUsers))\n\th.mux.Post(\"/db/:db/users\", http.HandlerFunc(h.serveCreateDBUser))\n\th.mux.Get(\"/db/:db/users/:user\", http.HandlerFunc(h.serveDBUser))\n\th.mux.Post(\"/db/:db/users/:user\", http.HandlerFunc(h.serveUpdateDBUser))\n\th.mux.Del(\"/db/:db/users/:user\", http.HandlerFunc(h.serveDeleteDBUser))\n\n\t// Utilities\n\th.mux.Get(\"/ping\", http.HandlerFunc(h.servePing))\n\th.mux.Get(\"/interfaces\", http.HandlerFunc(h.serveInterfaces))\n\n\t// Shard routes.\n\th.mux.Get(\"/cluster/shards\", http.HandlerFunc(h.serveShards))\n\th.mux.Post(\"/cluster/shards\", http.HandlerFunc(h.serveCreateShard))\n\th.mux.Del(\"/cluster/shards/:id\", http.HandlerFunc(h.serveDeleteShard))\n\n\t// Shard space routes.\n\th.mux.Get(\"/cluster/shard_spaces\", http.HandlerFunc(h.serveShardSpaces))\n\th.mux.Post(\"/cluster/shard_spaces/:db\", http.HandlerFunc(h.serveCreateShardSpace))\n\th.mux.Post(\"/cluster/shard_spaces/:db/:name\", http.HandlerFunc(h.serveUpdateShardSpace))\n\th.mux.Del(\"/cluster/shard_spaces/:db/:name\", http.HandlerFunc(h.serveDeleteShardSpace))\n\n\t// Cluster config endpoints\n\th.mux.Get(\"/cluster/servers\", http.HandlerFunc(h.serveServers))\n\th.mux.Del(\"/cluster/servers/:id\", http.HandlerFunc(h.serveDeleteServer))\n\n\treturn h\n}","func HttpRootHandler(w http.ResponseWriter, r *http.Request) {\n\tcmd := CommandStruct{}\n\n\tbs, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\treturn\n\t}\n\t_ = json.Unmarshal(bs, &cmd)\n\tswitch cmd.Command {\n\tcase \"check\":\n\t\tfmt.Fprintln(w, \"HTTP-Server is online.\")\n\t\tbreak\n\n\tcase \"disc\":\n\t\tfmt.Fprintln(w, \"HTTP-Server shutting down...\")\n\t\thttpAlive <- 1\n\t\tbreak\n\n\tcase \"lista\":\n\t\tfmt.Fprintln(w, \"List of all available NBD-devices:\")\n\t\tfor key, value := range AvailableList {\n\t\t\tif AvailableList[key] != \"\" {\n\t\t\t\tfmt.Fprintln(w, value)\n\t\t\t}\n\t\t}\n\t\tbreak\n\n\tcase \"listm\":\n\t\tfmt.Fprintln(w, \"List of all mounted NBD-devices:\")\n\t\tfor key, value := range MountedList {\n\t\t\tfmt.Fprintln(w, key+\"\\t\"+value)\n\t\t}\n\t\tbreak\n\n\tcase \"mount\":\n\t\tif strings.Contains(cmd.Device, \"/dev/nbd\") {\n\t\t\tfor i := 0; i < len(AvailableList); i++ {\n\t\t\t\tif AvailableList[i] == cmd.Device {\n\n\t\t\t\t\tLinkedLogins[len(LinkedLogins)+1], err = nethandler.SetupConnection(cmd.Image, cmd.User, cmd.Pass, cmd.Device)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Fprintf(w, \"Error: \", err)\n\t\t\t\t\t\tfmt.Fprintf(w, \"\\n\")\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tAddToMountedList(cmd.Device, cmd.Image)\n\t\t\t\t\tfmt.Fprintf(w, \"Successfully mounted \"+cmd.Image+\" to \"+cmd.Device+\"\\n\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tfor _, value := range AvailableList {\n\t\t\t\tif value != \"\" {\n\t\t\t\t\tAddToMountedList(value, cmd.Image)\n\t\t\t\t\tfmt.Fprintf(w, \"Device \"+cmd.Device+\" is already mounted.\\n\"+cmd.Image+\" has been mounted to \"+value+\" instead.\\n\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tfmt.Fprintf(w, \"No more devices available!\\n\")\n\t\t} else {\n\t\t\tfmt.Fprintf(w, \"Specified device not recognised.\\n\")\n\t\t}\n\t\tbreak\n\n\tcase \"unmount\":\n\t\t//TODO Real unmounting of NBD-devices\n\t\tfor key, _ := range AvailableList {\n\t\t\tif AvailableList[key] == \"\" {\n\t\t\t\tdelete(MountedList, cmd.Device)\n\t\t\t\tAvailableList[key] = cmd.Device\n\t\t\t\tfmt.Fprint(w, \"Successfully unmounted \"+cmd.Device)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tbreak\n\t}\n}","func HTTPHandler(router *mux.Router) {\n\trouter.Handle(\"/\", ImageHandler())\n\trouter.PathPrefix(\"/unsafe/\").Handler(ImageHandler())\n\trouter.Handle(\"/healthcheck\", HealthcheckHandler())\n}","func Mount(router *mux.Router) {\n\n}","func MountPhotoHandler(mux goahttp.Muxer, h http.Handler) {\n\tf, ok := handleStationOrigin(h).(http.HandlerFunc)\n\tif !ok {\n\t\tf = func(w http.ResponseWriter, r *http.Request) {\n\t\t\th.ServeHTTP(w, r)\n\t\t}\n\t}\n\tmux.Handle(\"GET\", \"/stations/{id}/photo\", f)\n}","func indexHandler(w http.ResponseWriter, r *http.Request) {\n\tif r.URL.Path != \"/\" {\n\t\thttp.NotFound(w, r)\n\t\treturn\n\t}\n\n\thttp.ServeFile(w, r, \"pages/main.html\")\n}","func pathHandler(w http.ResponseWriter, r *http.Request) {\n\tpath := \"/var/www/ear7h-net/\" + r.URL.Path[1:]\n\t//open file and send\n\tf, err := os.Open(path)\n\tif err != nil {\n\t\tfourOhFour(w, r)\n\t} else {\n\t\thttp.ServeContent(w, r, r.URL.Path, time.Now(), f)\n\t}\n}","func (r *Router) masterHandler(c *gin.Context) {\n\tklog.V(4).Infof(\"no router for method:%s, url:%s\", c.Request.Method, c.Request.URL.Path)\n\tc.JSON(404, gin.H{\n\t\t\"Method\": c.Request.Method,\n\t\t\"Path\": c.Request.URL.Path,\n\t\t\"error\": \"router not found\"})\n}","func Serve(root, iface string, port int) (err error) {\n\t// Root the path, and clean it if necessary.\n\n\t// 18/01/2013 It might make sense to move this to a helper routine\n\t// or further up in the stack.\n\tif !path.IsAbs(root) {\n\t\tvar wd string\n\t\twd, err = os.Getwd()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\troot = path.Join(wd, root)\n\t} else {\n\t\troot = path.Clean(root)\n\t}\n\tl.Printf(\"Starting http server %s:%d\\nRoot path %q\", iface, port, root)\n\trootpath = root\n\n\thttp.HandleFunc(\"/p/\", handle)\n\thttp.ListenAndServe(iface+\":\"+strconv.Itoa(port), nil)\n\treturn\n}","func (self *CentralBooking) InstallHandlers(router *mux.Router) {\n router.\n Methods(\"POST\").\n Path(\"/register/instance\").\n HandlerFunc(self.RegisterInstance)\n\n // apeing vault\n router.\n Methods(\"GET\").\n Path(\"/sys/health\").\n HandlerFunc(self.CheckHealth)\n}","func (r *SilencedRouter) Mount(parent *mux.Router) {\n\troutes := ResourceRoute{\n\t\tRouter: parent,\n\t\tPathPrefix: \"/namespaces/{namespace}/{resource:silenced}\",\n\t}\n\n\thandlers := handlers.NewHandlers[*corev2.Silenced](r.store)\n\n\troutes.Del(handlers.DeleteResource)\n\troutes.Get(r.get)\n\troutes.Post(r.create)\n\troutes.Put(r.createOrReplace)\n\troutes.List(r.listr, corev3.SilencedFields)\n\troutes.ListAllNamespaces(r.listr, \"/{resource:silenced}\", corev3.SilencedFields)\n\n\t// Custom routes for listing by subscription and checks for a specific\n\t// namespace, in addition to all namespaces for checks.\n\troutes.Router.HandleFunc(\"/{resource:silenced}/checks/{check}\", listHandler(r.list)).Methods(http.MethodGet)\n\troutes.Router.HandleFunc(routes.PathPrefix+\"/subscriptions/{subscription}\", listHandler(r.list)).Methods(http.MethodGet)\n\troutes.Router.HandleFunc(routes.PathPrefix+\"/checks/{check}\", listHandler(r.list)).Methods(http.MethodGet)\n}"],"string":"[\n \"func (mx *Mux) Mount(pattern string, handler interface{}) {\\n\\t// Provide runtime safety for ensuring a pattern isn't mounted on an existing\\n\\t// routing pattern.\\n\\tif mx.tree.findPattern(pattern+\\\"*\\\") || mx.tree.findPattern(pattern+\\\"/*\\\") {\\n\\t\\tpanic(fmt.Sprintf(\\\"chi: attempting to Mount() a handler on an existing path, '%s'\\\", pattern))\\n\\t}\\n\\n\\t// Assign Sub-Router's with the parent not found & method not allowed handler if not specified.\\n\\tsubr, ok := handler.(*Mux)\\n\\tif ok && subr.notFoundHandler == nil && mx.notFoundHandler != nil {\\n\\t\\tsubr.NotFound(mx.notFoundHandler)\\n\\t}\\n\\tif ok && subr.methodNotAllowedHandler == nil && mx.methodNotAllowedHandler != nil {\\n\\t\\tsubr.MethodNotAllowed(mx.methodNotAllowedHandler)\\n\\t}\\n\\n\\thttpHandler := HttpHandler(handler)\\n\\tvar mh ContextHandler\\n\\n\\tif mux, ok := handler.(Router); ok {\\n\\t\\tmux.SetPrefix(pattern)\\n\\t\\t// Wrap the Sub-router in a handlerFunc to scope the request path for routing.\\n\\t\\tmmx, ok := mux.(*Mux)\\n\\n\\t\\tif ok {\\n\\t\\t\\tmmx.parent = mx\\n\\t\\t}\\n\\n\\t\\tmh = &MountHandler{func(w http.ResponseWriter, r *http.Request, ctx *RouteContext) {\\n\\t\\t\\tctx.RoutePath = mx.nextRoutePath(ctx)\\n\\t\\t\\thttpHandler.ServeHTTPContext(w, r, ctx)\\n\\t\\t}, mux}\\n\\t} else {\\n\\t\\t// Wrap the Sub-router in a handlerFunc to scope the request path for routing.\\n\\t\\tmh = &MountHandler{func(w http.ResponseWriter, r *http.Request, ctx *RouteContext) {\\n\\t\\t\\tctx.RoutePath = mx.nextRoutePath(ctx)\\n\\t\\t\\thttpHandler.ServeHTTPContext(w, r, ctx)\\n\\t\\t}, handler}\\n\\t}\\n\\n\\tif pattern == \\\"\\\" || pattern[len(pattern)-1] != '/' {\\n\\t\\tnotFoundHandler := HttpHandler(func(w http.ResponseWriter, r *http.Request, arg *RouteContext) {\\n\\t\\t\\tif nfh := mx.NotFoundHandler(); nfh != nil {\\n\\t\\t\\t\\tnfh.ServeHTTPContext(w, r, arg)\\n\\t\\t\\t}\\n\\t\\t})\\n\\n\\t\\tmx.handle(ALL|STUB, pattern, mh)\\n\\t\\tmx.handle(ALL|STUB, pattern+\\\"/\\\", notFoundHandler)\\n\\t\\tpattern += \\\"/\\\"\\n\\t}\\n\\n\\tmethod := ALL\\n\\tsubroutes, _ := handler.(Routes)\\n\\tif subroutes != nil {\\n\\t\\tmethod |= STUB\\n\\t}\\n\\n\\tfor _, n := range mx.handle(method, pattern+\\\"*\\\", mh) {\\n\\t\\tif subroutes != nil {\\n\\t\\t\\tn.subroutes = subroutes\\n\\t\\t}\\n\\t}\\n}\",\n \"func Mount(r *mux.Router, path string, handler http.Handler) {\\n\\tr.PathPrefix(path).Handler(\\n\\t\\thttp.StripPrefix(\\n\\t\\t\\tstrings.TrimSuffix(path, \\\"/\\\"),\\n\\t\\t\\tAddSlashForRoot(handler),\\n\\t\\t),\\n\\t)\\n}\",\n \"func Mount(multiplexer multiplexer, pattern string, handler http.Handler) {\\n\\tpattern = pathkit.Clean(pattern)\\n\\thandler = http.StripPrefix(pattern, handler)\\n\\thandler = MountPoint(pattern, handler)\\n\\tmultiplexer.Handle(pattern, handler)\\n\\tmultiplexer.Handle(pattern+`/`, handler)\\n}\",\n \"func (s *server) Mount(path string, handler http.Handler) {\\n\\ts.Router.Mount(path, handler)\\n}\",\n \"func (a *App) MountHandler(verb, path string, handler Handler, mw ...Middleware) {\\n\\thandler = wrapMiddleware(mw, handler)\\n\\n\\t// wrap application level middlewares\\n\\thandler = wrapMiddleware(a.mw, handler)\\n\\n\\th := func(w http.ResponseWriter, r *http.Request) {\\n\\t\\t// TODO: start tracer span\\n\\n\\t\\t// add relevant values the context for propagation\\n\\t\\tv := Values{\\n\\t\\t\\tNow: time.Now(),\\n\\t\\t}\\n\\t\\tctx := context.WithValue(r.Context(), KeyValues, &v)\\n\\n\\t\\tif err := handler(ctx, w, r); err != nil {\\n\\t\\t\\tlog.Println(err)\\n\\t\\t\\ta.Shutdown()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t}\\n\\ta.HandleFunc(path, h).Methods(verb)\\n}\",\n \"func Mount(mux *http.ServeMux) {\\n\\tmux.HandleFunc(\\\"/\\\", index) // list all new\\n\\tmux.Handle(\\\"/upload/\\\", http.StripPrefix(\\\"/upload\\\", http.FileServer(http.Dir(\\\"upload\\\"))))\\n\\t//mux.HandleFunc(\\\"/news/\\\", newsView)\\n\\tmux.Handle(\\\"/news/\\\", http.StripPrefix(\\\"/news\\\", http.HandlerFunc(newsView)))\\n\\tmux.HandleFunc(\\\"/register\\\", adminRegister)\\n\\t// mux.Handle(\\\"/news/\\\", http.StripPrefix(\\\"/news\\\", http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\\n\\t// \\tid := r.URL.Path[1:]\\n\\t// \\tnewsView(id).ServeHTTP(w, r)\\n\\t// })))\\n\\n\\tadminMux := http.NewServeMux()\\n\\tadminMux.HandleFunc(\\\"/login\\\", adminLogin)\\n\\tadminMux.HandleFunc(\\\"/listTest\\\", adminListTest)\\n\\tadminMux.HandleFunc(\\\"/create\\\", adminCreate)\\n\\tadminMux.HandleFunc(\\\"/createTest\\\", adminCreateTest)\\n\\tadminMux.HandleFunc(\\\"/edit\\\", adminEdit)\\n\\n\\tadminMux.HandleFunc(\\\"/editTest\\\", adminEditTest)\\n\\tadminMux.HandleFunc(\\\"/runTest\\\", adminRunTest)\\n\\n\\tmux.Handle(\\\"/admin/\\\", http.StripPrefix(\\\"/admin\\\", onlyAdmin(adminMux)))\\n\\t// /news/:path\\n\\t// /admin/login/\\n\\t// /admin/list\\n\\t// /admin/create\\n\\t// /admin/edit\\n\\n}\",\n \"func Mount(mux *http.ServeMux) {\\n\\tmux.Handle(\\\"/\\\", fetchUser(http.HandlerFunc(index))) // list all news\\n\\tmux.Handle(\\\"/upload/\\\", http.StripPrefix(\\\"/upload\\\", http.FileServer(http.Dir(\\\"upload\\\"))))\\n\\tmux.Handle(\\\"/news/\\\", http.StripPrefix(\\\"/news\\\", http.HandlerFunc(newsView))) // /news/:id\\n\\n\\tmux.HandleFunc(\\\"/register\\\", adminRegister)\\n\\tmux.HandleFunc(\\\"/login\\\", adminLogin)\\n\\n\\tadminMux := http.NewServeMux()\\n\\tadminMux.HandleFunc(\\\"/logout\\\", adminLogout)\\n\\tadminMux.HandleFunc(\\\"/list\\\", adminList)\\n\\tadminMux.HandleFunc(\\\"/create\\\", adminCreate)\\n\\tadminMux.HandleFunc(\\\"/edit\\\", adminEdit)\\n\\n\\tmux.Handle(\\\"/admin/\\\", http.StripPrefix(\\\"/admin\\\", onlyAdmin(adminMux)))\\n}\",\n \"func Handler(basepath string, data io.ReadSeeker) http.Handler {\\n\\tif basepath == \\\"\\\" {\\n\\t\\tbasepath = \\\"/\\\"\\n\\t}\\n\\tas := &assetfs.AssetStore{\\n\\t\\tNames: internal.AssetNames,\\n\\t\\tData: internal.Asset,\\n\\t\\tInfo: internal.AssetInfo,\\n\\t}\\n\\tfs, err := assetfs.New(as)\\n\\tif err != nil {\\n\\t\\tpanic(fmt.Sprintf(\\\"failed to create static fs: %v\\\", err))\\n\\t}\\n\\tmux := http.NewServeMux()\\n\\tfsh := http.FileServer(http.FileSystem(fs))\\n\\tif basepath != \\\"/\\\" {\\n\\t\\tfsh = http.StripPrefix(basepath, fsh)\\n\\t}\\n\\tp := assetfs.AddPrefix(basepath, BasePath)\\n\\tf := assetfs.AddPrefix(basepath, SpecFile)\\n\\tmux.HandleFunc(basepath, func(w http.ResponseWriter, r *http.Request) {\\n\\t\\tif r.URL.Path == basepath {\\n\\t\\t\\thttp.Redirect(w, r, p+\\\"?url=\\\"+f, http.StatusSeeOther)\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\tfsh.ServeHTTP(w, r)\\n\\t})\\n\\tmux.Handle(f, &handler{modTime: time.Now(), body: data})\\n\\treturn mux\\n}\",\n \"func mount(e MountableResource, path string, handler MountHandler) {\\n\\thandler(e.Group(path))\\n}\",\n \"func (h *eventServiceHTTPHandler) Mount(group *echo.Group) {\\n\\tgroup.POST(\\\"/create\\\", h.CreateEvent)\\n\\tgroup.POST(\\\"/ticket/create\\\", h.CreateTicket)\\n\\tgroup.GET(\\\"/get_info\\\", h.GetEventInfo)\\n}\",\n \"func Mount(mux *http.ServeMux) {\\n\\tmiddleware := http.NewServeMux()\\n\\tmux.Handle(\\\"/\\\", checkPathCorrect(middleware))\\n\\tmiddleware.Handle(\\\"/\\\", http.FileServer(http.Dir(\\\"./static\\\")))\\n\\tmiddleware.HandleFunc(\\\"/movies\\\", list)\\n\\tmiddleware.Handle(\\\"/movie/\\\", http.StripPrefix(\\\"/movie\\\", http.HandlerFunc(get)))\\n\\tmiddleware.HandleFunc(\\\"/favorite\\\", favorite)\\n\\tmiddleware.Handle(\\\"/public/\\\", http.StripPrefix(\\\"/public\\\", http.FileServer(http.Dir(\\\"static\\\"))))\\n}\",\n \"func (h *todoHTTPHandler) Mount(group *echo.Group) {\\n group.POST(\\\"\\\", h.CreateTodo)\\n group.GET(\\\"\\\", h.GetAllTodo)\\n group.GET(\\\"/:id\\\", h.GetTodo)\\n group.PUT(\\\"/:id\\\", h.UpdateTodo)\\n group.DELETE(\\\"/:id\\\", h.DeleteTodo)\\n group.PUT(\\\"/status/:id\\\", h.UpdateStatusTodo)\\n}\",\n \"func handleMount(source, location, fstype string, flags uintptr, data string) error {\\n\\tif err := os.MkdirAll(location, os.FileMode(0755)); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\treturn syscall.Mount(source, location, fstype, flags, data)\\n}\",\n \"func SetRootHandler(path string) {\\n\\troothandlerpath = path\\n}\",\n \"func (p *Plugin) Mount(rt *h2o.Router) {\\n\\trt.GET(\\\"/site/info\\\", p.getSiteInfo)\\n\\trt.POST(\\\"/install\\\", p.mustDatabaseEmpty, p.postInstall)\\n\\n\\trt.Group(func(r *h2o.Router) {\\n\\t\\tr.GET(\\\"/users\\\", p.indexAdminUsers)\\n\\n\\t\\tr.GET(\\\"/locales\\\", p.getAdminLocales)\\n\\t\\tr.POST(\\\"/locales\\\", p.postAdminLocales)\\n\\t\\tr.DELETE(\\\"/locales/{:code}\\\", p.deleteAdminLocales)\\n\\n\\t}, \\\"/admin\\\", p.Jwt.MustAdminMiddleware)\\n\\n\\trt.Group(func(r *h2o.Router) {\\n\\t\\tr.GET(\\\"/status\\\", p.getAdminSiteStatus)\\n\\t\\tr.POST(\\\"/info\\\", p.postAdminSiteInfo)\\n\\t\\tr.POST(\\\"/author\\\", p.postAdminSiteAuthor)\\n\\t\\tr.GET(\\\"/seo\\\", p.getAdminSiteSeo)\\n\\t\\tr.POST(\\\"/seo\\\", p.postAdminSiteSeo)\\n\\t\\tr.GET(\\\"/smtp\\\", p.getAdminSiteSMTP)\\n\\t\\tr.POST(\\\"/smtp\\\", p.postAdminSiteSMTP)\\n\\t}, \\\"/site\\\", p.Jwt.MustAdminMiddleware)\\n\\n\\trt.Crud(\\n\\t\\t\\\"/notices\\\",\\n\\t\\t[]h2o.HandlerFunc{p.indexNotices},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createNotice},\\n\\t\\t[]h2o.HandlerFunc{p.showNotice},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateNotice},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyNotice},\\n\\t)\\n\\n\\trt.Crud(\\n\\t\\t\\\"/leave-words\\\",\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.indexLeaveWords},\\n\\t\\t[]h2o.HandlerFunc{p.createLeaveWord},\\n\\t\\tnil,\\n\\t\\tnil,\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyLeaveWord},\\n\\t)\\n\\n\\trt.Crud(\\n\\t\\t\\\"/friend-links\\\",\\n\\t\\t[]h2o.HandlerFunc{p.indexFriendLinks},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createFriendLink},\\n\\t\\t[]h2o.HandlerFunc{p.showFriendLink},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateFriendLink},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyFriendLink},\\n\\t)\\n\\n\\trt.Crud(\\n\\t\\t\\\"/links\\\",\\n\\t\\t[]h2o.HandlerFunc{p.indexLinks},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createLink},\\n\\t\\t[]h2o.HandlerFunc{p.showLink},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateLink},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyLink},\\n\\t)\\n\\n\\trt.Crud(\\n\\t\\t\\\"/cards\\\",\\n\\t\\t[]h2o.HandlerFunc{p.indexCards},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createCard},\\n\\t\\t[]h2o.HandlerFunc{p.showCard},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateCard},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyCard},\\n\\t)\\n\\n\\trt.Crud(\\n\\t\\t\\\"/posts\\\",\\n\\t\\t[]h2o.HandlerFunc{p.indexPosts},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createPost},\\n\\t\\t[]h2o.HandlerFunc{p.showPost},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updatePost},\\n\\t\\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyPost},\\n\\t)\\n}\",\n \"func Test_App_Mount(t *testing.T) {\\n\\tt.Parallel()\\n\\tmicro := New()\\n\\tmicro.Get(\\\"/doe\\\", func(c *Ctx) error {\\n\\t\\treturn c.SendStatus(StatusOK)\\n\\t})\\n\\n\\tapp := New()\\n\\tapp.Mount(\\\"/john\\\", micro)\\n\\tresp, err := app.Test(httptest.NewRequest(MethodGet, \\\"/john/doe\\\", http.NoBody))\\n\\tutils.AssertEqual(t, nil, err, \\\"app.Test(req)\\\")\\n\\tutils.AssertEqual(t, 200, resp.StatusCode, \\\"Status code\\\")\\n\\tutils.AssertEqual(t, uint32(2), app.handlersCount)\\n}\",\n \"func (h *PetHandler) Mount(r chi.Router, rs Routes) {\\n\\tif rs.has(PetCreate) {\\n\\t\\tr.Post(\\\"/\\\", h.Create)\\n\\t}\\n\\tif rs.has(PetRead) {\\n\\t\\tr.Get(\\\"/{id}\\\", h.Read)\\n\\t}\\n\\tif rs.has(PetUpdate) {\\n\\t\\tr.Patch(\\\"/{id}\\\", h.Update)\\n\\t}\\n\\tif rs.has(PetDelete) {\\n\\t\\tr.Delete(\\\"/{id}\\\", h.Delete)\\n\\t}\\n\\tif rs.has(PetList) {\\n\\t\\tr.Get(\\\"/\\\", h.List)\\n\\t}\\n\\tif rs.has(PetOwner) {\\n\\t\\tr.Get(\\\"/{id}/owner\\\", h.Owner)\\n\\t}\\n}\",\n \"func AddSlashForRoot(next http.Handler) http.Handler {\\n\\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\\n\\t\\t// @TODO MAYBE ENSURE SUFFIX DOESN'T ALREADY EXIST?\\n\\t\\tif r.URL.Path == \\\"\\\" {\\n\\t\\t\\tr.URL.Path = \\\"/\\\"\\n\\t\\t}\\n\\n\\t\\tnext.ServeHTTP(w, r)\\n\\t})\\n}\",\n \"func init() {\\n\\thttp.HandleFunc(\\\"/savecasedriver\\\",\\thandlerSaveCaseDriver)\\n\\thttp.HandleFunc(\\\"/saveimage\\\",\\t\\t\\t\\thandlerSaveImage)\\n\\thttp.HandleFunc(\\\"/sendemail\\\",\\t\\t\\t\\thandlerSendEmail)\\n\\thttp.HandleFunc(\\\"/serve/\\\",\\t\\t\\t\\t\\thandlerServe)\\n\\thttp.HandleFunc(\\\"/login\\\",\\t\\t\\t\\t\\t\\thandlerLogin)\\n\\n\\t// API Versions\\n\\thttp.HandleFunc(\\\"/api/1.0/\\\",\\t\\t\\t\\thandlerAPI10) // API version 1.0\\n\\n http.HandleFunc(\\\"/\\\",\\t\\t\\t\\t\\t\\t\\t\\thandlerRoot)\\n}\",\n \"func (service *LoanSrvc) Mount(ctx context.Context, mux goahttp.Muxer) goahttp.Server {\\n\\tendpoints := loan.NewEndpoints(service)\\n\\tsrv := server.New(endpoints, mux, goahttp.RequestDecoder, goahttp.ResponseEncoder, api.ErrorHandler, nil)\\n\\tserver.Mount(mux, srv)\\n\\n\\tfor _, m := range srv.Mounts {\\n\\t\\tlog.WithContext(ctx).Infof(\\\"%q mounted on %s %s\\\", m.Method, m.Verb, m.Pattern)\\n\\t}\\n\\treturn srv\\n}\",\n \"func RootHandler(w http.ResponseWriter, req *http.Request) {\\n\\tw.Header().Set(\\\"Content-type\\\", \\\"text/html\\\")\\n\\tif err := req.ParseForm(); err != nil {\\n\\t\\thttp.Error(w, fmt.Sprintf(\\\"error parsing url %s\\\", err), http.StatusInternalServerError)\\n\\t}\\n\\tpath := mux.Vars(req)[\\\"path\\\"]\\n\\tif path == \\\"\\\" || path == \\\"/\\\" {\\n\\t\\tpath = \\\"index.tpl\\\"\\n\\t}\\n\\tif !strings.HasSuffix(path, \\\".tpl\\\") {\\n\\t\\tpath += \\\".tpl\\\"\\n\\t}\\n\\tif _, ok := registeredTpl[path]; !ok {\\n\\t\\tw.WriteHeader(http.StatusNotFound)\\n\\t\\tfmt.Fprintf(w, \\\"Not found\\\")\\n\\t\\treturn\\n\\t}\\n\\tif err := templates.ExecuteTemplate(w, path, Page{\\n\\t\\tTitle: \\\"Home\\\",\\n\\t}); err != nil {\\n\\t\\tlog.Printf(\\\"Error executing template: %s\\\", err)\\n\\t\\thttp.Error(w, fmt.Sprintf(\\\"error parsing template: %s\\\", err), http.StatusInternalServerError)\\n\\t}\\n}\",\n \"func (h AppServer) Handler (w http.ResponseWriter, r *http.Request) {\\n\\twasHandled := false\\n\\turlPath := r.URL.Path\\n\\tl := len(urlPath)\\n\\tif l > 0 {\\n\\t\\tif urlPath[l-1:l] != \\\"/\\\" {\\n\\t\\t\\t// tack on a trailing slash\\n\\t\\t\\turlPath = urlPath + \\\"/\\\"\\n\\t\\t}\\n\\t\\tfmt.Println(\\\"appServer handler path=\\\", urlPath)\\n\\t\\t\\n\\t\\tfor p := range h.Handlers {\\n\\t\\t\\tif len(urlPath) >= len(p) &&\\turlPath[:len(p)] == p {\\n\\t\\t\\t\\twasHandled = true\\n\\t\\t\\t\\tphf := h.Handlers[p]\\n\\t\\t\\t\\tDispatchMethod(phf, w, r)\\n\\t\\t\\t} \\n\\t\\t}\\n\\t}\\n\\tif !wasHandled {\\n\\t\\t// not specific handler, assume it's a file\\n\\t\\tif h.FileServerInst != nil {\\n\\t\\t\\tDispatchMethod(h.FileServerInst, w, r)\\n\\t\\t} else {\\n\\t\\t\\thttp.Error(w, \\\"File not Found\\\", http.StatusNotFound)\\n\\t\\t}\\n\\t}\\n\\n}\",\n \"func RootHandler(w http.ResponseWriter, r *http.Request) {\\n\\tfilePath := r.URL.Path[len(\\\"/\\\"):]\\n\\n\\tsource, err := ioutil.ReadFile(root + filePath)\\n\\tif err != nil {\\n\\t\\tsource, err = ioutil.ReadFile(root + filePath + \\\"/index.html\\\")\\n\\t\\tif err != nil {\\n\\t\\t\\tw.WriteHeader(http.StatusNotFound)\\n\\t\\t\\tw.Write([]byte(err.Error()))\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\tfilePath += \\\"index.html\\\"\\n\\t}\\n\\n\\t// Set response headers\\n\\theaders.SetDefaultHeaders(w)\\n\\theaders.SetContentTypeHeader(w, filePath)\\n\\n\\t// Send the response\\n\\tw.Write(source)\\n\\n\\t//TODO: Log more detailed information.\\n\\tlog.Println(\\\"(rootHandler) The requested file has been sent: \\\", root+filePath)\\n}\",\n \"func (h *RestHandler) Mount(root *echo.Group) {\\n\\tv1Root := root.Group(helper.V1)\\n\\n\\tcustomer := v1Root.Group(\\\"/customer\\\")\\n\\tcustomer.GET(\\\"\\\", h.hello)\\n}\",\n \"func rootHandler(w http.ResponseWriter, r *http.Request) {\\n\\tparts := strings.Split(r.URL.Path, \\\"/\\\")\\n\\tif r.URL.Path == \\\"/igcinfo/api/\\\" {\\n\\t\\tinfoHandler(w, r)\\n\\t\\treturn\\n\\t} else if r.URL.Path == \\\"/igcinfo/api/igc/\\\" {\\n\\t\\tigcHandler(w, r)\\n\\t\\treturn\\n\\t} else if id, err := uuid.Parse(parts[4]); strings.HasPrefix(r.URL.Path, \\\"/igcinfo/api/igc/\\\") && err == nil && len(parts) < 6 {\\n\\t\\ttrackHandler(w, r, id)\\n\\t\\treturn\\n\\t} else if id, err := uuid.Parse(parts[4]); strings.HasPrefix(r.URL.Path, \\\"/igcinfo/api/igc/\\\") && err == nil && len(parts[5]) > 0 {\\n\\t\\ttrackFieldHandler(w, r, id, parts[5])\\n\\t\\treturn\\n\\t}\\n\\n\\thttp.NotFound(w, r)\\n}\",\n \"func Handle(pattern string, handler Handler) { DefaultServeMux.Handle(pattern, handler) }\",\n \"func Handle(pattern string, handler Handler) { DefaultServeMux.Handle(pattern, handler) }\",\n \"func Handle(pattern string, handler Handler) { DefaultServeMux.Handle(pattern, handler) }\",\n \"func catchAllHandler(w http.ResponseWriter, r *http.Request) {\\n\\tp := path.Clean(\\\"/\\\" + r.URL.Path)\\n\\tif p == \\\"/\\\" {\\n\\t\\tp += \\\"index\\\"\\n\\t}\\n\\tp = filepath.Join(config.Dir, filepath.FromSlash(p))\\n\\n\\tif _, err := os.Stat(p); os.IsNotExist(err) {\\n\\t\\tserveTemplate(w, r)\\n\\t\\treturn\\n\\t}\\n\\n\\thttp.ServeFile(w, r, p)\\n}\",\n \"func (d *Daemon) rootHandler(w *rest.ResponseWriter, r *rest.Request) {\\n\\tw.Header().Set(\\\"Access-Control-Allow-Origin\\\", \\\"*\\\")\\n\\tdata, _ := Asset(\\\"gui/index.html\\\")\\n\\tw.Write(data)\\n}\",\n \"func startHandlers(path string) {\\n\\n\\thttp.HandleFunc(\\\"/\\\", handleIndex)\\n\\thttp.HandleFunc(\\\"/login\\\", handleLogin)\\n\\thttp.HandleFunc(\\\"/logout\\\", handleLogout)\\n\\thttp.HandleFunc(\\\"/create_ticket\\\", handleCreateTicket)\\n\\thttp.HandleFunc(\\\"/holiday\\\", handleHoliday)\\n\\thttp.HandleFunc(\\\"/ticketSend\\\", handleTicketSent)\\n\\thttp.HandleFunc(\\\"/ticket\\\", handleTicket)\\n\\thttp.HandleFunc(\\\"/updateTicket\\\", handleUpdateTicket)\\n\\thttp.HandleFunc(\\\"/unassignTicket\\\", handleUnassignTicket)\\n\\thttp.HandleFunc(\\\"/assignTicket\\\", handleAssignTicket)\\n\\n\\t// Map the css, js and img folders to the location specified\\n\\thttp.Handle(\\\"/static/\\\", http.StripPrefix(\\\"/static/\\\", http.FileServer(http.Dir(path+\\\"/static\\\"))))\\n}\",\n \"func RootHandler(w http.ResponseWriter, r *http.Request){\\n\\t// We add this if statement to mention that the root path is explicitly the \\\"/\\\"\\n\\tif r.URL.Path != \\\"/\\\"{\\n\\t\\tw.WriteHeader(http.StatusNotFound)\\n\\t\\tw.Write([]byte(\\\"The Access is not Defined \\\\n\\\"))\\n\\t\\treturn\\n\\t}\\n\\t// if the path is correct then the code will be written to the server\\n\\tw.WriteHeader(http.StatusOK)\\n\\tw.Write([]byte(\\\"RUNNING API VERSION 1 \\\\n\\\"))\\n\\tfmt.Fprint(w, \\\"The easiest API \\\\n\\\")\\n\\n}\",\n \"func Mount(mux goahttp.Muxer, h *Server) {\\n\\tMountAddHandler(mux, h.Add)\\n\\tMountGetHandler(mux, h.Get)\\n\\tMountUpdateHandler(mux, h.Update)\\n\\tMountListMineHandler(mux, h.ListMine)\\n\\tMountListProjectHandler(mux, h.ListProject)\\n\\tMountPhotoHandler(mux, h.Photo)\\n\\tMountCORSHandler(mux, h.CORS)\\n}\",\n \"func root(ctx context.Context, d *aegis.HandlerDependencies, req *aegis.APIGatewayProxyRequest, res *aegis.APIGatewayProxyResponse, params url.Values) error {\\n\\thost := req.GetHeader(\\\"Host\\\")\\n\\tstage := req.RequestContext.Stage\\n\\tres.HTML(200, \\\"

Welcome

This is an unprotected route.
Click here to login.

\\\")\\n\\treturn nil\\n}\",\n \"func handleRoot(w http.ResponseWriter, r *http.Request) {\\n\\tfmt.Fprintf(w, \\\"hello from admission webhook server , you have hit : %q\\\", html.EscapeString(r.URL.Path))\\n}\",\n \"func init() {\\n\\tmux = http.NewServeMux()\\n\\tmux.Handle(\\\"/\\\", handlerFn(\\\"Hello World!\\\"))\\n\\tmux.Handle(\\\"/foo\\\", handlerFn(\\\"foo\\\"))\\n\\tmux.Handle(\\\"/bar\\\", handlerFn(\\\"bar\\\"))\\n\\tmux.Handle(\\\"/baz\\\", handlerFn(\\\"baz\\\"))\\n}\",\n \"func (s *Server) handleRootRequest(w http.ResponseWriter, req *http.Request) {\\n if req.URL.Path != \\\"/\\\" {\\n http.NotFound(w, req)\\n return\\n }\\n\\n http.Redirect(w, req, \\\"/ui/\\\", http.StatusFound)\\n}\",\n \"func defaultMount(ctx context.Context, c *Component, r *http.Request) error {\\n\\treturn nil\\n}\",\n \"func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\\n\\tinitService(service)\\n\\tvar h goa.Handler\\n\\n\\th = ctrl.FileHandler(\\\"/swagger-ui/*filepath\\\", \\\"swagger-ui/dist\\\")\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger-ui/*filepath\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger-ui/dist\\\", \\\"route\\\", \\\"GET /swagger-ui/*filepath\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swagger.json\\\", \\\"swagger/swagger.json\\\")\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger.json\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger/swagger.json\\\", \\\"route\\\", \\\"GET /swagger.json\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swagger-ui/\\\", \\\"swagger-ui/dist/index.html\\\")\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger-ui/\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger-ui/dist/index.html\\\", \\\"route\\\", \\\"GET /swagger-ui/\\\")\\n}\",\n \"func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\\n\\tinitService(service)\\n\\tvar h goa.Handler\\n\\n\\th = ctrl.FileHandler(\\\"/swagger-ui/*filepath\\\", \\\"swagger-ui/dist\\\")\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger-ui/*filepath\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger-ui/dist\\\", \\\"route\\\", \\\"GET /swagger-ui/*filepath\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swagger.json\\\", \\\"swagger/swagger.json\\\")\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger.json\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger/swagger.json\\\", \\\"route\\\", \\\"GET /swagger.json\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swagger-ui/\\\", \\\"swagger-ui/dist/index.html\\\")\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger-ui/\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger-ui/dist/index.html\\\", \\\"route\\\", \\\"GET /swagger-ui/\\\")\\n}\",\n \"func (r *Router) Mount(pattern string, router *Router, mws ...Middleware) {\\n\\trouter.parent = r\\n\\tr.subrouters = append(r.subrouters, router)\\n\\n\\tvar p string\\n\\tif pattern == \\\"/\\\" {\\n\\t\\tp = r.pattern\\n\\t} else {\\n\\t\\tp = r.pattern + pattern\\n\\t}\\n\\trouter.pattern = p\\n\\n\\thost := r.host\\n\\tfor i, route := range router.routes {\\n\\t\\trouter.Host(route.Host())\\n\\t\\tfor _, method := range route.Methods() {\\n\\t\\t\\trouter.Handle(method, route.Pattern(), route.Handler(method))\\n\\t\\t}\\n\\n\\t\\trouter.routes = append(router.routes[:i], router.routes[i+1:]...)\\n\\t}\\n\\t// Restore previous host\\n\\tr.host = host\\n}\",\n \"func rootHandler(c appengine.Context, w http.ResponseWriter, r *http.Request) *appError {\\n\\tlogoutURL, err := user.LogoutURL(c, \\\"/\\\")\\n\\tif err != nil {\\n\\t\\tc.Warningf(\\\"creating logout URL: %v\\\", err)\\n\\t\\tlogoutURL = \\\"/\\\"\\n\\t}\\n\\tuploadURL, err := blobstore.UploadURL(c, \\\"/upload\\\", nil)\\n\\tif err != nil {\\n\\t\\treturn appErrorf(err, \\\"could not create blobstore upload url\\\")\\n\\t}\\n\\tusername := \\\"none\\\"\\n\\tif u := user.Current(c); u != nil {\\n\\t\\tusername = u.String()\\n\\t}\\n\\terr = rootTemplate.Execute(w, &rootTemplateData{\\n\\t\\tLogoutURL: logoutURL,\\n\\t\\tUploadURL: uploadURL.String(),\\n\\t\\tUser: username,\\n\\t})\\n\\tif err != nil {\\n\\t\\treturn appErrorf(err, \\\"could not write template\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func init() {\\n\\thttp.HandleFunc(\\\"/\\\", playHtmlHandle)\\n\\thttp.HandleFunc(\\\"/runid\\\", runIdHandle)\\n\\thttp.HandleFunc(\\\"/img\\\", imgHandle)\\n\\thttp.HandleFunc(\\\"/clicked\\\", clickedHandle)\\n\\thttp.HandleFunc(\\\"/cheat\\\", cheatHandle)\\n\\thttp.HandleFunc(\\\"/new\\\", newGameHandle)\\n\\thttp.HandleFunc(\\\"/help\\\", helpHtmlHandle)\\n}\",\n \"func (e *Viewer) Mount(r chi.Router) {\\n\\tr.Get(\\\"/*\\\", e.serve)\\n\\tr.Get(\\\"/swagger.spec\\\", e.load)\\n}\",\n \"func (r *Router) Mount(src *Router) *Router {\\n\\tfor k, _ := range src.Router.Routes {\\n\\t\\troute := r.Route(k.Pattern).Name(k.Noun)\\n\\t\\tfor method, handler := range k.Handlers {\\n\\t\\t\\troute.Handle(handler, method)\\n\\t\\t}\\n\\t}\\n\\treturn r\\n}\",\n \"func (handlers *Handlers) rootHandler(w http.ResponseWriter, r *http.Request) {\\n\\t_, err := w.Write([]byte(\\\"OK!!\\\\n\\\"))\\n\\tif err != nil {\\n\\t\\tlog.Println(err.Error() + \\\" Failed to write response bytes in root handler\\\")\\n\\t}\\n}\",\n \"func (h *RestHandler) Mount(root *echo.Group) {\\n\\tbot := root.Group(\\\"/bot\\\")\\n\\n\\tbot.POST(\\\"/callback\\\", h.callback)\\n\\tbot.POST(\\\"/pushmessage\\\", h.pushMessage, h.mw.HTTPBasicAuth(false))\\n}\",\n \"func Serve() {\\n\\thttp.Handle(\\\"/\\\", Handler())\\n}\",\n \"func (fs *fileServer) Serve(root http.FileSystem) http.Handler {\\n return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\\n upath := r.URL.Path\\n if !strings.HasPrefix(upath, \\\"/\\\") {\\n upath = \\\"/\\\" + upath\\n r.URL.Path = upath\\n }\\n\\n name := path.Clean(upath)\\n\\n\\n // INFO: By having html this high up, it converts the document to HTML instread of go or some other rendering method\\n //fmt.Fprintf(w, \\\"%s\\\\n\\\", \\\"/search?q=\\\", htmlReplacer.Replace(\\\"pow\\\"))\\n\\n if file, ok := fs.cache[name]; ok {\\n fs.serveContent(w, r, file.name, file.modTime, file.size, bytes.NewReader(file.data))\\n // Serves all files as HTML\\n //io.Copy(w, bytes.NewReader(file.data))\\n return\\n }\\n\\n // redirect .../index.html to .../\\n // can't use Redirect() because that would make the path absolute,\\n // which would be a problem running under StripPrefix\\n if strings.HasSuffix(r.URL.Path, fs.indexPage) {\\n localRedirect(w, r, \\\"./\\\")\\n return\\n }\\n\\n f, err := root.Open(name)\\n if err != nil {\\n http.NotFound(w, r)\\n return\\n }\\n defer f.Close()\\n\\n d, err := f.Stat()\\n if err != nil {\\n http.NotFound(w, r)\\n return\\n }\\n\\n // redirect to canonical path: / at end of directory url\\n // r.URL.Path always begins with /\\n url := r.URL.Path\\n if d.IsDir() {\\n if url[len(url)-1] != '/' {\\n localRedirect(w, r, path.Base(url)+\\\"/\\\")\\n return\\n }\\n } else {\\n if url[len(url)-1] == '/' {\\n localRedirect(w, r, \\\"../\\\"+path.Base(url))\\n return\\n }\\n }\\n\\n \\n // use contents of index.html for directory, if present\\n if d.IsDir() {\\n index := name + fs.indexPage\\n ff, err := root.Open(index)\\n if err == nil {\\n defer ff.Close()\\n dd, err := ff.Stat()\\n if err == nil {\\n d = dd\\n f = ff\\n }\\n }\\n }\\n\\n // Still a directory? (we didn't find an index.html file)\\n if d.IsDir() {\\n if checkLastModified(w, r, d.ModTime()) {\\n return\\n }\\n dirList(w, f)\\n return\\n }\\n\\n fs.serveContent(w, r, d.Name(), d.ModTime(), d.Size(), f)\\n\\n var buf bytes.Buffer\\n io.Copy(&buf, f)\\n\\n fs.cache[name] = file{\\n name: d.Name(),\\n modTime: d.ModTime(),\\n data: buf.Bytes(),\\n }\\n })\\n}\",\n \"func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\\n\\tinitService(service)\\n\\tvar h goa.Handler\\n\\tservice.Mux.Handle(\\\"OPTIONS\\\", \\\"/swagger/*filepath\\\", ctrl.MuxHandler(\\\"preflight\\\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\\n\\n\\th = ctrl.FileHandler(\\\"/swagger/*filepath\\\", \\\"public/swagger/\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger/*filepath\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"public/swagger/\\\", \\\"route\\\", \\\"GET /swagger/*filepath\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swagger/\\\", \\\"public/swagger/index.html\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger/\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"public/swagger/index.html\\\", \\\"route\\\", \\\"GET /swagger/\\\")\\n}\",\n \"func Mount(m *httpmux.Mux) {\\n\\tm.Handle(\\\"/\\\", methodmux.Get(\\n\\t\\thime.Handler(index),\\n\\t))\\n\\tm.Handle(\\\"/signout\\\", methodmux.Post(\\n\\t\\thime.Handler(signOut),\\n\\t))\\n\\n\\tprofile := m.Group(\\\"/profile\\\", mustSignedIn)\\n\\tprofile.Handle(\\\"/\\\", methodmux.Get(\\n\\t\\thime.Handler(getProfile),\\n\\t))\\n\\tprofile.Handle(\\\"/edit\\\", methodmux.GetPost(\\n\\t\\thime.Handler(getProfileEdit),\\n\\t\\thime.Handler(postProfileEdit),\\n\\t))\\n\\n\\t// course\\n\\tm.Handle(\\\"/course/\\\", prefixhandler.New(\\\"/course\\\", courseIDKey{}, newCourseHandler()))\\n}\",\n \"func rootHandler(w http.ResponseWriter, r *http.Request) {\\r\\n\\t// Parsea la plantilla root.html \\r\\n\\tif t, err := template.ParseFiles(filepath.Join(templates, \\\"root.html\\\")); err != nil {\\r\\n\\t\\t// Se ha presentado un error\\r\\n\\t\\thttp.Error(w, err.Error(), http.StatusInternalServerError)\\r\\n\\t} else {\\r\\n\\t\\t// retorna la respuesta al cliente por medio de t.Execute\\r\\n\\t\\tt.Execute(w, nil)\\r\\n\\t}\\r\\n}\",\n \"func init() {\\n\\thttp.HandleFunc(\\\"/\\\", errorAdapter(rootHandler))\\n}\",\n \"func MountAddHandler(mux goahttp.Muxer, h http.Handler) {\\n\\tf, ok := handleStationOrigin(h).(http.HandlerFunc)\\n\\tif !ok {\\n\\t\\tf = func(w http.ResponseWriter, r *http.Request) {\\n\\t\\t\\th.ServeHTTP(w, r)\\n\\t\\t}\\n\\t}\\n\\tmux.Handle(\\\"POST\\\", \\\"/stations\\\", f)\\n}\",\n \"func (h *UserHandler) Mount(r chi.Router, rs Routes) {\\n\\tif rs.has(UserCreate) {\\n\\t\\tr.Post(\\\"/\\\", h.Create)\\n\\t}\\n\\tif rs.has(UserRead) {\\n\\t\\tr.Get(\\\"/{id}\\\", h.Read)\\n\\t}\\n\\tif rs.has(UserUpdate) {\\n\\t\\tr.Patch(\\\"/{id}\\\", h.Update)\\n\\t}\\n\\tif rs.has(UserDelete) {\\n\\t\\tr.Delete(\\\"/{id}\\\", h.Delete)\\n\\t}\\n\\tif rs.has(UserList) {\\n\\t\\tr.Get(\\\"/\\\", h.List)\\n\\t}\\n\\tif rs.has(UserPets) {\\n\\t\\tr.Get(\\\"/{id}/pets\\\", h.Pets)\\n\\t}\\n}\",\n \"func (h *HTTPHandler) Mount(g *echo.Group) {\\n\\tg.GET(\\\"/users\\\", h.FindAllUser)\\n}\",\n \"func setuphandlers() {\\n\\thttp.HandleFunc(\\\"/\\\", rootHandler)\\n\\thttp.HandleFunc(\\\"/status\\\", statusHandler)\\n\\thttp.HandleFunc(\\\"/stats\\\", statsHandler)\\n\\thttp.HandleFunc(\\\"/request\\\", requestHandler)\\n}\",\n \"func Handle(path string, v interface{}, options ...RestFunc) {\\n\\tDefaultServeMux.Handle(path, v, options...)\\n}\",\n \"func rootHandler(w http.ResponseWriter, r *http.Request) {\\n\\thttp.Error(w, \\\"Image Server\\\", http.StatusNotImplemented)\\n}\",\n \"func (s *Server) route() {\\n\\ts.r.HandleFunc(\\\"/\\\", s.HandleArticlesGetAll())\\n\\ts.r.HandleFunc(\\\"/article/{slug}\\\", s.HandleArticlesGetOne())\\n\\ts.r.HandleFunc(\\\"/talks\\\", s.HandleTalksGetOne())\\n\\ts.r.HandleFunc(\\\"/talks/{slug}\\\", s.HandleTalksGetOne())\\n\\ts.r.HandleFunc(\\\"/about\\\", s.HandleAboutGet())\\n}\",\n \"func main() {\\n\\tlog.SetOutput(os.Stdout)\\n\\tflag.Parse()\\n\\tvar router = mux.NewRouter()\\n\\tvar api = router.PathPrefix(\\\"/api\\\").Subrouter()\\n\\trouter.PathPrefix(\\\"/static/\\\").Handler(http.StripPrefix(\\\"/static/\\\", http.FileServer(http.Dir(\\\"./static\\\"))))\\n\\tapi.NotFoundHandler = http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\\n\\t\\tw.WriteHeader(http.StatusNotFound)\\n\\t})\\n\\n\\t//api.Use(func(next http.Handler) http.Handler {\\n\\t//\\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\\n\\t//\\t\\tif r.Header.Get(\\\"x-auth-token\\\") != \\\"admin\\\" {\\n\\t//\\t\\t\\tw.WriteHeader(http.StatusUnauthorized)\\n\\t//\\t\\t\\treturn\\n\\t//\\t\\t}\\n\\t//\\t\\tlog.Println(r.RequestURI)\\n\\t//\\t\\tnext.ServeHTTP(w, r)\\n\\t//\\t})\\n\\t//})\\n\\n\\tvar api1 = api.PathPrefix(\\\"/v1\\\").Subrouter()\\n\\tapi1.HandleFunc(\\\"/status\\\", func(w http.ResponseWriter, r *http.Request) {\\n\\t\\tw.WriteHeader(http.StatusOK)\\n\\t})\\n\\tapi1.NotFoundHandler = http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\\n\\t\\tw.WriteHeader(http.StatusNotFound)\\n\\t})\\n\\n\\tglueapi.Handlers(api1.PathPrefix(\\\"/glue/\\\").Subrouter())\\n\\ts3api.Handlers(api1.PathPrefix(\\\"/s3/\\\").Subrouter())\\n\\n\\tsrv := &http.Server{\\n\\t\\tHandler: router,\\n\\t\\tAddr: \\\":\\\" + *port,\\n\\t\\t// Good practice: enforce timeouts for servers you create!\\n\\t\\tWriteTimeout: 15 * time.Second,\\n\\t\\tReadTimeout: 15 * time.Second,\\n\\t}\\n\\n\\tlog.Fatal(srv.ListenAndServe())\\n\\t//http.ListenAndServe(\\\":\\\"+*port, router)\\n}\",\n \"func handleWebroot(webroot string) {\\n\\tdefer func() {\\n\\t\\tif r := recover(); r != nil {\\n\\t\\t\\tlog.Fatal(\\\"when using --webroot, --endpoint must not equal '/': \\\", r)\\n\\t\\t}\\n\\t}()\\n\\n\\t_, err := os.Stat(webroot)\\n\\tif err != nil && webroot != \\\"\\\" {\\n\\t\\tlog.Fatal(\\\"webroot '\\\", webroot, \\\"' inaccessible: \\\", err)\\n\\t}\\n\\n\\tif webroot != \\\"\\\" {\\n\\t\\thandler := responders.PublicResponder(webroot)\\n\\t\\thttp.Handle(\\\"/\\\", handler)\\n\\t}\\n}\",\n \"func main() {\\n\\n\\t// use custom handler\\n\\n\\txhandler := routes.RegexpHandler{}\\n\\n\\txhandler.HandleRegexp(\\\"/abc\\\\\\\\=.+\\\", handleAbc)\\n\\n\\txhandler.HandleRegexp(\\\"/\\\", handleRoot)\\n\\n\\thttp.Handle(\\\"/\\\", xhandler)\\n\\n\\t// use net/http handler\\n\\n\\thttp.HandleFunc(\\\"/zzz/\\\", handleSpec)\\n\\n\\thttp.HandleFunc(\\\"/sublvl\\\", func(w http.ResponseWriter, rq *http.Request) {\\n\\t\\thandleRoot(w, rq)\\n\\t})\\n\\n\\tlog.Fatal(http.ListenAndServe(\\\":8080\\\", nil))\\n}\",\n \"func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\\n\\tinitService(service)\\n\\tvar h goa.Handler\\n\\tservice.Mux.Handle(\\\"OPTIONS\\\", \\\"/swagger.json\\\", ctrl.MuxHandler(\\\"preflight\\\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\\n\\tservice.Mux.Handle(\\\"OPTIONS\\\", \\\"/swaggerui/*filepath\\\", ctrl.MuxHandler(\\\"preflight\\\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\\n\\n\\th = ctrl.FileHandler(\\\"/swagger.json\\\", \\\"swagger/swagger.json\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger.json\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger/swagger.json\\\", \\\"route\\\", \\\"GET /swagger.json\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swaggerui/*filepath\\\", \\\"swaggerui\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swaggerui/*filepath\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swaggerui\\\", \\\"route\\\", \\\"GET /swaggerui/*filepath\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swaggerui/\\\", \\\"swaggerui/index.html\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swaggerui/\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swaggerui/index.html\\\", \\\"route\\\", \\\"GET /swaggerui/\\\")\\n}\",\n \"func routeHandler(res http.ResponseWriter, req *http.Request) {\\n\\tpath := req.URL.Path\\n\\tswitch path {\\n\\tcase \\\"/\\\": HomeRoute(res, req)\\n\\tcase \\\"/test\\\": TestRoute(res, req)\\n\\tcase \\\"/login\\\": LoginRoute(res, req)\\n\\tcase \\\"/register\\\": RegisterRoute(res, req)\\n case \\\"/logout\\\": LogoutRoute(res, req)\\n\\tcase \\\"/library\\\": LibraryRoute(res, req)\\n\\tcase \\\"/upload\\\": UploadRoute(res, req)\\n\\tcase \\\"/read\\\": ReadRoute(res, req)\\n\\tdefault: data,err := ioutil.ReadFile(path[1:])\\n \\tif err != nil{\\n \\t\\t\\tNotFoundRoute(res, req)\\n \\t} else {\\n \\t\\t\\tres.Write(data)\\n \\t}\\n\\t}\\n}\",\n \"func (h *App) ServeHTTP(res http.ResponseWriter, req *http.Request) {\\n\\tvar head string\\n\\n\\tswitch req.URL.Path {\\n\\tcase \\\"/\\\", \\\"/orgs\\\", \\\"/users\\\", \\\"/devices\\\", \\\"/sign-in\\\", \\\"/groups\\\", \\\"/msg\\\":\\n\\t\\th.IndexHandler.ServeHTTP(res, req)\\n\\n\\tdefault:\\n\\t\\thead, req.URL.Path = ShiftPath(req.URL.Path)\\n\\t\\tswitch head {\\n\\t\\tcase \\\"public\\\":\\n\\t\\t\\th.PublicHandler.ServeHTTP(res, req)\\n\\t\\tcase \\\"v1\\\":\\n\\t\\t\\th.V1ApiHandler.ServeHTTP(res, req)\\n\\t\\tdefault:\\n\\t\\t\\thttp.Error(res, \\\"Not Found\\\", http.StatusNotFound)\\n\\t\\t}\\n\\t}\\n}\",\n \"func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\\n\\tinitService(service)\\n\\tvar h goa.Handler\\n\\tservice.Mux.Handle(\\\"OPTIONS\\\", \\\"/swagger-ui/*filepath\\\", ctrl.MuxHandler(\\\"preflight\\\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\\n\\tservice.Mux.Handle(\\\"OPTIONS\\\", \\\"/swagger.json\\\", ctrl.MuxHandler(\\\"preflight\\\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\\n\\n\\th = ctrl.FileHandler(\\\"/swagger-ui/*filepath\\\", \\\"swagger-ui/\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger-ui/*filepath\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger-ui/\\\", \\\"route\\\", \\\"GET /swagger-ui/*filepath\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swagger.json\\\", \\\"swagger/swagger.json\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger.json\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger/swagger.json\\\", \\\"route\\\", \\\"GET /swagger.json\\\")\\n\\n\\th = ctrl.FileHandler(\\\"/swagger-ui/\\\", \\\"swagger-ui/index.html\\\")\\n\\th = handleSwaggerOrigin(h)\\n\\tservice.Mux.Handle(\\\"GET\\\", \\\"/swagger-ui/\\\", ctrl.MuxHandler(\\\"serve\\\", h, nil))\\n\\tservice.LogInfo(\\\"mount\\\", \\\"ctrl\\\", \\\"Swagger\\\", \\\"files\\\", \\\"swagger-ui/index.html\\\", \\\"route\\\", \\\"GET /swagger-ui/\\\")\\n}\",\n \"func HandleRoot(w http.ResponseWriter, r *http.Request) (err error) {\\n\\tw.Header().Set(\\\"Content-Type\\\", \\\"text/html; charset=utf-8\\\")\\n\\tw.WriteHeader(http.StatusOK)\\n\\tw.Write([]byte(`Login using Twitch`))\\n\\n\\treturn\\n}\",\n \"func IndexHandler(ctx *Context, w http.ResponseWriter, r *http.Request) (int, error) {\\n\\tvar host string\\n\\tif r.Header.Get(\\\"X-Forwarded-Host\\\") != \\\"\\\" {\\n\\t\\thost = r.Header.Get(\\\"X-Forwarded-Host\\\")\\n\\t} else {\\n\\t\\thost = r.Host\\n\\t}\\n\\n\\tvar hostConfig *gabs.Container\\n\\tvar ok bool\\n\\n\\t// Check if host present in Config.\\n\\tchildren := ctx.Config.ChildrenMap()\\n\\tif hostConfig, ok = children[host]; !ok {\\n\\t\\treturn 404, fmt.Errorf(\\\"Shortcut '%s' not found in Config.\\\", host)\\n\\t}\\n\\n\\ttokens := tokenize(host + r.URL.Path)\\n\\n\\t// Set up handles on token and Config. We might need to skip ahead if there's a custom schema set.\\n\\ttokensStart := tokens.Front()\\n\\tconf := ctx.Config\\n\\n\\tvar path bytes.Buffer\\n\\tif s := hostConfig.Path(sslKey).Data(); s != nil && s.(bool) {\\n\\t\\tpath.WriteString(httpPrefix)\\n\\t} else if s := hostConfig.Path(schemaKey).Data(); s != nil && s.(string) != \\\"\\\" {\\n\\t\\tpath.WriteString(hostConfig.Path(schemaKey).Data().(string) + \\\":/\\\")\\n\\t\\t// move one token ahead to parse expansions correctly.\\n\\t\\tconf = conf.ChildrenMap()[tokensStart.Value.(string)]\\n\\t\\ttokensStart = tokensStart.Next()\\n\\t} else {\\n\\t\\t// Default to regular https prefix.\\n\\t\\tpath.WriteString(httpsPrefix)\\n\\t}\\n\\n\\tExpandPath(conf, tokensStart, &path)\\n\\n\\t// send result\\n\\thttp.Redirect(w, r, path.String(), http.StatusFound)\\n\\n\\treturn 302, nil\\n}\",\n \"func (srv *Server) ApplyHandlers() {\\n\\tsrv.router.Handle(\\\"/*\\\", http.FileServer(http.Dir(\\\"./web\\\")))\\n\\tsrv.router.Get(\\\"/socket\\\", srv.socketHandler)\\n}\",\n \"func homeHandler(env *Env, w http.ResponseWriter, r *http.Request) error {\\n return renderTemplate(w, \\\"index\\\", \\\"base\\\", nil)\\n}\",\n \"func Mount(mux goahttp.Muxer, h *Server) {\\n\\tMountRegisterHandler(mux, h.Register)\\n\\tMountRetrieveHandler(mux, h.Retrieve)\\n\\tMountHmrcCallbackHandler(mux, h.HmrcCallback)\\n}\",\n \"func (mux *ServeMux) handler(host, path string) (h http.Handler, pattern string) {\\n\\tmux.mu.RLock()\\n\\tdefer mux.mu.RUnlock()\\n\\n\\t// Host-specific pattern takes precedence over generic ones\\n\\tif mux.hosts {\\n\\t\\th, pattern = mux.match(host + path)\\n\\t}\\n\\tif h == nil {\\n\\t\\th, pattern = mux.match(path)\\n\\t}\\n\\tif h == nil {\\n\\t\\th, pattern = http.NotFoundHandler(), \\\"\\\"\\n\\t}\\n\\treturn\\n}\",\n \"func (mux *ServeMux) handler(host, path string) (h Handler, pattern string) {\\n\\tmux.mu.RLock()\\n\\tdefer mux.mu.RUnlock()\\n\\n\\t// Host-specific pattern takes precedence over generic ones\\n\\tif mux.hosts {\\n\\t\\th, pattern = mux.match(host + path)\\n\\t}\\n\\tif h == nil {\\n\\t\\th, pattern = mux.match(path)\\n\\t}\\n\\tif h == nil {\\n\\t\\th, pattern = NotFoundHandler(), \\\"\\\"\\n\\t}\\n\\treturn\\n}\",\n \"func (r *Router) handle(c *Ctx) {\\n\\tvar handler HandlerFunc\\n\\treq := c.Request()\\n\\tw := c.Writer()\\n\\tpath := req.URL.Path\\n\\tmethod := req.Method\\n\\tres := r.trie.Match(path)\\n\\n\\tif res.Node == nil {\\n\\t\\t// FixedPathRedirect or TrailingSlashRedirect\\n\\t\\tif res.TSR != \\\"\\\" || res.FPR != \\\"\\\" {\\n\\t\\t\\treq.URL.Path = res.TSR\\n\\t\\t\\tif res.FPR != \\\"\\\" {\\n\\t\\t\\t\\treq.URL.Path = res.FPR\\n\\t\\t\\t}\\n\\t\\t\\tcode := 301\\n\\t\\t\\tif method != \\\"GET\\\" {\\n\\t\\t\\t\\tcode = 307\\n\\t\\t\\t}\\n\\t\\t\\thttp.Redirect(w, req, req.URL.String(), code)\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\tif r.noRoute == nil {\\n\\t\\t\\thttp.Error(w, fmt.Sprintf(`\\\"%s\\\" not implemented`, path), 501)\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\thandler = r.noRoute\\n\\t} else {\\n\\t\\t// ok := false\\n\\t\\thd := res.Node.GetHandler(method)\\n\\t\\thandler, _ = hd.(HandlerFunc)\\n\\t\\t// handler = r.wrapHandler(hd)\\n\\t\\t// if !ok {\\n\\t\\t// \\tpanic(\\\"handler error\\\")\\n\\t\\t// }\\n\\t\\tif handler == nil {\\n\\t\\t\\t// OPTIONS support\\n\\t\\t\\tif method == http.MethodOptions {\\n\\t\\t\\t\\tw.Header().Set(\\\"Allow\\\", res.Node.GetAllow())\\n\\t\\t\\t\\tw.WriteHeader(204)\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\n\\t\\t\\tif r.noMethod == nil {\\n\\t\\t\\t\\t// If no route handler is returned, it's a 405 error\\n\\t\\t\\t\\tw.Header().Set(\\\"Allow\\\", res.Node.GetAllow())\\n\\t\\t\\t\\thttp.Error(w, fmt.Sprintf(`\\\"%s\\\" not allowed in \\\"%s\\\"`, method, path), 405)\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t\\thandler = r.noMethod\\n\\t\\t}\\n\\t}\\n\\n\\tif len(res.Params) != 0 {\\n\\t\\tc.params = res.Params\\n\\t}\\n\\tc.handlers = append(c.handlers, handler)\\n\\tc.Next()\\n}\",\n \"func MountApplication(mountpoints map[string]http.Handler, host *Host, app *Application) {\\n listing := &DirList{Host: host}\\n\\n\\t// Serve the static build files from the mountpoint path.\\n\\turl := app.PublishUrl()\\n\\tlog.Printf(\\\"Serving app %s from %s\\\", url, app.PublicDirectory())\\n fileserver := http.FileServer(http.Dir(app.PublicDirectory()))\\n mountpoints[url] = http.StripPrefix(url, PublicHandler{Listing: listing, App: app, FileServer: fileserver})\\n}\",\n \"func (s *Server) Start() *mux.Router {\\n\\trouter := mux.NewRouter()\\n\\n\\trouter.HandleFunc(\\\"/\\\", s.Config.Index)\\n\\trouter.HandleFunc(\\\"/host/{host}\\\", s.Config.HostSummary)\\n\\trouter.HandleFunc(\\\"/group/{group}\\\", s.Config.GroupSummary)\\n\\trouter.PathPrefix(\\\"/\\\").Handler(http.FileServer(http.Dir(\\\"./assets/\\\")))\\n\\n\\treturn router\\n}\",\n \"func rootHandler(w http.ResponseWriter, r *http.Request) {\\n\\t//fmt.Fprintf(w, \\\"

Hello All

\\\")\\n\\tt,_ := template.ParseFiles(\\\"root.html\\\")\\n\\tt.Execute(w, nil)\\n\\n}\",\n \"func (s *server) routes() {\\n s.router.HandleFunc(\\\"/\\\", s.homePageHandler)\\n s.router.HandleFunc(\\\"/signup/\\\", s.signupHandler)\\n s.router.HandleFunc(\\\"/signin/\\\", s.signinHandler)\\n s.router.HandleFunc(\\\"/signout/\\\", s.makeHandler(s.signoutHandler))\\n s.router.HandleFunc(\\\"/view/\\\", s.makeHandler(s.viewHandler))\\n s.router.HandleFunc(\\\"/save/\\\", s.makeHandler(s.saveHandler))\\n s.router.HandleFunc(\\\"/edit/\\\", s.makeHandler(s.editHandler))\\n s.router.HandleFunc(\\\"/delete/\\\", s.makeHandler(s.deleteHandler))\\n\\n s.validPath = regexp.MustCompile(\\n \\\"^/(new|view|save|edit|delete|signout)/([0-9]*)$\\\")\\n}\",\n \"func (mux *ServeMux) Handler(r *Request) (h Handler, pattern string) {\\n\\n\\t// CONNECT requests are not canonicalized.\\n\\tif r.Method == \\\"CONNECT\\\" {\\n\\t\\t// If r.URL.Path is /tree and its handler is not registered,\\n\\t\\t// the /tree -> /tree/ redirect applies to CONNECT requests\\n\\t\\t// but the path canonicalization does not.\\n\\t\\tif u, ok := mux.redirectToPathSlash(r.URL.Host, r.URL.Path, r.URL); ok {\\n\\t\\t\\treturn RedirectHandler(u.String(), StatusMovedPermanently), u.Path\\n\\t\\t}\\n\\n\\t\\treturn mux.handler(r.Host, r.URL.Path)\\n\\t}\\n\\n\\t// All other requests have any port stripped and path cleaned\\n\\t// before passing to mux.handler.\\n\\thost := stripHostPort(r.Host)\\n\\tpath := cleanPath(r.URL.Path)\\n\\n\\t// If the given path is /tree and its handler is not registered,\\n\\t// redirect for /tree/.\\n\\tif u, ok := mux.redirectToPathSlash(host, path, r.URL); ok {\\n\\t\\treturn RedirectHandler(u.String(), StatusMovedPermanently), u.Path\\n\\t}\\n\\n\\tif path != r.URL.Path {\\n\\t\\t_, pattern = mux.handler(host, path)\\n\\t\\tu := &url.URL{Path: path, RawQuery: r.URL.RawQuery}\\n\\t\\treturn RedirectHandler(u.String(), StatusMovedPermanently), pattern\\n\\t}\\n\\n\\treturn mux.handler(host, r.URL.Path)\\n}\",\n \"func (h *Host) initHostHandler() {\\n\\thttp.HandleFunc(h.url+\\\":\\\"+h.port+\\\"/\\\", func(w http.ResponseWriter, r *http.Request) {\\n\\t\\th.reqLog.Access(r)\\n\\t\\tw = h.addHeaders(w)\\n\\t\\tif h.static != \\\"\\\" {\\n\\t\\t\\tif strings.HasPrefix(r.URL.Path, h.static) {\\n\\t\\t\\t\\thttp.ServeFile(w, r, h.root+r.URL.Path)\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\t//get next proxy\\n\\t\\trewriteProxy := h.NextProxy()\\n\\t\\tr.RequestURI = strings.Replace(r.RequestURI, \\\"/\\\", \\\"\\\", 1)\\n\\t\\tr.URL.Path = strings.Replace(r.URL.Path, \\\"/\\\", \\\"\\\", 1)\\n\\t\\trewriteProxy.ServeHTTP(w, r)\\n\\t})\\n}\",\n \"func InitRouter() {\\n\\tr := mux.NewRouter()\\n\\n\\tr.HandleFunc(\\\"/login\\\", service.Session{}.Login)\\n\\tr.HandleFunc(\\\"/logmein\\\", service.Session{}.LogMeIn).Methods(\\\"GET\\\")\\n\\tr.HandleFunc(\\\"/logout\\\", service.Session{}.Logout).Methods(\\\"GET\\\")\\n\\n\\tr.HandleFunc(\\\"/faq\\\", func(w http.ResponseWriter, r *http.Request) {\\n\\t\\thc := &kinli.HttpContext{W: w, R: r}\\n\\t\\tpage := kinli.NewPage(hc, \\\"Frequently Asked Questions\\\", \\\"\\\", \\\"\\\", nil)\\n\\t\\tkinli.DisplayPage(w, \\\"faq\\\", page)\\n\\t}).Methods(\\\"GET\\\")\\n\\n\\tr.HandleFunc(\\\"/example\\\", func(w http.ResponseWriter, r *http.Request) {\\n\\t\\thc := &kinli.HttpContext{W: w, R: r}\\n\\t\\tpage := kinli.NewPage(hc, \\\"Example Form\\\", \\\"\\\", \\\"\\\", nil)\\n\\t\\tkinli.DisplayPage(w, \\\"example\\\", page)\\n\\t}).Methods(\\\"GET\\\")\\n\\n\\tr.HandleFunc(\\\"/\\\", func(w http.ResponseWriter, r *http.Request) {\\n\\t\\thc := &kinli.HttpContext{W: w, R: r}\\n\\t\\tpage := kinli.NewPage(hc, \\\"hello page\\\", \\\"\\\", \\\"\\\", nil)\\n\\t\\tkinli.DisplayPage(w, \\\"home\\\", page)\\n\\t}).Methods(\\\"GET\\\")\\n\\n\\tr.HandleFunc(\\\"/{uid}\\\", service.FormSubmissionRequest).Methods(\\\"POST\\\")\\n\\n\\tr.NotFoundHandler = http.HandlerFunc(notFound)\\n\\n\\tinitStatic(r)\\n\\n\\tsrv := &http.Server{\\n\\t\\tHandler: r,\\n\\t\\tAddr: common.Config.LocalServer,\\n\\t\\tWriteTimeout: 60 * time.Second,\\n\\t\\tReadTimeout: 60 * time.Second,\\n\\t}\\n\\tlog.Println(\\\"Starting server on\\\", common.Config.LocalServer)\\n\\tlog.Fatal(srv.ListenAndServe())\\n\\n}\",\n \"func (this *Router) Handle(title, path string, handlers ...HandlerFunc) *Router {\\n\\tif len(path) < 1 || path[0] != '/' || strings.Contains(path, \\\"//\\\") {\\n\\t\\tpanic(\\\"add router faild, invalid path \\\" + path)\\n\\t}\\n\\tif sepIndex := strings.Index(path[1:], \\\"/\\\") + 1; sepIndex > 1 {\\n\\t\\troot := path[:sepIndex]\\n\\t\\tsubpath := path[sepIndex:]\\n\\t\\tvar group *Router = nil\\n\\t\\tfor _, router := range this.children {\\n\\t\\t\\tif router.path == root {\\n\\t\\t\\t\\tgroup = router\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tif group == nil {\\n\\t\\t\\tgroup = this.Group(root)\\n\\t\\t}\\n\\t\\treturn group.Handle(title, subpath, handlers...)\\n\\t}\\n\\thandlerChain := append([]HandlerFunc{}, this.handlerChain...)\\n\\thandlerChain = append(handlerChain, handlers...)\\n\\trouter := &Router{\\n\\t\\ttitle: title,\\n\\t\\tpath: path,\\n\\t\\trealPath: this.realPath + path,\\n\\t\\thandlerChain: handlerChain,\\n\\t\\tchildren: []*Router{},\\n\\t}\\n\\tthis.children = append(this.children, router)\\n\\tlog.Log(\\\"DEBUG\\\", \\\"add router\\\", router.realPath)\\n\\treturn router\\n}\",\n \"func (_m *RESTHandler) Mount(group *echo.Group) {\\n\\t_m.Called(group)\\n}\",\n \"func (osh *SystemHandler) Mount(source string, target string, fsType string, flags uintptr, data string) error {\\n\\treturn syscall.Mount(source, target, fsType, flags, data)\\n}\",\n \"func (hr *httpRouter) Handler() http.Handler {\\n\\n\\tc, _ := console.New(console.Options{Color: true})\\n\\t_ = logger.Register(\\\"console\\\", logger.Config{Writer: c})\\n\\tcLogger, _ := logger.Get(\\\"console\\\")\\n\\tl := log.New(cLogger)\\n\\n\\tfmt.Print(\\\"Loading Routes...\\\")\\n\\t//add files in a directory\\n\\tro := newHttpRouterExtended(hr)\\n\\n\\tmw := middleware.Chain{}\\n\\n\\t//adding files\\n\\tfor path, file := range hr.file {\\n\\t\\tro.HandlerFunc(\\\"GET\\\", path, mw.Add(l.MW).Handle(\\n\\t\\t\\tfunc(w http.ResponseWriter, req *http.Request) {\\n\\t\\t\\t\\thttp.ServeFile(w, req, hr.file[req.Context().Value(router.PATTERN).(string)])\\n\\t\\t\\t}))\\n\\t\\tfmt.Printf(\\\"\\\\n\\\\x1b[32m %#v [GET]%v \\\\x1b[49m\\\\x1b[39m \\\", path, file)\\n\\t}\\n\\n\\t// adding directories\\n\\tfor k, path := range hr.dir {\\n\\t\\tfileServer := http.FileServer(http.Dir(path))\\n\\t\\tpattern := k + \\\"/*filepath\\\"\\n\\t\\tro.HandlerFunc(\\\"GET\\\", pattern, mw.Add(l.MW).Handle(\\n\\t\\t\\tfunc(w http.ResponseWriter, req *http.Request) {\\n\\t\\t\\t\\t//disable directory listing\\n\\t\\t\\t\\tif strings.HasSuffix(req.URL.Path, \\\"/\\\") {\\n\\t\\t\\t\\t\\thttp.NotFound(w, req)\\n\\t\\t\\t\\t\\treturn\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\tif val, ok := req.Context().Value(router.PARAMS).(map[string][]string)[\\\"filepath\\\"]; ok {\\n\\t\\t\\t\\t\\treq.URL.Path = val[0]\\n\\t\\t\\t\\t\\tfileServer.ServeHTTP(w, req)\\n\\t\\t\\t\\t\\treturn\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\thttp.NotFound(w, req)\\n\\t\\t\\t\\treturn\\n\\n\\t\\t\\t}))\\n\\t\\tfmt.Printf(\\\"\\\\n\\\\x1b[32m %#v [GET]%v \\\\x1b[49m\\\\x1b[39m \\\", pattern, http.Dir(path))\\n\\t}\\n\\n\\t//register all controller routes\\n\\tfor _, r := range hr.routes {\\n\\t\\tfmt.Printf(\\\"\\\\n\\\\x1b[32m %#v :name \\\\x1b[49m\\\\x1b[39m \\\", r.pattern)\\n\\t\\tfor method, fn := range r.controller.MappingBy(r.pattern) {\\n\\t\\t\\tif r.mws != nil {\\n\\t\\t\\t\\tro.HandlerFunc(strings.ToUpper(method), r.pattern, r.mws.Handle(r.controller.ServeHTTP)) //TODO ????? error no url pattern\\n\\t\\t\\t} else {\\n\\t\\t\\t\\tro.HandlerFunc(strings.ToUpper(method), r.pattern, r.controller.ServeHTTP)\\n\\t\\t\\t}\\n\\t\\t\\tfmt.Printf(\\\"\\\\x1b[32m [%v]%v name \\\\x1b[49m\\\\x1b[39m \\\", method, fn)\\n\\t\\t}\\n\\t}\\n\\n\\t//Not Found Handler\\n\\tif hr.notFound != nil {\\n\\t\\tro.NotFound = hr.notFound\\n\\t}\\n\\n\\treturn ro\\n}\",\n \"func rootHandler(w http.ResponseWriter, r *http.Request, title string) {\\n\\tp, err := loadRoot(title)\\n\\n\\tp.Body = template.HTML(blackfriday.MarkdownCommon([]byte(p.Body)))\\n\\tp.Body = template.HTML(convertWikiMarkup([]byte(p.Body)))\\n\\n\\terr = templates.ExecuteTemplate(w, \\\"root.html\\\", p)\\n\\tif err != nil {\\n\\t\\thttp.Error(w, err.Error(), http.StatusInternalServerError)\\n\\t}\\n}\",\n \"func (mux *Mux) Handler(r *Request) Handler {\\n\\tscheme := r.URL.Scheme\\n\\thost := r.URL.Hostname()\\n\\tpath := cleanPath(r.URL.Path)\\n\\n\\t// If the given path is /tree and its handler is not registered,\\n\\t// redirect for /tree/.\\n\\tif u, ok := mux.redirectToPathSlash(muxKey{scheme, host, path}, r.URL); ok {\\n\\t\\treturn StatusHandler(StatusPermanentRedirect, u.String())\\n\\t}\\n\\n\\tif path != r.URL.Path {\\n\\t\\tu := *r.URL\\n\\t\\tu.Path = path\\n\\t\\treturn StatusHandler(StatusPermanentRedirect, u.String())\\n\\t}\\n\\n\\tmux.mu.RLock()\\n\\tdefer mux.mu.RUnlock()\\n\\n\\th := mux.match(muxKey{scheme, host, path})\\n\\tif h == nil {\\n\\t\\t// Try wildcard\\n\\t\\tif wildcard, ok := getWildcard(host); ok {\\n\\t\\t\\th = mux.match(muxKey{scheme, wildcard, path})\\n\\t\\t}\\n\\t}\\n\\tif h == nil {\\n\\t\\th = NotFoundHandler()\\n\\t}\\n\\treturn h\\n}\",\n \"func MountCORSHandler(mux goahttp.Muxer, h http.Handler) {\\n\\th = handleStationOrigin(h)\\n\\tf, ok := h.(http.HandlerFunc)\\n\\tif !ok {\\n\\t\\tf = func(w http.ResponseWriter, r *http.Request) {\\n\\t\\t\\th.ServeHTTP(w, r)\\n\\t\\t}\\n\\t}\\n\\tmux.Handle(\\\"OPTIONS\\\", \\\"/stations\\\", f)\\n\\tmux.Handle(\\\"OPTIONS\\\", \\\"/stations/{id}\\\", f)\\n\\tmux.Handle(\\\"OPTIONS\\\", \\\"/stations/@/{id}\\\", f)\\n\\tmux.Handle(\\\"OPTIONS\\\", \\\"/projects/{id}/stations\\\", f)\\n\\tmux.Handle(\\\"OPTIONS\\\", \\\"/stations/{id}/photo\\\", f)\\n}\",\n \"func (e *Engine) setupFallback() {\\n\\te.srv.HTTPErrorHandler = func(err error, c echo.Context) {\\n\\t\\tif err != echo.ErrNotFound {\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\turi := genericPath(c.Request().RequestURI)\\n\\t\\tfs, err := os.Stat(uri)\\n\\t\\tif err != nil {\\n\\t\\t\\tc.Error(err)\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\tif fs.IsDir() {\\n\\t\\t\\t// if working at pwd, capture all dirs\\n\\t\\t\\tif len(e.dirs) == 1 && e.dirs[0] == \\\".\\\" {\\n\\t\\t\\t\\t// check if file excluded\\n\\t\\t\\t\\tif isExclude(uri) {\\n\\t\\t\\t\\t\\treturn\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\tif err := e.doDispatch(c, fs); err != nil {\\n\\t\\t\\t\\t\\tc.Error(err)\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\t\\t\\t// only capture missing dir when serving pwd\\n\\t\\t\\t// return for any other dirs\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\t// capture any files here\\n\\t\\terr = e.doServeFile(uri)(c)\\n\\t\\tif err != nil {\\n\\t\\t\\tc.Error(err)\\n\\t\\t}\\n\\t}\\n}\",\n \"func NewHandler(s *Server) *Handler {\\n\\th := &Handler{\\n\\t\\tserver: s,\\n\\t\\tmux: pat.New(),\\n\\t}\\n\\n\\t// Series routes.\\n\\th.mux.Get(\\\"/db/:db/series\\\", http.HandlerFunc(h.serveQuery))\\n\\th.mux.Post(\\\"/db/:db/series\\\", http.HandlerFunc(h.serveWriteSeries))\\n\\th.mux.Del(\\\"/db/:db/series/:series\\\", http.HandlerFunc(h.serveDeleteSeries))\\n\\th.mux.Get(\\\"/db\\\", http.HandlerFunc(h.serveDatabases))\\n\\th.mux.Post(\\\"/db\\\", http.HandlerFunc(h.serveCreateDatabase))\\n\\th.mux.Del(\\\"/db/:name\\\", http.HandlerFunc(h.serveDeleteDatabase))\\n\\n\\t// Cluster admins routes.\\n\\th.mux.Get(\\\"/cluster_admins/authenticate\\\", http.HandlerFunc(h.serveAuthenticateClusterAdmin))\\n\\th.mux.Get(\\\"/cluster_admins\\\", http.HandlerFunc(h.serveClusterAdmins))\\n\\th.mux.Post(\\\"/cluster_admins\\\", http.HandlerFunc(h.serveCreateClusterAdmin))\\n\\th.mux.Post(\\\"/cluster_admins/:user\\\", http.HandlerFunc(h.serveUpdateClusterAdmin))\\n\\th.mux.Del(\\\"/cluster_admins/:user\\\", http.HandlerFunc(h.serveDeleteClusterAdmin))\\n\\n\\t// Database users routes.\\n\\th.mux.Get(\\\"/db/:db/authenticate\\\", http.HandlerFunc(h.serveAuthenticateDBUser))\\n\\th.mux.Get(\\\"/db/:db/users\\\", http.HandlerFunc(h.serveDBUsers))\\n\\th.mux.Post(\\\"/db/:db/users\\\", http.HandlerFunc(h.serveCreateDBUser))\\n\\th.mux.Get(\\\"/db/:db/users/:user\\\", http.HandlerFunc(h.serveDBUser))\\n\\th.mux.Post(\\\"/db/:db/users/:user\\\", http.HandlerFunc(h.serveUpdateDBUser))\\n\\th.mux.Del(\\\"/db/:db/users/:user\\\", http.HandlerFunc(h.serveDeleteDBUser))\\n\\n\\t// Utilities\\n\\th.mux.Get(\\\"/ping\\\", http.HandlerFunc(h.servePing))\\n\\th.mux.Get(\\\"/interfaces\\\", http.HandlerFunc(h.serveInterfaces))\\n\\n\\t// Shard routes.\\n\\th.mux.Get(\\\"/cluster/shards\\\", http.HandlerFunc(h.serveShards))\\n\\th.mux.Post(\\\"/cluster/shards\\\", http.HandlerFunc(h.serveCreateShard))\\n\\th.mux.Del(\\\"/cluster/shards/:id\\\", http.HandlerFunc(h.serveDeleteShard))\\n\\n\\t// Shard space routes.\\n\\th.mux.Get(\\\"/cluster/shard_spaces\\\", http.HandlerFunc(h.serveShardSpaces))\\n\\th.mux.Post(\\\"/cluster/shard_spaces/:db\\\", http.HandlerFunc(h.serveCreateShardSpace))\\n\\th.mux.Post(\\\"/cluster/shard_spaces/:db/:name\\\", http.HandlerFunc(h.serveUpdateShardSpace))\\n\\th.mux.Del(\\\"/cluster/shard_spaces/:db/:name\\\", http.HandlerFunc(h.serveDeleteShardSpace))\\n\\n\\t// Cluster config endpoints\\n\\th.mux.Get(\\\"/cluster/servers\\\", http.HandlerFunc(h.serveServers))\\n\\th.mux.Del(\\\"/cluster/servers/:id\\\", http.HandlerFunc(h.serveDeleteServer))\\n\\n\\treturn h\\n}\",\n \"func HttpRootHandler(w http.ResponseWriter, r *http.Request) {\\n\\tcmd := CommandStruct{}\\n\\n\\tbs, err := ioutil.ReadAll(r.Body)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\t_ = json.Unmarshal(bs, &cmd)\\n\\tswitch cmd.Command {\\n\\tcase \\\"check\\\":\\n\\t\\tfmt.Fprintln(w, \\\"HTTP-Server is online.\\\")\\n\\t\\tbreak\\n\\n\\tcase \\\"disc\\\":\\n\\t\\tfmt.Fprintln(w, \\\"HTTP-Server shutting down...\\\")\\n\\t\\thttpAlive <- 1\\n\\t\\tbreak\\n\\n\\tcase \\\"lista\\\":\\n\\t\\tfmt.Fprintln(w, \\\"List of all available NBD-devices:\\\")\\n\\t\\tfor key, value := range AvailableList {\\n\\t\\t\\tif AvailableList[key] != \\\"\\\" {\\n\\t\\t\\t\\tfmt.Fprintln(w, value)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tbreak\\n\\n\\tcase \\\"listm\\\":\\n\\t\\tfmt.Fprintln(w, \\\"List of all mounted NBD-devices:\\\")\\n\\t\\tfor key, value := range MountedList {\\n\\t\\t\\tfmt.Fprintln(w, key+\\\"\\\\t\\\"+value)\\n\\t\\t}\\n\\t\\tbreak\\n\\n\\tcase \\\"mount\\\":\\n\\t\\tif strings.Contains(cmd.Device, \\\"/dev/nbd\\\") {\\n\\t\\t\\tfor i := 0; i < len(AvailableList); i++ {\\n\\t\\t\\t\\tif AvailableList[i] == cmd.Device {\\n\\n\\t\\t\\t\\t\\tLinkedLogins[len(LinkedLogins)+1], err = nethandler.SetupConnection(cmd.Image, cmd.User, cmd.Pass, cmd.Device)\\n\\t\\t\\t\\t\\tif err != nil {\\n\\t\\t\\t\\t\\t\\tfmt.Fprintf(w, \\\"Error: \\\", err)\\n\\t\\t\\t\\t\\t\\tfmt.Fprintf(w, \\\"\\\\n\\\")\\n\\t\\t\\t\\t\\t\\treturn\\n\\t\\t\\t\\t\\t}\\n\\n\\t\\t\\t\\t\\tAddToMountedList(cmd.Device, cmd.Image)\\n\\t\\t\\t\\t\\tfmt.Fprintf(w, \\\"Successfully mounted \\\"+cmd.Image+\\\" to \\\"+cmd.Device+\\\"\\\\n\\\")\\n\\t\\t\\t\\t\\treturn\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\t\\t\\tfor _, value := range AvailableList {\\n\\t\\t\\t\\tif value != \\\"\\\" {\\n\\t\\t\\t\\t\\tAddToMountedList(value, cmd.Image)\\n\\t\\t\\t\\t\\tfmt.Fprintf(w, \\\"Device \\\"+cmd.Device+\\\" is already mounted.\\\\n\\\"+cmd.Image+\\\" has been mounted to \\\"+value+\\\" instead.\\\\n\\\")\\n\\t\\t\\t\\t\\treturn\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\t\\t\\tfmt.Fprintf(w, \\\"No more devices available!\\\\n\\\")\\n\\t\\t} else {\\n\\t\\t\\tfmt.Fprintf(w, \\\"Specified device not recognised.\\\\n\\\")\\n\\t\\t}\\n\\t\\tbreak\\n\\n\\tcase \\\"unmount\\\":\\n\\t\\t//TODO Real unmounting of NBD-devices\\n\\t\\tfor key, _ := range AvailableList {\\n\\t\\t\\tif AvailableList[key] == \\\"\\\" {\\n\\t\\t\\t\\tdelete(MountedList, cmd.Device)\\n\\t\\t\\t\\tAvailableList[key] = cmd.Device\\n\\t\\t\\t\\tfmt.Fprint(w, \\\"Successfully unmounted \\\"+cmd.Device)\\n\\t\\t\\t\\tbreak\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tbreak\\n\\t}\\n}\",\n \"func HTTPHandler(router *mux.Router) {\\n\\trouter.Handle(\\\"/\\\", ImageHandler())\\n\\trouter.PathPrefix(\\\"/unsafe/\\\").Handler(ImageHandler())\\n\\trouter.Handle(\\\"/healthcheck\\\", HealthcheckHandler())\\n}\",\n \"func Mount(router *mux.Router) {\\n\\n}\",\n \"func MountPhotoHandler(mux goahttp.Muxer, h http.Handler) {\\n\\tf, ok := handleStationOrigin(h).(http.HandlerFunc)\\n\\tif !ok {\\n\\t\\tf = func(w http.ResponseWriter, r *http.Request) {\\n\\t\\t\\th.ServeHTTP(w, r)\\n\\t\\t}\\n\\t}\\n\\tmux.Handle(\\\"GET\\\", \\\"/stations/{id}/photo\\\", f)\\n}\",\n \"func indexHandler(w http.ResponseWriter, r *http.Request) {\\n\\tif r.URL.Path != \\\"/\\\" {\\n\\t\\thttp.NotFound(w, r)\\n\\t\\treturn\\n\\t}\\n\\n\\thttp.ServeFile(w, r, \\\"pages/main.html\\\")\\n}\",\n \"func pathHandler(w http.ResponseWriter, r *http.Request) {\\n\\tpath := \\\"/var/www/ear7h-net/\\\" + r.URL.Path[1:]\\n\\t//open file and send\\n\\tf, err := os.Open(path)\\n\\tif err != nil {\\n\\t\\tfourOhFour(w, r)\\n\\t} else {\\n\\t\\thttp.ServeContent(w, r, r.URL.Path, time.Now(), f)\\n\\t}\\n}\",\n \"func (r *Router) masterHandler(c *gin.Context) {\\n\\tklog.V(4).Infof(\\\"no router for method:%s, url:%s\\\", c.Request.Method, c.Request.URL.Path)\\n\\tc.JSON(404, gin.H{\\n\\t\\t\\\"Method\\\": c.Request.Method,\\n\\t\\t\\\"Path\\\": c.Request.URL.Path,\\n\\t\\t\\\"error\\\": \\\"router not found\\\"})\\n}\",\n \"func Serve(root, iface string, port int) (err error) {\\n\\t// Root the path, and clean it if necessary.\\n\\n\\t// 18/01/2013 It might make sense to move this to a helper routine\\n\\t// or further up in the stack.\\n\\tif !path.IsAbs(root) {\\n\\t\\tvar wd string\\n\\t\\twd, err = os.Getwd()\\n\\t\\tif err != nil {\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\troot = path.Join(wd, root)\\n\\t} else {\\n\\t\\troot = path.Clean(root)\\n\\t}\\n\\tl.Printf(\\\"Starting http server %s:%d\\\\nRoot path %q\\\", iface, port, root)\\n\\trootpath = root\\n\\n\\thttp.HandleFunc(\\\"/p/\\\", handle)\\n\\thttp.ListenAndServe(iface+\\\":\\\"+strconv.Itoa(port), nil)\\n\\treturn\\n}\",\n \"func (self *CentralBooking) InstallHandlers(router *mux.Router) {\\n router.\\n Methods(\\\"POST\\\").\\n Path(\\\"/register/instance\\\").\\n HandlerFunc(self.RegisterInstance)\\n\\n // apeing vault\\n router.\\n Methods(\\\"GET\\\").\\n Path(\\\"/sys/health\\\").\\n HandlerFunc(self.CheckHealth)\\n}\",\n \"func (r *SilencedRouter) Mount(parent *mux.Router) {\\n\\troutes := ResourceRoute{\\n\\t\\tRouter: parent,\\n\\t\\tPathPrefix: \\\"/namespaces/{namespace}/{resource:silenced}\\\",\\n\\t}\\n\\n\\thandlers := handlers.NewHandlers[*corev2.Silenced](r.store)\\n\\n\\troutes.Del(handlers.DeleteResource)\\n\\troutes.Get(r.get)\\n\\troutes.Post(r.create)\\n\\troutes.Put(r.createOrReplace)\\n\\troutes.List(r.listr, corev3.SilencedFields)\\n\\troutes.ListAllNamespaces(r.listr, \\\"/{resource:silenced}\\\", corev3.SilencedFields)\\n\\n\\t// Custom routes for listing by subscription and checks for a specific\\n\\t// namespace, in addition to all namespaces for checks.\\n\\troutes.Router.HandleFunc(\\\"/{resource:silenced}/checks/{check}\\\", listHandler(r.list)).Methods(http.MethodGet)\\n\\troutes.Router.HandleFunc(routes.PathPrefix+\\\"/subscriptions/{subscription}\\\", listHandler(r.list)).Methods(http.MethodGet)\\n\\troutes.Router.HandleFunc(routes.PathPrefix+\\\"/checks/{check}\\\", listHandler(r.list)).Methods(http.MethodGet)\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.702594","0.6997063","0.6954332","0.6820072","0.65756345","0.6451301","0.6383405","0.6365718","0.6285303","0.62754136","0.6251029","0.6157279","0.6116233","0.60878783","0.6076186","0.60656357","0.6065411","0.6024053","0.6022069","0.60135174","0.60126704","0.6005589","0.6004503","0.60038495","0.5991979","0.59673274","0.59673274","0.59673274","0.5966083","0.5946025","0.5934402","0.593274","0.59320784","0.5915197","0.5913959","0.5913508","0.5905052","0.5902104","0.5896477","0.5896477","0.58728826","0.5862679","0.5854989","0.58452594","0.5838185","0.58378357","0.58346957","0.582368","0.5816042","0.5792327","0.57894605","0.5784616","0.57818025","0.57736343","0.5773397","0.57706916","0.57699347","0.5760902","0.5760156","0.57574904","0.5721992","0.57217157","0.57118344","0.57063615","0.56997764","0.56936055","0.5681272","0.56789786","0.5678943","0.5676648","0.56746376","0.5670061","0.5669512","0.5663938","0.56606144","0.5650369","0.5646647","0.564429","0.5644125","0.56419843","0.5641957","0.5628667","0.56214565","0.5611402","0.5610784","0.56051207","0.56039834","0.55896366","0.5584176","0.5578367","0.5568307","0.556564","0.5561139","0.5555659","0.55513","0.5548231","0.5545511","0.55443543","0.5543333","0.5541932","0.5530312"],"string":"[\n \"0.702594\",\n \"0.6997063\",\n \"0.6954332\",\n \"0.6820072\",\n \"0.65756345\",\n \"0.6451301\",\n \"0.6383405\",\n \"0.6365718\",\n \"0.6285303\",\n \"0.62754136\",\n \"0.6251029\",\n \"0.6157279\",\n \"0.6116233\",\n \"0.60878783\",\n \"0.6076186\",\n \"0.60656357\",\n \"0.6065411\",\n \"0.6024053\",\n \"0.6022069\",\n \"0.60135174\",\n \"0.60126704\",\n \"0.6005589\",\n \"0.6004503\",\n \"0.60038495\",\n \"0.5991979\",\n \"0.59673274\",\n \"0.59673274\",\n \"0.59673274\",\n \"0.5966083\",\n \"0.5946025\",\n \"0.5934402\",\n \"0.593274\",\n \"0.59320784\",\n \"0.5915197\",\n \"0.5913959\",\n \"0.5913508\",\n \"0.5905052\",\n \"0.5902104\",\n \"0.5896477\",\n \"0.5896477\",\n \"0.58728826\",\n \"0.5862679\",\n \"0.5854989\",\n \"0.58452594\",\n \"0.5838185\",\n \"0.58378357\",\n \"0.58346957\",\n \"0.582368\",\n \"0.5816042\",\n \"0.5792327\",\n \"0.57894605\",\n \"0.5784616\",\n \"0.57818025\",\n \"0.57736343\",\n \"0.5773397\",\n \"0.57706916\",\n \"0.57699347\",\n \"0.5760902\",\n \"0.5760156\",\n \"0.57574904\",\n \"0.5721992\",\n \"0.57217157\",\n \"0.57118344\",\n \"0.57063615\",\n \"0.56997764\",\n \"0.56936055\",\n \"0.5681272\",\n \"0.56789786\",\n \"0.5678943\",\n \"0.5676648\",\n \"0.56746376\",\n \"0.5670061\",\n \"0.5669512\",\n \"0.5663938\",\n \"0.56606144\",\n \"0.5650369\",\n \"0.5646647\",\n \"0.564429\",\n \"0.5644125\",\n \"0.56419843\",\n \"0.5641957\",\n \"0.5628667\",\n \"0.56214565\",\n \"0.5611402\",\n \"0.5610784\",\n \"0.56051207\",\n \"0.56039834\",\n \"0.55896366\",\n \"0.5584176\",\n \"0.5578367\",\n \"0.5568307\",\n \"0.556564\",\n \"0.5561139\",\n \"0.5555659\",\n \"0.55513\",\n \"0.5548231\",\n \"0.5545511\",\n \"0.55443543\",\n \"0.5543333\",\n \"0.5541932\",\n \"0.5530312\"\n]"},"document_score":{"kind":"string","value":"0.0"},"document_rank":{"kind":"string","value":"-1"}}},{"rowIdx":104788,"cells":{"query":{"kind":"string","value":"SiaPath returns the siapath of a remote directory."},"document":{"kind":"string","value":"func (rd *RemoteDir) SiaPath() modules.SiaPath {\n\treturn rd.siapath\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (sd *SiaDir) SiaPath() modules.SiaPath {\n\tsd.mu.Lock()\n\tdefer sd.mu.Unlock()\n\treturn sd.siaPath\n}","func (rf *RemoteFile) SiaPath() modules.SiaPath {\n\trf.mu.Lock()\n\tdefer rf.mu.Unlock()\n\treturn rf.siaPath\n}","func (o *Object) remotePath() string {\n\treturn o.fs.slashRootSlash + o.remote\n}","func newSiaPath(path string) modules.SiaPath {\n\tsp, err := modules.NewSiaPath(path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn sp\n}","func (b *Binary) RemotePath() string {\n\treturn b.file.RemotePath()\n}","func (r *Repository) Path(id digest.Digest) (dir, path string) {\n\tdir = filepath.Join(r.Root, id.Hex()[:2])\n\treturn dir, filepath.Join(dir, id.Hex()[2:])\n}","func (proxy *remoteDriverProxy) Path(name string) (string, error) {\n\tvar req = remoteVolumePathReq{\n\t\tName: name,\n\t}\n\n\tvar resp remoteVolumePathResp\n\n\tif err := proxy.client.CallService(remoteVolumePathService, &req, &resp, true); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif resp.Err != \"\" {\n\t\treturn \"\", errors.New(resp.Err)\n\t}\n\n\treturn resp.Mountpoint, nil\n}","func LoadSiaDir(rootDir string, siaPath modules.SiaPath, deps modules.Dependencies, wal *writeaheadlog.WAL) (sd *SiaDir, err error) {\n\tsd = &SiaDir{\n\t\tdeps: deps,\n\t\tsiaPath: siaPath,\n\t\trootDir: rootDir,\n\t\twal: wal,\n\t}\n\tsd.metadata, err = loadSiaDirMetadata(siaPath.SiaDirMetadataSysPath(rootDir), modules.ProdDependencies)\n\treturn sd, err\n}","func (r *RepoRef) SkeletonsPath() string {\n\treturn filepath.Join(r.LocalPath(), SkeletonsDir)\n}","func (*HttpLocalSites) GetPath() string { return \"/api/objects/http/local_site/\" }","func (server *SingleInstance) Path() string {\n\treturn server.path\n}","func (d Driver) getSecureDiffPath(id, parent string, canBeRemote bool) string {\n\tvar diffDirName string\n\n\tif parent == \"\" || d.isParent(id, parent) {\n\t\tdiffDirName = \"diff\"\n\t} else {\n\t\tdiffDirName = fmt.Sprintf(\"%s-%s\", \"diff\", parent)\n\t}\n\n\tlocalSecureDiffPath := path.Join(d.dir(id), constSecureBaseDirName, diffDirName)\n\tremoteSecureDiffPath := path.Join(d.options.remoteDir, id, constSecureBaseDirName, diffDirName)\n\tlogrus.Debugf(\"secureoverlay2: getSecureDiffPath %s. localSecureDiffPath %s remoteSecureDiffPath\", localSecureDiffPath, remoteSecureDiffPath)\n\tdiffPath := localSecureDiffPath\n\t// remote only \"wins\" if local does not exist and remote exists\n\tif canBeRemote && d.options.remoteDir != \"\" {\n\t\tif b, _ := exists(localSecureDiffPath); !b {\n\t\t\tif b, _ := exists(remoteSecureDiffPath); b {\n\t\t\t\tdiffPath = remoteSecureDiffPath\n\t\t\t}\n\t\t}\n\t}\n\tlogrus.Debugf(\"secureoverlay2: getSecureDiffPath w. id: %s, parent: %s, canBeRemote: %v returns %s\",\n\t\tid, parent, canBeRemote, diffPath)\n\n\treturn diffPath\n}","func (d *driver) fullPath(path string) string {\n\treturn _path.Join(\"/ipfs\", d.roothash, path)\n}","func (o ApplicationStatusOperationStateSyncResultSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (d *Datastore) s3Path(p string) string {\n\treturn path.Join(d.RootDirectory, p)\n}","func (p *Part) RemotePath(prefix string) string {\n\tfor strings.HasSuffix(prefix, \"/\") {\n\t\tprefix = prefix[:len(prefix)-1]\n\t}\n\treturn fmt.Sprintf(\"%s/%s/%016X_%016X_%016X\", prefix, p.Path, p.FileSize, p.Offset, p.Size)\n}","func (o ApplicationStatusOperationStateOperationSyncSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (fi *fileInfo) Path() string {\n\treturn fi.fullPath\n}","func (ss *Sources) workspacePath(id consumerID) string {\n\treturn filepath.Join(ss.RootPath, \"workspace\", id.Namespace, id.Name, id.consumer)\n}","func (s *Store) Path() string { return s.path }","func (s *Store) Path() string { return s.path }","func BattleShiritoriPath(battleID string, userID string) string {\n\treturn fmt.Sprintf(\"/streams/battles/%v/%v\", battleID, userID)\n}","func (r *Resolver) Path(resource interface{}, id string) (path string, err error) {\n\tprovider := r.Provider\n\tswitch resource.(type) {\n\tcase *Provider:\n\t\tr := Provider{}\n\t\tr.UID = id\n\t\tr.Link()\n\t\tpath = r.SelfLink\n\tcase *Folder:\n\t\tr := Folder{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Datacenter:\n\t\tr := Datacenter{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Cluster:\n\t\tr := Cluster{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Host:\n\t\tr := Host{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Network:\n\t\tr := Network{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Datastore:\n\t\tr := Datastore{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *VM:\n\t\tr := VM{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Workload:\n\t\tr := Workload{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tdefault:\n\t\terr = liberr.Wrap(\n\t\t\tbase.ResourceNotResolvedError{\n\t\t\t\tObject: resource,\n\t\t\t})\n\t}\n\n\tpath = strings.TrimRight(path, \"/\")\n\n\treturn\n}","func (o ApplicationStatusOperationStateSyncResultSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusOperationStateSyncResultSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}","func (sds *SiaDirSet) NewSiaDir(siaPath string) (*SiaDirSetEntry, error) {\n\tsds.mu.Lock()\n\tdefer sds.mu.Unlock()\n\t// Check is SiaDir already exists\n\tsiaPath = strings.Trim(siaPath, \"/\")\n\texists, err := sds.exists(siaPath)\n\tif exists {\n\t\treturn nil, ErrPathOverload\n\t}\n\tif !os.IsNotExist(err) && err != nil {\n\t\treturn nil, err\n\t}\n\tsd, err := New(siaPath, sds.rootDir, sds.wal)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tentry := sds.newSiaDirSetEntry(sd)\n\tthreadUID := randomThreadUID()\n\tentry.threadMap[threadUID] = newThreadType()\n\tsds.siaDirMap[siaPath] = entry\n\treturn &SiaDirSetEntry{\n\t\tsiaDirSetEntry: entry,\n\t\tthreadUID: threadUID,\n\t}, nil\n}","func suriPath() (exists bool) {\n var err error\n path, err := utils.GetKeyValueString(\"suriPath\", \"path\")\n if err != nil {\n logs.Error(\"suriPath Error getting data from main.conf\")\n }\n\n if _, err := os.Stat(path); os.IsNotExist(err) {\n logs.Error(\"Suricata not installed, at least folder /etc/suricata dosn't exist\")\n return false\n }\n return true\n}","func (o TransferJobTransferSpecAwsS3DataSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecAwsS3DataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (driver *Driver) Path(volumeName, volumeID string) (string, error) {\n\tif volumeName == \"\" && volumeID == \"\" {\n\t\treturn \"\", errors.New(\"Missing volume name or ID\")\n\t}\n\n\tinstances, err := driver.sdm.GetInstance()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tswitch {\n\tcase len(instances) == 0:\n\t\treturn \"\", errors.New(\"No instances\")\n\tcase len(instances) > 1:\n\t\treturn \"\", errors.New(\"Too many instances returned, limit the storagedrivers\")\n\t}\n\n\tvolumes, err := driver.sdm.GetVolume(volumeID, volumeName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tswitch {\n\tcase len(volumes) == 0:\n\t\treturn \"\", errors.New(\"No volumes returned by name\")\n\tcase len(volumes) > 1:\n\t\treturn \"\", errors.New(\"Multiple volumes returned by name\")\n\t}\n\n\tvolumeAttachment, err := driver.sdm.GetVolumeAttach(volumes[0].VolumeID, instances[0].InstanceID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif len(volumeAttachment) == 0 {\n\t\treturn \"\", nil\n\t}\n\n\tmounts, err := driver.osdm.GetMounts(volumeAttachment[0].DeviceName, \"\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif len(mounts) == 0 {\n\t\treturn \"\", nil\n\t}\n\n\treturn mounts[0].Mountpoint, nil\n}","func (o IopingSpecVolumeVolumeSourceProjectedSourcesSecretItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesSecretItems) string { return v.Path }).(pulumi.StringOutput)\n}","func (i *Instance) Path() (string, error) {\n\tref, _, _, err := i.GetAsAny(WmiPathKey)\n\treturn ref.(string), err\n}","func (s *Store) IDPath() string { return filepath.Join(s.path, \"id\") }","func (d ImagefsDriver) Path(r *volume.PathRequest) (*volume.PathResponse, error) {\n\tgetReq := volume.GetRequest{\n\t\tName: r.Name,\n\t}\n\tret, err := d.Get(&getReq)\n\tvar _ret *volume.PathResponse\n\tif ret != nil {\n\t\t_ret = &volume.PathResponse{\n\t\t\tMountpoint: ret.Volume.Mountpoint,\n\t\t}\n\t}\n\treturn _ret, err\n}","func getRepoPath() (string, error) {\n\t// Set default base path and directory name\n\tdirectoryName := \".saturn\"\n\n\t// Join the path and directory name, then expand the home path\n\tfullPath, err := homedir.Expand(filepath.Join(\"~\", directoryName))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Return the shortest lexical representation of the path\n\treturn filepath.Clean(fullPath), nil\n}","func PantahubS3Path() string {\n\tif GetEnv(EnvPantahubStorageDriver) == \"s3\" {\n\t\treturn GetEnv(EnvPantahubStoragePath)\n\t}\n\n\tbasePath := path.Join(GetEnv(EnvPantahubS3Path), GetEnv(EnvPantahubStoragePath))\n\n\tif basePath == \"\" {\n\t\tbasePath = \".\"\n\t}\n\n\treturn basePath\n}","func (c *Credentials) path() (string, error) {\n\tusr, err := user.Current()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to get current user\")\n\t}\n\treturn path.Join(usr.HomeDir, *cred), nil\n}","func (ls *LocationStore) Path() string {\n\treturn fmt.Sprintf(\"env/locations/%s.yml\", ls.name)\n}","func TestNewSiaDir(t *testing.T) {\n\tif testing.Short() && !build.VLONG {\n\t\tt.SkipNow()\n\t}\n\tt.Parallel()\n\t// Create filesystem.\n\troot := filepath.Join(testDir(t.Name()), \"fs-root\")\n\tfs := newTestFileSystem(root)\n\t// Create dir /sub/foo\n\tsp := newSiaPath(\"sub/foo\")\n\tif err := fs.NewSiaDir(sp, modules.DefaultDirPerm); err != nil {\n\t\tt.Fatal(err)\n\t}\n\t// The whole path should exist.\n\tif _, err := os.Stat(filepath.Join(root, sp.String())); err != nil {\n\t\tt.Fatal(err)\n\t}\n}","func (s *session) Path() dbus.ObjectPath {\n\treturn s.path\n}","func (o ApplicationStatusOperationStateOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusOperationStateOperationSyncSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}","func Path(named Named) (name string) {\n\tif r, ok := named.(namedRepository); ok {\n\t\treturn r.Path()\n\t}\n\t_, path := splitDomain(named.Name())\n\treturn path\n}","func (c *Client) Path() (string, error) {\n\treturn c.GetProperty(\"path\")\n}","func (b *Bucket) Path() (string, error) {\n\tconf := b.conf.Viper.ConfigFileUsed()\n\tif conf == \"\" {\n\t\treturn b.cwd, nil\n\t}\n\treturn filepath.Dir(filepath.Dir(conf)), nil\n}","func (r *RemoteSyslogServer) GetPath() string {\n\treturn fmt.Sprintf(\"/api/objects/remote_syslog/server/%s\", r.Reference)\n}","func (h *HttpLocalSite) GetPath() string {\n\treturn fmt.Sprintf(\"/api/objects/http/local_site/%s\", h.Reference)\n}","func (r *ServiceLinkedRole) Path() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"path\"])\n}","func (k *Key) path() (string, error) {\n\tusr, err := user.Current()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to get current user\")\n\t}\n\treturn path.Join(usr.HomeDir, *identity), nil\n}","func (d *DirectHash) getPath() string {\n\tdomain := d.RootDomain\n\tif len(d.ReleaseBranch) > 0 {\n\t\tdomain = string(d.ReleaseBranch) + \".\" + d.RootDomain\n\t}\n\treturn fmt.Sprintf(\"https://%s/\"+strings.Trim(d.Path, \"/\"), domain, d.ReleaseBranch.String())\n}","func (id ID) Path() string {\n\treturn id.path\n}","func (p *SeriesPartition) Path() string { return p.path }","func (o *StorageNetAppCifsShareAllOf) GetPath() string {\n\tif o == nil || o.Path == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Path\n}","func (o FioSpecVolumeVolumeSourceProjectedSourcesSecretItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesSecretItems) string { return v.Path }).(pulumi.StringOutput)\n}","func (m *CgroupfsManager) SandboxCgroupPath(sbParent, sbID string) (cgParent, cgPath string, _ error) {\n\tif strings.HasSuffix(path.Base(sbParent), \".slice\") {\n\t\treturn \"\", \"\", fmt.Errorf(\"cri-o configured with cgroupfs cgroup manager, but received systemd slice as parent: %s\", sbParent)\n\t}\n\n\tif err := verifyCgroupHasEnoughMemory(sbParent, m.memoryPath, m.memoryMaxFile); err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\treturn sbParent, filepath.Join(sbParent, containerCgroupPath(sbID)), nil\n}","func (o ApplicationOperationSyncSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationOperationSyncSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (o IopingSpecVolumeVolumeSourceNfsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceNfs) string { return v.Path }).(pulumi.StringOutput)\n}","func (o TransferJobTransferSpecAwsS3DataSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TransferJobTransferSpecAwsS3DataSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}","func (d Driver) getSecureCryptMntPath(id string) string {\n\treturn path.Join(d.dir(id), constSecureBaseDirName, constSecureCryptMntDirName)\n}","func (n EpisodeDetails) GetSeriesPath(destination string) string {\n\treturn filepath.Join(destination, FileNameCleaner(n.Showtitle))\n}","func (*CaRsas) GetPath() string { return \"/api/objects/ca/rsa/\" }","func (i *Image) Path() string {\n\treturn i.p\n}","func (s *Store) Path() string {\n\treturn s.path\n}","func (s *Store) Path() string {\n\treturn s.path\n}","func (cs *ControllerServer) getVolumeSharePath(vol *nfsVolume) string {\n\treturn filepath.Join(string(filepath.Separator), vol.baseDir, vol.subDir)\n}","func (d *MinioDriver) Path(r volume.Request) volume.Response {\n\td.m.RLock()\n\tdefer d.m.RUnlock()\n\n\tv, exists := d.volumes[r.Name]\n\tif !exists {\n\t\treturn volumeResp(\"\", \"\", nil, capability, newErrVolNotFound(r.Name).Error())\n\t}\n\treturn volumeResp(v.mountpoint, r.Name, nil, capability, \"\")\n}","func Path(subpath string) (string, error) {\n\tif !isinit {\n\t\treturn \"\", fmt.Errorf(\"pathresolver not initialized\")\n\t}\n\n\tif subpath == \"\" {\n\t\treturn basedirectory, nil\n\t}\n\n\tif filepath.IsAbs(subpath) {\n\t\treturn \"\", fmt.Errorf(\"cannot use absolute path as subpath\")\n\t}\n\n\treturn filepath.Join(basedirectory, subpath), nil\n}","func (o ApplicationStatusSyncComparedToSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusSyncComparedToSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (*RemoteSyslogServers) GetPath() string { return \"/api/objects/remote_syslog/server/\" }","func (o IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItems) string { return v.Path }).(pulumi.StringOutput)\n}","func (o ArgoCDSpecServerIngressOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecServerIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (o IopingSpecVolumeVolumeSourceGlusterfsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceGlusterfs) string { return v.Path }).(pulumi.StringOutput)\n}","func (r *SysNet) getPath() (string, error) {\n\tvar procPath string\n\n\tswitch r.Path {\n\tcase sysNetPathCore:\n\t\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathCore), r.Property)\n\t\tbreak\n\tcase sysNetPathIPv4:\n\n\t\tif r.Link != \"\" {\n\t\t\tprocPath = path.Join(path.Join(path.Join(path.Join(sysNetPath, sysNetPathIPv4), \"conf\"), r.Link), r.Property)\n\t\t} else {\n\t\t\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathIPv4), r.Property)\n\t\t}\n\t\tbreak\n\tcase sysNetPathIPv6:\n\n\t\tif r.Link != \"\" {\n\t\t\tprocPath = path.Join(path.Join(path.Join(path.Join(sysNetPath, sysNetPathIPv6), \"conf\"), r.Link), r.Property)\n\t\t} else {\n\t\t\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathIPv6), r.Property)\n\t\t}\n\t\tbreak\n\tdefault:\n\t\treturn \"\", errors.New(\"Path not found\")\n\t}\n\n\treturn procPath, nil\n}","func (o IopingSpecVolumeVolumeSourceProjectedSourcesServiceAccountTokenOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesServiceAccountToken) string { return v.Path }).(pulumi.StringOutput)\n}","func (c *CaRsa) GetPath() string { return fmt.Sprintf(\"/api/objects/ca/rsa/%s\", c.Reference) }","func (c *Config) StoragePath() string {\n\tif c.options.StoragePath == \"\" {\n\t\tconst dirName = \"\"\n\n\t\tstorageDir := fs.FindDir(fs.StoragePaths)\n\t\tif fs.PathWritable(storageDir) && !c.ReadOnly() {\n\t\t\treturn storageDir\n\t\t}\n\t\t// Use .esp in home directory?\n\t\tif usr, _ := user.Current(); usr.HomeDir != \"\" {\n\t\t\tp := fs.Abs(filepath.Join(usr.HomeDir, fs.HiddenPath, dirName))\n\t\t\tif fs.PathWritable(p) || c.ReadOnly() {\n\t\t\t\treturn p\n\t\t\t}\n\t\t}\n\n\t\t// Fallback directory in case nothing else works.\n\t\tif c.ReadOnly() {\n\t\t\treturn fs.Abs(filepath.Join(fs.HiddenPath, dirName))\n\t\t}\n\n\t\treturn \"\"\n\t}\n\n\treturn fs.Abs(c.options.StoragePath)\n}","func (m *mountPoint) Path() string {\n\tif m.Volume != nil {\n\t\treturn m.Volume.Path()\n\t}\n\n\treturn m.Source\n}","func (p Pather) GetPath(svc addr.HostSVC, ps *seg.PathSegment) (*snet.SVCAddr, error) {\n\tif len(ps.ASEntries) == 0 {\n\t\treturn nil, serrors.New(\"empty path\")\n\t}\n\n\tbeta := ps.Info.SegmentID\n\t// The hop fields need to be in reversed order.\n\thopFields := make([]path.HopField, len(ps.ASEntries))\n\tfor i, entry := range ps.ASEntries {\n\t\thopFields[len(hopFields)-1-i] = path.HopField{\n\t\t\tConsIngress: entry.HopEntry.HopField.ConsIngress,\n\t\t\tConsEgress: entry.HopEntry.HopField.ConsEgress,\n\t\t\tExpTime: entry.HopEntry.HopField.ExpTime,\n\t\t\tMac: entry.HopEntry.HopField.MAC,\n\t\t}\n\t\t// the last AS entry is our AS for this we don't need to modify the beta.\n\t\tif i < len(ps.ASEntries)-1 {\n\t\t\tbeta = beta ^ binary.BigEndian.Uint16(entry.HopEntry.HopField.MAC[:2])\n\t\t}\n\t}\n\n\thops := len(hopFields)\n\tdec := scion.Decoded{\n\t\tBase: scion.Base{\n\t\t\tPathMeta: scion.MetaHdr{\n\t\t\t\tCurrHF: 0,\n\t\t\t\tCurrINF: 0,\n\t\t\t\tSegLen: [3]uint8{uint8(hops), 0, 0},\n\t\t\t},\n\t\t\tNumHops: hops,\n\t\t\tNumINF: 1,\n\t\t},\n\t\tInfoFields: []path.InfoField{{\n\t\t\tTimestamp: util.TimeToSecs(ps.Info.Timestamp),\n\t\t\tConsDir: false,\n\t\t\tSegID: beta,\n\t\t}},\n\t\tHopFields: hopFields,\n\t}\n\tpath, err := snetpath.NewSCIONFromDecoded(dec)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"serializing path\", err)\n\t}\n\tifID := dec.HopFields[0].ConsIngress\n\tnextHop := p.NextHopper.UnderlayNextHop(ifID)\n\tif nextHop == nil {\n\t\treturn nil, serrors.New(\"first-hop border router not found\", \"intf_id\", ifID)\n\t}\n\treturn &snet.SVCAddr{\n\t\tIA: ps.FirstIA(),\n\t\tPath: path,\n\t\tNextHop: nextHop,\n\t\tSVC: svc,\n\t}, nil\n\n}","func (c *Client) Path() string {\n\tc.mu.RLock()\n\tdefer c.mu.RUnlock()\n\treturn c.path\n}","func (store FileStore) infoPath(id string) string {\n\treturn filepath.Join(store.Path, id+\".info\")\n}","func (h Hashicorp) GetPath() string {\n\treturn h.FilePath\n}","func (o TransferJobTransferSpecGcsDataSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecGcsDataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (module *SdkLibrary) apiDistPath(apiScope *apiScope) string {\n\treturn path.Join(\"apistubs\", module.distGroup(), apiScope.name)\n}","func (o ApplicationOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationOperationSyncSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}","func (r *ManagerResource) Path() string {\n\treturn r.path\n}","func Path() string {\n\treturn c.Path\n}","func (o IopingSpecVolumeVolumeSourceHostPathOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceHostPath) string { return v.Path }).(pulumi.StringOutput)\n}","func (o ArgoCDSpecServerGrpcIngressOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecServerGrpcIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func SandboxConfigurationItemPath(id string, item Item) (string, error) {\n\tif id == \"\" {\n\t\treturn \"\", fmt.Errorf(\"Empty sandbox ID\")\n\t}\n\n\titemFile, err := itemToFile(item)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn filepath.Join(VCStorePrefix, ConfigStoragePath(), id, itemFile), nil\n}","func (o ArgoCDSpecServerIngressPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ArgoCDSpecServerIngress) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}","func (i *Index) securePath(path string) (cleaned string, hostpath string) {\n\tpath, err := url.QueryUnescape(path)\n\tif err != nil {\n\t\treturn \"/\", i.Root\n\t}\n\tcleaned = filepath.Clean(path)\n\thostpath = i.Root + cleaned\n\treturn\n}","func (o LocalCopyResponseOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LocalCopyResponse) string { return v.Path }).(pulumi.StringOutput)\n}","func (o ApplicationStatusOperationStateSyncResultSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func (o IopingSpecVolumeVolumeSourceNfsPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceNfs) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Path\n\t}).(pulumi.StringPtrOutput)\n}","func (o FioSpecVolumeVolumeSourceProjectedSourcesServiceAccountTokenOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesServiceAccountToken) string { return v.Path }).(pulumi.StringOutput)\n}","func (p *Plugin) ScopedPath(s string) string {\n\tif p.PluginObj.Config.PropagatedMount != \"\" && strings.HasPrefix(s, p.PluginObj.Config.PropagatedMount) {\n\t\t// re-scope to the propagated mount path on the host\n\t\treturn filepath.Join(filepath.Dir(p.Rootfs), \"propagated-mount\", strings.TrimPrefix(s, p.PluginObj.Config.PropagatedMount))\n\t}\n\treturn filepath.Join(p.Rootfs, s)\n}","func (o TransferJobTransferSpecAzureBlobStorageDataSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecAzureBlobStorageDataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}","func getPath(svc addr.HostSVC, ps *seg.PathSegment,\n\ttopoProv topology.Provider) (*snet.SVCAddr, error) {\n\n\tp, err := legacyPath(ps)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"constructing path from segment\", err)\n\t}\n\tif err := p.Reverse(); err != nil {\n\t\treturn nil, serrors.WrapStr(\"reversing path\", err)\n\t}\n\tif err := p.InitOffsets(); err != nil {\n\t\treturn nil, serrors.WrapStr(\"initializing offsets\", err)\n\t}\n\thopF, err := p.GetHopField(p.HopOff)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"extracting first hop field\", err)\n\t}\n\ttopo := topoProv.Get()\n\tifID := hopF.ConsIngress\n\tUnderlayNextHop, ok := topo.UnderlayNextHop2(ifID)\n\tif !ok {\n\t\treturn nil, serrors.New(\"first-hop border router not found\", \"intf_id\", ifID)\n\t}\n\treturn &snet.SVCAddr{IA: ps.FirstIA(), Path: p, NextHop: UnderlayNextHop, SVC: svc}, nil\n}","func (v VirtualSwitch) Path() (string, error) {\n\treturn v.virtualSwitch.Path()\n}","func (o FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItems) string { return v.Path }).(pulumi.StringOutput)\n}","func (o LocalCopyOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LocalCopy) string { return v.Path }).(pulumi.StringOutput)\n}","func (c *Client) Path() string {\n\treturn Path\n}","func (o ApplicationStatusOperationStateOperationSyncSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}"],"string":"[\n \"func (sd *SiaDir) SiaPath() modules.SiaPath {\\n\\tsd.mu.Lock()\\n\\tdefer sd.mu.Unlock()\\n\\treturn sd.siaPath\\n}\",\n \"func (rf *RemoteFile) SiaPath() modules.SiaPath {\\n\\trf.mu.Lock()\\n\\tdefer rf.mu.Unlock()\\n\\treturn rf.siaPath\\n}\",\n \"func (o *Object) remotePath() string {\\n\\treturn o.fs.slashRootSlash + o.remote\\n}\",\n \"func newSiaPath(path string) modules.SiaPath {\\n\\tsp, err := modules.NewSiaPath(path)\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\treturn sp\\n}\",\n \"func (b *Binary) RemotePath() string {\\n\\treturn b.file.RemotePath()\\n}\",\n \"func (r *Repository) Path(id digest.Digest) (dir, path string) {\\n\\tdir = filepath.Join(r.Root, id.Hex()[:2])\\n\\treturn dir, filepath.Join(dir, id.Hex()[2:])\\n}\",\n \"func (proxy *remoteDriverProxy) Path(name string) (string, error) {\\n\\tvar req = remoteVolumePathReq{\\n\\t\\tName: name,\\n\\t}\\n\\n\\tvar resp remoteVolumePathResp\\n\\n\\tif err := proxy.client.CallService(remoteVolumePathService, &req, &resp, true); err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\tif resp.Err != \\\"\\\" {\\n\\t\\treturn \\\"\\\", errors.New(resp.Err)\\n\\t}\\n\\n\\treturn resp.Mountpoint, nil\\n}\",\n \"func LoadSiaDir(rootDir string, siaPath modules.SiaPath, deps modules.Dependencies, wal *writeaheadlog.WAL) (sd *SiaDir, err error) {\\n\\tsd = &SiaDir{\\n\\t\\tdeps: deps,\\n\\t\\tsiaPath: siaPath,\\n\\t\\trootDir: rootDir,\\n\\t\\twal: wal,\\n\\t}\\n\\tsd.metadata, err = loadSiaDirMetadata(siaPath.SiaDirMetadataSysPath(rootDir), modules.ProdDependencies)\\n\\treturn sd, err\\n}\",\n \"func (r *RepoRef) SkeletonsPath() string {\\n\\treturn filepath.Join(r.LocalPath(), SkeletonsDir)\\n}\",\n \"func (*HttpLocalSites) GetPath() string { return \\\"/api/objects/http/local_site/\\\" }\",\n \"func (server *SingleInstance) Path() string {\\n\\treturn server.path\\n}\",\n \"func (d Driver) getSecureDiffPath(id, parent string, canBeRemote bool) string {\\n\\tvar diffDirName string\\n\\n\\tif parent == \\\"\\\" || d.isParent(id, parent) {\\n\\t\\tdiffDirName = \\\"diff\\\"\\n\\t} else {\\n\\t\\tdiffDirName = fmt.Sprintf(\\\"%s-%s\\\", \\\"diff\\\", parent)\\n\\t}\\n\\n\\tlocalSecureDiffPath := path.Join(d.dir(id), constSecureBaseDirName, diffDirName)\\n\\tremoteSecureDiffPath := path.Join(d.options.remoteDir, id, constSecureBaseDirName, diffDirName)\\n\\tlogrus.Debugf(\\\"secureoverlay2: getSecureDiffPath %s. localSecureDiffPath %s remoteSecureDiffPath\\\", localSecureDiffPath, remoteSecureDiffPath)\\n\\tdiffPath := localSecureDiffPath\\n\\t// remote only \\\"wins\\\" if local does not exist and remote exists\\n\\tif canBeRemote && d.options.remoteDir != \\\"\\\" {\\n\\t\\tif b, _ := exists(localSecureDiffPath); !b {\\n\\t\\t\\tif b, _ := exists(remoteSecureDiffPath); b {\\n\\t\\t\\t\\tdiffPath = remoteSecureDiffPath\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\tlogrus.Debugf(\\\"secureoverlay2: getSecureDiffPath w. id: %s, parent: %s, canBeRemote: %v returns %s\\\",\\n\\t\\tid, parent, canBeRemote, diffPath)\\n\\n\\treturn diffPath\\n}\",\n \"func (d *driver) fullPath(path string) string {\\n\\treturn _path.Join(\\\"/ipfs\\\", d.roothash, path)\\n}\",\n \"func (o ApplicationStatusOperationStateSyncResultSourceOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultSource) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (d *Datastore) s3Path(p string) string {\\n\\treturn path.Join(d.RootDirectory, p)\\n}\",\n \"func (p *Part) RemotePath(prefix string) string {\\n\\tfor strings.HasSuffix(prefix, \\\"/\\\") {\\n\\t\\tprefix = prefix[:len(prefix)-1]\\n\\t}\\n\\treturn fmt.Sprintf(\\\"%s/%s/%016X_%016X_%016X\\\", prefix, p.Path, p.FileSize, p.Offset, p.Size)\\n}\",\n \"func (o ApplicationStatusOperationStateOperationSyncSourceOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSource) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (fi *fileInfo) Path() string {\\n\\treturn fi.fullPath\\n}\",\n \"func (ss *Sources) workspacePath(id consumerID) string {\\n\\treturn filepath.Join(ss.RootPath, \\\"workspace\\\", id.Namespace, id.Name, id.consumer)\\n}\",\n \"func (s *Store) Path() string { return s.path }\",\n \"func (s *Store) Path() string { return s.path }\",\n \"func BattleShiritoriPath(battleID string, userID string) string {\\n\\treturn fmt.Sprintf(\\\"/streams/battles/%v/%v\\\", battleID, userID)\\n}\",\n \"func (r *Resolver) Path(resource interface{}, id string) (path string, err error) {\\n\\tprovider := r.Provider\\n\\tswitch resource.(type) {\\n\\tcase *Provider:\\n\\t\\tr := Provider{}\\n\\t\\tr.UID = id\\n\\t\\tr.Link()\\n\\t\\tpath = r.SelfLink\\n\\tcase *Folder:\\n\\t\\tr := Folder{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tcase *Datacenter:\\n\\t\\tr := Datacenter{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tcase *Cluster:\\n\\t\\tr := Cluster{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tcase *Host:\\n\\t\\tr := Host{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tcase *Network:\\n\\t\\tr := Network{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tcase *Datastore:\\n\\t\\tr := Datastore{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tcase *VM:\\n\\t\\tr := VM{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tcase *Workload:\\n\\t\\tr := Workload{}\\n\\t\\tr.ID = id\\n\\t\\tr.Link(provider)\\n\\t\\tpath = r.SelfLink\\n\\tdefault:\\n\\t\\terr = liberr.Wrap(\\n\\t\\t\\tbase.ResourceNotResolvedError{\\n\\t\\t\\t\\tObject: resource,\\n\\t\\t\\t})\\n\\t}\\n\\n\\tpath = strings.TrimRight(path, \\\"/\\\")\\n\\n\\treturn\\n}\",\n \"func (o ApplicationStatusOperationStateSyncResultSourcePtrOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v *ApplicationStatusOperationStateSyncResultSource) *string {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.Path\\n\\t}).(pulumi.StringPtrOutput)\\n}\",\n \"func (sds *SiaDirSet) NewSiaDir(siaPath string) (*SiaDirSetEntry, error) {\\n\\tsds.mu.Lock()\\n\\tdefer sds.mu.Unlock()\\n\\t// Check is SiaDir already exists\\n\\tsiaPath = strings.Trim(siaPath, \\\"/\\\")\\n\\texists, err := sds.exists(siaPath)\\n\\tif exists {\\n\\t\\treturn nil, ErrPathOverload\\n\\t}\\n\\tif !os.IsNotExist(err) && err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tsd, err := New(siaPath, sds.rootDir, sds.wal)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tentry := sds.newSiaDirSetEntry(sd)\\n\\tthreadUID := randomThreadUID()\\n\\tentry.threadMap[threadUID] = newThreadType()\\n\\tsds.siaDirMap[siaPath] = entry\\n\\treturn &SiaDirSetEntry{\\n\\t\\tsiaDirSetEntry: entry,\\n\\t\\tthreadUID: threadUID,\\n\\t}, nil\\n}\",\n \"func suriPath() (exists bool) {\\n var err error\\n path, err := utils.GetKeyValueString(\\\"suriPath\\\", \\\"path\\\")\\n if err != nil {\\n logs.Error(\\\"suriPath Error getting data from main.conf\\\")\\n }\\n\\n if _, err := os.Stat(path); os.IsNotExist(err) {\\n logs.Error(\\\"Suricata not installed, at least folder /etc/suricata dosn't exist\\\")\\n return false\\n }\\n return true\\n}\",\n \"func (o TransferJobTransferSpecAwsS3DataSourceOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v TransferJobTransferSpecAwsS3DataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (driver *Driver) Path(volumeName, volumeID string) (string, error) {\\n\\tif volumeName == \\\"\\\" && volumeID == \\\"\\\" {\\n\\t\\treturn \\\"\\\", errors.New(\\\"Missing volume name or ID\\\")\\n\\t}\\n\\n\\tinstances, err := driver.sdm.GetInstance()\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\tswitch {\\n\\tcase len(instances) == 0:\\n\\t\\treturn \\\"\\\", errors.New(\\\"No instances\\\")\\n\\tcase len(instances) > 1:\\n\\t\\treturn \\\"\\\", errors.New(\\\"Too many instances returned, limit the storagedrivers\\\")\\n\\t}\\n\\n\\tvolumes, err := driver.sdm.GetVolume(volumeID, volumeName)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\tswitch {\\n\\tcase len(volumes) == 0:\\n\\t\\treturn \\\"\\\", errors.New(\\\"No volumes returned by name\\\")\\n\\tcase len(volumes) > 1:\\n\\t\\treturn \\\"\\\", errors.New(\\\"Multiple volumes returned by name\\\")\\n\\t}\\n\\n\\tvolumeAttachment, err := driver.sdm.GetVolumeAttach(volumes[0].VolumeID, instances[0].InstanceID)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\tif len(volumeAttachment) == 0 {\\n\\t\\treturn \\\"\\\", nil\\n\\t}\\n\\n\\tmounts, err := driver.osdm.GetMounts(volumeAttachment[0].DeviceName, \\\"\\\")\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\tif len(mounts) == 0 {\\n\\t\\treturn \\\"\\\", nil\\n\\t}\\n\\n\\treturn mounts[0].Mountpoint, nil\\n}\",\n \"func (o IopingSpecVolumeVolumeSourceProjectedSourcesSecretItemsOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesSecretItems) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (i *Instance) Path() (string, error) {\\n\\tref, _, _, err := i.GetAsAny(WmiPathKey)\\n\\treturn ref.(string), err\\n}\",\n \"func (s *Store) IDPath() string { return filepath.Join(s.path, \\\"id\\\") }\",\n \"func (d ImagefsDriver) Path(r *volume.PathRequest) (*volume.PathResponse, error) {\\n\\tgetReq := volume.GetRequest{\\n\\t\\tName: r.Name,\\n\\t}\\n\\tret, err := d.Get(&getReq)\\n\\tvar _ret *volume.PathResponse\\n\\tif ret != nil {\\n\\t\\t_ret = &volume.PathResponse{\\n\\t\\t\\tMountpoint: ret.Volume.Mountpoint,\\n\\t\\t}\\n\\t}\\n\\treturn _ret, err\\n}\",\n \"func getRepoPath() (string, error) {\\n\\t// Set default base path and directory name\\n\\tdirectoryName := \\\".saturn\\\"\\n\\n\\t// Join the path and directory name, then expand the home path\\n\\tfullPath, err := homedir.Expand(filepath.Join(\\\"~\\\", directoryName))\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\t// Return the shortest lexical representation of the path\\n\\treturn filepath.Clean(fullPath), nil\\n}\",\n \"func PantahubS3Path() string {\\n\\tif GetEnv(EnvPantahubStorageDriver) == \\\"s3\\\" {\\n\\t\\treturn GetEnv(EnvPantahubStoragePath)\\n\\t}\\n\\n\\tbasePath := path.Join(GetEnv(EnvPantahubS3Path), GetEnv(EnvPantahubStoragePath))\\n\\n\\tif basePath == \\\"\\\" {\\n\\t\\tbasePath = \\\".\\\"\\n\\t}\\n\\n\\treturn basePath\\n}\",\n \"func (c *Credentials) path() (string, error) {\\n\\tusr, err := user.Current()\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", fmt.Errorf(\\\"failed to get current user\\\")\\n\\t}\\n\\treturn path.Join(usr.HomeDir, *cred), nil\\n}\",\n \"func (ls *LocationStore) Path() string {\\n\\treturn fmt.Sprintf(\\\"env/locations/%s.yml\\\", ls.name)\\n}\",\n \"func TestNewSiaDir(t *testing.T) {\\n\\tif testing.Short() && !build.VLONG {\\n\\t\\tt.SkipNow()\\n\\t}\\n\\tt.Parallel()\\n\\t// Create filesystem.\\n\\troot := filepath.Join(testDir(t.Name()), \\\"fs-root\\\")\\n\\tfs := newTestFileSystem(root)\\n\\t// Create dir /sub/foo\\n\\tsp := newSiaPath(\\\"sub/foo\\\")\\n\\tif err := fs.NewSiaDir(sp, modules.DefaultDirPerm); err != nil {\\n\\t\\tt.Fatal(err)\\n\\t}\\n\\t// The whole path should exist.\\n\\tif _, err := os.Stat(filepath.Join(root, sp.String())); err != nil {\\n\\t\\tt.Fatal(err)\\n\\t}\\n}\",\n \"func (s *session) Path() dbus.ObjectPath {\\n\\treturn s.path\\n}\",\n \"func (o ApplicationStatusOperationStateOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v *ApplicationStatusOperationStateOperationSyncSource) *string {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.Path\\n\\t}).(pulumi.StringPtrOutput)\\n}\",\n \"func Path(named Named) (name string) {\\n\\tif r, ok := named.(namedRepository); ok {\\n\\t\\treturn r.Path()\\n\\t}\\n\\t_, path := splitDomain(named.Name())\\n\\treturn path\\n}\",\n \"func (c *Client) Path() (string, error) {\\n\\treturn c.GetProperty(\\\"path\\\")\\n}\",\n \"func (b *Bucket) Path() (string, error) {\\n\\tconf := b.conf.Viper.ConfigFileUsed()\\n\\tif conf == \\\"\\\" {\\n\\t\\treturn b.cwd, nil\\n\\t}\\n\\treturn filepath.Dir(filepath.Dir(conf)), nil\\n}\",\n \"func (r *RemoteSyslogServer) GetPath() string {\\n\\treturn fmt.Sprintf(\\\"/api/objects/remote_syslog/server/%s\\\", r.Reference)\\n}\",\n \"func (h *HttpLocalSite) GetPath() string {\\n\\treturn fmt.Sprintf(\\\"/api/objects/http/local_site/%s\\\", h.Reference)\\n}\",\n \"func (r *ServiceLinkedRole) Path() pulumi.StringOutput {\\n\\treturn (pulumi.StringOutput)(r.s.State[\\\"path\\\"])\\n}\",\n \"func (k *Key) path() (string, error) {\\n\\tusr, err := user.Current()\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", fmt.Errorf(\\\"failed to get current user\\\")\\n\\t}\\n\\treturn path.Join(usr.HomeDir, *identity), nil\\n}\",\n \"func (d *DirectHash) getPath() string {\\n\\tdomain := d.RootDomain\\n\\tif len(d.ReleaseBranch) > 0 {\\n\\t\\tdomain = string(d.ReleaseBranch) + \\\".\\\" + d.RootDomain\\n\\t}\\n\\treturn fmt.Sprintf(\\\"https://%s/\\\"+strings.Trim(d.Path, \\\"/\\\"), domain, d.ReleaseBranch.String())\\n}\",\n \"func (id ID) Path() string {\\n\\treturn id.path\\n}\",\n \"func (p *SeriesPartition) Path() string { return p.path }\",\n \"func (o *StorageNetAppCifsShareAllOf) GetPath() string {\\n\\tif o == nil || o.Path == nil {\\n\\t\\tvar ret string\\n\\t\\treturn ret\\n\\t}\\n\\treturn *o.Path\\n}\",\n \"func (o FioSpecVolumeVolumeSourceProjectedSourcesSecretItemsOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesSecretItems) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (m *CgroupfsManager) SandboxCgroupPath(sbParent, sbID string) (cgParent, cgPath string, _ error) {\\n\\tif strings.HasSuffix(path.Base(sbParent), \\\".slice\\\") {\\n\\t\\treturn \\\"\\\", \\\"\\\", fmt.Errorf(\\\"cri-o configured with cgroupfs cgroup manager, but received systemd slice as parent: %s\\\", sbParent)\\n\\t}\\n\\n\\tif err := verifyCgroupHasEnoughMemory(sbParent, m.memoryPath, m.memoryMaxFile); err != nil {\\n\\t\\treturn \\\"\\\", \\\"\\\", err\\n\\t}\\n\\n\\treturn sbParent, filepath.Join(sbParent, containerCgroupPath(sbID)), nil\\n}\",\n \"func (o ApplicationOperationSyncSourceOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ApplicationOperationSyncSource) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (o IopingSpecVolumeVolumeSourceNfsOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceNfs) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (o TransferJobTransferSpecAwsS3DataSourcePtrOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v *TransferJobTransferSpecAwsS3DataSource) *string {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.Path\\n\\t}).(pulumi.StringPtrOutput)\\n}\",\n \"func (d Driver) getSecureCryptMntPath(id string) string {\\n\\treturn path.Join(d.dir(id), constSecureBaseDirName, constSecureCryptMntDirName)\\n}\",\n \"func (n EpisodeDetails) GetSeriesPath(destination string) string {\\n\\treturn filepath.Join(destination, FileNameCleaner(n.Showtitle))\\n}\",\n \"func (*CaRsas) GetPath() string { return \\\"/api/objects/ca/rsa/\\\" }\",\n \"func (i *Image) Path() string {\\n\\treturn i.p\\n}\",\n \"func (s *Store) Path() string {\\n\\treturn s.path\\n}\",\n \"func (s *Store) Path() string {\\n\\treturn s.path\\n}\",\n \"func (cs *ControllerServer) getVolumeSharePath(vol *nfsVolume) string {\\n\\treturn filepath.Join(string(filepath.Separator), vol.baseDir, vol.subDir)\\n}\",\n \"func (d *MinioDriver) Path(r volume.Request) volume.Response {\\n\\td.m.RLock()\\n\\tdefer d.m.RUnlock()\\n\\n\\tv, exists := d.volumes[r.Name]\\n\\tif !exists {\\n\\t\\treturn volumeResp(\\\"\\\", \\\"\\\", nil, capability, newErrVolNotFound(r.Name).Error())\\n\\t}\\n\\treturn volumeResp(v.mountpoint, r.Name, nil, capability, \\\"\\\")\\n}\",\n \"func Path(subpath string) (string, error) {\\n\\tif !isinit {\\n\\t\\treturn \\\"\\\", fmt.Errorf(\\\"pathresolver not initialized\\\")\\n\\t}\\n\\n\\tif subpath == \\\"\\\" {\\n\\t\\treturn basedirectory, nil\\n\\t}\\n\\n\\tif filepath.IsAbs(subpath) {\\n\\t\\treturn \\\"\\\", fmt.Errorf(\\\"cannot use absolute path as subpath\\\")\\n\\t}\\n\\n\\treturn filepath.Join(basedirectory, subpath), nil\\n}\",\n \"func (o ApplicationStatusSyncComparedToSourceOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ApplicationStatusSyncComparedToSource) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (*RemoteSyslogServers) GetPath() string { return \\\"/api/objects/remote_syslog/server/\\\" }\",\n \"func (o IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItems) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (o ArgoCDSpecServerIngressOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ArgoCDSpecServerIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (o IopingSpecVolumeVolumeSourceGlusterfsOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceGlusterfs) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (r *SysNet) getPath() (string, error) {\\n\\tvar procPath string\\n\\n\\tswitch r.Path {\\n\\tcase sysNetPathCore:\\n\\t\\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathCore), r.Property)\\n\\t\\tbreak\\n\\tcase sysNetPathIPv4:\\n\\n\\t\\tif r.Link != \\\"\\\" {\\n\\t\\t\\tprocPath = path.Join(path.Join(path.Join(path.Join(sysNetPath, sysNetPathIPv4), \\\"conf\\\"), r.Link), r.Property)\\n\\t\\t} else {\\n\\t\\t\\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathIPv4), r.Property)\\n\\t\\t}\\n\\t\\tbreak\\n\\tcase sysNetPathIPv6:\\n\\n\\t\\tif r.Link != \\\"\\\" {\\n\\t\\t\\tprocPath = path.Join(path.Join(path.Join(path.Join(sysNetPath, sysNetPathIPv6), \\\"conf\\\"), r.Link), r.Property)\\n\\t\\t} else {\\n\\t\\t\\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathIPv6), r.Property)\\n\\t\\t}\\n\\t\\tbreak\\n\\tdefault:\\n\\t\\treturn \\\"\\\", errors.New(\\\"Path not found\\\")\\n\\t}\\n\\n\\treturn procPath, nil\\n}\",\n \"func (o IopingSpecVolumeVolumeSourceProjectedSourcesServiceAccountTokenOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesServiceAccountToken) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (c *CaRsa) GetPath() string { return fmt.Sprintf(\\\"/api/objects/ca/rsa/%s\\\", c.Reference) }\",\n \"func (c *Config) StoragePath() string {\\n\\tif c.options.StoragePath == \\\"\\\" {\\n\\t\\tconst dirName = \\\"\\\"\\n\\n\\t\\tstorageDir := fs.FindDir(fs.StoragePaths)\\n\\t\\tif fs.PathWritable(storageDir) && !c.ReadOnly() {\\n\\t\\t\\treturn storageDir\\n\\t\\t}\\n\\t\\t// Use .esp in home directory?\\n\\t\\tif usr, _ := user.Current(); usr.HomeDir != \\\"\\\" {\\n\\t\\t\\tp := fs.Abs(filepath.Join(usr.HomeDir, fs.HiddenPath, dirName))\\n\\t\\t\\tif fs.PathWritable(p) || c.ReadOnly() {\\n\\t\\t\\t\\treturn p\\n\\t\\t\\t}\\n\\t\\t}\\n\\n\\t\\t// Fallback directory in case nothing else works.\\n\\t\\tif c.ReadOnly() {\\n\\t\\t\\treturn fs.Abs(filepath.Join(fs.HiddenPath, dirName))\\n\\t\\t}\\n\\n\\t\\treturn \\\"\\\"\\n\\t}\\n\\n\\treturn fs.Abs(c.options.StoragePath)\\n}\",\n \"func (m *mountPoint) Path() string {\\n\\tif m.Volume != nil {\\n\\t\\treturn m.Volume.Path()\\n\\t}\\n\\n\\treturn m.Source\\n}\",\n \"func (p Pather) GetPath(svc addr.HostSVC, ps *seg.PathSegment) (*snet.SVCAddr, error) {\\n\\tif len(ps.ASEntries) == 0 {\\n\\t\\treturn nil, serrors.New(\\\"empty path\\\")\\n\\t}\\n\\n\\tbeta := ps.Info.SegmentID\\n\\t// The hop fields need to be in reversed order.\\n\\thopFields := make([]path.HopField, len(ps.ASEntries))\\n\\tfor i, entry := range ps.ASEntries {\\n\\t\\thopFields[len(hopFields)-1-i] = path.HopField{\\n\\t\\t\\tConsIngress: entry.HopEntry.HopField.ConsIngress,\\n\\t\\t\\tConsEgress: entry.HopEntry.HopField.ConsEgress,\\n\\t\\t\\tExpTime: entry.HopEntry.HopField.ExpTime,\\n\\t\\t\\tMac: entry.HopEntry.HopField.MAC,\\n\\t\\t}\\n\\t\\t// the last AS entry is our AS for this we don't need to modify the beta.\\n\\t\\tif i < len(ps.ASEntries)-1 {\\n\\t\\t\\tbeta = beta ^ binary.BigEndian.Uint16(entry.HopEntry.HopField.MAC[:2])\\n\\t\\t}\\n\\t}\\n\\n\\thops := len(hopFields)\\n\\tdec := scion.Decoded{\\n\\t\\tBase: scion.Base{\\n\\t\\t\\tPathMeta: scion.MetaHdr{\\n\\t\\t\\t\\tCurrHF: 0,\\n\\t\\t\\t\\tCurrINF: 0,\\n\\t\\t\\t\\tSegLen: [3]uint8{uint8(hops), 0, 0},\\n\\t\\t\\t},\\n\\t\\t\\tNumHops: hops,\\n\\t\\t\\tNumINF: 1,\\n\\t\\t},\\n\\t\\tInfoFields: []path.InfoField{{\\n\\t\\t\\tTimestamp: util.TimeToSecs(ps.Info.Timestamp),\\n\\t\\t\\tConsDir: false,\\n\\t\\t\\tSegID: beta,\\n\\t\\t}},\\n\\t\\tHopFields: hopFields,\\n\\t}\\n\\tpath, err := snetpath.NewSCIONFromDecoded(dec)\\n\\tif err != nil {\\n\\t\\treturn nil, serrors.WrapStr(\\\"serializing path\\\", err)\\n\\t}\\n\\tifID := dec.HopFields[0].ConsIngress\\n\\tnextHop := p.NextHopper.UnderlayNextHop(ifID)\\n\\tif nextHop == nil {\\n\\t\\treturn nil, serrors.New(\\\"first-hop border router not found\\\", \\\"intf_id\\\", ifID)\\n\\t}\\n\\treturn &snet.SVCAddr{\\n\\t\\tIA: ps.FirstIA(),\\n\\t\\tPath: path,\\n\\t\\tNextHop: nextHop,\\n\\t\\tSVC: svc,\\n\\t}, nil\\n\\n}\",\n \"func (c *Client) Path() string {\\n\\tc.mu.RLock()\\n\\tdefer c.mu.RUnlock()\\n\\treturn c.path\\n}\",\n \"func (store FileStore) infoPath(id string) string {\\n\\treturn filepath.Join(store.Path, id+\\\".info\\\")\\n}\",\n \"func (h Hashicorp) GetPath() string {\\n\\treturn h.FilePath\\n}\",\n \"func (o TransferJobTransferSpecGcsDataSourceOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v TransferJobTransferSpecGcsDataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (module *SdkLibrary) apiDistPath(apiScope *apiScope) string {\\n\\treturn path.Join(\\\"apistubs\\\", module.distGroup(), apiScope.name)\\n}\",\n \"func (o ApplicationOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v *ApplicationOperationSyncSource) *string {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.Path\\n\\t}).(pulumi.StringPtrOutput)\\n}\",\n \"func (r *ManagerResource) Path() string {\\n\\treturn r.path\\n}\",\n \"func Path() string {\\n\\treturn c.Path\\n}\",\n \"func (o IopingSpecVolumeVolumeSourceHostPathOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceHostPath) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (o ArgoCDSpecServerGrpcIngressOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ArgoCDSpecServerGrpcIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func SandboxConfigurationItemPath(id string, item Item) (string, error) {\\n\\tif id == \\\"\\\" {\\n\\t\\treturn \\\"\\\", fmt.Errorf(\\\"Empty sandbox ID\\\")\\n\\t}\\n\\n\\titemFile, err := itemToFile(item)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\treturn filepath.Join(VCStorePrefix, ConfigStoragePath(), id, itemFile), nil\\n}\",\n \"func (o ArgoCDSpecServerIngressPtrOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v *ArgoCDSpecServerIngress) *string {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.Path\\n\\t}).(pulumi.StringPtrOutput)\\n}\",\n \"func (i *Index) securePath(path string) (cleaned string, hostpath string) {\\n\\tpath, err := url.QueryUnescape(path)\\n\\tif err != nil {\\n\\t\\treturn \\\"/\\\", i.Root\\n\\t}\\n\\tcleaned = filepath.Clean(path)\\n\\thostpath = i.Root + cleaned\\n\\treturn\\n}\",\n \"func (o LocalCopyResponseOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v LocalCopyResponse) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (o ApplicationStatusOperationStateSyncResultSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func (o IopingSpecVolumeVolumeSourceNfsPtrOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceNfs) *string {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn &v.Path\\n\\t}).(pulumi.StringPtrOutput)\\n}\",\n \"func (o FioSpecVolumeVolumeSourceProjectedSourcesServiceAccountTokenOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesServiceAccountToken) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (p *Plugin) ScopedPath(s string) string {\\n\\tif p.PluginObj.Config.PropagatedMount != \\\"\\\" && strings.HasPrefix(s, p.PluginObj.Config.PropagatedMount) {\\n\\t\\t// re-scope to the propagated mount path on the host\\n\\t\\treturn filepath.Join(filepath.Dir(p.Rootfs), \\\"propagated-mount\\\", strings.TrimPrefix(s, p.PluginObj.Config.PropagatedMount))\\n\\t}\\n\\treturn filepath.Join(p.Rootfs, s)\\n}\",\n \"func (o TransferJobTransferSpecAzureBlobStorageDataSourceOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v TransferJobTransferSpecAzureBlobStorageDataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\",\n \"func getPath(svc addr.HostSVC, ps *seg.PathSegment,\\n\\ttopoProv topology.Provider) (*snet.SVCAddr, error) {\\n\\n\\tp, err := legacyPath(ps)\\n\\tif err != nil {\\n\\t\\treturn nil, serrors.WrapStr(\\\"constructing path from segment\\\", err)\\n\\t}\\n\\tif err := p.Reverse(); err != nil {\\n\\t\\treturn nil, serrors.WrapStr(\\\"reversing path\\\", err)\\n\\t}\\n\\tif err := p.InitOffsets(); err != nil {\\n\\t\\treturn nil, serrors.WrapStr(\\\"initializing offsets\\\", err)\\n\\t}\\n\\thopF, err := p.GetHopField(p.HopOff)\\n\\tif err != nil {\\n\\t\\treturn nil, serrors.WrapStr(\\\"extracting first hop field\\\", err)\\n\\t}\\n\\ttopo := topoProv.Get()\\n\\tifID := hopF.ConsIngress\\n\\tUnderlayNextHop, ok := topo.UnderlayNextHop2(ifID)\\n\\tif !ok {\\n\\t\\treturn nil, serrors.New(\\\"first-hop border router not found\\\", \\\"intf_id\\\", ifID)\\n\\t}\\n\\treturn &snet.SVCAddr{IA: ps.FirstIA(), Path: p, NextHop: UnderlayNextHop, SVC: svc}, nil\\n}\",\n \"func (v VirtualSwitch) Path() (string, error) {\\n\\treturn v.virtualSwitch.Path()\\n}\",\n \"func (o FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItems) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (o LocalCopyOutput) Path() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v LocalCopy) string { return v.Path }).(pulumi.StringOutput)\\n}\",\n \"func (c *Client) Path() string {\\n\\treturn Path\\n}\",\n \"func (o ApplicationStatusOperationStateOperationSyncSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.74364364","0.7279818","0.57709485","0.56959575","0.5652686","0.5584653","0.55214775","0.5389453","0.53671575","0.53460824","0.533256","0.52875644","0.52684146","0.5267466","0.5247035","0.52464074","0.51968586","0.5180792","0.5175526","0.5171551","0.5171551","0.5149007","0.5134776","0.5132581","0.5127024","0.5122917","0.5081935","0.5072549","0.5065655","0.5044717","0.5043704","0.5033308","0.50218445","0.5018088","0.50139695","0.50037605","0.4998196","0.49921614","0.49828467","0.49802768","0.49619955","0.4961612","0.49522248","0.4951539","0.49467763","0.4939332","0.49350157","0.49290487","0.4928961","0.49207598","0.49056935","0.49028972","0.49003157","0.48988467","0.48890874","0.4870896","0.48647636","0.48580962","0.485039","0.4848557","0.4848557","0.48399955","0.4831611","0.48266563","0.48237526","0.4818253","0.48071283","0.4803512","0.48011374","0.47945023","0.47937265","0.4788762","0.4780677","0.4777919","0.47764984","0.47762063","0.4754437","0.4750877","0.4749953","0.4744859","0.47386527","0.47355336","0.47309208","0.4729769","0.4719194","0.47187114","0.47182325","0.4704324","0.47033232","0.47016972","0.4697198","0.46923193","0.46872768","0.4684552","0.4679754","0.46779963","0.4675425","0.46706453","0.4667558","0.4665334"],"string":"[\n \"0.74364364\",\n \"0.7279818\",\n \"0.57709485\",\n \"0.56959575\",\n \"0.5652686\",\n \"0.5584653\",\n \"0.55214775\",\n \"0.5389453\",\n \"0.53671575\",\n \"0.53460824\",\n \"0.533256\",\n \"0.52875644\",\n \"0.52684146\",\n \"0.5267466\",\n \"0.5247035\",\n \"0.52464074\",\n \"0.51968586\",\n \"0.5180792\",\n \"0.5175526\",\n \"0.5171551\",\n \"0.5171551\",\n \"0.5149007\",\n \"0.5134776\",\n \"0.5132581\",\n \"0.5127024\",\n \"0.5122917\",\n \"0.5081935\",\n \"0.5072549\",\n \"0.5065655\",\n \"0.5044717\",\n \"0.5043704\",\n \"0.5033308\",\n \"0.50218445\",\n \"0.5018088\",\n \"0.50139695\",\n \"0.50037605\",\n \"0.4998196\",\n \"0.49921614\",\n \"0.49828467\",\n \"0.49802768\",\n \"0.49619955\",\n \"0.4961612\",\n \"0.49522248\",\n \"0.4951539\",\n \"0.49467763\",\n \"0.4939332\",\n \"0.49350157\",\n \"0.49290487\",\n \"0.4928961\",\n \"0.49207598\",\n \"0.49056935\",\n \"0.49028972\",\n \"0.49003157\",\n \"0.48988467\",\n \"0.48890874\",\n \"0.4870896\",\n \"0.48647636\",\n \"0.48580962\",\n \"0.485039\",\n \"0.4848557\",\n \"0.4848557\",\n \"0.48399955\",\n \"0.4831611\",\n \"0.48266563\",\n \"0.48237526\",\n \"0.4818253\",\n \"0.48071283\",\n \"0.4803512\",\n \"0.48011374\",\n \"0.47945023\",\n \"0.47937265\",\n \"0.4788762\",\n \"0.4780677\",\n \"0.4777919\",\n \"0.47764984\",\n \"0.47762063\",\n \"0.4754437\",\n \"0.4750877\",\n \"0.4749953\",\n \"0.4744859\",\n \"0.47386527\",\n \"0.47355336\",\n \"0.47309208\",\n \"0.4729769\",\n \"0.4719194\",\n \"0.47187114\",\n \"0.47182325\",\n \"0.4704324\",\n \"0.47033232\",\n \"0.47016972\",\n \"0.4697198\",\n \"0.46923193\",\n \"0.46872768\",\n \"0.4684552\",\n \"0.4679754\",\n \"0.46779963\",\n \"0.4675425\",\n \"0.46706453\",\n \"0.4667558\",\n \"0.4665334\"\n]"},"document_score":{"kind":"string","value":"0.85262775"},"document_rank":{"kind":"string","value":"0"}}},{"rowIdx":104789,"cells":{"query":{"kind":"string","value":"update fs block size to BlkSize"},"document":{"kind":"string","value":"func getBlkSize(p string) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(p, &fs)\n\tif err != nil {\n\t\tlog.Fatal(\"get block size error with:\", err)\n\t}\n\tBlockSize = uint64(fs.Bsize)\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (sm3 *SM3) BlockSize() int { return 64 }","func (s *State) BlockSize() int { return 2 }","func (s *State) BlockSize() int { return 8 }","func (d *state) BlockSize() int { return d.rate }","func (s *State) BlockSize() int { return 1 }","func (h *Hash) BlockSize() int { return len(h.buf) }","func (d *digest) BlockSize() int { return 1 }","func (d *RabinKarp64) BlockSize() int { return 1 }","func (bdi *Info) BlockSize() int {\n\treturn int(C.spdk_bdev_get_block_size(bdi.ptr()))\n}","func (c Zfs) UpdateBlock(b *i3barjson.Block) {\n\tb.Color = c.Color\n\tfullTextFmt := fmt.Sprintf(\"%s%%s\", c.Label)\n\n\tzpoolCmd := exec.Command(\"sudo\", \"zpool\", \"status\", c.PoolName)\n\tout, err := zpoolCmd.Output()\n\n\tif err != nil {\n\t\tb.Urgent = true\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\treturn\n\t}\n\n\tzpoolLines := strings.Split(string(out), \"\\n\")\n\tfor _, zpoolLine := range zpoolLines {\n\t\tline := strings.TrimSpace(zpoolLine)\n\t\tif strings.HasPrefix(line, \"state\") {\n\t\t\tsplit := strings.Split(line, \":\")\n\t\t\tstatus := strings.TrimSpace(split[1])\n\n\t\t\tif status == \"ONLINE\" {\n\t\t\t\tb.Urgent = false\n\t\t\t} else {\n\t\t\t\tb.Urgent = true\n\t\t\t}\n\t\t\tb.FullText = fmt.Sprintf(fullTextFmt, status)\n\t\t\treturn\n\t\t}\n\t}\n\n\tb.Urgent = true\n\tb.FullText = fmt.Sprintf(fullTextFmt, \"NOT FOUND\")\n\treturn\n}","func (h *ihash) BlockSize() int { return h.blockSize }","func (syncer *MerkleSyncer) updateHeight() {\n\tatomic.AddUint64(&syncer.height, 1)\n}","func (k Keeper) BlockByteSize(ctx sdk.Ctx) (res int64) {\n\tk.Paramstore.Get(ctx, types.KeyBlockByteSize, &res)\n\treturn\n}","func (ob *Observer) updateBlock(curHeight, nextHeight int64, curBlockHash string) error {\n\tblock, err := ob.deps.Recorder.Block(nextHeight)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"[Observer.updateBlock]: failed to get block info, height=%d\", nextHeight)\n\t}\n\n\tif curHeight != 0 && block.ParentBlockHash != curBlockHash {\n\t\tif err := ob.DeleteBlock(curHeight); err != nil {\n\t\t\treturn errors.Wrap(err, \"[Observer.updateBlock]: failed to delete a forked block\")\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tif err := ob.RecordBlockAndTxs(block); err != nil {\n\t\treturn errors.Wrap(err, \"[Observer.updateBlock]: failed to save and process block\")\n\t}\n\n\treturn nil\n}","func (xxh *xxHash) BlockSize() int {\n\treturn 1\n}","func (d *btrfs) Update(changedConfig map[string]string) error {\n\t// We only care about btrfs.mount_options.\n\tval, ok := changedConfig[\"btrfs.mount_options\"]\n\tif ok {\n\t\t// Custom mount options don't work inside containers\n\t\tif d.state.OS.RunningInUserNS {\n\t\t\treturn nil\n\t\t}\n\n\t\t// Trigger a re-mount.\n\t\td.config[\"btrfs.mount_options\"] = val\n\t\tmntFlags, mntOptions := filesystem.ResolveMountOptions(strings.Split(d.getMountOptions(), \",\"))\n\t\tmntFlags |= unix.MS_REMOUNT\n\n\t\terr := TryMount(\"\", GetPoolMountPath(d.name), \"none\", mntFlags, mntOptions)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tsize, ok := changedConfig[\"size\"]\n\tif ok {\n\t\t// Figure out loop path\n\t\tloopPath := loopFilePath(d.name)\n\n\t\tif d.config[\"source\"] != loopPath {\n\t\t\treturn fmt.Errorf(\"Cannot resize non-loopback pools\")\n\t\t}\n\n\t\t// Resize loop file\n\t\tf, err := os.OpenFile(loopPath, os.O_RDWR, 0600)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdefer func() { _ = f.Close() }()\n\n\t\tsizeBytes, _ := units.ParseByteSizeString(size)\n\n\t\terr = f.Truncate(sizeBytes)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tloopDevPath, err := loopDeviceSetup(loopPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdefer func() { _ = loopDeviceAutoDetach(loopDevPath) }()\n\n\t\terr = loopDeviceSetCapacity(loopDevPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = shared.RunCommand(\"btrfs\", \"filesystem\", \"resize\", \"max\", GetPoolMountPath(d.name))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}","func (ns *EsIndexer) UpdateLastBlockHeightFromDb() {\n\tbestBlock, err := ns.GetBestBlockFromDb()\n\tif err != nil {\n\t\treturn\n\t}\n\tns.lastBlockHeight = bestBlock.BlockNo\n\tns.lastBlockHash = bestBlock.id\n}","func (s *StatFS) BlockFiles(size int) int {\n\treturn size / int(s.stat.Bsize)\n}","func (digest *sm3Digest) BlockSize() int {\n\treturn BlockSize\n}","func (d Avx512Digest) BlockSize() int { return BlockSize }","func (nsc *NilConsumerStatsCollector) UpdateCheckpointSize(int) {}","func (s *Store) UpdateTreeStoreSize(key string, newSize int64) error {\n\treturn s.db.Do(func(tx *sql.Tx) error {\n\t\t_, err := tx.Exec(\"UPDATE aciinfo SET treestoresize = $1 WHERE blobkey == $2\", newSize, key)\n\t\treturn err\n\t})\n}","func (fs *FileSystem) UpdateFst(file inode) error {\n\tfs.nextFreeInode = fs.nextFreeInode[1:]\n\tfs.nextFreeDataBlock = fs.nextFreeDataBlock[len(file.dataList):]\n\treturn nil\n}","func (fsys *FS) Statfs(path string, stat *fuse.Statfs_t) (errc int) {\n\tdefer fs.Trace(path, \"\")(\"stat=%+v, errc=%d\", stat, &errc)\n\tconst blockSize = 4096\n\tfsBlocks := uint64(1 << 50)\n\tif runtime.GOOS == \"windows\" {\n\t\tfsBlocks = (1 << 43) - 1\n\t}\n\tstat.Blocks = fsBlocks // Total data blocks in file system.\n\tstat.Bfree = fsBlocks // Free blocks in file system.\n\tstat.Bavail = fsBlocks // Free blocks in file system if you're not root.\n\tstat.Files = 1E9 // Total files in file system.\n\tstat.Ffree = 1E9 // Free files in file system.\n\tstat.Bsize = blockSize // Block size\n\tstat.Namemax = 255 // Maximum file name length?\n\tstat.Frsize = blockSize // Fragment size, smallest addressable data size in the file system.\n\treturn 0\n}","func getSizeForPath(path string) (uint64, error) {\n\ts := syscall.Statfs_t{}\n\tif err := syscall.Statfs(path, &s); err != nil {\n\t\treturn 0, fmt.Errorf(\"failed to statfs on %s, %+v\", path, err)\n\t}\n\n\treturn s.Blocks * uint64(s.Bsize), nil\n}","func (b *Block) Size() common.StorageSize {\n\tif size := b.size.Load(); size != nil {\n\t\treturn size.(common.StorageSize)\n\t}\n\tc := writeCounter(0)\n\trlp.Encode(&c, b)\n\tb.size.Store(common.StorageSize(c))\n\treturn common.StorageSize(c)\n}","func blocklistUpdate() error {\n\tctx, cancel := context.WithTimeout(context.TODO(), ezhttp.DefaultTimeout10s)\n\tdefer cancel()\n\tres, err := ezhttp.Get(\n\t\tctx,\n\t\t\"https://download.dnscrypt.info/blacklists/domains/mybase.txt\",\n\t\tezhttp.Header(\"User-Agent\", \"github.com/function61/function53\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn atomicfilewrite.Write(blocklistFilename, func(blocklist io.Writer) error {\n\t\t_, err := io.Copy(blocklist, res.Body)\n\t\treturn err\n\t})\n}","func (leaf *Node) update(newInfo os.FileInfo, withContent bool) (err error) {\n\tif newInfo == nil {\n\t\tnewInfo, err = os.Stat(leaf.SysPath)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"lib/memfs: Node.update %q: %s\",\n\t\t\t\tleaf.Path, err.Error())\n\t\t}\n\t}\n\n\tif leaf.Mode != newInfo.Mode() {\n\t\tleaf.Mode = newInfo.Mode()\n\t\treturn nil\n\t}\n\n\tleaf.ModTime = newInfo.ModTime()\n\tleaf.Size = newInfo.Size()\n\n\tif !withContent || newInfo.IsDir() {\n\t\treturn nil\n\t}\n\n\treturn leaf.updateContent()\n}","func (f FormatHeader) BlockSize() uint16 {\n\treturn (f.BitsPerSample / 8) * f.NumChannels\n}","func (x *gcm) BlockSize() int { return x.blockSize }","func (h *blockHeader) setSize(v uint32) {\n\tconst mask = 7\n\t*h = (*h)&mask | blockHeader(v<<3)\n}","func (c *BlockCache) setDbFiles(height int) {\n\tif height <= c.nextBlock {\n\t\tif height < c.firstBlock {\n\t\t\theight = c.firstBlock\n\t\t}\n\t\tindex := height - c.firstBlock\n\t\tif err := c.lengthsFile.Truncate(int64(index * 4)); err != nil {\n\t\t\tLog.Fatal(\"truncate lengths file failed: \", err)\n\t\t}\n\t\tif err := c.blocksFile.Truncate(c.starts[index]); err != nil {\n\t\t\tLog.Fatal(\"truncate blocks file failed: \", err)\n\t\t}\n\t\tc.Sync()\n\t\tc.starts = c.starts[:index+1]\n\t\tc.nextBlock = height\n\t\tc.setLatestHash()\n\t}\n}","func Diskuse(path string) (cap uint64, used uint64) {\n\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\tcap = fs.Blocks * uint64(fs.Bsize)\n\tfree := fs.Bfree * uint64(fs.Bsize) // yup, I just did that\n\tused = cap - free\n\treturn cap, used\n}","func (leaf *Node) updateContent() (err error) {\n\tif leaf.Size > MaxFileSize {\n\t\treturn nil\n\t}\n\n\tleaf.V, err = ioutil.ReadFile(leaf.SysPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}","func (fup *folderUpdatePrepper) updateResolutionUsageLockedCache(\n\tctx context.Context, lState *kbfssync.LockState, md *RootMetadata,\n\tbps blockPutState, unmergedChains, mergedChains *crChains,\n\tmostRecentMergedMD ImmutableRootMetadata,\n\trefs, unrefs map[data.BlockPointer]bool) error {\n\tmd.SetRefBytes(0)\n\tmd.SetUnrefBytes(0)\n\tmd.SetMDRefBytes(0)\n\tmd.SetDiskUsage(mostRecentMergedMD.DiskUsage())\n\tmd.SetMDDiskUsage(mostRecentMergedMD.MDDiskUsage())\n\n\tlocalBlocks := make(map[data.BlockPointer]data.Block)\n\tfor _, ptr := range bps.Ptrs() {\n\t\tif block, err := bps.GetBlock(ctx, ptr); err == nil && block != nil {\n\t\t\tlocalBlocks[ptr] = block\n\t\t}\n\t}\n\n\t// Add bytes for every ref'd block.\n\trefPtrsToFetch := make([]data.BlockPointer, 0, len(refs))\n\tvar refSum uint64\n\tfor ptr := range refs {\n\t\tif block, ok := localBlocks[ptr]; ok {\n\t\t\trefSum += uint64(block.GetEncodedSize())\n\t\t} else {\n\t\t\trefPtrsToFetch = append(refPtrsToFetch, ptr)\n\t\t}\n\t\tfup.vlog.CLogf(ctx, libkb.VLog1, \"Ref'ing block %v\", ptr)\n\t}\n\n\t// Look up the total sum of the ref blocks in parallel to get\n\t// their sizes.\n\t//\n\t// TODO: If the blocks weren't already in the cache, this call\n\t// won't cache them, so it's kind of wasting work. Furthermore,\n\t// we might be able to get the encoded size from other sources as\n\t// well (such as its directory entry or its indirect file block)\n\t// if we happened to have come across it before.\n\trefSumFetched, err := fup.blocks.GetCleanEncodedBlocksSizeSum(\n\t\tctx, lState, md.ReadOnly(), refPtrsToFetch, nil, fup.branch(), false)\n\tif err != nil {\n\t\treturn err\n\t}\n\trefSum += refSumFetched\n\n\tfup.vlog.CLogf(ctx, libkb.VLog1, \"Ref'ing a total of %d bytes\", refSum)\n\tmd.AddRefBytes(refSum)\n\tmd.AddDiskUsage(refSum)\n\n\tunrefPtrsToFetch := make([]data.BlockPointer, 0, len(unrefs))\n\tvar unrefSum uint64\n\tfor ptr := range unrefs {\n\t\toriginal, ok := unmergedChains.originals[ptr]\n\t\tif !ok {\n\t\t\toriginal = ptr\n\t\t}\n\t\tif original != ptr || unmergedChains.isCreated(original) {\n\t\t\t// Only unref pointers that weren't created as part of the\n\t\t\t// unmerged branch. Either they existed already or they\n\t\t\t// were created as part of the merged branch.\n\t\t\tcontinue\n\t\t}\n\t\t// Also make sure this wasn't already removed or overwritten\n\t\t// on the merged branch.\n\t\toriginal, ok = mergedChains.originals[ptr]\n\t\tif !ok {\n\t\t\toriginal = ptr\n\t\t}\n\t\tmergedChain, ok := mergedChains.byOriginal[original]\n\t\tif (ok && original != mergedChain.mostRecent && original == ptr) ||\n\t\t\tmergedChains.isDeleted(original) {\n\t\t\tcontinue\n\t\t}\n\n\t\tif info, ok := fup.cachedInfos[ptr]; ok {\n\t\t\tunrefSum += uint64(info.EncodedSize)\n\t\t} else {\n\t\t\tunrefPtrsToFetch = append(unrefPtrsToFetch, ptr)\n\t\t}\n\t}\n\n\t// Look up the unref blocks in parallel to get their sizes. Since\n\t// we don't know whether these are files or directories, just look\n\t// them up generically. Ignore any recoverable errors for unrefs.\n\t// Note that we can't combine these with the above ref fetches\n\t// since they require a different MD. If the merged changes\n\t// didn't change any blocks (in particular, the root block), we\n\t// can assume all the blocks we are unreferencing were live;\n\t// otherwise, we need to check with the server to make sure.\n\tonlyCountIfLive := len(mergedChains.byOriginal) != 0\n\tunrefSumFetched, err := fup.blocks.GetCleanEncodedBlocksSizeSum(\n\t\tctx, lState, mostRecentMergedMD, unrefPtrsToFetch, unrefs,\n\t\tfup.branch(), onlyCountIfLive)\n\tif err != nil {\n\t\treturn err\n\t}\n\tunrefSum += unrefSumFetched\n\n\t// Subtract bytes for every unref'd block that wasn't created in\n\t// the unmerged branch.\n\tfup.vlog.CLogf(ctx, libkb.VLog1, \"Unref'ing a total of %d bytes\", unrefSum)\n\tmd.AddUnrefBytes(unrefSum)\n\tmd.SetDiskUsage(md.DiskUsage() - unrefSum)\n\treturn nil\n}","func (b *SnailBlock) Size() common.StorageSize {\n\tif size := b.size.Load(); size != nil {\n\t\treturn size.(common.StorageSize)\n\t}\n\tc := writeCounter(0)\n\trlp.Encode(&c, b)\n\tb.size.Store(common.StorageSize(c))\n\treturn common.StorageSize(c)\n}","func SetBpfBuflen(fd, l int) (int, error) {\n\terr := ioctlPtr(fd, BIOCSBLEN, unsafe.Pointer(&l))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn l, nil\n}","func UpdateClusterMemoryCurrentBytes(memoryCount int64) {\n\tmemoryCurrentBytes.Set(float64(memoryCount))\n}","func (u *uploader) initSize() {\n\tu.totalSize = -1\n\n\tswitch r := u.in.Body.(type) {\n\tcase io.Seeker:\n\t\tn, err := aws.SeekerLen(r)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tu.totalSize = n\n\n\t\t// Try to adjust partSize if it is too small and account for\n\t\t// integer division truncation.\n\t\tif u.totalSize/u.cfg.PartSize >= int64(u.cfg.MaxUploadParts) {\n\t\t\t// Add one to the part size to account for remainders\n\t\t\t// during the size calculation. e.g odd number of bytes.\n\t\t\tu.cfg.PartSize = (u.totalSize / int64(u.cfg.MaxUploadParts)) + 1\n\t\t}\n\t}\n}","func (layout Layout) hashesPerBlock() int64 {\n\treturn layout.blockSize / layout.digestSize\n}","func (m *metricSshcheckSftpStatus) updateCapacity() {\n\tif m.data.Sum().DataPoints().Len() > m.capacity {\n\t\tm.capacity = m.data.Sum().DataPoints().Len()\n\t}\n}","func (*digest) BlockSize() int {\n\treturn int(BlockSize)\n}","func (*digest) BlockSize() int {\n\treturn int(BlockSize)\n}","func newBlockfileMgr(id string, conf *Conf, indexConfig *blkstorage.IndexConfig, indexStore *leveldbhelper.DBHandle) *blockfileMgr {\n\tlogger.Debugf(\"newBlockfileMgr() initializing file-based block storage for ledger: %s \", id)\n\tvar rwMutexs []*sync.RWMutex\n\n\t//Determine the root directory for the blockfile storage, if it does not exist create it\n\trootDir := conf.getLedgerBlockDir(id)\n\t_, err := util.CreateDirIfMissing(rootDir)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Error: %s\", err))\n\t}\n\t// Instantiate the manager, i.e. blockFileMgr structure\n\tmgr := &blockfileMgr{rootDir: rootDir, conf: conf, db: indexStore, rwMutexs: rwMutexs}\n\n\t// cp = checkpointInfo, retrieve from the database the file suffix or number of where blocks were stored.\n\t// It also retrieves the current size of that file and the last block number that was written to that file.\n\t// At init checkpointInfo:latestFileChunkSuffixNum=[0], latestFileChunksize=[0], lastBlockNumber=[0]\n\tcpInfo, err := mgr.loadCurrentInfo()\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not get block file info for current block file from db: %s\", err))\n\t}\n\tif cpInfo == nil {\n\t\tlogger.Info(`Getting block information from block storage`)\n\t\tif cpInfo, err = constructCheckpointInfoFromBlockFiles(rootDir); err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Could not build checkpoint info from block files: %s\", err))\n\t\t}\n\t\tlogger.Debugf(\"Info constructed by scanning the blocks dir = %s\", spew.Sdump(cpInfo))\n\t} else {\n\t\tlogger.Debug(`Synching block information from block storage (if needed)`)\n\t\tsyncCPInfoFromFS(rootDir, cpInfo)\n\t}\n\terr = mgr.saveCurrentInfo(cpInfo, true)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not save next block file info to db: %s\", err))\n\t}\n\n\tmgr.oldestFileChunkSuffixNum = syncOldestFileNum(rootDir)\n\t//If start up is a restart of an existing storage,new the rwMutex for the files\n\tif conf.dumpConf.Enabled {\n\t\tfor i := 0; i <= cpInfo.latestFileChunkSuffixNum; i++ {\n\t\t\trwMutex := new(sync.RWMutex)\n\t\t\tmgr.rwMutexs = append(mgr.rwMutexs, rwMutex)\n\t\t}\n\t}\n\tmgr.dumpMutex = new(sync.Mutex)\n\n\t//Open a writer to the file identified by the number and truncate it to only contain the latest block\n\t// that was completely saved (file system, index, cpinfo, etc)\n\tcurrentFileWriter, err := newBlockfileWriter(deriveBlockfilePath(rootDir, cpInfo.latestFileChunkSuffixNum))\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not open writer to current file: %s\", err))\n\t}\n\t//Truncate the file to remove excess past last block\n\terr = currentFileWriter.truncateFile(cpInfo.latestFileChunksize)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not truncate current file to known size in db: %s\", err))\n\t}\n\n\t// Create a new KeyValue store database handler for the blocks index in the keyvalue database\n\tmgr.index = newBlockIndex(indexConfig, indexStore)\n\n\t// Update the manager with the checkpoint info and the file writer\n\tmgr.cpInfo = cpInfo\n\tmgr.currentFileWriter = currentFileWriter\n\t// Create a checkpoint condition (event) variable, for the goroutine waiting for\n\t// or announcing the occurrence of an event.\n\tmgr.cpInfoCond = sync.NewCond(&sync.Mutex{})\n\n\t// init BlockchainInfo for external API's\n\tbcInfo := &common.BlockchainInfo{\n\t\tHeight: 0,\n\t\tCurrentBlockHash: nil,\n\t\tPreviousBlockHash: nil}\n\n\tif !cpInfo.isChainEmpty {\n\t\t//If start up is a restart of an existing storage, sync the index from block storage and update BlockchainInfo for external API's\n\t\tmgr.syncIndex()\n\t\tlastBlockHeader, err := mgr.retrieveBlockHeaderByNumber(cpInfo.lastBlockNumber)\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Could not retrieve header of the last block form file: %s\", err))\n\t\t}\n\t\tlastBlockHash := lastBlockHeader.Hash()\n\t\tpreviousBlockHash := lastBlockHeader.PreviousHash\n\t\tbcInfo = &common.BlockchainInfo{\n\t\t\tHeight: cpInfo.lastBlockNumber + 1,\n\t\t\tCurrentBlockHash: lastBlockHash,\n\t\t\tPreviousBlockHash: previousBlockHash}\n\t}\n\tmgr.bcInfo.Store(bcInfo)\n\treturn mgr\n}","func DirSizeByte(path string) uint64 {\n\tdirSize = 0\n\tfilepath.Walk(path, readSize)\n\treturn dirSize\n}","func calculateBufferSize(blocks uint64) uint64 {\n\tif nbb := NBufferBlocks; blocks < nbb {\n\t\treturn blocks\n\t} else {\n\t\treturn nbb\n\t}\n}","func (b *BlockSplitterSimple) SetMaxDirEntriesByBlockSize(\n\tcodec kbfscodec.Codec) error {\n\tdirEnv := os.Getenv(\"KEYBASE_BSPLIT_MAX_DIR_ENTRIES\")\n\tif len(dirEnv) > 0 {\n\t\t// Don't override the environment variable.\n\t\treturn nil\n\t}\n\n\tblock := NewDirBlock().(*DirBlock)\n\tbigName := strings.Repeat(\"a\", MaxNameBytesDefault)\n\t// Make \"typical\" DirEntry, though the max dir entry is a bit\n\t// bigger than this (can contain a variable-length symlink path,\n\t// for example).\n\tde := DirEntry{\n\t\tBlockInfo: BlockInfo{\n\t\t\tBlockPointer: BlockPointer{\n\t\t\t\tDirectType: DirectBlock,\n\t\t\t},\n\t\t},\n\t\tEntryInfo: EntryInfo{\n\t\t\tPrevRevisions: PrevRevisions{\n\t\t\t\t{Revision: 0, Count: 0},\n\t\t\t\t{Revision: 1, Count: 1},\n\t\t\t\t{Revision: 2, Count: 2},\n\t\t\t\t{Revision: 3, Count: 3},\n\t\t\t\t{Revision: 4, Count: 4},\n\t\t\t},\n\t\t},\n\t}\n\tblock.Children[bigName] = de\n\tencodedBlock, err := codec.Encode(block)\n\tif err != nil {\n\t\treturn err\n\t}\n\toneEntrySize := int64(len(encodedBlock))\n\tb.maxDirEntriesPerBlock = int(b.maxSize / oneEntrySize)\n\tif b.maxDirEntriesPerBlock == 0 {\n\t\tb.maxDirEntriesPerBlock = 1\n\t}\n\treturn nil\n}","func (bdi *Info) CountBlocks() int {\n\treturn int(C.spdk_bdev_get_num_blocks(bdi.ptr()))\n}","func updateTxSize(tx *transaction.Transaction) (*transaction.Transaction, error) {\n\tbw := io.NewBufBinWriter()\n\ttx.EncodeBinary(bw.BinWriter)\n\tif bw.Err != nil {\n\t\treturn nil, fmt.Errorf(\"encode binary: %w\", bw.Err)\n\t}\n\treturn transaction.NewTransactionFromBytes(tx.Bytes())\n}","func (c *Container) rootFsSize() (int64, error) {\n\tcontainer, err := c.runtime.store.Container(c.ID())\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Ignore the size of the top layer. The top layer is a mutable RW layer\n\t// and is not considered a part of the rootfs\n\trwLayer, err := c.runtime.store.Layer(container.LayerID)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tlayer, err := c.runtime.store.Layer(rwLayer.Parent)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tsize := int64(0)\n\tfor layer.Parent != \"\" {\n\t\tlayerSize, err := c.runtime.store.DiffSize(layer.Parent, layer.ID)\n\t\tif err != nil {\n\t\t\treturn size, errors.Wrapf(err, \"getting diffsize of layer %q and its parent %q\", layer.ID, layer.Parent)\n\t\t}\n\t\tsize += layerSize\n\t\tlayer, err = c.runtime.store.Layer(layer.Parent)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\t// Get the size of the last layer. Has to be outside of the loop\n\t// because the parent of the last layer is \"\", andlstore.Get(\"\")\n\t// will return an error.\n\tlayerSize, err := c.runtime.store.DiffSize(layer.Parent, layer.ID)\n\treturn size + layerSize, err\n}","func (o KubernetesClusterNodePoolLinuxOsConfigOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfig) *int { return v.SwapFileSizeMb }).(pulumi.IntPtrOutput)\n}","func (req *DronaRequest) updateOsize(size int64) {\n\treq.Lock()\n\tdefer req.Unlock()\n\treq.objectSize = size\n}","func statfsImpl(t *kernel.Task, d *fs.Dirent, addr usermem.Addr) error {\n\tinfo, err := d.Inode.StatFS(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Construct the statfs structure and copy it out.\n\tstatfs := linux.Statfs{\n\t\tType: info.Type,\n\t\t// Treat block size and fragment size as the same, as\n\t\t// most consumers of this structure will expect one\n\t\t// or the other to be filled in.\n\t\tBlockSize: d.Inode.StableAttr.BlockSize,\n\t\tBlocks: info.TotalBlocks,\n\t\t// We don't have the concept of reserved blocks, so\n\t\t// report blocks free the same as available blocks.\n\t\t// This is a normal thing for filesystems, to do, see\n\t\t// udf, hugetlbfs, tmpfs, among others.\n\t\tBlocksFree: info.FreeBlocks,\n\t\tBlocksAvailable: info.FreeBlocks,\n\t\tFiles: info.TotalFiles,\n\t\tFilesFree: info.FreeFiles,\n\t\t// Same as Linux for simple_statfs, see fs/libfs.c.\n\t\tNameLength: linux.NAME_MAX,\n\t\tFragmentSize: d.Inode.StableAttr.BlockSize,\n\t\t// Leave other fields 0 like simple_statfs does.\n\t}\n\t_, err = t.CopyOut(addr, &statfs)\n\treturn err\n}","func (c Temperature) UpdateBlock(b *i3barjson.Block) {\n\tb.Color = c.Color\n\tfullTextFmt := fmt.Sprintf(\"%s%%s\", c.Label)\n\ttotalTemp := 0\n\tprocs := 0\n\tsysFileDirList, err := ioutil.ReadDir(c.CpuTempPath)\n\tif err != nil {\n\t\tb.Urgent = true\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\treturn\n\t}\n\tif len(sysFileDirList) != 1 {\n\t\tb.Urgent = true\n\t\tmsg := fmt.Sprintf(\n\t\t\t\"in %s, expected 1 file, got %d\",\n\t\t\tc.CpuTempPath,\n\t\t\tlen(sysFileDirList),\n\t\t)\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, msg)\n\t\treturn\n\t}\n\tsysFileDirPath := fmt.Sprintf(\n\t\t\"%s/%s\",\n\t\tc.CpuTempPath,\n\t\tsysFileDirList[0].Name(),\n\t)\n\tsysFileNameFmt := fmt.Sprintf(\"%s/%%s\", sysFileDirPath)\n\tsysFiles, err := ioutil.ReadDir(sysFileDirPath)\n\tif err != nil {\n\t\tb.Urgent = true\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\treturn\n\t}\n\tfor _, sysFile := range sysFiles {\n\t\tsysFileName := sysFile.Name()\n\t\tif !strings.HasSuffix(sysFileName, \"input\") {\n\t\t\tcontinue\n\t\t}\n\t\tr, err := os.Open(fmt.Sprintf(sysFileNameFmt, sysFileName))\n\t\tif err != nil {\n\t\t\tb.Urgent = true\n\t\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\t\treturn\n\t\t}\n\t\tdefer r.Close()\n\t\tvar temp int\n\t\t_, err = fmt.Fscanf(r, \"%d\", &temp)\n\t\tif err != nil {\n\t\t\tb.Urgent = true\n\t\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\t\treturn\n\t\t}\n\t\ttotalTemp += temp\n\t\tprocs++\n\t}\n\tavgTemp := float64(totalTemp) / float64(procs*1000)\n\tif avgTemp >= c.CritTemp {\n\t\tb.Urgent = true\n\t} else {\n\t\tb.Urgent = false\n\t}\n\tb.FullText = fmt.Sprintf(\"%s%.2f°C\", c.Label, avgTemp)\n}","func updateRemainingSpace(ctx *downloaderContext) {\n\n\tctx.globalStatus.RemainingSpace = ctx.globalConfig.MaxSpace -\n\t\tctx.globalStatus.UsedSpace - ctx.globalStatus.ReservedSpace\n\n\tlog.Infof(\"RemainingSpace %d, maxspace %d, usedspace %d, reserved %d\\n\",\n\t\tctx.globalStatus.RemainingSpace, ctx.globalConfig.MaxSpace,\n\t\tctx.globalStatus.UsedSpace, ctx.globalStatus.ReservedSpace)\n}","func (f *FileSystem) Statfs(ctx context.Context, req *fuse.StatfsRequest, resp *fuse.StatfsResponse) error {\n\t// each block size is 4096 bytes by default.\n\tconst unit = uint64(4096)\n\n\tresp.Bsize = uint32(unit)\n\tresp.Blocks = uint64(f.account.Disk.Size) / unit\n\tresp.Bavail = uint64(f.account.Disk.Avail) / unit\n\tresp.Bfree = uint64(f.account.Disk.Avail) / unit\n\n\treturn nil\n}","func (lf *ListFile) NumBytes() int64 {\n\t// NOTE: here we don't use IsClosed() because\n\t// it uses the mutex; Size() is used in noMutexIterateLines\n\t// which is called after another mutex is locked,\n\t// making IsClosed() wait forever for the mutex unlock.\n\tif lf.isClosed {\n\t\treturn 0\n\t}\n\n\terr := lf.file.Sync()\n\tif err != nil {\n\t\t// TODO: not panic??\n\t\tpanic(err)\n\t}\n\n\tinfo, err := lf.file.Stat()\n\tif err != nil {\n\t\t// TODO: not panic??\n\t\tpanic(err)\n\t}\n\n\treturn info.Size()\n}","func (b *Block) Size() metric.StorageSize {\n\tif cached := b.cache.size.Load(); cached != nil {\n\t\treturn cached.(metric.StorageSize)\n\t}\n\tvar size metric.StorageSize\n\trlp.Encode(&size, b)\n\tb.cache.size.Store(size)\n\treturn size\n}","func (pb *PutBlock) Height() uint64 { return pb.height }","func (n *nodeBlock) sync() (err error) {\n\tmm := false\n\tif n.offset == 0 {\n\t\tif n._super.mmapSizeUsed+nodeBlockSize <= n._super.mmapSize {\n\t\t\tn.offset = int64(n._super.mmapSizeUsed)\n\t\t\tn._super.mmapSizeUsed += nodeBlockSize\n\t\t\tmm = true\n\t\t} else {\n\t\t\tn.offset, err = n._super._fd.Seek(0, os.SEEK_END)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\n\tx := *(*[nodeBlockSize]byte)(unsafe.Pointer(n))\n\n\tif mm {\n\t\tcopy(n._super._mmap[n.offset:], x[:])\n\t} else {\n\t\t_, err = n._super._fd.Seek(n.offset, 0)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\t_, err = n._super._fd.Write(x[:])\n\t}\n\n\tif err == nil {\n\t\tn._super._snapshotChPending[n] = x\n\t}\n\treturn\n}","func (fm *FinalModelStructBytes) FBEAllocationSize(fbeValue *StructBytes) int {\n fbeResult := 0 +\n fm.F1.FBEAllocationSize(fbeValue.F1) +\n fm.F2.FBEAllocationSize(fbeValue.F2) +\n fm.F3.FBEAllocationSize(fbeValue.F3) +\n 0\n return fbeResult\n}","func (d *Directory) updateDirectorySize(p string) {\n\tpathParams := strings.Split(p, \"/\")\n\td.Size = 0\n\td.lock()\n\tdefer d.unlock()\n\tfor i := 0; i < len(d.INodes); i++ {\n\t\tswitch d.INodes[i].(type) {\n\t\tcase *Directory:\n\t\t\tif d.INodes[i].GetName() == pathParams[1] {\n\t\t\t\tsubPath := strings.Join(pathParams[2:], \"/\")\n\t\t\t\tsubPath = \"/\" + subPath\n\t\t\t\td.INodes[i].(*Directory).updateDirectorySize(subPath)\n\t\t\t}\n\t\t\td.Size += d.INodes[i].GetSize()\n\t\tcase *File:\n\t\t\td.Size += d.INodes[i].GetSize()\n\t\t}\n\t}\n}","func (cs *ConsensusState) updateHeight(height int64) {\n\tcs.Height = height\n}","func (g *testGenerator) updateBlockState(oldBlockName string, oldBlockHash chainhash.Hash, newBlockName string, newBlock *wire.MsgBlock) {\n\t// Remove existing entries.\n\tdelete(g.blocks, oldBlockHash)\n\tdelete(g.blocksByName, oldBlockName)\n\n\t// Add new entries.\n\tnewBlockHash := newBlock.BlockHash()\n\tg.blocks[newBlockHash] = newBlock\n\tg.blocksByName[newBlockName] = newBlock\n}","func (b *Buffer) update() {\n\tb.NumLines = len(b.lines)\n}","func (o KubernetesClusterDefaultNodePoolLinuxOsConfigOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfig) *int { return v.SwapFileSizeMb }).(pulumi.IntPtrOutput)\n}","func (p *pbkdf2) getHashLengthBitBlockCount() int {\n\treturn int(math.Ceil(float64(p.size) / float64(p.pseudoRandomFunction.Size())))\n}","func (x *ecbEncrypter) BlockSize() int { return x.blockSize }","func (trd *trxDispatcher) updateLastSeenBlock() {\n\t// get the current value\n\tlsb := trd.blkObserver.Load()\n\tlog.Noticef(\"last seen block is #%d\", lsb)\n\n\t// make the change in the database so the progress persists\n\terr := repo.UpdateLastKnownBlock((*hexutil.Uint64)(&lsb))\n\tif err != nil {\n\t\tlog.Errorf(\"could not update last seen block; %s\", err.Error())\n\t}\n}","func UpdateCgroupDeviceWeight(pid, innerPath, value string) error {\n\tif pid == \"0\" {\n\t\treturn nil\n\t}\n\n\tcgroupPath, err := FindCgroupPath(pid, \"blkio\", innerPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tpath := filepath.Join(cgroupPath, \"blkio.weight_device\")\n\tif err := ioutil.WriteFile(path, []byte(value), 0600); err != nil {\n\t\treturn fmt.Errorf(\"%s, please check whether current OS support blkio weight device configuration for bfq scheduler\", err)\n\t}\n\treturn nil\n}","func (req *DronaRequest) updateAsize(size int64) {\n\treq.Lock()\n\tdefer req.Unlock()\n\treq.asize = size\n}","func (fileSystem *FileSystem) Put(fileName string) {\n\tif (*fileSystem).PfsFile == nil {\n\t\tfmt.Println(\"You must open pfs first\")\n\t\treturn\n\t}\n\n\tdata, err := ioutil.ReadFile(fileName)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n\n\tif len(data) > totalDataBlocks*dataBlockSize {\n\t\tfmt.Println(\"File is too big to be inserted in pfs\")\n\t\treturn\n\t}\n\n\tfileSize := uint16(len(data)) // Get the file size\n\n\t// Finding how many blocks is easy enough\n\tnecesaryBlocks := uint16(math.Ceil(float64(fileSize) / float64(dataBlockSize)))\n\tfoundBlocks := uint16(0)\n\tlocation := -1\n\tfor idx, availability := range (*fileSystem).Directory.FreeDataBlockArray {\n\t\tif foundBlocks > 0 && !availability {\n\t\t\tfoundBlocks = 0\n\t\t\tcontinue\n\t\t}\n\n\t\tif availability {\n\t\t\tfoundBlocks++\n\t\t}\n\t\tif foundBlocks == necesaryBlocks {\n\t\t\tlocation = idx\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif location == -1 {\n\t\tfmt.Println(\"Not enough free blocks to store file.\")\n\t\treturn\n\t}\n\n\t// We have found the location of data that should be written\n\tblockID := location - int(necesaryBlocks) + 1\n\toffset := dataAddress + blockID*dataBlockSize\n\t(*fileSystem).PfsFile.Seek(int64(offset), 0)\n\t(*fileSystem).PfsFile.Write(data)\n\n\t// Now we need to find a FCB to store the records of the value\n\tfor idx, fcb := range (*fileSystem).Directory.FCBArray {\n\t\tif !fcb.ContainsValidData {\n\t\t\tblock, _ := NewFCB(fileName, fileSize, uint8(blockID))\n\t\t\t(*fileSystem).Directory.FCBArray[idx] = block\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// Now we also need to set the free data blocks as ocupied\n\tfor i := 0; i < int(necesaryBlocks); i++ {\n\t\t(*fileSystem).Directory.FreeDataBlockArray[blockID+i] = false\n\t}\n\n\t// We know update the directory\n\t(*fileSystem).Directory.Metadata.NumberOfFilesStored++\n}","func (o KubernetesClusterNodePoolLinuxOsConfigPtrOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SwapFileSizeMb\n\t}).(pulumi.IntPtrOutput)\n}","func UpdateDumpFile(fileInode inode, data string, fileName string, parentInode inode, parentInodeNum int) error {\n\t// open dump file\n\tfile, err := os.OpenFile(DumpFile, os.O_WRONLY, os.ModeAppend)\n\tif err != nil {\n\t\tfmt.Println(\"Error in opening file while loading the filesystem: \", err)\n\t\tos.Exit(1)\n\t}\n\tdefer file.Close()\n\n\toffset := OffsetInodeTable + fileInode.inodeNum\n\t// update inode bitmap\n\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(strconv.Itoa(SetBit))); err != nil {\n\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\treturn err\n\t}\n\n\t// update inode block\n\toffset = OffsetInodeBlock + (fileInode.inodeNum * InodeBlockSize)\n\tinodeInfo := frameInodeString(fileInode)\n\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(inodeInfo)); err != nil {\n\t\tfmt.Println(\"Error: unable to update inode block\", err)\n\t\treturn err\n\t}\n\n\t//update parent inode block\n\tif parentInodeNum != -1 { // skipping it for root directory\n\t\toffset = OffsetInodeBlock + (parentInodeNum * InodeBlockSize)\n\t\tinodeInfo = frameInodeString(parentInode)\n\t\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(inodeInfo)); err != nil {\n\t\t\tfmt.Println(\"Error: unable to update parent inode block\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// update data bitmap and fill the data block as well ...\n\toffset = OffsetDataTable\n\tlistByte := []byte(data)\n\tcount := 0\n\tfor _, val := range fileInode.dataList {\n\t\t// update inode bitmap\n\t\tif err := disklib.WriteBlockOnDisk(file, offset+val, []byte(strconv.Itoa(SetBit))); err != nil {\n\t\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t// update data block in disk\n\t\tif err := disklib.WriteBlockOnDisk(file, OffsetDataBlock+(val*DataBlockSize), []byte(listByte[count:count+DataBlockSize])); err != nil {\n\t\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\t\treturn err\n\t\t}\n\t\tcount += DataBlockSize\n\t}\n\n\t// put fst in disk\n\toffset = OffsetFST + (fileInode.inodeNum * CellFSTSize)\n\tstr := fmt.Sprintf(\"%3d\", fileInode.inodeNum) + fmt.Sprintf(\"%10s\", fileName)\n\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(str)); err != nil {\n\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\treturn err\n\t}\n\treturn nil\n\n}","func DirSizeB(path string) (int64, error) {\n\tvar size int64\n\terr := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error {\n\t\tif !info.IsDir() {\n\t\t\tsize += info.Size()\n\t\t}\n\t\treturn err\n\t})\n\treturn size, err\n}","func UpdateCapacity(cap int) {\n\ttype requestInfo struct {\n\t\tNode string `json:\"node\" binding:\"required\"`\n\t\tCapacity int `json:\"capability\" binding:\"required\"`\n\t}\n\tgo func() {\n\t\tlog.InfoF(\"node %s capacity updated to %d\", conf.Get(\"nodeLanHost\").(string), cap)\n\t\turl := fmt.Sprintf(\"http://%s/nodeUpdateCapacity\", conf.Get(\"registerCenterHost\"))\n\t\t_, _, errs := gorequest.New().Post(url).Send(requestInfo{Node: conf.Get(\"nodeLanHost\").(string), Capacity: cap}).End()\n\t\tif errs != nil && len(errs) > 0 {\n\t\t\tlog.InfoF(\"requestGroupDispatchLoop error:%s\", errs[0])\n\t\t}\n\t}()\n}","func (fs *FS) fsInfo(ctx context.Context, path string) (int64, int64, int64, int64, int64, int64, error) {\n\tstatfs := &unix.Statfs_t{}\n\terr := unix.Statfs(path, statfs)\n\tif err != nil {\n\t\treturn 0, 0, 0, 0, 0, 0, err\n\t}\n\n\t// Available is blocks available * fragment size\n\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\n\n\t// Capacity is total block count * fragment size\n\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\n\n\t// Usage is block being used * fragment size (aka block size).\n\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\n\n\tinodes := int64(statfs.Files)\n\tinodesFree := int64(statfs.Ffree)\n\tinodesUsed := inodes - inodesFree\n\n\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\n}","func BlockSize() int {\n\treturn config.Record.BlockSize\n}","func (b *Block) Size() int {\n\tpbb, err := b.ToProto()\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn pbb.Size()\n}","func (b *Block) Size() int {\n\tpbb, err := b.ToProto()\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn pbb.Size()\n}","func (s *hashBLAKE2s) BlockLen() int {\n\treturn blake2s.BlockSize\n}","func (x *ecbDecrypter) BlockSize() int { return x.blockSize }","func vmUpdateAdditionalBandwidth(db *Database, vm *VirtualMachine) {\n\t// determine how much of the plan bandwidth to add to the user's bandwidth pool for current month\n\tnow := time.Now()\n\tmonthStart := time.Date(now.Year(), now.Month(), 1, 0, 0, 0, 0, time.UTC)\n\tmonthEnd := monthStart.AddDate(0, 1, 0)\n\tvar factor float64\n\n\tif vm.CreatedTime.Before(monthStart) {\n\t\tfactor = float64(now.Sub(monthStart))\n\t} else {\n\t\tfactor = float64(now.Sub(vm.CreatedTime))\n\t}\n\n\tfactor /= float64(monthEnd.Sub(monthStart))\n\tif factor > 1 {\n\t\tfactor = 1\n\t}\n\n\tadditionalBandwidth := int64((factor * float64(vm.Plan.Bandwidth) + 15) * 1024 * 1024 * 1024)\n\trows := db.Query(\"SELECT id FROM region_bandwidth WHERE region = ? AND user_id = ?\", vm.Region, vm.UserId)\n\tif rows.Next() {\n\t\tvar rowId int\n\t\trows.Scan(&rowId)\n\t\trows.Close()\n\t\tdb.Exec(\"UPDATE region_bandwidth SET bandwidth_additional = bandwidth_additional + ? WHERE id = ?\", additionalBandwidth, rowId)\n\t} else {\n\t\tdb.Exec(\"INSERT INTO region_bandwidth (user_id, region, bandwidth_additional) VALUES (?, ?, ?)\", vm.UserId, vm.Region, additionalBandwidth)\n\t}\n}","func (b *B) SetBytes(n int64) {}","func (n *TreeNode) AddSize(by uint64) {\n\tn.mutex.RLock()\n\tn.size += by\n\tn.mutex.RUnlock()\n\treturn\n}","func (cipher *Simon96Cipher) BlockSize() int {\n\treturn 12\n}","func (fm FinalModelTimestamp) FBEAllocationSize(value time.Time) int { return fm.FBESize() }","func (db *DB) grow(sz int) error {\n\t// Ignore if the new size is less than available file size.\n\tif sz <= db.filesz {\n\t\treturn nil\n\t}\n\n\t// If the data is smaller than the alloc size then only allocate what's needed.\n\t// Once it goes over the allocation size then allocate in chunks.\n\tif db.datasz <= db.AllocSize {\n\t\tsz = db.datasz\n\t} else {\n\t\tsz += db.AllocSize\n\t}\n\n\t// Truncate and fsync to ensure file size metadata is flushed.\n\t// https://github.com/boltdb/bolt/issues/284\n\tif !db.NoGrowSync && !db.readOnly {\n\t\tif runtime.GOOS != \"windows\" {\n\t\t\tif err := db.file.Truncate(int64(sz)); err != nil {\n\t\t\t\treturn fmt.Errorf(\"file resize error: %s\", err)\n\t\t\t}\n\t\t}\n\t\tif err := db.file.Sync(); err != nil {\n\t\t\treturn fmt.Errorf(\"file sync error: %s\", err)\n\t\t}\n\t\tif db.Mlock {\n\t\t\t// unlock old file and lock new one\n\t\t\tif err := db.mrelock(db.filesz, sz); err != nil {\n\t\t\t\treturn fmt.Errorf(\"mlock/munlock error: %s\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\tdb.filesz = sz\n\treturn nil\n}","func (b *BaseNode) updateBytes(n Node) {\n\tbw := io.NewBufBinWriter()\n\tbw.Grow(1 + n.Size())\n\tencodeNodeWithType(n, bw.BinWriter)\n\tb.bytes = bw.Bytes()\n\tb.bytesValid = true\n}","func (b *bpfHandle) SetReadBufferSize(size int) error {\n\tif ok, err := b.ioctlWithInt(ioctlBIOCSBLEN, size); !ok {\n\t\treturn err\n\t}\n\n\tnumData := make([]byte, ioctlIntegerSize)\n\tb.ioctlWithData(ioctlBIOCGBLEN, numData)\n\tif binary.LittleEndian.Uint32(numData) < uint32(size) {\n\t\treturn errors.New(\"unsupported buffer size\")\n\t}\n\n\tb.readBuffer = make([]byte, size)\n\treturn nil\n}","func (c *BlockCache) Add(height int, block *walletrpc.CompactBlock) error {\n\t// Invariant: m[firstBlock..nextBlock) are valid.\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tif height > c.nextBlock {\n\t\t// Cache has been reset (for example, checksum error)\n\t\treturn nil\n\t}\n\tif height < c.firstBlock {\n\t\t// Should never try to add a block before Sapling activation height\n\t\tLog.Fatal(\"cache.Add height below Sapling: \", height)\n\t\treturn nil\n\t}\n\tif height < c.nextBlock {\n\t\t// Should never try to \"backup\" (call Reorg() instead).\n\t\tLog.Fatal(\"cache.Add height going backwards: \", height)\n\t\treturn nil\n\t}\n\tbheight := int(block.Height)\n\n\tif bheight != height {\n\t\t// This could only happen if zcashd returned the wrong\n\t\t// block (not the height we requested).\n\t\tLog.Fatal(\"cache.Add wrong height: \", bheight, \" expecting: \", height)\n\t\treturn nil\n\t}\n\n\t// Add the new block and its length to the db files.\n\tdata, err := proto.Marshal(block)\n\tif err != nil {\n\t\treturn err\n\t}\n\tb := append(checksum(height, data), data...)\n\tn, err := c.blocksFile.Write(b)\n\tif err != nil {\n\t\tLog.Fatal(\"blocks write failed: \", err)\n\t}\n\tif n != len(b) {\n\t\tLog.Fatal(\"blocks write incorrect length: expected: \", len(b), \"written: \", n)\n\t}\n\tb = make([]byte, 4)\n\tbinary.LittleEndian.PutUint32(b, uint32(len(data)))\n\tn, err = c.lengthsFile.Write(b)\n\tif err != nil {\n\t\tLog.Fatal(\"lengths write failed: \", err)\n\t}\n\tif n != len(b) {\n\t\tLog.Fatal(\"lengths write incorrect length: expected: \", len(b), \"written: \", n)\n\t}\n\n\t// update the in-memory variables\n\toffset := c.starts[len(c.starts)-1]\n\tc.starts = append(c.starts, offset+int64(len(data)+8))\n\n\tif c.latestHash == nil {\n\t\tc.latestHash = make([]byte, len(block.Hash))\n\t}\n\tcopy(c.latestHash, block.Hash)\n\tc.nextBlock++\n\t// Invariant: m[firstBlock..nextBlock) are valid.\n\treturn nil\n}","func GetFSInfo(path string) (total, available int) {\n\ttotal = -1\n\tavailable = -1\n\tvar buf syscall.Statfs_t\n\n\tif syscall.Statfs(path, &buf) != nil {\n\t\treturn\n\t}\n\n\tif buf.Bsize <= 0 {\n\t\treturn\n\t}\n\n\ttotal = int((uint64(buf.Bsize) * buf.Blocks) / (1000 * 1000))\n\tavailable = int((uint64(buf.Bsize) * buf.Bavail) / (1000 * 1000))\n\n\treturn\n}","func (p *Peer) UpdateLastBlockHeight(newHeight int32) {\n\tp.statsMtx.Lock()\n\tif newHeight <= p.lastBlock {\n\t\tp.statsMtx.Unlock()\n\t\treturn\n\t}\n\tlog.Tracef(\"Updating last block height of peer %v from %v to %v\",\n\t\tp.addr, p.lastBlock, newHeight)\n\tp.lastBlock = newHeight\n\tp.statsMtx.Unlock()\n}","func (fm FinalModelEnumUInt32) FBEAllocationSize(value EnumUInt32) int { return fm.FBESize() }","func convertSize(size uint64) C.ulonglong {\n\treturn C.ulonglong(size * 2) // kB to 512-byte sectors\n}","func (mio *Mio) getOptimalBlockSz(bufSz int) (bsz, gsz int) {\n if bufSz > maxM0BufSz {\n bufSz = maxM0BufSz\n }\n pver := C.m0_pool_version_find(&C.instance.m0c_pools_common,\n &mio.obj.ob_attr.oa_pver)\n if pver == nil {\n log.Panic(\"cannot find the object's pool version\")\n }\n pa := &pver.pv_attr\n if pa.pa_P < pa.pa_N + pa.pa_K + pa.pa_S {\n log.Panic(\"pool width (%v) is less than the parity group size\" +\n \" (%v + %v + %v == %v), check pool parity configuration\",\n pa.pa_P, pa.pa_N, pa.pa_K, pa.pa_S,\n pa.pa_N + pa.pa_K + pa.pa_S)\n }\n\n usz := int(C.m0_obj_layout_id_to_unit_size(mio.objLid))\n gsz = usz * int(pa.pa_N) // group size in data units only\n\n // bs should be max 4-times pool-width deep counting by 1MB units, or\n // 8-times deep counting by 512K units, 16-times deep by 256K units,\n // and so on. Several units to one target will be aggregated to make\n // fewer network RPCs, disk i/o operations and BE transactions.\n // For unit sizes of 32K or less, the koefficient (k) is 128, which\n // makes it 32K * 128 == 4MB - the maximum amount per target when\n // the performance is still good on LNet (which has max 1MB frames).\n // XXX: it may be different on libfabric, should be re-measured.\n k := C.uint(128 / ((usz + 0x7fff) / 0x8000))\n if k == 0 {\n k = 1\n }\n // P * N / (N + K + S) - number of data units to span the pool-width\n maxBs := int(k * C.uint(usz) * pa.pa_P * pa.pa_N /\n (pa.pa_N + pa.pa_K + pa.pa_S))\n maxBs = roundup(maxBs, gsz) // multiple of group size\n\n if bufSz >= maxBs {\n return maxBs, gsz\n } else if bufSz <= gsz {\n return gsz, gsz\n } else {\n return roundup(bufSz, gsz), gsz\n }\n}","func (constr Construction) BlockSize() int { return 16 }","func (o KubernetesClusterDefaultNodePoolLinuxOsConfigPtrOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SwapFileSizeMb\n\t}).(pulumi.IntPtrOutput)\n}","func (b *Block) Size() int {\n\treturn len(b.wire)\n}","func (devices *DeviceSet) ResizePool(size int64) error {\n\tdirname := devices.loopbackDir()\n\tdatafilename := path.Join(dirname, \"data\")\n\tif len(devices.dataDevice) > 0 {\n\t\tdatafilename = devices.dataDevice\n\t}\n\tmetadatafilename := path.Join(dirname, \"metadata\")\n\tif len(devices.metadataDevice) > 0 {\n\t\tmetadatafilename = devices.metadataDevice\n\t}\n\n\tdatafile, err := os.OpenFile(datafilename, os.O_RDWR, 0)\n\tif datafile == nil {\n\t\treturn err\n\t}\n\tdefer datafile.Close()\n\n\tfi, err := datafile.Stat()\n\tif fi == nil {\n\t\treturn err\n\t}\n\n\tif fi.Size() > size {\n\t\treturn fmt.Errorf(\"devmapper: Can't shrink file\")\n\t}\n\n\tdataloopback := loopback.FindLoopDeviceFor(datafile)\n\tif dataloopback == nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to find loopback mount for: %s\", datafilename)\n\t}\n\tdefer dataloopback.Close()\n\n\tmetadatafile, err := os.OpenFile(metadatafilename, os.O_RDWR, 0)\n\tif metadatafile == nil {\n\t\treturn err\n\t}\n\tdefer metadatafile.Close()\n\n\tmetadataloopback := loopback.FindLoopDeviceFor(metadatafile)\n\tif metadataloopback == nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to find loopback mount for: %s\", metadatafilename)\n\t}\n\tdefer metadataloopback.Close()\n\n\t// Grow loopback file\n\tif err := datafile.Truncate(size); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to grow loopback file: %s\", err)\n\t}\n\n\t// Reload size for loopback device\n\tif err := loopback.SetCapacity(dataloopback); err != nil {\n\t\treturn fmt.Errorf(\"unable to update loopback capacity: %s\", err)\n\t}\n\n\t// Suspend the pool\n\tif err := devicemapper.SuspendDevice(devices.getPoolName()); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to suspend pool: %s\", err)\n\t}\n\n\t// Reload with the new block sizes\n\tif err := devicemapper.ReloadPool(devices.getPoolName(), dataloopback, metadataloopback, devices.thinpBlockSize); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to reload pool: %s\", err)\n\t}\n\n\t// Resume the pool\n\tif err := devicemapper.ResumeDevice(devices.getPoolName()); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to resume pool: %s\", err)\n\t}\n\n\treturn nil\n}"],"string":"[\n \"func (sm3 *SM3) BlockSize() int { return 64 }\",\n \"func (s *State) BlockSize() int { return 2 }\",\n \"func (s *State) BlockSize() int { return 8 }\",\n \"func (d *state) BlockSize() int { return d.rate }\",\n \"func (s *State) BlockSize() int { return 1 }\",\n \"func (h *Hash) BlockSize() int { return len(h.buf) }\",\n \"func (d *digest) BlockSize() int { return 1 }\",\n \"func (d *RabinKarp64) BlockSize() int { return 1 }\",\n \"func (bdi *Info) BlockSize() int {\\n\\treturn int(C.spdk_bdev_get_block_size(bdi.ptr()))\\n}\",\n \"func (c Zfs) UpdateBlock(b *i3barjson.Block) {\\n\\tb.Color = c.Color\\n\\tfullTextFmt := fmt.Sprintf(\\\"%s%%s\\\", c.Label)\\n\\n\\tzpoolCmd := exec.Command(\\\"sudo\\\", \\\"zpool\\\", \\\"status\\\", c.PoolName)\\n\\tout, err := zpoolCmd.Output()\\n\\n\\tif err != nil {\\n\\t\\tb.Urgent = true\\n\\t\\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\\n\\t\\treturn\\n\\t}\\n\\n\\tzpoolLines := strings.Split(string(out), \\\"\\\\n\\\")\\n\\tfor _, zpoolLine := range zpoolLines {\\n\\t\\tline := strings.TrimSpace(zpoolLine)\\n\\t\\tif strings.HasPrefix(line, \\\"state\\\") {\\n\\t\\t\\tsplit := strings.Split(line, \\\":\\\")\\n\\t\\t\\tstatus := strings.TrimSpace(split[1])\\n\\n\\t\\t\\tif status == \\\"ONLINE\\\" {\\n\\t\\t\\t\\tb.Urgent = false\\n\\t\\t\\t} else {\\n\\t\\t\\t\\tb.Urgent = true\\n\\t\\t\\t}\\n\\t\\t\\tb.FullText = fmt.Sprintf(fullTextFmt, status)\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t}\\n\\n\\tb.Urgent = true\\n\\tb.FullText = fmt.Sprintf(fullTextFmt, \\\"NOT FOUND\\\")\\n\\treturn\\n}\",\n \"func (h *ihash) BlockSize() int { return h.blockSize }\",\n \"func (syncer *MerkleSyncer) updateHeight() {\\n\\tatomic.AddUint64(&syncer.height, 1)\\n}\",\n \"func (k Keeper) BlockByteSize(ctx sdk.Ctx) (res int64) {\\n\\tk.Paramstore.Get(ctx, types.KeyBlockByteSize, &res)\\n\\treturn\\n}\",\n \"func (ob *Observer) updateBlock(curHeight, nextHeight int64, curBlockHash string) error {\\n\\tblock, err := ob.deps.Recorder.Block(nextHeight)\\n\\tif err != nil {\\n\\t\\treturn errors.Wrapf(err, \\\"[Observer.updateBlock]: failed to get block info, height=%d\\\", nextHeight)\\n\\t}\\n\\n\\tif curHeight != 0 && block.ParentBlockHash != curBlockHash {\\n\\t\\tif err := ob.DeleteBlock(curHeight); err != nil {\\n\\t\\t\\treturn errors.Wrap(err, \\\"[Observer.updateBlock]: failed to delete a forked block\\\")\\n\\t\\t}\\n\\n\\t\\treturn nil\\n\\t}\\n\\n\\tif err := ob.RecordBlockAndTxs(block); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"[Observer.updateBlock]: failed to save and process block\\\")\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (xxh *xxHash) BlockSize() int {\\n\\treturn 1\\n}\",\n \"func (d *btrfs) Update(changedConfig map[string]string) error {\\n\\t// We only care about btrfs.mount_options.\\n\\tval, ok := changedConfig[\\\"btrfs.mount_options\\\"]\\n\\tif ok {\\n\\t\\t// Custom mount options don't work inside containers\\n\\t\\tif d.state.OS.RunningInUserNS {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\n\\t\\t// Trigger a re-mount.\\n\\t\\td.config[\\\"btrfs.mount_options\\\"] = val\\n\\t\\tmntFlags, mntOptions := filesystem.ResolveMountOptions(strings.Split(d.getMountOptions(), \\\",\\\"))\\n\\t\\tmntFlags |= unix.MS_REMOUNT\\n\\n\\t\\terr := TryMount(\\\"\\\", GetPoolMountPath(d.name), \\\"none\\\", mntFlags, mntOptions)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\n\\tsize, ok := changedConfig[\\\"size\\\"]\\n\\tif ok {\\n\\t\\t// Figure out loop path\\n\\t\\tloopPath := loopFilePath(d.name)\\n\\n\\t\\tif d.config[\\\"source\\\"] != loopPath {\\n\\t\\t\\treturn fmt.Errorf(\\\"Cannot resize non-loopback pools\\\")\\n\\t\\t}\\n\\n\\t\\t// Resize loop file\\n\\t\\tf, err := os.OpenFile(loopPath, os.O_RDWR, 0600)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\tdefer func() { _ = f.Close() }()\\n\\n\\t\\tsizeBytes, _ := units.ParseByteSizeString(size)\\n\\n\\t\\terr = f.Truncate(sizeBytes)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\tloopDevPath, err := loopDeviceSetup(loopPath)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\tdefer func() { _ = loopDeviceAutoDetach(loopDevPath) }()\\n\\n\\t\\terr = loopDeviceSetCapacity(loopDevPath)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\t_, err = shared.RunCommand(\\\"btrfs\\\", \\\"filesystem\\\", \\\"resize\\\", \\\"max\\\", GetPoolMountPath(d.name))\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (ns *EsIndexer) UpdateLastBlockHeightFromDb() {\\n\\tbestBlock, err := ns.GetBestBlockFromDb()\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tns.lastBlockHeight = bestBlock.BlockNo\\n\\tns.lastBlockHash = bestBlock.id\\n}\",\n \"func (s *StatFS) BlockFiles(size int) int {\\n\\treturn size / int(s.stat.Bsize)\\n}\",\n \"func (digest *sm3Digest) BlockSize() int {\\n\\treturn BlockSize\\n}\",\n \"func (d Avx512Digest) BlockSize() int { return BlockSize }\",\n \"func (nsc *NilConsumerStatsCollector) UpdateCheckpointSize(int) {}\",\n \"func (s *Store) UpdateTreeStoreSize(key string, newSize int64) error {\\n\\treturn s.db.Do(func(tx *sql.Tx) error {\\n\\t\\t_, err := tx.Exec(\\\"UPDATE aciinfo SET treestoresize = $1 WHERE blobkey == $2\\\", newSize, key)\\n\\t\\treturn err\\n\\t})\\n}\",\n \"func (fs *FileSystem) UpdateFst(file inode) error {\\n\\tfs.nextFreeInode = fs.nextFreeInode[1:]\\n\\tfs.nextFreeDataBlock = fs.nextFreeDataBlock[len(file.dataList):]\\n\\treturn nil\\n}\",\n \"func (fsys *FS) Statfs(path string, stat *fuse.Statfs_t) (errc int) {\\n\\tdefer fs.Trace(path, \\\"\\\")(\\\"stat=%+v, errc=%d\\\", stat, &errc)\\n\\tconst blockSize = 4096\\n\\tfsBlocks := uint64(1 << 50)\\n\\tif runtime.GOOS == \\\"windows\\\" {\\n\\t\\tfsBlocks = (1 << 43) - 1\\n\\t}\\n\\tstat.Blocks = fsBlocks // Total data blocks in file system.\\n\\tstat.Bfree = fsBlocks // Free blocks in file system.\\n\\tstat.Bavail = fsBlocks // Free blocks in file system if you're not root.\\n\\tstat.Files = 1E9 // Total files in file system.\\n\\tstat.Ffree = 1E9 // Free files in file system.\\n\\tstat.Bsize = blockSize // Block size\\n\\tstat.Namemax = 255 // Maximum file name length?\\n\\tstat.Frsize = blockSize // Fragment size, smallest addressable data size in the file system.\\n\\treturn 0\\n}\",\n \"func getSizeForPath(path string) (uint64, error) {\\n\\ts := syscall.Statfs_t{}\\n\\tif err := syscall.Statfs(path, &s); err != nil {\\n\\t\\treturn 0, fmt.Errorf(\\\"failed to statfs on %s, %+v\\\", path, err)\\n\\t}\\n\\n\\treturn s.Blocks * uint64(s.Bsize), nil\\n}\",\n \"func (b *Block) Size() common.StorageSize {\\n\\tif size := b.size.Load(); size != nil {\\n\\t\\treturn size.(common.StorageSize)\\n\\t}\\n\\tc := writeCounter(0)\\n\\trlp.Encode(&c, b)\\n\\tb.size.Store(common.StorageSize(c))\\n\\treturn common.StorageSize(c)\\n}\",\n \"func blocklistUpdate() error {\\n\\tctx, cancel := context.WithTimeout(context.TODO(), ezhttp.DefaultTimeout10s)\\n\\tdefer cancel()\\n\\tres, err := ezhttp.Get(\\n\\t\\tctx,\\n\\t\\t\\\"https://download.dnscrypt.info/blacklists/domains/mybase.txt\\\",\\n\\t\\tezhttp.Header(\\\"User-Agent\\\", \\\"github.com/function61/function53\\\"))\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn atomicfilewrite.Write(blocklistFilename, func(blocklist io.Writer) error {\\n\\t\\t_, err := io.Copy(blocklist, res.Body)\\n\\t\\treturn err\\n\\t})\\n}\",\n \"func (leaf *Node) update(newInfo os.FileInfo, withContent bool) (err error) {\\n\\tif newInfo == nil {\\n\\t\\tnewInfo, err = os.Stat(leaf.SysPath)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"lib/memfs: Node.update %q: %s\\\",\\n\\t\\t\\t\\tleaf.Path, err.Error())\\n\\t\\t}\\n\\t}\\n\\n\\tif leaf.Mode != newInfo.Mode() {\\n\\t\\tleaf.Mode = newInfo.Mode()\\n\\t\\treturn nil\\n\\t}\\n\\n\\tleaf.ModTime = newInfo.ModTime()\\n\\tleaf.Size = newInfo.Size()\\n\\n\\tif !withContent || newInfo.IsDir() {\\n\\t\\treturn nil\\n\\t}\\n\\n\\treturn leaf.updateContent()\\n}\",\n \"func (f FormatHeader) BlockSize() uint16 {\\n\\treturn (f.BitsPerSample / 8) * f.NumChannels\\n}\",\n \"func (x *gcm) BlockSize() int { return x.blockSize }\",\n \"func (h *blockHeader) setSize(v uint32) {\\n\\tconst mask = 7\\n\\t*h = (*h)&mask | blockHeader(v<<3)\\n}\",\n \"func (c *BlockCache) setDbFiles(height int) {\\n\\tif height <= c.nextBlock {\\n\\t\\tif height < c.firstBlock {\\n\\t\\t\\theight = c.firstBlock\\n\\t\\t}\\n\\t\\tindex := height - c.firstBlock\\n\\t\\tif err := c.lengthsFile.Truncate(int64(index * 4)); err != nil {\\n\\t\\t\\tLog.Fatal(\\\"truncate lengths file failed: \\\", err)\\n\\t\\t}\\n\\t\\tif err := c.blocksFile.Truncate(c.starts[index]); err != nil {\\n\\t\\t\\tLog.Fatal(\\\"truncate blocks file failed: \\\", err)\\n\\t\\t}\\n\\t\\tc.Sync()\\n\\t\\tc.starts = c.starts[:index+1]\\n\\t\\tc.nextBlock = height\\n\\t\\tc.setLatestHash()\\n\\t}\\n}\",\n \"func Diskuse(path string) (cap uint64, used uint64) {\\n\\n\\tfs := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\tfmt.Println(err)\\n\\t}\\n\\n\\tcap = fs.Blocks * uint64(fs.Bsize)\\n\\tfree := fs.Bfree * uint64(fs.Bsize) // yup, I just did that\\n\\tused = cap - free\\n\\treturn cap, used\\n}\",\n \"func (leaf *Node) updateContent() (err error) {\\n\\tif leaf.Size > MaxFileSize {\\n\\t\\treturn nil\\n\\t}\\n\\n\\tleaf.V, err = ioutil.ReadFile(leaf.SysPath)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (fup *folderUpdatePrepper) updateResolutionUsageLockedCache(\\n\\tctx context.Context, lState *kbfssync.LockState, md *RootMetadata,\\n\\tbps blockPutState, unmergedChains, mergedChains *crChains,\\n\\tmostRecentMergedMD ImmutableRootMetadata,\\n\\trefs, unrefs map[data.BlockPointer]bool) error {\\n\\tmd.SetRefBytes(0)\\n\\tmd.SetUnrefBytes(0)\\n\\tmd.SetMDRefBytes(0)\\n\\tmd.SetDiskUsage(mostRecentMergedMD.DiskUsage())\\n\\tmd.SetMDDiskUsage(mostRecentMergedMD.MDDiskUsage())\\n\\n\\tlocalBlocks := make(map[data.BlockPointer]data.Block)\\n\\tfor _, ptr := range bps.Ptrs() {\\n\\t\\tif block, err := bps.GetBlock(ctx, ptr); err == nil && block != nil {\\n\\t\\t\\tlocalBlocks[ptr] = block\\n\\t\\t}\\n\\t}\\n\\n\\t// Add bytes for every ref'd block.\\n\\trefPtrsToFetch := make([]data.BlockPointer, 0, len(refs))\\n\\tvar refSum uint64\\n\\tfor ptr := range refs {\\n\\t\\tif block, ok := localBlocks[ptr]; ok {\\n\\t\\t\\trefSum += uint64(block.GetEncodedSize())\\n\\t\\t} else {\\n\\t\\t\\trefPtrsToFetch = append(refPtrsToFetch, ptr)\\n\\t\\t}\\n\\t\\tfup.vlog.CLogf(ctx, libkb.VLog1, \\\"Ref'ing block %v\\\", ptr)\\n\\t}\\n\\n\\t// Look up the total sum of the ref blocks in parallel to get\\n\\t// their sizes.\\n\\t//\\n\\t// TODO: If the blocks weren't already in the cache, this call\\n\\t// won't cache them, so it's kind of wasting work. Furthermore,\\n\\t// we might be able to get the encoded size from other sources as\\n\\t// well (such as its directory entry or its indirect file block)\\n\\t// if we happened to have come across it before.\\n\\trefSumFetched, err := fup.blocks.GetCleanEncodedBlocksSizeSum(\\n\\t\\tctx, lState, md.ReadOnly(), refPtrsToFetch, nil, fup.branch(), false)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\trefSum += refSumFetched\\n\\n\\tfup.vlog.CLogf(ctx, libkb.VLog1, \\\"Ref'ing a total of %d bytes\\\", refSum)\\n\\tmd.AddRefBytes(refSum)\\n\\tmd.AddDiskUsage(refSum)\\n\\n\\tunrefPtrsToFetch := make([]data.BlockPointer, 0, len(unrefs))\\n\\tvar unrefSum uint64\\n\\tfor ptr := range unrefs {\\n\\t\\toriginal, ok := unmergedChains.originals[ptr]\\n\\t\\tif !ok {\\n\\t\\t\\toriginal = ptr\\n\\t\\t}\\n\\t\\tif original != ptr || unmergedChains.isCreated(original) {\\n\\t\\t\\t// Only unref pointers that weren't created as part of the\\n\\t\\t\\t// unmerged branch. Either they existed already or they\\n\\t\\t\\t// were created as part of the merged branch.\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\t// Also make sure this wasn't already removed or overwritten\\n\\t\\t// on the merged branch.\\n\\t\\toriginal, ok = mergedChains.originals[ptr]\\n\\t\\tif !ok {\\n\\t\\t\\toriginal = ptr\\n\\t\\t}\\n\\t\\tmergedChain, ok := mergedChains.byOriginal[original]\\n\\t\\tif (ok && original != mergedChain.mostRecent && original == ptr) ||\\n\\t\\t\\tmergedChains.isDeleted(original) {\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\n\\t\\tif info, ok := fup.cachedInfos[ptr]; ok {\\n\\t\\t\\tunrefSum += uint64(info.EncodedSize)\\n\\t\\t} else {\\n\\t\\t\\tunrefPtrsToFetch = append(unrefPtrsToFetch, ptr)\\n\\t\\t}\\n\\t}\\n\\n\\t// Look up the unref blocks in parallel to get their sizes. Since\\n\\t// we don't know whether these are files or directories, just look\\n\\t// them up generically. Ignore any recoverable errors for unrefs.\\n\\t// Note that we can't combine these with the above ref fetches\\n\\t// since they require a different MD. If the merged changes\\n\\t// didn't change any blocks (in particular, the root block), we\\n\\t// can assume all the blocks we are unreferencing were live;\\n\\t// otherwise, we need to check with the server to make sure.\\n\\tonlyCountIfLive := len(mergedChains.byOriginal) != 0\\n\\tunrefSumFetched, err := fup.blocks.GetCleanEncodedBlocksSizeSum(\\n\\t\\tctx, lState, mostRecentMergedMD, unrefPtrsToFetch, unrefs,\\n\\t\\tfup.branch(), onlyCountIfLive)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tunrefSum += unrefSumFetched\\n\\n\\t// Subtract bytes for every unref'd block that wasn't created in\\n\\t// the unmerged branch.\\n\\tfup.vlog.CLogf(ctx, libkb.VLog1, \\\"Unref'ing a total of %d bytes\\\", unrefSum)\\n\\tmd.AddUnrefBytes(unrefSum)\\n\\tmd.SetDiskUsage(md.DiskUsage() - unrefSum)\\n\\treturn nil\\n}\",\n \"func (b *SnailBlock) Size() common.StorageSize {\\n\\tif size := b.size.Load(); size != nil {\\n\\t\\treturn size.(common.StorageSize)\\n\\t}\\n\\tc := writeCounter(0)\\n\\trlp.Encode(&c, b)\\n\\tb.size.Store(common.StorageSize(c))\\n\\treturn common.StorageSize(c)\\n}\",\n \"func SetBpfBuflen(fd, l int) (int, error) {\\n\\terr := ioctlPtr(fd, BIOCSBLEN, unsafe.Pointer(&l))\\n\\tif err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\treturn l, nil\\n}\",\n \"func UpdateClusterMemoryCurrentBytes(memoryCount int64) {\\n\\tmemoryCurrentBytes.Set(float64(memoryCount))\\n}\",\n \"func (u *uploader) initSize() {\\n\\tu.totalSize = -1\\n\\n\\tswitch r := u.in.Body.(type) {\\n\\tcase io.Seeker:\\n\\t\\tn, err := aws.SeekerLen(r)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\tu.totalSize = n\\n\\n\\t\\t// Try to adjust partSize if it is too small and account for\\n\\t\\t// integer division truncation.\\n\\t\\tif u.totalSize/u.cfg.PartSize >= int64(u.cfg.MaxUploadParts) {\\n\\t\\t\\t// Add one to the part size to account for remainders\\n\\t\\t\\t// during the size calculation. e.g odd number of bytes.\\n\\t\\t\\tu.cfg.PartSize = (u.totalSize / int64(u.cfg.MaxUploadParts)) + 1\\n\\t\\t}\\n\\t}\\n}\",\n \"func (layout Layout) hashesPerBlock() int64 {\\n\\treturn layout.blockSize / layout.digestSize\\n}\",\n \"func (m *metricSshcheckSftpStatus) updateCapacity() {\\n\\tif m.data.Sum().DataPoints().Len() > m.capacity {\\n\\t\\tm.capacity = m.data.Sum().DataPoints().Len()\\n\\t}\\n}\",\n \"func (*digest) BlockSize() int {\\n\\treturn int(BlockSize)\\n}\",\n \"func (*digest) BlockSize() int {\\n\\treturn int(BlockSize)\\n}\",\n \"func newBlockfileMgr(id string, conf *Conf, indexConfig *blkstorage.IndexConfig, indexStore *leveldbhelper.DBHandle) *blockfileMgr {\\n\\tlogger.Debugf(\\\"newBlockfileMgr() initializing file-based block storage for ledger: %s \\\", id)\\n\\tvar rwMutexs []*sync.RWMutex\\n\\n\\t//Determine the root directory for the blockfile storage, if it does not exist create it\\n\\trootDir := conf.getLedgerBlockDir(id)\\n\\t_, err := util.CreateDirIfMissing(rootDir)\\n\\tif err != nil {\\n\\t\\tpanic(fmt.Sprintf(\\\"Error: %s\\\", err))\\n\\t}\\n\\t// Instantiate the manager, i.e. blockFileMgr structure\\n\\tmgr := &blockfileMgr{rootDir: rootDir, conf: conf, db: indexStore, rwMutexs: rwMutexs}\\n\\n\\t// cp = checkpointInfo, retrieve from the database the file suffix or number of where blocks were stored.\\n\\t// It also retrieves the current size of that file and the last block number that was written to that file.\\n\\t// At init checkpointInfo:latestFileChunkSuffixNum=[0], latestFileChunksize=[0], lastBlockNumber=[0]\\n\\tcpInfo, err := mgr.loadCurrentInfo()\\n\\tif err != nil {\\n\\t\\tpanic(fmt.Sprintf(\\\"Could not get block file info for current block file from db: %s\\\", err))\\n\\t}\\n\\tif cpInfo == nil {\\n\\t\\tlogger.Info(`Getting block information from block storage`)\\n\\t\\tif cpInfo, err = constructCheckpointInfoFromBlockFiles(rootDir); err != nil {\\n\\t\\t\\tpanic(fmt.Sprintf(\\\"Could not build checkpoint info from block files: %s\\\", err))\\n\\t\\t}\\n\\t\\tlogger.Debugf(\\\"Info constructed by scanning the blocks dir = %s\\\", spew.Sdump(cpInfo))\\n\\t} else {\\n\\t\\tlogger.Debug(`Synching block information from block storage (if needed)`)\\n\\t\\tsyncCPInfoFromFS(rootDir, cpInfo)\\n\\t}\\n\\terr = mgr.saveCurrentInfo(cpInfo, true)\\n\\tif err != nil {\\n\\t\\tpanic(fmt.Sprintf(\\\"Could not save next block file info to db: %s\\\", err))\\n\\t}\\n\\n\\tmgr.oldestFileChunkSuffixNum = syncOldestFileNum(rootDir)\\n\\t//If start up is a restart of an existing storage,new the rwMutex for the files\\n\\tif conf.dumpConf.Enabled {\\n\\t\\tfor i := 0; i <= cpInfo.latestFileChunkSuffixNum; i++ {\\n\\t\\t\\trwMutex := new(sync.RWMutex)\\n\\t\\t\\tmgr.rwMutexs = append(mgr.rwMutexs, rwMutex)\\n\\t\\t}\\n\\t}\\n\\tmgr.dumpMutex = new(sync.Mutex)\\n\\n\\t//Open a writer to the file identified by the number and truncate it to only contain the latest block\\n\\t// that was completely saved (file system, index, cpinfo, etc)\\n\\tcurrentFileWriter, err := newBlockfileWriter(deriveBlockfilePath(rootDir, cpInfo.latestFileChunkSuffixNum))\\n\\tif err != nil {\\n\\t\\tpanic(fmt.Sprintf(\\\"Could not open writer to current file: %s\\\", err))\\n\\t}\\n\\t//Truncate the file to remove excess past last block\\n\\terr = currentFileWriter.truncateFile(cpInfo.latestFileChunksize)\\n\\tif err != nil {\\n\\t\\tpanic(fmt.Sprintf(\\\"Could not truncate current file to known size in db: %s\\\", err))\\n\\t}\\n\\n\\t// Create a new KeyValue store database handler for the blocks index in the keyvalue database\\n\\tmgr.index = newBlockIndex(indexConfig, indexStore)\\n\\n\\t// Update the manager with the checkpoint info and the file writer\\n\\tmgr.cpInfo = cpInfo\\n\\tmgr.currentFileWriter = currentFileWriter\\n\\t// Create a checkpoint condition (event) variable, for the goroutine waiting for\\n\\t// or announcing the occurrence of an event.\\n\\tmgr.cpInfoCond = sync.NewCond(&sync.Mutex{})\\n\\n\\t// init BlockchainInfo for external API's\\n\\tbcInfo := &common.BlockchainInfo{\\n\\t\\tHeight: 0,\\n\\t\\tCurrentBlockHash: nil,\\n\\t\\tPreviousBlockHash: nil}\\n\\n\\tif !cpInfo.isChainEmpty {\\n\\t\\t//If start up is a restart of an existing storage, sync the index from block storage and update BlockchainInfo for external API's\\n\\t\\tmgr.syncIndex()\\n\\t\\tlastBlockHeader, err := mgr.retrieveBlockHeaderByNumber(cpInfo.lastBlockNumber)\\n\\t\\tif err != nil {\\n\\t\\t\\tpanic(fmt.Sprintf(\\\"Could not retrieve header of the last block form file: %s\\\", err))\\n\\t\\t}\\n\\t\\tlastBlockHash := lastBlockHeader.Hash()\\n\\t\\tpreviousBlockHash := lastBlockHeader.PreviousHash\\n\\t\\tbcInfo = &common.BlockchainInfo{\\n\\t\\t\\tHeight: cpInfo.lastBlockNumber + 1,\\n\\t\\t\\tCurrentBlockHash: lastBlockHash,\\n\\t\\t\\tPreviousBlockHash: previousBlockHash}\\n\\t}\\n\\tmgr.bcInfo.Store(bcInfo)\\n\\treturn mgr\\n}\",\n \"func DirSizeByte(path string) uint64 {\\n\\tdirSize = 0\\n\\tfilepath.Walk(path, readSize)\\n\\treturn dirSize\\n}\",\n \"func calculateBufferSize(blocks uint64) uint64 {\\n\\tif nbb := NBufferBlocks; blocks < nbb {\\n\\t\\treturn blocks\\n\\t} else {\\n\\t\\treturn nbb\\n\\t}\\n}\",\n \"func (b *BlockSplitterSimple) SetMaxDirEntriesByBlockSize(\\n\\tcodec kbfscodec.Codec) error {\\n\\tdirEnv := os.Getenv(\\\"KEYBASE_BSPLIT_MAX_DIR_ENTRIES\\\")\\n\\tif len(dirEnv) > 0 {\\n\\t\\t// Don't override the environment variable.\\n\\t\\treturn nil\\n\\t}\\n\\n\\tblock := NewDirBlock().(*DirBlock)\\n\\tbigName := strings.Repeat(\\\"a\\\", MaxNameBytesDefault)\\n\\t// Make \\\"typical\\\" DirEntry, though the max dir entry is a bit\\n\\t// bigger than this (can contain a variable-length symlink path,\\n\\t// for example).\\n\\tde := DirEntry{\\n\\t\\tBlockInfo: BlockInfo{\\n\\t\\t\\tBlockPointer: BlockPointer{\\n\\t\\t\\t\\tDirectType: DirectBlock,\\n\\t\\t\\t},\\n\\t\\t},\\n\\t\\tEntryInfo: EntryInfo{\\n\\t\\t\\tPrevRevisions: PrevRevisions{\\n\\t\\t\\t\\t{Revision: 0, Count: 0},\\n\\t\\t\\t\\t{Revision: 1, Count: 1},\\n\\t\\t\\t\\t{Revision: 2, Count: 2},\\n\\t\\t\\t\\t{Revision: 3, Count: 3},\\n\\t\\t\\t\\t{Revision: 4, Count: 4},\\n\\t\\t\\t},\\n\\t\\t},\\n\\t}\\n\\tblock.Children[bigName] = de\\n\\tencodedBlock, err := codec.Encode(block)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\toneEntrySize := int64(len(encodedBlock))\\n\\tb.maxDirEntriesPerBlock = int(b.maxSize / oneEntrySize)\\n\\tif b.maxDirEntriesPerBlock == 0 {\\n\\t\\tb.maxDirEntriesPerBlock = 1\\n\\t}\\n\\treturn nil\\n}\",\n \"func (bdi *Info) CountBlocks() int {\\n\\treturn int(C.spdk_bdev_get_num_blocks(bdi.ptr()))\\n}\",\n \"func updateTxSize(tx *transaction.Transaction) (*transaction.Transaction, error) {\\n\\tbw := io.NewBufBinWriter()\\n\\ttx.EncodeBinary(bw.BinWriter)\\n\\tif bw.Err != nil {\\n\\t\\treturn nil, fmt.Errorf(\\\"encode binary: %w\\\", bw.Err)\\n\\t}\\n\\treturn transaction.NewTransactionFromBytes(tx.Bytes())\\n}\",\n \"func (c *Container) rootFsSize() (int64, error) {\\n\\tcontainer, err := c.runtime.store.Container(c.ID())\\n\\tif err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\t// Ignore the size of the top layer. The top layer is a mutable RW layer\\n\\t// and is not considered a part of the rootfs\\n\\trwLayer, err := c.runtime.store.Layer(container.LayerID)\\n\\tif err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\tlayer, err := c.runtime.store.Layer(rwLayer.Parent)\\n\\tif err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\tsize := int64(0)\\n\\tfor layer.Parent != \\\"\\\" {\\n\\t\\tlayerSize, err := c.runtime.store.DiffSize(layer.Parent, layer.ID)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn size, errors.Wrapf(err, \\\"getting diffsize of layer %q and its parent %q\\\", layer.ID, layer.Parent)\\n\\t\\t}\\n\\t\\tsize += layerSize\\n\\t\\tlayer, err = c.runtime.store.Layer(layer.Parent)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn 0, err\\n\\t\\t}\\n\\t}\\n\\t// Get the size of the last layer. Has to be outside of the loop\\n\\t// because the parent of the last layer is \\\"\\\", andlstore.Get(\\\"\\\")\\n\\t// will return an error.\\n\\tlayerSize, err := c.runtime.store.DiffSize(layer.Parent, layer.ID)\\n\\treturn size + layerSize, err\\n}\",\n \"func (o KubernetesClusterNodePoolLinuxOsConfigOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\\n\\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfig) *int { return v.SwapFileSizeMb }).(pulumi.IntPtrOutput)\\n}\",\n \"func (req *DronaRequest) updateOsize(size int64) {\\n\\treq.Lock()\\n\\tdefer req.Unlock()\\n\\treq.objectSize = size\\n}\",\n \"func statfsImpl(t *kernel.Task, d *fs.Dirent, addr usermem.Addr) error {\\n\\tinfo, err := d.Inode.StatFS(t)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\t// Construct the statfs structure and copy it out.\\n\\tstatfs := linux.Statfs{\\n\\t\\tType: info.Type,\\n\\t\\t// Treat block size and fragment size as the same, as\\n\\t\\t// most consumers of this structure will expect one\\n\\t\\t// or the other to be filled in.\\n\\t\\tBlockSize: d.Inode.StableAttr.BlockSize,\\n\\t\\tBlocks: info.TotalBlocks,\\n\\t\\t// We don't have the concept of reserved blocks, so\\n\\t\\t// report blocks free the same as available blocks.\\n\\t\\t// This is a normal thing for filesystems, to do, see\\n\\t\\t// udf, hugetlbfs, tmpfs, among others.\\n\\t\\tBlocksFree: info.FreeBlocks,\\n\\t\\tBlocksAvailable: info.FreeBlocks,\\n\\t\\tFiles: info.TotalFiles,\\n\\t\\tFilesFree: info.FreeFiles,\\n\\t\\t// Same as Linux for simple_statfs, see fs/libfs.c.\\n\\t\\tNameLength: linux.NAME_MAX,\\n\\t\\tFragmentSize: d.Inode.StableAttr.BlockSize,\\n\\t\\t// Leave other fields 0 like simple_statfs does.\\n\\t}\\n\\t_, err = t.CopyOut(addr, &statfs)\\n\\treturn err\\n}\",\n \"func (c Temperature) UpdateBlock(b *i3barjson.Block) {\\n\\tb.Color = c.Color\\n\\tfullTextFmt := fmt.Sprintf(\\\"%s%%s\\\", c.Label)\\n\\ttotalTemp := 0\\n\\tprocs := 0\\n\\tsysFileDirList, err := ioutil.ReadDir(c.CpuTempPath)\\n\\tif err != nil {\\n\\t\\tb.Urgent = true\\n\\t\\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\\n\\t\\treturn\\n\\t}\\n\\tif len(sysFileDirList) != 1 {\\n\\t\\tb.Urgent = true\\n\\t\\tmsg := fmt.Sprintf(\\n\\t\\t\\t\\\"in %s, expected 1 file, got %d\\\",\\n\\t\\t\\tc.CpuTempPath,\\n\\t\\t\\tlen(sysFileDirList),\\n\\t\\t)\\n\\t\\tb.FullText = fmt.Sprintf(fullTextFmt, msg)\\n\\t\\treturn\\n\\t}\\n\\tsysFileDirPath := fmt.Sprintf(\\n\\t\\t\\\"%s/%s\\\",\\n\\t\\tc.CpuTempPath,\\n\\t\\tsysFileDirList[0].Name(),\\n\\t)\\n\\tsysFileNameFmt := fmt.Sprintf(\\\"%s/%%s\\\", sysFileDirPath)\\n\\tsysFiles, err := ioutil.ReadDir(sysFileDirPath)\\n\\tif err != nil {\\n\\t\\tb.Urgent = true\\n\\t\\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\\n\\t\\treturn\\n\\t}\\n\\tfor _, sysFile := range sysFiles {\\n\\t\\tsysFileName := sysFile.Name()\\n\\t\\tif !strings.HasSuffix(sysFileName, \\\"input\\\") {\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\tr, err := os.Open(fmt.Sprintf(sysFileNameFmt, sysFileName))\\n\\t\\tif err != nil {\\n\\t\\t\\tb.Urgent = true\\n\\t\\t\\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\tdefer r.Close()\\n\\t\\tvar temp int\\n\\t\\t_, err = fmt.Fscanf(r, \\\"%d\\\", &temp)\\n\\t\\tif err != nil {\\n\\t\\t\\tb.Urgent = true\\n\\t\\t\\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\ttotalTemp += temp\\n\\t\\tprocs++\\n\\t}\\n\\tavgTemp := float64(totalTemp) / float64(procs*1000)\\n\\tif avgTemp >= c.CritTemp {\\n\\t\\tb.Urgent = true\\n\\t} else {\\n\\t\\tb.Urgent = false\\n\\t}\\n\\tb.FullText = fmt.Sprintf(\\\"%s%.2f°C\\\", c.Label, avgTemp)\\n}\",\n \"func updateRemainingSpace(ctx *downloaderContext) {\\n\\n\\tctx.globalStatus.RemainingSpace = ctx.globalConfig.MaxSpace -\\n\\t\\tctx.globalStatus.UsedSpace - ctx.globalStatus.ReservedSpace\\n\\n\\tlog.Infof(\\\"RemainingSpace %d, maxspace %d, usedspace %d, reserved %d\\\\n\\\",\\n\\t\\tctx.globalStatus.RemainingSpace, ctx.globalConfig.MaxSpace,\\n\\t\\tctx.globalStatus.UsedSpace, ctx.globalStatus.ReservedSpace)\\n}\",\n \"func (f *FileSystem) Statfs(ctx context.Context, req *fuse.StatfsRequest, resp *fuse.StatfsResponse) error {\\n\\t// each block size is 4096 bytes by default.\\n\\tconst unit = uint64(4096)\\n\\n\\tresp.Bsize = uint32(unit)\\n\\tresp.Blocks = uint64(f.account.Disk.Size) / unit\\n\\tresp.Bavail = uint64(f.account.Disk.Avail) / unit\\n\\tresp.Bfree = uint64(f.account.Disk.Avail) / unit\\n\\n\\treturn nil\\n}\",\n \"func (lf *ListFile) NumBytes() int64 {\\n\\t// NOTE: here we don't use IsClosed() because\\n\\t// it uses the mutex; Size() is used in noMutexIterateLines\\n\\t// which is called after another mutex is locked,\\n\\t// making IsClosed() wait forever for the mutex unlock.\\n\\tif lf.isClosed {\\n\\t\\treturn 0\\n\\t}\\n\\n\\terr := lf.file.Sync()\\n\\tif err != nil {\\n\\t\\t// TODO: not panic??\\n\\t\\tpanic(err)\\n\\t}\\n\\n\\tinfo, err := lf.file.Stat()\\n\\tif err != nil {\\n\\t\\t// TODO: not panic??\\n\\t\\tpanic(err)\\n\\t}\\n\\n\\treturn info.Size()\\n}\",\n \"func (b *Block) Size() metric.StorageSize {\\n\\tif cached := b.cache.size.Load(); cached != nil {\\n\\t\\treturn cached.(metric.StorageSize)\\n\\t}\\n\\tvar size metric.StorageSize\\n\\trlp.Encode(&size, b)\\n\\tb.cache.size.Store(size)\\n\\treturn size\\n}\",\n \"func (pb *PutBlock) Height() uint64 { return pb.height }\",\n \"func (n *nodeBlock) sync() (err error) {\\n\\tmm := false\\n\\tif n.offset == 0 {\\n\\t\\tif n._super.mmapSizeUsed+nodeBlockSize <= n._super.mmapSize {\\n\\t\\t\\tn.offset = int64(n._super.mmapSizeUsed)\\n\\t\\t\\tn._super.mmapSizeUsed += nodeBlockSize\\n\\t\\t\\tmm = true\\n\\t\\t} else {\\n\\t\\t\\tn.offset, err = n._super._fd.Seek(0, os.SEEK_END)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\n\\tx := *(*[nodeBlockSize]byte)(unsafe.Pointer(n))\\n\\n\\tif mm {\\n\\t\\tcopy(n._super._mmap[n.offset:], x[:])\\n\\t} else {\\n\\t\\t_, err = n._super._fd.Seek(n.offset, 0)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\t_, err = n._super._fd.Write(x[:])\\n\\t}\\n\\n\\tif err == nil {\\n\\t\\tn._super._snapshotChPending[n] = x\\n\\t}\\n\\treturn\\n}\",\n \"func (fm *FinalModelStructBytes) FBEAllocationSize(fbeValue *StructBytes) int {\\n fbeResult := 0 +\\n fm.F1.FBEAllocationSize(fbeValue.F1) +\\n fm.F2.FBEAllocationSize(fbeValue.F2) +\\n fm.F3.FBEAllocationSize(fbeValue.F3) +\\n 0\\n return fbeResult\\n}\",\n \"func (d *Directory) updateDirectorySize(p string) {\\n\\tpathParams := strings.Split(p, \\\"/\\\")\\n\\td.Size = 0\\n\\td.lock()\\n\\tdefer d.unlock()\\n\\tfor i := 0; i < len(d.INodes); i++ {\\n\\t\\tswitch d.INodes[i].(type) {\\n\\t\\tcase *Directory:\\n\\t\\t\\tif d.INodes[i].GetName() == pathParams[1] {\\n\\t\\t\\t\\tsubPath := strings.Join(pathParams[2:], \\\"/\\\")\\n\\t\\t\\t\\tsubPath = \\\"/\\\" + subPath\\n\\t\\t\\t\\td.INodes[i].(*Directory).updateDirectorySize(subPath)\\n\\t\\t\\t}\\n\\t\\t\\td.Size += d.INodes[i].GetSize()\\n\\t\\tcase *File:\\n\\t\\t\\td.Size += d.INodes[i].GetSize()\\n\\t\\t}\\n\\t}\\n}\",\n \"func (cs *ConsensusState) updateHeight(height int64) {\\n\\tcs.Height = height\\n}\",\n \"func (g *testGenerator) updateBlockState(oldBlockName string, oldBlockHash chainhash.Hash, newBlockName string, newBlock *wire.MsgBlock) {\\n\\t// Remove existing entries.\\n\\tdelete(g.blocks, oldBlockHash)\\n\\tdelete(g.blocksByName, oldBlockName)\\n\\n\\t// Add new entries.\\n\\tnewBlockHash := newBlock.BlockHash()\\n\\tg.blocks[newBlockHash] = newBlock\\n\\tg.blocksByName[newBlockName] = newBlock\\n}\",\n \"func (b *Buffer) update() {\\n\\tb.NumLines = len(b.lines)\\n}\",\n \"func (o KubernetesClusterDefaultNodePoolLinuxOsConfigOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\\n\\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfig) *int { return v.SwapFileSizeMb }).(pulumi.IntPtrOutput)\\n}\",\n \"func (p *pbkdf2) getHashLengthBitBlockCount() int {\\n\\treturn int(math.Ceil(float64(p.size) / float64(p.pseudoRandomFunction.Size())))\\n}\",\n \"func (x *ecbEncrypter) BlockSize() int { return x.blockSize }\",\n \"func (trd *trxDispatcher) updateLastSeenBlock() {\\n\\t// get the current value\\n\\tlsb := trd.blkObserver.Load()\\n\\tlog.Noticef(\\\"last seen block is #%d\\\", lsb)\\n\\n\\t// make the change in the database so the progress persists\\n\\terr := repo.UpdateLastKnownBlock((*hexutil.Uint64)(&lsb))\\n\\tif err != nil {\\n\\t\\tlog.Errorf(\\\"could not update last seen block; %s\\\", err.Error())\\n\\t}\\n}\",\n \"func UpdateCgroupDeviceWeight(pid, innerPath, value string) error {\\n\\tif pid == \\\"0\\\" {\\n\\t\\treturn nil\\n\\t}\\n\\n\\tcgroupPath, err := FindCgroupPath(pid, \\\"blkio\\\", innerPath)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tpath := filepath.Join(cgroupPath, \\\"blkio.weight_device\\\")\\n\\tif err := ioutil.WriteFile(path, []byte(value), 0600); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"%s, please check whether current OS support blkio weight device configuration for bfq scheduler\\\", err)\\n\\t}\\n\\treturn nil\\n}\",\n \"func (req *DronaRequest) updateAsize(size int64) {\\n\\treq.Lock()\\n\\tdefer req.Unlock()\\n\\treq.asize = size\\n}\",\n \"func (fileSystem *FileSystem) Put(fileName string) {\\n\\tif (*fileSystem).PfsFile == nil {\\n\\t\\tfmt.Println(\\\"You must open pfs first\\\")\\n\\t\\treturn\\n\\t}\\n\\n\\tdata, err := ioutil.ReadFile(fileName)\\n\\tif err != nil {\\n\\t\\tfmt.Println(err)\\n\\t\\treturn\\n\\t}\\n\\n\\tif len(data) > totalDataBlocks*dataBlockSize {\\n\\t\\tfmt.Println(\\\"File is too big to be inserted in pfs\\\")\\n\\t\\treturn\\n\\t}\\n\\n\\tfileSize := uint16(len(data)) // Get the file size\\n\\n\\t// Finding how many blocks is easy enough\\n\\tnecesaryBlocks := uint16(math.Ceil(float64(fileSize) / float64(dataBlockSize)))\\n\\tfoundBlocks := uint16(0)\\n\\tlocation := -1\\n\\tfor idx, availability := range (*fileSystem).Directory.FreeDataBlockArray {\\n\\t\\tif foundBlocks > 0 && !availability {\\n\\t\\t\\tfoundBlocks = 0\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\n\\t\\tif availability {\\n\\t\\t\\tfoundBlocks++\\n\\t\\t}\\n\\t\\tif foundBlocks == necesaryBlocks {\\n\\t\\t\\tlocation = idx\\n\\t\\t\\tbreak\\n\\t\\t}\\n\\t}\\n\\n\\tif location == -1 {\\n\\t\\tfmt.Println(\\\"Not enough free blocks to store file.\\\")\\n\\t\\treturn\\n\\t}\\n\\n\\t// We have found the location of data that should be written\\n\\tblockID := location - int(necesaryBlocks) + 1\\n\\toffset := dataAddress + blockID*dataBlockSize\\n\\t(*fileSystem).PfsFile.Seek(int64(offset), 0)\\n\\t(*fileSystem).PfsFile.Write(data)\\n\\n\\t// Now we need to find a FCB to store the records of the value\\n\\tfor idx, fcb := range (*fileSystem).Directory.FCBArray {\\n\\t\\tif !fcb.ContainsValidData {\\n\\t\\t\\tblock, _ := NewFCB(fileName, fileSize, uint8(blockID))\\n\\t\\t\\t(*fileSystem).Directory.FCBArray[idx] = block\\n\\t\\t\\tbreak\\n\\t\\t}\\n\\t}\\n\\n\\t// Now we also need to set the free data blocks as ocupied\\n\\tfor i := 0; i < int(necesaryBlocks); i++ {\\n\\t\\t(*fileSystem).Directory.FreeDataBlockArray[blockID+i] = false\\n\\t}\\n\\n\\t// We know update the directory\\n\\t(*fileSystem).Directory.Metadata.NumberOfFilesStored++\\n}\",\n \"func (o KubernetesClusterNodePoolLinuxOsConfigPtrOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\\n\\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfig) *int {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.SwapFileSizeMb\\n\\t}).(pulumi.IntPtrOutput)\\n}\",\n \"func UpdateDumpFile(fileInode inode, data string, fileName string, parentInode inode, parentInodeNum int) error {\\n\\t// open dump file\\n\\tfile, err := os.OpenFile(DumpFile, os.O_WRONLY, os.ModeAppend)\\n\\tif err != nil {\\n\\t\\tfmt.Println(\\\"Error in opening file while loading the filesystem: \\\", err)\\n\\t\\tos.Exit(1)\\n\\t}\\n\\tdefer file.Close()\\n\\n\\toffset := OffsetInodeTable + fileInode.inodeNum\\n\\t// update inode bitmap\\n\\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(strconv.Itoa(SetBit))); err != nil {\\n\\t\\tfmt.Println(\\\"Error: unable to update inode bitmap\\\", err)\\n\\t\\treturn err\\n\\t}\\n\\n\\t// update inode block\\n\\toffset = OffsetInodeBlock + (fileInode.inodeNum * InodeBlockSize)\\n\\tinodeInfo := frameInodeString(fileInode)\\n\\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(inodeInfo)); err != nil {\\n\\t\\tfmt.Println(\\\"Error: unable to update inode block\\\", err)\\n\\t\\treturn err\\n\\t}\\n\\n\\t//update parent inode block\\n\\tif parentInodeNum != -1 { // skipping it for root directory\\n\\t\\toffset = OffsetInodeBlock + (parentInodeNum * InodeBlockSize)\\n\\t\\tinodeInfo = frameInodeString(parentInode)\\n\\t\\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(inodeInfo)); err != nil {\\n\\t\\t\\tfmt.Println(\\\"Error: unable to update parent inode block\\\", err)\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\n\\t// update data bitmap and fill the data block as well ...\\n\\toffset = OffsetDataTable\\n\\tlistByte := []byte(data)\\n\\tcount := 0\\n\\tfor _, val := range fileInode.dataList {\\n\\t\\t// update inode bitmap\\n\\t\\tif err := disklib.WriteBlockOnDisk(file, offset+val, []byte(strconv.Itoa(SetBit))); err != nil {\\n\\t\\t\\tfmt.Println(\\\"Error: unable to update inode bitmap\\\", err)\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\t// update data block in disk\\n\\t\\tif err := disklib.WriteBlockOnDisk(file, OffsetDataBlock+(val*DataBlockSize), []byte(listByte[count:count+DataBlockSize])); err != nil {\\n\\t\\t\\tfmt.Println(\\\"Error: unable to update inode bitmap\\\", err)\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\tcount += DataBlockSize\\n\\t}\\n\\n\\t// put fst in disk\\n\\toffset = OffsetFST + (fileInode.inodeNum * CellFSTSize)\\n\\tstr := fmt.Sprintf(\\\"%3d\\\", fileInode.inodeNum) + fmt.Sprintf(\\\"%10s\\\", fileName)\\n\\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(str)); err != nil {\\n\\t\\tfmt.Println(\\\"Error: unable to update inode bitmap\\\", err)\\n\\t\\treturn err\\n\\t}\\n\\treturn nil\\n\\n}\",\n \"func DirSizeB(path string) (int64, error) {\\n\\tvar size int64\\n\\terr := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error {\\n\\t\\tif !info.IsDir() {\\n\\t\\t\\tsize += info.Size()\\n\\t\\t}\\n\\t\\treturn err\\n\\t})\\n\\treturn size, err\\n}\",\n \"func UpdateCapacity(cap int) {\\n\\ttype requestInfo struct {\\n\\t\\tNode string `json:\\\"node\\\" binding:\\\"required\\\"`\\n\\t\\tCapacity int `json:\\\"capability\\\" binding:\\\"required\\\"`\\n\\t}\\n\\tgo func() {\\n\\t\\tlog.InfoF(\\\"node %s capacity updated to %d\\\", conf.Get(\\\"nodeLanHost\\\").(string), cap)\\n\\t\\turl := fmt.Sprintf(\\\"http://%s/nodeUpdateCapacity\\\", conf.Get(\\\"registerCenterHost\\\"))\\n\\t\\t_, _, errs := gorequest.New().Post(url).Send(requestInfo{Node: conf.Get(\\\"nodeLanHost\\\").(string), Capacity: cap}).End()\\n\\t\\tif errs != nil && len(errs) > 0 {\\n\\t\\t\\tlog.InfoF(\\\"requestGroupDispatchLoop error:%s\\\", errs[0])\\n\\t\\t}\\n\\t}()\\n}\",\n \"func (fs *FS) fsInfo(ctx context.Context, path string) (int64, int64, int64, int64, int64, int64, error) {\\n\\tstatfs := &unix.Statfs_t{}\\n\\terr := unix.Statfs(path, statfs)\\n\\tif err != nil {\\n\\t\\treturn 0, 0, 0, 0, 0, 0, err\\n\\t}\\n\\n\\t// Available is blocks available * fragment size\\n\\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\\n\\n\\t// Capacity is total block count * fragment size\\n\\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\\n\\n\\t// Usage is block being used * fragment size (aka block size).\\n\\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\\n\\n\\tinodes := int64(statfs.Files)\\n\\tinodesFree := int64(statfs.Ffree)\\n\\tinodesUsed := inodes - inodesFree\\n\\n\\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\\n}\",\n \"func BlockSize() int {\\n\\treturn config.Record.BlockSize\\n}\",\n \"func (b *Block) Size() int {\\n\\tpbb, err := b.ToProto()\\n\\tif err != nil {\\n\\t\\treturn 0\\n\\t}\\n\\n\\treturn pbb.Size()\\n}\",\n \"func (b *Block) Size() int {\\n\\tpbb, err := b.ToProto()\\n\\tif err != nil {\\n\\t\\treturn 0\\n\\t}\\n\\n\\treturn pbb.Size()\\n}\",\n \"func (s *hashBLAKE2s) BlockLen() int {\\n\\treturn blake2s.BlockSize\\n}\",\n \"func (x *ecbDecrypter) BlockSize() int { return x.blockSize }\",\n \"func vmUpdateAdditionalBandwidth(db *Database, vm *VirtualMachine) {\\n\\t// determine how much of the plan bandwidth to add to the user's bandwidth pool for current month\\n\\tnow := time.Now()\\n\\tmonthStart := time.Date(now.Year(), now.Month(), 1, 0, 0, 0, 0, time.UTC)\\n\\tmonthEnd := monthStart.AddDate(0, 1, 0)\\n\\tvar factor float64\\n\\n\\tif vm.CreatedTime.Before(monthStart) {\\n\\t\\tfactor = float64(now.Sub(monthStart))\\n\\t} else {\\n\\t\\tfactor = float64(now.Sub(vm.CreatedTime))\\n\\t}\\n\\n\\tfactor /= float64(monthEnd.Sub(monthStart))\\n\\tif factor > 1 {\\n\\t\\tfactor = 1\\n\\t}\\n\\n\\tadditionalBandwidth := int64((factor * float64(vm.Plan.Bandwidth) + 15) * 1024 * 1024 * 1024)\\n\\trows := db.Query(\\\"SELECT id FROM region_bandwidth WHERE region = ? AND user_id = ?\\\", vm.Region, vm.UserId)\\n\\tif rows.Next() {\\n\\t\\tvar rowId int\\n\\t\\trows.Scan(&rowId)\\n\\t\\trows.Close()\\n\\t\\tdb.Exec(\\\"UPDATE region_bandwidth SET bandwidth_additional = bandwidth_additional + ? WHERE id = ?\\\", additionalBandwidth, rowId)\\n\\t} else {\\n\\t\\tdb.Exec(\\\"INSERT INTO region_bandwidth (user_id, region, bandwidth_additional) VALUES (?, ?, ?)\\\", vm.UserId, vm.Region, additionalBandwidth)\\n\\t}\\n}\",\n \"func (b *B) SetBytes(n int64) {}\",\n \"func (n *TreeNode) AddSize(by uint64) {\\n\\tn.mutex.RLock()\\n\\tn.size += by\\n\\tn.mutex.RUnlock()\\n\\treturn\\n}\",\n \"func (cipher *Simon96Cipher) BlockSize() int {\\n\\treturn 12\\n}\",\n \"func (fm FinalModelTimestamp) FBEAllocationSize(value time.Time) int { return fm.FBESize() }\",\n \"func (db *DB) grow(sz int) error {\\n\\t// Ignore if the new size is less than available file size.\\n\\tif sz <= db.filesz {\\n\\t\\treturn nil\\n\\t}\\n\\n\\t// If the data is smaller than the alloc size then only allocate what's needed.\\n\\t// Once it goes over the allocation size then allocate in chunks.\\n\\tif db.datasz <= db.AllocSize {\\n\\t\\tsz = db.datasz\\n\\t} else {\\n\\t\\tsz += db.AllocSize\\n\\t}\\n\\n\\t// Truncate and fsync to ensure file size metadata is flushed.\\n\\t// https://github.com/boltdb/bolt/issues/284\\n\\tif !db.NoGrowSync && !db.readOnly {\\n\\t\\tif runtime.GOOS != \\\"windows\\\" {\\n\\t\\t\\tif err := db.file.Truncate(int64(sz)); err != nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"file resize error: %s\\\", err)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tif err := db.file.Sync(); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"file sync error: %s\\\", err)\\n\\t\\t}\\n\\t\\tif db.Mlock {\\n\\t\\t\\t// unlock old file and lock new one\\n\\t\\t\\tif err := db.mrelock(db.filesz, sz); err != nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"mlock/munlock error: %s\\\", err)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\n\\tdb.filesz = sz\\n\\treturn nil\\n}\",\n \"func (b *BaseNode) updateBytes(n Node) {\\n\\tbw := io.NewBufBinWriter()\\n\\tbw.Grow(1 + n.Size())\\n\\tencodeNodeWithType(n, bw.BinWriter)\\n\\tb.bytes = bw.Bytes()\\n\\tb.bytesValid = true\\n}\",\n \"func (b *bpfHandle) SetReadBufferSize(size int) error {\\n\\tif ok, err := b.ioctlWithInt(ioctlBIOCSBLEN, size); !ok {\\n\\t\\treturn err\\n\\t}\\n\\n\\tnumData := make([]byte, ioctlIntegerSize)\\n\\tb.ioctlWithData(ioctlBIOCGBLEN, numData)\\n\\tif binary.LittleEndian.Uint32(numData) < uint32(size) {\\n\\t\\treturn errors.New(\\\"unsupported buffer size\\\")\\n\\t}\\n\\n\\tb.readBuffer = make([]byte, size)\\n\\treturn nil\\n}\",\n \"func (c *BlockCache) Add(height int, block *walletrpc.CompactBlock) error {\\n\\t// Invariant: m[firstBlock..nextBlock) are valid.\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\tif height > c.nextBlock {\\n\\t\\t// Cache has been reset (for example, checksum error)\\n\\t\\treturn nil\\n\\t}\\n\\tif height < c.firstBlock {\\n\\t\\t// Should never try to add a block before Sapling activation height\\n\\t\\tLog.Fatal(\\\"cache.Add height below Sapling: \\\", height)\\n\\t\\treturn nil\\n\\t}\\n\\tif height < c.nextBlock {\\n\\t\\t// Should never try to \\\"backup\\\" (call Reorg() instead).\\n\\t\\tLog.Fatal(\\\"cache.Add height going backwards: \\\", height)\\n\\t\\treturn nil\\n\\t}\\n\\tbheight := int(block.Height)\\n\\n\\tif bheight != height {\\n\\t\\t// This could only happen if zcashd returned the wrong\\n\\t\\t// block (not the height we requested).\\n\\t\\tLog.Fatal(\\\"cache.Add wrong height: \\\", bheight, \\\" expecting: \\\", height)\\n\\t\\treturn nil\\n\\t}\\n\\n\\t// Add the new block and its length to the db files.\\n\\tdata, err := proto.Marshal(block)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tb := append(checksum(height, data), data...)\\n\\tn, err := c.blocksFile.Write(b)\\n\\tif err != nil {\\n\\t\\tLog.Fatal(\\\"blocks write failed: \\\", err)\\n\\t}\\n\\tif n != len(b) {\\n\\t\\tLog.Fatal(\\\"blocks write incorrect length: expected: \\\", len(b), \\\"written: \\\", n)\\n\\t}\\n\\tb = make([]byte, 4)\\n\\tbinary.LittleEndian.PutUint32(b, uint32(len(data)))\\n\\tn, err = c.lengthsFile.Write(b)\\n\\tif err != nil {\\n\\t\\tLog.Fatal(\\\"lengths write failed: \\\", err)\\n\\t}\\n\\tif n != len(b) {\\n\\t\\tLog.Fatal(\\\"lengths write incorrect length: expected: \\\", len(b), \\\"written: \\\", n)\\n\\t}\\n\\n\\t// update the in-memory variables\\n\\toffset := c.starts[len(c.starts)-1]\\n\\tc.starts = append(c.starts, offset+int64(len(data)+8))\\n\\n\\tif c.latestHash == nil {\\n\\t\\tc.latestHash = make([]byte, len(block.Hash))\\n\\t}\\n\\tcopy(c.latestHash, block.Hash)\\n\\tc.nextBlock++\\n\\t// Invariant: m[firstBlock..nextBlock) are valid.\\n\\treturn nil\\n}\",\n \"func GetFSInfo(path string) (total, available int) {\\n\\ttotal = -1\\n\\tavailable = -1\\n\\tvar buf syscall.Statfs_t\\n\\n\\tif syscall.Statfs(path, &buf) != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\tif buf.Bsize <= 0 {\\n\\t\\treturn\\n\\t}\\n\\n\\ttotal = int((uint64(buf.Bsize) * buf.Blocks) / (1000 * 1000))\\n\\tavailable = int((uint64(buf.Bsize) * buf.Bavail) / (1000 * 1000))\\n\\n\\treturn\\n}\",\n \"func (p *Peer) UpdateLastBlockHeight(newHeight int32) {\\n\\tp.statsMtx.Lock()\\n\\tif newHeight <= p.lastBlock {\\n\\t\\tp.statsMtx.Unlock()\\n\\t\\treturn\\n\\t}\\n\\tlog.Tracef(\\\"Updating last block height of peer %v from %v to %v\\\",\\n\\t\\tp.addr, p.lastBlock, newHeight)\\n\\tp.lastBlock = newHeight\\n\\tp.statsMtx.Unlock()\\n}\",\n \"func (fm FinalModelEnumUInt32) FBEAllocationSize(value EnumUInt32) int { return fm.FBESize() }\",\n \"func convertSize(size uint64) C.ulonglong {\\n\\treturn C.ulonglong(size * 2) // kB to 512-byte sectors\\n}\",\n \"func (mio *Mio) getOptimalBlockSz(bufSz int) (bsz, gsz int) {\\n if bufSz > maxM0BufSz {\\n bufSz = maxM0BufSz\\n }\\n pver := C.m0_pool_version_find(&C.instance.m0c_pools_common,\\n &mio.obj.ob_attr.oa_pver)\\n if pver == nil {\\n log.Panic(\\\"cannot find the object's pool version\\\")\\n }\\n pa := &pver.pv_attr\\n if pa.pa_P < pa.pa_N + pa.pa_K + pa.pa_S {\\n log.Panic(\\\"pool width (%v) is less than the parity group size\\\" +\\n \\\" (%v + %v + %v == %v), check pool parity configuration\\\",\\n pa.pa_P, pa.pa_N, pa.pa_K, pa.pa_S,\\n pa.pa_N + pa.pa_K + pa.pa_S)\\n }\\n\\n usz := int(C.m0_obj_layout_id_to_unit_size(mio.objLid))\\n gsz = usz * int(pa.pa_N) // group size in data units only\\n\\n // bs should be max 4-times pool-width deep counting by 1MB units, or\\n // 8-times deep counting by 512K units, 16-times deep by 256K units,\\n // and so on. Several units to one target will be aggregated to make\\n // fewer network RPCs, disk i/o operations and BE transactions.\\n // For unit sizes of 32K or less, the koefficient (k) is 128, which\\n // makes it 32K * 128 == 4MB - the maximum amount per target when\\n // the performance is still good on LNet (which has max 1MB frames).\\n // XXX: it may be different on libfabric, should be re-measured.\\n k := C.uint(128 / ((usz + 0x7fff) / 0x8000))\\n if k == 0 {\\n k = 1\\n }\\n // P * N / (N + K + S) - number of data units to span the pool-width\\n maxBs := int(k * C.uint(usz) * pa.pa_P * pa.pa_N /\\n (pa.pa_N + pa.pa_K + pa.pa_S))\\n maxBs = roundup(maxBs, gsz) // multiple of group size\\n\\n if bufSz >= maxBs {\\n return maxBs, gsz\\n } else if bufSz <= gsz {\\n return gsz, gsz\\n } else {\\n return roundup(bufSz, gsz), gsz\\n }\\n}\",\n \"func (constr Construction) BlockSize() int { return 16 }\",\n \"func (o KubernetesClusterDefaultNodePoolLinuxOsConfigPtrOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\\n\\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfig) *int {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.SwapFileSizeMb\\n\\t}).(pulumi.IntPtrOutput)\\n}\",\n \"func (b *Block) Size() int {\\n\\treturn len(b.wire)\\n}\",\n \"func (devices *DeviceSet) ResizePool(size int64) error {\\n\\tdirname := devices.loopbackDir()\\n\\tdatafilename := path.Join(dirname, \\\"data\\\")\\n\\tif len(devices.dataDevice) > 0 {\\n\\t\\tdatafilename = devices.dataDevice\\n\\t}\\n\\tmetadatafilename := path.Join(dirname, \\\"metadata\\\")\\n\\tif len(devices.metadataDevice) > 0 {\\n\\t\\tmetadatafilename = devices.metadataDevice\\n\\t}\\n\\n\\tdatafile, err := os.OpenFile(datafilename, os.O_RDWR, 0)\\n\\tif datafile == nil {\\n\\t\\treturn err\\n\\t}\\n\\tdefer datafile.Close()\\n\\n\\tfi, err := datafile.Stat()\\n\\tif fi == nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif fi.Size() > size {\\n\\t\\treturn fmt.Errorf(\\\"devmapper: Can't shrink file\\\")\\n\\t}\\n\\n\\tdataloopback := loopback.FindLoopDeviceFor(datafile)\\n\\tif dataloopback == nil {\\n\\t\\treturn fmt.Errorf(\\\"devmapper: Unable to find loopback mount for: %s\\\", datafilename)\\n\\t}\\n\\tdefer dataloopback.Close()\\n\\n\\tmetadatafile, err := os.OpenFile(metadatafilename, os.O_RDWR, 0)\\n\\tif metadatafile == nil {\\n\\t\\treturn err\\n\\t}\\n\\tdefer metadatafile.Close()\\n\\n\\tmetadataloopback := loopback.FindLoopDeviceFor(metadatafile)\\n\\tif metadataloopback == nil {\\n\\t\\treturn fmt.Errorf(\\\"devmapper: Unable to find loopback mount for: %s\\\", metadatafilename)\\n\\t}\\n\\tdefer metadataloopback.Close()\\n\\n\\t// Grow loopback file\\n\\tif err := datafile.Truncate(size); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"devmapper: Unable to grow loopback file: %s\\\", err)\\n\\t}\\n\\n\\t// Reload size for loopback device\\n\\tif err := loopback.SetCapacity(dataloopback); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"unable to update loopback capacity: %s\\\", err)\\n\\t}\\n\\n\\t// Suspend the pool\\n\\tif err := devicemapper.SuspendDevice(devices.getPoolName()); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"devmapper: Unable to suspend pool: %s\\\", err)\\n\\t}\\n\\n\\t// Reload with the new block sizes\\n\\tif err := devicemapper.ReloadPool(devices.getPoolName(), dataloopback, metadataloopback, devices.thinpBlockSize); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"devmapper: Unable to reload pool: %s\\\", err)\\n\\t}\\n\\n\\t// Resume the pool\\n\\tif err := devicemapper.ResumeDevice(devices.getPoolName()); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"devmapper: Unable to resume pool: %s\\\", err)\\n\\t}\\n\\n\\treturn nil\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.6398401","0.6159081","0.61049414","0.6102407","0.60230005","0.59387565","0.5909034","0.58386946","0.5696438","0.55819315","0.55369323","0.55319744","0.5520192","0.5518587","0.54999447","0.5494222","0.54749113","0.546401","0.54373175","0.5408651","0.5381527","0.53756344","0.53732264","0.53259104","0.5310815","0.5305676","0.5295887","0.52399707","0.52352995","0.5229974","0.5185518","0.5182078","0.5159868","0.5159796","0.51534414","0.5148623","0.51304775","0.51213604","0.51156986","0.5092056","0.5065336","0.50630146","0.50630146","0.50522894","0.505215","0.50508386","0.50464","0.5038033","0.502839","0.50281864","0.50243306","0.50187206","0.5018458","0.49985227","0.49827752","0.4981893","0.49743578","0.49689808","0.49616164","0.4960614","0.49561852","0.4955078","0.49526072","0.49506283","0.49401417","0.49354264","0.4932481","0.49295735","0.49248686","0.49222845","0.49207133","0.49203518","0.49185267","0.49168202","0.49126175","0.49089667","0.49028394","0.49024814","0.48985076","0.48985076","0.48923045","0.4891238","0.48876664","0.4886914","0.48844817","0.4883141","0.48818976","0.48676148","0.48655823","0.48626712","0.48552784","0.48511532","0.48387212","0.48260468","0.48243353","0.4823574","0.4821207","0.48165795","0.48079115","0.48067385"],"string":"[\n \"0.6398401\",\n \"0.6159081\",\n \"0.61049414\",\n \"0.6102407\",\n \"0.60230005\",\n \"0.59387565\",\n \"0.5909034\",\n \"0.58386946\",\n \"0.5696438\",\n \"0.55819315\",\n \"0.55369323\",\n \"0.55319744\",\n \"0.5520192\",\n \"0.5518587\",\n \"0.54999447\",\n \"0.5494222\",\n \"0.54749113\",\n \"0.546401\",\n \"0.54373175\",\n \"0.5408651\",\n \"0.5381527\",\n \"0.53756344\",\n \"0.53732264\",\n \"0.53259104\",\n \"0.5310815\",\n \"0.5305676\",\n \"0.5295887\",\n \"0.52399707\",\n \"0.52352995\",\n \"0.5229974\",\n \"0.5185518\",\n \"0.5182078\",\n \"0.5159868\",\n \"0.5159796\",\n \"0.51534414\",\n \"0.5148623\",\n \"0.51304775\",\n \"0.51213604\",\n \"0.51156986\",\n \"0.5092056\",\n \"0.5065336\",\n \"0.50630146\",\n \"0.50630146\",\n \"0.50522894\",\n \"0.505215\",\n \"0.50508386\",\n \"0.50464\",\n \"0.5038033\",\n \"0.502839\",\n \"0.50281864\",\n \"0.50243306\",\n \"0.50187206\",\n \"0.5018458\",\n \"0.49985227\",\n \"0.49827752\",\n \"0.4981893\",\n \"0.49743578\",\n \"0.49689808\",\n \"0.49616164\",\n \"0.4960614\",\n \"0.49561852\",\n \"0.4955078\",\n \"0.49526072\",\n \"0.49506283\",\n \"0.49401417\",\n \"0.49354264\",\n \"0.4932481\",\n \"0.49295735\",\n \"0.49248686\",\n \"0.49222845\",\n \"0.49207133\",\n \"0.49203518\",\n \"0.49185267\",\n \"0.49168202\",\n \"0.49126175\",\n \"0.49089667\",\n \"0.49028394\",\n \"0.49024814\",\n \"0.48985076\",\n \"0.48985076\",\n \"0.48923045\",\n \"0.4891238\",\n \"0.48876664\",\n \"0.4886914\",\n \"0.48844817\",\n \"0.4883141\",\n \"0.48818976\",\n \"0.48676148\",\n \"0.48655823\",\n \"0.48626712\",\n \"0.48552784\",\n \"0.48511532\",\n \"0.48387212\",\n \"0.48260468\",\n \"0.48243353\",\n \"0.4823574\",\n \"0.4821207\",\n \"0.48165795\",\n \"0.48079115\",\n \"0.48067385\"\n]"},"document_score":{"kind":"string","value":"0.65449804"},"document_rank":{"kind":"string","value":"0"}}},{"rowIdx":104790,"cells":{"query":{"kind":"string","value":"Du stat disk usage like linux du cmd"},"document":{"kind":"string","value":"func Du(paths []string) {\n\tfor _, path := range paths {\n\t\t// check whether path is a normal dir or file.\n\t\tfd, err := os.Open(path)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tfd.Close()\n\n\t\tgetBlkSize(paths[0])\n\t\tResult = PathStatus{}\n\t\tResult.Name = path\n\t\tfilepath.Walk(path, walkFn)\n\t\tfmt.Printf(\"%+v\\n\", Result)\n\t}\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func DiskUsage(path string) (UsageInfo, error) {\n\tvar usage UsageInfo\n\n\tif path == \"\" {\n\t\treturn usage, fmt.Errorf(\"invalid directory\")\n\t}\n\n\t// First check whether the quota system knows about this directory\n\t// A nil quantity or error means that the path does not support quotas\n\t// or xfs_quota tool is missing and we should use other mechanisms.\n\tstartTime := time.Now()\n\tconsumption, _ := fsquota.GetConsumption(path)\n\tif consumption != nil {\n\t\tusage.Bytes = consumption.Value()\n\t\tdefer servermetrics.CollectVolumeStatCalDuration(\"fsquota\", startTime)\n\t} else {\n\t\tdefer servermetrics.CollectVolumeStatCalDuration(\"du\", startTime)\n\t}\n\n\tinodes, _ := fsquota.GetInodes(path)\n\tif inodes != nil {\n\t\tusage.Inodes = inodes.Value()\n\t}\n\n\tif inodes != nil && consumption != nil {\n\t\treturn usage, nil\n\t}\n\n\ttopLevelStat := &unix.Stat_t{}\n\terr := unix.Stat(path, topLevelStat)\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\t// dedupedInode stores inodes that could be duplicates (nlink > 1)\n\tdedupedInodes := make(map[uint64]struct{})\n\n\terr = filepath.Walk(path, func(path string, info os.FileInfo, err error) error {\n\t\t// ignore files that have been deleted after directory was read\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to count inodes for %s: %s\", path, err)\n\t\t}\n\n\t\t// according to the docs, Sys can be nil\n\t\tif info.Sys() == nil {\n\t\t\treturn fmt.Errorf(\"fileinfo Sys is nil\")\n\t\t}\n\n\t\ts, ok := info.Sys().(*syscall.Stat_t)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"unsupported fileinfo; could not convert to stat_t\")\n\t\t}\n\n\t\tif s.Dev != topLevelStat.Dev {\n\t\t\t// don't descend into directories on other devices\n\t\t\treturn filepath.SkipDir\n\t\t}\n\n\t\t// Dedupe hardlinks\n\t\tif s.Nlink > 1 {\n\t\t\tif _, ok := dedupedInodes[s.Ino]; !ok {\n\t\t\t\tdedupedInodes[s.Ino] = struct{}{}\n\t\t\t} else {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\n\t\tif consumption == nil {\n\t\t\tusage.Bytes += int64(s.Blocks) * int64(512) // blocksize in bytes\n\t\t}\n\n\t\tif inodes == nil {\n\t\t\tusage.Inodes++\n\t\t}\n\n\t\treturn nil\n\t})\n\n\treturn usage, err\n}","func DiskUsage(path string) (disk DiskStatus, err error) {\n\tfs := syscall.Statfs_t{}\n\terr = syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}","func DiskUsage(path string) (disk DiskStatus) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}","func DiskUsage(path string) (disk DiskStatus) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}","func DiskUsage(path string) (disk DiskStatus) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}","func DiskUsage(path string) (disk DiskStatus) {\n\n\tif path == \"\" {\n\t\treturn\n\t}\n\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn disk\n}","func Usage(path string) (*Status, error) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"get disk status for path '%s' failed\", path)\n\t}\n\tvar disk Status\n\tdisk.BytesTotal = fs.Blocks * uint64(fs.Bsize)\n\tdisk.BytesFree = fs.Bfree * uint64(fs.Bsize)\n\tdisk.BytesUsed = disk.BytesTotal - disk.BytesFree\n\n\tdisk.InodesTotal = uint64(fs.Files)\n\tdisk.InodesFree = uint64(fs.Ffree)\n\tdisk.InodesUsed = disk.InodesTotal - disk.InodesFree\n\treturn &disk, nil\n}","func du() {\n\trf, err := os.Stat(\"/\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\trdev := rf.Sys().(*syscall.Stat_t).Dev\n\tf := os.DirFS(\"/\")\n\terr = fs.WalkDir(f, \".\", func(p string, d fs.DirEntry, err error) error {\n\t\tif err != nil {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\ti, err := d.Info()\n\t\tif err != nil {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\tsys := i.Sys()\n\t\tif sys == nil {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\tdev := sys.(*syscall.Stat_t).Dev\n\t\tfmt.Printf(\"p=%v size=%v dev=%v\\n\", p, i.Size(), dev)\n\t\tif dev != rdev {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n}","func (c *DefaultChecker) DiskPerf() (info DiskPerfInfo, err error) {\n\tvar data []byte\n\n\texe := execCommand(\"dd\", \"bs=1M\", \"count=256\", \"if=/dev/zero\", fmt.Sprintf(\"of=%v\", path.Join(c.WritePath, \"test\")), \"conv=fdatasync\")\n\tdata, err = exe.CombinedOutput()\n\tgohtypes.PanicIfError(\"Not possible to execute the 'dd' command.\", 500, err)\n\n\tout := string(data)\n\tlogrus.Info(out)\n\n\trate := strings.Split(strings.Split(strings.Replace(out, \"\\n\", \"\", -1), \" s, \")[1], \" \")\n\n\tinfo = DiskPerfInfo{\n\t\tWriteSpeed: rate[0],\n\t\tUnit: rate[1],\n\t}\n\treturn\n}","func DiskUsage(path string) (disk DiskStatus) {\n\th := syscall.MustLoadDLL(\"kernel32.dll\")\n\tc := h.MustFindProc(\"GetDiskFreeSpaceExW\")\n\tlpFreeBytesAvailable := int64(0)\n\tlpTotalNumberOfBytes := int64(0)\n\tlpTotalNumberOfFreeBytes := int64(0)\n\t_, _, err := c.Call(uintptr(unsafe.Pointer(syscall.StringToUTF16Ptr(path[:2]))),\n\t\tuintptr(unsafe.Pointer(&lpFreeBytesAvailable)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfBytes)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfFreeBytes)))\n\tif err != nil {\n\t\tif !strings.Contains(fmt.Sprint(err), \"successfully\") {\n\t\t\tlogger.Error(\"Error during retrieving memory statistic:\", err)\n\t\t\treturn\n\t\t}\n\t}\n\tdisk.All = uint64(lpTotalNumberOfBytes)\n\tdisk.Free = uint64(lpTotalNumberOfFreeBytes)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}","func GetDiskUsage(path string) ([]float64, error) {\n\tvar usage []float64 = make([]float64, 6) // Data to return\n\tvar sc syscall.Statfs_t // Filesystem stat\n\tvar err error // Error catching\n\tvar buffer bytes.Buffer // Buffer for json indent\n\tvar content []byte // Json's content\n\n\t// Get filesystem stat\n\terr = syscall.Statfs(path, &sc)\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\t// Convert structure => json\n\tcontent, err = json.Marshal(sc)\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\t// Indent json\n\tjson.Indent(&buffer, content, \"\", \" \")\n\n\t// Set data to return\n\tusage[0] = float64(sc.Bsize) * float64(sc.Blocks) // TotalStorage\n\tusage[1] = float64(sc.Files) // TotalFileNodes\n\tusage[3] = float64(sc.Bsize) * float64(sc.Bfree) // FreeStorage\n\tusage[2] = usage[0] - usage[3] // UsedStorage\n\tusage[5] = float64(sc.Ffree) // FreeFileNodes\n\tusage[4] = usage[1] - usage[5] // UsedFileNodes\n\n\treturn usage, nil\n}","func DiskUsage(path string) (*DiskStatus, error) {\n\tlpDirectoryName, err := syscall.UTF16PtrFromString(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlpFreeBytesAvailable := int64(0)\n\tlpTotalNumberOfBytes := int64(0)\n\tlpTotalNumberOfFreeBytes := int64(0)\n\t_, _, e := syscall.Syscall6(pGetDiskFreeSpaceEx, 4,\n\t\tuintptr(unsafe.Pointer(lpDirectoryName)),\n\t\tuintptr(unsafe.Pointer(&lpFreeBytesAvailable)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfBytes)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfFreeBytes)), 0, 0)\n\tif e != 0 {\n\t\treturn nil, errors.New(e.Error())\n\t}\n\tstatus := &DiskStatus{\n\t\tAll: lpTotalNumberOfBytes,\n\t\tFree: lpFreeBytesAvailable,\n\t}\n\tstatus.Used = status.All - status.Free\n\treturn status, nil\n}","func GetDiskUsage(rw http.ResponseWriter) error {\n\tu, r := disk.Usage(\"/\")\n\tif r != nil {\n\t\treturn r\n\t}\n\n\treturn share.JSONResponse(u, rw)\n}","func getDiskInfo(path string) (uint64, uint64, error) {\n\ts := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &s)\n\tif err != nil {\n\t\treturn 0,0, err\n\t}\n\treservedBlocks := s.Bfree - s.Bavail\n\n\ttotal := uint64(s.Frsize) * (s.Blocks - reservedBlocks)\n\tfree := uint64(s.Frsize) * s.Bavail\n\t// Check for overflows.\n\t// https://github.com/minio/minio/issues/8035\n\t// XFS can show wrong values at times error out\n\t// in such scenarios.\n\tif free > total {\n\t\treturn 0, 0, fmt.Errorf(\"detected free space (%d) > total disk space (%d), fs corruption at (%s). please run 'fsck'\", free, total, path)\n\t}\n\n\treturn total, free, nil\n}","func Diskuse(path string) (cap uint64, used uint64) {\n\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\tcap = fs.Blocks * uint64(fs.Bsize)\n\tfree := fs.Bfree * uint64(fs.Bsize) // yup, I just did that\n\tused = cap - free\n\treturn cap, used\n}","func (d *Datastore) DiskUsage() (uint64, error) {\n\tvar total uint64 = 0\n\n\terr := d.meta.ListKeys(func(key []byte) error {\n\t\tmeta, err := d.meta.GetMetadata(key)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\ttotal += uint64(meta.StorageSize)\n\t\treturn nil\n\t})\n\n\treturn total, err\n}","func Disk(pt ...bool) ([]*disk.UsageStat, error) {\n\tvar ptBool bool\n\tif len(pt) > 0 {\n\t\tptBool = pt[0]\n\t}\n\n\tvar usage []*disk.UsageStat\n\tparts, err := disk.Partitions(ptBool)\n\n\tfor _, part := range parts {\n\t\tuse, err := disk.Usage(part.Mountpoint)\n\t\tif err != nil {\n\t\t\treturn usage, err\n\t\t}\n\t\tusage = append(usage, use)\n\t\t// printUsage(use)\n\t}\n\n\treturn usage, err\n}","func DiskStat() DiskInfo {\n\tusage, _ := disk.Usage(\"./\")\n\tpartitions, _ := disk.Partitions(true)\n\tiOCounters, _ := disk.IOCounters()\n\n\tue := usage.String()\n\n\tdiskstat := DiskInfo{\n\t\tUsage: ue,\n\t\tPartition: partitions,\n\t\tIOCounters: iOCounters,\n\t}\n\n\treturn diskstat\n}","func PrintDiskStats() {\n\tfmt.Println(DiskStats)\n}","func (p *btrfsPool) Usage() (usage Usage, err error) {\n\tmnt, err := p.Mounted()\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\tvolumes, err := p.Volumes()\n\n\tif err != nil {\n\t\treturn usage, errors.Wrapf(err, \"failed to list pool '%s' volumes\", mnt)\n\t}\n\n\tusage.Size = p.device.Size\n\n\tfor _, volume := range volumes {\n\t\tvol, err := volume.Usage()\n\t\tif err != nil {\n\t\t\treturn Usage{}, errors.Wrapf(err, \"failed to calculate volume '%s' usage\", volume.Path())\n\t\t}\n\n\t\tusage.Used += vol.Used\n\t\tusage.Excl += vol.Excl\n\t}\n\n\treturn\n}","func (s *SigarChecker) GetDiskUsage(path string) (usePercent float64, avail uint64) {\n\tvar fsu sigar.FileSystemUsage\n\n\terr := fsu.Get(path)\n\n\tutil.Must(err)\n\n\treturn fsu.UsePercent(), fsu.Avail\n}","func DirSize(dir string, excludes []string) (int, error) {\n\targs := []string{\"-bs\", dir}\n\tfor _, s := range excludes {\n\t\targs = append(args, fmt.Sprintf(\"--exclude=\\\"%s\\\"\", s))\n\t}\n\n\tcmd := exec.Command(\"/usr/bin/du\", args...)\n\tdata, err := cmd.Output()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn strconv.Atoi(strings.Split(string(data), \"\\t\")[0])\n}","func testDiskUsageDoubleCount(dirFunc mkShardFunc, t *testing.T) {\n\ttemp, cleanup := tempdir(t)\n\tdefer cleanup()\n\tdefer checkTemp(t, temp)\n\n\tfs, err := flatfs.CreateOrOpen(temp, dirFunc(2), false)\n\tif err != nil {\n\t\tt.Fatalf(\"New fail: %v\\n\", err)\n\t}\n\tdefer fs.Close()\n\n\tvar count int\n\tvar wg sync.WaitGroup\n\ttestKey := datastore.NewKey(\"TEST\")\n\n\tput := func() {\n\t\tdefer wg.Done()\n\t\tfor i := 0; i < count; i++ {\n\t\t\tv := []byte(\"10bytes---\")\n\t\t\terr := fs.Put(bg, testKey, v)\n\t\t\tif err != nil {\n\t\t\t\tt.Errorf(\"Put fail: %v\\n\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\tdel := func() {\n\t\tdefer wg.Done()\n\t\tfor i := 0; i < count; i++ {\n\t\t\terr := fs.Delete(bg, testKey)\n\t\t\tif err != nil && !strings.Contains(err.Error(), \"key not found\") {\n\t\t\t\tt.Errorf(\"Delete fail: %v\\n\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Add one element and then remove it and check disk usage\n\t// makes sense\n\tcount = 1\n\twg.Add(2)\n\tput()\n\tdu, _ := fs.DiskUsage(bg)\n\tdel()\n\tdu2, _ := fs.DiskUsage(bg)\n\tif du-10 != du2 {\n\t\tt.Error(\"should have deleted exactly 10 bytes:\", du, du2)\n\t}\n\n\t// Add and remove many times at the same time\n\tcount = 200\n\twg.Add(4)\n\tgo put()\n\tgo del()\n\tgo put()\n\tgo del()\n\twg.Wait()\n\n\tdu3, _ := fs.DiskUsage(bg)\n\thas, err := fs.Has(bg, testKey)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif has { // put came last\n\t\tif du3 != du {\n\t\t\tt.Error(\"du should be the same as after first put:\", du, du3)\n\t\t}\n\t} else { // delete came last\n\t\tif du3 != du2 {\n\t\t\tt.Error(\"du should be the same as after first delete:\", du2, du3)\n\t\t}\n\t}\n}","func getDiskSize(path string) (uint64, error) {\n\tdiskInfo := exec.Command(\"qemu-img\", \"info\", \"--output\", \"json\", path)\n\tstdout, err := diskInfo.StdoutPipe()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tif err := diskInfo.Start(); err != nil {\n\t\treturn 0, err\n\t}\n\ttmpInfo := struct {\n\t\tVirtualSize uint64 `json:\"virtual-size\"`\n\t\tFilename string `json:\"filename\"`\n\t\tClusterSize int64 `json:\"cluster-size\"`\n\t\tFormat string `json:\"format\"`\n\t\tFormatSpecific struct {\n\t\t\tType string `json:\"type\"`\n\t\t\tData map[string]string `json:\"data\"`\n\t\t}\n\t\tDirtyFlag bool `json:\"dirty-flag\"`\n\t}{}\n\tif err := json.NewDecoder(stdout).Decode(&tmpInfo); err != nil {\n\t\treturn 0, err\n\t}\n\tif err := diskInfo.Wait(); err != nil {\n\t\treturn 0, err\n\t}\n\treturn tmpInfo.VirtualSize, nil\n}","func Stats(c *libvirt.Connect, uuid string) error {\n\t//Check exists\n\td, err := c.LookupDomainByUUIDString(uuid)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to lookup: %s\", err)\n\t}\n\n\t//Check is running\n\ts, _, err := d.GetState()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed check state: %s\", err)\n\t}\n\tif s != libvirt.DOMAIN_RUNNING {\n\t\treturn fmt.Errorf(\"domain not running: %d\", s)\n\t}\n\n\tmemStats, err := memStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"STAT: %+v\\n\", memStats)\n\tfmt.Printf(\"STAT Used: %+v\\n\", memStats.Available-memStats.Unused)\n\tfmt.Printf(\"STAT Last: %s\\n\", time.Unix(int64(memStats.LastUpdate), 0))\n\n\tcpuStats, total, err := cpuStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"%+v\\n\", cpuStats)\n\tfmt.Printf(\"Total: %+#v\\n\", total)\n\n\tnetStats, err := netStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"NET: %+v\\n\", netStats)\n\n\t_, dTotal, err := diskStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"DISK: %+v\\n\", dTotal)\n\n\treturn nil\n}","func Stat(path, index string, raw bool) (value string) {\n\tvar row = map[string]int{\"quota\": 3, \"usage\": 2}\n\n\targs := []string{\"qgroup\", \"show\", \"-re\", config.Agent.LxcPrefix}\n\tif raw {\n\t\targs = []string{\"qgroup\", \"show\", \"-re\", \"--raw\", config.Agent.LxcPrefix}\n\t}\n\tout, err := exec.Command(\"btrfs\", args...).Output()\n\tlog.Check(log.FatalLevel, \"Getting btrfs stats\", err)\n\tind := id(path)\n\tscanner := bufio.NewScanner(bytes.NewReader(out))\n\tfor scanner.Scan() {\n\t\tif line := strings.Fields(scanner.Text()); len(line) > 3 && strings.HasSuffix(line[0], \"/\"+ind) {\n\t\t\tvalue = line[row[index]]\n\t\t}\n\t}\n\treturn value\n}","func (f *StorageFingerprint) diskFree(path string) (volume string, total, free uint64, err error) {\n\tabsPath, err := filepath.Abs(path)\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to determine absolute path for %s\", path)\n\t}\n\n\t// Use -k to standardize the output values between darwin and linux\n\tvar dfArgs string\n\tif runtime.GOOS == \"linux\" {\n\t\t// df on linux needs the -P option to prevent linebreaks on long filesystem paths\n\t\tdfArgs = \"-kP\"\n\t} else {\n\t\tdfArgs = \"-k\"\n\t}\n\n\tmountOutput, err := exec.Command(\"df\", dfArgs, absPath).Output()\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to determine mount point for %s\", absPath)\n\t}\n\t// Output looks something like:\n\t//\tFilesystem 1024-blocks Used Available Capacity iused ifree %iused Mounted on\n\t//\t/dev/disk1 487385240 423722532 63406708 87% 105994631 15851677 87% /\n\t//\t[0] volume [1] capacity [2] SKIP [3] free\n\tlines := strings.Split(string(mountOutput), \"\\n\")\n\tif len(lines) < 2 {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse `df` output; expected at least 2 lines\")\n\t}\n\tfields := strings.Fields(lines[1])\n\tif len(fields) < 4 {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse `df` output; expected at least 4 columns\")\n\t}\n\tvolume = fields[0]\n\n\ttotal, err = strconv.ParseUint(fields[1], 10, 64)\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse storage.bytestotal size in kilobytes\")\n\t}\n\t// convert to bytes\n\ttotal *= 1024\n\n\tfree, err = strconv.ParseUint(fields[3], 10, 64)\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse storage.bytesfree size in kilobytes\")\n\t}\n\t// convert to bytes\n\tfree *= 1024\n\n\treturn volume, total, free, nil\n}","func GetDiskUsageStats(path string) (uint64, uint64, error) {\n\tvar dirSize, inodeCount uint64\n\n\terr := filepath.Walk(path, func(path string, info os.FileInfo, err error) error {\n\t\tfileStat, error := os.Lstat(path)\n\t\tif error != nil {\n\t\t\tif fileStat.Mode()&os.ModeSymlink != 0 {\n\t\t\t\t// Is a symlink; no error should be returned\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn error\n\t\t}\n\n\t\tdirSize += uint64(info.Size())\n\t\tinodeCount++\n\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\n\treturn dirSize, inodeCount, err\n}","func (v *btrfsVolume) Usage() (usage Usage, err error) {\n\tctx := context.TODO()\n\tinfo, err := v.utils.SubvolumeInfo(ctx, v.Path())\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\tgroups, err := v.utils.QGroupList(ctx, v.Path())\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\tgroup, ok := groups[fmt.Sprintf(\"0/%d\", info.ID)]\n\tif !ok {\n\t\t// no qgroup associated with the subvolume id! means no limit, but we also\n\t\t// cannot read the usage.\n\t\treturn\n\t}\n\n\t// used is basically amount of space reserved for this\n\t// volume. We assume that's total usage of the volume\n\tused := group.MaxRfer\n\n\tif used == 0 {\n\t\t// in case no limit is set on the subvolume, we assume\n\t\t// it's size is the size of the files on that volumes\n\t\t// or a special case when the volume is a zdb volume\n\t\tused, err = volumeUsage(v.Path())\n\t\tif err != nil {\n\t\t\treturn usage, errors.Wrap(err, \"failed to get subvolume usage\")\n\t\t}\n\t}\n\n\treturn Usage{Used: used, Size: group.MaxRfer, Excl: group.Excl}, nil\n}","func (f *Fs) About(ctx context.Context) (*fs.Usage, error) {\n\tvar available, total, free int64\n\troot, e := syscall.UTF16PtrFromString(f.root)\n\tif e != nil {\n\t\treturn nil, fmt.Errorf(\"failed to read disk usage: %w\", e)\n\t}\n\t_, _, e1 := getFreeDiskSpace.Call(\n\t\tuintptr(unsafe.Pointer(root)),\n\t\tuintptr(unsafe.Pointer(&available)), // lpFreeBytesAvailable - for this user\n\t\tuintptr(unsafe.Pointer(&total)), // lpTotalNumberOfBytes\n\t\tuintptr(unsafe.Pointer(&free)), // lpTotalNumberOfFreeBytes\n\t)\n\tif e1 != syscall.Errno(0) {\n\t\treturn nil, fmt.Errorf(\"failed to read disk usage: %w\", e1)\n\t}\n\tusage := &fs.Usage{\n\t\tTotal: fs.NewUsageValue(total), // quota of bytes that can be used\n\t\tUsed: fs.NewUsageValue(total - free), // bytes in use\n\t\tFree: fs.NewUsageValue(available), // bytes which can be uploaded before reaching the quota\n\t}\n\treturn usage, nil\n}","func print_stats(){\nfmt.Print(\"\\nMemory usage statistics:\\n\")\nfmt.Printf(\"%v names\\n\",len(name_dir))\nfmt.Printf(\"%v replacement texts\\n\",len(text_info))\n}","func printIostat(v *gocui.View, s stat.Diskstats) error {\n\t// print header\n\t_, err := fmt.Fprintf(v, \"\\033[30;47m Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await r_await w_await %%util\\033[0m\\n\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor i := 0; i < len(s); i++ {\n\t\t// skip devices which never do IOs\n\t\tif s[i].Completed == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// print stats\n\t\t_, err := fmt.Fprintf(v, \"%20s\\t%10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f\\n\",\n\t\t\ts[i].Device,\n\t\t\ts[i].Rmerged, s[i].Wmerged, s[i].Rcompleted, s[i].Wcompleted,\n\t\t\ts[i].Rsectors, s[i].Wsectors, s[i].Arqsz, s[i].Tweighted,\n\t\t\ts[i].Await, s[i].Rawait, s[i].Wawait, s[i].Util,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}","func stat(t *kernel.Task, d *fs.Dirent, dirPath bool, statAddr usermem.Addr) error {\n\tif dirPath && !fs.IsDir(d.Inode.StableAttr) {\n\t\treturn syserror.ENOTDIR\n\t}\n\tuattr, err := d.Inode.UnstableAttr(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\ts := statFromAttrs(t, d.Inode.StableAttr, uattr)\n\treturn s.CopyOut(t, statAddr)\n}","func statfsImpl(t *kernel.Task, d *fs.Dirent, addr usermem.Addr) error {\n\tinfo, err := d.Inode.StatFS(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Construct the statfs structure and copy it out.\n\tstatfs := linux.Statfs{\n\t\tType: info.Type,\n\t\t// Treat block size and fragment size as the same, as\n\t\t// most consumers of this structure will expect one\n\t\t// or the other to be filled in.\n\t\tBlockSize: d.Inode.StableAttr.BlockSize,\n\t\tBlocks: info.TotalBlocks,\n\t\t// We don't have the concept of reserved blocks, so\n\t\t// report blocks free the same as available blocks.\n\t\t// This is a normal thing for filesystems, to do, see\n\t\t// udf, hugetlbfs, tmpfs, among others.\n\t\tBlocksFree: info.FreeBlocks,\n\t\tBlocksAvailable: info.FreeBlocks,\n\t\tFiles: info.TotalFiles,\n\t\tFilesFree: info.FreeFiles,\n\t\t// Same as Linux for simple_statfs, see fs/libfs.c.\n\t\tNameLength: linux.NAME_MAX,\n\t\tFragmentSize: d.Inode.StableAttr.BlockSize,\n\t\t// Leave other fields 0 like simple_statfs does.\n\t}\n\t_, err = t.CopyOut(addr, &statfs)\n\treturn err\n}","func TestDiskUsage(t *testing.T) {\n\tt.Parallel()\n\twinners := []parameters{[]string{\"/\", \"99\"}, []string{\"/\", \"98\"}}\n\tlosers := []parameters{[]string{\"/\", \"1\"}, []string{\"/\", \"2\"}}\n\ttestInputs(t, diskUsage, winners, losers)\n}","func (fsys *FS) stat(node mountlib.Node, stat *fuse.Stat_t) (errc int) {\n\tvar Size uint64\n\tvar Blocks uint64\n\tvar modTime time.Time\n\tvar Mode os.FileMode\n\tswitch x := node.(type) {\n\tcase *mountlib.Dir:\n\t\tmodTime = x.ModTime()\n\t\tMode = mountlib.DirPerms | fuse.S_IFDIR\n\tcase *mountlib.File:\n\t\tvar err error\n\t\tmodTime, Size, Blocks, err = x.Attr(mountlib.NoModTime)\n\t\tif err != nil {\n\t\t\treturn translateError(err)\n\t\t}\n\t\tMode = mountlib.FilePerms | fuse.S_IFREG\n\t}\n\t//stat.Dev = 1\n\tstat.Ino = node.Inode() // FIXME do we need to set the inode number?\n\tstat.Mode = uint32(Mode)\n\tstat.Nlink = 1\n\tstat.Uid = mountlib.UID\n\tstat.Gid = mountlib.GID\n\t//stat.Rdev\n\tstat.Size = int64(Size)\n\tt := fuse.NewTimespec(modTime)\n\tstat.Atim = t\n\tstat.Mtim = t\n\tstat.Ctim = t\n\tstat.Blksize = 512\n\tstat.Blocks = int64(Blocks)\n\tstat.Birthtim = t\n\t// fs.Debugf(nil, \"stat = %+v\", *stat)\n\treturn 0\n}","func Info(path string) (int64, int64, int64, int64, int64, int64, error) {\n\tstatfs := &unix.Statfs_t{}\n\terr := unix.Statfs(path, statfs)\n\tif err != nil {\n\t\treturn 0, 0, 0, 0, 0, 0, err\n\t}\n\n\t// Available is blocks available * fragment size\n\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\n\n\t// Capacity is total block count * fragment size\n\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\n\n\t// Usage is block being used * fragment size (aka block size).\n\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\n\n\tinodes := int64(statfs.Files)\n\tinodesFree := int64(statfs.Ffree)\n\tinodesUsed := inodes - inodesFree\n\n\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\n}","func getFSSizeMb(f *framework.Framework, pod *v1.Pod) (int64, error) {\n\tvar output string\n\tvar err error\n\n\tif supervisorCluster {\n\t\tnamespace := getNamespaceToRunTests(f)\n\t\tcmd := []string{\"exec\", pod.Name, \"--namespace=\" + namespace, \"--\", \"/bin/sh\", \"-c\", \"df -Tkm | grep /mnt/volume1\"}\n\t\toutput = framework.RunKubectlOrDie(namespace, cmd...)\n\t\tgomega.Expect(strings.Contains(output, ext4FSType)).NotTo(gomega.BeFalse())\n\t} else {\n\t\toutput, _, err = fvolume.PodExec(f, pod, \"df -T -m | grep /mnt/volume1\")\n\t\tif err != nil {\n\t\t\treturn -1, fmt.Errorf(\"unable to find mount path via `df -T`: %v\", err)\n\t\t}\n\t}\n\n\tarrMountOut := strings.Fields(string(output))\n\tif len(arrMountOut) <= 0 {\n\t\treturn -1, fmt.Errorf(\"error when parsing output of `df -T`. output: %s\", string(output))\n\t}\n\tvar devicePath, strSize string\n\tdevicePath = arrMountOut[0]\n\tif devicePath == \"\" {\n\t\treturn -1, fmt.Errorf(\"error when parsing output of `df -T` to find out devicePath of /mnt/volume1. output: %s\",\n\t\t\tstring(output))\n\t}\n\tstrSize = arrMountOut[2]\n\tif strSize == \"\" {\n\t\treturn -1, fmt.Errorf(\"error when parsing output of `df -T` to find out size of /mnt/volume1: output: %s\",\n\t\t\tstring(output))\n\t}\n\n\tintSizeInMb, err := strconv.ParseInt(strSize, 10, 64)\n\tif err != nil {\n\t\treturn -1, fmt.Errorf(\"failed to parse size %s into int size\", strSize)\n\t}\n\n\treturn intSizeInMb, nil\n}","func getMemUsage(memStat types.MemoryStats) uint64 {\n\t// Version 1 of the Linux cgroup API uses total_inactive_file\n\tif v, ok := memStat.Stats[\"total_inactive_file\"]; ok && v < memStat.Usage {\n\t\treturn memStat.Usage - v\n\t}\n\n\t// Version 2 of the Linux cgroup API uses inactive_file\n\tif v := memStat.Stats[\"inactive_file\"]; v < memStat.Usage {\n\t\treturn memStat.Usage - v\n\t}\n\n\treturn memStat.Usage\n}","func DiskUsed() (uint64, error) {\n\tusage, err := Disk()\n\t// for i := 0; i < len(usage); i++ {\n\tif len(usage) > 0 {\n\t\tuseDisk := usage[0].Used\n\t\treturn useDisk, err\n\t}\n\n\treturn 0, err\n}","func stat(name string) (*info, error) {\n\tf, err := os.Open(filepath.Clean(name))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar di info\n\n\tvar buf bytes.Buffer\n\n\t_, err = io.CopyN(&buf, f, int64(binary.Size(di.Header)))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfi, err := f.Stat()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = f.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = binary.Read(&buf, binary.LittleEndian, &di.Header)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif di.Header.MagicNumber != 0x564d444b { // SPARSE_MAGICNUMBER\n\t\treturn nil, ErrInvalidFormat\n\t}\n\n\tif di.Header.Flags&(1<<16) == 0 { // SPARSEFLAG_COMPRESSED\n\t\t// Needs to be converted, for example:\n\t\t// vmware-vdiskmanager -r src.vmdk -t 5 dst.vmdk\n\t\t// qemu-img convert -O vmdk -o subformat=streamOptimized src.vmdk dst.vmdk\n\t\treturn nil, ErrInvalidFormat\n\t}\n\n\tdi.Capacity = di.Header.Capacity * 512 // VMDK_SECTOR_SIZE\n\tdi.Size = fi.Size()\n\tdi.Name = filepath.Base(name)\n\tdi.ImportName = strings.TrimSuffix(di.Name, \".vmdk\")\n\n\treturn &di, nil\n}","func (c *Cache) updateStats() error {\n\tvar newUsed int64\n\terr := c.walk(func(osPath string, fi os.FileInfo, name string) error {\n\t\tif !fi.IsDir() {\n\t\t\t// Update the atime with that of the file\n\t\t\tatime := times.Get(fi).AccessTime()\n\t\t\tc.updateStat(name, atime, fi.Size())\n\t\t\tnewUsed += fi.Size()\n\t\t} else {\n\t\t\tc.cacheDir(name)\n\t\t}\n\t\treturn nil\n\t})\n\tc.itemMu.Lock()\n\tc.used = newUsed\n\tc.itemMu.Unlock()\n\treturn err\n}","func (a *app) printStat() {\n\tfmt.Printf(\"Total capacity %vgb \\n\", a.sq.Limit/1024/1024/1024)\n\tfmt.Printf(\"Usage %vmb \\n\", a.sq.Usage/1024/1024)\n\tfmt.Printf(\"In Drive %vmb \\n\", a.sq.UsageInDrive/1024/1024)\n\tfmt.Printf(\"In Trash %vmb \\n\", a.sq.UsageInDriveTrash/1024/1024)\n\n\tperOfUse := float64(a.sq.UsageInDrive) / float64(a.sq.Limit) * 100\n\tfmt.Println(\"Current percentage of use:\", perOfUse)\n}","func (s *SystemMetrics) GetDiskIOStats(c chan *DiskStats) {\n\tbefore, err := disk.Get()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\ttime.Sleep(time.Duration(time.Second * 3))\n\n\tnow, err := disk.Get()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tc <- &DiskStats{\n\t\t// default is always in bytes. hence, convert into the required format.\n\t\tDiskIO: (int(now.Used) - int(before.Used)) / 1000, // in kilo-bytes\n\t\tCached: int(now.Cached / 1000000), // mega-bytes\n\t}\n}","func DiskTotal() (uint64, error) {\n\tusage, err := Disk()\n\t// for i := 0; i < len(usage); i++ {\n\tif len(usage) > 0 {\n\t\ttotalDisk := usage[0].Total\n\t\treturn totalDisk, err\n\t}\n\n\treturn 0, err\n}","func GetFSInfo(ch chan metrics.Metric) {\n\tmountedFS, err := disk.Partitions(false)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"collector\": \"filesystem\",\n\t\t\t\"os\": \"linux\",\n\t\t\t\"action\": \"GetFileSystems\",\n\t\t}).Errorf(\"Unable to find mounted filesystems: %+v\", err)\n\t}\n\tfor _, FSs := range mountedFS {\n\t\tfsStats, err := disk.Usage(FSs.Mountpoint)\n\t\tif err != nil {\n\t\t\tlog.WithFields(log.Fields{\n\t\t\t\t\"collector\": \"filesystem\",\n\t\t\t\t\"os\": \"linux\",\n\t\t\t\t\"action\": \"GetFSStats\",\n\t\t\t}).Errorf(\"Unable to get stats from mounted filesystem: %+v\", err)\n\t\t}\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"collector\": \"filesystem\",\n\t\t\t\"os\": \"linux\",\n\t\t}).Debug(fsStats)\n\t\tfsStat := metrics.FileSystem{\n\t\t\tFileSystem: *fsStats,\n\t\t}\n\t\tch <- fsStat\n\t}\n}","func (i *ImageService) LayerDiskUsage(ctx context.Context) (int64, error) {\n\tvar allLayersSize int64\n\t// TODO(thaJeztah): do we need to take multiple snapshotters into account? See https://github.com/moby/moby/issues/45273\n\tsnapshotter := i.client.SnapshotService(i.snapshotter)\n\tsnapshotter.Walk(ctx, func(ctx context.Context, info snapshots.Info) error {\n\t\tusage, err := snapshotter.Usage(ctx, info.Name)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tallLayersSize += usage.Size\n\t\treturn nil\n\t})\n\treturn allLayersSize, nil\n}","func DiskPercent() (string, error) {\n\tusage, err := Disk()\n\tif len(usage) > 0 {\n\t\tuseDisk := fmt.Sprintf(\"%.2f\", usage[0].UsedPercent)\n\t\treturn useDisk, err\n\t}\n\n\treturn \"0.00\", err\n}","func getRemainingDiskSpace(ctxPtr *volumemgrContext) (uint64, error) {\n\n\tvar totalDiskSize uint64\n\n\tpubContentTree := ctxPtr.pubContentTreeStatus\n\titemsContentTree := pubContentTree.GetAll()\n\tfor _, iterContentTreeStatusJSON := range itemsContentTree {\n\t\titerContentTreeStatus := iterContentTreeStatusJSON.(types.ContentTreeStatus)\n\t\tif iterContentTreeStatus.State < types.LOADED {\n\t\t\tlog.Tracef(\"Content tree %s State %d < LOADED\",\n\t\t\t\titerContentTreeStatus.Key(), iterContentTreeStatus.State)\n\t\t\tcontinue\n\t\t}\n\t\ttotalDiskSize += uint64(iterContentTreeStatus.CurrentSize)\n\t}\n\n\tpubVolume := ctxPtr.pubVolumeStatus\n\titemsVolume := pubVolume.GetAll()\n\tfor _, iterVolumeStatusJSON := range itemsVolume {\n\t\titerVolumeStatus := iterVolumeStatusJSON.(types.VolumeStatus)\n\t\t// we start consume space when moving into CREATING_VOLUME state\n\t\tif iterVolumeStatus.State < types.CREATING_VOLUME {\n\t\t\tlog.Tracef(\"Volume %s State %d < CREATING_VOLUME\",\n\t\t\t\titerVolumeStatus.Key(), iterVolumeStatus.State)\n\t\t\tcontinue\n\t\t}\n\t\ttotalDiskSize += volumehandlers.GetVolumeHandler(log, ctxPtr, &iterVolumeStatus).UsageFromStatus()\n\t}\n\tdeviceDiskUsage, err := diskmetrics.PersistUsageStat(log)\n\tif err != nil {\n\t\terr := fmt.Errorf(\"Failed to get diskUsage for /persist. err: %s\", err)\n\t\tlog.Error(err)\n\t\treturn 0, err\n\t}\n\tdeviceDiskSize := deviceDiskUsage.Total\n\tdiskReservedForDom0 := diskmetrics.Dom0DiskReservedSize(log, ctxPtr.globalConfig, deviceDiskSize)\n\tvar allowedDeviceDiskSize uint64\n\tif deviceDiskSize < diskReservedForDom0 {\n\t\terr = fmt.Errorf(\"Total Disk Size(%d) <= diskReservedForDom0(%d)\",\n\t\t\tdeviceDiskSize, diskReservedForDom0)\n\t\tlog.Errorf(\"***getRemainingDiskSpace: err: %s\", err)\n\t\treturn uint64(0), err\n\t}\n\tallowedDeviceDiskSize = deviceDiskSize - diskReservedForDom0\n\tif allowedDeviceDiskSize < totalDiskSize {\n\t\treturn 0, nil\n\t} else {\n\t\treturn allowedDeviceDiskSize - totalDiskSize, nil\n\t}\n}","func (s *Server) statproc() {\n\tvar (\n\t\tv *volume.Volume\n\t\tolds *stat.Stats\n\t\tnews = new(stat.Stats)\n\t)\n\tfor {\n\t\tolds = s.info.Stats\n\t\t*news = *olds\n\t\ts.info.Stats = news // use news instead, for current display\n\t\tolds.Reset()\n\t\tfor _, v = range s.store.Volumes {\n\t\t\tv.Stats.Calc()\n\t\t\tolds.Merge(v.Stats)\n\t\t}\n\t\tolds.Calc()\n\t\ts.info.Stats = olds\n\t\ttime.Sleep(statDuration)\n\t}\n}","func ShowDiskStatus(c *gin.Context) {\n\tdiskinfo, err := diskInfo()\n\tif err != nil {\n\t\tlog.Fatal(\"Get Disk information error: \", err.Error())\n\t}\n\tutils.Render(\n\t\tc,\n\t\tgin.H{\n\t\t\t\"title\": \"System Status\",\n\t\t\t\"Name\": \"System base information\",\n\t\t\t\"payload\": diskinfo,\n\t\t},\n\t\t\"stat.html\",\n\t)\n}","func TestParseDiskUsageBasic(t *testing.T) {\n\tcontent, err := ioutil.ReadFile(\"../test_resources/stdout\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\ts := string(content)\n\n\tdiskUsage, _ := parseDiskUsage(&s)\n\tif diskUsage != 1111.0 {\n\t\tt.Errorf(\"json value returned unexpected: got %f want %f\",\n\t\t\tdiskUsage, 1111.0)\n\t}\n}","func DiskFree() (uint64, error) {\n\tusage, err := Disk()\n\t// for i := 0; i < len(usage); i++ {\n\tif len(usage) > 0 {\n\t\tfreeDisk := usage[0].Free\n\t\treturn freeDisk, err\n\t}\n\n\treturn 0, err\n}","func (c *CryptohomeBinary) GetAccountDiskUsage(ctx context.Context, username string) ([]byte, error) {\n\treturn c.call(ctx, \"--action=get_account_disk_usage\", \"--user=\"+username)\n}","func getHDDString() string {\n\tpartitions, _ := disk.Partitions(false)\n\tbiggestDiskSize := uint64(0)\n\tbiggestDiskUsed := uint64(0)\n\tbiggestDiskName := \"\"\n\tfor _, partition := range partitions {\n\t\td, _ := disk.Usage(partition.Mountpoint)\n\t\tif d.Total > biggestDiskSize {\n\t\t\tbiggestDiskName = partition.Mountpoint\n\t\t\tbiggestDiskUsed = d.Used\n\t\t\tbiggestDiskSize = d.Total\n\t\t}\n\t}\n\treturn fmt.Sprintf(\"%v: %v/%v\", biggestDiskName, formatSize(biggestDiskUsed, 1000), formatSize(biggestDiskSize, 1000))\n}","func GetDiskStats(disks cmn.SimpleKVs) DiskStats {\n\tif len(disks) < largeNumDisks {\n\t\toutput := make(DiskStats, len(disks))\n\n\t\tfor disk := range disks {\n\t\t\tstat, ok := readSingleDiskStat(disk)\n\t\t\tif !ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\toutput[disk] = stat\n\t\t}\n\t\treturn output\n\t}\n\n\treturn readMultipleDiskStats(disks)\n}","func (fs *FS) fsInfo(ctx context.Context, path string) (int64, int64, int64, int64, int64, int64, error) {\n\tstatfs := &unix.Statfs_t{}\n\terr := unix.Statfs(path, statfs)\n\tif err != nil {\n\t\treturn 0, 0, 0, 0, 0, 0, err\n\t}\n\n\t// Available is blocks available * fragment size\n\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\n\n\t// Capacity is total block count * fragment size\n\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\n\n\t// Usage is block being used * fragment size (aka block size).\n\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\n\n\tinodes := int64(statfs.Files)\n\tinodesFree := int64(statfs.Ffree)\n\tinodesUsed := inodes - inodesFree\n\n\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\n}","func (c *Container) statInsideMount(containerPath string) (*copier.StatForItem, string, string, error) {\n\tresolvedRoot := \"/\"\n\tresolvedPath := c.pathAbs(containerPath)\n\tvar statInfo *copier.StatForItem\n\n\terr := c.joinMountAndExec(\n\t\tfunc() error {\n\t\t\tvar statErr error\n\t\t\tstatInfo, statErr = secureStat(resolvedRoot, resolvedPath)\n\t\t\treturn statErr\n\t\t},\n\t)\n\n\treturn statInfo, resolvedRoot, resolvedPath, err\n}","func RunListDisk() {\n\n\t// dir, err := filepath.Abs(filepath.Dir(os.Args[0]))\n\t// if err != nil {\n\t// \tlog.Fatal(err)\n\t// \treturn\n\t// }\n\n\t// lsscsipath := path.Join(dir, \"lsscsi\")\n\t// if _, err := os.Stat(lsscsipath); os.IsNotExist(err) {\n\t// \tlsscsipath = \"lsscsi\"\n\t// }\n\tlsscsipath := \"lsscsi\"\n\tcmd := exec.Command(lsscsipath, \"-s\", \"-g\")\n\tstdout, err := cmd.StdoutPipe()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif err := cmd.Start(); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\ttimer := time.AfterFunc(10*time.Second, func() {\n\t\tcmd.Process.Kill()\n\t})\n\n\tscanner := bufio.NewScanner(stdout)\n\tvar hddinfo []string\n\tvar hddchanged bool\n\tvar wg sync.WaitGroup\n\tfor scanner.Scan() {\n\t\tss := scanner.Text()\n\t\tfmt.Println(ss)\n\t\thddinfo = append(hddinfo, ss)\n\t\tif !DetectData.MatchKey(ss) {\n\t\t\thddchanged = true\n\t\t}\n\t\tif !DetectData.ContainsKey(ss) {\n\t\t\t//\\s Matches any white-space character.\n\t\t\tr := regexp.MustCompile(`^([\\s\\S]{13})(disk[\\s\\S]{4})([\\s\\S]{9})([\\s\\S]{17})([\\s\\S]{6})([\\s\\S]{11})([\\s\\S]{11})([\\s\\S]+)$`)\n\t\t\tdiskinfos := r.FindStringSubmatch(ss)\n\t\t\tif len(diskinfos) == 9 {\n\t\t\t\tvar dddect = NewSyncDataDetect()\n\t\t\t\tdddect.detectHDD.Locpath = strings.Trim(diskinfos[1], \" \")\n\t\t\t\tdddect.detectHDD.Type = strings.Trim(diskinfos[2], \" \")\n\t\t\t\tdddect.detectHDD.Manufacture = strings.Trim(diskinfos[3], \" \")\n\t\t\t\tdddect.detectHDD.Model = strings.Trim(diskinfos[4], \" \")\n\t\t\t\tdddect.detectHDD.Version = strings.Trim(diskinfos[5], \" \")\n\t\t\t\tdddect.detectHDD.LinuxName = strings.Trim(diskinfos[6], \" \")\n\t\t\t\tdddect.detectHDD.SGLibName = strings.Trim(diskinfos[7], \" \")\n\t\t\t\tdddect.detectHDD.Size = strings.Trim(diskinfos[8], \" \")\n\n\t\t\t\tif strings.Index(dddect.detectHDD.LinuxName, `/dev/`) == -1 {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\t//hddchanged = true\n\t\t\t\tDetectData.AddValue(ss, dddect)\n\t\t\t\twg.Add(1)\n\t\t\t\tgo dddect.ReadDataFromSmartCtl(&wg)\n\t\t\t}\n\t\t} else {\n\t\t\tif vv, ok := DetectData.Get(ss); ok {\n\t\t\t\tif len(vv.detectHDD.UILabel) == 0 && len(vv.detectHDD.Otherinfo) == 0 {\n\t\t\t\t\twg.Add(1)\n\t\t\t\t\tgo vv.ReadDataFromSmartCtl(&wg)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\ttimer.Stop()\n\tDetectData.RemoveOld(hddinfo)\n\n\ttime.Sleep(4 * time.Second)\n\n\tif hddchanged {\n\t\tfmt.Print(\"changed!\")\n\t\tcclist, err := configxmldata.Conf.GetCardListIndex()\n\t\tif err == nil {\n\t\t\tfor _, i := range cclist {\n\t\t\t\twg.Add(1)\n\t\t\t\tgo SASHDDinfo.RunCardInfo(i, &wg)\n\t\t\t}\n\t\t}\n\t\tfor i := 0; i < 30; i++ {\n\t\t\tif waitTimeout(&wg, 10*time.Second) {\n\t\t\t\tfmt.Println(\"Timed out waiting for wait group\")\n\t\t\t\tMergeCalibration()\n\t\t\t} else {\n\t\t\t\tfmt.Println(\"Wait group finished\")\n\t\t\t\tMergeCalibration()\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t} else {\n\t\twaitTimeout(&wg, 300*time.Second)\n\t}\n\n}","func (o *OSS) ReturnSize(groupID int64) error {\n\n partLine := partLine()\n\n totalData := map[string]map[string]int{}\n wg := &sync.WaitGroup{}\n ch := make(chan base.BaseInfo, 1000)\n wg.Add(2)\n go register(groupID, ch, wg)\n go fileCalc(groupID, ch, wg, o, totalData)\n\n time.Sleep(2 * time.Second)\n wg.Wait()\n\n for t := range totalData {\n ts := strconv.Itoa(totalData[t][\"totalSize\"])\n\n write.CreateFile(t, partLine + \"\\n\")\n write.CreateFile(t, fmt.Sprintf(\"Total: RecordCount: %d ; FileCount: %d ; FileSize: %s .\\n\",totalData[t][\"RecordCount\"],totalData[t][\"totalCount\"], utils.FormatSize(ts) ))\n }\n return nil\n}","func umountProc(syncFD int) {\n\tsyncFile := os.NewFile(uintptr(syncFD), \"procfs umount sync FD\")\n\tbuf := make([]byte, 1)\n\tif w, err := syncFile.Write(buf); err != nil || w != 1 {\n\t\tutil.Fatalf(\"unable to write into the proc umounter descriptor: %v\", err)\n\t}\n\tsyncFile.Close()\n\n\tvar waitStatus unix.WaitStatus\n\tif _, err := unix.Wait4(0, &waitStatus, 0, nil); err != nil {\n\t\tutil.Fatalf(\"error waiting for the proc umounter process: %v\", err)\n\t}\n\tif !waitStatus.Exited() || waitStatus.ExitStatus() != 0 {\n\t\tutil.Fatalf(\"the proc umounter process failed: %v\", waitStatus)\n\t}\n\tif err := unix.Access(\"/proc/self\", unix.F_OK); err != unix.ENOENT {\n\t\tutil.Fatalf(\"/proc is still accessible\")\n\t}\n}","func Space(inputs SpaceInputs) error {\n\tfilesystemPath := inputs.FilesystemPath\n\tdb[\"filter\"] = inputs.Filter\n\tvar err error\n\tif !verify(filesystemPath) {\n\t\terr = errors.New(\"Path is not abosule path\")\n\t\treturn err\n\t}\n\tstart := time.Now()\n\terr = filepath.Walk(filesystemPath, process)\n\troot := getDiskUsage(\"/\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tt := time.Now()\n\telapsed := t.Sub(start)\n\telapsedInMin := elapsed.Minutes()\n\tlargeFileProcessed, _ := db[largeFilenameKey]\n\tlargeSizeProcessed, _ := db[largeSizeKey]\n\ttotalSizeProcessed, _ := db[totalSizeKey]\n\tlargeSize, _ := strconv.ParseUint(largeSizeProcessed, base, bitSize)\n\ttotalSize, _ := strconv.ParseUint(totalSizeProcessed, base, bitSize)\n\tif root.all > 0 {\n\t\ttotalSize = root.all\n\t}\n\tlargeFileSizePercentage := (float64(largeSize) / float64(totalSize)) * 100.00\n\tdata := [][]string{\n\t\t[]string{\"LARGE_FILE_NAME\", largeFileProcessed, \"NA\"},\n\t\t[]string{\"LARGE_FILE_SIZE\", bytefmt.ByteSize(largeSize), strconv.FormatFloat(largeFileSizePercentage, 'f', floatPrec, bitSize)},\n\t\t[]string{\"DISK_TOTAL_SIZE\", bytefmt.ByteSize(root.all), \"NA\"},\n\t\t[]string{\"DISK_USED_PERCENTAGE\", \"--\", strconv.FormatFloat(root.usedPer, 'f', floatPrec, bitSize)},\n\t\t[]string{\"PROCESSING_TIME\", strconv.FormatFloat(elapsedInMin, 'f', floatPrec, bitSize) + \" min(s)\", \"NA\"},\n\t}\n\tviews.Print(data)\n\treturn err\n}","func (b *Bucket) Statd(path string) (file *File) {\n\treq, err := http.NewRequest(\"HEAD\", b.URL+path+\" d\", nil)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tresp, err := b.run(req, 0)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tresp.Body.Close()\n\tif resp.StatusCode == http.StatusOK {\n\t\tfile = &File{\n\t\t\tPath: path,\n\t\t\tLastModified: resp.Header.Get(\"Last-Modified\"),\n\t\t\tETag: resp.Header.Get(\"ETag\"),\n\t\t\tSize: 0,\n\t\t\tIsDir: true,\n\t\t}\n\t}\n\treturn\n}","func (filterdev *NetworkTap) Stats() (*syscall.BpfStat, error) {\n\tvar s syscall.BpfStat\n\t_, _, err := syscall.Syscall(syscall.SYS_IOCTL, uintptr(filterdev.device.Fd()), syscall.BIOCGSTATS, uintptr(unsafe.Pointer(&s)))\n\tif err != 0 {\n\t\treturn nil, syscall.Errno(err)\n\t}\n\treturn &s, nil\n}","func Deserialize(p []byte) *structs.DiskStats {\n\tstts := &structs.DiskStats{}\n\tdevF := &flat.Device{}\n\tstatsFlat := flat.GetRootAsDiskStats(p, 0)\n\tstts.Timestamp = statsFlat.Timestamp()\n\tlen := statsFlat.DeviceLength()\n\tstts.Device = make([]structs.Device, len)\n\tfor i := 0; i < len; i++ {\n\t\tvar dev structs.Device\n\t\tif statsFlat.Device(devF, i) {\n\t\t\tdev.Major = devF.Major()\n\t\t\tdev.Minor = devF.Minor()\n\t\t\tdev.Name = string(devF.Name())\n\t\t\tdev.ReadsCompleted = devF.ReadsCompleted()\n\t\t\tdev.ReadsMerged = devF.ReadsMerged()\n\t\t\tdev.ReadSectors = devF.ReadSectors()\n\t\t\tdev.ReadingTime = devF.ReadingTime()\n\t\t\tdev.WritesCompleted = devF.WritesCompleted()\n\t\t\tdev.WritesMerged = devF.WritesMerged()\n\t\t\tdev.WrittenSectors = devF.WrittenSectors()\n\t\t\tdev.WritingTime = devF.WritingTime()\n\t\t\tdev.IOInProgress = devF.IOInProgress()\n\t\t\tdev.IOTime = devF.IOTime()\n\t\t\tdev.WeightedIOTime = devF.WeightedIOTime()\n\t\t}\n\t\tstts.Device[i] = dev\n\t}\n\treturn stts\n}","func DiskQuota(path string, size ...string) string {\n\tparent := id(path)\n\texec.Command(\"btrfs\", \"qgroup\", \"create\", \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/opt\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/var\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/home\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/rootfs\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\n\tif len(size) > 0 && len(size[0]) > 0 {\n\t\tout, err := exec.Command(\"btrfs\", \"qgroup\", \"limit\", size[0]+\"G\", \"1/\"+parent, config.Agent.LxcPrefix+path).CombinedOutput()\n\t\tlog.Check(log.ErrorLevel, \"Limiting BTRFS group 1/\"+parent+\" \"+string(out), err)\n\t\texec.Command(\"btrfs\", \"quota\", \"rescan\", \"-w\", config.Agent.LxcPrefix).Run()\n\t}\n\treturn Stat(path, \"quota\", false)\n}","func dirSize(path string) (float64, error) {\n\tvar size int64\n\terr := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\tsize += 0\n\t\t} else {\n\t\t\tif !info.IsDir() {\n\t\t\t\tsize += info.Size()\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n\tr, _ := decimal.NewFromFloat(float64(size) / (1024 * 1024 * 1024)).Round(2).Float64()\n\treturn r, err\n}","func (o ClusterNodeGroupSystemDiskOutput) Count() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v ClusterNodeGroupSystemDisk) *int { return v.Count }).(pulumi.IntPtrOutput)\n}","func (c *cpusetHandler) Stat(ctr *CgroupControl, m *Metrics) error {\n\treturn nil\n}","func (a *app) gatherStat() {\n\tabout, err := a.srv.About.Get().Fields(\"storageQuota\").Do()\n\tif err != nil {\n\t\tlog.Fatalf(\"Unable to execute an about request: %v\", err)\n\t}\n\n\ta.sq = about.StorageQuota\n}","func (s Sysinfo) MemUsedPercent() uint64 {\n\treturn s.MemUsed() * 100 / s.MemTotal()\n}","func (c *Client) Stat() (count, size int, err error) {\n\tline, err := c.Cmd(\"%s\\r\\n\", STAT)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Number of messages in maildrop\n\tcount, err = strconv.Atoi(strings.Fields(line)[1])\n\tif err != nil {\n\t\treturn\n\t}\n\tif count == 0 {\n\t\treturn\n\t}\n\n\t// Total size of messages in bytes\n\tsize, err = strconv.Atoi(strings.Fields(line)[2])\n\tif err != nil {\n\t\treturn\n\t}\n\tif size == 0 {\n\t\treturn\n\t}\n\treturn\n}","func diskInfo(y *ec2.Instance) diskInfoType {\n\n\tdeviceMap := map[string]int64{}\n\n\trootdevice := *y.RootDeviceName\n\tfor _, f := range y.BlockDeviceMappings {\n\t\tdeviceMap[aws.StringValue(f.DeviceName)] = f.Ebs.AttachTime.Unix()\n\t}\n\tdd := time.Unix(int64(deviceMap[rootdevice]), 0)\n\tsplitdate := strings.Split(dd.Format(time.RFC850), \" \")[1]\n\tdiskAttached := len(deviceMap)\n\tdItype := diskInfoType{date: splitdate, diskCount: diskAttached}\n\treturn dItype\n\n}","func (c *cpuMeasure) update() error {\n\tfile, err := os.Open(stat)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\tvar lineName string\n\tvar n, cpuCount int\n\n\tscanner := bufio.NewScanner(file)\n\tfor scanner.Scan() {\n\t\tline := scanner.Text()\n\n\t\tif strings.Contains(line, \"cpu\") {\n\t\t\tn, err = fmt.Sscanf(line,\n\t\t\t\t\"%s %d %d %d %d %d %d %d %d %d %d\", &lineName,\n\t\t\t\t&c.cores[cpuCount][0], &c.cores[cpuCount][1],\n\t\t\t\t&c.cores[cpuCount][2], &c.cores[cpuCount][3],\n\t\t\t\t&c.cores[cpuCount][4], &c.cores[cpuCount][5],\n\t\t\t\t&c.cores[cpuCount][6], &c.cores[cpuCount][7],\n\t\t\t\t&c.cores[cpuCount][8], &c.cores[cpuCount][9],\n\t\t\t)\n\t\t\tcheckSscanf(lineName, err, n, 11)\n\t\t\tcpuCount++\n\t\t} else if strings.Contains(line, \"ctxt\") {\n\t\t\tn, err = fmt.Sscanf(line, \"ctxt %d\", &c.SwitchCtxt)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"btime\") {\n\t\t\tn, err = fmt.Sscanf(line, \"btime %d\", &c.BootTime)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"processes\") {\n\t\t\tn, err = fmt.Sscanf(line, \"processes %d\", &c.Processes)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"procs_running\") {\n\t\t\tn, err = fmt.Sscanf(line, \"procs_running %d\", &c.ProcsRunning)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"procs_blocked\") {\n\t\t\tn, err = fmt.Sscanf(line, \"procs_blocked %d\", &c.ProcsBlocked)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t}\n\t}\n\n\treturn nil\n}","func VolumeScanner(pathToScan string) error {\n\tresult := make(map[string]DirStats)\n\tvar largestDir string\n\tlargestSize := int64(0)\n\ttotalSize := int64(0)\n\tvar cwd string\n\terr := filepath.Walk(pathToScan, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif info.IsDir() {\n\t\t\tcwd = path\n\t\t\tresult[cwd] = DirStats{\n\t\t\t\tPath: cwd,\n\t\t\t\tTotalSize: int64(0),\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t\tdir := filepath.Dir(path)\n\t\tdirStat, _ := result[dir]\n\t\tdirStat.TotalSize += info.Size()\n\t\ttotalSize += info.Size()\n\t\tif dirStat.TotalSize > largestSize {\n\t\t\tlargestSize = dirStat.TotalSize\n\t\t\tlargestDir = dir\n\t\t}\n\t\treturn err\n\n\t})\n\tdata := [][]string{\n\t\t[]string{\"TOTAL_SIZE\", bytefmt.ByteSize(uint64(totalSize)), \"\"},\n\t\t[]string{\"LARGEST_DIR\", largestDir, \"\"},\n\t\t[]string{\"LARGEST_DIR_SIZE\", bytefmt.ByteSize(uint64(largestSize)), \"\"},\n\t}\n\n\tviews.Print(data)\n\treturn err\n\n}","func (u *User) GetDiskUsage() int {\n\tif u == nil || u.DiskUsage == nil {\n\t\treturn 0\n\t}\n\treturn *u.DiskUsage\n}","func report(p *rc.Process, wallTime time.Duration) {\n\tsv, err := p.SystemVersion()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\n\tss, err := p.SystemStatus()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\n\tproc, err := p.Stop()\n\tif err != nil {\n\t\treturn\n\t}\n\n\trusage, ok := proc.SysUsage().(*syscall.Rusage)\n\tif !ok {\n\t\treturn\n\t}\n\n\tlog.Println(\"Version:\", sv.Version)\n\tlog.Println(\"Alloc:\", ss.Alloc/1024, \"KiB\")\n\tlog.Println(\"Sys:\", ss.Sys/1024, \"KiB\")\n\tlog.Println(\"Goroutines:\", ss.Goroutines)\n\tlog.Println(\"Wall time:\", wallTime)\n\tlog.Println(\"Utime:\", time.Duration(rusage.Utime.Nano()))\n\tlog.Println(\"Stime:\", time.Duration(rusage.Stime.Nano()))\n\tif runtime.GOOS == \"darwin\" {\n\t\t// Darwin reports in bytes, Linux seems to report in KiB even\n\t\t// though the manpage says otherwise.\n\t\trusage.Maxrss /= 1024\n\t}\n\tlog.Println(\"MaxRSS:\", rusage.Maxrss, \"KiB\")\n\n\tfmt.Printf(\"%s,%d,%d,%d,%.02f,%.02f,%.02f,%d\\n\",\n\t\tsv.Version,\n\t\tss.Alloc/1024,\n\t\tss.Sys/1024,\n\t\tss.Goroutines,\n\t\twallTime.Seconds(),\n\t\ttime.Duration(rusage.Utime.Nano()).Seconds(),\n\t\ttime.Duration(rusage.Stime.Nano()).Seconds(),\n\t\trusage.Maxrss)\n}","func Stat(t *kernel.Task, args arch.SyscallArguments) (uintptr, *kernel.SyscallControl, error) {\n\taddr := args[0].Pointer()\n\tstatAddr := args[1].Pointer()\n\n\tpath, dirPath, err := copyInPath(t, addr, false /* allowEmpty */)\n\tif err != nil {\n\t\treturn 0, nil, err\n\t}\n\n\treturn 0, nil, fileOpOn(t, linux.AT_FDCWD, path, true /* resolve */, func(root *fs.Dirent, d *fs.Dirent, _ uint) error {\n\t\treturn stat(t, d, dirPath, statAddr)\n\t})\n}","func (rso *RadosStripedObject) Stat() (size uint64, pmtime uint64, err error) {\n\tobj := C.CString(rso.ObjectName)\n\tdefer C.free(unsafe.Pointer(obj))\n\n\tvar c_size C.uint64_t\n\tvar c_time_t C.time_t\n\tret := C.rados_striper_stat(rso.Striper, obj, &c_size, &c_time_t)\n\tif ret < 0 {\n\t\treturn 0, 0, errors.New(\"get stat failed\")\n\t}\n\treturn uint64(c_size), uint64(C.uint64_t(c_time_t)), nil\n}","func (s *Simple) DiskInfo(req *acomm.Request) (interface{}, *url.URL, error) {\n\tvar args CPUInfoArgs\n\tif err := req.UnmarshalArgs(&args); err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif args.GuestID == \"\" {\n\t\treturn nil, nil, errors.New(\"missing guest_id\")\n\t}\n\n\tresult := &DiskInfoResult{\n\t\t&DiskInfo{\n\t\t\tDevice: \"vda1\",\n\t\t\tSize: 10 * (1024 * 1024 * 1024), // 10 GB in bytes\n\t\t},\n\t}\n\n\treturn result, nil, nil\n}","func MBUsed() float64 {\n var m runtime.MemStats\n runtime.ReadMemStats(&m)\n return float64(m.TotalAlloc) / BytesPerMBF \n}","func (s *DataNode) incDiskErrCnt(partitionID uint64, err error, flag uint8) {\n\tif err == nil {\n\t\treturn\n\t}\n\tdp := s.space.Partition(partitionID)\n\tif dp == nil {\n\t\treturn\n\t}\n\td := dp.Disk()\n\tif d == nil {\n\t\treturn\n\t}\n\tif !IsDiskErr(err.Error()) {\n\t\treturn\n\t}\n\tif flag == WriteFlag {\n\t\td.incWriteErrCnt()\n\t} else if flag == ReadFlag {\n\t\td.incReadErrCnt()\n\t}\n}","func ProcStat(c *gin.Context) {\n\tres := CmdExec(\"cat /proc/stat | head -n 1 | awk '{$1=\\\"\\\";print}'\")\n\tresArray := strings.Split(res[0], \" \")\n\tvar cpu []int64\n\tvar totalcpu, idlecpu int64\n\tfor _, v := range resArray {\n\t\ttemp, err := strconv.ParseInt(v, 10, 64)\n\t\tif err == nil {\n\t\t\tcpu = append(cpu, temp)\n\t\t\ttotalcpu = totalcpu + temp\n\t\t}\n\t}\n\tidlecpu = cpu[3]\n\tc.JSON(http.StatusOK, gin.H{\n\t\t\"totalcpu\": totalcpu,\n\t\t\"idlecpu\": idlecpu,\n\t})\n}","func dirents(dir string) []os.FileInfo {\n\tentries, err := ioutil.ReadDir(dir)\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"du1: %v\\n\", err)\n\t\treturn nil\n\t}\n\treturn entries\n}","func (s *Store) scrubDisk(diskno int, d Disk) {\n\t// We use 0 as capacity for token bucket limiter, which means we'll wait\n\t// for \"s.config.ScrubRate\" tokens to be refilled every time we call \"Take\",\n\t// which is fine for our usage here.\n\ttb := tokenbucket.New(0, 0)\n\n\tfor {\n\t\t// In case of errors, empty disk, etc., don't spin. PL-1113\n\t\ttime.Sleep(5 * time.Minute)\n\n\t\trate := s.Config().ScrubRate\n\t\tif rate < 1024 {\n\t\t\tcontinue\n\t\t}\n\t\ttb.SetRate(float32(rate), 0)\n\n\t\tdir, err := d.OpenDir()\n\t\tif err == core.ErrDiskRemoved {\n\t\t\treturn\n\t\t}\n\t\tif err != core.NoError {\n\t\t\tlog.Errorf(\"aborting disk scrub for disk %d, failed to open dir, err=%s\", diskno, err)\n\t\t\tcontinue\n\t\t}\n\n\t\tlog.Infof(\"scrub of disk %d starting\", diskno)\n\n\t\tvar scrubbed, ttracts, ok, bad int\n\t\tvar bytes int64\n\t\tstart := time.Now()\n\n\t\tfor {\n\t\t\ttracts, terr := d.ReadDir(dir)\n\t\t\tif terr != core.NoError {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tttracts += len(tracts)\n\t\t\tfor _, tract := range tracts {\n\t\t\t\t// If we can't lock it someone else is probably scrubbing it by virtue of reading or writing it.\n\t\t\t\tif !s.tryLockTract(tract, READ) {\n\t\t\t\t\tlog.V(5).Infof(\"tract %s is busy, won't scrub this iteration\", tract)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// Scrub returns how many bytes it read. We use this to throttle scrubbing to s.config.ScrubRate bytes/sec.\n\t\t\t\tn, err := d.Scrub(tract)\n\t\t\t\ts.unlock(tract, READ)\n\n\t\t\t\tif err == core.ErrDiskRemoved {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t// This might sleep so we want to unlock the tract before calling it.\n\t\t\t\ttb.Take(float32(n))\n\n\t\t\t\t// Collect and log some stats.\n\t\t\t\tif s.maybeReportError(tract, err) {\n\t\t\t\t\tbad++\n\t\t\t\t} else {\n\t\t\t\t\tok++\n\t\t\t\t}\n\t\t\t\tscrubbed++\n\t\t\t\tbytes += n\n\t\t\t\tif scrubbed%10 == 0 {\n\t\t\t\t\tlogStats(diskno, start, scrubbed, ok, bad, ttracts, bytes, 2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\td.CloseDir(dir)\n\t\tlogStats(diskno, start, scrubbed, ok, bad, ttracts, bytes, 0)\n\t}\n}","func GetFSInfo(path string) (total, available int) {\n\ttotal = -1\n\tavailable = -1\n\tvar buf syscall.Statfs_t\n\n\tif syscall.Statfs(path, &buf) != nil {\n\t\treturn\n\t}\n\n\tif buf.Bsize <= 0 {\n\t\treturn\n\t}\n\n\ttotal = int((uint64(buf.Bsize) * buf.Blocks) / (1000 * 1000))\n\tavailable = int((uint64(buf.Bsize) * buf.Bavail) / (1000 * 1000))\n\n\treturn\n}","func GetStat() (uint64, uint64) {\n\treturn filesAmount, dirsAmount\n}","func (d *dataUsageCache) dui(path string, buckets []BucketInfo) madmin.DataUsageInfo {\n\te := d.find(path)\n\tif e == nil {\n\t\t// No entry found, return empty.\n\t\treturn madmin.DataUsageInfo{}\n\t}\n\tflat := d.flatten(*e)\n\tdui := madmin.DataUsageInfo{\n\t\tLastUpdate: d.Info.LastUpdate,\n\t\tObjectsTotalCount: flat.Objects,\n\t\tObjectsTotalSize: uint64(flat.Size),\n\t\tBucketsCount: uint64(len(e.Children)),\n\t\tBucketsUsage: d.bucketsUsageInfo(buckets),\n\t}\n\tif flat.ReplicationStats != nil {\n\t\tdui.ReplicationPendingSize = flat.ReplicationStats.PendingSize\n\t\tdui.ReplicatedSize = flat.ReplicationStats.ReplicatedSize\n\t\tdui.ReplicationFailedSize = flat.ReplicationStats.FailedSize\n\t\tdui.ReplicationPendingCount = flat.ReplicationStats.PendingCount\n\t\tdui.ReplicationFailedCount = flat.ReplicationStats.FailedCount\n\t\tdui.ReplicaSize = flat.ReplicationStats.ReplicaSize\n\t}\n\treturn dui\n}","func (s *CPUStat) Usage() float64 {\n\treturn s.All.Usage() * float64(len(s.cpus))\n}","func (p *xlStorageDiskIDCheck) monitorDiskStatus() {\n\tt := time.NewTicker(5 * time.Second)\n\tdefer t.Stop()\n\tfn := mustGetUUID()\n\tfor range t.C {\n\t\tif len(p.health.tokens) == 0 {\n\t\t\t// Queue is still full, no need to check.\n\t\t\tcontinue\n\t\t}\n\t\terr := p.storage.WriteAll(context.Background(), minioMetaTmpBucket, fn, []byte{10000: 42})\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\t\tb, err := p.storage.ReadAll(context.Background(), minioMetaTmpBucket, fn)\n\t\tif err != nil || len(b) != 10001 {\n\t\t\tcontinue\n\t\t}\n\t\terr = p.storage.Delete(context.Background(), minioMetaTmpBucket, fn, DeleteOptions{\n\t\t\tRecursive: false,\n\t\t\tForce: false,\n\t\t})\n\t\tif err == nil {\n\t\t\tlogger.Info(\"node(%s): Read/Write/Delete successful, bringing drive %s online. Drive was offline for %s.\", globalLocalNodeName, p.storage.String(),\n\t\t\t\ttime.Since(time.Unix(0, atomic.LoadInt64(&p.health.lastSuccess))))\n\t\t\tatomic.StoreInt32(&p.health.status, diskHealthOK)\n\t\t\treturn\n\t\t}\n\t}\n}","func (s *CRDBStorage) Stat(key string) (certmagic.KeyInfo, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), defaultQueryTimeout)\n\tdefer cancel()\n\trow := s.DB.QueryRowContext(ctx, `select length(value), updated from certmagic_values where \"key\" = $1`, key)\n\tinfo := certmagic.KeyInfo{\n\t\tKey: key,\n\t\tIsTerminal: true,\n\t}\n\tif err := row.Scan(&info.Size, &info.Modified); err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\treturn certmagic.KeyInfo{}, certmagic.ErrNotExist(err)\n\t\t}\n\t\treturn certmagic.KeyInfo{}, err\n\t}\n\treturn info, nil\n}","func dirents(dir string) []os.FileInfo {\r\n\tsema <- struct{}{} // acquire token\r\n\tdefer func() { <-sema }() // release token\r\n\r\n\tentries, err := ioutil.ReadDir(dir)\r\n\tif err != nil {\r\n\t\tfmt.Fprintf(os.Stderr, \"du: %v\\n\", err)\r\n\t\treturn nil\r\n\t}\r\n\treturn entries\r\n}","func (o LocalDiskOutput) DiskCount() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v LocalDisk) *int { return v.DiskCount }).(pulumi.IntPtrOutput)\n}","func (e *BackupEnv) GetDirSize(source string) (int64) {\n directory, _ := os.Open(source);\n var sum int64 = 0;\n defer directory.Close();\n\n objects, _ := directory.Readdir(-1)\n for _, obj := range objects {\n if obj.IsDir() {\n sum += e.GetDirSize(source + \"/\" + obj.Name());\n } else {\n stat, _ := os.Stat(source + \"/\" + obj.Name());\n sum += stat.Size();\n }\n }\n\n return sum;\n}","func RamUsage() {\r\n v, _ := mem.VirtualMemory()\r\n fmt.Printf(\"RAM{ Total: %v, Free:%v, UsedPercent:%f%%}\\n\", v.Total, v.Free, v.UsedPercent)\r\n}","func measureStorageDevice(blkDevicePath string) error {\n\tlog.Printf(\"Storage Collector: Measuring block device %s\\n\", blkDevicePath)\n\tfile, err := os.Open(blkDevicePath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"couldn't open disk=%s err=%v\", blkDevicePath, err)\n\t}\n\n\teventDesc := fmt.Sprintf(\"Storage Collector: Measured %s\", blkDevicePath)\n\treturn tpm.ExtendPCRDebug(pcr, file, eventDesc)\n}","func printSysstat(v *gocui.View, s stat.Stat) error {\n\tvar err error\n\n\t/* line1: current time and load average */\n\t_, err = fmt.Fprintf(v, \"pgcenter: %s, load average: %.2f, %.2f, %.2f\\n\",\n\t\ttime.Now().Format(\"2006-01-02 15:04:05\"),\n\t\ts.LoadAvg.One, s.LoadAvg.Five, s.LoadAvg.Fifteen)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t/* line2: cpu usage */\n\t_, err = fmt.Fprintf(v, \" %%cpu: \\033[37;1m%4.1f\\033[0m us, \\033[37;1m%4.1f\\033[0m sy, \\033[37;1m%4.1f\\033[0m ni, \\033[37;1m%4.1f\\033[0m id, \\033[37;1m%4.1f\\033[0m wa, \\033[37;1m%4.1f\\033[0m hi, \\033[37;1m%4.1f\\033[0m si, \\033[37;1m%4.1f\\033[0m st\\n\",\n\t\ts.CpuStat.User, s.CpuStat.Sys, s.CpuStat.Nice, s.CpuStat.Idle,\n\t\ts.CpuStat.Iowait, s.CpuStat.Irq, s.CpuStat.Softirq, s.CpuStat.Steal)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t/* line3: memory usage */\n\t_, err = fmt.Fprintf(v, \" MiB mem: \\033[37;1m%6d\\033[0m total, \\033[37;1m%6d\\033[0m free, \\033[37;1m%6d\\033[0m used, \\033[37;1m%8d\\033[0m buff/cached\\n\",\n\t\ts.Meminfo.MemTotal, s.Meminfo.MemFree, s.Meminfo.MemUsed,\n\t\ts.Meminfo.MemCached+s.Meminfo.MemBuffers+s.Meminfo.MemSlab)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t/* line4: swap usage, dirty and writeback */\n\t_, err = fmt.Fprintf(v, \"MiB swap: \\033[37;1m%6d\\033[0m total, \\033[37;1m%6d\\033[0m free, \\033[37;1m%6d\\033[0m used, \\033[37;1m%6d/%d\\033[0m dirty/writeback\\n\",\n\t\ts.Meminfo.SwapTotal, s.Meminfo.SwapFree, s.Meminfo.SwapUsed,\n\t\ts.Meminfo.MemDirty, s.Meminfo.MemWriteback)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}","func (c *NFSCollector) Describe(ch chan<- *prometheus.Desc) {\n\tch <- c.Up\n\tch <- c.NFSInfo\n\tch <- c.DiskFree\n\tch <- c.NICInfo\n\tch <- c.NICReceive\n\tch <- c.NICSend\n\n\tch <- c.MaintenanceScheduled\n\tch <- c.MaintenanceInfo\n\tch <- c.MaintenanceStartTime\n\tch <- c.MaintenanceEndTime\n}","func (f *Fs) About(ctx context.Context) (*fs.Usage, error) {\n\topts := rest.Opts{\n\t\tMethod: \"PROPFIND\",\n\t\tPath: \"\",\n\t\tExtraHeaders: map[string]string{\n\t\t\t\"Depth\": \"0\",\n\t\t},\n\t}\n\topts.Body = bytes.NewBuffer([]byte(`\n\n \n \n \n \n\n`))\n\tvar q api.Quota\n\tvar resp *http.Response\n\tvar err error\n\terr = f.pacer.Call(func() (bool, error) {\n\t\tresp, err = f.srv.CallXML(ctx, &opts, nil, &q)\n\t\treturn f.shouldRetry(ctx, resp, err)\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tusage := &fs.Usage{}\n\tif i, err := strconv.ParseInt(q.Used, 10, 64); err == nil && i >= 0 {\n\t\tusage.Used = fs.NewUsageValue(i)\n\t}\n\tif i, err := strconv.ParseInt(q.Available, 10, 64); err == nil && i >= 0 {\n\t\tusage.Free = fs.NewUsageValue(i)\n\t}\n\tif usage.Used != nil && usage.Free != nil {\n\t\tusage.Total = fs.NewUsageValue(*usage.Used + *usage.Free)\n\t}\n\treturn usage, nil\n}","func updateMetrics(interval time.Duration, diskRoot string) {\n\tlogger := logrus.WithField(\"sync-loop\", \"updateMetrics\")\n\tticker := time.NewTicker(interval)\n\tfor ; true; <-ticker.C {\n\t\tlogger.Info(\"tick\")\n\t\t_, bytesFree, bytesUsed, _, _, _, err := diskutil.GetDiskUsage(diskRoot)\n\t\tif err != nil {\n\t\t\tlogger.WithError(err).Error(\"Failed to get disk metrics\")\n\t\t} else {\n\t\t\tpromMetrics.DiskFree.Set(float64(bytesFree) / 1e9)\n\t\t\tpromMetrics.DiskUsed.Set(float64(bytesUsed) / 1e9)\n\t\t\tpromMetrics.DiskTotal.Set(float64(bytesFree+bytesUsed) / 1e9)\n\t\t}\n\t}\n}"],"string":"[\n \"func DiskUsage(path string) (UsageInfo, error) {\\n\\tvar usage UsageInfo\\n\\n\\tif path == \\\"\\\" {\\n\\t\\treturn usage, fmt.Errorf(\\\"invalid directory\\\")\\n\\t}\\n\\n\\t// First check whether the quota system knows about this directory\\n\\t// A nil quantity or error means that the path does not support quotas\\n\\t// or xfs_quota tool is missing and we should use other mechanisms.\\n\\tstartTime := time.Now()\\n\\tconsumption, _ := fsquota.GetConsumption(path)\\n\\tif consumption != nil {\\n\\t\\tusage.Bytes = consumption.Value()\\n\\t\\tdefer servermetrics.CollectVolumeStatCalDuration(\\\"fsquota\\\", startTime)\\n\\t} else {\\n\\t\\tdefer servermetrics.CollectVolumeStatCalDuration(\\\"du\\\", startTime)\\n\\t}\\n\\n\\tinodes, _ := fsquota.GetInodes(path)\\n\\tif inodes != nil {\\n\\t\\tusage.Inodes = inodes.Value()\\n\\t}\\n\\n\\tif inodes != nil && consumption != nil {\\n\\t\\treturn usage, nil\\n\\t}\\n\\n\\ttopLevelStat := &unix.Stat_t{}\\n\\terr := unix.Stat(path, topLevelStat)\\n\\tif err != nil {\\n\\t\\treturn usage, err\\n\\t}\\n\\n\\t// dedupedInode stores inodes that could be duplicates (nlink > 1)\\n\\tdedupedInodes := make(map[uint64]struct{})\\n\\n\\terr = filepath.Walk(path, func(path string, info os.FileInfo, err error) error {\\n\\t\\t// ignore files that have been deleted after directory was read\\n\\t\\tif os.IsNotExist(err) {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"unable to count inodes for %s: %s\\\", path, err)\\n\\t\\t}\\n\\n\\t\\t// according to the docs, Sys can be nil\\n\\t\\tif info.Sys() == nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"fileinfo Sys is nil\\\")\\n\\t\\t}\\n\\n\\t\\ts, ok := info.Sys().(*syscall.Stat_t)\\n\\t\\tif !ok {\\n\\t\\t\\treturn fmt.Errorf(\\\"unsupported fileinfo; could not convert to stat_t\\\")\\n\\t\\t}\\n\\n\\t\\tif s.Dev != topLevelStat.Dev {\\n\\t\\t\\t// don't descend into directories on other devices\\n\\t\\t\\treturn filepath.SkipDir\\n\\t\\t}\\n\\n\\t\\t// Dedupe hardlinks\\n\\t\\tif s.Nlink > 1 {\\n\\t\\t\\tif _, ok := dedupedInodes[s.Ino]; !ok {\\n\\t\\t\\t\\tdedupedInodes[s.Ino] = struct{}{}\\n\\t\\t\\t} else {\\n\\t\\t\\t\\treturn nil\\n\\t\\t\\t}\\n\\t\\t}\\n\\n\\t\\tif consumption == nil {\\n\\t\\t\\tusage.Bytes += int64(s.Blocks) * int64(512) // blocksize in bytes\\n\\t\\t}\\n\\n\\t\\tif inodes == nil {\\n\\t\\t\\tusage.Inodes++\\n\\t\\t}\\n\\n\\t\\treturn nil\\n\\t})\\n\\n\\treturn usage, err\\n}\",\n \"func DiskUsage(path string) (disk DiskStatus, err error) {\\n\\tfs := syscall.Statfs_t{}\\n\\terr = syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tdisk.All = fs.Blocks * uint64(fs.Bsize)\\n\\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\\n\\tdisk.Used = disk.All - disk.Free\\n\\treturn\\n}\",\n \"func DiskUsage(path string) (disk DiskStatus) {\\n\\tfs := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tdisk.All = fs.Blocks * uint64(fs.Bsize)\\n\\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\\n\\tdisk.Used = disk.All - disk.Free\\n\\treturn\\n}\",\n \"func DiskUsage(path string) (disk DiskStatus) {\\n\\tfs := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tdisk.All = fs.Blocks * uint64(fs.Bsize)\\n\\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\\n\\tdisk.Used = disk.All - disk.Free\\n\\treturn\\n}\",\n \"func DiskUsage(path string) (disk DiskStatus) {\\n\\tfs := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tdisk.All = fs.Blocks * uint64(fs.Bsize)\\n\\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\\n\\tdisk.Used = disk.All - disk.Free\\n\\treturn\\n}\",\n \"func DiskUsage(path string) (disk DiskStatus) {\\n\\n\\tif path == \\\"\\\" {\\n\\t\\treturn\\n\\t}\\n\\n\\tfs := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\tlog.Fatal(err)\\n\\t}\\n\\tdisk.All = fs.Blocks * uint64(fs.Bsize)\\n\\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\\n\\tdisk.Used = disk.All - disk.Free\\n\\treturn disk\\n}\",\n \"func Usage(path string) (*Status, error) {\\n\\tfs := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\treturn nil, errors.Wrapf(err, \\\"get disk status for path '%s' failed\\\", path)\\n\\t}\\n\\tvar disk Status\\n\\tdisk.BytesTotal = fs.Blocks * uint64(fs.Bsize)\\n\\tdisk.BytesFree = fs.Bfree * uint64(fs.Bsize)\\n\\tdisk.BytesUsed = disk.BytesTotal - disk.BytesFree\\n\\n\\tdisk.InodesTotal = uint64(fs.Files)\\n\\tdisk.InodesFree = uint64(fs.Ffree)\\n\\tdisk.InodesUsed = disk.InodesTotal - disk.InodesFree\\n\\treturn &disk, nil\\n}\",\n \"func du() {\\n\\trf, err := os.Stat(\\\"/\\\")\\n\\tif err != nil {\\n\\t\\tlog.Fatal(err)\\n\\t}\\n\\trdev := rf.Sys().(*syscall.Stat_t).Dev\\n\\tf := os.DirFS(\\\"/\\\")\\n\\terr = fs.WalkDir(f, \\\".\\\", func(p string, d fs.DirEntry, err error) error {\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fs.SkipDir\\n\\t\\t}\\n\\t\\ti, err := d.Info()\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fs.SkipDir\\n\\t\\t}\\n\\t\\tsys := i.Sys()\\n\\t\\tif sys == nil {\\n\\t\\t\\treturn fs.SkipDir\\n\\t\\t}\\n\\t\\tdev := sys.(*syscall.Stat_t).Dev\\n\\t\\tfmt.Printf(\\\"p=%v size=%v dev=%v\\\\n\\\", p, i.Size(), dev)\\n\\t\\tif dev != rdev {\\n\\t\\t\\treturn fs.SkipDir\\n\\t\\t}\\n\\t\\treturn nil\\n\\t})\\n\\tif err != nil {\\n\\t\\tlog.Print(err)\\n\\t}\\n}\",\n \"func (c *DefaultChecker) DiskPerf() (info DiskPerfInfo, err error) {\\n\\tvar data []byte\\n\\n\\texe := execCommand(\\\"dd\\\", \\\"bs=1M\\\", \\\"count=256\\\", \\\"if=/dev/zero\\\", fmt.Sprintf(\\\"of=%v\\\", path.Join(c.WritePath, \\\"test\\\")), \\\"conv=fdatasync\\\")\\n\\tdata, err = exe.CombinedOutput()\\n\\tgohtypes.PanicIfError(\\\"Not possible to execute the 'dd' command.\\\", 500, err)\\n\\n\\tout := string(data)\\n\\tlogrus.Info(out)\\n\\n\\trate := strings.Split(strings.Split(strings.Replace(out, \\\"\\\\n\\\", \\\"\\\", -1), \\\" s, \\\")[1], \\\" \\\")\\n\\n\\tinfo = DiskPerfInfo{\\n\\t\\tWriteSpeed: rate[0],\\n\\t\\tUnit: rate[1],\\n\\t}\\n\\treturn\\n}\",\n \"func DiskUsage(path string) (disk DiskStatus) {\\n\\th := syscall.MustLoadDLL(\\\"kernel32.dll\\\")\\n\\tc := h.MustFindProc(\\\"GetDiskFreeSpaceExW\\\")\\n\\tlpFreeBytesAvailable := int64(0)\\n\\tlpTotalNumberOfBytes := int64(0)\\n\\tlpTotalNumberOfFreeBytes := int64(0)\\n\\t_, _, err := c.Call(uintptr(unsafe.Pointer(syscall.StringToUTF16Ptr(path[:2]))),\\n\\t\\tuintptr(unsafe.Pointer(&lpFreeBytesAvailable)),\\n\\t\\tuintptr(unsafe.Pointer(&lpTotalNumberOfBytes)),\\n\\t\\tuintptr(unsafe.Pointer(&lpTotalNumberOfFreeBytes)))\\n\\tif err != nil {\\n\\t\\tif !strings.Contains(fmt.Sprint(err), \\\"successfully\\\") {\\n\\t\\t\\tlogger.Error(\\\"Error during retrieving memory statistic:\\\", err)\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t}\\n\\tdisk.All = uint64(lpTotalNumberOfBytes)\\n\\tdisk.Free = uint64(lpTotalNumberOfFreeBytes)\\n\\tdisk.Used = disk.All - disk.Free\\n\\treturn\\n}\",\n \"func GetDiskUsage(path string) ([]float64, error) {\\n\\tvar usage []float64 = make([]float64, 6) // Data to return\\n\\tvar sc syscall.Statfs_t // Filesystem stat\\n\\tvar err error // Error catching\\n\\tvar buffer bytes.Buffer // Buffer for json indent\\n\\tvar content []byte // Json's content\\n\\n\\t// Get filesystem stat\\n\\terr = syscall.Statfs(path, &sc)\\n\\tif err != nil {\\n\\t\\treturn usage, err\\n\\t}\\n\\n\\t// Convert structure => json\\n\\tcontent, err = json.Marshal(sc)\\n\\tif err != nil {\\n\\t\\treturn usage, err\\n\\t}\\n\\n\\t// Indent json\\n\\tjson.Indent(&buffer, content, \\\"\\\", \\\" \\\")\\n\\n\\t// Set data to return\\n\\tusage[0] = float64(sc.Bsize) * float64(sc.Blocks) // TotalStorage\\n\\tusage[1] = float64(sc.Files) // TotalFileNodes\\n\\tusage[3] = float64(sc.Bsize) * float64(sc.Bfree) // FreeStorage\\n\\tusage[2] = usage[0] - usage[3] // UsedStorage\\n\\tusage[5] = float64(sc.Ffree) // FreeFileNodes\\n\\tusage[4] = usage[1] - usage[5] // UsedFileNodes\\n\\n\\treturn usage, nil\\n}\",\n \"func DiskUsage(path string) (*DiskStatus, error) {\\n\\tlpDirectoryName, err := syscall.UTF16PtrFromString(path)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tlpFreeBytesAvailable := int64(0)\\n\\tlpTotalNumberOfBytes := int64(0)\\n\\tlpTotalNumberOfFreeBytes := int64(0)\\n\\t_, _, e := syscall.Syscall6(pGetDiskFreeSpaceEx, 4,\\n\\t\\tuintptr(unsafe.Pointer(lpDirectoryName)),\\n\\t\\tuintptr(unsafe.Pointer(&lpFreeBytesAvailable)),\\n\\t\\tuintptr(unsafe.Pointer(&lpTotalNumberOfBytes)),\\n\\t\\tuintptr(unsafe.Pointer(&lpTotalNumberOfFreeBytes)), 0, 0)\\n\\tif e != 0 {\\n\\t\\treturn nil, errors.New(e.Error())\\n\\t}\\n\\tstatus := &DiskStatus{\\n\\t\\tAll: lpTotalNumberOfBytes,\\n\\t\\tFree: lpFreeBytesAvailable,\\n\\t}\\n\\tstatus.Used = status.All - status.Free\\n\\treturn status, nil\\n}\",\n \"func GetDiskUsage(rw http.ResponseWriter) error {\\n\\tu, r := disk.Usage(\\\"/\\\")\\n\\tif r != nil {\\n\\t\\treturn r\\n\\t}\\n\\n\\treturn share.JSONResponse(u, rw)\\n}\",\n \"func getDiskInfo(path string) (uint64, uint64, error) {\\n\\ts := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &s)\\n\\tif err != nil {\\n\\t\\treturn 0,0, err\\n\\t}\\n\\treservedBlocks := s.Bfree - s.Bavail\\n\\n\\ttotal := uint64(s.Frsize) * (s.Blocks - reservedBlocks)\\n\\tfree := uint64(s.Frsize) * s.Bavail\\n\\t// Check for overflows.\\n\\t// https://github.com/minio/minio/issues/8035\\n\\t// XFS can show wrong values at times error out\\n\\t// in such scenarios.\\n\\tif free > total {\\n\\t\\treturn 0, 0, fmt.Errorf(\\\"detected free space (%d) > total disk space (%d), fs corruption at (%s). please run 'fsck'\\\", free, total, path)\\n\\t}\\n\\n\\treturn total, free, nil\\n}\",\n \"func Diskuse(path string) (cap uint64, used uint64) {\\n\\n\\tfs := syscall.Statfs_t{}\\n\\terr := syscall.Statfs(path, &fs)\\n\\tif err != nil {\\n\\t\\tfmt.Println(err)\\n\\t}\\n\\n\\tcap = fs.Blocks * uint64(fs.Bsize)\\n\\tfree := fs.Bfree * uint64(fs.Bsize) // yup, I just did that\\n\\tused = cap - free\\n\\treturn cap, used\\n}\",\n \"func (d *Datastore) DiskUsage() (uint64, error) {\\n\\tvar total uint64 = 0\\n\\n\\terr := d.meta.ListKeys(func(key []byte) error {\\n\\t\\tmeta, err := d.meta.GetMetadata(key)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\ttotal += uint64(meta.StorageSize)\\n\\t\\treturn nil\\n\\t})\\n\\n\\treturn total, err\\n}\",\n \"func Disk(pt ...bool) ([]*disk.UsageStat, error) {\\n\\tvar ptBool bool\\n\\tif len(pt) > 0 {\\n\\t\\tptBool = pt[0]\\n\\t}\\n\\n\\tvar usage []*disk.UsageStat\\n\\tparts, err := disk.Partitions(ptBool)\\n\\n\\tfor _, part := range parts {\\n\\t\\tuse, err := disk.Usage(part.Mountpoint)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn usage, err\\n\\t\\t}\\n\\t\\tusage = append(usage, use)\\n\\t\\t// printUsage(use)\\n\\t}\\n\\n\\treturn usage, err\\n}\",\n \"func DiskStat() DiskInfo {\\n\\tusage, _ := disk.Usage(\\\"./\\\")\\n\\tpartitions, _ := disk.Partitions(true)\\n\\tiOCounters, _ := disk.IOCounters()\\n\\n\\tue := usage.String()\\n\\n\\tdiskstat := DiskInfo{\\n\\t\\tUsage: ue,\\n\\t\\tPartition: partitions,\\n\\t\\tIOCounters: iOCounters,\\n\\t}\\n\\n\\treturn diskstat\\n}\",\n \"func PrintDiskStats() {\\n\\tfmt.Println(DiskStats)\\n}\",\n \"func (p *btrfsPool) Usage() (usage Usage, err error) {\\n\\tmnt, err := p.Mounted()\\n\\tif err != nil {\\n\\t\\treturn usage, err\\n\\t}\\n\\n\\tvolumes, err := p.Volumes()\\n\\n\\tif err != nil {\\n\\t\\treturn usage, errors.Wrapf(err, \\\"failed to list pool '%s' volumes\\\", mnt)\\n\\t}\\n\\n\\tusage.Size = p.device.Size\\n\\n\\tfor _, volume := range volumes {\\n\\t\\tvol, err := volume.Usage()\\n\\t\\tif err != nil {\\n\\t\\t\\treturn Usage{}, errors.Wrapf(err, \\\"failed to calculate volume '%s' usage\\\", volume.Path())\\n\\t\\t}\\n\\n\\t\\tusage.Used += vol.Used\\n\\t\\tusage.Excl += vol.Excl\\n\\t}\\n\\n\\treturn\\n}\",\n \"func (s *SigarChecker) GetDiskUsage(path string) (usePercent float64, avail uint64) {\\n\\tvar fsu sigar.FileSystemUsage\\n\\n\\terr := fsu.Get(path)\\n\\n\\tutil.Must(err)\\n\\n\\treturn fsu.UsePercent(), fsu.Avail\\n}\",\n \"func DirSize(dir string, excludes []string) (int, error) {\\n\\targs := []string{\\\"-bs\\\", dir}\\n\\tfor _, s := range excludes {\\n\\t\\targs = append(args, fmt.Sprintf(\\\"--exclude=\\\\\\\"%s\\\\\\\"\\\", s))\\n\\t}\\n\\n\\tcmd := exec.Command(\\\"/usr/bin/du\\\", args...)\\n\\tdata, err := cmd.Output()\\n\\tif err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\treturn strconv.Atoi(strings.Split(string(data), \\\"\\\\t\\\")[0])\\n}\",\n \"func testDiskUsageDoubleCount(dirFunc mkShardFunc, t *testing.T) {\\n\\ttemp, cleanup := tempdir(t)\\n\\tdefer cleanup()\\n\\tdefer checkTemp(t, temp)\\n\\n\\tfs, err := flatfs.CreateOrOpen(temp, dirFunc(2), false)\\n\\tif err != nil {\\n\\t\\tt.Fatalf(\\\"New fail: %v\\\\n\\\", err)\\n\\t}\\n\\tdefer fs.Close()\\n\\n\\tvar count int\\n\\tvar wg sync.WaitGroup\\n\\ttestKey := datastore.NewKey(\\\"TEST\\\")\\n\\n\\tput := func() {\\n\\t\\tdefer wg.Done()\\n\\t\\tfor i := 0; i < count; i++ {\\n\\t\\t\\tv := []byte(\\\"10bytes---\\\")\\n\\t\\t\\terr := fs.Put(bg, testKey, v)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\tt.Errorf(\\\"Put fail: %v\\\\n\\\", err)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\n\\tdel := func() {\\n\\t\\tdefer wg.Done()\\n\\t\\tfor i := 0; i < count; i++ {\\n\\t\\t\\terr := fs.Delete(bg, testKey)\\n\\t\\t\\tif err != nil && !strings.Contains(err.Error(), \\\"key not found\\\") {\\n\\t\\t\\t\\tt.Errorf(\\\"Delete fail: %v\\\\n\\\", err)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\n\\t// Add one element and then remove it and check disk usage\\n\\t// makes sense\\n\\tcount = 1\\n\\twg.Add(2)\\n\\tput()\\n\\tdu, _ := fs.DiskUsage(bg)\\n\\tdel()\\n\\tdu2, _ := fs.DiskUsage(bg)\\n\\tif du-10 != du2 {\\n\\t\\tt.Error(\\\"should have deleted exactly 10 bytes:\\\", du, du2)\\n\\t}\\n\\n\\t// Add and remove many times at the same time\\n\\tcount = 200\\n\\twg.Add(4)\\n\\tgo put()\\n\\tgo del()\\n\\tgo put()\\n\\tgo del()\\n\\twg.Wait()\\n\\n\\tdu3, _ := fs.DiskUsage(bg)\\n\\thas, err := fs.Has(bg, testKey)\\n\\tif err != nil {\\n\\t\\tt.Fatal(err)\\n\\t}\\n\\n\\tif has { // put came last\\n\\t\\tif du3 != du {\\n\\t\\t\\tt.Error(\\\"du should be the same as after first put:\\\", du, du3)\\n\\t\\t}\\n\\t} else { // delete came last\\n\\t\\tif du3 != du2 {\\n\\t\\t\\tt.Error(\\\"du should be the same as after first delete:\\\", du2, du3)\\n\\t\\t}\\n\\t}\\n}\",\n \"func getDiskSize(path string) (uint64, error) {\\n\\tdiskInfo := exec.Command(\\\"qemu-img\\\", \\\"info\\\", \\\"--output\\\", \\\"json\\\", path)\\n\\tstdout, err := diskInfo.StdoutPipe()\\n\\tif err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\tif err := diskInfo.Start(); err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\ttmpInfo := struct {\\n\\t\\tVirtualSize uint64 `json:\\\"virtual-size\\\"`\\n\\t\\tFilename string `json:\\\"filename\\\"`\\n\\t\\tClusterSize int64 `json:\\\"cluster-size\\\"`\\n\\t\\tFormat string `json:\\\"format\\\"`\\n\\t\\tFormatSpecific struct {\\n\\t\\t\\tType string `json:\\\"type\\\"`\\n\\t\\t\\tData map[string]string `json:\\\"data\\\"`\\n\\t\\t}\\n\\t\\tDirtyFlag bool `json:\\\"dirty-flag\\\"`\\n\\t}{}\\n\\tif err := json.NewDecoder(stdout).Decode(&tmpInfo); err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\tif err := diskInfo.Wait(); err != nil {\\n\\t\\treturn 0, err\\n\\t}\\n\\treturn tmpInfo.VirtualSize, nil\\n}\",\n \"func Stats(c *libvirt.Connect, uuid string) error {\\n\\t//Check exists\\n\\td, err := c.LookupDomainByUUIDString(uuid)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"failed to lookup: %s\\\", err)\\n\\t}\\n\\n\\t//Check is running\\n\\ts, _, err := d.GetState()\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"failed check state: %s\\\", err)\\n\\t}\\n\\tif s != libvirt.DOMAIN_RUNNING {\\n\\t\\treturn fmt.Errorf(\\\"domain not running: %d\\\", s)\\n\\t}\\n\\n\\tmemStats, err := memStats(d)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tfmt.Printf(\\\"STAT: %+v\\\\n\\\", memStats)\\n\\tfmt.Printf(\\\"STAT Used: %+v\\\\n\\\", memStats.Available-memStats.Unused)\\n\\tfmt.Printf(\\\"STAT Last: %s\\\\n\\\", time.Unix(int64(memStats.LastUpdate), 0))\\n\\n\\tcpuStats, total, err := cpuStats(d)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tfmt.Printf(\\\"%+v\\\\n\\\", cpuStats)\\n\\tfmt.Printf(\\\"Total: %+#v\\\\n\\\", total)\\n\\n\\tnetStats, err := netStats(d)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tfmt.Printf(\\\"NET: %+v\\\\n\\\", netStats)\\n\\n\\t_, dTotal, err := diskStats(d)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tfmt.Printf(\\\"DISK: %+v\\\\n\\\", dTotal)\\n\\n\\treturn nil\\n}\",\n \"func Stat(path, index string, raw bool) (value string) {\\n\\tvar row = map[string]int{\\\"quota\\\": 3, \\\"usage\\\": 2}\\n\\n\\targs := []string{\\\"qgroup\\\", \\\"show\\\", \\\"-re\\\", config.Agent.LxcPrefix}\\n\\tif raw {\\n\\t\\targs = []string{\\\"qgroup\\\", \\\"show\\\", \\\"-re\\\", \\\"--raw\\\", config.Agent.LxcPrefix}\\n\\t}\\n\\tout, err := exec.Command(\\\"btrfs\\\", args...).Output()\\n\\tlog.Check(log.FatalLevel, \\\"Getting btrfs stats\\\", err)\\n\\tind := id(path)\\n\\tscanner := bufio.NewScanner(bytes.NewReader(out))\\n\\tfor scanner.Scan() {\\n\\t\\tif line := strings.Fields(scanner.Text()); len(line) > 3 && strings.HasSuffix(line[0], \\\"/\\\"+ind) {\\n\\t\\t\\tvalue = line[row[index]]\\n\\t\\t}\\n\\t}\\n\\treturn value\\n}\",\n \"func (f *StorageFingerprint) diskFree(path string) (volume string, total, free uint64, err error) {\\n\\tabsPath, err := filepath.Abs(path)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", 0, 0, fmt.Errorf(\\\"failed to determine absolute path for %s\\\", path)\\n\\t}\\n\\n\\t// Use -k to standardize the output values between darwin and linux\\n\\tvar dfArgs string\\n\\tif runtime.GOOS == \\\"linux\\\" {\\n\\t\\t// df on linux needs the -P option to prevent linebreaks on long filesystem paths\\n\\t\\tdfArgs = \\\"-kP\\\"\\n\\t} else {\\n\\t\\tdfArgs = \\\"-k\\\"\\n\\t}\\n\\n\\tmountOutput, err := exec.Command(\\\"df\\\", dfArgs, absPath).Output()\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", 0, 0, fmt.Errorf(\\\"failed to determine mount point for %s\\\", absPath)\\n\\t}\\n\\t// Output looks something like:\\n\\t//\\tFilesystem 1024-blocks Used Available Capacity iused ifree %iused Mounted on\\n\\t//\\t/dev/disk1 487385240 423722532 63406708 87% 105994631 15851677 87% /\\n\\t//\\t[0] volume [1] capacity [2] SKIP [3] free\\n\\tlines := strings.Split(string(mountOutput), \\\"\\\\n\\\")\\n\\tif len(lines) < 2 {\\n\\t\\treturn \\\"\\\", 0, 0, fmt.Errorf(\\\"failed to parse `df` output; expected at least 2 lines\\\")\\n\\t}\\n\\tfields := strings.Fields(lines[1])\\n\\tif len(fields) < 4 {\\n\\t\\treturn \\\"\\\", 0, 0, fmt.Errorf(\\\"failed to parse `df` output; expected at least 4 columns\\\")\\n\\t}\\n\\tvolume = fields[0]\\n\\n\\ttotal, err = strconv.ParseUint(fields[1], 10, 64)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", 0, 0, fmt.Errorf(\\\"failed to parse storage.bytestotal size in kilobytes\\\")\\n\\t}\\n\\t// convert to bytes\\n\\ttotal *= 1024\\n\\n\\tfree, err = strconv.ParseUint(fields[3], 10, 64)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", 0, 0, fmt.Errorf(\\\"failed to parse storage.bytesfree size in kilobytes\\\")\\n\\t}\\n\\t// convert to bytes\\n\\tfree *= 1024\\n\\n\\treturn volume, total, free, nil\\n}\",\n \"func GetDiskUsageStats(path string) (uint64, uint64, error) {\\n\\tvar dirSize, inodeCount uint64\\n\\n\\terr := filepath.Walk(path, func(path string, info os.FileInfo, err error) error {\\n\\t\\tfileStat, error := os.Lstat(path)\\n\\t\\tif error != nil {\\n\\t\\t\\tif fileStat.Mode()&os.ModeSymlink != 0 {\\n\\t\\t\\t\\t// Is a symlink; no error should be returned\\n\\t\\t\\t\\treturn nil\\n\\t\\t\\t}\\n\\t\\t\\treturn error\\n\\t\\t}\\n\\n\\t\\tdirSize += uint64(info.Size())\\n\\t\\tinodeCount++\\n\\n\\t\\treturn nil\\n\\t})\\n\\n\\tif err != nil {\\n\\t\\treturn 0, 0, err\\n\\t}\\n\\n\\treturn dirSize, inodeCount, err\\n}\",\n \"func (v *btrfsVolume) Usage() (usage Usage, err error) {\\n\\tctx := context.TODO()\\n\\tinfo, err := v.utils.SubvolumeInfo(ctx, v.Path())\\n\\tif err != nil {\\n\\t\\treturn usage, err\\n\\t}\\n\\n\\tgroups, err := v.utils.QGroupList(ctx, v.Path())\\n\\tif err != nil {\\n\\t\\treturn usage, err\\n\\t}\\n\\n\\tgroup, ok := groups[fmt.Sprintf(\\\"0/%d\\\", info.ID)]\\n\\tif !ok {\\n\\t\\t// no qgroup associated with the subvolume id! means no limit, but we also\\n\\t\\t// cannot read the usage.\\n\\t\\treturn\\n\\t}\\n\\n\\t// used is basically amount of space reserved for this\\n\\t// volume. We assume that's total usage of the volume\\n\\tused := group.MaxRfer\\n\\n\\tif used == 0 {\\n\\t\\t// in case no limit is set on the subvolume, we assume\\n\\t\\t// it's size is the size of the files on that volumes\\n\\t\\t// or a special case when the volume is a zdb volume\\n\\t\\tused, err = volumeUsage(v.Path())\\n\\t\\tif err != nil {\\n\\t\\t\\treturn usage, errors.Wrap(err, \\\"failed to get subvolume usage\\\")\\n\\t\\t}\\n\\t}\\n\\n\\treturn Usage{Used: used, Size: group.MaxRfer, Excl: group.Excl}, nil\\n}\",\n \"func (f *Fs) About(ctx context.Context) (*fs.Usage, error) {\\n\\tvar available, total, free int64\\n\\troot, e := syscall.UTF16PtrFromString(f.root)\\n\\tif e != nil {\\n\\t\\treturn nil, fmt.Errorf(\\\"failed to read disk usage: %w\\\", e)\\n\\t}\\n\\t_, _, e1 := getFreeDiskSpace.Call(\\n\\t\\tuintptr(unsafe.Pointer(root)),\\n\\t\\tuintptr(unsafe.Pointer(&available)), // lpFreeBytesAvailable - for this user\\n\\t\\tuintptr(unsafe.Pointer(&total)), // lpTotalNumberOfBytes\\n\\t\\tuintptr(unsafe.Pointer(&free)), // lpTotalNumberOfFreeBytes\\n\\t)\\n\\tif e1 != syscall.Errno(0) {\\n\\t\\treturn nil, fmt.Errorf(\\\"failed to read disk usage: %w\\\", e1)\\n\\t}\\n\\tusage := &fs.Usage{\\n\\t\\tTotal: fs.NewUsageValue(total), // quota of bytes that can be used\\n\\t\\tUsed: fs.NewUsageValue(total - free), // bytes in use\\n\\t\\tFree: fs.NewUsageValue(available), // bytes which can be uploaded before reaching the quota\\n\\t}\\n\\treturn usage, nil\\n}\",\n \"func print_stats(){\\nfmt.Print(\\\"\\\\nMemory usage statistics:\\\\n\\\")\\nfmt.Printf(\\\"%v names\\\\n\\\",len(name_dir))\\nfmt.Printf(\\\"%v replacement texts\\\\n\\\",len(text_info))\\n}\",\n \"func printIostat(v *gocui.View, s stat.Diskstats) error {\\n\\t// print header\\n\\t_, err := fmt.Fprintf(v, \\\"\\\\033[30;47m Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await r_await w_await %%util\\\\033[0m\\\\n\\\")\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tfor i := 0; i < len(s); i++ {\\n\\t\\t// skip devices which never do IOs\\n\\t\\tif s[i].Completed == 0 {\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\n\\t\\t// print stats\\n\\t\\t_, err := fmt.Fprintf(v, \\\"%20s\\\\t%10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f\\\\n\\\",\\n\\t\\t\\ts[i].Device,\\n\\t\\t\\ts[i].Rmerged, s[i].Wmerged, s[i].Rcompleted, s[i].Wcompleted,\\n\\t\\t\\ts[i].Rsectors, s[i].Wsectors, s[i].Arqsz, s[i].Tweighted,\\n\\t\\t\\ts[i].Await, s[i].Rawait, s[i].Wawait, s[i].Util,\\n\\t\\t)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func stat(t *kernel.Task, d *fs.Dirent, dirPath bool, statAddr usermem.Addr) error {\\n\\tif dirPath && !fs.IsDir(d.Inode.StableAttr) {\\n\\t\\treturn syserror.ENOTDIR\\n\\t}\\n\\tuattr, err := d.Inode.UnstableAttr(t)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\ts := statFromAttrs(t, d.Inode.StableAttr, uattr)\\n\\treturn s.CopyOut(t, statAddr)\\n}\",\n \"func statfsImpl(t *kernel.Task, d *fs.Dirent, addr usermem.Addr) error {\\n\\tinfo, err := d.Inode.StatFS(t)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\t// Construct the statfs structure and copy it out.\\n\\tstatfs := linux.Statfs{\\n\\t\\tType: info.Type,\\n\\t\\t// Treat block size and fragment size as the same, as\\n\\t\\t// most consumers of this structure will expect one\\n\\t\\t// or the other to be filled in.\\n\\t\\tBlockSize: d.Inode.StableAttr.BlockSize,\\n\\t\\tBlocks: info.TotalBlocks,\\n\\t\\t// We don't have the concept of reserved blocks, so\\n\\t\\t// report blocks free the same as available blocks.\\n\\t\\t// This is a normal thing for filesystems, to do, see\\n\\t\\t// udf, hugetlbfs, tmpfs, among others.\\n\\t\\tBlocksFree: info.FreeBlocks,\\n\\t\\tBlocksAvailable: info.FreeBlocks,\\n\\t\\tFiles: info.TotalFiles,\\n\\t\\tFilesFree: info.FreeFiles,\\n\\t\\t// Same as Linux for simple_statfs, see fs/libfs.c.\\n\\t\\tNameLength: linux.NAME_MAX,\\n\\t\\tFragmentSize: d.Inode.StableAttr.BlockSize,\\n\\t\\t// Leave other fields 0 like simple_statfs does.\\n\\t}\\n\\t_, err = t.CopyOut(addr, &statfs)\\n\\treturn err\\n}\",\n \"func TestDiskUsage(t *testing.T) {\\n\\tt.Parallel()\\n\\twinners := []parameters{[]string{\\\"/\\\", \\\"99\\\"}, []string{\\\"/\\\", \\\"98\\\"}}\\n\\tlosers := []parameters{[]string{\\\"/\\\", \\\"1\\\"}, []string{\\\"/\\\", \\\"2\\\"}}\\n\\ttestInputs(t, diskUsage, winners, losers)\\n}\",\n \"func (fsys *FS) stat(node mountlib.Node, stat *fuse.Stat_t) (errc int) {\\n\\tvar Size uint64\\n\\tvar Blocks uint64\\n\\tvar modTime time.Time\\n\\tvar Mode os.FileMode\\n\\tswitch x := node.(type) {\\n\\tcase *mountlib.Dir:\\n\\t\\tmodTime = x.ModTime()\\n\\t\\tMode = mountlib.DirPerms | fuse.S_IFDIR\\n\\tcase *mountlib.File:\\n\\t\\tvar err error\\n\\t\\tmodTime, Size, Blocks, err = x.Attr(mountlib.NoModTime)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn translateError(err)\\n\\t\\t}\\n\\t\\tMode = mountlib.FilePerms | fuse.S_IFREG\\n\\t}\\n\\t//stat.Dev = 1\\n\\tstat.Ino = node.Inode() // FIXME do we need to set the inode number?\\n\\tstat.Mode = uint32(Mode)\\n\\tstat.Nlink = 1\\n\\tstat.Uid = mountlib.UID\\n\\tstat.Gid = mountlib.GID\\n\\t//stat.Rdev\\n\\tstat.Size = int64(Size)\\n\\tt := fuse.NewTimespec(modTime)\\n\\tstat.Atim = t\\n\\tstat.Mtim = t\\n\\tstat.Ctim = t\\n\\tstat.Blksize = 512\\n\\tstat.Blocks = int64(Blocks)\\n\\tstat.Birthtim = t\\n\\t// fs.Debugf(nil, \\\"stat = %+v\\\", *stat)\\n\\treturn 0\\n}\",\n \"func Info(path string) (int64, int64, int64, int64, int64, int64, error) {\\n\\tstatfs := &unix.Statfs_t{}\\n\\terr := unix.Statfs(path, statfs)\\n\\tif err != nil {\\n\\t\\treturn 0, 0, 0, 0, 0, 0, err\\n\\t}\\n\\n\\t// Available is blocks available * fragment size\\n\\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\\n\\n\\t// Capacity is total block count * fragment size\\n\\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\\n\\n\\t// Usage is block being used * fragment size (aka block size).\\n\\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\\n\\n\\tinodes := int64(statfs.Files)\\n\\tinodesFree := int64(statfs.Ffree)\\n\\tinodesUsed := inodes - inodesFree\\n\\n\\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\\n}\",\n \"func getFSSizeMb(f *framework.Framework, pod *v1.Pod) (int64, error) {\\n\\tvar output string\\n\\tvar err error\\n\\n\\tif supervisorCluster {\\n\\t\\tnamespace := getNamespaceToRunTests(f)\\n\\t\\tcmd := []string{\\\"exec\\\", pod.Name, \\\"--namespace=\\\" + namespace, \\\"--\\\", \\\"/bin/sh\\\", \\\"-c\\\", \\\"df -Tkm | grep /mnt/volume1\\\"}\\n\\t\\toutput = framework.RunKubectlOrDie(namespace, cmd...)\\n\\t\\tgomega.Expect(strings.Contains(output, ext4FSType)).NotTo(gomega.BeFalse())\\n\\t} else {\\n\\t\\toutput, _, err = fvolume.PodExec(f, pod, \\\"df -T -m | grep /mnt/volume1\\\")\\n\\t\\tif err != nil {\\n\\t\\t\\treturn -1, fmt.Errorf(\\\"unable to find mount path via `df -T`: %v\\\", err)\\n\\t\\t}\\n\\t}\\n\\n\\tarrMountOut := strings.Fields(string(output))\\n\\tif len(arrMountOut) <= 0 {\\n\\t\\treturn -1, fmt.Errorf(\\\"error when parsing output of `df -T`. output: %s\\\", string(output))\\n\\t}\\n\\tvar devicePath, strSize string\\n\\tdevicePath = arrMountOut[0]\\n\\tif devicePath == \\\"\\\" {\\n\\t\\treturn -1, fmt.Errorf(\\\"error when parsing output of `df -T` to find out devicePath of /mnt/volume1. output: %s\\\",\\n\\t\\t\\tstring(output))\\n\\t}\\n\\tstrSize = arrMountOut[2]\\n\\tif strSize == \\\"\\\" {\\n\\t\\treturn -1, fmt.Errorf(\\\"error when parsing output of `df -T` to find out size of /mnt/volume1: output: %s\\\",\\n\\t\\t\\tstring(output))\\n\\t}\\n\\n\\tintSizeInMb, err := strconv.ParseInt(strSize, 10, 64)\\n\\tif err != nil {\\n\\t\\treturn -1, fmt.Errorf(\\\"failed to parse size %s into int size\\\", strSize)\\n\\t}\\n\\n\\treturn intSizeInMb, nil\\n}\",\n \"func getMemUsage(memStat types.MemoryStats) uint64 {\\n\\t// Version 1 of the Linux cgroup API uses total_inactive_file\\n\\tif v, ok := memStat.Stats[\\\"total_inactive_file\\\"]; ok && v < memStat.Usage {\\n\\t\\treturn memStat.Usage - v\\n\\t}\\n\\n\\t// Version 2 of the Linux cgroup API uses inactive_file\\n\\tif v := memStat.Stats[\\\"inactive_file\\\"]; v < memStat.Usage {\\n\\t\\treturn memStat.Usage - v\\n\\t}\\n\\n\\treturn memStat.Usage\\n}\",\n \"func DiskUsed() (uint64, error) {\\n\\tusage, err := Disk()\\n\\t// for i := 0; i < len(usage); i++ {\\n\\tif len(usage) > 0 {\\n\\t\\tuseDisk := usage[0].Used\\n\\t\\treturn useDisk, err\\n\\t}\\n\\n\\treturn 0, err\\n}\",\n \"func stat(name string) (*info, error) {\\n\\tf, err := os.Open(filepath.Clean(name))\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tvar di info\\n\\n\\tvar buf bytes.Buffer\\n\\n\\t_, err = io.CopyN(&buf, f, int64(binary.Size(di.Header)))\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tfi, err := f.Stat()\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\terr = f.Close()\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\terr = binary.Read(&buf, binary.LittleEndian, &di.Header)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tif di.Header.MagicNumber != 0x564d444b { // SPARSE_MAGICNUMBER\\n\\t\\treturn nil, ErrInvalidFormat\\n\\t}\\n\\n\\tif di.Header.Flags&(1<<16) == 0 { // SPARSEFLAG_COMPRESSED\\n\\t\\t// Needs to be converted, for example:\\n\\t\\t// vmware-vdiskmanager -r src.vmdk -t 5 dst.vmdk\\n\\t\\t// qemu-img convert -O vmdk -o subformat=streamOptimized src.vmdk dst.vmdk\\n\\t\\treturn nil, ErrInvalidFormat\\n\\t}\\n\\n\\tdi.Capacity = di.Header.Capacity * 512 // VMDK_SECTOR_SIZE\\n\\tdi.Size = fi.Size()\\n\\tdi.Name = filepath.Base(name)\\n\\tdi.ImportName = strings.TrimSuffix(di.Name, \\\".vmdk\\\")\\n\\n\\treturn &di, nil\\n}\",\n \"func (c *Cache) updateStats() error {\\n\\tvar newUsed int64\\n\\terr := c.walk(func(osPath string, fi os.FileInfo, name string) error {\\n\\t\\tif !fi.IsDir() {\\n\\t\\t\\t// Update the atime with that of the file\\n\\t\\t\\tatime := times.Get(fi).AccessTime()\\n\\t\\t\\tc.updateStat(name, atime, fi.Size())\\n\\t\\t\\tnewUsed += fi.Size()\\n\\t\\t} else {\\n\\t\\t\\tc.cacheDir(name)\\n\\t\\t}\\n\\t\\treturn nil\\n\\t})\\n\\tc.itemMu.Lock()\\n\\tc.used = newUsed\\n\\tc.itemMu.Unlock()\\n\\treturn err\\n}\",\n \"func (a *app) printStat() {\\n\\tfmt.Printf(\\\"Total capacity %vgb \\\\n\\\", a.sq.Limit/1024/1024/1024)\\n\\tfmt.Printf(\\\"Usage %vmb \\\\n\\\", a.sq.Usage/1024/1024)\\n\\tfmt.Printf(\\\"In Drive %vmb \\\\n\\\", a.sq.UsageInDrive/1024/1024)\\n\\tfmt.Printf(\\\"In Trash %vmb \\\\n\\\", a.sq.UsageInDriveTrash/1024/1024)\\n\\n\\tperOfUse := float64(a.sq.UsageInDrive) / float64(a.sq.Limit) * 100\\n\\tfmt.Println(\\\"Current percentage of use:\\\", perOfUse)\\n}\",\n \"func (s *SystemMetrics) GetDiskIOStats(c chan *DiskStats) {\\n\\tbefore, err := disk.Get()\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\n\\ttime.Sleep(time.Duration(time.Second * 3))\\n\\n\\tnow, err := disk.Get()\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\n\\tc <- &DiskStats{\\n\\t\\t// default is always in bytes. hence, convert into the required format.\\n\\t\\tDiskIO: (int(now.Used) - int(before.Used)) / 1000, // in kilo-bytes\\n\\t\\tCached: int(now.Cached / 1000000), // mega-bytes\\n\\t}\\n}\",\n \"func DiskTotal() (uint64, error) {\\n\\tusage, err := Disk()\\n\\t// for i := 0; i < len(usage); i++ {\\n\\tif len(usage) > 0 {\\n\\t\\ttotalDisk := usage[0].Total\\n\\t\\treturn totalDisk, err\\n\\t}\\n\\n\\treturn 0, err\\n}\",\n \"func GetFSInfo(ch chan metrics.Metric) {\\n\\tmountedFS, err := disk.Partitions(false)\\n\\tif err != nil {\\n\\t\\tlog.WithFields(log.Fields{\\n\\t\\t\\t\\\"collector\\\": \\\"filesystem\\\",\\n\\t\\t\\t\\\"os\\\": \\\"linux\\\",\\n\\t\\t\\t\\\"action\\\": \\\"GetFileSystems\\\",\\n\\t\\t}).Errorf(\\\"Unable to find mounted filesystems: %+v\\\", err)\\n\\t}\\n\\tfor _, FSs := range mountedFS {\\n\\t\\tfsStats, err := disk.Usage(FSs.Mountpoint)\\n\\t\\tif err != nil {\\n\\t\\t\\tlog.WithFields(log.Fields{\\n\\t\\t\\t\\t\\\"collector\\\": \\\"filesystem\\\",\\n\\t\\t\\t\\t\\\"os\\\": \\\"linux\\\",\\n\\t\\t\\t\\t\\\"action\\\": \\\"GetFSStats\\\",\\n\\t\\t\\t}).Errorf(\\\"Unable to get stats from mounted filesystem: %+v\\\", err)\\n\\t\\t}\\n\\t\\tlog.WithFields(log.Fields{\\n\\t\\t\\t\\\"collector\\\": \\\"filesystem\\\",\\n\\t\\t\\t\\\"os\\\": \\\"linux\\\",\\n\\t\\t}).Debug(fsStats)\\n\\t\\tfsStat := metrics.FileSystem{\\n\\t\\t\\tFileSystem: *fsStats,\\n\\t\\t}\\n\\t\\tch <- fsStat\\n\\t}\\n}\",\n \"func (i *ImageService) LayerDiskUsage(ctx context.Context) (int64, error) {\\n\\tvar allLayersSize int64\\n\\t// TODO(thaJeztah): do we need to take multiple snapshotters into account? See https://github.com/moby/moby/issues/45273\\n\\tsnapshotter := i.client.SnapshotService(i.snapshotter)\\n\\tsnapshotter.Walk(ctx, func(ctx context.Context, info snapshots.Info) error {\\n\\t\\tusage, err := snapshotter.Usage(ctx, info.Name)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\tallLayersSize += usage.Size\\n\\t\\treturn nil\\n\\t})\\n\\treturn allLayersSize, nil\\n}\",\n \"func DiskPercent() (string, error) {\\n\\tusage, err := Disk()\\n\\tif len(usage) > 0 {\\n\\t\\tuseDisk := fmt.Sprintf(\\\"%.2f\\\", usage[0].UsedPercent)\\n\\t\\treturn useDisk, err\\n\\t}\\n\\n\\treturn \\\"0.00\\\", err\\n}\",\n \"func getRemainingDiskSpace(ctxPtr *volumemgrContext) (uint64, error) {\\n\\n\\tvar totalDiskSize uint64\\n\\n\\tpubContentTree := ctxPtr.pubContentTreeStatus\\n\\titemsContentTree := pubContentTree.GetAll()\\n\\tfor _, iterContentTreeStatusJSON := range itemsContentTree {\\n\\t\\titerContentTreeStatus := iterContentTreeStatusJSON.(types.ContentTreeStatus)\\n\\t\\tif iterContentTreeStatus.State < types.LOADED {\\n\\t\\t\\tlog.Tracef(\\\"Content tree %s State %d < LOADED\\\",\\n\\t\\t\\t\\titerContentTreeStatus.Key(), iterContentTreeStatus.State)\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\ttotalDiskSize += uint64(iterContentTreeStatus.CurrentSize)\\n\\t}\\n\\n\\tpubVolume := ctxPtr.pubVolumeStatus\\n\\titemsVolume := pubVolume.GetAll()\\n\\tfor _, iterVolumeStatusJSON := range itemsVolume {\\n\\t\\titerVolumeStatus := iterVolumeStatusJSON.(types.VolumeStatus)\\n\\t\\t// we start consume space when moving into CREATING_VOLUME state\\n\\t\\tif iterVolumeStatus.State < types.CREATING_VOLUME {\\n\\t\\t\\tlog.Tracef(\\\"Volume %s State %d < CREATING_VOLUME\\\",\\n\\t\\t\\t\\titerVolumeStatus.Key(), iterVolumeStatus.State)\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\ttotalDiskSize += volumehandlers.GetVolumeHandler(log, ctxPtr, &iterVolumeStatus).UsageFromStatus()\\n\\t}\\n\\tdeviceDiskUsage, err := diskmetrics.PersistUsageStat(log)\\n\\tif err != nil {\\n\\t\\terr := fmt.Errorf(\\\"Failed to get diskUsage for /persist. err: %s\\\", err)\\n\\t\\tlog.Error(err)\\n\\t\\treturn 0, err\\n\\t}\\n\\tdeviceDiskSize := deviceDiskUsage.Total\\n\\tdiskReservedForDom0 := diskmetrics.Dom0DiskReservedSize(log, ctxPtr.globalConfig, deviceDiskSize)\\n\\tvar allowedDeviceDiskSize uint64\\n\\tif deviceDiskSize < diskReservedForDom0 {\\n\\t\\terr = fmt.Errorf(\\\"Total Disk Size(%d) <= diskReservedForDom0(%d)\\\",\\n\\t\\t\\tdeviceDiskSize, diskReservedForDom0)\\n\\t\\tlog.Errorf(\\\"***getRemainingDiskSpace: err: %s\\\", err)\\n\\t\\treturn uint64(0), err\\n\\t}\\n\\tallowedDeviceDiskSize = deviceDiskSize - diskReservedForDom0\\n\\tif allowedDeviceDiskSize < totalDiskSize {\\n\\t\\treturn 0, nil\\n\\t} else {\\n\\t\\treturn allowedDeviceDiskSize - totalDiskSize, nil\\n\\t}\\n}\",\n \"func (s *Server) statproc() {\\n\\tvar (\\n\\t\\tv *volume.Volume\\n\\t\\tolds *stat.Stats\\n\\t\\tnews = new(stat.Stats)\\n\\t)\\n\\tfor {\\n\\t\\tolds = s.info.Stats\\n\\t\\t*news = *olds\\n\\t\\ts.info.Stats = news // use news instead, for current display\\n\\t\\tolds.Reset()\\n\\t\\tfor _, v = range s.store.Volumes {\\n\\t\\t\\tv.Stats.Calc()\\n\\t\\t\\tolds.Merge(v.Stats)\\n\\t\\t}\\n\\t\\tolds.Calc()\\n\\t\\ts.info.Stats = olds\\n\\t\\ttime.Sleep(statDuration)\\n\\t}\\n}\",\n \"func ShowDiskStatus(c *gin.Context) {\\n\\tdiskinfo, err := diskInfo()\\n\\tif err != nil {\\n\\t\\tlog.Fatal(\\\"Get Disk information error: \\\", err.Error())\\n\\t}\\n\\tutils.Render(\\n\\t\\tc,\\n\\t\\tgin.H{\\n\\t\\t\\t\\\"title\\\": \\\"System Status\\\",\\n\\t\\t\\t\\\"Name\\\": \\\"System base information\\\",\\n\\t\\t\\t\\\"payload\\\": diskinfo,\\n\\t\\t},\\n\\t\\t\\\"stat.html\\\",\\n\\t)\\n}\",\n \"func TestParseDiskUsageBasic(t *testing.T) {\\n\\tcontent, err := ioutil.ReadFile(\\\"../test_resources/stdout\\\")\\n\\tif err != nil {\\n\\t\\tt.Fatal(err)\\n\\t}\\n\\ts := string(content)\\n\\n\\tdiskUsage, _ := parseDiskUsage(&s)\\n\\tif diskUsage != 1111.0 {\\n\\t\\tt.Errorf(\\\"json value returned unexpected: got %f want %f\\\",\\n\\t\\t\\tdiskUsage, 1111.0)\\n\\t}\\n}\",\n \"func DiskFree() (uint64, error) {\\n\\tusage, err := Disk()\\n\\t// for i := 0; i < len(usage); i++ {\\n\\tif len(usage) > 0 {\\n\\t\\tfreeDisk := usage[0].Free\\n\\t\\treturn freeDisk, err\\n\\t}\\n\\n\\treturn 0, err\\n}\",\n \"func (c *CryptohomeBinary) GetAccountDiskUsage(ctx context.Context, username string) ([]byte, error) {\\n\\treturn c.call(ctx, \\\"--action=get_account_disk_usage\\\", \\\"--user=\\\"+username)\\n}\",\n \"func getHDDString() string {\\n\\tpartitions, _ := disk.Partitions(false)\\n\\tbiggestDiskSize := uint64(0)\\n\\tbiggestDiskUsed := uint64(0)\\n\\tbiggestDiskName := \\\"\\\"\\n\\tfor _, partition := range partitions {\\n\\t\\td, _ := disk.Usage(partition.Mountpoint)\\n\\t\\tif d.Total > biggestDiskSize {\\n\\t\\t\\tbiggestDiskName = partition.Mountpoint\\n\\t\\t\\tbiggestDiskUsed = d.Used\\n\\t\\t\\tbiggestDiskSize = d.Total\\n\\t\\t}\\n\\t}\\n\\treturn fmt.Sprintf(\\\"%v: %v/%v\\\", biggestDiskName, formatSize(biggestDiskUsed, 1000), formatSize(biggestDiskSize, 1000))\\n}\",\n \"func GetDiskStats(disks cmn.SimpleKVs) DiskStats {\\n\\tif len(disks) < largeNumDisks {\\n\\t\\toutput := make(DiskStats, len(disks))\\n\\n\\t\\tfor disk := range disks {\\n\\t\\t\\tstat, ok := readSingleDiskStat(disk)\\n\\t\\t\\tif !ok {\\n\\t\\t\\t\\tcontinue\\n\\t\\t\\t}\\n\\t\\t\\toutput[disk] = stat\\n\\t\\t}\\n\\t\\treturn output\\n\\t}\\n\\n\\treturn readMultipleDiskStats(disks)\\n}\",\n \"func (fs *FS) fsInfo(ctx context.Context, path string) (int64, int64, int64, int64, int64, int64, error) {\\n\\tstatfs := &unix.Statfs_t{}\\n\\terr := unix.Statfs(path, statfs)\\n\\tif err != nil {\\n\\t\\treturn 0, 0, 0, 0, 0, 0, err\\n\\t}\\n\\n\\t// Available is blocks available * fragment size\\n\\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\\n\\n\\t// Capacity is total block count * fragment size\\n\\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\\n\\n\\t// Usage is block being used * fragment size (aka block size).\\n\\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\\n\\n\\tinodes := int64(statfs.Files)\\n\\tinodesFree := int64(statfs.Ffree)\\n\\tinodesUsed := inodes - inodesFree\\n\\n\\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\\n}\",\n \"func (c *Container) statInsideMount(containerPath string) (*copier.StatForItem, string, string, error) {\\n\\tresolvedRoot := \\\"/\\\"\\n\\tresolvedPath := c.pathAbs(containerPath)\\n\\tvar statInfo *copier.StatForItem\\n\\n\\terr := c.joinMountAndExec(\\n\\t\\tfunc() error {\\n\\t\\t\\tvar statErr error\\n\\t\\t\\tstatInfo, statErr = secureStat(resolvedRoot, resolvedPath)\\n\\t\\t\\treturn statErr\\n\\t\\t},\\n\\t)\\n\\n\\treturn statInfo, resolvedRoot, resolvedPath, err\\n}\",\n \"func RunListDisk() {\\n\\n\\t// dir, err := filepath.Abs(filepath.Dir(os.Args[0]))\\n\\t// if err != nil {\\n\\t// \\tlog.Fatal(err)\\n\\t// \\treturn\\n\\t// }\\n\\n\\t// lsscsipath := path.Join(dir, \\\"lsscsi\\\")\\n\\t// if _, err := os.Stat(lsscsipath); os.IsNotExist(err) {\\n\\t// \\tlsscsipath = \\\"lsscsi\\\"\\n\\t// }\\n\\tlsscsipath := \\\"lsscsi\\\"\\n\\tcmd := exec.Command(lsscsipath, \\\"-s\\\", \\\"-g\\\")\\n\\tstdout, err := cmd.StdoutPipe()\\n\\tif err != nil {\\n\\t\\tlog.Fatal(err)\\n\\t}\\n\\tif err := cmd.Start(); err != nil {\\n\\t\\tlog.Fatal(err)\\n\\t}\\n\\n\\ttimer := time.AfterFunc(10*time.Second, func() {\\n\\t\\tcmd.Process.Kill()\\n\\t})\\n\\n\\tscanner := bufio.NewScanner(stdout)\\n\\tvar hddinfo []string\\n\\tvar hddchanged bool\\n\\tvar wg sync.WaitGroup\\n\\tfor scanner.Scan() {\\n\\t\\tss := scanner.Text()\\n\\t\\tfmt.Println(ss)\\n\\t\\thddinfo = append(hddinfo, ss)\\n\\t\\tif !DetectData.MatchKey(ss) {\\n\\t\\t\\thddchanged = true\\n\\t\\t}\\n\\t\\tif !DetectData.ContainsKey(ss) {\\n\\t\\t\\t//\\\\s Matches any white-space character.\\n\\t\\t\\tr := regexp.MustCompile(`^([\\\\s\\\\S]{13})(disk[\\\\s\\\\S]{4})([\\\\s\\\\S]{9})([\\\\s\\\\S]{17})([\\\\s\\\\S]{6})([\\\\s\\\\S]{11})([\\\\s\\\\S]{11})([\\\\s\\\\S]+)$`)\\n\\t\\t\\tdiskinfos := r.FindStringSubmatch(ss)\\n\\t\\t\\tif len(diskinfos) == 9 {\\n\\t\\t\\t\\tvar dddect = NewSyncDataDetect()\\n\\t\\t\\t\\tdddect.detectHDD.Locpath = strings.Trim(diskinfos[1], \\\" \\\")\\n\\t\\t\\t\\tdddect.detectHDD.Type = strings.Trim(diskinfos[2], \\\" \\\")\\n\\t\\t\\t\\tdddect.detectHDD.Manufacture = strings.Trim(diskinfos[3], \\\" \\\")\\n\\t\\t\\t\\tdddect.detectHDD.Model = strings.Trim(diskinfos[4], \\\" \\\")\\n\\t\\t\\t\\tdddect.detectHDD.Version = strings.Trim(diskinfos[5], \\\" \\\")\\n\\t\\t\\t\\tdddect.detectHDD.LinuxName = strings.Trim(diskinfos[6], \\\" \\\")\\n\\t\\t\\t\\tdddect.detectHDD.SGLibName = strings.Trim(diskinfos[7], \\\" \\\")\\n\\t\\t\\t\\tdddect.detectHDD.Size = strings.Trim(diskinfos[8], \\\" \\\")\\n\\n\\t\\t\\t\\tif strings.Index(dddect.detectHDD.LinuxName, `/dev/`) == -1 {\\n\\t\\t\\t\\t\\tcontinue\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\t//hddchanged = true\\n\\t\\t\\t\\tDetectData.AddValue(ss, dddect)\\n\\t\\t\\t\\twg.Add(1)\\n\\t\\t\\t\\tgo dddect.ReadDataFromSmartCtl(&wg)\\n\\t\\t\\t}\\n\\t\\t} else {\\n\\t\\t\\tif vv, ok := DetectData.Get(ss); ok {\\n\\t\\t\\t\\tif len(vv.detectHDD.UILabel) == 0 && len(vv.detectHDD.Otherinfo) == 0 {\\n\\t\\t\\t\\t\\twg.Add(1)\\n\\t\\t\\t\\t\\tgo vv.ReadDataFromSmartCtl(&wg)\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\ttimer.Stop()\\n\\tDetectData.RemoveOld(hddinfo)\\n\\n\\ttime.Sleep(4 * time.Second)\\n\\n\\tif hddchanged {\\n\\t\\tfmt.Print(\\\"changed!\\\")\\n\\t\\tcclist, err := configxmldata.Conf.GetCardListIndex()\\n\\t\\tif err == nil {\\n\\t\\t\\tfor _, i := range cclist {\\n\\t\\t\\t\\twg.Add(1)\\n\\t\\t\\t\\tgo SASHDDinfo.RunCardInfo(i, &wg)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tfor i := 0; i < 30; i++ {\\n\\t\\t\\tif waitTimeout(&wg, 10*time.Second) {\\n\\t\\t\\t\\tfmt.Println(\\\"Timed out waiting for wait group\\\")\\n\\t\\t\\t\\tMergeCalibration()\\n\\t\\t\\t} else {\\n\\t\\t\\t\\tfmt.Println(\\\"Wait group finished\\\")\\n\\t\\t\\t\\tMergeCalibration()\\n\\t\\t\\t\\tbreak\\n\\t\\t\\t}\\n\\t\\t}\\n\\t} else {\\n\\t\\twaitTimeout(&wg, 300*time.Second)\\n\\t}\\n\\n}\",\n \"func (o *OSS) ReturnSize(groupID int64) error {\\n\\n partLine := partLine()\\n\\n totalData := map[string]map[string]int{}\\n wg := &sync.WaitGroup{}\\n ch := make(chan base.BaseInfo, 1000)\\n wg.Add(2)\\n go register(groupID, ch, wg)\\n go fileCalc(groupID, ch, wg, o, totalData)\\n\\n time.Sleep(2 * time.Second)\\n wg.Wait()\\n\\n for t := range totalData {\\n ts := strconv.Itoa(totalData[t][\\\"totalSize\\\"])\\n\\n write.CreateFile(t, partLine + \\\"\\\\n\\\")\\n write.CreateFile(t, fmt.Sprintf(\\\"Total: RecordCount: %d ; FileCount: %d ; FileSize: %s .\\\\n\\\",totalData[t][\\\"RecordCount\\\"],totalData[t][\\\"totalCount\\\"], utils.FormatSize(ts) ))\\n }\\n return nil\\n}\",\n \"func umountProc(syncFD int) {\\n\\tsyncFile := os.NewFile(uintptr(syncFD), \\\"procfs umount sync FD\\\")\\n\\tbuf := make([]byte, 1)\\n\\tif w, err := syncFile.Write(buf); err != nil || w != 1 {\\n\\t\\tutil.Fatalf(\\\"unable to write into the proc umounter descriptor: %v\\\", err)\\n\\t}\\n\\tsyncFile.Close()\\n\\n\\tvar waitStatus unix.WaitStatus\\n\\tif _, err := unix.Wait4(0, &waitStatus, 0, nil); err != nil {\\n\\t\\tutil.Fatalf(\\\"error waiting for the proc umounter process: %v\\\", err)\\n\\t}\\n\\tif !waitStatus.Exited() || waitStatus.ExitStatus() != 0 {\\n\\t\\tutil.Fatalf(\\\"the proc umounter process failed: %v\\\", waitStatus)\\n\\t}\\n\\tif err := unix.Access(\\\"/proc/self\\\", unix.F_OK); err != unix.ENOENT {\\n\\t\\tutil.Fatalf(\\\"/proc is still accessible\\\")\\n\\t}\\n}\",\n \"func Space(inputs SpaceInputs) error {\\n\\tfilesystemPath := inputs.FilesystemPath\\n\\tdb[\\\"filter\\\"] = inputs.Filter\\n\\tvar err error\\n\\tif !verify(filesystemPath) {\\n\\t\\terr = errors.New(\\\"Path is not abosule path\\\")\\n\\t\\treturn err\\n\\t}\\n\\tstart := time.Now()\\n\\terr = filepath.Walk(filesystemPath, process)\\n\\troot := getDiskUsage(\\\"/\\\")\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tt := time.Now()\\n\\telapsed := t.Sub(start)\\n\\telapsedInMin := elapsed.Minutes()\\n\\tlargeFileProcessed, _ := db[largeFilenameKey]\\n\\tlargeSizeProcessed, _ := db[largeSizeKey]\\n\\ttotalSizeProcessed, _ := db[totalSizeKey]\\n\\tlargeSize, _ := strconv.ParseUint(largeSizeProcessed, base, bitSize)\\n\\ttotalSize, _ := strconv.ParseUint(totalSizeProcessed, base, bitSize)\\n\\tif root.all > 0 {\\n\\t\\ttotalSize = root.all\\n\\t}\\n\\tlargeFileSizePercentage := (float64(largeSize) / float64(totalSize)) * 100.00\\n\\tdata := [][]string{\\n\\t\\t[]string{\\\"LARGE_FILE_NAME\\\", largeFileProcessed, \\\"NA\\\"},\\n\\t\\t[]string{\\\"LARGE_FILE_SIZE\\\", bytefmt.ByteSize(largeSize), strconv.FormatFloat(largeFileSizePercentage, 'f', floatPrec, bitSize)},\\n\\t\\t[]string{\\\"DISK_TOTAL_SIZE\\\", bytefmt.ByteSize(root.all), \\\"NA\\\"},\\n\\t\\t[]string{\\\"DISK_USED_PERCENTAGE\\\", \\\"--\\\", strconv.FormatFloat(root.usedPer, 'f', floatPrec, bitSize)},\\n\\t\\t[]string{\\\"PROCESSING_TIME\\\", strconv.FormatFloat(elapsedInMin, 'f', floatPrec, bitSize) + \\\" min(s)\\\", \\\"NA\\\"},\\n\\t}\\n\\tviews.Print(data)\\n\\treturn err\\n}\",\n \"func (b *Bucket) Statd(path string) (file *File) {\\n\\treq, err := http.NewRequest(\\\"HEAD\\\", b.URL+path+\\\" d\\\", nil)\\n\\tif err != nil {\\n\\t\\tlog.Println(err)\\n\\t\\treturn\\n\\t}\\n\\tresp, err := b.run(req, 0)\\n\\tif err != nil {\\n\\t\\tlog.Println(err)\\n\\t\\treturn\\n\\t}\\n\\tresp.Body.Close()\\n\\tif resp.StatusCode == http.StatusOK {\\n\\t\\tfile = &File{\\n\\t\\t\\tPath: path,\\n\\t\\t\\tLastModified: resp.Header.Get(\\\"Last-Modified\\\"),\\n\\t\\t\\tETag: resp.Header.Get(\\\"ETag\\\"),\\n\\t\\t\\tSize: 0,\\n\\t\\t\\tIsDir: true,\\n\\t\\t}\\n\\t}\\n\\treturn\\n}\",\n \"func (filterdev *NetworkTap) Stats() (*syscall.BpfStat, error) {\\n\\tvar s syscall.BpfStat\\n\\t_, _, err := syscall.Syscall(syscall.SYS_IOCTL, uintptr(filterdev.device.Fd()), syscall.BIOCGSTATS, uintptr(unsafe.Pointer(&s)))\\n\\tif err != 0 {\\n\\t\\treturn nil, syscall.Errno(err)\\n\\t}\\n\\treturn &s, nil\\n}\",\n \"func Deserialize(p []byte) *structs.DiskStats {\\n\\tstts := &structs.DiskStats{}\\n\\tdevF := &flat.Device{}\\n\\tstatsFlat := flat.GetRootAsDiskStats(p, 0)\\n\\tstts.Timestamp = statsFlat.Timestamp()\\n\\tlen := statsFlat.DeviceLength()\\n\\tstts.Device = make([]structs.Device, len)\\n\\tfor i := 0; i < len; i++ {\\n\\t\\tvar dev structs.Device\\n\\t\\tif statsFlat.Device(devF, i) {\\n\\t\\t\\tdev.Major = devF.Major()\\n\\t\\t\\tdev.Minor = devF.Minor()\\n\\t\\t\\tdev.Name = string(devF.Name())\\n\\t\\t\\tdev.ReadsCompleted = devF.ReadsCompleted()\\n\\t\\t\\tdev.ReadsMerged = devF.ReadsMerged()\\n\\t\\t\\tdev.ReadSectors = devF.ReadSectors()\\n\\t\\t\\tdev.ReadingTime = devF.ReadingTime()\\n\\t\\t\\tdev.WritesCompleted = devF.WritesCompleted()\\n\\t\\t\\tdev.WritesMerged = devF.WritesMerged()\\n\\t\\t\\tdev.WrittenSectors = devF.WrittenSectors()\\n\\t\\t\\tdev.WritingTime = devF.WritingTime()\\n\\t\\t\\tdev.IOInProgress = devF.IOInProgress()\\n\\t\\t\\tdev.IOTime = devF.IOTime()\\n\\t\\t\\tdev.WeightedIOTime = devF.WeightedIOTime()\\n\\t\\t}\\n\\t\\tstts.Device[i] = dev\\n\\t}\\n\\treturn stts\\n}\",\n \"func DiskQuota(path string, size ...string) string {\\n\\tparent := id(path)\\n\\texec.Command(\\\"btrfs\\\", \\\"qgroup\\\", \\\"create\\\", \\\"1/\\\"+parent, config.Agent.LxcPrefix+path).Run()\\n\\texec.Command(\\\"btrfs\\\", \\\"qgroup\\\", \\\"assign\\\", \\\"0/\\\"+id(path+\\\"/opt\\\"), \\\"1/\\\"+parent, config.Agent.LxcPrefix+path).Run()\\n\\texec.Command(\\\"btrfs\\\", \\\"qgroup\\\", \\\"assign\\\", \\\"0/\\\"+id(path+\\\"/var\\\"), \\\"1/\\\"+parent, config.Agent.LxcPrefix+path).Run()\\n\\texec.Command(\\\"btrfs\\\", \\\"qgroup\\\", \\\"assign\\\", \\\"0/\\\"+id(path+\\\"/home\\\"), \\\"1/\\\"+parent, config.Agent.LxcPrefix+path).Run()\\n\\texec.Command(\\\"btrfs\\\", \\\"qgroup\\\", \\\"assign\\\", \\\"0/\\\"+id(path+\\\"/rootfs\\\"), \\\"1/\\\"+parent, config.Agent.LxcPrefix+path).Run()\\n\\n\\tif len(size) > 0 && len(size[0]) > 0 {\\n\\t\\tout, err := exec.Command(\\\"btrfs\\\", \\\"qgroup\\\", \\\"limit\\\", size[0]+\\\"G\\\", \\\"1/\\\"+parent, config.Agent.LxcPrefix+path).CombinedOutput()\\n\\t\\tlog.Check(log.ErrorLevel, \\\"Limiting BTRFS group 1/\\\"+parent+\\\" \\\"+string(out), err)\\n\\t\\texec.Command(\\\"btrfs\\\", \\\"quota\\\", \\\"rescan\\\", \\\"-w\\\", config.Agent.LxcPrefix).Run()\\n\\t}\\n\\treturn Stat(path, \\\"quota\\\", false)\\n}\",\n \"func dirSize(path string) (float64, error) {\\n\\tvar size int64\\n\\terr := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error {\\n\\t\\tif err != nil {\\n\\t\\t\\tsize += 0\\n\\t\\t} else {\\n\\t\\t\\tif !info.IsDir() {\\n\\t\\t\\t\\tsize += info.Size()\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\treturn nil\\n\\t})\\n\\tr, _ := decimal.NewFromFloat(float64(size) / (1024 * 1024 * 1024)).Round(2).Float64()\\n\\treturn r, err\\n}\",\n \"func (o ClusterNodeGroupSystemDiskOutput) Count() pulumi.IntPtrOutput {\\n\\treturn o.ApplyT(func(v ClusterNodeGroupSystemDisk) *int { return v.Count }).(pulumi.IntPtrOutput)\\n}\",\n \"func (c *cpusetHandler) Stat(ctr *CgroupControl, m *Metrics) error {\\n\\treturn nil\\n}\",\n \"func (a *app) gatherStat() {\\n\\tabout, err := a.srv.About.Get().Fields(\\\"storageQuota\\\").Do()\\n\\tif err != nil {\\n\\t\\tlog.Fatalf(\\\"Unable to execute an about request: %v\\\", err)\\n\\t}\\n\\n\\ta.sq = about.StorageQuota\\n}\",\n \"func (s Sysinfo) MemUsedPercent() uint64 {\\n\\treturn s.MemUsed() * 100 / s.MemTotal()\\n}\",\n \"func (c *Client) Stat() (count, size int, err error) {\\n\\tline, err := c.Cmd(\\\"%s\\\\r\\\\n\\\", STAT)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\t// Number of messages in maildrop\\n\\tcount, err = strconv.Atoi(strings.Fields(line)[1])\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif count == 0 {\\n\\t\\treturn\\n\\t}\\n\\n\\t// Total size of messages in bytes\\n\\tsize, err = strconv.Atoi(strings.Fields(line)[2])\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif size == 0 {\\n\\t\\treturn\\n\\t}\\n\\treturn\\n}\",\n \"func diskInfo(y *ec2.Instance) diskInfoType {\\n\\n\\tdeviceMap := map[string]int64{}\\n\\n\\trootdevice := *y.RootDeviceName\\n\\tfor _, f := range y.BlockDeviceMappings {\\n\\t\\tdeviceMap[aws.StringValue(f.DeviceName)] = f.Ebs.AttachTime.Unix()\\n\\t}\\n\\tdd := time.Unix(int64(deviceMap[rootdevice]), 0)\\n\\tsplitdate := strings.Split(dd.Format(time.RFC850), \\\" \\\")[1]\\n\\tdiskAttached := len(deviceMap)\\n\\tdItype := diskInfoType{date: splitdate, diskCount: diskAttached}\\n\\treturn dItype\\n\\n}\",\n \"func (c *cpuMeasure) update() error {\\n\\tfile, err := os.Open(stat)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tdefer file.Close()\\n\\n\\tvar lineName string\\n\\tvar n, cpuCount int\\n\\n\\tscanner := bufio.NewScanner(file)\\n\\tfor scanner.Scan() {\\n\\t\\tline := scanner.Text()\\n\\n\\t\\tif strings.Contains(line, \\\"cpu\\\") {\\n\\t\\t\\tn, err = fmt.Sscanf(line,\\n\\t\\t\\t\\t\\\"%s %d %d %d %d %d %d %d %d %d %d\\\", &lineName,\\n\\t\\t\\t\\t&c.cores[cpuCount][0], &c.cores[cpuCount][1],\\n\\t\\t\\t\\t&c.cores[cpuCount][2], &c.cores[cpuCount][3],\\n\\t\\t\\t\\t&c.cores[cpuCount][4], &c.cores[cpuCount][5],\\n\\t\\t\\t\\t&c.cores[cpuCount][6], &c.cores[cpuCount][7],\\n\\t\\t\\t\\t&c.cores[cpuCount][8], &c.cores[cpuCount][9],\\n\\t\\t\\t)\\n\\t\\t\\tcheckSscanf(lineName, err, n, 11)\\n\\t\\t\\tcpuCount++\\n\\t\\t} else if strings.Contains(line, \\\"ctxt\\\") {\\n\\t\\t\\tn, err = fmt.Sscanf(line, \\\"ctxt %d\\\", &c.SwitchCtxt)\\n\\t\\t\\tcheckSscanf(\\\"ctxt\\\", err, n, 1)\\n\\t\\t} else if strings.Contains(line, \\\"btime\\\") {\\n\\t\\t\\tn, err = fmt.Sscanf(line, \\\"btime %d\\\", &c.BootTime)\\n\\t\\t\\tcheckSscanf(\\\"ctxt\\\", err, n, 1)\\n\\t\\t} else if strings.Contains(line, \\\"processes\\\") {\\n\\t\\t\\tn, err = fmt.Sscanf(line, \\\"processes %d\\\", &c.Processes)\\n\\t\\t\\tcheckSscanf(\\\"ctxt\\\", err, n, 1)\\n\\t\\t} else if strings.Contains(line, \\\"procs_running\\\") {\\n\\t\\t\\tn, err = fmt.Sscanf(line, \\\"procs_running %d\\\", &c.ProcsRunning)\\n\\t\\t\\tcheckSscanf(\\\"ctxt\\\", err, n, 1)\\n\\t\\t} else if strings.Contains(line, \\\"procs_blocked\\\") {\\n\\t\\t\\tn, err = fmt.Sscanf(line, \\\"procs_blocked %d\\\", &c.ProcsBlocked)\\n\\t\\t\\tcheckSscanf(\\\"ctxt\\\", err, n, 1)\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func VolumeScanner(pathToScan string) error {\\n\\tresult := make(map[string]DirStats)\\n\\tvar largestDir string\\n\\tlargestSize := int64(0)\\n\\ttotalSize := int64(0)\\n\\tvar cwd string\\n\\terr := filepath.Walk(pathToScan, func(path string, info os.FileInfo, err error) error {\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\tif info.IsDir() {\\n\\t\\t\\tcwd = path\\n\\t\\t\\tresult[cwd] = DirStats{\\n\\t\\t\\t\\tPath: cwd,\\n\\t\\t\\t\\tTotalSize: int64(0),\\n\\t\\t\\t}\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\tdir := filepath.Dir(path)\\n\\t\\tdirStat, _ := result[dir]\\n\\t\\tdirStat.TotalSize += info.Size()\\n\\t\\ttotalSize += info.Size()\\n\\t\\tif dirStat.TotalSize > largestSize {\\n\\t\\t\\tlargestSize = dirStat.TotalSize\\n\\t\\t\\tlargestDir = dir\\n\\t\\t}\\n\\t\\treturn err\\n\\n\\t})\\n\\tdata := [][]string{\\n\\t\\t[]string{\\\"TOTAL_SIZE\\\", bytefmt.ByteSize(uint64(totalSize)), \\\"\\\"},\\n\\t\\t[]string{\\\"LARGEST_DIR\\\", largestDir, \\\"\\\"},\\n\\t\\t[]string{\\\"LARGEST_DIR_SIZE\\\", bytefmt.ByteSize(uint64(largestSize)), \\\"\\\"},\\n\\t}\\n\\n\\tviews.Print(data)\\n\\treturn err\\n\\n}\",\n \"func (u *User) GetDiskUsage() int {\\n\\tif u == nil || u.DiskUsage == nil {\\n\\t\\treturn 0\\n\\t}\\n\\treturn *u.DiskUsage\\n}\",\n \"func report(p *rc.Process, wallTime time.Duration) {\\n\\tsv, err := p.SystemVersion()\\n\\tif err != nil {\\n\\t\\tlog.Println(err)\\n\\t\\treturn\\n\\t}\\n\\n\\tss, err := p.SystemStatus()\\n\\tif err != nil {\\n\\t\\tlog.Println(err)\\n\\t\\treturn\\n\\t}\\n\\n\\tproc, err := p.Stop()\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\trusage, ok := proc.SysUsage().(*syscall.Rusage)\\n\\tif !ok {\\n\\t\\treturn\\n\\t}\\n\\n\\tlog.Println(\\\"Version:\\\", sv.Version)\\n\\tlog.Println(\\\"Alloc:\\\", ss.Alloc/1024, \\\"KiB\\\")\\n\\tlog.Println(\\\"Sys:\\\", ss.Sys/1024, \\\"KiB\\\")\\n\\tlog.Println(\\\"Goroutines:\\\", ss.Goroutines)\\n\\tlog.Println(\\\"Wall time:\\\", wallTime)\\n\\tlog.Println(\\\"Utime:\\\", time.Duration(rusage.Utime.Nano()))\\n\\tlog.Println(\\\"Stime:\\\", time.Duration(rusage.Stime.Nano()))\\n\\tif runtime.GOOS == \\\"darwin\\\" {\\n\\t\\t// Darwin reports in bytes, Linux seems to report in KiB even\\n\\t\\t// though the manpage says otherwise.\\n\\t\\trusage.Maxrss /= 1024\\n\\t}\\n\\tlog.Println(\\\"MaxRSS:\\\", rusage.Maxrss, \\\"KiB\\\")\\n\\n\\tfmt.Printf(\\\"%s,%d,%d,%d,%.02f,%.02f,%.02f,%d\\\\n\\\",\\n\\t\\tsv.Version,\\n\\t\\tss.Alloc/1024,\\n\\t\\tss.Sys/1024,\\n\\t\\tss.Goroutines,\\n\\t\\twallTime.Seconds(),\\n\\t\\ttime.Duration(rusage.Utime.Nano()).Seconds(),\\n\\t\\ttime.Duration(rusage.Stime.Nano()).Seconds(),\\n\\t\\trusage.Maxrss)\\n}\",\n \"func Stat(t *kernel.Task, args arch.SyscallArguments) (uintptr, *kernel.SyscallControl, error) {\\n\\taddr := args[0].Pointer()\\n\\tstatAddr := args[1].Pointer()\\n\\n\\tpath, dirPath, err := copyInPath(t, addr, false /* allowEmpty */)\\n\\tif err != nil {\\n\\t\\treturn 0, nil, err\\n\\t}\\n\\n\\treturn 0, nil, fileOpOn(t, linux.AT_FDCWD, path, true /* resolve */, func(root *fs.Dirent, d *fs.Dirent, _ uint) error {\\n\\t\\treturn stat(t, d, dirPath, statAddr)\\n\\t})\\n}\",\n \"func (rso *RadosStripedObject) Stat() (size uint64, pmtime uint64, err error) {\\n\\tobj := C.CString(rso.ObjectName)\\n\\tdefer C.free(unsafe.Pointer(obj))\\n\\n\\tvar c_size C.uint64_t\\n\\tvar c_time_t C.time_t\\n\\tret := C.rados_striper_stat(rso.Striper, obj, &c_size, &c_time_t)\\n\\tif ret < 0 {\\n\\t\\treturn 0, 0, errors.New(\\\"get stat failed\\\")\\n\\t}\\n\\treturn uint64(c_size), uint64(C.uint64_t(c_time_t)), nil\\n}\",\n \"func (s *Simple) DiskInfo(req *acomm.Request) (interface{}, *url.URL, error) {\\n\\tvar args CPUInfoArgs\\n\\tif err := req.UnmarshalArgs(&args); err != nil {\\n\\t\\treturn nil, nil, err\\n\\t}\\n\\tif args.GuestID == \\\"\\\" {\\n\\t\\treturn nil, nil, errors.New(\\\"missing guest_id\\\")\\n\\t}\\n\\n\\tresult := &DiskInfoResult{\\n\\t\\t&DiskInfo{\\n\\t\\t\\tDevice: \\\"vda1\\\",\\n\\t\\t\\tSize: 10 * (1024 * 1024 * 1024), // 10 GB in bytes\\n\\t\\t},\\n\\t}\\n\\n\\treturn result, nil, nil\\n}\",\n \"func MBUsed() float64 {\\n var m runtime.MemStats\\n runtime.ReadMemStats(&m)\\n return float64(m.TotalAlloc) / BytesPerMBF \\n}\",\n \"func (s *DataNode) incDiskErrCnt(partitionID uint64, err error, flag uint8) {\\n\\tif err == nil {\\n\\t\\treturn\\n\\t}\\n\\tdp := s.space.Partition(partitionID)\\n\\tif dp == nil {\\n\\t\\treturn\\n\\t}\\n\\td := dp.Disk()\\n\\tif d == nil {\\n\\t\\treturn\\n\\t}\\n\\tif !IsDiskErr(err.Error()) {\\n\\t\\treturn\\n\\t}\\n\\tif flag == WriteFlag {\\n\\t\\td.incWriteErrCnt()\\n\\t} else if flag == ReadFlag {\\n\\t\\td.incReadErrCnt()\\n\\t}\\n}\",\n \"func ProcStat(c *gin.Context) {\\n\\tres := CmdExec(\\\"cat /proc/stat | head -n 1 | awk '{$1=\\\\\\\"\\\\\\\";print}'\\\")\\n\\tresArray := strings.Split(res[0], \\\" \\\")\\n\\tvar cpu []int64\\n\\tvar totalcpu, idlecpu int64\\n\\tfor _, v := range resArray {\\n\\t\\ttemp, err := strconv.ParseInt(v, 10, 64)\\n\\t\\tif err == nil {\\n\\t\\t\\tcpu = append(cpu, temp)\\n\\t\\t\\ttotalcpu = totalcpu + temp\\n\\t\\t}\\n\\t}\\n\\tidlecpu = cpu[3]\\n\\tc.JSON(http.StatusOK, gin.H{\\n\\t\\t\\\"totalcpu\\\": totalcpu,\\n\\t\\t\\\"idlecpu\\\": idlecpu,\\n\\t})\\n}\",\n \"func dirents(dir string) []os.FileInfo {\\n\\tentries, err := ioutil.ReadDir(dir)\\n\\tif err != nil {\\n\\t\\tfmt.Fprintf(os.Stderr, \\\"du1: %v\\\\n\\\", err)\\n\\t\\treturn nil\\n\\t}\\n\\treturn entries\\n}\",\n \"func (s *Store) scrubDisk(diskno int, d Disk) {\\n\\t// We use 0 as capacity for token bucket limiter, which means we'll wait\\n\\t// for \\\"s.config.ScrubRate\\\" tokens to be refilled every time we call \\\"Take\\\",\\n\\t// which is fine for our usage here.\\n\\ttb := tokenbucket.New(0, 0)\\n\\n\\tfor {\\n\\t\\t// In case of errors, empty disk, etc., don't spin. PL-1113\\n\\t\\ttime.Sleep(5 * time.Minute)\\n\\n\\t\\trate := s.Config().ScrubRate\\n\\t\\tif rate < 1024 {\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\ttb.SetRate(float32(rate), 0)\\n\\n\\t\\tdir, err := d.OpenDir()\\n\\t\\tif err == core.ErrDiskRemoved {\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\tif err != core.NoError {\\n\\t\\t\\tlog.Errorf(\\\"aborting disk scrub for disk %d, failed to open dir, err=%s\\\", diskno, err)\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\n\\t\\tlog.Infof(\\\"scrub of disk %d starting\\\", diskno)\\n\\n\\t\\tvar scrubbed, ttracts, ok, bad int\\n\\t\\tvar bytes int64\\n\\t\\tstart := time.Now()\\n\\n\\t\\tfor {\\n\\t\\t\\ttracts, terr := d.ReadDir(dir)\\n\\t\\t\\tif terr != core.NoError {\\n\\t\\t\\t\\tbreak\\n\\t\\t\\t}\\n\\t\\t\\tttracts += len(tracts)\\n\\t\\t\\tfor _, tract := range tracts {\\n\\t\\t\\t\\t// If we can't lock it someone else is probably scrubbing it by virtue of reading or writing it.\\n\\t\\t\\t\\tif !s.tryLockTract(tract, READ) {\\n\\t\\t\\t\\t\\tlog.V(5).Infof(\\\"tract %s is busy, won't scrub this iteration\\\", tract)\\n\\t\\t\\t\\t\\tcontinue\\n\\t\\t\\t\\t}\\n\\n\\t\\t\\t\\t// Scrub returns how many bytes it read. We use this to throttle scrubbing to s.config.ScrubRate bytes/sec.\\n\\t\\t\\t\\tn, err := d.Scrub(tract)\\n\\t\\t\\t\\ts.unlock(tract, READ)\\n\\n\\t\\t\\t\\tif err == core.ErrDiskRemoved {\\n\\t\\t\\t\\t\\treturn\\n\\t\\t\\t\\t}\\n\\n\\t\\t\\t\\t// This might sleep so we want to unlock the tract before calling it.\\n\\t\\t\\t\\ttb.Take(float32(n))\\n\\n\\t\\t\\t\\t// Collect and log some stats.\\n\\t\\t\\t\\tif s.maybeReportError(tract, err) {\\n\\t\\t\\t\\t\\tbad++\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\tok++\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\tscrubbed++\\n\\t\\t\\t\\tbytes += n\\n\\t\\t\\t\\tif scrubbed%10 == 0 {\\n\\t\\t\\t\\t\\tlogStats(diskno, start, scrubbed, ok, bad, ttracts, bytes, 2)\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\td.CloseDir(dir)\\n\\t\\tlogStats(diskno, start, scrubbed, ok, bad, ttracts, bytes, 0)\\n\\t}\\n}\",\n \"func GetFSInfo(path string) (total, available int) {\\n\\ttotal = -1\\n\\tavailable = -1\\n\\tvar buf syscall.Statfs_t\\n\\n\\tif syscall.Statfs(path, &buf) != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\tif buf.Bsize <= 0 {\\n\\t\\treturn\\n\\t}\\n\\n\\ttotal = int((uint64(buf.Bsize) * buf.Blocks) / (1000 * 1000))\\n\\tavailable = int((uint64(buf.Bsize) * buf.Bavail) / (1000 * 1000))\\n\\n\\treturn\\n}\",\n \"func GetStat() (uint64, uint64) {\\n\\treturn filesAmount, dirsAmount\\n}\",\n \"func (d *dataUsageCache) dui(path string, buckets []BucketInfo) madmin.DataUsageInfo {\\n\\te := d.find(path)\\n\\tif e == nil {\\n\\t\\t// No entry found, return empty.\\n\\t\\treturn madmin.DataUsageInfo{}\\n\\t}\\n\\tflat := d.flatten(*e)\\n\\tdui := madmin.DataUsageInfo{\\n\\t\\tLastUpdate: d.Info.LastUpdate,\\n\\t\\tObjectsTotalCount: flat.Objects,\\n\\t\\tObjectsTotalSize: uint64(flat.Size),\\n\\t\\tBucketsCount: uint64(len(e.Children)),\\n\\t\\tBucketsUsage: d.bucketsUsageInfo(buckets),\\n\\t}\\n\\tif flat.ReplicationStats != nil {\\n\\t\\tdui.ReplicationPendingSize = flat.ReplicationStats.PendingSize\\n\\t\\tdui.ReplicatedSize = flat.ReplicationStats.ReplicatedSize\\n\\t\\tdui.ReplicationFailedSize = flat.ReplicationStats.FailedSize\\n\\t\\tdui.ReplicationPendingCount = flat.ReplicationStats.PendingCount\\n\\t\\tdui.ReplicationFailedCount = flat.ReplicationStats.FailedCount\\n\\t\\tdui.ReplicaSize = flat.ReplicationStats.ReplicaSize\\n\\t}\\n\\treturn dui\\n}\",\n \"func (s *CPUStat) Usage() float64 {\\n\\treturn s.All.Usage() * float64(len(s.cpus))\\n}\",\n \"func (p *xlStorageDiskIDCheck) monitorDiskStatus() {\\n\\tt := time.NewTicker(5 * time.Second)\\n\\tdefer t.Stop()\\n\\tfn := mustGetUUID()\\n\\tfor range t.C {\\n\\t\\tif len(p.health.tokens) == 0 {\\n\\t\\t\\t// Queue is still full, no need to check.\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\terr := p.storage.WriteAll(context.Background(), minioMetaTmpBucket, fn, []byte{10000: 42})\\n\\t\\tif err != nil {\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\tb, err := p.storage.ReadAll(context.Background(), minioMetaTmpBucket, fn)\\n\\t\\tif err != nil || len(b) != 10001 {\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\terr = p.storage.Delete(context.Background(), minioMetaTmpBucket, fn, DeleteOptions{\\n\\t\\t\\tRecursive: false,\\n\\t\\t\\tForce: false,\\n\\t\\t})\\n\\t\\tif err == nil {\\n\\t\\t\\tlogger.Info(\\\"node(%s): Read/Write/Delete successful, bringing drive %s online. Drive was offline for %s.\\\", globalLocalNodeName, p.storage.String(),\\n\\t\\t\\t\\ttime.Since(time.Unix(0, atomic.LoadInt64(&p.health.lastSuccess))))\\n\\t\\t\\tatomic.StoreInt32(&p.health.status, diskHealthOK)\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t}\\n}\",\n \"func (s *CRDBStorage) Stat(key string) (certmagic.KeyInfo, error) {\\n\\tctx, cancel := context.WithTimeout(context.Background(), defaultQueryTimeout)\\n\\tdefer cancel()\\n\\trow := s.DB.QueryRowContext(ctx, `select length(value), updated from certmagic_values where \\\"key\\\" = $1`, key)\\n\\tinfo := certmagic.KeyInfo{\\n\\t\\tKey: key,\\n\\t\\tIsTerminal: true,\\n\\t}\\n\\tif err := row.Scan(&info.Size, &info.Modified); err != nil {\\n\\t\\tif errors.Is(err, sql.ErrNoRows) {\\n\\t\\t\\treturn certmagic.KeyInfo{}, certmagic.ErrNotExist(err)\\n\\t\\t}\\n\\t\\treturn certmagic.KeyInfo{}, err\\n\\t}\\n\\treturn info, nil\\n}\",\n \"func dirents(dir string) []os.FileInfo {\\r\\n\\tsema <- struct{}{} // acquire token\\r\\n\\tdefer func() { <-sema }() // release token\\r\\n\\r\\n\\tentries, err := ioutil.ReadDir(dir)\\r\\n\\tif err != nil {\\r\\n\\t\\tfmt.Fprintf(os.Stderr, \\\"du: %v\\\\n\\\", err)\\r\\n\\t\\treturn nil\\r\\n\\t}\\r\\n\\treturn entries\\r\\n}\",\n \"func (o LocalDiskOutput) DiskCount() pulumi.IntPtrOutput {\\n\\treturn o.ApplyT(func(v LocalDisk) *int { return v.DiskCount }).(pulumi.IntPtrOutput)\\n}\",\n \"func (e *BackupEnv) GetDirSize(source string) (int64) {\\n directory, _ := os.Open(source);\\n var sum int64 = 0;\\n defer directory.Close();\\n\\n objects, _ := directory.Readdir(-1)\\n for _, obj := range objects {\\n if obj.IsDir() {\\n sum += e.GetDirSize(source + \\\"/\\\" + obj.Name());\\n } else {\\n stat, _ := os.Stat(source + \\\"/\\\" + obj.Name());\\n sum += stat.Size();\\n }\\n }\\n\\n return sum;\\n}\",\n \"func RamUsage() {\\r\\n v, _ := mem.VirtualMemory()\\r\\n fmt.Printf(\\\"RAM{ Total: %v, Free:%v, UsedPercent:%f%%}\\\\n\\\", v.Total, v.Free, v.UsedPercent)\\r\\n}\",\n \"func measureStorageDevice(blkDevicePath string) error {\\n\\tlog.Printf(\\\"Storage Collector: Measuring block device %s\\\\n\\\", blkDevicePath)\\n\\tfile, err := os.Open(blkDevicePath)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"couldn't open disk=%s err=%v\\\", blkDevicePath, err)\\n\\t}\\n\\n\\teventDesc := fmt.Sprintf(\\\"Storage Collector: Measured %s\\\", blkDevicePath)\\n\\treturn tpm.ExtendPCRDebug(pcr, file, eventDesc)\\n}\",\n \"func printSysstat(v *gocui.View, s stat.Stat) error {\\n\\tvar err error\\n\\n\\t/* line1: current time and load average */\\n\\t_, err = fmt.Fprintf(v, \\\"pgcenter: %s, load average: %.2f, %.2f, %.2f\\\\n\\\",\\n\\t\\ttime.Now().Format(\\\"2006-01-02 15:04:05\\\"),\\n\\t\\ts.LoadAvg.One, s.LoadAvg.Five, s.LoadAvg.Fifteen)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t/* line2: cpu usage */\\n\\t_, err = fmt.Fprintf(v, \\\" %%cpu: \\\\033[37;1m%4.1f\\\\033[0m us, \\\\033[37;1m%4.1f\\\\033[0m sy, \\\\033[37;1m%4.1f\\\\033[0m ni, \\\\033[37;1m%4.1f\\\\033[0m id, \\\\033[37;1m%4.1f\\\\033[0m wa, \\\\033[37;1m%4.1f\\\\033[0m hi, \\\\033[37;1m%4.1f\\\\033[0m si, \\\\033[37;1m%4.1f\\\\033[0m st\\\\n\\\",\\n\\t\\ts.CpuStat.User, s.CpuStat.Sys, s.CpuStat.Nice, s.CpuStat.Idle,\\n\\t\\ts.CpuStat.Iowait, s.CpuStat.Irq, s.CpuStat.Softirq, s.CpuStat.Steal)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t/* line3: memory usage */\\n\\t_, err = fmt.Fprintf(v, \\\" MiB mem: \\\\033[37;1m%6d\\\\033[0m total, \\\\033[37;1m%6d\\\\033[0m free, \\\\033[37;1m%6d\\\\033[0m used, \\\\033[37;1m%8d\\\\033[0m buff/cached\\\\n\\\",\\n\\t\\ts.Meminfo.MemTotal, s.Meminfo.MemFree, s.Meminfo.MemUsed,\\n\\t\\ts.Meminfo.MemCached+s.Meminfo.MemBuffers+s.Meminfo.MemSlab)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t/* line4: swap usage, dirty and writeback */\\n\\t_, err = fmt.Fprintf(v, \\\"MiB swap: \\\\033[37;1m%6d\\\\033[0m total, \\\\033[37;1m%6d\\\\033[0m free, \\\\033[37;1m%6d\\\\033[0m used, \\\\033[37;1m%6d/%d\\\\033[0m dirty/writeback\\\\n\\\",\\n\\t\\ts.Meminfo.SwapTotal, s.Meminfo.SwapFree, s.Meminfo.SwapUsed,\\n\\t\\ts.Meminfo.MemDirty, s.Meminfo.MemWriteback)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *NFSCollector) Describe(ch chan<- *prometheus.Desc) {\\n\\tch <- c.Up\\n\\tch <- c.NFSInfo\\n\\tch <- c.DiskFree\\n\\tch <- c.NICInfo\\n\\tch <- c.NICReceive\\n\\tch <- c.NICSend\\n\\n\\tch <- c.MaintenanceScheduled\\n\\tch <- c.MaintenanceInfo\\n\\tch <- c.MaintenanceStartTime\\n\\tch <- c.MaintenanceEndTime\\n}\",\n \"func (f *Fs) About(ctx context.Context) (*fs.Usage, error) {\\n\\topts := rest.Opts{\\n\\t\\tMethod: \\\"PROPFIND\\\",\\n\\t\\tPath: \\\"\\\",\\n\\t\\tExtraHeaders: map[string]string{\\n\\t\\t\\t\\\"Depth\\\": \\\"0\\\",\\n\\t\\t},\\n\\t}\\n\\topts.Body = bytes.NewBuffer([]byte(`\\n\\n \\n \\n \\n \\n\\n`))\\n\\tvar q api.Quota\\n\\tvar resp *http.Response\\n\\tvar err error\\n\\terr = f.pacer.Call(func() (bool, error) {\\n\\t\\tresp, err = f.srv.CallXML(ctx, &opts, nil, &q)\\n\\t\\treturn f.shouldRetry(ctx, resp, err)\\n\\t})\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tusage := &fs.Usage{}\\n\\tif i, err := strconv.ParseInt(q.Used, 10, 64); err == nil && i >= 0 {\\n\\t\\tusage.Used = fs.NewUsageValue(i)\\n\\t}\\n\\tif i, err := strconv.ParseInt(q.Available, 10, 64); err == nil && i >= 0 {\\n\\t\\tusage.Free = fs.NewUsageValue(i)\\n\\t}\\n\\tif usage.Used != nil && usage.Free != nil {\\n\\t\\tusage.Total = fs.NewUsageValue(*usage.Used + *usage.Free)\\n\\t}\\n\\treturn usage, nil\\n}\",\n \"func updateMetrics(interval time.Duration, diskRoot string) {\\n\\tlogger := logrus.WithField(\\\"sync-loop\\\", \\\"updateMetrics\\\")\\n\\tticker := time.NewTicker(interval)\\n\\tfor ; true; <-ticker.C {\\n\\t\\tlogger.Info(\\\"tick\\\")\\n\\t\\t_, bytesFree, bytesUsed, _, _, _, err := diskutil.GetDiskUsage(diskRoot)\\n\\t\\tif err != nil {\\n\\t\\t\\tlogger.WithError(err).Error(\\\"Failed to get disk metrics\\\")\\n\\t\\t} else {\\n\\t\\t\\tpromMetrics.DiskFree.Set(float64(bytesFree) / 1e9)\\n\\t\\t\\tpromMetrics.DiskUsed.Set(float64(bytesUsed) / 1e9)\\n\\t\\t\\tpromMetrics.DiskTotal.Set(float64(bytesFree+bytesUsed) / 1e9)\\n\\t\\t}\\n\\t}\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.7427519","0.6913226","0.6908944","0.6908944","0.6908944","0.6850082","0.6845622","0.68086916","0.67880803","0.6696417","0.66945654","0.66048414","0.6564812","0.64671034","0.6415186","0.63966525","0.6337184","0.6207931","0.6153916","0.6146029","0.6145258","0.6116076","0.6114989","0.60801154","0.6067873","0.6029154","0.6003057","0.5980582","0.5962313","0.5911625","0.5880454","0.58735424","0.58672833","0.58442676","0.5832619","0.57554525","0.5737849","0.57301164","0.5717931","0.571664","0.571028","0.56971127","0.5675387","0.56729925","0.5661262","0.5639872","0.5628183","0.56103593","0.55855405","0.5575156","0.5534724","0.5530005","0.5512552","0.54882336","0.54822147","0.5462636","0.5426426","0.5410591","0.54026526","0.5390695","0.53873783","0.53803426","0.53712326","0.53668964","0.5360813","0.53590465","0.53558","0.5354744","0.5354114","0.535246","0.5343847","0.53411627","0.5308106","0.52986884","0.52955246","0.529139","0.5272785","0.52600473","0.5257732","0.52522355","0.5250704","0.5249574","0.524948","0.52467304","0.52450097","0.5244901","0.5224547","0.52129126","0.52082425","0.52064675","0.52015716","0.51909375","0.5186846","0.5185947","0.5174491","0.517434","0.51715106","0.51713604","0.51673776","0.51664156"],"string":"[\n \"0.7427519\",\n \"0.6913226\",\n \"0.6908944\",\n \"0.6908944\",\n \"0.6908944\",\n \"0.6850082\",\n \"0.6845622\",\n \"0.68086916\",\n \"0.67880803\",\n \"0.6696417\",\n \"0.66945654\",\n \"0.66048414\",\n \"0.6564812\",\n \"0.64671034\",\n \"0.6415186\",\n \"0.63966525\",\n \"0.6337184\",\n \"0.6207931\",\n \"0.6153916\",\n \"0.6146029\",\n \"0.6145258\",\n \"0.6116076\",\n \"0.6114989\",\n \"0.60801154\",\n \"0.6067873\",\n \"0.6029154\",\n \"0.6003057\",\n \"0.5980582\",\n \"0.5962313\",\n \"0.5911625\",\n \"0.5880454\",\n \"0.58735424\",\n \"0.58672833\",\n \"0.58442676\",\n \"0.5832619\",\n \"0.57554525\",\n \"0.5737849\",\n \"0.57301164\",\n \"0.5717931\",\n \"0.571664\",\n \"0.571028\",\n \"0.56971127\",\n \"0.5675387\",\n \"0.56729925\",\n \"0.5661262\",\n \"0.5639872\",\n \"0.5628183\",\n \"0.56103593\",\n \"0.55855405\",\n \"0.5575156\",\n \"0.5534724\",\n \"0.5530005\",\n \"0.5512552\",\n \"0.54882336\",\n \"0.54822147\",\n \"0.5462636\",\n \"0.5426426\",\n \"0.5410591\",\n \"0.54026526\",\n \"0.5390695\",\n \"0.53873783\",\n \"0.53803426\",\n \"0.53712326\",\n \"0.53668964\",\n \"0.5360813\",\n \"0.53590465\",\n \"0.53558\",\n \"0.5354744\",\n \"0.5354114\",\n \"0.535246\",\n \"0.5343847\",\n \"0.53411627\",\n \"0.5308106\",\n \"0.52986884\",\n \"0.52955246\",\n \"0.529139\",\n \"0.5272785\",\n \"0.52600473\",\n \"0.5257732\",\n \"0.52522355\",\n \"0.5250704\",\n \"0.5249574\",\n \"0.524948\",\n \"0.52467304\",\n \"0.52450097\",\n \"0.5244901\",\n \"0.5224547\",\n \"0.52129126\",\n \"0.52082425\",\n \"0.52064675\",\n \"0.52015716\",\n \"0.51909375\",\n \"0.5186846\",\n \"0.5185947\",\n \"0.5174491\",\n \"0.517434\",\n \"0.51715106\",\n \"0.51713604\",\n \"0.51673776\",\n \"0.51664156\"\n]"},"document_score":{"kind":"string","value":"0.56129587"},"document_rank":{"kind":"string","value":"47"}}},{"rowIdx":104791,"cells":{"query":{"kind":"string","value":"Max returns the larger of x or y."},"document":{"kind":"string","value":"func Max(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func Max(x, y int64) int64 {\n if x > y {\n return x\n }\n return y\n}","func Max(x, y float64) float64 {\n\n\treturn math.Max(x, y)\n}","func max(x, y int64) int64 {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}","func Max[T comdef.XintOrFloat](x, y T) T {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func Max(x int, y int) int {\n\tif x > y { return x }\n\treturn y\n}","func Max(x, y int) int {\n if x < y {\n return y\n }\n return x\n}","func Max(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}","func Max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func Max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func Max(x, y int) int {\n\tif y > x {\n\t\treturn y\n\t}\n\treturn x\n}","func max(x int, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y float64) float64 {\n\tswitch {\n\tcase math.IsNaN(x) || math.IsNaN(y):\n\t\treturn math.NaN()\n\tcase math.IsInf(x, 1) || math.IsInf(y, 1):\n\t\treturn math.Inf(1)\n\n\tcase x == 0 && x == y:\n\t\tif math.Signbit(x) {\n\t\t\treturn y\n\t\t}\n\t\treturn x\n\t}\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}","func Max[T constraints.Ordered](x T, y T) T {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t} else {\n\t\treturn y\n\t}\n}","func Max(x, y int64) int64 {\n\treturn x ^ ((x ^ y) & ((x - y) >> 63))\n}","func (a Vec2) Max(b Vec2) Vec2 {\n\tvar r Vec2\n\tif a.X > b.X {\n\t\tr.X = a.X\n\t} else {\n\t\tr.X = b.X\n\t}\n\tif a.Y > b.Y {\n\t\tr.Y = a.Y\n\t} else {\n\t\tr.Y = b.Y\n\t}\n\treturn r\n}","func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func maxi(x int, y int) int {\n if x >= y {\n return x\n } else {\n return y\n }\n}","func Max(x, y *big.Int) *big.Int {\n\tif x.Cmp(y) == 1 {\n\t\treturn x\n\t}\n\treturn y\n}","func Max(arg, arg2 float64) float64 {\n\treturn math.Max(arg, arg2)\n}","func Max(a uint64, b uint64) uint64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(x, y int32) int32 {\n\treturn x - (((x - y) >> 31) & (x - y))\n}","func Max(x, y string) string {\n\tif Compare(x, y) < 0 {\n\t\treturn y\n\t}\n\treturn x\n}","func MaxFloat64(x, y float64) float64 {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}","func max(a, b int64) int64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int64) int64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a int, b int) int {\n\tif a >= b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\r\n\tif a > b {\r\n\t\treturn a\r\n\t}\r\n\treturn b\r\n}","func Max(a int, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a interface{}, b interface{}) interface{} {\n\tif a == nil {\n\t\treturn b\n\t}\n\tif b == nil {\n\t\treturn a\n\t}\n\tif Cmp(a, b) > 0 {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}","func Max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}","func (m *Arg) GetMax() (val float64, set bool) {\n\tif m.Max == nil {\n\t\treturn\n\t}\n\n\treturn *m.Max, true\n}","func maxPoints(max, other point) point {\n\tif other.x > max.x {\n\t\tmax.x = other.x\n\t}\n\tif other.y > max.y {\n\t\tmax.y = other.y\n\t}\n\treturn max\n}","func Max(a int, b int) int {\n if (b > a) {\n return b;\n }\n\n return a;\n}","func Max(a, b int) int {\n\treturn neogointernal.Opcode2(\"MAX\", a, b).(int)\n}","func Max(valueA gcv.Value, valueB gcv.Value) (gcv.Value, error) {\n\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\n\t\treturn nil, errors.New(\"Max is not supported for Complex numbers\")\n\t}\n\treturn gcv.MakeValue(math.Max(valueA.Real(), valueB.Real())), nil\n}","func Max(i, j int) int {\n\tif i > j {\n\t\treturn i\n\t}\n\treturn j\n}","func max(a int, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(a, b int) int {\n\treturn int(math.Max(float64(a), float64(b)))\n}","func Max(a, b int) int {\n\tif a-b > 0 {\n\t\treturn a\n\t}\n\n\treturn b\n}","func max(a, b int) int {\nif a < b {\nreturn b\n}\nreturn a\n}","func Max(val, max any) bool { return valueCompare(val, max, \"<=\") }","func Max(a, b uint32) uint32 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func Max(numbers ...cty.Value) (cty.Value, error) {\n\treturn MaxFunc.Call(numbers)\n}","func (self *State)Max(a,b any)any{\n self.IncOperations(self.coeff[\"max\"]+self.off[\"max\"])\n return wrap2(a,b,math.Max)\n}","func Max(a, operand int) int {\n\tif a > operand {\n\t\treturn a\n\t}\n\treturn operand\n}","func max(first int, second int) int {\n\tif first >= second {\n\t\treturn first\n\t} else {\n\t\treturn second\n\t}\n}","func max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}","func Max(vals ...float64) float64 {\n\tmax := vals[0]\n\tfor _, v := range vals {\n\t\tif v > max {\n\t\t\tmax = v\n\t\t}\n\t}\n\n\treturn max\n}","func IntMax(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}","func (c *Capsule) Max() Coord {\n\treturn c.P1.Max(c.P2).AddScalar(c.Radius)\n}","func max(a, b int) int {\n\tif b > a {\n\t\treturn b\n\t}\n\treturn a\n}","func Max(xs ...float64) float64 {\n\tm := xs[0]\n\tfor i := 1; i < len(xs); i++ {\n\t\tif m < xs[i] {\n\t\t\tm = xs[i]\n\t\t}\n\t}\n\n\treturn m\n}","func max(a, b int32) int32 {\n\tif a >= b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b ImpactAmount) ImpactAmount {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func max(a, b int32) int32 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func MaxFloat(x, y float64) float64 {\n\treturn math.Max(x, y)\n}","func max(a, b int32) int32 {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}","func MaxI(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func SwapMax[T comdef.XintOrFloat](x, y T) (T, T) {\n\tif x > y {\n\t\treturn x, y\n\t}\n\treturn y, x\n}","func (g *Graph) Max(x1 Node, x2 Node) Node {\n\treturn g.NewOperator(fn.NewMax(x1, x2), x1, x2)\n}","func MinMax(x, min, max int) int { return x }","func MaxInt64(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}","func MaxInt(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}","func max(num1, num2 int) int {\nresult int\n\n\tif (num1 > num2){\n\t\tresult = num1\n\t} else{\n\t\tresult = num2\n\t}\nreturn result\n}","func MaxI64(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func MaxI64(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func (o *Range) GetMax() gdnative.Real {\n\t//log.Println(\"Calling Range.GetMax()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"Range\", \"get_max\")\n\n\t// Call the parent method.\n\t// float\n\tretPtr := gdnative.NewEmptyReal()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n\t// If we have a return type, convert it from a pointer into its actual object.\n\tret := gdnative.NewRealFromPointer(retPtr)\n\treturn ret\n}","func (px *Paxos) Max() int {\n\t// Your code here.\n\treturn px.max\n}","func MaxInt(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func MAX(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}","func MaxInt(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}","func MaxLamport(x, y Lamport) Lamport {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}","func (v *Vector) Max(m *Vector) {\n\tif m.X > v.X {\n\t\tv.X = m.X\n\t}\n\tif m.Y > v.Y {\n\t\tv.Y = m.Y\n\t}\n\tif m.Z > v.Z {\n\t\tv.Z = m.Z\n\t}\n}"],"string":"[\n \"func Max(x, y int64) int64 {\\n if x > y {\\n return x\\n }\\n return y\\n}\",\n \"func Max(x, y float64) float64 {\\n\\n\\treturn math.Max(x, y)\\n}\",\n \"func max(x, y int64) int64 {\\n\\tif x < y {\\n\\t\\treturn y\\n\\t}\\n\\treturn x\\n}\",\n \"func Max[T comdef.XintOrFloat](x, y T) T {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func Max(x int, y int) int {\\n\\tif x > y { return x }\\n\\treturn y\\n}\",\n \"func Max(x, y int) int {\\n if x < y {\\n return y\\n }\\n return x\\n}\",\n \"func Max(x, y int) int {\\n\\tif x < y {\\n\\t\\treturn y\\n\\t}\\n\\treturn x\\n}\",\n \"func Max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func Max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func Max(x, y int) int {\\n\\tif y > x {\\n\\t\\treturn y\\n\\t}\\n\\treturn x\\n}\",\n \"func max(x int, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y float64) float64 {\\n\\tswitch {\\n\\tcase math.IsNaN(x) || math.IsNaN(y):\\n\\t\\treturn math.NaN()\\n\\tcase math.IsInf(x, 1) || math.IsInf(y, 1):\\n\\t\\treturn math.Inf(1)\\n\\n\\tcase x == 0 && x == y:\\n\\t\\tif math.Signbit(x) {\\n\\t\\t\\treturn y\\n\\t\\t}\\n\\t\\treturn x\\n\\t}\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y int) int {\\n\\tif x < y {\\n\\t\\treturn y\\n\\t}\\n\\treturn x\\n}\",\n \"func Max[T constraints.Ordered](x T, y T) T {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func max(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t} else {\\n\\t\\treturn y\\n\\t}\\n}\",\n \"func Max(x, y int64) int64 {\\n\\treturn x ^ ((x ^ y) & ((x - y) >> 63))\\n}\",\n \"func (a Vec2) Max(b Vec2) Vec2 {\\n\\tvar r Vec2\\n\\tif a.X > b.X {\\n\\t\\tr.X = a.X\\n\\t} else {\\n\\t\\tr.X = b.X\\n\\t}\\n\\tif a.Y > b.Y {\\n\\t\\tr.Y = a.Y\\n\\t} else {\\n\\t\\tr.Y = b.Y\\n\\t}\\n\\treturn r\\n}\",\n \"func Max(a, b float64) float64 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b float64) float64 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b float64) float64 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func maxi(x int, y int) int {\\n if x >= y {\\n return x\\n } else {\\n return y\\n }\\n}\",\n \"func Max(x, y *big.Int) *big.Int {\\n\\tif x.Cmp(y) == 1 {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func Max(arg, arg2 float64) float64 {\\n\\treturn math.Max(arg, arg2)\\n}\",\n \"func Max(a uint64, b uint64) uint64 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b float64) float64 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(x, y int32) int32 {\\n\\treturn x - (((x - y) >> 31) & (x - y))\\n}\",\n \"func Max(x, y string) string {\\n\\tif Compare(x, y) < 0 {\\n\\t\\treturn y\\n\\t}\\n\\treturn x\\n}\",\n \"func MaxFloat64(x, y float64) float64 {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\n\\treturn y\\n}\",\n \"func max(a, b int64) int64 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int64) int64 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a int, b int) int {\\n\\tif a >= b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\r\\n\\tif a > b {\\r\\n\\t\\treturn a\\r\\n\\t}\\r\\n\\treturn b\\r\\n}\",\n \"func Max(a int, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a interface{}, b interface{}) interface{} {\\n\\tif a == nil {\\n\\t\\treturn b\\n\\t}\\n\\tif b == nil {\\n\\t\\treturn a\\n\\t}\\n\\tif Cmp(a, b) > 0 {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\n\\tif a < b {\\n\\t\\treturn b\\n\\t}\\n\\treturn a\\n}\",\n \"func Max(a, b int) int {\\n\\tif a < b {\\n\\t\\treturn b\\n\\t}\\n\\treturn a\\n}\",\n \"func (m *Arg) GetMax() (val float64, set bool) {\\n\\tif m.Max == nil {\\n\\t\\treturn\\n\\t}\\n\\n\\treturn *m.Max, true\\n}\",\n \"func maxPoints(max, other point) point {\\n\\tif other.x > max.x {\\n\\t\\tmax.x = other.x\\n\\t}\\n\\tif other.y > max.y {\\n\\t\\tmax.y = other.y\\n\\t}\\n\\treturn max\\n}\",\n \"func Max(a int, b int) int {\\n if (b > a) {\\n return b;\\n }\\n\\n return a;\\n}\",\n \"func Max(a, b int) int {\\n\\treturn neogointernal.Opcode2(\\\"MAX\\\", a, b).(int)\\n}\",\n \"func Max(valueA gcv.Value, valueB gcv.Value) (gcv.Value, error) {\\n\\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\\n\\t\\treturn nil, errors.New(\\\"Max is not supported for Complex numbers\\\")\\n\\t}\\n\\treturn gcv.MakeValue(math.Max(valueA.Real(), valueB.Real())), nil\\n}\",\n \"func Max(i, j int) int {\\n\\tif i > j {\\n\\t\\treturn i\\n\\t}\\n\\treturn j\\n}\",\n \"func max(a int, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(a, b int) int {\\n\\treturn int(math.Max(float64(a), float64(b)))\\n}\",\n \"func Max(a, b int) int {\\n\\tif a-b > 0 {\\n\\t\\treturn a\\n\\t}\\n\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\nif a < b {\\nreturn b\\n}\\nreturn a\\n}\",\n \"func Max(val, max any) bool { return valueCompare(val, max, \\\"<=\\\") }\",\n \"func Max(a, b uint32) uint32 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func Max(numbers ...cty.Value) (cty.Value, error) {\\n\\treturn MaxFunc.Call(numbers)\\n}\",\n \"func (self *State)Max(a,b any)any{\\n self.IncOperations(self.coeff[\\\"max\\\"]+self.off[\\\"max\\\"])\\n return wrap2(a,b,math.Max)\\n}\",\n \"func Max(a, operand int) int {\\n\\tif a > operand {\\n\\t\\treturn a\\n\\t}\\n\\treturn operand\\n}\",\n \"func max(first int, second int) int {\\n\\tif first >= second {\\n\\t\\treturn first\\n\\t} else {\\n\\t\\treturn second\\n\\t}\\n}\",\n \"func max(a, b int) int {\\n\\tif a < b {\\n\\t\\treturn b\\n\\t}\\n\\treturn a\\n}\",\n \"func Max(vals ...float64) float64 {\\n\\tmax := vals[0]\\n\\tfor _, v := range vals {\\n\\t\\tif v > max {\\n\\t\\t\\tmax = v\\n\\t\\t}\\n\\t}\\n\\n\\treturn max\\n}\",\n \"func IntMax(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\n\\treturn y\\n}\",\n \"func (c *Capsule) Max() Coord {\\n\\treturn c.P1.Max(c.P2).AddScalar(c.Radius)\\n}\",\n \"func max(a, b int) int {\\n\\tif b > a {\\n\\t\\treturn b\\n\\t}\\n\\treturn a\\n}\",\n \"func Max(xs ...float64) float64 {\\n\\tm := xs[0]\\n\\tfor i := 1; i < len(xs); i++ {\\n\\t\\tif m < xs[i] {\\n\\t\\t\\tm = xs[i]\\n\\t\\t}\\n\\t}\\n\\n\\treturn m\\n}\",\n \"func max(a, b int32) int32 {\\n\\tif a >= b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b ImpactAmount) ImpactAmount {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func max(a, b int32) int32 {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func MaxFloat(x, y float64) float64 {\\n\\treturn math.Max(x, y)\\n}\",\n \"func max(a, b int32) int32 {\\n\\tif a < b {\\n\\t\\treturn b\\n\\t}\\n\\treturn a\\n}\",\n \"func MaxI(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func SwapMax[T comdef.XintOrFloat](x, y T) (T, T) {\\n\\tif x > y {\\n\\t\\treturn x, y\\n\\t}\\n\\treturn y, x\\n}\",\n \"func (g *Graph) Max(x1 Node, x2 Node) Node {\\n\\treturn g.NewOperator(fn.NewMax(x1, x2), x1, x2)\\n}\",\n \"func MinMax(x, min, max int) int { return x }\",\n \"func MaxInt64(x, y int64) int64 {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\n\\treturn y\\n}\",\n \"func MaxInt(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\n\\treturn y\\n}\",\n \"func max(num1, num2 int) int {\\nresult int\\n\\n\\tif (num1 > num2){\\n\\t\\tresult = num1\\n\\t} else{\\n\\t\\tresult = num2\\n\\t}\\nreturn result\\n}\",\n \"func MaxI64(x, y int64) int64 {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func MaxI64(x, y int64) int64 {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func (o *Range) GetMax() gdnative.Real {\\n\\t//log.Println(\\\"Calling Range.GetMax()\\\")\\n\\n\\t// Build out the method's arguments\\n\\tptrArguments := make([]gdnative.Pointer, 0, 0)\\n\\n\\t// Get the method bind\\n\\tmethodBind := gdnative.NewMethodBind(\\\"Range\\\", \\\"get_max\\\")\\n\\n\\t// Call the parent method.\\n\\t// float\\n\\tretPtr := gdnative.NewEmptyReal()\\n\\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\\n\\n\\t// If we have a return type, convert it from a pointer into its actual object.\\n\\tret := gdnative.NewRealFromPointer(retPtr)\\n\\treturn ret\\n}\",\n \"func (px *Paxos) Max() int {\\n\\t// Your code here.\\n\\treturn px.max\\n}\",\n \"func MaxInt(x, y int) int {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func MAX(a, b int) int {\\n\\tif a > b {\\n\\t\\treturn a\\n\\t}\\n\\treturn b\\n}\",\n \"func MaxInt(x, y int) int {\\n\\tif x < y {\\n\\t\\treturn y\\n\\t}\\n\\treturn x\\n}\",\n \"func MaxLamport(x, y Lamport) Lamport {\\n\\tif x > y {\\n\\t\\treturn x\\n\\t}\\n\\treturn y\\n}\",\n \"func (v *Vector) Max(m *Vector) {\\n\\tif m.X > v.X {\\n\\t\\tv.X = m.X\\n\\t}\\n\\tif m.Y > v.Y {\\n\\t\\tv.Y = m.Y\\n\\t}\\n\\tif m.Z > v.Z {\\n\\t\\tv.Z = m.Z\\n\\t}\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.8431474","0.8313492","0.82121664","0.8178514","0.81000376","0.807142","0.80567896","0.80367714","0.80367714","0.8026792","0.79476416","0.79179174","0.7908209","0.7897175","0.78907746","0.78907746","0.78907746","0.78907746","0.78907746","0.78712374","0.76928234","0.76893675","0.7688665","0.7688665","0.7688665","0.76567596","0.75705755","0.7567119","0.75659263","0.741002","0.74039495","0.738016","0.73081607","0.72831273","0.72831273","0.7208672","0.7188059","0.71852016","0.71818304","0.71231407","0.71231407","0.71231407","0.71231407","0.71231407","0.70794576","0.70794576","0.7033021","0.7031511","0.7029756","0.70153874","0.69883513","0.6983458","0.6975979","0.6967625","0.69609183","0.6928662","0.69091177","0.6897687","0.68943536","0.68943536","0.68943536","0.68943536","0.68943536","0.68943536","0.68943536","0.68943536","0.68943536","0.68943536","0.68943536","0.6879462","0.6868003","0.6855652","0.6851298","0.6848685","0.6840217","0.6829919","0.6829149","0.68278414","0.68239653","0.68156725","0.6804227","0.6799002","0.6787549","0.67735755","0.67688906","0.67611164","0.67587006","0.6712741","0.6702978","0.6693654","0.66725016","0.6670033","0.6670033","0.6646335","0.66434634","0.6640711","0.6636861","0.6622298","0.66192377","0.6599641"],"string":"[\n \"0.8431474\",\n \"0.8313492\",\n \"0.82121664\",\n \"0.8178514\",\n \"0.81000376\",\n \"0.807142\",\n \"0.80567896\",\n \"0.80367714\",\n \"0.80367714\",\n \"0.8026792\",\n \"0.79476416\",\n \"0.79179174\",\n \"0.7908209\",\n \"0.7897175\",\n \"0.78907746\",\n \"0.78907746\",\n \"0.78907746\",\n \"0.78907746\",\n \"0.78907746\",\n \"0.78712374\",\n \"0.76928234\",\n \"0.76893675\",\n \"0.7688665\",\n \"0.7688665\",\n \"0.7688665\",\n \"0.76567596\",\n \"0.75705755\",\n \"0.7567119\",\n \"0.75659263\",\n \"0.741002\",\n \"0.74039495\",\n \"0.738016\",\n \"0.73081607\",\n \"0.72831273\",\n \"0.72831273\",\n \"0.7208672\",\n \"0.7188059\",\n \"0.71852016\",\n \"0.71818304\",\n \"0.71231407\",\n \"0.71231407\",\n \"0.71231407\",\n \"0.71231407\",\n \"0.71231407\",\n \"0.70794576\",\n \"0.70794576\",\n \"0.7033021\",\n \"0.7031511\",\n \"0.7029756\",\n \"0.70153874\",\n \"0.69883513\",\n \"0.6983458\",\n \"0.6975979\",\n \"0.6967625\",\n \"0.69609183\",\n \"0.6928662\",\n \"0.69091177\",\n \"0.6897687\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.68943536\",\n \"0.6879462\",\n \"0.6868003\",\n \"0.6855652\",\n \"0.6851298\",\n \"0.6848685\",\n \"0.6840217\",\n \"0.6829919\",\n \"0.6829149\",\n \"0.68278414\",\n \"0.68239653\",\n \"0.68156725\",\n \"0.6804227\",\n \"0.6799002\",\n \"0.6787549\",\n \"0.67735755\",\n \"0.67688906\",\n \"0.67611164\",\n \"0.67587006\",\n \"0.6712741\",\n \"0.6702978\",\n \"0.6693654\",\n \"0.66725016\",\n \"0.6670033\",\n \"0.6670033\",\n \"0.6646335\",\n \"0.66434634\",\n \"0.6640711\",\n \"0.6636861\",\n \"0.6622298\",\n \"0.66192377\",\n \"0.6599641\"\n]"},"document_score":{"kind":"string","value":"0.8420511"},"document_rank":{"kind":"string","value":"1"}}},{"rowIdx":104792,"cells":{"query":{"kind":"string","value":"Now() should replace usage of time.Now(). If chrono has been paused, Now() returns the time when it was paused. If there is any skew (due to forwarding or reversing), this is always added to the end time."},"document":{"kind":"string","value":"func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (u *Util) Now() time.Time {\n\tif controlDuration != 0 {\n\t\treturn time.Now().Add(controlDuration).UTC()\n\t}\n\treturn time.Now().UTC()\n}","func (c *Clock) Now() time.Time {\n\treturn time.Now().UTC().Truncate(time.Second)\n}","func (realClocker) Now() time.Time { return time.Now() }","func (c *Clock) Now() time.Time { return time.Now() }","func (c *StoppedClock) Now() time.Time {\n\treturn c.t\n}","func Now() time.Time { return now() }","func Now() time.Time { return time.Now().UTC() }","func Now() time.Time { return time.Now().UTC() }","func (c *RunningClock) Now() time.Time {\n\treturn time.Now()\n}","func (c *StoppedClock) Now() time.Time {\n\treturn c.time\n}","func (c *FakeClock) Now() time.Time {\n\tc.steps++\n\treturn c.Time(c.steps)\n}","func (c stubClocker) Now() time.Time { return c.t }","func (c *Clock) Now() time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tstep := c.step\n\tif c.skipStep {\n\t\tstep = 0\n\t\tc.skipStep = false\n\t}\n\tc.advanceLocked(rt, step)\n\n\treturn c.present\n}","func (p *PredefinedFake) Now() time.Time {\n\tadjustedTime := p.Base.Add(p.Delays[p.Next])\n\tp.Next++\n\treturn adjustedTime\n}","func Now() time.Time {\n\treturn Clock.Now()\n}","func now() time.Duration { return time.Since(startTime) }","func (clock *Clock) Now() time.Time {\n\tvar now time.Time\n\tclock.do(func(s *state) {\n\t\tnow = s.t\n\t\ts.t = s.t.Add(clock.step)\n\t})\n\treturn now\n}","func Now() time.Time { return time.Now() }","func Now() time.Time {\n\tif IsTest() {\n\t\treturn now\n\t}\n\n\treturn time.Now()\n}","func (p PT) Now() int64 {\n\tif p.Seconds {\n\t\treturn time.Now().Unix()\n\t}\n\treturn time.Now().UnixNano()\n}","func Now() time.Time {\n\treturn CurrentClock().Now()\n}","func (Clock) Now() time.Time {\n\treturn time.Now()\n}","func (t *Time) Now() time.Time {\n\treturn t.current\n}","func (fc *fakeClock) Now() time.Time {\n\treturn fc.time\n}","func (defaultClock) Now() time.Time {\n\treturn time.Now()\n}","func (c *CumulativeClock) Now() time.Time {\n\treturn c.current\n}","func (d *dummyClock) Now() time.Time {\n\treturn time.Time{}\n}","func (t *DefaultClock) Now() time.Time {\n\treturn time.Now()\n}","func Now() time.Time {\n\treturn time.Unix(0, time.Now().UnixNano()/1e6*1e6)\n}","func Now() Time {\n\treturn DefaultScheduler.Now()\n}","func Now() time.Time {\n\treturn now()\n}","func Now() time.Time {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tvar now time.Time\n\tif testMode {\n\t\tnow = testNow()\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.In(localtz.Get())\n}","func (RealClock) Now() time.Time {\n\treturn time.Now()\n}","func (t *Timer) Now() time.Duration {\n\treturn nowFunc().Sub(t.Zero)\n}","func (c *MockClock) Now() time.Time {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\treturn c.now\n}","func Now() time.Time {\n\treturn Work.Now()\n}","func (wc WallClock) Now() time.Time {\n\treturn time.Now()\n}","func (tc *TestClock) Now() time.Time {\n\ttc.l.RLock()\n\tdefer tc.l.RUnlock()\n\treturn tc.t\n}","func (a *IncrementingFakeTimeSource) Now() time.Time {\n\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\n\ta.NextIncrement++\n\n\treturn adjustedTime\n}","func now() time.Time {\n\treturn time.Now().UTC()\n}","func Now() time.Time {\n\treturn c.Now()\n}","func (c *Context) Now() time.Time {\n\treturn c.currentTime\n}","func Now(ctx Context) time.Time {\n\treturn getWorkflowEnvironment(ctx).Now()\n}","func Now() Time {\n\treturn Time(time.Now().UnixNano() / 1000)\n}","func Now() time.Time {\n\treturn time.Now().In(_defaultLocation)\n}","func (s *Scheduler) now() Time {\n\treturn s.provider.Now().Add(s.timeOffset)\n}","func nowTime() time.Time {\n\treturn time.Now().UTC()\n}","func Now() Time {\n\treturn Time{format(time.Now())}\n}","func Now() Time {\n\treturn NewTime(time.Now())\n}","func (c *webCtx) Now() time.Time {\n\treturn c.now\n}","func (f FakeTimeSource) Now() time.Time {\n\treturn f.FakeTime\n}","func now() int64 {\n\treturn time.Nanoseconds()\n}","func (c *ClockVal) Now() {\n\tc.CurrentTime = NowTime()\n}","func (m *Mock) Now() time.Time {\n\tm.Lock()\n\tdefer m.Unlock()\n\treturn m.now\n}","func Now() *time.Time {\n\tt := time.Now()\n\treturn &t\n}","func (t *TimeTravelCtx) now() time.Time {\n\tt.mutex.RLock()\n\tdefer t.mutex.RUnlock()\n\treturn t.ts\n}","func Now() Time {\n\t/*\n\t\tft := windows.Filetime{}\n\t\twindows.GetSystemTimeAsFileTime(&ft)\n\t\treturn Time(ft.Nanoseconds() / nanosecondsScale)\n\t*/\n\t// On Windows, it turns out to be much faster to simply call time.Now().Unix(), so do it\n\treturn Time(time.Now().Unix())\n}","func Now() Time {\n\tvar t Time\n\tt.FromNSec(uint64(gotime.Now().UnixNano()))\n\treturn t\n}","func TimeNow() time.Time {\n\treturn time.Now().UTC()\n}","func (m *Mock) Now() time.Time {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\treturn m.now\n}","func (m *timeSource) Now() mstime.Time {\n\treturn mstime.Now()\n}","func (f *FakeTimeSource) Now() time.Time {\n\tf.mu.RLock()\n\tdefer f.mu.RUnlock()\n\treturn f.now\n}","func GetCurrentTime() uint64 {\n\treturn ComposeTSByTime(time.Now(), 0)\n}","func (f *FixedTimeSource) Now() time.Time {\n\treturn f.fakeTime\n}","func TestClock_Now(t *testing.T) {\n\tnow := raft.NewClock().Now()\n\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\n\t\tt.Fatalf(\"clock time is different than wall time: exp=%v, got=%v\", exp, now)\n\t}\n}","func (s systemTimeSource) Now() time.Time {\n\treturn time.Now()\n}","func (s *Scheduler) now() time.Time {\n\tif s.nowTimeFn == nil {\n\t\treturn time.Now()\n\t} else {\n\t\treturn s.nowTimeFn()\n\t}\n}","func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}","func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}","func TimeNow() Time {\n\treturn Time{\n\t\ttime.Now(),\n\t}\n}","func (ns *Namespace) Now() _time.Time {\n\treturn _time.Now()\n}","func Now() time.Time {\n\treturn time.Date(int(Year.Status().(uint16)), time.Month(Month.Status().(uint8)), int(Day.Status().(uint16)), int(Hour.Status().(uint8)), int(Minute.Status().(uint8)), int(Second.Status().(uint8)), 0, time.FixedZone(\"custom\", int(Tz.Status().(int8)) * 3600))\n}","func stoppableNow() time.Time {\n\tstopMux.RLock()\n\tdefer stopMux.RUnlock()\n\n\tif stoppedAt != nil {\n\t\treturn *stoppedAt\n\t}\n\treturn time.Now().Add(-stoppedFor)\n}","func (s *Scheduler) Now() Time {\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\treturn s.now()\n}","func (t Time) GetNow() time.Time {\n\treturn now()\n}","func CurrentTime() Time {\n\tif atomic.LoadUint32(&isRunning) != 0 {\n\t\treturn currentTime\n\t}\n\n\treturn Time(time.Now().UnixNano() / 1e6)\n}","func Now() Timespec {\n\treturn NewTimespec(time.Now())\n}","func Now() DateTime {\n\treturn DateTimeFromTime(time.Now())\n}","func Now(upToSecond ...bool) *TimeStamp {\n\treturn TimeFrom(time.Now(), upToSecond...)\n}","func (em *eventManager) Now() time.Time {\n\tem.mu.Lock()\n\tdefer em.mu.Unlock()\n\treturn em.now\n}","func Now() time.Time {\n\treturn time.Now().In(LOCATION)\n}","func (rs *requestContext) Now() time.Time {\n\treturn rs.now\n}","func GetNow() time.Time {\n\treturn time.Now().UTC()\n}","func GetNow() time.Time {\n\treturn time.Now().UTC()\n}","func NowTime() time.Time {\n\treturn ExtractTimeFromDatetime(time.Now())\n}","func CurrentTime() time.Time {\n\treturn time.Now()\n}","func ExampleNow() {\n\tt := gtime.Now()\n\tfmt.Println(t)\n\n\t// May Output:\n\t// 2021-11-06 13:41:08\n}","func (tr *TextRegion) TimeNow() {\n\ttr.Time.Now()\n}","func Now() int64 {\n\treturn time.Now().Unix()\n}","func Now() int64 {\n\treturn time.Now().Unix()\n}","func (t *TimeService) Now(request *NowRequest) (*NowResponse, error) {\n\trsp := &NowResponse{}\n\treturn rsp, t.client.Call(\"time\", \"Now\", request, rsp)\n}","func GetCurrentTime() typeutil.Timestamp {\n\treturn ComposeTSByTime(time.Now(), 0)\n}","func Now(ctx context.Context) time.Time {\n\tif ts := ctx.Value(ContextKey); ts != nil {\n\t\tswitch v := ts.(type) {\n\t\tcase NowProvider:\n\t\t\treturn v()\n\t\tcase time.Time:\n\t\t\treturn v\n\t\tdefault:\n\t\t\tpanic(fmt.Sprintf(\"Unknown value for ContextKey: %v\", v))\n\t\t}\n\t}\n\treturn time.Now()\n}","func (t *TOTP) Now() string {\n\treturn t.At(currentTimestamp())\n}","func dbNow() time.Time {\n\treturn roundTime(time.Now())\n}","func Now() Date {\n\tn := time.Now()\n\treturn Of(n)\n}","func tNow() uint64 {\n\treturn uint64(time.Now().In(time.UTC).UnixNano())\n}","func (c *Cron) now() time.Time {\n\treturn time.Now().In(c.location)\n}","func makeFakeNow() func() time.Time {\n\tvar t time.Time\n\treturn func() time.Time {\n\t\tt = t.Add(time.Second)\n\t\treturn t\n\t}\n}","func defaultNow() time.Time {\n\treturn time.Now().UTC()\n}"],"string":"[\n \"func (u *Util) Now() time.Time {\\n\\tif controlDuration != 0 {\\n\\t\\treturn time.Now().Add(controlDuration).UTC()\\n\\t}\\n\\treturn time.Now().UTC()\\n}\",\n \"func (c *Clock) Now() time.Time {\\n\\treturn time.Now().UTC().Truncate(time.Second)\\n}\",\n \"func (realClocker) Now() time.Time { return time.Now() }\",\n \"func (c *Clock) Now() time.Time { return time.Now() }\",\n \"func (c *StoppedClock) Now() time.Time {\\n\\treturn c.t\\n}\",\n \"func Now() time.Time { return now() }\",\n \"func Now() time.Time { return time.Now().UTC() }\",\n \"func Now() time.Time { return time.Now().UTC() }\",\n \"func (c *RunningClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (c *StoppedClock) Now() time.Time {\\n\\treturn c.time\\n}\",\n \"func (c *FakeClock) Now() time.Time {\\n\\tc.steps++\\n\\treturn c.Time(c.steps)\\n}\",\n \"func (c stubClocker) Now() time.Time { return c.t }\",\n \"func (c *Clock) Now() time.Time {\\n\\tc.init()\\n\\trt := c.maybeGetRealTime()\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\n\\tstep := c.step\\n\\tif c.skipStep {\\n\\t\\tstep = 0\\n\\t\\tc.skipStep = false\\n\\t}\\n\\tc.advanceLocked(rt, step)\\n\\n\\treturn c.present\\n}\",\n \"func (p *PredefinedFake) Now() time.Time {\\n\\tadjustedTime := p.Base.Add(p.Delays[p.Next])\\n\\tp.Next++\\n\\treturn adjustedTime\\n}\",\n \"func Now() time.Time {\\n\\treturn Clock.Now()\\n}\",\n \"func now() time.Duration { return time.Since(startTime) }\",\n \"func (clock *Clock) Now() time.Time {\\n\\tvar now time.Time\\n\\tclock.do(func(s *state) {\\n\\t\\tnow = s.t\\n\\t\\ts.t = s.t.Add(clock.step)\\n\\t})\\n\\treturn now\\n}\",\n \"func Now() time.Time { return time.Now() }\",\n \"func Now() time.Time {\\n\\tif IsTest() {\\n\\t\\treturn now\\n\\t}\\n\\n\\treturn time.Now()\\n}\",\n \"func (p PT) Now() int64 {\\n\\tif p.Seconds {\\n\\t\\treturn time.Now().Unix()\\n\\t}\\n\\treturn time.Now().UnixNano()\\n}\",\n \"func Now() time.Time {\\n\\treturn CurrentClock().Now()\\n}\",\n \"func (Clock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (t *Time) Now() time.Time {\\n\\treturn t.current\\n}\",\n \"func (fc *fakeClock) Now() time.Time {\\n\\treturn fc.time\\n}\",\n \"func (defaultClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (c *CumulativeClock) Now() time.Time {\\n\\treturn c.current\\n}\",\n \"func (d *dummyClock) Now() time.Time {\\n\\treturn time.Time{}\\n}\",\n \"func (t *DefaultClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Unix(0, time.Now().UnixNano()/1e6*1e6)\\n}\",\n \"func Now() Time {\\n\\treturn DefaultScheduler.Now()\\n}\",\n \"func Now() time.Time {\\n\\treturn now()\\n}\",\n \"func Now() time.Time {\\n\\tmu.Lock()\\n\\tdefer mu.Unlock()\\n\\tvar now time.Time\\n\\tif testMode {\\n\\t\\tnow = testNow()\\n\\t} else {\\n\\t\\tnow = time.Now()\\n\\t}\\n\\treturn now.In(localtz.Get())\\n}\",\n \"func (RealClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (t *Timer) Now() time.Duration {\\n\\treturn nowFunc().Sub(t.Zero)\\n}\",\n \"func (c *MockClock) Now() time.Time {\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\treturn c.now\\n}\",\n \"func Now() time.Time {\\n\\treturn Work.Now()\\n}\",\n \"func (wc WallClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (tc *TestClock) Now() time.Time {\\n\\ttc.l.RLock()\\n\\tdefer tc.l.RUnlock()\\n\\treturn tc.t\\n}\",\n \"func (a *IncrementingFakeTimeSource) Now() time.Time {\\n\\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\\n\\ta.NextIncrement++\\n\\n\\treturn adjustedTime\\n}\",\n \"func now() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func Now() time.Time {\\n\\treturn c.Now()\\n}\",\n \"func (c *Context) Now() time.Time {\\n\\treturn c.currentTime\\n}\",\n \"func Now(ctx Context) time.Time {\\n\\treturn getWorkflowEnvironment(ctx).Now()\\n}\",\n \"func Now() Time {\\n\\treturn Time(time.Now().UnixNano() / 1000)\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Now().In(_defaultLocation)\\n}\",\n \"func (s *Scheduler) now() Time {\\n\\treturn s.provider.Now().Add(s.timeOffset)\\n}\",\n \"func nowTime() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func Now() Time {\\n\\treturn Time{format(time.Now())}\\n}\",\n \"func Now() Time {\\n\\treturn NewTime(time.Now())\\n}\",\n \"func (c *webCtx) Now() time.Time {\\n\\treturn c.now\\n}\",\n \"func (f FakeTimeSource) Now() time.Time {\\n\\treturn f.FakeTime\\n}\",\n \"func now() int64 {\\n\\treturn time.Nanoseconds()\\n}\",\n \"func (c *ClockVal) Now() {\\n\\tc.CurrentTime = NowTime()\\n}\",\n \"func (m *Mock) Now() time.Time {\\n\\tm.Lock()\\n\\tdefer m.Unlock()\\n\\treturn m.now\\n}\",\n \"func Now() *time.Time {\\n\\tt := time.Now()\\n\\treturn &t\\n}\",\n \"func (t *TimeTravelCtx) now() time.Time {\\n\\tt.mutex.RLock()\\n\\tdefer t.mutex.RUnlock()\\n\\treturn t.ts\\n}\",\n \"func Now() Time {\\n\\t/*\\n\\t\\tft := windows.Filetime{}\\n\\t\\twindows.GetSystemTimeAsFileTime(&ft)\\n\\t\\treturn Time(ft.Nanoseconds() / nanosecondsScale)\\n\\t*/\\n\\t// On Windows, it turns out to be much faster to simply call time.Now().Unix(), so do it\\n\\treturn Time(time.Now().Unix())\\n}\",\n \"func Now() Time {\\n\\tvar t Time\\n\\tt.FromNSec(uint64(gotime.Now().UnixNano()))\\n\\treturn t\\n}\",\n \"func TimeNow() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func (m *Mock) Now() time.Time {\\n\\tm.mu.Lock()\\n\\tdefer m.mu.Unlock()\\n\\treturn m.now\\n}\",\n \"func (m *timeSource) Now() mstime.Time {\\n\\treturn mstime.Now()\\n}\",\n \"func (f *FakeTimeSource) Now() time.Time {\\n\\tf.mu.RLock()\\n\\tdefer f.mu.RUnlock()\\n\\treturn f.now\\n}\",\n \"func GetCurrentTime() uint64 {\\n\\treturn ComposeTSByTime(time.Now(), 0)\\n}\",\n \"func (f *FixedTimeSource) Now() time.Time {\\n\\treturn f.fakeTime\\n}\",\n \"func TestClock_Now(t *testing.T) {\\n\\tnow := raft.NewClock().Now()\\n\\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\\n\\t\\tt.Fatalf(\\\"clock time is different than wall time: exp=%v, got=%v\\\", exp, now)\\n\\t}\\n}\",\n \"func (s systemTimeSource) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (s *Scheduler) now() time.Time {\\n\\tif s.nowTimeFn == nil {\\n\\t\\treturn time.Now()\\n\\t} else {\\n\\t\\treturn s.nowTimeFn()\\n\\t}\\n}\",\n \"func (s SystemTimeSource) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (s SystemTimeSource) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func TimeNow() Time {\\n\\treturn Time{\\n\\t\\ttime.Now(),\\n\\t}\\n}\",\n \"func (ns *Namespace) Now() _time.Time {\\n\\treturn _time.Now()\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Date(int(Year.Status().(uint16)), time.Month(Month.Status().(uint8)), int(Day.Status().(uint16)), int(Hour.Status().(uint8)), int(Minute.Status().(uint8)), int(Second.Status().(uint8)), 0, time.FixedZone(\\\"custom\\\", int(Tz.Status().(int8)) * 3600))\\n}\",\n \"func stoppableNow() time.Time {\\n\\tstopMux.RLock()\\n\\tdefer stopMux.RUnlock()\\n\\n\\tif stoppedAt != nil {\\n\\t\\treturn *stoppedAt\\n\\t}\\n\\treturn time.Now().Add(-stoppedFor)\\n}\",\n \"func (s *Scheduler) Now() Time {\\n\\ts.lock.RLock()\\n\\tdefer s.lock.RUnlock()\\n\\treturn s.now()\\n}\",\n \"func (t Time) GetNow() time.Time {\\n\\treturn now()\\n}\",\n \"func CurrentTime() Time {\\n\\tif atomic.LoadUint32(&isRunning) != 0 {\\n\\t\\treturn currentTime\\n\\t}\\n\\n\\treturn Time(time.Now().UnixNano() / 1e6)\\n}\",\n \"func Now() Timespec {\\n\\treturn NewTimespec(time.Now())\\n}\",\n \"func Now() DateTime {\\n\\treturn DateTimeFromTime(time.Now())\\n}\",\n \"func Now(upToSecond ...bool) *TimeStamp {\\n\\treturn TimeFrom(time.Now(), upToSecond...)\\n}\",\n \"func (em *eventManager) Now() time.Time {\\n\\tem.mu.Lock()\\n\\tdefer em.mu.Unlock()\\n\\treturn em.now\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Now().In(LOCATION)\\n}\",\n \"func (rs *requestContext) Now() time.Time {\\n\\treturn rs.now\\n}\",\n \"func GetNow() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func GetNow() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func NowTime() time.Time {\\n\\treturn ExtractTimeFromDatetime(time.Now())\\n}\",\n \"func CurrentTime() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func ExampleNow() {\\n\\tt := gtime.Now()\\n\\tfmt.Println(t)\\n\\n\\t// May Output:\\n\\t// 2021-11-06 13:41:08\\n}\",\n \"func (tr *TextRegion) TimeNow() {\\n\\ttr.Time.Now()\\n}\",\n \"func Now() int64 {\\n\\treturn time.Now().Unix()\\n}\",\n \"func Now() int64 {\\n\\treturn time.Now().Unix()\\n}\",\n \"func (t *TimeService) Now(request *NowRequest) (*NowResponse, error) {\\n\\trsp := &NowResponse{}\\n\\treturn rsp, t.client.Call(\\\"time\\\", \\\"Now\\\", request, rsp)\\n}\",\n \"func GetCurrentTime() typeutil.Timestamp {\\n\\treturn ComposeTSByTime(time.Now(), 0)\\n}\",\n \"func Now(ctx context.Context) time.Time {\\n\\tif ts := ctx.Value(ContextKey); ts != nil {\\n\\t\\tswitch v := ts.(type) {\\n\\t\\tcase NowProvider:\\n\\t\\t\\treturn v()\\n\\t\\tcase time.Time:\\n\\t\\t\\treturn v\\n\\t\\tdefault:\\n\\t\\t\\tpanic(fmt.Sprintf(\\\"Unknown value for ContextKey: %v\\\", v))\\n\\t\\t}\\n\\t}\\n\\treturn time.Now()\\n}\",\n \"func (t *TOTP) Now() string {\\n\\treturn t.At(currentTimestamp())\\n}\",\n \"func dbNow() time.Time {\\n\\treturn roundTime(time.Now())\\n}\",\n \"func Now() Date {\\n\\tn := time.Now()\\n\\treturn Of(n)\\n}\",\n \"func tNow() uint64 {\\n\\treturn uint64(time.Now().In(time.UTC).UnixNano())\\n}\",\n \"func (c *Cron) now() time.Time {\\n\\treturn time.Now().In(c.location)\\n}\",\n \"func makeFakeNow() func() time.Time {\\n\\tvar t time.Time\\n\\treturn func() time.Time {\\n\\t\\tt = t.Add(time.Second)\\n\\t\\treturn t\\n\\t}\\n}\",\n \"func defaultNow() time.Time {\\n\\treturn time.Now().UTC()\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.7707415","0.7624919","0.7593016","0.7560435","0.7556393","0.75166434","0.750841","0.750841","0.75061864","0.75025976","0.74946547","0.7469169","0.74512213","0.7446948","0.74375767","0.7418802","0.7408312","0.7363236","0.73555875","0.73452246","0.7335264","0.7324511","0.73211706","0.72900397","0.727462","0.72717696","0.7267732","0.7267188","0.7235727","0.71980613","0.7178802","0.7170116","0.71680427","0.7163473","0.7153584","0.71471274","0.7145478","0.71379447","0.71215934","0.706913","0.70586264","0.69819075","0.6975091","0.6967875","0.69553274","0.69470847","0.692535","0.6902087","0.68836117","0.68544614","0.68344927","0.6816418","0.6807244","0.6792128","0.6790532","0.67905074","0.6772249","0.67709064","0.6756","0.67420495","0.6730117","0.67076683","0.6687628","0.66865784","0.66746855","0.66665804","0.6655333","0.6634851","0.6634851","0.66326106","0.65709186","0.65596914","0.65103865","0.64965206","0.6493475","0.6489606","0.6461745","0.6433064","0.6424062","0.64114916","0.6406821","0.63955444","0.6391871","0.6391871","0.6343702","0.63428","0.6324995","0.63193285","0.62775713","0.62775713","0.62687683","0.62623066","0.6259163","0.623397","0.6224072","0.6217173","0.6191839","0.6184208","0.61661226","0.61654395"],"string":"[\n \"0.7707415\",\n \"0.7624919\",\n \"0.7593016\",\n \"0.7560435\",\n \"0.7556393\",\n \"0.75166434\",\n \"0.750841\",\n \"0.750841\",\n \"0.75061864\",\n \"0.75025976\",\n \"0.74946547\",\n \"0.7469169\",\n \"0.74512213\",\n \"0.7446948\",\n \"0.74375767\",\n \"0.7418802\",\n \"0.7408312\",\n \"0.7363236\",\n \"0.73555875\",\n \"0.73452246\",\n \"0.7335264\",\n \"0.7324511\",\n \"0.73211706\",\n \"0.72900397\",\n \"0.727462\",\n \"0.72717696\",\n \"0.7267732\",\n \"0.7267188\",\n \"0.7235727\",\n \"0.71980613\",\n \"0.7178802\",\n \"0.7170116\",\n \"0.71680427\",\n \"0.7163473\",\n \"0.7153584\",\n \"0.71471274\",\n \"0.7145478\",\n \"0.71379447\",\n \"0.71215934\",\n \"0.706913\",\n \"0.70586264\",\n \"0.69819075\",\n \"0.6975091\",\n \"0.6967875\",\n \"0.69553274\",\n \"0.69470847\",\n \"0.692535\",\n \"0.6902087\",\n \"0.68836117\",\n \"0.68544614\",\n \"0.68344927\",\n \"0.6816418\",\n \"0.6807244\",\n \"0.6792128\",\n \"0.6790532\",\n \"0.67905074\",\n \"0.6772249\",\n \"0.67709064\",\n \"0.6756\",\n \"0.67420495\",\n \"0.6730117\",\n \"0.67076683\",\n \"0.6687628\",\n \"0.66865784\",\n \"0.66746855\",\n \"0.66665804\",\n \"0.6655333\",\n \"0.6634851\",\n \"0.6634851\",\n \"0.66326106\",\n \"0.65709186\",\n \"0.65596914\",\n \"0.65103865\",\n \"0.64965206\",\n \"0.6493475\",\n \"0.6489606\",\n \"0.6461745\",\n \"0.6433064\",\n \"0.6424062\",\n \"0.64114916\",\n \"0.6406821\",\n \"0.63955444\",\n \"0.6391871\",\n \"0.6391871\",\n \"0.6343702\",\n \"0.63428\",\n \"0.6324995\",\n \"0.63193285\",\n \"0.62775713\",\n \"0.62775713\",\n \"0.62687683\",\n \"0.62623066\",\n \"0.6259163\",\n \"0.623397\",\n \"0.6224072\",\n \"0.6217173\",\n \"0.6191839\",\n \"0.6184208\",\n \"0.61661226\",\n \"0.61654395\"\n]"},"document_score":{"kind":"string","value":"0.84008926"},"document_rank":{"kind":"string","value":"0"}}},{"rowIdx":104793,"cells":{"query":{"kind":"string","value":"Forwards time of chrono with skew time. This can be used in both running and paused mode."},"document":{"kind":"string","value":"func (c *chrono) Forward(skew time.Duration) {\n\tc.skew = skew\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}","func Timeskew(cctx *testcontext.Context, name, offset string, pods ...string) (Teardown, error) {\n\ttc := chaos.TimeChaos{}\n\ttc.Name = name\n\ttc.Namespace = cctx.Namespace\n\n\ttc.Spec.Mode = chaos.AllMode\n\ttc.Spec.Selector = selectPods(pods)\n\ttc.Spec.TimeOffset = offset\n\n\tif err := cctx.Generic.Create(cctx, &tc); err != nil {\n\t\treturn nil, err\n\t}\n\treturn func(ctx context.Context) error {\n\t\treturn cctx.Generic.Delete(ctx, &tc)\n\t}, nil\n}","func (c *chrono) Reset() {\n\tc.skew = 0\n}","func (Order) Clockwise() Winding { return Clockwise }","func (c *Clock) AdvanceTo(t time.Time) {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\tc.realTime = rt\n\tc.present = t\n\tc.events.AdvanceTo(c.present)\n}","func AdvanceTime(by Duration) (newOffset Duration) {\n\treturn DefaultScheduler.AdvanceTime(by)\n}","func UpTime() func() int64 {\n\tvar startTime int64 = timeStamp()\n\treturn func() int64 {\n\t\treturn timeStamp() - startTime\n\t}\n}","func (c *MockClock) AdvanceTo(t time.Time) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tc.advanceTo(t)\n}","func (c stubClocker) Now() time.Time { return c.t }","func (c *Client) ClockSkew() time.Duration {\n\treturn c.clockSkew\n}","func (c *Clock) Now() time.Time {\n\treturn time.Now().UTC().Truncate(time.Second)\n}","func (s *Scheduler) AdvanceTime(by Duration) (newOffset Duration) {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ts.timeOffset += by\n\n\tif s.scheduled.Timer != nil {\n\t\ts.scheduled.Timer.Reset(s.scheduled.ID.Due().Sub(s.now()))\n\t}\n\treturn s.timeOffset\n}","func (tt *TimeLatencyTracking) ComputeSkew() (offset int64, latency int64, _ error) {\n\tif tt.PollerSendTimestamp == 0 || tt.PollerRecvTimestamp == 0 || tt.ServerRecvTimestamp == 0 || tt.ServerRespTimestamp == 0 {\n\t\treturn 0, 0, errors.New(\"Unable to compute with any unset timestamp\")\n\t}\n\n\t// Variable aliases for the timeline\n\tT1 := tt.PollerSendTimestamp\n\tT2 := tt.ServerRecvTimestamp\n\tT3 := tt.ServerRespTimestamp\n\tT4 := tt.PollerRecvTimestamp\n\n\toffset = ((T2 - T1) + (T3 - T4)) / 2\n\tlatency = ((T4 - T1) + (T3 - T2))\n\n\treturn\n}","func AdvanceBy(duration time.Duration) {\n\tAdvanceTo(Now().Add(duration))\n}","func (Order) CounterClockwise() Winding { return CounterClockwise }","func now() time.Duration { return time.Since(startTime) }","func monotonicTimeSince(start time.Time) (end time.Time) {\n\treturn start.Add(time.Since(start))\n}","func (p *PredefinedFake) Now() time.Time {\n\tadjustedTime := p.Base.Add(p.Delays[p.Next])\n\tp.Next++\n\treturn adjustedTime\n}","func AdvanceTo(newTime time.Time) {\n\tnow := Now()\n\tfound := false\n\tfor _, s := range sortedSchedulers() {\n\t\tnextTick := s.tickAfter(now)\n\t\tif nextTick.After(now) && !nextTick.After(newTime) {\n\t\t\tfound = true\n\t\t\tnowInTest.Store(nextTick)\n\t\t\ts.Lock()\n\t\t\tif s.paused {\n\t\t\t\ts.fireOnResume = true\n\t\t\t} else {\n\t\t\t\ts.notifyFn()\n\t\t\t}\n\t\t\ts.Unlock()\n\t\t}\n\t}\n\tif !found {\n\t\tnowInTest.Store(newTime)\n\t\treturn\n\t}\n\tif Now() != newTime {\n\t\tAdvanceTo(newTime)\n\t}\n}","func (c *FakeClock) Now() time.Time {\n\tc.steps++\n\treturn c.Time(c.steps)\n}","func (u *Util) Now() time.Time {\n\tif controlDuration != 0 {\n\t\treturn time.Now().Add(controlDuration).UTC()\n\t}\n\treturn time.Now().UTC()\n}","func sleep(delay uint) {\n\tfmt.Println(\"Time Now is:\",time.Now());\n\t//Time.After returns current time\n\tdelayTime := <-time.After(time.Second * time.Duration(delay))\n\tfmt.Println(\"Timestamp after delay:\",delay,\" by time.After is:\", delayTime);\n}","func (realClocker) Now() time.Time { return time.Now() }","func main() {\n\n\t//t1 := time.Now()\nt1 := time.Now()\n\nfmt.Printf(\"%T\\n\", t1)\n\nfmt.Println(t1.Format(time.StampMilli))\n\ntime.Sleep(33*time.Millisecond)\n\n\tt2 := time.Now()\nfmt.Println(t2.Format(time.StampMilli))\n\nduration := t2.Sub(t1)\nfmt.Println(duration.Seconds())\n\n}","func TimeSinceSmoothStep(d0 time.Duration, p0 float64, d1 time.Duration, p1 float64) TimePriority {\n\tx0 := float64(d0)\n\tx1 := float64(d1)\n\treturn func(t time.Time) float64 {\n\t\tx := float64(time.Since(t))\n\t\treturn smoothstep(x, x0, p0, x1, p1)\n\t}\n}","func (c *MockClock) Set(t time.Time) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tif t.After(c.now) {\n\t\tc.advanceTo(t)\n\t} else {\n\t\tc.now = t // move back in time\n\t}\n}","func (ms *MVCCStats) Forward(nowNanos int64) {\n\tif ms.LastUpdateNanos >= nowNanos {\n\t\treturn\n\t}\n\tms.AgeTo(nowNanos)\n}","func (c *StoppedClock) Now() time.Time {\n\treturn c.t\n}","func Until(gameTime time.Time, fromTime time.Time) uint16 {\n\tfmt.Println(gameTime)\n\treturn uint16(math.RoundToEven(gameTime.Sub(fromTime).Seconds()))\n}","func (c *Clock) Now() time.Time { return time.Now() }","func SawTooth(x, period float64) float64 {\n\tx += period / 2\n\tt := x / period\n\treturn period*(t-math.Floor(t)) - period/2\n}","func (c *MockClock) advanceTo(end time.Time) {\n\tfor {\n\t\tt := c.nextWakeupTime()\n\t\tif t == (time.Time{}) || t.After(end) {\n\t\t\tbreak\n\t\t}\n\t\tc.advanceToNextWakeup()\n\t}\n\tc.now = end\n}","func toDelay(apiTime int64) time.Duration {\n\treturn time.Now().Sub(time.Unix(apiTime, 10))\n}","func roundUp(t time.Time, d time.Duration) time.Time {\n\tr := t.Round(d)\n\tif t.After(r) {\n\t\tr = r.Add(d)\n\t}\n\treturn r\n}","func (c *StoppedClock) Now() time.Time {\n\treturn c.time\n}","func Time(then time.Time) string {\n\treturn RelTime(then, time.Now(), \"ago\", \"from now\")\n}","func (c *CumulativeClock) Now() time.Time {\n\treturn c.current\n}","func (a *IncrementingFakeTimeSource) Now() time.Time {\n\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\n\ta.NextIncrement++\n\n\treturn adjustedTime\n}","func (o BucketAutoclassOutput) ToggleTime() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketAutoclass) *string { return v.ToggleTime }).(pulumi.StringPtrOutput)\n}","func (c FakeClock) Time(step int) time.Time {\n\t// We truncate to microsecond precision because Postgres' timestamptz type\n\t// doesn't handle nanoseconds.\n\treturn c.epoch.Add(time.Duration(step) * c.step).UTC().Truncate(time.Microsecond)\n}","func stoppableNow() time.Time {\n\tstopMux.RLock()\n\tdefer stopMux.RUnlock()\n\n\tif stoppedAt != nil {\n\t\treturn *stoppedAt\n\t}\n\treturn time.Now().Add(-stoppedFor)\n}","func AnounanceTime(){\n\tfor {\n\t\tfmt.Println(time.Now())\n\t\ttime.Sleep(5*time.Second)\n\t}\n}","func Now(ctx Context) time.Time {\n\treturn getWorkflowEnvironment(ctx).Now()\n}","func (t *Time) Sleep(s time.Duration) {\n\tt.current = t.current.Add(s)\n}","func makeFakeNow() func() time.Time {\n\tvar t time.Time\n\treturn func() time.Time {\n\t\tt = t.Add(time.Second)\n\t\treturn t\n\t}\n}","func (t *TimeLine) forward(num, denom uint32, runCallbacks bool) {\n\tend := t.cursor + t.Ticks(num, denom)\n\tif runCallbacks {\n\t\tt.lastDelta = t.runCallbacks(t.cursor, end)\n\t}\n\tt.cursor = end\n}","func Since(t Time) Duration {\n\treturn Now().Sub(t)\n}","func (t *Time) Since(position Position) Distance {\n\treturn Distance(t.Current() - position)\n}","func (c *chrono) Pause() {\n\tc.pausedAt = c.Now()\n\tc.paused = true\n}","func (wc WallClock) Now() time.Time {\n\treturn time.Now()\n}","func Tomorrow() time.Time {\n\treturn time.Now().AddDate(0, 0, 1)\n}","func Now() time.Time { return time.Now().UTC() }","func Now() time.Time { return time.Now().UTC() }","func Time() time.Time {\n\tnow := time.Now().UTC()\n\tdif := time.Duration(rand.Int())\n\tstart := now.Add(dif * -1)\n\tend := now.Add(dif)\n\treturn TimeSpan(start, end)\n}","func (fc *fakeClock) Now() time.Time {\n\treturn fc.time\n}","func (c *MockClock) advanceToNextWakeup() {\n\tif len(c.wakeups) < 1 {\n\t\treturn\n\t}\n\tw := heap.Pop(&c.wakeups).(*wakeup)\n\tc.logf(\"MockClock: Advancing time from now=%s to next wakeup time %s.\",\n\t\ttsStr(c.now), tsStr(w.t))\n\tc.now = w.t\n\tselect {\n\tcase w.c <- w.t:\n\tdefault:\n\t}\n\t// give things a chance to run\n\truntime.Gosched()\n\tc.logf(\"MockClock: Advanced time, now=%s.\", tsStr(c.now))\n}","func (c *MockClock) Advance(delta time.Duration) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tend := c.now.Add(delta)\n\tc.advanceTo(end)\n}","func (t *Time) Sleep(distance Distance) {\n\ttime.Sleep(time.Duration(distance))\n}","func (o BucketAutoclassPtrOutput) ToggleTime() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketAutoclass) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ToggleTime\n\t}).(pulumi.StringPtrOutput)\n}","func Thursday(t time.Time) time.Time {\n\treturn t.AddDate(0, 0, -Weekday(t)+4)\n}","func (s *Stopwatch) Handover(s2 *Stopwatch) time.Duration {\n\ts2.t = time.Now()\n\td := s2.t.Sub(s.t)\n\ts.acc += d\n\treturn d\n}","func (c *Clock) Since(t time.Time) time.Duration {\n\treturn c.Now().Sub(t)\n}","func roundup(t time.Time, d time.Duration) time.Time {\n\trounded := t.Round(d)\n\tif rounded.Before(t) {\n\t\trounded = rounded.Add(d)\n\t}\n\treturn rounded\n}","func (s *Scheduler) now() Time {\n\treturn s.provider.Now().Add(s.timeOffset)\n}","func (ship *Ship) fwdThrustersOff() {\n\tship.fwdThrusters = false\n\tif !ship.isThrusting() {\n\t\tstopRcsSound()\n\t}\n}","func (c *MockClock) Since(t time.Time) time.Duration {\n\treturn c.Now().Sub(t)\n}","func (c *RunningClock) Now() time.Time {\n\treturn time.Now()\n}","func Until(t Time) Duration {}","func (c *chrono) Continue() {\n\tc.paused = false\n}","func (t Time) After(u Time) bool {}","func TestClock_Now(t *testing.T) {\n\tnow := raft.NewClock().Now()\n\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\n\t\tt.Fatalf(\"clock time is different than wall time: exp=%v, got=%v\", exp, now)\n\t}\n}","func loop() {\n\n\tdelta := 0 * time.Nanosecond\n\n\tlast := time.Now()\n\n\tfor true {\n\t\tcur := time.Now()\n\t\tdelta += cur.Sub(last)\n\t\tlast = cur\n\n\t\tfor delta >= 15*time.Millisecond {\n\t\t\tdelta -= time.Millisecond\n\t\t\t//fmt.Println(\"Up\")\n\t\t}\n\t\t//fmt.Println(\"Re\")\n\t}\n\n}","func FakeClockNowWithExtraTime(d time.Duration) {\n\tclock.Now = func() time.Time {\n\t\treturn originalMockedTime.Add(d)\n\t}\n}","func (clock *Clock) Now() time.Time {\n\tvar now time.Time\n\tclock.do(func(s *state) {\n\t\tnow = s.t\n\t\ts.t = s.t.Add(clock.step)\n\t})\n\treturn now\n}","func (ft *Time) Next() {\n\ta := rand.Float64()\n\n\t// Ensure first time doesn't have any variance to respect the start time parameter\n\tif ft.firstVal {\n\t\tft.firstVal = false\n\t\tft.v = ft.ts\n\n\t\tif ft.keepStats {\n\t\t\tft.Stats.Add(ft.v)\n\t\t}\n\n\t\treturn\n\t}\n\n\tft.ts = ft.ts.Add(time.Duration(ft.increment) * time.Millisecond)\n\ttmp := (float64(ft.variance) * a) - float64(int64(float64(ft.variance)*a))\n\ttmp2 := float64(-1)\n\n\tif ft.direction < 0 {\n\t\ttmp2 = float64(-1)\n\t} else if ft.direction > 0 {\n\t\ttmp2 = float64(1)\n\t} else if tmp > 0.5 {\n\t\ttmp2 = float64(1)\n\t}\n\n\tc := int64(round(float64(ft.variance)*a, 0.0000000005) * tmp2)\n\tft.v = ft.ts.Add(time.Duration(c) * time.Millisecond)\n\n\tif ft.keepStats {\n\t\tft.Stats.Add(ft.v)\n\t}\n}","func (c *ClockVal) Now() {\n\tc.CurrentTime = NowTime()\n}","func (p PT) Now() int64 {\n\tif p.Seconds {\n\t\treturn time.Now().Unix()\n\t}\n\treturn time.Now().UnixNano()\n}","func (c *Clock) Now() time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tstep := c.step\n\tif c.skipStep {\n\t\tstep = 0\n\t\tc.skipStep = false\n\t}\n\tc.advanceLocked(rt, step)\n\n\treturn c.present\n}","func OutputContinueTo(t time.Duration) {\n\tdeltaDur := t - outputToDur\n\tdeltaTz := Tz(masterFreq * float64((deltaDur)/time.Second))\n\t//debug.Printf(\"mix.OutputContinueTo(%+v) deltaDur:%+v nowTz:%+v deltaTz:%+v begin...\", t, deltaDur, nowTz, deltaTz)\n\tApiOutputNext(deltaTz)\n\toutputToDur = t\n\t//debug.Printf(\"mix.OutputContinueTo(%+v) ...done! nowTz:%+v outputToDur:%+v\", t, nowTz, outputToDur)\n}","func (o BucketAutoclassResponseOutput) ToggleTime() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketAutoclassResponse) string { return v.ToggleTime }).(pulumi.StringOutput)\n}","func Now() time.Time { return now() }","func (d *MonotonicRandomWalkDistribution) Advance() {\n\td.Step.Advance()\n\td.State += math.Abs(d.Step.Get())\n}","func TestCap(t *testing.T) {\n\tbacko := NewBacko(milliseconds(100), 2, 0, milliseconds(600))\n\n\tassert.Equal(t, milliseconds(100), backo.Duration(0))\n\tassert.Equal(t, milliseconds(200), backo.Duration(1))\n\tassert.Equal(t, milliseconds(400), backo.Duration(2))\n\tassert.Equal(t, milliseconds(600), backo.Duration(3))\n}","func (tl *TimeLord) warp() {\n\ttl.guard = monkey.Patch(time.Now, func() time.Time {\n\t\ttl.guard.Unpatch()\n\t\tdefer tl.guard.Restore()\n\n\t\treturn time.Now().Add(tl.offset)\n\t})\n}","func (ct *ClockTimestamper) Stamp(height abi.ChainEpoch) uint64 {\n\tstartTime := ct.c.StartTimeOfEpoch(height)\n\n\treturn uint64(startTime.Unix())\n}","func (fc *FakeClock) Sleep(d time.Duration) {\n\t<-fc.After(d)\n}","func (t *TimeTravelCtx) now() time.Time {\n\tt.mutex.RLock()\n\tdefer t.mutex.RUnlock()\n\treturn t.ts\n}","func sinceDate(t time.Time) string { return time.Since(t).Truncate(time.Second).String() }","func StubNow(stub func() time.Time) { now = stub }","func NewTimeOver(side Colour) Outcome { return Outcome{Winner: side.Invert(), Reason: timeOver} }","func (ship *Ship) cwThrustersOff() {\n\tship.cwThrusters = false\n\tif !ship.isThrusting() {\n\t\tstopRcsSound()\n\t}\n}","func TimeShrinker(v interface{}) gopter.Shrink {\n\tt := v.(time.Time)\n\tsec := t.Unix()\n\tnsec := int64(t.Nanosecond())\n\tsecShrink := int64Shrink{\n\t\toriginal: sec,\n\t\thalf: sec,\n\t}\n\tnsecShrink := int64Shrink{\n\t\toriginal: nsec,\n\t\thalf: nsec,\n\t}\n\treturn gopter.Shrink(secShrink.Next).Map(func(v int64) time.Time {\n\t\treturn time.Unix(v, nsec)\n\t}).Interleave(gopter.Shrink(nsecShrink.Next).Map(func(v int64) time.Time {\n\t\treturn time.Unix(sec, v)\n\t}))\n}","func StandardFor(t time.Time) times.Object {\n\treturn Standard(t.Second(), t.Minute(), t.Hour(), t.Day())\n}","func TimeFrom(t time.Time, upToSecond ...bool) *TimeStamp {\n\tts := &TimeStamp{t, t.Unix(), 0}\n\tif len(upToSecond) > 0 && upToSecond[0] == true {\n\t\tts.Time = time.Unix(ts.UnixSecond, 0).Local()\n\t\tts.UnixNanoSecond = ts.UnixSecond * 1e9\n\t} else {\n\t\tts.UnixNanoSecond = t.UnixNano()\n\t}\n\treturn ts\n}","func TestMonotonic(t *testing.T) {\n\told := Now()\n\tfor i := 0; i < 50; i++ {\n\t\tnext := Now()\n\t\tif next.After(old) {\n\t\t\tt.Error(\"Whitening insufficient\")\n\t\t}\n\t\ttime.Sleep(time.Duration(whitenerMask)/time.Nanosecond + 1)\n\t\tnext = Now()\n\t\tif !next.After(old) {\n\t\t\tt.Error(\"Not monotonically increasing on whitened nano-second scale\")\n\t\t}\n\t\told = next\n\t}\n}","func (c *Clock) Advance(d time.Duration) time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\n\tc.advanceLocked(rt, d)\n\treturn c.present\n}","func (l *limiter) floor(mark time.Time, now time.Time) time.Time {\n\tif t := now.Add(-l.quantum); !mark.After(t) {\n\t\treturn t\n\t}\n\treturn mark\n}","func (d *MonotonicUpDownRandomWalkDistribution) Advance() {\n\td.Step.Advance()\n\td.State += d.Step.Get() * float64(d.direction)\n\tif d.State < d.Min {\n\t\td.State = d.Min\n\t\td.direction = 1\n\t} else if d.State > d.Max {\n\t\td.State = d.Max\n\t\td.direction = -1\n\t}\n}","func (c *T) SkipNow()","func GetCurrentSegment() time.Time {\n\treturn time.Now().UTC().Add(-(SegmentDur / 2)).Round(SegmentDur)\n}"],"string":"[\n \"func (c *chrono) Now() time.Time {\\n\\tvar now time.Time\\n\\tif c.paused {\\n\\t\\tnow = c.pausedAt\\n\\t} else {\\n\\t\\tnow = time.Now()\\n\\t}\\n\\treturn now.Add(c.skew)\\n}\",\n \"func Timeskew(cctx *testcontext.Context, name, offset string, pods ...string) (Teardown, error) {\\n\\ttc := chaos.TimeChaos{}\\n\\ttc.Name = name\\n\\ttc.Namespace = cctx.Namespace\\n\\n\\ttc.Spec.Mode = chaos.AllMode\\n\\ttc.Spec.Selector = selectPods(pods)\\n\\ttc.Spec.TimeOffset = offset\\n\\n\\tif err := cctx.Generic.Create(cctx, &tc); err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\treturn func(ctx context.Context) error {\\n\\t\\treturn cctx.Generic.Delete(ctx, &tc)\\n\\t}, nil\\n}\",\n \"func (c *chrono) Reset() {\\n\\tc.skew = 0\\n}\",\n \"func (Order) Clockwise() Winding { return Clockwise }\",\n \"func (c *Clock) AdvanceTo(t time.Time) {\\n\\tc.init()\\n\\trt := c.maybeGetRealTime()\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\tc.skipStep = true\\n\\tc.realTime = rt\\n\\tc.present = t\\n\\tc.events.AdvanceTo(c.present)\\n}\",\n \"func AdvanceTime(by Duration) (newOffset Duration) {\\n\\treturn DefaultScheduler.AdvanceTime(by)\\n}\",\n \"func UpTime() func() int64 {\\n\\tvar startTime int64 = timeStamp()\\n\\treturn func() int64 {\\n\\t\\treturn timeStamp() - startTime\\n\\t}\\n}\",\n \"func (c *MockClock) AdvanceTo(t time.Time) {\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\tc.advanceTo(t)\\n}\",\n \"func (c stubClocker) Now() time.Time { return c.t }\",\n \"func (c *Client) ClockSkew() time.Duration {\\n\\treturn c.clockSkew\\n}\",\n \"func (c *Clock) Now() time.Time {\\n\\treturn time.Now().UTC().Truncate(time.Second)\\n}\",\n \"func (s *Scheduler) AdvanceTime(by Duration) (newOffset Duration) {\\n\\ts.lock.Lock()\\n\\tdefer s.lock.Unlock()\\n\\n\\ts.timeOffset += by\\n\\n\\tif s.scheduled.Timer != nil {\\n\\t\\ts.scheduled.Timer.Reset(s.scheduled.ID.Due().Sub(s.now()))\\n\\t}\\n\\treturn s.timeOffset\\n}\",\n \"func (tt *TimeLatencyTracking) ComputeSkew() (offset int64, latency int64, _ error) {\\n\\tif tt.PollerSendTimestamp == 0 || tt.PollerRecvTimestamp == 0 || tt.ServerRecvTimestamp == 0 || tt.ServerRespTimestamp == 0 {\\n\\t\\treturn 0, 0, errors.New(\\\"Unable to compute with any unset timestamp\\\")\\n\\t}\\n\\n\\t// Variable aliases for the timeline\\n\\tT1 := tt.PollerSendTimestamp\\n\\tT2 := tt.ServerRecvTimestamp\\n\\tT3 := tt.ServerRespTimestamp\\n\\tT4 := tt.PollerRecvTimestamp\\n\\n\\toffset = ((T2 - T1) + (T3 - T4)) / 2\\n\\tlatency = ((T4 - T1) + (T3 - T2))\\n\\n\\treturn\\n}\",\n \"func AdvanceBy(duration time.Duration) {\\n\\tAdvanceTo(Now().Add(duration))\\n}\",\n \"func (Order) CounterClockwise() Winding { return CounterClockwise }\",\n \"func now() time.Duration { return time.Since(startTime) }\",\n \"func monotonicTimeSince(start time.Time) (end time.Time) {\\n\\treturn start.Add(time.Since(start))\\n}\",\n \"func (p *PredefinedFake) Now() time.Time {\\n\\tadjustedTime := p.Base.Add(p.Delays[p.Next])\\n\\tp.Next++\\n\\treturn adjustedTime\\n}\",\n \"func AdvanceTo(newTime time.Time) {\\n\\tnow := Now()\\n\\tfound := false\\n\\tfor _, s := range sortedSchedulers() {\\n\\t\\tnextTick := s.tickAfter(now)\\n\\t\\tif nextTick.After(now) && !nextTick.After(newTime) {\\n\\t\\t\\tfound = true\\n\\t\\t\\tnowInTest.Store(nextTick)\\n\\t\\t\\ts.Lock()\\n\\t\\t\\tif s.paused {\\n\\t\\t\\t\\ts.fireOnResume = true\\n\\t\\t\\t} else {\\n\\t\\t\\t\\ts.notifyFn()\\n\\t\\t\\t}\\n\\t\\t\\ts.Unlock()\\n\\t\\t}\\n\\t}\\n\\tif !found {\\n\\t\\tnowInTest.Store(newTime)\\n\\t\\treturn\\n\\t}\\n\\tif Now() != newTime {\\n\\t\\tAdvanceTo(newTime)\\n\\t}\\n}\",\n \"func (c *FakeClock) Now() time.Time {\\n\\tc.steps++\\n\\treturn c.Time(c.steps)\\n}\",\n \"func (u *Util) Now() time.Time {\\n\\tif controlDuration != 0 {\\n\\t\\treturn time.Now().Add(controlDuration).UTC()\\n\\t}\\n\\treturn time.Now().UTC()\\n}\",\n \"func sleep(delay uint) {\\n\\tfmt.Println(\\\"Time Now is:\\\",time.Now());\\n\\t//Time.After returns current time\\n\\tdelayTime := <-time.After(time.Second * time.Duration(delay))\\n\\tfmt.Println(\\\"Timestamp after delay:\\\",delay,\\\" by time.After is:\\\", delayTime);\\n}\",\n \"func (realClocker) Now() time.Time { return time.Now() }\",\n \"func main() {\\n\\n\\t//t1 := time.Now()\\nt1 := time.Now()\\n\\nfmt.Printf(\\\"%T\\\\n\\\", t1)\\n\\nfmt.Println(t1.Format(time.StampMilli))\\n\\ntime.Sleep(33*time.Millisecond)\\n\\n\\tt2 := time.Now()\\nfmt.Println(t2.Format(time.StampMilli))\\n\\nduration := t2.Sub(t1)\\nfmt.Println(duration.Seconds())\\n\\n}\",\n \"func TimeSinceSmoothStep(d0 time.Duration, p0 float64, d1 time.Duration, p1 float64) TimePriority {\\n\\tx0 := float64(d0)\\n\\tx1 := float64(d1)\\n\\treturn func(t time.Time) float64 {\\n\\t\\tx := float64(time.Since(t))\\n\\t\\treturn smoothstep(x, x0, p0, x1, p1)\\n\\t}\\n}\",\n \"func (c *MockClock) Set(t time.Time) {\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\tif t.After(c.now) {\\n\\t\\tc.advanceTo(t)\\n\\t} else {\\n\\t\\tc.now = t // move back in time\\n\\t}\\n}\",\n \"func (ms *MVCCStats) Forward(nowNanos int64) {\\n\\tif ms.LastUpdateNanos >= nowNanos {\\n\\t\\treturn\\n\\t}\\n\\tms.AgeTo(nowNanos)\\n}\",\n \"func (c *StoppedClock) Now() time.Time {\\n\\treturn c.t\\n}\",\n \"func Until(gameTime time.Time, fromTime time.Time) uint16 {\\n\\tfmt.Println(gameTime)\\n\\treturn uint16(math.RoundToEven(gameTime.Sub(fromTime).Seconds()))\\n}\",\n \"func (c *Clock) Now() time.Time { return time.Now() }\",\n \"func SawTooth(x, period float64) float64 {\\n\\tx += period / 2\\n\\tt := x / period\\n\\treturn period*(t-math.Floor(t)) - period/2\\n}\",\n \"func (c *MockClock) advanceTo(end time.Time) {\\n\\tfor {\\n\\t\\tt := c.nextWakeupTime()\\n\\t\\tif t == (time.Time{}) || t.After(end) {\\n\\t\\t\\tbreak\\n\\t\\t}\\n\\t\\tc.advanceToNextWakeup()\\n\\t}\\n\\tc.now = end\\n}\",\n \"func toDelay(apiTime int64) time.Duration {\\n\\treturn time.Now().Sub(time.Unix(apiTime, 10))\\n}\",\n \"func roundUp(t time.Time, d time.Duration) time.Time {\\n\\tr := t.Round(d)\\n\\tif t.After(r) {\\n\\t\\tr = r.Add(d)\\n\\t}\\n\\treturn r\\n}\",\n \"func (c *StoppedClock) Now() time.Time {\\n\\treturn c.time\\n}\",\n \"func Time(then time.Time) string {\\n\\treturn RelTime(then, time.Now(), \\\"ago\\\", \\\"from now\\\")\\n}\",\n \"func (c *CumulativeClock) Now() time.Time {\\n\\treturn c.current\\n}\",\n \"func (a *IncrementingFakeTimeSource) Now() time.Time {\\n\\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\\n\\ta.NextIncrement++\\n\\n\\treturn adjustedTime\\n}\",\n \"func (o BucketAutoclassOutput) ToggleTime() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v BucketAutoclass) *string { return v.ToggleTime }).(pulumi.StringPtrOutput)\\n}\",\n \"func (c FakeClock) Time(step int) time.Time {\\n\\t// We truncate to microsecond precision because Postgres' timestamptz type\\n\\t// doesn't handle nanoseconds.\\n\\treturn c.epoch.Add(time.Duration(step) * c.step).UTC().Truncate(time.Microsecond)\\n}\",\n \"func stoppableNow() time.Time {\\n\\tstopMux.RLock()\\n\\tdefer stopMux.RUnlock()\\n\\n\\tif stoppedAt != nil {\\n\\t\\treturn *stoppedAt\\n\\t}\\n\\treturn time.Now().Add(-stoppedFor)\\n}\",\n \"func AnounanceTime(){\\n\\tfor {\\n\\t\\tfmt.Println(time.Now())\\n\\t\\ttime.Sleep(5*time.Second)\\n\\t}\\n}\",\n \"func Now(ctx Context) time.Time {\\n\\treturn getWorkflowEnvironment(ctx).Now()\\n}\",\n \"func (t *Time) Sleep(s time.Duration) {\\n\\tt.current = t.current.Add(s)\\n}\",\n \"func makeFakeNow() func() time.Time {\\n\\tvar t time.Time\\n\\treturn func() time.Time {\\n\\t\\tt = t.Add(time.Second)\\n\\t\\treturn t\\n\\t}\\n}\",\n \"func (t *TimeLine) forward(num, denom uint32, runCallbacks bool) {\\n\\tend := t.cursor + t.Ticks(num, denom)\\n\\tif runCallbacks {\\n\\t\\tt.lastDelta = t.runCallbacks(t.cursor, end)\\n\\t}\\n\\tt.cursor = end\\n}\",\n \"func Since(t Time) Duration {\\n\\treturn Now().Sub(t)\\n}\",\n \"func (t *Time) Since(position Position) Distance {\\n\\treturn Distance(t.Current() - position)\\n}\",\n \"func (c *chrono) Pause() {\\n\\tc.pausedAt = c.Now()\\n\\tc.paused = true\\n}\",\n \"func (wc WallClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func Tomorrow() time.Time {\\n\\treturn time.Now().AddDate(0, 0, 1)\\n}\",\n \"func Now() time.Time { return time.Now().UTC() }\",\n \"func Now() time.Time { return time.Now().UTC() }\",\n \"func Time() time.Time {\\n\\tnow := time.Now().UTC()\\n\\tdif := time.Duration(rand.Int())\\n\\tstart := now.Add(dif * -1)\\n\\tend := now.Add(dif)\\n\\treturn TimeSpan(start, end)\\n}\",\n \"func (fc *fakeClock) Now() time.Time {\\n\\treturn fc.time\\n}\",\n \"func (c *MockClock) advanceToNextWakeup() {\\n\\tif len(c.wakeups) < 1 {\\n\\t\\treturn\\n\\t}\\n\\tw := heap.Pop(&c.wakeups).(*wakeup)\\n\\tc.logf(\\\"MockClock: Advancing time from now=%s to next wakeup time %s.\\\",\\n\\t\\ttsStr(c.now), tsStr(w.t))\\n\\tc.now = w.t\\n\\tselect {\\n\\tcase w.c <- w.t:\\n\\tdefault:\\n\\t}\\n\\t// give things a chance to run\\n\\truntime.Gosched()\\n\\tc.logf(\\\"MockClock: Advanced time, now=%s.\\\", tsStr(c.now))\\n}\",\n \"func (c *MockClock) Advance(delta time.Duration) {\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\tend := c.now.Add(delta)\\n\\tc.advanceTo(end)\\n}\",\n \"func (t *Time) Sleep(distance Distance) {\\n\\ttime.Sleep(time.Duration(distance))\\n}\",\n \"func (o BucketAutoclassPtrOutput) ToggleTime() pulumi.StringPtrOutput {\\n\\treturn o.ApplyT(func(v *BucketAutoclass) *string {\\n\\t\\tif v == nil {\\n\\t\\t\\treturn nil\\n\\t\\t}\\n\\t\\treturn v.ToggleTime\\n\\t}).(pulumi.StringPtrOutput)\\n}\",\n \"func Thursday(t time.Time) time.Time {\\n\\treturn t.AddDate(0, 0, -Weekday(t)+4)\\n}\",\n \"func (s *Stopwatch) Handover(s2 *Stopwatch) time.Duration {\\n\\ts2.t = time.Now()\\n\\td := s2.t.Sub(s.t)\\n\\ts.acc += d\\n\\treturn d\\n}\",\n \"func (c *Clock) Since(t time.Time) time.Duration {\\n\\treturn c.Now().Sub(t)\\n}\",\n \"func roundup(t time.Time, d time.Duration) time.Time {\\n\\trounded := t.Round(d)\\n\\tif rounded.Before(t) {\\n\\t\\trounded = rounded.Add(d)\\n\\t}\\n\\treturn rounded\\n}\",\n \"func (s *Scheduler) now() Time {\\n\\treturn s.provider.Now().Add(s.timeOffset)\\n}\",\n \"func (ship *Ship) fwdThrustersOff() {\\n\\tship.fwdThrusters = false\\n\\tif !ship.isThrusting() {\\n\\t\\tstopRcsSound()\\n\\t}\\n}\",\n \"func (c *MockClock) Since(t time.Time) time.Duration {\\n\\treturn c.Now().Sub(t)\\n}\",\n \"func (c *RunningClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func Until(t Time) Duration {}\",\n \"func (c *chrono) Continue() {\\n\\tc.paused = false\\n}\",\n \"func (t Time) After(u Time) bool {}\",\n \"func TestClock_Now(t *testing.T) {\\n\\tnow := raft.NewClock().Now()\\n\\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\\n\\t\\tt.Fatalf(\\\"clock time is different than wall time: exp=%v, got=%v\\\", exp, now)\\n\\t}\\n}\",\n \"func loop() {\\n\\n\\tdelta := 0 * time.Nanosecond\\n\\n\\tlast := time.Now()\\n\\n\\tfor true {\\n\\t\\tcur := time.Now()\\n\\t\\tdelta += cur.Sub(last)\\n\\t\\tlast = cur\\n\\n\\t\\tfor delta >= 15*time.Millisecond {\\n\\t\\t\\tdelta -= time.Millisecond\\n\\t\\t\\t//fmt.Println(\\\"Up\\\")\\n\\t\\t}\\n\\t\\t//fmt.Println(\\\"Re\\\")\\n\\t}\\n\\n}\",\n \"func FakeClockNowWithExtraTime(d time.Duration) {\\n\\tclock.Now = func() time.Time {\\n\\t\\treturn originalMockedTime.Add(d)\\n\\t}\\n}\",\n \"func (clock *Clock) Now() time.Time {\\n\\tvar now time.Time\\n\\tclock.do(func(s *state) {\\n\\t\\tnow = s.t\\n\\t\\ts.t = s.t.Add(clock.step)\\n\\t})\\n\\treturn now\\n}\",\n \"func (ft *Time) Next() {\\n\\ta := rand.Float64()\\n\\n\\t// Ensure first time doesn't have any variance to respect the start time parameter\\n\\tif ft.firstVal {\\n\\t\\tft.firstVal = false\\n\\t\\tft.v = ft.ts\\n\\n\\t\\tif ft.keepStats {\\n\\t\\t\\tft.Stats.Add(ft.v)\\n\\t\\t}\\n\\n\\t\\treturn\\n\\t}\\n\\n\\tft.ts = ft.ts.Add(time.Duration(ft.increment) * time.Millisecond)\\n\\ttmp := (float64(ft.variance) * a) - float64(int64(float64(ft.variance)*a))\\n\\ttmp2 := float64(-1)\\n\\n\\tif ft.direction < 0 {\\n\\t\\ttmp2 = float64(-1)\\n\\t} else if ft.direction > 0 {\\n\\t\\ttmp2 = float64(1)\\n\\t} else if tmp > 0.5 {\\n\\t\\ttmp2 = float64(1)\\n\\t}\\n\\n\\tc := int64(round(float64(ft.variance)*a, 0.0000000005) * tmp2)\\n\\tft.v = ft.ts.Add(time.Duration(c) * time.Millisecond)\\n\\n\\tif ft.keepStats {\\n\\t\\tft.Stats.Add(ft.v)\\n\\t}\\n}\",\n \"func (c *ClockVal) Now() {\\n\\tc.CurrentTime = NowTime()\\n}\",\n \"func (p PT) Now() int64 {\\n\\tif p.Seconds {\\n\\t\\treturn time.Now().Unix()\\n\\t}\\n\\treturn time.Now().UnixNano()\\n}\",\n \"func (c *Clock) Now() time.Time {\\n\\tc.init()\\n\\trt := c.maybeGetRealTime()\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\n\\tstep := c.step\\n\\tif c.skipStep {\\n\\t\\tstep = 0\\n\\t\\tc.skipStep = false\\n\\t}\\n\\tc.advanceLocked(rt, step)\\n\\n\\treturn c.present\\n}\",\n \"func OutputContinueTo(t time.Duration) {\\n\\tdeltaDur := t - outputToDur\\n\\tdeltaTz := Tz(masterFreq * float64((deltaDur)/time.Second))\\n\\t//debug.Printf(\\\"mix.OutputContinueTo(%+v) deltaDur:%+v nowTz:%+v deltaTz:%+v begin...\\\", t, deltaDur, nowTz, deltaTz)\\n\\tApiOutputNext(deltaTz)\\n\\toutputToDur = t\\n\\t//debug.Printf(\\\"mix.OutputContinueTo(%+v) ...done! nowTz:%+v outputToDur:%+v\\\", t, nowTz, outputToDur)\\n}\",\n \"func (o BucketAutoclassResponseOutput) ToggleTime() pulumi.StringOutput {\\n\\treturn o.ApplyT(func(v BucketAutoclassResponse) string { return v.ToggleTime }).(pulumi.StringOutput)\\n}\",\n \"func Now() time.Time { return now() }\",\n \"func (d *MonotonicRandomWalkDistribution) Advance() {\\n\\td.Step.Advance()\\n\\td.State += math.Abs(d.Step.Get())\\n}\",\n \"func TestCap(t *testing.T) {\\n\\tbacko := NewBacko(milliseconds(100), 2, 0, milliseconds(600))\\n\\n\\tassert.Equal(t, milliseconds(100), backo.Duration(0))\\n\\tassert.Equal(t, milliseconds(200), backo.Duration(1))\\n\\tassert.Equal(t, milliseconds(400), backo.Duration(2))\\n\\tassert.Equal(t, milliseconds(600), backo.Duration(3))\\n}\",\n \"func (tl *TimeLord) warp() {\\n\\ttl.guard = monkey.Patch(time.Now, func() time.Time {\\n\\t\\ttl.guard.Unpatch()\\n\\t\\tdefer tl.guard.Restore()\\n\\n\\t\\treturn time.Now().Add(tl.offset)\\n\\t})\\n}\",\n \"func (ct *ClockTimestamper) Stamp(height abi.ChainEpoch) uint64 {\\n\\tstartTime := ct.c.StartTimeOfEpoch(height)\\n\\n\\treturn uint64(startTime.Unix())\\n}\",\n \"func (fc *FakeClock) Sleep(d time.Duration) {\\n\\t<-fc.After(d)\\n}\",\n \"func (t *TimeTravelCtx) now() time.Time {\\n\\tt.mutex.RLock()\\n\\tdefer t.mutex.RUnlock()\\n\\treturn t.ts\\n}\",\n \"func sinceDate(t time.Time) string { return time.Since(t).Truncate(time.Second).String() }\",\n \"func StubNow(stub func() time.Time) { now = stub }\",\n \"func NewTimeOver(side Colour) Outcome { return Outcome{Winner: side.Invert(), Reason: timeOver} }\",\n \"func (ship *Ship) cwThrustersOff() {\\n\\tship.cwThrusters = false\\n\\tif !ship.isThrusting() {\\n\\t\\tstopRcsSound()\\n\\t}\\n}\",\n \"func TimeShrinker(v interface{}) gopter.Shrink {\\n\\tt := v.(time.Time)\\n\\tsec := t.Unix()\\n\\tnsec := int64(t.Nanosecond())\\n\\tsecShrink := int64Shrink{\\n\\t\\toriginal: sec,\\n\\t\\thalf: sec,\\n\\t}\\n\\tnsecShrink := int64Shrink{\\n\\t\\toriginal: nsec,\\n\\t\\thalf: nsec,\\n\\t}\\n\\treturn gopter.Shrink(secShrink.Next).Map(func(v int64) time.Time {\\n\\t\\treturn time.Unix(v, nsec)\\n\\t}).Interleave(gopter.Shrink(nsecShrink.Next).Map(func(v int64) time.Time {\\n\\t\\treturn time.Unix(sec, v)\\n\\t}))\\n}\",\n \"func StandardFor(t time.Time) times.Object {\\n\\treturn Standard(t.Second(), t.Minute(), t.Hour(), t.Day())\\n}\",\n \"func TimeFrom(t time.Time, upToSecond ...bool) *TimeStamp {\\n\\tts := &TimeStamp{t, t.Unix(), 0}\\n\\tif len(upToSecond) > 0 && upToSecond[0] == true {\\n\\t\\tts.Time = time.Unix(ts.UnixSecond, 0).Local()\\n\\t\\tts.UnixNanoSecond = ts.UnixSecond * 1e9\\n\\t} else {\\n\\t\\tts.UnixNanoSecond = t.UnixNano()\\n\\t}\\n\\treturn ts\\n}\",\n \"func TestMonotonic(t *testing.T) {\\n\\told := Now()\\n\\tfor i := 0; i < 50; i++ {\\n\\t\\tnext := Now()\\n\\t\\tif next.After(old) {\\n\\t\\t\\tt.Error(\\\"Whitening insufficient\\\")\\n\\t\\t}\\n\\t\\ttime.Sleep(time.Duration(whitenerMask)/time.Nanosecond + 1)\\n\\t\\tnext = Now()\\n\\t\\tif !next.After(old) {\\n\\t\\t\\tt.Error(\\\"Not monotonically increasing on whitened nano-second scale\\\")\\n\\t\\t}\\n\\t\\told = next\\n\\t}\\n}\",\n \"func (c *Clock) Advance(d time.Duration) time.Time {\\n\\tc.init()\\n\\trt := c.maybeGetRealTime()\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\tc.skipStep = true\\n\\n\\tc.advanceLocked(rt, d)\\n\\treturn c.present\\n}\",\n \"func (l *limiter) floor(mark time.Time, now time.Time) time.Time {\\n\\tif t := now.Add(-l.quantum); !mark.After(t) {\\n\\t\\treturn t\\n\\t}\\n\\treturn mark\\n}\",\n \"func (d *MonotonicUpDownRandomWalkDistribution) Advance() {\\n\\td.Step.Advance()\\n\\td.State += d.Step.Get() * float64(d.direction)\\n\\tif d.State < d.Min {\\n\\t\\td.State = d.Min\\n\\t\\td.direction = 1\\n\\t} else if d.State > d.Max {\\n\\t\\td.State = d.Max\\n\\t\\td.direction = -1\\n\\t}\\n}\",\n \"func (c *T) SkipNow()\",\n \"func GetCurrentSegment() time.Time {\\n\\treturn time.Now().UTC().Add(-(SegmentDur / 2)).Round(SegmentDur)\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.6377099","0.57420695","0.55626804","0.52779603","0.5123773","0.50687426","0.5051111","0.5007728","0.49764827","0.49639702","0.49189162","0.48556083","0.4847172","0.48447493","0.48446742","0.48409605","0.4827538","0.48221955","0.47696277","0.4712712","0.46684983","0.46447614","0.4639994","0.46390235","0.46322784","0.46279216","0.46159396","0.46143475","0.45857015","0.458469","0.4575363","0.45659685","0.4560727","0.4547705","0.45358622","0.45318428","0.4515254","0.4500201","0.44983536","0.44778347","0.44768745","0.44742745","0.4463998","0.44606817","0.44397876","0.4438386","0.44350022","0.44242248","0.44209474","0.44130993","0.4408065","0.4403237","0.4403237","0.4397341","0.43957078","0.43931156","0.43916333","0.43915167","0.43913218","0.4389962","0.43754026","0.4374432","0.43721148","0.43689024","0.43624708","0.4346447","0.43423003","0.43350083","0.43214065","0.43194172","0.4317958","0.43114153","0.43071178","0.43022218","0.4280368","0.4261439","0.4258817","0.4252349","0.42464542","0.42438734","0.42428762","0.42419526","0.42364377","0.42238113","0.42218247","0.42206946","0.42198256","0.42174518","0.42142338","0.42142045","0.42106998","0.42076206","0.42072263","0.42063943","0.42023113","0.42003888","0.41998807","0.41986006","0.41964772","0.41958177"],"string":"[\n \"0.6377099\",\n \"0.57420695\",\n \"0.55626804\",\n \"0.52779603\",\n \"0.5123773\",\n \"0.50687426\",\n \"0.5051111\",\n \"0.5007728\",\n \"0.49764827\",\n \"0.49639702\",\n \"0.49189162\",\n \"0.48556083\",\n \"0.4847172\",\n \"0.48447493\",\n \"0.48446742\",\n \"0.48409605\",\n \"0.4827538\",\n \"0.48221955\",\n \"0.47696277\",\n \"0.4712712\",\n \"0.46684983\",\n \"0.46447614\",\n \"0.4639994\",\n \"0.46390235\",\n \"0.46322784\",\n \"0.46279216\",\n \"0.46159396\",\n \"0.46143475\",\n \"0.45857015\",\n \"0.458469\",\n \"0.4575363\",\n \"0.45659685\",\n \"0.4560727\",\n \"0.4547705\",\n \"0.45358622\",\n \"0.45318428\",\n \"0.4515254\",\n \"0.4500201\",\n \"0.44983536\",\n \"0.44778347\",\n \"0.44768745\",\n \"0.44742745\",\n \"0.4463998\",\n \"0.44606817\",\n \"0.44397876\",\n \"0.4438386\",\n \"0.44350022\",\n \"0.44242248\",\n \"0.44209474\",\n \"0.44130993\",\n \"0.4408065\",\n \"0.4403237\",\n \"0.4403237\",\n \"0.4397341\",\n \"0.43957078\",\n \"0.43931156\",\n \"0.43916333\",\n \"0.43915167\",\n \"0.43913218\",\n \"0.4389962\",\n \"0.43754026\",\n \"0.4374432\",\n \"0.43721148\",\n \"0.43689024\",\n \"0.43624708\",\n \"0.4346447\",\n \"0.43423003\",\n \"0.43350083\",\n \"0.43214065\",\n \"0.43194172\",\n \"0.4317958\",\n \"0.43114153\",\n \"0.43071178\",\n \"0.43022218\",\n \"0.4280368\",\n \"0.4261439\",\n \"0.4258817\",\n \"0.4252349\",\n \"0.42464542\",\n \"0.42438734\",\n \"0.42428762\",\n \"0.42419526\",\n \"0.42364377\",\n \"0.42238113\",\n \"0.42218247\",\n \"0.42206946\",\n \"0.42198256\",\n \"0.42174518\",\n \"0.42142338\",\n \"0.42142045\",\n \"0.42106998\",\n \"0.42076206\",\n \"0.42072263\",\n \"0.42063943\",\n \"0.42023113\",\n \"0.42003888\",\n \"0.41998807\",\n \"0.41986006\",\n \"0.41964772\",\n \"0.41958177\"\n]"},"document_score":{"kind":"string","value":"0.6801785"},"document_rank":{"kind":"string","value":"0"}}},{"rowIdx":104794,"cells":{"query":{"kind":"string","value":"Resets any previous set clock skew."},"document":{"kind":"string","value":"func (c *chrono) Reset() {\n\tc.skew = 0\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (sw *Stopwatch) Reset() {\n\tsw.t = time.Now()\n}","func (rw *RandW) Reset() {\n\trw.r = rand.New(rand.NewSource(uint64(time.Now().UnixNano())))\n}","func Reset() {\n\tmutex.Lock()\n\taverages = make(map[string]*currentAverage)\n\tmutex.Unlock()\n}","func (s *Stopwatch) Reset() {\n\t*s = Stopwatch{}\n}","func (d *RabinKarp64) Reset() {\n\td.tables = nil\n\td.value = 0\n\td.window = d.window[:0]\n\td.oldest = 0\n\td.updateTables()\n}","func ResetNow() { now = time.Now }","func (rd *ratedisp) reset() {\n\trd.start = time.Now()\n\trd.cnt = 0\n\trd.size = 0\n}","func (e *Timing) Reset() {\n\te.Min = 0\n\te.Max = 0\n\te.Value = 0\n\te.Values = make(float64Slice, 0)\n\te.Count = 0\n}","func resetTimeNow() {\n\ttimeNow = time.Now\n}","func (t *TimeLine) Reset() {\n\tt.cursor = 0\n\tt.lastDelta = 0\n}","func (s *Stopwatch) Reset() {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.isRunning() {\n\t\tfmt.Printf(\"WARNING: Stopwatch.Reset() isRunning is true\\n\")\n\t}\n\ts.refTime = time.Time{}\n\ts.elapsedTime = 0\n}","func (ts *TextState) Reset() {\n\tts.Tm = transform.IdentityMatrix()\n\tts.Tlm = transform.IdentityMatrix()\n}","func (c *TimeAvgAggregator) Reset(w Window) {\n\tc.integral = 0\n\tif c.initialized {\n\t\tc.startTime = w.StartTime\n\t\tc.startValue = c.endValue\n\t\tc.endTime = w.EndTime\n\t}\n}","func Timeskew(cctx *testcontext.Context, name, offset string, pods ...string) (Teardown, error) {\n\ttc := chaos.TimeChaos{}\n\ttc.Name = name\n\ttc.Namespace = cctx.Namespace\n\n\ttc.Spec.Mode = chaos.AllMode\n\ttc.Spec.Selector = selectPods(pods)\n\ttc.Spec.TimeOffset = offset\n\n\tif err := cctx.Generic.Create(cctx, &tc); err != nil {\n\t\treturn nil, err\n\t}\n\treturn func(ctx context.Context) error {\n\t\treturn cctx.Generic.Delete(ctx, &tc)\n\t}, nil\n}","func (tw *TimingsWrapper) Reset() {\n\ttw.timings.Reset()\n}","func (lp *Loadpoint) resetMeasuredPhases() {\n\tlp.Lock()\n\tlp.measuredPhases = 0\n\tlp.Unlock()\n\n\tlp.publish(phasesActive, lp.activePhases())\n}","func (tw *MultiTimingsWrapper) Reset() {\n\ttw.timings.Reset()\n}","func (s *Stopwatch) Reset(offset time.Duration, active bool) {\n\tnow := time.Now()\n\ts.Lock()\n\tdefer s.Unlock()\n\ts.start = now.Add(-offset)\n\tif active {\n\t\ts.stop = time.Time{}\n\t} else {\n\t\ts.stop = now\n\t}\n\ts.mark = 0\n\ts.laps = nil\n}","func (t *Transform) Reset() {\n\tt.access.Lock()\n\tt.parent = nil\n\tt.built = nil\n\tt.localToWorld = nil\n\tt.worldToLocal = nil\n\tt.quat = nil\n\tt.pos = lmath.Vec3Zero\n\tt.rot = lmath.Vec3Zero\n\tt.scale = lmath.Vec3One\n\tt.shear = lmath.Vec3Zero\n\tt.access.Unlock()\n}","func ResetBoard() {\n\tfor i := 0; i < (Scale * Scale); i++ {\n\t\tBoard[i].Fill = false\n\t\tBoard[i].Slogic = strconv.Itoa(i + 1)\n\t}\n}","func (p *TimePanel) Reset() {\n}","func (w *Week) reset() {\n\tw.days = nil\n}","func (mock *Serf) Reset() {\n\tlockSerfBootstrap.Lock()\n\tmock.calls.Bootstrap = nil\n\tlockSerfBootstrap.Unlock()\n\tlockSerfCluster.Lock()\n\tmock.calls.Cluster = nil\n\tlockSerfCluster.Unlock()\n\tlockSerfID.Lock()\n\tmock.calls.ID = nil\n\tlockSerfID.Unlock()\n\tlockSerfJoin.Lock()\n\tmock.calls.Join = nil\n\tlockSerfJoin.Unlock()\n\tlockSerfMember.Lock()\n\tmock.calls.Member = nil\n\tlockSerfMember.Unlock()\n\tlockSerfShutdown.Lock()\n\tmock.calls.Shutdown = nil\n\tlockSerfShutdown.Unlock()\n}","func Reset() {\n\tC.yices_reset()\n}","func (s *Spinner) Reset() {\n\ts.step = 0\n\ts.createFrames()\n}","func (c *Clock) Reset(last UUID) error {\n\tif err := last.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\t// Reset clock internals to match uuid standards.\n\tc.id = last.ID\n\tc.tickT = last.Type\n\tc.origin = last.Origin\n\n\t// reset last tick of clock.\n\tc.last = &last\n\n\treturn nil\n}","func (a *MovAvg) Reset() {\n\ta.r, a.w = 0, 0\n\tfor i := range a.sum {\n\t\ta.sum[i] = 0\n\t\ta.v[i] = 0\n\t}\n}","func reseed() {\n\trand.Seed(time.Now().UTC().UnixNano())\n}","func (c *CTR) Reset() {\n\tc.blocks = 0\n\tc.ks = nil\n}","func (s *Surface) ResetTransform() {\n\ts.Ctx.Call(\"resetTransform\")\n}","func (u *Util) ResetControlDuration() {\n\tcontrolDuration = 0\n}","func (t Time) ResetNow() time.Time {\n\tnow = defaultNow\n\treturn now()\n}","func (tr *trooper) reset() {\n\ttr.trash()\n\ttr.addCenter()\n\tfor cnt, b := range tr.bits {\n\t\tb.reset(tr.ipos[cnt])\n\t}\n\ttr.healthChanged(tr.health())\n}","func (context *context) ResetTimings() {\n\tcontext.model.ctx.Whisper_reset_timings()\n}","func reset() {\n\tterm.Sync()\n}","func (self *Weights) reset() {\n\tif len(self.Scale)>0 {\n\t\tself.Scale = self.Scale[:0]\n\t}\n}","func (wpr *Wrapper) Reset() {\n\twpr.O = 0\n\twpr.L = 0\n}","func (tr *trooper) resetEnergy() {\n\ttr.teleportEnergy = tr.temax\n\ttr.cloakEnergy = 1000\n}","func (c *Context) Reset() {\n\tc.sp = &storagePointer{0, 0, 0}\n\tfor _, storage := range c.storages {\n\t\tstorage.Disk.Format()\n\t}\n}","func (c *CumulativeClock) Reset() {\n\tc.Set(c.start)\n}","func UnfreezeClock(t *testing.T) {\n\tif t == nil {\n\t\tpanic(\"nice try\")\n\t}\n\tc = &DefaultClock{}\n}","func (sc *Scavenger) Reset() {\n\tsc.mu.Lock()\n\tsc.entries = nil\n\tsc.mu.Unlock()\n}","func (t *TriDense) Reset() {\n\t// N and Stride must be zeroed in unison.\n\tt.mat.N, t.mat.Stride = 0, 0\n\t// Defensively zero Uplo to ensure\n\t// it is set correctly later.\n\tt.mat.Uplo = 0\n\tt.mat.Data = t.mat.Data[:0]\n}","func (gta *GlobalTSOAllocator) Reset() {\n\ttsoAllocatorRole.WithLabelValues(gta.timestampOracle.dcLocation).Set(0)\n\tgta.timestampOracle.ResetTimestamp()\n}","func Unset() {\n\tmock = time.Time{}\n}","func (b *blockEnc) reset(prev *blockEnc) {\n\tb.extraLits = 0\n\tb.literals = b.literals[:0]\n\tb.size = 0\n\tb.sequences = b.sequences[:0]\n\tb.output = b.output[:0]\n\tb.last = false\n\tif prev != nil {\n\t\tb.recentOffsets = prev.prevRecentOffsets\n\t}\n\tb.dictLitEnc = nil\n}","func (c *Canvas) Reset() {\n\tfor y := 0; uint8(y) < canvasHeight; y++ {\n\t\tfor x := 0; uint8(x) < canvasWidth; x++ {\n\t\t\t(*c)[y][x] = 0\n\t\t}\n\t}\n}","func (f *factory) ResetSecretsLocation() {\n\tf.secretLocation = nil\n}","func Reset() {\n\tstats.Reset()\n}","func (e *Zero) Reset() {}","func (ssys *Ankiddie) FullReset() {\n\tssys.m.Lock()\n\tdefer ssys.m.Unlock()\n\tfor _, env := range ssys.envs {\n\t\tenv.cancel()\n\t}\n\tssys.envs = make(map[uint]*Environment)\n}","func (s *Suite) Reset() {\n\tfor _, set := range s.sets {\n\t\tset.Reset()\n\t}\n}","func (t Time) ResetTime() Time {\n\treturn Date(t.Year(), t.Month(), t.Day(), 0, 0, 0, 0, time.UTC).In(t.Location())\n}","func (c *Cipher) Reset() {\n\tfor i := range c.state {\n\t\tc.state[i] = 0\n\t}\n\tfor i := range c.buf {\n\t\tc.buf[i] = 0\n\t}\n}","func (r *randList) Reset() {\n\tr.offset = 0\n\tr.perm = rand.Perm(len(r.list))\n}","func (m *Mouse) Reset(monWidth, monHeight uint32) {\n\tm.PosX = monWidth / 2\n\tm.PosY = monHeight / 2\n\tm.LeftBtnUp()\n\tm.RightBtnUp()\n\tm.Wheel.ScrollVal = 5\n\tm.Sensitivity(1)\n\tm.WriteJSON()\n}","func (sl *StagesLatency) ResetStatistics() {\n\tsl.first = duplicateSlice(sl.last)\n\tsl.FirstCollected = sl.LastCollected\n\n\tsl.calculate()\n}","func Reset() {\n\tstopMux.Lock()\n\tstoppedAt = nil\n\tstoppedFor = 0\n\tstopMux.Unlock()\n}","func (r *Raft) ResetTimer(){\n\t//fmt.Println(\"Election TImer Reset\")\n\tif r.Id==0 {\n \tElectionTimer.Reset(time.Millisecond*10000) \t\n\t}else if r.Id==1 {\n \tElectionTimer.Reset(time.Millisecond*3000)\n }else if r.Id==2 {\n \tElectionTimer.Reset(time.Millisecond*12000)\n\t}else if r.Id==3 {\n \tElectionTimer.Reset(time.Millisecond*14000)\n }else if r.Id==4 {\n \tElectionTimer.Reset(time.Millisecond*16000)\n\t}else {\n\tElectionTimer.Reset(time.Millisecond*18000)\n\t}\n\n}","func (w *denseRankWindow) Reset(context.Context) {\n\tw.denseRank = 0\n\tw.peerRes = nil\n}","func (t *Timer) Reset() {\n\tt.goalTime = math.MaxFloat64\n\tt.startTime = time.Now()\n}","func (s *SimPDF) ResetMargins() {\n\ts.AddMargins(s.Margin)\n\ts.PDF.SetLeftMargin(s.Margin.Left)\n\ts.PDF.SetRightMargin(s.Margin.Right)\n}","func (mu *MuHash) Reset() {\n\tmu.numerator.SetToOne()\n\tmu.denominator.SetToOne()\n}","func resetStart() {\n\tstart = time.Now()\n}","func (imd *IMDraw) Reset() {\n\timd.points = imd.points[:0]\n\timd.Color = pixel.Alpha(1)\n\timd.Picture = pixel.ZV\n\timd.Intensity = 0\n\timd.Precision = 64\n\timd.EndShape = NoEndShape\n}","func (st *SlimTrie) Reset() {\n\tst.Children.Array32.Reset()\n\tst.Steps.Array32.Reset()\n\tst.Leaves.Array32.Reset()\n}","func (t *MCTS) Reset() {\n\tt.Lock()\n\tdefer t.Unlock()\n\n\tt.freelist = t.freelist[:0]\n\tt.freeables = t.freeables[:0]\n\tfor i := range t.nodes {\n\t\tt.nodes[i].move = -1\n\t\tt.nodes[i].visits = 0\n\t\tt.nodes[i].status = 0\n\t\tt.nodes[i].psa = 0\n\t\tt.nodes[i].hasChildren = false\n\t\tt.nodes[i].qsa = 0\n\t\tt.freelist = append(t.freelist, t.nodes[i].id)\n\t}\n\n\tfor i := range t.children {\n\t\tt.children[i] = t.children[i][:0]\n\t}\n\n\tt.nodes = t.nodes[:0]\n\tt.policies = nil\n\truntime.GC()\n}","func (s *sum64_128) Reset() {\n\ts.h1, s.h2, s.k1, s.k2, s.length, s.offset = 0, 0, 0, 0, 0, 0\n}","func (m *ZoneproductMutation) ResetZonestock() {\n\tm.zonestock = nil\n\tm.removedzonestock = nil\n}","func (c *canvasRenderer) ResetTransform() {\n\tif c.currentLayer == c.topLayer {\n\t\tc.currentLayer.Transform = sprec.IdentityMat4()\n\t} else {\n\t\tc.currentLayer.Transform = c.currentLayer.previous.Transform\n\t}\n}","func (a *Agent) Reset() {\n\ta.Sketch.Reset()\n\ta.Buf = nil // TODO: pool\n}","func (m *IntervalMutation) ResetTrades() {\n\tm.trades = nil\n\tm.clearedtrades = false\n\tm.removedtrades = nil\n}","func (es *ExponentialSleeper) Reset() {\n\tes.delay = es.initDelay\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}","func (d *Detector) Reset() {\n\tC.fvad_reset(d.fvad)\n}","func (g *Game) Reset() {\n\tg.prepareGame()\n\n\tg.GreedsReleased = false\n\tg.TimeElapsed = 0\n\tg.KillsCount = 0\n\tg.Over = false\n}","func (t *Timer) Reset() {\n\tt.currentTime = t.getCurrentTimeMs()\n\tt.lastTime = t.currentTime\n\tt.tick = 0\n}","func (g *Float32s) Reset() {\r\n\t// No state\r\n}","func (c *SwitchTicker) Reset() {\n\tatomic.StoreInt64(&c.failCount, 0)\n}","func (h *handlerImpl) reset(adjMatrix adjacencyMatrix) {\n\tfor from := range adjMatrix {\n\t\tfor to := range adjMatrix[from] {\n\t\t\tadjMatrix[from][to].disabled = false\n\t\t}\n\t}\n}","func (el *Election) Reset(validators *pos.Validators, frameToDecide idx.Frame) {\n\tel.validators = validators\n\tel.frameToDecide = frameToDecide\n\tel.votes = make(map[voteID]voteValue)\n\tel.decidedRoots = make(map[idx.StakerID]voteValue)\n}","func (sm3 *SM3) Reset() {\n\t// Reset digest\n\tsm3.digest[0] = 0x7380166f\n\tsm3.digest[1] = 0x4914b2b9\n\tsm3.digest[2] = 0x172442d7\n\tsm3.digest[3] = 0xda8a0600\n\tsm3.digest[4] = 0xa96f30bc\n\tsm3.digest[5] = 0x163138aa\n\tsm3.digest[6] = 0xe38dee4d\n\tsm3.digest[7] = 0xb0fb0e4e\n\n\tsm3.length = 0 // Reset numberic states\n\tsm3.unhandleMsg = []byte{}\n}","func (s *Snowball) Reset() {\n\ts.Preferred = nil\n\ts.Last = nil\n\n\ts.Counts = make(map[[blake2b.Size256]byte]uint16)\n\ts.Count = 0\n\n\ts.Done = false\n}","func (r *Robot) Reset() {\n\t*r = Robot{}\n}","func (v *mandelbrotViewer) reset() {\n\tv.redraw = true\n\tv.maxIterations = defaultIterations\n\tv.rMin = rMin\n\tv.rMax = rMax\n\tv.iMin = iMin\n\tv.iMax = iMax\n\tv.zoom = zoom\n}","func (d *state) Reset() {\n\t// Zero the permutation's state.\n\tfor i := range d.a {\n\t\td.a[i] = 0\n\t}\n\td.state = spongeAbsorbing\n\td.buf = d.storage[:0]\n}","func ResetScreen() {\n\tClearScreen()\n\tResetAttrs()\n\tCursorHome()\n}","func (g *Game) resetGame() {\n\tg.board = make([][]int, boardHeight)\n\tfor y := 0; y < boardHeight; y++ {\n\t\tg.board[y] = make([]int, boardWidth)\n\t\tfor x := 0; x < boardWidth; x++ {\n\t\t\tg.board[y][x] = 0\n\t\t}\n\t}\n\n\tg.state = gameIntro\n\tg.dots = 1\n\tg.d = 0\n\tg.x = 5\n\tg.y = 5\n\tg.direction = 3\n\tg.dotLocation = []int{0, 0}\n\tg.prevLocations = [][]int{{0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}}\n\tg.dx = []int{1, 0, 0, 0}\n\tg.dy = []int{1, 0, 0, 0}\n\tg.dotX = -500\n\tg.dotY = -500\n\n\tg.fallingTimer = time.NewTimer(time.Duration(1000000 * time.Second))\n\tg.fallingTimer.Stop()\n\n}","func (vm *BFVM) Reset() {\n\tvm.position = 0\n\n\tfor _, c := range vm.tape {\n\t\tc.Zero()\n\t}\n}","func (m *ram) Reset() (err error) {\n\tfor i := 0; i < len(m.b); i++ {\n\t\tm.b[i] = 0\n\t}\n\treturn\n}","func (rf *Raft) reset(term int) {\n\trf.term = term\n\tfor idx := range rf.votes {\n\t\trf.votes[idx] = -1\n\t}\n\trf.lastHeartBeat = time.Now()\n\trf.lastElection = time.Now()\n\trf.vote = -1\n}","func (w *Window) Reset() {\n\tw.highest = 0\n\t// this is fine because higher blocks are cleared during Check()\n\tw.blocks[0] = 0\n}","func Reset() string {\n\treturn csi(\"m\")\n}"],"string":"[\n \"func (sw *Stopwatch) Reset() {\\n\\tsw.t = time.Now()\\n}\",\n \"func (rw *RandW) Reset() {\\n\\trw.r = rand.New(rand.NewSource(uint64(time.Now().UnixNano())))\\n}\",\n \"func Reset() {\\n\\tmutex.Lock()\\n\\taverages = make(map[string]*currentAverage)\\n\\tmutex.Unlock()\\n}\",\n \"func (s *Stopwatch) Reset() {\\n\\t*s = Stopwatch{}\\n}\",\n \"func (d *RabinKarp64) Reset() {\\n\\td.tables = nil\\n\\td.value = 0\\n\\td.window = d.window[:0]\\n\\td.oldest = 0\\n\\td.updateTables()\\n}\",\n \"func ResetNow() { now = time.Now }\",\n \"func (rd *ratedisp) reset() {\\n\\trd.start = time.Now()\\n\\trd.cnt = 0\\n\\trd.size = 0\\n}\",\n \"func (e *Timing) Reset() {\\n\\te.Min = 0\\n\\te.Max = 0\\n\\te.Value = 0\\n\\te.Values = make(float64Slice, 0)\\n\\te.Count = 0\\n}\",\n \"func resetTimeNow() {\\n\\ttimeNow = time.Now\\n}\",\n \"func (t *TimeLine) Reset() {\\n\\tt.cursor = 0\\n\\tt.lastDelta = 0\\n}\",\n \"func (s *Stopwatch) Reset() {\\n\\ts.Lock()\\n\\tdefer s.Unlock()\\n\\n\\tif s.isRunning() {\\n\\t\\tfmt.Printf(\\\"WARNING: Stopwatch.Reset() isRunning is true\\\\n\\\")\\n\\t}\\n\\ts.refTime = time.Time{}\\n\\ts.elapsedTime = 0\\n}\",\n \"func (ts *TextState) Reset() {\\n\\tts.Tm = transform.IdentityMatrix()\\n\\tts.Tlm = transform.IdentityMatrix()\\n}\",\n \"func (c *TimeAvgAggregator) Reset(w Window) {\\n\\tc.integral = 0\\n\\tif c.initialized {\\n\\t\\tc.startTime = w.StartTime\\n\\t\\tc.startValue = c.endValue\\n\\t\\tc.endTime = w.EndTime\\n\\t}\\n}\",\n \"func Timeskew(cctx *testcontext.Context, name, offset string, pods ...string) (Teardown, error) {\\n\\ttc := chaos.TimeChaos{}\\n\\ttc.Name = name\\n\\ttc.Namespace = cctx.Namespace\\n\\n\\ttc.Spec.Mode = chaos.AllMode\\n\\ttc.Spec.Selector = selectPods(pods)\\n\\ttc.Spec.TimeOffset = offset\\n\\n\\tif err := cctx.Generic.Create(cctx, &tc); err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\treturn func(ctx context.Context) error {\\n\\t\\treturn cctx.Generic.Delete(ctx, &tc)\\n\\t}, nil\\n}\",\n \"func (tw *TimingsWrapper) Reset() {\\n\\ttw.timings.Reset()\\n}\",\n \"func (lp *Loadpoint) resetMeasuredPhases() {\\n\\tlp.Lock()\\n\\tlp.measuredPhases = 0\\n\\tlp.Unlock()\\n\\n\\tlp.publish(phasesActive, lp.activePhases())\\n}\",\n \"func (tw *MultiTimingsWrapper) Reset() {\\n\\ttw.timings.Reset()\\n}\",\n \"func (s *Stopwatch) Reset(offset time.Duration, active bool) {\\n\\tnow := time.Now()\\n\\ts.Lock()\\n\\tdefer s.Unlock()\\n\\ts.start = now.Add(-offset)\\n\\tif active {\\n\\t\\ts.stop = time.Time{}\\n\\t} else {\\n\\t\\ts.stop = now\\n\\t}\\n\\ts.mark = 0\\n\\ts.laps = nil\\n}\",\n \"func (t *Transform) Reset() {\\n\\tt.access.Lock()\\n\\tt.parent = nil\\n\\tt.built = nil\\n\\tt.localToWorld = nil\\n\\tt.worldToLocal = nil\\n\\tt.quat = nil\\n\\tt.pos = lmath.Vec3Zero\\n\\tt.rot = lmath.Vec3Zero\\n\\tt.scale = lmath.Vec3One\\n\\tt.shear = lmath.Vec3Zero\\n\\tt.access.Unlock()\\n}\",\n \"func ResetBoard() {\\n\\tfor i := 0; i < (Scale * Scale); i++ {\\n\\t\\tBoard[i].Fill = false\\n\\t\\tBoard[i].Slogic = strconv.Itoa(i + 1)\\n\\t}\\n}\",\n \"func (p *TimePanel) Reset() {\\n}\",\n \"func (w *Week) reset() {\\n\\tw.days = nil\\n}\",\n \"func (mock *Serf) Reset() {\\n\\tlockSerfBootstrap.Lock()\\n\\tmock.calls.Bootstrap = nil\\n\\tlockSerfBootstrap.Unlock()\\n\\tlockSerfCluster.Lock()\\n\\tmock.calls.Cluster = nil\\n\\tlockSerfCluster.Unlock()\\n\\tlockSerfID.Lock()\\n\\tmock.calls.ID = nil\\n\\tlockSerfID.Unlock()\\n\\tlockSerfJoin.Lock()\\n\\tmock.calls.Join = nil\\n\\tlockSerfJoin.Unlock()\\n\\tlockSerfMember.Lock()\\n\\tmock.calls.Member = nil\\n\\tlockSerfMember.Unlock()\\n\\tlockSerfShutdown.Lock()\\n\\tmock.calls.Shutdown = nil\\n\\tlockSerfShutdown.Unlock()\\n}\",\n \"func Reset() {\\n\\tC.yices_reset()\\n}\",\n \"func (s *Spinner) Reset() {\\n\\ts.step = 0\\n\\ts.createFrames()\\n}\",\n \"func (c *Clock) Reset(last UUID) error {\\n\\tif err := last.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\n\\t// Reset clock internals to match uuid standards.\\n\\tc.id = last.ID\\n\\tc.tickT = last.Type\\n\\tc.origin = last.Origin\\n\\n\\t// reset last tick of clock.\\n\\tc.last = &last\\n\\n\\treturn nil\\n}\",\n \"func (a *MovAvg) Reset() {\\n\\ta.r, a.w = 0, 0\\n\\tfor i := range a.sum {\\n\\t\\ta.sum[i] = 0\\n\\t\\ta.v[i] = 0\\n\\t}\\n}\",\n \"func reseed() {\\n\\trand.Seed(time.Now().UTC().UnixNano())\\n}\",\n \"func (c *CTR) Reset() {\\n\\tc.blocks = 0\\n\\tc.ks = nil\\n}\",\n \"func (s *Surface) ResetTransform() {\\n\\ts.Ctx.Call(\\\"resetTransform\\\")\\n}\",\n \"func (u *Util) ResetControlDuration() {\\n\\tcontrolDuration = 0\\n}\",\n \"func (t Time) ResetNow() time.Time {\\n\\tnow = defaultNow\\n\\treturn now()\\n}\",\n \"func (tr *trooper) reset() {\\n\\ttr.trash()\\n\\ttr.addCenter()\\n\\tfor cnt, b := range tr.bits {\\n\\t\\tb.reset(tr.ipos[cnt])\\n\\t}\\n\\ttr.healthChanged(tr.health())\\n}\",\n \"func (context *context) ResetTimings() {\\n\\tcontext.model.ctx.Whisper_reset_timings()\\n}\",\n \"func reset() {\\n\\tterm.Sync()\\n}\",\n \"func (self *Weights) reset() {\\n\\tif len(self.Scale)>0 {\\n\\t\\tself.Scale = self.Scale[:0]\\n\\t}\\n}\",\n \"func (wpr *Wrapper) Reset() {\\n\\twpr.O = 0\\n\\twpr.L = 0\\n}\",\n \"func (tr *trooper) resetEnergy() {\\n\\ttr.teleportEnergy = tr.temax\\n\\ttr.cloakEnergy = 1000\\n}\",\n \"func (c *Context) Reset() {\\n\\tc.sp = &storagePointer{0, 0, 0}\\n\\tfor _, storage := range c.storages {\\n\\t\\tstorage.Disk.Format()\\n\\t}\\n}\",\n \"func (c *CumulativeClock) Reset() {\\n\\tc.Set(c.start)\\n}\",\n \"func UnfreezeClock(t *testing.T) {\\n\\tif t == nil {\\n\\t\\tpanic(\\\"nice try\\\")\\n\\t}\\n\\tc = &DefaultClock{}\\n}\",\n \"func (sc *Scavenger) Reset() {\\n\\tsc.mu.Lock()\\n\\tsc.entries = nil\\n\\tsc.mu.Unlock()\\n}\",\n \"func (t *TriDense) Reset() {\\n\\t// N and Stride must be zeroed in unison.\\n\\tt.mat.N, t.mat.Stride = 0, 0\\n\\t// Defensively zero Uplo to ensure\\n\\t// it is set correctly later.\\n\\tt.mat.Uplo = 0\\n\\tt.mat.Data = t.mat.Data[:0]\\n}\",\n \"func (gta *GlobalTSOAllocator) Reset() {\\n\\ttsoAllocatorRole.WithLabelValues(gta.timestampOracle.dcLocation).Set(0)\\n\\tgta.timestampOracle.ResetTimestamp()\\n}\",\n \"func Unset() {\\n\\tmock = time.Time{}\\n}\",\n \"func (b *blockEnc) reset(prev *blockEnc) {\\n\\tb.extraLits = 0\\n\\tb.literals = b.literals[:0]\\n\\tb.size = 0\\n\\tb.sequences = b.sequences[:0]\\n\\tb.output = b.output[:0]\\n\\tb.last = false\\n\\tif prev != nil {\\n\\t\\tb.recentOffsets = prev.prevRecentOffsets\\n\\t}\\n\\tb.dictLitEnc = nil\\n}\",\n \"func (c *Canvas) Reset() {\\n\\tfor y := 0; uint8(y) < canvasHeight; y++ {\\n\\t\\tfor x := 0; uint8(x) < canvasWidth; x++ {\\n\\t\\t\\t(*c)[y][x] = 0\\n\\t\\t}\\n\\t}\\n}\",\n \"func (f *factory) ResetSecretsLocation() {\\n\\tf.secretLocation = nil\\n}\",\n \"func Reset() {\\n\\tstats.Reset()\\n}\",\n \"func (e *Zero) Reset() {}\",\n \"func (ssys *Ankiddie) FullReset() {\\n\\tssys.m.Lock()\\n\\tdefer ssys.m.Unlock()\\n\\tfor _, env := range ssys.envs {\\n\\t\\tenv.cancel()\\n\\t}\\n\\tssys.envs = make(map[uint]*Environment)\\n}\",\n \"func (s *Suite) Reset() {\\n\\tfor _, set := range s.sets {\\n\\t\\tset.Reset()\\n\\t}\\n}\",\n \"func (t Time) ResetTime() Time {\\n\\treturn Date(t.Year(), t.Month(), t.Day(), 0, 0, 0, 0, time.UTC).In(t.Location())\\n}\",\n \"func (c *Cipher) Reset() {\\n\\tfor i := range c.state {\\n\\t\\tc.state[i] = 0\\n\\t}\\n\\tfor i := range c.buf {\\n\\t\\tc.buf[i] = 0\\n\\t}\\n}\",\n \"func (r *randList) Reset() {\\n\\tr.offset = 0\\n\\tr.perm = rand.Perm(len(r.list))\\n}\",\n \"func (m *Mouse) Reset(monWidth, monHeight uint32) {\\n\\tm.PosX = monWidth / 2\\n\\tm.PosY = monHeight / 2\\n\\tm.LeftBtnUp()\\n\\tm.RightBtnUp()\\n\\tm.Wheel.ScrollVal = 5\\n\\tm.Sensitivity(1)\\n\\tm.WriteJSON()\\n}\",\n \"func (sl *StagesLatency) ResetStatistics() {\\n\\tsl.first = duplicateSlice(sl.last)\\n\\tsl.FirstCollected = sl.LastCollected\\n\\n\\tsl.calculate()\\n}\",\n \"func Reset() {\\n\\tstopMux.Lock()\\n\\tstoppedAt = nil\\n\\tstoppedFor = 0\\n\\tstopMux.Unlock()\\n}\",\n \"func (r *Raft) ResetTimer(){\\n\\t//fmt.Println(\\\"Election TImer Reset\\\")\\n\\tif r.Id==0 {\\n \\tElectionTimer.Reset(time.Millisecond*10000) \\t\\n\\t}else if r.Id==1 {\\n \\tElectionTimer.Reset(time.Millisecond*3000)\\n }else if r.Id==2 {\\n \\tElectionTimer.Reset(time.Millisecond*12000)\\n\\t}else if r.Id==3 {\\n \\tElectionTimer.Reset(time.Millisecond*14000)\\n }else if r.Id==4 {\\n \\tElectionTimer.Reset(time.Millisecond*16000)\\n\\t}else {\\n\\tElectionTimer.Reset(time.Millisecond*18000)\\n\\t}\\n\\n}\",\n \"func (w *denseRankWindow) Reset(context.Context) {\\n\\tw.denseRank = 0\\n\\tw.peerRes = nil\\n}\",\n \"func (t *Timer) Reset() {\\n\\tt.goalTime = math.MaxFloat64\\n\\tt.startTime = time.Now()\\n}\",\n \"func (s *SimPDF) ResetMargins() {\\n\\ts.AddMargins(s.Margin)\\n\\ts.PDF.SetLeftMargin(s.Margin.Left)\\n\\ts.PDF.SetRightMargin(s.Margin.Right)\\n}\",\n \"func (mu *MuHash) Reset() {\\n\\tmu.numerator.SetToOne()\\n\\tmu.denominator.SetToOne()\\n}\",\n \"func resetStart() {\\n\\tstart = time.Now()\\n}\",\n \"func (imd *IMDraw) Reset() {\\n\\timd.points = imd.points[:0]\\n\\timd.Color = pixel.Alpha(1)\\n\\timd.Picture = pixel.ZV\\n\\timd.Intensity = 0\\n\\timd.Precision = 64\\n\\timd.EndShape = NoEndShape\\n}\",\n \"func (st *SlimTrie) Reset() {\\n\\tst.Children.Array32.Reset()\\n\\tst.Steps.Array32.Reset()\\n\\tst.Leaves.Array32.Reset()\\n}\",\n \"func (t *MCTS) Reset() {\\n\\tt.Lock()\\n\\tdefer t.Unlock()\\n\\n\\tt.freelist = t.freelist[:0]\\n\\tt.freeables = t.freeables[:0]\\n\\tfor i := range t.nodes {\\n\\t\\tt.nodes[i].move = -1\\n\\t\\tt.nodes[i].visits = 0\\n\\t\\tt.nodes[i].status = 0\\n\\t\\tt.nodes[i].psa = 0\\n\\t\\tt.nodes[i].hasChildren = false\\n\\t\\tt.nodes[i].qsa = 0\\n\\t\\tt.freelist = append(t.freelist, t.nodes[i].id)\\n\\t}\\n\\n\\tfor i := range t.children {\\n\\t\\tt.children[i] = t.children[i][:0]\\n\\t}\\n\\n\\tt.nodes = t.nodes[:0]\\n\\tt.policies = nil\\n\\truntime.GC()\\n}\",\n \"func (s *sum64_128) Reset() {\\n\\ts.h1, s.h2, s.k1, s.k2, s.length, s.offset = 0, 0, 0, 0, 0, 0\\n}\",\n \"func (m *ZoneproductMutation) ResetZonestock() {\\n\\tm.zonestock = nil\\n\\tm.removedzonestock = nil\\n}\",\n \"func (c *canvasRenderer) ResetTransform() {\\n\\tif c.currentLayer == c.topLayer {\\n\\t\\tc.currentLayer.Transform = sprec.IdentityMat4()\\n\\t} else {\\n\\t\\tc.currentLayer.Transform = c.currentLayer.previous.Transform\\n\\t}\\n}\",\n \"func (a *Agent) Reset() {\\n\\ta.Sketch.Reset()\\n\\ta.Buf = nil // TODO: pool\\n}\",\n \"func (m *IntervalMutation) ResetTrades() {\\n\\tm.trades = nil\\n\\tm.clearedtrades = false\\n\\tm.removedtrades = nil\\n}\",\n \"func (es *ExponentialSleeper) Reset() {\\n\\tes.delay = es.initDelay\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\\n\\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\\n\\tfor _, op := range options {\\n\\t\\top(mb)\\n\\t}\\n}\",\n \"func (d *Detector) Reset() {\\n\\tC.fvad_reset(d.fvad)\\n}\",\n \"func (g *Game) Reset() {\\n\\tg.prepareGame()\\n\\n\\tg.GreedsReleased = false\\n\\tg.TimeElapsed = 0\\n\\tg.KillsCount = 0\\n\\tg.Over = false\\n}\",\n \"func (t *Timer) Reset() {\\n\\tt.currentTime = t.getCurrentTimeMs()\\n\\tt.lastTime = t.currentTime\\n\\tt.tick = 0\\n}\",\n \"func (g *Float32s) Reset() {\\r\\n\\t// No state\\r\\n}\",\n \"func (c *SwitchTicker) Reset() {\\n\\tatomic.StoreInt64(&c.failCount, 0)\\n}\",\n \"func (h *handlerImpl) reset(adjMatrix adjacencyMatrix) {\\n\\tfor from := range adjMatrix {\\n\\t\\tfor to := range adjMatrix[from] {\\n\\t\\t\\tadjMatrix[from][to].disabled = false\\n\\t\\t}\\n\\t}\\n}\",\n \"func (el *Election) Reset(validators *pos.Validators, frameToDecide idx.Frame) {\\n\\tel.validators = validators\\n\\tel.frameToDecide = frameToDecide\\n\\tel.votes = make(map[voteID]voteValue)\\n\\tel.decidedRoots = make(map[idx.StakerID]voteValue)\\n}\",\n \"func (sm3 *SM3) Reset() {\\n\\t// Reset digest\\n\\tsm3.digest[0] = 0x7380166f\\n\\tsm3.digest[1] = 0x4914b2b9\\n\\tsm3.digest[2] = 0x172442d7\\n\\tsm3.digest[3] = 0xda8a0600\\n\\tsm3.digest[4] = 0xa96f30bc\\n\\tsm3.digest[5] = 0x163138aa\\n\\tsm3.digest[6] = 0xe38dee4d\\n\\tsm3.digest[7] = 0xb0fb0e4e\\n\\n\\tsm3.length = 0 // Reset numberic states\\n\\tsm3.unhandleMsg = []byte{}\\n}\",\n \"func (s *Snowball) Reset() {\\n\\ts.Preferred = nil\\n\\ts.Last = nil\\n\\n\\ts.Counts = make(map[[blake2b.Size256]byte]uint16)\\n\\ts.Count = 0\\n\\n\\ts.Done = false\\n}\",\n \"func (r *Robot) Reset() {\\n\\t*r = Robot{}\\n}\",\n \"func (v *mandelbrotViewer) reset() {\\n\\tv.redraw = true\\n\\tv.maxIterations = defaultIterations\\n\\tv.rMin = rMin\\n\\tv.rMax = rMax\\n\\tv.iMin = iMin\\n\\tv.iMax = iMax\\n\\tv.zoom = zoom\\n}\",\n \"func (d *state) Reset() {\\n\\t// Zero the permutation's state.\\n\\tfor i := range d.a {\\n\\t\\td.a[i] = 0\\n\\t}\\n\\td.state = spongeAbsorbing\\n\\td.buf = d.storage[:0]\\n}\",\n \"func ResetScreen() {\\n\\tClearScreen()\\n\\tResetAttrs()\\n\\tCursorHome()\\n}\",\n \"func (g *Game) resetGame() {\\n\\tg.board = make([][]int, boardHeight)\\n\\tfor y := 0; y < boardHeight; y++ {\\n\\t\\tg.board[y] = make([]int, boardWidth)\\n\\t\\tfor x := 0; x < boardWidth; x++ {\\n\\t\\t\\tg.board[y][x] = 0\\n\\t\\t}\\n\\t}\\n\\n\\tg.state = gameIntro\\n\\tg.dots = 1\\n\\tg.d = 0\\n\\tg.x = 5\\n\\tg.y = 5\\n\\tg.direction = 3\\n\\tg.dotLocation = []int{0, 0}\\n\\tg.prevLocations = [][]int{{0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}}\\n\\tg.dx = []int{1, 0, 0, 0}\\n\\tg.dy = []int{1, 0, 0, 0}\\n\\tg.dotX = -500\\n\\tg.dotY = -500\\n\\n\\tg.fallingTimer = time.NewTimer(time.Duration(1000000 * time.Second))\\n\\tg.fallingTimer.Stop()\\n\\n}\",\n \"func (vm *BFVM) Reset() {\\n\\tvm.position = 0\\n\\n\\tfor _, c := range vm.tape {\\n\\t\\tc.Zero()\\n\\t}\\n}\",\n \"func (m *ram) Reset() (err error) {\\n\\tfor i := 0; i < len(m.b); i++ {\\n\\t\\tm.b[i] = 0\\n\\t}\\n\\treturn\\n}\",\n \"func (rf *Raft) reset(term int) {\\n\\trf.term = term\\n\\tfor idx := range rf.votes {\\n\\t\\trf.votes[idx] = -1\\n\\t}\\n\\trf.lastHeartBeat = time.Now()\\n\\trf.lastElection = time.Now()\\n\\trf.vote = -1\\n}\",\n \"func (w *Window) Reset() {\\n\\tw.highest = 0\\n\\t// this is fine because higher blocks are cleared during Check()\\n\\tw.blocks[0] = 0\\n}\",\n \"func Reset() string {\\n\\treturn csi(\\\"m\\\")\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.61890984","0.60074925","0.5998274","0.5847359","0.5753193","0.5735723","0.5669589","0.56392","0.5606564","0.5587392","0.5555873","0.55551696","0.55442333","0.550156","0.54839456","0.54712147","0.5442829","0.5439494","0.5414043","0.5409556","0.5399395","0.5383782","0.5365399","0.53583115","0.5349352","0.5347378","0.53031236","0.52984345","0.5296025","0.52884495","0.5281995","0.52719116","0.525522","0.5254747","0.5231971","0.521111","0.5178473","0.51774335","0.516984","0.51593214","0.5156711","0.51559037","0.51389563","0.5127372","0.51192427","0.5111568","0.5101283","0.5090413","0.50880367","0.5079992","0.5079832","0.507684","0.50753546","0.5073108","0.50718445","0.5061409","0.5044037","0.5034725","0.5030166","0.5022912","0.5019557","0.5017229","0.5016622","0.5016191","0.50030184","0.49928075","0.4990893","0.4980057","0.4979305","0.49771127","0.49754313","0.49633843","0.4963343","0.49567586","0.49567586","0.49567586","0.49567586","0.49567586","0.49567586","0.49567586","0.49567586","0.49556893","0.49528968","0.4952801","0.49465436","0.4938712","0.49375966","0.49341542","0.49333543","0.492955","0.49293807","0.49221754","0.4922105","0.492001","0.49161577","0.4913444","0.49099994","0.49099702","0.49066544","0.48975143"],"string":"[\n \"0.61890984\",\n \"0.60074925\",\n \"0.5998274\",\n \"0.5847359\",\n \"0.5753193\",\n \"0.5735723\",\n \"0.5669589\",\n \"0.56392\",\n \"0.5606564\",\n \"0.5587392\",\n \"0.5555873\",\n \"0.55551696\",\n \"0.55442333\",\n \"0.550156\",\n \"0.54839456\",\n \"0.54712147\",\n \"0.5442829\",\n \"0.5439494\",\n \"0.5414043\",\n \"0.5409556\",\n \"0.5399395\",\n \"0.5383782\",\n \"0.5365399\",\n \"0.53583115\",\n \"0.5349352\",\n \"0.5347378\",\n \"0.53031236\",\n \"0.52984345\",\n \"0.5296025\",\n \"0.52884495\",\n \"0.5281995\",\n \"0.52719116\",\n \"0.525522\",\n \"0.5254747\",\n \"0.5231971\",\n \"0.521111\",\n \"0.5178473\",\n \"0.51774335\",\n \"0.516984\",\n \"0.51593214\",\n \"0.5156711\",\n \"0.51559037\",\n \"0.51389563\",\n \"0.5127372\",\n \"0.51192427\",\n \"0.5111568\",\n \"0.5101283\",\n \"0.5090413\",\n \"0.50880367\",\n \"0.5079992\",\n \"0.5079832\",\n \"0.507684\",\n \"0.50753546\",\n \"0.5073108\",\n \"0.50718445\",\n \"0.5061409\",\n \"0.5044037\",\n \"0.5034725\",\n \"0.5030166\",\n \"0.5022912\",\n \"0.5019557\",\n \"0.5017229\",\n \"0.5016622\",\n \"0.5016191\",\n \"0.50030184\",\n \"0.49928075\",\n \"0.4990893\",\n \"0.4980057\",\n \"0.4979305\",\n \"0.49771127\",\n \"0.49754313\",\n \"0.49633843\",\n \"0.4963343\",\n \"0.49567586\",\n \"0.49567586\",\n \"0.49567586\",\n \"0.49567586\",\n \"0.49567586\",\n \"0.49567586\",\n \"0.49567586\",\n \"0.49567586\",\n \"0.49556893\",\n \"0.49528968\",\n \"0.4952801\",\n \"0.49465436\",\n \"0.4938712\",\n \"0.49375966\",\n \"0.49341542\",\n \"0.49333543\",\n \"0.492955\",\n \"0.49293807\",\n \"0.49221754\",\n \"0.4922105\",\n \"0.492001\",\n \"0.49161577\",\n \"0.4913444\",\n \"0.49099994\",\n \"0.49099702\",\n \"0.49066544\",\n \"0.48975143\"\n]"},"document_score":{"kind":"string","value":"0.7780016"},"document_rank":{"kind":"string","value":"0"}}},{"rowIdx":104795,"cells":{"query":{"kind":"string","value":"Pause \"Stops\" time by recording current time and shortcircuit Now() to return this time instead of the actual time (plus skew)."},"document":{"kind":"string","value":"func (c *chrono) Pause() {\n\tc.pausedAt = c.Now()\n\tc.paused = true\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}","func stoppableNow() time.Time {\n\tstopMux.RLock()\n\tdefer stopMux.RUnlock()\n\n\tif stoppedAt != nil {\n\t\treturn *stoppedAt\n\t}\n\treturn time.Now().Add(-stoppedFor)\n}","func (c *StoppedClock) Now() time.Time {\n\treturn c.t\n}","func (c *StoppedClock) Now() time.Time {\n\treturn c.time\n}","func (c *Clock) Now() time.Time {\n\treturn time.Now().UTC().Truncate(time.Second)\n}","func Stop() {\n\tstopMux.Lock()\n\tif stoppedAt != nil {\n\t\tpanic(\"Time was already stopped\")\n\t}\n\tnow := Now()\n\tstoppedAt = &now\n\tstopMux.Unlock()\n}","func (c *Clock) Now() time.Time { return time.Now() }","func (u *Util) Now() time.Time {\n\tif controlDuration != 0 {\n\t\treturn time.Now().Add(controlDuration).UTC()\n\t}\n\treturn time.Now().UTC()\n}","func (c stubClocker) Now() time.Time { return c.t }","func (realClocker) Now() time.Time { return time.Now() }","func (t *Time) Now() time.Time {\n\treturn t.current\n}","func (c *ClockVal) Now() {\n\tc.CurrentTime = NowTime()\n}","func Now() time.Time { return now() }","func (c *FakeClock) Now() time.Time {\n\tc.steps++\n\treturn c.Time(c.steps)\n}","func (c *RunningClock) Now() time.Time {\n\treturn time.Now()\n}","func (d *dummyClock) Now() time.Time {\n\treturn time.Time{}\n}","func Now() time.Time { return time.Now().UTC() }","func Now() time.Time { return time.Now().UTC() }","func now() time.Duration { return time.Since(startTime) }","func (clock *Clock) Now() time.Time {\n\tvar now time.Time\n\tclock.do(func(s *state) {\n\t\tnow = s.t\n\t\ts.t = s.t.Add(clock.step)\n\t})\n\treturn now\n}","func (p *PredefinedFake) Now() time.Time {\n\tadjustedTime := p.Base.Add(p.Delays[p.Next])\n\tp.Next++\n\treturn adjustedTime\n}","func Now() time.Time { return time.Now() }","func Now() time.Time {\n\treturn CurrentClock().Now()\n}","func (Clock) Now() time.Time {\n\treturn time.Now()\n}","func (c *Clock) Now() time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tstep := c.step\n\tif c.skipStep {\n\t\tstep = 0\n\t\tc.skipStep = false\n\t}\n\tc.advanceLocked(rt, step)\n\n\treturn c.present\n}","func (p PT) Now() int64 {\n\tif p.Seconds {\n\t\treturn time.Now().Unix()\n\t}\n\treturn time.Now().UnixNano()\n}","func (s *Stopwatch) Stop() {\n\ts.Lock()\n\tdefer s.Unlock()\n\tif s.active() {\n\t\ts.stop = time.Now()\n\t}\n}","func (t *DefaultClock) Now() time.Time {\n\treturn time.Now()\n}","func (defaultClock) Now() time.Time {\n\treturn time.Now()\n}","func Now() time.Time {\n\treturn Clock.Now()\n}","func (fc *fakeClock) Now() time.Time {\n\treturn fc.time\n}","func (tc *TestClock) Now() time.Time {\n\ttc.l.RLock()\n\tdefer tc.l.RUnlock()\n\treturn tc.t\n}","func Now() Time {\n\treturn DefaultScheduler.Now()\n}","func (c *CumulativeClock) Now() time.Time {\n\treturn c.current\n}","func Now() Time {\n\treturn Time{format(time.Now())}\n}","func Now(upToSecond ...bool) *TimeStamp {\n\treturn TimeFrom(time.Now(), upToSecond...)\n}","func InstrumentNow() {\n\tNow = stoppableNow\n}","func (c *MockClock) Now() time.Time {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\treturn c.now\n}","func Now() *time.Time {\n\tt := time.Now()\n\treturn &t\n}","func (t *Timer) Now() time.Duration {\n\treturn nowFunc().Sub(t.Zero)\n}","func Now() Time {\n\treturn NewTime(time.Now())\n}","func Now() time.Time {\n\treturn Work.Now()\n}","func Now() time.Time {\n\treturn now()\n}","func (tr *TextRegion) TimeNow() {\n\ttr.Time.Now()\n}","func CurrentTime() Time {\n\tif atomic.LoadUint32(&isRunning) != 0 {\n\t\treturn currentTime\n\t}\n\n\treturn Time(time.Now().UnixNano() / 1e6)\n}","func Now() time.Time {\n\tif IsTest() {\n\t\treturn now\n\t}\n\n\treturn time.Now()\n}","func Now() time.Time {\n\treturn c.Now()\n}","func (c *Context) Now() time.Time {\n\treturn c.currentTime\n}","func (t Time) ResetNow() time.Time {\n\tnow = defaultNow\n\treturn now()\n}","func Now() time.Time {\n\treturn time.Unix(0, time.Now().UnixNano()/1e6*1e6)\n}","func (a *IncrementingFakeTimeSource) Now() time.Time {\n\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\n\ta.NextIncrement++\n\n\treturn adjustedTime\n}","func Now() Time {\n\treturn Time(time.Now().UnixNano() / 1000)\n}","func (RealClock) Now() time.Time {\n\treturn time.Now()\n}","func nowTime() time.Time {\n\treturn time.Now().UTC()\n}","func resetTimeNow() {\n\ttimeNow = time.Now\n}","func (wc WallClock) Now() time.Time {\n\treturn time.Now()\n}","func Now() time.Time {\n\treturn time.Now().In(_defaultLocation)\n}","func GetCurrentTime() uint64 {\n\treturn ComposeTSByTime(time.Now(), 0)\n}","func Resume() {\n\tstopMux.Lock()\n\tif stoppedAt == nil {\n\t\tpanic(\"Time was not stopped\")\n\t}\n\tstoppedFor += time.Since(*stoppedAt)\n\tstoppedAt = nil\n\tstopMux.Unlock()\n}","func ResetNow() { now = time.Now }","func TimeNow() Time {\n\treturn Time{\n\t\ttime.Now(),\n\t}\n}","func (s *Stopwatch) Stop() {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.isRunning() {\n\t\ts.elapsedTime += time.Since(s.refTime)\n\t\ts.refTime = time.Time{}\n\t} else {\n\t\tfmt.Printf(\"WARNING: Stopwatch.Stop() isRunning is false\\n\")\n\t}\n}","func Now() Time {\n\t/*\n\t\tft := windows.Filetime{}\n\t\twindows.GetSystemTimeAsFileTime(&ft)\n\t\treturn Time(ft.Nanoseconds() / nanosecondsScale)\n\t*/\n\t// On Windows, it turns out to be much faster to simply call time.Now().Unix(), so do it\n\treturn Time(time.Now().Unix())\n}","func CurrentTime() time.Time {\n\treturn time.Now()\n}","func Now() Time {\n\tvar t Time\n\tt.FromNSec(uint64(gotime.Now().UnixNano()))\n\treturn t\n}","func now() time.Time {\n\treturn time.Now().UTC()\n}","func Now() time.Time {\n\treturn time.Date(int(Year.Status().(uint16)), time.Month(Month.Status().(uint8)), int(Day.Status().(uint16)), int(Hour.Status().(uint8)), int(Minute.Status().(uint8)), int(Second.Status().(uint8)), 0, time.FixedZone(\"custom\", int(Tz.Status().(int8)) * 3600))\n}","func GetCurrentTime() typeutil.Timestamp {\n\treturn ComposeTSByTime(time.Now(), 0)\n}","func (s *Scheduler) now() Time {\n\treturn s.provider.Now().Add(s.timeOffset)\n}","func (t *Tracker) Stop() {\n\tt.Finish = time.Now()\n\tt.Duration = time.Since(t.Run)\n}","func (sit *SyncIntervalTimer) Pause() {\n\t<-sit.sync\n}","func TimeNow() time.Time {\n\treturn time.Now().UTC()\n}","func (m *timeSource) Now() mstime.Time {\n\treturn mstime.Now()\n}","func (f FakeTimeSource) Now() time.Time {\n\treturn f.FakeTime\n}","func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}","func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}","func Now() time.Time {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tvar now time.Time\n\tif testMode {\n\t\tnow = testNow()\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.In(localtz.Get())\n}","func (s systemTimeSource) Now() time.Time {\n\treturn time.Now()\n}","func (t Time) GetNow() time.Time {\n\treturn now()\n}","func SetNow(n time.Time) {\n\tnow = n\n}","func Now() time.Time {\n\treturn time.Now().In(LOCATION)\n}","func TestClock_Now(t *testing.T) {\n\tnow := raft.NewClock().Now()\n\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\n\t\tt.Fatalf(\"clock time is different than wall time: exp=%v, got=%v\", exp, now)\n\t}\n}","func Now(ctx Context) time.Time {\n\treturn getWorkflowEnvironment(ctx).Now()\n}","func SetTime(ct int) {\n\tcurrenttime = ct\n}","func RestoreNow() {\n\tNow = time.Now\n}","func (tr *TimedRun) Stop(t *Timer) {\n\tstop := currentTimeFunc()\n\ttr.cl.Lock()\n\tdefer tr.cl.Unlock()\n\tif _, ok := tr.categories[t.category]; !ok {\n\t\ttr.categories[t.category] = 0\n\t}\n\ttr.categories[t.category] += stop.Sub(t.startTime)\n}","func (m *Mock) Now() time.Time {\n\tm.Lock()\n\tdefer m.Unlock()\n\treturn m.now\n}","func (sw *Stopwatch) Reset() {\n\tsw.t = time.Now()\n}","func (c *Timer) Pause() {\n\tc.ticker.Stop()\n\tc.passed += time.Now().Sub(c.lastTick)\n\tc.lastTick = time.Now()\n\tc.options.OnPaused()\n}","func makeFakeNow() func() time.Time {\n\tvar t time.Time\n\treturn func() time.Time {\n\t\tt = t.Add(time.Second)\n\t\treturn t\n\t}\n}","func (j *ScheduledJob) Pause() {\n\tj.rec.NextRun = time.Time{}\n\tj.markDirty(\"next_run\")\n}","func (gores *Gores) CurrentTime() int64 {\n\ttimestamp := time.Now().Unix()\n\treturn timestamp\n}","func newFakeTime() {\n\tfakeCurrentTime = fakeTime().Add(time.Hour * 24 * 2)\n}","func Time() time.Time {\n\tnow := time.Now().UTC()\n\tdif := time.Duration(rand.Int())\n\tstart := now.Add(dif * -1)\n\tend := now.Add(dif)\n\treturn TimeSpan(start, end)\n}","func NowTime() time.Time {\n\treturn ExtractTimeFromDatetime(time.Now())\n}","func (f *FakeTimeSource) Now() time.Time {\n\tf.mu.RLock()\n\tdefer f.mu.RUnlock()\n\treturn f.now\n}","func (tx *tx) SetNow(now time.Time) { tx.now = now }","func StopCurrent() error {\n\tsession := toggl.OpenSession(viper.GetString(\"token\"))\n\taccount, err := session.GetAccount()\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar timeEntry *toggl.TimeEntry\n\ttimeEntry, _ = getCurrentTimeEntry(account)\n\tif nil != timeEntry {\n\t\tsession.StopTimeEntry(*timeEntry)\n\t}\n\n\treturn nil\n}","func Now() Timespec {\n\treturn NewTimespec(time.Now())\n}","func (m *Mock) Now() time.Time {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\treturn m.now\n}"],"string":"[\n \"func (c *chrono) Now() time.Time {\\n\\tvar now time.Time\\n\\tif c.paused {\\n\\t\\tnow = c.pausedAt\\n\\t} else {\\n\\t\\tnow = time.Now()\\n\\t}\\n\\treturn now.Add(c.skew)\\n}\",\n \"func stoppableNow() time.Time {\\n\\tstopMux.RLock()\\n\\tdefer stopMux.RUnlock()\\n\\n\\tif stoppedAt != nil {\\n\\t\\treturn *stoppedAt\\n\\t}\\n\\treturn time.Now().Add(-stoppedFor)\\n}\",\n \"func (c *StoppedClock) Now() time.Time {\\n\\treturn c.t\\n}\",\n \"func (c *StoppedClock) Now() time.Time {\\n\\treturn c.time\\n}\",\n \"func (c *Clock) Now() time.Time {\\n\\treturn time.Now().UTC().Truncate(time.Second)\\n}\",\n \"func Stop() {\\n\\tstopMux.Lock()\\n\\tif stoppedAt != nil {\\n\\t\\tpanic(\\\"Time was already stopped\\\")\\n\\t}\\n\\tnow := Now()\\n\\tstoppedAt = &now\\n\\tstopMux.Unlock()\\n}\",\n \"func (c *Clock) Now() time.Time { return time.Now() }\",\n \"func (u *Util) Now() time.Time {\\n\\tif controlDuration != 0 {\\n\\t\\treturn time.Now().Add(controlDuration).UTC()\\n\\t}\\n\\treturn time.Now().UTC()\\n}\",\n \"func (c stubClocker) Now() time.Time { return c.t }\",\n \"func (realClocker) Now() time.Time { return time.Now() }\",\n \"func (t *Time) Now() time.Time {\\n\\treturn t.current\\n}\",\n \"func (c *ClockVal) Now() {\\n\\tc.CurrentTime = NowTime()\\n}\",\n \"func Now() time.Time { return now() }\",\n \"func (c *FakeClock) Now() time.Time {\\n\\tc.steps++\\n\\treturn c.Time(c.steps)\\n}\",\n \"func (c *RunningClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (d *dummyClock) Now() time.Time {\\n\\treturn time.Time{}\\n}\",\n \"func Now() time.Time { return time.Now().UTC() }\",\n \"func Now() time.Time { return time.Now().UTC() }\",\n \"func now() time.Duration { return time.Since(startTime) }\",\n \"func (clock *Clock) Now() time.Time {\\n\\tvar now time.Time\\n\\tclock.do(func(s *state) {\\n\\t\\tnow = s.t\\n\\t\\ts.t = s.t.Add(clock.step)\\n\\t})\\n\\treturn now\\n}\",\n \"func (p *PredefinedFake) Now() time.Time {\\n\\tadjustedTime := p.Base.Add(p.Delays[p.Next])\\n\\tp.Next++\\n\\treturn adjustedTime\\n}\",\n \"func Now() time.Time { return time.Now() }\",\n \"func Now() time.Time {\\n\\treturn CurrentClock().Now()\\n}\",\n \"func (Clock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (c *Clock) Now() time.Time {\\n\\tc.init()\\n\\trt := c.maybeGetRealTime()\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\n\\tstep := c.step\\n\\tif c.skipStep {\\n\\t\\tstep = 0\\n\\t\\tc.skipStep = false\\n\\t}\\n\\tc.advanceLocked(rt, step)\\n\\n\\treturn c.present\\n}\",\n \"func (p PT) Now() int64 {\\n\\tif p.Seconds {\\n\\t\\treturn time.Now().Unix()\\n\\t}\\n\\treturn time.Now().UnixNano()\\n}\",\n \"func (s *Stopwatch) Stop() {\\n\\ts.Lock()\\n\\tdefer s.Unlock()\\n\\tif s.active() {\\n\\t\\ts.stop = time.Now()\\n\\t}\\n}\",\n \"func (t *DefaultClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (defaultClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func Now() time.Time {\\n\\treturn Clock.Now()\\n}\",\n \"func (fc *fakeClock) Now() time.Time {\\n\\treturn fc.time\\n}\",\n \"func (tc *TestClock) Now() time.Time {\\n\\ttc.l.RLock()\\n\\tdefer tc.l.RUnlock()\\n\\treturn tc.t\\n}\",\n \"func Now() Time {\\n\\treturn DefaultScheduler.Now()\\n}\",\n \"func (c *CumulativeClock) Now() time.Time {\\n\\treturn c.current\\n}\",\n \"func Now() Time {\\n\\treturn Time{format(time.Now())}\\n}\",\n \"func Now(upToSecond ...bool) *TimeStamp {\\n\\treturn TimeFrom(time.Now(), upToSecond...)\\n}\",\n \"func InstrumentNow() {\\n\\tNow = stoppableNow\\n}\",\n \"func (c *MockClock) Now() time.Time {\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\treturn c.now\\n}\",\n \"func Now() *time.Time {\\n\\tt := time.Now()\\n\\treturn &t\\n}\",\n \"func (t *Timer) Now() time.Duration {\\n\\treturn nowFunc().Sub(t.Zero)\\n}\",\n \"func Now() Time {\\n\\treturn NewTime(time.Now())\\n}\",\n \"func Now() time.Time {\\n\\treturn Work.Now()\\n}\",\n \"func Now() time.Time {\\n\\treturn now()\\n}\",\n \"func (tr *TextRegion) TimeNow() {\\n\\ttr.Time.Now()\\n}\",\n \"func CurrentTime() Time {\\n\\tif atomic.LoadUint32(&isRunning) != 0 {\\n\\t\\treturn currentTime\\n\\t}\\n\\n\\treturn Time(time.Now().UnixNano() / 1e6)\\n}\",\n \"func Now() time.Time {\\n\\tif IsTest() {\\n\\t\\treturn now\\n\\t}\\n\\n\\treturn time.Now()\\n}\",\n \"func Now() time.Time {\\n\\treturn c.Now()\\n}\",\n \"func (c *Context) Now() time.Time {\\n\\treturn c.currentTime\\n}\",\n \"func (t Time) ResetNow() time.Time {\\n\\tnow = defaultNow\\n\\treturn now()\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Unix(0, time.Now().UnixNano()/1e6*1e6)\\n}\",\n \"func (a *IncrementingFakeTimeSource) Now() time.Time {\\n\\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\\n\\ta.NextIncrement++\\n\\n\\treturn adjustedTime\\n}\",\n \"func Now() Time {\\n\\treturn Time(time.Now().UnixNano() / 1000)\\n}\",\n \"func (RealClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func nowTime() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func resetTimeNow() {\\n\\ttimeNow = time.Now\\n}\",\n \"func (wc WallClock) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Now().In(_defaultLocation)\\n}\",\n \"func GetCurrentTime() uint64 {\\n\\treturn ComposeTSByTime(time.Now(), 0)\\n}\",\n \"func Resume() {\\n\\tstopMux.Lock()\\n\\tif stoppedAt == nil {\\n\\t\\tpanic(\\\"Time was not stopped\\\")\\n\\t}\\n\\tstoppedFor += time.Since(*stoppedAt)\\n\\tstoppedAt = nil\\n\\tstopMux.Unlock()\\n}\",\n \"func ResetNow() { now = time.Now }\",\n \"func TimeNow() Time {\\n\\treturn Time{\\n\\t\\ttime.Now(),\\n\\t}\\n}\",\n \"func (s *Stopwatch) Stop() {\\n\\ts.Lock()\\n\\tdefer s.Unlock()\\n\\n\\tif s.isRunning() {\\n\\t\\ts.elapsedTime += time.Since(s.refTime)\\n\\t\\ts.refTime = time.Time{}\\n\\t} else {\\n\\t\\tfmt.Printf(\\\"WARNING: Stopwatch.Stop() isRunning is false\\\\n\\\")\\n\\t}\\n}\",\n \"func Now() Time {\\n\\t/*\\n\\t\\tft := windows.Filetime{}\\n\\t\\twindows.GetSystemTimeAsFileTime(&ft)\\n\\t\\treturn Time(ft.Nanoseconds() / nanosecondsScale)\\n\\t*/\\n\\t// On Windows, it turns out to be much faster to simply call time.Now().Unix(), so do it\\n\\treturn Time(time.Now().Unix())\\n}\",\n \"func CurrentTime() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func Now() Time {\\n\\tvar t Time\\n\\tt.FromNSec(uint64(gotime.Now().UnixNano()))\\n\\treturn t\\n}\",\n \"func now() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Date(int(Year.Status().(uint16)), time.Month(Month.Status().(uint8)), int(Day.Status().(uint16)), int(Hour.Status().(uint8)), int(Minute.Status().(uint8)), int(Second.Status().(uint8)), 0, time.FixedZone(\\\"custom\\\", int(Tz.Status().(int8)) * 3600))\\n}\",\n \"func GetCurrentTime() typeutil.Timestamp {\\n\\treturn ComposeTSByTime(time.Now(), 0)\\n}\",\n \"func (s *Scheduler) now() Time {\\n\\treturn s.provider.Now().Add(s.timeOffset)\\n}\",\n \"func (t *Tracker) Stop() {\\n\\tt.Finish = time.Now()\\n\\tt.Duration = time.Since(t.Run)\\n}\",\n \"func (sit *SyncIntervalTimer) Pause() {\\n\\t<-sit.sync\\n}\",\n \"func TimeNow() time.Time {\\n\\treturn time.Now().UTC()\\n}\",\n \"func (m *timeSource) Now() mstime.Time {\\n\\treturn mstime.Now()\\n}\",\n \"func (f FakeTimeSource) Now() time.Time {\\n\\treturn f.FakeTime\\n}\",\n \"func (s SystemTimeSource) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (s SystemTimeSource) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func Now() time.Time {\\n\\tmu.Lock()\\n\\tdefer mu.Unlock()\\n\\tvar now time.Time\\n\\tif testMode {\\n\\t\\tnow = testNow()\\n\\t} else {\\n\\t\\tnow = time.Now()\\n\\t}\\n\\treturn now.In(localtz.Get())\\n}\",\n \"func (s systemTimeSource) Now() time.Time {\\n\\treturn time.Now()\\n}\",\n \"func (t Time) GetNow() time.Time {\\n\\treturn now()\\n}\",\n \"func SetNow(n time.Time) {\\n\\tnow = n\\n}\",\n \"func Now() time.Time {\\n\\treturn time.Now().In(LOCATION)\\n}\",\n \"func TestClock_Now(t *testing.T) {\\n\\tnow := raft.NewClock().Now()\\n\\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\\n\\t\\tt.Fatalf(\\\"clock time is different than wall time: exp=%v, got=%v\\\", exp, now)\\n\\t}\\n}\",\n \"func Now(ctx Context) time.Time {\\n\\treturn getWorkflowEnvironment(ctx).Now()\\n}\",\n \"func SetTime(ct int) {\\n\\tcurrenttime = ct\\n}\",\n \"func RestoreNow() {\\n\\tNow = time.Now\\n}\",\n \"func (tr *TimedRun) Stop(t *Timer) {\\n\\tstop := currentTimeFunc()\\n\\ttr.cl.Lock()\\n\\tdefer tr.cl.Unlock()\\n\\tif _, ok := tr.categories[t.category]; !ok {\\n\\t\\ttr.categories[t.category] = 0\\n\\t}\\n\\ttr.categories[t.category] += stop.Sub(t.startTime)\\n}\",\n \"func (m *Mock) Now() time.Time {\\n\\tm.Lock()\\n\\tdefer m.Unlock()\\n\\treturn m.now\\n}\",\n \"func (sw *Stopwatch) Reset() {\\n\\tsw.t = time.Now()\\n}\",\n \"func (c *Timer) Pause() {\\n\\tc.ticker.Stop()\\n\\tc.passed += time.Now().Sub(c.lastTick)\\n\\tc.lastTick = time.Now()\\n\\tc.options.OnPaused()\\n}\",\n \"func makeFakeNow() func() time.Time {\\n\\tvar t time.Time\\n\\treturn func() time.Time {\\n\\t\\tt = t.Add(time.Second)\\n\\t\\treturn t\\n\\t}\\n}\",\n \"func (j *ScheduledJob) Pause() {\\n\\tj.rec.NextRun = time.Time{}\\n\\tj.markDirty(\\\"next_run\\\")\\n}\",\n \"func (gores *Gores) CurrentTime() int64 {\\n\\ttimestamp := time.Now().Unix()\\n\\treturn timestamp\\n}\",\n \"func newFakeTime() {\\n\\tfakeCurrentTime = fakeTime().Add(time.Hour * 24 * 2)\\n}\",\n \"func Time() time.Time {\\n\\tnow := time.Now().UTC()\\n\\tdif := time.Duration(rand.Int())\\n\\tstart := now.Add(dif * -1)\\n\\tend := now.Add(dif)\\n\\treturn TimeSpan(start, end)\\n}\",\n \"func NowTime() time.Time {\\n\\treturn ExtractTimeFromDatetime(time.Now())\\n}\",\n \"func (f *FakeTimeSource) Now() time.Time {\\n\\tf.mu.RLock()\\n\\tdefer f.mu.RUnlock()\\n\\treturn f.now\\n}\",\n \"func (tx *tx) SetNow(now time.Time) { tx.now = now }\",\n \"func StopCurrent() error {\\n\\tsession := toggl.OpenSession(viper.GetString(\\\"token\\\"))\\n\\taccount, err := session.GetAccount()\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tvar timeEntry *toggl.TimeEntry\\n\\ttimeEntry, _ = getCurrentTimeEntry(account)\\n\\tif nil != timeEntry {\\n\\t\\tsession.StopTimeEntry(*timeEntry)\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func Now() Timespec {\\n\\treturn NewTimespec(time.Now())\\n}\",\n \"func (m *Mock) Now() time.Time {\\n\\tm.mu.Lock()\\n\\tdefer m.mu.Unlock()\\n\\treturn m.now\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.7088481","0.6919379","0.67659324","0.67019147","0.6559164","0.6402291","0.63203806","0.6272591","0.6245372","0.61367035","0.61324173","0.61246","0.6115938","0.60733736","0.6040197","0.60223395","0.6020786","0.6020786","0.60146177","0.6012215","0.59966373","0.5959776","0.594986","0.59412545","0.5932818","0.5914386","0.5879223","0.5872968","0.58612317","0.5852448","0.58428824","0.582453","0.5794403","0.579429","0.5771547","0.57653993","0.57504296","0.57231873","0.56907177","0.56901085","0.56872255","0.5681397","0.56751794","0.56665623","0.5663568","0.5654582","0.5639455","0.5635226","0.5632688","0.56285256","0.5625493","0.56074226","0.5604195","0.55978066","0.5587286","0.5576184","0.5575942","0.5544697","0.5544119","0.55364","0.55070937","0.5504631","0.547006","0.5465696","0.54551595","0.54522413","0.5430223","0.5428441","0.54235643","0.54144454","0.5408864","0.539088","0.5390319","0.5384731","0.5373761","0.5373761","0.537357","0.5373175","0.5362128","0.5352885","0.5338776","0.5330385","0.5323998","0.53135836","0.5288172","0.5284446","0.52778125","0.5271241","0.52665097","0.5262797","0.5260879","0.5251372","0.5250993","0.5248289","0.52457446","0.52367693","0.523104","0.5229294","0.5227222","0.5227182"],"string":"[\n \"0.7088481\",\n \"0.6919379\",\n \"0.67659324\",\n \"0.67019147\",\n \"0.6559164\",\n \"0.6402291\",\n \"0.63203806\",\n \"0.6272591\",\n \"0.6245372\",\n \"0.61367035\",\n \"0.61324173\",\n \"0.61246\",\n \"0.6115938\",\n \"0.60733736\",\n \"0.6040197\",\n \"0.60223395\",\n \"0.6020786\",\n \"0.6020786\",\n \"0.60146177\",\n \"0.6012215\",\n \"0.59966373\",\n \"0.5959776\",\n \"0.594986\",\n \"0.59412545\",\n \"0.5932818\",\n \"0.5914386\",\n \"0.5879223\",\n \"0.5872968\",\n \"0.58612317\",\n \"0.5852448\",\n \"0.58428824\",\n \"0.582453\",\n \"0.5794403\",\n \"0.579429\",\n \"0.5771547\",\n \"0.57653993\",\n \"0.57504296\",\n \"0.57231873\",\n \"0.56907177\",\n \"0.56901085\",\n \"0.56872255\",\n \"0.5681397\",\n \"0.56751794\",\n \"0.56665623\",\n \"0.5663568\",\n \"0.5654582\",\n \"0.5639455\",\n \"0.5635226\",\n \"0.5632688\",\n \"0.56285256\",\n \"0.5625493\",\n \"0.56074226\",\n \"0.5604195\",\n \"0.55978066\",\n \"0.5587286\",\n \"0.5576184\",\n \"0.5575942\",\n \"0.5544697\",\n \"0.5544119\",\n \"0.55364\",\n \"0.55070937\",\n \"0.5504631\",\n \"0.547006\",\n \"0.5465696\",\n \"0.54551595\",\n \"0.54522413\",\n \"0.5430223\",\n \"0.5428441\",\n \"0.54235643\",\n \"0.54144454\",\n \"0.5408864\",\n \"0.539088\",\n \"0.5390319\",\n \"0.5384731\",\n \"0.5373761\",\n \"0.5373761\",\n \"0.537357\",\n \"0.5373175\",\n \"0.5362128\",\n \"0.5352885\",\n \"0.5338776\",\n \"0.5330385\",\n \"0.5323998\",\n \"0.53135836\",\n \"0.5288172\",\n \"0.5284446\",\n \"0.52778125\",\n \"0.5271241\",\n \"0.52665097\",\n \"0.5262797\",\n \"0.5260879\",\n \"0.5251372\",\n \"0.5250993\",\n \"0.5248289\",\n \"0.52457446\",\n \"0.52367693\",\n \"0.523104\",\n \"0.5229294\",\n \"0.5227222\",\n \"0.5227182\"\n]"},"document_score":{"kind":"string","value":"0.5683553"},"document_rank":{"kind":"string","value":"41"}}},{"rowIdx":104796,"cells":{"query":{"kind":"string","value":"Continues time after having been paused. This has no effect if clock is already running."},"document":{"kind":"string","value":"func (c *chrono) Continue() {\n\tc.paused = false\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (c *chrono) Pause() {\n\tc.pausedAt = c.Now()\n\tc.paused = true\n}","func (c *Timer) Pause() {\n\tc.ticker.Stop()\n\tc.passed += time.Now().Sub(c.lastTick)\n\tc.lastTick = time.Now()\n\tc.options.OnPaused()\n}","func (sit *SyncIntervalTimer) Pause() {\n\t<-sit.sync\n}","func (j *ScheduledJob) Pause() {\n\tj.rec.NextRun = time.Time{}\n\tj.markDirty(\"next_run\")\n}","func (e *Emulator) Pause() {}","func (c *Context) PAUSE() {\n\tc.addinstruction(x86.PAUSE())\n}","func (clk *Clock) Continue() {\n\tclk.midiOut <- midiContinue\n}","func (timer *WallclockTimer) Resume() error {\n\ttimer.command <- \"resume\"\n\treturn nil\n}","func Resume() {\n\tstopMux.Lock()\n\tif stoppedAt == nil {\n\t\tpanic(\"Time was not stopped\")\n\t}\n\tstoppedFor += time.Since(*stoppedAt)\n\tstoppedAt = nil\n\tstopMux.Unlock()\n}","func pause() {\n\ttime.Sleep(500 * time.Millisecond)\n}","func PAUSE() { ctx.PAUSE() }","func (s *sched) pause() {\n\told := atomic.LoadPointer(&s.timer)\n\t// if old is nil then there is someone who tries to stop the timer.\n\tif old != nil {\n\t\t(*time.Timer)(old).Stop()\n\t}\n}","func Pause() {\n\tatomic.AddUint64(&sched0.pausing, 1)\n\tsched0.pause()\n}","func (p *Peer) pause() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Pause()\n}","func (p *Peer) pause() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Pause()\n}","func (sit *SyncIntervalTimer) Resume() {\n\tsit.sync <- struct{}{}\n}","func pause() tea.Msg {\n\ttime.Sleep(time.Millisecond * 600)\n\treturn DoneMsg{}\n}","func Clock(ctx context.Context) {\n\tfor {\n\t\tselect {\n\t\tcase <- ctx.Done():\n\t\t\tfmt.Printf(\"\\r[ # ]\")\n\t\t\treturn\n\t\tdefault:\n\t\t\tfor _, c := range `-\\|/` {\n\t\t\t\tfmt.Printf(\"\\r[ %c ]\", c)\n\t\t\t\ttime.Sleep(200 * time.Millisecond)\n\t\t\t}\n\t\t}\n\t}\n}","func (p *Pacer) Pause() {\n\tp.pause <- struct{}{} // block this channel\n\tp.paused <- struct{}{} // set flag to indicate paused\n}","func (p *Peer) resume() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Reset()\n}","func (p *Peer) resume() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Reset()\n}","func (p *ProgressMeter) Pause() {\n\tif atomic.CompareAndSwapInt32(&p.started, 1, 0) {\n\t\tp.finished <- true\n\t}\n}","func (doDebugger Debugger) Pause() (err error) {\n\tb := debugger.Pause()\n\treturn b.Do(doDebugger.ctxWithExecutor)\n}","func (cpu *Mos6502) Clock() {\n\t// if current instruction complete, read and execute next instruction\n\tif cpu.cycles == 0 {\n\t\topcode := cpu.read(cpu.pc)\n\t\tinstruction := cpu.lookup[opcode]\n\n\t\tcpu.setStatusFlag(U, true)\n\t\tcpu.pc++\n\n\t\tcpu.cycles = instruction.cycles\n\t\tadditionalCycleAddr := instruction.setAddressMode()\n\t\tadditionalCycleOp := instruction.performOp()\n\t\tcpu.cycles += additionalCycleAddr & additionalCycleOp\n\n\t\tcpu.setStatusFlag(U, true)\n\t}\n\n\tcpu.cycles--\n}","func (t *task) Pause(_ context.Context) error {\n\tif t.ctr.ociSpec.Windows.HyperV == nil {\n\t\treturn cerrdefs.ErrNotImplemented\n\t}\n\n\tt.ctr.mu.Lock()\n\tdefer t.ctr.mu.Unlock()\n\n\tif err := t.assertIsCurrentTask(); err != nil {\n\t\treturn err\n\t}\n\tif t.ctr.hcsContainer == nil {\n\t\treturn errdefs.NotFound(errors.WithStack(fmt.Errorf(\"container %q not found\", t.ctr.id)))\n\t}\n\tif err := t.ctr.hcsContainer.Pause(); err != nil {\n\t\treturn err\n\t}\n\n\tt.ctr.isPaused = true\n\n\tt.ctr.client.eventQ.Append(t.ctr.id, func() {\n\t\terr := t.ctr.client.backend.ProcessEvent(t.ctr.id, libcontainerdtypes.EventPaused, libcontainerdtypes.EventInfo{\n\t\t\tContainerID: t.ctr.id,\n\t\t\tProcessID: t.id,\n\t\t})\n\t\tt.ctr.client.logger.WithFields(log.Fields{\n\t\t\t\"container\": t.ctr.id,\n\t\t\t\"event\": libcontainerdtypes.EventPaused,\n\t\t}).Info(\"sending event\")\n\t\tif err != nil {\n\t\t\tt.ctr.client.logger.WithError(err).WithFields(log.Fields{\n\t\t\t\t\"container\": t.ctr.id,\n\t\t\t\t\"event\": libcontainerdtypes.EventPaused,\n\t\t\t}).Error(\"failed to process event\")\n\t\t}\n\t})\n\n\treturn nil\n}","func (helper *testHelper) Pause(secs int) {}","func (o *OMXPlayer) Pause() error {\n\tif o.status == statusPaused {\n\t\treturn nil\n\t}\n\n\t// TODO: test this properly by mocking dbus package.\n\tif o.testing {\n\t\to.status = statusPaused\n\t\treturn nil\n\t}\n\n\t_, err := o.dbusSend(\"Pause\", dbus.FlagNoAutoStart)\n\tif err != nil {\n\t\treturn err\n\t}\n\to.status = statusPaused\n\treturn nil\n}","func (c *Clock) AdvanceTo(t time.Time) {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\tc.realTime = rt\n\tc.present = t\n\tc.events.AdvanceTo(c.present)\n}","func (s *Scheduler) Pause() error {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\ts.paused = true\n\treturn nil\n}","func (c *Clock) Advance(d time.Duration) time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\n\tc.advanceLocked(rt, d)\n\treturn c.present\n}","func (e *Enumeration) Pause() {\n\te.pause <- struct{}{}\n}","func (e *Enumeration) Pause() {\n\te.pause <- struct{}{}\n}","func (ft *fakeTimer) advance(d time.Duration) {\n\tft.lock.Lock()\n\tdefer ft.lock.Unlock()\n\n\tft.now = ft.now.Add(d)\n\tif ft.active && !ft.now.Before(ft.timeout) {\n\t\tft.active = false\n\t\tft.c <- ft.timeout\n\t}\n}","func (timer *Timer) Pause() error {\n\tch := make(chan error)\n\ttimer.pause <- ch\n\treturn <-ch\n}","func (c *Switch) Pause() {\n\tc.lock.RLock()\n\tif cp, ok := c.subRenderables[c.curRenderable].(CanPause); ok {\n\t\tcp.Pause()\n\t}\n\tc.lock.RUnlock()\n}","func (_Cakevault *CakevaultTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Cakevault.contract.Transact(opts, \"pause\")\n}","func (w *Walker) Pause() {\n\tw.pauseRequested = true\n}","func (p *Pacer) Resume() {\n\t<-p.paused // clear flag to indicate paused\n\t<-p.pause // unblock this channel\n}","func (p *Playlist) Pause() {\n\tp.ch <- \"pause\"\n}","func (rep *Reporter) Continue() {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tif rep == nil {\n\t\treturn\n\t}\n\n\tif !rep.metricsCh.IsPaused() {\n\t\treturn\n\t}\n\n\trep.metricsCh.Continue()\n}","func AdvanceTo(newTime time.Time) {\n\tnow := Now()\n\tfound := false\n\tfor _, s := range sortedSchedulers() {\n\t\tnextTick := s.tickAfter(now)\n\t\tif nextTick.After(now) && !nextTick.After(newTime) {\n\t\t\tfound = true\n\t\t\tnowInTest.Store(nextTick)\n\t\t\ts.Lock()\n\t\t\tif s.paused {\n\t\t\t\ts.fireOnResume = true\n\t\t\t} else {\n\t\t\t\ts.notifyFn()\n\t\t\t}\n\t\t\ts.Unlock()\n\t\t}\n\t}\n\tif !found {\n\t\tnowInTest.Store(newTime)\n\t\treturn\n\t}\n\tif Now() != newTime {\n\t\tAdvanceTo(newTime)\n\t}\n}","func (j Jibi) Pause() {\n\tj.RunCommand(CmdPause, nil)\n}","func (c *StoppedClock) SetTime(time time.Time) {\n\tc.time = time\n}","func main() {\n var working float64\n var relax float64\n var switchact bool = false\n var switch2relax bool = true\n var switch2working bool = false\n\n fmt.Println(\"Type any key for start the Workday\")\n fmt.Println(\"The counter starts with relaxing type any key for start working\")\n fmt.Println(\"You start workint at \" ,time.Now())\n ch := make(chan string)\n go func(ch chan string) {\n // disable input buffering\n exec.Command(\"stty\", \"-F\", \"/dev/tty\", \"cbreak\", \"min\", \"1\").Run()\n // do not display entered characters on the screen\n exec.Command(\"stty\", \"-F\", \"/dev/tty\", \"-echo\").Run()\n var b []byte = make([]byte, 1)\n //Routine for check for standard input with no wait for a key\n for {\n os.Stdin.Read(b)\n ch <- string(b)\n }\n }(ch)\n\n //The intention with this fork was do a Thiker and with this controll the flow of time\n for {\n select {\n \n //In case of a signal by key lanch this case, change status of working to relax\n //and show time elapsed\n case stdin, _ := <-ch:\n fmt.Println(\"\\n \\n Keys pressed:\", stdin)\n if switch2relax == true {\n fmt.Println(\"\\n Now you are relaxing\\n \\n\")\n }else {\n fmt.Println(\"\\n Now you are working \\n \\n\")\n }\n fmt.Printf(\"working for %f Hours \\n\", ((working/60.0)/60.0))\n fmt.Printf(\"Relaxing for %f Hours \\n\", (relax/60.0)/60.0)\n\n switch2relax = !switch2relax\n switch2working = !switch2working\n switchact = true\n \n //By default the program is count time, if the routin is not counting time\n //working count time relaxing, this script has no pause\n default:\n if switchact == true {\n if switch2relax == true {\n working=1.0+working\n if ((working/60.0)/60.0) == 8.00000 {\n exec.Command(\"mplayer\", \"stopwork.mp3\").Run()\n }\n //fmt.Println(\"Working for \", working)\n }else{\n relax=1.0+relax\n //fmt.Println(\"relaxing for \", relax)\n }\n }\n }\n time.Sleep(time.Second * 1)\n }\n}","func (_Cakevault *CakevaultSession) Pause() (*types.Transaction, error) {\n\treturn _Cakevault.Contract.Pause(&_Cakevault.TransactOpts)\n}","func (p *Profile) Pause() error {\n\tif !p.started {\n\t\treturn errors.New(\"mxnet profile was not started\")\n\t}\n\tif p.stopped == true || p.paused == true {\n\t\treturn nil\n\t}\n\tdefer func() {\n\t\tp.paused = true\n\t}()\n\tp.lastPauseTime = time.Now()\n\tsuccess, err := C.MXProfilePause(C.int(1))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif success != 0 {\n\t\treturn GetLastError()\n\t}\n\n\treturn nil\n}","func (m *Machine) Pause() error {\n\tm.State = driver.Paused\n\tfmt.Printf(\"Pause %s: %s\\n\", m.Name, m.State)\n\treturn nil\n}","func (t Track) Pause(bool) {\n\tpanic(\"implement me\")\n}","func (_Cakevault *CakevaultTransactorSession) Pause() (*types.Transaction, error) {\n\treturn _Cakevault.Contract.Pause(&_Cakevault.TransactOpts)\n}","func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}","func (s *SlaveHealthChecker) Pause() {\n\ts.Lock()\n\tdefer s.Unlock()\n\ts.paused = true\n}","func (_OracleMgr *OracleMgrTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _OracleMgr.contract.Transact(opts, \"pause\")\n}","func (_ERC20Pausable *ERC20PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _ERC20Pausable.contract.Transact(opts, \"pause\")\n}","func Pause(msg string) {\n\tfmt.Print(msg, \"[Press Enter to Continue]: \")\n\tvar s string\n\t_, err := fmt.Scan(&s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}","func (n *BaseNode) Continue() {\n\tn.continueFunc(func() {\n\t\tif n.cancelPause != nil {\n\t\t\tn.cancelPause()\n\t\t\tn.cancelPause = nil\n\t\t}\n\t})\n}","func (d *Device) Pause() {\n\tdefer d.mu.Unlock()\n\td.mu.Lock()\n\td.paused = true\n}","func (_DelegateProfile *DelegateProfileTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _DelegateProfile.contract.Transact(opts, \"pause\")\n}","func UpdateTime(dt int) {\n\tcurrenttime = currenttime + dt\n\tif currenttime >= 2400 {\n\t\tcurrenttime = currenttime - 2400\n\t\tdaycount = daycount + 1\n\t}\n}","func (a *Agent) PauseSync() {\n\t// Do this outside of lock as it has it's own locking\n\ta.sync.Pause()\n\n\t// Coordinate local state watchers\n\ta.syncMu.Lock()\n\tdefer a.syncMu.Unlock()\n\tif a.syncCh == nil {\n\t\ta.syncCh = make(chan struct{})\n\t}\n}","func (c *MockClock) Set(t time.Time) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tif t.After(c.now) {\n\t\tc.advanceTo(t)\n\t} else {\n\t\tc.now = t // move back in time\n\t}\n}","func (_MonsterAccessControl *MonsterAccessControlTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _MonsterAccessControl.contract.Transact(opts, \"pause\")\n}","func (p *Proxy) Pause() {\n\tp.configMutex.Lock()\n\tif p.state == proxyActive {\n\t\tp.state = proxyPaused\n\t\tp.connMutex.Unlock()\n\t}\n\tp.configMutex.Unlock()\n}","func (s *ExperimentSpec) Pause() bool {\n\tif s.ManualOverride != nil && s.ManualOverride.Action == ActionPause {\n\t\treturn true\n\t}\n\treturn false\n}","func (c *MockClock) advanceToNextWakeup() {\n\tif len(c.wakeups) < 1 {\n\t\treturn\n\t}\n\tw := heap.Pop(&c.wakeups).(*wakeup)\n\tc.logf(\"MockClock: Advancing time from now=%s to next wakeup time %s.\",\n\t\ttsStr(c.now), tsStr(w.t))\n\tc.now = w.t\n\tselect {\n\tcase w.c <- w.t:\n\tdefault:\n\t}\n\t// give things a chance to run\n\truntime.Gosched()\n\tc.logf(\"MockClock: Advanced time, now=%s.\", tsStr(c.now))\n}","func (c *MockClock) advanceTo(end time.Time) {\n\tfor {\n\t\tt := c.nextWakeupTime()\n\t\tif t == (time.Time{}) || t.After(end) {\n\t\t\tbreak\n\t\t}\n\t\tc.advanceToNextWakeup()\n\t}\n\tc.now = end\n}","func SetTime(ct int) {\n\tcurrenttime = ct\n}","func (cg *CgroupImpl) Pause() error {\n\treturn cg.setFreezeState(1)\n}","func (l *LightningLoader) Pause() {\n\tl.Lock()\n\tdefer l.Unlock()\n\tif l.isClosed() {\n\t\tl.logger.Warn(\"try to pause, but already closed\")\n\t\treturn\n\t}\n\tif l.cancel != nil {\n\t\tl.cancel()\n\t}\n\tl.core.Stop()\n}","func AnounanceTime(){\n\tfor {\n\t\tfmt.Println(time.Now())\n\t\ttime.Sleep(5*time.Second)\n\t}\n}","func (_Token *TokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Token.contract.Transact(opts, \"pause\")\n}","func (_ChpRegistry *ChpRegistryTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _ChpRegistry.contract.Transact(opts, \"pause\")\n}","func (_Cakevault *CakevaultCaller) Paused(opts *bind.CallOpts) (bool, error) {\n\tvar out []interface{}\n\terr := _Cakevault.contract.Call(opts, &out, \"paused\")\n\n\tif err != nil {\n\t\treturn *new(bool), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(bool)).(*bool)\n\n\treturn out0, err\n\n}","func (_ElvToken *ElvTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _ElvToken.contract.Transact(opts, \"pause\")\n}","func (t Time) Clock() (hour, min, sec int) {}","func (_MonsterOwnership *MonsterOwnershipTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _MonsterOwnership.contract.Transact(opts, \"pause\")\n}","func (c *Compound) Pause() {\n\tc.lock.RLock()\n\tif cp, ok := c.subRenderables[c.curRenderable].(CanPause); ok {\n\t\tcp.Pause()\n\t}\n\tc.lock.RUnlock()\n}","func Pause() {\n\tNamedLoggers.Range(func(key, value interface{}) bool {\n\t\tlogger := value.(*Logger)\n\t\tlogger.Pause()\n\t\tlogger.Flush()\n\t\treturn true\n\t})\n}","func (d *Debugger) Continue() {\n\tif d.BreakpointActive {\n\t\td.Cont <- true\n\t}\n}","func (_LifToken *LifTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _LifToken.contract.Transact(opts, \"pause\")\n}","func (_Pausable *PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Pausable.contract.Transact(opts, \"pause\")\n}","func (_Pausable *PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Pausable.contract.Transact(opts, \"pause\")\n}","func (c *MockClock) Advance(delta time.Duration) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tend := c.now.Add(delta)\n\tc.advanceTo(end)\n}","func (cr *Playback) Pause() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tcurrentParty, err := session.CurrentParty(c)\n\t\tif err != nil {\n\t\t\tif err != sql.ErrNoRows {\n\t\t\t\tc.Error(e.Internal.CausedBy(err))\n\t\t\t\tc.Abort()\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tc.Error(e.NotFound.WithMessage(\"Host has not started a party\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\t// It is a bad request to pause when no track is playing\n\t\tif !currentParty.CurrentTrack.Valid {\n\t\t\tc.Error(e.BadRequest.WithDetail(\"Host is not playing music\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\t// Pause the current track\n\t\tvar currentTrack models.PlayingTrack\n\t\terr = currentParty.CurrentTrack.Unmarshal(currentTrack)\n\t\tif err != nil {\n\t\t\tc.Error(e.Internal.CausedBy(err))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\telapsedParam, exists := c.GetQuery(\"elapsed\")\n\t\tif !exists {\n\t\t\tc.Error(e.BadRequest.WithDetail(\"Must provide elapsed duration to pause playback\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\t\telapsed, err := strconv.ParseUint(elapsedParam, 10, 64)\n\t\tif err != nil {\n\t\t\tc.Error(\n\t\t\t\te.BadRequest.\n\t\t\t\t\tWithDetail(\"Must provide valid elapsed duration to pause playback\").\n\t\t\t\t\tCausedBy(err),\n\t\t\t)\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\t// It is a bad request to pause a track that is already paused\n\t\tif currentTrack.Paused {\n\t\t\tc.Error(e.BadRequest.WithDetail(\"Current track is already paused\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\tcurrentTrack.Paused = true\n\t\tcurrentTrack.Elapsed = uint(elapsed)\n\t\terr = cr.updateTrackAndBroadcast(currentParty, &currentTrack)\n\t\tif err != nil {\n\t\t\tc.Error(e.Internal.CausedBy(err))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\tc.JSON(http.StatusOK, models.EmptyRespose)\n\t}\n}","func (_MonsterAccessControl *MonsterAccessControlTransactorSession) Pause() (*types.Transaction, error) {\n\treturn _MonsterAccessControl.Contract.Pause(&_MonsterAccessControl.TransactOpts)\n}","func (t *Time) Sleep(s time.Duration) {\n\tt.current = t.current.Add(s)\n}","func main() {\n\tp := fmt.Println\n\t\n\tp(\"------- Alarm Clock - Main initialized -------\")\n\tvar wg sync.WaitGroup\n\t\n\n\t//start := time.Now()\n\tcurrent := time.Now()\n\tfmt.Print(\"Klockan är: \", current.Hour(),\":\", current.Minute(),\".\",current.Second(),\"\\n\")\n\tp(\"---------------------------------\")\n\t//diff := current.Sub(start)\n\t//fmt.Println(\"DIFF:\",diff)\n\tp(\"START TIME:\",current)\n\t\n\t//add one to waitgroup object for each new goroutine started\n\twg.Add(1)\n\tgo Remind(\"Dags att äta\",3*time.Second)\n\twg.Add(1)\n\tgo Remind(\"Dags att arbeta\",8*time.Second)\n\twg.Add(1)\n\tgo Remind(\"Dags att sova\",24*time.Second)\n\t\n\twg.Wait()\n\t//p(currTime.Format(\"15:04:05 (2006-01-02)\"))\n\t//p(\"TIME DIFF: \",currTime.Sub(startTime))\n\t\n\n}","func (_DelegateProfile *DelegateProfileTransactorSession) Pause() (*types.Transaction, error) {\n\treturn _DelegateProfile.Contract.Pause(&_DelegateProfile.TransactOpts)\n}","func (c *Client) Pause() error {\n\treturn nil\n}","func TestProgressResume(t *testing.T) {\n\tp := &Progress{\n\t\tNext: 2,\n\t\tPaused: true,\n\t}\n\tp.maybeDecrTo(1, 1)\n\tif p.Paused {\n\t\tt.Errorf(\"paused= %v, want false\", p.Paused)\n\t}\n\tp.Paused = true\n\tp.maybeUpdate(2)\n\tif p.Paused {\n\t\tt.Errorf(\"paused= %v, want false\", p.Paused)\n\t}\n}","func (n *BaseNode) Pause() {\n\tn.pauseFunc(func() {\n\t\tn.ctxPause, n.cancelPause = context.WithCancel(n.ctx)\n\t})\n}","func (container *container) Pause() error {\r\n\treturn convertSystemError(container.system.Pause(context.Background()), container)\r\n}","func NextPause() func() (pause time.Duration) {\n rando := rand.New(rand.NewSource(time.Now().UnixNano()))\n return func() (pause time.Duration) {\n pause = time.Duration((rando.Intn(500)) * int(time.Millisecond))\n return\n }\n}","func (_PausableToken *PausableTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _PausableToken.contract.Transact(opts, \"pause\")\n}","func (o RunnerOutput) Paused() pulumi.BoolOutput {\n\treturn o.ApplyT(func(v *Runner) pulumi.BoolOutput { return v.Paused }).(pulumi.BoolOutput)\n}","func (t *SpotifyDevice) Pause() {\n\tif t.client == nil {\n\t\treturn\n\t}\n\tlog.Println(\"Pausing Spotify device\")\n\tif err := t.client.PauseOpt(t.playOpts); err != nil {\n\t\tlog.Println(err)\n\t}\n}","func (_MonsterAccessControl *MonsterAccessControlSession) Pause() (*types.Transaction, error) {\n\treturn _MonsterAccessControl.Contract.Pause(&_MonsterAccessControl.TransactOpts)\n}","func (st *LevelCompleteState) OnPause(world w.World) {}","func (s *SlaveHealthChecker) Continue(slaveUPID *upid.UPID) {\n\ts.Lock()\n\tdefer s.Unlock()\n\ts.paused = false\n\ts.slaveUPID = slaveUPID\n}","func (_DelegateProfile *DelegateProfileSession) Pause() (*types.Transaction, error) {\n\treturn _DelegateProfile.Contract.Pause(&_DelegateProfile.TransactOpts)\n}","func (fgs *FakeGraphSync) Pause(ctx context.Context, requestID graphsync.RequestID) error {\n\tfgs.pauses <- requestID\n\treturn fgs.ReturnedPauseError\n}"],"string":"[\n \"func (c *chrono) Pause() {\\n\\tc.pausedAt = c.Now()\\n\\tc.paused = true\\n}\",\n \"func (c *Timer) Pause() {\\n\\tc.ticker.Stop()\\n\\tc.passed += time.Now().Sub(c.lastTick)\\n\\tc.lastTick = time.Now()\\n\\tc.options.OnPaused()\\n}\",\n \"func (sit *SyncIntervalTimer) Pause() {\\n\\t<-sit.sync\\n}\",\n \"func (j *ScheduledJob) Pause() {\\n\\tj.rec.NextRun = time.Time{}\\n\\tj.markDirty(\\\"next_run\\\")\\n}\",\n \"func (e *Emulator) Pause() {}\",\n \"func (c *Context) PAUSE() {\\n\\tc.addinstruction(x86.PAUSE())\\n}\",\n \"func (clk *Clock) Continue() {\\n\\tclk.midiOut <- midiContinue\\n}\",\n \"func (timer *WallclockTimer) Resume() error {\\n\\ttimer.command <- \\\"resume\\\"\\n\\treturn nil\\n}\",\n \"func Resume() {\\n\\tstopMux.Lock()\\n\\tif stoppedAt == nil {\\n\\t\\tpanic(\\\"Time was not stopped\\\")\\n\\t}\\n\\tstoppedFor += time.Since(*stoppedAt)\\n\\tstoppedAt = nil\\n\\tstopMux.Unlock()\\n}\",\n \"func pause() {\\n\\ttime.Sleep(500 * time.Millisecond)\\n}\",\n \"func PAUSE() { ctx.PAUSE() }\",\n \"func (s *sched) pause() {\\n\\told := atomic.LoadPointer(&s.timer)\\n\\t// if old is nil then there is someone who tries to stop the timer.\\n\\tif old != nil {\\n\\t\\t(*time.Timer)(old).Stop()\\n\\t}\\n}\",\n \"func Pause() {\\n\\tatomic.AddUint64(&sched0.pausing, 1)\\n\\tsched0.pause()\\n}\",\n \"func (p *Peer) pause() {\\n\\tp.mutex.Lock()\\n\\tdefer p.mutex.Unlock()\\n\\tp.heartbeatTimer.Pause()\\n}\",\n \"func (p *Peer) pause() {\\n\\tp.mutex.Lock()\\n\\tdefer p.mutex.Unlock()\\n\\tp.heartbeatTimer.Pause()\\n}\",\n \"func (sit *SyncIntervalTimer) Resume() {\\n\\tsit.sync <- struct{}{}\\n}\",\n \"func pause() tea.Msg {\\n\\ttime.Sleep(time.Millisecond * 600)\\n\\treturn DoneMsg{}\\n}\",\n \"func Clock(ctx context.Context) {\\n\\tfor {\\n\\t\\tselect {\\n\\t\\tcase <- ctx.Done():\\n\\t\\t\\tfmt.Printf(\\\"\\\\r[ # ]\\\")\\n\\t\\t\\treturn\\n\\t\\tdefault:\\n\\t\\t\\tfor _, c := range `-\\\\|/` {\\n\\t\\t\\t\\tfmt.Printf(\\\"\\\\r[ %c ]\\\", c)\\n\\t\\t\\t\\ttime.Sleep(200 * time.Millisecond)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n}\",\n \"func (p *Pacer) Pause() {\\n\\tp.pause <- struct{}{} // block this channel\\n\\tp.paused <- struct{}{} // set flag to indicate paused\\n}\",\n \"func (p *Peer) resume() {\\n\\tp.mutex.Lock()\\n\\tdefer p.mutex.Unlock()\\n\\tp.heartbeatTimer.Reset()\\n}\",\n \"func (p *Peer) resume() {\\n\\tp.mutex.Lock()\\n\\tdefer p.mutex.Unlock()\\n\\tp.heartbeatTimer.Reset()\\n}\",\n \"func (p *ProgressMeter) Pause() {\\n\\tif atomic.CompareAndSwapInt32(&p.started, 1, 0) {\\n\\t\\tp.finished <- true\\n\\t}\\n}\",\n \"func (doDebugger Debugger) Pause() (err error) {\\n\\tb := debugger.Pause()\\n\\treturn b.Do(doDebugger.ctxWithExecutor)\\n}\",\n \"func (cpu *Mos6502) Clock() {\\n\\t// if current instruction complete, read and execute next instruction\\n\\tif cpu.cycles == 0 {\\n\\t\\topcode := cpu.read(cpu.pc)\\n\\t\\tinstruction := cpu.lookup[opcode]\\n\\n\\t\\tcpu.setStatusFlag(U, true)\\n\\t\\tcpu.pc++\\n\\n\\t\\tcpu.cycles = instruction.cycles\\n\\t\\tadditionalCycleAddr := instruction.setAddressMode()\\n\\t\\tadditionalCycleOp := instruction.performOp()\\n\\t\\tcpu.cycles += additionalCycleAddr & additionalCycleOp\\n\\n\\t\\tcpu.setStatusFlag(U, true)\\n\\t}\\n\\n\\tcpu.cycles--\\n}\",\n \"func (t *task) Pause(_ context.Context) error {\\n\\tif t.ctr.ociSpec.Windows.HyperV == nil {\\n\\t\\treturn cerrdefs.ErrNotImplemented\\n\\t}\\n\\n\\tt.ctr.mu.Lock()\\n\\tdefer t.ctr.mu.Unlock()\\n\\n\\tif err := t.assertIsCurrentTask(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tif t.ctr.hcsContainer == nil {\\n\\t\\treturn errdefs.NotFound(errors.WithStack(fmt.Errorf(\\\"container %q not found\\\", t.ctr.id)))\\n\\t}\\n\\tif err := t.ctr.hcsContainer.Pause(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tt.ctr.isPaused = true\\n\\n\\tt.ctr.client.eventQ.Append(t.ctr.id, func() {\\n\\t\\terr := t.ctr.client.backend.ProcessEvent(t.ctr.id, libcontainerdtypes.EventPaused, libcontainerdtypes.EventInfo{\\n\\t\\t\\tContainerID: t.ctr.id,\\n\\t\\t\\tProcessID: t.id,\\n\\t\\t})\\n\\t\\tt.ctr.client.logger.WithFields(log.Fields{\\n\\t\\t\\t\\\"container\\\": t.ctr.id,\\n\\t\\t\\t\\\"event\\\": libcontainerdtypes.EventPaused,\\n\\t\\t}).Info(\\\"sending event\\\")\\n\\t\\tif err != nil {\\n\\t\\t\\tt.ctr.client.logger.WithError(err).WithFields(log.Fields{\\n\\t\\t\\t\\t\\\"container\\\": t.ctr.id,\\n\\t\\t\\t\\t\\\"event\\\": libcontainerdtypes.EventPaused,\\n\\t\\t\\t}).Error(\\\"failed to process event\\\")\\n\\t\\t}\\n\\t})\\n\\n\\treturn nil\\n}\",\n \"func (helper *testHelper) Pause(secs int) {}\",\n \"func (o *OMXPlayer) Pause() error {\\n\\tif o.status == statusPaused {\\n\\t\\treturn nil\\n\\t}\\n\\n\\t// TODO: test this properly by mocking dbus package.\\n\\tif o.testing {\\n\\t\\to.status = statusPaused\\n\\t\\treturn nil\\n\\t}\\n\\n\\t_, err := o.dbusSend(\\\"Pause\\\", dbus.FlagNoAutoStart)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\to.status = statusPaused\\n\\treturn nil\\n}\",\n \"func (c *Clock) AdvanceTo(t time.Time) {\\n\\tc.init()\\n\\trt := c.maybeGetRealTime()\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\tc.skipStep = true\\n\\tc.realTime = rt\\n\\tc.present = t\\n\\tc.events.AdvanceTo(c.present)\\n}\",\n \"func (s *Scheduler) Pause() error {\\n\\ts.mu.Lock()\\n\\tdefer s.mu.Unlock()\\n\\ts.paused = true\\n\\treturn nil\\n}\",\n \"func (c *Clock) Advance(d time.Duration) time.Time {\\n\\tc.init()\\n\\trt := c.maybeGetRealTime()\\n\\n\\tc.mu.Lock()\\n\\tdefer c.mu.Unlock()\\n\\tc.skipStep = true\\n\\n\\tc.advanceLocked(rt, d)\\n\\treturn c.present\\n}\",\n \"func (e *Enumeration) Pause() {\\n\\te.pause <- struct{}{}\\n}\",\n \"func (e *Enumeration) Pause() {\\n\\te.pause <- struct{}{}\\n}\",\n \"func (ft *fakeTimer) advance(d time.Duration) {\\n\\tft.lock.Lock()\\n\\tdefer ft.lock.Unlock()\\n\\n\\tft.now = ft.now.Add(d)\\n\\tif ft.active && !ft.now.Before(ft.timeout) {\\n\\t\\tft.active = false\\n\\t\\tft.c <- ft.timeout\\n\\t}\\n}\",\n \"func (timer *Timer) Pause() error {\\n\\tch := make(chan error)\\n\\ttimer.pause <- ch\\n\\treturn <-ch\\n}\",\n \"func (c *Switch) Pause() {\\n\\tc.lock.RLock()\\n\\tif cp, ok := c.subRenderables[c.curRenderable].(CanPause); ok {\\n\\t\\tcp.Pause()\\n\\t}\\n\\tc.lock.RUnlock()\\n}\",\n \"func (_Cakevault *CakevaultTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _Cakevault.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (w *Walker) Pause() {\\n\\tw.pauseRequested = true\\n}\",\n \"func (p *Pacer) Resume() {\\n\\t<-p.paused // clear flag to indicate paused\\n\\t<-p.pause // unblock this channel\\n}\",\n \"func (p *Playlist) Pause() {\\n\\tp.ch <- \\\"pause\\\"\\n}\",\n \"func (rep *Reporter) Continue() {\\n\\tlock.Lock()\\n\\tdefer lock.Unlock()\\n\\tif rep == nil {\\n\\t\\treturn\\n\\t}\\n\\n\\tif !rep.metricsCh.IsPaused() {\\n\\t\\treturn\\n\\t}\\n\\n\\trep.metricsCh.Continue()\\n}\",\n \"func AdvanceTo(newTime time.Time) {\\n\\tnow := Now()\\n\\tfound := false\\n\\tfor _, s := range sortedSchedulers() {\\n\\t\\tnextTick := s.tickAfter(now)\\n\\t\\tif nextTick.After(now) && !nextTick.After(newTime) {\\n\\t\\t\\tfound = true\\n\\t\\t\\tnowInTest.Store(nextTick)\\n\\t\\t\\ts.Lock()\\n\\t\\t\\tif s.paused {\\n\\t\\t\\t\\ts.fireOnResume = true\\n\\t\\t\\t} else {\\n\\t\\t\\t\\ts.notifyFn()\\n\\t\\t\\t}\\n\\t\\t\\ts.Unlock()\\n\\t\\t}\\n\\t}\\n\\tif !found {\\n\\t\\tnowInTest.Store(newTime)\\n\\t\\treturn\\n\\t}\\n\\tif Now() != newTime {\\n\\t\\tAdvanceTo(newTime)\\n\\t}\\n}\",\n \"func (j Jibi) Pause() {\\n\\tj.RunCommand(CmdPause, nil)\\n}\",\n \"func (c *StoppedClock) SetTime(time time.Time) {\\n\\tc.time = time\\n}\",\n \"func main() {\\n var working float64\\n var relax float64\\n var switchact bool = false\\n var switch2relax bool = true\\n var switch2working bool = false\\n\\n fmt.Println(\\\"Type any key for start the Workday\\\")\\n fmt.Println(\\\"The counter starts with relaxing type any key for start working\\\")\\n fmt.Println(\\\"You start workint at \\\" ,time.Now())\\n ch := make(chan string)\\n go func(ch chan string) {\\n // disable input buffering\\n exec.Command(\\\"stty\\\", \\\"-F\\\", \\\"/dev/tty\\\", \\\"cbreak\\\", \\\"min\\\", \\\"1\\\").Run()\\n // do not display entered characters on the screen\\n exec.Command(\\\"stty\\\", \\\"-F\\\", \\\"/dev/tty\\\", \\\"-echo\\\").Run()\\n var b []byte = make([]byte, 1)\\n //Routine for check for standard input with no wait for a key\\n for {\\n os.Stdin.Read(b)\\n ch <- string(b)\\n }\\n }(ch)\\n\\n //The intention with this fork was do a Thiker and with this controll the flow of time\\n for {\\n select {\\n \\n //In case of a signal by key lanch this case, change status of working to relax\\n //and show time elapsed\\n case stdin, _ := <-ch:\\n fmt.Println(\\\"\\\\n \\\\n Keys pressed:\\\", stdin)\\n if switch2relax == true {\\n fmt.Println(\\\"\\\\n Now you are relaxing\\\\n \\\\n\\\")\\n }else {\\n fmt.Println(\\\"\\\\n Now you are working \\\\n \\\\n\\\")\\n }\\n fmt.Printf(\\\"working for %f Hours \\\\n\\\", ((working/60.0)/60.0))\\n fmt.Printf(\\\"Relaxing for %f Hours \\\\n\\\", (relax/60.0)/60.0)\\n\\n switch2relax = !switch2relax\\n switch2working = !switch2working\\n switchact = true\\n \\n //By default the program is count time, if the routin is not counting time\\n //working count time relaxing, this script has no pause\\n default:\\n if switchact == true {\\n if switch2relax == true {\\n working=1.0+working\\n if ((working/60.0)/60.0) == 8.00000 {\\n exec.Command(\\\"mplayer\\\", \\\"stopwork.mp3\\\").Run()\\n }\\n //fmt.Println(\\\"Working for \\\", working)\\n }else{\\n relax=1.0+relax\\n //fmt.Println(\\\"relaxing for \\\", relax)\\n }\\n }\\n }\\n time.Sleep(time.Second * 1)\\n }\\n}\",\n \"func (_Cakevault *CakevaultSession) Pause() (*types.Transaction, error) {\\n\\treturn _Cakevault.Contract.Pause(&_Cakevault.TransactOpts)\\n}\",\n \"func (p *Profile) Pause() error {\\n\\tif !p.started {\\n\\t\\treturn errors.New(\\\"mxnet profile was not started\\\")\\n\\t}\\n\\tif p.stopped == true || p.paused == true {\\n\\t\\treturn nil\\n\\t}\\n\\tdefer func() {\\n\\t\\tp.paused = true\\n\\t}()\\n\\tp.lastPauseTime = time.Now()\\n\\tsuccess, err := C.MXProfilePause(C.int(1))\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tif success != 0 {\\n\\t\\treturn GetLastError()\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (m *Machine) Pause() error {\\n\\tm.State = driver.Paused\\n\\tfmt.Printf(\\\"Pause %s: %s\\\\n\\\", m.Name, m.State)\\n\\treturn nil\\n}\",\n \"func (t Track) Pause(bool) {\\n\\tpanic(\\\"implement me\\\")\\n}\",\n \"func (_Cakevault *CakevaultTransactorSession) Pause() (*types.Transaction, error) {\\n\\treturn _Cakevault.Contract.Pause(&_Cakevault.TransactOpts)\\n}\",\n \"func (c *chrono) Now() time.Time {\\n\\tvar now time.Time\\n\\tif c.paused {\\n\\t\\tnow = c.pausedAt\\n\\t} else {\\n\\t\\tnow = time.Now()\\n\\t}\\n\\treturn now.Add(c.skew)\\n}\",\n \"func (s *SlaveHealthChecker) Pause() {\\n\\ts.Lock()\\n\\tdefer s.Unlock()\\n\\ts.paused = true\\n}\",\n \"func (_OracleMgr *OracleMgrTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _OracleMgr.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (_ERC20Pausable *ERC20PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _ERC20Pausable.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func Pause(msg string) {\\n\\tfmt.Print(msg, \\\"[Press Enter to Continue]: \\\")\\n\\tvar s string\\n\\t_, err := fmt.Scan(&s)\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n}\",\n \"func (n *BaseNode) Continue() {\\n\\tn.continueFunc(func() {\\n\\t\\tif n.cancelPause != nil {\\n\\t\\t\\tn.cancelPause()\\n\\t\\t\\tn.cancelPause = nil\\n\\t\\t}\\n\\t})\\n}\",\n \"func (d *Device) Pause() {\\n\\tdefer d.mu.Unlock()\\n\\td.mu.Lock()\\n\\td.paused = true\\n}\",\n \"func (_DelegateProfile *DelegateProfileTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _DelegateProfile.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func UpdateTime(dt int) {\\n\\tcurrenttime = currenttime + dt\\n\\tif currenttime >= 2400 {\\n\\t\\tcurrenttime = currenttime - 2400\\n\\t\\tdaycount = daycount + 1\\n\\t}\\n}\",\n \"func (a *Agent) PauseSync() {\\n\\t// Do this outside of lock as it has it's own locking\\n\\ta.sync.Pause()\\n\\n\\t// Coordinate local state watchers\\n\\ta.syncMu.Lock()\\n\\tdefer a.syncMu.Unlock()\\n\\tif a.syncCh == nil {\\n\\t\\ta.syncCh = make(chan struct{})\\n\\t}\\n}\",\n \"func (c *MockClock) Set(t time.Time) {\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\tif t.After(c.now) {\\n\\t\\tc.advanceTo(t)\\n\\t} else {\\n\\t\\tc.now = t // move back in time\\n\\t}\\n}\",\n \"func (_MonsterAccessControl *MonsterAccessControlTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _MonsterAccessControl.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (p *Proxy) Pause() {\\n\\tp.configMutex.Lock()\\n\\tif p.state == proxyActive {\\n\\t\\tp.state = proxyPaused\\n\\t\\tp.connMutex.Unlock()\\n\\t}\\n\\tp.configMutex.Unlock()\\n}\",\n \"func (s *ExperimentSpec) Pause() bool {\\n\\tif s.ManualOverride != nil && s.ManualOverride.Action == ActionPause {\\n\\t\\treturn true\\n\\t}\\n\\treturn false\\n}\",\n \"func (c *MockClock) advanceToNextWakeup() {\\n\\tif len(c.wakeups) < 1 {\\n\\t\\treturn\\n\\t}\\n\\tw := heap.Pop(&c.wakeups).(*wakeup)\\n\\tc.logf(\\\"MockClock: Advancing time from now=%s to next wakeup time %s.\\\",\\n\\t\\ttsStr(c.now), tsStr(w.t))\\n\\tc.now = w.t\\n\\tselect {\\n\\tcase w.c <- w.t:\\n\\tdefault:\\n\\t}\\n\\t// give things a chance to run\\n\\truntime.Gosched()\\n\\tc.logf(\\\"MockClock: Advanced time, now=%s.\\\", tsStr(c.now))\\n}\",\n \"func (c *MockClock) advanceTo(end time.Time) {\\n\\tfor {\\n\\t\\tt := c.nextWakeupTime()\\n\\t\\tif t == (time.Time{}) || t.After(end) {\\n\\t\\t\\tbreak\\n\\t\\t}\\n\\t\\tc.advanceToNextWakeup()\\n\\t}\\n\\tc.now = end\\n}\",\n \"func SetTime(ct int) {\\n\\tcurrenttime = ct\\n}\",\n \"func (cg *CgroupImpl) Pause() error {\\n\\treturn cg.setFreezeState(1)\\n}\",\n \"func (l *LightningLoader) Pause() {\\n\\tl.Lock()\\n\\tdefer l.Unlock()\\n\\tif l.isClosed() {\\n\\t\\tl.logger.Warn(\\\"try to pause, but already closed\\\")\\n\\t\\treturn\\n\\t}\\n\\tif l.cancel != nil {\\n\\t\\tl.cancel()\\n\\t}\\n\\tl.core.Stop()\\n}\",\n \"func AnounanceTime(){\\n\\tfor {\\n\\t\\tfmt.Println(time.Now())\\n\\t\\ttime.Sleep(5*time.Second)\\n\\t}\\n}\",\n \"func (_Token *TokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _Token.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (_ChpRegistry *ChpRegistryTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _ChpRegistry.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (_Cakevault *CakevaultCaller) Paused(opts *bind.CallOpts) (bool, error) {\\n\\tvar out []interface{}\\n\\terr := _Cakevault.contract.Call(opts, &out, \\\"paused\\\")\\n\\n\\tif err != nil {\\n\\t\\treturn *new(bool), err\\n\\t}\\n\\n\\tout0 := *abi.ConvertType(out[0], new(bool)).(*bool)\\n\\n\\treturn out0, err\\n\\n}\",\n \"func (_ElvToken *ElvTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _ElvToken.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (t Time) Clock() (hour, min, sec int) {}\",\n \"func (_MonsterOwnership *MonsterOwnershipTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _MonsterOwnership.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (c *Compound) Pause() {\\n\\tc.lock.RLock()\\n\\tif cp, ok := c.subRenderables[c.curRenderable].(CanPause); ok {\\n\\t\\tcp.Pause()\\n\\t}\\n\\tc.lock.RUnlock()\\n}\",\n \"func Pause() {\\n\\tNamedLoggers.Range(func(key, value interface{}) bool {\\n\\t\\tlogger := value.(*Logger)\\n\\t\\tlogger.Pause()\\n\\t\\tlogger.Flush()\\n\\t\\treturn true\\n\\t})\\n}\",\n \"func (d *Debugger) Continue() {\\n\\tif d.BreakpointActive {\\n\\t\\td.Cont <- true\\n\\t}\\n}\",\n \"func (_LifToken *LifTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _LifToken.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (_Pausable *PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _Pausable.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (_Pausable *PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _Pausable.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (c *MockClock) Advance(delta time.Duration) {\\n\\tc.mutex.Lock()\\n\\tdefer c.mutex.Unlock()\\n\\n\\tend := c.now.Add(delta)\\n\\tc.advanceTo(end)\\n}\",\n \"func (cr *Playback) Pause() gin.HandlerFunc {\\n\\treturn func(c *gin.Context) {\\n\\t\\tcurrentParty, err := session.CurrentParty(c)\\n\\t\\tif err != nil {\\n\\t\\t\\tif err != sql.ErrNoRows {\\n\\t\\t\\t\\tc.Error(e.Internal.CausedBy(err))\\n\\t\\t\\t\\tc.Abort()\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\n\\t\\t\\tc.Error(e.NotFound.WithMessage(\\\"Host has not started a party\\\"))\\n\\t\\t\\tc.Abort()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\t// It is a bad request to pause when no track is playing\\n\\t\\tif !currentParty.CurrentTrack.Valid {\\n\\t\\t\\tc.Error(e.BadRequest.WithDetail(\\\"Host is not playing music\\\"))\\n\\t\\t\\tc.Abort()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\t// Pause the current track\\n\\t\\tvar currentTrack models.PlayingTrack\\n\\t\\terr = currentParty.CurrentTrack.Unmarshal(currentTrack)\\n\\t\\tif err != nil {\\n\\t\\t\\tc.Error(e.Internal.CausedBy(err))\\n\\t\\t\\tc.Abort()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\telapsedParam, exists := c.GetQuery(\\\"elapsed\\\")\\n\\t\\tif !exists {\\n\\t\\t\\tc.Error(e.BadRequest.WithDetail(\\\"Must provide elapsed duration to pause playback\\\"))\\n\\t\\t\\tc.Abort()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\t\\telapsed, err := strconv.ParseUint(elapsedParam, 10, 64)\\n\\t\\tif err != nil {\\n\\t\\t\\tc.Error(\\n\\t\\t\\t\\te.BadRequest.\\n\\t\\t\\t\\t\\tWithDetail(\\\"Must provide valid elapsed duration to pause playback\\\").\\n\\t\\t\\t\\t\\tCausedBy(err),\\n\\t\\t\\t)\\n\\t\\t\\tc.Abort()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\t// It is a bad request to pause a track that is already paused\\n\\t\\tif currentTrack.Paused {\\n\\t\\t\\tc.Error(e.BadRequest.WithDetail(\\\"Current track is already paused\\\"))\\n\\t\\t\\tc.Abort()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\tcurrentTrack.Paused = true\\n\\t\\tcurrentTrack.Elapsed = uint(elapsed)\\n\\t\\terr = cr.updateTrackAndBroadcast(currentParty, &currentTrack)\\n\\t\\tif err != nil {\\n\\t\\t\\tc.Error(e.Internal.CausedBy(err))\\n\\t\\t\\tc.Abort()\\n\\t\\t\\treturn\\n\\t\\t}\\n\\n\\t\\tc.JSON(http.StatusOK, models.EmptyRespose)\\n\\t}\\n}\",\n \"func (_MonsterAccessControl *MonsterAccessControlTransactorSession) Pause() (*types.Transaction, error) {\\n\\treturn _MonsterAccessControl.Contract.Pause(&_MonsterAccessControl.TransactOpts)\\n}\",\n \"func (t *Time) Sleep(s time.Duration) {\\n\\tt.current = t.current.Add(s)\\n}\",\n \"func main() {\\n\\tp := fmt.Println\\n\\t\\n\\tp(\\\"------- Alarm Clock - Main initialized -------\\\")\\n\\tvar wg sync.WaitGroup\\n\\t\\n\\n\\t//start := time.Now()\\n\\tcurrent := time.Now()\\n\\tfmt.Print(\\\"Klockan är: \\\", current.Hour(),\\\":\\\", current.Minute(),\\\".\\\",current.Second(),\\\"\\\\n\\\")\\n\\tp(\\\"---------------------------------\\\")\\n\\t//diff := current.Sub(start)\\n\\t//fmt.Println(\\\"DIFF:\\\",diff)\\n\\tp(\\\"START TIME:\\\",current)\\n\\t\\n\\t//add one to waitgroup object for each new goroutine started\\n\\twg.Add(1)\\n\\tgo Remind(\\\"Dags att äta\\\",3*time.Second)\\n\\twg.Add(1)\\n\\tgo Remind(\\\"Dags att arbeta\\\",8*time.Second)\\n\\twg.Add(1)\\n\\tgo Remind(\\\"Dags att sova\\\",24*time.Second)\\n\\t\\n\\twg.Wait()\\n\\t//p(currTime.Format(\\\"15:04:05 (2006-01-02)\\\"))\\n\\t//p(\\\"TIME DIFF: \\\",currTime.Sub(startTime))\\n\\t\\n\\n}\",\n \"func (_DelegateProfile *DelegateProfileTransactorSession) Pause() (*types.Transaction, error) {\\n\\treturn _DelegateProfile.Contract.Pause(&_DelegateProfile.TransactOpts)\\n}\",\n \"func (c *Client) Pause() error {\\n\\treturn nil\\n}\",\n \"func TestProgressResume(t *testing.T) {\\n\\tp := &Progress{\\n\\t\\tNext: 2,\\n\\t\\tPaused: true,\\n\\t}\\n\\tp.maybeDecrTo(1, 1)\\n\\tif p.Paused {\\n\\t\\tt.Errorf(\\\"paused= %v, want false\\\", p.Paused)\\n\\t}\\n\\tp.Paused = true\\n\\tp.maybeUpdate(2)\\n\\tif p.Paused {\\n\\t\\tt.Errorf(\\\"paused= %v, want false\\\", p.Paused)\\n\\t}\\n}\",\n \"func (n *BaseNode) Pause() {\\n\\tn.pauseFunc(func() {\\n\\t\\tn.ctxPause, n.cancelPause = context.WithCancel(n.ctx)\\n\\t})\\n}\",\n \"func (container *container) Pause() error {\\r\\n\\treturn convertSystemError(container.system.Pause(context.Background()), container)\\r\\n}\",\n \"func NextPause() func() (pause time.Duration) {\\n rando := rand.New(rand.NewSource(time.Now().UnixNano()))\\n return func() (pause time.Duration) {\\n pause = time.Duration((rando.Intn(500)) * int(time.Millisecond))\\n return\\n }\\n}\",\n \"func (_PausableToken *PausableTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _PausableToken.contract.Transact(opts, \\\"pause\\\")\\n}\",\n \"func (o RunnerOutput) Paused() pulumi.BoolOutput {\\n\\treturn o.ApplyT(func(v *Runner) pulumi.BoolOutput { return v.Paused }).(pulumi.BoolOutput)\\n}\",\n \"func (t *SpotifyDevice) Pause() {\\n\\tif t.client == nil {\\n\\t\\treturn\\n\\t}\\n\\tlog.Println(\\\"Pausing Spotify device\\\")\\n\\tif err := t.client.PauseOpt(t.playOpts); err != nil {\\n\\t\\tlog.Println(err)\\n\\t}\\n}\",\n \"func (_MonsterAccessControl *MonsterAccessControlSession) Pause() (*types.Transaction, error) {\\n\\treturn _MonsterAccessControl.Contract.Pause(&_MonsterAccessControl.TransactOpts)\\n}\",\n \"func (st *LevelCompleteState) OnPause(world w.World) {}\",\n \"func (s *SlaveHealthChecker) Continue(slaveUPID *upid.UPID) {\\n\\ts.Lock()\\n\\tdefer s.Unlock()\\n\\ts.paused = false\\n\\ts.slaveUPID = slaveUPID\\n}\",\n \"func (_DelegateProfile *DelegateProfileSession) Pause() (*types.Transaction, error) {\\n\\treturn _DelegateProfile.Contract.Pause(&_DelegateProfile.TransactOpts)\\n}\",\n \"func (fgs *FakeGraphSync) Pause(ctx context.Context, requestID graphsync.RequestID) error {\\n\\tfgs.pauses <- requestID\\n\\treturn fgs.ReturnedPauseError\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.67877674","0.653856","0.6279754","0.6165107","0.6063943","0.6050643","0.6011128","0.6008339","0.5997136","0.59503627","0.59455216","0.5916581","0.58829933","0.5872751","0.5872751","0.58583057","0.57535905","0.57200044","0.56405187","0.5613111","0.5613111","0.56104696","0.56087905","0.55646914","0.5563314","0.550896","0.55066174","0.54984736","0.5495456","0.5481841","0.5450739","0.5450739","0.5447858","0.54471993","0.54377466","0.5433592","0.5430805","0.53919566","0.53906846","0.5390298","0.53871286","0.53265625","0.53138626","0.52900684","0.5288507","0.52870095","0.5281264","0.52765656","0.5261484","0.52584726","0.5239021","0.5234105","0.5219808","0.5209068","0.52085245","0.51810974","0.51807606","0.51799476","0.5178544","0.51756287","0.5173574","0.5166772","0.5154566","0.51509416","0.5147486","0.51215285","0.510734","0.51034176","0.5097587","0.50900215","0.50880927","0.5082865","0.50586","0.505711","0.5044188","0.5040813","0.50352293","0.50329983","0.5029847","0.50236815","0.50236815","0.5004772","0.49994218","0.49925694","0.49838725","0.49754855","0.49745432","0.49652076","0.49650443","0.4963122","0.49604392","0.49596387","0.49585953","0.49580398","0.49558106","0.49533692","0.49461812","0.49450165","0.49448425","0.49348"],"string":"[\n \"0.67877674\",\n \"0.653856\",\n \"0.6279754\",\n \"0.6165107\",\n \"0.6063943\",\n \"0.6050643\",\n \"0.6011128\",\n \"0.6008339\",\n \"0.5997136\",\n \"0.59503627\",\n \"0.59455216\",\n \"0.5916581\",\n \"0.58829933\",\n \"0.5872751\",\n \"0.5872751\",\n \"0.58583057\",\n \"0.57535905\",\n \"0.57200044\",\n \"0.56405187\",\n \"0.5613111\",\n \"0.5613111\",\n \"0.56104696\",\n \"0.56087905\",\n \"0.55646914\",\n \"0.5563314\",\n \"0.550896\",\n \"0.55066174\",\n \"0.54984736\",\n \"0.5495456\",\n \"0.5481841\",\n \"0.5450739\",\n \"0.5450739\",\n \"0.5447858\",\n \"0.54471993\",\n \"0.54377466\",\n \"0.5433592\",\n \"0.5430805\",\n \"0.53919566\",\n \"0.53906846\",\n \"0.5390298\",\n \"0.53871286\",\n \"0.53265625\",\n \"0.53138626\",\n \"0.52900684\",\n \"0.5288507\",\n \"0.52870095\",\n \"0.5281264\",\n \"0.52765656\",\n \"0.5261484\",\n \"0.52584726\",\n \"0.5239021\",\n \"0.5234105\",\n \"0.5219808\",\n \"0.5209068\",\n \"0.52085245\",\n \"0.51810974\",\n \"0.51807606\",\n \"0.51799476\",\n \"0.5178544\",\n \"0.51756287\",\n \"0.5173574\",\n \"0.5166772\",\n \"0.5154566\",\n \"0.51509416\",\n \"0.5147486\",\n \"0.51215285\",\n \"0.510734\",\n \"0.51034176\",\n \"0.5097587\",\n \"0.50900215\",\n \"0.50880927\",\n \"0.5082865\",\n \"0.50586\",\n \"0.505711\",\n \"0.5044188\",\n \"0.5040813\",\n \"0.50352293\",\n \"0.50329983\",\n \"0.5029847\",\n \"0.50236815\",\n \"0.50236815\",\n \"0.5004772\",\n \"0.49994218\",\n \"0.49925694\",\n \"0.49838725\",\n \"0.49754855\",\n \"0.49745432\",\n \"0.49652076\",\n \"0.49650443\",\n \"0.4963122\",\n \"0.49604392\",\n \"0.49596387\",\n \"0.49585953\",\n \"0.49580398\",\n \"0.49558106\",\n \"0.49533692\",\n \"0.49461812\",\n \"0.49450165\",\n \"0.49448425\",\n \"0.49348\"\n]"},"document_score":{"kind":"string","value":"0.7427754"},"document_rank":{"kind":"string","value":"0"}}},{"rowIdx":104797,"cells":{"query":{"kind":"string","value":"ValidConfig will return true if there are values set for each Property of the Azure config object"},"document":{"kind":"string","value":"func (a *Azure) ValidConfig() bool {\n\treturn a.SubscriptionID != \"\" && a.ClientID != \"\" && a.ClientSecret != \"\" && a.TenantID != \"\"\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (g *Google) ValidConfig() bool {\n\treturn g.AccountFileJSON != \"\" && g.ProjectID != \"\" && g.Zone != \"\"\n}","func (c *Config) Valid() error {\n\tif len(c.Servers) == 0 {\n\t\treturn fmt.Errorf(\"No servers in config\")\n\t}\n\tfor _, s := range c.Servers {\n\t\terr := s.Valid()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}","func (c *Config) IsValid() bool {\n\tif len(c.GithubAccessToken) != 40 {\n\t\treturn false\n\t}\n\tif len(c.GitName) == 0 {\n\t\treturn false\n\t}\n\tif len(c.GitEmail) == 0 {\n\t\treturn false\n\t}\n\n\treturn true\n}","func (config *Config) IsValid() bool {\n\tif config.Basic.Enabled && config.OAuth2.Enabled {\n\t\treturn false\n\t}\n\n\tif config.Basic.Enabled && config.X509.Enabled {\n\t\treturn false\n\t}\n\n\tif config.OAuth2.Enabled && config.X509.Enabled {\n\t\treturn false\n\t}\n\n\tif !config.Basic.Enabled && !config.OAuth2.Enabled && !config.X509.Enabled {\n\t\treturn false\n\t}\n\n\treturn true\n}","func (sc *StartupConfig) Validate() (bool, error) {\n\tif sc.Meta == nil {\n\t\treturn false, fmt.Errorf(\"meta object is nil\")\n\t}\n\n\tif sc.Server == nil {\n\t\t//return false, fmt.Errorf(\"Server is nil\")\n\t}\n\n\tif sc.Database == nil {\n\t\treturn false, fmt.Errorf(\"database object is nil\")\n\t}\n\n\tif sc.Session == nil {\n\t\treturn false, fmt.Errorf(\"session object is nil\")\n\t}\n\n\tif sc.Crypto == nil {\n\t\treturn false, fmt.Errorf(\"crypto object is nil\")\n\t}\n\n\tif sc.Secrets == nil {\n\t\treturn false, fmt.Errorf(\"secrets object is nil\")\n\t}\n\n\tif sc.Capsul == nil {\n\t\treturn false, fmt.Errorf(\"capsul object is nil\")\n\t}\n\n\tif sc.CustomCapsul == nil {\n\t\treturn false, fmt.Errorf(\"custom capsul object is nil\")\n\t}\n\n\tif sc.Create == nil {\n\t\treturn false, fmt.Errorf(\"create object is nil\")\n\t}\n\n\treturn true, nil\n}","func (o *Config) IsValid() *AppError {\n\tif err := o.ServiceSettings.isValid(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}","func (c *configuration) validateConfig() error {\n\tif c.Provider == \"\" {\n\t\treturn errors.New(msgConfigNoProvider)\n\t}\n\n\tif len(c.Servers) == 0 {\n\t\treturn errors.New(msgConfigNoServers)\n\t}\n\n\tfor i, srv := range c.Servers {\n\t\tif srv.FabricIface == \"\" {\n\t\t\treturn errors.Errorf(\n\t\t\t\tmsgConfigServerNoIface+\" for I/O service %d\", i)\n\t\t}\n\t}\n\n\treturn nil\n}","func (c Config) IsValid() bool {\n\tif c.Port == \"\" || c.DSN == \"\" {\n\t\treturn false\n\t}\n\treturn true\n}","func (c *Config) IsValid() error {\n\tif len(c.NetworkServices) == 0 {\n\t\treturn errors.New(\"no network services are specified\")\n\t}\n\tif c.Name == \"\" {\n\t\treturn errors.New(\"no cleint name specified\")\n\t}\n\tif c.ConnectTo.String() == \"\" {\n\t\treturn errors.New(\"no NSMGr ConnectTO URL are sepecified\")\n\t}\n\treturn nil\n}","func (c *Config) IsValid() bool {\n\tif c.DiffType == \"schema\" || c.DiffType == \"data\" {\n\t\treturn c.Source != \"\" && c.Target != \"\"\n\t} else if c.DiffType == \"md\" || c.DiffType == \"wiki\" || c.DiffType == \"sql\" {\n\t\treturn c.Source != \"\"\n\t}\n\treturn false\n}","func (c Config) Validate() error {\n\tif len(c.Project) == 0 {\n\t\treturn errors.New(\"missing project in configuration\")\n\t}\n\tif len(c.Bucket) == 0 {\n\t\treturn errors.New(\"missing bucket in configuration\")\n\t}\n\tif len(c.LastMigrationObjectName) == 0 {\n\t\treturn errors.New(\"missing state name in configuration\")\n\t}\n\treturn nil\n}","func (c Config) IsValid() bool {\n\treturn c.RandomTemperature > 0 && c.NumSimulation > 0\n}","func (c *Config) Validate() (err error) {\n\tcon := *c\n\tfor i, v := range con {\n\t\tk, err := registry.OpenKey(v.GetScope(), v.Path, registry.ALL_ACCESS)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer k.Close()\n\t\tfor n, p := range v.Properties {\n\t\t\tval := &con[i].Properties[n].PrevValue\n\t\t\tswitch p.Type {\n\t\t\tcase \"DWord\", \"QWord\":\n\t\t\t\tif s, _, err := k.GetIntegerValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tcase \"String\":\n\t\t\t\tif s, _, err := k.GetStringValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tcase \"Strings\":\n\t\t\t\tif s, _, err := k.GetStringsValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tcase \"Binary\":\n\t\t\t\tif s, _, err := k.GetBinaryValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tdefault:\n\t\t\t\tvar buf []byte\n\t\t\t\tif _, _, err := k.GetValue(p.Name, buf); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"%s of %s path in %s scope returned code %d.\") // TODO: Convert const int representation of value types to explicitly match what the user should type into their JSON config.\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}","func (c AuthConfig) Validate() []error {\n\tvar errs []error\n\n\tif len(c.JwksURI) == 0 {\n\t\terrs = append(errs, errors.Errorf(\"AuthConfig requires a non-empty JwksURI config value\"))\n\t}\n\n\treturn errs\n}","func ValidateConfig(configMap *corev1.ConfigMap) (*kle.Config, error) {\n\tconfig, err := kle.NewConfigFromMap(configMap.Data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor component := range config.EnabledComponents {\n\t\tif !validComponents.Has(component) {\n\t\t\treturn nil, fmt.Errorf(\"invalid enabledComponent %q: valid values are %q\", component, validComponents.List())\n\t\t}\n\t}\n\n\treturn config, nil\n}","func (c *Config) IsValid() bool {\n\t_, latErr := strconv.ParseFloat(c.Latitude, 10)\n\t_, lonErr := strconv.ParseFloat(c.Longitude, 10)\n\treturn c.APIKey != \"\" && c.BaseUrl != \"\" && latErr == nil && lonErr == nil\n}","func (c *config) Validate() []error {\n\tvar errs []error\n\tif value, err := validateProvider(c.Provider); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.Provider = value\n\t}\n\tif value, err := validateAccountName(c.AccountName); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.AccountName = value\n\t}\n\tif value, err := validateAccountSecret(c.AccountSecret); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.AccountSecret = value\n\t}\n\tif value, err := validateDNSContent(c.DNSContent); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.DNSContent = value\n\t}\n\tif value, err := validateDockerLabel(c.DockerLabel); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.DockerLabel = value\n\t}\n\tif value, err := validateStore(c.Store); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.Store = value\n\t}\n\tif value, err := validateDataDirectory(c.DataDirectory); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.DataDirectory = value\n\t}\n\treturn errs\n}","func (m *Manager) IsDBConfigValid(config config.DatabaseConfigs) error {\n\tm.lock.RLock()\n\tdefer m.lock.RUnlock()\n\n\t// Only count the length of enabled databases\n\tvar length int\n\tfor _, c := range config {\n\t\tif c.Enabled {\n\t\t\tlength++\n\t\t}\n\t}\n\n\treturn nil\n}","func (sc *StoreConfig) Valid() bool {\n\treturn sc.Clock != nil && sc.Transport != nil &&\n\t\tsc.RaftTickInterval != 0 && sc.RaftHeartbeatIntervalTicks > 0 &&\n\t\tsc.RaftElectionTimeoutTicks > 0 && sc.ScanInterval >= 0 &&\n\t\tsc.AmbientCtx.Tracer != nil\n}","func (c *Config) Validate() error {\n\tvar errs error\n\tif c.Tenant.RemoteServiceAPI == \"\" {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.remote_service_api is required\"))\n\t}\n\tif c.Tenant.InternalAPI == \"\" && c.Analytics.FluentdEndpoint == \"\" && c.Analytics.CredentialsJSON == nil {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.internal_api or tenant.analytics.fluentd_endpoint is required if no service account\"))\n\t}\n\tif c.Tenant.OrgName == \"\" {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.org_name is required\"))\n\t}\n\tif c.Tenant.EnvName == \"\" {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.env_name is required\"))\n\t}\n\tif (c.Global.TLS.CertFile != \"\" || c.Global.TLS.KeyFile != \"\") &&\n\t\t(c.Global.TLS.CertFile == \"\" || c.Global.TLS.KeyFile == \"\") {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"global.tls.cert_file and global.tls.key_file are both required if either are present\"))\n\t}\n\tif (c.Analytics.TLS.CAFile != \"\" || c.Analytics.TLS.CertFile != \"\" || c.Analytics.TLS.KeyFile != \"\") &&\n\t\t(c.Analytics.TLS.CAFile == \"\" || c.Analytics.TLS.CertFile == \"\" || c.Analytics.TLS.KeyFile == \"\") {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"all analytics.tls options are required if any are present\"))\n\t}\n\treturn errs\n}","func (c Config) Validate() error {\n\t// Check DbConfig\n\tif err := c.Db.Validate(); err != nil {\n\t\treturn fmt.Errorf(\"error verifying db config: %s\", err.Error())\n\t}\n\n\t// Check AuthConfig\n\tif err := c.Auth.Validate(); err != nil {\n\t\treturn fmt.Errorf(\"error verifying auth config: %s\", err.Error())\n\t}\n\n\t// All good\n\treturn nil\n}","func (c *Config) Validate() error {\n\n\tif err := c.Data.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\t//if err := c.HintedHandoff.Validate(); err != nil {\n\t//\treturn err\n\t//}\n\tfor _, graphite := range c.GraphiteInputs {\n\t\tif err := graphite.Validate(); err != nil {\n\t\t\treturn fmt.Errorf(\"invalid graphite config: %v\", err)\n\t\t}\n\t}\n\n\tif err := c.Monitor.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.ContinuousQuery.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.Retention.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.Precreator.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.Subscriber.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, collectd := range c.CollectdInputs {\n\t\tif err := collectd.Validate(); err != nil {\n\t\t\treturn fmt.Errorf(\"invalid collectd config: %v\", err)\n\t\t}\n\t}\n\n\tif err := c.TLS.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}","func IsConfigPresent(resources []ResourceConfig, resourceConfig ResourceConfig) bool {\n\tfor _, resource := range resources {\n\t\tif resource.Name == resourceConfig.Name && resource.Source == resourceConfig.Source {\n\t\t\tif reflect.DeepEqual(resource.Config, resourceConfig.Config) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}","func (config *Config) Validate() error {\n\tif len(config.NsqLookupdAddress) == 0 {\n\t\treturn fmt.Errorf(\"parameter NsqLookupdAddress missing\")\n\t}\n\tif len(config.NsqdAddress) == 0 {\n\t\treturn fmt.Errorf(\"parameter NsqdAddress missing\")\n\t}\n\tif len(config.BotName) == 0 {\n\t\treturn fmt.Errorf(\"parameter BotName missing\")\n\t}\n\tif len(config.BambooUrl) == 0 {\n\t\treturn fmt.Errorf(\"parameter BambooUrl missing\")\n\t}\n\tif len(config.BambooUsername) == 0 {\n\t\treturn fmt.Errorf(\"parameter BambooUsername missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter AuthUrl missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter AuthApplicationName missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter AuthApplicationPassword missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter BambooPassword missing\")\n\t}\n\tif len(config.Prefix) == 0 {\n\t\treturn fmt.Errorf(\"parameter Prefix missing\")\n\t}\n\treturn nil\n}","func (o *OAuth2Config) IsValid() bool {\n\treturn o.TokenURL != \"\" && o.AuthURL != \"\" && len(o.Scopes) != 0\n}","func validateConfig(config ProvisionerConfig) error {\n\tswitch config.CloudProvider {\n\tcase constants.AWS, constants.Azure, constants.GCE, constants.Ops:\n\tdefault:\n\t\treturn trace.BadParameter(\"unknown cloud provider %s\", config.CloudProvider)\n\t}\n\n\terr := validator.New().Struct(&config)\n\tif err == nil {\n\t\treturn nil\n\t}\n\n\tvar errs []error\n\tif validationErrors, ok := err.(validator.ValidationErrors); ok {\n\t\tfor _, fieldError := range validationErrors {\n\t\t\terrs = append(errs,\n\t\t\t\ttrace.BadParameter(` * %s=\"%v\" fails \"%s\"`,\n\t\t\t\t\tfieldError.Field(), fieldError.Value(), fieldError.Tag()))\n\t\t}\n\t}\n\treturn trace.NewAggregate(errs...)\n}","func (_ runtimeConfigValidator) ValidateConfig(config *stiapi.Config) []validation.ValidationError {\n\treturn validation.ValidateConfig(config)\n}","func (cfg *Config) Validate() error {\n\tchecks := []struct {\n\t\tbad bool\n\t\terrMsg string\n\t}{\n\t\t{cfg.AuthorizeURL == \"\", \"no authorizeURL specified\"},\n\t\t{cfg.TokenURL == \"\", \"no tokenURL specified\"},\n\t\t{cfg.ClientID == \"\", \"no clientID specified\"},\n\t\t{cfg.ClientSecret == \"\" && !cfg.AllowEmptyClientSecret, \"no clientSecret specified\"},\n\t\t{cfg.RedirectURL == \"\", \"no redirectURL specified\"},\n\t\t{cfg.SessionSecurityKey == \"\", \"no SessionSecurityKey specified\"},\n\t\t{cfg.APIServerURL == \"\", \"no apiServerURL specified\"},\n\t}\n\n\tfor _, check := range checks {\n\t\tif check.bad {\n\t\t\treturn fmt.Errorf(\"invalid config: %s\", check.errMsg)\n\t\t}\n\t}\n\treturn nil\n}","func (c Config) Validate() error {\n\treturn validation.ValidateStruct(&c,\n\t\tvalidation.Field(&c.AppMode, validation.Required),\n\t\tvalidation.Field(&c.AppName, validation.Required),\n\t\tvalidation.Field(&c.DBType, validation.Required),\n\t\tvalidation.Field(&c.DSN, validation.Required),\n\t\tvalidation.Field(&c.JWTSigningKey, validation.Required),\n\t\tvalidation.Field(&c.JWTExpiration, validation.Required),\n\t\tvalidation.Field(&c.MailSmtphost, validation.Required),\n\t\tvalidation.Field(&c.MailSmtpport, validation.Required),\n\t\tvalidation.Field(&c.MailUsername, validation.Required),\n\t\tvalidation.Field(&c.MailPassword, validation.Required),\n\t\tvalidation.Field(&c.AppFqdn, validation.Required),\n\t\tvalidation.Field(&c.HttpEntrypoint, validation.Required),\n\t\tvalidation.Field(&c.WebservName, validation.Required),\n\t\tvalidation.Field(&c.GoogleCredentialFile, validation.Required),\n\t\tvalidation.Field(&c.GoogleRedirectPath, validation.Required),\n\t\tvalidation.Field(&c.AppSecretKey, validation.Required),\n\t\tvalidation.Field(&c.BizName, validation.Required),\n\t\tvalidation.Field(&c.BizShortname, validation.Required),\n\t\tvalidation.Field(&c.BizEmail, validation.Required),\n\t\tvalidation.Field(&c.BizPhone),\n\t\tvalidation.Field(&c.BizPhone2),\n\t\tvalidation.Field(&c.BizLogo, validation.Required),\n\t)\n}","func (c Config) Validate() (err error) {\n\tvar fi os.FileInfo\n\n\t// validate key fingerprint\n\t_, err = HexStringToFingerprint(c.MasterKeyFingerprint)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// validate TLSCert\n\tif len(c.TLSCert) == 0 {\n\t\treturn errors.New(\"Missing config param: TLSCert\")\n\t}\n\tfi, err = os.Stat(c.TLSCert)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in TLSCert '%s': %s\", c.TLSCert, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in TLSCert '%s': expected file path, got directory\", c.TLSCert)\n\t}\n\n\t// validate TLSKey\n\tif len(c.TLSKey) == 0 {\n\t\treturn errors.New(\"Missing config param: TLSKey\")\n\t}\n\tfi, err = os.Stat(c.TLSKey)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in TLSKey '%s': %s\", c.TLSKey, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in TLSKey '%s': expected file path, got directory\", c.TLSKey)\n\t}\n\n\t// validate SecRing\n\tif len(c.SecRing) == 0 {\n\t\treturn errors.New(\"Missing config param: SecRing\")\n\t}\n\tfi, err = os.Stat(c.SecRing)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in SecRing '%s': %s\", c.SecRing, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in SecRing '%s': expected file path, got directory\", c.SecRing)\n\t}\n\n\t// validate ProdSupportPubRing\n\tif len(c.ProdSupportPubRing) == 0 {\n\t\treturn errors.New(\"Missing config param: ProdSupportPubRing\")\n\t}\n\tfi, err = os.Stat(c.ProdSupportPubRing)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in ProdSupportPubRing '%s': %s\", c.ProdSupportPubRing, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in ProdSupportPubRing '%s': expected file path, got directory\", c.ProdSupportPubRing)\n\t}\n\n\t// validate DataRoot\n\tif len(c.DataRoot) == 0 {\n\t\treturn errors.New(\"Missing config param: DataRoot\")\n\t}\n\tfi, err = os.Stat(c.DataRoot)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.DataRoot, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in DataRoot '%s': %s\", c.DataRoot, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in DataRoot '%s': expected directory, got file path\", c.DataRoot)\n\t\t}\n\t}\n\n\t// validate ProdSupportDir\n\tif len(c.ProdSupportDir) == 0 {\n\t\treturn errors.New(\"Missing config param: ProdSupportDir\")\n\t}\n\tfi, err = os.Stat(c.ProdSupportDir)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.ProdSupportDir, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in ProdSupportDir '%s': %s\", c.ProdSupportDir, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in ProdSupportDir '%s': expected directory, got file path\", c.ProdSupportDir)\n\t\t}\n\t}\n\n\t// validate KeyRoot\n\tif len(c.KeyRoot) == 0 {\n\t\treturn errors.New(\"Missing config param: KeyRoot\")\n\t}\n\tfi, err = os.Stat(c.KeyRoot)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.KeyRoot, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in KeyRoot '%s': %s\", c.KeyRoot, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in KeyRoot '%s': expected directory, got file path\", c.KeyRoot)\n\t\t}\n\t}\n\n\t// validate MetaRoot\n\tif len(c.MetaRoot) == 0 {\n\t\treturn errors.New(\"Missing config param: MetaRoot\")\n\t}\n\tfi, err = os.Stat(c.MetaRoot)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.MetaRoot, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in MetaRoot '%s': %s\", c.MetaRoot, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in MetaRoot '%s': expected directory, got file path\", c.MetaRoot)\n\t\t}\n\t}\n\n\t// validate HTTPLog\n\tif len(c.HTTPLog) > 0 {\n\t\tfi, err = os.Stat(filepath.Dir(c.HTTPLog))\n\t\tif err != nil {\n\t\t\t// doesn't exist... can we create it?\n\t\t\tif err = os.MkdirAll(filepath.Dir(c.HTTPLog), 0744); err != nil {\n\t\t\t\treturn fmt.Errorf(\"Config error in HTTPLog '%s': %s\", c.HTTPLog, err)\n\t\t\t}\n\t\t}\n\t}\n\n\t// validate HtpasswdFile\n\tif len(c.HtpasswdFile) == 0 {\n\t\treturn errors.New(\"Missing config param: HtpasswdFile\")\n\t}\n\tfi, err = os.Stat(c.HtpasswdFile)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in HtpasswdFile '%s': %s\", c.HtpasswdFile, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in HtpasswdFile '%s': expected file path, got directory\", c.HtpasswdFile)\n\t}\n\n\tif len(c.MasterKeyPassphrase) == 0 {\n\t\tlog.Println(\"no passphrase specified for secure keyring\")\n\t}\n\n\treturn nil\n}","func (c *Config) Validate(log log.Logger) error {\n\tif err := c.SchemaConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid schema config\")\n\t}\n\tif err := c.StorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid storage config\")\n\t}\n\tif err := c.QueryRange.Validate(log); err != nil {\n\t\treturn errors.Wrap(err, \"invalid queryrange config\")\n\t}\n\tif err := c.TableManager.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid tablemanager config\")\n\t}\n\treturn nil\n}","func (c *Config) Validate() error {\n\tif c.ServerConf.RestAPIPort < 0 || c.ServerConf.RestAPIPort > 65535 {\n\t\treturn errors.New(\"invalid restful port: \" + strconv.Itoa(c.ServerConf.RestAPIPort))\n\t}\n\n\t// remove \"/\"\n\tc.ZeekConf.LogDir = processDirName(c.ZeekConf.LogDir)\n\tc.ZeekConf.PcapDir = processDirName(c.ZeekConf.PcapDir)\n\tc.ZeekConf.ExtractedFileDir = processDirName(c.ZeekConf.ExtractedFileDir)\n\tc.SeaweedfsConf.ExtractedFileDir = processDirName(c.SeaweedfsConf.ExtractedFileDir)\n\n\tif c.SeaweedfsConf.Retries < 1 {\n\t\treturn errors.New(\"seaweedfs post retry times < 1\")\n\t}\n\treturn nil\n}","func (config *Config) validateConfig() error {\n\tfor _, element := range config.AllowedEntries {\n\t\tif element.Width <= 0 && element.Height <= 0 {\n\t\t\treturn fmt.Errorf(\"The width and height of the configuration element with name \\\"%s\\\" are invalid.\", element.Name)\n\t\t}\n\n\t\tif element.Name == \"\" {\n\t\t\treturn fmt.Errorf(\"Name must be set\")\n\t\t}\n\n\t\tif element.Type == \"\" {\n\t\t\treturn fmt.Errorf(\"Type must be set\")\n\t\t}\n\n\t\ttypes := paint.GetAvailableTypes()\n\t\tif _, found := types[element.Type]; !found {\n\t\t\treturn fmt.Errorf(\"Type must be either %s, %s or %s at element \\\"%s\\\"\", paint.TypeCrop, paint.TypeResize, paint.TypeFit, element.Name)\n\t\t}\n\t}\n\n\treturn nil\n}","func (c *Config) Validate() error {\n\tvar errs []error\n\n\tif len(c.Hosts) == 0 {\n\t\terrs = append(errs, fmt.Errorf(\"missing hosts\"))\n\t}\n\tif c.Port == \"\" {\n\t\terrs = append(errs, fmt.Errorf(\"missing port\"))\n\t}\n\n\treturn apierrors.NewAggregate(errs)\n}","func (cfg config) Validate() {\n\tif v, exists := cfg[\"tls_min_version\"]; exists {\n\t\tvar t tlsMinVersionOption\n\t\terr := t.Set(fmt.Sprintf(\"%v\", v))\n\t\tif err == nil {\n\t\t\tnewVal := fmt.Sprintf(\"%v\", t.Get())\n\t\t\tif newVal != \"0\" {\n\t\t\t\tcfg[\"tls_min_version\"] = newVal\n\t\t\t} else {\n\t\t\t\tdelete(cfg, \"tls_min_version\")\n\t\t\t}\n\t\t} else {\n\t\t\tlogFatal(\"failed parsing tls_min_version %+v\", v)\n\t\t}\n\t}\n\tif v, exists := cfg[\"log_level\"]; exists {\n\t\tvar t lg.LogLevel\n\t\terr := t.Set(fmt.Sprintf(\"%v\", v))\n\t\tif err == nil {\n\t\t\tcfg[\"log_level\"] = t\n\t\t} else {\n\t\t\tlogFatal(\"failed parsing log_level %+v\", v)\n\t\t}\n\t}\n}","func (c Config) Validate() error {\n\tncNames := map[string]bool{}\n\tfor _, nc := range c.NotificationConfig {\n\t\tif nc.Name == nil {\n\t\t\treturn fmt.Errorf(\"Missing name in notification config: %s\", proto.MarshalTextString(nc))\n\t\t}\n\t\tfor _, pdc := range nc.PagerdutyConfig {\n\t\t\tif pdc.ServiceKey == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing service key in PagerDuty notification config: %s\", proto.MarshalTextString(pdc))\n\t\t\t}\n\t\t}\n\t\tfor _, ec := range nc.EmailConfig {\n\t\t\tif ec.Email == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing email address in email notification config: %s\", proto.MarshalTextString(ec))\n\t\t\t}\n\t\t}\n\t\tfor _, ec := range nc.PushoverConfig {\n\t\t\tif ec.Token == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing token in Pushover notification config: %s\", proto.MarshalTextString(ec))\n\t\t\t}\n\t\t\tif ec.UserKey == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing user key in Pushover notification config: %s\", proto.MarshalTextString(ec))\n\t\t\t}\n\t\t}\n\t\tfor _, hcc := range nc.HipchatConfig {\n\t\t\tif hcc.AuthToken == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing token in HipChat config: %s\", proto.MarshalTextString(hcc))\n\t\t\t}\n\t\t\tif hcc.RoomId == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing room in HipChat config: %s\", proto.MarshalTextString(hcc))\n\t\t\t}\n\t\t}\n\t\tfor _, sc := range nc.SlackConfig {\n\t\t\tif sc.WebhookUrl == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing webhook URL in Slack config: %s\", proto.MarshalTextString(sc))\n\t\t\t}\n\t\t}\n\t\tfor _, fc := range nc.FlowdockConfig {\n\t\t\tif fc.ApiToken == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing API token in Flowdock config: %s\", proto.MarshalTextString(fc))\n\t\t\t}\n\t\t\tif fc.FromAddress == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing from_address Flowdock config: %s\", proto.MarshalTextString(fc))\n\t\t\t}\n\t\t}\n\n\t\tif _, ok := ncNames[nc.GetName()]; ok {\n\t\t\treturn fmt.Errorf(\"Notification config name not unique: %s\", nc.GetName())\n\t\t}\n\n\t\tncNames[nc.GetName()] = true\n\t}\n\n\tfor _, a := range c.AggregationRule {\n\t\tfor _, f := range a.Filter {\n\t\t\tif f.NameRe == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing name pattern (name_re) in filter definition: %s\", proto.MarshalTextString(f))\n\t\t\t}\n\t\t\tif f.ValueRe == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing value pattern (value_re) in filter definition: %s\", proto.MarshalTextString(f))\n\t\t\t}\n\t\t}\n\n\t\tif _, ok := ncNames[a.GetNotificationConfigName()]; !ok {\n\t\t\treturn fmt.Errorf(\"No such notification config: %s\", a.GetNotificationConfigName())\n\t\t}\n\t}\n\n\treturn nil\n}","func (cfg Config) Validate() error {\n\tvar errs []string\n\tif cfg.Releaser == nil {\n\t\terrs = append(errs, \"releaser not supplied\")\n\t}\n\tif cfg.History == nil {\n\t\terrs = append(errs, \"history DB not supplied\")\n\t}\n\tif len(errs) > 0 {\n\t\treturn errors.New(\"invalid: \" + strings.Join(errs, \"; \"))\n\t}\n\treturn nil\n}","func (c *Config) Validate() error {\n\tnames := make(map[string]struct{}, len(c.Configs))\n\tfor idx, c := range c.Configs {\n\t\tif c.Name == \"\" {\n\t\t\treturn fmt.Errorf(\"tempo config at index %d is missing a name\", idx)\n\t\t}\n\t\tif _, exist := names[c.Name]; exist {\n\t\t\treturn fmt.Errorf(\"found multiple tempo configs with name %s\", c.Name)\n\t\t}\n\t\tnames[c.Name] = struct{}{}\n\t}\n\n\treturn nil\n}","func validateConfig(cfg YamlConfig) error {\n\n\t// If Directory is nil, try to retrieve from env var, if not then error\n\tif cfg.Directory == \"\" {\n\t\treturn errors.New(\"Did not find logs directory in YAML configuration\")\n\t}\n\n\t// If Directory resolved, check if directory exists, if not then error\n\tif _, err := os.Stat(cfg.Directory); os.IsNotExist(err) {\n\t\treturn errors.New(fmt.Sprintf(\"Resolved logs directory %s does not exist, exiting\", cfg.Directory))\n\t}\n\n\t// If Name or LogPattern missing, error\n\tif cfg.ConfigName == \"\" {\n\t\treturn errors.New(\"YAML configuration missing required 'ConfigName' key, exiting\")\n\t} else if cfg.LogPattern == \"\" {\n\t\treturn errors.New(\"YAML configuration missing required 'LogPattern' key, exiting\")\n\t}\n\n\tconnectorsConfigs := getConnectorsConfigs(cfg)\n\n\tfor _, connCfg := range connectorsConfigs {\n\t\t// Assert connectors have valid common fields values\n\t\terr := validateConnectorsCommonFields(connCfg)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Assert connectors have valid S3 connector fields\n\t\terr = connCfg.validate()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}","func (m *Config) Validate() error {\n\treturn m.validate(false)\n}","func (cc *Config) Validate() error {\n\tif len(cc.BrokerList) == 0 {\n\t\treturn ErrConfigNoBrokers\n\t}\n\n\tif cc.ReadTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidReadTimeout\n\t}\n\n\tif cc.WriteTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidWriteTimeout\n\t}\n\n\tif cc.ConnectTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidConnectTimeout\n\t}\n\n\tif cc.KeepAliveTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidKeepAliveTimeout\n\t}\n\n\tif cc.FetchSize < 1 {\n\t\treturn ErrConfigInvalidFetchSize\n\t}\n\n\tif cc.MetadataRetries < 0 {\n\t\treturn ErrConfigInvalidMetadataRetries\n\t}\n\n\tif cc.MetadataBackoff < time.Millisecond {\n\t\treturn ErrConfigInvalidMetadataBackoff\n\t}\n\n\tif cc.MetadataTTL < time.Millisecond {\n\t\treturn ErrConfigInvalidMetadataTTL\n\t}\n\n\tif cc.CommitOffsetRetries < 0 {\n\t\treturn ErrConfigInvalidCommitOffsetRetries\n\t}\n\n\tif cc.CommitOffsetBackoff < time.Millisecond {\n\t\treturn ErrConfigInvalidCommitOffsetBackoff\n\t}\n\n\tif cc.ConsumerMetadataRetries < 0 {\n\t\treturn ErrConfigInvalidConsumerMetadataRetries\n\t}\n\n\tif cc.ConsumerMetadataBackoff < time.Millisecond {\n\t\treturn ErrConfigInvalidConsumerMetadataBackoff\n\t}\n\n\tif cc.ClientID == \"\" {\n\t\treturn ErrConfigEmptyClientID\n\t}\n\n\treturn nil\n}","func (m *TunnelConfig) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.validateIPAddress(formats); err != nil {\n\t\t// prop\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateSharedSecret(formats); err != nil {\n\t\t// prop\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}","func (mgc Config) Validate() error {\n\tif mgc.User == \"\" {\n\t\treturn errors.New(\"Config.User is required\")\n\t}\n\tif mgc.Password == \"\" {\n\t\treturn errors.New(\"Config.Password is required\")\n\t}\n\tif mgc.AuthDB == \"\" {\n\t\treturn errors.New(\"Config.AuthDB is required\")\n\t}\n\tif mgc.Host == \"\" {\n\t\treturn errors.New(\"Config.Host is required\")\n\t}\n\tif mgc.DB == \"\" {\n\t\treturn errors.New(\"Config.DB is required\")\n\t}\n\treturn nil\n}","func (m *EncryptionAtRestConfig) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.validateKmsConfigUUID(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateOpType(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateType(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}","func (config Configuration) Valid() error {\n\tif config.Issuer == \"\" {\n\t\treturn errors.New(\"authproxy/oidc: issuer is empty\")\n\t}\n\tif config.AuthorizationEndpoint == \"\" {\n\t\treturn errors.New(\"authproxy/oidc: authorization_endpoint is empty\")\n\t}\n\tif config.JWKSURI == \"\" {\n\t\treturn errors.New(\"authproxy/oidc: jwks_uri is empty\")\n\t}\n\tif len(config.ResponseTypesSupported) == 0 {\n\t\treturn errors.New(\"authproxy/oidc: response_types_supported is empty\")\n\t}\n\tif len(config.SubjectTypesSupported) == 0 {\n\t\treturn errors.New(\"authproxy/oidc: subject_types_supported is empty\")\n\t}\n\tif len(config.IDTokenSigningAlgValuesSupported) == 0 {\n\t\treturn errors.New(\"authproxy/oidc: id_token_signing_alg_values_supported is empty\")\n\t}\n\n\treturn nil\n}","func (c *Config) Validate() error {\n\tvar err error\n\tif c.Endpoint == \"\" {\n\t\terr = multierr.Append(err, errors.New(\"no manager endpoint was specified\"))\n\t\treturn err\n\t}\n\n\tres, err := url.Parse(c.Endpoint)\n\tif err != nil {\n\t\terr = multierr.Append(err, fmt.Errorf(\"unable to parse url %s: %w\", c.Endpoint, err))\n\t\treturn err\n\t}\n\n\tif res.Scheme != \"http\" && res.Scheme != \"https\" {\n\t\terr = multierr.Append(err, errors.New(\"url scheme must be http or https\"))\n\t}\n\n\tif c.Username == \"\" {\n\t\terr = multierr.Append(err, errors.New(\"username not provided and is required\"))\n\t}\n\n\tif c.Password == \"\" {\n\t\terr = multierr.Append(err, errors.New(\"password not provided and is required\"))\n\t}\n\treturn err\n}","func (cfg *Config) Validate() error {\n\tif !cfg.LogDataEnabled && !cfg.ProfilingDataEnabled {\n\t\treturn errors.New(`either \"log_data_enabled\" or \"profiling_data_enabled\" has to be true`)\n\t}\n\tif cfg.HTTPClientSettings.Endpoint == \"\" {\n\t\treturn errors.New(`requires a non-empty \"endpoint\"`)\n\t}\n\t_, err := cfg.getURL()\n\tif err != nil {\n\t\treturn fmt.Errorf(`invalid \"endpoint\": %w`, err)\n\t}\n\tif cfg.Token == \"\" {\n\t\treturn errors.New(`requires a non-empty \"token\"`)\n\t}\n\n\tif cfg.MaxContentLengthLogs > maxContentLengthLogsLimit {\n\t\treturn fmt.Errorf(`requires \"max_content_length_logs\" <= %d`, maxContentLengthLogsLimit)\n\t}\n\n\tif cfg.MaxContentLengthMetrics > maxContentLengthMetricsLimit {\n\t\treturn fmt.Errorf(`requires \"max_content_length_metrics\" <= %d`, maxContentLengthMetricsLimit)\n\t}\n\n\tif cfg.MaxContentLengthTraces > maxContentLengthTracesLimit {\n\t\treturn fmt.Errorf(`requires \"max_content_length_traces\" <= %d`, maxContentLengthTracesLimit)\n\t}\n\n\tif cfg.MaxEventSize > maxMaxEventSize {\n\t\treturn fmt.Errorf(`requires \"max_event_size\" <= %d`, maxMaxEventSize)\n\t}\n\n\tif err := cfg.QueueSettings.Validate(); err != nil {\n\t\treturn fmt.Errorf(\"sending_queue settings has invalid configuration: %w\", err)\n\t}\n\treturn nil\n}","func (c *Config) Validate() bool {\n\tswitch c.Ops {\n\tcase 0, 1, 2, 4, 8:\n\t\treturn true\n\t}\n\treturn false\n}","func ValidateConfig() {\n\tconfig := GetConfig()\n\n\t// list of required keys\n\t// if these do not exist, prevent running the app\n\trequiredKeys := []string{\n\t\t\"app.core.root_directory\",\n\t\t\"app.core.log_level\",\n\t\t\"app.core.rest.url\",\n\t\t\"app.core.rest.user\",\n\t\t\"app.core.rest.xor_key\",\n\t\t\"app.core.rest.password\",\n\t\t\"app.core.rest.masked\",\n\t\t\"app.core.db.initialised\",\n\t\t\"app.core.db.path\",\n\t}\n\n\tfor _, key := range requiredKeys {\n\t\tif !config.IsSet(key) {\n\t\t\tlog.WithFields(log.Fields{\"error\": fmt.Sprintf(\"The key %s is required, but could not be found! See the sample file for reference!\", key)}).Error(\"Invalid config file detected!\")\n\t\t\tos.Exit(1)\n\t\t}\n\t}\n}","func (cfg *Config) Validate() error {\n\tif cfg.AssumeRole.STSRegion == \"\" && cfg.Region != \"\" {\n\t\tcfg.AssumeRole.STSRegion = cfg.Region\n\t}\n\n\tcredsProvider, err := getCredsProviderFromConfig(cfg)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not retrieve credential provider: %w\", err)\n\t}\n\tif credsProvider == nil {\n\t\treturn fmt.Errorf(\"credsProvider cannot be nil\")\n\t}\n\tcfg.credsProvider = credsProvider\n\n\treturn nil\n}","func (c *Config) Validate() error {\n\tvar errs multierror.Errors\n\n\tfor _, ht := range c.HashTypes {\n\t\tif !ht.IsValid() {\n\t\t\terrs = append(errs, errors.Errorf(\"invalid hash_types value '%v'\", ht))\n\t\t}\n\t}\n\n\tvar err error\n\n\tc.MaxFileSizeBytes, err = humanize.ParseBytes(c.MaxFileSize)\n\tif err != nil {\n\t\terrs = append(errs, errors.Wrap(err, \"invalid max_file_size value\"))\n\t} else if c.MaxFileSizeBytes <= 0 {\n\t\terrs = append(errs, errors.Errorf(\"max_file_size value (%v) must be positive\", c.MaxFileSize))\n\t}\n\n\tc.ScanRateBytesPerSec, err = humanize.ParseBytes(c.ScanRatePerSec)\n\tif err != nil {\n\t\terrs = append(errs, errors.Wrap(err, \"invalid scan_rate_per_sec value\"))\n\t}\n\n\treturn errs.Err()\n}","func (c Config) Validate() error {\n\tif c.ApplicationID != \"\" && c.AlgoliaAPIKey == \"\" {\n\t\treturn errors.New(\"API key must not be empty if indexer is enabled\")\n\t}\n\tif c.ApplicationID != \"\" && c.AlgoliaSearchKey == \"\" {\n\t\treturn errors.New(\"Search key must not be empty if indexer is enabled\")\n\t}\n\tif c.ApplicationID != \"\" && c.IndexName == \"\" {\n\t\treturn errors.New(\"Index name must not be empty if indexer is enabled\")\n\t}\n\treturn nil\n}","func (cfg *NetworkServiceConfig) IsValid() error {\n\tif cfg.Mechanism == \"\" {\n\t\treturn errors.New(\"invalid mechanism specified\")\n\t}\n\tswitch cfg.Mechanism {\n\tcase memif.MECHANISM:\n\t\t// Verify folder for memif file exists and writable.\n\t\t//TODO: Add support of this validation.\n\tcase kernel.MECHANISM:\n\t\t// Verify interface name\n\t\tif len(cfg.Path) > 1 {\n\t\t\treturn errors.New(\"invalid client interface name specified\")\n\t\t}\n\t\tif len(cfg.Path[0]) > 15 {\n\t\t\treturn errors.New(\"interface part cannot exceed 15 characters\")\n\t\t}\n\t}\n\treturn nil\n}","func (c *Config) Validate() error {\n\treturn nil\n}","func (c *Config) Validate() error {\n\treturn nil\n}","func (c *Config) Validate() error {\n\treturn nil\n}","func (c *Config) Validate() error {\n\t// TODO(felix): complete validates\n\treturn nil\n}","func (cfg *Config) Validate() error {\n\tif err := cfg.CassandraStorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Cassandra Storage config\")\n\t}\n\tif err := cfg.GCPStorageConfig.Validate(util_log.Logger); err != nil {\n\t\treturn errors.Wrap(err, \"invalid GCP Storage Storage config\")\n\t}\n\tif err := cfg.Swift.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Swift Storage config\")\n\t}\n\tif err := cfg.IndexQueriesCacheConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Index Queries Cache config\")\n\t}\n\tif err := cfg.AzureStorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Azure Storage config\")\n\t}\n\tif err := cfg.AWSStorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid AWS Storage config\")\n\t}\n\tif err := cfg.BoltDBShipperConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid boltdb-shipper config\")\n\t}\n\tif err := cfg.TSDBShipperConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid tsdb config\")\n\t}\n\n\treturn cfg.NamedStores.validate()\n}","func (cfg Config) Validate() error {\n\treturn validation.ValidateStruct(\n\t\t&cfg,\n\t\tvalidation.Field(&cfg.NodeID, validation.Required),\n\t\tvalidation.Field(&cfg.ListenAddr, validation.Required, is.Host),\n\t\tvalidation.Field(&cfg.DataDir, validation.Required),\n\t\tvalidation.Field(&cfg.CompactionEnabled, validation.Required),\n\t\tvalidation.Field(&cfg.Peers),\n\t)\n}","func (cfg *Config) Validate() error {\n\tif cfg.HTTPServerSettings == nil && cfg.GRPCServerSettings == nil {\n\t\treturn errAtLeastOneProtocol\n\t}\n\n\tif cfg.Source.File != \"\" && cfg.Source.Remote != nil {\n\t\treturn errTooManySources\n\t}\n\n\tif cfg.Source.File == \"\" && cfg.Source.Remote == nil {\n\t\treturn errNoSources\n\t}\n\n\treturn nil\n}","func (cfg *Config) Validate() error {\n\tif cfg.HTTPServerSettings == nil && cfg.GRPCServerSettings == nil {\n\t\treturn errAtLeastOneProtocol\n\t}\n\n\tif cfg.Source.File != \"\" && cfg.Source.Remote != nil {\n\t\treturn errTooManySources\n\t}\n\n\tif cfg.Source.File == \"\" && cfg.Source.Remote == nil {\n\t\treturn errNoSources\n\t}\n\n\treturn nil\n}","func ValidateConfig(c Config) error {\n\tif c.Stack.ID == \"\" {\n\t\treturn errors.New(\"stack.id is required\")\n\t}\n\n\tif c.Stack.BuildImage == \"\" {\n\t\treturn errors.New(\"stack.build-image is required\")\n\t}\n\n\tif c.Stack.RunImage == \"\" {\n\t\treturn errors.New(\"stack.run-image is required\")\n\t}\n\n\treturn nil\n}","func ValidateConfig(Validations Validations, config *types.Config) error {\n\tfor _, Service := range config.Services {\n\n\t\terr := ValidateVolumes(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = ValidateSecrets(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = ValidateNetworks(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = ValidateResources(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Println(Service.NetworkMode)\n\n\t}\n\n\treturn nil\n}","func validateConfigValue(value interface{}) error {\n\tvar err error\n\tswitch v := reflect.ValueOf(value); v.Kind() {\n\tcase reflect.Slice:\n\t\tlog.Infof(\">> checking config: found slice %+v\\n\", v.Slice(0, v.Cap()))\n\t\tfor i := 0; i < v.Len(); i++ {\n\t\t\terr = validateConfigValue(v.Index(i).Interface())\n\t\t}\n\tcase reflect.Map:\n\t\tlog.Infof(\">> checking config: found map %+v\\n\", v.MapKeys())\n\t\tfor _, key := range v.MapKeys() {\n\t\t\terr = validateConfigValue(v.MapIndex(key).Interface())\n\t\t}\n\tcase reflect.String:\n\t\tlog.Infof(\">> checking config: found string \\n\")\n\tdefault:\n\t\terr = fmt.Errorf(\"tfs doesn't handle %s type in the configuration file\", v.Kind())\n\t}\n\treturn err\n}","func (c *hostNameFormatConfig) IsValid(name string) bool {\n\tfor _, validator := range c.validators {\n\t\tif !validator.IsValid(name) {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}","func validateConfig(appConfig AppConfig) error {\n\n\treturn nil\n}","func (c Config) Validate() error {\n\tif c.CollectorEndpoint == \"\" && c.AgentEndpoint == \"\" {\n\t\treturn xerrors.New(\"jaeger: either collector endpoint or agent endpoint must be configured\")\n\t}\n\tif c.ServiceName == \"\" {\n\t\treturn xerrors.New(\"jaeger: service name must not be blank\")\n\t}\n\n\treturn nil\n}","func (c *Config) valid() error {\n\tif c.Score == nil {\n\t\treturn errors.New(\"Expected Score to not be nil\")\n\t}\n\tif c.Sampler == nil {\n\t\treturn errors.New(\"Expected Sampler to not be nil\")\n\t}\n\treturn nil\n}","func (c *Config) Validate() error {\n\tvar allErrs error\n\n\tif c.Endpoint == \"\" {\n\t\treturn multierr.Append(allErrs, errEmptyEndpoint)\n\t}\n\n\thost, portStr, err := net.SplitHostPort(c.Endpoint)\n\tif err != nil {\n\t\treturn multierr.Append(allErrs, fmt.Errorf(\"%w: %s\", errBadEndpoint, err.Error()))\n\t}\n\n\tif host == \"\" {\n\t\tallErrs = multierr.Append(allErrs, errBadEndpoint)\n\t}\n\n\tport, err := strconv.ParseInt(portStr, 10, 32)\n\tif err != nil {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: %s\", errBadPort, err.Error()))\n\t}\n\n\tif port < 0 || port > 65535 {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: %d\", errBadPort, port))\n\t}\n\n\tif c.Username != \"\" && c.Password == \"\" {\n\t\tallErrs = multierr.Append(allErrs, errEmptyPassword)\n\t}\n\n\tif c.Password != \"\" && c.Username == \"\" {\n\t\tallErrs = multierr.Append(allErrs, errEmptyUsername)\n\t}\n\tif c.Timeout.Milliseconds() < 0 {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: must be positive\", errNegativeTimeout))\n\t}\n\n\tif c.TLS != nil {\n\t\t_, err := c.TLS.LoadTLSConfig()\n\t\tif err != nil {\n\t\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: %s\", errFailedTLSLoad, err.Error()))\n\t\t}\n\t}\n\n\tif c.TLS != nil && c.TLSName == \"\" {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: when using TLS\", errEmptyEndpointTLSName))\n\t}\n\n\treturn allErrs\n}","func (c CopyConfig) Validate() error {\n\tif err := c.GitilesConfig.Validate(); err != nil {\n\t\treturn skerr.Wrap(err)\n\t}\n\tif len(c.Copies) == 0 {\n\t\treturn skerr.Fmt(\"Copies are required\")\n\t}\n\tfor _, copy := range c.Copies {\n\t\tif err := copy.Validate(); err != nil {\n\t\t\treturn skerr.Wrap(err)\n\t\t}\n\t}\n\treturn nil\n}","func (o *IamLdapConfigParamsAllOf) HasConfigHost() bool {\n\tif o != nil && o.ConfigHost != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}","func (d *common) validateConfig(info *api.NetworkACLPut) error {\n\terr := d.validateConfigMap(info.Config, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Normalise rules before validation for duplicate detection.\n\tfor i := range info.Ingress {\n\t\tinfo.Ingress[i].Normalise()\n\t}\n\n\tfor i := range info.Egress {\n\t\tinfo.Egress[i].Normalise()\n\t}\n\n\t// Validate each ingress rule.\n\tfor i, ingressRule := range info.Ingress {\n\t\terr := d.validateRule(ruleDirectionIngress, ingressRule)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Invalid ingress rule %d: %w\", i, err)\n\t\t}\n\n\t\t// Check for duplicates.\n\t\tfor ri, r := range info.Ingress {\n\t\t\tif ri == i {\n\t\t\t\tcontinue // Skip ourselves.\n\t\t\t}\n\n\t\t\tif r == ingressRule {\n\t\t\t\treturn fmt.Errorf(\"Duplicate of ingress rule %d\", i)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Validate each egress rule.\n\tfor i, egressRule := range info.Egress {\n\t\terr := d.validateRule(ruleDirectionEgress, egressRule)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Invalid egress rule %d: %w\", i, err)\n\t\t}\n\n\t\t// Check for duplicates.\n\t\tfor ri, r := range info.Egress {\n\t\t\tif ri == i {\n\t\t\t\tcontinue // Skip ourselves.\n\t\t\t}\n\n\t\t\tif r == egressRule {\n\t\t\t\treturn fmt.Errorf(\"Duplicate of egress rule %d\", i)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}","func ValidateConfig(config *Config) error {\n\tif config.Gcal.ClientID == \"\" {\n\t\treturn errors.New(\"config validattion error: gcal.client_id is missing\")\n\t}\n\tif config.Gcal.ClientSecret == \"\" {\n\t\treturn errors.New(\"config validattion error: gcal.client_secret is missing\")\n\t}\n\n\treturn nil\n}","func (config appConfig) Validate() error {\n\treturn validator.New().Struct(&config)\n}","func (tc *TestConfig) IsValid() bool {\n\treturn tc.APIConfig != nil && tc.err == nil\n}","func Validate(cfg *Config) error {\n\tif cfg == nil {\n\t\treturn errors.New(\"config needs to be defined\")\n\t}\n\n\tvar result *multierror.Error\n\tif cfg.HostProvider == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"a host provider needs to be provided\"))\n\t}\n\tif cfg.BaseClusterCloudprovider == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"the cloudprovider of the hostcluster needs to be defined\"))\n\t}\n\tif cfg.Shoots.DefaultTest == nil {\n\t\tresult = multierror.Append(result, errors.New(\"a default test needs to be defined\"))\n\t}\n\n\tif cfg.Shoots.Namespace == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"the shoot project namespace has to be defined\"))\n\t}\n\n\tif cfg.Gardener.Version == \"\" && cfg.Gardener.Commit == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"a gardener version or commit has to be defined\"))\n\t}\n\n\tif len(cfg.GardenerExtensions) == 0 {\n\t\tresult = multierror.Append(result, errors.New(\"the gardener extensions have to be defined\"))\n\t}\n\n\treturn util.ReturnMultiError(result)\n}","func (o *RemoteEnvironmentConfigListDto) HasValues() bool {\n\tif o != nil && o.Values != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}","func (e ExternalService) validateConfig() error {\n\tsl := gojsonschema.NewSchemaLoader()\n\tsc, err := sl.Compile(gojsonschema.NewStringLoader(e.schema()))\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to compile schema for external service of kind %q\", e.Kind)\n\t}\n\n\tnormalized, err := jsonc.Parse(e.Config)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to normalize JSON\")\n\t}\n\n\tres, err := sc.Validate(gojsonschema.NewBytesLoader(normalized))\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to validate config against schema\")\n\t}\n\n\terrs := new(multierror.Error)\n\tfor _, err := range res.Errors() {\n\t\terrs = multierror.Append(errs, errors.New(err.String()))\n\t}\n\n\treturn errs.ErrorOrNil()\n}","func (cm ConfigMessage) IsValid() errors.Error {\n\tif cm.MspID == \"\" {\n\t\treturn errors.New(errors.InvalidConfigMessage, \"MSPID cannot be empty\")\n\t}\n\n\tif len(cm.Peers) == 0 && len(cm.Apps) == 0 {\n\t\treturn errors.New(errors.InvalidConfigMessage, \"Either peers or apps should be set\")\n\t}\n\n\tif len(cm.Peers) > 0 {\n\t\tfor _, config := range cm.Peers {\n\t\t\tif err := config.IsValid(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn nil\n}","func (c *config) checkValidConfig() hcl.Diagnostics {\n\tvar diagnostics hcl.Diagnostics\n\n\tdiagnostics = append(diagnostics, c.checkNotEmptyWorkers()...)\n\tdiagnostics = append(diagnostics, c.checkWorkerPoolNamesUnique()...)\n\tdiagnostics = append(diagnostics, c.checkReservationIDs()...)\n\tdiagnostics = append(diagnostics, c.validateOSVersion()...)\n\tdiagnostics = append(diagnostics, c.checkCPUManagerPolicy()...)\n\n\tif c.ConntrackMaxPerCore < 0 {\n\t\tdiagnostics = append(diagnostics, &hcl.Diagnostic{\n\t\t\tSeverity: hcl.DiagError,\n\t\t\tSummary: \"conntrack_max_per_core can't be negative value\",\n\t\t\tDetail: fmt.Sprintf(\"'conntrack_max_per_core' value is %d\", c.ConntrackMaxPerCore),\n\t\t})\n\t}\n\n\tif c.OIDC != nil {\n\t\t_, diags := c.OIDC.ToKubeAPIServerFlags(c.clusterDomain())\n\t\tdiagnostics = append(diagnostics, diags...)\n\t}\n\n\tif _, diags := c.resolveNodePrivateCIDRs(); diags != nil {\n\t\tdiagnostics = append(diagnostics, diags...)\n\t}\n\n\treturn diagnostics\n}","func (pc PeerConfig) IsValid() errors.Error {\n\tif pc.PeerID == \"\" {\n\t\treturn errors.New(errors.InvalidPeerConfig, \"PeerID cannot be empty\")\n\t}\n\tif len(pc.App) == 0 {\n\t\treturn errors.New(errors.InvalidPeerConfig, \"App cannot be empty\")\n\t}\n\t//App is required\n\tfor _, appConfig := range pc.App {\n\t\tif err := appConfig.IsValid(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}","func (c *Config) Validate() error {\n\tfor _, ci := range c.Converters {\n\t\tif ci.Pattern == \"\" {\n\t\t\treturn errors.New(\"[pattern] is required for a converter\")\n\t\t}\n\t\tif ci.MetricName == \"\" {\n\t\t\treturn errors.New(\"[metricName] is required for a converter\")\n\t\t}\n\t}\n\n\treturn nil\n}","func (kce *KafkaConfigExecutor) Validate(config *gateways.ConfigContext) error {\n\tkafkaConfig, err := parseConfig(config.Data.Config)\n\tif err != nil {\n\t\treturn gateways.ErrConfigParseFailed\n\t}\n\tif kafkaConfig == nil {\n\t\treturn fmt.Errorf(\"%+v, configuration must be non empty\", gateways.ErrInvalidConfig)\n\t}\n\tif kafkaConfig.URL == \"\" {\n\t\treturn fmt.Errorf(\"%+v, url must be specified\", gateways.ErrInvalidConfig)\n\t}\n\tif kafkaConfig.Topic == \"\" {\n\t\treturn fmt.Errorf(\"%+v, topic must be specified\", gateways.ErrInvalidConfig)\n\t}\n\tif kafkaConfig.Partition == \"\" {\n\t\treturn fmt.Errorf(\"%+v, partition must be specified\", gateways.ErrInvalidConfig)\n\t}\n\treturn nil\n}","func (config *Config) valid() error {\n\t// check minlevel validation\n\tif \"\" != config.MinLevel && !LevelFromString(config.MinLevel).valid() {\n\t\treturn ErrConfigBadAttributes\n\t}\n\n\t// check filters len\n\tif len(config.Filters) < 1 {\n\t\treturn ErrConfigFiltersNotFound\n\t}\n\n\t// check filter one by one\n\tfor _, filter := range config.Filters {\n\t\tif \"\" == filter.Levels {\n\t\t\treturn ErrConfigLevelsNotFound\n\t\t}\n\n\t\tif (file{}) != filter.File {\n\t\t\t// seem not needed now\n\t\t\t//if \"\" == filter.File.Path {\n\t\t\t//return ErrConfigFilePathNotFound\n\t\t\t//}\n\t\t} else if (rotateFile{}) != filter.RotateFile {\n\t\t\tif \"\" == filter.RotateFile.Path {\n\t\t\t\treturn ErrConfigFilePathNotFound\n\t\t\t}\n\n\t\t\tif \"\" == filter.RotateFile.Type {\n\t\t\t\treturn ErrConfigFileRotateTypeNotFound\n\t\t\t}\n\t\t} else if (socket{}) != filter.Socket {\n\t\t\tif \"\" == filter.Socket.Address {\n\t\t\t\treturn ErrConfigSocketAddressNotFound\n\t\t\t}\n\n\t\t\tif \"\" == filter.Socket.Network {\n\t\t\t\treturn ErrConfigSocketNetworkNotFound\n\t\t\t}\n\t\t} else {\n\t\t\treturn ErrConfigMissingFilterType\n\t\t}\n\t}\n\n\treturn nil\n}","func (config *Config) Validate() error {\n\n\tif _, err := os.Stat(filepath.Join(config.KirdPath, config.KernelFile)); os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"kernel '%s' not found\", filepath.Join(config.KirdPath, config.KernelFile))\n\t}\n\tif _, err := os.Stat(filepath.Join(config.KirdPath, config.InitrdFile)); os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"initrd '%s' not found\", filepath.Join(config.KirdPath, config.InitrdFile))\n\t}\n\n\t// Ensure all the MappedVirtualDisks exist on the host\n\tfor _, mvd := range config.MappedVirtualDisks {\n\t\tif _, err := os.Stat(mvd.HostPath); err != nil {\n\t\t\treturn fmt.Errorf(\"mapped virtual disk '%s' not found\", mvd.HostPath)\n\t\t}\n\t\tif mvd.ContainerPath == \"\" {\n\t\t\treturn fmt.Errorf(\"mapped virtual disk '%s' requested without a container path\", mvd.HostPath)\n\t\t}\n\t}\n\n\treturn nil\n}","func ValidateConfig(fldPath *field.Path, config tmv1beta1.ConfigElement) field.ErrorList {\n\tvar allErrs field.ErrorList\n\tif config.Name == \"\" {\n\t\tallErrs = append(allErrs, field.Required(fldPath.Child(\"name\"), \"Required value\"))\n\t}\n\n\t// configmaps should either have a value or a value from defined\n\tif len(config.Value) == 0 && config.ValueFrom == nil {\n\t\tallErrs = append(allErrs, field.Required(fldPath.Child(\"value/valueFrom\"), \"A config must consist of a value or a reference to a value\"))\n\t}\n\n\t// if a valuefrom is defined then a configmap or a secret reference should be defined\n\tif config.ValueFrom != nil {\n\t\tallErrs = append(allErrs, strconf.Validate(fldPath.Child(\"valueFrom\"), config.ValueFrom)...)\n\t}\n\n\tif config.Type != tmv1beta1.ConfigTypeEnv && config.Type != tmv1beta1.ConfigTypeFile {\n\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"type\"), config.Type, \"unknown config type\"))\n\t\treturn allErrs\n\t}\n\n\tif config.Type == tmv1beta1.ConfigTypeEnv {\n\t\tif errs := validation.IsEnvVarName(config.Name); len(errs) != 0 {\n\t\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"name\"), config.Name, strings.Join(errs, \":\")))\n\t\t}\n\t\tif errs := validation.IsCIdentifier(config.Name); len(errs) != 0 {\n\t\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"name\"), config.Name, strings.Join(errs, \":\")))\n\t\t}\n\t}\n\n\tif config.Type == tmv1beta1.ConfigTypeFile {\n\t\tif config.Path == \"\" {\n\t\t\tallErrs = append(allErrs, field.Required(fldPath.Child(\"path\"), fmt.Sprintf(\"path is required for configtype %q\", tmv1beta1.ConfigTypeFile)))\n\t\t}\n\t\t// check if value is base64 encoded\n\t\tif config.Value != \"\" {\n\t\t\tif _, err := base64.StdEncoding.DecodeString(config.Value); err != nil {\n\t\t\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"value\"), config.Value, \"Value must be base64 encoded\"))\n\t\t\t}\n\t\t}\n\t}\n\treturn allErrs\n}","func (c DbConfig) Validate() error {\n\t// Holds empty config fields\n\tempty := []string{}\n\n\t// Host\n\tif len(c.Host) == 0 {\n\t\tempty = append(empty, \"Host\")\n\t}\n\n\t// User\n\tif len(c.User) == 0 {\n\t\tempty = append(empty, \"User\")\n\t}\n\n\t// Password\n\tif len(c.Password) == 0 {\n\t\tempty = append(empty, \"Password\")\n\t}\n\n\t// Db\n\tif len(c.Db) == 0 {\n\t\tempty = append(empty, \"Db\")\n\t}\n\n\t// If any empty\n\tif len(empty) > 0 {\n\t\treturn fmt.Errorf(\"the db config fields: %s, were empty\",\n\t\t\tstrings.Join(empty, \",\"))\n\t}\n\n\t// All good\n\treturn nil\n}","func (scc *ServiceCustomConfig) Validate() error {\n\n\tif len(scc.MyserviceInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for service not configured\")\n\t}\n\tif scc.MyserviceInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for service not configured\")\n\t}\n\n\tif len(scc.CommandClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Core Command client not configured\")\n\t}\n\tif scc.CommandClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Core Command client not configured\")\n\t}\n\n\tif len(scc.NotificationClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Core Command client not configured\")\n\t}\n\tif scc.NotificationClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Core Command client not configured\")\n\t}\n\n\tif len(scc.SchedulerClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Scheduler client not configured\")\n\t}\n\tif scc.SchedulerClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Scheduler client not configured\")\n\t}\n\n\tif len(scc.RuleEngineClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Rule Engine client not configured\")\n\t}\n\tif scc.RuleEngineClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Rule Engine client not configured\")\n\t}\n\n\treturn nil\n}","func (c *Config) validate() error {\n\tif err := validation.ValidateStruct(c); err != nil {\n\t\treturn err\n\t}\n\n\tif c.EnableBuiltInFiltering != nil && !*c.EnableBuiltInFiltering {\n\t\treturn errors.New(\"enableBuiltInFiltering must be true or unset, false is no longer supported\")\n\t}\n\n\tif _, err := url.Parse(c.IngestURL); err != nil {\n\t\treturn fmt.Errorf(\"%s is not a valid ingest URL: %v\", c.IngestURL, err)\n\t}\n\n\tif _, err := url.Parse(c.APIURL); err != nil {\n\t\treturn fmt.Errorf(\"%s is not a valid API URL: %v\", c.APIURL, err)\n\t}\n\n\tif _, err := url.Parse(c.EventEndpointURL); err != nil {\n\t\treturn fmt.Errorf(\"%s is not a valid event endpoint URL: %v\", c.EventEndpointURL, err)\n\t}\n\n\tif c.TraceEndpointURL != \"\" {\n\t\tif _, err := url.Parse(c.TraceEndpointURL); err != nil {\n\t\t\treturn fmt.Errorf(\"%s is not a valid trace endpoint URL: %v\", c.TraceEndpointURL, err)\n\t\t}\n\t}\n\n\tif err := c.Collectd.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tfor i := range c.Monitors {\n\t\tif err := c.Monitors[i].Validate(); err != nil {\n\t\t\treturn fmt.Errorf(\"monitor config for type '%s' is invalid: %v\", c.Monitors[i].Type, err)\n\t\t}\n\t}\n\n\treturn c.Writer.Validate()\n}","func (cfg *HTTPConfig) Valid() error {\n\t// TODO: more validation, for example, the file existence\n\tif cfg.Schema != \"https\" && cfg.Schema != \"http\" {\n\t\treturn fmt.Errorf(\"invalid schema '%s', only 'http' and 'https' are supported \", cfg.Schema)\n\t}\n\n\treturn nil\n}","func (v ConfigValueFilter) IsValid() bool {\n\tfor _, existing := range allowedConfigValueFilterEnumValues {\n\t\tif existing == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}","func (c *DataGeneratorConfig) Validate() error {\n\terr := c.BaseConfig.Validate()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif c.InitialScale == 0 {\n\t\tc.InitialScale = c.BaseConfig.Scale\n\t}\n\n\tif c.LogInterval == 0 {\n\t\treturn fmt.Errorf(errLogIntervalZero)\n\t}\n\n\terr = utils.ValidateGroups(c.InterleavedGroupID, c.InterleavedNumGroups)\n\n\tif c.Use == UseCaseDevopsGeneric && c.MaxMetricCountPerHost < 1 {\n\t\treturn fmt.Errorf(errMaxMetricCountValue)\n\t}\n\n\treturn err\n}","func (v *Config) Validate() error {\n\tctx := v.ctx\n\n\tif v.Log.Level == \"info\" {\n\t\tWarn.Println(ctx, \"info level hurts performance\")\n\t}\n\n\tif len(v.Stat.Disks) > 0 {\n\t\tWarn.Println(ctx, \"stat disks not support\")\n\t}\n\n\tif v.Workers < 0 || v.Workers > 64 {\n\t\treturn fmt.Errorf(\"workers must in [0, 64], actual is %v\", v.Workers)\n\t}\n\tif v.Listen <= 0 || v.Listen > 65535 {\n\t\treturn fmt.Errorf(\"listen must in (0, 65535], actual is %v\", v.Listen)\n\t}\n\tif v.ChunkSize < 128 || v.ChunkSize > 65535 {\n\t\treturn fmt.Errorf(\"chunk_size must in [128, 65535], actual is %v\", v.ChunkSize)\n\t}\n\n\tif v.Go.GcInterval < 0 || v.Go.GcInterval > 24*3600 {\n\t\treturn fmt.Errorf(\"go gc_interval must in [0, 24*3600], actual is %v\", v.Go.GcInterval)\n\t}\n\n\tif v.Log.Level != \"info\" && v.Log.Level != \"trace\" && v.Log.Level != \"warn\" && v.Log.Level != \"error\" {\n\t\treturn fmt.Errorf(\"log.leve must be info/trace/warn/error, actual is %v\", v.Log.Level)\n\t}\n\tif v.Log.Tank != \"console\" && v.Log.Tank != \"file\" {\n\t\treturn fmt.Errorf(\"log.tank must be console/file, actual is %v\", v.Log.Tank)\n\t}\n\tif v.Log.Tank == \"file\" && len(v.Log.File) == 0 {\n\t\treturn errors.New(\"log.file must not be empty for file tank\")\n\t}\n\n\tfor i, p := range v.Vhosts {\n\t\tif p.Name == \"\" {\n\t\t\treturn fmt.Errorf(\"the %v vhost is empty\", i)\n\t\t}\n\t}\n\n\treturn nil\n}","func ConfigOK(config Config) bool {\n\treturn config.Location != \"\" && config.User != \"\" && config.password != \"\"\n}","func (m *LolClubsClubsConfig) Validate(formats strfmt.Registry) error {\n\treturn nil\n}","func (c Config) Validate() (bool, error) {\n\t// First do some sanity checks\n\thostPorts := map[string]interface{}{}\n\n\tfor container, portForwards := range c.Forwards {\n\t\tfor _, portForward := range portForwards {\n\n\t\t\t// Make sure that port lists were actually provided\n\t\t\tif len(portForward.Ports) == 0 {\n\t\t\t\treturn false, fmt.Errorf(\"No ports provided for container %s\", container)\n\t\t\t}\n\t\t\tfor hPort := range portForward.Ports {\n\t\t\t\t_, err := strconv.Atoi(hPort)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn false, fmt.Errorf(\"Invalid port %s provided for container %s\", hPort, container)\n\t\t\t\t}\n\n\t\t\t\t// Can only forward a port from the host to one container, check to ensure no duplicate host ports\n\t\t\t\tfullPort := portForward.Protocol + \":\" + hPort\n\t\t\t\t_, ok := hostPorts[fullPort]\n\t\t\t\tif ok {\n\t\t\t\t\treturn false, fmt.Errorf(\"Port %s has already been mapped\", fullPort)\n\t\t\t\t}\n\t\t\t\thostPorts[fullPort] = nil\n\t\t\t\tportForward.Name = container\n\t\t\t}\n\t\t}\n\t}\n\treturn true, nil\n}","func (c *Config) Validate() (err error) {\n\t// Handle queue size\n\tif c.QueueSize <= 0 {\n\t\tc.QueueSize = defaultQueueSize\n\t}\n\n\t// Handle the number of workers\n\tif c.Workers <= 0 {\n\t\tc.Workers = runtime.NumCPU()\n\t}\n\n\t// Handle the addr\n\tif c.Addr == \"\" {\n\t\tc.Addr = defaultAddr\n\t}\n\n\t// Handle the metrics addr\n\tif c.MetricsAddr == \"\" {\n\t\tc.MetricsAddr = defaultMetricsAddr\n\t}\n\n\t// Handle the log level\n\tif c.LogLevel == \"\" {\n\t\tc.LogLevel = \"info\"\n\t} else {\n\t\tc.LogLevel = strings.ToLower(c.LogLevel)\n\t\tif _, ok := logLevels[c.LogLevel]; !ok {\n\t\t\treturn Errorf(ErrInvalidConfig, \"%q is an invalid log level, use trace, debug, info, caution, status, warn, or silent\", c.LogLevel)\n\t\t}\n\t}\n\tc.setLogLevel()\n\n\t// Handle the caution threshold\n\tif c.CautionThreshold == 0 {\n\t\tc.CautionThreshold = out.DefaultCautionThreshold\n\t}\n\tc.setCautionThreshold()\n\n\treturn nil\n}","func validateConfig(args ArgCollection) error {\n\texpectedKeys := []string{}\n\tconfigFileRequired := false\n\tif args.ProviderFlag == \"custom\" {\n\t\tconfigFileRequired = true\n\t\texpectedKeys = append(expectedKeys, []string{\"providers.custom.insecure\", \"providers.custom.endpoint_format\", \"providers.custom.regions\", \"providers.custom.address_style\"}...)\n\t}\n\tif args.WriteToDB {\n\t\tconfigFileRequired = true\n\t\texpectedKeys = append(expectedKeys, []string{\"db.uri\"}...)\n\t}\n\tif args.UseMq {\n\t\tconfigFileRequired = true\n\t\texpectedKeys = append(expectedKeys, []string{\"mq.queue_name\", \"mq.uri\"}...)\n\t}\n\t// User didn't give any arguments that require the config file\n\tif !configFileRequired {\n\t\treturn nil\n\t}\n\n\t// Try to find and read config file\n\tif err := viper.ReadInConfig(); err != nil {\n\t\tif _, ok := err.(viper.ConfigFileNotFoundError); ok {\n\t\t\treturn errors.New(\"config file not found\")\n\t\t} else {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Verify all expected keys are in the config file\n\tfor _, k := range expectedKeys {\n\t\tif !viper.IsSet(k) {\n\t\t\treturn fmt.Errorf(\"config file missing key: %s\", k)\n\t\t}\n\t}\n\treturn nil\n}","func (c configuration) Validate() error {\n\tvar errs error\n\n\terrs = errors.Append(errs, c.Auth.Validate())\n\terrs = errors.Append(errs, c.Config.Validate())\n\n\tif c.Environment == \"\" {\n\t\terrs = errors.Append(errs, errors.New(\"environment is required\"))\n\t}\n\n\t// TODO: this config is only used here, so the validation is here too. Either the config or the validation should be moved somewhere else.\n\tif c.Distribution.PKE.Amazon.GlobalRegion == \"\" {\n\t\terrs = errors.Append(errs, errors.New(\"pke amazon global region is required\"))\n\t}\n\n\treturn errs\n}","func (config StorageConfigStruct) Validate() error {\n\treturn nil\n}"],"string":"[\n \"func (g *Google) ValidConfig() bool {\\n\\treturn g.AccountFileJSON != \\\"\\\" && g.ProjectID != \\\"\\\" && g.Zone != \\\"\\\"\\n}\",\n \"func (c *Config) Valid() error {\\n\\tif len(c.Servers) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"No servers in config\\\")\\n\\t}\\n\\tfor _, s := range c.Servers {\\n\\t\\terr := s.Valid()\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c *Config) IsValid() bool {\\n\\tif len(c.GithubAccessToken) != 40 {\\n\\t\\treturn false\\n\\t}\\n\\tif len(c.GitName) == 0 {\\n\\t\\treturn false\\n\\t}\\n\\tif len(c.GitEmail) == 0 {\\n\\t\\treturn false\\n\\t}\\n\\n\\treturn true\\n}\",\n \"func (config *Config) IsValid() bool {\\n\\tif config.Basic.Enabled && config.OAuth2.Enabled {\\n\\t\\treturn false\\n\\t}\\n\\n\\tif config.Basic.Enabled && config.X509.Enabled {\\n\\t\\treturn false\\n\\t}\\n\\n\\tif config.OAuth2.Enabled && config.X509.Enabled {\\n\\t\\treturn false\\n\\t}\\n\\n\\tif !config.Basic.Enabled && !config.OAuth2.Enabled && !config.X509.Enabled {\\n\\t\\treturn false\\n\\t}\\n\\n\\treturn true\\n}\",\n \"func (sc *StartupConfig) Validate() (bool, error) {\\n\\tif sc.Meta == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"meta object is nil\\\")\\n\\t}\\n\\n\\tif sc.Server == nil {\\n\\t\\t//return false, fmt.Errorf(\\\"Server is nil\\\")\\n\\t}\\n\\n\\tif sc.Database == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"database object is nil\\\")\\n\\t}\\n\\n\\tif sc.Session == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"session object is nil\\\")\\n\\t}\\n\\n\\tif sc.Crypto == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"crypto object is nil\\\")\\n\\t}\\n\\n\\tif sc.Secrets == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"secrets object is nil\\\")\\n\\t}\\n\\n\\tif sc.Capsul == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"capsul object is nil\\\")\\n\\t}\\n\\n\\tif sc.CustomCapsul == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"custom capsul object is nil\\\")\\n\\t}\\n\\n\\tif sc.Create == nil {\\n\\t\\treturn false, fmt.Errorf(\\\"create object is nil\\\")\\n\\t}\\n\\n\\treturn true, nil\\n}\",\n \"func (o *Config) IsValid() *AppError {\\n\\tif err := o.ServiceSettings.isValid(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *configuration) validateConfig() error {\\n\\tif c.Provider == \\\"\\\" {\\n\\t\\treturn errors.New(msgConfigNoProvider)\\n\\t}\\n\\n\\tif len(c.Servers) == 0 {\\n\\t\\treturn errors.New(msgConfigNoServers)\\n\\t}\\n\\n\\tfor i, srv := range c.Servers {\\n\\t\\tif srv.FabricIface == \\\"\\\" {\\n\\t\\t\\treturn errors.Errorf(\\n\\t\\t\\t\\tmsgConfigServerNoIface+\\\" for I/O service %d\\\", i)\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c Config) IsValid() bool {\\n\\tif c.Port == \\\"\\\" || c.DSN == \\\"\\\" {\\n\\t\\treturn false\\n\\t}\\n\\treturn true\\n}\",\n \"func (c *Config) IsValid() error {\\n\\tif len(c.NetworkServices) == 0 {\\n\\t\\treturn errors.New(\\\"no network services are specified\\\")\\n\\t}\\n\\tif c.Name == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"no cleint name specified\\\")\\n\\t}\\n\\tif c.ConnectTo.String() == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"no NSMGr ConnectTO URL are sepecified\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c *Config) IsValid() bool {\\n\\tif c.DiffType == \\\"schema\\\" || c.DiffType == \\\"data\\\" {\\n\\t\\treturn c.Source != \\\"\\\" && c.Target != \\\"\\\"\\n\\t} else if c.DiffType == \\\"md\\\" || c.DiffType == \\\"wiki\\\" || c.DiffType == \\\"sql\\\" {\\n\\t\\treturn c.Source != \\\"\\\"\\n\\t}\\n\\treturn false\\n}\",\n \"func (c Config) Validate() error {\\n\\tif len(c.Project) == 0 {\\n\\t\\treturn errors.New(\\\"missing project in configuration\\\")\\n\\t}\\n\\tif len(c.Bucket) == 0 {\\n\\t\\treturn errors.New(\\\"missing bucket in configuration\\\")\\n\\t}\\n\\tif len(c.LastMigrationObjectName) == 0 {\\n\\t\\treturn errors.New(\\\"missing state name in configuration\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c Config) IsValid() bool {\\n\\treturn c.RandomTemperature > 0 && c.NumSimulation > 0\\n}\",\n \"func (c *Config) Validate() (err error) {\\n\\tcon := *c\\n\\tfor i, v := range con {\\n\\t\\tk, err := registry.OpenKey(v.GetScope(), v.Path, registry.ALL_ACCESS)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\tdefer k.Close()\\n\\t\\tfor n, p := range v.Properties {\\n\\t\\t\\tval := &con[i].Properties[n].PrevValue\\n\\t\\t\\tswitch p.Type {\\n\\t\\t\\tcase \\\"DWord\\\", \\\"QWord\\\":\\n\\t\\t\\t\\tif s, _, err := k.GetIntegerValue(p.Name); err == nil {\\n\\t\\t\\t\\t\\t*val = s\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\treturn err\\n\\t\\t\\t\\t}\\n\\t\\t\\tcase \\\"String\\\":\\n\\t\\t\\t\\tif s, _, err := k.GetStringValue(p.Name); err == nil {\\n\\t\\t\\t\\t\\t*val = s\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\treturn err\\n\\t\\t\\t\\t}\\n\\t\\t\\tcase \\\"Strings\\\":\\n\\t\\t\\t\\tif s, _, err := k.GetStringsValue(p.Name); err == nil {\\n\\t\\t\\t\\t\\t*val = s\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\treturn err\\n\\t\\t\\t\\t}\\n\\t\\t\\tcase \\\"Binary\\\":\\n\\t\\t\\t\\tif s, _, err := k.GetBinaryValue(p.Name); err == nil {\\n\\t\\t\\t\\t\\t*val = s\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\treturn err\\n\\t\\t\\t\\t}\\n\\t\\t\\tdefault:\\n\\t\\t\\t\\tvar buf []byte\\n\\t\\t\\t\\tif _, _, err := k.GetValue(p.Name, buf); err != nil {\\n\\t\\t\\t\\t\\treturn err\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"%s of %s path in %s scope returned code %d.\\\") // TODO: Convert const int representation of value types to explicitly match what the user should type into their JSON config.\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c AuthConfig) Validate() []error {\\n\\tvar errs []error\\n\\n\\tif len(c.JwksURI) == 0 {\\n\\t\\terrs = append(errs, errors.Errorf(\\\"AuthConfig requires a non-empty JwksURI config value\\\"))\\n\\t}\\n\\n\\treturn errs\\n}\",\n \"func ValidateConfig(configMap *corev1.ConfigMap) (*kle.Config, error) {\\n\\tconfig, err := kle.NewConfigFromMap(configMap.Data)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tfor component := range config.EnabledComponents {\\n\\t\\tif !validComponents.Has(component) {\\n\\t\\t\\treturn nil, fmt.Errorf(\\\"invalid enabledComponent %q: valid values are %q\\\", component, validComponents.List())\\n\\t\\t}\\n\\t}\\n\\n\\treturn config, nil\\n}\",\n \"func (c *Config) IsValid() bool {\\n\\t_, latErr := strconv.ParseFloat(c.Latitude, 10)\\n\\t_, lonErr := strconv.ParseFloat(c.Longitude, 10)\\n\\treturn c.APIKey != \\\"\\\" && c.BaseUrl != \\\"\\\" && latErr == nil && lonErr == nil\\n}\",\n \"func (c *config) Validate() []error {\\n\\tvar errs []error\\n\\tif value, err := validateProvider(c.Provider); err != nil {\\n\\t\\terrs = append(errs, err)\\n\\t} else {\\n\\t\\tc.Provider = value\\n\\t}\\n\\tif value, err := validateAccountName(c.AccountName); err != nil {\\n\\t\\terrs = append(errs, err)\\n\\t} else {\\n\\t\\tc.AccountName = value\\n\\t}\\n\\tif value, err := validateAccountSecret(c.AccountSecret); err != nil {\\n\\t\\terrs = append(errs, err)\\n\\t} else {\\n\\t\\tc.AccountSecret = value\\n\\t}\\n\\tif value, err := validateDNSContent(c.DNSContent); err != nil {\\n\\t\\terrs = append(errs, err)\\n\\t} else {\\n\\t\\tc.DNSContent = value\\n\\t}\\n\\tif value, err := validateDockerLabel(c.DockerLabel); err != nil {\\n\\t\\terrs = append(errs, err)\\n\\t} else {\\n\\t\\tc.DockerLabel = value\\n\\t}\\n\\tif value, err := validateStore(c.Store); err != nil {\\n\\t\\terrs = append(errs, err)\\n\\t} else {\\n\\t\\tc.Store = value\\n\\t}\\n\\tif value, err := validateDataDirectory(c.DataDirectory); err != nil {\\n\\t\\terrs = append(errs, err)\\n\\t} else {\\n\\t\\tc.DataDirectory = value\\n\\t}\\n\\treturn errs\\n}\",\n \"func (m *Manager) IsDBConfigValid(config config.DatabaseConfigs) error {\\n\\tm.lock.RLock()\\n\\tdefer m.lock.RUnlock()\\n\\n\\t// Only count the length of enabled databases\\n\\tvar length int\\n\\tfor _, c := range config {\\n\\t\\tif c.Enabled {\\n\\t\\t\\tlength++\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (sc *StoreConfig) Valid() bool {\\n\\treturn sc.Clock != nil && sc.Transport != nil &&\\n\\t\\tsc.RaftTickInterval != 0 && sc.RaftHeartbeatIntervalTicks > 0 &&\\n\\t\\tsc.RaftElectionTimeoutTicks > 0 && sc.ScanInterval >= 0 &&\\n\\t\\tsc.AmbientCtx.Tracer != nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tvar errs error\\n\\tif c.Tenant.RemoteServiceAPI == \\\"\\\" {\\n\\t\\terrs = multierror.Append(errs, fmt.Errorf(\\\"tenant.remote_service_api is required\\\"))\\n\\t}\\n\\tif c.Tenant.InternalAPI == \\\"\\\" && c.Analytics.FluentdEndpoint == \\\"\\\" && c.Analytics.CredentialsJSON == nil {\\n\\t\\terrs = multierror.Append(errs, fmt.Errorf(\\\"tenant.internal_api or tenant.analytics.fluentd_endpoint is required if no service account\\\"))\\n\\t}\\n\\tif c.Tenant.OrgName == \\\"\\\" {\\n\\t\\terrs = multierror.Append(errs, fmt.Errorf(\\\"tenant.org_name is required\\\"))\\n\\t}\\n\\tif c.Tenant.EnvName == \\\"\\\" {\\n\\t\\terrs = multierror.Append(errs, fmt.Errorf(\\\"tenant.env_name is required\\\"))\\n\\t}\\n\\tif (c.Global.TLS.CertFile != \\\"\\\" || c.Global.TLS.KeyFile != \\\"\\\") &&\\n\\t\\t(c.Global.TLS.CertFile == \\\"\\\" || c.Global.TLS.KeyFile == \\\"\\\") {\\n\\t\\terrs = multierror.Append(errs, fmt.Errorf(\\\"global.tls.cert_file and global.tls.key_file are both required if either are present\\\"))\\n\\t}\\n\\tif (c.Analytics.TLS.CAFile != \\\"\\\" || c.Analytics.TLS.CertFile != \\\"\\\" || c.Analytics.TLS.KeyFile != \\\"\\\") &&\\n\\t\\t(c.Analytics.TLS.CAFile == \\\"\\\" || c.Analytics.TLS.CertFile == \\\"\\\" || c.Analytics.TLS.KeyFile == \\\"\\\") {\\n\\t\\terrs = multierror.Append(errs, fmt.Errorf(\\\"all analytics.tls options are required if any are present\\\"))\\n\\t}\\n\\treturn errs\\n}\",\n \"func (c Config) Validate() error {\\n\\t// Check DbConfig\\n\\tif err := c.Db.Validate(); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"error verifying db config: %s\\\", err.Error())\\n\\t}\\n\\n\\t// Check AuthConfig\\n\\tif err := c.Auth.Validate(); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"error verifying auth config: %s\\\", err.Error())\\n\\t}\\n\\n\\t// All good\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\n\\tif err := c.Data.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t//if err := c.HintedHandoff.Validate(); err != nil {\\n\\t//\\treturn err\\n\\t//}\\n\\tfor _, graphite := range c.GraphiteInputs {\\n\\t\\tif err := graphite.Validate(); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"invalid graphite config: %v\\\", err)\\n\\t\\t}\\n\\t}\\n\\n\\tif err := c.Monitor.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif err := c.ContinuousQuery.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif err := c.Retention.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif err := c.Precreator.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif err := c.Subscriber.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tfor _, collectd := range c.CollectdInputs {\\n\\t\\tif err := collectd.Validate(); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"invalid collectd config: %v\\\", err)\\n\\t\\t}\\n\\t}\\n\\n\\tif err := c.TLS.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func IsConfigPresent(resources []ResourceConfig, resourceConfig ResourceConfig) bool {\\n\\tfor _, resource := range resources {\\n\\t\\tif resource.Name == resourceConfig.Name && resource.Source == resourceConfig.Source {\\n\\t\\t\\tif reflect.DeepEqual(resource.Config, resourceConfig.Config) {\\n\\t\\t\\t\\treturn true\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\treturn false\\n}\",\n \"func (config *Config) Validate() error {\\n\\tif len(config.NsqLookupdAddress) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter NsqLookupdAddress missing\\\")\\n\\t}\\n\\tif len(config.NsqdAddress) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter NsqdAddress missing\\\")\\n\\t}\\n\\tif len(config.BotName) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter BotName missing\\\")\\n\\t}\\n\\tif len(config.BambooUrl) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter BambooUrl missing\\\")\\n\\t}\\n\\tif len(config.BambooUsername) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter BambooUsername missing\\\")\\n\\t}\\n\\tif len(config.BambooPassword) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter AuthUrl missing\\\")\\n\\t}\\n\\tif len(config.BambooPassword) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter AuthApplicationName missing\\\")\\n\\t}\\n\\tif len(config.BambooPassword) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter AuthApplicationPassword missing\\\")\\n\\t}\\n\\tif len(config.BambooPassword) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter BambooPassword missing\\\")\\n\\t}\\n\\tif len(config.Prefix) == 0 {\\n\\t\\treturn fmt.Errorf(\\\"parameter Prefix missing\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (o *OAuth2Config) IsValid() bool {\\n\\treturn o.TokenURL != \\\"\\\" && o.AuthURL != \\\"\\\" && len(o.Scopes) != 0\\n}\",\n \"func validateConfig(config ProvisionerConfig) error {\\n\\tswitch config.CloudProvider {\\n\\tcase constants.AWS, constants.Azure, constants.GCE, constants.Ops:\\n\\tdefault:\\n\\t\\treturn trace.BadParameter(\\\"unknown cloud provider %s\\\", config.CloudProvider)\\n\\t}\\n\\n\\terr := validator.New().Struct(&config)\\n\\tif err == nil {\\n\\t\\treturn nil\\n\\t}\\n\\n\\tvar errs []error\\n\\tif validationErrors, ok := err.(validator.ValidationErrors); ok {\\n\\t\\tfor _, fieldError := range validationErrors {\\n\\t\\t\\terrs = append(errs,\\n\\t\\t\\t\\ttrace.BadParameter(` * %s=\\\"%v\\\" fails \\\"%s\\\"`,\\n\\t\\t\\t\\t\\tfieldError.Field(), fieldError.Value(), fieldError.Tag()))\\n\\t\\t}\\n\\t}\\n\\treturn trace.NewAggregate(errs...)\\n}\",\n \"func (_ runtimeConfigValidator) ValidateConfig(config *stiapi.Config) []validation.ValidationError {\\n\\treturn validation.ValidateConfig(config)\\n}\",\n \"func (cfg *Config) Validate() error {\\n\\tchecks := []struct {\\n\\t\\tbad bool\\n\\t\\terrMsg string\\n\\t}{\\n\\t\\t{cfg.AuthorizeURL == \\\"\\\", \\\"no authorizeURL specified\\\"},\\n\\t\\t{cfg.TokenURL == \\\"\\\", \\\"no tokenURL specified\\\"},\\n\\t\\t{cfg.ClientID == \\\"\\\", \\\"no clientID specified\\\"},\\n\\t\\t{cfg.ClientSecret == \\\"\\\" && !cfg.AllowEmptyClientSecret, \\\"no clientSecret specified\\\"},\\n\\t\\t{cfg.RedirectURL == \\\"\\\", \\\"no redirectURL specified\\\"},\\n\\t\\t{cfg.SessionSecurityKey == \\\"\\\", \\\"no SessionSecurityKey specified\\\"},\\n\\t\\t{cfg.APIServerURL == \\\"\\\", \\\"no apiServerURL specified\\\"},\\n\\t}\\n\\n\\tfor _, check := range checks {\\n\\t\\tif check.bad {\\n\\t\\t\\treturn fmt.Errorf(\\\"invalid config: %s\\\", check.errMsg)\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c Config) Validate() error {\\n\\treturn validation.ValidateStruct(&c,\\n\\t\\tvalidation.Field(&c.AppMode, validation.Required),\\n\\t\\tvalidation.Field(&c.AppName, validation.Required),\\n\\t\\tvalidation.Field(&c.DBType, validation.Required),\\n\\t\\tvalidation.Field(&c.DSN, validation.Required),\\n\\t\\tvalidation.Field(&c.JWTSigningKey, validation.Required),\\n\\t\\tvalidation.Field(&c.JWTExpiration, validation.Required),\\n\\t\\tvalidation.Field(&c.MailSmtphost, validation.Required),\\n\\t\\tvalidation.Field(&c.MailSmtpport, validation.Required),\\n\\t\\tvalidation.Field(&c.MailUsername, validation.Required),\\n\\t\\tvalidation.Field(&c.MailPassword, validation.Required),\\n\\t\\tvalidation.Field(&c.AppFqdn, validation.Required),\\n\\t\\tvalidation.Field(&c.HttpEntrypoint, validation.Required),\\n\\t\\tvalidation.Field(&c.WebservName, validation.Required),\\n\\t\\tvalidation.Field(&c.GoogleCredentialFile, validation.Required),\\n\\t\\tvalidation.Field(&c.GoogleRedirectPath, validation.Required),\\n\\t\\tvalidation.Field(&c.AppSecretKey, validation.Required),\\n\\t\\tvalidation.Field(&c.BizName, validation.Required),\\n\\t\\tvalidation.Field(&c.BizShortname, validation.Required),\\n\\t\\tvalidation.Field(&c.BizEmail, validation.Required),\\n\\t\\tvalidation.Field(&c.BizPhone),\\n\\t\\tvalidation.Field(&c.BizPhone2),\\n\\t\\tvalidation.Field(&c.BizLogo, validation.Required),\\n\\t)\\n}\",\n \"func (c Config) Validate() (err error) {\\n\\tvar fi os.FileInfo\\n\\n\\t// validate key fingerprint\\n\\t_, err = HexStringToFingerprint(c.MasterKeyFingerprint)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\t// validate TLSCert\\n\\tif len(c.TLSCert) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: TLSCert\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.TLSCert)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"Config error in TLSCert '%s': %s\\\", c.TLSCert, err)\\n\\t}\\n\\tif fi.IsDir() {\\n\\t\\treturn fmt.Errorf(\\\"Config error in TLSCert '%s': expected file path, got directory\\\", c.TLSCert)\\n\\t}\\n\\n\\t// validate TLSKey\\n\\tif len(c.TLSKey) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: TLSKey\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.TLSKey)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"Config error in TLSKey '%s': %s\\\", c.TLSKey, err)\\n\\t}\\n\\tif fi.IsDir() {\\n\\t\\treturn fmt.Errorf(\\\"Config error in TLSKey '%s': expected file path, got directory\\\", c.TLSKey)\\n\\t}\\n\\n\\t// validate SecRing\\n\\tif len(c.SecRing) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: SecRing\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.SecRing)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"Config error in SecRing '%s': %s\\\", c.SecRing, err)\\n\\t}\\n\\tif fi.IsDir() {\\n\\t\\treturn fmt.Errorf(\\\"Config error in SecRing '%s': expected file path, got directory\\\", c.SecRing)\\n\\t}\\n\\n\\t// validate ProdSupportPubRing\\n\\tif len(c.ProdSupportPubRing) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: ProdSupportPubRing\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.ProdSupportPubRing)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"Config error in ProdSupportPubRing '%s': %s\\\", c.ProdSupportPubRing, err)\\n\\t}\\n\\tif fi.IsDir() {\\n\\t\\treturn fmt.Errorf(\\\"Config error in ProdSupportPubRing '%s': expected file path, got directory\\\", c.ProdSupportPubRing)\\n\\t}\\n\\n\\t// validate DataRoot\\n\\tif len(c.DataRoot) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: DataRoot\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.DataRoot)\\n\\tif err != nil {\\n\\t\\t// doesn't exist... can we create it?\\n\\t\\tif err = os.MkdirAll(c.DataRoot, 0744); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in DataRoot '%s': %s\\\", c.DataRoot, err)\\n\\t\\t}\\n\\t} else {\\n\\t\\tif !fi.IsDir() {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in DataRoot '%s': expected directory, got file path\\\", c.DataRoot)\\n\\t\\t}\\n\\t}\\n\\n\\t// validate ProdSupportDir\\n\\tif len(c.ProdSupportDir) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: ProdSupportDir\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.ProdSupportDir)\\n\\tif err != nil {\\n\\t\\t// doesn't exist... can we create it?\\n\\t\\tif err = os.MkdirAll(c.ProdSupportDir, 0744); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in ProdSupportDir '%s': %s\\\", c.ProdSupportDir, err)\\n\\t\\t}\\n\\t} else {\\n\\t\\tif !fi.IsDir() {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in ProdSupportDir '%s': expected directory, got file path\\\", c.ProdSupportDir)\\n\\t\\t}\\n\\t}\\n\\n\\t// validate KeyRoot\\n\\tif len(c.KeyRoot) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: KeyRoot\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.KeyRoot)\\n\\tif err != nil {\\n\\t\\t// doesn't exist... can we create it?\\n\\t\\tif err = os.MkdirAll(c.KeyRoot, 0744); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in KeyRoot '%s': %s\\\", c.KeyRoot, err)\\n\\t\\t}\\n\\t} else {\\n\\t\\tif !fi.IsDir() {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in KeyRoot '%s': expected directory, got file path\\\", c.KeyRoot)\\n\\t\\t}\\n\\t}\\n\\n\\t// validate MetaRoot\\n\\tif len(c.MetaRoot) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: MetaRoot\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.MetaRoot)\\n\\tif err != nil {\\n\\t\\t// doesn't exist... can we create it?\\n\\t\\tif err = os.MkdirAll(c.MetaRoot, 0744); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in MetaRoot '%s': %s\\\", c.MetaRoot, err)\\n\\t\\t}\\n\\t} else {\\n\\t\\tif !fi.IsDir() {\\n\\t\\t\\treturn fmt.Errorf(\\\"Config error in MetaRoot '%s': expected directory, got file path\\\", c.MetaRoot)\\n\\t\\t}\\n\\t}\\n\\n\\t// validate HTTPLog\\n\\tif len(c.HTTPLog) > 0 {\\n\\t\\tfi, err = os.Stat(filepath.Dir(c.HTTPLog))\\n\\t\\tif err != nil {\\n\\t\\t\\t// doesn't exist... can we create it?\\n\\t\\t\\tif err = os.MkdirAll(filepath.Dir(c.HTTPLog), 0744); err != nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Config error in HTTPLog '%s': %s\\\", c.HTTPLog, err)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\n\\t// validate HtpasswdFile\\n\\tif len(c.HtpasswdFile) == 0 {\\n\\t\\treturn errors.New(\\\"Missing config param: HtpasswdFile\\\")\\n\\t}\\n\\tfi, err = os.Stat(c.HtpasswdFile)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"Config error in HtpasswdFile '%s': %s\\\", c.HtpasswdFile, err)\\n\\t}\\n\\tif fi.IsDir() {\\n\\t\\treturn fmt.Errorf(\\\"Config error in HtpasswdFile '%s': expected file path, got directory\\\", c.HtpasswdFile)\\n\\t}\\n\\n\\tif len(c.MasterKeyPassphrase) == 0 {\\n\\t\\tlog.Println(\\\"no passphrase specified for secure keyring\\\")\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate(log log.Logger) error {\\n\\tif err := c.SchemaConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid schema config\\\")\\n\\t}\\n\\tif err := c.StorageConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid storage config\\\")\\n\\t}\\n\\tif err := c.QueryRange.Validate(log); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid queryrange config\\\")\\n\\t}\\n\\tif err := c.TableManager.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid tablemanager config\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tif c.ServerConf.RestAPIPort < 0 || c.ServerConf.RestAPIPort > 65535 {\\n\\t\\treturn errors.New(\\\"invalid restful port: \\\" + strconv.Itoa(c.ServerConf.RestAPIPort))\\n\\t}\\n\\n\\t// remove \\\"/\\\"\\n\\tc.ZeekConf.LogDir = processDirName(c.ZeekConf.LogDir)\\n\\tc.ZeekConf.PcapDir = processDirName(c.ZeekConf.PcapDir)\\n\\tc.ZeekConf.ExtractedFileDir = processDirName(c.ZeekConf.ExtractedFileDir)\\n\\tc.SeaweedfsConf.ExtractedFileDir = processDirName(c.SeaweedfsConf.ExtractedFileDir)\\n\\n\\tif c.SeaweedfsConf.Retries < 1 {\\n\\t\\treturn errors.New(\\\"seaweedfs post retry times < 1\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (config *Config) validateConfig() error {\\n\\tfor _, element := range config.AllowedEntries {\\n\\t\\tif element.Width <= 0 && element.Height <= 0 {\\n\\t\\t\\treturn fmt.Errorf(\\\"The width and height of the configuration element with name \\\\\\\"%s\\\\\\\" are invalid.\\\", element.Name)\\n\\t\\t}\\n\\n\\t\\tif element.Name == \\\"\\\" {\\n\\t\\t\\treturn fmt.Errorf(\\\"Name must be set\\\")\\n\\t\\t}\\n\\n\\t\\tif element.Type == \\\"\\\" {\\n\\t\\t\\treturn fmt.Errorf(\\\"Type must be set\\\")\\n\\t\\t}\\n\\n\\t\\ttypes := paint.GetAvailableTypes()\\n\\t\\tif _, found := types[element.Type]; !found {\\n\\t\\t\\treturn fmt.Errorf(\\\"Type must be either %s, %s or %s at element \\\\\\\"%s\\\\\\\"\\\", paint.TypeCrop, paint.TypeResize, paint.TypeFit, element.Name)\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tvar errs []error\\n\\n\\tif len(c.Hosts) == 0 {\\n\\t\\terrs = append(errs, fmt.Errorf(\\\"missing hosts\\\"))\\n\\t}\\n\\tif c.Port == \\\"\\\" {\\n\\t\\terrs = append(errs, fmt.Errorf(\\\"missing port\\\"))\\n\\t}\\n\\n\\treturn apierrors.NewAggregate(errs)\\n}\",\n \"func (cfg config) Validate() {\\n\\tif v, exists := cfg[\\\"tls_min_version\\\"]; exists {\\n\\t\\tvar t tlsMinVersionOption\\n\\t\\terr := t.Set(fmt.Sprintf(\\\"%v\\\", v))\\n\\t\\tif err == nil {\\n\\t\\t\\tnewVal := fmt.Sprintf(\\\"%v\\\", t.Get())\\n\\t\\t\\tif newVal != \\\"0\\\" {\\n\\t\\t\\t\\tcfg[\\\"tls_min_version\\\"] = newVal\\n\\t\\t\\t} else {\\n\\t\\t\\t\\tdelete(cfg, \\\"tls_min_version\\\")\\n\\t\\t\\t}\\n\\t\\t} else {\\n\\t\\t\\tlogFatal(\\\"failed parsing tls_min_version %+v\\\", v)\\n\\t\\t}\\n\\t}\\n\\tif v, exists := cfg[\\\"log_level\\\"]; exists {\\n\\t\\tvar t lg.LogLevel\\n\\t\\terr := t.Set(fmt.Sprintf(\\\"%v\\\", v))\\n\\t\\tif err == nil {\\n\\t\\t\\tcfg[\\\"log_level\\\"] = t\\n\\t\\t} else {\\n\\t\\t\\tlogFatal(\\\"failed parsing log_level %+v\\\", v)\\n\\t\\t}\\n\\t}\\n}\",\n \"func (c Config) Validate() error {\\n\\tncNames := map[string]bool{}\\n\\tfor _, nc := range c.NotificationConfig {\\n\\t\\tif nc.Name == nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"Missing name in notification config: %s\\\", proto.MarshalTextString(nc))\\n\\t\\t}\\n\\t\\tfor _, pdc := range nc.PagerdutyConfig {\\n\\t\\t\\tif pdc.ServiceKey == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing service key in PagerDuty notification config: %s\\\", proto.MarshalTextString(pdc))\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tfor _, ec := range nc.EmailConfig {\\n\\t\\t\\tif ec.Email == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing email address in email notification config: %s\\\", proto.MarshalTextString(ec))\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tfor _, ec := range nc.PushoverConfig {\\n\\t\\t\\tif ec.Token == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing token in Pushover notification config: %s\\\", proto.MarshalTextString(ec))\\n\\t\\t\\t}\\n\\t\\t\\tif ec.UserKey == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing user key in Pushover notification config: %s\\\", proto.MarshalTextString(ec))\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tfor _, hcc := range nc.HipchatConfig {\\n\\t\\t\\tif hcc.AuthToken == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing token in HipChat config: %s\\\", proto.MarshalTextString(hcc))\\n\\t\\t\\t}\\n\\t\\t\\tif hcc.RoomId == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing room in HipChat config: %s\\\", proto.MarshalTextString(hcc))\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tfor _, sc := range nc.SlackConfig {\\n\\t\\t\\tif sc.WebhookUrl == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing webhook URL in Slack config: %s\\\", proto.MarshalTextString(sc))\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\tfor _, fc := range nc.FlowdockConfig {\\n\\t\\t\\tif fc.ApiToken == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing API token in Flowdock config: %s\\\", proto.MarshalTextString(fc))\\n\\t\\t\\t}\\n\\t\\t\\tif fc.FromAddress == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing from_address Flowdock config: %s\\\", proto.MarshalTextString(fc))\\n\\t\\t\\t}\\n\\t\\t}\\n\\n\\t\\tif _, ok := ncNames[nc.GetName()]; ok {\\n\\t\\t\\treturn fmt.Errorf(\\\"Notification config name not unique: %s\\\", nc.GetName())\\n\\t\\t}\\n\\n\\t\\tncNames[nc.GetName()] = true\\n\\t}\\n\\n\\tfor _, a := range c.AggregationRule {\\n\\t\\tfor _, f := range a.Filter {\\n\\t\\t\\tif f.NameRe == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing name pattern (name_re) in filter definition: %s\\\", proto.MarshalTextString(f))\\n\\t\\t\\t}\\n\\t\\t\\tif f.ValueRe == nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Missing value pattern (value_re) in filter definition: %s\\\", proto.MarshalTextString(f))\\n\\t\\t\\t}\\n\\t\\t}\\n\\n\\t\\tif _, ok := ncNames[a.GetNotificationConfigName()]; !ok {\\n\\t\\t\\treturn fmt.Errorf(\\\"No such notification config: %s\\\", a.GetNotificationConfigName())\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (cfg Config) Validate() error {\\n\\tvar errs []string\\n\\tif cfg.Releaser == nil {\\n\\t\\terrs = append(errs, \\\"releaser not supplied\\\")\\n\\t}\\n\\tif cfg.History == nil {\\n\\t\\terrs = append(errs, \\\"history DB not supplied\\\")\\n\\t}\\n\\tif len(errs) > 0 {\\n\\t\\treturn errors.New(\\\"invalid: \\\" + strings.Join(errs, \\\"; \\\"))\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tnames := make(map[string]struct{}, len(c.Configs))\\n\\tfor idx, c := range c.Configs {\\n\\t\\tif c.Name == \\\"\\\" {\\n\\t\\t\\treturn fmt.Errorf(\\\"tempo config at index %d is missing a name\\\", idx)\\n\\t\\t}\\n\\t\\tif _, exist := names[c.Name]; exist {\\n\\t\\t\\treturn fmt.Errorf(\\\"found multiple tempo configs with name %s\\\", c.Name)\\n\\t\\t}\\n\\t\\tnames[c.Name] = struct{}{}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func validateConfig(cfg YamlConfig) error {\\n\\n\\t// If Directory is nil, try to retrieve from env var, if not then error\\n\\tif cfg.Directory == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Did not find logs directory in YAML configuration\\\")\\n\\t}\\n\\n\\t// If Directory resolved, check if directory exists, if not then error\\n\\tif _, err := os.Stat(cfg.Directory); os.IsNotExist(err) {\\n\\t\\treturn errors.New(fmt.Sprintf(\\\"Resolved logs directory %s does not exist, exiting\\\", cfg.Directory))\\n\\t}\\n\\n\\t// If Name or LogPattern missing, error\\n\\tif cfg.ConfigName == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"YAML configuration missing required 'ConfigName' key, exiting\\\")\\n\\t} else if cfg.LogPattern == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"YAML configuration missing required 'LogPattern' key, exiting\\\")\\n\\t}\\n\\n\\tconnectorsConfigs := getConnectorsConfigs(cfg)\\n\\n\\tfor _, connCfg := range connectorsConfigs {\\n\\t\\t// Assert connectors have valid common fields values\\n\\t\\terr := validateConnectorsCommonFields(connCfg)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\t// Assert connectors have valid S3 connector fields\\n\\t\\terr = connCfg.validate()\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (m *Config) Validate() error {\\n\\treturn m.validate(false)\\n}\",\n \"func (cc *Config) Validate() error {\\n\\tif len(cc.BrokerList) == 0 {\\n\\t\\treturn ErrConfigNoBrokers\\n\\t}\\n\\n\\tif cc.ReadTimeout < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidReadTimeout\\n\\t}\\n\\n\\tif cc.WriteTimeout < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidWriteTimeout\\n\\t}\\n\\n\\tif cc.ConnectTimeout < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidConnectTimeout\\n\\t}\\n\\n\\tif cc.KeepAliveTimeout < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidKeepAliveTimeout\\n\\t}\\n\\n\\tif cc.FetchSize < 1 {\\n\\t\\treturn ErrConfigInvalidFetchSize\\n\\t}\\n\\n\\tif cc.MetadataRetries < 0 {\\n\\t\\treturn ErrConfigInvalidMetadataRetries\\n\\t}\\n\\n\\tif cc.MetadataBackoff < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidMetadataBackoff\\n\\t}\\n\\n\\tif cc.MetadataTTL < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidMetadataTTL\\n\\t}\\n\\n\\tif cc.CommitOffsetRetries < 0 {\\n\\t\\treturn ErrConfigInvalidCommitOffsetRetries\\n\\t}\\n\\n\\tif cc.CommitOffsetBackoff < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidCommitOffsetBackoff\\n\\t}\\n\\n\\tif cc.ConsumerMetadataRetries < 0 {\\n\\t\\treturn ErrConfigInvalidConsumerMetadataRetries\\n\\t}\\n\\n\\tif cc.ConsumerMetadataBackoff < time.Millisecond {\\n\\t\\treturn ErrConfigInvalidConsumerMetadataBackoff\\n\\t}\\n\\n\\tif cc.ClientID == \\\"\\\" {\\n\\t\\treturn ErrConfigEmptyClientID\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (m *TunnelConfig) Validate(formats strfmt.Registry) error {\\n\\tvar res []error\\n\\n\\tif err := m.validateIPAddress(formats); err != nil {\\n\\t\\t// prop\\n\\t\\tres = append(res, err)\\n\\t}\\n\\n\\tif err := m.validateSharedSecret(formats); err != nil {\\n\\t\\t// prop\\n\\t\\tres = append(res, err)\\n\\t}\\n\\n\\tif len(res) > 0 {\\n\\t\\treturn errors.CompositeValidationError(res...)\\n\\t}\\n\\treturn nil\\n}\",\n \"func (mgc Config) Validate() error {\\n\\tif mgc.User == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Config.User is required\\\")\\n\\t}\\n\\tif mgc.Password == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Config.Password is required\\\")\\n\\t}\\n\\tif mgc.AuthDB == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Config.AuthDB is required\\\")\\n\\t}\\n\\tif mgc.Host == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Config.Host is required\\\")\\n\\t}\\n\\tif mgc.DB == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Config.DB is required\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (m *EncryptionAtRestConfig) Validate(formats strfmt.Registry) error {\\n\\tvar res []error\\n\\n\\tif err := m.validateKmsConfigUUID(formats); err != nil {\\n\\t\\tres = append(res, err)\\n\\t}\\n\\n\\tif err := m.validateOpType(formats); err != nil {\\n\\t\\tres = append(res, err)\\n\\t}\\n\\n\\tif err := m.validateType(formats); err != nil {\\n\\t\\tres = append(res, err)\\n\\t}\\n\\n\\tif len(res) > 0 {\\n\\t\\treturn errors.CompositeValidationError(res...)\\n\\t}\\n\\treturn nil\\n}\",\n \"func (config Configuration) Valid() error {\\n\\tif config.Issuer == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"authproxy/oidc: issuer is empty\\\")\\n\\t}\\n\\tif config.AuthorizationEndpoint == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"authproxy/oidc: authorization_endpoint is empty\\\")\\n\\t}\\n\\tif config.JWKSURI == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"authproxy/oidc: jwks_uri is empty\\\")\\n\\t}\\n\\tif len(config.ResponseTypesSupported) == 0 {\\n\\t\\treturn errors.New(\\\"authproxy/oidc: response_types_supported is empty\\\")\\n\\t}\\n\\tif len(config.SubjectTypesSupported) == 0 {\\n\\t\\treturn errors.New(\\\"authproxy/oidc: subject_types_supported is empty\\\")\\n\\t}\\n\\tif len(config.IDTokenSigningAlgValuesSupported) == 0 {\\n\\t\\treturn errors.New(\\\"authproxy/oidc: id_token_signing_alg_values_supported is empty\\\")\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tvar err error\\n\\tif c.Endpoint == \\\"\\\" {\\n\\t\\terr = multierr.Append(err, errors.New(\\\"no manager endpoint was specified\\\"))\\n\\t\\treturn err\\n\\t}\\n\\n\\tres, err := url.Parse(c.Endpoint)\\n\\tif err != nil {\\n\\t\\terr = multierr.Append(err, fmt.Errorf(\\\"unable to parse url %s: %w\\\", c.Endpoint, err))\\n\\t\\treturn err\\n\\t}\\n\\n\\tif res.Scheme != \\\"http\\\" && res.Scheme != \\\"https\\\" {\\n\\t\\terr = multierr.Append(err, errors.New(\\\"url scheme must be http or https\\\"))\\n\\t}\\n\\n\\tif c.Username == \\\"\\\" {\\n\\t\\terr = multierr.Append(err, errors.New(\\\"username not provided and is required\\\"))\\n\\t}\\n\\n\\tif c.Password == \\\"\\\" {\\n\\t\\terr = multierr.Append(err, errors.New(\\\"password not provided and is required\\\"))\\n\\t}\\n\\treturn err\\n}\",\n \"func (cfg *Config) Validate() error {\\n\\tif !cfg.LogDataEnabled && !cfg.ProfilingDataEnabled {\\n\\t\\treturn errors.New(`either \\\"log_data_enabled\\\" or \\\"profiling_data_enabled\\\" has to be true`)\\n\\t}\\n\\tif cfg.HTTPClientSettings.Endpoint == \\\"\\\" {\\n\\t\\treturn errors.New(`requires a non-empty \\\"endpoint\\\"`)\\n\\t}\\n\\t_, err := cfg.getURL()\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(`invalid \\\"endpoint\\\": %w`, err)\\n\\t}\\n\\tif cfg.Token == \\\"\\\" {\\n\\t\\treturn errors.New(`requires a non-empty \\\"token\\\"`)\\n\\t}\\n\\n\\tif cfg.MaxContentLengthLogs > maxContentLengthLogsLimit {\\n\\t\\treturn fmt.Errorf(`requires \\\"max_content_length_logs\\\" <= %d`, maxContentLengthLogsLimit)\\n\\t}\\n\\n\\tif cfg.MaxContentLengthMetrics > maxContentLengthMetricsLimit {\\n\\t\\treturn fmt.Errorf(`requires \\\"max_content_length_metrics\\\" <= %d`, maxContentLengthMetricsLimit)\\n\\t}\\n\\n\\tif cfg.MaxContentLengthTraces > maxContentLengthTracesLimit {\\n\\t\\treturn fmt.Errorf(`requires \\\"max_content_length_traces\\\" <= %d`, maxContentLengthTracesLimit)\\n\\t}\\n\\n\\tif cfg.MaxEventSize > maxMaxEventSize {\\n\\t\\treturn fmt.Errorf(`requires \\\"max_event_size\\\" <= %d`, maxMaxEventSize)\\n\\t}\\n\\n\\tif err := cfg.QueueSettings.Validate(); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"sending_queue settings has invalid configuration: %w\\\", err)\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() bool {\\n\\tswitch c.Ops {\\n\\tcase 0, 1, 2, 4, 8:\\n\\t\\treturn true\\n\\t}\\n\\treturn false\\n}\",\n \"func ValidateConfig() {\\n\\tconfig := GetConfig()\\n\\n\\t// list of required keys\\n\\t// if these do not exist, prevent running the app\\n\\trequiredKeys := []string{\\n\\t\\t\\\"app.core.root_directory\\\",\\n\\t\\t\\\"app.core.log_level\\\",\\n\\t\\t\\\"app.core.rest.url\\\",\\n\\t\\t\\\"app.core.rest.user\\\",\\n\\t\\t\\\"app.core.rest.xor_key\\\",\\n\\t\\t\\\"app.core.rest.password\\\",\\n\\t\\t\\\"app.core.rest.masked\\\",\\n\\t\\t\\\"app.core.db.initialised\\\",\\n\\t\\t\\\"app.core.db.path\\\",\\n\\t}\\n\\n\\tfor _, key := range requiredKeys {\\n\\t\\tif !config.IsSet(key) {\\n\\t\\t\\tlog.WithFields(log.Fields{\\\"error\\\": fmt.Sprintf(\\\"The key %s is required, but could not be found! See the sample file for reference!\\\", key)}).Error(\\\"Invalid config file detected!\\\")\\n\\t\\t\\tos.Exit(1)\\n\\t\\t}\\n\\t}\\n}\",\n \"func (cfg *Config) Validate() error {\\n\\tif cfg.AssumeRole.STSRegion == \\\"\\\" && cfg.Region != \\\"\\\" {\\n\\t\\tcfg.AssumeRole.STSRegion = cfg.Region\\n\\t}\\n\\n\\tcredsProvider, err := getCredsProviderFromConfig(cfg)\\n\\tif err != nil {\\n\\t\\treturn fmt.Errorf(\\\"could not retrieve credential provider: %w\\\", err)\\n\\t}\\n\\tif credsProvider == nil {\\n\\t\\treturn fmt.Errorf(\\\"credsProvider cannot be nil\\\")\\n\\t}\\n\\tcfg.credsProvider = credsProvider\\n\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tvar errs multierror.Errors\\n\\n\\tfor _, ht := range c.HashTypes {\\n\\t\\tif !ht.IsValid() {\\n\\t\\t\\terrs = append(errs, errors.Errorf(\\\"invalid hash_types value '%v'\\\", ht))\\n\\t\\t}\\n\\t}\\n\\n\\tvar err error\\n\\n\\tc.MaxFileSizeBytes, err = humanize.ParseBytes(c.MaxFileSize)\\n\\tif err != nil {\\n\\t\\terrs = append(errs, errors.Wrap(err, \\\"invalid max_file_size value\\\"))\\n\\t} else if c.MaxFileSizeBytes <= 0 {\\n\\t\\terrs = append(errs, errors.Errorf(\\\"max_file_size value (%v) must be positive\\\", c.MaxFileSize))\\n\\t}\\n\\n\\tc.ScanRateBytesPerSec, err = humanize.ParseBytes(c.ScanRatePerSec)\\n\\tif err != nil {\\n\\t\\terrs = append(errs, errors.Wrap(err, \\\"invalid scan_rate_per_sec value\\\"))\\n\\t}\\n\\n\\treturn errs.Err()\\n}\",\n \"func (c Config) Validate() error {\\n\\tif c.ApplicationID != \\\"\\\" && c.AlgoliaAPIKey == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"API key must not be empty if indexer is enabled\\\")\\n\\t}\\n\\tif c.ApplicationID != \\\"\\\" && c.AlgoliaSearchKey == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Search key must not be empty if indexer is enabled\\\")\\n\\t}\\n\\tif c.ApplicationID != \\\"\\\" && c.IndexName == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"Index name must not be empty if indexer is enabled\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (cfg *NetworkServiceConfig) IsValid() error {\\n\\tif cfg.Mechanism == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"invalid mechanism specified\\\")\\n\\t}\\n\\tswitch cfg.Mechanism {\\n\\tcase memif.MECHANISM:\\n\\t\\t// Verify folder for memif file exists and writable.\\n\\t\\t//TODO: Add support of this validation.\\n\\tcase kernel.MECHANISM:\\n\\t\\t// Verify interface name\\n\\t\\tif len(cfg.Path) > 1 {\\n\\t\\t\\treturn errors.New(\\\"invalid client interface name specified\\\")\\n\\t\\t}\\n\\t\\tif len(cfg.Path[0]) > 15 {\\n\\t\\t\\treturn errors.New(\\\"interface part cannot exceed 15 characters\\\")\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\t// TODO(felix): complete validates\\n\\treturn nil\\n}\",\n \"func (cfg *Config) Validate() error {\\n\\tif err := cfg.CassandraStorageConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid Cassandra Storage config\\\")\\n\\t}\\n\\tif err := cfg.GCPStorageConfig.Validate(util_log.Logger); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid GCP Storage Storage config\\\")\\n\\t}\\n\\tif err := cfg.Swift.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid Swift Storage config\\\")\\n\\t}\\n\\tif err := cfg.IndexQueriesCacheConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid Index Queries Cache config\\\")\\n\\t}\\n\\tif err := cfg.AzureStorageConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid Azure Storage config\\\")\\n\\t}\\n\\tif err := cfg.AWSStorageConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid AWS Storage config\\\")\\n\\t}\\n\\tif err := cfg.BoltDBShipperConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid boltdb-shipper config\\\")\\n\\t}\\n\\tif err := cfg.TSDBShipperConfig.Validate(); err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"invalid tsdb config\\\")\\n\\t}\\n\\n\\treturn cfg.NamedStores.validate()\\n}\",\n \"func (cfg Config) Validate() error {\\n\\treturn validation.ValidateStruct(\\n\\t\\t&cfg,\\n\\t\\tvalidation.Field(&cfg.NodeID, validation.Required),\\n\\t\\tvalidation.Field(&cfg.ListenAddr, validation.Required, is.Host),\\n\\t\\tvalidation.Field(&cfg.DataDir, validation.Required),\\n\\t\\tvalidation.Field(&cfg.CompactionEnabled, validation.Required),\\n\\t\\tvalidation.Field(&cfg.Peers),\\n\\t)\\n}\",\n \"func (cfg *Config) Validate() error {\\n\\tif cfg.HTTPServerSettings == nil && cfg.GRPCServerSettings == nil {\\n\\t\\treturn errAtLeastOneProtocol\\n\\t}\\n\\n\\tif cfg.Source.File != \\\"\\\" && cfg.Source.Remote != nil {\\n\\t\\treturn errTooManySources\\n\\t}\\n\\n\\tif cfg.Source.File == \\\"\\\" && cfg.Source.Remote == nil {\\n\\t\\treturn errNoSources\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (cfg *Config) Validate() error {\\n\\tif cfg.HTTPServerSettings == nil && cfg.GRPCServerSettings == nil {\\n\\t\\treturn errAtLeastOneProtocol\\n\\t}\\n\\n\\tif cfg.Source.File != \\\"\\\" && cfg.Source.Remote != nil {\\n\\t\\treturn errTooManySources\\n\\t}\\n\\n\\tif cfg.Source.File == \\\"\\\" && cfg.Source.Remote == nil {\\n\\t\\treturn errNoSources\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func ValidateConfig(c Config) error {\\n\\tif c.Stack.ID == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"stack.id is required\\\")\\n\\t}\\n\\n\\tif c.Stack.BuildImage == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"stack.build-image is required\\\")\\n\\t}\\n\\n\\tif c.Stack.RunImage == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"stack.run-image is required\\\")\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func ValidateConfig(Validations Validations, config *types.Config) error {\\n\\tfor _, Service := range config.Services {\\n\\n\\t\\terr := ValidateVolumes(Validations, Service)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\terr = ValidateSecrets(Validations, Service)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\terr = ValidateNetworks(Validations, Service)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\terr = ValidateResources(Validations, Service)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\tfmt.Println(Service.NetworkMode)\\n\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func validateConfigValue(value interface{}) error {\\n\\tvar err error\\n\\tswitch v := reflect.ValueOf(value); v.Kind() {\\n\\tcase reflect.Slice:\\n\\t\\tlog.Infof(\\\">> checking config: found slice %+v\\\\n\\\", v.Slice(0, v.Cap()))\\n\\t\\tfor i := 0; i < v.Len(); i++ {\\n\\t\\t\\terr = validateConfigValue(v.Index(i).Interface())\\n\\t\\t}\\n\\tcase reflect.Map:\\n\\t\\tlog.Infof(\\\">> checking config: found map %+v\\\\n\\\", v.MapKeys())\\n\\t\\tfor _, key := range v.MapKeys() {\\n\\t\\t\\terr = validateConfigValue(v.MapIndex(key).Interface())\\n\\t\\t}\\n\\tcase reflect.String:\\n\\t\\tlog.Infof(\\\">> checking config: found string \\\\n\\\")\\n\\tdefault:\\n\\t\\terr = fmt.Errorf(\\\"tfs doesn't handle %s type in the configuration file\\\", v.Kind())\\n\\t}\\n\\treturn err\\n}\",\n \"func (c *hostNameFormatConfig) IsValid(name string) bool {\\n\\tfor _, validator := range c.validators {\\n\\t\\tif !validator.IsValid(name) {\\n\\t\\t\\treturn false\\n\\t\\t}\\n\\t}\\n\\n\\treturn true\\n}\",\n \"func validateConfig(appConfig AppConfig) error {\\n\\n\\treturn nil\\n}\",\n \"func (c Config) Validate() error {\\n\\tif c.CollectorEndpoint == \\\"\\\" && c.AgentEndpoint == \\\"\\\" {\\n\\t\\treturn xerrors.New(\\\"jaeger: either collector endpoint or agent endpoint must be configured\\\")\\n\\t}\\n\\tif c.ServiceName == \\\"\\\" {\\n\\t\\treturn xerrors.New(\\\"jaeger: service name must not be blank\\\")\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *Config) valid() error {\\n\\tif c.Score == nil {\\n\\t\\treturn errors.New(\\\"Expected Score to not be nil\\\")\\n\\t}\\n\\tif c.Sampler == nil {\\n\\t\\treturn errors.New(\\\"Expected Sampler to not be nil\\\")\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tvar allErrs error\\n\\n\\tif c.Endpoint == \\\"\\\" {\\n\\t\\treturn multierr.Append(allErrs, errEmptyEndpoint)\\n\\t}\\n\\n\\thost, portStr, err := net.SplitHostPort(c.Endpoint)\\n\\tif err != nil {\\n\\t\\treturn multierr.Append(allErrs, fmt.Errorf(\\\"%w: %s\\\", errBadEndpoint, err.Error()))\\n\\t}\\n\\n\\tif host == \\\"\\\" {\\n\\t\\tallErrs = multierr.Append(allErrs, errBadEndpoint)\\n\\t}\\n\\n\\tport, err := strconv.ParseInt(portStr, 10, 32)\\n\\tif err != nil {\\n\\t\\tallErrs = multierr.Append(allErrs, fmt.Errorf(\\\"%w: %s\\\", errBadPort, err.Error()))\\n\\t}\\n\\n\\tif port < 0 || port > 65535 {\\n\\t\\tallErrs = multierr.Append(allErrs, fmt.Errorf(\\\"%w: %d\\\", errBadPort, port))\\n\\t}\\n\\n\\tif c.Username != \\\"\\\" && c.Password == \\\"\\\" {\\n\\t\\tallErrs = multierr.Append(allErrs, errEmptyPassword)\\n\\t}\\n\\n\\tif c.Password != \\\"\\\" && c.Username == \\\"\\\" {\\n\\t\\tallErrs = multierr.Append(allErrs, errEmptyUsername)\\n\\t}\\n\\tif c.Timeout.Milliseconds() < 0 {\\n\\t\\tallErrs = multierr.Append(allErrs, fmt.Errorf(\\\"%w: must be positive\\\", errNegativeTimeout))\\n\\t}\\n\\n\\tif c.TLS != nil {\\n\\t\\t_, err := c.TLS.LoadTLSConfig()\\n\\t\\tif err != nil {\\n\\t\\t\\tallErrs = multierr.Append(allErrs, fmt.Errorf(\\\"%w: %s\\\", errFailedTLSLoad, err.Error()))\\n\\t\\t}\\n\\t}\\n\\n\\tif c.TLS != nil && c.TLSName == \\\"\\\" {\\n\\t\\tallErrs = multierr.Append(allErrs, fmt.Errorf(\\\"%w: when using TLS\\\", errEmptyEndpointTLSName))\\n\\t}\\n\\n\\treturn allErrs\\n}\",\n \"func (c CopyConfig) Validate() error {\\n\\tif err := c.GitilesConfig.Validate(); err != nil {\\n\\t\\treturn skerr.Wrap(err)\\n\\t}\\n\\tif len(c.Copies) == 0 {\\n\\t\\treturn skerr.Fmt(\\\"Copies are required\\\")\\n\\t}\\n\\tfor _, copy := range c.Copies {\\n\\t\\tif err := copy.Validate(); err != nil {\\n\\t\\t\\treturn skerr.Wrap(err)\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func (o *IamLdapConfigParamsAllOf) HasConfigHost() bool {\\n\\tif o != nil && o.ConfigHost != nil {\\n\\t\\treturn true\\n\\t}\\n\\n\\treturn false\\n}\",\n \"func (d *common) validateConfig(info *api.NetworkACLPut) error {\\n\\terr := d.validateConfigMap(info.Config, nil)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t// Normalise rules before validation for duplicate detection.\\n\\tfor i := range info.Ingress {\\n\\t\\tinfo.Ingress[i].Normalise()\\n\\t}\\n\\n\\tfor i := range info.Egress {\\n\\t\\tinfo.Egress[i].Normalise()\\n\\t}\\n\\n\\t// Validate each ingress rule.\\n\\tfor i, ingressRule := range info.Ingress {\\n\\t\\terr := d.validateRule(ruleDirectionIngress, ingressRule)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"Invalid ingress rule %d: %w\\\", i, err)\\n\\t\\t}\\n\\n\\t\\t// Check for duplicates.\\n\\t\\tfor ri, r := range info.Ingress {\\n\\t\\t\\tif ri == i {\\n\\t\\t\\t\\tcontinue // Skip ourselves.\\n\\t\\t\\t}\\n\\n\\t\\t\\tif r == ingressRule {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Duplicate of ingress rule %d\\\", i)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\n\\t// Validate each egress rule.\\n\\tfor i, egressRule := range info.Egress {\\n\\t\\terr := d.validateRule(ruleDirectionEgress, egressRule)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"Invalid egress rule %d: %w\\\", i, err)\\n\\t\\t}\\n\\n\\t\\t// Check for duplicates.\\n\\t\\tfor ri, r := range info.Egress {\\n\\t\\t\\tif ri == i {\\n\\t\\t\\t\\tcontinue // Skip ourselves.\\n\\t\\t\\t}\\n\\n\\t\\t\\tif r == egressRule {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"Duplicate of egress rule %d\\\", i)\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func ValidateConfig(config *Config) error {\\n\\tif config.Gcal.ClientID == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"config validattion error: gcal.client_id is missing\\\")\\n\\t}\\n\\tif config.Gcal.ClientSecret == \\\"\\\" {\\n\\t\\treturn errors.New(\\\"config validattion error: gcal.client_secret is missing\\\")\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (config appConfig) Validate() error {\\n\\treturn validator.New().Struct(&config)\\n}\",\n \"func (tc *TestConfig) IsValid() bool {\\n\\treturn tc.APIConfig != nil && tc.err == nil\\n}\",\n \"func Validate(cfg *Config) error {\\n\\tif cfg == nil {\\n\\t\\treturn errors.New(\\\"config needs to be defined\\\")\\n\\t}\\n\\n\\tvar result *multierror.Error\\n\\tif cfg.HostProvider == \\\"\\\" {\\n\\t\\tresult = multierror.Append(result, errors.New(\\\"a host provider needs to be provided\\\"))\\n\\t}\\n\\tif cfg.BaseClusterCloudprovider == \\\"\\\" {\\n\\t\\tresult = multierror.Append(result, errors.New(\\\"the cloudprovider of the hostcluster needs to be defined\\\"))\\n\\t}\\n\\tif cfg.Shoots.DefaultTest == nil {\\n\\t\\tresult = multierror.Append(result, errors.New(\\\"a default test needs to be defined\\\"))\\n\\t}\\n\\n\\tif cfg.Shoots.Namespace == \\\"\\\" {\\n\\t\\tresult = multierror.Append(result, errors.New(\\\"the shoot project namespace has to be defined\\\"))\\n\\t}\\n\\n\\tif cfg.Gardener.Version == \\\"\\\" && cfg.Gardener.Commit == \\\"\\\" {\\n\\t\\tresult = multierror.Append(result, errors.New(\\\"a gardener version or commit has to be defined\\\"))\\n\\t}\\n\\n\\tif len(cfg.GardenerExtensions) == 0 {\\n\\t\\tresult = multierror.Append(result, errors.New(\\\"the gardener extensions have to be defined\\\"))\\n\\t}\\n\\n\\treturn util.ReturnMultiError(result)\\n}\",\n \"func (o *RemoteEnvironmentConfigListDto) HasValues() bool {\\n\\tif o != nil && o.Values != nil {\\n\\t\\treturn true\\n\\t}\\n\\n\\treturn false\\n}\",\n \"func (e ExternalService) validateConfig() error {\\n\\tsl := gojsonschema.NewSchemaLoader()\\n\\tsc, err := sl.Compile(gojsonschema.NewStringLoader(e.schema()))\\n\\tif err != nil {\\n\\t\\treturn errors.Wrapf(err, \\\"failed to compile schema for external service of kind %q\\\", e.Kind)\\n\\t}\\n\\n\\tnormalized, err := jsonc.Parse(e.Config)\\n\\tif err != nil {\\n\\t\\treturn errors.Wrapf(err, \\\"failed to normalize JSON\\\")\\n\\t}\\n\\n\\tres, err := sc.Validate(gojsonschema.NewBytesLoader(normalized))\\n\\tif err != nil {\\n\\t\\treturn errors.Wrap(err, \\\"failed to validate config against schema\\\")\\n\\t}\\n\\n\\terrs := new(multierror.Error)\\n\\tfor _, err := range res.Errors() {\\n\\t\\terrs = multierror.Append(errs, errors.New(err.String()))\\n\\t}\\n\\n\\treturn errs.ErrorOrNil()\\n}\",\n \"func (cm ConfigMessage) IsValid() errors.Error {\\n\\tif cm.MspID == \\\"\\\" {\\n\\t\\treturn errors.New(errors.InvalidConfigMessage, \\\"MSPID cannot be empty\\\")\\n\\t}\\n\\n\\tif len(cm.Peers) == 0 && len(cm.Apps) == 0 {\\n\\t\\treturn errors.New(errors.InvalidConfigMessage, \\\"Either peers or apps should be set\\\")\\n\\t}\\n\\n\\tif len(cm.Peers) > 0 {\\n\\t\\tfor _, config := range cm.Peers {\\n\\t\\t\\tif err := config.IsValid(); err != nil {\\n\\t\\t\\t\\treturn err\\n\\t\\t\\t}\\n\\t\\t}\\n\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *config) checkValidConfig() hcl.Diagnostics {\\n\\tvar diagnostics hcl.Diagnostics\\n\\n\\tdiagnostics = append(diagnostics, c.checkNotEmptyWorkers()...)\\n\\tdiagnostics = append(diagnostics, c.checkWorkerPoolNamesUnique()...)\\n\\tdiagnostics = append(diagnostics, c.checkReservationIDs()...)\\n\\tdiagnostics = append(diagnostics, c.validateOSVersion()...)\\n\\tdiagnostics = append(diagnostics, c.checkCPUManagerPolicy()...)\\n\\n\\tif c.ConntrackMaxPerCore < 0 {\\n\\t\\tdiagnostics = append(diagnostics, &hcl.Diagnostic{\\n\\t\\t\\tSeverity: hcl.DiagError,\\n\\t\\t\\tSummary: \\\"conntrack_max_per_core can't be negative value\\\",\\n\\t\\t\\tDetail: fmt.Sprintf(\\\"'conntrack_max_per_core' value is %d\\\", c.ConntrackMaxPerCore),\\n\\t\\t})\\n\\t}\\n\\n\\tif c.OIDC != nil {\\n\\t\\t_, diags := c.OIDC.ToKubeAPIServerFlags(c.clusterDomain())\\n\\t\\tdiagnostics = append(diagnostics, diags...)\\n\\t}\\n\\n\\tif _, diags := c.resolveNodePrivateCIDRs(); diags != nil {\\n\\t\\tdiagnostics = append(diagnostics, diags...)\\n\\t}\\n\\n\\treturn diagnostics\\n}\",\n \"func (pc PeerConfig) IsValid() errors.Error {\\n\\tif pc.PeerID == \\\"\\\" {\\n\\t\\treturn errors.New(errors.InvalidPeerConfig, \\\"PeerID cannot be empty\\\")\\n\\t}\\n\\tif len(pc.App) == 0 {\\n\\t\\treturn errors.New(errors.InvalidPeerConfig, \\\"App cannot be empty\\\")\\n\\t}\\n\\t//App is required\\n\\tfor _, appConfig := range pc.App {\\n\\t\\tif err := appConfig.IsValid(); err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *Config) Validate() error {\\n\\tfor _, ci := range c.Converters {\\n\\t\\tif ci.Pattern == \\\"\\\" {\\n\\t\\t\\treturn errors.New(\\\"[pattern] is required for a converter\\\")\\n\\t\\t}\\n\\t\\tif ci.MetricName == \\\"\\\" {\\n\\t\\t\\treturn errors.New(\\\"[metricName] is required for a converter\\\")\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (kce *KafkaConfigExecutor) Validate(config *gateways.ConfigContext) error {\\n\\tkafkaConfig, err := parseConfig(config.Data.Config)\\n\\tif err != nil {\\n\\t\\treturn gateways.ErrConfigParseFailed\\n\\t}\\n\\tif kafkaConfig == nil {\\n\\t\\treturn fmt.Errorf(\\\"%+v, configuration must be non empty\\\", gateways.ErrInvalidConfig)\\n\\t}\\n\\tif kafkaConfig.URL == \\\"\\\" {\\n\\t\\treturn fmt.Errorf(\\\"%+v, url must be specified\\\", gateways.ErrInvalidConfig)\\n\\t}\\n\\tif kafkaConfig.Topic == \\\"\\\" {\\n\\t\\treturn fmt.Errorf(\\\"%+v, topic must be specified\\\", gateways.ErrInvalidConfig)\\n\\t}\\n\\tif kafkaConfig.Partition == \\\"\\\" {\\n\\t\\treturn fmt.Errorf(\\\"%+v, partition must be specified\\\", gateways.ErrInvalidConfig)\\n\\t}\\n\\treturn nil\\n}\",\n \"func (config *Config) valid() error {\\n\\t// check minlevel validation\\n\\tif \\\"\\\" != config.MinLevel && !LevelFromString(config.MinLevel).valid() {\\n\\t\\treturn ErrConfigBadAttributes\\n\\t}\\n\\n\\t// check filters len\\n\\tif len(config.Filters) < 1 {\\n\\t\\treturn ErrConfigFiltersNotFound\\n\\t}\\n\\n\\t// check filter one by one\\n\\tfor _, filter := range config.Filters {\\n\\t\\tif \\\"\\\" == filter.Levels {\\n\\t\\t\\treturn ErrConfigLevelsNotFound\\n\\t\\t}\\n\\n\\t\\tif (file{}) != filter.File {\\n\\t\\t\\t// seem not needed now\\n\\t\\t\\t//if \\\"\\\" == filter.File.Path {\\n\\t\\t\\t//return ErrConfigFilePathNotFound\\n\\t\\t\\t//}\\n\\t\\t} else if (rotateFile{}) != filter.RotateFile {\\n\\t\\t\\tif \\\"\\\" == filter.RotateFile.Path {\\n\\t\\t\\t\\treturn ErrConfigFilePathNotFound\\n\\t\\t\\t}\\n\\n\\t\\t\\tif \\\"\\\" == filter.RotateFile.Type {\\n\\t\\t\\t\\treturn ErrConfigFileRotateTypeNotFound\\n\\t\\t\\t}\\n\\t\\t} else if (socket{}) != filter.Socket {\\n\\t\\t\\tif \\\"\\\" == filter.Socket.Address {\\n\\t\\t\\t\\treturn ErrConfigSocketAddressNotFound\\n\\t\\t\\t}\\n\\n\\t\\t\\tif \\\"\\\" == filter.Socket.Network {\\n\\t\\t\\t\\treturn ErrConfigSocketNetworkNotFound\\n\\t\\t\\t}\\n\\t\\t} else {\\n\\t\\t\\treturn ErrConfigMissingFilterType\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (config *Config) Validate() error {\\n\\n\\tif _, err := os.Stat(filepath.Join(config.KirdPath, config.KernelFile)); os.IsNotExist(err) {\\n\\t\\treturn fmt.Errorf(\\\"kernel '%s' not found\\\", filepath.Join(config.KirdPath, config.KernelFile))\\n\\t}\\n\\tif _, err := os.Stat(filepath.Join(config.KirdPath, config.InitrdFile)); os.IsNotExist(err) {\\n\\t\\treturn fmt.Errorf(\\\"initrd '%s' not found\\\", filepath.Join(config.KirdPath, config.InitrdFile))\\n\\t}\\n\\n\\t// Ensure all the MappedVirtualDisks exist on the host\\n\\tfor _, mvd := range config.MappedVirtualDisks {\\n\\t\\tif _, err := os.Stat(mvd.HostPath); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"mapped virtual disk '%s' not found\\\", mvd.HostPath)\\n\\t\\t}\\n\\t\\tif mvd.ContainerPath == \\\"\\\" {\\n\\t\\t\\treturn fmt.Errorf(\\\"mapped virtual disk '%s' requested without a container path\\\", mvd.HostPath)\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func ValidateConfig(fldPath *field.Path, config tmv1beta1.ConfigElement) field.ErrorList {\\n\\tvar allErrs field.ErrorList\\n\\tif config.Name == \\\"\\\" {\\n\\t\\tallErrs = append(allErrs, field.Required(fldPath.Child(\\\"name\\\"), \\\"Required value\\\"))\\n\\t}\\n\\n\\t// configmaps should either have a value or a value from defined\\n\\tif len(config.Value) == 0 && config.ValueFrom == nil {\\n\\t\\tallErrs = append(allErrs, field.Required(fldPath.Child(\\\"value/valueFrom\\\"), \\\"A config must consist of a value or a reference to a value\\\"))\\n\\t}\\n\\n\\t// if a valuefrom is defined then a configmap or a secret reference should be defined\\n\\tif config.ValueFrom != nil {\\n\\t\\tallErrs = append(allErrs, strconf.Validate(fldPath.Child(\\\"valueFrom\\\"), config.ValueFrom)...)\\n\\t}\\n\\n\\tif config.Type != tmv1beta1.ConfigTypeEnv && config.Type != tmv1beta1.ConfigTypeFile {\\n\\t\\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\\\"type\\\"), config.Type, \\\"unknown config type\\\"))\\n\\t\\treturn allErrs\\n\\t}\\n\\n\\tif config.Type == tmv1beta1.ConfigTypeEnv {\\n\\t\\tif errs := validation.IsEnvVarName(config.Name); len(errs) != 0 {\\n\\t\\t\\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\\\"name\\\"), config.Name, strings.Join(errs, \\\":\\\")))\\n\\t\\t}\\n\\t\\tif errs := validation.IsCIdentifier(config.Name); len(errs) != 0 {\\n\\t\\t\\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\\\"name\\\"), config.Name, strings.Join(errs, \\\":\\\")))\\n\\t\\t}\\n\\t}\\n\\n\\tif config.Type == tmv1beta1.ConfigTypeFile {\\n\\t\\tif config.Path == \\\"\\\" {\\n\\t\\t\\tallErrs = append(allErrs, field.Required(fldPath.Child(\\\"path\\\"), fmt.Sprintf(\\\"path is required for configtype %q\\\", tmv1beta1.ConfigTypeFile)))\\n\\t\\t}\\n\\t\\t// check if value is base64 encoded\\n\\t\\tif config.Value != \\\"\\\" {\\n\\t\\t\\tif _, err := base64.StdEncoding.DecodeString(config.Value); err != nil {\\n\\t\\t\\t\\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\\\"value\\\"), config.Value, \\\"Value must be base64 encoded\\\"))\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\treturn allErrs\\n}\",\n \"func (c DbConfig) Validate() error {\\n\\t// Holds empty config fields\\n\\tempty := []string{}\\n\\n\\t// Host\\n\\tif len(c.Host) == 0 {\\n\\t\\tempty = append(empty, \\\"Host\\\")\\n\\t}\\n\\n\\t// User\\n\\tif len(c.User) == 0 {\\n\\t\\tempty = append(empty, \\\"User\\\")\\n\\t}\\n\\n\\t// Password\\n\\tif len(c.Password) == 0 {\\n\\t\\tempty = append(empty, \\\"Password\\\")\\n\\t}\\n\\n\\t// Db\\n\\tif len(c.Db) == 0 {\\n\\t\\tempty = append(empty, \\\"Db\\\")\\n\\t}\\n\\n\\t// If any empty\\n\\tif len(empty) > 0 {\\n\\t\\treturn fmt.Errorf(\\\"the db config fields: %s, were empty\\\",\\n\\t\\t\\tstrings.Join(empty, \\\",\\\"))\\n\\t}\\n\\n\\t// All good\\n\\treturn nil\\n}\",\n \"func (scc *ServiceCustomConfig) Validate() error {\\n\\n\\tif len(scc.MyserviceInfo.Host) == 0 {\\n\\t\\treturn errors.New(\\\"host setting for service not configured\\\")\\n\\t}\\n\\tif scc.MyserviceInfo.Port == 0 {\\n\\t\\treturn errors.New(\\\"port setting for service not configured\\\")\\n\\t}\\n\\n\\tif len(scc.CommandClientInfo.Host) == 0 {\\n\\t\\treturn errors.New(\\\"host setting for Core Command client not configured\\\")\\n\\t}\\n\\tif scc.CommandClientInfo.Port == 0 {\\n\\t\\treturn errors.New(\\\"port setting for Core Command client not configured\\\")\\n\\t}\\n\\n\\tif len(scc.NotificationClientInfo.Host) == 0 {\\n\\t\\treturn errors.New(\\\"host setting for Core Command client not configured\\\")\\n\\t}\\n\\tif scc.NotificationClientInfo.Port == 0 {\\n\\t\\treturn errors.New(\\\"port setting for Core Command client not configured\\\")\\n\\t}\\n\\n\\tif len(scc.SchedulerClientInfo.Host) == 0 {\\n\\t\\treturn errors.New(\\\"host setting for Scheduler client not configured\\\")\\n\\t}\\n\\tif scc.SchedulerClientInfo.Port == 0 {\\n\\t\\treturn errors.New(\\\"port setting for Scheduler client not configured\\\")\\n\\t}\\n\\n\\tif len(scc.RuleEngineClientInfo.Host) == 0 {\\n\\t\\treturn errors.New(\\\"host setting for Rule Engine client not configured\\\")\\n\\t}\\n\\tif scc.RuleEngineClientInfo.Port == 0 {\\n\\t\\treturn errors.New(\\\"port setting for Rule Engine client not configured\\\")\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (c *Config) validate() error {\\n\\tif err := validation.ValidateStruct(c); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif c.EnableBuiltInFiltering != nil && !*c.EnableBuiltInFiltering {\\n\\t\\treturn errors.New(\\\"enableBuiltInFiltering must be true or unset, false is no longer supported\\\")\\n\\t}\\n\\n\\tif _, err := url.Parse(c.IngestURL); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"%s is not a valid ingest URL: %v\\\", c.IngestURL, err)\\n\\t}\\n\\n\\tif _, err := url.Parse(c.APIURL); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"%s is not a valid API URL: %v\\\", c.APIURL, err)\\n\\t}\\n\\n\\tif _, err := url.Parse(c.EventEndpointURL); err != nil {\\n\\t\\treturn fmt.Errorf(\\\"%s is not a valid event endpoint URL: %v\\\", c.EventEndpointURL, err)\\n\\t}\\n\\n\\tif c.TraceEndpointURL != \\\"\\\" {\\n\\t\\tif _, err := url.Parse(c.TraceEndpointURL); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"%s is not a valid trace endpoint URL: %v\\\", c.TraceEndpointURL, err)\\n\\t\\t}\\n\\t}\\n\\n\\tif err := c.Collectd.Validate(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tfor i := range c.Monitors {\\n\\t\\tif err := c.Monitors[i].Validate(); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"monitor config for type '%s' is invalid: %v\\\", c.Monitors[i].Type, err)\\n\\t\\t}\\n\\t}\\n\\n\\treturn c.Writer.Validate()\\n}\",\n \"func (cfg *HTTPConfig) Valid() error {\\n\\t// TODO: more validation, for example, the file existence\\n\\tif cfg.Schema != \\\"https\\\" && cfg.Schema != \\\"http\\\" {\\n\\t\\treturn fmt.Errorf(\\\"invalid schema '%s', only 'http' and 'https' are supported \\\", cfg.Schema)\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (v ConfigValueFilter) IsValid() bool {\\n\\tfor _, existing := range allowedConfigValueFilterEnumValues {\\n\\t\\tif existing == v {\\n\\t\\t\\treturn true\\n\\t\\t}\\n\\t}\\n\\treturn false\\n}\",\n \"func (c *DataGeneratorConfig) Validate() error {\\n\\terr := c.BaseConfig.Validate()\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif c.InitialScale == 0 {\\n\\t\\tc.InitialScale = c.BaseConfig.Scale\\n\\t}\\n\\n\\tif c.LogInterval == 0 {\\n\\t\\treturn fmt.Errorf(errLogIntervalZero)\\n\\t}\\n\\n\\terr = utils.ValidateGroups(c.InterleavedGroupID, c.InterleavedNumGroups)\\n\\n\\tif c.Use == UseCaseDevopsGeneric && c.MaxMetricCountPerHost < 1 {\\n\\t\\treturn fmt.Errorf(errMaxMetricCountValue)\\n\\t}\\n\\n\\treturn err\\n}\",\n \"func (v *Config) Validate() error {\\n\\tctx := v.ctx\\n\\n\\tif v.Log.Level == \\\"info\\\" {\\n\\t\\tWarn.Println(ctx, \\\"info level hurts performance\\\")\\n\\t}\\n\\n\\tif len(v.Stat.Disks) > 0 {\\n\\t\\tWarn.Println(ctx, \\\"stat disks not support\\\")\\n\\t}\\n\\n\\tif v.Workers < 0 || v.Workers > 64 {\\n\\t\\treturn fmt.Errorf(\\\"workers must in [0, 64], actual is %v\\\", v.Workers)\\n\\t}\\n\\tif v.Listen <= 0 || v.Listen > 65535 {\\n\\t\\treturn fmt.Errorf(\\\"listen must in (0, 65535], actual is %v\\\", v.Listen)\\n\\t}\\n\\tif v.ChunkSize < 128 || v.ChunkSize > 65535 {\\n\\t\\treturn fmt.Errorf(\\\"chunk_size must in [128, 65535], actual is %v\\\", v.ChunkSize)\\n\\t}\\n\\n\\tif v.Go.GcInterval < 0 || v.Go.GcInterval > 24*3600 {\\n\\t\\treturn fmt.Errorf(\\\"go gc_interval must in [0, 24*3600], actual is %v\\\", v.Go.GcInterval)\\n\\t}\\n\\n\\tif v.Log.Level != \\\"info\\\" && v.Log.Level != \\\"trace\\\" && v.Log.Level != \\\"warn\\\" && v.Log.Level != \\\"error\\\" {\\n\\t\\treturn fmt.Errorf(\\\"log.leve must be info/trace/warn/error, actual is %v\\\", v.Log.Level)\\n\\t}\\n\\tif v.Log.Tank != \\\"console\\\" && v.Log.Tank != \\\"file\\\" {\\n\\t\\treturn fmt.Errorf(\\\"log.tank must be console/file, actual is %v\\\", v.Log.Tank)\\n\\t}\\n\\tif v.Log.Tank == \\\"file\\\" && len(v.Log.File) == 0 {\\n\\t\\treturn errors.New(\\\"log.file must not be empty for file tank\\\")\\n\\t}\\n\\n\\tfor i, p := range v.Vhosts {\\n\\t\\tif p.Name == \\\"\\\" {\\n\\t\\t\\treturn fmt.Errorf(\\\"the %v vhost is empty\\\", i)\\n\\t\\t}\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func ConfigOK(config Config) bool {\\n\\treturn config.Location != \\\"\\\" && config.User != \\\"\\\" && config.password != \\\"\\\"\\n}\",\n \"func (m *LolClubsClubsConfig) Validate(formats strfmt.Registry) error {\\n\\treturn nil\\n}\",\n \"func (c Config) Validate() (bool, error) {\\n\\t// First do some sanity checks\\n\\thostPorts := map[string]interface{}{}\\n\\n\\tfor container, portForwards := range c.Forwards {\\n\\t\\tfor _, portForward := range portForwards {\\n\\n\\t\\t\\t// Make sure that port lists were actually provided\\n\\t\\t\\tif len(portForward.Ports) == 0 {\\n\\t\\t\\t\\treturn false, fmt.Errorf(\\\"No ports provided for container %s\\\", container)\\n\\t\\t\\t}\\n\\t\\t\\tfor hPort := range portForward.Ports {\\n\\t\\t\\t\\t_, err := strconv.Atoi(hPort)\\n\\t\\t\\t\\tif err != nil {\\n\\t\\t\\t\\t\\treturn false, fmt.Errorf(\\\"Invalid port %s provided for container %s\\\", hPort, container)\\n\\t\\t\\t\\t}\\n\\n\\t\\t\\t\\t// Can only forward a port from the host to one container, check to ensure no duplicate host ports\\n\\t\\t\\t\\tfullPort := portForward.Protocol + \\\":\\\" + hPort\\n\\t\\t\\t\\t_, ok := hostPorts[fullPort]\\n\\t\\t\\t\\tif ok {\\n\\t\\t\\t\\t\\treturn false, fmt.Errorf(\\\"Port %s has already been mapped\\\", fullPort)\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\thostPorts[fullPort] = nil\\n\\t\\t\\t\\tportForward.Name = container\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\treturn true, nil\\n}\",\n \"func (c *Config) Validate() (err error) {\\n\\t// Handle queue size\\n\\tif c.QueueSize <= 0 {\\n\\t\\tc.QueueSize = defaultQueueSize\\n\\t}\\n\\n\\t// Handle the number of workers\\n\\tif c.Workers <= 0 {\\n\\t\\tc.Workers = runtime.NumCPU()\\n\\t}\\n\\n\\t// Handle the addr\\n\\tif c.Addr == \\\"\\\" {\\n\\t\\tc.Addr = defaultAddr\\n\\t}\\n\\n\\t// Handle the metrics addr\\n\\tif c.MetricsAddr == \\\"\\\" {\\n\\t\\tc.MetricsAddr = defaultMetricsAddr\\n\\t}\\n\\n\\t// Handle the log level\\n\\tif c.LogLevel == \\\"\\\" {\\n\\t\\tc.LogLevel = \\\"info\\\"\\n\\t} else {\\n\\t\\tc.LogLevel = strings.ToLower(c.LogLevel)\\n\\t\\tif _, ok := logLevels[c.LogLevel]; !ok {\\n\\t\\t\\treturn Errorf(ErrInvalidConfig, \\\"%q is an invalid log level, use trace, debug, info, caution, status, warn, or silent\\\", c.LogLevel)\\n\\t\\t}\\n\\t}\\n\\tc.setLogLevel()\\n\\n\\t// Handle the caution threshold\\n\\tif c.CautionThreshold == 0 {\\n\\t\\tc.CautionThreshold = out.DefaultCautionThreshold\\n\\t}\\n\\tc.setCautionThreshold()\\n\\n\\treturn nil\\n}\",\n \"func validateConfig(args ArgCollection) error {\\n\\texpectedKeys := []string{}\\n\\tconfigFileRequired := false\\n\\tif args.ProviderFlag == \\\"custom\\\" {\\n\\t\\tconfigFileRequired = true\\n\\t\\texpectedKeys = append(expectedKeys, []string{\\\"providers.custom.insecure\\\", \\\"providers.custom.endpoint_format\\\", \\\"providers.custom.regions\\\", \\\"providers.custom.address_style\\\"}...)\\n\\t}\\n\\tif args.WriteToDB {\\n\\t\\tconfigFileRequired = true\\n\\t\\texpectedKeys = append(expectedKeys, []string{\\\"db.uri\\\"}...)\\n\\t}\\n\\tif args.UseMq {\\n\\t\\tconfigFileRequired = true\\n\\t\\texpectedKeys = append(expectedKeys, []string{\\\"mq.queue_name\\\", \\\"mq.uri\\\"}...)\\n\\t}\\n\\t// User didn't give any arguments that require the config file\\n\\tif !configFileRequired {\\n\\t\\treturn nil\\n\\t}\\n\\n\\t// Try to find and read config file\\n\\tif err := viper.ReadInConfig(); err != nil {\\n\\t\\tif _, ok := err.(viper.ConfigFileNotFoundError); ok {\\n\\t\\t\\treturn errors.New(\\\"config file not found\\\")\\n\\t\\t} else {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\n\\t// Verify all expected keys are in the config file\\n\\tfor _, k := range expectedKeys {\\n\\t\\tif !viper.IsSet(k) {\\n\\t\\t\\treturn fmt.Errorf(\\\"config file missing key: %s\\\", k)\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func (c configuration) Validate() error {\\n\\tvar errs error\\n\\n\\terrs = errors.Append(errs, c.Auth.Validate())\\n\\terrs = errors.Append(errs, c.Config.Validate())\\n\\n\\tif c.Environment == \\\"\\\" {\\n\\t\\terrs = errors.Append(errs, errors.New(\\\"environment is required\\\"))\\n\\t}\\n\\n\\t// TODO: this config is only used here, so the validation is here too. Either the config or the validation should be moved somewhere else.\\n\\tif c.Distribution.PKE.Amazon.GlobalRegion == \\\"\\\" {\\n\\t\\terrs = errors.Append(errs, errors.New(\\\"pke amazon global region is required\\\"))\\n\\t}\\n\\n\\treturn errs\\n}\",\n \"func (config StorageConfigStruct) Validate() error {\\n\\treturn nil\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.6982893","0.67954654","0.67327917","0.65675837","0.6560407","0.6475244","0.64741606","0.6416732","0.6389593","0.63582075","0.6342166","0.6282044","0.6264434","0.62190115","0.6190914","0.61886424","0.61752874","0.61643165","0.61314374","0.6065363","0.6061954","0.6032592","0.6018692","0.601484","0.6006803","0.6006371","0.6003116","0.6001448","0.5997107","0.59835184","0.59798545","0.5973025","0.59688985","0.5968867","0.5968517","0.59621924","0.5957821","0.59565663","0.59427094","0.5931806","0.5930262","0.5929543","0.5921107","0.59152794","0.5913778","0.59125125","0.59114504","0.590905","0.5904003","0.5900556","0.5898378","0.58954227","0.58811885","0.5879261","0.5879261","0.5879261","0.5873312","0.58590055","0.58581686","0.584782","0.584782","0.58375776","0.58356726","0.58304083","0.58110565","0.58109397","0.5809816","0.580944","0.58093315","0.5800911","0.5796492","0.5794249","0.5785868","0.57725745","0.5772552","0.5771563","0.57696015","0.5768524","0.57637054","0.57627934","0.57596225","0.575839","0.5757816","0.575475","0.57406676","0.573856","0.57382","0.57305086","0.57250303","0.5724823","0.5706314","0.570441","0.5687705","0.5682664","0.5672296","0.56670696","0.56621116","0.566108","0.56584996","0.56527084"],"string":"[\n \"0.6982893\",\n \"0.67954654\",\n \"0.67327917\",\n \"0.65675837\",\n \"0.6560407\",\n \"0.6475244\",\n \"0.64741606\",\n \"0.6416732\",\n \"0.6389593\",\n \"0.63582075\",\n \"0.6342166\",\n \"0.6282044\",\n \"0.6264434\",\n \"0.62190115\",\n \"0.6190914\",\n \"0.61886424\",\n \"0.61752874\",\n \"0.61643165\",\n \"0.61314374\",\n \"0.6065363\",\n \"0.6061954\",\n \"0.6032592\",\n \"0.6018692\",\n \"0.601484\",\n \"0.6006803\",\n \"0.6006371\",\n \"0.6003116\",\n \"0.6001448\",\n \"0.5997107\",\n \"0.59835184\",\n \"0.59798545\",\n \"0.5973025\",\n \"0.59688985\",\n \"0.5968867\",\n \"0.5968517\",\n \"0.59621924\",\n \"0.5957821\",\n \"0.59565663\",\n \"0.59427094\",\n \"0.5931806\",\n \"0.5930262\",\n \"0.5929543\",\n \"0.5921107\",\n \"0.59152794\",\n \"0.5913778\",\n \"0.59125125\",\n \"0.59114504\",\n \"0.590905\",\n \"0.5904003\",\n \"0.5900556\",\n \"0.5898378\",\n \"0.58954227\",\n \"0.58811885\",\n \"0.5879261\",\n \"0.5879261\",\n \"0.5879261\",\n \"0.5873312\",\n \"0.58590055\",\n \"0.58581686\",\n \"0.584782\",\n \"0.584782\",\n \"0.58375776\",\n \"0.58356726\",\n \"0.58304083\",\n \"0.58110565\",\n \"0.58109397\",\n \"0.5809816\",\n \"0.580944\",\n \"0.58093315\",\n \"0.5800911\",\n \"0.5796492\",\n \"0.5794249\",\n \"0.5785868\",\n \"0.57725745\",\n \"0.5772552\",\n \"0.5771563\",\n \"0.57696015\",\n \"0.5768524\",\n \"0.57637054\",\n \"0.57627934\",\n \"0.57596225\",\n \"0.575839\",\n \"0.5757816\",\n \"0.575475\",\n \"0.57406676\",\n \"0.573856\",\n \"0.57382\",\n \"0.57305086\",\n \"0.57250303\",\n \"0.5724823\",\n \"0.5706314\",\n \"0.570441\",\n \"0.5687705\",\n \"0.5682664\",\n \"0.5672296\",\n \"0.56670696\",\n \"0.56621116\",\n \"0.566108\",\n \"0.56584996\",\n \"0.56527084\"\n]"},"document_score":{"kind":"string","value":"0.80435836"},"document_rank":{"kind":"string","value":"0"}}},{"rowIdx":104798,"cells":{"query":{"kind":"string","value":"MAIN FUNCTIONS Handle Transaction Create"},"document":{"kind":"string","value":"func (main *Main) Create(e echo.Context) (err error) {\n\n\t// get request and validate\n\treq := new(request.Create)\n\te.Bind(req)\n\tif err = e.Validate(req); err != nil {\n\t\treturn rest.ConstructErrorResponse(e, exception.NewInputValidationFailed(err.Error()))\n\t}\n\t// map req to input data\n\treqData := input.NewNewTransactionCreate(\n\t\tmap[string]interface{}{\n\t\t\t\"Name\": req.Name,\n\t\t\t\"Qty\": req.Qty,\n\t\t\t\"Price\": req.Price,\n\t\t\t\"Weight\": req.Weight,\n\t\t\t\"Images\": req.Images,\n\t\t\t\"Description\": req.Description,\n\t\t},\n\t)\n\t//insert data to db\n\ttransaction, exc := TransactionModel.Create(reqData)\n\tif exc != nil {\n\t\treturn rest.ConstructErrorResponse(e, exc)\n\t}\n\tdata := map[string]contract.Model{\n\t\t\"created_transaction\": transaction,\n\t}\n\treturn rest.ConstructSuccessResponse(e, data)\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func TransactionCreate(c *gin.Context) {\n\tvar t models.Transaction\n\tbuffer, err := ioutil.ReadAll(c.Request.Body)\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusNotAcceptable, err)\n\t}\n\n\terr2 := jsonapi.Unmarshal(buffer, &t)\n\n\tif err2 != nil {\n\t\tparseFail := appError.JSONParseFailure\n\t\tparseFail.Detail = err2.Error()\n\t\tc.AbortWithError(http.StatusMethodNotAllowed, err2).\n\t\t\tSetMeta(parseFail)\n\t\treturn\n\t}\n\n\tt.CreatorID = c.Keys[\"CurrentUserID\"].(uint)\n\n\t// Validate our new transaction\n\tisValid, errApp := t.Validate()\n\n\tif isValid == false {\n\t\tc.AbortWithError(errApp.Status, errApp).\n\t\t\tSetMeta(errApp)\n\t\treturn\n\t}\n\n\tdatabase.DBCon.Create(&t)\n\n\tdatabase.DBCon.First(&t.Recipient, t.RecipientID)\n\tdatabase.DBCon.First(&t.Sender, t.SenderID)\n\tdatabase.DBCon.First(&t.Creator, t.CreatorID)\n\n\tdata, err := jsonapi.Marshal(&t)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tc.Data(http.StatusCreated, \"application/vnd.api+json\", data)\n}","func CreateTransaction(resp http.ResponseWriter, req *http.Request) {\n\tresp.Header().Add(\"Content-Type\", \"application/json\")\n\tvar transaction domain.Transaction\n\tdecoder := json.NewDecoder(req.Body)\n\n\terr := decoder.Decode(&transaction)\n\n\tif err != nil {\n\t\thttp.Error(resp, \"Bad Request\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tapiErr := service.CreateTransaction(transaction)\n\n\tif apiErr != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tencoder := json.NewEncoder(resp)\n\t\tencoder.Encode(apiErr)\n\t}\n\n\tresp.WriteHeader(http.StatusOK)\n\n\tjsonValue, _ := json.Marshal(utils.AppMsgs{http.StatusOK, \"Transaction was recorded\"})\n\n\tresp.Write(jsonValue)\n\n}","func (client *Client) CreateTransaction(txn *CreateTransaction) (_ *Response, err error) {\n\tpath := \"/transaction\"\n\turi := fmt.Sprintf(\"%s%s\", client.apiBaseURL, path)\n\n\tb, err := json.Marshal(txn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(http.MethodPost, uri, bytes.NewBuffer(b))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.performRequest(req, string(b))\n\treturn resp, err\n}","func (db *Datastore) Create(txn transaction.Transaction) error {\n\tfmt.Println(`Creating txn with data:`, txn)\n\t// your DB operations to transactions ...\n\treturn nil\n}","func createTransaction(request Request) (Transaction, error) {\n\tamount, err := strconv.ParseFloat(request.Load_amount[1:], 64)\n\ttransaction := Transaction{\n\t\tId: TransactionId(request.Id),\n\t\tCustomer_id: CustomerId(request.Customer_id),\n\t\tAmount: amount,\n\t\tTime: request.Time,\n\t}\n\n\treturn transaction, err\n}","func (db *MySQLDB) CreateTransaction(ctx context.Context, rec *TransactionRecord) (int, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"CreateTransaction\")\n\n\t// start db transaction\n\ttx, err := db.instance.BeginTx(ctx, nil)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.BeginTx got %s\", err.Error())\n\t\treturn 0, err\n\t}\n\n\t// create transaction record\n\ttrans, err := tx.ExecContext(ctx, \"INSERT INTO transactions(user_id, date, grand_total, discount, reason) VALUES(?,?,?,?,?)\", rec.UserID, rec.Date, 0, rec.Discount, rec.Reason)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\ttID, err := trans.LastInsertId()\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\tgrandTotal := 0\n\n\t//loop tx detail\n\tfor i := 0; i < len(rec.TransactionDetail); i++ {\n\t\tdetail := rec.TransactionDetail[i]\n\n\t\t//get product stock from products table\n\t\tp, err := db.GetProductByID(ctx, detail.ProductID)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, err\n\t\t}\n\n\t\t//check qty\n\t\tif p.Qty-detail.Qty < 0 {\n\t\t\tfLog.Errorf(\"product qty is not enough\")\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, fmt.Errorf(\"product qty is not enough\")\n\t\t}\n\n\t\tqty := p.Qty - detail.Qty\n\t\tsubTotal := p.Price * detail.Qty\n\t\tgrandTotal = grandTotal + subTotal\n\n\t\t//update qty from products table\n\t\t_, err = tx.ExecContext(ctx, \"UPDATE products SET qty=? WHERE id=?\", qty, detail.ProductID)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, err\n\t\t}\n\n\t\t//insert transaction detail\n\t\t_, err = tx.ExecContext(ctx, \"INSERT INTO transaction_detail(transaction_id, product_id, price, qty, sub_total) VALUES(?,?,?,?,?)\", tID, detail.ProductID, p.Price, detail.Qty, subTotal)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tgrandTotal = grandTotal - rec.Discount\n\n\t// update transaction grand total\n\t_, err = tx.ExecContext(ctx, \"UPDATE transactions SET grand_total=? WHERE id=?\", grandTotal, tID)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\t// commit transaction\n\terr = tx.Commit()\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\treturn int(tID), nil\n}","func (facade *BranchManagerFacade) createTransaction(srcAccountId, destAccountId string, amount float32) *Transaction {\n\n\tvar transaction = facade.transaction.create(srcAccountId, destAccountId, amount)\n\n\treturn transaction\n}","func CreateTransaction(transaction *models.Transaction) error {\n\tsession, collection := service.Connect(collectionName)\n\tdefer session.Close()\n\n\tif transaction.ID == \"\" {\n\t\ttransaction.ID = bson.NewObjectId()\n\t}\n\n\terr := collection.Insert(transaction)\n\n\treturn err\n}","func (repository *transactionRepository) CreateTransaction(transaction model.Transaction) (err error) {\n\ttx, err := repository.dbAdapter.Begin()\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\t_, err = tx.Exec(\n\t\t\"INSERT INTO transactions(id, account_id, operation_type, amount) VALUES(?,?,?,?)\",\n\t\ttransaction.GetID().GetValue(),\n\t\ttransaction.GetAccount().GetID().GetValue(),\n\t\ttransaction.GetOperationType().GetValue(),\n\t\ttransaction.GetAmountValueByOperationType(),\n\t)\n\n\tif err != nil {\n\t\ttx.Rollback()\n\n\t\treturn\n\t}\n\n\t_, err = tx.Exec(\n\t\t\"UPDATE accounts set credit_limit = ? where id = ?\",\n\t\ttransaction.GetAccount().GetAvailableCreditLimit().GetValue(),\n\t\ttransaction.GetAccount().GetID().GetValue(),\n\t)\n\n\tif err != nil {\n\t\ttx.Rollback()\n\n\t\treturn\n\t}\n\n\terr = tx.Commit()\n\n\treturn\n}","func CreateTransaction(from uint, to uint, amount int) {\n\tdb := helpers.ConnectDB().Debug()\n\tdefer db.Close()\n\n\ttransaction := &interfaces.Transaction{From: from, To: to, Amount: amount}\n\tdb.Create(&transaction)\n}","func (a API) CreateRawTransaction(cmd *btcjson.CreateRawTransactionCmd) (e error) {\n\tRPCHandlers[\"createrawtransaction\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}","func CreateTransactions(\n\trouter *mux.Router,\n\tpk cryptography.PrivateKey,\n\tuser users.User,\n\tserverSDK sdks.Servers,\n\troutePrefix string,\n\tsignedTrsBufferSize int,\n\tatomicTrsBufferSize int,\n\taggregatedTrsBufferSize int,\n\ttrsAggregationDelay time.Duration,\n) *Transactions {\n\n\t//channels:\n\tnewSignedTrs := make(chan signed_transactions.Transaction, signedTrsBufferSize)\n\tnewAtomicSignedTrs := make(chan signed_transactions.AtomicTransaction, atomicTrsBufferSize)\n\tnewAggregatedTrs := make(chan aggregated_transactions.Transactions, aggregatedTrsBufferSize)\n\n\t//factories:\n\tmetaDataBuilderFactory := concrete_metadata.CreateBuilderFactory()\n\thtBuilderFactory := concrete_hashtrees.CreateHashTreeBuilderFactory()\n\tpublicKeyBuilderFactory := concrete_cryptography.CreatePublicKeyBuilderFactory()\n\tsigBuilderFactory := concrete_cryptography.CreateSignatureBuilderFactory(publicKeyBuilderFactory)\n\tuserSigBuilderFactory := concrete_users.CreateSignatureBuilderFactory(sigBuilderFactory, htBuilderFactory, metaDataBuilderFactory)\n\n\t//transactions and blocks factories:\n\tblockChainMetaDataBuilderFactory := concrete_blockchain_metadata.CreateBuilderFactory()\n\tsignedTrsBuilderFactory := concrete_signed_transactions.CreateTransactionBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tsignedTransBuilderFactory := concrete_signed_transactions.CreateTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tsignedAtomicTransBuilderFactory := concrete_signed_transactions.CreateAtomicTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tatomicSignedTrsBuilderFactory := concrete_signed_transactions.CreateAtomicTransactionBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tsignedAggregatedTrsBuilderFactory := concrete_aggregated_transactions.CreateTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\n\t//create the leader SDK:\n\tleadSDK := concrete_sdks.CreateLeaders(userSigBuilderFactory, routePrefix, pk, user)\n\n\t//create the transaction API:\n\ttransactionsAPI := apis.CreateTransactions(\n\t\troutePrefix,\n\t\trouter,\n\t\tsignedTrsBuilderFactory,\n\t\tatomicSignedTrsBuilderFactory,\n\t\tnewSignedTrs,\n\t\tnewAtomicSignedTrs,\n\t)\n\n\t//create the transaction agent:\n\ttrsAgent := agents.CreatePushTransactionsToLeaders(\n\t\tsignedAggregatedTrsBuilderFactory,\n\t\tsignedTransBuilderFactory,\n\t\tsignedAtomicTransBuilderFactory,\n\t\ttrsAggregationDelay,\n\t\tnewSignedTrs,\n\t\tnewAtomicSignedTrs,\n\t\tnewAggregatedTrs,\n\t)\n\n\tout := Transactions{\n\t\tapi: transactionsAPI,\n\t\tagent: trsAgent,\n\t\tleadSDK: leadSDK,\n\t\tservSDK: serverSDK,\n\t\tnewAggregatedTrs: newAggregatedTrs,\n\t}\n\n\treturn &out\n}","func (u *User) CreateTransaction(nodeID, data string, idempotencyKey ...string) (map[string]interface{}, error) {\n\tlog.info(\"========== CREATE TRANSACTION ==========\")\n\turl := buildURL(path[\"users\"], u.UserID, path[\"nodes\"], nodeID, path[\"transactions\"])\n\n\treturn u.do(\"POST\", url, data, idempotencyKey)\n}","func (f formatter) CreateTransaction(t bill.Transaction, ins, outs []string) (string, error) {\n\titem := transactionItem{\n\t\tRaw: t.Raw,\n\t\tAt: t.At.Format(\"2006-01-02\"),\n\t\tPayee: t.Payee,\n\t\tTitle: t.Title,\n\t\tPayeeAccounts: ins,\n\t\tAccounts: outs,\n\t\tAmount: t.Amount,\n\t\tCurrency: t.Currency,\n\t}\n\tif item.PayeeAccounts == nil || len(item.PayeeAccounts) == 0 {\n\t\titem.PayeeAccounts = []string{\"TODO\"}\n\t}\n\tif item.Accounts == nil || len(item.Accounts) == 0 {\n\t\titem.Accounts = []string{\"TODO\"}\n\t}\n\n\tvar tplBytes bytes.Buffer\n\ttpl, err := template.New(\"transaction\").Parse(transactionTpl)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"template error, should be a bug: %w\", err)\n\t}\n\n\terr = tpl.Execute(&tplBytes, item)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"template exec error, should be a bug: %w\", err)\n\t}\n\n\treturn tplBytes.String(), nil\n}","func (transaction *Transaction) create(srcAccountId, destAccountId string, amount float32) *Transaction {\n\n\tfmt.Println(\"creating transaction\")\n\ttransaction.srcAccountId = srcAccountId\n\ttransaction.destAccountId = destAccountId\n\ttransaction.amount = amount\n\treturn transaction\n}","func createTransaction(\n\tctx context.Context,\n\tdb storage.Database,\n\tappserviceID string,\n) (\n\ttransactionJSON []byte,\n\ttxnID, maxID int,\n\teventsRemaining bool,\n\terr error,\n) {\n\t// Retrieve the latest events from the DB (will return old events if they weren't successfully sent)\n\ttxnID, maxID, events, eventsRemaining, err := db.GetEventsWithAppServiceID(ctx, appserviceID, transactionBatchSize)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"appservice\": appserviceID,\n\t\t}).WithError(err).Fatalf(\"appservice worker unable to read queued events from DB\")\n\n\t\treturn\n\t}\n\n\t// Check if these events do not already have a transaction ID\n\tif txnID == -1 {\n\t\t// If not, grab next available ID from the DB\n\t\ttxnID, err = db.GetLatestTxnID(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, 0, 0, false, err\n\t\t}\n\n\t\t// Mark new events with current transactionID\n\t\tif err = db.UpdateTxnIDForEvents(ctx, appserviceID, maxID, txnID); err != nil {\n\t\t\treturn nil, 0, 0, false, err\n\t\t}\n\t}\n\n\tvar ev []*gomatrixserverlib.HeaderedEvent\n\tfor i := range events {\n\t\tev = append(ev, &events[i])\n\t}\n\n\t// Create a transaction and store the events inside\n\ttransaction := gomatrixserverlib.ApplicationServiceTransaction{\n\t\tEvents: gomatrixserverlib.HeaderedToClientEvents(ev, gomatrixserverlib.FormatAll),\n\t}\n\n\ttransactionJSON, err = json.Marshal(transaction)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}","func (tm *TransactionMutex) Create(transactionType TransactionType, fileUUID string) {\n\ttm.mu.Lock()\n\tdefer tm.mu.Unlock()\n\tnewTransaction := Transaction{\n\t\tUUID: NewUUID(),\n\t\tCreationTimestamp: time.Now().Unix(),\n\t\tType: transactionType,\n\t\tTargetFileUUID: fileUUID,\n\t\tVersion: config.Version,\n\t}\n\ttm.Transactions = append(tm.Transactions, newTransaction)\n}","func (tr *Transactor) CreateTransaction(ctx context.Context, ipfsHash string, tags []Tag, w WalletSigner, amount string, data []byte, target string) (*Transaction, error) {\n\tlastTx, err := tr.Client.TxAnchor(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tprice, err := tr.Client.GetReward(ctx, data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Non encoded transaction fields\n\ttx := NewTransaction(\n\t\ttags,\n\t\tipfsHash,\n\t\tlastTx,\n\t\tw.PubKeyModulus(),\n\t\tamount,\n\t\ttarget,\n\t\tdata,\n\t\tprice,\n\t)\n\n\treturn tx, nil\n}","func CreateTransaction(session gocqlx.Session, transaction *Transaction) (*Transaction, error) {\n\tuuid, _ := gocql.RandomUUID()\n\ttransaction.ID = uuid\n\ttransaction.CreatedAt = time.Now()\n\n\tlog.Print(transaction)\n\n\tq := session.Query(transactionTable.Insert()).BindStruct(transaction)\n\tif err := q.ExecRelease(); err != nil {\n\t\treturn transaction, err\n\t}\n\treturn transaction, nil\n}","func (u *createTrans) CreateTransaction(transaction *models.Transaction) error {\n\n\tacc, err := u.accountServ.Get(transaction.Account.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnewCredit := acc.AvailableCreditLimit + transaction.Amount\n\tif newCredit < 0 {\n\t\treturn models.ErrAmountBeyondCreditLimit\n\t}\n\n\terr = u.tranServ.Create(transaction)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tacc.AvailableCreditLimit = newCredit\n\treturn u.accountServ.UpdateCreditLimit(acc)\n}","func (s *TransactionService) Create(walletID string, tp *TransactionParams) (*Transaction, error) {\n\tu := fmt.Sprintf(\"/kms/wallets/%s/transactions/\", walletID)\n\ttransaction := &Transaction{}\n\tp := &Params{}\n\tp.SetAuthProvider(s.auth)\n\terr := s.client.Call(http.MethodPost, u, tp, transaction, p)\n\treturn transaction, err\n}","func (s *Service) CreateTx(tx *sql.Tx, headerID uint, ms Models) error {\n\treturn s.storage.CreateTx(tx, headerID, ms)\n}","func addTransaction(db meddler.DB, lender *User, debtor *User, amount int, expense *Expense) (*Transaction, error) {\n trans := new(Transaction)\n trans.LenderId = lender.Id\n trans.DebtorId = debtor.Id\n trans.Amount = amount\n trans.Date = expense.Date\n trans.ExpenseId = expense.Id\n\n err := meddler.Insert(db, \"transactions\", trans)\n if err != nil {\n return nil, err\n }\n\n lender.UpdateBalance(db, amount)\n debtor.UpdateBalance(db, -amount)\n\n return trans, nil\n}","func (c Client) createTransaction(merchantID string, body io.Reader) (*TransactionID, error) {\n\tpath := fmt.Sprintf(\"/merchants/%s/transactions\", merchantID)\n\treq, err := http.NewRequest(\"POST\", c.getURL(path), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar marshalled map[string]*TransactionID\n\treturn marshalled[\"transaction\"], c.executeRequestAndMarshal(req, &marshalled)\n}","func (h *Harness) CreateTransaction(\n\ttargetOutputs []*wire.TxOut,\n\tfeeRate amt.Amount, change bool,\n) (*wire.MsgTx, error) {\n\treturn h.wallet.CreateTransaction(targetOutputs, feeRate, change)\n}","func CreateTransaction(from, to, amount string) (txeB64 string) {\n\n\ttx, err := build.Transaction(\n\t\tbuild.SourceAccount{AddressOrSeed: from},\n\t\tbuild.PublicNetwork,\n\t\t//b.AutoSequence{SequenceProvider: horizon.DefaultPublicNetClient}, ???\n\t\tbuild.Payment(\n\t\t\tbuild.Destination{AddressOrSeed: to},\n\t\t\tbuild.NativeAmount{Amount: amount},\n\t\t),\n\t)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\ttxe, err := tx.Sign(from)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\ttxeB64, err = txe.Base64()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn\n}","func Transaction(c *gin.Context) {\n\n\tt_type,_ := strconv.Atoi(c.PostForm(\"transaction_type\")) // 1 : sales , 2 : missing products (hilang)\n\tstatus := 1\n\tmessage := \"Success\"\n var responseTransaction ResponseTransaction\n\tvar newstocks int\n\tvar products Products\n\tvar products_arr []Products\n\tvar stock_ins_arr []Stock_Ins\n\tvar stock_outs Stock_Outs\n\tvar stock_ins Stock_Ins\n\tvar note string\n\ttransaction_id := \"\"\n\tsellPrice,_ := strconv.Atoi(c.PostForm(\"sell_price\"))\n\tvar buyPrice int\n\tqtY,_ := strconv.Atoi(c.PostForm(\"qty\"))\n\tcurrentdatetime := time.Now().Format(\"2006-01-02 15:04:05\")\n\tdb := InitDb() //db intiate\n\t//get data products\n\tdb.Where(\"sku = ?\", c.PostForm(\"sku\")).First(&products).Limit(1).Scan(&products_arr)\n\n\t//check if the sku is exist?\n\tif(len(products_arr) > 0) {\n\t\ttx := db.Begin()\n\n\t\t/**\n\t * Identify product is gone / transaction by sales\n\t */\n\n\t\tif (t_type == 1) {\n\n\t\t\ttransaction_id = generateTransactionID()\n\n\t\t\t//get data products\n\t\t\tdb.Where(\"sku = ?\", c.PostForm(\"sku\")).First(&stock_ins).Limit(1).Scan(&stock_ins_arr)\n\n\t\t\t// get the data stock after transaction\n\t\t\tfor i,element := range stock_ins_arr{\n\t\t\t\tif (i == 0) {\n\t\t\t\t\tbuyPrice = element.Buy_Price\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tnote = \"Pesanan \"+transaction_id\n\t\t\ttransactions := Transactions{Id:transaction_id,Buy_Price:buyPrice,Sell_Price:sellPrice,Qty:qtY,Sku:c.PostForm(\"sku\"),Created_Date:currentdatetime}\n\t\t\tif err := tx.Create(&transactions).Error; err != nil {\n\t\t\t\ttx.Rollback()\n\t\t\t\tstatus = 0\n\t\t\t\tmessage = \"failed to insert data transaction\"\n\t\t\t}\n\n\n\t\t} else if (t_type == 2) {\n\n\t\t\tnote = \"Barang Hilang\"\n\n\t\t}\n\t\t//insert data to stock_outs\n\t\tstock_outs = Stock_Outs{Sku:c.PostForm(\"sku\"),Created_Date:currentdatetime,Qty:qtY,Note:note,Transaction_Id:transaction_id}\n\t\tif err := tx.Create(&stock_outs).Error; err != nil {\n\t\t\ttx.Rollback()\n\t\t\tstatus = 0\n\t\t\tmessage = \"failed to insert data stocks_outs\"\n\t\t}\n\n\t\t// get the data stock after transaction\n\t\tfor i,element := range products_arr{\n\t\t\tif (i == 0) {\n\t\t\t\tnewstocks = element.Stocks - qtY\n\t\t\t}\n\t\t}\n\n\t\t//update product stocks in table products\n\t\tif err := tx.Model(&products).Where(\"sku = ?\", c.PostForm(\"sku\")).Update(\"stocks\", newstocks).Error; err != nil {\n\t\t\ttx.Rollback()\n\t\t\tstatus = 0\n\t\t\tmessage = \"failed to update data products\"\n\t\t}\n\n\n\t\t//transaction commit\n\t\ttx.Commit()\n\t}else{\n\t\tstatus = 0\n\t\tmessage = \"SKU Not found!\"\n\t}\n\n\tif status == 1{\n\t\tresponseTransaction = ResponseTransaction{Status:status,Message:message,Data:DataTransaction{Sku:c.PostForm(\"sku\"),Buy_Price:buyPrice,Sell_Price:sellPrice,Created_Date:currentdatetime,Product_name:c.PostForm(\"product_name\"),Stocks:newstocks,Transaction_Id:transaction_id}}\n\t}else{\n\t\tresponseTransaction = ResponseTransaction{Status:status,Message:message}\n\t}\n\n\t// Close connection database\n\tdefer db.Close()\n\tc.JSON(200, responseTransaction)\n}","func (c *Constructor) createTransaction(\n\tctx context.Context,\n\tintent []*types.Operation,\n) (*types.TransactionIdentifier, string, error) {\n\tmetadataRequest, err := c.helper.Preprocess(\n\t\tctx,\n\t\tc.network,\n\t\tintent,\n\t\tnil,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to preprocess\", err)\n\t}\n\n\trequiredMetadata, err := c.helper.Metadata(\n\t\tctx,\n\t\tc.network,\n\t\tmetadataRequest,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to construct metadata\", err)\n\t}\n\n\tunsignedTransaction, payloads, err := c.helper.Payloads(\n\t\tctx,\n\t\tc.network,\n\t\tintent,\n\t\trequiredMetadata,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to construct payloads\", err)\n\t}\n\n\tparsedOps, signers, _, err := c.helper.Parse(\n\t\tctx,\n\t\tc.network,\n\t\tfalse,\n\t\tunsignedTransaction,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to parse unsigned transaction\", err)\n\t}\n\n\tif len(signers) != 0 {\n\t\treturn nil, \"\", fmt.Errorf(\n\t\t\t\"signers should be empty in unsigned transaction but found %d\",\n\t\t\tlen(signers),\n\t\t)\n\t}\n\n\tif err := c.parser.ExpectedOperations(intent, parsedOps, false, false); err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unsigned parsed ops do not match intent\", err)\n\t}\n\n\tsignatures, err := c.helper.Sign(ctx, payloads)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to sign payloads\", err)\n\t}\n\n\tnetworkTransaction, err := c.helper.Combine(\n\t\tctx,\n\t\tc.network,\n\t\tunsignedTransaction,\n\t\tsignatures,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to combine signatures\", err)\n\t}\n\n\tsignedParsedOps, signers, _, err := c.helper.Parse(\n\t\tctx,\n\t\tc.network,\n\t\ttrue,\n\t\tnetworkTransaction,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to parse signed transaction\", err)\n\t}\n\n\tif err := c.parser.ExpectedOperations(intent, signedParsedOps, false, false); err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: signed parsed ops do not match intent\", err)\n\t}\n\n\tif err := parser.ExpectedSigners(payloads, signers); err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: signed transactions signers do not match intent\", err)\n\t}\n\n\ttransactionIdentifier, err := c.helper.Hash(\n\t\tctx,\n\t\tc.network,\n\t\tnetworkTransaction,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to get transaction hash\", err)\n\t}\n\n\treturn transactionIdentifier, networkTransaction, nil\n}","func (service ServiceTx) Create(edgeJob *portainer.EdgeJob) error {\n\treturn service.CreateWithID(portainer.EdgeJobID(service.GetNextIdentifier()), edgeJob)\n}","func (c *TransactionClient) Create() *TransactionCreate {\n\tmutation := newTransactionMutation(c.config, OpCreate)\n\treturn &TransactionCreate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}","func (gw *Gateway) CreateTransaction(params wallet.CreateTransactionParams) (*coin.Transaction, []wallet.UxBalance, error) {\n\tif !gw.Config.EnableWalletAPI {\n\t\treturn nil, nil, wallet.ErrWalletAPIDisabled\n\t}\n\n\tvar txn *coin.Transaction\n\tvar inputs []wallet.UxBalance\n\tvar err error\n\tgw.strand(\"CreateTransaction\", func() {\n\t\ttxn, inputs, err = gw.v.CreateTransaction(params)\n\t})\n\treturn txn, inputs, err\n}","func (p *PsqlInvoiceHeader) CreateTx(tx *sql.Tx, m *invoiceheader.Model) error {\n\tstmt, err := tx.Prepare(psqlCreateInvoiceHeader)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer stmt.Close()\n\n\treturn stmt.QueryRow(m.Client).Scan(&m.ID, &m.CreatedAt)\n}","func (c Client) CreateTransaction(merchantID string, dto TransactionDTO) (*TransactionID, error) {\n\tb, err := json.Marshal(dto)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.createTransaction(merchantID, bytes.NewBuffer(b))\n}","func (t *explainTablet) CreateTransaction(ctx context.Context, target *querypb.Target, dtid string, participants []*querypb.Target) (err error) {\n\tt.mu.Lock()\n\tt.currentTime = t.vte.batchTime.Wait()\n\tt.mu.Unlock()\n\treturn t.tsv.CreateTransaction(ctx, target, dtid, participants)\n}","func (u *User) CreateDummyTransaction(nodeID string, queryParams ...string) (map[string]interface{}, error) {\n\tlog.info(\"========== CREATE DUMMY TRANSACTION ==========\")\n\turl := buildURL(path[\"users\"], u.UserID, path[\"nodes\"], nodeID) + \"/dummy-tran\"\n\n\treturn u.do(\"GET\", url, \"\", queryParams)\n}","func (tx *Tx) Create(ls Lease) error {\n\ttx.leases = append(tx.leases, ls)\n\ttx.ops = append(tx.ops, Op{\n\t\tType: Create,\n\t\tLease: ls,\n\t})\n\tsort.Sort(tx.leases)\n\treturn nil\n}","func (m *Mysql) CreateTX(tx *sql.Tx, paramAudit *paramAuditDom.ParamsAudit) error {\n\tstmt, err := tx.Prepare(mysqlInsert)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\tnow := utils.Now()\n\terr = mysql.ExecAffectingOneRow(\n\t\tstmt,\n\t\tparamAudit.PrevParam.ParamID,\n\t\tparamAudit.PrevParam.Name,\n\t\tparamAudit.PrevParam.Value,\n\t\tparamAudit.PrevParam.Active,\n\t\tparamAudit.Param.Name,\n\t\tparamAudit.Param.Value,\n\t\tparamAudit.Param.Active,\n\t\tparamAudit.SetUserID,\n\t\tnow[\"date\"], now[\"time\"], now[\"unix\"],\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}","func (nos *NewOrderServiceImpl) ProcessTransaction(req *models.NewOrder) (*models.NewOrderOutput, error) {\n\tlog.Printf(\"Starting the New Order Transaction for row: c=%d w=%d d=%d n=%d\", req.CustomerID, req.WarehouseID, req.DistrictID, req.UniqueItems)\n\n\tresult, err := nos.execute(req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error occured while executing the new order transaction. Err: %v\", err)\n\t}\n\n\tlog.Printf(\"Completed the New Order Transaction for row: c=%d w=%d d=%d n=%d\", req.CustomerID, req.WarehouseID, req.DistrictID, req.UniqueItems)\n\treturn result, nil\n}","func (c *Jrpc) CreateRawTransaction(in *pty.ReqCreatePrivacyTx, result *interface{}) error {\n\treply, err := c.cli.CreateRawTransaction(context.Background(), in)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*result = hex.EncodeToString(types.Encode(reply))\n\treturn err\n}","func (b *Store) CreateTx(ctx context.Context, tx *sql.Tx, userID, username, password string) error {\n\terr := permission.LimitCheckAny(ctx, permission.System, permission.Admin, permission.MatchUser(userID))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = validate.Many(\n\t\tvalidate.UUID(\"UserID\", userID),\n\t\tvalidate.UserName(\"UserName\", username),\n\t\tvalidate.Text(\"Password\", password, 8, 200),\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thashedPassword, err := bcrypt.GenerateFromPassword([]byte(password), passCost)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = tx.StmtContext(ctx, b.insert).ExecContext(ctx, userID, username, string(hashedPassword))\n\treturn err\n}","func (s *transactionStore) Create(ctx context.Context, transaction *configapi.Transaction) error {\n\tif transaction.ID == \"\" {\n\t\ttransaction.ID = newTransactionID()\n\t}\n\tif transaction.Version != 0 {\n\t\treturn errors.NewInvalid(\"not a new object\")\n\t}\n\tif transaction.Revision != 0 {\n\t\treturn errors.NewInvalid(\"not a new object\")\n\t}\n\ttransaction.Revision = 1\n\ttransaction.Created = time.Now()\n\ttransaction.Updated = time.Now()\n\n\t// Append a new entry to the transaction log.\n\tentry, err := s.transactions.Append(ctx, transaction.ID, transaction)\n\tif err != nil {\n\t\treturn errors.FromAtomix(err)\n\t}\n\ttransaction.Index = configapi.Index(entry.Index)\n\ttransaction.Version = uint64(entry.Version)\n\treturn nil\n}","func (c *Constructor) CreateTransactions(\n\tctx context.Context,\n\tclearBroadcasts bool,\n) error {\n\t// Before starting loop, delete any pending broadcasts if configuration\n\t// indicates to do so.\n\tif clearBroadcasts {\n\t\tbroadcasts, err := c.helper.ClearBroadcasts(ctx)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to clear broadcasts\", err)\n\t\t}\n\n\t\tlog.Printf(\n\t\t\t\"Cleared pending %d broadcasts: %s\\n\",\n\t\t\tlen(broadcasts),\n\t\t\ttypes.PrettyPrintStruct(broadcasts),\n\t\t)\n\t}\n\n\tfor ctx.Err() == nil {\n\t\tsender, balance, coinIdentifier, err := c.findSender(ctx)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to find sender\", err)\n\t\t}\n\n\t\t// Determine Action\n\t\tscenarioCtx, scenarioOps, err := c.generateScenario(\n\t\t\tctx,\n\t\t\tsender,\n\t\t\tbalance,\n\t\t\tcoinIdentifier,\n\t\t)\n\t\tif errors.Is(err, ErrInsufficientFunds) {\n\t\t\tbroadcasts, err := c.helper.AllBroadcasts(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"%w: unable to get broadcasts\", err)\n\t\t\t}\n\n\t\t\tif len(broadcasts) > 0 {\n\t\t\t\t// we will wait for in-flight to process\n\t\t\t\ttime.Sleep(defaultSleepTime * time.Second)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err := c.generateNewAndRequest(ctx); err != nil {\n\t\t\t\treturn fmt.Errorf(\"%w: unable to generate new address\", err)\n\t\t\t}\n\n\t\t\tcontinue\n\t\t} else if err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to generate intent\", err)\n\t\t}\n\n\t\tintent, err := scenario.PopulateScenario(ctx, scenarioCtx, scenarioOps)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to populate scenario\", err)\n\t\t}\n\n\t\t// Create transaction\n\t\ttransactionIdentifier, networkTransaction, err := c.createTransaction(ctx, intent)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\n\t\t\t\t\"%w: unable to create transaction with operations %s\",\n\t\t\t\terr,\n\t\t\t\ttypes.PrettyPrintStruct(intent),\n\t\t\t)\n\t\t}\n\n\t\tlogger.LogScenario(scenarioCtx, transactionIdentifier, c.currency)\n\n\t\t// Broadcast Transaction\n\t\terr = c.helper.Broadcast(\n\t\t\tctx,\n\t\t\tsender,\n\t\t\tintent,\n\t\t\ttransactionIdentifier,\n\t\t\tnetworkTransaction,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to enqueue transaction for broadcast\", err)\n\t\t}\n\n\t\tif err := c.handler.TransactionCreated(ctx, sender, transactionIdentifier); err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to handle transaction creation\", err)\n\t\t}\n\t}\n\n\treturn ctx.Err()\n}","func Transaction(rt *Runtime, c chan goengage.Fundraise) (err error) {\n\trt.Log.Println(\"Transaction: start\")\n\tfor true {\n\t\tr, ok := <-c\n\t\tif !ok {\n\t\t\tbreak\n\t\t}\n\n\t\trt.Log.Printf(\"%v Transaction\\n\", r.ActivityID)\n\t\tif rt.GoodYear(r.ActivityDate) {\n\t\t\tif len(r.Transactions) != 0 {\n\t\t\t\tfor _, c := range r.Transactions {\n\t\t\t\t\tc.ActivityID = r.ActivityID\n\t\t\t\t\trt.DB.Create(&c)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\trt.Log.Println(\"Transaction: end\")\n\treturn nil\n}","func (a *ChainAdaptor) CreateUtxoTransaction(req *proto.CreateUtxoTransactionRequest) (*proto.CreateUtxoTransactionReply, error) {\n\tvinNum := len(req.Vins)\n\tvar totalAmountIn, totalAmountOut int64\n\n\tif vinNum == 0 {\n\t\terr := fmt.Errorf(\"no Vin in req:%v\", req)\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\t// check the Fee\n\tfee, ok := big.NewInt(0).SetString(req.Fee, 0)\n\tif !ok {\n\t\terr := errors.New(\"CreateTransaction, fail to get fee\")\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\tfor _, in := range req.Vins {\n\t\ttotalAmountIn += in.Amount\n\t}\n\n\tfor _, out := range req.Vouts {\n\t\ttotalAmountOut += out.Amount\n\t}\n\n\tif totalAmountIn != totalAmountOut+fee.Int64() {\n\t\terr := errors.New(\"CreateTransaction, total amount in != total amount out + fee\")\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\trawTx, err := a.createRawTx(req.Vins, req.Vouts)\n\tif err != nil {\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\tbuf := bytes.NewBuffer(make([]byte, 0, rawTx.SerializeSize()))\n\terr = rawTx.Serialize(buf)\n\tif err != nil {\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\t// build the pkScript and Generate signhash for each Vin,\n\tsignHashes, err := a.calcSignHashes(req.Vins, req.Vouts)\n\tif err != nil {\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\tlog.Info(\"CreateTransaction\", \"usigned tx\", hex.EncodeToString(buf.Bytes()))\n\n\treturn &proto.CreateUtxoTransactionReply{\n\t\tCode: proto.ReturnCode_SUCCESS,\n\t\tTxData: buf.Bytes(),\n\t\tSignHashes: signHashes,\n\t}, nil\n}","func (c *RPC) CreateTransaction(recvAddr string, amount uint64) (*coin.Transaction, error) {\n\t// TODO -- this can support sending to multiple receivers at once,\n\t// which would be necessary if the exchange was busy\n\tsendAmount := cli.SendAmount{\n\t\tAddr: recvAddr,\n\t\tCoins: amount,\n\t}\n\n\tif err := validateSendAmount(sendAmount); err != nil {\n\t\treturn nil, err\n\t}\n\n\ttxn, err := cli.CreateRawTxFromWallet(c.rpcClient, c.walletFile, c.changeAddr, []cli.SendAmount{sendAmount})\n\tif err != nil {\n\t\treturn nil, RPCError{err}\n\t}\n\n\treturn txn, nil\n}","func Db_create(db_name string) string{\n\t// Open the my.db data_backup file in your current directory.\n\t// It will be created if it doesn't exist.\n\tdb, err := bolt.Open(db_name, 0600, nil)\n\t//defer db.Close()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\ttx, err := db.Begin(true)\n\tif err != nil {\n\t\tfmt.Println(\"Error : \",err)\n\t\tdb.Close()\n\t\treturn \"ERROR,DB_COMMIT\"\n\t}\n\tfmt.Printf(\"Created DB with ID %d\", tx.ID())\n\tdb.Close()\n\treturn \"OK,DB_COMMIT\"\n}","func (t *SimpleChaincode) create_event(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\n\tvar tEvent TransactionEvent\n\t\n\ttranID \t\t\t:= \"\\\"TranID\\\":\\\"\"+args[0]+\"\\\", \"\n\tsenderName \t\t:= \"\\\"SenderName\\\":\\\"\"+args[1]+\"\\\", \"\n\tsenderCountry := \"\\\"SenderCountry\\\":\\\"\"+args[2]+\"\\\", \"\n\treceiverName \t:= \"\\\"ReceiverName\\\":\\\"\"+args[3]+\"\\\", \"\n\treceiverCountry := \"\\\"ReceiverCountry\\\":\\\"\"+args[4]+\"\\\", \"\n\tamount \t\t\t:= \"\\\"Amount\\\":\\\"\"+args[5]+\"\\\", \"\n\n // Concatenates the variables to create the total JSON object\n\tevent_json := \"{\"+tranID+senderName+senderCountry+receiverName+receiverCountry+amount+\"}\" \t\t\n\t// Convert the JSON defined above into a TransactionEvent object for go\n\terr := json.Unmarshal([]byte(event_json), &tEvent)\t\t\t\t\t\t\t\t\t\t\n\tif err != nil { \n\t\treturn nil, errors.New(\"Invalid JSON object\") \n\t}\n\n\tbytes, err := json.Marshal(tEvent)\n\tif err != nil { \n\t\treturn nil, errors.New(\"Error converting transaction event\") \n\t}\n\n\t// Save new tran event record\n\terr = stub.PutState(tEvent.TranID, bytes)\n\tif err != nil { \n\t\tfmt.Printf(\"create_event: Error storing transaction event: %s\", err); \n\t\treturn nil, errors.New(\"Error storing transaction event\") \n\t}\n\n\t// Update tranIDs with newly created ID and store it in chain.\n\tbytes, err = stub.GetState(\"tranIDs\")\n\tif err != nil { \n\t\treturn nil, errors.New(\"Unable to get tranIDs\") \n\t}\n\n\tvar tranHld TRAN_Holder\n\terr = json.Unmarshal(bytes, &tranHld)\n\tif err != nil {\t\n\t\treturn nil, errors.New(\"Corrupt TRAN_Holder record\") \n\t}\n\n\ttranHld.TranIDs = append(tranHld.TranIDs, args[0])\n\tbytes, err = json.Marshal(tranHld)\n\n\terr = stub.PutState(\"tranIDs\", bytes)\n\tif err != nil { \n\t\tfmt.Printf(\"create_event: Error storing TranIDs: %s\", err); \n\t\treturn nil, errors.New(\"Error storing TranIDs\") \n\t}\n\n\treturn nil, nil \n}","func (t *Procure2Pay) CreatePurchaseOrder(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n var objpurchaseOrder purchaseOrder\n\tvar objitem item\n\tvar err error\n\t\n\tfmt.Println(\"Entering CreatePurchaseOrder\")\n\n\tif len(args) < 1 {\n\t\tfmt.Println(\"Invalid number of args\")\n\t\treturn shim.Error(err.Error())\n\t}\n\n\tfmt.Println(\"Args [0] is : %v\\n\", args[0])\n\n\t//unmarshal customerInfo data from UI to \"customerInfo\" struct\n\terr = json.Unmarshal([]byte(args[0]), &objpurchaseOrder)\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to unmarshal CreatePurchaseOrder input purchaseOrder: %s\\n\", err)\n\t\treturn shim.Error(err.Error())\n\t\t}\n\n\tfmt.Println(\"purchase order object PO ID variable value is : %s\\n\", objpurchaseOrder.POID)\n\tfmt.Println(\"purchase order object PO ID variable value is : %s\\n\", objpurchaseOrder.Quantity)\n\n\t// Data insertion for Couch DB starts here \n\ttransJSONasBytes, err := json.Marshal(objpurchaseOrder)\n\terr = stub.PutState(objpurchaseOrder.POID, transJSONasBytes)\n\t// Data insertion for Couch DB ends here\n\n\t//unmarshal LoanTransactions data from UI to \"LoanTransactions\" struct\n\terr = json.Unmarshal([]byte(args[0]), &objitem)\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to unmarshal CreatePurchaseOrder input purchaseOrder: %s\\n\", err)\n\t\treturn shim.Error(err.Error())\n\t\t}\n\n\tfmt.Println(\"item object Item ID variable value is : %s\\n\", objitem.ItemID)\n\n\t// Data insertion for Couch DB starts here \n\ttransJSONasBytesLoan, err := json.Marshal(objitem)\n\terr = stub.PutState(objitem.ItemID, transJSONasBytesLoan)\n\t// Data insertion for Couch DB ends here\n\n\tfmt.Println(\"Create Purchase Order Successfully Done\")\n\n\tif err != nil {\n\t\tfmt.Printf(\"\\nUnable to make transevent inputs : %v \", err)\n\t\treturn shim.Error(err.Error())\n\t\t//return nil,nil\n\t}\n\treturn shim.Success(nil)\n}","func (r RepresentativeRepo) CreateTx(tx *sql.Tx, repr d.Representative) error {\n\terr := r.UserRepo.CreateTx(tx, repr.User)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconst insertQuery = `INSERT INTO \"Representative\"(representative_id, job_title, ref_user, ref_company)\n\tVALUES ($1, $2, $3, $4);`\n\t_, err = tx.Exec(insertQuery,\n\t\trepr.ID,\n\t\trepr.JobTitle,\n\t\trepr.User.ID,\n\t\trepr.CompanyID,\n\t)\n\n\tif err != nil {\n\t\t_ = tx.Rollback()\n\t\treturn err\n\t}\n\n\treturn nil\n}","func (p *PgDb) CreateEmpTx(r *model.Emp, tx *sqlx.Tx, isValidate bool) (*model.Emp, error) {\n\tcnt := \"postgres (p *PgDb) CreateEmpTx\" // Имя текущего метода для логирования\n\t//mylog.PrintfDebug(\"[DEBUG] %v - START, param: '%+v'\", cnt, r)\n\n\t// Проверяем определен ли контекст транзакции\n\tif tx == nil {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - tx *sqlx.Tx is NULL\", cnt)\n\t\tlog.Printf(errM)\n\t\treturn nil, errors.New(errM)\n\t}\n\n\t//=====================================================================\n\t// Добавить валидацию входной структуры\n\t//=====================================================================\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\t// Если запускаем с проверками\n\tif isValidate {\n\t\t{ //Если Dept NULL или НЕ существует, то ошибка\n\t\t\tif !r.Deptno.Valid {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - r.Deptno is NULL\", cnt)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.New(errM)\n\t\t\t}\n\t\t\tdeptno := int(r.Deptno.Int64)\n\t\t\t// Запрос в транзакции\n\t\t\tdeptExists, err := p.DeptExistsTx(deptno, tx)\n\t\t\tif err != nil {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.DeptExistsTx(deptno, tx), args = '%v'\", cnt, deptno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t\t}\n\t\t\tif !deptExists {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - dept '%v' does not exist\", cnt, deptno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.New(errM)\n\t\t\t}\n\t\t\t//mylog.PrintfDebug(\"[DEBUG] %v - dept %v exists\", cnt, deptno)\n\t\t}\n\t\t{ //Если Emp существует, то игнорируем\n\t\t\texists, err := p.EmpExistsTx(r.Empno, tx)\n\t\t\tif err != nil {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.EmpExistsTx(r.Empno, tx), args = '%v'\", cnt, r.Empno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t\t}\n\t\t\t// Если запись существует, то ни чего не делаем, возвращем, что пришло на вход\n\t\t\tif exists {\n\t\t\t\terrM := fmt.Sprintf(\"[WARN] %v - WARN - emp '%v' already exist - nothing to do\", cnt, r.Empno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, nil\n\t\t\t}\n\t\t\t//mylog.PrintfDebug(\"[DEBUG] %v - emp '%v' does not exist - can be created\", cnt, r.Empno)\n\t\t}\n\t}\n\t// =====================================================================\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\tstmText := sqlInsertEmpText\n\t// =====================================================================\n\n\t//Выполняем команду\n\tres, err := tx.NamedExec(stmText, r)\n\tif err != nil {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - tx.NamedExec(stmText, r), args = '%+v'\", cnt, r)\n\t\tlog.Printf(errM)\n\t\treturn nil, errors.Wrap(err, errM)\n\t}\n\n\t{ // Необязательная часть - можно удалить в последствии\n\t\t// Проверим количество обработанных строк\n\t\trowCount, err := res.RowsAffected()\n\t\t_ = rowCount\n\t\t//mylog.PrintfDebug(\"[DEBUG] %v -- process %v rows\", cnt, rowCount)\n\t\tif err != nil {\n\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - res.RowsAffected()\", cnt)\n\t\t\tlog.Printf(errM)\n\t\t\treturn nil, errors.Wrap(err, errM)\n\t\t}\n\t}\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\t// считаем данные обновленные данные - в БД могли быть тригера, которые поменяли данные\n\t// если запустили без проверок, то можно не возвращать результат - он будет запрошен уровнем выше\n\tif isValidate {\n\t\tv, err := p.GetEmpTx(r.Empno, tx)\n\t\tif err != nil {\n\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.GetEmpTx(r.Empno, tx), args = '%v'\", cnt, r.Empno)\n\t\t\tlog.Printf(errM)\n\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t}\n\t\tr = v\n\t}\n\t// =====================================================================\n\n\t//mylog.PrintfDebug(\"[DEBUG] %v - SUCCESS\", cnt)\n\n\treturn r, nil\n}","func (gt GameType) CreateTx(action string, message json.RawMessage) (*types.Transaction, error) {\n\ttlog.Debug(\"Game.CreateTx\", \"action\", action)\n\tif action == ActionCreateGame {\n\t\tvar param GamePreCreateTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreCreateTx(&param)\n\t} else if action == ActionMatchGame {\n\t\tvar param GamePreMatchTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreMatchTx(&param)\n\t} else if action == ActionCancelGame {\n\t\tvar param GamePreCancelTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreCancelTx(&param)\n\t} else if action == ActionCloseGame {\n\t\tvar param GamePreCloseTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreCloseTx(&param)\n\t}\n\treturn nil, types.ErrNotSupport\n}","func Create(reqDto *ReqCreateDto, custDto *ReqCustomerDto) (statusCode int, code string, respDto *RespBase, err error) {\r\n\treqDto.ServiceType = \"TMS_CREATE_ORDER\"\r\n\tbizData, err := xml.Marshal(reqDto.BizData)\r\n\tif err != nil {\r\n\t\tcode = E02\r\n\t\treturn\r\n\t}\r\n\tsignParam := string(bizData) + reqDto.PartnerKey\r\n\treqDto.Sign, err = sign.GetMD5Hash(signParam, true)\r\n\tif err != nil {\r\n\t\tcode = E02\r\n\t\treturn\r\n\t}\r\n\treqMap := make(map[string]string, 0)\r\n\treqMap[\"serviceType\"] = reqDto.ServiceType\r\n\treqMap[\"partnerID\"] = reqDto.PartnerID\r\n\treqMap[\"bizData\"] = string(bizData)\r\n\treqMap[\"sign\"] = reqDto.Sign\r\n\r\n\tdata := base.JoinMapString(reqMap)\r\n\r\n\treq := httpreq.New(http.MethodPost, custDto.Url, data, func(httpReq *httpreq.HttpReq) error {\r\n\t\thttpReq.ReqDataType = httpreq.FormType\r\n\t\thttpReq.RespDataType = httpreq.XmlType\r\n\t\treturn nil\r\n\t})\r\n\tstatusCode, err = req.Call(&respDto)\r\n\tif err != nil {\r\n\t\tcode = E01\r\n\t\treturn\r\n\t}\r\n\tif statusCode != http.StatusOK {\r\n\t\tcode = E01\r\n\t\terr = fmt.Errorf(\"http status exp:200,act:%v\", statusCode)\r\n\t\treturn\r\n\t}\r\n\tif respDto.Result != true {\r\n\t\tcode = E03\r\n\t\terr = fmt.Errorf(\"%v-%v\", respDto.ErrorCode, respDto.ErrorDescription)\r\n\t\treturn\r\n\t}\r\n\tcode = SUC\r\n\treturn\r\n}","func (t *Procure2Pay) CreateItems(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n\tvar objitem item\t\n\tvar err error\n\n\tfmt.Println(\"Entering CreateItems\")\n\n\tif (len(args) < 1) {\n\t\tfmt.Println(\"Invalid number of args\")\n\t\treturn shim.Error(err.Error())\n\t\t//return nil, errors.New(\"Expected atleast one arguments for initiate Transaction\")\n\t}\n\n\tfmt.Println(\"Args [0] is : %v\\n\",args[0])\n\t\n\t//unmarshal item data from UI to \"item\" struct\n\terr = json.Unmarshal([]byte(args[0]), &objitem)\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to unmarshal CreateItem input item: %s\\n\", err)\n\t\treturn shim.Error(err.Error())\n\t\t//return nil, nil\n\t}\n\n\tfmt.Println(\"item object ItemID variable value is : %s\\n\",objitem.ItemID);\n\t\n\t\t// Data insertion for Couch DB starts here \n\t\ttransJSONasBytes, err := json.Marshal(objitem)\n\t\terr = stub.PutState(objitem.ItemID, transJSONasBytes)\n\t\t// Data insertion for Couch DB ends here \n\n\t\tfmt.Println(\"Create items Successfully Done\")\t\n\t\n\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"\\nUnable to make transevent inputs : %v \", err)\n\t\t\t\treturn shim.Error(err.Error())\n\t\t\t\t//return nil,nil\n\t\t\t}\n\treturn shim.Success(nil)\n\t//return nil, nil\n}","func Transaction(next http.Handler) http.Handler {\n\tfn := func(w http.ResponseWriter, r *http.Request) {\n\t\tt, ctx := orm.NewTransaction(r.Context())\n\t\tdefer func() {\n\t\t\tif rec := recover(); rec != nil {\n\t\t\t\tt.Rollback()\n\t\t\t\t// Panic to let recoverer handle 500\n\t\t\t\tpanic(rec)\n\t\t\t} else {\n\t\t\t\terr := t.Commit()\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t}\n\treturn http.HandlerFunc(fn)\n}","func (ingest *Ingestion) Transaction(\n\tid int64,\n\ttx *core.Transaction,\n\tfee *core.TransactionFee,\n) error {\n\n\tsql := ingest.transactionInsertBuilder(id, tx, fee)\n\t_, err := ingest.DB.Exec(sql)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}","func MakeResTransaction(creator utils.Addr,key utils.Key,\n inps []TrInput,out TrOutput,\n jobblock,jobtrans,hashsol string,\n evaluation float64,isMin bool)*ResTransaction{\n tr := new(ResTransaction)\n tr.Timestamp = time.Now()\n tr.Output = out\n tr.Inputs = inps\n tr.Creator = creator\n tr.JobBlock = jobblock\n tr.JobTrans = jobtrans\n tr.HashSol = hashsol\n tr.Evaluation = evaluation\n tr.IsMin = isMin\n tr.Hash = tr.GetHash()\n tr.Signature = fmt.Sprintf(\"%x\",utils.GetSignatureFromHash(tr.Hash,key))\n return tr\n}","func (c CreateTransactionService) Execute(newTransaction entities.Transaction) (*entities.Transaction, error) {\n\ttransactType := newTransaction.Type\n\tif transactType != \"income\" && transactType != \"outcome\" {\n\t\treturn nil, errors.New(\"Cannot create transaction type different fom income or outcome\")\n\t}\n\n\tif transactType == \"outcome\" && !c.checksValidBalance(newTransaction) {\n\t\treturn nil, errors.New(\"Cannot create transaction with invalid balance\")\n\t}\n\n\tcreatedTransaction, err := c.Repo.Create(&newTransaction)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn createdTransaction, nil\n}","func TransactionUpdate(c *gin.Context) {\n\tvar t models.Transaction\n\tvar newT models.Transaction\n\n\tif database.DBCon.First(&t, c.Param(\"id\")).RecordNotFound() {\n\t\tc.AbortWithError(http.StatusNotFound, appError.RecordNotFound).\n\t\t\tSetMeta(appError.RecordNotFound)\n\t\treturn\n\t}\n\n\t// Ensure current user is creator of transaction\n\tif t.CreatorID != c.Keys[\"CurrentUserID\"].(uint) {\n\t\tc.AbortWithError(appError.InsufficientPermission.Status, appError.InsufficientPermission).\n\t\t\tSetMeta(appError.InsufficientPermission)\n\t\treturn\n\t}\n\n\tbuffer, err := ioutil.ReadAll(c.Request.Body)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusNotAcceptable, err)\n\t}\n\n\terr2 := jsonapi.Unmarshal(buffer, &newT)\n\n\tif err2 != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tt.Type = newT.Type\n\tt.Amount = newT.Amount\n\tt.Memo = newT.Memo\n\tt.RecipientID = newT.RecipientID\n\tt.SenderID = newT.SenderID\n\n\t// Validate our new transaction\n\tisValid, errApp := t.Validate()\n\n\tif isValid == false {\n\t\tc.AbortWithError(errApp.Status, errApp).\n\t\t\tSetMeta(errApp)\n\t\treturn\n\t}\n\n\tdatabase.DBCon.Save(&t)\n\n\tdatabase.DBCon.First(&t.Recipient, t.RecipientID)\n\tdatabase.DBCon.First(&t.Sender, t.SenderID)\n\tdatabase.DBCon.First(&t.Creator, t.CreatorID)\n\n\tdata, err := jsonapi.Marshal(&t)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tc.Data(http.StatusOK, \"application/vnd.api+json\", data)\n}","func (t *Transaction) createID() (string, error) {\n\n\t// Strip ID of txn\n\ttn := &Transaction{\n\t\tID: nil,\n\t\tVersion: t.Version,\n\t\tInputs: t.Inputs,\n\t\tOutputs: t.Outputs,\n\t\tOperation: t.Operation,\n\t\tAsset: t.Asset,\n\t\tMetadata: t.Metadata,\n\t}\n\t// Serialize transaction - encoding/json follows RFC7159 and BDB marshalling\n\tdbytes, err := tn.JSON()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Return hash of serialized txn object\n\th := sha3.Sum256(dbytes)\n\treturn hex.EncodeToString(h[:]), nil\n}","func (_Content *ContentTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Content.contract.Transact(opts, \"runCreate\")\n}","func (tm *TransactionManager) CreateContract(sk string, data []byte, gasPrice uint64, nonce uint64, gasLimit uint64) (string, error) {\n\treturn tm.SendTx(sk, \"\", nil, data, gasPrice, nonce, gasLimit)\n}","func (_LvRecording *LvRecordingTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _LvRecording.contract.Transact(opts, \"runCreate\")\n}","func (bi *Blockchainidentifier) SaveTransaction(w http.ResponseWriter, r *http.Request) {\n\tlog.Println(r.RemoteAddr + \" POST /transactions/new\")\n\n\tvar t transaction\n\n\terr := json.NewDecoder(r.Body).Decode(&t)\n\tif err != nil {\n\t\thttp.Error(w, \"ERROR: \"+err.Error(), 500)\n\t\treturn\n\t}\n\n\tif bi.isValid(t) == false {\n\t\thttp.Error(w, \"ERROR: Missing values in transaction\", 400)\n\t\treturn\n\t}\n\n\tt.Timestamp = time.Now().UTC().Format(\"2006-01-02 15:04:05\")\n\n\tnewblockindex := bi.newTransaction(t)\n\n\tresponseMessage := map[string]string{\n\t\t\"message\": \"Transaction will be added in Block#\" + strconv.Itoa(newblockindex),\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\tw.WriteHeader(http.StatusCreated)\n\tjson.NewEncoder(w).Encode(responseMessage)\n}","func NewTransaction(p *requestParams) {\n\tw, r, c, u := p.w, p.r, p.c, p.u\n\n\td := json.NewDecoder(r.Body)\n\tvar request TransactionRequest\n\tif err := d.Decode(&request); err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tif len(request.Amounts) != len(request.Accounts) {\n\t\thttp.Error(w, \"Amounts and accounts of different lengths\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tdate, err := time.Parse(dateStringFormat, request.Date)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tuserKey := userKey(c, u)\n\ttransactionId := uuid.NewRandom().String()\n\taccountKeys := make([]*datastore.Key, len(request.Accounts))\n\tsplitKeys := make([]*datastore.Key, len(request.Accounts))\n\tsplits := make([]*transaction.Split, len(request.Accounts))\n\n\tfor i := range request.Accounts {\n\t\taccountKeys[i] = datastore.NewKey(c, \"Account\", \"\", request.Accounts[i], userKey)\n\t\tsplitKeys[i] = datastore.NewKey(c, \"Split\", transactionId, 0, accountKeys[i])\n\t\tsplits[i] = &transaction.Split{\n\t\t\tAmount: request.Amounts[i],\n\t\t\tAccount: request.Accounts[i],\n\t\t\tMemo: request.Memo,\n\t\t\tDate: date,\n\t\t}\n\t}\n\n\tx := transaction.NewTransaction()\n\tx.AddSplits(splits)\n\n\tif err := x.ValidateAmount(); err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\terr = datastore.RunInTransaction(c, func(c appengine.Context) error {\n\t\taccounts := make([]transaction.Account, len(accountKeys))\n\t\tif err := datastore.GetMulti(c, accountKeys, accounts); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor i := range accounts {\n\t\t\tx.AddAccount(&accounts[i], accountKeys[i].IntID())\n\t\t}\n\n\t\tif err := x.Commit(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tputStatus := make(chan error)\n\n\t\tgo func() {\n\t\t\t_, err := datastore.PutMulti(c, accountKeys, accounts)\n\t\t\tputStatus <- err\n\t\t}()\n\t\tgo func() {\n\t\t\t_, err := datastore.PutMulti(c, splitKeys, splits)\n\t\t\tputStatus <- err\n\t\t}()\n\n\t\terr := <-putStatus\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn <-putStatus\n\t}, nil)\n\tif err != nil {\n\t\t// TODO(cjc25): This might not be a 400: if e.g. datastore failed it should\n\t\t// be a 500. Interpret err and return the right thing.\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n}","func (s *dbTestSuite) CreateTestTx() (*sql.Tx, func()) {\n\ttx, err := s.db.DB().Begin()\n\ts.Nil(err)\n\tcommit := func() {\n\t\ts.Nil(tx.Commit())\n\t}\n\n\treturn tx, commit\n}","func (p *KVServer) NewTransaction(req bool, resp *NewTransactionResp) error {\n\ttxID := nextTransactionID\n\tnextTransactionID = nextTransactionID + 10\n\tmutex.Lock()\n\tkvStore := keyValueStore\n\tmutex.Unlock()\n\t*resp = NewTransactionResp{txID, kvStore}\n\treturn nil\n}","func CreateServiceSubscriptionTransaction(db *sql.DB, params url.Values) (data []byte, err error) {\n\tvar (\n\t\tresult sql.Result\n\t\tresponse = SQLResponse{}\n\t\ttx *sql.Tx\n\t\tserviceTypeStr string\n\t\tserviceType int\n\t\tserviceName string\n\t\tdescription string\n\t\torganizationName string\n\t\tactiveUntil string\n\t\tamountPaidStr string\n\t\tamountPaid int\n\t\tprocessedTimeStamp string\n\t)\n\n\tif tx, err = db.Begin(); err != nil {\n\t\treturn\n\t}\n\tif amountPaidStr, err = common.GetRequiredParam(params, \"amountPaid\"); err != nil {\n\t\treturn\n\t}\n\tif amountPaid, err = strconv.Atoi(amountPaidStr); err != nil {\n\t\treturn\n\t}\n\tif processedTimeStamp, err = common.GetRequiredParam(params, \"processedTimestamp\"); err != nil {\n\t\treturn\n\t}\n\tif serviceTypeStr, err = common.GetRequiredParam(params, \"serviceType\"); err != nil {\n\t\treturn\n\t}\n\tif serviceType, err = strconv.Atoi(serviceTypeStr); err != nil {\n\t\treturn\n\t}\n\tif serviceName, err = common.GetRequiredParam(params, \"serviceName\"); err != nil {\n\t\treturn\n\t}\n\tif description, err = common.GetRequiredParam(params, \"description\"); err != nil {\n\t\treturn\n\t}\n\tif organizationName, err = common.GetRequiredParam(params, \"organizationName\"); err != nil {\n\t\treturn\n\t}\n\tif activeUntil, err = common.GetRequiredParam(params, \"activeUntil\"); err != nil {\n\t\treturn\n\t}\n\tif result, err = tx.Exec(\"INSERT INTO ServiceSubscriptionTransaction (amountPaid, processedTimestamp,type,serviceName,description, organizationName, activeUntil) VALUES(?,?,?,?,?,?,?);\",\n\t\tamountPaid, processedTimeStamp, serviceType, serviceName, description, organizationName, activeUntil); err != nil {\n\t\ttx.Rollback()\n\t\treturn\n\t}\n\tif err = tx.Commit(); err != nil {\n\t\treturn\n\t}\n\tif response.AffectedRows, err = result.RowsAffected(); err != nil {\n\t\treturn\n\t}\n\tdata, err = json.Marshal(response)\n\treturn\n}","func (_LvRecordableStream *LvRecordableStreamTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _LvRecordableStream.contract.Transact(opts, \"runCreate\")\n}","func (m *MySQLModelAudit) CreateTx(tx *sql.Tx, modelAudit *modelaudit.Model) error {\n\tstmt, err := tx.Prepare(mysqlCreateModelAudit)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\tnow := utils.Now()\n\treturn mysql.ExecAffectingOneRow(\n\t\tstmt,\n\t\tmodelAudit.PrevModel.ModelID,\n\t\tmodelAudit.PrevModel.Mjml,\n\t\tmodelAudit.PrevModel.Html,\n\t\tmodelAudit.PrevModel.Variables,\n\t\tmodelAudit.Model.Mjml,\n\t\tmodelAudit.Model.Html,\n\t\tmodelAudit.Model.Variables,\n\t\tmodelAudit.SetUserID,\n\t\tnow[\"date\"], now[\"time\"], now[\"unix\"],\n\t)\n}","func (w *WalletChaincode) create(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tif len(args) < 6 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting at least 2\")\n\t}\n\taddress := args[2]\n\taccountKey := buildAccountKey(address)\n\tif err := stub.PutState(accountKey, []byte(address)); err != nil {\n\t\treturn shim.Error(fmt.Sprintf(\"Error putting data for key [%s]: %s\", accountKey, err))\n\t}\n\tfmt.Println(\"create an account: \", accountKey)\n\n\twalletKey := buildWalletKey(args[0], args[1], address)\n\tif err := stub.PutState(walletKey, []byte(args[5])); err != nil {\n\t\treturn shim.Error(fmt.Sprintf(\"Error putting data for key [%s]: %s\", walletKey, err))\n\t}\n\tfmt.Println(\"create a wallet: \", walletKey)\n\n\t// seqBytes, err := stub.GetState(SEQUENCE)\n\t// if err != nil {\n\t// \treturn shim.Error(\"Failed to get state\")\n\t// }\n\t// if seqBytes == nil {\n\t// \treturn shim.Error(\"Entity not found\")\n\t// }\n\t// seq, _ := strconv.ParseInt(string(seqBytes), 10, 64)\n\tseq := atomic.AddUint64(&w.Sequence, 1)\n\tsequenceKey := buildSequenceKey(seq)\n\tjsonTx := \"{\\\"sequence\\\":\\\"\" + strconv.FormatUint(seq, 10) + \"\\\",\\\"txid\\\":\\\"\" + string(stub.GetTxID()) + \"\\\"}\"\n\tif err := stub.PutState(sequenceKey, []byte(jsonTx)); err != nil {\n\t\treturn shim.Error(fmt.Sprintf(\"Error putting data for key [%s]: %s\", walletKey, err))\n\t}\n\n\tfmt.Println(\"create success: \", stub.GetTxID())\n\treturn shim.Success([]byte(fmt.Sprintf(\"{\\\"wallet\\\":\\\"%s\\\", \\\"txid\\\":\\\"%s\\\"}\", walletKey, stub.GetTxID())))\n}","func (tx *Transaction) Create(entity interface{}) error {\n\treturn createFunc(tx.Query, entity)\n}","func Create(t *contract.SQLTable) error {\n\tsqlstr := t.GetCreateSQL()\n\tif t.Database == \"\" {\n\t\tt.Database = \"default\"\n\t}\n\tdb, err := dblist.Get(t.Database)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = db.Exec(sqlstr)\n\treturn err\n}","func (s *PaymentStorage) Create(\n\tctx context.Context,\n\tid aggregate.ID,\n\tversion transaction.Version,\n\torganisationID transaction.OrganisationID,\n\tattributes interface{},\n) error {\n\tlogger := log.FromContext(ctx)\n\n\tquery := `INSERT INTO %[1]s (id, version, organisation_id, attributes) VALUES ($1, $2, $3, $4)`\n\tquery = fmt.Sprintf(query, s.table)\n\n\tif logger != nil {\n\t\tlogger.Debugf(\"exec in transaction sql %s, values %+v\", query, []interface{}{\n\t\t\tid,\n\t\t\tversion,\n\t\t\torganisationID,\n\t\t\tattributes,\n\t\t})\n\t}\n\n\treturn execInTransaction(s.db, func(tx *sqlx.Tx) error {\n\t\t_, err := tx.ExecContext(ctx, query, id, version, organisationID, attributes)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n}","func (_TableManager *TableManagerTransactor) CreateTable(opts *bind.TransactOpts, path string, tableInfo TableInfo) (*types.Transaction, *types.Receipt, error) {\n\treturn _TableManager.contract.Transact(opts, \"createTable\", path, tableInfo)\n}","func (s *SmartContract) CreateContract(ctx contractapi.TransactionContextInterface, id string, sellerID string, consumerID string, created string, contract string, signature string) error {\n\tctc := Contract{\n\t\tContext: \"http://wuldid.ddns.net\",\n\t\tSellerID: sellerID,\n\t\tConsumerID: consumerID,\n\t\tCreated: created,\n\t\tContract: contract,\n\t\tSignature: signature,\n\t}\n\n\texists, err := s.DidExists(ctx, id)\n\tif err != nil {\n\t\tfmt.Errorf(\"Unexpected error!! : %q\", err)\n\t}\n\tif !exists {\n\t\tctcJSON, _ := json.Marshal(ctc)\n\t\treturn ctx.GetStub().PutState(id, ctcJSON)\n\t} else {\n\t\treturn fmt.Errorf(\"Don't exsit did!\")\n\t}\n\n}","func (client *Client) CreateTransactionType(transactionType string, customIndexes []CustomIndexStructure) (_ *Response, err error) {\n\tpath := \"/transaction-type\"\n\turi := fmt.Sprintf(\"%s%s\", client.apiBaseURL, path)\n\tvar params TransactionType\n\tparams.Version = \"1\"\n\tparams.Type = transactionType\n\tparams.CustomIndexes = customIndexes\n\n\tb, err := json.Marshal(params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := client.httpClient.Post(uri, \"content/json\", bytes.NewBuffer(b))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer func() {\n\t\t_ = resp.Body.Close()\n\t}()\n\tvar chainResp Response\n\trespBytes, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := json.Unmarshal(respBytes, &chainResp.Response); err != nil {\n\t\treturn nil, err\n\t}\n\tchainResp.Status = resp.StatusCode\n\tif 200 <= resp.StatusCode && resp.StatusCode < 300 {\n\t\tchainResp.OK = true\n\t}\n\treturn &chainResp, err\n}","func (r Virtual_Guest) CreateArchiveTransaction(groupName *string, blockDevices []datatypes.Virtual_Guest_Block_Device, note *string) (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\tparams := []interface{}{\n\t\tgroupName,\n\t\tblockDevices,\n\t\tnote,\n\t}\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"createArchiveTransaction\", params, &r.Options, &resp)\n\treturn\n}","func (t *TxCreate) CreatePaymentTx() (string, string, error) {\n\tsender := t.paymentSender\n\treceiver := account.AccountTypeAnonymous\n\ttargetAction := action.ActionTypePayment\n\tt.logger.Debug(\"account\",\n\t\tzap.String(\"sender\", sender.String()),\n\t\tzap.String(\"receiver\", receiver.String()),\n\t)\n\n\t// get payment data from payment_request\n\tuserPayments, totalAmount, paymentRequestIds, err := t.createUserPayment()\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\tif len(userPayments) == 0 {\n\t\tt.logger.Debug(\"no data in userPayments\")\n\t\t// no data\n\t\treturn \"\", \"\", nil\n\t}\n\n\t// check sender's total balance\n\t// GetOneUnAllocated\n\taddrItem, err := t.addrRepo.GetOneUnAllocated(sender)\n\tif err != nil {\n\t\treturn \"\", \"\", errors.Wrap(err, \"fail to call addrRepo.GetAll(account.AccountTypeClient)\")\n\t}\n\tsenderBalance, err := t.xrp.GetBalance(addrItem.WalletAddress)\n\tif err != nil {\n\t\treturn \"\", \"\", errors.Wrap(err, \"fail to call xrp.GetAccountInfo()\")\n\t}\n\n\tif senderBalance <= totalAmount {\n\t\treturn \"\", \"\", errors.New(\"sender balance is insufficient to send\")\n\t}\n\n\t// create raw transaction each address\n\tserializedTxs := make([]string, 0, len(userPayments))\n\ttxDetailItems := make([]*models.XRPDetailTX, 0, len(userPayments))\n\tvar sequence uint64\n\tfor _, userPayment := range userPayments {\n\t\t// call CreateRawTransaction\n\t\tinstructions := &pb.Instructions{\n\t\t\tMaxLedgerVersionOffset: xrp.MaxLedgerVersionOffset,\n\t\t}\n\t\tif sequence != 0 {\n\t\t\tinstructions.Sequence = sequence\n\t\t}\n\t\ttxJSON, rawTxString, err := t.xrp.CreateRawTransaction(addrItem.WalletAddress, userPayment.receiverAddr, userPayment.floatAmount, instructions)\n\t\tif err != nil {\n\t\t\tt.logger.Warn(\"fail to call xrp.CreateRawTransaction()\", zap.Error(err))\n\t\t\tcontinue\n\t\t}\n\t\tt.logger.Debug(\"txJSON\", zap.Any(\"txJSON\", txJSON))\n\t\tgrok.Value(txJSON)\n\n\t\t// sequence for next rawTransaction\n\t\tsequence = txJSON.Sequence + 1\n\n\t\t// generate UUID to trace transaction because unsignedTx is not unique\n\t\tuid := uuid.NewV4().String()\n\n\t\tserializedTxs = append(serializedTxs, fmt.Sprintf(\"%s,%s\", uid, rawTxString))\n\n\t\t// create insert data for eth_detail_tx\n\t\ttxDetailItem := &models.XRPDetailTX{\n\t\t\tUUID: uid,\n\t\t\tCurrentTXType: tx.TxTypeUnsigned.Int8(),\n\t\t\tSenderAccount: sender.String(),\n\t\t\tSenderAddress: addrItem.WalletAddress,\n\t\t\tReceiverAccount: receiver.String(),\n\t\t\tReceiverAddress: userPayment.receiverAddr,\n\t\t\tAmount: txJSON.Amount,\n\t\t\tXRPTXType: txJSON.TransactionType,\n\t\t\tFee: txJSON.Fee,\n\t\t\tFlags: txJSON.Flags,\n\t\t\tLastLedgerSequence: txJSON.LastLedgerSequence,\n\t\t\tSequence: txJSON.Sequence,\n\t\t\t// SigningPubkey: txJSON.SigningPubKey,\n\t\t\t// TXNSignature: txJSON.TxnSignature,\n\t\t\t// Hash: txJSON.Hash,\n\t\t}\n\t\ttxDetailItems = append(txDetailItems, txDetailItem)\n\t}\n\n\treturn t.afterTxCreation(targetAction, sender, serializedTxs, txDetailItems, paymentRequestIds)\n}","func (allocateIncome *AllocateIncome) Create() (err error) {\n\t//插入数据\n\tstmt, err := Db.Prepare(\"INSERT INTO allocateIncome(UDID, pool_id,mining_id,income,fee,date,status) values(?,?,?,?,?,?,?)\")\n\tif err != nil {\n\t\treturn\n\t}\n\tres, err := stmt.Exec(allocateIncome.UDID, allocateIncome.PoolID, allocateIncome.MiningId, allocateIncome.Income, allocateIncome.Fee, time.Now().Unix(), allocateIncome.Status)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tid, err := res.LastInsertId()\n\tlog.Info(id)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}","func (w *Wallet) txCreator() {\n\tquit := w.quitChan()\nout:\n\tfor {\n\t\tselect {\n\t\tcase txr := <-w.createTxRequests:\n\t\t\theldUnlock, err := w.holdUnlock()\n\t\t\tif err != nil {\n\t\t\t\ttxr.resp <- createTxResponse{nil, err}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ttx, err := w.txToOutputs(txr.req)\n\t\t\theldUnlock.release()\n\t\t\ttxr.resp <- createTxResponse{tx, err}\n\t\tcase <-quit:\n\t\t\tbreak out\n\t\t}\n\t}\n\tw.wg.Done()\n}","func (w *Wallet) CreateSimpleTx(r CreateTxReq) (*txauthor.AuthoredTx, er.R) {\n\treq := createTxRequest{\n\t\treq: r,\n\t\tresp: make(chan createTxResponse),\n\t}\n\tw.createTxRequests <- req\n\tresp := <-req.resp\n\treturn resp.tx, resp.err\n}","func (db *Database) NewTransaction() *Transaction{\r\n\treturn &Transaction{db,make(map[string]*record)}\r\n}","func (p *psqlInvoiceItem) CreateTx(tx *sql.Tx, headerID uint, ms invoiceItem.Models) error {\n\tstmt, err := tx.Prepare(psqlCreateInvoiceItem)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\tfor _, item := range ms {\n\t\terr = stmt.QueryRow(headerID, item.ProductID).Scan(&item.ID, &item.CreatedAt)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n\t// return stmt.QueryRow(m.).Scan(&m.ID, &m.CreatedAt)\n}","func (tr *Transactor) CreateTransactionArweave(ctx context.Context, tags []Tag, w WalletSigner, amount string, data []byte, target string) (*Transaction, error) {\n\tlastTx, err := tr.Client.TxAnchor(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tprice, err := tr.Client.GetReward(ctx, data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Non encoded transaction fields\n\ttx := NewTransactionArweave(\n\t\ttags,\n\t\tlastTx,\n\t\tw.PubKeyModulus(),\n\t\tamount,\n\t\ttarget,\n\t\tdata,\n\t\tprice,\n\t)\n\n\treturn tx, nil\n}","func create(\n\ttrx storage.Transaction,\n\ttxId merkle.Digest,\n\townerData OwnerData,\n\towner *account.Account,\n) {\n\t// increment the count for owner\n\tnKey := owner.Bytes()\n\tcount := trx.Get(storage.Pool.OwnerNextCount, nKey)\n\tif nil == count {\n\t\tcount = []byte{0, 0, 0, 0, 0, 0, 0, 0}\n\t} else if uint64ByteSize != len(count) {\n\t\tlogger.Panic(\"OwnerNextCount database corrupt\")\n\t}\n\tnewCount := make([]byte, uint64ByteSize)\n\tbinary.BigEndian.PutUint64(newCount, binary.BigEndian.Uint64(count)+1)\n\ttrx.Put(storage.Pool.OwnerNextCount, nKey, newCount, []byte{})\n\n\t// write to the owner list\n\toKey := append(owner.Bytes(), count...)\n\ttrx.Put(storage.Pool.OwnerList, oKey, txId[:], []byte{})\n\n\t// write new index record\n\tdKey := append(owner.Bytes(), txId[:]...)\n\ttrx.Put(storage.Pool.OwnerTxIndex, dKey, count, []byte{})\n\n\t// save owner data record\n\ttrx.Put(storage.Pool.OwnerData, txId[:], ownerData.Pack(), []byte{})\n}","func CreateTestContractCreationTransaction(\n\tsigner hmytypes.Signer, shard uint32, nonce, gasLimit uint64, gasPrice, amount *big.Int, data []byte,\n) (*hmytypes.Transaction, error) {\n\tfromKey, err := crypto.GenerateKey()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttx := hmytypes.NewContractCreation(nonce, shard, amount, gasLimit, gasPrice, data)\n\treturn hmytypes.SignTx(tx, signer, fromKey)\n}","func Create(responseWriter http.ResponseWriter, request *http.Request) {\n\tfmt.Println(\"[ CreateOrder ]\")\n\tbody, _ := json.Marshal(request.Body)\n\tfmt.Println(\"[ CreateOrder ] Body=\" + string(body))\n\t//params := mux.Vars(request)\n\tvar orderEntity OrderEntity\n\t_ = json.NewDecoder(request.Body).Decode(&orderEntity)\n\n\tvar result OrderEntity = Insert(orderEntity)\n\n\tWriteMessages(result, Topic.TOPIC_SUCCESS)\n\n\tjson.NewEncoder(responseWriter).Encode(result)\n}","func (s *Service) CreateClientTx(tx *gorm.DB, clientID, secret, redirectURI string) (*models.OauthClient, error) {\n\treturn s.createClientCommon(tx, clientID, secret, redirectURI)\n}","func (dt DefaultTransactor) Create(ctx context.Context, m *movie.Movie) error {\n\ttx, err := dt.datastorer.BeginTx(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Prepare the sql statement using bind variables\n\tstmt, err := tx.PrepareContext(ctx, `\n\tselect o_create_timestamp,\n\t\t o_update_timestamp\n\t from demo.create_movie (\n\t\tp_id => $1,\n\t\tp_extl_id => $2,\n\t\tp_title => $3,\n\t\tp_rated => $4,\n\t\tp_released => $5,\n\t\tp_run_time => $6,\n\t\tp_director => $7,\n\t\tp_writer => $8,\n\t\tp_create_client_id => $9,\n\t\tp_create_username => $10)`)\n\n\tif err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\tdefer stmt.Close()\n\n\t// At some point, I will add a whole client flow, but for now\n\t// faking a client uuid....\n\tfakeClientID := uuid.New()\n\n\t// Execute stored function that returns the create_date timestamp,\n\t// hence the use of QueryContext instead of Exec\n\trows, err := stmt.QueryContext(ctx,\n\t\tm.ID, //$1\n\t\tm.ExternalID, //$2\n\t\tm.Title, //$3\n\t\tm.Rated, //$4\n\t\tm.Released, //$5\n\t\tm.RunTime, //$6\n\t\tm.Director, //$7\n\t\tm.Writer, //$8\n\t\tfakeClientID, //$9\n\t\tm.CreateUser.Email) //$10\n\n\tif err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\tdefer rows.Close()\n\n\t// Iterate through the returned record(s)\n\tfor rows.Next() {\n\t\tif err := rows.Scan(&m.CreateTime, &m.UpdateTime); err != nil {\n\t\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t\t}\n\t}\n\n\t// If any error was encountered while iterating through rows.Next above\n\t// it will be returned here\n\tif err := rows.Err(); err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\n\t// Commit the Transaction\n\tif err := dt.datastorer.CommitTx(tx); err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\n\treturn nil\n}","func createOrder(writer http.ResponseWriter, request *http.Request) {\n\ttransactionId := request.Header.Get(\"transactionId\")\n\torder := <-orderService.GetOrder(transactionId)\n\tif order.Id.Value == \"\" {\n\t\tlog.Printf(\"Creating order for transactionId :%s......\", transactionId)\n\t\torderHandler.CreateOrder(transactionId, command.CreateOrderCommand{Id: transactionId})\n\t}\n\trenderResponse(writer, []byte(transactionId))\n}","func createSignedTransaction(fromAddress common.Address, toAddress *common.Address, amount *big.Int, gasLimit uint64, data []byte) (signedTx *types.Transaction, err error) {\n\t// Create the transaction\n\ttx, err := createTransaction(fromAddress, toAddress, amount, gasLimit, data)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Sign the transaction\n\tsignedTx, err = signTransaction(fromAddress, tx)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Failed to sign transaction: %v\", err)\n\t\treturn\n\t}\n\n\t// Increment the nonce for the next transaction\n\tnextNonce(fromAddress)\n\n\treturn\n}","func (r *Renter) createContractTransaction(terms modules.ContractTerms, merkleRoot crypto.Hash) (txn consensus.Transaction, id string, err error) {\n\t// Get the payout as set by the missed proofs, and the client fund as determined by the terms.\n\tvar payout consensus.Currency\n\tfor _, output := range terms.MissedProofOutputs {\n\t\tpayout = payout.Add(output.Value)\n\t}\n\n\t// Get the cost to the client as per the terms in the contract.\n\tsizeCurrency := consensus.NewCurrency64(terms.FileSize)\n\tdurationCurrency := consensus.NewCurrency64(uint64(terms.Duration))\n\tclientCost := terms.Price.Mul(sizeCurrency).Mul(durationCurrency)\n\n\t// Fill out the contract.\n\tcontract := consensus.FileContract{\n\t\tFileMerkleRoot: merkleRoot,\n\t\tFileSize: terms.FileSize,\n\t\tStart: terms.DurationStart + terms.Duration,\n\t\tExpiration: terms.DurationStart + terms.Duration + terms.WindowSize,\n\t\tPayout: payout,\n\t\tValidProofOutputs: terms.ValidProofOutputs,\n\t\tMissedProofOutputs: terms.MissedProofOutputs,\n\t}\n\n\t// Create the transaction.\n\tid, err = r.wallet.RegisterTransaction(txn)\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = r.wallet.FundTransaction(id, clientCost)\n\tif err != nil {\n\t\treturn\n\t}\n\ttxn, _, err = r.wallet.AddFileContract(id, contract)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}","func TransactionHandler(w http.ResponseWriter, r *http.Request) {\n\taction := r.URL.Path[len(\"/api/transactions\"):]\n\n\tlog.Println(\"Handling method\", r.Method, \"with action\", action)\n\n\tswitch r.Method {\n\tcase \"POST\":\n\t\tswitch action {\n\t\tcase \"\": // Create new transaction\n\t\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorForm, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar t TransactionResponse\n\n\t\t\terr = json.Unmarshal(body, &t)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpass := []byte(t.Password)\n\n\t\t\ttx, e := createAndBroadcastTx(t.Recipient, *big.NewInt(t.Amount), pass)\n\n\t\t\tvar jsonResponse TransactionResponse\n\t\t\tif e == nil {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseOk, Hash: hex.EncodeToString(tx.Hash())}\n\t\t\t} else {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseFailed, ErrorText: e.Error()}\n\t\t\t}\n\n\t\t\tres, err := json.Marshal(jsonResponse)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfmt.Fprintf(w, string(res))\n\t\tdefault:\n\t\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\t\t}\n\tcase \"GET\":\n\t\tswitch action {\n\t\tcase \"\":\n\t\t\tvar txs []TransactionJson\n\t\t\tfor _, tx := range Config.DeserializedTxs {\n\t\t\t\ttxs = append(txs, EncodeToFriendlyStruct(tx))\n\n\t\t\t}\n\n\t\t\tif len(txs) == 0 {\n\t\t\t\tfmt.Fprintf(w, string(\"[]\"))\n\t\t\t} else {\n\n\t\t\t\tres, err := json.Marshal(txs)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Println(\"Nope\", err.Error())\n\t\t\t\t} else {\n\t\t\t\t\tfmt.Fprintf(w, string(res))\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\tdefault:\n\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\n\t}\n}","func (p *PsqlInvoiceItem) CreateTx(tx *sql.Tx, headerID uint, ms invoiceitem.Models) error {\n\tstmt, err := tx.Prepare(psqlCreateInvoiceItem)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\tfor _, item := range ms {\n\t\terr = stmt.QueryRow(headerID, item.ProductID).Scan(\n\t\t\t&item.ID,\n\t\t\t&item.CreatedAt,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn err\n\n}","func (a *ChainAdaptor) CreateUtxoSignedTransaction(req *proto.CreateUtxoSignedTransactionRequest) (*proto.CreateSignedTransactionReply, error) {\n\tr := bytes.NewReader(req.TxData)\n\tvar msgTx wire.MsgTx\n\terr := msgTx.Deserialize(r)\n\tif err != nil {\n\t\tlog.Error(\"CreateSignedTransaction msgTx.Deserialize\", \"err\", err)\n\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\tif len(req.Signatures) != len(msgTx.TxIn) {\n\t\tlog.Error(\"CreateSignedTransaction invalid params\", \"err\", \"Signature number mismatch Txin number\")\n\t\terr = errors.New(\"Signature number != Txin number\")\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\tif len(req.PublicKeys) != len(msgTx.TxIn) {\n\t\tlog.Error(\"CreateSignedTransaction invalid params\", \"err\", \"Pubkey number mismatch Txin number\")\n\t\terr = errors.New(\"Pubkey number != Txin number\")\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\t// assemble signatures\n\tfor i, in := range msgTx.TxIn {\n\t\tbtcecPub, err2 := btcec.ParsePubKey(req.PublicKeys[i], btcec.S256())\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction ParsePubKey\", \"err\", err2)\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tvar pkData []byte\n\t\tif btcec.IsCompressedPubKey(req.PublicKeys[i]) {\n\t\t\tpkData = btcecPub.SerializeCompressed()\n\t\t} else {\n\t\t\tpkData = btcecPub.SerializeUncompressed()\n\t\t}\n\n\t\t// verify transaction\n\t\tpreTx, err2 := a.getClient().GetRawTransactionVerbose(&in.PreviousOutPoint.Hash)\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction GetRawTransactionVerbose\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tlog.Info(\"CreateSignedTransaction \", \"from address\", preTx.Vout[in.PreviousOutPoint.Index].ScriptPubKey.Addresses[0])\n\n\t\tfromAddress, err2 := btcutil.DecodeAddress(preTx.Vout[in.PreviousOutPoint.Index].ScriptPubKey.Addresses[0], a.getClient().GetNetwork())\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction DecodeAddress\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tfromPkScript, err2 := txscript.PayToAddrScript(fromAddress)\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction PayToAddrScript\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\t// creat sigscript and verify\n\t\tif len(req.Signatures[i]) < 64 {\n\t\t\terr2 = errors.New(\"Invalid signature length\")\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\t\tr := new(big.Int).SetBytes(req.Signatures[i][0:32])\n\t\ts := new(big.Int).SetBytes(req.Signatures[i][32:64])\n\n\t\tbtcecSig := &btcec.Signature{\n\t\t\tR: r,\n\t\t\tS: s,\n\t\t}\n\t\tsig := append(btcecSig.Serialize(), byte(txscript.SigHashAll))\n\t\tsigScript, err2 := txscript.NewScriptBuilder().AddData(sig).AddData(pkData).Script()\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction NewScriptBuilder\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tmsgTx.TxIn[i].SignatureScript = sigScript\n\t\tamount := btcToSatoshi(preTx.Vout[in.PreviousOutPoint.Index].Value).Int64()\n\t\tlog.Info(\"CreateSignedTransaction \", \"amount\", preTx.Vout[in.PreviousOutPoint.Index].Value, \"int amount\", amount)\n\n\t\tvm, err2 := txscript.NewEngine(fromPkScript, &msgTx, i, txscript.StandardVerifyFlags, nil, nil, amount)\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction NewEngine\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\t\tif err3 := vm.Execute(); err3 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction NewEngine Execute\", \"err\", err3)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err3.Error(),\n\t\t\t}, err3\n\t\t}\n\n\t}\n\n\t// serialize tx\n\tbuf := bytes.NewBuffer(make([]byte, 0, msgTx.SerializeSize()))\n\n\terr = msgTx.Serialize(buf)\n\tif err != nil {\n\t\tlog.Error(\"CreateSignedTransaction tx Serialize\", \"err\", err)\n\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\thash := msgTx.TxHash()\n\treturn &proto.CreateSignedTransactionReply{\n\t\tCode: proto.ReturnCode_SUCCESS,\n\t\tSignedTxData: buf.Bytes(),\n\t\tHash: (&hash).CloneBytes(),\n\t}, nil\n}","func (cs *CustStoreSqlite) AddTransaction(\n id string,\n customerId string,\n loadAmountCents int64,\n time time.Time,\n accepted bool,\n) error {\n // Check transaction id\n isDuplicate, err := cs.isDuplicate(id, customerId)\n if err != nil {\n return err\n }\n if isDuplicate {\n return custstore.DuplicateError\n }\n // Add to transactions\n err = cs.addTransaction(id, customerId, loadAmountCents, time, accepted)\n return err\n}","func CreateContractObject(args []string) (SalesContractObject, error) {\n\t// S001 LHTMO bosch\n\tvar err error\n\tvar myContract SalesContractObject\n\n\t// Check there are 3 Arguments provided as per the the struct\n\tif len(args) != 8 {\n\t\tfmt.Println(\"CreateContractObject(): Incorrect number of arguments. Expecting 8 \")\n\t\treturn myContract, errors.New(\"CreateContractObject(): Incorrect number of arguments. Expecting 8 \")\n\t}\n\n\t// Validate Serialno is an integer\n\n\tstage, err := strconv.Atoi(args[1])\n\tif err != nil {\n\t\tfmt.Println(\"CreateAssetObject(): Stage should be an integer create failed! \")\n\t\treturn myContract, errors.New(\"CreateAssetbject(): Stage should be an integer create failed. \")\n\t}\n\tif stage != 0 {\n\t\tfmt.Println(\"CreateAssetObject(): Stage should be set as open \")\n\t\treturn myContract, errors.New(\"CreateAssetbject(): Stage should be set as open\")\n\t}\n\n\tmyContract = SalesContractObject{args[0], STATE_OPEN, args[2], args[3], args[4], args[5], args[6], time.Now().Format(\"20060102150405\")}\n\n\tfmt.Println(\"CreateContractObject(): Contract Object created: \", myContract.Contractid, myContract.Stage, myContract.Buyer, myContract.Transporter, myContract.Seller, myContract.AssetID, myContract.DocumentID, time.Now().Format(\"20060102150405\"))\n\treturn myContract, nil\n}","func (a *ApiDB) CreateContract(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\tp := MODELS.CREATE_UPDATE_CONTRACT_REQUEST{}\n\terr := json.NewDecoder(r.Body).Decode(&p)\n\tif err != nil {\n\t\tio.WriteString(w, `{\"message\": \"wrong format!\"}`)\n\t\treturn\n\t}\n\n\tresult := BUSINESS.CreateContract(a.Db, p)\n\tif result {\n\t\tio.WriteString(w, ` { \"status\": 200,\n \"message\": \"Create contract success\",\n \"data\": {\n \"status\": 1\n }\n}\n`)\n\t} else {\n\t\tio.WriteString(w, `{ \"message\": \"Can’t create contract\"}`)\n\t}\n}","func (d Database) CreateRoot() error {\n\tdb, connectErr := d.Connect()\n\tdefer db.Close()\n\tutil.Handle(\"Error creating a DB connection: \", connectErr)\n\tvar count int\n\terr := db.QueryRow(\"SELECT COUNT(*) FROM \"+d.Cf.GetTableName()+\" ORDER BY $1 DESC\", \"tx_time\").Scan(&count)\n\tswitch {\n\tcase err != nil:\n\t\tutil.Handle(\"There was a problem counting database transactions: \", err)\n\t\treturn err\n\tdefault:\n\t\t// fmt.Printf(\"Found %v transactions in the db.\", count)\n\t\tif count == 0 {\n\t\t\ttxTime := \"1603203489229912200\"\n\t\t\ttxType := \"1\"\n\t\t\ttxSubg := \"0\"\n\t\t\ttxPrnt := \"0\"\n\t\t\ttxData := \"8d3729b91a13878508c564fbf410ae4f33fcb4cfdb99677f4b23d4c4adb447650964b4fe9da16299831b9cc17aaabd5b8d81fb05460be92af99d128584101a30\" // ?\n\t\t\ttxPrev := \"c66f4851618cd53104d4a395212958abf88d96962c0c298a0c7a7c1242fac5c2ee616c8c4f140a2e199558ead6d18ae263b2311b590b0d7bf3777be5b3623d9c\" // RockSteady was here\n\t\t\thash := sha512.Sum512([]byte(txTime + txType + txData + txPrev))\n\t\t\ttxHash := hex.EncodeToString(hash[:])\n\t\t\ttxMile := true\n\t\t\ttxLead := false\n\t\t\ttxEpoc := txHash\n\t\t\ttx := db.MustBegin()\n\t\t\ttx.MustExec(\"INSERT INTO \"+d.Cf.GetTableName()+\" (tx_time, tx_type, tx_hash, tx_data, tx_prev, tx_epoc, tx_subg, tx_prnt, tx_mile, tx_lead, tx_height ) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11)\", txTime, txType, txHash, txData, txPrev, txEpoc, txSubg, txPrnt, txMile, txLead, 0)\n\t\t\ttx.Commit()\n\t\t\treturn nil\n\t\t} else if count > 0 {\n\t\t\treturn errors.New(\"Root tx already present. \")\n\t\t}\n\t}\n\treturn nil\n}","func CreateTransfer(w http.ResponseWriter, r *http.Request) {\r\n\tparams := mux.Vars(r)\r\n\tvar v = params[\"id\"] //We use a simpler variable as the two 'Transfer' functions are fairly long\r\n\t\r\n\tfor _, item := range certif {\r\n\t\t//Let's first check the user has the right to transfer this certificate\r\n if item.Id == v && item.Ownerid==clientnum { \r\n\t\t\tcheckk:=\"ok\"\r\n\t\t\t\r\n\t\t\t//We browse the transfers list to verify there is no transfer for this certificate\r\n\t\t\tfor _, itemm := range trans {\r\n\t\t\t\tif itemm.Id == v {\r\n\t\t\t\t\tcheckk=\"not ok\"\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t\t//If so, we inform the user\r\n\t\t\tif checkk==\"not ok\" {\r\n\t\t\t\tfmt.Fprintf(w, \"There is already a transfer for this certificate which is pending.\")\r\n\t\t\t} else {\r\n\t\t\t\tvar transcertif Transfer //Creation of a new transfer\r\n\t\t\t\t\r\n\t\t\t\t//We collect the receiving party's email address \r\n\t\t\t\t_ = json.NewDecoder(r.Body).Decode(&transcertif) \r\n\t\t\t\t\r\n\t\t\t\ttranscertif.Id=v //Certificate ID stays unchanged \r\n\t\t\t\ttranscertif.Status=\"Pending\" //The transfer status is now changed until the answer of the receiving party\r\n\t\t\t\t\r\n\t\t\t\t//We verify if the email address of the receiver is correct and is not the current user's \r\n\t\t\t\tcheck:=\"not ok\" \r\n\t\t\t\tfor _, ite := range owner {\r\n\t\t\t\t\tif ite.Email==transcertif.To && ite.Userid!=clientnum {\r\n\t\t\t\t\t\tcheck=\"ok\"\r\n\t\t\t\t\t}\r\n\t\t\t\t}\r\n\t\t\t\t\r\n\t\t\t\tif check==\"ok\" { //Then, if the email is valid, the transfer proposal is sent\r\n\t\t\t\t\t\ttrans=append(trans, transcertif) //The transfer is added to the transfers list\r\n\t\t\t\t\t\t//The 'transfer' status of the certificate is changed\r\n\t\t\t\t\t\tfor index, itp := range certif {\r\n\t\t\t\t\t\t\tif itp.Id==v {\r\n\t\t\t\t\t\t\t\tcertif[index].Transfer=\"Transfer pending\"\r\n\t\t\t\t\t\t\t}\r\n\t\t\t\t\t\t}\r\n\t\t\t\t\t\tfmt.Fprintf(w, \"Transfer proposal sent.\")\r\n\t\t\t\t\t\t\r\n\t\t\t\t}\telse { //If the email is not valid, the user is informed\r\n\t\t\t\t\t\tfmt.Fprintf(w, \"This email is not valid.\")\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t} else if item.Id == v && item.Ownerid!=clientnum {\r\n\t\t\tfmt.Fprintf(w, \"You are not allowed to transfer this certificate.\")\r\n\t\t}\r\n\t}\r\n}"],"string":"[\n \"func TransactionCreate(c *gin.Context) {\\n\\tvar t models.Transaction\\n\\tbuffer, err := ioutil.ReadAll(c.Request.Body)\\n\\tif err != nil {\\n\\t\\tc.AbortWithError(http.StatusNotAcceptable, err)\\n\\t}\\n\\n\\terr2 := jsonapi.Unmarshal(buffer, &t)\\n\\n\\tif err2 != nil {\\n\\t\\tparseFail := appError.JSONParseFailure\\n\\t\\tparseFail.Detail = err2.Error()\\n\\t\\tc.AbortWithError(http.StatusMethodNotAllowed, err2).\\n\\t\\t\\tSetMeta(parseFail)\\n\\t\\treturn\\n\\t}\\n\\n\\tt.CreatorID = c.Keys[\\\"CurrentUserID\\\"].(uint)\\n\\n\\t// Validate our new transaction\\n\\tisValid, errApp := t.Validate()\\n\\n\\tif isValid == false {\\n\\t\\tc.AbortWithError(errApp.Status, errApp).\\n\\t\\t\\tSetMeta(errApp)\\n\\t\\treturn\\n\\t}\\n\\n\\tdatabase.DBCon.Create(&t)\\n\\n\\tdatabase.DBCon.First(&t.Recipient, t.RecipientID)\\n\\tdatabase.DBCon.First(&t.Sender, t.SenderID)\\n\\tdatabase.DBCon.First(&t.Creator, t.CreatorID)\\n\\n\\tdata, err := jsonapi.Marshal(&t)\\n\\n\\tif err != nil {\\n\\t\\tc.AbortWithError(http.StatusInternalServerError, err).\\n\\t\\t\\tSetMeta(appError.JSONParseFailure)\\n\\t\\treturn\\n\\t}\\n\\n\\tc.Data(http.StatusCreated, \\\"application/vnd.api+json\\\", data)\\n}\",\n \"func CreateTransaction(resp http.ResponseWriter, req *http.Request) {\\n\\tresp.Header().Add(\\\"Content-Type\\\", \\\"application/json\\\")\\n\\tvar transaction domain.Transaction\\n\\tdecoder := json.NewDecoder(req.Body)\\n\\n\\terr := decoder.Decode(&transaction)\\n\\n\\tif err != nil {\\n\\t\\thttp.Error(resp, \\\"Bad Request\\\", http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\n\\tapiErr := service.CreateTransaction(transaction)\\n\\n\\tif apiErr != nil {\\n\\t\\tresp.WriteHeader(http.StatusInternalServerError)\\n\\t\\tencoder := json.NewEncoder(resp)\\n\\t\\tencoder.Encode(apiErr)\\n\\t}\\n\\n\\tresp.WriteHeader(http.StatusOK)\\n\\n\\tjsonValue, _ := json.Marshal(utils.AppMsgs{http.StatusOK, \\\"Transaction was recorded\\\"})\\n\\n\\tresp.Write(jsonValue)\\n\\n}\",\n \"func (client *Client) CreateTransaction(txn *CreateTransaction) (_ *Response, err error) {\\n\\tpath := \\\"/transaction\\\"\\n\\turi := fmt.Sprintf(\\\"%s%s\\\", client.apiBaseURL, path)\\n\\n\\tb, err := json.Marshal(txn)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\treq, err := http.NewRequest(http.MethodPost, uri, bytes.NewBuffer(b))\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tresp, err := client.performRequest(req, string(b))\\n\\treturn resp, err\\n}\",\n \"func (db *Datastore) Create(txn transaction.Transaction) error {\\n\\tfmt.Println(`Creating txn with data:`, txn)\\n\\t// your DB operations to transactions ...\\n\\treturn nil\\n}\",\n \"func createTransaction(request Request) (Transaction, error) {\\n\\tamount, err := strconv.ParseFloat(request.Load_amount[1:], 64)\\n\\ttransaction := Transaction{\\n\\t\\tId: TransactionId(request.Id),\\n\\t\\tCustomer_id: CustomerId(request.Customer_id),\\n\\t\\tAmount: amount,\\n\\t\\tTime: request.Time,\\n\\t}\\n\\n\\treturn transaction, err\\n}\",\n \"func (db *MySQLDB) CreateTransaction(ctx context.Context, rec *TransactionRecord) (int, error) {\\n\\tfLog := mysqlLog.WithField(\\\"func\\\", \\\"CreateTransaction\\\")\\n\\n\\t// start db transaction\\n\\ttx, err := db.instance.BeginTx(ctx, nil)\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"db.instance.BeginTx got %s\\\", err.Error())\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\t// create transaction record\\n\\ttrans, err := tx.ExecContext(ctx, \\\"INSERT INTO transactions(user_id, date, grand_total, discount, reason) VALUES(?,?,?,?,?)\\\", rec.UserID, rec.Date, 0, rec.Discount, rec.Reason)\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"db.tx.ExecContext got %s\\\", err.Error())\\n\\t\\terrRollback := tx.Rollback()\\n\\t\\tif errRollback != nil {\\n\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\treturn 0, errRollback\\n\\t\\t}\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\ttID, err := trans.LastInsertId()\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"db.tx.ExecContext got %s\\\", err.Error())\\n\\t\\terrRollback := tx.Rollback()\\n\\t\\tif errRollback != nil {\\n\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\treturn 0, errRollback\\n\\t\\t}\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\tgrandTotal := 0\\n\\n\\t//loop tx detail\\n\\tfor i := 0; i < len(rec.TransactionDetail); i++ {\\n\\t\\tdetail := rec.TransactionDetail[i]\\n\\n\\t\\t//get product stock from products table\\n\\t\\tp, err := db.GetProductByID(ctx, detail.ProductID)\\n\\t\\tif err != nil {\\n\\t\\t\\tfLog.Errorf(\\\"db.tx.ExecContext got %s\\\", err.Error())\\n\\t\\t\\terrRollback := tx.Rollback()\\n\\t\\t\\tif errRollback != nil {\\n\\t\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\t\\treturn 0, errRollback\\n\\t\\t\\t}\\n\\t\\t\\treturn 0, err\\n\\t\\t}\\n\\n\\t\\t//check qty\\n\\t\\tif p.Qty-detail.Qty < 0 {\\n\\t\\t\\tfLog.Errorf(\\\"product qty is not enough\\\")\\n\\t\\t\\terrRollback := tx.Rollback()\\n\\t\\t\\tif errRollback != nil {\\n\\t\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\t\\treturn 0, errRollback\\n\\t\\t\\t}\\n\\t\\t\\treturn 0, fmt.Errorf(\\\"product qty is not enough\\\")\\n\\t\\t}\\n\\n\\t\\tqty := p.Qty - detail.Qty\\n\\t\\tsubTotal := p.Price * detail.Qty\\n\\t\\tgrandTotal = grandTotal + subTotal\\n\\n\\t\\t//update qty from products table\\n\\t\\t_, err = tx.ExecContext(ctx, \\\"UPDATE products SET qty=? WHERE id=?\\\", qty, detail.ProductID)\\n\\t\\tif err != nil {\\n\\t\\t\\tfLog.Errorf(\\\"db.tx.ExecContext got %s\\\", err.Error())\\n\\t\\t\\terrRollback := tx.Rollback()\\n\\t\\t\\tif errRollback != nil {\\n\\t\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\t\\treturn 0, errRollback\\n\\t\\t\\t}\\n\\t\\t\\treturn 0, err\\n\\t\\t}\\n\\n\\t\\t//insert transaction detail\\n\\t\\t_, err = tx.ExecContext(ctx, \\\"INSERT INTO transaction_detail(transaction_id, product_id, price, qty, sub_total) VALUES(?,?,?,?,?)\\\", tID, detail.ProductID, p.Price, detail.Qty, subTotal)\\n\\t\\tif err != nil {\\n\\t\\t\\tfLog.Errorf(\\\"db.tx.ExecContext got %s\\\", err.Error())\\n\\t\\t\\terrRollback := tx.Rollback()\\n\\t\\t\\tif errRollback != nil {\\n\\t\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\t\\treturn 0, errRollback\\n\\t\\t\\t}\\n\\t\\t\\treturn 0, err\\n\\t\\t}\\n\\t}\\n\\n\\tgrandTotal = grandTotal - rec.Discount\\n\\n\\t// update transaction grand total\\n\\t_, err = tx.ExecContext(ctx, \\\"UPDATE transactions SET grand_total=? WHERE id=?\\\", grandTotal, tID)\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"db.tx.ExecContext got %s\\\", err.Error())\\n\\t\\terrRollback := tx.Rollback()\\n\\t\\tif errRollback != nil {\\n\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\treturn 0, errRollback\\n\\t\\t}\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\t// commit transaction\\n\\terr = tx.Commit()\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"db.tx.ExecContext got %s\\\", err.Error())\\n\\t\\terrRollback := tx.Rollback()\\n\\t\\tif errRollback != nil {\\n\\t\\t\\tfLog.Errorf(\\\"error rollback, got %s\\\", err.Error())\\n\\t\\t\\treturn 0, errRollback\\n\\t\\t}\\n\\t\\treturn 0, err\\n\\t}\\n\\n\\treturn int(tID), nil\\n}\",\n \"func (facade *BranchManagerFacade) createTransaction(srcAccountId, destAccountId string, amount float32) *Transaction {\\n\\n\\tvar transaction = facade.transaction.create(srcAccountId, destAccountId, amount)\\n\\n\\treturn transaction\\n}\",\n \"func CreateTransaction(transaction *models.Transaction) error {\\n\\tsession, collection := service.Connect(collectionName)\\n\\tdefer session.Close()\\n\\n\\tif transaction.ID == \\\"\\\" {\\n\\t\\ttransaction.ID = bson.NewObjectId()\\n\\t}\\n\\n\\terr := collection.Insert(transaction)\\n\\n\\treturn err\\n}\",\n \"func (repository *transactionRepository) CreateTransaction(transaction model.Transaction) (err error) {\\n\\ttx, err := repository.dbAdapter.Begin()\\n\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\t_, err = tx.Exec(\\n\\t\\t\\\"INSERT INTO transactions(id, account_id, operation_type, amount) VALUES(?,?,?,?)\\\",\\n\\t\\ttransaction.GetID().GetValue(),\\n\\t\\ttransaction.GetAccount().GetID().GetValue(),\\n\\t\\ttransaction.GetOperationType().GetValue(),\\n\\t\\ttransaction.GetAmountValueByOperationType(),\\n\\t)\\n\\n\\tif err != nil {\\n\\t\\ttx.Rollback()\\n\\n\\t\\treturn\\n\\t}\\n\\n\\t_, err = tx.Exec(\\n\\t\\t\\\"UPDATE accounts set credit_limit = ? where id = ?\\\",\\n\\t\\ttransaction.GetAccount().GetAvailableCreditLimit().GetValue(),\\n\\t\\ttransaction.GetAccount().GetID().GetValue(),\\n\\t)\\n\\n\\tif err != nil {\\n\\t\\ttx.Rollback()\\n\\n\\t\\treturn\\n\\t}\\n\\n\\terr = tx.Commit()\\n\\n\\treturn\\n}\",\n \"func CreateTransaction(from uint, to uint, amount int) {\\n\\tdb := helpers.ConnectDB().Debug()\\n\\tdefer db.Close()\\n\\n\\ttransaction := &interfaces.Transaction{From: from, To: to, Amount: amount}\\n\\tdb.Create(&transaction)\\n}\",\n \"func (a API) CreateRawTransaction(cmd *btcjson.CreateRawTransactionCmd) (e error) {\\n\\tRPCHandlers[\\\"createrawtransaction\\\"].Call <-API{a.Ch, cmd, nil}\\n\\treturn\\n}\",\n \"func CreateTransactions(\\n\\trouter *mux.Router,\\n\\tpk cryptography.PrivateKey,\\n\\tuser users.User,\\n\\tserverSDK sdks.Servers,\\n\\troutePrefix string,\\n\\tsignedTrsBufferSize int,\\n\\tatomicTrsBufferSize int,\\n\\taggregatedTrsBufferSize int,\\n\\ttrsAggregationDelay time.Duration,\\n) *Transactions {\\n\\n\\t//channels:\\n\\tnewSignedTrs := make(chan signed_transactions.Transaction, signedTrsBufferSize)\\n\\tnewAtomicSignedTrs := make(chan signed_transactions.AtomicTransaction, atomicTrsBufferSize)\\n\\tnewAggregatedTrs := make(chan aggregated_transactions.Transactions, aggregatedTrsBufferSize)\\n\\n\\t//factories:\\n\\tmetaDataBuilderFactory := concrete_metadata.CreateBuilderFactory()\\n\\thtBuilderFactory := concrete_hashtrees.CreateHashTreeBuilderFactory()\\n\\tpublicKeyBuilderFactory := concrete_cryptography.CreatePublicKeyBuilderFactory()\\n\\tsigBuilderFactory := concrete_cryptography.CreateSignatureBuilderFactory(publicKeyBuilderFactory)\\n\\tuserSigBuilderFactory := concrete_users.CreateSignatureBuilderFactory(sigBuilderFactory, htBuilderFactory, metaDataBuilderFactory)\\n\\n\\t//transactions and blocks factories:\\n\\tblockChainMetaDataBuilderFactory := concrete_blockchain_metadata.CreateBuilderFactory()\\n\\tsignedTrsBuilderFactory := concrete_signed_transactions.CreateTransactionBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\\n\\tsignedTransBuilderFactory := concrete_signed_transactions.CreateTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\\n\\tsignedAtomicTransBuilderFactory := concrete_signed_transactions.CreateAtomicTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\\n\\tatomicSignedTrsBuilderFactory := concrete_signed_transactions.CreateAtomicTransactionBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\\n\\tsignedAggregatedTrsBuilderFactory := concrete_aggregated_transactions.CreateTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\\n\\n\\t//create the leader SDK:\\n\\tleadSDK := concrete_sdks.CreateLeaders(userSigBuilderFactory, routePrefix, pk, user)\\n\\n\\t//create the transaction API:\\n\\ttransactionsAPI := apis.CreateTransactions(\\n\\t\\troutePrefix,\\n\\t\\trouter,\\n\\t\\tsignedTrsBuilderFactory,\\n\\t\\tatomicSignedTrsBuilderFactory,\\n\\t\\tnewSignedTrs,\\n\\t\\tnewAtomicSignedTrs,\\n\\t)\\n\\n\\t//create the transaction agent:\\n\\ttrsAgent := agents.CreatePushTransactionsToLeaders(\\n\\t\\tsignedAggregatedTrsBuilderFactory,\\n\\t\\tsignedTransBuilderFactory,\\n\\t\\tsignedAtomicTransBuilderFactory,\\n\\t\\ttrsAggregationDelay,\\n\\t\\tnewSignedTrs,\\n\\t\\tnewAtomicSignedTrs,\\n\\t\\tnewAggregatedTrs,\\n\\t)\\n\\n\\tout := Transactions{\\n\\t\\tapi: transactionsAPI,\\n\\t\\tagent: trsAgent,\\n\\t\\tleadSDK: leadSDK,\\n\\t\\tservSDK: serverSDK,\\n\\t\\tnewAggregatedTrs: newAggregatedTrs,\\n\\t}\\n\\n\\treturn &out\\n}\",\n \"func (u *User) CreateTransaction(nodeID, data string, idempotencyKey ...string) (map[string]interface{}, error) {\\n\\tlog.info(\\\"========== CREATE TRANSACTION ==========\\\")\\n\\turl := buildURL(path[\\\"users\\\"], u.UserID, path[\\\"nodes\\\"], nodeID, path[\\\"transactions\\\"])\\n\\n\\treturn u.do(\\\"POST\\\", url, data, idempotencyKey)\\n}\",\n \"func (f formatter) CreateTransaction(t bill.Transaction, ins, outs []string) (string, error) {\\n\\titem := transactionItem{\\n\\t\\tRaw: t.Raw,\\n\\t\\tAt: t.At.Format(\\\"2006-01-02\\\"),\\n\\t\\tPayee: t.Payee,\\n\\t\\tTitle: t.Title,\\n\\t\\tPayeeAccounts: ins,\\n\\t\\tAccounts: outs,\\n\\t\\tAmount: t.Amount,\\n\\t\\tCurrency: t.Currency,\\n\\t}\\n\\tif item.PayeeAccounts == nil || len(item.PayeeAccounts) == 0 {\\n\\t\\titem.PayeeAccounts = []string{\\\"TODO\\\"}\\n\\t}\\n\\tif item.Accounts == nil || len(item.Accounts) == 0 {\\n\\t\\titem.Accounts = []string{\\\"TODO\\\"}\\n\\t}\\n\\n\\tvar tplBytes bytes.Buffer\\n\\ttpl, err := template.New(\\\"transaction\\\").Parse(transactionTpl)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", fmt.Errorf(\\\"template error, should be a bug: %w\\\", err)\\n\\t}\\n\\n\\terr = tpl.Execute(&tplBytes, item)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", fmt.Errorf(\\\"template exec error, should be a bug: %w\\\", err)\\n\\t}\\n\\n\\treturn tplBytes.String(), nil\\n}\",\n \"func (transaction *Transaction) create(srcAccountId, destAccountId string, amount float32) *Transaction {\\n\\n\\tfmt.Println(\\\"creating transaction\\\")\\n\\ttransaction.srcAccountId = srcAccountId\\n\\ttransaction.destAccountId = destAccountId\\n\\ttransaction.amount = amount\\n\\treturn transaction\\n}\",\n \"func createTransaction(\\n\\tctx context.Context,\\n\\tdb storage.Database,\\n\\tappserviceID string,\\n) (\\n\\ttransactionJSON []byte,\\n\\ttxnID, maxID int,\\n\\teventsRemaining bool,\\n\\terr error,\\n) {\\n\\t// Retrieve the latest events from the DB (will return old events if they weren't successfully sent)\\n\\ttxnID, maxID, events, eventsRemaining, err := db.GetEventsWithAppServiceID(ctx, appserviceID, transactionBatchSize)\\n\\tif err != nil {\\n\\t\\tlog.WithFields(log.Fields{\\n\\t\\t\\t\\\"appservice\\\": appserviceID,\\n\\t\\t}).WithError(err).Fatalf(\\\"appservice worker unable to read queued events from DB\\\")\\n\\n\\t\\treturn\\n\\t}\\n\\n\\t// Check if these events do not already have a transaction ID\\n\\tif txnID == -1 {\\n\\t\\t// If not, grab next available ID from the DB\\n\\t\\ttxnID, err = db.GetLatestTxnID(ctx)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn nil, 0, 0, false, err\\n\\t\\t}\\n\\n\\t\\t// Mark new events with current transactionID\\n\\t\\tif err = db.UpdateTxnIDForEvents(ctx, appserviceID, maxID, txnID); err != nil {\\n\\t\\t\\treturn nil, 0, 0, false, err\\n\\t\\t}\\n\\t}\\n\\n\\tvar ev []*gomatrixserverlib.HeaderedEvent\\n\\tfor i := range events {\\n\\t\\tev = append(ev, &events[i])\\n\\t}\\n\\n\\t// Create a transaction and store the events inside\\n\\ttransaction := gomatrixserverlib.ApplicationServiceTransaction{\\n\\t\\tEvents: gomatrixserverlib.HeaderedToClientEvents(ev, gomatrixserverlib.FormatAll),\\n\\t}\\n\\n\\ttransactionJSON, err = json.Marshal(transaction)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\treturn\\n}\",\n \"func (tm *TransactionMutex) Create(transactionType TransactionType, fileUUID string) {\\n\\ttm.mu.Lock()\\n\\tdefer tm.mu.Unlock()\\n\\tnewTransaction := Transaction{\\n\\t\\tUUID: NewUUID(),\\n\\t\\tCreationTimestamp: time.Now().Unix(),\\n\\t\\tType: transactionType,\\n\\t\\tTargetFileUUID: fileUUID,\\n\\t\\tVersion: config.Version,\\n\\t}\\n\\ttm.Transactions = append(tm.Transactions, newTransaction)\\n}\",\n \"func (tr *Transactor) CreateTransaction(ctx context.Context, ipfsHash string, tags []Tag, w WalletSigner, amount string, data []byte, target string) (*Transaction, error) {\\n\\tlastTx, err := tr.Client.TxAnchor(ctx)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tprice, err := tr.Client.GetReward(ctx, data)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\t// Non encoded transaction fields\\n\\ttx := NewTransaction(\\n\\t\\ttags,\\n\\t\\tipfsHash,\\n\\t\\tlastTx,\\n\\t\\tw.PubKeyModulus(),\\n\\t\\tamount,\\n\\t\\ttarget,\\n\\t\\tdata,\\n\\t\\tprice,\\n\\t)\\n\\n\\treturn tx, nil\\n}\",\n \"func CreateTransaction(session gocqlx.Session, transaction *Transaction) (*Transaction, error) {\\n\\tuuid, _ := gocql.RandomUUID()\\n\\ttransaction.ID = uuid\\n\\ttransaction.CreatedAt = time.Now()\\n\\n\\tlog.Print(transaction)\\n\\n\\tq := session.Query(transactionTable.Insert()).BindStruct(transaction)\\n\\tif err := q.ExecRelease(); err != nil {\\n\\t\\treturn transaction, err\\n\\t}\\n\\treturn transaction, nil\\n}\",\n \"func (u *createTrans) CreateTransaction(transaction *models.Transaction) error {\\n\\n\\tacc, err := u.accountServ.Get(transaction.Account.ID)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tnewCredit := acc.AvailableCreditLimit + transaction.Amount\\n\\tif newCredit < 0 {\\n\\t\\treturn models.ErrAmountBeyondCreditLimit\\n\\t}\\n\\n\\terr = u.tranServ.Create(transaction)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tacc.AvailableCreditLimit = newCredit\\n\\treturn u.accountServ.UpdateCreditLimit(acc)\\n}\",\n \"func (s *TransactionService) Create(walletID string, tp *TransactionParams) (*Transaction, error) {\\n\\tu := fmt.Sprintf(\\\"/kms/wallets/%s/transactions/\\\", walletID)\\n\\ttransaction := &Transaction{}\\n\\tp := &Params{}\\n\\tp.SetAuthProvider(s.auth)\\n\\terr := s.client.Call(http.MethodPost, u, tp, transaction, p)\\n\\treturn transaction, err\\n}\",\n \"func (s *Service) CreateTx(tx *sql.Tx, headerID uint, ms Models) error {\\n\\treturn s.storage.CreateTx(tx, headerID, ms)\\n}\",\n \"func addTransaction(db meddler.DB, lender *User, debtor *User, amount int, expense *Expense) (*Transaction, error) {\\n trans := new(Transaction)\\n trans.LenderId = lender.Id\\n trans.DebtorId = debtor.Id\\n trans.Amount = amount\\n trans.Date = expense.Date\\n trans.ExpenseId = expense.Id\\n\\n err := meddler.Insert(db, \\\"transactions\\\", trans)\\n if err != nil {\\n return nil, err\\n }\\n\\n lender.UpdateBalance(db, amount)\\n debtor.UpdateBalance(db, -amount)\\n\\n return trans, nil\\n}\",\n \"func (c Client) createTransaction(merchantID string, body io.Reader) (*TransactionID, error) {\\n\\tpath := fmt.Sprintf(\\\"/merchants/%s/transactions\\\", merchantID)\\n\\treq, err := http.NewRequest(\\\"POST\\\", c.getURL(path), body)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tvar marshalled map[string]*TransactionID\\n\\treturn marshalled[\\\"transaction\\\"], c.executeRequestAndMarshal(req, &marshalled)\\n}\",\n \"func (h *Harness) CreateTransaction(\\n\\ttargetOutputs []*wire.TxOut,\\n\\tfeeRate amt.Amount, change bool,\\n) (*wire.MsgTx, error) {\\n\\treturn h.wallet.CreateTransaction(targetOutputs, feeRate, change)\\n}\",\n \"func CreateTransaction(from, to, amount string) (txeB64 string) {\\n\\n\\ttx, err := build.Transaction(\\n\\t\\tbuild.SourceAccount{AddressOrSeed: from},\\n\\t\\tbuild.PublicNetwork,\\n\\t\\t//b.AutoSequence{SequenceProvider: horizon.DefaultPublicNetClient}, ???\\n\\t\\tbuild.Payment(\\n\\t\\t\\tbuild.Destination{AddressOrSeed: to},\\n\\t\\t\\tbuild.NativeAmount{Amount: amount},\\n\\t\\t),\\n\\t)\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\n\\ttxe, err := tx.Sign(from)\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\n\\ttxeB64, err = txe.Base64()\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\n\\treturn\\n}\",\n \"func Transaction(c *gin.Context) {\\n\\n\\tt_type,_ := strconv.Atoi(c.PostForm(\\\"transaction_type\\\")) // 1 : sales , 2 : missing products (hilang)\\n\\tstatus := 1\\n\\tmessage := \\\"Success\\\"\\n var responseTransaction ResponseTransaction\\n\\tvar newstocks int\\n\\tvar products Products\\n\\tvar products_arr []Products\\n\\tvar stock_ins_arr []Stock_Ins\\n\\tvar stock_outs Stock_Outs\\n\\tvar stock_ins Stock_Ins\\n\\tvar note string\\n\\ttransaction_id := \\\"\\\"\\n\\tsellPrice,_ := strconv.Atoi(c.PostForm(\\\"sell_price\\\"))\\n\\tvar buyPrice int\\n\\tqtY,_ := strconv.Atoi(c.PostForm(\\\"qty\\\"))\\n\\tcurrentdatetime := time.Now().Format(\\\"2006-01-02 15:04:05\\\")\\n\\tdb := InitDb() //db intiate\\n\\t//get data products\\n\\tdb.Where(\\\"sku = ?\\\", c.PostForm(\\\"sku\\\")).First(&products).Limit(1).Scan(&products_arr)\\n\\n\\t//check if the sku is exist?\\n\\tif(len(products_arr) > 0) {\\n\\t\\ttx := db.Begin()\\n\\n\\t\\t/**\\n\\t * Identify product is gone / transaction by sales\\n\\t */\\n\\n\\t\\tif (t_type == 1) {\\n\\n\\t\\t\\ttransaction_id = generateTransactionID()\\n\\n\\t\\t\\t//get data products\\n\\t\\t\\tdb.Where(\\\"sku = ?\\\", c.PostForm(\\\"sku\\\")).First(&stock_ins).Limit(1).Scan(&stock_ins_arr)\\n\\n\\t\\t\\t// get the data stock after transaction\\n\\t\\t\\tfor i,element := range stock_ins_arr{\\n\\t\\t\\t\\tif (i == 0) {\\n\\t\\t\\t\\t\\tbuyPrice = element.Buy_Price\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\n\\t\\t\\tnote = \\\"Pesanan \\\"+transaction_id\\n\\t\\t\\ttransactions := Transactions{Id:transaction_id,Buy_Price:buyPrice,Sell_Price:sellPrice,Qty:qtY,Sku:c.PostForm(\\\"sku\\\"),Created_Date:currentdatetime}\\n\\t\\t\\tif err := tx.Create(&transactions).Error; err != nil {\\n\\t\\t\\t\\ttx.Rollback()\\n\\t\\t\\t\\tstatus = 0\\n\\t\\t\\t\\tmessage = \\\"failed to insert data transaction\\\"\\n\\t\\t\\t}\\n\\n\\n\\t\\t} else if (t_type == 2) {\\n\\n\\t\\t\\tnote = \\\"Barang Hilang\\\"\\n\\n\\t\\t}\\n\\t\\t//insert data to stock_outs\\n\\t\\tstock_outs = Stock_Outs{Sku:c.PostForm(\\\"sku\\\"),Created_Date:currentdatetime,Qty:qtY,Note:note,Transaction_Id:transaction_id}\\n\\t\\tif err := tx.Create(&stock_outs).Error; err != nil {\\n\\t\\t\\ttx.Rollback()\\n\\t\\t\\tstatus = 0\\n\\t\\t\\tmessage = \\\"failed to insert data stocks_outs\\\"\\n\\t\\t}\\n\\n\\t\\t// get the data stock after transaction\\n\\t\\tfor i,element := range products_arr{\\n\\t\\t\\tif (i == 0) {\\n\\t\\t\\t\\tnewstocks = element.Stocks - qtY\\n\\t\\t\\t}\\n\\t\\t}\\n\\n\\t\\t//update product stocks in table products\\n\\t\\tif err := tx.Model(&products).Where(\\\"sku = ?\\\", c.PostForm(\\\"sku\\\")).Update(\\\"stocks\\\", newstocks).Error; err != nil {\\n\\t\\t\\ttx.Rollback()\\n\\t\\t\\tstatus = 0\\n\\t\\t\\tmessage = \\\"failed to update data products\\\"\\n\\t\\t}\\n\\n\\n\\t\\t//transaction commit\\n\\t\\ttx.Commit()\\n\\t}else{\\n\\t\\tstatus = 0\\n\\t\\tmessage = \\\"SKU Not found!\\\"\\n\\t}\\n\\n\\tif status == 1{\\n\\t\\tresponseTransaction = ResponseTransaction{Status:status,Message:message,Data:DataTransaction{Sku:c.PostForm(\\\"sku\\\"),Buy_Price:buyPrice,Sell_Price:sellPrice,Created_Date:currentdatetime,Product_name:c.PostForm(\\\"product_name\\\"),Stocks:newstocks,Transaction_Id:transaction_id}}\\n\\t}else{\\n\\t\\tresponseTransaction = ResponseTransaction{Status:status,Message:message}\\n\\t}\\n\\n\\t// Close connection database\\n\\tdefer db.Close()\\n\\tc.JSON(200, responseTransaction)\\n}\",\n \"func (c *Constructor) createTransaction(\\n\\tctx context.Context,\\n\\tintent []*types.Operation,\\n) (*types.TransactionIdentifier, string, error) {\\n\\tmetadataRequest, err := c.helper.Preprocess(\\n\\t\\tctx,\\n\\t\\tc.network,\\n\\t\\tintent,\\n\\t\\tnil,\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to preprocess\\\", err)\\n\\t}\\n\\n\\trequiredMetadata, err := c.helper.Metadata(\\n\\t\\tctx,\\n\\t\\tc.network,\\n\\t\\tmetadataRequest,\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to construct metadata\\\", err)\\n\\t}\\n\\n\\tunsignedTransaction, payloads, err := c.helper.Payloads(\\n\\t\\tctx,\\n\\t\\tc.network,\\n\\t\\tintent,\\n\\t\\trequiredMetadata,\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to construct payloads\\\", err)\\n\\t}\\n\\n\\tparsedOps, signers, _, err := c.helper.Parse(\\n\\t\\tctx,\\n\\t\\tc.network,\\n\\t\\tfalse,\\n\\t\\tunsignedTransaction,\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to parse unsigned transaction\\\", err)\\n\\t}\\n\\n\\tif len(signers) != 0 {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\n\\t\\t\\t\\\"signers should be empty in unsigned transaction but found %d\\\",\\n\\t\\t\\tlen(signers),\\n\\t\\t)\\n\\t}\\n\\n\\tif err := c.parser.ExpectedOperations(intent, parsedOps, false, false); err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unsigned parsed ops do not match intent\\\", err)\\n\\t}\\n\\n\\tsignatures, err := c.helper.Sign(ctx, payloads)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to sign payloads\\\", err)\\n\\t}\\n\\n\\tnetworkTransaction, err := c.helper.Combine(\\n\\t\\tctx,\\n\\t\\tc.network,\\n\\t\\tunsignedTransaction,\\n\\t\\tsignatures,\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to combine signatures\\\", err)\\n\\t}\\n\\n\\tsignedParsedOps, signers, _, err := c.helper.Parse(\\n\\t\\tctx,\\n\\t\\tc.network,\\n\\t\\ttrue,\\n\\t\\tnetworkTransaction,\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to parse signed transaction\\\", err)\\n\\t}\\n\\n\\tif err := c.parser.ExpectedOperations(intent, signedParsedOps, false, false); err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: signed parsed ops do not match intent\\\", err)\\n\\t}\\n\\n\\tif err := parser.ExpectedSigners(payloads, signers); err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: signed transactions signers do not match intent\\\", err)\\n\\t}\\n\\n\\ttransactionIdentifier, err := c.helper.Hash(\\n\\t\\tctx,\\n\\t\\tc.network,\\n\\t\\tnetworkTransaction,\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn nil, \\\"\\\", fmt.Errorf(\\\"%w: unable to get transaction hash\\\", err)\\n\\t}\\n\\n\\treturn transactionIdentifier, networkTransaction, nil\\n}\",\n \"func (service ServiceTx) Create(edgeJob *portainer.EdgeJob) error {\\n\\treturn service.CreateWithID(portainer.EdgeJobID(service.GetNextIdentifier()), edgeJob)\\n}\",\n \"func (c *TransactionClient) Create() *TransactionCreate {\\n\\tmutation := newTransactionMutation(c.config, OpCreate)\\n\\treturn &TransactionCreate{config: c.config, hooks: c.Hooks(), mutation: mutation}\\n}\",\n \"func (gw *Gateway) CreateTransaction(params wallet.CreateTransactionParams) (*coin.Transaction, []wallet.UxBalance, error) {\\n\\tif !gw.Config.EnableWalletAPI {\\n\\t\\treturn nil, nil, wallet.ErrWalletAPIDisabled\\n\\t}\\n\\n\\tvar txn *coin.Transaction\\n\\tvar inputs []wallet.UxBalance\\n\\tvar err error\\n\\tgw.strand(\\\"CreateTransaction\\\", func() {\\n\\t\\ttxn, inputs, err = gw.v.CreateTransaction(params)\\n\\t})\\n\\treturn txn, inputs, err\\n}\",\n \"func (p *PsqlInvoiceHeader) CreateTx(tx *sql.Tx, m *invoiceheader.Model) error {\\n\\tstmt, err := tx.Prepare(psqlCreateInvoiceHeader)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tdefer stmt.Close()\\n\\n\\treturn stmt.QueryRow(m.Client).Scan(&m.ID, &m.CreatedAt)\\n}\",\n \"func (c Client) CreateTransaction(merchantID string, dto TransactionDTO) (*TransactionID, error) {\\n\\tb, err := json.Marshal(dto)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\treturn c.createTransaction(merchantID, bytes.NewBuffer(b))\\n}\",\n \"func (t *explainTablet) CreateTransaction(ctx context.Context, target *querypb.Target, dtid string, participants []*querypb.Target) (err error) {\\n\\tt.mu.Lock()\\n\\tt.currentTime = t.vte.batchTime.Wait()\\n\\tt.mu.Unlock()\\n\\treturn t.tsv.CreateTransaction(ctx, target, dtid, participants)\\n}\",\n \"func (u *User) CreateDummyTransaction(nodeID string, queryParams ...string) (map[string]interface{}, error) {\\n\\tlog.info(\\\"========== CREATE DUMMY TRANSACTION ==========\\\")\\n\\turl := buildURL(path[\\\"users\\\"], u.UserID, path[\\\"nodes\\\"], nodeID) + \\\"/dummy-tran\\\"\\n\\n\\treturn u.do(\\\"GET\\\", url, \\\"\\\", queryParams)\\n}\",\n \"func (tx *Tx) Create(ls Lease) error {\\n\\ttx.leases = append(tx.leases, ls)\\n\\ttx.ops = append(tx.ops, Op{\\n\\t\\tType: Create,\\n\\t\\tLease: ls,\\n\\t})\\n\\tsort.Sort(tx.leases)\\n\\treturn nil\\n}\",\n \"func (m *Mysql) CreateTX(tx *sql.Tx, paramAudit *paramAuditDom.ParamsAudit) error {\\n\\tstmt, err := tx.Prepare(mysqlInsert)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tdefer stmt.Close()\\n\\n\\tnow := utils.Now()\\n\\terr = mysql.ExecAffectingOneRow(\\n\\t\\tstmt,\\n\\t\\tparamAudit.PrevParam.ParamID,\\n\\t\\tparamAudit.PrevParam.Name,\\n\\t\\tparamAudit.PrevParam.Value,\\n\\t\\tparamAudit.PrevParam.Active,\\n\\t\\tparamAudit.Param.Name,\\n\\t\\tparamAudit.Param.Value,\\n\\t\\tparamAudit.Param.Active,\\n\\t\\tparamAudit.SetUserID,\\n\\t\\tnow[\\\"date\\\"], now[\\\"time\\\"], now[\\\"unix\\\"],\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (nos *NewOrderServiceImpl) ProcessTransaction(req *models.NewOrder) (*models.NewOrderOutput, error) {\\n\\tlog.Printf(\\\"Starting the New Order Transaction for row: c=%d w=%d d=%d n=%d\\\", req.CustomerID, req.WarehouseID, req.DistrictID, req.UniqueItems)\\n\\n\\tresult, err := nos.execute(req)\\n\\tif err != nil {\\n\\t\\treturn nil, fmt.Errorf(\\\"error occured while executing the new order transaction. Err: %v\\\", err)\\n\\t}\\n\\n\\tlog.Printf(\\\"Completed the New Order Transaction for row: c=%d w=%d d=%d n=%d\\\", req.CustomerID, req.WarehouseID, req.DistrictID, req.UniqueItems)\\n\\treturn result, nil\\n}\",\n \"func (c *Jrpc) CreateRawTransaction(in *pty.ReqCreatePrivacyTx, result *interface{}) error {\\n\\treply, err := c.cli.CreateRawTransaction(context.Background(), in)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t*result = hex.EncodeToString(types.Encode(reply))\\n\\treturn err\\n}\",\n \"func (b *Store) CreateTx(ctx context.Context, tx *sql.Tx, userID, username, password string) error {\\n\\terr := permission.LimitCheckAny(ctx, permission.System, permission.Admin, permission.MatchUser(userID))\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\terr = validate.Many(\\n\\t\\tvalidate.UUID(\\\"UserID\\\", userID),\\n\\t\\tvalidate.UserName(\\\"UserName\\\", username),\\n\\t\\tvalidate.Text(\\\"Password\\\", password, 8, 200),\\n\\t)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\thashedPassword, err := bcrypt.GenerateFromPassword([]byte(password), passCost)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\t_, err = tx.StmtContext(ctx, b.insert).ExecContext(ctx, userID, username, string(hashedPassword))\\n\\treturn err\\n}\",\n \"func (s *transactionStore) Create(ctx context.Context, transaction *configapi.Transaction) error {\\n\\tif transaction.ID == \\\"\\\" {\\n\\t\\ttransaction.ID = newTransactionID()\\n\\t}\\n\\tif transaction.Version != 0 {\\n\\t\\treturn errors.NewInvalid(\\\"not a new object\\\")\\n\\t}\\n\\tif transaction.Revision != 0 {\\n\\t\\treturn errors.NewInvalid(\\\"not a new object\\\")\\n\\t}\\n\\ttransaction.Revision = 1\\n\\ttransaction.Created = time.Now()\\n\\ttransaction.Updated = time.Now()\\n\\n\\t// Append a new entry to the transaction log.\\n\\tentry, err := s.transactions.Append(ctx, transaction.ID, transaction)\\n\\tif err != nil {\\n\\t\\treturn errors.FromAtomix(err)\\n\\t}\\n\\ttransaction.Index = configapi.Index(entry.Index)\\n\\ttransaction.Version = uint64(entry.Version)\\n\\treturn nil\\n}\",\n \"func (c *Constructor) CreateTransactions(\\n\\tctx context.Context,\\n\\tclearBroadcasts bool,\\n) error {\\n\\t// Before starting loop, delete any pending broadcasts if configuration\\n\\t// indicates to do so.\\n\\tif clearBroadcasts {\\n\\t\\tbroadcasts, err := c.helper.ClearBroadcasts(ctx)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to clear broadcasts\\\", err)\\n\\t\\t}\\n\\n\\t\\tlog.Printf(\\n\\t\\t\\t\\\"Cleared pending %d broadcasts: %s\\\\n\\\",\\n\\t\\t\\tlen(broadcasts),\\n\\t\\t\\ttypes.PrettyPrintStruct(broadcasts),\\n\\t\\t)\\n\\t}\\n\\n\\tfor ctx.Err() == nil {\\n\\t\\tsender, balance, coinIdentifier, err := c.findSender(ctx)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to find sender\\\", err)\\n\\t\\t}\\n\\n\\t\\t// Determine Action\\n\\t\\tscenarioCtx, scenarioOps, err := c.generateScenario(\\n\\t\\t\\tctx,\\n\\t\\t\\tsender,\\n\\t\\t\\tbalance,\\n\\t\\t\\tcoinIdentifier,\\n\\t\\t)\\n\\t\\tif errors.Is(err, ErrInsufficientFunds) {\\n\\t\\t\\tbroadcasts, err := c.helper.AllBroadcasts(ctx)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to get broadcasts\\\", err)\\n\\t\\t\\t}\\n\\n\\t\\t\\tif len(broadcasts) > 0 {\\n\\t\\t\\t\\t// we will wait for in-flight to process\\n\\t\\t\\t\\ttime.Sleep(defaultSleepTime * time.Second)\\n\\t\\t\\t\\tcontinue\\n\\t\\t\\t}\\n\\n\\t\\t\\tif err := c.generateNewAndRequest(ctx); err != nil {\\n\\t\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to generate new address\\\", err)\\n\\t\\t\\t}\\n\\n\\t\\t\\tcontinue\\n\\t\\t} else if err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to generate intent\\\", err)\\n\\t\\t}\\n\\n\\t\\tintent, err := scenario.PopulateScenario(ctx, scenarioCtx, scenarioOps)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to populate scenario\\\", err)\\n\\t\\t}\\n\\n\\t\\t// Create transaction\\n\\t\\ttransactionIdentifier, networkTransaction, err := c.createTransaction(ctx, intent)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\n\\t\\t\\t\\t\\\"%w: unable to create transaction with operations %s\\\",\\n\\t\\t\\t\\terr,\\n\\t\\t\\t\\ttypes.PrettyPrintStruct(intent),\\n\\t\\t\\t)\\n\\t\\t}\\n\\n\\t\\tlogger.LogScenario(scenarioCtx, transactionIdentifier, c.currency)\\n\\n\\t\\t// Broadcast Transaction\\n\\t\\terr = c.helper.Broadcast(\\n\\t\\t\\tctx,\\n\\t\\t\\tsender,\\n\\t\\t\\tintent,\\n\\t\\t\\ttransactionIdentifier,\\n\\t\\t\\tnetworkTransaction,\\n\\t\\t)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to enqueue transaction for broadcast\\\", err)\\n\\t\\t}\\n\\n\\t\\tif err := c.handler.TransactionCreated(ctx, sender, transactionIdentifier); err != nil {\\n\\t\\t\\treturn fmt.Errorf(\\\"%w: unable to handle transaction creation\\\", err)\\n\\t\\t}\\n\\t}\\n\\n\\treturn ctx.Err()\\n}\",\n \"func Transaction(rt *Runtime, c chan goengage.Fundraise) (err error) {\\n\\trt.Log.Println(\\\"Transaction: start\\\")\\n\\tfor true {\\n\\t\\tr, ok := <-c\\n\\t\\tif !ok {\\n\\t\\t\\tbreak\\n\\t\\t}\\n\\n\\t\\trt.Log.Printf(\\\"%v Transaction\\\\n\\\", r.ActivityID)\\n\\t\\tif rt.GoodYear(r.ActivityDate) {\\n\\t\\t\\tif len(r.Transactions) != 0 {\\n\\t\\t\\t\\tfor _, c := range r.Transactions {\\n\\t\\t\\t\\t\\tc.ActivityID = r.ActivityID\\n\\t\\t\\t\\t\\trt.DB.Create(&c)\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\t\\t}\\n\\t}\\n\\trt.Log.Println(\\\"Transaction: end\\\")\\n\\treturn nil\\n}\",\n \"func (a *ChainAdaptor) CreateUtxoTransaction(req *proto.CreateUtxoTransactionRequest) (*proto.CreateUtxoTransactionReply, error) {\\n\\tvinNum := len(req.Vins)\\n\\tvar totalAmountIn, totalAmountOut int64\\n\\n\\tif vinNum == 0 {\\n\\t\\terr := fmt.Errorf(\\\"no Vin in req:%v\\\", req)\\n\\t\\treturn &proto.CreateUtxoTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\t// check the Fee\\n\\tfee, ok := big.NewInt(0).SetString(req.Fee, 0)\\n\\tif !ok {\\n\\t\\terr := errors.New(\\\"CreateTransaction, fail to get fee\\\")\\n\\t\\treturn &proto.CreateUtxoTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\tfor _, in := range req.Vins {\\n\\t\\ttotalAmountIn += in.Amount\\n\\t}\\n\\n\\tfor _, out := range req.Vouts {\\n\\t\\ttotalAmountOut += out.Amount\\n\\t}\\n\\n\\tif totalAmountIn != totalAmountOut+fee.Int64() {\\n\\t\\terr := errors.New(\\\"CreateTransaction, total amount in != total amount out + fee\\\")\\n\\t\\treturn &proto.CreateUtxoTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\trawTx, err := a.createRawTx(req.Vins, req.Vouts)\\n\\tif err != nil {\\n\\t\\treturn &proto.CreateUtxoTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\tbuf := bytes.NewBuffer(make([]byte, 0, rawTx.SerializeSize()))\\n\\terr = rawTx.Serialize(buf)\\n\\tif err != nil {\\n\\t\\treturn &proto.CreateUtxoTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\t// build the pkScript and Generate signhash for each Vin,\\n\\tsignHashes, err := a.calcSignHashes(req.Vins, req.Vouts)\\n\\tif err != nil {\\n\\t\\treturn &proto.CreateUtxoTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\tlog.Info(\\\"CreateTransaction\\\", \\\"usigned tx\\\", hex.EncodeToString(buf.Bytes()))\\n\\n\\treturn &proto.CreateUtxoTransactionReply{\\n\\t\\tCode: proto.ReturnCode_SUCCESS,\\n\\t\\tTxData: buf.Bytes(),\\n\\t\\tSignHashes: signHashes,\\n\\t}, nil\\n}\",\n \"func (c *RPC) CreateTransaction(recvAddr string, amount uint64) (*coin.Transaction, error) {\\n\\t// TODO -- this can support sending to multiple receivers at once,\\n\\t// which would be necessary if the exchange was busy\\n\\tsendAmount := cli.SendAmount{\\n\\t\\tAddr: recvAddr,\\n\\t\\tCoins: amount,\\n\\t}\\n\\n\\tif err := validateSendAmount(sendAmount); err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\ttxn, err := cli.CreateRawTxFromWallet(c.rpcClient, c.walletFile, c.changeAddr, []cli.SendAmount{sendAmount})\\n\\tif err != nil {\\n\\t\\treturn nil, RPCError{err}\\n\\t}\\n\\n\\treturn txn, nil\\n}\",\n \"func Db_create(db_name string) string{\\n\\t// Open the my.db data_backup file in your current directory.\\n\\t// It will be created if it doesn't exist.\\n\\tdb, err := bolt.Open(db_name, 0600, nil)\\n\\t//defer db.Close()\\n\\tif err != nil {\\n\\t\\tlog.Fatal(err)\\n\\t}\\n\\ttx, err := db.Begin(true)\\n\\tif err != nil {\\n\\t\\tfmt.Println(\\\"Error : \\\",err)\\n\\t\\tdb.Close()\\n\\t\\treturn \\\"ERROR,DB_COMMIT\\\"\\n\\t}\\n\\tfmt.Printf(\\\"Created DB with ID %d\\\", tx.ID())\\n\\tdb.Close()\\n\\treturn \\\"OK,DB_COMMIT\\\"\\n}\",\n \"func (t *SimpleChaincode) create_event(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\\n\\tvar tEvent TransactionEvent\\n\\t\\n\\ttranID \\t\\t\\t:= \\\"\\\\\\\"TranID\\\\\\\":\\\\\\\"\\\"+args[0]+\\\"\\\\\\\", \\\"\\n\\tsenderName \\t\\t:= \\\"\\\\\\\"SenderName\\\\\\\":\\\\\\\"\\\"+args[1]+\\\"\\\\\\\", \\\"\\n\\tsenderCountry := \\\"\\\\\\\"SenderCountry\\\\\\\":\\\\\\\"\\\"+args[2]+\\\"\\\\\\\", \\\"\\n\\treceiverName \\t:= \\\"\\\\\\\"ReceiverName\\\\\\\":\\\\\\\"\\\"+args[3]+\\\"\\\\\\\", \\\"\\n\\treceiverCountry := \\\"\\\\\\\"ReceiverCountry\\\\\\\":\\\\\\\"\\\"+args[4]+\\\"\\\\\\\", \\\"\\n\\tamount \\t\\t\\t:= \\\"\\\\\\\"Amount\\\\\\\":\\\\\\\"\\\"+args[5]+\\\"\\\\\\\", \\\"\\n\\n // Concatenates the variables to create the total JSON object\\n\\tevent_json := \\\"{\\\"+tranID+senderName+senderCountry+receiverName+receiverCountry+amount+\\\"}\\\" \\t\\t\\n\\t// Convert the JSON defined above into a TransactionEvent object for go\\n\\terr := json.Unmarshal([]byte(event_json), &tEvent)\\t\\t\\t\\t\\t\\t\\t\\t\\t\\t\\n\\tif err != nil { \\n\\t\\treturn nil, errors.New(\\\"Invalid JSON object\\\") \\n\\t}\\n\\n\\tbytes, err := json.Marshal(tEvent)\\n\\tif err != nil { \\n\\t\\treturn nil, errors.New(\\\"Error converting transaction event\\\") \\n\\t}\\n\\n\\t// Save new tran event record\\n\\terr = stub.PutState(tEvent.TranID, bytes)\\n\\tif err != nil { \\n\\t\\tfmt.Printf(\\\"create_event: Error storing transaction event: %s\\\", err); \\n\\t\\treturn nil, errors.New(\\\"Error storing transaction event\\\") \\n\\t}\\n\\n\\t// Update tranIDs with newly created ID and store it in chain.\\n\\tbytes, err = stub.GetState(\\\"tranIDs\\\")\\n\\tif err != nil { \\n\\t\\treturn nil, errors.New(\\\"Unable to get tranIDs\\\") \\n\\t}\\n\\n\\tvar tranHld TRAN_Holder\\n\\terr = json.Unmarshal(bytes, &tranHld)\\n\\tif err != nil {\\t\\n\\t\\treturn nil, errors.New(\\\"Corrupt TRAN_Holder record\\\") \\n\\t}\\n\\n\\ttranHld.TranIDs = append(tranHld.TranIDs, args[0])\\n\\tbytes, err = json.Marshal(tranHld)\\n\\n\\terr = stub.PutState(\\\"tranIDs\\\", bytes)\\n\\tif err != nil { \\n\\t\\tfmt.Printf(\\\"create_event: Error storing TranIDs: %s\\\", err); \\n\\t\\treturn nil, errors.New(\\\"Error storing TranIDs\\\") \\n\\t}\\n\\n\\treturn nil, nil \\n}\",\n \"func (t *Procure2Pay) CreatePurchaseOrder(stub shim.ChaincodeStubInterface, args []string) pb.Response {\\n\\n var objpurchaseOrder purchaseOrder\\n\\tvar objitem item\\n\\tvar err error\\n\\t\\n\\tfmt.Println(\\\"Entering CreatePurchaseOrder\\\")\\n\\n\\tif len(args) < 1 {\\n\\t\\tfmt.Println(\\\"Invalid number of args\\\")\\n\\t\\treturn shim.Error(err.Error())\\n\\t}\\n\\n\\tfmt.Println(\\\"Args [0] is : %v\\\\n\\\", args[0])\\n\\n\\t//unmarshal customerInfo data from UI to \\\"customerInfo\\\" struct\\n\\terr = json.Unmarshal([]byte(args[0]), &objpurchaseOrder)\\n\\tif err != nil {\\n\\t\\tfmt.Printf(\\\"Unable to unmarshal CreatePurchaseOrder input purchaseOrder: %s\\\\n\\\", err)\\n\\t\\treturn shim.Error(err.Error())\\n\\t\\t}\\n\\n\\tfmt.Println(\\\"purchase order object PO ID variable value is : %s\\\\n\\\", objpurchaseOrder.POID)\\n\\tfmt.Println(\\\"purchase order object PO ID variable value is : %s\\\\n\\\", objpurchaseOrder.Quantity)\\n\\n\\t// Data insertion for Couch DB starts here \\n\\ttransJSONasBytes, err := json.Marshal(objpurchaseOrder)\\n\\terr = stub.PutState(objpurchaseOrder.POID, transJSONasBytes)\\n\\t// Data insertion for Couch DB ends here\\n\\n\\t//unmarshal LoanTransactions data from UI to \\\"LoanTransactions\\\" struct\\n\\terr = json.Unmarshal([]byte(args[0]), &objitem)\\n\\tif err != nil {\\n\\t\\tfmt.Printf(\\\"Unable to unmarshal CreatePurchaseOrder input purchaseOrder: %s\\\\n\\\", err)\\n\\t\\treturn shim.Error(err.Error())\\n\\t\\t}\\n\\n\\tfmt.Println(\\\"item object Item ID variable value is : %s\\\\n\\\", objitem.ItemID)\\n\\n\\t// Data insertion for Couch DB starts here \\n\\ttransJSONasBytesLoan, err := json.Marshal(objitem)\\n\\terr = stub.PutState(objitem.ItemID, transJSONasBytesLoan)\\n\\t// Data insertion for Couch DB ends here\\n\\n\\tfmt.Println(\\\"Create Purchase Order Successfully Done\\\")\\n\\n\\tif err != nil {\\n\\t\\tfmt.Printf(\\\"\\\\nUnable to make transevent inputs : %v \\\", err)\\n\\t\\treturn shim.Error(err.Error())\\n\\t\\t//return nil,nil\\n\\t}\\n\\treturn shim.Success(nil)\\n}\",\n \"func (r RepresentativeRepo) CreateTx(tx *sql.Tx, repr d.Representative) error {\\n\\terr := r.UserRepo.CreateTx(tx, repr.User)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tconst insertQuery = `INSERT INTO \\\"Representative\\\"(representative_id, job_title, ref_user, ref_company)\\n\\tVALUES ($1, $2, $3, $4);`\\n\\t_, err = tx.Exec(insertQuery,\\n\\t\\trepr.ID,\\n\\t\\trepr.JobTitle,\\n\\t\\trepr.User.ID,\\n\\t\\trepr.CompanyID,\\n\\t)\\n\\n\\tif err != nil {\\n\\t\\t_ = tx.Rollback()\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (p *PgDb) CreateEmpTx(r *model.Emp, tx *sqlx.Tx, isValidate bool) (*model.Emp, error) {\\n\\tcnt := \\\"postgres (p *PgDb) CreateEmpTx\\\" // Имя текущего метода для логирования\\n\\t//mylog.PrintfDebug(\\\"[DEBUG] %v - START, param: '%+v'\\\", cnt, r)\\n\\n\\t// Проверяем определен ли контекст транзакции\\n\\tif tx == nil {\\n\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - tx *sqlx.Tx is NULL\\\", cnt)\\n\\t\\tlog.Printf(errM)\\n\\t\\treturn nil, errors.New(errM)\\n\\t}\\n\\n\\t//=====================================================================\\n\\t// Добавить валидацию входной структуры\\n\\t//=====================================================================\\n\\n\\t// =====================================================================\\n\\t// Переменная часть кода\\n\\t// =====================================================================\\n\\t// Если запускаем с проверками\\n\\tif isValidate {\\n\\t\\t{ //Если Dept NULL или НЕ существует, то ошибка\\n\\t\\t\\tif !r.Deptno.Valid {\\n\\t\\t\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - r.Deptno is NULL\\\", cnt)\\n\\t\\t\\t\\tlog.Printf(errM)\\n\\t\\t\\t\\treturn nil, errors.New(errM)\\n\\t\\t\\t}\\n\\t\\t\\tdeptno := int(r.Deptno.Int64)\\n\\t\\t\\t// Запрос в транзакции\\n\\t\\t\\tdeptExists, err := p.DeptExistsTx(deptno, tx)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - p.DeptExistsTx(deptno, tx), args = '%v'\\\", cnt, deptno)\\n\\t\\t\\t\\tlog.Printf(errM)\\n\\t\\t\\t\\treturn nil, errors.WithMessage(err, errM)\\n\\t\\t\\t}\\n\\t\\t\\tif !deptExists {\\n\\t\\t\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - dept '%v' does not exist\\\", cnt, deptno)\\n\\t\\t\\t\\tlog.Printf(errM)\\n\\t\\t\\t\\treturn nil, errors.New(errM)\\n\\t\\t\\t}\\n\\t\\t\\t//mylog.PrintfDebug(\\\"[DEBUG] %v - dept %v exists\\\", cnt, deptno)\\n\\t\\t}\\n\\t\\t{ //Если Emp существует, то игнорируем\\n\\t\\t\\texists, err := p.EmpExistsTx(r.Empno, tx)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - p.EmpExistsTx(r.Empno, tx), args = '%v'\\\", cnt, r.Empno)\\n\\t\\t\\t\\tlog.Printf(errM)\\n\\t\\t\\t\\treturn nil, errors.WithMessage(err, errM)\\n\\t\\t\\t}\\n\\t\\t\\t// Если запись существует, то ни чего не делаем, возвращем, что пришло на вход\\n\\t\\t\\tif exists {\\n\\t\\t\\t\\terrM := fmt.Sprintf(\\\"[WARN] %v - WARN - emp '%v' already exist - nothing to do\\\", cnt, r.Empno)\\n\\t\\t\\t\\tlog.Printf(errM)\\n\\t\\t\\t\\treturn nil, nil\\n\\t\\t\\t}\\n\\t\\t\\t//mylog.PrintfDebug(\\\"[DEBUG] %v - emp '%v' does not exist - can be created\\\", cnt, r.Empno)\\n\\t\\t}\\n\\t}\\n\\t// =====================================================================\\n\\n\\t// =====================================================================\\n\\t// Переменная часть кода\\n\\t// =====================================================================\\n\\tstmText := sqlInsertEmpText\\n\\t// =====================================================================\\n\\n\\t//Выполняем команду\\n\\tres, err := tx.NamedExec(stmText, r)\\n\\tif err != nil {\\n\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - tx.NamedExec(stmText, r), args = '%+v'\\\", cnt, r)\\n\\t\\tlog.Printf(errM)\\n\\t\\treturn nil, errors.Wrap(err, errM)\\n\\t}\\n\\n\\t{ // Необязательная часть - можно удалить в последствии\\n\\t\\t// Проверим количество обработанных строк\\n\\t\\trowCount, err := res.RowsAffected()\\n\\t\\t_ = rowCount\\n\\t\\t//mylog.PrintfDebug(\\\"[DEBUG] %v -- process %v rows\\\", cnt, rowCount)\\n\\t\\tif err != nil {\\n\\t\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - res.RowsAffected()\\\", cnt)\\n\\t\\t\\tlog.Printf(errM)\\n\\t\\t\\treturn nil, errors.Wrap(err, errM)\\n\\t\\t}\\n\\t}\\n\\n\\t// =====================================================================\\n\\t// Переменная часть кода\\n\\t// =====================================================================\\n\\t// считаем данные обновленные данные - в БД могли быть тригера, которые поменяли данные\\n\\t// если запустили без проверок, то можно не возвращать результат - он будет запрошен уровнем выше\\n\\tif isValidate {\\n\\t\\tv, err := p.GetEmpTx(r.Empno, tx)\\n\\t\\tif err != nil {\\n\\t\\t\\terrM := fmt.Sprintf(\\\"[ERROR] %v - ERROR - p.GetEmpTx(r.Empno, tx), args = '%v'\\\", cnt, r.Empno)\\n\\t\\t\\tlog.Printf(errM)\\n\\t\\t\\treturn nil, errors.WithMessage(err, errM)\\n\\t\\t}\\n\\t\\tr = v\\n\\t}\\n\\t// =====================================================================\\n\\n\\t//mylog.PrintfDebug(\\\"[DEBUG] %v - SUCCESS\\\", cnt)\\n\\n\\treturn r, nil\\n}\",\n \"func (gt GameType) CreateTx(action string, message json.RawMessage) (*types.Transaction, error) {\\n\\ttlog.Debug(\\\"Game.CreateTx\\\", \\\"action\\\", action)\\n\\tif action == ActionCreateGame {\\n\\t\\tvar param GamePreCreateTx\\n\\t\\terr := json.Unmarshal(message, &param)\\n\\t\\tif err != nil {\\n\\t\\t\\ttlog.Error(\\\"CreateTx\\\", \\\"Error\\\", err)\\n\\t\\t\\treturn nil, types.ErrInvalidParam\\n\\t\\t}\\n\\n\\t\\treturn CreateRawGamePreCreateTx(&param)\\n\\t} else if action == ActionMatchGame {\\n\\t\\tvar param GamePreMatchTx\\n\\t\\terr := json.Unmarshal(message, &param)\\n\\t\\tif err != nil {\\n\\t\\t\\ttlog.Error(\\\"CreateTx\\\", \\\"Error\\\", err)\\n\\t\\t\\treturn nil, types.ErrInvalidParam\\n\\t\\t}\\n\\n\\t\\treturn CreateRawGamePreMatchTx(&param)\\n\\t} else if action == ActionCancelGame {\\n\\t\\tvar param GamePreCancelTx\\n\\t\\terr := json.Unmarshal(message, &param)\\n\\t\\tif err != nil {\\n\\t\\t\\ttlog.Error(\\\"CreateTx\\\", \\\"Error\\\", err)\\n\\t\\t\\treturn nil, types.ErrInvalidParam\\n\\t\\t}\\n\\n\\t\\treturn CreateRawGamePreCancelTx(&param)\\n\\t} else if action == ActionCloseGame {\\n\\t\\tvar param GamePreCloseTx\\n\\t\\terr := json.Unmarshal(message, &param)\\n\\t\\tif err != nil {\\n\\t\\t\\ttlog.Error(\\\"CreateTx\\\", \\\"Error\\\", err)\\n\\t\\t\\treturn nil, types.ErrInvalidParam\\n\\t\\t}\\n\\n\\t\\treturn CreateRawGamePreCloseTx(&param)\\n\\t}\\n\\treturn nil, types.ErrNotSupport\\n}\",\n \"func Create(reqDto *ReqCreateDto, custDto *ReqCustomerDto) (statusCode int, code string, respDto *RespBase, err error) {\\r\\n\\treqDto.ServiceType = \\\"TMS_CREATE_ORDER\\\"\\r\\n\\tbizData, err := xml.Marshal(reqDto.BizData)\\r\\n\\tif err != nil {\\r\\n\\t\\tcode = E02\\r\\n\\t\\treturn\\r\\n\\t}\\r\\n\\tsignParam := string(bizData) + reqDto.PartnerKey\\r\\n\\treqDto.Sign, err = sign.GetMD5Hash(signParam, true)\\r\\n\\tif err != nil {\\r\\n\\t\\tcode = E02\\r\\n\\t\\treturn\\r\\n\\t}\\r\\n\\treqMap := make(map[string]string, 0)\\r\\n\\treqMap[\\\"serviceType\\\"] = reqDto.ServiceType\\r\\n\\treqMap[\\\"partnerID\\\"] = reqDto.PartnerID\\r\\n\\treqMap[\\\"bizData\\\"] = string(bizData)\\r\\n\\treqMap[\\\"sign\\\"] = reqDto.Sign\\r\\n\\r\\n\\tdata := base.JoinMapString(reqMap)\\r\\n\\r\\n\\treq := httpreq.New(http.MethodPost, custDto.Url, data, func(httpReq *httpreq.HttpReq) error {\\r\\n\\t\\thttpReq.ReqDataType = httpreq.FormType\\r\\n\\t\\thttpReq.RespDataType = httpreq.XmlType\\r\\n\\t\\treturn nil\\r\\n\\t})\\r\\n\\tstatusCode, err = req.Call(&respDto)\\r\\n\\tif err != nil {\\r\\n\\t\\tcode = E01\\r\\n\\t\\treturn\\r\\n\\t}\\r\\n\\tif statusCode != http.StatusOK {\\r\\n\\t\\tcode = E01\\r\\n\\t\\terr = fmt.Errorf(\\\"http status exp:200,act:%v\\\", statusCode)\\r\\n\\t\\treturn\\r\\n\\t}\\r\\n\\tif respDto.Result != true {\\r\\n\\t\\tcode = E03\\r\\n\\t\\terr = fmt.Errorf(\\\"%v-%v\\\", respDto.ErrorCode, respDto.ErrorDescription)\\r\\n\\t\\treturn\\r\\n\\t}\\r\\n\\tcode = SUC\\r\\n\\treturn\\r\\n}\",\n \"func (t *Procure2Pay) CreateItems(stub shim.ChaincodeStubInterface, args []string) pb.Response {\\n\\n\\tvar objitem item\\t\\n\\tvar err error\\n\\n\\tfmt.Println(\\\"Entering CreateItems\\\")\\n\\n\\tif (len(args) < 1) {\\n\\t\\tfmt.Println(\\\"Invalid number of args\\\")\\n\\t\\treturn shim.Error(err.Error())\\n\\t\\t//return nil, errors.New(\\\"Expected atleast one arguments for initiate Transaction\\\")\\n\\t}\\n\\n\\tfmt.Println(\\\"Args [0] is : %v\\\\n\\\",args[0])\\n\\t\\n\\t//unmarshal item data from UI to \\\"item\\\" struct\\n\\terr = json.Unmarshal([]byte(args[0]), &objitem)\\n\\tif err != nil {\\n\\t\\tfmt.Printf(\\\"Unable to unmarshal CreateItem input item: %s\\\\n\\\", err)\\n\\t\\treturn shim.Error(err.Error())\\n\\t\\t//return nil, nil\\n\\t}\\n\\n\\tfmt.Println(\\\"item object ItemID variable value is : %s\\\\n\\\",objitem.ItemID);\\n\\t\\n\\t\\t// Data insertion for Couch DB starts here \\n\\t\\ttransJSONasBytes, err := json.Marshal(objitem)\\n\\t\\terr = stub.PutState(objitem.ItemID, transJSONasBytes)\\n\\t\\t// Data insertion for Couch DB ends here \\n\\n\\t\\tfmt.Println(\\\"Create items Successfully Done\\\")\\t\\n\\t\\n\\t\\tif err != nil {\\n\\t\\t\\t\\tfmt.Printf(\\\"\\\\nUnable to make transevent inputs : %v \\\", err)\\n\\t\\t\\t\\treturn shim.Error(err.Error())\\n\\t\\t\\t\\t//return nil,nil\\n\\t\\t\\t}\\n\\treturn shim.Success(nil)\\n\\t//return nil, nil\\n}\",\n \"func Transaction(next http.Handler) http.Handler {\\n\\tfn := func(w http.ResponseWriter, r *http.Request) {\\n\\t\\tt, ctx := orm.NewTransaction(r.Context())\\n\\t\\tdefer func() {\\n\\t\\t\\tif rec := recover(); rec != nil {\\n\\t\\t\\t\\tt.Rollback()\\n\\t\\t\\t\\t// Panic to let recoverer handle 500\\n\\t\\t\\t\\tpanic(rec)\\n\\t\\t\\t} else {\\n\\t\\t\\t\\terr := t.Commit()\\n\\t\\t\\t\\tif err != nil {\\n\\t\\t\\t\\t\\tpanic(err)\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\t\\t}()\\n\\t\\tnext.ServeHTTP(w, r.WithContext(ctx))\\n\\t}\\n\\treturn http.HandlerFunc(fn)\\n}\",\n \"func (ingest *Ingestion) Transaction(\\n\\tid int64,\\n\\ttx *core.Transaction,\\n\\tfee *core.TransactionFee,\\n) error {\\n\\n\\tsql := ingest.transactionInsertBuilder(id, tx, fee)\\n\\t_, err := ingest.DB.Exec(sql)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func MakeResTransaction(creator utils.Addr,key utils.Key,\\n inps []TrInput,out TrOutput,\\n jobblock,jobtrans,hashsol string,\\n evaluation float64,isMin bool)*ResTransaction{\\n tr := new(ResTransaction)\\n tr.Timestamp = time.Now()\\n tr.Output = out\\n tr.Inputs = inps\\n tr.Creator = creator\\n tr.JobBlock = jobblock\\n tr.JobTrans = jobtrans\\n tr.HashSol = hashsol\\n tr.Evaluation = evaluation\\n tr.IsMin = isMin\\n tr.Hash = tr.GetHash()\\n tr.Signature = fmt.Sprintf(\\\"%x\\\",utils.GetSignatureFromHash(tr.Hash,key))\\n return tr\\n}\",\n \"func (c CreateTransactionService) Execute(newTransaction entities.Transaction) (*entities.Transaction, error) {\\n\\ttransactType := newTransaction.Type\\n\\tif transactType != \\\"income\\\" && transactType != \\\"outcome\\\" {\\n\\t\\treturn nil, errors.New(\\\"Cannot create transaction type different fom income or outcome\\\")\\n\\t}\\n\\n\\tif transactType == \\\"outcome\\\" && !c.checksValidBalance(newTransaction) {\\n\\t\\treturn nil, errors.New(\\\"Cannot create transaction with invalid balance\\\")\\n\\t}\\n\\n\\tcreatedTransaction, err := c.Repo.Create(&newTransaction)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\treturn createdTransaction, nil\\n}\",\n \"func TransactionUpdate(c *gin.Context) {\\n\\tvar t models.Transaction\\n\\tvar newT models.Transaction\\n\\n\\tif database.DBCon.First(&t, c.Param(\\\"id\\\")).RecordNotFound() {\\n\\t\\tc.AbortWithError(http.StatusNotFound, appError.RecordNotFound).\\n\\t\\t\\tSetMeta(appError.RecordNotFound)\\n\\t\\treturn\\n\\t}\\n\\n\\t// Ensure current user is creator of transaction\\n\\tif t.CreatorID != c.Keys[\\\"CurrentUserID\\\"].(uint) {\\n\\t\\tc.AbortWithError(appError.InsufficientPermission.Status, appError.InsufficientPermission).\\n\\t\\t\\tSetMeta(appError.InsufficientPermission)\\n\\t\\treturn\\n\\t}\\n\\n\\tbuffer, err := ioutil.ReadAll(c.Request.Body)\\n\\n\\tif err != nil {\\n\\t\\tc.AbortWithError(http.StatusNotAcceptable, err)\\n\\t}\\n\\n\\terr2 := jsonapi.Unmarshal(buffer, &newT)\\n\\n\\tif err2 != nil {\\n\\t\\tc.AbortWithError(http.StatusInternalServerError, err).\\n\\t\\t\\tSetMeta(appError.JSONParseFailure)\\n\\t\\treturn\\n\\t}\\n\\n\\tt.Type = newT.Type\\n\\tt.Amount = newT.Amount\\n\\tt.Memo = newT.Memo\\n\\tt.RecipientID = newT.RecipientID\\n\\tt.SenderID = newT.SenderID\\n\\n\\t// Validate our new transaction\\n\\tisValid, errApp := t.Validate()\\n\\n\\tif isValid == false {\\n\\t\\tc.AbortWithError(errApp.Status, errApp).\\n\\t\\t\\tSetMeta(errApp)\\n\\t\\treturn\\n\\t}\\n\\n\\tdatabase.DBCon.Save(&t)\\n\\n\\tdatabase.DBCon.First(&t.Recipient, t.RecipientID)\\n\\tdatabase.DBCon.First(&t.Sender, t.SenderID)\\n\\tdatabase.DBCon.First(&t.Creator, t.CreatorID)\\n\\n\\tdata, err := jsonapi.Marshal(&t)\\n\\n\\tif err != nil {\\n\\t\\tc.AbortWithError(http.StatusInternalServerError, err).\\n\\t\\t\\tSetMeta(appError.JSONParseFailure)\\n\\t\\treturn\\n\\t}\\n\\n\\tc.Data(http.StatusOK, \\\"application/vnd.api+json\\\", data)\\n}\",\n \"func (t *Transaction) createID() (string, error) {\\n\\n\\t// Strip ID of txn\\n\\ttn := &Transaction{\\n\\t\\tID: nil,\\n\\t\\tVersion: t.Version,\\n\\t\\tInputs: t.Inputs,\\n\\t\\tOutputs: t.Outputs,\\n\\t\\tOperation: t.Operation,\\n\\t\\tAsset: t.Asset,\\n\\t\\tMetadata: t.Metadata,\\n\\t}\\n\\t// Serialize transaction - encoding/json follows RFC7159 and BDB marshalling\\n\\tdbytes, err := tn.JSON()\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", err\\n\\t}\\n\\n\\t// Return hash of serialized txn object\\n\\th := sha3.Sum256(dbytes)\\n\\treturn hex.EncodeToString(h[:]), nil\\n}\",\n \"func (_Content *ContentTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _Content.contract.Transact(opts, \\\"runCreate\\\")\\n}\",\n \"func (tm *TransactionManager) CreateContract(sk string, data []byte, gasPrice uint64, nonce uint64, gasLimit uint64) (string, error) {\\n\\treturn tm.SendTx(sk, \\\"\\\", nil, data, gasPrice, nonce, gasLimit)\\n}\",\n \"func (_LvRecording *LvRecordingTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _LvRecording.contract.Transact(opts, \\\"runCreate\\\")\\n}\",\n \"func (bi *Blockchainidentifier) SaveTransaction(w http.ResponseWriter, r *http.Request) {\\n\\tlog.Println(r.RemoteAddr + \\\" POST /transactions/new\\\")\\n\\n\\tvar t transaction\\n\\n\\terr := json.NewDecoder(r.Body).Decode(&t)\\n\\tif err != nil {\\n\\t\\thttp.Error(w, \\\"ERROR: \\\"+err.Error(), 500)\\n\\t\\treturn\\n\\t}\\n\\n\\tif bi.isValid(t) == false {\\n\\t\\thttp.Error(w, \\\"ERROR: Missing values in transaction\\\", 400)\\n\\t\\treturn\\n\\t}\\n\\n\\tt.Timestamp = time.Now().UTC().Format(\\\"2006-01-02 15:04:05\\\")\\n\\n\\tnewblockindex := bi.newTransaction(t)\\n\\n\\tresponseMessage := map[string]string{\\n\\t\\t\\\"message\\\": \\\"Transaction will be added in Block#\\\" + strconv.Itoa(newblockindex),\\n\\t}\\n\\n\\tw.Header().Set(\\\"Content-Type\\\", \\\"application/json; charset=utf-8\\\")\\n\\tw.WriteHeader(http.StatusCreated)\\n\\tjson.NewEncoder(w).Encode(responseMessage)\\n}\",\n \"func NewTransaction(p *requestParams) {\\n\\tw, r, c, u := p.w, p.r, p.c, p.u\\n\\n\\td := json.NewDecoder(r.Body)\\n\\tvar request TransactionRequest\\n\\tif err := d.Decode(&request); err != nil {\\n\\t\\thttp.Error(w, err.Error(), http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\n\\tif len(request.Amounts) != len(request.Accounts) {\\n\\t\\thttp.Error(w, \\\"Amounts and accounts of different lengths\\\", http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\n\\tdate, err := time.Parse(dateStringFormat, request.Date)\\n\\tif err != nil {\\n\\t\\thttp.Error(w, err.Error(), http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\n\\tuserKey := userKey(c, u)\\n\\ttransactionId := uuid.NewRandom().String()\\n\\taccountKeys := make([]*datastore.Key, len(request.Accounts))\\n\\tsplitKeys := make([]*datastore.Key, len(request.Accounts))\\n\\tsplits := make([]*transaction.Split, len(request.Accounts))\\n\\n\\tfor i := range request.Accounts {\\n\\t\\taccountKeys[i] = datastore.NewKey(c, \\\"Account\\\", \\\"\\\", request.Accounts[i], userKey)\\n\\t\\tsplitKeys[i] = datastore.NewKey(c, \\\"Split\\\", transactionId, 0, accountKeys[i])\\n\\t\\tsplits[i] = &transaction.Split{\\n\\t\\t\\tAmount: request.Amounts[i],\\n\\t\\t\\tAccount: request.Accounts[i],\\n\\t\\t\\tMemo: request.Memo,\\n\\t\\t\\tDate: date,\\n\\t\\t}\\n\\t}\\n\\n\\tx := transaction.NewTransaction()\\n\\tx.AddSplits(splits)\\n\\n\\tif err := x.ValidateAmount(); err != nil {\\n\\t\\thttp.Error(w, err.Error(), http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\n\\terr = datastore.RunInTransaction(c, func(c appengine.Context) error {\\n\\t\\taccounts := make([]transaction.Account, len(accountKeys))\\n\\t\\tif err := datastore.GetMulti(c, accountKeys, accounts); err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\tfor i := range accounts {\\n\\t\\t\\tx.AddAccount(&accounts[i], accountKeys[i].IntID())\\n\\t\\t}\\n\\n\\t\\tif err := x.Commit(); err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\tputStatus := make(chan error)\\n\\n\\t\\tgo func() {\\n\\t\\t\\t_, err := datastore.PutMulti(c, accountKeys, accounts)\\n\\t\\t\\tputStatus <- err\\n\\t\\t}()\\n\\t\\tgo func() {\\n\\t\\t\\t_, err := datastore.PutMulti(c, splitKeys, splits)\\n\\t\\t\\tputStatus <- err\\n\\t\\t}()\\n\\n\\t\\terr := <-putStatus\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\treturn <-putStatus\\n\\t}, nil)\\n\\tif err != nil {\\n\\t\\t// TODO(cjc25): This might not be a 400: if e.g. datastore failed it should\\n\\t\\t// be a 500. Interpret err and return the right thing.\\n\\t\\thttp.Error(w, err.Error(), http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n}\",\n \"func (s *dbTestSuite) CreateTestTx() (*sql.Tx, func()) {\\n\\ttx, err := s.db.DB().Begin()\\n\\ts.Nil(err)\\n\\tcommit := func() {\\n\\t\\ts.Nil(tx.Commit())\\n\\t}\\n\\n\\treturn tx, commit\\n}\",\n \"func (p *KVServer) NewTransaction(req bool, resp *NewTransactionResp) error {\\n\\ttxID := nextTransactionID\\n\\tnextTransactionID = nextTransactionID + 10\\n\\tmutex.Lock()\\n\\tkvStore := keyValueStore\\n\\tmutex.Unlock()\\n\\t*resp = NewTransactionResp{txID, kvStore}\\n\\treturn nil\\n}\",\n \"func CreateServiceSubscriptionTransaction(db *sql.DB, params url.Values) (data []byte, err error) {\\n\\tvar (\\n\\t\\tresult sql.Result\\n\\t\\tresponse = SQLResponse{}\\n\\t\\ttx *sql.Tx\\n\\t\\tserviceTypeStr string\\n\\t\\tserviceType int\\n\\t\\tserviceName string\\n\\t\\tdescription string\\n\\t\\torganizationName string\\n\\t\\tactiveUntil string\\n\\t\\tamountPaidStr string\\n\\t\\tamountPaid int\\n\\t\\tprocessedTimeStamp string\\n\\t)\\n\\n\\tif tx, err = db.Begin(); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif amountPaidStr, err = common.GetRequiredParam(params, \\\"amountPaid\\\"); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif amountPaid, err = strconv.Atoi(amountPaidStr); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif processedTimeStamp, err = common.GetRequiredParam(params, \\\"processedTimestamp\\\"); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif serviceTypeStr, err = common.GetRequiredParam(params, \\\"serviceType\\\"); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif serviceType, err = strconv.Atoi(serviceTypeStr); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif serviceName, err = common.GetRequiredParam(params, \\\"serviceName\\\"); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif description, err = common.GetRequiredParam(params, \\\"description\\\"); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif organizationName, err = common.GetRequiredParam(params, \\\"organizationName\\\"); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif activeUntil, err = common.GetRequiredParam(params, \\\"activeUntil\\\"); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif result, err = tx.Exec(\\\"INSERT INTO ServiceSubscriptionTransaction (amountPaid, processedTimestamp,type,serviceName,description, organizationName, activeUntil) VALUES(?,?,?,?,?,?,?);\\\",\\n\\t\\tamountPaid, processedTimeStamp, serviceType, serviceName, description, organizationName, activeUntil); err != nil {\\n\\t\\ttx.Rollback()\\n\\t\\treturn\\n\\t}\\n\\tif err = tx.Commit(); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tif response.AffectedRows, err = result.RowsAffected(); err != nil {\\n\\t\\treturn\\n\\t}\\n\\tdata, err = json.Marshal(response)\\n\\treturn\\n}\",\n \"func (_LvRecordableStream *LvRecordableStreamTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\\n\\treturn _LvRecordableStream.contract.Transact(opts, \\\"runCreate\\\")\\n}\",\n \"func (m *MySQLModelAudit) CreateTx(tx *sql.Tx, modelAudit *modelaudit.Model) error {\\n\\tstmt, err := tx.Prepare(mysqlCreateModelAudit)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tdefer stmt.Close()\\n\\n\\tnow := utils.Now()\\n\\treturn mysql.ExecAffectingOneRow(\\n\\t\\tstmt,\\n\\t\\tmodelAudit.PrevModel.ModelID,\\n\\t\\tmodelAudit.PrevModel.Mjml,\\n\\t\\tmodelAudit.PrevModel.Html,\\n\\t\\tmodelAudit.PrevModel.Variables,\\n\\t\\tmodelAudit.Model.Mjml,\\n\\t\\tmodelAudit.Model.Html,\\n\\t\\tmodelAudit.Model.Variables,\\n\\t\\tmodelAudit.SetUserID,\\n\\t\\tnow[\\\"date\\\"], now[\\\"time\\\"], now[\\\"unix\\\"],\\n\\t)\\n}\",\n \"func (w *WalletChaincode) create(stub shim.ChaincodeStubInterface, args []string) pb.Response {\\n\\tif len(args) < 6 {\\n\\t\\treturn shim.Error(\\\"Incorrect number of arguments. Expecting at least 2\\\")\\n\\t}\\n\\taddress := args[2]\\n\\taccountKey := buildAccountKey(address)\\n\\tif err := stub.PutState(accountKey, []byte(address)); err != nil {\\n\\t\\treturn shim.Error(fmt.Sprintf(\\\"Error putting data for key [%s]: %s\\\", accountKey, err))\\n\\t}\\n\\tfmt.Println(\\\"create an account: \\\", accountKey)\\n\\n\\twalletKey := buildWalletKey(args[0], args[1], address)\\n\\tif err := stub.PutState(walletKey, []byte(args[5])); err != nil {\\n\\t\\treturn shim.Error(fmt.Sprintf(\\\"Error putting data for key [%s]: %s\\\", walletKey, err))\\n\\t}\\n\\tfmt.Println(\\\"create a wallet: \\\", walletKey)\\n\\n\\t// seqBytes, err := stub.GetState(SEQUENCE)\\n\\t// if err != nil {\\n\\t// \\treturn shim.Error(\\\"Failed to get state\\\")\\n\\t// }\\n\\t// if seqBytes == nil {\\n\\t// \\treturn shim.Error(\\\"Entity not found\\\")\\n\\t// }\\n\\t// seq, _ := strconv.ParseInt(string(seqBytes), 10, 64)\\n\\tseq := atomic.AddUint64(&w.Sequence, 1)\\n\\tsequenceKey := buildSequenceKey(seq)\\n\\tjsonTx := \\\"{\\\\\\\"sequence\\\\\\\":\\\\\\\"\\\" + strconv.FormatUint(seq, 10) + \\\"\\\\\\\",\\\\\\\"txid\\\\\\\":\\\\\\\"\\\" + string(stub.GetTxID()) + \\\"\\\\\\\"}\\\"\\n\\tif err := stub.PutState(sequenceKey, []byte(jsonTx)); err != nil {\\n\\t\\treturn shim.Error(fmt.Sprintf(\\\"Error putting data for key [%s]: %s\\\", walletKey, err))\\n\\t}\\n\\n\\tfmt.Println(\\\"create success: \\\", stub.GetTxID())\\n\\treturn shim.Success([]byte(fmt.Sprintf(\\\"{\\\\\\\"wallet\\\\\\\":\\\\\\\"%s\\\\\\\", \\\\\\\"txid\\\\\\\":\\\\\\\"%s\\\\\\\"}\\\", walletKey, stub.GetTxID())))\\n}\",\n \"func (tx *Transaction) Create(entity interface{}) error {\\n\\treturn createFunc(tx.Query, entity)\\n}\",\n \"func Create(t *contract.SQLTable) error {\\n\\tsqlstr := t.GetCreateSQL()\\n\\tif t.Database == \\\"\\\" {\\n\\t\\tt.Database = \\\"default\\\"\\n\\t}\\n\\tdb, err := dblist.Get(t.Database)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\t_, err = db.Exec(sqlstr)\\n\\treturn err\\n}\",\n \"func (s *PaymentStorage) Create(\\n\\tctx context.Context,\\n\\tid aggregate.ID,\\n\\tversion transaction.Version,\\n\\torganisationID transaction.OrganisationID,\\n\\tattributes interface{},\\n) error {\\n\\tlogger := log.FromContext(ctx)\\n\\n\\tquery := `INSERT INTO %[1]s (id, version, organisation_id, attributes) VALUES ($1, $2, $3, $4)`\\n\\tquery = fmt.Sprintf(query, s.table)\\n\\n\\tif logger != nil {\\n\\t\\tlogger.Debugf(\\\"exec in transaction sql %s, values %+v\\\", query, []interface{}{\\n\\t\\t\\tid,\\n\\t\\t\\tversion,\\n\\t\\t\\torganisationID,\\n\\t\\t\\tattributes,\\n\\t\\t})\\n\\t}\\n\\n\\treturn execInTransaction(s.db, func(tx *sqlx.Tx) error {\\n\\t\\t_, err := tx.ExecContext(ctx, query, id, version, organisationID, attributes)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\treturn nil\\n\\t})\\n}\",\n \"func (_TableManager *TableManagerTransactor) CreateTable(opts *bind.TransactOpts, path string, tableInfo TableInfo) (*types.Transaction, *types.Receipt, error) {\\n\\treturn _TableManager.contract.Transact(opts, \\\"createTable\\\", path, tableInfo)\\n}\",\n \"func (s *SmartContract) CreateContract(ctx contractapi.TransactionContextInterface, id string, sellerID string, consumerID string, created string, contract string, signature string) error {\\n\\tctc := Contract{\\n\\t\\tContext: \\\"http://wuldid.ddns.net\\\",\\n\\t\\tSellerID: sellerID,\\n\\t\\tConsumerID: consumerID,\\n\\t\\tCreated: created,\\n\\t\\tContract: contract,\\n\\t\\tSignature: signature,\\n\\t}\\n\\n\\texists, err := s.DidExists(ctx, id)\\n\\tif err != nil {\\n\\t\\tfmt.Errorf(\\\"Unexpected error!! : %q\\\", err)\\n\\t}\\n\\tif !exists {\\n\\t\\tctcJSON, _ := json.Marshal(ctc)\\n\\t\\treturn ctx.GetStub().PutState(id, ctcJSON)\\n\\t} else {\\n\\t\\treturn fmt.Errorf(\\\"Don't exsit did!\\\")\\n\\t}\\n\\n}\",\n \"func (client *Client) CreateTransactionType(transactionType string, customIndexes []CustomIndexStructure) (_ *Response, err error) {\\n\\tpath := \\\"/transaction-type\\\"\\n\\turi := fmt.Sprintf(\\\"%s%s\\\", client.apiBaseURL, path)\\n\\tvar params TransactionType\\n\\tparams.Version = \\\"1\\\"\\n\\tparams.Type = transactionType\\n\\tparams.CustomIndexes = customIndexes\\n\\n\\tb, err := json.Marshal(params)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tresp, err := client.httpClient.Post(uri, \\\"content/json\\\", bytes.NewBuffer(b))\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tdefer func() {\\n\\t\\t_ = resp.Body.Close()\\n\\t}()\\n\\tvar chainResp Response\\n\\trespBytes, err := ioutil.ReadAll(resp.Body)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tif err := json.Unmarshal(respBytes, &chainResp.Response); err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\tchainResp.Status = resp.StatusCode\\n\\tif 200 <= resp.StatusCode && resp.StatusCode < 300 {\\n\\t\\tchainResp.OK = true\\n\\t}\\n\\treturn &chainResp, err\\n}\",\n \"func (r Virtual_Guest) CreateArchiveTransaction(groupName *string, blockDevices []datatypes.Virtual_Guest_Block_Device, note *string) (resp datatypes.Provisioning_Version1_Transaction, err error) {\\n\\tparams := []interface{}{\\n\\t\\tgroupName,\\n\\t\\tblockDevices,\\n\\t\\tnote,\\n\\t}\\n\\terr = r.Session.DoRequest(\\\"SoftLayer_Virtual_Guest\\\", \\\"createArchiveTransaction\\\", params, &r.Options, &resp)\\n\\treturn\\n}\",\n \"func (t *TxCreate) CreatePaymentTx() (string, string, error) {\\n\\tsender := t.paymentSender\\n\\treceiver := account.AccountTypeAnonymous\\n\\ttargetAction := action.ActionTypePayment\\n\\tt.logger.Debug(\\\"account\\\",\\n\\t\\tzap.String(\\\"sender\\\", sender.String()),\\n\\t\\tzap.String(\\\"receiver\\\", receiver.String()),\\n\\t)\\n\\n\\t// get payment data from payment_request\\n\\tuserPayments, totalAmount, paymentRequestIds, err := t.createUserPayment()\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", \\\"\\\", err\\n\\t}\\n\\tif len(userPayments) == 0 {\\n\\t\\tt.logger.Debug(\\\"no data in userPayments\\\")\\n\\t\\t// no data\\n\\t\\treturn \\\"\\\", \\\"\\\", nil\\n\\t}\\n\\n\\t// check sender's total balance\\n\\t// GetOneUnAllocated\\n\\taddrItem, err := t.addrRepo.GetOneUnAllocated(sender)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", \\\"\\\", errors.Wrap(err, \\\"fail to call addrRepo.GetAll(account.AccountTypeClient)\\\")\\n\\t}\\n\\tsenderBalance, err := t.xrp.GetBalance(addrItem.WalletAddress)\\n\\tif err != nil {\\n\\t\\treturn \\\"\\\", \\\"\\\", errors.Wrap(err, \\\"fail to call xrp.GetAccountInfo()\\\")\\n\\t}\\n\\n\\tif senderBalance <= totalAmount {\\n\\t\\treturn \\\"\\\", \\\"\\\", errors.New(\\\"sender balance is insufficient to send\\\")\\n\\t}\\n\\n\\t// create raw transaction each address\\n\\tserializedTxs := make([]string, 0, len(userPayments))\\n\\ttxDetailItems := make([]*models.XRPDetailTX, 0, len(userPayments))\\n\\tvar sequence uint64\\n\\tfor _, userPayment := range userPayments {\\n\\t\\t// call CreateRawTransaction\\n\\t\\tinstructions := &pb.Instructions{\\n\\t\\t\\tMaxLedgerVersionOffset: xrp.MaxLedgerVersionOffset,\\n\\t\\t}\\n\\t\\tif sequence != 0 {\\n\\t\\t\\tinstructions.Sequence = sequence\\n\\t\\t}\\n\\t\\ttxJSON, rawTxString, err := t.xrp.CreateRawTransaction(addrItem.WalletAddress, userPayment.receiverAddr, userPayment.floatAmount, instructions)\\n\\t\\tif err != nil {\\n\\t\\t\\tt.logger.Warn(\\\"fail to call xrp.CreateRawTransaction()\\\", zap.Error(err))\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\t\\tt.logger.Debug(\\\"txJSON\\\", zap.Any(\\\"txJSON\\\", txJSON))\\n\\t\\tgrok.Value(txJSON)\\n\\n\\t\\t// sequence for next rawTransaction\\n\\t\\tsequence = txJSON.Sequence + 1\\n\\n\\t\\t// generate UUID to trace transaction because unsignedTx is not unique\\n\\t\\tuid := uuid.NewV4().String()\\n\\n\\t\\tserializedTxs = append(serializedTxs, fmt.Sprintf(\\\"%s,%s\\\", uid, rawTxString))\\n\\n\\t\\t// create insert data for eth_detail_tx\\n\\t\\ttxDetailItem := &models.XRPDetailTX{\\n\\t\\t\\tUUID: uid,\\n\\t\\t\\tCurrentTXType: tx.TxTypeUnsigned.Int8(),\\n\\t\\t\\tSenderAccount: sender.String(),\\n\\t\\t\\tSenderAddress: addrItem.WalletAddress,\\n\\t\\t\\tReceiverAccount: receiver.String(),\\n\\t\\t\\tReceiverAddress: userPayment.receiverAddr,\\n\\t\\t\\tAmount: txJSON.Amount,\\n\\t\\t\\tXRPTXType: txJSON.TransactionType,\\n\\t\\t\\tFee: txJSON.Fee,\\n\\t\\t\\tFlags: txJSON.Flags,\\n\\t\\t\\tLastLedgerSequence: txJSON.LastLedgerSequence,\\n\\t\\t\\tSequence: txJSON.Sequence,\\n\\t\\t\\t// SigningPubkey: txJSON.SigningPubKey,\\n\\t\\t\\t// TXNSignature: txJSON.TxnSignature,\\n\\t\\t\\t// Hash: txJSON.Hash,\\n\\t\\t}\\n\\t\\ttxDetailItems = append(txDetailItems, txDetailItem)\\n\\t}\\n\\n\\treturn t.afterTxCreation(targetAction, sender, serializedTxs, txDetailItems, paymentRequestIds)\\n}\",\n \"func (allocateIncome *AllocateIncome) Create() (err error) {\\n\\t//插入数据\\n\\tstmt, err := Db.Prepare(\\\"INSERT INTO allocateIncome(UDID, pool_id,mining_id,income,fee,date,status) values(?,?,?,?,?,?,?)\\\")\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\tres, err := stmt.Exec(allocateIncome.UDID, allocateIncome.PoolID, allocateIncome.MiningId, allocateIncome.Income, allocateIncome.Fee, time.Now().Unix(), allocateIncome.Status)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\tid, err := res.LastInsertId()\\n\\tlog.Info(id)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\treturn\\n}\",\n \"func (w *Wallet) txCreator() {\\n\\tquit := w.quitChan()\\nout:\\n\\tfor {\\n\\t\\tselect {\\n\\t\\tcase txr := <-w.createTxRequests:\\n\\t\\t\\theldUnlock, err := w.holdUnlock()\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\ttxr.resp <- createTxResponse{nil, err}\\n\\t\\t\\t\\tcontinue\\n\\t\\t\\t}\\n\\t\\t\\ttx, err := w.txToOutputs(txr.req)\\n\\t\\t\\theldUnlock.release()\\n\\t\\t\\ttxr.resp <- createTxResponse{tx, err}\\n\\t\\tcase <-quit:\\n\\t\\t\\tbreak out\\n\\t\\t}\\n\\t}\\n\\tw.wg.Done()\\n}\",\n \"func (w *Wallet) CreateSimpleTx(r CreateTxReq) (*txauthor.AuthoredTx, er.R) {\\n\\treq := createTxRequest{\\n\\t\\treq: r,\\n\\t\\tresp: make(chan createTxResponse),\\n\\t}\\n\\tw.createTxRequests <- req\\n\\tresp := <-req.resp\\n\\treturn resp.tx, resp.err\\n}\",\n \"func (db *Database) NewTransaction() *Transaction{\\r\\n\\treturn &Transaction{db,make(map[string]*record)}\\r\\n}\",\n \"func (p *psqlInvoiceItem) CreateTx(tx *sql.Tx, headerID uint, ms invoiceItem.Models) error {\\n\\tstmt, err := tx.Prepare(psqlCreateInvoiceItem)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tdefer stmt.Close()\\n\\tfor _, item := range ms {\\n\\t\\terr = stmt.QueryRow(headerID, item.ProductID).Scan(&item.ID, &item.CreatedAt)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n\\t// return stmt.QueryRow(m.).Scan(&m.ID, &m.CreatedAt)\\n}\",\n \"func (tr *Transactor) CreateTransactionArweave(ctx context.Context, tags []Tag, w WalletSigner, amount string, data []byte, target string) (*Transaction, error) {\\n\\tlastTx, err := tr.Client.TxAnchor(ctx)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tprice, err := tr.Client.GetReward(ctx, data)\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\t// Non encoded transaction fields\\n\\ttx := NewTransactionArweave(\\n\\t\\ttags,\\n\\t\\tlastTx,\\n\\t\\tw.PubKeyModulus(),\\n\\t\\tamount,\\n\\t\\ttarget,\\n\\t\\tdata,\\n\\t\\tprice,\\n\\t)\\n\\n\\treturn tx, nil\\n}\",\n \"func create(\\n\\ttrx storage.Transaction,\\n\\ttxId merkle.Digest,\\n\\townerData OwnerData,\\n\\towner *account.Account,\\n) {\\n\\t// increment the count for owner\\n\\tnKey := owner.Bytes()\\n\\tcount := trx.Get(storage.Pool.OwnerNextCount, nKey)\\n\\tif nil == count {\\n\\t\\tcount = []byte{0, 0, 0, 0, 0, 0, 0, 0}\\n\\t} else if uint64ByteSize != len(count) {\\n\\t\\tlogger.Panic(\\\"OwnerNextCount database corrupt\\\")\\n\\t}\\n\\tnewCount := make([]byte, uint64ByteSize)\\n\\tbinary.BigEndian.PutUint64(newCount, binary.BigEndian.Uint64(count)+1)\\n\\ttrx.Put(storage.Pool.OwnerNextCount, nKey, newCount, []byte{})\\n\\n\\t// write to the owner list\\n\\toKey := append(owner.Bytes(), count...)\\n\\ttrx.Put(storage.Pool.OwnerList, oKey, txId[:], []byte{})\\n\\n\\t// write new index record\\n\\tdKey := append(owner.Bytes(), txId[:]...)\\n\\ttrx.Put(storage.Pool.OwnerTxIndex, dKey, count, []byte{})\\n\\n\\t// save owner data record\\n\\ttrx.Put(storage.Pool.OwnerData, txId[:], ownerData.Pack(), []byte{})\\n}\",\n \"func CreateTestContractCreationTransaction(\\n\\tsigner hmytypes.Signer, shard uint32, nonce, gasLimit uint64, gasPrice, amount *big.Int, data []byte,\\n) (*hmytypes.Transaction, error) {\\n\\tfromKey, err := crypto.GenerateKey()\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\ttx := hmytypes.NewContractCreation(nonce, shard, amount, gasLimit, gasPrice, data)\\n\\treturn hmytypes.SignTx(tx, signer, fromKey)\\n}\",\n \"func Create(responseWriter http.ResponseWriter, request *http.Request) {\\n\\tfmt.Println(\\\"[ CreateOrder ]\\\")\\n\\tbody, _ := json.Marshal(request.Body)\\n\\tfmt.Println(\\\"[ CreateOrder ] Body=\\\" + string(body))\\n\\t//params := mux.Vars(request)\\n\\tvar orderEntity OrderEntity\\n\\t_ = json.NewDecoder(request.Body).Decode(&orderEntity)\\n\\n\\tvar result OrderEntity = Insert(orderEntity)\\n\\n\\tWriteMessages(result, Topic.TOPIC_SUCCESS)\\n\\n\\tjson.NewEncoder(responseWriter).Encode(result)\\n}\",\n \"func (s *Service) CreateClientTx(tx *gorm.DB, clientID, secret, redirectURI string) (*models.OauthClient, error) {\\n\\treturn s.createClientCommon(tx, clientID, secret, redirectURI)\\n}\",\n \"func (dt DefaultTransactor) Create(ctx context.Context, m *movie.Movie) error {\\n\\ttx, err := dt.datastorer.BeginTx(ctx)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t// Prepare the sql statement using bind variables\\n\\tstmt, err := tx.PrepareContext(ctx, `\\n\\tselect o_create_timestamp,\\n\\t\\t o_update_timestamp\\n\\t from demo.create_movie (\\n\\t\\tp_id => $1,\\n\\t\\tp_extl_id => $2,\\n\\t\\tp_title => $3,\\n\\t\\tp_rated => $4,\\n\\t\\tp_released => $5,\\n\\t\\tp_run_time => $6,\\n\\t\\tp_director => $7,\\n\\t\\tp_writer => $8,\\n\\t\\tp_create_client_id => $9,\\n\\t\\tp_create_username => $10)`)\\n\\n\\tif err != nil {\\n\\t\\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\\n\\t}\\n\\tdefer stmt.Close()\\n\\n\\t// At some point, I will add a whole client flow, but for now\\n\\t// faking a client uuid....\\n\\tfakeClientID := uuid.New()\\n\\n\\t// Execute stored function that returns the create_date timestamp,\\n\\t// hence the use of QueryContext instead of Exec\\n\\trows, err := stmt.QueryContext(ctx,\\n\\t\\tm.ID, //$1\\n\\t\\tm.ExternalID, //$2\\n\\t\\tm.Title, //$3\\n\\t\\tm.Rated, //$4\\n\\t\\tm.Released, //$5\\n\\t\\tm.RunTime, //$6\\n\\t\\tm.Director, //$7\\n\\t\\tm.Writer, //$8\\n\\t\\tfakeClientID, //$9\\n\\t\\tm.CreateUser.Email) //$10\\n\\n\\tif err != nil {\\n\\t\\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\\n\\t}\\n\\tdefer rows.Close()\\n\\n\\t// Iterate through the returned record(s)\\n\\tfor rows.Next() {\\n\\t\\tif err := rows.Scan(&m.CreateTime, &m.UpdateTime); err != nil {\\n\\t\\t\\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\\n\\t\\t}\\n\\t}\\n\\n\\t// If any error was encountered while iterating through rows.Next above\\n\\t// it will be returned here\\n\\tif err := rows.Err(); err != nil {\\n\\t\\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\\n\\t}\\n\\n\\t// Commit the Transaction\\n\\tif err := dt.datastorer.CommitTx(tx); err != nil {\\n\\t\\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func createOrder(writer http.ResponseWriter, request *http.Request) {\\n\\ttransactionId := request.Header.Get(\\\"transactionId\\\")\\n\\torder := <-orderService.GetOrder(transactionId)\\n\\tif order.Id.Value == \\\"\\\" {\\n\\t\\tlog.Printf(\\\"Creating order for transactionId :%s......\\\", transactionId)\\n\\t\\torderHandler.CreateOrder(transactionId, command.CreateOrderCommand{Id: transactionId})\\n\\t}\\n\\trenderResponse(writer, []byte(transactionId))\\n}\",\n \"func createSignedTransaction(fromAddress common.Address, toAddress *common.Address, amount *big.Int, gasLimit uint64, data []byte) (signedTx *types.Transaction, err error) {\\n\\t// Create the transaction\\n\\ttx, err := createTransaction(fromAddress, toAddress, amount, gasLimit, data)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\t// Sign the transaction\\n\\tsignedTx, err = signTransaction(fromAddress, tx)\\n\\tif err != nil {\\n\\t\\terr = fmt.Errorf(\\\"Failed to sign transaction: %v\\\", err)\\n\\t\\treturn\\n\\t}\\n\\n\\t// Increment the nonce for the next transaction\\n\\tnextNonce(fromAddress)\\n\\n\\treturn\\n}\",\n \"func (r *Renter) createContractTransaction(terms modules.ContractTerms, merkleRoot crypto.Hash) (txn consensus.Transaction, id string, err error) {\\n\\t// Get the payout as set by the missed proofs, and the client fund as determined by the terms.\\n\\tvar payout consensus.Currency\\n\\tfor _, output := range terms.MissedProofOutputs {\\n\\t\\tpayout = payout.Add(output.Value)\\n\\t}\\n\\n\\t// Get the cost to the client as per the terms in the contract.\\n\\tsizeCurrency := consensus.NewCurrency64(terms.FileSize)\\n\\tdurationCurrency := consensus.NewCurrency64(uint64(terms.Duration))\\n\\tclientCost := terms.Price.Mul(sizeCurrency).Mul(durationCurrency)\\n\\n\\t// Fill out the contract.\\n\\tcontract := consensus.FileContract{\\n\\t\\tFileMerkleRoot: merkleRoot,\\n\\t\\tFileSize: terms.FileSize,\\n\\t\\tStart: terms.DurationStart + terms.Duration,\\n\\t\\tExpiration: terms.DurationStart + terms.Duration + terms.WindowSize,\\n\\t\\tPayout: payout,\\n\\t\\tValidProofOutputs: terms.ValidProofOutputs,\\n\\t\\tMissedProofOutputs: terms.MissedProofOutputs,\\n\\t}\\n\\n\\t// Create the transaction.\\n\\tid, err = r.wallet.RegisterTransaction(txn)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\t_, err = r.wallet.FundTransaction(id, clientCost)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\ttxn, _, err = r.wallet.AddFileContract(id, contract)\\n\\tif err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\treturn\\n}\",\n \"func TransactionHandler(w http.ResponseWriter, r *http.Request) {\\n\\taction := r.URL.Path[len(\\\"/api/transactions\\\"):]\\n\\n\\tlog.Println(\\\"Handling method\\\", r.Method, \\\"with action\\\", action)\\n\\n\\tswitch r.Method {\\n\\tcase \\\"POST\\\":\\n\\t\\tswitch action {\\n\\t\\tcase \\\"\\\": // Create new transaction\\n\\t\\t\\tbody, err := ioutil.ReadAll(r.Body)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorForm, err.Error())\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t\\tvar t TransactionResponse\\n\\n\\t\\t\\terr = json.Unmarshal(body, &t)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t\\tpass := []byte(t.Password)\\n\\n\\t\\t\\ttx, e := createAndBroadcastTx(t.Recipient, *big.NewInt(t.Amount), pass)\\n\\n\\t\\t\\tvar jsonResponse TransactionResponse\\n\\t\\t\\tif e == nil {\\n\\t\\t\\t\\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseOk, Hash: hex.EncodeToString(tx.Hash())}\\n\\t\\t\\t} else {\\n\\t\\t\\t\\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseFailed, ErrorText: e.Error()}\\n\\t\\t\\t}\\n\\n\\t\\t\\tres, err := json.Marshal(jsonResponse)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t\\tfmt.Fprintf(w, string(res))\\n\\t\\tdefault:\\n\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\\\"No action: \\\", r.Method, action))\\n\\t\\t}\\n\\tcase \\\"GET\\\":\\n\\t\\tswitch action {\\n\\t\\tcase \\\"\\\":\\n\\t\\t\\tvar txs []TransactionJson\\n\\t\\t\\tfor _, tx := range Config.DeserializedTxs {\\n\\t\\t\\t\\ttxs = append(txs, EncodeToFriendlyStruct(tx))\\n\\n\\t\\t\\t}\\n\\n\\t\\t\\tif len(txs) == 0 {\\n\\t\\t\\t\\tfmt.Fprintf(w, string(\\\"[]\\\"))\\n\\t\\t\\t} else {\\n\\n\\t\\t\\t\\tres, err := json.Marshal(txs)\\n\\t\\t\\t\\tif err != nil {\\n\\t\\t\\t\\t\\tfmt.Println(\\\"Nope\\\", err.Error())\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\tfmt.Fprintf(w, string(res))\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\n\\t\\t}\\n\\tdefault:\\n\\t\\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\\\"No action: \\\", r.Method, action))\\n\\n\\t}\\n}\",\n \"func (p *PsqlInvoiceItem) CreateTx(tx *sql.Tx, headerID uint, ms invoiceitem.Models) error {\\n\\tstmt, err := tx.Prepare(psqlCreateInvoiceItem)\\n\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\tdefer stmt.Close()\\n\\n\\tfor _, item := range ms {\\n\\t\\terr = stmt.QueryRow(headerID, item.ProductID).Scan(\\n\\t\\t\\t&item.ID,\\n\\t\\t\\t&item.CreatedAt,\\n\\t\\t)\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\n\\treturn err\\n\\n}\",\n \"func (a *ChainAdaptor) CreateUtxoSignedTransaction(req *proto.CreateUtxoSignedTransactionRequest) (*proto.CreateSignedTransactionReply, error) {\\n\\tr := bytes.NewReader(req.TxData)\\n\\tvar msgTx wire.MsgTx\\n\\terr := msgTx.Deserialize(r)\\n\\tif err != nil {\\n\\t\\tlog.Error(\\\"CreateSignedTransaction msgTx.Deserialize\\\", \\\"err\\\", err)\\n\\n\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\tif len(req.Signatures) != len(msgTx.TxIn) {\\n\\t\\tlog.Error(\\\"CreateSignedTransaction invalid params\\\", \\\"err\\\", \\\"Signature number mismatch Txin number\\\")\\n\\t\\terr = errors.New(\\\"Signature number != Txin number\\\")\\n\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\tif len(req.PublicKeys) != len(msgTx.TxIn) {\\n\\t\\tlog.Error(\\\"CreateSignedTransaction invalid params\\\", \\\"err\\\", \\\"Pubkey number mismatch Txin number\\\")\\n\\t\\terr = errors.New(\\\"Pubkey number != Txin number\\\")\\n\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\t// assemble signatures\\n\\tfor i, in := range msgTx.TxIn {\\n\\t\\tbtcecPub, err2 := btcec.ParsePubKey(req.PublicKeys[i], btcec.S256())\\n\\t\\tif err2 != nil {\\n\\t\\t\\tlog.Error(\\\"CreateSignedTransaction ParsePubKey\\\", \\\"err\\\", err2)\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err2.Error(),\\n\\t\\t\\t}, err2\\n\\t\\t}\\n\\n\\t\\tvar pkData []byte\\n\\t\\tif btcec.IsCompressedPubKey(req.PublicKeys[i]) {\\n\\t\\t\\tpkData = btcecPub.SerializeCompressed()\\n\\t\\t} else {\\n\\t\\t\\tpkData = btcecPub.SerializeUncompressed()\\n\\t\\t}\\n\\n\\t\\t// verify transaction\\n\\t\\tpreTx, err2 := a.getClient().GetRawTransactionVerbose(&in.PreviousOutPoint.Hash)\\n\\t\\tif err2 != nil {\\n\\t\\t\\tlog.Error(\\\"CreateSignedTransaction GetRawTransactionVerbose\\\", \\\"err\\\", err2)\\n\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err2.Error(),\\n\\t\\t\\t}, err2\\n\\t\\t}\\n\\n\\t\\tlog.Info(\\\"CreateSignedTransaction \\\", \\\"from address\\\", preTx.Vout[in.PreviousOutPoint.Index].ScriptPubKey.Addresses[0])\\n\\n\\t\\tfromAddress, err2 := btcutil.DecodeAddress(preTx.Vout[in.PreviousOutPoint.Index].ScriptPubKey.Addresses[0], a.getClient().GetNetwork())\\n\\t\\tif err2 != nil {\\n\\t\\t\\tlog.Error(\\\"CreateSignedTransaction DecodeAddress\\\", \\\"err\\\", err2)\\n\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err2.Error(),\\n\\t\\t\\t}, err2\\n\\t\\t}\\n\\n\\t\\tfromPkScript, err2 := txscript.PayToAddrScript(fromAddress)\\n\\t\\tif err2 != nil {\\n\\t\\t\\tlog.Error(\\\"CreateSignedTransaction PayToAddrScript\\\", \\\"err\\\", err2)\\n\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err2.Error(),\\n\\t\\t\\t}, err2\\n\\t\\t}\\n\\n\\t\\t// creat sigscript and verify\\n\\t\\tif len(req.Signatures[i]) < 64 {\\n\\t\\t\\terr2 = errors.New(\\\"Invalid signature length\\\")\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err2.Error(),\\n\\t\\t\\t}, err2\\n\\t\\t}\\n\\t\\tr := new(big.Int).SetBytes(req.Signatures[i][0:32])\\n\\t\\ts := new(big.Int).SetBytes(req.Signatures[i][32:64])\\n\\n\\t\\tbtcecSig := &btcec.Signature{\\n\\t\\t\\tR: r,\\n\\t\\t\\tS: s,\\n\\t\\t}\\n\\t\\tsig := append(btcecSig.Serialize(), byte(txscript.SigHashAll))\\n\\t\\tsigScript, err2 := txscript.NewScriptBuilder().AddData(sig).AddData(pkData).Script()\\n\\t\\tif err2 != nil {\\n\\t\\t\\tlog.Error(\\\"CreateSignedTransaction NewScriptBuilder\\\", \\\"err\\\", err2)\\n\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err2.Error(),\\n\\t\\t\\t}, err2\\n\\t\\t}\\n\\n\\t\\tmsgTx.TxIn[i].SignatureScript = sigScript\\n\\t\\tamount := btcToSatoshi(preTx.Vout[in.PreviousOutPoint.Index].Value).Int64()\\n\\t\\tlog.Info(\\\"CreateSignedTransaction \\\", \\\"amount\\\", preTx.Vout[in.PreviousOutPoint.Index].Value, \\\"int amount\\\", amount)\\n\\n\\t\\tvm, err2 := txscript.NewEngine(fromPkScript, &msgTx, i, txscript.StandardVerifyFlags, nil, nil, amount)\\n\\t\\tif err2 != nil {\\n\\t\\t\\tlog.Error(\\\"CreateSignedTransaction NewEngine\\\", \\\"err\\\", err2)\\n\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err2.Error(),\\n\\t\\t\\t}, err2\\n\\t\\t}\\n\\t\\tif err3 := vm.Execute(); err3 != nil {\\n\\t\\t\\tlog.Error(\\\"CreateSignedTransaction NewEngine Execute\\\", \\\"err\\\", err3)\\n\\n\\t\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\t\\tMsg: err3.Error(),\\n\\t\\t\\t}, err3\\n\\t\\t}\\n\\n\\t}\\n\\n\\t// serialize tx\\n\\tbuf := bytes.NewBuffer(make([]byte, 0, msgTx.SerializeSize()))\\n\\n\\terr = msgTx.Serialize(buf)\\n\\tif err != nil {\\n\\t\\tlog.Error(\\\"CreateSignedTransaction tx Serialize\\\", \\\"err\\\", err)\\n\\n\\t\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\t\\tCode: proto.ReturnCode_ERROR,\\n\\t\\t\\tMsg: err.Error(),\\n\\t\\t}, err\\n\\t}\\n\\n\\thash := msgTx.TxHash()\\n\\treturn &proto.CreateSignedTransactionReply{\\n\\t\\tCode: proto.ReturnCode_SUCCESS,\\n\\t\\tSignedTxData: buf.Bytes(),\\n\\t\\tHash: (&hash).CloneBytes(),\\n\\t}, nil\\n}\",\n \"func (cs *CustStoreSqlite) AddTransaction(\\n id string,\\n customerId string,\\n loadAmountCents int64,\\n time time.Time,\\n accepted bool,\\n) error {\\n // Check transaction id\\n isDuplicate, err := cs.isDuplicate(id, customerId)\\n if err != nil {\\n return err\\n }\\n if isDuplicate {\\n return custstore.DuplicateError\\n }\\n // Add to transactions\\n err = cs.addTransaction(id, customerId, loadAmountCents, time, accepted)\\n return err\\n}\",\n \"func CreateContractObject(args []string) (SalesContractObject, error) {\\n\\t// S001 LHTMO bosch\\n\\tvar err error\\n\\tvar myContract SalesContractObject\\n\\n\\t// Check there are 3 Arguments provided as per the the struct\\n\\tif len(args) != 8 {\\n\\t\\tfmt.Println(\\\"CreateContractObject(): Incorrect number of arguments. Expecting 8 \\\")\\n\\t\\treturn myContract, errors.New(\\\"CreateContractObject(): Incorrect number of arguments. Expecting 8 \\\")\\n\\t}\\n\\n\\t// Validate Serialno is an integer\\n\\n\\tstage, err := strconv.Atoi(args[1])\\n\\tif err != nil {\\n\\t\\tfmt.Println(\\\"CreateAssetObject(): Stage should be an integer create failed! \\\")\\n\\t\\treturn myContract, errors.New(\\\"CreateAssetbject(): Stage should be an integer create failed. \\\")\\n\\t}\\n\\tif stage != 0 {\\n\\t\\tfmt.Println(\\\"CreateAssetObject(): Stage should be set as open \\\")\\n\\t\\treturn myContract, errors.New(\\\"CreateAssetbject(): Stage should be set as open\\\")\\n\\t}\\n\\n\\tmyContract = SalesContractObject{args[0], STATE_OPEN, args[2], args[3], args[4], args[5], args[6], time.Now().Format(\\\"20060102150405\\\")}\\n\\n\\tfmt.Println(\\\"CreateContractObject(): Contract Object created: \\\", myContract.Contractid, myContract.Stage, myContract.Buyer, myContract.Transporter, myContract.Seller, myContract.AssetID, myContract.DocumentID, time.Now().Format(\\\"20060102150405\\\"))\\n\\treturn myContract, nil\\n}\",\n \"func (a *ApiDB) CreateContract(w http.ResponseWriter, r *http.Request) {\\n\\tw.Header().Add(\\\"Content-Type\\\", \\\"application/json\\\")\\n\\tp := MODELS.CREATE_UPDATE_CONTRACT_REQUEST{}\\n\\terr := json.NewDecoder(r.Body).Decode(&p)\\n\\tif err != nil {\\n\\t\\tio.WriteString(w, `{\\\"message\\\": \\\"wrong format!\\\"}`)\\n\\t\\treturn\\n\\t}\\n\\n\\tresult := BUSINESS.CreateContract(a.Db, p)\\n\\tif result {\\n\\t\\tio.WriteString(w, ` { \\\"status\\\": 200,\\n \\\"message\\\": \\\"Create contract success\\\",\\n \\\"data\\\": {\\n \\\"status\\\": 1\\n }\\n}\\n`)\\n\\t} else {\\n\\t\\tio.WriteString(w, `{ \\\"message\\\": \\\"Can’t create contract\\\"}`)\\n\\t}\\n}\",\n \"func (d Database) CreateRoot() error {\\n\\tdb, connectErr := d.Connect()\\n\\tdefer db.Close()\\n\\tutil.Handle(\\\"Error creating a DB connection: \\\", connectErr)\\n\\tvar count int\\n\\terr := db.QueryRow(\\\"SELECT COUNT(*) FROM \\\"+d.Cf.GetTableName()+\\\" ORDER BY $1 DESC\\\", \\\"tx_time\\\").Scan(&count)\\n\\tswitch {\\n\\tcase err != nil:\\n\\t\\tutil.Handle(\\\"There was a problem counting database transactions: \\\", err)\\n\\t\\treturn err\\n\\tdefault:\\n\\t\\t// fmt.Printf(\\\"Found %v transactions in the db.\\\", count)\\n\\t\\tif count == 0 {\\n\\t\\t\\ttxTime := \\\"1603203489229912200\\\"\\n\\t\\t\\ttxType := \\\"1\\\"\\n\\t\\t\\ttxSubg := \\\"0\\\"\\n\\t\\t\\ttxPrnt := \\\"0\\\"\\n\\t\\t\\ttxData := \\\"8d3729b91a13878508c564fbf410ae4f33fcb4cfdb99677f4b23d4c4adb447650964b4fe9da16299831b9cc17aaabd5b8d81fb05460be92af99d128584101a30\\\" // ?\\n\\t\\t\\ttxPrev := \\\"c66f4851618cd53104d4a395212958abf88d96962c0c298a0c7a7c1242fac5c2ee616c8c4f140a2e199558ead6d18ae263b2311b590b0d7bf3777be5b3623d9c\\\" // RockSteady was here\\n\\t\\t\\thash := sha512.Sum512([]byte(txTime + txType + txData + txPrev))\\n\\t\\t\\ttxHash := hex.EncodeToString(hash[:])\\n\\t\\t\\ttxMile := true\\n\\t\\t\\ttxLead := false\\n\\t\\t\\ttxEpoc := txHash\\n\\t\\t\\ttx := db.MustBegin()\\n\\t\\t\\ttx.MustExec(\\\"INSERT INTO \\\"+d.Cf.GetTableName()+\\\" (tx_time, tx_type, tx_hash, tx_data, tx_prev, tx_epoc, tx_subg, tx_prnt, tx_mile, tx_lead, tx_height ) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11)\\\", txTime, txType, txHash, txData, txPrev, txEpoc, txSubg, txPrnt, txMile, txLead, 0)\\n\\t\\t\\ttx.Commit()\\n\\t\\t\\treturn nil\\n\\t\\t} else if count > 0 {\\n\\t\\t\\treturn errors.New(\\\"Root tx already present. \\\")\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func CreateTransfer(w http.ResponseWriter, r *http.Request) {\\r\\n\\tparams := mux.Vars(r)\\r\\n\\tvar v = params[\\\"id\\\"] //We use a simpler variable as the two 'Transfer' functions are fairly long\\r\\n\\t\\r\\n\\tfor _, item := range certif {\\r\\n\\t\\t//Let's first check the user has the right to transfer this certificate\\r\\n if item.Id == v && item.Ownerid==clientnum { \\r\\n\\t\\t\\tcheckk:=\\\"ok\\\"\\r\\n\\t\\t\\t\\r\\n\\t\\t\\t//We browse the transfers list to verify there is no transfer for this certificate\\r\\n\\t\\t\\tfor _, itemm := range trans {\\r\\n\\t\\t\\t\\tif itemm.Id == v {\\r\\n\\t\\t\\t\\t\\tcheckk=\\\"not ok\\\"\\r\\n\\t\\t\\t\\t}\\r\\n\\t\\t\\t}\\r\\n\\t\\t\\t//If so, we inform the user\\r\\n\\t\\t\\tif checkk==\\\"not ok\\\" {\\r\\n\\t\\t\\t\\tfmt.Fprintf(w, \\\"There is already a transfer for this certificate which is pending.\\\")\\r\\n\\t\\t\\t} else {\\r\\n\\t\\t\\t\\tvar transcertif Transfer //Creation of a new transfer\\r\\n\\t\\t\\t\\t\\r\\n\\t\\t\\t\\t//We collect the receiving party's email address \\r\\n\\t\\t\\t\\t_ = json.NewDecoder(r.Body).Decode(&transcertif) \\r\\n\\t\\t\\t\\t\\r\\n\\t\\t\\t\\ttranscertif.Id=v //Certificate ID stays unchanged \\r\\n\\t\\t\\t\\ttranscertif.Status=\\\"Pending\\\" //The transfer status is now changed until the answer of the receiving party\\r\\n\\t\\t\\t\\t\\r\\n\\t\\t\\t\\t//We verify if the email address of the receiver is correct and is not the current user's \\r\\n\\t\\t\\t\\tcheck:=\\\"not ok\\\" \\r\\n\\t\\t\\t\\tfor _, ite := range owner {\\r\\n\\t\\t\\t\\t\\tif ite.Email==transcertif.To && ite.Userid!=clientnum {\\r\\n\\t\\t\\t\\t\\t\\tcheck=\\\"ok\\\"\\r\\n\\t\\t\\t\\t\\t}\\r\\n\\t\\t\\t\\t}\\r\\n\\t\\t\\t\\t\\r\\n\\t\\t\\t\\tif check==\\\"ok\\\" { //Then, if the email is valid, the transfer proposal is sent\\r\\n\\t\\t\\t\\t\\t\\ttrans=append(trans, transcertif) //The transfer is added to the transfers list\\r\\n\\t\\t\\t\\t\\t\\t//The 'transfer' status of the certificate is changed\\r\\n\\t\\t\\t\\t\\t\\tfor index, itp := range certif {\\r\\n\\t\\t\\t\\t\\t\\t\\tif itp.Id==v {\\r\\n\\t\\t\\t\\t\\t\\t\\t\\tcertif[index].Transfer=\\\"Transfer pending\\\"\\r\\n\\t\\t\\t\\t\\t\\t\\t}\\r\\n\\t\\t\\t\\t\\t\\t}\\r\\n\\t\\t\\t\\t\\t\\tfmt.Fprintf(w, \\\"Transfer proposal sent.\\\")\\r\\n\\t\\t\\t\\t\\t\\t\\r\\n\\t\\t\\t\\t}\\telse { //If the email is not valid, the user is informed\\r\\n\\t\\t\\t\\t\\t\\tfmt.Fprintf(w, \\\"This email is not valid.\\\")\\r\\n\\t\\t\\t\\t}\\r\\n\\t\\t\\t}\\r\\n\\t\\t} else if item.Id == v && item.Ownerid!=clientnum {\\r\\n\\t\\t\\tfmt.Fprintf(w, \\\"You are not allowed to transfer this certificate.\\\")\\r\\n\\t\\t}\\r\\n\\t}\\r\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.7724118","0.7264912","0.70775414","0.6888156","0.68655086","0.68548596","0.6757293","0.6735288","0.67024076","0.6690598","0.66825473","0.66563475","0.66508013","0.6650783","0.6646745","0.6620966","0.651692","0.6492073","0.64902925","0.64711386","0.64431155","0.6413492","0.6349267","0.6324566","0.6317189","0.6317171","0.63057953","0.6300514","0.629059","0.62549233","0.624068","0.62293965","0.6223501","0.6218654","0.6213387","0.6183986","0.6165097","0.61645687","0.6162605","0.6137689","0.61369556","0.6129068","0.60602653","0.60508245","0.6050218","0.60404915","0.6036547","0.60299987","0.60229063","0.6019304","0.60180694","0.6013867","0.6009765","0.60041904","0.59954953","0.599169","0.59661","0.59639084","0.594549","0.59435207","0.59420186","0.5926969","0.5926498","0.5924816","0.5915466","0.590202","0.589787","0.5877139","0.5876875","0.58694655","0.5852218","0.58480793","0.5842476","0.5840367","0.5835763","0.58305955","0.5823804","0.58232665","0.57938266","0.57832116","0.57825506","0.57759595","0.57736665","0.57582563","0.5751235","0.575008","0.57492316","0.5744091","0.574083","0.57363623","0.57355344","0.57302046","0.57297397","0.57266575","0.5724789","0.57241297","0.5717956","0.5708074","0.57025945","0.57023287"],"string":"[\n \"0.7724118\",\n \"0.7264912\",\n \"0.70775414\",\n \"0.6888156\",\n \"0.68655086\",\n \"0.68548596\",\n \"0.6757293\",\n \"0.6735288\",\n \"0.67024076\",\n \"0.6690598\",\n \"0.66825473\",\n \"0.66563475\",\n \"0.66508013\",\n \"0.6650783\",\n \"0.6646745\",\n \"0.6620966\",\n \"0.651692\",\n \"0.6492073\",\n \"0.64902925\",\n \"0.64711386\",\n \"0.64431155\",\n \"0.6413492\",\n \"0.6349267\",\n \"0.6324566\",\n \"0.6317189\",\n \"0.6317171\",\n \"0.63057953\",\n \"0.6300514\",\n \"0.629059\",\n \"0.62549233\",\n \"0.624068\",\n \"0.62293965\",\n \"0.6223501\",\n \"0.6218654\",\n \"0.6213387\",\n \"0.6183986\",\n \"0.6165097\",\n \"0.61645687\",\n \"0.6162605\",\n \"0.6137689\",\n \"0.61369556\",\n \"0.6129068\",\n \"0.60602653\",\n \"0.60508245\",\n \"0.6050218\",\n \"0.60404915\",\n \"0.6036547\",\n \"0.60299987\",\n \"0.60229063\",\n \"0.6019304\",\n \"0.60180694\",\n \"0.6013867\",\n \"0.6009765\",\n \"0.60041904\",\n \"0.59954953\",\n \"0.599169\",\n \"0.59661\",\n \"0.59639084\",\n \"0.594549\",\n \"0.59435207\",\n \"0.59420186\",\n \"0.5926969\",\n \"0.5926498\",\n \"0.5924816\",\n \"0.5915466\",\n \"0.590202\",\n \"0.589787\",\n \"0.5877139\",\n \"0.5876875\",\n \"0.58694655\",\n \"0.5852218\",\n \"0.58480793\",\n \"0.5842476\",\n \"0.5840367\",\n \"0.5835763\",\n \"0.58305955\",\n \"0.5823804\",\n \"0.58232665\",\n \"0.57938266\",\n \"0.57832116\",\n \"0.57825506\",\n \"0.57759595\",\n \"0.57736665\",\n \"0.57582563\",\n \"0.5751235\",\n \"0.575008\",\n \"0.57492316\",\n \"0.5744091\",\n \"0.574083\",\n \"0.57363623\",\n \"0.57355344\",\n \"0.57302046\",\n \"0.57297397\",\n \"0.57266575\",\n \"0.5724789\",\n \"0.57241297\",\n \"0.5717956\",\n \"0.5708074\",\n \"0.57025945\",\n \"0.57023287\"\n]"},"document_score":{"kind":"string","value":"0.6715117"},"document_rank":{"kind":"string","value":"8"}}},{"rowIdx":104799,"cells":{"query":{"kind":"string","value":"Handle Transaction Detail Show"},"document":{"kind":"string","value":"func (main *Main) GetDetail(e echo.Context) (err error) {\n\t// get path parameter\n\ttransactionCode := e.Param(\"code\")\n\n\t// get transaction details\n\ttransactionDetail, exc := TransactionModel.Get(\"code\", transactionCode)\n\tif exc != nil {\n\t\treturn rest.ConstructErrorResponse(e, exc)\n\t}\n\n\t// prepare data\n\tdata := map[string]contract.Model{\n\t\t\"transaction_detail\": output.NewTransactionDetail(transactionDetail),\n\t}\n\n\treturn rest.ConstructSuccessResponse(e, data)\n}"},"metadata":{"kind":"string","value":"{\n \"objective\": {\n \"self\": [],\n \"paired\": [],\n \"triplet\": [\n [\n \"query\",\n \"document\",\n \"negatives\"\n ]\n ]\n }\n}"},"negatives":{"kind":"list like","value":["func (tc *TransactionsController) Show(c *gin.Context) {\n\thash := common.HexToHash(c.Param(\"TxHash\"))\n\tif tx, err := tc.App.GetStore().FindTxByAttempt(hash); err == orm.ErrorNotFound {\n\t\tpublicError(c, http.StatusNotFound, errors.New(\"Transaction not found\"))\n\t} else if err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else if doc, err := jsonapi.Marshal(presenters.NewTx(tx)); err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else {\n\t\tc.Data(http.StatusOK, MediaType, doc)\n\t}\n}","func ViewTransaction(rw http.ResponseWriter, r *http.Request) {\n\t// get the token\n\treqToken := r.Header.Get(\"Authorization\")\n\n\t// get the claims\n\tclaims, isNotValid := GetClaims(reqToken, rw)\n\tif isNotValid {\n\t\treturn\n\t}\n\n\tdt, err := db.GetUserTransaction(claims.Roll)\n\tif err != nil {\n\t\trw.WriteHeader(http.StatusInternalServerError)\n\t\trw.Write(Rsp(err.Error(), \"Server Error\"))\n\t\treturn\n\t}\n\trw.WriteHeader(http.StatusOK)\n\tres := c.RespData{\n\t\tMessage: \"All data\",\n\t\tData: dt,\n\t}\n\tjson.NewEncoder(rw).Encode(res)\n}","func (c *PurchaseController) Show(ctx *app.ShowPurchaseContext) error {\n\n\tsession := Database.Session.Copy()\n\tdefer session.Close()\n\n\tresult := app.Purchase{}\n\n\terr := session.DB(\"services-pos\").C(\"Purchase\").FindId(bson.ObjectIdHex(ctx.TransactionID)).One(&result)\n\n\tif err != nil {\n\t\treturn ctx.NotFound()\n\t}\n\n\tresult.TransactionID = ctx.TransactionID\n\tresult.Href = app.PurchaseHref(ctx.TransactionID)\n\n\treturn ctx.OK(&result)\n}","func (g *Goods) Detail(c Context) {\n\t// TODO\n\tc.String(http.StatusOK, \"get goods detail\")\n}","func (this *FamilyAccount) showDetails() {\n\tfmt.Println(\"------My Income and Expense Detail-------\")\n\tif this.flag {\n\t\t//因为我们用的是FamilyAccount结构体里传过来的字段,所以不能直接yongflag, 要用this. , 表示调用这个方法的结构体变量里面的字段\n\t\tfmt.Println(this.details)\n\t} else {\n\t\tfmt.Println(\"No current income and expenditure details!\")\n\t}\n}","func (r *BTCRPC) GetTransactionDetail(txhash string) ([]byte, error) {\n\tvar (\n\t\ttx []byte\n\t\terr error\n\t)\n\n\terr = r.Client.Call(\"getrawtransaction\", jsonrpc.Params{txhash, 1}, &tx)\n\treturn tx, err\n}","func (h *Handler) show(c echo.Context) (e error) {\n\tctx := c.(*cuxs.Context)\n\n\tvar id int64\n\tvar as *model.SalesReturn\n\tif id, e = common.Decrypt(ctx.Param(\"id\")); e == nil {\n\t\tif as, e = ShowSalesReturn(\"id\", id); e == nil {\n\t\t\tctx.Data(as)\n\t\t} else {\n\t\t\te = echo.ErrNotFound\n\t\t}\n\t}\n\treturn ctx.Serve(e)\n}","func (p *Proxy) handleShowTxnz(session *driver.Session, query string, node sqlparser.Statement) (*sqltypes.Result, error) {\n\treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// privilegePlug := spanner.plugins.PlugPrivilege()\n\t// if !privilegePlug.IsSuperPriv(session.User()) {\n\t// \treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// }\n\n\t// qr := &sqltypes.Result{}\n\t// qr.Fields = []*querypb.Field{\n\t// \t{Name: \"TxnID\", Type: querypb.Type_INT64},\n\t// \t{Name: \"Start\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"Duration\", Type: querypb.Type_INT32},\n\t// \t{Name: \"XaState\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"TxnState\", Type: querypb.Type_VARCHAR},\n\t// }\n\n\t// rows := spanner.scatter.Txnz().GetTxnzRows()\n\t// for _, row := range rows {\n\t// \trow := []sqltypes.Value{\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT64, []byte(fmt.Sprintf(\"%v\", uint64(row.TxnID)))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.Start.Format(\"20060102150405.000\"))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT32, []byte(fmt.Sprintf(\"%v\", row.Duration))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.XaState)),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.State)),\n\t// \t}\n\t// \tqr.Rows = append(qr.Rows, row)\n\t// }\n\t// return qr, nil\n}","func (showTxCommand ShowTransactionCommand) Run(ctx context.Context, wallet walletcore.Wallet) error {\n\ttransaction, err := wallet.GetTransaction(showTxCommand.Args.TxHash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbasicOutput := \"Hash\\t%s\\n\" +\n\t\t\"Confirmations\\t%d\\n\" +\n\t\t\"Included in block\\t%d\\n\" +\n\t\t\"Type\\t%s\\n\" +\n\t\t\"Amount %s\\t%s\\n\" +\n\t\t\"Date\\t%s\\n\" +\n\t\t\"Size\\t%s\\n\" +\n\t\t\"Fee\\t%s\\n\" +\n\t\t\"Rate\\t%s/kB\\n\"\n\n\ttxDirection := strings.ToLower(transaction.Direction.String())\n\ttxSize := fmt.Sprintf(\"%.1f kB\", float64(transaction.Size)/1000)\n\tbasicOutput = fmt.Sprintf(basicOutput,\n\t\ttransaction.Hash,\n\t\ttransaction.Confirmations,\n\t\ttransaction.BlockHeight,\n\t\ttransaction.Type,\n\t\ttxDirection, transaction.Amount,\n\t\ttransaction.FormattedTime,\n\t\ttxSize,\n\t\ttransaction.Fee,\n\t\ttransaction.FeeRate)\n\n\tif showTxCommand.Detailed {\n\t\tdetailedOutput := strings.Builder{}\n\t\tdetailedOutput.WriteString(\"General Info\\n\")\n\t\tdetailedOutput.WriteString(basicOutput)\n\t\tdetailedOutput.WriteString(\"\\nInputs\\n\")\n\t\tfor _, input := range transaction.Inputs {\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t%s\\n\", dcrutil.Amount(input.AmountIn).String(), input.PreviousOutpoint))\n\t\t}\n\t\tdetailedOutput.WriteString(\"\\nOutputs\\n\")\n\t\tfor _, out := range transaction.Outputs {\n\t\t\tif len(out.Addresses) == 0 {\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t (no address)\\n\", dcrutil.Amount(out.Value).String()))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\", dcrutil.Amount(out.Value).String()))\n\t\t\tfor _, address := range out.Addresses {\n\t\t\t\taccountName := address.AccountName\n\t\t\t\tif !address.IsMine {\n\t\t\t\t\taccountName = \"external\"\n\t\t\t\t}\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"\\t%s (%s)\\n\", address.Address, accountName))\n\t\t\t}\n\t\t}\n\t\ttermio.PrintStringResult(strings.TrimRight(detailedOutput.String(), \" \\n\\r\"))\n\t} else {\n\t\ttermio.PrintStringResult(basicOutput)\n\t}\n\treturn nil\n}","func viewAnyOrderGet(c *gin.Context) { //admin also have the same view , later combine those two func TBD\n\tOrdID := c.Request.URL.Query()[\"ordid\"][0] // Getting Order ID passed with URL\n\t_, usrName := session.SessinStatus(c, \"user_session_cookie\")\n\tfmt.Println(\"Wnat to see the order details of order number \", OrdID)\n\toK, itemsList, date, status, PayMode, amt := db.GetSingleOredrDetails(OrdID)\n\tif !oK {\n\t\tfmt.Println(\"Something went wrong while picking Single Order Deatils ..Please have a look\")\n\t}\n\tfmt.Println(oK, itemsList, date, status, PayMode, amt)\n\t//\t\tsubTotalToFloat, _ := strconv.ParseFloat(singleCartItem.SubTotal, 64)\n\t//\t\tTotalAmt = TotalAmt + subTotalToFloat\n\t//\tTotalAmtInPaisa := TotalAmt * 100 // This is required while initate for payment in Razorpay\n\n\t//\tTotalAmtString := fmt.Sprintf(\"%.2f\", TotalAmt)\n\n\tc.HTML(\n\t\thttp.StatusOK,\n\t\t\"view_particular_order.html\",\n\t\tgin.H{\"title\": \"OrderDetail\",\n\t\t\t\"ItemsOrdered\": itemsList,\n\t\t\t\"OrdID\": OrdID,\n\t\t\t\"date\": date,\n\t\t\t\"PayMode\": PayMode,\n\t\t\t\"amt\": amt,\n\t\t\t\"OrdStatus\": status,\n\t\t\t\"usrName\": usrName,\n\n\t\t\t// \"TotalAmt\": TotalAmtString,\n\t\t\t// \"TotalAmtInPaisa\": TotalAmtInPaisa,\n\t\t},\n\t)\n\n}","func (s *Service) GetExplorerTransaction(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tid := r.FormValue(\"id\")\n\n\tdata := &Data{}\n\tdefer func() {\n\t\tif err := json.NewEncoder(w).Encode(data.TX); err != nil {\n\t\t\tutils.Logger().Warn().Err(err).Msg(\"cannot JSON-encode TX\")\n\t\t}\n\t}()\n\tif id == \"\" {\n\t\tutils.Logger().Warn().Msg(\"invalid id parameter\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\tdb := s.Storage.GetDB()\n\tbytes, err := db.Get([]byte(GetTXKey(id)))\n\tif err != nil {\n\t\tutils.Logger().Warn().Err(err).Str(\"id\", id).Msg(\"cannot read TX\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\ttx := new(Transaction)\n\tif rlp.DecodeBytes(bytes, tx) != nil {\n\t\tutils.Logger().Warn().Str(\"id\", id).Msg(\"cannot convert data from DB\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\tdata.TX = *tx\n}","func (ctl *SaleCounterProductController) Detail() {\n\t//获取信息一样,直接调用Edit\n\tctl.Edit()\n\tctl.Data[\"Readonly\"] = true\n\tctl.Data[\"Action\"] = \"detail\"\n}","func orderAdminApproveGet(c *gin.Context) {\n\tOrdID := c.Request.URL.Query()[\"ordid\"][0] // Getting Order ID passed with URL\n\t_, usrName := session.SessinStatus(c, \"user_session_cookie\")\n\tfmt.Println(\"Wnat to see the order details of order number \", OrdID)\n\toK, itemsList, date, status, PayMode, amt := db.GetSingleOredrDetails(OrdID)\n\tif !oK {\n\t\tfmt.Println(\"Something went wrong while picking Single Order Deatils ..Please have a look\")\n\t}\n\tfmt.Println(oK, itemsList, date, status, PayMode, amt)\n\tc.HTML(\n\t\thttp.StatusOK,\n\t\t\"order_adminview.html\",\n\t\tgin.H{\"title\": \"OrderDetail\",\n\t\t\t\"ItemsOrdered\": itemsList,\n\t\t\t\"OrdID\": OrdID,\n\t\t\t\"date\": date,\n\t\t\t\"PayMode\": PayMode,\n\t\t\t\"amt\": amt,\n\t\t\t\"OrdStatus\": status,\n\t\t\t\"usrName\": usrName,\n\n\t\t\t// \"TotalAmt\": TotalAmtString,\n\t\t\t// \"TotalAmtInPaisa\": TotalAmtInPaisa,\n\t\t},\n\t)\n}","func (c *TradeController) Show(ctx *app.ShowTradeContext) error {\n\t// TradeController_Show: start_implement\n\n\t// Put your logic here\n\tt, ok := tradeRegistry[ctx.TradeID]\n\tif !ok {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.GoaTrade{\n\t\tTradeID: t.TradeID,\n\t\tContractID: t.ContractID,\n\t\tCounterpartyID: t.CounterpartyID,\n\t}\n\treturn ctx.OK(res)\n\t// TradeController_Show: end_implement\n}","func (action *TransactionShowAction) JSON() {\n\taction.Do(\n\t\taction.EnsureHistoryFreshness,\n\t\taction.loadParams,\n\t\taction.checkAllowed,\n\t\taction.loadRecord,\n\t\taction.loadResource,\n\t\tfunc() { hal.Render(action.W, action.Resource) },\n\t)\n}","func view(ctx context.Context, db transactor, fn func(*bolt.Tx) error) error {\n\ttx, ok := ctx.Value(transactionKey{}).(*bolt.Tx)\n\tif !ok {\n\t\treturn db.View(fn)\n\t}\n\treturn fn(tx)\n}","func (db *DB) View(ctx context.Context, fn func(*TransactionManager) error) error {\n\ttx := db.BeginTransaction(false)\n\tdefer tx.Discard()\n\treturn fn(tx)\n}","func displayTransactions(scid string, option int, ID string) {\n\t\t\n\tscKeys:= []string{\"numberOfOwners\", \"txCount\"}\n\tresult:= getKeysFromDaemon(scid, scKeys)\n\tif result == \"\" {return}\n\n\n\t//Response ok, extract keys from JSON\n\t\n\n\ttxCount := gjson.Get(result, \"txs.#.sc_keys.txCount\")\n\ttxCountArray:= txCount.Array()[0]\n\ttxCountInt:= txCountArray.Int()\n\t//fmt.Printf(\"Tx Count: %d\\n\", txCountInt)\n\n\t//Make a slice of keys so we can request in RPC call\n\tx:= int(txCountInt) //txCount in wallet smart contract is always 1 ahead of actual number of transactions\t\n\tx4:= x * 4\t\n\tkeySlice:= make([]string, x4) \n\t\n\tfor i:=0; i%.2f\", urlCustomersTransactionsView(q.CustomerID, q.AccountID, q.ID), q.Amount)\n\t\t\tcase \"created_at\":\n\t\t\t\tdate := web.NewTimeResponse(ctx, time.Unix(q.CreatedAt, 0))\n\t\t\t\tv.Value = date.LocalDate\n\t\t\t\tv.Formatted = date.LocalDate\n\t\t\tcase \"narration\":\n\t\t\t\tvalues := strings.Split(q.Narration, \":\")\n\t\t\t\tif len(values) > 1 {\n\t\t\t\t\tif values[0] == \"sale\" {\n\t\t\t\t\t\tv.Value = values[1]\n\t\t\t\t\t\tv.Formatted = fmt.Sprintf(\"%s\", urlSalesView(values[2]), v.Value)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tv.Value = q.Narration\n\t\t\t\t\tv.Formatted = q.Narration\n\t\t\t\t}\n\t\t\tcase \"payment_method\":\n\t\t\t\tv.Value = q.PaymentMethod\n\t\t\t\tv.Formatted = q.PaymentMethod\n\t\t\tcase \"customer_name\":\n\t\t\t\tv.Value = q.CustomerName\n\t\t\t\tv.Formatted = fmt.Sprintf(\"%s\", urlCustomersView(q.CustomerID), v.Value)\n\t\t\tcase \"account\":\n\t\t\t\tv.Value = q.AccountNumber\n\t\t\t\tv.Formatted = fmt.Sprintf(\"%s\", urlCustomersAccountsView(q.CustomerID, q.AccountID), v.Value)\n\t\t\tcase \"sales_rep_id\":\n\t\t\t\tv.Value = q.SalesRepID\n\t\t\t\tv.Formatted = fmt.Sprintf(\"%s\", urlUsersView(q.SalesRepID), q.SalesRep)\n\t\t\tdefault:\n\t\t\t\treturn resp, errors.Errorf(\"Failed to map value for %s.\", col.Field)\n\t\t\t}\n\t\t\tresp = append(resp, v)\n\t\t}\n\n\t\treturn resp, nil\n\t}\n\n\tvar txWhere = []string{\"tx_type = 'deposit'\"}\n\tvar txArgs []interface{}\n\t// todo sales rep filtering\n\tif v := r.URL.Query().Get(\"sales_rep_id\"); v != \"\" {\n\t\ttxWhere = append(txWhere, \"sales_rep_id = $1\")\n\t\ttxArgs = append(txArgs, v)\n\t\tdata[\"salesRepID\"] = v\n\t}\n\n\tif v := r.URL.Query().Get(\"payment_method\"); v != \"\" {\n\t\ttxWhere = append(txWhere, fmt.Sprintf(\"payment_method = $%d\", len(txArgs)+1))\n\t\ttxArgs = append(txArgs, v)\n\t\tdata[\"paymentMethod\"] = v\n\t}\n\n\tvar date = time.Now()\n\tif v := r.URL.Query().Get(\"start_date\"); v != \"\" {\n\t\tdate, err = time.Parse(\"01/02/2006\", v)\n\t\tif err != nil {\n\t\t\tdate = time.Now()\n\t\t\treturn err\n\t\t}\n\t}\n\tdate = date.Truncate(time.Millisecond)\n\tdate = now.New(date).BeginningOfDay().Add(-1 * time.Hour)\n\ttxWhere = append(txWhere, fmt.Sprintf(\"created_at >= $%d\", len(txArgs)+1))\n\ttxArgs = append(txArgs, date.UTC().Unix())\n\tdata[\"startDate\"] = date.Format(\"01/02/2006\")\n\n\tdate = time.Now()\n\tif v := r.URL.Query().Get(\"end_date\"); v != \"\" {\n\t\tdate, err = time.Parse(\"01/02/2006\", v)\n\t\tif err != nil {\n\t\t\tdate = time.Now()\n\t\t\treturn err\n\t\t}\n\n\t}\n\tdate = date.Truncate(time.Millisecond)\n\tdate = now.New(date).EndOfDay().Add(-1 * time.Hour)\n\ttxWhere = append(txWhere, fmt.Sprintf(\"created_at <= $%d\", len(txArgs)+1))\n\ttxArgs = append(txArgs, date.Unix())\n\tdata[\"endDate\"] = date.Format(\"01/02/2006\")\n\n\tloadFunc := func(ctx context.Context, sorting string, fields []datatable.DisplayField) (resp [][]datatable.ColumnValue, err error) {\n\n\t\tvar order []string\n\t\tif len(sorting) > 0 {\n\t\t\torder = strings.Split(sorting, \",\")\n\t\t}\n\n\t\tfor i := range txWhere {\n\t\t\ttxWhere[i] = \"tx.\" + txWhere[i]\n\t\t}\n\t\tres, err := h.TransactionRepo.TxReport(ctx, claims, transaction.FindRequest{\n\t\t\tOrder: order, Where: strings.Join(txWhere, \" AND \"), Args: txArgs,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn resp, err\n\t\t}\n\n\t\tfor _, a := range res {\n\t\t\tl, err := mapFunc(a, fields)\n\t\t\tif err != nil {\n\t\t\t\treturn resp, errors.Wrapf(err, \"Failed to map brand for display.\")\n\t\t\t}\n\n\t\t\tresp = append(resp, l)\n\t\t}\n\n\t\treturn resp, nil\n\t}\n\n\tdt, err := datatable.New(ctx, w, r, h.Redis, fields, loadFunc)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif dt.HasCache() {\n\t\treturn nil\n\t}\n\n\tif ok, err := dt.Render(); ok {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\n\tusers, err := h.UserRepos.Find(ctx, claims, user.UserFindRequest{\n\t\tOrder: []string{\"first_name\", \"last_name\"},\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttotal, err = h.TransactionRepo.DepositAmountByWhere(ctx, strings.Join(txWhere, \" and \"), txArgs)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata[\"paymentMethods\"] = transaction.PaymentMethods\n\tdata[\"users\"] = users\n\tdata[\"total\"] = total\n\tdata[\"datatable\"] = dt.Response()\n\n\treturn h.Renderer.Render(ctx, w, r, TmplLayoutBase, \"report-transactions.gohtml\", web.MIMETextHTMLCharsetUTF8, http.StatusOK, data)\n}","func viewandedititemGet(c *gin.Context) {\n\tIsSectionActive, _ := session.SessinStatus(c, \"admin_session_cookie\")\n\tif !IsSectionActive {\n\t\tfmt.Println(\"No Active Sessions found \")\n\t\t// c.HTML(http.StatusOK, \"admin_login.html\", []string{\"a\", \"b\", \"c\"})\n\t\tc.HTML(\n\t\t\thttp.StatusOK,\n\t\t\t\"admin_login.html\",\n\t\t\tgin.H{\"title\": \"success login\",\n\t\t\t\t\"diplay\": \"none\",\n\t\t\t},\n\t\t)\n\t} else {\n\t\titemID := c.Request.URL.Query()[\"itemid\"][0] // Getting Order ID passed with URL\n\t\tfmt.Println(\"Initiating to View/Edit item ,having ID\", itemID)\n\t\t//populateCategoryItems(c, itemID)\n\t\t//GetItemDetails(itemID string) (itemDesc string, itemRate float64, unit string,itmID,itmStock int,itmBuyRate float64) {\n\t\t//Don't Confuse above function will redirect\n\t\t//to edit page, usual practice is giving here\n\t\t//but we achived this by modifying the existing\n\t\t//code so it happend so..\n\t\titmDesc, itmSelRate, itmUnit, itmID, itmStock, itmBuyPrice := db.GetItemDetails(itemID)\n\t\tc.HTML(\n\t\t\thttp.StatusOK,\n\t\t\t\"edit_item.html\", gin.H{\n\t\t\t\t\"delWarning\": \"none\",\n\t\t\t\t\"updateSucess\": \"none\",\n\t\t\t\t\"title\": \"Edit Item\",\n\t\t\t\t\"itmID\": itmID,\n\t\t\t\t\"itmDesc\": itmDesc,\n\t\t\t\t\"itmUnit\": itmUnit,\n\t\t\t\t\"itmBuyPrice\": itmBuyPrice,\n\t\t\t\t\"itmSelRate\": itmSelRate,\n\t\t\t\t\"itmStock\": itmStock,\n\t\t\t})\n\t}\n}","func (db *DB) View(fn func(*TransactionManager) error) error {\n\ttx := db.BeginTransaction(false)\n\tdefer tx.Discard()\n\treturn fn(tx)\n}","func transactionSummary(ctx *quorumContext, w http.ResponseWriter, r *http.Request) (int, error) {\n\tif ctx.conn == nil {\n\t\tfmt.Fprintf(w, \"Cannot get transaction before deploying contract\\n\")\n\t\treturn 400, nil\n\t}\n\n\t// parse hash from request URL\n\tkeys := r.URL.Query()[\"hash\"]\n\tif len(keys) < 1 {\n\t\tfmt.Fprintf(w, \"Invalid parameter, require 'hash'\")\n\t\treturn 400, nil\n\t}\n\tlog.Println(\"Hash supplied :\", keys[0])\n\thash := common.HexToHash(keys[0])\n\n\tbasic_context := context.Background()\n\ttx, pending, err := ctx.conn.TransactionByHash(basic_context, hash)\n\t//common.HexToHash(\"0x378674bebd1430d9ce63adc792c573da56e69b8d6c97174c93a43c5991ae0d61\"))\n\tif err != nil {\n\t\tfmt.Fprintf(w, \"Failed to get transaction details: %v\", err)\n\t\treturn 500, err\n\t}\n\tfmt.Fprintf(w, \"Transaction pending? %v; details: %v\\n\",\n\t\tpending, tx.String())\n\treturn 200, nil\n}","func (self* userRestAPI) transactions(w http.ResponseWriter, r *http.Request) {\n\n\t// Read arguments\n\tband,number,err := self.extractBandAndNumber(r)\n\tif err != nil {\n\t\tlogError(err)\n\t\thttp.Error(w, fmt.Sprintf(\"\\nFailed to parse arguments '%s'\\n\",err), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t// Retrieve transactions for specified traveller\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\thistory,err := self.engine.TransactionsAsJSON(band,number)\n\tif err != nil {\n\t\tlogError(err)\n\t\thttp.Error(w, fmt.Sprintf(\"\\nFailed to retrieve transaction history with error '%s'\\n\",err), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tio.WriteString(w,history)\n\n}","func (c *PlanController) PlanDetails() {\n\tplanViewModel := viewmodels.Plan{}\n\tr := c.Ctx.Request\n\tw := c.Ctx.ResponseWriter\n\tsessionValues, sessionStatus := SessionForPlan(w,r)\n\tplanViewModel.SessionFlag = sessionStatus\n\tplanViewModel.CompanyPlan = sessionValues.CompanyPlan\n\tplanViewModel.CompanyTeamName =sessionValues.CompanyTeamName\n\tc.Data[\"vm\"] = planViewModel\n\tc.TplName = \"template/plan.html\"\n}","func (action TransactionShowAction) ServeHTTPC(c web.C, w http.ResponseWriter, r *http.Request) {\n\tap := &action.Action\n\tap.Prepare(c, w, r)\n\taction.Log = action.Log.WithField(\"action\", \"TransactionShowAction\")\n\tap.Execute(&action)\n}","func (api *API) Get(tid string) (*pagarme.Response, *pagarme.Transaction, error) {\n\tresp, err := api.Config.Do(http.MethodGet, \"/transactions/\"+tid, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif werr := www.ExtractError(resp); werr != nil {\n\t\treturn werr, nil, nil\n\t}\n\tresult := &pagarme.Transaction{}\n\tif err := www.Unmarshal(resp, result); err != nil {\n\t\tapi.Config.Logger.Error(\"could not unmarshal transaction [Get]: \" + err.Error())\n\t\treturn nil, nil, err\n\t}\n\n\treturn www.Ok(), result, nil\n}","func (tb *transactionBuilder) View() (types.Transaction, []types.Transaction) {\n\treturn tb.transaction, tb.parents\n}","func (tb *transactionBuilder) View() (types.Transaction, []types.Transaction) {\n\treturn tb.transaction, tb.parents\n}","func ShowDeposit2ChannelMsgDetail(Status string) {\n\tswitch Status {\n\tcase \"200 OK\":\n\t\tlog.Println(\"Success Deposit\")\n\tcase \"400 Bad Request\":\n\t\tlog.Println(\"The provided json is in some way malformed!\")\n\tcase \"402 Payment required\":\n\t\tlog.Println(\"Insufficient balance to do a deposit\")\n\tcase \"408 Request Timeout\":\n\t\tlog.Println(\"The deposit event was not read in time by the ethereum node\")\n\tcase \"409 Conflict\":\n\t\tlog.Println(\"Provided channel does not exist\")\n\tcase \"500 Server Error\":\n\t\tlog.Println(\"Internal Raiden node error\")\n\tcase \"504 TimeOut\":\n\t\tlog.Println(\"No response,timeout\")\n\tdefault:\n\t\tfmt.Printf(\"Unknown error,Deposit Failure! %s\\n\", Status)\n\t}\n}","func (account *Account) ShowTx(P string) bool {\n\tfor i := 0; i < account.WalletIndex; i++ {\n\t\tif P == account.Wallet[i].P {\n\t\t\tJSON, _ := json.MarshalIndent(account.Wallet[i], \"\", \" \")\n\n\t\t\tfmt.Println(string(JSON))\n\t\t\treturn true\n\n\t\t}\n\t}\n\treturn false\n}","func (s *searcher) Transaction(resp http.ResponseWriter, req *http.Request) {\n\tsearchTerms := mux.Vars(req)\n\n\ttransactionID := searchTerms[\"transaction_id\"]\n\tif len(transactionID) == 0 {\n\t\tresp.WriteHeader(http.StatusBadRequest)\n\t\tresp.Write([]byte(\"transaction ID is empty\"))\n\t\treturn\n\t}\n\n\tif len(transactionID) != 64 {\n\t\tresp.WriteHeader(http.StatusBadRequest)\n\t\tresp.Write([]byte(\"transaction ID is not 64 characters\"))\n\t\treturn\n\t}\n\n\tfileName, transactionIndex, err := s.searchIndex.GetTransactionPathByID(transactionID)\n\tif err != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tresp.Write([]byte(fmt.Sprintf(\"error finding transaction: %s\", err.Error())))\n\t\treturn\n\t}\n\n\ttransactions, err := s.searchIndex.GetTransactionsFromSingleFile(fileName, []int{transactionIndex})\n\tif err != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tresp.Write([]byte(fmt.Sprintf(\"error finding transaction: %s\", err.Error())))\n\t\treturn\n\t}\n\n\tresultBytes, err := json.Marshal(transactions)\n\tif err != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tresp.Write([]byte(fmt.Sprintf(\"error marshallig transaction to json: %s\", err.Error())))\n\t\treturn\n\t}\n\n\tresp.WriteHeader(http.StatusOK)\n\tresp.Write(resultBytes)\n}","func (db *MySQLDB) GetTransactionDetailByTransactionID(ctx context.Context, transactionID int) ([]*TransactionDetailRecord, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"GetTransactionByTransactionID\")\n\ttransaction := &TransactionRecord{}\n\n\trow := db.instance.QueryRowContext(ctx, \"SELECT id FROM transactions WHERE id = ?\", transactionID)\n\terr := row.Scan(&transaction.ID)\n\tif err != nil {\n\t\tfLog.Errorf(\"row.Scan got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\tq := fmt.Sprintf(\"SELECT td.transaction_id, td.product_id, td.qty, td.sub_total, p.sku FROM transaction_detail td INNER JOIN products p ON td.product_id = p.id WHERE transaction_id = %v\", transactionID)\n\trows, err := db.instance.QueryContext(ctx, q)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.QueryContext got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\ttDetail := make([]*TransactionDetailRecord, 0)\n\tfor rows.Next() {\n\t\ttD := &TransactionDetailRecord{}\n\t\terr := rows.Scan(&tD.TransactionID, &tD.ProductID, &tD.Qty, &tD.SubTotal, tD.SKU)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"rows.Scan got %s\", err.Error())\n\t\t} else {\n\t\t\ttDetail = append(tDetail, tD)\n\t\t}\n\t}\n\n\treturn tDetail, nil\n}","func (res BackTestResult) Show() TradestatPort {\n\t//\tp := NewPortfolio()\n\tmds := bean.NewRPCMDSConnC(\"tcp\", res.dbhost+\":\"+res.dbport)\n\tratesbook := make(ReferenceRateBook)\n\n\t// FIXME: think about how to show multi pair result\n\tvar stat TradestatPort\n\tif len(res.pairs) > 0 {\n\t\tp := res.pairs[0]\n\t\ttxn, _ := mds.GetTransactions(p, res.start, res.end)\n\t\tratesbook[p] = RefRatesFromTxn(txn)\n\t\t//\t\tsnapts.Print()\n\t\t//\t\tperfts.Print()\n\n\t\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\n\t\tstat.Print()\n\t}\n\treturn stat\n}","func (s *Store) TxDetails(ns walletdb.ReadBucket, txHash *chainhash.Hash) (*TxDetails, error) {\n\t// First, check whether there exists an unmined transaction with this\n\t// hash. Use it if found.\n\tv := existsRawUnmined(ns, txHash[:])\n\tif v != nil {\n\t\treturn s.unminedTxDetails(ns, txHash, v)\n\t}\n\n\t// Otherwise, if there exists a mined transaction with this matching\n\t// hash, skip over to the newest and begin fetching all details.\n\tk, v := latestTxRecord(ns, txHash)\n\tif v == nil {\n\t\t// not found\n\t\treturn nil, nil\n\t}\n\treturn s.minedTxDetails(ns, txHash, k, v)\n}","func (dao *InfoDao) Transaction(ctx context.Context, f func(ctx context.Context, tx *gdb.TX) error) (err error) {\n\treturn dao.Ctx(ctx).Transaction(ctx, f)\n}","func Show(c *gin.Context) {\r\n\tpost := getById(c)\r\n\tc.JSON(http.StatusOK, gin.H{\r\n\t\t\"messege\": \"\",\r\n\t\t\"data\": post,\r\n\t})\r\n}","func tableInfoFlowShow(meta *halproto.TableMetadata) {\n\tfMeta := meta.GetFlowMeta()\n\tinsStr := fmt.Sprintf(\"%d[%d]\", fMeta.GetNumInserts(), fMeta.GetNumInsertFailures())\n\tupdStr := fmt.Sprintf(\"%d[%d]\", fMeta.GetNumUpdates(), fMeta.GetNumUpdateFailures())\n\tdelStr := fmt.Sprintf(\"%d[%d]\", fMeta.GetNumDeletes(), fMeta.GetNumDeleteFailures())\n\tfmt.Printf(\"%-30s%-10d%-10s%-10d%-10d%-10d%-10d%-10s%-10s%-10s\\n\",\n\t\tmeta.GetTableName(),\n\t\tmeta.GetTableId(),\n\t\tutils.TableKindToStr(meta.GetKind()),\n\t\tfMeta.GetCapacity(),\n\t\tfMeta.GetCollCapacity(),\n\t\tfMeta.GetHashUsage(),\n\t\tfMeta.GetCollUsage(),\n\t\tinsStr,\n\t\tupdStr,\n\t\tdelStr)\n}","func (c *BalanceClient) RetrieveTransaction(id string) (*BalanceTransaction, error) {\n\tbalanceTransaction := BalanceTransaction{}\n\terr := c.client.get(\"/balance/history/\"+id, nil, &balanceTransaction)\n\treturn &balanceTransaction, err\n}","func (s *Server) handleTransaction(client string, req *pb.Command) (err error) {\n\t// Get the transfer from the original command, will panic if nil\n\ttransfer := req.GetTransfer()\n\tmsg := fmt.Sprintf(\"starting transaction of %0.2f from %s to %s\", transfer.Amount, transfer.Account, transfer.Beneficiary)\n\ts.updates.Broadcast(req.Id, msg, pb.MessageCategory_LEDGER)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Handle Demo UI errors before the account lookup\n\tif transfer.OriginatingVasp != \"\" && transfer.OriginatingVasp != s.vasp.Name {\n\t\tlog.Info().Str(\"requested\", transfer.OriginatingVasp).Str(\"local\", s.vasp.Name).Msg(\"requested originator does not match local VASP\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrWrongVASP, \"message sent to the wrong originator VASP\"),\n\t\t)\n\t}\n\n\t// Lookup the account associated with the transfer originator\n\tvar account Account\n\tif err = LookupAccount(s.db, transfer.Account).First(&account).Error; err != nil {\n\t\tif errors.Is(err, gorm.ErrRecordNotFound) {\n\t\t\tlog.Info().Str(\"account\", transfer.Account).Msg(\"not found\")\n\t\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\t\tpb.Errorf(pb.ErrNotFound, \"account not found\"),\n\t\t\t)\n\t\t}\n\t\treturn fmt.Errorf(\"could not fetch account: %s\", err)\n\t}\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"account %04d accessed successfully\", account.ID), pb.MessageCategory_LEDGER)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Lookup the wallet of the beneficiary\n\tvar beneficiary Wallet\n\tif err = LookupBeneficiary(s.db, transfer.Beneficiary).First(&beneficiary).Error; err != nil {\n\t\tif errors.Is(err, gorm.ErrRecordNotFound) {\n\t\t\tlog.Info().Str(\"beneficiary\", transfer.Beneficiary).Msg(\"not found\")\n\t\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\t\tpb.Errorf(pb.ErrNotFound, \"beneficiary wallet not found\"),\n\t\t\t)\n\t\t}\n\t\treturn fmt.Errorf(\"could not fetch beneficiary wallet: %s\", err)\n\t}\n\n\tif transfer.CheckBeneficiary {\n\t\tif transfer.BeneficiaryVasp != beneficiary.Provider.Name {\n\t\t\tlog.Info().\n\t\t\t\tStr(\"expected\", transfer.BeneficiaryVasp).\n\t\t\t\tStr(\"actual\", beneficiary.Provider.Name).\n\t\t\t\tMsg(\"check beneficiary failed\")\n\t\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\t\tpb.Errorf(pb.ErrWrongVASP, \"beneficiary wallet does not match beneficiary vasp\"),\n\t\t\t)\n\t\t}\n\t}\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"wallet %s provided by %s\", beneficiary.Address, beneficiary.Provider.Name), pb.MessageCategory_BLOCKCHAIN)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// TODO: lookup peer from cache rather than always doing a directory service lookup\n\tvar peer *peers.Peer\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"search for %s in directory service\", beneficiary.Provider.Name), pb.MessageCategory_TRISADS)\n\tif peer, err = s.peers.Search(beneficiary.Provider.Name); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not search peer from directory service\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not search peer from directory service\"),\n\t\t)\n\t}\n\tinfo := peer.Info()\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"identified TRISA remote peer %s at %s via directory service\", info.ID, info.Endpoint), pb.MessageCategory_TRISADS)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\tvar signKey *rsa.PublicKey\n\ts.updates.Broadcast(req.Id, \"exchanging peer signing keys\", pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\tif signKey, err = peer.ExchangeKeys(true); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not exchange keys with remote peer\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not exchange keyrs with remote peer\"),\n\t\t)\n\t}\n\n\t// Prepare the transaction\n\t// Save the pending transaction and increment the accounts pending field\n\txfer := Transaction{\n\t\tEnvelope: uuid.New().String(),\n\t\tAccount: account,\n\t\tAmount: decimal.NewFromFloat32(transfer.Amount),\n\t\tDebit: true,\n\t\tCompleted: false,\n\t}\n\n\tif err = s.db.Save(&xfer).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not save transaction\"),\n\t\t)\n\t}\n\n\t// Save the pending transaction on the account\n\t// TODO: remove pending transactions\n\taccount.Pending++\n\tif err = s.db.Save(&account).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save originator account\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not save originator account\"),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, \"ready to execute transaction\", pb.MessageCategory_BLOCKCHAIN)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Create an identity and transaction payload for TRISA exchange\n\ttransaction := &generic.Transaction{\n\t\tTxid: fmt.Sprintf(\"%d\", xfer.ID),\n\t\tOriginator: account.WalletAddress,\n\t\tBeneficiary: beneficiary.Address,\n\t\tAmount: float64(transfer.Amount),\n\t\tNetwork: \"TestNet\",\n\t\tTimestamp: xfer.Timestamp.Format(time.RFC3339),\n\t}\n\tidentity := &ivms101.IdentityPayload{\n\t\tOriginator: &ivms101.Originator{},\n\t\tOriginatingVasp: &ivms101.OriginatingVasp{},\n\t}\n\tif identity.OriginatingVasp.OriginatingVasp, err = s.vasp.LoadIdentity(); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not load originator vasp\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not load originator vasp\"),\n\t\t)\n\t}\n\n\tidentity.Originator = &ivms101.Originator{\n\t\tOriginatorPersons: make([]*ivms101.Person, 0, 1),\n\t\tAccountNumbers: []string{account.WalletAddress},\n\t}\n\tvar originator *ivms101.Person\n\tif originator, err = account.LoadIdentity(); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not load originator identity\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not load originator identity\"),\n\t\t)\n\t}\n\tidentity.Originator.OriginatorPersons = append(identity.Originator.OriginatorPersons, originator)\n\n\tpayload := &protocol.Payload{}\n\tif payload.Transaction, err = anypb.New(transaction); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not serialize transaction payload\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not serialize transaction payload\"),\n\t\t)\n\t}\n\tif payload.Identity, err = anypb.New(identity); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not serialize identity payload\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not serialize identity payload\"),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, \"transaction and identity payload constructed\", pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Secure the envelope with the remote beneficiary's signing keys\n\tvar envelope *protocol.SecureEnvelope\n\tif envelope, err = handler.New(xfer.Envelope, payload, nil).Seal(signKey); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not create or sign secure envelope\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not create or sign secure envelope\"),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"secure envelope %s sealed: encrypted with AES-GCM and RSA - sending ...\", envelope.Id), pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Conduct the TRISA transaction, handle errors and send back to user\n\tif envelope, err = peer.Transfer(envelope); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not perform TRISA exchange\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"received %s information exchange reply from %s\", envelope.Id, peer.String()), pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Open the response envelope with local private keys\n\tvar opened *handler.Envelope\n\tif opened, err = handler.Open(envelope, s.trisa.sign); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not unseal TRISA response\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\t// Verify the contents of the response\n\tpayload = opened.Payload\n\tif payload.Identity.TypeUrl != \"type.googleapis.com/ivms101.IdentityPayload\" {\n\t\tlog.Warn().Str(\"type\", payload.Identity.TypeUrl).Msg(\"unsupported identity type\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"unsupported identity type\", payload.Identity.TypeUrl),\n\t\t)\n\t}\n\n\tif payload.Transaction.TypeUrl != \"type.googleapis.com/trisa.data.generic.v1beta1.Transaction\" {\n\t\tlog.Warn().Str(\"type\", payload.Transaction.TypeUrl).Msg(\"unsupported transaction type\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"unsupported transaction type\", payload.Transaction.TypeUrl),\n\t\t)\n\t}\n\n\tidentity = &ivms101.IdentityPayload{}\n\ttransaction = &generic.Transaction{}\n\tif err = payload.Identity.UnmarshalTo(identity); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not unmarshal identity\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\tif err = payload.Transaction.UnmarshalTo(transaction); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not unmarshal transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, \"successfully decrypted and parsed secure envelope\", pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Update the completed transaction and save to disk\n\txfer.Beneficiary = Identity{\n\t\tWalletAddress: transaction.Beneficiary,\n\t}\n\txfer.Completed = true\n\txfer.Timestamp, _ = time.Parse(time.RFC3339, transaction.Timestamp)\n\n\t// Serialize the identity information as JSON data\n\tvar data []byte\n\tif data, err = json.Marshal(identity); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not marshal IVMS 101 identity\"),\n\t\t)\n\t}\n\txfer.Identity = string(data)\n\n\tif err = s.db.Save(&xfer).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\t// Save the pending transaction on the account\n\t// TODO: remove pending transactions\n\taccount.Pending--\n\taccount.Completed++\n\taccount.Balance.Sub(xfer.Amount)\n\tif err = s.db.Save(&account).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\tmsg = fmt.Sprintf(\"transaction %04d complete: %s transfered from %s to %s\", xfer.ID, xfer.Amount.String(), xfer.Originator.WalletAddress, xfer.Beneficiary.WalletAddress)\n\ts.updates.Broadcast(req.Id, msg, pb.MessageCategory_BLOCKCHAIN)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"%04d new account balance: %s\", account.ID, account.Balance), pb.MessageCategory_LEDGER)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\trep := &pb.Message{\n\t\tType: pb.RPC_TRANSFER,\n\t\tId: req.Id,\n\t\tTimestamp: time.Now().Format(time.RFC3339),\n\t\tCategory: pb.MessageCategory_LEDGER,\n\t\tReply: &pb.Message_Transfer{Transfer: &pb.TransferReply{\n\t\t\tTransaction: xfer.Proto(),\n\t\t}},\n\t}\n\n\treturn s.updates.Send(client, rep)\n}","func (eobj emp)details(){\n\tfmt.Println(\"Employee details are \",eobj.id,\" \",eobj.name,\" \",eobj.age)\n}","func (c *TrackerController) Show(ctx *app.ShowTrackerContext) error {\n\treturn application.Transactional(c.db, func(appl application.Application) error {\n\t\tt, err := appl.Trackers().Load(ctx.Context, ctx.ID)\n\t\tif err != nil {\n\t\t\tswitch err.(type) {\n\t\t\tcase remoteworkitem.NotFoundError:\n\t\t\t\tlog.Printf(\"not found, id=%s\", ctx.ID)\n\t\t\t\treturn goa.ErrNotFound(err.Error())\n\t\t\tdefault:\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn ctx.OK(t)\n\t})\n}","func (srv *Server) walletTransactionHandler(w http.ResponseWriter, req *http.Request, ps httprouter.Params) {\n\t// Parse the id from the url.\n\tvar id types.TransactionID\n\tjsonID := \"\\\"\" + ps.ByName(\"id\") + \"\\\"\"\n\terr := id.UnmarshalJSON([]byte(jsonID))\n\tif err != nil {\n\t\twriteError(w, \"error after call to /wallet/history: \"+err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\ttxn, ok := srv.wallet.Transaction(id)\n\tif !ok {\n\t\twriteError(w, \"error when calling /wallet/transaction/$(id): transaction not found\", http.StatusBadRequest)\n\t\treturn\n\t}\n\twriteJSON(w, WalletTransactionGETid{\n\t\tTransaction: txn,\n\t})\n}","func GetTransactionHandler(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\t// retrieve the parameters\n\tparam := make(map[string]uint64)\n\tfor _, key := range []string{\"blockId\", \"txId\"} {\n\t\tparam[key], _ = strconv.ParseUint(vars[\"blockId\"], 10, 64)\n\t}\n\n\ttmp := atomic.LoadUint64(&lastBlock)\n\tif param[\"blockId\"] > tmp {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\terr := fmt.Errorf(\"requested id %d latest %d\", param[\"blockId\"], lastBlock)\n\t\tlog.Println(err.Error())\n\t\t_, _ = w.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\t// retuning anything in the body regardless of any error code\n\t// it may contain\n\t_, _, body, _ := dataCollection.GetTransaction(param[\"blockId\"], param[\"txId\"], config.DefaultRequestsTimeout)\n\twriteResponse(body, &w)\n}","func Show(w http.ResponseWriter, r *http.Request) {\n\tc := flight.Context(w, r)\n\n\titem, _, err := summary.ByID(c.DB, c.Param(\"id\"))\n\tif err != nil {\n\t\tc.FlashErrorGeneric(err)\n\t\tc.Redirect(uri)\n\t\treturn\n\t}\n\n\tv := c.View.New(\"summary/show\")\n\tv.Vars[\"item\"] = item\n\tv.Render(w, r)\n}","func show(req events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// Get the `employeeid` query string parameter from the request and\n\t// validate it.\n\temployeeid := req.QueryStringParameters[\"employeeid\"]\n\tif !employeeidRegexp.MatchString(employeeid) {\n\t\treturn clientError(http.StatusBadRequest)\n\t} //fetch a specific employee record from dynamodb in this case\n\n\t// Fetch the employee record from the database based on the employeeid value.\n\temp, err := getItem(employeeid)\n\tif err != nil {\n\t\treturn serverError(err)\n\t}\n\tif emp == nil {\n\t\treturn clientError(http.StatusNotFound)\n\t}\n\n\t// The APIGatewayProxyResponse.Body field needs to be a string, so\n\t// we marshal the employee record into JSON.\n\tjs, err := json.Marshal(emp)\n\tif err != nil {\n\t\treturn serverError(err)\n\t}\n\n\t// Return a response with a 200 OK status and the JSON employee record\n\t// as the body.\n\treturn events.APIGatewayProxyResponse{\n\t\tStatusCode: http.StatusOK,\n\t\tBody: string(js),\n\t}, nil\n}","func (_VinTracker *VinTrackerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _VinTracker.Contract.VinTrackerTransactor.contract.Transact(opts, method, params...)\n}","func (db *MySQLDB) GetTransactionByTransactionID(ctx context.Context, transactionID int) (*TransactionRecord, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"GetTransactionByTransactionID\")\n\ttransaction := &TransactionRecord{}\n\n\trow := db.instance.QueryRowContext(ctx, \"SELECT id, user_id, date, grand_total, discount, reason FROM transactions WHERE id = ?\", transactionID)\n\terr := row.Scan(&transaction.ID, &transaction.UserID, &transaction.Date, &transaction.GrandTotal, &transaction.Discount, &transaction.Reason)\n\tif err != nil {\n\t\tfLog.Errorf(\"row.Scan got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\tq := fmt.Sprintf(\"SELECT td.transaction_id, td.product_id, td.qty, td.sub_total, p.sku FROM transaction_detail td INNER JOIN products p ON td.product_id = p.id WHERE transaction_id = %v\", transactionID)\n\trows, err := db.instance.QueryContext(ctx, q)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.QueryContext got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\ttDetail := make([]*TransactionDetailRecord, 0)\n\tfor rows.Next() {\n\t\ttD := &TransactionDetailRecord{}\n\t\terr := rows.Scan(&tD.TransactionID, &tD.ProductID, &tD.Qty, &tD.SubTotal, &tD.SKU)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"rows.Scan got %s\", err.Error())\n\t\t} else {\n\t\t\ttDetail = append(tDetail, tD)\n\t\t}\n\t}\n\n\ttransaction.TransactionDetail = tDetail\n\n\treturn transaction, nil\n}","func (_DetailedERC20 *DetailedERC20Raw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.DetailedERC20Transactor.contract.Transact(opts, method, params...)\n}","func (_DetailedERC20 *DetailedERC20Raw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.DetailedERC20Transactor.contract.Transact(opts, method, params...)\n}","func TransactionIndex(c *gin.Context) {\n\trelatedObjectID := c.Query(\"relatedObjectId\")\n\trelatedObjectType := c.Query(\"relatedObjectType\")\n\tisSettledQuery := c.Query(\"isSettled\")\n\tstatusQuery := c.Query(\"status\")\n\tcurUserID := c.Keys[\"CurrentUserID\"]\n\n\tvar transactions []models.Transaction\n\n\tquery := database.DBCon\n\n\tisSettled, err := strconv.ParseBool(isSettledQuery)\n\tif isSettledQuery != \"\" && err == nil {\n\t\tquery = query.Where(\"is_settled = ?\", isSettled)\n\t}\n\n\t// TODO: Check that statusQuery is a valid status\n\tif statusQuery != \"\" {\n\t\tquery = query.Where(\"status = ?\", statusQuery)\n\t}\n\n\tif relatedObjectID != \"\" && relatedObjectType != \"\" {\n\t\tquery.\n\t\t\tWhere(\"related_object_id = ? AND related_object_type = ?\", relatedObjectID, relatedObjectType).\n\t\t\tOrder(\"created_at desc\").\n\t\t\tFind(&transactions)\n\t} else {\n\t\tquery.\n\t\t\tWhere(\"creator_id = ?\", curUserID).\n\t\t\tFind(&transactions)\n\t}\n\n\t// Get creator and relatedUser\n\t// TODO: n + 1 query problem here, so we'll figure this out later\n\tfor i := range transactions {\n\t\tdatabase.DBCon.First(&transactions[i].Recipient, transactions[i].RecipientID)\n\t\tdatabase.DBCon.First(&transactions[i].Sender, transactions[i].SenderID)\n\t\tdatabase.DBCon.First(&transactions[i].Creator, transactions[i].CreatorID)\n\t}\n\n\tdata, err := jsonapi.Marshal(transactions)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tc.Data(http.StatusOK, \"application/vnd.api+json\", data)\n}","func displayTransferDetails(xmlMessage string){\n // Replace all &quot; with single quote\n strings.ReplaceAll(xmlMessage, \"&quot;\", \"'\")\n // Create an parsed XML document\n doc, err := xmlquery.Parse(strings.NewReader(xmlMessage))\n if err != nil {\n panic(err)\n }\n\n // Get required 'transaction' element from Xml message\n transaction := xmlquery.FindOne(doc, \"//transaction\")\n if transaction != nil {\n transferId := transaction.SelectAttr(\"ID\")\n if action := transaction.SelectElement(\"action\"); action != nil {\n if strings.EqualFold(action.InnerText(),\"completed\") {\n // Process transfer complete Xml message\n var supplementMsg string\n status := transaction.SelectElement(\"status\")\n if status != nil {\n supplementMsg = status.SelectElement(\"supplement\").InnerText()\n fmt.Printf(\"\\n[%s] TransferID: %s Status: %s\\n \\tSupplement: %s\\n\",\n action.SelectAttr(\"time\"),\n strings.ToUpper(transferId),\n action.InnerText(),\n supplementMsg)\n }\n\n destAgent := transaction.SelectElement(\"destinationAgent\")\n statistics := transaction.SelectElement(\"statistics\")\n // Retrieve statistics\n var actualStartTimeText = \"\"\n var retryCount string\n var numFileFailures string\n var numFileWarnings string\n if statistics != nil {\n actualStartTime := statistics.SelectElement(\"actualStartTime\")\n if actualStartTime != nil {\n actualStartTimeText = actualStartTime.InnerText()\n }\n if statistics.SelectElement(\"retryCount\") != nil {\n retryCount = statistics.SelectElement(\"retryCount\").InnerText()\n }\n if statistics.SelectElement(\"numFileFailures\") != nil {\n numFileFailures = statistics.SelectElement(\"numFileFailures\").InnerText()\n }\n if statistics.SelectElement(\"numFileWarnings\") != nil {\n numFileWarnings = statistics.SelectElement(\"numFileWarnings\").InnerText()\n }\n }\n var elapsedTime time.Duration\n if actualStartTimeText != \"\" {\n startTime := getFormattedTime(actualStartTimeText)\n completePublishTIme := getFormattedTime(action.SelectAttr(\"time\"))\n elapsedTime = completePublishTIme.Sub(startTime)\n }\n\n fmt.Printf(\"\\tDestination Agent: %s\\n\\tStart time: %s\\n\\tCompletion Time: %s\\n\\tElapsed time: %s\\n\\tRetry Count: %s\\n\\tFailures:%s\\n\\tWarnings:%s\\n\\n\",\n destAgent.SelectAttr(\"agent\"),\n actualStartTimeText,\n action.SelectAttr(\"time\"),\n elapsedTime,\n retryCount,\n numFileFailures,\n numFileWarnings)\n } else if strings.EqualFold(action.InnerText(),\"progress\") {\n // Process transfer progress Xml message\n destAgent := transaction.SelectElement(\"destinationAgent\")\n progressPublishTimeText := action.SelectAttr(\"time\")\n fmt.Printf(\"\\n[%s] %s Status: %s Destination: %s \\n\", progressPublishTimeText,\n strings.ToUpper(transferId),\n action.InnerText(),\n destAgent.SelectAttr(\"agent\"))\n transferSet := transaction.SelectElement(\"transferSet\")\n startTimeText := transferSet.SelectAttr(\"startTime\")\n //startTime := getFormattedTime(startTimeText)\n //progressPublishTime := getFormattedTime(progressPublishTimeText)\n //elapsedTime := progressPublishTime.Sub(startTime)\n fmt.Printf(\"\\tStart time: %s\\n\\tTotal items in transfer request: %s\\n\\tBytes sent: %s\\n\",\n startTimeText,\n transferSet.SelectAttr(\"total\"),\n transferSet.SelectAttr(\"bytesSent\"))\n\n // Loop through all items in the progress message and display details.\n items := transferSet.SelectElements(\"item\")\n for i := 0 ; i < len(items); i++ {\n status := items[i].SelectElement(\"status\")\n resultCode := status.SelectAttr(\"resultCode\")\n var sourceName string\n var sourceSize = \"-1\"\n queueSource := items[i].SelectElement(\"source/queue\")\n if queueSource != nil {\n sourceName = queueSource.InnerText()\n } else {\n fileName := items[i].SelectElement(\"source/file\")\n if fileName != nil {\n sourceName = fileName.InnerText()\n sourceSize = fileName.SelectAttr(\"size\")\n }\n }\n\n var destinationName string\n queueDest := items[i].SelectElement(\"destination/queue\")\n var destinationSize = \"-1\"\n if queueDest != nil {\n destinationName = queueDest.InnerText()\n } else {\n fileName := items[i].SelectElement(\"destination/file\")\n if fileName != nil {\n destinationName = fileName.InnerText()\n destinationSize = fileName.SelectAttr(\"size\")\n }\n }\n\n // Display details of each item\n fmt.Printf(\"\\tItem # %d\\n\\t\\tSource: %s\\tSize: %s bytes\\n\\t\\tDestination: %s\\tSize: %s bytes\\n\",\n i+1,\n sourceName, sourceSize,\n destinationName, destinationSize)\n // Process result code and append any supplement\n if resultCode != \"0\" {\n supplement := status.SelectElement(\"supplement\")\n if supplement != nil {\n fmt.Printf(\"\\t\\tResult code %s Supplement %s\\n\", resultCode, supplement.InnerText())\n } else {\n fmt.Printf(\"\\t\\tResult code %s\\n\", resultCode)\n }\n } else {\n fmt.Printf(\"\\t\\tResult code %s\\n\", resultCode)\n }\n }\n } else if strings.EqualFold(action.InnerText(),\"started\") {\n // Process transfer started Xml message\n destAgent := transaction.SelectElement(\"destinationAgent\")\n destinationAgentName := destAgent.SelectAttr(\"agent\")\n transferSet := transaction.SelectElement(\"transferSet\")\n startTime := \"\"\n if transferSet != nil {\n startTime = transferSet.SelectAttr(\"startTime\")\n } else {\n startTime = action.SelectAttr(\"time\")\n }\n fmt.Printf(\"[%s] TransferID: %s Status: %s Destination: %s\\n\",\n startTime,\n strings.ToUpper(transferId),\n action.InnerText(),\n destinationAgentName)\n }\n }\n }\n}","func Show(w http.ResponseWriter, r *http.Request) {\r\n\tdb := dbConn()\r\n\tnId := r.URL.Query().Get(\"id\")\r\n\tselDB, err := db.Query(\"SELECT * FROM Pegawai WHERE id=?\", nId)\r\n\tif err != nil {\r\n\t\tpanic(err.Error())\r\n\t}\r\n\temp := Pegawai{}\r\n\tfor selDB.Next() {\r\n\t\t//buat variabel untuk menampung data\r\n\t\t//sesuaikan sama nama kolom database (huruf kecil)\r\n\t\tvar id int\r\n\t\tvar nama, alamat, jabatan string\r\n\r\n\t\terr = selDB.Scan(&id, &nama, &alamat, &jabatan)\r\n\t\tif err != nil {\r\n\t\t\tpanic(err.Error())\r\n\t\t}\r\n\r\n\t\t//kanan nama var struct - kiri nama kolom database yang diinisialisasikan diatas\r\n\t\temp.Id = id\r\n\t\temp.Nama = nama\r\n\t\temp.Alamat = alamat\r\n\t\temp.Jabatan = jabatan\r\n\t}\r\n\ttmpl.ExecuteTemplate(w, \"Show\", emp)\r\n\tdefer db.Close()\r\n}","func (t *Transaction) String() string {\n\treturn t.From + \" -> \" + t.To + \" : \" + strconv.Itoa(t.Amount)\n}","func (T transaction) String() string {\n\treturn fmt.Sprintf(\"{\\n\\t\\t\\tsender:%v,\\n\\t\\t\\treceiver:%v,\\n\\t\\t\\tamount:%v\\n\\t\\t}\", T.sender, T.receiver, T.amount)\n}","func (t *Transaction) String() string {\n\tswitch t.Type {\n\tcase TIN, TEP:\n\t\treturn fmt.Sprintf(\"%22s {%d} %s %.3f %s [%s]\", t.Datetime, t.Index, t.Category.FullName, t.Amount, t.AmountCurrency, t.Type.Name)\n\tcase TOB:\n\t\treturn fmt.Sprintf(\"%22s {%d} Opening new Account\", t.Datetime, t.Index)\n\tcase TBJ:\n\t\treturn fmt.Sprintf(\"%22s {%d} Update balance by %.3f %s\", t.Datetime, t.Index, t.Amount, t.AmountCurrency)\n\tcase TMT:\n\t\treturn fmt.Sprintf(\"%22s {%d} Move %.3f %s to '%s' (%.3f %s)\", t.Datetime, t.Index, t.Amount, t.AmountCurrency, t.AccountTo, t.AmountTo, t.AmountToCurrency)\n\tdefault:\n\t\treturn fmt.Sprintf(\"%22s {%d} [%s] TODO: DIDN'T IMPLEMENT THIS TYPE YET\", t.Datetime, t.Index, t.Type.Name)\n\t}\n}","func (_CommitteeManager *CommitteeManagerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _CommitteeManager.Contract.CommitteeManagerTransactor.contract.Transact(opts, method, params...)\n}","func ShowTxStatusTracker(stdout io.Writer, hash string, rpcClient types2.Client) error {\n\ts := spinner.New(spinner.CharSets[14], 100*time.Millisecond)\n\ts.Writer = stdout\n\ts.Prefix = \" \"\n\ts.Start()\n\tlastStatus := \"\"\n\n\tvar err error\n\tvar resp *api2.ResultTx\n\tattempts := 0\n\tfor {\n\t\tattempts += 1\n\t\tif attempts == 3 {\n\t\t\treturn err\n\t\t}\n\n\t\ttime.Sleep(1 * time.Second)\n\t\tresp, err = api.GetTransaction(hash, rpcClient)\n\t\tif err != nil {\n\t\t\ts.Stop()\n\t\t\tcontinue\n\t\t}\n\n\t\tif lastStatus == resp.Status {\n\t\t\tcontinue\n\t\t}\n\n\t\tlastStatus = resp.Status\n\t\tif resp.Status == types3.TxStatusInMempool {\n\t\t\ts.Suffix = colorfmt.YellowStringf(\" In mempool\")\n\t\t} else if resp.Status == types3.TxStatusInPushpool {\n\t\t\ts.Suffix = colorfmt.YellowStringf(\" In pushpool\")\n\t\t} else {\n\t\t\ts.FinalMSG = colorfmt.GreenString(\" Confirmed!\\n\")\n\t\t\ts.Stop()\n\t\t\tbreak\n\t\t}\n\t}\n\treturn nil\n}","func (res BackTestResult) Show() TradestatPort {\n\t//\tp := NewPortfolio()\n\t// mds := bean.NewRPCMDSConnC(\"tcp\", res.dbhost+\":\"+res.dbport)\n\tratesbook := make(ReferenceRateBook)\n\n\t// FIXME: think about how to show multi pair result\n\tvar stat TradestatPort\n\tif len(res.pairs) > 0 {\n\t\tp := res.pairs[0]\n\t\t//\t\ttxn, _ := mds.GetTransactions2(NameFcoin, p, res.start, res.end)\n\t\tratesbook[p] = RefRatesFromTxn(res.Txn)\n\t\t//\t\tsnapts.Print()\n\t\t//\t\tperfts.Print()\n\n\t\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\n\t\tssTS := GenerateSnapshotTS(res.Txn, NewPortfolio())\n\t\tmaxPos := 0.0\n\t\tmaxNeg := 0.0\n\t\tfor _, ss := range ssTS {\n\t\t\tmaxPos = math.Max(maxPos, ss.Port.Balance(p.Coin))\n\t\t\tmaxNeg = math.Min(maxNeg, ss.Port.Balance(p.Coin))\n\t\t}\n\t\tfmt.Println(\"max position\", maxPos, maxNeg)\n\t\tstat.Print()\n\t}\n\treturn stat\n}","func (ok OKHandler) DeliverTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.DeliverResult, err error) {\n\treturn sdk.DeliverResult{Log: ok.Log}, nil\n}","func Index(txn *cheshire.Txn) {\n\t//create a context map to be passed to the template\n\tcontext := make(map[string]interface{})\n\tcontext[\"services\"] = Servs.RouterTables()\n\tcheshire.RenderInLayout(txn, \"/index.html\", \"/template.html\", context)\n}","func (c *dummyWavesMDLrpcclient) GetTransaction(txid string) (*model.Transactions, error) {\n\ttransaction, _, err := client.NewTransactionsService(c.MainNET).GetTransactionsInfoID(txid)\n\treturn transaction, err\n}","func (_VinTracker *VinTrackerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _VinTracker.Contract.contract.Transact(opts, method, params...)\n}","func (action *LedgerShowAction) JSON() {\n\tquery := action.Query()\n\n\tif action.Err != nil {\n\t\treturn\n\t}\n\n\taction.Err = db.Get(action.Ctx, query, &action.Record)\n\n\tif action.Err != nil {\n\t\treturn\n\t}\n\n\thal.Render(action.W, NewLedgerResource(action.Record))\n}","func (_CommitteeManager *CommitteeManagerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _CommitteeManager.Contract.contract.Transact(opts, method, params...)\n}","func (t *Transaction) String() string {\n\tvar builder strings.Builder\n\tbuilder.WriteString(\"Transaction(\")\n\tbuilder.WriteString(fmt.Sprintf(\"id=%v\", t.ID))\n\tbuilder.WriteString(\", sequenceInDay=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.SequenceInDay))\n\tbuilder.WriteString(\", status=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Status))\n\tbuilder.WriteString(\", executedAmount=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.ExecutedAmount))\n\tbuilder.WriteString(\", executedCurrencyCode=\")\n\tbuilder.WriteString(t.ExecutedCurrencyCode)\n\tbuilder.WriteString(\", exchangeRate=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.ExchangeRate))\n\tbuilder.WriteString(\", originatingAmount=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.OriginatingAmount))\n\tbuilder.WriteString(\", originatingCurrencyCode=\")\n\tbuilder.WriteString(t.OriginatingCurrencyCode)\n\tbuilder.WriteString(\", direction=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Direction))\n\tbuilder.WriteString(\", runningBalance=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.RunningBalance))\n\tbuilder.WriteString(\", createdDate=\")\n\tbuilder.WriteString(t.CreatedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", postedDate=\")\n\tbuilder.WriteString(t.PostedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", executedDate=\")\n\tbuilder.WriteString(t.ExecutedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", updatedDate=\")\n\tbuilder.WriteString(t.UpdatedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", description=\")\n\tbuilder.WriteString(t.Description)\n\tbuilder.WriteString(\", memo=\")\n\tbuilder.WriteString(t.Memo)\n\tbuilder.WriteString(\", group=\")\n\tbuilder.WriteString(t.Group)\n\tbuilder.WriteString(\", type=\")\n\tbuilder.WriteString(t.Type)\n\tbuilder.WriteString(\", mainCategory=\")\n\tbuilder.WriteString(t.MainCategory)\n\tbuilder.WriteString(\", subCategory=\")\n\tbuilder.WriteString(t.SubCategory)\n\tbuilder.WriteString(\", checkNumber=\")\n\tbuilder.WriteString(t.CheckNumber)\n\tbuilder.WriteString(\", latitude=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Latitude))\n\tbuilder.WriteString(\", longitude=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Longitude))\n\tbuilder.WriteString(\", merchantCode=\")\n\tbuilder.WriteString(t.MerchantCode)\n\tbuilder.WriteString(\", reversal=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Reversal))\n\tbuilder.WriteString(\", reversalFor=\")\n\tbuilder.WriteString(t.ReversalFor)\n\tbuilder.WriteString(\", reversed=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Reversed))\n\tbuilder.WriteString(\", reversedBy=\")\n\tbuilder.WriteString(t.ReversedBy)\n\tbuilder.WriteString(\", url=\")\n\tbuilder.WriteString(t.URL)\n\tbuilder.WriteByte(')')\n\treturn builder.String()\n}","func (controller *AccountController) DoTransaction(ctx *gin.Context) {\n\tsourceID, ok := ctx.GetPostForm(\"sourceid\")\n\tif !ok {\n\t\tlog.WithFields(log.Fields{\"URL\": ctx.Request.URL.String()}).Warn(\"No SourceID found in postform\")\n\n\t\terrResp, _ := restapi.NewErrorResponse(\"No sourceID given\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(errResp))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\ttargetID, ok := ctx.GetPostForm(\"targetid\")\n\tif !ok {\n\t\tlog.WithFields(log.Fields{\"URL\": ctx.Request.URL.String()}).Warn(\"No TargetID found in postform\")\n\n\t\terrResp, _ := restapi.NewErrorResponse(\"No targetID given\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(errResp))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tamount, err := strconv.Atoi(ctx.PostForm(\"amount\"))\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\"URL\": ctx.Request.URL.String()}).Warn(\"No int amount found in postform\")\n\n\t\tresponse, _ := restapi.NewErrorResponse(\"No valid diff value\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tinfo, err := authStuff.GetLoginInfoFromCtx(ctx)\n\tif err != nil {\n\t\tresponse, _ := restapi.NewErrorResponse(err.Error()).Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tif err := controller.service.Transaction(sourceID, targetID, info.Name, amount); err == nil {\n\t\tresponse, _ := restapi.NewOkResponse(\"\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Next()\n\t} else {\n\t\tlog.WithFields(log.Fields{\"user\": info.Name}).WithError(err).Error(\"Transaction Error\")\n\n\t\tresponse, _ := restapi.NewErrorResponse(err.Error()).Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n}","func (_TableManager *TableManagerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _TableManager.Contract.TableManagerTransactor.contract.Transact(opts, method, params...)\n}","func (_TableManager *TableManagerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _TableManager.Contract.contract.Transact(opts, method, params...)\n}","func (_ArbGasInfo *ArbGasInfoRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _ArbGasInfo.Contract.ArbGasInfoTransactor.contract.Transact(opts, method, params...)\n}","func (_DetailedERC20 *DetailedERC20TransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.contract.Transact(opts, method, params...)\n}","func (_DetailedERC20 *DetailedERC20TransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.contract.Transact(opts, method, params...)\n}","func TransactionScreen(w fyne.Window) fyne.CanvasObject {\n\treturn widget.NewTabContainer(\n\t\twidget.NewTabItem(res.GetLocalString(\"Transfer\"), makeTransferTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Move\"), makeMoveTransTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Vote\"), makeVoteTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Unvote\"), makeUnvoteTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Register Miner\"), makeMinerTab(w)),\n\t)\n}","func ShowTask(w http.ResponseWriter, r *http.Request, repo *tasks.TaskRepository) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tparams := mux.Vars(r)\n\ttaskID, err := strconv.Atoi(params[\"id\"])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ttask, err := repo.GetTask(taskID)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tjson.NewEncoder(w).Encode(apiIndexTask(task))\n}","func (api *DeprecatedApiService) getTransactionIntro(params map[string]string) map[string]string {\n\tresult := make(map[string]string)\n\n\t_, is_only_check_exist := params[\"only_check_exist\"]\n\n\ttrsid, ok1 := params[\"id\"]\n\tif !ok1 {\n\t\tresult[\"err\"] = \"param id must.\"\n\t\treturn result\n\t}\n\n\tvar trshx []byte\n\tif txhx, e := hex.DecodeString(trsid); e == nil && len(txhx) == 32 {\n\t\ttrshx = txhx\n\t} else {\n\t\tresult[\"err\"] = \"transaction hash error.\"\n\t\treturn result\n\t}\n\n\t// Query transaction\n\tblkhei, trsresbytes, err := api.blockchain.GetChainEngineKernel().StateRead().ReadTransactionBytesByHash(trshx)\n\tif err != nil {\n\t\tresult[\"err\"] = err.Error()\n\t\treturn result\n\t}\n\tif trsresbytes == nil {\n\t\tresult[\"err\"] = \"transaction not fond.\"\n\t\treturn result\n\t}\n\n\t// Whether to just judge whether it exists\n\tif is_only_check_exist && len(trsresbytes) > 0 {\n\t\tresult[\"ret\"] = \"0\"\n\t\tresult[\"exist\"] = \"yes\"\n\t\treturn result\n\t}\n\n\ttrsres, _, err := transactions.ParseTransaction(trsresbytes, 0)\n\tif err != nil {\n\t\tresult[\"err\"] = err.Error()\n\t\treturn result\n\t}\n\n\t// Resolve actions\n\tvar allactions = trsres.GetActionList()\n\tvar actions_ary []string\n\tvar actions_strings = \"\"\n\tfor _, act := range allactions {\n\t\tvar kind = act.Kind()\n\t\tactstr := fmt.Sprintf(`{\"k\":%d`, kind)\n\t\tif kind == 1 {\n\t\t\tacc := act.(*actions.Action_1_SimpleToTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"to\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 13 {\n\t\t\tacc := act.(*actions.Action_13_FromTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 14 {\n\t\t\tacc := act.(*actions.Action_14_FromToTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"to\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 2 {\n\t\t\tacc := act.(*actions.Action_2_OpenPaymentChannel)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"left_addr\":\"%s\",\"left_amt\":\"%s\",\"right_addr\":\"%s\",\"right_amt\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t\tacc.LeftAddress.ToReadable(),\n\t\t\t\tacc.LeftAmount.ToFinString(),\n\t\t\t\tacc.RightAddress.ToReadable(),\n\t\t\t\tacc.RightAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 3 {\n\t\t\tacc := act.(*actions.Action_3_ClosePaymentChannel)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 4 {\n\t\t\tacc := act.(*actions.Action_4_DiamondCreate)\n\t\t\tactstr += fmt.Sprintf(`,\"number\":\"%d\",\"name\":\"%s\",\"address\":\"%s\"`,\n\t\t\t\tacc.Number,\n\t\t\t\tacc.Diamond,\n\t\t\t\tacc.Address.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 5 {\n\t\t\tacc := act.(*actions.Action_5_DiamondTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"count\":1,\"names\":\"%s\",\"from\":\"%s\",\"to\":\"%s\"`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Diamond,\n\t\t\t\ttrsres.GetAddress().ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 6 {\n\t\t\tacc := act.(*actions.Action_6_OutfeeQuantityDiamondTransfer)\n\t\t\tdmds := make([]string, len(acc.DiamondList.Diamonds))\n\t\t\tfor i, v := range acc.DiamondList.Diamonds {\n\t\t\t\tdmds[i] = string(v)\n\t\t\t}\n\t\t\tactstr += fmt.Sprintf(`,\"count\":%d,\"names\":\"%s\",\"from\":\"%s\",\"to\":\"%s\"`,\n\t\t\t\tacc.DiamondList.Count,\n\t\t\t\tstrings.Join(dmds, \",\"),\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 7 {\n\t\t\tacc := act.(*actions.Action_7_SatoshiGenesis)\n\t\t\tactstr += fmt.Sprintf(`,\"trs_no\":%d,\"btc_num\":%d,\"hac_subsidy\":%d,\"address\":\"%s\",\"lockbls_id\":\"%s\"`,\n\t\t\t\tacc.TransferNo,\n\t\t\t\tacc.BitcoinQuantity,\n\t\t\t\tacc.AdditionalTotalHacAmount,\n\t\t\t\tacc.OriginAddress.ToReadable(),\n\t\t\t\thex.EncodeToString(actions.GainLockblsIdByBtcMove(uint32(acc.TransferNo))),\n\t\t\t)\n\t\t} else if kind == 8 {\n\t\t\tacc := act.(*actions.Action_8_SimpleSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"to\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 9 {\n\t\t\tacc := act.(*actions.Action_9_LockblsCreate)\n\t\t\tactstr += fmt.Sprintf(`,\"lockbls_id\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.LockblsId),\n\t\t\t\tacc.TotalStockAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 10 {\n\t\t\tacc := act.(*actions.Action_10_LockblsRelease)\n\t\t\tactstr += fmt.Sprintf(`,\"lockbls_id\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.LockblsId),\n\t\t\t\tacc.ReleaseAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 11 {\n\t\t\tacc := act.(*actions.Action_11_FromToSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"to\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 28 {\n\t\t\tacc := act.(*actions.Action_28_FromSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 12 {\n\t\t\tacc := act.(*actions.Action_12_ClosePaymentChannelBySetupAmount)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 21 {\n\t\t\tacc := act.(*actions.Action_21_ClosePaymentChannelBySetupOnlyLeftAmount)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 22 {\n\t\t\tacc := act.(*actions.Action_22_UnilateralClosePaymentChannelByNothing)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\"\":\"%s\",\"bill_number\"\":0`,\n\t\t\t\thex.EncodeToString(acc.ChannelId), acc.AssertCloseAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 23 {\n\t\t\tacc := act.(*actions.Action_23_UnilateralCloseOrRespondChallengePaymentChannelByRealtimeReconciliation)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"%s\",\"bill_number\":%d`,\n\t\t\t\thex.EncodeToString(acc.Reconciliation.GetChannelId()), acc.AssertAddress.ToReadable(), acc.Reconciliation.GetAutoNumber(),\n\t\t\t)\n\t\t} else if kind == 24 {\n\t\t\tacc := act.(*actions.Action_24_UnilateralCloseOrRespondChallengePaymentChannelByChannelChainTransferBody)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"%s\",\"bill_number\":%d`,\n\t\t\t\thex.EncodeToString(acc.ChannelChainTransferTargetProveBody.GetChannelId()), acc.AssertAddress.ToReadable(), acc.ChannelChainTransferTargetProveBody.GetAutoNumber(),\n\t\t\t)\n\t\t} else if kind == 27 {\n\t\t\tacc := act.(*actions.Action_27_ClosePaymentChannelByClaimDistribution)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"any\",\"bill_number\"\":\"closed\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t}\n\t\tactstr += \"}\"\n\t\tactions_ary = append(actions_ary, actstr)\n\t}\n\tactions_strings = strings.Join(actions_ary, \",\")\n\n\t// Transaction return data\n\ttxaddr := fields.Address(trsres.GetAddress())\n\tvar txfee = trsres.GetFee()\n\tvar txfeeminergot = trsres.GetFeeOfMinerRealReceived()\n\tresult[\"jsondata\"] = fmt.Sprintf(\n\t\t`{\"block\":{\"height\":%d,\"timestamp\":%d},\"type\":%d,\"address\":\"%s\",\"fee\":\"%s\",\"feeminergot\":\"%s\",\"timestamp\":%d,\"actioncount\":%d,\"actions\":[%s]`,\n\t\tblkhei,\n\t\ttrsres.GetTimestamp(),\n\t\ttrsres.Type(),\n\t\ttxaddr.ToReadable(), // Primary address\n\t\ttxfee.ToFinString(),\n\t\ttxfeeminergot.ToFinString(),\n\t\ttrsres.GetTimestamp(),\n\t\tlen(allactions),\n\t\tactions_strings,\n\t)\n\n\tif _, ok := params[\"txbodyhex\"]; ok {\n\t\tresult[\"jsondata\"] += fmt.Sprintf(`,\"txbodyhex\":\"%s\"`,\n\t\t\thex.EncodeToString(trsresbytes))\n\t}\n\n\t// Wrap up and return\n\tresult[\"jsondata\"] += \"}\"\n\treturn result\n}","func TransactionHandler(w http.ResponseWriter, r *http.Request) {\n\taction := r.URL.Path[len(\"/api/transactions\"):]\n\n\tlog.Println(\"Handling method\", r.Method, \"with action\", action)\n\n\tswitch r.Method {\n\tcase \"POST\":\n\t\tswitch action {\n\t\tcase \"\": // Create new transaction\n\t\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorForm, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar t TransactionResponse\n\n\t\t\terr = json.Unmarshal(body, &t)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpass := []byte(t.Password)\n\n\t\t\ttx, e := createAndBroadcastTx(t.Recipient, *big.NewInt(t.Amount), pass)\n\n\t\t\tvar jsonResponse TransactionResponse\n\t\t\tif e == nil {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseOk, Hash: hex.EncodeToString(tx.Hash())}\n\t\t\t} else {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseFailed, ErrorText: e.Error()}\n\t\t\t}\n\n\t\t\tres, err := json.Marshal(jsonResponse)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfmt.Fprintf(w, string(res))\n\t\tdefault:\n\t\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\t\t}\n\tcase \"GET\":\n\t\tswitch action {\n\t\tcase \"\":\n\t\t\tvar txs []TransactionJson\n\t\t\tfor _, tx := range Config.DeserializedTxs {\n\t\t\t\ttxs = append(txs, EncodeToFriendlyStruct(tx))\n\n\t\t\t}\n\n\t\t\tif len(txs) == 0 {\n\t\t\t\tfmt.Fprintf(w, string(\"[]\"))\n\t\t\t} else {\n\n\t\t\t\tres, err := json.Marshal(txs)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Println(\"Nope\", err.Error())\n\t\t\t\t} else {\n\t\t\t\t\tfmt.Fprintf(w, string(res))\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\tdefault:\n\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\n\t}\n}","func (action *TransactionIndexAction) JSON() {\n\taction.Do(\n\t\taction.EnsureHistoryFreshness,\n\t\taction.loadParams,\n\t\taction.checkAllowed,\n\t\taction.ValidateCursorWithinHistory,\n\t\taction.loadRecords,\n\t\taction.loadPage,\n\t\tfunc() {\n\t\t\thal.Render(action.W, action.Page)\n\t\t},\n\t)\n}","func (EchoHandler) DeliverTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.DeliverResult, err error) {\n\tdata, err := data.ToWire(tx)\n\treturn sdk.DeliverResult{Data: data}, err\n}","func (s *Service) HandleShow(w http.ResponseWriter, r *http.Request) {\n\tsubs, err := s.subscriptionRepository.FindOne(\n\t\tr.Context(), s.getResourceID(r), s.getSubscriptionID(r),\n\t)\n\tif err != nil {\n\t\tstatus := http.StatusInternalServerError\n\t\tif errRepo, ok := err.(flare.SubscriptionRepositoryError); ok && errRepo.NotFound() {\n\t\t\tstatus = http.StatusNotFound\n\t\t}\n\n\t\ts.writer.Error(w, \"error during subscription search\", err, status)\n\t\treturn\n\t}\n\n\ts.writer.Response(w, transformSubscription(subs), http.StatusOK, nil)\n}","func (me TSAFPTTransactionID) String() string { return xsdt.String(me).String() }","func (u *UseCase) ShowOrder(w http.ResponseWriter, r *http.Request) {\n\n\tmerchant := r.Header.Get(\"merchant_id\")\n\tuuid := mux.Vars(r)[\"order_id\"]\n\n\torder, err := cache.ShowOrder(merchant, uuid)\n\tif err == nil && order != nil {\n\t\trespondWithJSON(w, http.StatusOK, order)\n\t\treturn\n\t}\n\n\tvar dbOrders models.OrderPg\n\tu.DB.Conn.Table(\"orders\").Where(\"uuid = ?\", uuid).First(&dbOrders)\n\n\trespondWithJSON(w, http.StatusOK, dbOrders.Payload)\n}","func tableInfoShow(meta *halproto.TableMetadata) {\n\tswitch meta.GetKind() {\n\tcase halproto.TableKind_TABLE_INDEX:\n\t\ttableInfoIndexShow(meta)\n\tcase halproto.TableKind_TABLE_TCAM:\n\t\ttableInfoTcamShow(meta)\n\tcase halproto.TableKind_TABLE_HASH:\n\t\ttableInfoHashShow(meta)\n\tcase halproto.TableKind_TABLE_FLOW:\n\t\ttableInfoFlowShow(meta)\n\tcase halproto.TableKind_TABLE_MET:\n\t\ttableInfoMetShow(meta)\n\tdefault:\n\t\tfmt.Printf(\"Invalid table type: %d\", meta.GetKind())\n\t}\n}","func (r Virtual_Guest) GetActiveTransaction() (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getActiveTransaction\", nil, &r.Options, &resp)\n\treturn\n}","func (t *Trade) View(c echo.Context, id int) (*gorsk.Trade, error) {\n\treturn t.tdb.View(t.db, id)\n}","func GetTransaction() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tsugar, _ := item.New(\"Sugar\", map[string]float64{\"Kabras\": 110, \"Mumias\": 110}, \"kg(s)\")\n\t\tpurchase, message, err := transaction.New(sugar, map[string]float64{\"Nzoia\": 150}, 3)\n\t\tc.JSON(\n\t\t\thttp.StatusOK,\n\t\t\tGetResponse{GetData{purchase}, message, responseerr.GetStrErr(err)},\n\t\t)\n\t}\n}","func (as *AddrServer) HandleTxGet(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\ttxid := mux.Vars(r)[\"txid\"]\n\n\t// paginate through transactions\n\ttxns, err := as.GetRawTransaction(txid)\n\tif err != nil {\n\t\tw.WriteHeader(400)\n\t\tw.Write(NewPostError(\"error fetching all transactions for address\", err))\n\t\treturn\n\t}\n\tout, _ := json.Marshal(txns.Result)\n\tw.Write(out)\n}","func (t *trip) Show(ctx context.Context, id int32) (*model.Trip, error) {\n\trow := t.conn.QueryRow(ctx, `SELECT \n\t\t\t\ttrips.id, trips.dates, trips.price, origin.name, destination.name\n\t\t\t\tFROM trips \n\t\t\t\tINNER JOIN cities AS origin ON trips.origin_id = origin.id\n\t\t\t\tINNER JOIN cities AS destination ON trips.destination_id = destination.id\n\t\t\t\tWHERE trips.id = $1`, id)\n\n\tvar trip model.Trip\n\n\tvar origin, destination model.City\n\n\tif err := row.Scan(&trip.ID, &trip.Dates, &trip.Price, &origin.Name, &destination.Name); err != nil {\n\t\treturn nil, err\n\t}\n\n\ttrip.Origin = &origin\n\ttrip.Destination = &destination\n\n\treturn &trip, nil\n}","func (core *coreService) TraceTransaction(ctx context.Context, actHash string, config *logger.Config) ([]byte, *action.Receipt, *logger.StructLogger, error) {\n\tactInfo, err := core.Action(util.Remove0xPrefix(actHash), false)\n\tif err != nil {\n\t\treturn nil, nil, nil, err\n\t}\n\tact, err := (&action.Deserializer{}).SetEvmNetworkID(core.EVMNetworkID()).ActionToSealedEnvelope(actInfo.Action)\n\tif err != nil {\n\t\treturn nil, nil, nil, err\n\t}\n\tsc, ok := act.Action().(*action.Execution)\n\tif !ok {\n\t\treturn nil, nil, nil, errors.New(\"the type of action is not supported\")\n\t}\n\ttraces := logger.NewStructLogger(config)\n\tctx = protocol.WithVMConfigCtx(ctx, vm.Config{\n\t\tDebug: true,\n\t\tTracer: traces,\n\t\tNoBaseFee: true,\n\t})\n\taddr, _ := address.FromString(address.ZeroAddress)\n\tretval, receipt, err := core.SimulateExecution(ctx, addr, sc)\n\treturn retval, receipt, traces, err\n}","func (c *BulkDeletesController) Show(ctx *gin.Context) {\n\tid := ctx.Param(\"taskID\")\n\ttask := models.BulkDeleteRunTask{}\n\n\tif err := c.App.GetStore().One(\"ID\", id, &task); err == storm.ErrNotFound {\n\t\tctx.AbortWithError(404, errors.New(\"Bulk delete task not found\"))\n\t} else if err != nil {\n\t\tctx.AbortWithError(500, err)\n\t} else if doc, err := jsonapi.Marshal(&task); err != nil {\n\t\tctx.AbortWithError(500, err)\n\t} else {\n\t\tctx.Data(200, MediaType, doc)\n\t}\n}","func (gw *Gateway) GetTransaction(txid cipher.SHA256) (*visor.Transaction, error) {\n\tvar txn *visor.Transaction\n\tvar err error\n\n\tgw.strand(\"GetTransaction\", func() {\n\t\ttxn, err = gw.v.GetTransaction(txid)\n\t})\n\n\treturn txn, err\n}","func (_ArbGasInfo *ArbGasInfoTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _ArbGasInfo.Contract.contract.Transact(opts, method, params...)\n}","func (t *Transaction) DisplayInline() string {\n\treturn fmt.Sprintf(\"%s - - Operation: %s - Status: %s - Side: %s - Amount(cts): %d\", t.ID, t.OperationType, t.Status, t.Side, t.AmountCents)\n}","func tableInfoHashShow(meta *halproto.TableMetadata) {\n\thMeta := meta.GetHashMeta()\n\tinsStr := fmt.Sprintf(\"%d[%d]\", hMeta.GetNumInserts(), hMeta.GetNumInsertFailures())\n\tupdStr := fmt.Sprintf(\"%d[%d]\", hMeta.GetNumUpdates(), hMeta.GetNumUpdateFailures())\n\tdelStr := fmt.Sprintf(\"%d[%d]\", hMeta.GetNumDeletes(), hMeta.GetNumDeleteFailures())\n\tfmt.Printf(\"%-30s%-10d%-10s%-10d%-10d%-10d%-10d%-10s%-10s%-10s\\n\",\n\t\tmeta.GetTableName(),\n\t\tmeta.GetTableId(),\n\t\tutils.TableKindToStr(meta.GetKind()),\n\t\thMeta.GetCapacity(),\n\t\thMeta.GetOtcamCapacity(),\n\t\thMeta.GetHashUsage(),\n\t\thMeta.GetOtcamUsage(),\n\t\tinsStr,\n\t\tupdStr,\n\t\tdelStr)\n}","func Details(attrs []htmlgo.Attribute, children ...HTML) HTML {\n\treturn &htmlgo.Tree{Tag: \"details\", Attributes: attrs, Children: children}\n}","func (db *DB) View(fn func(*Tx) error) error {\n\tt, err := db.Begin(false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Make sure the transaction rolls back in the event of a panic.\n\tdefer func() {\n\t\tif t.db != nil {\n\t\t\tt.rollback()\n\t\t}\n\t}()\n\n\t// Mark as a managed tx so that the inner function cannot manually rollback.\n\tt.managed = true\n\n\t// If an error is returned from the function then pass it through.\n\terr = fn(t)\n\tt.managed = false\n\tif err != nil {\n\t\t_ = t.Rollback()\n\t\treturn err\n\t}\n\n\tif err := t.Rollback(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}","func (HTTPOperation) GetDetails() (string, string, string) {\n\treturn \"update\", \"updated\", \"vaccine availability \" + id\n}","func (tc *TransactionsController) Index(c *gin.Context, size, page, offset int) {\n\ttxs, count, err := tc.App.GetStore().Transactions(offset, size)\n\tptxs := make([]presenters.Tx, len(txs))\n\tfor i, tx := range txs {\n\t\tptxs[i] = presenters.NewTx(&tx)\n\t}\n\tpaginatedResponse(c, \"Transactions\", size, page, ptxs, count, err)\n}","func (_Onesplitaudit *OnesplitauditTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _Onesplitaudit.Contract.contract.Transact(opts, method, params...)\n}","func tableInfoIndexShow(meta *halproto.TableMetadata) {\n\tiMeta := meta.GetIndexMeta()\n\tinsStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumInserts(), iMeta.GetNumInsertFailures())\n\tupdStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumUpdates(), iMeta.GetNumUpdateFailures())\n\tdelStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumDeletes(), iMeta.GetNumDeleteFailures())\n\tfmt.Printf(\"%-30s%-10d%-10s%-10d%-10s%-10d%-10s%-10s%-10s%-10s\\n\",\n\t\tmeta.GetTableName(),\n\t\tmeta.GetTableId(),\n\t\tutils.TableKindToStr(meta.GetKind()),\n\t\tiMeta.GetCapacity(),\n\t\t\"-\",\n\t\tiMeta.GetUsage(),\n\t\t\"-\",\n\t\tinsStr,\n\t\tupdStr,\n\t\tdelStr)\n}","func (resolver *ResolverTODO) Show(params graphql.ResolveParams) (interface{}, error) {\n\ttodo, err := resolver.Db.Show(params.Args[\"id\"].(string))\n\n\tif err != nil {\n\t\treturn nil, err\n\t} else if todo.ID == \"\" {\n\t\treturn nil, errors.New(\"todo not found\")\n\t}\n\n\treturn todo, nil\n}"],"string":"[\n \"func (tc *TransactionsController) Show(c *gin.Context) {\\n\\thash := common.HexToHash(c.Param(\\\"TxHash\\\"))\\n\\tif tx, err := tc.App.GetStore().FindTxByAttempt(hash); err == orm.ErrorNotFound {\\n\\t\\tpublicError(c, http.StatusNotFound, errors.New(\\\"Transaction not found\\\"))\\n\\t} else if err != nil {\\n\\t\\tc.AbortWithError(http.StatusInternalServerError, err)\\n\\t} else if doc, err := jsonapi.Marshal(presenters.NewTx(tx)); err != nil {\\n\\t\\tc.AbortWithError(http.StatusInternalServerError, err)\\n\\t} else {\\n\\t\\tc.Data(http.StatusOK, MediaType, doc)\\n\\t}\\n}\",\n \"func ViewTransaction(rw http.ResponseWriter, r *http.Request) {\\n\\t// get the token\\n\\treqToken := r.Header.Get(\\\"Authorization\\\")\\n\\n\\t// get the claims\\n\\tclaims, isNotValid := GetClaims(reqToken, rw)\\n\\tif isNotValid {\\n\\t\\treturn\\n\\t}\\n\\n\\tdt, err := db.GetUserTransaction(claims.Roll)\\n\\tif err != nil {\\n\\t\\trw.WriteHeader(http.StatusInternalServerError)\\n\\t\\trw.Write(Rsp(err.Error(), \\\"Server Error\\\"))\\n\\t\\treturn\\n\\t}\\n\\trw.WriteHeader(http.StatusOK)\\n\\tres := c.RespData{\\n\\t\\tMessage: \\\"All data\\\",\\n\\t\\tData: dt,\\n\\t}\\n\\tjson.NewEncoder(rw).Encode(res)\\n}\",\n \"func (c *PurchaseController) Show(ctx *app.ShowPurchaseContext) error {\\n\\n\\tsession := Database.Session.Copy()\\n\\tdefer session.Close()\\n\\n\\tresult := app.Purchase{}\\n\\n\\terr := session.DB(\\\"services-pos\\\").C(\\\"Purchase\\\").FindId(bson.ObjectIdHex(ctx.TransactionID)).One(&result)\\n\\n\\tif err != nil {\\n\\t\\treturn ctx.NotFound()\\n\\t}\\n\\n\\tresult.TransactionID = ctx.TransactionID\\n\\tresult.Href = app.PurchaseHref(ctx.TransactionID)\\n\\n\\treturn ctx.OK(&result)\\n}\",\n \"func (g *Goods) Detail(c Context) {\\n\\t// TODO\\n\\tc.String(http.StatusOK, \\\"get goods detail\\\")\\n}\",\n \"func (this *FamilyAccount) showDetails() {\\n\\tfmt.Println(\\\"------My Income and Expense Detail-------\\\")\\n\\tif this.flag {\\n\\t\\t//因为我们用的是FamilyAccount结构体里传过来的字段,所以不能直接yongflag, 要用this. , 表示调用这个方法的结构体变量里面的字段\\n\\t\\tfmt.Println(this.details)\\n\\t} else {\\n\\t\\tfmt.Println(\\\"No current income and expenditure details!\\\")\\n\\t}\\n}\",\n \"func (r *BTCRPC) GetTransactionDetail(txhash string) ([]byte, error) {\\n\\tvar (\\n\\t\\ttx []byte\\n\\t\\terr error\\n\\t)\\n\\n\\terr = r.Client.Call(\\\"getrawtransaction\\\", jsonrpc.Params{txhash, 1}, &tx)\\n\\treturn tx, err\\n}\",\n \"func (h *Handler) show(c echo.Context) (e error) {\\n\\tctx := c.(*cuxs.Context)\\n\\n\\tvar id int64\\n\\tvar as *model.SalesReturn\\n\\tif id, e = common.Decrypt(ctx.Param(\\\"id\\\")); e == nil {\\n\\t\\tif as, e = ShowSalesReturn(\\\"id\\\", id); e == nil {\\n\\t\\t\\tctx.Data(as)\\n\\t\\t} else {\\n\\t\\t\\te = echo.ErrNotFound\\n\\t\\t}\\n\\t}\\n\\treturn ctx.Serve(e)\\n}\",\n \"func (p *Proxy) handleShowTxnz(session *driver.Session, query string, node sqlparser.Statement) (*sqltypes.Result, error) {\\n\\treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \\\"Access denied; lacking super privilege for the operation\\\")\\n\\t// privilegePlug := spanner.plugins.PlugPrivilege()\\n\\t// if !privilegePlug.IsSuperPriv(session.User()) {\\n\\t// \\treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \\\"Access denied; lacking super privilege for the operation\\\")\\n\\t// }\\n\\n\\t// qr := &sqltypes.Result{}\\n\\t// qr.Fields = []*querypb.Field{\\n\\t// \\t{Name: \\\"TxnID\\\", Type: querypb.Type_INT64},\\n\\t// \\t{Name: \\\"Start\\\", Type: querypb.Type_VARCHAR},\\n\\t// \\t{Name: \\\"Duration\\\", Type: querypb.Type_INT32},\\n\\t// \\t{Name: \\\"XaState\\\", Type: querypb.Type_VARCHAR},\\n\\t// \\t{Name: \\\"TxnState\\\", Type: querypb.Type_VARCHAR},\\n\\t// }\\n\\n\\t// rows := spanner.scatter.Txnz().GetTxnzRows()\\n\\t// for _, row := range rows {\\n\\t// \\trow := []sqltypes.Value{\\n\\t// \\t\\tsqltypes.MakeTrusted(querypb.Type_INT64, []byte(fmt.Sprintf(\\\"%v\\\", uint64(row.TxnID)))),\\n\\t// \\t\\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.Start.Format(\\\"20060102150405.000\\\"))),\\n\\t// \\t\\tsqltypes.MakeTrusted(querypb.Type_INT32, []byte(fmt.Sprintf(\\\"%v\\\", row.Duration))),\\n\\t// \\t\\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.XaState)),\\n\\t// \\t\\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.State)),\\n\\t// \\t}\\n\\t// \\tqr.Rows = append(qr.Rows, row)\\n\\t// }\\n\\t// return qr, nil\\n}\",\n \"func (showTxCommand ShowTransactionCommand) Run(ctx context.Context, wallet walletcore.Wallet) error {\\n\\ttransaction, err := wallet.GetTransaction(showTxCommand.Args.TxHash)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tbasicOutput := \\\"Hash\\\\t%s\\\\n\\\" +\\n\\t\\t\\\"Confirmations\\\\t%d\\\\n\\\" +\\n\\t\\t\\\"Included in block\\\\t%d\\\\n\\\" +\\n\\t\\t\\\"Type\\\\t%s\\\\n\\\" +\\n\\t\\t\\\"Amount %s\\\\t%s\\\\n\\\" +\\n\\t\\t\\\"Date\\\\t%s\\\\n\\\" +\\n\\t\\t\\\"Size\\\\t%s\\\\n\\\" +\\n\\t\\t\\\"Fee\\\\t%s\\\\n\\\" +\\n\\t\\t\\\"Rate\\\\t%s/kB\\\\n\\\"\\n\\n\\ttxDirection := strings.ToLower(transaction.Direction.String())\\n\\ttxSize := fmt.Sprintf(\\\"%.1f kB\\\", float64(transaction.Size)/1000)\\n\\tbasicOutput = fmt.Sprintf(basicOutput,\\n\\t\\ttransaction.Hash,\\n\\t\\ttransaction.Confirmations,\\n\\t\\ttransaction.BlockHeight,\\n\\t\\ttransaction.Type,\\n\\t\\ttxDirection, transaction.Amount,\\n\\t\\ttransaction.FormattedTime,\\n\\t\\ttxSize,\\n\\t\\ttransaction.Fee,\\n\\t\\ttransaction.FeeRate)\\n\\n\\tif showTxCommand.Detailed {\\n\\t\\tdetailedOutput := strings.Builder{}\\n\\t\\tdetailedOutput.WriteString(\\\"General Info\\\\n\\\")\\n\\t\\tdetailedOutput.WriteString(basicOutput)\\n\\t\\tdetailedOutput.WriteString(\\\"\\\\nInputs\\\\n\\\")\\n\\t\\tfor _, input := range transaction.Inputs {\\n\\t\\t\\tdetailedOutput.WriteString(fmt.Sprintf(\\\"%s\\\\t%s\\\\n\\\", dcrutil.Amount(input.AmountIn).String(), input.PreviousOutpoint))\\n\\t\\t}\\n\\t\\tdetailedOutput.WriteString(\\\"\\\\nOutputs\\\\n\\\")\\n\\t\\tfor _, out := range transaction.Outputs {\\n\\t\\t\\tif len(out.Addresses) == 0 {\\n\\t\\t\\t\\tdetailedOutput.WriteString(fmt.Sprintf(\\\"%s\\\\t (no address)\\\\n\\\", dcrutil.Amount(out.Value).String()))\\n\\t\\t\\t\\tcontinue\\n\\t\\t\\t}\\n\\n\\t\\t\\tdetailedOutput.WriteString(fmt.Sprintf(\\\"%s\\\", dcrutil.Amount(out.Value).String()))\\n\\t\\t\\tfor _, address := range out.Addresses {\\n\\t\\t\\t\\taccountName := address.AccountName\\n\\t\\t\\t\\tif !address.IsMine {\\n\\t\\t\\t\\t\\taccountName = \\\"external\\\"\\n\\t\\t\\t\\t}\\n\\t\\t\\t\\tdetailedOutput.WriteString(fmt.Sprintf(\\\"\\\\t%s (%s)\\\\n\\\", address.Address, accountName))\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\ttermio.PrintStringResult(strings.TrimRight(detailedOutput.String(), \\\" \\\\n\\\\r\\\"))\\n\\t} else {\\n\\t\\ttermio.PrintStringResult(basicOutput)\\n\\t}\\n\\treturn nil\\n}\",\n \"func viewAnyOrderGet(c *gin.Context) { //admin also have the same view , later combine those two func TBD\\n\\tOrdID := c.Request.URL.Query()[\\\"ordid\\\"][0] // Getting Order ID passed with URL\\n\\t_, usrName := session.SessinStatus(c, \\\"user_session_cookie\\\")\\n\\tfmt.Println(\\\"Wnat to see the order details of order number \\\", OrdID)\\n\\toK, itemsList, date, status, PayMode, amt := db.GetSingleOredrDetails(OrdID)\\n\\tif !oK {\\n\\t\\tfmt.Println(\\\"Something went wrong while picking Single Order Deatils ..Please have a look\\\")\\n\\t}\\n\\tfmt.Println(oK, itemsList, date, status, PayMode, amt)\\n\\t//\\t\\tsubTotalToFloat, _ := strconv.ParseFloat(singleCartItem.SubTotal, 64)\\n\\t//\\t\\tTotalAmt = TotalAmt + subTotalToFloat\\n\\t//\\tTotalAmtInPaisa := TotalAmt * 100 // This is required while initate for payment in Razorpay\\n\\n\\t//\\tTotalAmtString := fmt.Sprintf(\\\"%.2f\\\", TotalAmt)\\n\\n\\tc.HTML(\\n\\t\\thttp.StatusOK,\\n\\t\\t\\\"view_particular_order.html\\\",\\n\\t\\tgin.H{\\\"title\\\": \\\"OrderDetail\\\",\\n\\t\\t\\t\\\"ItemsOrdered\\\": itemsList,\\n\\t\\t\\t\\\"OrdID\\\": OrdID,\\n\\t\\t\\t\\\"date\\\": date,\\n\\t\\t\\t\\\"PayMode\\\": PayMode,\\n\\t\\t\\t\\\"amt\\\": amt,\\n\\t\\t\\t\\\"OrdStatus\\\": status,\\n\\t\\t\\t\\\"usrName\\\": usrName,\\n\\n\\t\\t\\t// \\\"TotalAmt\\\": TotalAmtString,\\n\\t\\t\\t// \\\"TotalAmtInPaisa\\\": TotalAmtInPaisa,\\n\\t\\t},\\n\\t)\\n\\n}\",\n \"func (s *Service) GetExplorerTransaction(w http.ResponseWriter, r *http.Request) {\\n\\tw.Header().Set(\\\"Content-Type\\\", \\\"application/json\\\")\\n\\tid := r.FormValue(\\\"id\\\")\\n\\n\\tdata := &Data{}\\n\\tdefer func() {\\n\\t\\tif err := json.NewEncoder(w).Encode(data.TX); err != nil {\\n\\t\\t\\tutils.Logger().Warn().Err(err).Msg(\\\"cannot JSON-encode TX\\\")\\n\\t\\t}\\n\\t}()\\n\\tif id == \\\"\\\" {\\n\\t\\tutils.Logger().Warn().Msg(\\\"invalid id parameter\\\")\\n\\t\\tw.WriteHeader(http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\tdb := s.Storage.GetDB()\\n\\tbytes, err := db.Get([]byte(GetTXKey(id)))\\n\\tif err != nil {\\n\\t\\tutils.Logger().Warn().Err(err).Str(\\\"id\\\", id).Msg(\\\"cannot read TX\\\")\\n\\t\\tw.WriteHeader(http.StatusInternalServerError)\\n\\t\\treturn\\n\\t}\\n\\ttx := new(Transaction)\\n\\tif rlp.DecodeBytes(bytes, tx) != nil {\\n\\t\\tutils.Logger().Warn().Str(\\\"id\\\", id).Msg(\\\"cannot convert data from DB\\\")\\n\\t\\tw.WriteHeader(http.StatusInternalServerError)\\n\\t\\treturn\\n\\t}\\n\\tdata.TX = *tx\\n}\",\n \"func (ctl *SaleCounterProductController) Detail() {\\n\\t//获取信息一样,直接调用Edit\\n\\tctl.Edit()\\n\\tctl.Data[\\\"Readonly\\\"] = true\\n\\tctl.Data[\\\"Action\\\"] = \\\"detail\\\"\\n}\",\n \"func orderAdminApproveGet(c *gin.Context) {\\n\\tOrdID := c.Request.URL.Query()[\\\"ordid\\\"][0] // Getting Order ID passed with URL\\n\\t_, usrName := session.SessinStatus(c, \\\"user_session_cookie\\\")\\n\\tfmt.Println(\\\"Wnat to see the order details of order number \\\", OrdID)\\n\\toK, itemsList, date, status, PayMode, amt := db.GetSingleOredrDetails(OrdID)\\n\\tif !oK {\\n\\t\\tfmt.Println(\\\"Something went wrong while picking Single Order Deatils ..Please have a look\\\")\\n\\t}\\n\\tfmt.Println(oK, itemsList, date, status, PayMode, amt)\\n\\tc.HTML(\\n\\t\\thttp.StatusOK,\\n\\t\\t\\\"order_adminview.html\\\",\\n\\t\\tgin.H{\\\"title\\\": \\\"OrderDetail\\\",\\n\\t\\t\\t\\\"ItemsOrdered\\\": itemsList,\\n\\t\\t\\t\\\"OrdID\\\": OrdID,\\n\\t\\t\\t\\\"date\\\": date,\\n\\t\\t\\t\\\"PayMode\\\": PayMode,\\n\\t\\t\\t\\\"amt\\\": amt,\\n\\t\\t\\t\\\"OrdStatus\\\": status,\\n\\t\\t\\t\\\"usrName\\\": usrName,\\n\\n\\t\\t\\t// \\\"TotalAmt\\\": TotalAmtString,\\n\\t\\t\\t// \\\"TotalAmtInPaisa\\\": TotalAmtInPaisa,\\n\\t\\t},\\n\\t)\\n}\",\n \"func (c *TradeController) Show(ctx *app.ShowTradeContext) error {\\n\\t// TradeController_Show: start_implement\\n\\n\\t// Put your logic here\\n\\tt, ok := tradeRegistry[ctx.TradeID]\\n\\tif !ok {\\n\\t\\treturn ctx.NotFound()\\n\\t}\\n\\tres := &app.GoaTrade{\\n\\t\\tTradeID: t.TradeID,\\n\\t\\tContractID: t.ContractID,\\n\\t\\tCounterpartyID: t.CounterpartyID,\\n\\t}\\n\\treturn ctx.OK(res)\\n\\t// TradeController_Show: end_implement\\n}\",\n \"func (action *TransactionShowAction) JSON() {\\n\\taction.Do(\\n\\t\\taction.EnsureHistoryFreshness,\\n\\t\\taction.loadParams,\\n\\t\\taction.checkAllowed,\\n\\t\\taction.loadRecord,\\n\\t\\taction.loadResource,\\n\\t\\tfunc() { hal.Render(action.W, action.Resource) },\\n\\t)\\n}\",\n \"func view(ctx context.Context, db transactor, fn func(*bolt.Tx) error) error {\\n\\ttx, ok := ctx.Value(transactionKey{}).(*bolt.Tx)\\n\\tif !ok {\\n\\t\\treturn db.View(fn)\\n\\t}\\n\\treturn fn(tx)\\n}\",\n \"func (db *DB) View(ctx context.Context, fn func(*TransactionManager) error) error {\\n\\ttx := db.BeginTransaction(false)\\n\\tdefer tx.Discard()\\n\\treturn fn(tx)\\n}\",\n \"func displayTransactions(scid string, option int, ID string) {\\n\\t\\t\\n\\tscKeys:= []string{\\\"numberOfOwners\\\", \\\"txCount\\\"}\\n\\tresult:= getKeysFromDaemon(scid, scKeys)\\n\\tif result == \\\"\\\" {return}\\n\\n\\n\\t//Response ok, extract keys from JSON\\n\\t\\n\\n\\ttxCount := gjson.Get(result, \\\"txs.#.sc_keys.txCount\\\")\\n\\ttxCountArray:= txCount.Array()[0]\\n\\ttxCountInt:= txCountArray.Int()\\n\\t//fmt.Printf(\\\"Tx Count: %d\\\\n\\\", txCountInt)\\n\\n\\t//Make a slice of keys so we can request in RPC call\\n\\tx:= int(txCountInt) //txCount in wallet smart contract is always 1 ahead of actual number of transactions\\t\\n\\tx4:= x * 4\\t\\n\\tkeySlice:= make([]string, x4) \\n\\t\\n\\tfor i:=0; i%.2f\\\", urlCustomersTransactionsView(q.CustomerID, q.AccountID, q.ID), q.Amount)\\n\\t\\t\\tcase \\\"created_at\\\":\\n\\t\\t\\t\\tdate := web.NewTimeResponse(ctx, time.Unix(q.CreatedAt, 0))\\n\\t\\t\\t\\tv.Value = date.LocalDate\\n\\t\\t\\t\\tv.Formatted = date.LocalDate\\n\\t\\t\\tcase \\\"narration\\\":\\n\\t\\t\\t\\tvalues := strings.Split(q.Narration, \\\":\\\")\\n\\t\\t\\t\\tif len(values) > 1 {\\n\\t\\t\\t\\t\\tif values[0] == \\\"sale\\\" {\\n\\t\\t\\t\\t\\t\\tv.Value = values[1]\\n\\t\\t\\t\\t\\t\\tv.Formatted = fmt.Sprintf(\\\"%s\\\", urlSalesView(values[2]), v.Value)\\n\\t\\t\\t\\t\\t}\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\tv.Value = q.Narration\\n\\t\\t\\t\\t\\tv.Formatted = q.Narration\\n\\t\\t\\t\\t}\\n\\t\\t\\tcase \\\"payment_method\\\":\\n\\t\\t\\t\\tv.Value = q.PaymentMethod\\n\\t\\t\\t\\tv.Formatted = q.PaymentMethod\\n\\t\\t\\tcase \\\"customer_name\\\":\\n\\t\\t\\t\\tv.Value = q.CustomerName\\n\\t\\t\\t\\tv.Formatted = fmt.Sprintf(\\\"%s\\\", urlCustomersView(q.CustomerID), v.Value)\\n\\t\\t\\tcase \\\"account\\\":\\n\\t\\t\\t\\tv.Value = q.AccountNumber\\n\\t\\t\\t\\tv.Formatted = fmt.Sprintf(\\\"%s\\\", urlCustomersAccountsView(q.CustomerID, q.AccountID), v.Value)\\n\\t\\t\\tcase \\\"sales_rep_id\\\":\\n\\t\\t\\t\\tv.Value = q.SalesRepID\\n\\t\\t\\t\\tv.Formatted = fmt.Sprintf(\\\"%s\\\", urlUsersView(q.SalesRepID), q.SalesRep)\\n\\t\\t\\tdefault:\\n\\t\\t\\t\\treturn resp, errors.Errorf(\\\"Failed to map value for %s.\\\", col.Field)\\n\\t\\t\\t}\\n\\t\\t\\tresp = append(resp, v)\\n\\t\\t}\\n\\n\\t\\treturn resp, nil\\n\\t}\\n\\n\\tvar txWhere = []string{\\\"tx_type = 'deposit'\\\"}\\n\\tvar txArgs []interface{}\\n\\t// todo sales rep filtering\\n\\tif v := r.URL.Query().Get(\\\"sales_rep_id\\\"); v != \\\"\\\" {\\n\\t\\ttxWhere = append(txWhere, \\\"sales_rep_id = $1\\\")\\n\\t\\ttxArgs = append(txArgs, v)\\n\\t\\tdata[\\\"salesRepID\\\"] = v\\n\\t}\\n\\n\\tif v := r.URL.Query().Get(\\\"payment_method\\\"); v != \\\"\\\" {\\n\\t\\ttxWhere = append(txWhere, fmt.Sprintf(\\\"payment_method = $%d\\\", len(txArgs)+1))\\n\\t\\ttxArgs = append(txArgs, v)\\n\\t\\tdata[\\\"paymentMethod\\\"] = v\\n\\t}\\n\\n\\tvar date = time.Now()\\n\\tif v := r.URL.Query().Get(\\\"start_date\\\"); v != \\\"\\\" {\\n\\t\\tdate, err = time.Parse(\\\"01/02/2006\\\", v)\\n\\t\\tif err != nil {\\n\\t\\t\\tdate = time.Now()\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t}\\n\\tdate = date.Truncate(time.Millisecond)\\n\\tdate = now.New(date).BeginningOfDay().Add(-1 * time.Hour)\\n\\ttxWhere = append(txWhere, fmt.Sprintf(\\\"created_at >= $%d\\\", len(txArgs)+1))\\n\\ttxArgs = append(txArgs, date.UTC().Unix())\\n\\tdata[\\\"startDate\\\"] = date.Format(\\\"01/02/2006\\\")\\n\\n\\tdate = time.Now()\\n\\tif v := r.URL.Query().Get(\\\"end_date\\\"); v != \\\"\\\" {\\n\\t\\tdate, err = time.Parse(\\\"01/02/2006\\\", v)\\n\\t\\tif err != nil {\\n\\t\\t\\tdate = time.Now()\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t}\\n\\tdate = date.Truncate(time.Millisecond)\\n\\tdate = now.New(date).EndOfDay().Add(-1 * time.Hour)\\n\\ttxWhere = append(txWhere, fmt.Sprintf(\\\"created_at <= $%d\\\", len(txArgs)+1))\\n\\ttxArgs = append(txArgs, date.Unix())\\n\\tdata[\\\"endDate\\\"] = date.Format(\\\"01/02/2006\\\")\\n\\n\\tloadFunc := func(ctx context.Context, sorting string, fields []datatable.DisplayField) (resp [][]datatable.ColumnValue, err error) {\\n\\n\\t\\tvar order []string\\n\\t\\tif len(sorting) > 0 {\\n\\t\\t\\torder = strings.Split(sorting, \\\",\\\")\\n\\t\\t}\\n\\n\\t\\tfor i := range txWhere {\\n\\t\\t\\ttxWhere[i] = \\\"tx.\\\" + txWhere[i]\\n\\t\\t}\\n\\t\\tres, err := h.TransactionRepo.TxReport(ctx, claims, transaction.FindRequest{\\n\\t\\t\\tOrder: order, Where: strings.Join(txWhere, \\\" AND \\\"), Args: txArgs,\\n\\t\\t})\\n\\t\\tif err != nil {\\n\\t\\t\\treturn resp, err\\n\\t\\t}\\n\\n\\t\\tfor _, a := range res {\\n\\t\\t\\tl, err := mapFunc(a, fields)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\treturn resp, errors.Wrapf(err, \\\"Failed to map brand for display.\\\")\\n\\t\\t\\t}\\n\\n\\t\\t\\tresp = append(resp, l)\\n\\t\\t}\\n\\n\\t\\treturn resp, nil\\n\\t}\\n\\n\\tdt, err := datatable.New(ctx, w, r, h.Redis, fields, loadFunc)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tif dt.HasCache() {\\n\\t\\treturn nil\\n\\t}\\n\\n\\tif ok, err := dt.Render(); ok {\\n\\t\\tif err != nil {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\t\\treturn nil\\n\\t}\\n\\n\\tusers, err := h.UserRepos.Find(ctx, claims, user.UserFindRequest{\\n\\t\\tOrder: []string{\\\"first_name\\\", \\\"last_name\\\"},\\n\\t})\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\ttotal, err = h.TransactionRepo.DepositAmountByWhere(ctx, strings.Join(txWhere, \\\" and \\\"), txArgs)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\tdata[\\\"paymentMethods\\\"] = transaction.PaymentMethods\\n\\tdata[\\\"users\\\"] = users\\n\\tdata[\\\"total\\\"] = total\\n\\tdata[\\\"datatable\\\"] = dt.Response()\\n\\n\\treturn h.Renderer.Render(ctx, w, r, TmplLayoutBase, \\\"report-transactions.gohtml\\\", web.MIMETextHTMLCharsetUTF8, http.StatusOK, data)\\n}\",\n \"func viewandedititemGet(c *gin.Context) {\\n\\tIsSectionActive, _ := session.SessinStatus(c, \\\"admin_session_cookie\\\")\\n\\tif !IsSectionActive {\\n\\t\\tfmt.Println(\\\"No Active Sessions found \\\")\\n\\t\\t// c.HTML(http.StatusOK, \\\"admin_login.html\\\", []string{\\\"a\\\", \\\"b\\\", \\\"c\\\"})\\n\\t\\tc.HTML(\\n\\t\\t\\thttp.StatusOK,\\n\\t\\t\\t\\\"admin_login.html\\\",\\n\\t\\t\\tgin.H{\\\"title\\\": \\\"success login\\\",\\n\\t\\t\\t\\t\\\"diplay\\\": \\\"none\\\",\\n\\t\\t\\t},\\n\\t\\t)\\n\\t} else {\\n\\t\\titemID := c.Request.URL.Query()[\\\"itemid\\\"][0] // Getting Order ID passed with URL\\n\\t\\tfmt.Println(\\\"Initiating to View/Edit item ,having ID\\\", itemID)\\n\\t\\t//populateCategoryItems(c, itemID)\\n\\t\\t//GetItemDetails(itemID string) (itemDesc string, itemRate float64, unit string,itmID,itmStock int,itmBuyRate float64) {\\n\\t\\t//Don't Confuse above function will redirect\\n\\t\\t//to edit page, usual practice is giving here\\n\\t\\t//but we achived this by modifying the existing\\n\\t\\t//code so it happend so..\\n\\t\\titmDesc, itmSelRate, itmUnit, itmID, itmStock, itmBuyPrice := db.GetItemDetails(itemID)\\n\\t\\tc.HTML(\\n\\t\\t\\thttp.StatusOK,\\n\\t\\t\\t\\\"edit_item.html\\\", gin.H{\\n\\t\\t\\t\\t\\\"delWarning\\\": \\\"none\\\",\\n\\t\\t\\t\\t\\\"updateSucess\\\": \\\"none\\\",\\n\\t\\t\\t\\t\\\"title\\\": \\\"Edit Item\\\",\\n\\t\\t\\t\\t\\\"itmID\\\": itmID,\\n\\t\\t\\t\\t\\\"itmDesc\\\": itmDesc,\\n\\t\\t\\t\\t\\\"itmUnit\\\": itmUnit,\\n\\t\\t\\t\\t\\\"itmBuyPrice\\\": itmBuyPrice,\\n\\t\\t\\t\\t\\\"itmSelRate\\\": itmSelRate,\\n\\t\\t\\t\\t\\\"itmStock\\\": itmStock,\\n\\t\\t\\t})\\n\\t}\\n}\",\n \"func (db *DB) View(fn func(*TransactionManager) error) error {\\n\\ttx := db.BeginTransaction(false)\\n\\tdefer tx.Discard()\\n\\treturn fn(tx)\\n}\",\n \"func transactionSummary(ctx *quorumContext, w http.ResponseWriter, r *http.Request) (int, error) {\\n\\tif ctx.conn == nil {\\n\\t\\tfmt.Fprintf(w, \\\"Cannot get transaction before deploying contract\\\\n\\\")\\n\\t\\treturn 400, nil\\n\\t}\\n\\n\\t// parse hash from request URL\\n\\tkeys := r.URL.Query()[\\\"hash\\\"]\\n\\tif len(keys) < 1 {\\n\\t\\tfmt.Fprintf(w, \\\"Invalid parameter, require 'hash'\\\")\\n\\t\\treturn 400, nil\\n\\t}\\n\\tlog.Println(\\\"Hash supplied :\\\", keys[0])\\n\\thash := common.HexToHash(keys[0])\\n\\n\\tbasic_context := context.Background()\\n\\ttx, pending, err := ctx.conn.TransactionByHash(basic_context, hash)\\n\\t//common.HexToHash(\\\"0x378674bebd1430d9ce63adc792c573da56e69b8d6c97174c93a43c5991ae0d61\\\"))\\n\\tif err != nil {\\n\\t\\tfmt.Fprintf(w, \\\"Failed to get transaction details: %v\\\", err)\\n\\t\\treturn 500, err\\n\\t}\\n\\tfmt.Fprintf(w, \\\"Transaction pending? %v; details: %v\\\\n\\\",\\n\\t\\tpending, tx.String())\\n\\treturn 200, nil\\n}\",\n \"func (self* userRestAPI) transactions(w http.ResponseWriter, r *http.Request) {\\n\\n\\t// Read arguments\\n\\tband,number,err := self.extractBandAndNumber(r)\\n\\tif err != nil {\\n\\t\\tlogError(err)\\n\\t\\thttp.Error(w, fmt.Sprintf(\\\"\\\\nFailed to parse arguments '%s'\\\\n\\\",err), http.StatusInternalServerError)\\n\\t\\treturn\\n\\t}\\n\\n\\t// Retrieve transactions for specified traveller\\n\\tw.Header().Set(\\\"Content-Type\\\", \\\"application/json\\\")\\n\\thistory,err := self.engine.TransactionsAsJSON(band,number)\\n\\tif err != nil {\\n\\t\\tlogError(err)\\n\\t\\thttp.Error(w, fmt.Sprintf(\\\"\\\\nFailed to retrieve transaction history with error '%s'\\\\n\\\",err), http.StatusInternalServerError)\\n\\t\\treturn\\n\\t}\\n\\tio.WriteString(w,history)\\n\\n}\",\n \"func (c *PlanController) PlanDetails() {\\n\\tplanViewModel := viewmodels.Plan{}\\n\\tr := c.Ctx.Request\\n\\tw := c.Ctx.ResponseWriter\\n\\tsessionValues, sessionStatus := SessionForPlan(w,r)\\n\\tplanViewModel.SessionFlag = sessionStatus\\n\\tplanViewModel.CompanyPlan = sessionValues.CompanyPlan\\n\\tplanViewModel.CompanyTeamName =sessionValues.CompanyTeamName\\n\\tc.Data[\\\"vm\\\"] = planViewModel\\n\\tc.TplName = \\\"template/plan.html\\\"\\n}\",\n \"func (action TransactionShowAction) ServeHTTPC(c web.C, w http.ResponseWriter, r *http.Request) {\\n\\tap := &action.Action\\n\\tap.Prepare(c, w, r)\\n\\taction.Log = action.Log.WithField(\\\"action\\\", \\\"TransactionShowAction\\\")\\n\\tap.Execute(&action)\\n}\",\n \"func (api *API) Get(tid string) (*pagarme.Response, *pagarme.Transaction, error) {\\n\\tresp, err := api.Config.Do(http.MethodGet, \\\"/transactions/\\\"+tid, nil)\\n\\tif err != nil {\\n\\t\\treturn nil, nil, err\\n\\t}\\n\\tif werr := www.ExtractError(resp); werr != nil {\\n\\t\\treturn werr, nil, nil\\n\\t}\\n\\tresult := &pagarme.Transaction{}\\n\\tif err := www.Unmarshal(resp, result); err != nil {\\n\\t\\tapi.Config.Logger.Error(\\\"could not unmarshal transaction [Get]: \\\" + err.Error())\\n\\t\\treturn nil, nil, err\\n\\t}\\n\\n\\treturn www.Ok(), result, nil\\n}\",\n \"func (tb *transactionBuilder) View() (types.Transaction, []types.Transaction) {\\n\\treturn tb.transaction, tb.parents\\n}\",\n \"func (tb *transactionBuilder) View() (types.Transaction, []types.Transaction) {\\n\\treturn tb.transaction, tb.parents\\n}\",\n \"func ShowDeposit2ChannelMsgDetail(Status string) {\\n\\tswitch Status {\\n\\tcase \\\"200 OK\\\":\\n\\t\\tlog.Println(\\\"Success Deposit\\\")\\n\\tcase \\\"400 Bad Request\\\":\\n\\t\\tlog.Println(\\\"The provided json is in some way malformed!\\\")\\n\\tcase \\\"402 Payment required\\\":\\n\\t\\tlog.Println(\\\"Insufficient balance to do a deposit\\\")\\n\\tcase \\\"408 Request Timeout\\\":\\n\\t\\tlog.Println(\\\"The deposit event was not read in time by the ethereum node\\\")\\n\\tcase \\\"409 Conflict\\\":\\n\\t\\tlog.Println(\\\"Provided channel does not exist\\\")\\n\\tcase \\\"500 Server Error\\\":\\n\\t\\tlog.Println(\\\"Internal Raiden node error\\\")\\n\\tcase \\\"504 TimeOut\\\":\\n\\t\\tlog.Println(\\\"No response,timeout\\\")\\n\\tdefault:\\n\\t\\tfmt.Printf(\\\"Unknown error,Deposit Failure! %s\\\\n\\\", Status)\\n\\t}\\n}\",\n \"func (account *Account) ShowTx(P string) bool {\\n\\tfor i := 0; i < account.WalletIndex; i++ {\\n\\t\\tif P == account.Wallet[i].P {\\n\\t\\t\\tJSON, _ := json.MarshalIndent(account.Wallet[i], \\\"\\\", \\\" \\\")\\n\\n\\t\\t\\tfmt.Println(string(JSON))\\n\\t\\t\\treturn true\\n\\n\\t\\t}\\n\\t}\\n\\treturn false\\n}\",\n \"func (s *searcher) Transaction(resp http.ResponseWriter, req *http.Request) {\\n\\tsearchTerms := mux.Vars(req)\\n\\n\\ttransactionID := searchTerms[\\\"transaction_id\\\"]\\n\\tif len(transactionID) == 0 {\\n\\t\\tresp.WriteHeader(http.StatusBadRequest)\\n\\t\\tresp.Write([]byte(\\\"transaction ID is empty\\\"))\\n\\t\\treturn\\n\\t}\\n\\n\\tif len(transactionID) != 64 {\\n\\t\\tresp.WriteHeader(http.StatusBadRequest)\\n\\t\\tresp.Write([]byte(\\\"transaction ID is not 64 characters\\\"))\\n\\t\\treturn\\n\\t}\\n\\n\\tfileName, transactionIndex, err := s.searchIndex.GetTransactionPathByID(transactionID)\\n\\tif err != nil {\\n\\t\\tresp.WriteHeader(http.StatusInternalServerError)\\n\\t\\tresp.Write([]byte(fmt.Sprintf(\\\"error finding transaction: %s\\\", err.Error())))\\n\\t\\treturn\\n\\t}\\n\\n\\ttransactions, err := s.searchIndex.GetTransactionsFromSingleFile(fileName, []int{transactionIndex})\\n\\tif err != nil {\\n\\t\\tresp.WriteHeader(http.StatusInternalServerError)\\n\\t\\tresp.Write([]byte(fmt.Sprintf(\\\"error finding transaction: %s\\\", err.Error())))\\n\\t\\treturn\\n\\t}\\n\\n\\tresultBytes, err := json.Marshal(transactions)\\n\\tif err != nil {\\n\\t\\tresp.WriteHeader(http.StatusInternalServerError)\\n\\t\\tresp.Write([]byte(fmt.Sprintf(\\\"error marshallig transaction to json: %s\\\", err.Error())))\\n\\t\\treturn\\n\\t}\\n\\n\\tresp.WriteHeader(http.StatusOK)\\n\\tresp.Write(resultBytes)\\n}\",\n \"func (db *MySQLDB) GetTransactionDetailByTransactionID(ctx context.Context, transactionID int) ([]*TransactionDetailRecord, error) {\\n\\tfLog := mysqlLog.WithField(\\\"func\\\", \\\"GetTransactionByTransactionID\\\")\\n\\ttransaction := &TransactionRecord{}\\n\\n\\trow := db.instance.QueryRowContext(ctx, \\\"SELECT id FROM transactions WHERE id = ?\\\", transactionID)\\n\\terr := row.Scan(&transaction.ID)\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"row.Scan got %s\\\", err.Error())\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tq := fmt.Sprintf(\\\"SELECT td.transaction_id, td.product_id, td.qty, td.sub_total, p.sku FROM transaction_detail td INNER JOIN products p ON td.product_id = p.id WHERE transaction_id = %v\\\", transactionID)\\n\\trows, err := db.instance.QueryContext(ctx, q)\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"db.instance.QueryContext got %s\\\", err.Error())\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\ttDetail := make([]*TransactionDetailRecord, 0)\\n\\tfor rows.Next() {\\n\\t\\ttD := &TransactionDetailRecord{}\\n\\t\\terr := rows.Scan(&tD.TransactionID, &tD.ProductID, &tD.Qty, &tD.SubTotal, tD.SKU)\\n\\t\\tif err != nil {\\n\\t\\t\\tfLog.Errorf(\\\"rows.Scan got %s\\\", err.Error())\\n\\t\\t} else {\\n\\t\\t\\ttDetail = append(tDetail, tD)\\n\\t\\t}\\n\\t}\\n\\n\\treturn tDetail, nil\\n}\",\n \"func (res BackTestResult) Show() TradestatPort {\\n\\t//\\tp := NewPortfolio()\\n\\tmds := bean.NewRPCMDSConnC(\\\"tcp\\\", res.dbhost+\\\":\\\"+res.dbport)\\n\\tratesbook := make(ReferenceRateBook)\\n\\n\\t// FIXME: think about how to show multi pair result\\n\\tvar stat TradestatPort\\n\\tif len(res.pairs) > 0 {\\n\\t\\tp := res.pairs[0]\\n\\t\\ttxn, _ := mds.GetTransactions(p, res.start, res.end)\\n\\t\\tratesbook[p] = RefRatesFromTxn(txn)\\n\\t\\t//\\t\\tsnapts.Print()\\n\\t\\t//\\t\\tperfts.Print()\\n\\n\\t\\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\\n\\t\\tstat.Print()\\n\\t}\\n\\treturn stat\\n}\",\n \"func (s *Store) TxDetails(ns walletdb.ReadBucket, txHash *chainhash.Hash) (*TxDetails, error) {\\n\\t// First, check whether there exists an unmined transaction with this\\n\\t// hash. Use it if found.\\n\\tv := existsRawUnmined(ns, txHash[:])\\n\\tif v != nil {\\n\\t\\treturn s.unminedTxDetails(ns, txHash, v)\\n\\t}\\n\\n\\t// Otherwise, if there exists a mined transaction with this matching\\n\\t// hash, skip over to the newest and begin fetching all details.\\n\\tk, v := latestTxRecord(ns, txHash)\\n\\tif v == nil {\\n\\t\\t// not found\\n\\t\\treturn nil, nil\\n\\t}\\n\\treturn s.minedTxDetails(ns, txHash, k, v)\\n}\",\n \"func (dao *InfoDao) Transaction(ctx context.Context, f func(ctx context.Context, tx *gdb.TX) error) (err error) {\\n\\treturn dao.Ctx(ctx).Transaction(ctx, f)\\n}\",\n \"func Show(c *gin.Context) {\\r\\n\\tpost := getById(c)\\r\\n\\tc.JSON(http.StatusOK, gin.H{\\r\\n\\t\\t\\\"messege\\\": \\\"\\\",\\r\\n\\t\\t\\\"data\\\": post,\\r\\n\\t})\\r\\n}\",\n \"func tableInfoFlowShow(meta *halproto.TableMetadata) {\\n\\tfMeta := meta.GetFlowMeta()\\n\\tinsStr := fmt.Sprintf(\\\"%d[%d]\\\", fMeta.GetNumInserts(), fMeta.GetNumInsertFailures())\\n\\tupdStr := fmt.Sprintf(\\\"%d[%d]\\\", fMeta.GetNumUpdates(), fMeta.GetNumUpdateFailures())\\n\\tdelStr := fmt.Sprintf(\\\"%d[%d]\\\", fMeta.GetNumDeletes(), fMeta.GetNumDeleteFailures())\\n\\tfmt.Printf(\\\"%-30s%-10d%-10s%-10d%-10d%-10d%-10d%-10s%-10s%-10s\\\\n\\\",\\n\\t\\tmeta.GetTableName(),\\n\\t\\tmeta.GetTableId(),\\n\\t\\tutils.TableKindToStr(meta.GetKind()),\\n\\t\\tfMeta.GetCapacity(),\\n\\t\\tfMeta.GetCollCapacity(),\\n\\t\\tfMeta.GetHashUsage(),\\n\\t\\tfMeta.GetCollUsage(),\\n\\t\\tinsStr,\\n\\t\\tupdStr,\\n\\t\\tdelStr)\\n}\",\n \"func (c *BalanceClient) RetrieveTransaction(id string) (*BalanceTransaction, error) {\\n\\tbalanceTransaction := BalanceTransaction{}\\n\\terr := c.client.get(\\\"/balance/history/\\\"+id, nil, &balanceTransaction)\\n\\treturn &balanceTransaction, err\\n}\",\n \"func (s *Server) handleTransaction(client string, req *pb.Command) (err error) {\\n\\t// Get the transfer from the original command, will panic if nil\\n\\ttransfer := req.GetTransfer()\\n\\tmsg := fmt.Sprintf(\\\"starting transaction of %0.2f from %s to %s\\\", transfer.Amount, transfer.Account, transfer.Beneficiary)\\n\\ts.updates.Broadcast(req.Id, msg, pb.MessageCategory_LEDGER)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// Handle Demo UI errors before the account lookup\\n\\tif transfer.OriginatingVasp != \\\"\\\" && transfer.OriginatingVasp != s.vasp.Name {\\n\\t\\tlog.Info().Str(\\\"requested\\\", transfer.OriginatingVasp).Str(\\\"local\\\", s.vasp.Name).Msg(\\\"requested originator does not match local VASP\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrWrongVASP, \\\"message sent to the wrong originator VASP\\\"),\\n\\t\\t)\\n\\t}\\n\\n\\t// Lookup the account associated with the transfer originator\\n\\tvar account Account\\n\\tif err = LookupAccount(s.db, transfer.Account).First(&account).Error; err != nil {\\n\\t\\tif errors.Is(err, gorm.ErrRecordNotFound) {\\n\\t\\t\\tlog.Info().Str(\\\"account\\\", transfer.Account).Msg(\\\"not found\\\")\\n\\t\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\t\\tpb.Errorf(pb.ErrNotFound, \\\"account not found\\\"),\\n\\t\\t\\t)\\n\\t\\t}\\n\\t\\treturn fmt.Errorf(\\\"could not fetch account: %s\\\", err)\\n\\t}\\n\\ts.updates.Broadcast(req.Id, fmt.Sprintf(\\\"account %04d accessed successfully\\\", account.ID), pb.MessageCategory_LEDGER)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// Lookup the wallet of the beneficiary\\n\\tvar beneficiary Wallet\\n\\tif err = LookupBeneficiary(s.db, transfer.Beneficiary).First(&beneficiary).Error; err != nil {\\n\\t\\tif errors.Is(err, gorm.ErrRecordNotFound) {\\n\\t\\t\\tlog.Info().Str(\\\"beneficiary\\\", transfer.Beneficiary).Msg(\\\"not found\\\")\\n\\t\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\t\\tpb.Errorf(pb.ErrNotFound, \\\"beneficiary wallet not found\\\"),\\n\\t\\t\\t)\\n\\t\\t}\\n\\t\\treturn fmt.Errorf(\\\"could not fetch beneficiary wallet: %s\\\", err)\\n\\t}\\n\\n\\tif transfer.CheckBeneficiary {\\n\\t\\tif transfer.BeneficiaryVasp != beneficiary.Provider.Name {\\n\\t\\t\\tlog.Info().\\n\\t\\t\\t\\tStr(\\\"expected\\\", transfer.BeneficiaryVasp).\\n\\t\\t\\t\\tStr(\\\"actual\\\", beneficiary.Provider.Name).\\n\\t\\t\\t\\tMsg(\\\"check beneficiary failed\\\")\\n\\t\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\t\\tpb.Errorf(pb.ErrWrongVASP, \\\"beneficiary wallet does not match beneficiary vasp\\\"),\\n\\t\\t\\t)\\n\\t\\t}\\n\\t}\\n\\ts.updates.Broadcast(req.Id, fmt.Sprintf(\\\"wallet %s provided by %s\\\", beneficiary.Address, beneficiary.Provider.Name), pb.MessageCategory_BLOCKCHAIN)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// TODO: lookup peer from cache rather than always doing a directory service lookup\\n\\tvar peer *peers.Peer\\n\\ts.updates.Broadcast(req.Id, fmt.Sprintf(\\\"search for %s in directory service\\\", beneficiary.Provider.Name), pb.MessageCategory_TRISADS)\\n\\tif peer, err = s.peers.Search(beneficiary.Provider.Name); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not search peer from directory service\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not search peer from directory service\\\"),\\n\\t\\t)\\n\\t}\\n\\tinfo := peer.Info()\\n\\ts.updates.Broadcast(req.Id, fmt.Sprintf(\\\"identified TRISA remote peer %s at %s via directory service\\\", info.ID, info.Endpoint), pb.MessageCategory_TRISADS)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\tvar signKey *rsa.PublicKey\\n\\ts.updates.Broadcast(req.Id, \\\"exchanging peer signing keys\\\", pb.MessageCategory_TRISAP2P)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\tif signKey, err = peer.ExchangeKeys(true); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not exchange keys with remote peer\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not exchange keyrs with remote peer\\\"),\\n\\t\\t)\\n\\t}\\n\\n\\t// Prepare the transaction\\n\\t// Save the pending transaction and increment the accounts pending field\\n\\txfer := Transaction{\\n\\t\\tEnvelope: uuid.New().String(),\\n\\t\\tAccount: account,\\n\\t\\tAmount: decimal.NewFromFloat32(transfer.Amount),\\n\\t\\tDebit: true,\\n\\t\\tCompleted: false,\\n\\t}\\n\\n\\tif err = s.db.Save(&xfer).Error; err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not save transaction\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not save transaction\\\"),\\n\\t\\t)\\n\\t}\\n\\n\\t// Save the pending transaction on the account\\n\\t// TODO: remove pending transactions\\n\\taccount.Pending++\\n\\tif err = s.db.Save(&account).Error; err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not save originator account\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not save originator account\\\"),\\n\\t\\t)\\n\\t}\\n\\n\\ts.updates.Broadcast(req.Id, \\\"ready to execute transaction\\\", pb.MessageCategory_BLOCKCHAIN)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// Create an identity and transaction payload for TRISA exchange\\n\\ttransaction := &generic.Transaction{\\n\\t\\tTxid: fmt.Sprintf(\\\"%d\\\", xfer.ID),\\n\\t\\tOriginator: account.WalletAddress,\\n\\t\\tBeneficiary: beneficiary.Address,\\n\\t\\tAmount: float64(transfer.Amount),\\n\\t\\tNetwork: \\\"TestNet\\\",\\n\\t\\tTimestamp: xfer.Timestamp.Format(time.RFC3339),\\n\\t}\\n\\tidentity := &ivms101.IdentityPayload{\\n\\t\\tOriginator: &ivms101.Originator{},\\n\\t\\tOriginatingVasp: &ivms101.OriginatingVasp{},\\n\\t}\\n\\tif identity.OriginatingVasp.OriginatingVasp, err = s.vasp.LoadIdentity(); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not load originator vasp\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not load originator vasp\\\"),\\n\\t\\t)\\n\\t}\\n\\n\\tidentity.Originator = &ivms101.Originator{\\n\\t\\tOriginatorPersons: make([]*ivms101.Person, 0, 1),\\n\\t\\tAccountNumbers: []string{account.WalletAddress},\\n\\t}\\n\\tvar originator *ivms101.Person\\n\\tif originator, err = account.LoadIdentity(); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not load originator identity\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not load originator identity\\\"),\\n\\t\\t)\\n\\t}\\n\\tidentity.Originator.OriginatorPersons = append(identity.Originator.OriginatorPersons, originator)\\n\\n\\tpayload := &protocol.Payload{}\\n\\tif payload.Transaction, err = anypb.New(transaction); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not serialize transaction payload\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not serialize transaction payload\\\"),\\n\\t\\t)\\n\\t}\\n\\tif payload.Identity, err = anypb.New(identity); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not serialize identity payload\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not serialize identity payload\\\"),\\n\\t\\t)\\n\\t}\\n\\n\\ts.updates.Broadcast(req.Id, \\\"transaction and identity payload constructed\\\", pb.MessageCategory_TRISAP2P)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// Secure the envelope with the remote beneficiary's signing keys\\n\\tvar envelope *protocol.SecureEnvelope\\n\\tif envelope, err = handler.New(xfer.Envelope, payload, nil).Seal(signKey); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not create or sign secure envelope\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not create or sign secure envelope\\\"),\\n\\t\\t)\\n\\t}\\n\\n\\ts.updates.Broadcast(req.Id, fmt.Sprintf(\\\"secure envelope %s sealed: encrypted with AES-GCM and RSA - sending ...\\\", envelope.Id), pb.MessageCategory_TRISAP2P)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// Conduct the TRISA transaction, handle errors and send back to user\\n\\tif envelope, err = peer.Transfer(envelope); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not perform TRISA exchange\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, err.Error()),\\n\\t\\t)\\n\\t}\\n\\n\\ts.updates.Broadcast(req.Id, fmt.Sprintf(\\\"received %s information exchange reply from %s\\\", envelope.Id, peer.String()), pb.MessageCategory_TRISAP2P)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// Open the response envelope with local private keys\\n\\tvar opened *handler.Envelope\\n\\tif opened, err = handler.Open(envelope, s.trisa.sign); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not unseal TRISA response\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, err.Error()),\\n\\t\\t)\\n\\t}\\n\\n\\t// Verify the contents of the response\\n\\tpayload = opened.Payload\\n\\tif payload.Identity.TypeUrl != \\\"type.googleapis.com/ivms101.IdentityPayload\\\" {\\n\\t\\tlog.Warn().Str(\\\"type\\\", payload.Identity.TypeUrl).Msg(\\\"unsupported identity type\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"unsupported identity type\\\", payload.Identity.TypeUrl),\\n\\t\\t)\\n\\t}\\n\\n\\tif payload.Transaction.TypeUrl != \\\"type.googleapis.com/trisa.data.generic.v1beta1.Transaction\\\" {\\n\\t\\tlog.Warn().Str(\\\"type\\\", payload.Transaction.TypeUrl).Msg(\\\"unsupported transaction type\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"unsupported transaction type\\\", payload.Transaction.TypeUrl),\\n\\t\\t)\\n\\t}\\n\\n\\tidentity = &ivms101.IdentityPayload{}\\n\\ttransaction = &generic.Transaction{}\\n\\tif err = payload.Identity.UnmarshalTo(identity); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not unmarshal identity\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, err.Error()),\\n\\t\\t)\\n\\t}\\n\\tif err = payload.Transaction.UnmarshalTo(transaction); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not unmarshal transaction\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, err.Error()),\\n\\t\\t)\\n\\t}\\n\\n\\ts.updates.Broadcast(req.Id, \\\"successfully decrypted and parsed secure envelope\\\", pb.MessageCategory_TRISAP2P)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\t// Update the completed transaction and save to disk\\n\\txfer.Beneficiary = Identity{\\n\\t\\tWalletAddress: transaction.Beneficiary,\\n\\t}\\n\\txfer.Completed = true\\n\\txfer.Timestamp, _ = time.Parse(time.RFC3339, transaction.Timestamp)\\n\\n\\t// Serialize the identity information as JSON data\\n\\tvar data []byte\\n\\tif data, err = json.Marshal(identity); err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not save transaction\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, \\\"could not marshal IVMS 101 identity\\\"),\\n\\t\\t)\\n\\t}\\n\\txfer.Identity = string(data)\\n\\n\\tif err = s.db.Save(&xfer).Error; err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not save transaction\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, err.Error()),\\n\\t\\t)\\n\\t}\\n\\n\\t// Save the pending transaction on the account\\n\\t// TODO: remove pending transactions\\n\\taccount.Pending--\\n\\taccount.Completed++\\n\\taccount.Balance.Sub(xfer.Amount)\\n\\tif err = s.db.Save(&account).Error; err != nil {\\n\\t\\tlog.Error().Err(err).Msg(\\\"could not save transaction\\\")\\n\\t\\treturn s.updates.SendTransferError(client, req.Id,\\n\\t\\t\\tpb.Errorf(pb.ErrInternal, err.Error()),\\n\\t\\t)\\n\\t}\\n\\n\\tmsg = fmt.Sprintf(\\\"transaction %04d complete: %s transfered from %s to %s\\\", xfer.ID, xfer.Amount.String(), xfer.Originator.WalletAddress, xfer.Beneficiary.WalletAddress)\\n\\ts.updates.Broadcast(req.Id, msg, pb.MessageCategory_BLOCKCHAIN)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\ts.updates.Broadcast(req.Id, fmt.Sprintf(\\\"%04d new account balance: %s\\\", account.ID, account.Balance), pb.MessageCategory_LEDGER)\\n\\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\\n\\n\\trep := &pb.Message{\\n\\t\\tType: pb.RPC_TRANSFER,\\n\\t\\tId: req.Id,\\n\\t\\tTimestamp: time.Now().Format(time.RFC3339),\\n\\t\\tCategory: pb.MessageCategory_LEDGER,\\n\\t\\tReply: &pb.Message_Transfer{Transfer: &pb.TransferReply{\\n\\t\\t\\tTransaction: xfer.Proto(),\\n\\t\\t}},\\n\\t}\\n\\n\\treturn s.updates.Send(client, rep)\\n}\",\n \"func (eobj emp)details(){\\n\\tfmt.Println(\\\"Employee details are \\\",eobj.id,\\\" \\\",eobj.name,\\\" \\\",eobj.age)\\n}\",\n \"func (c *TrackerController) Show(ctx *app.ShowTrackerContext) error {\\n\\treturn application.Transactional(c.db, func(appl application.Application) error {\\n\\t\\tt, err := appl.Trackers().Load(ctx.Context, ctx.ID)\\n\\t\\tif err != nil {\\n\\t\\t\\tswitch err.(type) {\\n\\t\\t\\tcase remoteworkitem.NotFoundError:\\n\\t\\t\\t\\tlog.Printf(\\\"not found, id=%s\\\", ctx.ID)\\n\\t\\t\\t\\treturn goa.ErrNotFound(err.Error())\\n\\t\\t\\tdefault:\\n\\t\\t\\t\\treturn err\\n\\t\\t\\t}\\n\\t\\t}\\n\\t\\treturn ctx.OK(t)\\n\\t})\\n}\",\n \"func (srv *Server) walletTransactionHandler(w http.ResponseWriter, req *http.Request, ps httprouter.Params) {\\n\\t// Parse the id from the url.\\n\\tvar id types.TransactionID\\n\\tjsonID := \\\"\\\\\\\"\\\" + ps.ByName(\\\"id\\\") + \\\"\\\\\\\"\\\"\\n\\terr := id.UnmarshalJSON([]byte(jsonID))\\n\\tif err != nil {\\n\\t\\twriteError(w, \\\"error after call to /wallet/history: \\\"+err.Error(), http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\n\\ttxn, ok := srv.wallet.Transaction(id)\\n\\tif !ok {\\n\\t\\twriteError(w, \\\"error when calling /wallet/transaction/$(id): transaction not found\\\", http.StatusBadRequest)\\n\\t\\treturn\\n\\t}\\n\\twriteJSON(w, WalletTransactionGETid{\\n\\t\\tTransaction: txn,\\n\\t})\\n}\",\n \"func GetTransactionHandler(w http.ResponseWriter, r *http.Request) {\\n\\tvars := mux.Vars(r)\\n\\t// retrieve the parameters\\n\\tparam := make(map[string]uint64)\\n\\tfor _, key := range []string{\\\"blockId\\\", \\\"txId\\\"} {\\n\\t\\tparam[key], _ = strconv.ParseUint(vars[\\\"blockId\\\"], 10, 64)\\n\\t}\\n\\n\\ttmp := atomic.LoadUint64(&lastBlock)\\n\\tif param[\\\"blockId\\\"] > tmp {\\n\\t\\tw.WriteHeader(http.StatusBadRequest)\\n\\t\\terr := fmt.Errorf(\\\"requested id %d latest %d\\\", param[\\\"blockId\\\"], lastBlock)\\n\\t\\tlog.Println(err.Error())\\n\\t\\t_, _ = w.Write([]byte(err.Error()))\\n\\t\\treturn\\n\\t}\\n\\t// retuning anything in the body regardless of any error code\\n\\t// it may contain\\n\\t_, _, body, _ := dataCollection.GetTransaction(param[\\\"blockId\\\"], param[\\\"txId\\\"], config.DefaultRequestsTimeout)\\n\\twriteResponse(body, &w)\\n}\",\n \"func Show(w http.ResponseWriter, r *http.Request) {\\n\\tc := flight.Context(w, r)\\n\\n\\titem, _, err := summary.ByID(c.DB, c.Param(\\\"id\\\"))\\n\\tif err != nil {\\n\\t\\tc.FlashErrorGeneric(err)\\n\\t\\tc.Redirect(uri)\\n\\t\\treturn\\n\\t}\\n\\n\\tv := c.View.New(\\\"summary/show\\\")\\n\\tv.Vars[\\\"item\\\"] = item\\n\\tv.Render(w, r)\\n}\",\n \"func show(req events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\\n\\t// Get the `employeeid` query string parameter from the request and\\n\\t// validate it.\\n\\temployeeid := req.QueryStringParameters[\\\"employeeid\\\"]\\n\\tif !employeeidRegexp.MatchString(employeeid) {\\n\\t\\treturn clientError(http.StatusBadRequest)\\n\\t} //fetch a specific employee record from dynamodb in this case\\n\\n\\t// Fetch the employee record from the database based on the employeeid value.\\n\\temp, err := getItem(employeeid)\\n\\tif err != nil {\\n\\t\\treturn serverError(err)\\n\\t}\\n\\tif emp == nil {\\n\\t\\treturn clientError(http.StatusNotFound)\\n\\t}\\n\\n\\t// The APIGatewayProxyResponse.Body field needs to be a string, so\\n\\t// we marshal the employee record into JSON.\\n\\tjs, err := json.Marshal(emp)\\n\\tif err != nil {\\n\\t\\treturn serverError(err)\\n\\t}\\n\\n\\t// Return a response with a 200 OK status and the JSON employee record\\n\\t// as the body.\\n\\treturn events.APIGatewayProxyResponse{\\n\\t\\tStatusCode: http.StatusOK,\\n\\t\\tBody: string(js),\\n\\t}, nil\\n}\",\n \"func (_VinTracker *VinTrackerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _VinTracker.Contract.VinTrackerTransactor.contract.Transact(opts, method, params...)\\n}\",\n \"func (db *MySQLDB) GetTransactionByTransactionID(ctx context.Context, transactionID int) (*TransactionRecord, error) {\\n\\tfLog := mysqlLog.WithField(\\\"func\\\", \\\"GetTransactionByTransactionID\\\")\\n\\ttransaction := &TransactionRecord{}\\n\\n\\trow := db.instance.QueryRowContext(ctx, \\\"SELECT id, user_id, date, grand_total, discount, reason FROM transactions WHERE id = ?\\\", transactionID)\\n\\terr := row.Scan(&transaction.ID, &transaction.UserID, &transaction.Date, &transaction.GrandTotal, &transaction.Discount, &transaction.Reason)\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"row.Scan got %s\\\", err.Error())\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\tq := fmt.Sprintf(\\\"SELECT td.transaction_id, td.product_id, td.qty, td.sub_total, p.sku FROM transaction_detail td INNER JOIN products p ON td.product_id = p.id WHERE transaction_id = %v\\\", transactionID)\\n\\trows, err := db.instance.QueryContext(ctx, q)\\n\\tif err != nil {\\n\\t\\tfLog.Errorf(\\\"db.instance.QueryContext got %s\\\", err.Error())\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\ttDetail := make([]*TransactionDetailRecord, 0)\\n\\tfor rows.Next() {\\n\\t\\ttD := &TransactionDetailRecord{}\\n\\t\\terr := rows.Scan(&tD.TransactionID, &tD.ProductID, &tD.Qty, &tD.SubTotal, &tD.SKU)\\n\\t\\tif err != nil {\\n\\t\\t\\tfLog.Errorf(\\\"rows.Scan got %s\\\", err.Error())\\n\\t\\t} else {\\n\\t\\t\\ttDetail = append(tDetail, tD)\\n\\t\\t}\\n\\t}\\n\\n\\ttransaction.TransactionDetail = tDetail\\n\\n\\treturn transaction, nil\\n}\",\n \"func (_DetailedERC20 *DetailedERC20Raw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _DetailedERC20.Contract.DetailedERC20Transactor.contract.Transact(opts, method, params...)\\n}\",\n \"func (_DetailedERC20 *DetailedERC20Raw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _DetailedERC20.Contract.DetailedERC20Transactor.contract.Transact(opts, method, params...)\\n}\",\n \"func TransactionIndex(c *gin.Context) {\\n\\trelatedObjectID := c.Query(\\\"relatedObjectId\\\")\\n\\trelatedObjectType := c.Query(\\\"relatedObjectType\\\")\\n\\tisSettledQuery := c.Query(\\\"isSettled\\\")\\n\\tstatusQuery := c.Query(\\\"status\\\")\\n\\tcurUserID := c.Keys[\\\"CurrentUserID\\\"]\\n\\n\\tvar transactions []models.Transaction\\n\\n\\tquery := database.DBCon\\n\\n\\tisSettled, err := strconv.ParseBool(isSettledQuery)\\n\\tif isSettledQuery != \\\"\\\" && err == nil {\\n\\t\\tquery = query.Where(\\\"is_settled = ?\\\", isSettled)\\n\\t}\\n\\n\\t// TODO: Check that statusQuery is a valid status\\n\\tif statusQuery != \\\"\\\" {\\n\\t\\tquery = query.Where(\\\"status = ?\\\", statusQuery)\\n\\t}\\n\\n\\tif relatedObjectID != \\\"\\\" && relatedObjectType != \\\"\\\" {\\n\\t\\tquery.\\n\\t\\t\\tWhere(\\\"related_object_id = ? AND related_object_type = ?\\\", relatedObjectID, relatedObjectType).\\n\\t\\t\\tOrder(\\\"created_at desc\\\").\\n\\t\\t\\tFind(&transactions)\\n\\t} else {\\n\\t\\tquery.\\n\\t\\t\\tWhere(\\\"creator_id = ?\\\", curUserID).\\n\\t\\t\\tFind(&transactions)\\n\\t}\\n\\n\\t// Get creator and relatedUser\\n\\t// TODO: n + 1 query problem here, so we'll figure this out later\\n\\tfor i := range transactions {\\n\\t\\tdatabase.DBCon.First(&transactions[i].Recipient, transactions[i].RecipientID)\\n\\t\\tdatabase.DBCon.First(&transactions[i].Sender, transactions[i].SenderID)\\n\\t\\tdatabase.DBCon.First(&transactions[i].Creator, transactions[i].CreatorID)\\n\\t}\\n\\n\\tdata, err := jsonapi.Marshal(transactions)\\n\\n\\tif err != nil {\\n\\t\\tc.AbortWithError(http.StatusInternalServerError, err).\\n\\t\\t\\tSetMeta(appError.JSONParseFailure)\\n\\t\\treturn\\n\\t}\\n\\n\\tc.Data(http.StatusOK, \\\"application/vnd.api+json\\\", data)\\n}\",\n \"func displayTransferDetails(xmlMessage string){\\n // Replace all &quot; with single quote\\n strings.ReplaceAll(xmlMessage, \\\"&quot;\\\", \\\"'\\\")\\n // Create an parsed XML document\\n doc, err := xmlquery.Parse(strings.NewReader(xmlMessage))\\n if err != nil {\\n panic(err)\\n }\\n\\n // Get required 'transaction' element from Xml message\\n transaction := xmlquery.FindOne(doc, \\\"//transaction\\\")\\n if transaction != nil {\\n transferId := transaction.SelectAttr(\\\"ID\\\")\\n if action := transaction.SelectElement(\\\"action\\\"); action != nil {\\n if strings.EqualFold(action.InnerText(),\\\"completed\\\") {\\n // Process transfer complete Xml message\\n var supplementMsg string\\n status := transaction.SelectElement(\\\"status\\\")\\n if status != nil {\\n supplementMsg = status.SelectElement(\\\"supplement\\\").InnerText()\\n fmt.Printf(\\\"\\\\n[%s] TransferID: %s Status: %s\\\\n \\\\tSupplement: %s\\\\n\\\",\\n action.SelectAttr(\\\"time\\\"),\\n strings.ToUpper(transferId),\\n action.InnerText(),\\n supplementMsg)\\n }\\n\\n destAgent := transaction.SelectElement(\\\"destinationAgent\\\")\\n statistics := transaction.SelectElement(\\\"statistics\\\")\\n // Retrieve statistics\\n var actualStartTimeText = \\\"\\\"\\n var retryCount string\\n var numFileFailures string\\n var numFileWarnings string\\n if statistics != nil {\\n actualStartTime := statistics.SelectElement(\\\"actualStartTime\\\")\\n if actualStartTime != nil {\\n actualStartTimeText = actualStartTime.InnerText()\\n }\\n if statistics.SelectElement(\\\"retryCount\\\") != nil {\\n retryCount = statistics.SelectElement(\\\"retryCount\\\").InnerText()\\n }\\n if statistics.SelectElement(\\\"numFileFailures\\\") != nil {\\n numFileFailures = statistics.SelectElement(\\\"numFileFailures\\\").InnerText()\\n }\\n if statistics.SelectElement(\\\"numFileWarnings\\\") != nil {\\n numFileWarnings = statistics.SelectElement(\\\"numFileWarnings\\\").InnerText()\\n }\\n }\\n var elapsedTime time.Duration\\n if actualStartTimeText != \\\"\\\" {\\n startTime := getFormattedTime(actualStartTimeText)\\n completePublishTIme := getFormattedTime(action.SelectAttr(\\\"time\\\"))\\n elapsedTime = completePublishTIme.Sub(startTime)\\n }\\n\\n fmt.Printf(\\\"\\\\tDestination Agent: %s\\\\n\\\\tStart time: %s\\\\n\\\\tCompletion Time: %s\\\\n\\\\tElapsed time: %s\\\\n\\\\tRetry Count: %s\\\\n\\\\tFailures:%s\\\\n\\\\tWarnings:%s\\\\n\\\\n\\\",\\n destAgent.SelectAttr(\\\"agent\\\"),\\n actualStartTimeText,\\n action.SelectAttr(\\\"time\\\"),\\n elapsedTime,\\n retryCount,\\n numFileFailures,\\n numFileWarnings)\\n } else if strings.EqualFold(action.InnerText(),\\\"progress\\\") {\\n // Process transfer progress Xml message\\n destAgent := transaction.SelectElement(\\\"destinationAgent\\\")\\n progressPublishTimeText := action.SelectAttr(\\\"time\\\")\\n fmt.Printf(\\\"\\\\n[%s] %s Status: %s Destination: %s \\\\n\\\", progressPublishTimeText,\\n strings.ToUpper(transferId),\\n action.InnerText(),\\n destAgent.SelectAttr(\\\"agent\\\"))\\n transferSet := transaction.SelectElement(\\\"transferSet\\\")\\n startTimeText := transferSet.SelectAttr(\\\"startTime\\\")\\n //startTime := getFormattedTime(startTimeText)\\n //progressPublishTime := getFormattedTime(progressPublishTimeText)\\n //elapsedTime := progressPublishTime.Sub(startTime)\\n fmt.Printf(\\\"\\\\tStart time: %s\\\\n\\\\tTotal items in transfer request: %s\\\\n\\\\tBytes sent: %s\\\\n\\\",\\n startTimeText,\\n transferSet.SelectAttr(\\\"total\\\"),\\n transferSet.SelectAttr(\\\"bytesSent\\\"))\\n\\n // Loop through all items in the progress message and display details.\\n items := transferSet.SelectElements(\\\"item\\\")\\n for i := 0 ; i < len(items); i++ {\\n status := items[i].SelectElement(\\\"status\\\")\\n resultCode := status.SelectAttr(\\\"resultCode\\\")\\n var sourceName string\\n var sourceSize = \\\"-1\\\"\\n queueSource := items[i].SelectElement(\\\"source/queue\\\")\\n if queueSource != nil {\\n sourceName = queueSource.InnerText()\\n } else {\\n fileName := items[i].SelectElement(\\\"source/file\\\")\\n if fileName != nil {\\n sourceName = fileName.InnerText()\\n sourceSize = fileName.SelectAttr(\\\"size\\\")\\n }\\n }\\n\\n var destinationName string\\n queueDest := items[i].SelectElement(\\\"destination/queue\\\")\\n var destinationSize = \\\"-1\\\"\\n if queueDest != nil {\\n destinationName = queueDest.InnerText()\\n } else {\\n fileName := items[i].SelectElement(\\\"destination/file\\\")\\n if fileName != nil {\\n destinationName = fileName.InnerText()\\n destinationSize = fileName.SelectAttr(\\\"size\\\")\\n }\\n }\\n\\n // Display details of each item\\n fmt.Printf(\\\"\\\\tItem # %d\\\\n\\\\t\\\\tSource: %s\\\\tSize: %s bytes\\\\n\\\\t\\\\tDestination: %s\\\\tSize: %s bytes\\\\n\\\",\\n i+1,\\n sourceName, sourceSize,\\n destinationName, destinationSize)\\n // Process result code and append any supplement\\n if resultCode != \\\"0\\\" {\\n supplement := status.SelectElement(\\\"supplement\\\")\\n if supplement != nil {\\n fmt.Printf(\\\"\\\\t\\\\tResult code %s Supplement %s\\\\n\\\", resultCode, supplement.InnerText())\\n } else {\\n fmt.Printf(\\\"\\\\t\\\\tResult code %s\\\\n\\\", resultCode)\\n }\\n } else {\\n fmt.Printf(\\\"\\\\t\\\\tResult code %s\\\\n\\\", resultCode)\\n }\\n }\\n } else if strings.EqualFold(action.InnerText(),\\\"started\\\") {\\n // Process transfer started Xml message\\n destAgent := transaction.SelectElement(\\\"destinationAgent\\\")\\n destinationAgentName := destAgent.SelectAttr(\\\"agent\\\")\\n transferSet := transaction.SelectElement(\\\"transferSet\\\")\\n startTime := \\\"\\\"\\n if transferSet != nil {\\n startTime = transferSet.SelectAttr(\\\"startTime\\\")\\n } else {\\n startTime = action.SelectAttr(\\\"time\\\")\\n }\\n fmt.Printf(\\\"[%s] TransferID: %s Status: %s Destination: %s\\\\n\\\",\\n startTime,\\n strings.ToUpper(transferId),\\n action.InnerText(),\\n destinationAgentName)\\n }\\n }\\n }\\n}\",\n \"func Show(w http.ResponseWriter, r *http.Request) {\\r\\n\\tdb := dbConn()\\r\\n\\tnId := r.URL.Query().Get(\\\"id\\\")\\r\\n\\tselDB, err := db.Query(\\\"SELECT * FROM Pegawai WHERE id=?\\\", nId)\\r\\n\\tif err != nil {\\r\\n\\t\\tpanic(err.Error())\\r\\n\\t}\\r\\n\\temp := Pegawai{}\\r\\n\\tfor selDB.Next() {\\r\\n\\t\\t//buat variabel untuk menampung data\\r\\n\\t\\t//sesuaikan sama nama kolom database (huruf kecil)\\r\\n\\t\\tvar id int\\r\\n\\t\\tvar nama, alamat, jabatan string\\r\\n\\r\\n\\t\\terr = selDB.Scan(&id, &nama, &alamat, &jabatan)\\r\\n\\t\\tif err != nil {\\r\\n\\t\\t\\tpanic(err.Error())\\r\\n\\t\\t}\\r\\n\\r\\n\\t\\t//kanan nama var struct - kiri nama kolom database yang diinisialisasikan diatas\\r\\n\\t\\temp.Id = id\\r\\n\\t\\temp.Nama = nama\\r\\n\\t\\temp.Alamat = alamat\\r\\n\\t\\temp.Jabatan = jabatan\\r\\n\\t}\\r\\n\\ttmpl.ExecuteTemplate(w, \\\"Show\\\", emp)\\r\\n\\tdefer db.Close()\\r\\n}\",\n \"func (t *Transaction) String() string {\\n\\treturn t.From + \\\" -> \\\" + t.To + \\\" : \\\" + strconv.Itoa(t.Amount)\\n}\",\n \"func (T transaction) String() string {\\n\\treturn fmt.Sprintf(\\\"{\\\\n\\\\t\\\\t\\\\tsender:%v,\\\\n\\\\t\\\\t\\\\treceiver:%v,\\\\n\\\\t\\\\t\\\\tamount:%v\\\\n\\\\t\\\\t}\\\", T.sender, T.receiver, T.amount)\\n}\",\n \"func (t *Transaction) String() string {\\n\\tswitch t.Type {\\n\\tcase TIN, TEP:\\n\\t\\treturn fmt.Sprintf(\\\"%22s {%d} %s %.3f %s [%s]\\\", t.Datetime, t.Index, t.Category.FullName, t.Amount, t.AmountCurrency, t.Type.Name)\\n\\tcase TOB:\\n\\t\\treturn fmt.Sprintf(\\\"%22s {%d} Opening new Account\\\", t.Datetime, t.Index)\\n\\tcase TBJ:\\n\\t\\treturn fmt.Sprintf(\\\"%22s {%d} Update balance by %.3f %s\\\", t.Datetime, t.Index, t.Amount, t.AmountCurrency)\\n\\tcase TMT:\\n\\t\\treturn fmt.Sprintf(\\\"%22s {%d} Move %.3f %s to '%s' (%.3f %s)\\\", t.Datetime, t.Index, t.Amount, t.AmountCurrency, t.AccountTo, t.AmountTo, t.AmountToCurrency)\\n\\tdefault:\\n\\t\\treturn fmt.Sprintf(\\\"%22s {%d} [%s] TODO: DIDN'T IMPLEMENT THIS TYPE YET\\\", t.Datetime, t.Index, t.Type.Name)\\n\\t}\\n}\",\n \"func (_CommitteeManager *CommitteeManagerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\\n\\treturn _CommitteeManager.Contract.CommitteeManagerTransactor.contract.Transact(opts, method, params...)\\n}\",\n \"func ShowTxStatusTracker(stdout io.Writer, hash string, rpcClient types2.Client) error {\\n\\ts := spinner.New(spinner.CharSets[14], 100*time.Millisecond)\\n\\ts.Writer = stdout\\n\\ts.Prefix = \\\" \\\"\\n\\ts.Start()\\n\\tlastStatus := \\\"\\\"\\n\\n\\tvar err error\\n\\tvar resp *api2.ResultTx\\n\\tattempts := 0\\n\\tfor {\\n\\t\\tattempts += 1\\n\\t\\tif attempts == 3 {\\n\\t\\t\\treturn err\\n\\t\\t}\\n\\n\\t\\ttime.Sleep(1 * time.Second)\\n\\t\\tresp, err = api.GetTransaction(hash, rpcClient)\\n\\t\\tif err != nil {\\n\\t\\t\\ts.Stop()\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\n\\t\\tif lastStatus == resp.Status {\\n\\t\\t\\tcontinue\\n\\t\\t}\\n\\n\\t\\tlastStatus = resp.Status\\n\\t\\tif resp.Status == types3.TxStatusInMempool {\\n\\t\\t\\ts.Suffix = colorfmt.YellowStringf(\\\" In mempool\\\")\\n\\t\\t} else if resp.Status == types3.TxStatusInPushpool {\\n\\t\\t\\ts.Suffix = colorfmt.YellowStringf(\\\" In pushpool\\\")\\n\\t\\t} else {\\n\\t\\t\\ts.FinalMSG = colorfmt.GreenString(\\\" Confirmed!\\\\n\\\")\\n\\t\\t\\ts.Stop()\\n\\t\\t\\tbreak\\n\\t\\t}\\n\\t}\\n\\treturn nil\\n}\",\n \"func (res BackTestResult) Show() TradestatPort {\\n\\t//\\tp := NewPortfolio()\\n\\t// mds := bean.NewRPCMDSConnC(\\\"tcp\\\", res.dbhost+\\\":\\\"+res.dbport)\\n\\tratesbook := make(ReferenceRateBook)\\n\\n\\t// FIXME: think about how to show multi pair result\\n\\tvar stat TradestatPort\\n\\tif len(res.pairs) > 0 {\\n\\t\\tp := res.pairs[0]\\n\\t\\t//\\t\\ttxn, _ := mds.GetTransactions2(NameFcoin, p, res.start, res.end)\\n\\t\\tratesbook[p] = RefRatesFromTxn(res.Txn)\\n\\t\\t//\\t\\tsnapts.Print()\\n\\t\\t//\\t\\tperfts.Print()\\n\\n\\t\\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\\n\\t\\tssTS := GenerateSnapshotTS(res.Txn, NewPortfolio())\\n\\t\\tmaxPos := 0.0\\n\\t\\tmaxNeg := 0.0\\n\\t\\tfor _, ss := range ssTS {\\n\\t\\t\\tmaxPos = math.Max(maxPos, ss.Port.Balance(p.Coin))\\n\\t\\t\\tmaxNeg = math.Min(maxNeg, ss.Port.Balance(p.Coin))\\n\\t\\t}\\n\\t\\tfmt.Println(\\\"max position\\\", maxPos, maxNeg)\\n\\t\\tstat.Print()\\n\\t}\\n\\treturn stat\\n}\",\n \"func (ok OKHandler) DeliverTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.DeliverResult, err error) {\\n\\treturn sdk.DeliverResult{Log: ok.Log}, nil\\n}\",\n \"func Index(txn *cheshire.Txn) {\\n\\t//create a context map to be passed to the template\\n\\tcontext := make(map[string]interface{})\\n\\tcontext[\\\"services\\\"] = Servs.RouterTables()\\n\\tcheshire.RenderInLayout(txn, \\\"/index.html\\\", \\\"/template.html\\\", context)\\n}\",\n \"func (c *dummyWavesMDLrpcclient) GetTransaction(txid string) (*model.Transactions, error) {\\n\\ttransaction, _, err := client.NewTransactionsService(c.MainNET).GetTransactionsInfoID(txid)\\n\\treturn transaction, err\\n}\",\n \"func (_VinTracker *VinTrackerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _VinTracker.Contract.contract.Transact(opts, method, params...)\\n}\",\n \"func (action *LedgerShowAction) JSON() {\\n\\tquery := action.Query()\\n\\n\\tif action.Err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\taction.Err = db.Get(action.Ctx, query, &action.Record)\\n\\n\\tif action.Err != nil {\\n\\t\\treturn\\n\\t}\\n\\n\\thal.Render(action.W, NewLedgerResource(action.Record))\\n}\",\n \"func (_CommitteeManager *CommitteeManagerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\\n\\treturn _CommitteeManager.Contract.contract.Transact(opts, method, params...)\\n}\",\n \"func (t *Transaction) String() string {\\n\\tvar builder strings.Builder\\n\\tbuilder.WriteString(\\\"Transaction(\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"id=%v\\\", t.ID))\\n\\tbuilder.WriteString(\\\", sequenceInDay=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.SequenceInDay))\\n\\tbuilder.WriteString(\\\", status=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.Status))\\n\\tbuilder.WriteString(\\\", executedAmount=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.ExecutedAmount))\\n\\tbuilder.WriteString(\\\", executedCurrencyCode=\\\")\\n\\tbuilder.WriteString(t.ExecutedCurrencyCode)\\n\\tbuilder.WriteString(\\\", exchangeRate=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.ExchangeRate))\\n\\tbuilder.WriteString(\\\", originatingAmount=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.OriginatingAmount))\\n\\tbuilder.WriteString(\\\", originatingCurrencyCode=\\\")\\n\\tbuilder.WriteString(t.OriginatingCurrencyCode)\\n\\tbuilder.WriteString(\\\", direction=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.Direction))\\n\\tbuilder.WriteString(\\\", runningBalance=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.RunningBalance))\\n\\tbuilder.WriteString(\\\", createdDate=\\\")\\n\\tbuilder.WriteString(t.CreatedDate.Format(time.ANSIC))\\n\\tbuilder.WriteString(\\\", postedDate=\\\")\\n\\tbuilder.WriteString(t.PostedDate.Format(time.ANSIC))\\n\\tbuilder.WriteString(\\\", executedDate=\\\")\\n\\tbuilder.WriteString(t.ExecutedDate.Format(time.ANSIC))\\n\\tbuilder.WriteString(\\\", updatedDate=\\\")\\n\\tbuilder.WriteString(t.UpdatedDate.Format(time.ANSIC))\\n\\tbuilder.WriteString(\\\", description=\\\")\\n\\tbuilder.WriteString(t.Description)\\n\\tbuilder.WriteString(\\\", memo=\\\")\\n\\tbuilder.WriteString(t.Memo)\\n\\tbuilder.WriteString(\\\", group=\\\")\\n\\tbuilder.WriteString(t.Group)\\n\\tbuilder.WriteString(\\\", type=\\\")\\n\\tbuilder.WriteString(t.Type)\\n\\tbuilder.WriteString(\\\", mainCategory=\\\")\\n\\tbuilder.WriteString(t.MainCategory)\\n\\tbuilder.WriteString(\\\", subCategory=\\\")\\n\\tbuilder.WriteString(t.SubCategory)\\n\\tbuilder.WriteString(\\\", checkNumber=\\\")\\n\\tbuilder.WriteString(t.CheckNumber)\\n\\tbuilder.WriteString(\\\", latitude=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.Latitude))\\n\\tbuilder.WriteString(\\\", longitude=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.Longitude))\\n\\tbuilder.WriteString(\\\", merchantCode=\\\")\\n\\tbuilder.WriteString(t.MerchantCode)\\n\\tbuilder.WriteString(\\\", reversal=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.Reversal))\\n\\tbuilder.WriteString(\\\", reversalFor=\\\")\\n\\tbuilder.WriteString(t.ReversalFor)\\n\\tbuilder.WriteString(\\\", reversed=\\\")\\n\\tbuilder.WriteString(fmt.Sprintf(\\\"%v\\\", t.Reversed))\\n\\tbuilder.WriteString(\\\", reversedBy=\\\")\\n\\tbuilder.WriteString(t.ReversedBy)\\n\\tbuilder.WriteString(\\\", url=\\\")\\n\\tbuilder.WriteString(t.URL)\\n\\tbuilder.WriteByte(')')\\n\\treturn builder.String()\\n}\",\n \"func (controller *AccountController) DoTransaction(ctx *gin.Context) {\\n\\tsourceID, ok := ctx.GetPostForm(\\\"sourceid\\\")\\n\\tif !ok {\\n\\t\\tlog.WithFields(log.Fields{\\\"URL\\\": ctx.Request.URL.String()}).Warn(\\\"No SourceID found in postform\\\")\\n\\n\\t\\terrResp, _ := restapi.NewErrorResponse(\\\"No sourceID given\\\").Marshal()\\n\\t\\tfmt.Fprint(ctx.Writer, string(errResp))\\n\\t\\tctx.Abort()\\n\\t\\treturn\\n\\t}\\n\\n\\ttargetID, ok := ctx.GetPostForm(\\\"targetid\\\")\\n\\tif !ok {\\n\\t\\tlog.WithFields(log.Fields{\\\"URL\\\": ctx.Request.URL.String()}).Warn(\\\"No TargetID found in postform\\\")\\n\\n\\t\\terrResp, _ := restapi.NewErrorResponse(\\\"No targetID given\\\").Marshal()\\n\\t\\tfmt.Fprint(ctx.Writer, string(errResp))\\n\\t\\tctx.Abort()\\n\\t\\treturn\\n\\t}\\n\\n\\tamount, err := strconv.Atoi(ctx.PostForm(\\\"amount\\\"))\\n\\tif err != nil {\\n\\t\\tlog.WithFields(log.Fields{\\\"URL\\\": ctx.Request.URL.String()}).Warn(\\\"No int amount found in postform\\\")\\n\\n\\t\\tresponse, _ := restapi.NewErrorResponse(\\\"No valid diff value\\\").Marshal()\\n\\t\\tfmt.Fprint(ctx.Writer, string(response))\\n\\t\\tctx.Abort()\\n\\t\\treturn\\n\\t}\\n\\n\\tinfo, err := authStuff.GetLoginInfoFromCtx(ctx)\\n\\tif err != nil {\\n\\t\\tresponse, _ := restapi.NewErrorResponse(err.Error()).Marshal()\\n\\t\\tfmt.Fprint(ctx.Writer, string(response))\\n\\t\\tctx.Abort()\\n\\t\\treturn\\n\\t}\\n\\n\\tif err := controller.service.Transaction(sourceID, targetID, info.Name, amount); err == nil {\\n\\t\\tresponse, _ := restapi.NewOkResponse(\\\"\\\").Marshal()\\n\\t\\tfmt.Fprint(ctx.Writer, string(response))\\n\\t\\tctx.Next()\\n\\t} else {\\n\\t\\tlog.WithFields(log.Fields{\\\"user\\\": info.Name}).WithError(err).Error(\\\"Transaction Error\\\")\\n\\n\\t\\tresponse, _ := restapi.NewErrorResponse(err.Error()).Marshal()\\n\\t\\tfmt.Fprint(ctx.Writer, string(response))\\n\\t\\tctx.Abort()\\n\\t\\treturn\\n\\t}\\n}\",\n \"func (_TableManager *TableManagerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\\n\\treturn _TableManager.Contract.TableManagerTransactor.contract.Transact(opts, method, params...)\\n}\",\n \"func (_TableManager *TableManagerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\\n\\treturn _TableManager.Contract.contract.Transact(opts, method, params...)\\n}\",\n \"func (_ArbGasInfo *ArbGasInfoRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _ArbGasInfo.Contract.ArbGasInfoTransactor.contract.Transact(opts, method, params...)\\n}\",\n \"func (_DetailedERC20 *DetailedERC20TransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _DetailedERC20.Contract.contract.Transact(opts, method, params...)\\n}\",\n \"func (_DetailedERC20 *DetailedERC20TransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _DetailedERC20.Contract.contract.Transact(opts, method, params...)\\n}\",\n \"func TransactionScreen(w fyne.Window) fyne.CanvasObject {\\n\\treturn widget.NewTabContainer(\\n\\t\\twidget.NewTabItem(res.GetLocalString(\\\"Transfer\\\"), makeTransferTab(w)),\\n\\t\\twidget.NewTabItem(res.GetLocalString(\\\"Move\\\"), makeMoveTransTab(w)),\\n\\t\\twidget.NewTabItem(res.GetLocalString(\\\"Vote\\\"), makeVoteTab(w)),\\n\\t\\twidget.NewTabItem(res.GetLocalString(\\\"Unvote\\\"), makeUnvoteTab(w)),\\n\\t\\twidget.NewTabItem(res.GetLocalString(\\\"Register Miner\\\"), makeMinerTab(w)),\\n\\t)\\n}\",\n \"func ShowTask(w http.ResponseWriter, r *http.Request, repo *tasks.TaskRepository) {\\n\\tw.Header().Set(\\\"Content-Type\\\", \\\"application/json\\\")\\n\\tparams := mux.Vars(r)\\n\\ttaskID, err := strconv.Atoi(params[\\\"id\\\"])\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\ttask, err := repo.GetTask(taskID)\\n\\tif err != nil {\\n\\t\\tpanic(err)\\n\\t}\\n\\tjson.NewEncoder(w).Encode(apiIndexTask(task))\\n}\",\n \"func (api *DeprecatedApiService) getTransactionIntro(params map[string]string) map[string]string {\\n\\tresult := make(map[string]string)\\n\\n\\t_, is_only_check_exist := params[\\\"only_check_exist\\\"]\\n\\n\\ttrsid, ok1 := params[\\\"id\\\"]\\n\\tif !ok1 {\\n\\t\\tresult[\\\"err\\\"] = \\\"param id must.\\\"\\n\\t\\treturn result\\n\\t}\\n\\n\\tvar trshx []byte\\n\\tif txhx, e := hex.DecodeString(trsid); e == nil && len(txhx) == 32 {\\n\\t\\ttrshx = txhx\\n\\t} else {\\n\\t\\tresult[\\\"err\\\"] = \\\"transaction hash error.\\\"\\n\\t\\treturn result\\n\\t}\\n\\n\\t// Query transaction\\n\\tblkhei, trsresbytes, err := api.blockchain.GetChainEngineKernel().StateRead().ReadTransactionBytesByHash(trshx)\\n\\tif err != nil {\\n\\t\\tresult[\\\"err\\\"] = err.Error()\\n\\t\\treturn result\\n\\t}\\n\\tif trsresbytes == nil {\\n\\t\\tresult[\\\"err\\\"] = \\\"transaction not fond.\\\"\\n\\t\\treturn result\\n\\t}\\n\\n\\t// Whether to just judge whether it exists\\n\\tif is_only_check_exist && len(trsresbytes) > 0 {\\n\\t\\tresult[\\\"ret\\\"] = \\\"0\\\"\\n\\t\\tresult[\\\"exist\\\"] = \\\"yes\\\"\\n\\t\\treturn result\\n\\t}\\n\\n\\ttrsres, _, err := transactions.ParseTransaction(trsresbytes, 0)\\n\\tif err != nil {\\n\\t\\tresult[\\\"err\\\"] = err.Error()\\n\\t\\treturn result\\n\\t}\\n\\n\\t// Resolve actions\\n\\tvar allactions = trsres.GetActionList()\\n\\tvar actions_ary []string\\n\\tvar actions_strings = \\\"\\\"\\n\\tfor _, act := range allactions {\\n\\t\\tvar kind = act.Kind()\\n\\t\\tactstr := fmt.Sprintf(`{\\\"k\\\":%d`, kind)\\n\\t\\tif kind == 1 {\\n\\t\\t\\tacc := act.(*actions.Action_1_SimpleToTransfer)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"to\\\":\\\"%s\\\",\\\"amount\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\tacc.ToAddress.ToReadable(),\\n\\t\\t\\t\\tacc.Amount.ToFinString(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 13 {\\n\\t\\t\\tacc := act.(*actions.Action_13_FromTransfer)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"from\\\":\\\"%s\\\",\\\"amount\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\tacc.FromAddress.ToReadable(),\\n\\t\\t\\t\\tacc.Amount.ToFinString(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 14 {\\n\\t\\t\\tacc := act.(*actions.Action_14_FromToTransfer)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"from\\\":\\\"%s\\\",\\\"to\\\":\\\"%s\\\",\\\"amount\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\tacc.FromAddress.ToReadable(),\\n\\t\\t\\t\\tacc.ToAddress.ToReadable(),\\n\\t\\t\\t\\tacc.Amount.ToFinString(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 2 {\\n\\t\\t\\tacc := act.(*actions.Action_2_OpenPaymentChannel)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\",\\\"left_addr\\\":\\\"%s\\\",\\\"left_amt\\\":\\\"%s\\\",\\\"right_addr\\\":\\\"%s\\\",\\\"right_amt\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\thex.EncodeToString(acc.ChannelId),\\n\\t\\t\\t\\tacc.LeftAddress.ToReadable(),\\n\\t\\t\\t\\tacc.LeftAmount.ToFinString(),\\n\\t\\t\\t\\tacc.RightAddress.ToReadable(),\\n\\t\\t\\t\\tacc.RightAmount.ToFinString(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 3 {\\n\\t\\t\\tacc := act.(*actions.Action_3_ClosePaymentChannel)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\thex.EncodeToString(acc.ChannelId),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 4 {\\n\\t\\t\\tacc := act.(*actions.Action_4_DiamondCreate)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"number\\\":\\\"%d\\\",\\\"name\\\":\\\"%s\\\",\\\"address\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\tacc.Number,\\n\\t\\t\\t\\tacc.Diamond,\\n\\t\\t\\t\\tacc.Address.ToReadable(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 5 {\\n\\t\\t\\tacc := act.(*actions.Action_5_DiamondTransfer)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"count\\\":1,\\\"names\\\":\\\"%s\\\",\\\"from\\\":\\\"%s\\\",\\\"to\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\tacc.ToAddress.ToReadable(),\\n\\t\\t\\t\\tacc.Diamond,\\n\\t\\t\\t\\ttrsres.GetAddress().ToReadable(),\\n\\t\\t\\t\\tacc.ToAddress.ToReadable(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 6 {\\n\\t\\t\\tacc := act.(*actions.Action_6_OutfeeQuantityDiamondTransfer)\\n\\t\\t\\tdmds := make([]string, len(acc.DiamondList.Diamonds))\\n\\t\\t\\tfor i, v := range acc.DiamondList.Diamonds {\\n\\t\\t\\t\\tdmds[i] = string(v)\\n\\t\\t\\t}\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"count\\\":%d,\\\"names\\\":\\\"%s\\\",\\\"from\\\":\\\"%s\\\",\\\"to\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\tacc.DiamondList.Count,\\n\\t\\t\\t\\tstrings.Join(dmds, \\\",\\\"),\\n\\t\\t\\t\\tacc.FromAddress.ToReadable(),\\n\\t\\t\\t\\tacc.ToAddress.ToReadable(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 7 {\\n\\t\\t\\tacc := act.(*actions.Action_7_SatoshiGenesis)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"trs_no\\\":%d,\\\"btc_num\\\":%d,\\\"hac_subsidy\\\":%d,\\\"address\\\":\\\"%s\\\",\\\"lockbls_id\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\tacc.TransferNo,\\n\\t\\t\\t\\tacc.BitcoinQuantity,\\n\\t\\t\\t\\tacc.AdditionalTotalHacAmount,\\n\\t\\t\\t\\tacc.OriginAddress.ToReadable(),\\n\\t\\t\\t\\thex.EncodeToString(actions.GainLockblsIdByBtcMove(uint32(acc.TransferNo))),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 8 {\\n\\t\\t\\tacc := act.(*actions.Action_8_SimpleSatoshiTransfer)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"to\\\":\\\"%s\\\",\\\"amount\\\":%d`,\\n\\t\\t\\t\\tacc.ToAddress.ToReadable(),\\n\\t\\t\\t\\tacc.Amount,\\n\\t\\t\\t)\\n\\t\\t} else if kind == 9 {\\n\\t\\t\\tacc := act.(*actions.Action_9_LockblsCreate)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"lockbls_id\\\":\\\"%s\\\",\\\"amount\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\thex.EncodeToString(acc.LockblsId),\\n\\t\\t\\t\\tacc.TotalStockAmount.ToFinString(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 10 {\\n\\t\\t\\tacc := act.(*actions.Action_10_LockblsRelease)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"lockbls_id\\\":\\\"%s\\\",\\\"amount\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\thex.EncodeToString(acc.LockblsId),\\n\\t\\t\\t\\tacc.ReleaseAmount.ToFinString(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 11 {\\n\\t\\t\\tacc := act.(*actions.Action_11_FromToSatoshiTransfer)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"from\\\":\\\"%s\\\",\\\"to\\\":\\\"%s\\\",\\\"amount\\\":%d`,\\n\\t\\t\\t\\tacc.FromAddress.ToReadable(),\\n\\t\\t\\t\\tacc.ToAddress.ToReadable(),\\n\\t\\t\\t\\tacc.Amount,\\n\\t\\t\\t)\\n\\t\\t} else if kind == 28 {\\n\\t\\t\\tacc := act.(*actions.Action_28_FromSatoshiTransfer)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"from\\\":\\\"%s\\\",\\\"amount\\\":%d`,\\n\\t\\t\\t\\tacc.FromAddress.ToReadable(),\\n\\t\\t\\t\\tacc.Amount,\\n\\t\\t\\t)\\n\\t\\t} else if kind == 12 {\\n\\t\\t\\tacc := act.(*actions.Action_12_ClosePaymentChannelBySetupAmount)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\thex.EncodeToString(acc.ChannelId),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 21 {\\n\\t\\t\\tacc := act.(*actions.Action_21_ClosePaymentChannelBySetupOnlyLeftAmount)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\"`,\\n\\t\\t\\t\\thex.EncodeToString(acc.ChannelId),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 22 {\\n\\t\\t\\tacc := act.(*actions.Action_22_UnilateralClosePaymentChannelByNothing)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\",\\\"assert_address\\\"\\\":\\\"%s\\\",\\\"bill_number\\\"\\\":0`,\\n\\t\\t\\t\\thex.EncodeToString(acc.ChannelId), acc.AssertCloseAddress.ToReadable(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 23 {\\n\\t\\t\\tacc := act.(*actions.Action_23_UnilateralCloseOrRespondChallengePaymentChannelByRealtimeReconciliation)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\",\\\"assert_address\\\":\\\"%s\\\",\\\"bill_number\\\":%d`,\\n\\t\\t\\t\\thex.EncodeToString(acc.Reconciliation.GetChannelId()), acc.AssertAddress.ToReadable(), acc.Reconciliation.GetAutoNumber(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 24 {\\n\\t\\t\\tacc := act.(*actions.Action_24_UnilateralCloseOrRespondChallengePaymentChannelByChannelChainTransferBody)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\",\\\"assert_address\\\":\\\"%s\\\",\\\"bill_number\\\":%d`,\\n\\t\\t\\t\\thex.EncodeToString(acc.ChannelChainTransferTargetProveBody.GetChannelId()), acc.AssertAddress.ToReadable(), acc.ChannelChainTransferTargetProveBody.GetAutoNumber(),\\n\\t\\t\\t)\\n\\t\\t} else if kind == 27 {\\n\\t\\t\\tacc := act.(*actions.Action_27_ClosePaymentChannelByClaimDistribution)\\n\\t\\t\\tactstr += fmt.Sprintf(`,\\\"channel_id\\\":\\\"%s\\\",\\\"assert_address\\\":\\\"any\\\",\\\"bill_number\\\"\\\":\\\"closed\\\"`,\\n\\t\\t\\t\\thex.EncodeToString(acc.ChannelId),\\n\\t\\t\\t)\\n\\t\\t}\\n\\t\\tactstr += \\\"}\\\"\\n\\t\\tactions_ary = append(actions_ary, actstr)\\n\\t}\\n\\tactions_strings = strings.Join(actions_ary, \\\",\\\")\\n\\n\\t// Transaction return data\\n\\ttxaddr := fields.Address(trsres.GetAddress())\\n\\tvar txfee = trsres.GetFee()\\n\\tvar txfeeminergot = trsres.GetFeeOfMinerRealReceived()\\n\\tresult[\\\"jsondata\\\"] = fmt.Sprintf(\\n\\t\\t`{\\\"block\\\":{\\\"height\\\":%d,\\\"timestamp\\\":%d},\\\"type\\\":%d,\\\"address\\\":\\\"%s\\\",\\\"fee\\\":\\\"%s\\\",\\\"feeminergot\\\":\\\"%s\\\",\\\"timestamp\\\":%d,\\\"actioncount\\\":%d,\\\"actions\\\":[%s]`,\\n\\t\\tblkhei,\\n\\t\\ttrsres.GetTimestamp(),\\n\\t\\ttrsres.Type(),\\n\\t\\ttxaddr.ToReadable(), // Primary address\\n\\t\\ttxfee.ToFinString(),\\n\\t\\ttxfeeminergot.ToFinString(),\\n\\t\\ttrsres.GetTimestamp(),\\n\\t\\tlen(allactions),\\n\\t\\tactions_strings,\\n\\t)\\n\\n\\tif _, ok := params[\\\"txbodyhex\\\"]; ok {\\n\\t\\tresult[\\\"jsondata\\\"] += fmt.Sprintf(`,\\\"txbodyhex\\\":\\\"%s\\\"`,\\n\\t\\t\\thex.EncodeToString(trsresbytes))\\n\\t}\\n\\n\\t// Wrap up and return\\n\\tresult[\\\"jsondata\\\"] += \\\"}\\\"\\n\\treturn result\\n}\",\n \"func TransactionHandler(w http.ResponseWriter, r *http.Request) {\\n\\taction := r.URL.Path[len(\\\"/api/transactions\\\"):]\\n\\n\\tlog.Println(\\\"Handling method\\\", r.Method, \\\"with action\\\", action)\\n\\n\\tswitch r.Method {\\n\\tcase \\\"POST\\\":\\n\\t\\tswitch action {\\n\\t\\tcase \\\"\\\": // Create new transaction\\n\\t\\t\\tbody, err := ioutil.ReadAll(r.Body)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorForm, err.Error())\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t\\tvar t TransactionResponse\\n\\n\\t\\t\\terr = json.Unmarshal(body, &t)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t\\tpass := []byte(t.Password)\\n\\n\\t\\t\\ttx, e := createAndBroadcastTx(t.Recipient, *big.NewInt(t.Amount), pass)\\n\\n\\t\\t\\tvar jsonResponse TransactionResponse\\n\\t\\t\\tif e == nil {\\n\\t\\t\\t\\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseOk, Hash: hex.EncodeToString(tx.Hash())}\\n\\t\\t\\t} else {\\n\\t\\t\\t\\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseFailed, ErrorText: e.Error()}\\n\\t\\t\\t}\\n\\n\\t\\t\\tres, err := json.Marshal(jsonResponse)\\n\\t\\t\\tif err != nil {\\n\\t\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\\n\\t\\t\\t\\treturn\\n\\t\\t\\t}\\n\\t\\t\\tfmt.Fprintf(w, string(res))\\n\\t\\tdefault:\\n\\t\\t\\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\\\"No action: \\\", r.Method, action))\\n\\t\\t}\\n\\tcase \\\"GET\\\":\\n\\t\\tswitch action {\\n\\t\\tcase \\\"\\\":\\n\\t\\t\\tvar txs []TransactionJson\\n\\t\\t\\tfor _, tx := range Config.DeserializedTxs {\\n\\t\\t\\t\\ttxs = append(txs, EncodeToFriendlyStruct(tx))\\n\\n\\t\\t\\t}\\n\\n\\t\\t\\tif len(txs) == 0 {\\n\\t\\t\\t\\tfmt.Fprintf(w, string(\\\"[]\\\"))\\n\\t\\t\\t} else {\\n\\n\\t\\t\\t\\tres, err := json.Marshal(txs)\\n\\t\\t\\t\\tif err != nil {\\n\\t\\t\\t\\t\\tfmt.Println(\\\"Nope\\\", err.Error())\\n\\t\\t\\t\\t} else {\\n\\t\\t\\t\\t\\tfmt.Fprintf(w, string(res))\\n\\t\\t\\t\\t}\\n\\t\\t\\t}\\n\\n\\t\\t}\\n\\tdefault:\\n\\t\\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\\\"No action: \\\", r.Method, action))\\n\\n\\t}\\n}\",\n \"func (action *TransactionIndexAction) JSON() {\\n\\taction.Do(\\n\\t\\taction.EnsureHistoryFreshness,\\n\\t\\taction.loadParams,\\n\\t\\taction.checkAllowed,\\n\\t\\taction.ValidateCursorWithinHistory,\\n\\t\\taction.loadRecords,\\n\\t\\taction.loadPage,\\n\\t\\tfunc() {\\n\\t\\t\\thal.Render(action.W, action.Page)\\n\\t\\t},\\n\\t)\\n}\",\n \"func (EchoHandler) DeliverTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.DeliverResult, err error) {\\n\\tdata, err := data.ToWire(tx)\\n\\treturn sdk.DeliverResult{Data: data}, err\\n}\",\n \"func (s *Service) HandleShow(w http.ResponseWriter, r *http.Request) {\\n\\tsubs, err := s.subscriptionRepository.FindOne(\\n\\t\\tr.Context(), s.getResourceID(r), s.getSubscriptionID(r),\\n\\t)\\n\\tif err != nil {\\n\\t\\tstatus := http.StatusInternalServerError\\n\\t\\tif errRepo, ok := err.(flare.SubscriptionRepositoryError); ok && errRepo.NotFound() {\\n\\t\\t\\tstatus = http.StatusNotFound\\n\\t\\t}\\n\\n\\t\\ts.writer.Error(w, \\\"error during subscription search\\\", err, status)\\n\\t\\treturn\\n\\t}\\n\\n\\ts.writer.Response(w, transformSubscription(subs), http.StatusOK, nil)\\n}\",\n \"func (me TSAFPTTransactionID) String() string { return xsdt.String(me).String() }\",\n \"func (u *UseCase) ShowOrder(w http.ResponseWriter, r *http.Request) {\\n\\n\\tmerchant := r.Header.Get(\\\"merchant_id\\\")\\n\\tuuid := mux.Vars(r)[\\\"order_id\\\"]\\n\\n\\torder, err := cache.ShowOrder(merchant, uuid)\\n\\tif err == nil && order != nil {\\n\\t\\trespondWithJSON(w, http.StatusOK, order)\\n\\t\\treturn\\n\\t}\\n\\n\\tvar dbOrders models.OrderPg\\n\\tu.DB.Conn.Table(\\\"orders\\\").Where(\\\"uuid = ?\\\", uuid).First(&dbOrders)\\n\\n\\trespondWithJSON(w, http.StatusOK, dbOrders.Payload)\\n}\",\n \"func tableInfoShow(meta *halproto.TableMetadata) {\\n\\tswitch meta.GetKind() {\\n\\tcase halproto.TableKind_TABLE_INDEX:\\n\\t\\ttableInfoIndexShow(meta)\\n\\tcase halproto.TableKind_TABLE_TCAM:\\n\\t\\ttableInfoTcamShow(meta)\\n\\tcase halproto.TableKind_TABLE_HASH:\\n\\t\\ttableInfoHashShow(meta)\\n\\tcase halproto.TableKind_TABLE_FLOW:\\n\\t\\ttableInfoFlowShow(meta)\\n\\tcase halproto.TableKind_TABLE_MET:\\n\\t\\ttableInfoMetShow(meta)\\n\\tdefault:\\n\\t\\tfmt.Printf(\\\"Invalid table type: %d\\\", meta.GetKind())\\n\\t}\\n}\",\n \"func (r Virtual_Guest) GetActiveTransaction() (resp datatypes.Provisioning_Version1_Transaction, err error) {\\n\\terr = r.Session.DoRequest(\\\"SoftLayer_Virtual_Guest\\\", \\\"getActiveTransaction\\\", nil, &r.Options, &resp)\\n\\treturn\\n}\",\n \"func (t *Trade) View(c echo.Context, id int) (*gorsk.Trade, error) {\\n\\treturn t.tdb.View(t.db, id)\\n}\",\n \"func GetTransaction() gin.HandlerFunc {\\n\\treturn func(c *gin.Context) {\\n\\t\\tsugar, _ := item.New(\\\"Sugar\\\", map[string]float64{\\\"Kabras\\\": 110, \\\"Mumias\\\": 110}, \\\"kg(s)\\\")\\n\\t\\tpurchase, message, err := transaction.New(sugar, map[string]float64{\\\"Nzoia\\\": 150}, 3)\\n\\t\\tc.JSON(\\n\\t\\t\\thttp.StatusOK,\\n\\t\\t\\tGetResponse{GetData{purchase}, message, responseerr.GetStrErr(err)},\\n\\t\\t)\\n\\t}\\n}\",\n \"func (as *AddrServer) HandleTxGet(w http.ResponseWriter, r *http.Request) {\\n\\tw.Header().Set(\\\"Content-Type\\\", \\\"application/json\\\")\\n\\ttxid := mux.Vars(r)[\\\"txid\\\"]\\n\\n\\t// paginate through transactions\\n\\ttxns, err := as.GetRawTransaction(txid)\\n\\tif err != nil {\\n\\t\\tw.WriteHeader(400)\\n\\t\\tw.Write(NewPostError(\\\"error fetching all transactions for address\\\", err))\\n\\t\\treturn\\n\\t}\\n\\tout, _ := json.Marshal(txns.Result)\\n\\tw.Write(out)\\n}\",\n \"func (t *trip) Show(ctx context.Context, id int32) (*model.Trip, error) {\\n\\trow := t.conn.QueryRow(ctx, `SELECT \\n\\t\\t\\t\\ttrips.id, trips.dates, trips.price, origin.name, destination.name\\n\\t\\t\\t\\tFROM trips \\n\\t\\t\\t\\tINNER JOIN cities AS origin ON trips.origin_id = origin.id\\n\\t\\t\\t\\tINNER JOIN cities AS destination ON trips.destination_id = destination.id\\n\\t\\t\\t\\tWHERE trips.id = $1`, id)\\n\\n\\tvar trip model.Trip\\n\\n\\tvar origin, destination model.City\\n\\n\\tif err := row.Scan(&trip.ID, &trip.Dates, &trip.Price, &origin.Name, &destination.Name); err != nil {\\n\\t\\treturn nil, err\\n\\t}\\n\\n\\ttrip.Origin = &origin\\n\\ttrip.Destination = &destination\\n\\n\\treturn &trip, nil\\n}\",\n \"func (core *coreService) TraceTransaction(ctx context.Context, actHash string, config *logger.Config) ([]byte, *action.Receipt, *logger.StructLogger, error) {\\n\\tactInfo, err := core.Action(util.Remove0xPrefix(actHash), false)\\n\\tif err != nil {\\n\\t\\treturn nil, nil, nil, err\\n\\t}\\n\\tact, err := (&action.Deserializer{}).SetEvmNetworkID(core.EVMNetworkID()).ActionToSealedEnvelope(actInfo.Action)\\n\\tif err != nil {\\n\\t\\treturn nil, nil, nil, err\\n\\t}\\n\\tsc, ok := act.Action().(*action.Execution)\\n\\tif !ok {\\n\\t\\treturn nil, nil, nil, errors.New(\\\"the type of action is not supported\\\")\\n\\t}\\n\\ttraces := logger.NewStructLogger(config)\\n\\tctx = protocol.WithVMConfigCtx(ctx, vm.Config{\\n\\t\\tDebug: true,\\n\\t\\tTracer: traces,\\n\\t\\tNoBaseFee: true,\\n\\t})\\n\\taddr, _ := address.FromString(address.ZeroAddress)\\n\\tretval, receipt, err := core.SimulateExecution(ctx, addr, sc)\\n\\treturn retval, receipt, traces, err\\n}\",\n \"func (c *BulkDeletesController) Show(ctx *gin.Context) {\\n\\tid := ctx.Param(\\\"taskID\\\")\\n\\ttask := models.BulkDeleteRunTask{}\\n\\n\\tif err := c.App.GetStore().One(\\\"ID\\\", id, &task); err == storm.ErrNotFound {\\n\\t\\tctx.AbortWithError(404, errors.New(\\\"Bulk delete task not found\\\"))\\n\\t} else if err != nil {\\n\\t\\tctx.AbortWithError(500, err)\\n\\t} else if doc, err := jsonapi.Marshal(&task); err != nil {\\n\\t\\tctx.AbortWithError(500, err)\\n\\t} else {\\n\\t\\tctx.Data(200, MediaType, doc)\\n\\t}\\n}\",\n \"func (gw *Gateway) GetTransaction(txid cipher.SHA256) (*visor.Transaction, error) {\\n\\tvar txn *visor.Transaction\\n\\tvar err error\\n\\n\\tgw.strand(\\\"GetTransaction\\\", func() {\\n\\t\\ttxn, err = gw.v.GetTransaction(txid)\\n\\t})\\n\\n\\treturn txn, err\\n}\",\n \"func (_ArbGasInfo *ArbGasInfoTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _ArbGasInfo.Contract.contract.Transact(opts, method, params...)\\n}\",\n \"func (t *Transaction) DisplayInline() string {\\n\\treturn fmt.Sprintf(\\\"%s - - Operation: %s - Status: %s - Side: %s - Amount(cts): %d\\\", t.ID, t.OperationType, t.Status, t.Side, t.AmountCents)\\n}\",\n \"func tableInfoHashShow(meta *halproto.TableMetadata) {\\n\\thMeta := meta.GetHashMeta()\\n\\tinsStr := fmt.Sprintf(\\\"%d[%d]\\\", hMeta.GetNumInserts(), hMeta.GetNumInsertFailures())\\n\\tupdStr := fmt.Sprintf(\\\"%d[%d]\\\", hMeta.GetNumUpdates(), hMeta.GetNumUpdateFailures())\\n\\tdelStr := fmt.Sprintf(\\\"%d[%d]\\\", hMeta.GetNumDeletes(), hMeta.GetNumDeleteFailures())\\n\\tfmt.Printf(\\\"%-30s%-10d%-10s%-10d%-10d%-10d%-10d%-10s%-10s%-10s\\\\n\\\",\\n\\t\\tmeta.GetTableName(),\\n\\t\\tmeta.GetTableId(),\\n\\t\\tutils.TableKindToStr(meta.GetKind()),\\n\\t\\thMeta.GetCapacity(),\\n\\t\\thMeta.GetOtcamCapacity(),\\n\\t\\thMeta.GetHashUsage(),\\n\\t\\thMeta.GetOtcamUsage(),\\n\\t\\tinsStr,\\n\\t\\tupdStr,\\n\\t\\tdelStr)\\n}\",\n \"func Details(attrs []htmlgo.Attribute, children ...HTML) HTML {\\n\\treturn &htmlgo.Tree{Tag: \\\"details\\\", Attributes: attrs, Children: children}\\n}\",\n \"func (db *DB) View(fn func(*Tx) error) error {\\n\\tt, err := db.Begin(false)\\n\\tif err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\t// Make sure the transaction rolls back in the event of a panic.\\n\\tdefer func() {\\n\\t\\tif t.db != nil {\\n\\t\\t\\tt.rollback()\\n\\t\\t}\\n\\t}()\\n\\n\\t// Mark as a managed tx so that the inner function cannot manually rollback.\\n\\tt.managed = true\\n\\n\\t// If an error is returned from the function then pass it through.\\n\\terr = fn(t)\\n\\tt.managed = false\\n\\tif err != nil {\\n\\t\\t_ = t.Rollback()\\n\\t\\treturn err\\n\\t}\\n\\n\\tif err := t.Rollback(); err != nil {\\n\\t\\treturn err\\n\\t}\\n\\n\\treturn nil\\n}\",\n \"func (HTTPOperation) GetDetails() (string, string, string) {\\n\\treturn \\\"update\\\", \\\"updated\\\", \\\"vaccine availability \\\" + id\\n}\",\n \"func (tc *TransactionsController) Index(c *gin.Context, size, page, offset int) {\\n\\ttxs, count, err := tc.App.GetStore().Transactions(offset, size)\\n\\tptxs := make([]presenters.Tx, len(txs))\\n\\tfor i, tx := range txs {\\n\\t\\tptxs[i] = presenters.NewTx(&tx)\\n\\t}\\n\\tpaginatedResponse(c, \\\"Transactions\\\", size, page, ptxs, count, err)\\n}\",\n \"func (_Onesplitaudit *OnesplitauditTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\\n\\treturn _Onesplitaudit.Contract.contract.Transact(opts, method, params...)\\n}\",\n \"func tableInfoIndexShow(meta *halproto.TableMetadata) {\\n\\tiMeta := meta.GetIndexMeta()\\n\\tinsStr := fmt.Sprintf(\\\"%d[%d]\\\", iMeta.GetNumInserts(), iMeta.GetNumInsertFailures())\\n\\tupdStr := fmt.Sprintf(\\\"%d[%d]\\\", iMeta.GetNumUpdates(), iMeta.GetNumUpdateFailures())\\n\\tdelStr := fmt.Sprintf(\\\"%d[%d]\\\", iMeta.GetNumDeletes(), iMeta.GetNumDeleteFailures())\\n\\tfmt.Printf(\\\"%-30s%-10d%-10s%-10d%-10s%-10d%-10s%-10s%-10s%-10s\\\\n\\\",\\n\\t\\tmeta.GetTableName(),\\n\\t\\tmeta.GetTableId(),\\n\\t\\tutils.TableKindToStr(meta.GetKind()),\\n\\t\\tiMeta.GetCapacity(),\\n\\t\\t\\\"-\\\",\\n\\t\\tiMeta.GetUsage(),\\n\\t\\t\\\"-\\\",\\n\\t\\tinsStr,\\n\\t\\tupdStr,\\n\\t\\tdelStr)\\n}\",\n \"func (resolver *ResolverTODO) Show(params graphql.ResolveParams) (interface{}, error) {\\n\\ttodo, err := resolver.Db.Show(params.Args[\\\"id\\\"].(string))\\n\\n\\tif err != nil {\\n\\t\\treturn nil, err\\n\\t} else if todo.ID == \\\"\\\" {\\n\\t\\treturn nil, errors.New(\\\"todo not found\\\")\\n\\t}\\n\\n\\treturn todo, nil\\n}\"\n]"},"negative_scores":{"kind":"list like","value":["0.7065399","0.6323545","0.6089757","0.58380103","0.5829757","0.5806046","0.5743745","0.5695702","0.5692406","0.5600741","0.55515116","0.55509126","0.55220157","0.54999244","0.53384566","0.53074944","0.52982825","0.52785516","0.5276625","0.52471644","0.5238717","0.52065337","0.51978594","0.51897407","0.51788306","0.5154507","0.5149336","0.5148842","0.5148842","0.5124913","0.5117847","0.5116096","0.50937843","0.5069765","0.5049436","0.50435567","0.5006893","0.49923867","0.4981022","0.49754423","0.4921016","0.4898289","0.4889206","0.48864287","0.48829022","0.48819882","0.486109","0.48377797","0.48292157","0.48292157","0.4825918","0.48157176","0.48137054","0.48055518","0.48034588","0.480199","0.47995147","0.47936705","0.47911152","0.47875905","0.47867686","0.47867665","0.47850013","0.4778883","0.47752193","0.47638208","0.47636342","0.47625333","0.47538182","0.4751926","0.47489476","0.47489476","0.47414246","0.47372147","0.4732912","0.47310746","0.47151774","0.4705625","0.47027025","0.46943593","0.46935827","0.4692121","0.4691574","0.46858248","0.4676262","0.46733868","0.46709293","0.46631366","0.46594134","0.46525756","0.46511626","0.46491662","0.46372992","0.46351305","0.4628336","0.46106213","0.46100897","0.46069738","0.46058023","0.4605689"],"string":"[\n \"0.7065399\",\n \"0.6323545\",\n \"0.6089757\",\n \"0.58380103\",\n \"0.5829757\",\n \"0.5806046\",\n \"0.5743745\",\n \"0.5695702\",\n \"0.5692406\",\n \"0.5600741\",\n \"0.55515116\",\n \"0.55509126\",\n \"0.55220157\",\n \"0.54999244\",\n \"0.53384566\",\n \"0.53074944\",\n \"0.52982825\",\n \"0.52785516\",\n \"0.5276625\",\n \"0.52471644\",\n \"0.5238717\",\n \"0.52065337\",\n \"0.51978594\",\n \"0.51897407\",\n \"0.51788306\",\n \"0.5154507\",\n \"0.5149336\",\n \"0.5148842\",\n \"0.5148842\",\n \"0.5124913\",\n \"0.5117847\",\n \"0.5116096\",\n \"0.50937843\",\n \"0.5069765\",\n \"0.5049436\",\n \"0.50435567\",\n \"0.5006893\",\n \"0.49923867\",\n \"0.4981022\",\n \"0.49754423\",\n \"0.4921016\",\n \"0.4898289\",\n \"0.4889206\",\n \"0.48864287\",\n \"0.48829022\",\n \"0.48819882\",\n \"0.486109\",\n \"0.48377797\",\n \"0.48292157\",\n \"0.48292157\",\n \"0.4825918\",\n \"0.48157176\",\n \"0.48137054\",\n \"0.48055518\",\n \"0.48034588\",\n \"0.480199\",\n \"0.47995147\",\n \"0.47936705\",\n \"0.47911152\",\n \"0.47875905\",\n \"0.47867686\",\n \"0.47867665\",\n \"0.47850013\",\n \"0.4778883\",\n \"0.47752193\",\n \"0.47638208\",\n \"0.47636342\",\n \"0.47625333\",\n \"0.47538182\",\n \"0.4751926\",\n \"0.47489476\",\n \"0.47489476\",\n \"0.47414246\",\n \"0.47372147\",\n \"0.4732912\",\n \"0.47310746\",\n \"0.47151774\",\n \"0.4705625\",\n \"0.47027025\",\n \"0.46943593\",\n \"0.46935827\",\n \"0.4692121\",\n \"0.4691574\",\n \"0.46858248\",\n \"0.4676262\",\n \"0.46733868\",\n \"0.46709293\",\n \"0.46631366\",\n \"0.46594134\",\n \"0.46525756\",\n \"0.46511626\",\n \"0.46491662\",\n \"0.46372992\",\n \"0.46351305\",\n \"0.4628336\",\n \"0.46106213\",\n \"0.46100897\",\n \"0.46069738\",\n \"0.46058023\",\n \"0.4605689\"\n]"},"document_score":{"kind":"string","value":"0.59414774"},"document_rank":{"kind":"string","value":"3"}}}],"truncated":false,"partial":true},"paginationData":{"pageIndex":1047,"numItemsPerPage":100,"numTotalItems":107511,"offset":104700,"length":100}},"jwt":"eyJhbGciOiJFZERTQSJ9.eyJyZWFkIjp0cnVlLCJwZXJtaXNzaW9ucyI6eyJyZXBvLmNvbnRlbnQucmVhZCI6dHJ1ZX0sImlhdCI6MTc1ODIyNzM1OCwic3ViIjoiL2RhdGFzZXRzL25vbWljLWFpL2Nvcm5zdGFjay1nby12MSIsImV4cCI6MTc1ODIzMDk1OCwiaXNzIjoiaHR0cHM6Ly9odWdnaW5nZmFjZS5jbyJ9.6Py_es5g7uKPF43s3Jbl3nJLb59lI4iXgkeZQNaEsXh951qAzZ69TtZiH9zT4ilerjvtVhOin0H0HguUKsGSBQ","displayUrls":true},"discussionsStats":{"closed":0,"open":0,"total":0},"fullWidth":true,"hasGatedAccess":true,"hasFullAccess":true,"isEmbedded":false,"savedQueries":{"community":[],"user":[]}}">
query
stringlengths
7
3.85k
document
stringlengths
11
430k
metadata
dict
negatives
listlengths
0
101
negative_scores
listlengths
0
101
document_score
stringlengths
3
10
document_rank
stringclasses
102 values
encodeGRPCConcatResponse is a transport/grpc.EncodeResponseFunc that converts a userdomain concat response to a gRPC concat reply. Primarily useful in a server.
func encodeGRPCNameResponse(_ context.Context, response interface{}) (interface{}, error) { resp := response.(loginendpoint.LoginResponse) return &pb.NameReply{V: resp.V, Err: err2str(resp.Err)}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func encodeGRPCConcatResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\treply := grpcReply.(endpoints.ConcatResponse)\n\treturn &pb.ConcatReply{Rs: reply.Rs}, def.GrpcEncodeError(reply.Err)\n}", "func encodeGRPCConcatRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(endpoints.ConcatRequest)\n\treturn &pb.ConcatRequest{A: req.A, B: req.B}, nil\n}", "func decodeGRPCConcatResponse(_ context.Context, grpcReply interface{}) (interface{}, error) {\n\treply := grpcReply.(*pb.ConcatReply)\n\treturn endpoints.ConcatResponse{Rs: reply.Rs}, nil\n}", "func EncodeGRPCResponse(_ context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeGRPCConcatRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(loginendpoint.LoginRequest)\n\treturn &pb.NameRequest{N: req.N}, nil\n}", "func decodeGRPCConcatRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.ConcatRequest)\n\treturn endpoints.ConcatRequest{A: req.A, B: req.B}, nil\n}", "func encodeGRPCGenerateReportResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(generateReportResponse)\n\treturn &pb.GenerateReportReply{V: int64(resp.V), Err: err2str(resp.Err)}, nil\n}", "func encodeGRPCSumResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\treply := grpcReply.(endpoints.SumResponse)\n\treturn &pb.SumReply{Rs: reply.Rs}, def.GrpcEncodeError(reply.Err)\n}", "func encodeGRPCGreetingResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(greeterendpoint.GreetingResponse)\n\treturn &pb.GreetingResponse{Greeting: res.Greeting}, nil\n}", "func encodeGRPCEchoResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\n\treply := grpcReply.(endpoints.EchoResponse)\n\treturn &pb.EchoReply{Rs: reply.Rs}, def.GrpcEncodeError(reply.Err)\n}", "func EncodeGrpcRespResultSeries(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeGrpcRespWorkload(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeGRPCQueryResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.QueryResponse)\n\treturn resp, nil\n}", "func encodeGRPCTacResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\treply := grpcReply.(endpoints.TacResponse)\n\treturn &pb.TacResponse{Res: reply.Res}, grpcEncodeError(errors.Cast(reply.Err))\n}", "func encodeGRPCTicResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\treply := grpcReply.(endpoints.TicResponse)\n\treturn &pb.TicResponse{}, grpcEncodeError(errors.Cast(reply.Err))\n}", "func EncodeGRPCLogoutResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.LogoutResponse)\n\treturn resp, nil\n}", "func EncodeGRPCNewUserResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresp := r.(newUserResponse)\n\treturn &pb.NewUserResponse{\n\t\tId: resp.Id,\n\t\tErr: resp.Err,\n\t}, nil\n}", "func encodeRegisterResponse(_ context.Context, r interface{}) (interface{}, error) {\n\trs := r.(endpoint.RegisterResponse)\n\n\tif rs.Err != nil {\n\t\treturn &pb.RegisterReply{\n\t\t\tMessage: rs.Err.Error(),\n\t\t\tStatus: \"ERROR\",\n\t\t}, fmt.Errorf(\"Message: %v With Status: %v\", rs.Err.Error(), rs.Err)\n\t}\n\n\treturn &pb.RegisterReply{\n\t\tMessage: fmt.Sprintf(\"Hi %s %s We Send a SMS for verify your Phone!\", rs.Response.Name, rs.Response.LastName),\n\t\tStatus: \"SUCCESS\",\n\t}, nil\n}", "func encodeUpdateCompanyResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(endpoints.UpdateCompanyResponse)\n\terr := getError(res.Err)\n\tif err == nil {\n\t\treturn res.Company.ToProto(), nil\n\t}\n\treturn nil, err\n}", "func encodeGRPCMathOpResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(mathendpoint2.MathOpResponse)\n\treturn &pb.MathOpReply{V: resp.V, Err: err2str(resp.Err)}, nil\n}", "func EncodeGRPCGetUserConfigResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.UserCofing)\n\treturn resp, nil\n}", "func EncodeGrpcRespRDSpec(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeGrpcRespFeature(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeGrpcRespDSCProfile(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func decodeHTTPConcatResponse(_ context.Context, r *http.Response) (interface{}, error) {\n\tif r.StatusCode != http.StatusOK {\n\t\treturn nil, JSONErrorDecoder(r)\n\t}\n\tvar resp endpoints.ConcatResponse\n\terr := json.NewDecoder(r.Body).Decode(&resp)\n\treturn resp, err\n}", "func encodeGRPCSayHelloResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\treply := grpcReply.(endpoints.SayHelloResponse)\n\treturn &pb.SayHelloReply{Rs: reply.Rs}, def.GrpcEncodeError(reply.Err)\n}", "func EncodeGRPCUpdateUserInfoResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.ErrCode)\n\treturn resp, nil\n}", "func EncodeGRPCGetUserByEmailResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresp := r.(getUserByEmailResponse)\n\tu := &pb.User{\n\t\tId: resp.User.Id,\n\t\tName: resp.User.Name,\n\t\tLastName: resp.User.LastName,\n\t\tEmail: resp.User.Email,\n\t\tRole: resp.User.Role,\n\t}\n\treturn &pb.GetUserByEmailResponse{\n\t\tUser: u,\n\t\tErr: resp.Err,\n\t}, nil\n}", "func EncodeGRPCRefreshTokenResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.RefreshTokenResp)\n\treturn resp, nil\n}", "func EncodeGRPCChangePasswordResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresp := r.(changePasswordResponse)\n\treturn &pb.ChangePasswordResponse{\n\t\tSuccess: resp.Success,\n\t\tErr: resp.Err,\n\t}, nil\n}", "func encodeRegisterResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresponse := r.(endpoint.RegisterResponse)\n\tif response.U0 == nil {\n\t\treturn nil,response.E1\n\t}\n\t//进行数据的转换\n\tvar user = &pb.UserInfo{\n\t\tId:response.U0.Id,\n\t\tPhone:response.U0.Phone,\n\t\tPassword:response.U0.Password,\n\t\tAge:response.U0.Age,\n\t}\n\treturn &pb.RegisterReply{\n\t\tUser:user,\n\t},response.E1\n\n}", "func encodeGetAllCompaniesResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(endpoints.GetAllCompaniesResponse)\n\terr := getError(res.Err)\n\tif err == nil {\n\t\tvar companies []*pb.JobCompany\n\t\tfor _, company := range res.Companies {\n\t\t\tcompanies = append(companies, company.ToProto())\n\t\t}\n\t\treturn &pb.GetAllJobCompaniesResponse{Companies: companies}, nil\n\t}\n\treturn nil, err\n}", "func EncodeGrpcRespRoute(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeCreateCompanyResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(endpoints.CreateCompanyResponse)\n\terr := getError(res.Err)\n\tif err == nil {\n\t\treturn res.Company.ToProto(), nil\n\t}\n\treturn nil, err\n}", "func EncodeGRPCGetUserInfoResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.UserInfo)\n\treturn resp, nil\n}", "func EncodeAddResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) kithttp.EncodeResponseFunc {\n\treturn server.EncodeAddResponse(encoder)\n}", "func EncodeGRPCLoginResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.LoginResponse)\n\treturn resp, nil\n}", "func EncodeGRPCStatusResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.StatusResponse)\n\treturn resp, nil\n}", "func EncodeGRPCDistanceResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.DistanceResponse)\n\treturn resp, nil\n}", "func EncodeGRPCPingResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.PingResponse)\n\treturn resp, nil\n}", "func EncodeGrpcRespMetricsQueryResponse(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeResponse(resp *plugin.CodeGeneratorResponse, w io.Writer) {\n\toutBytes, err := proto.Marshal(resp)\n\tif err != nil {\n\t\tlog.Fatal(\"unable to marshal response to protobuf: \" + err.Error())\n\t}\n\n\tif _, err := w.Write(outBytes); err != nil {\n\t\tlog.Fatal(\"unable to write protobuf to stdout: \" + err.Error())\n\t}\n}", "func (c *Client) Concat(ctx context.Context, p *ConcatPayload) (res string, err error) {\n\tvar ires any\n\tires, err = c.ConcatEndpoint(ctx, p)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn ires.(string), nil\n}", "func EncodeGrpcRespPaginationSpec(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeAddResponse(ctx context.Context, w http1.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func encodeAddResponse(ctx context.Context, w http1.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func EncodeGRPCGetCodeResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.GetCodeResp)\n\treturn resp, nil\n}", "func encodeGRPCGenerateReportRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(generateReportRequest)\n\treturn &pb.GenerateReportRequest{A: int64(req.A), B: int64(req.B)}, nil\n}", "func EncodeGrpcReqMetricsQueryResponse(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQueryResponse)\n\treturn req, nil\n}", "func EncodeGRPCAllowConnectionResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(AllowConnectionResponse)\n\tgRPCRes := &pb.AllowConnectionResponse{}\n\tif res.Error != nil {\n\t\tgRPCRes.Error = res.Error.Error()\n\t}\n\treturn gRPCRes, nil\n}", "func EncodeGrpcRespDSCProfileSpec(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeGrpcReqResultSeries(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*ResultSeries)\n\treturn req, nil\n}", "func EncodeGRPCBlockConnectionResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(BlockConnectionResponse)\n\tgRPCRes := &pb.BlockConnectionResponse{}\n\tif res.Error != nil {\n\t\tgRPCRes.Error = res.Error.Error()\n\t}\n\treturn gRPCRes, nil\n}", "func encodeGRPCSumRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(endpoints.SumRequest)\n\t\n\treturn &pb.SumRequest{A: req.A, B: req.B}, nil\n}", "func EncodeGRPCLoginResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresp := r.(loginResponse)\n\treturn &pb.LoginResponse{\n\t\tToken: resp.Token,\n\t\tRefreshToken: resp.RefreshToken,\n\t\tErr: resp.Err,\n\t}, nil\n}", "func EncodeGRPCBanTokenResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.Empty)\n\treturn resp, nil\n}", "func EncodeGRPCSetUserConfigResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.ErrCode)\n\treturn resp, nil\n}", "func encodeHTTPConcatRequest(_ context.Context, r *http.Request, request interface{}) (err error) {\n\tvar buf bytes.Buffer\n\tif err := json.NewEncoder(&buf).Encode(request); err != nil {\n\t\treturn err\n\t}\n\tr.Body = ioutil.NopCloser(&buf)\n\treturn nil\n}", "func EncodeGrpcRespDSCProfileStatus(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeDeleteCompanyResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(endpoints.DeleteCompanyResponse)\n\terr := getError(res.Err)\n\tif err == nil {\n\t\treturn &pb.DeleteJobCompanyResponse{}, nil\n\t}\n\treturn nil, err\n}", "func EncodeGrpcRespWorkloadSpec(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeAddResponse(_ context.Context, w http.ResponseWriter, response interface{}) (err error) {\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\te := json.NewEncoder(w)\n\te.SetIndent(\"\", \"\\t\")\n\terr = e.Encode(response)\n\treturn err\n}", "func EncodeGrpcRespRouteDistinguisher(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeGetUserResponse(ctx context.Context, w http1.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func EncodeGrpcRespWorkloadStatus(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeGetAllJobFunctionsResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(endpoints.GetAllJobFunctionsResponse)\n\terr := getError(res.Err)\n\tif err == nil {\n\t\tvar jobFunctions []*pb.JobFunction\n\t\tfor _, jobFunction := range res.JobFunctions {\n\t\t\tjobFunctions = append(jobFunctions, jobFunction.ToProto())\n\t\t}\n\t\treturn &pb.GetAllJobFunctionsResponse{JobFunctions: jobFunctions}, nil\n\t}\n\treturn nil, err\n}", "func EncodeGRPCGetTokenResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.GetTokenResp)\n\treturn resp, nil\n}", "func EncodeGRPCLoginResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.Session)\n\treturn resp, nil\n}", "func EncodeGRPCExecuteRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.ExecuteRequest)\n\treturn &containerPB.ExecuteRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t\tCmd: req.CMD,\n\t}, nil\n}", "func MarshalConcat(va interface{}, vbs ...interface{}) ([]byte, error) {\n\tunique, err := marshalConcat(nil, va, vbs...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn json.Marshal(unique)\n}", "func EncodeTxResponse(res *MsgEthereumTxResponse) ([]byte, error) {\n\treturn proto.Marshal(res)\n}", "func EncodeGrpcRespBGPAuthStatus(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeConfirmResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresp := r.(ConfirmResponse)\n\treturn &pb.ConfirmResponse{\n\t\tOk: resp.Ok,\n\t}, nil\n}", "func EncodeGrpcRespLicense(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeGRPCAllowPortResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(AllowPortResponse)\n\tgRPCRes := &pb.AllowPortResponse{}\n\tif res.Error != nil {\n\t\tgRPCRes.Error = res.Error.Error()\n\t}\n\treturn gRPCRes, nil\n}", "func EncodeGrpcRespBGPConfig(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func EncodeGrpcRespLicenseSpec(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func DecodeBatchGRPCResponse(ctx context.Context, v interface{}, hdr, trlr metadata.MD) (interface{}, error) {\n\tmessage, ok := v.(*publicpb.BatchGRPCResponse)\n\tif !ok {\n\t\treturn nil, goagrpc.ErrInvalidType(\"public\", \"batchGRPC\", \"*publicpb.BatchGRPCResponse\", v)\n\t}\n\tres := NewBatchGRPCResult(message)\n\treturn res, nil\n}", "func EncodeGrpcRespFeatureStatus(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeCreateJobFunctionResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(endpoints.CreateJobFunctionResponse)\n\terr := getError(res.Err)\n\tif err == nil {\n\t\treturn res.JobFunction.ToProto(), nil\n\t}\n\treturn nil, err\n}", "func EncodeGRPCBlockPortResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(BlockPortResponse)\n\tgRPCRes := &pb.BlockPortResponse{}\n\tif res.Error != nil {\n\t\tgRPCRes.Error = res.Error.Error()\n\t}\n\treturn gRPCRes, nil\n}", "func UnaryCall(request *connectorpb.UnaryCallRequest) *connectorpb.UnaryCallResponse {\n\tif grpcServer == nil {\n\t\treturn makeErrorResponse(codes.Unavailable, \"call the initialize function first\")\n\t}\n\n\tr, err := makeHTTPRequest(request)\n\tif err != nil {\n\t\treturn makeErrorResponse(codes.Internal, \"could not prepare an HTTP request\")\n\t}\n\n\tw := &BodyWriter{\n\t\theader: make(http.Header),\n\t}\n\n\tgrpcServer.ServeHTTP(w, r)\n\n\t// While the server sets common HTTP header (e.g., Content-Type) we only care\n\t// about Grpc-Status and Grpc-Message.\n\tgrpcStatus := codes.Internal\n\tgrpcMessage := \"grpc-message field not found\"\n\tfor h, v := range w.header {\n\t\tif len(v) == 1 {\n\t\t\tif strings.ToLower(h) == \"grpc-status\" {\n\t\t\t\ti, err := strconv.Atoi(v[0])\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn makeErrorResponse(codes.Internal, \"could not parse Grpc-Status header\")\n\t\t\t\t}\n\t\t\t\tgrpcStatus = codes.Code((i))\n\t\t\t}\n\t\t\tif strings.ToLower(h) == \"grpc-message\" {\n\t\t\t\tgrpcMessage = v[0]\n\t\t\t}\n\t\t}\n\t}\n\n\tif grpcStatus != codes.OK {\n\t\treturn makeErrorResponse(grpcStatus, grpcMessage)\n\t}\n\n\t// Parse the length-prefixed message response body:\n\t// <1 byte compression flag><4 byte big-endian length><serialized response>\n\tlpm := w.b.Bytes()\n\tif len(lpm) < 5 {\n\t\treturn makeErrorResponse(codes.Internal, \"missing response body\")\n\t}\n\n\tlength := (int(lpm[1]) << 24) | (int(lpm[2]) << 16) | (int(lpm[3]) << 8) | int(lpm[4])\n\tif len(lpm) < 5+length {\n\t\treturn makeErrorResponse(codes.Internal, \"truncated response body\")\n\t}\n\n\treturn connectorpb.UnaryCallResponse_builder{\n\t\tResponse: lpm[5:],\n\t\tStatus: &statuspb.Status{\n\t\t\tCode: int32(codes.OK),\n\t\t\tMessage: grpcMessage,\n\t\t},\n\t}.Build()\n}", "func EncodeDoublySecureResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) func(context.Context, http.ResponseWriter, any) error {\n\treturn func(ctx context.Context, w http.ResponseWriter, v any) error {\n\t\tres, _ := v.(string)\n\t\tenc := encoder(ctx, w)\n\t\tbody := res\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn enc.Encode(body)\n\t}\n}", "func EncodeFetcherResponse(ctx context.Context, w http.ResponseWriter, response interface{}) error {\n\tfetcherResponse, ok := response.(FetchResponser)\n\tif !ok {\n\t\t//return errors.New(\"invalid Fetcher Response\")\n\t\te := errs.BadGateway{What: \"response\"}\n\t\tencodeError(ctx, &e, w)\n\t\treturn nil\n\t}\n\t//statusCode := fetcherResponse.GetStatusCode()\n\t//if statusCode != 200 {\n\t//\treturn errors.New(strconv.Itoa(statusCode))\n\t//}\n\t//if fetcherResponse.Error != \"\" {\n\t//\treturn errors.New(fetcherResponse.Error)\n\t//}\n\n\tdata, err := json.Marshal(fetcherResponse)\n\tif err != nil {\n\t\tencodeError(ctx, err, w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\t_, err = w.Write(data)\n\n\tif err != nil {\n\t\tencodeError(ctx, err, w)\n\t\treturn nil\n\t}\n\treturn nil\n}", "func EncodeGRPCInitBridgeResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(InitBridgeResponse)\n\tgRPCRes := &pb.InitBridgeResponse{}\n\tif res.Error != nil {\n\t\tgRPCRes.Error = res.Error.Error()\n\t}\n\treturn gRPCRes, nil\n}", "func EncodeGrpcRespEVPNConfig(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeUploadResponse(ctx context.Context, w http.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\tif err != nil {\n\t\t\tlogrus.Warn(err.Error())\n\t\t}\n\treturn\n}", "func (ad *stubAddService) Concat(ctx context.Context, a string, b string) (res string, err error) {\n\tres = a + b\n\terr = ad.nc.Publish(addTopic, []byte(res))\n\tif err != nil {\n\t\tlevel.Error(ad.logger).Log(\"method\", \"ad.nc.Publish\", \"value\", res, \"err\", err)\n\t\treturn\n\t}\n\n\treturn res, err\n}", "func DecodeRegisterResponse(ctx context.Context, v interface{}, hdr, trlr metadata.MD) (interface{}, error) {\n\tmessage, ok := v.(*user_methodpb.RegisterResponse)\n\tif !ok {\n\t\treturn nil, goagrpc.ErrInvalidType(\"userMethod\", \"register\", \"*user_methodpb.RegisterResponse\", v)\n\t}\n\tres := NewRegisterResult(message)\n\treturn res, nil\n}", "func EncodeSecureResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) func(context.Context, http.ResponseWriter, any) error {\n\treturn func(ctx context.Context, w http.ResponseWriter, v any) error {\n\t\tres, _ := v.(string)\n\t\tenc := encoder(ctx, w)\n\t\tbody := res\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn enc.Encode(body)\n\t}\n}", "func EncodeGrpcRespMetricsQueryResult(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeCreateUserResponse(ctx context.Context, w http1.ResponseWriter, response interface{}) (err error) {\n\tif f, ok := response.(endpoint.Failure); ok && f.Failed() != nil {\n\t\tErrorEncoder(ctx, f.Failed(), w)\n\t\treturn nil\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\terr = json.NewEncoder(w).Encode(response)\n\treturn\n}", "func EncodeGrpcRespLicenseStatus(ctx context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeResponse(ctx context.Context, w http.ResponseWriter, response interface{}) (err error) {\n\tresp := response.(*common.XmidtResponse)\n\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.Header().Set(common.HeaderWPATID, ctx.Value(common.ContextKeyRequestTID).(string))\n\tcommon.ForwardHeadersByPrefix(\"\", resp.ForwardedHeaders, w.Header())\n\n\tw.WriteHeader(resp.Code)\n\t_, err = w.Write(resp.Body)\n\treturn\n}", "func encodeVerifyResponse(_ context.Context, r interface{}) (interface{}, error) {\n\trs := r.(endpoint.VerifyResponse)\n\n\tif rs.Err != nil {\n\t\treturn &pb.VerifyReply{}, fmt.Errorf(\"Error: %s \", rs.Err.Error())\n\t}\n\n\tpermissions := make([]*pb.Permission, len(rs.Response.Role.Permissions))\n\tfor _, v := range rs.Response.Role.Permissions {\n\t\tpermissions = append(permissions, &pb.Permission{Name: v.Name})\n\t}\n\n\treturn &pb.VerifyReply{\n\t\tUsername: rs.Response.Username,\n\t\tName: rs.Response.Name,\n\t\tLastName: rs.Response.LastName,\n\t\tPhone: rs.Response.Phone,\n\t\tEmail: rs.Response.Email,\n\t\tToken: rs.Response.Token,\n\t\tRole: &pb.Role{\n\t\t\tName: rs.Response.Role.Name,\n\t\t\tPermissions: permissions,\n\t\t},\n\t}, nil\n}", "func EncodeRegisterResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) func(context.Context, http.ResponseWriter, interface{}) error {\n\treturn func(ctx context.Context, w http.ResponseWriter, v interface{}) error {\n\t\tres := v.(int)\n\t\tenc := encoder(ctx, w)\n\t\tbody := res\n\t\tw.WriteHeader(http.StatusCreated)\n\t\treturn enc.Encode(body)\n\t}\n}", "func DecodeStreamedBatchGRPCResponse(ctx context.Context, v interface{}, hdr, trlr metadata.MD) (interface{}, error) {\n\treturn &StreamedBatchGRPCClientStream{\n\t\tstream: v.(publicpb.Public_StreamedBatchGRPCClient),\n\t}, nil\n}", "func EncodeAlsoDoublySecureResponse(encoder func(context.Context, http.ResponseWriter) goahttp.Encoder) func(context.Context, http.ResponseWriter, any) error {\n\treturn func(ctx context.Context, w http.ResponseWriter, v any) error {\n\t\tres, _ := v.(string)\n\t\tenc := encoder(ctx, w)\n\t\tbody := res\n\t\tw.WriteHeader(http.StatusOK)\n\t\treturn enc.Encode(body)\n\t}\n}", "func encodeLoginUPResponse(_ context.Context, r interface{}) (interface{}, error) {\n\trs := r.(endpoint.LoginUPResponse)\n\n\tif rs.Err != nil {\n\t\treturn &pb.LoginUPReply{}, fmt.Errorf(\"Error: %s \", rs.Err.Error())\n\t}\n\n\tpermissions := make([]*pb.Permission, len(rs.Response.Role.Permissions))\n\tfor _, v := range rs.Response.Role.Permissions {\n\t\tpermissions = append(permissions, &pb.Permission{Name: v.Name})\n\t}\n\n\treturn &pb.LoginUPReply{\n\t\tUsername: rs.Response.Username,\n\t\tName: rs.Response.Name,\n\t\tLastName: rs.Response.LastName,\n\t\tPhone: rs.Response.Phone,\n\t\tEmail: rs.Response.Email,\n\t\tToken: rs.Response.Token,\n\t\tRole: &pb.Role{\n\t\t\tName: rs.Response.Role.Name,\n\t\t\tPermissions: permissions,\n\t\t},\n\t}, nil\n}", "func encodeDeleteJobFunctionResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(endpoints.DeleteJobFunctionResponse)\n\terr := getError(res.Err)\n\tif err == nil {\n\t\treturn &pb.DeleteJobFunctionResponse{}, nil\n\t}\n\treturn nil, err\n}" ]
[ "0.83516836", "0.7371681", "0.7335246", "0.66589355", "0.6655732", "0.6442083", "0.6404204", "0.6344658", "0.6250281", "0.62424475", "0.60453826", "0.60057443", "0.5990996", "0.5986964", "0.58899724", "0.58410585", "0.58371496", "0.5823894", "0.57436144", "0.5743134", "0.57214195", "0.56942207", "0.5691651", "0.5689629", "0.5670009", "0.5661112", "0.56402934", "0.5639709", "0.5593553", "0.5583798", "0.5541167", "0.5534244", "0.5510351", "0.5493254", "0.54931706", "0.5484242", "0.54707867", "0.5453279", "0.54488486", "0.5448321", "0.544256", "0.54407567", "0.5437183", "0.54284096", "0.5388301", "0.5388301", "0.53712595", "0.53504443", "0.5331835", "0.53104156", "0.5308071", "0.5304077", "0.53021896", "0.5299803", "0.52856153", "0.5272309", "0.52593136", "0.5243112", "0.52357924", "0.5228699", "0.52241874", "0.521608", "0.5203552", "0.5194393", "0.5193286", "0.5177681", "0.51646996", "0.514631", "0.51332664", "0.51212025", "0.5114406", "0.50854295", "0.5080333", "0.5078802", "0.50749886", "0.5066669", "0.5056919", "0.5051972", "0.5050317", "0.5049527", "0.50494176", "0.50431895", "0.5033729", "0.5032804", "0.49705452", "0.49679962", "0.49636945", "0.49599093", "0.49585757", "0.49570182", "0.49487355", "0.4948297", "0.4928771", "0.49169028", "0.48973578", "0.48914102", "0.48836097", "0.48678133", "0.48648474", "0.48624492" ]
0.56325334
28
encodeGRPCNameRequest is a transport/grpc.EncodeRequestFunc that converts a userdomain name request to a gRPC name request. Primarily useful in a client.
func encodeGRPCNameRequest(_ context.Context, request interface{}) (interface{}, error) { req := request.(loginendpoint.LoginRequest) return &pb.NameRequest{N: string(req.N)}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func encodeGRPCConcatRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(loginendpoint.LoginRequest)\n\treturn &pb.NameRequest{N: req.N}, nil\n}", "func EncodeGRPCIDForNameRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.IDForNameRequest)\n\treturn &containerPB.IDForNameRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tName: req.Name,\n\t}, nil\n}", "func decodeGRPCNameRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.NameRequest)\n\treturn loginendpoint.LoginRequest{N: string(req.N)}, nil\n}", "func encodeGRPCNameResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(loginendpoint.LoginResponse)\n\treturn &pb.NameReply{V: resp.V, Err: err2str(resp.Err)}, nil\n}", "func EncodeGRPCGetEnvRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.GetEnvRequest)\n\treturn &containerPB.GetEnvRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t\tKey: req.Key,\n\t}, nil\n}", "func EncodeGRPCSetEnvRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.SetEnvRequest)\n\treturn &containerPB.SetEnvRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t\tKey: req.Key,\n\t\tValue: req.Value,\n\t}, nil\n}", "func EncodeGRPCExecuteRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.ExecuteRequest)\n\treturn &containerPB.ExecuteRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t\tCmd: req.CMD,\n\t}, nil\n}", "func encodeGRPCMathOpRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(mathendpoint2.MathOpRequest)\n\treturn &pb.MathOpRequest{A: req.A, B: req.B}, nil\n}", "func encodeGRPCGenerateReportRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(generateReportRequest)\n\treturn &pb.GenerateReportRequest{A: int64(req.A), B: int64(req.B)}, nil\n}", "func EncodeGRPCGetLinksRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.GetLinksRequest)\n\treturn &containerPB.GetLinksRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tContainerID: req.ContainerID,\n\t}, nil\n}", "func encodeGRPCSumRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(endpoints.SumRequest)\n\t\n\treturn &pb.SumRequest{A: req.A, B: req.B}, nil\n}", "func encodeGRPCTicRequest(_ context.Context, request interface{}) (interface{}, error) {\n\t_ = request.(endpoints.TicRequest)\n\treturn &pb.TicRequest{}, nil\n}", "func EncodeUsernameTakenRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(UsernameTakenRequest)\n\treturn &pb.UsernameTakenRequest{\n\t\tUsername: req.Username,\n\t}, nil\n}", "func encodeGRPCTacRequest(_ context.Context, request interface{}) (interface{}, error) {\n\t_ = request.(endpoints.TacRequest)\n\treturn &pb.TacRequest{}, nil\n}", "func EncodeByCatalogKindNameRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*resource.ByCatalogKindNamePayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"resource\", \"ByCatalogKindName\", \"*resource.ByCatalogKindNamePayload\", v)\n\t\t}\n\t\tvalues := req.URL.Query()\n\t\tif p.Pipelinesversion != nil {\n\t\t\tvalues.Add(\"pipelinesversion\", *p.Pipelinesversion)\n\t\t}\n\t\treq.URL.RawQuery = values.Encode()\n\t\treturn nil\n\t}\n}", "func encodeGRPCConcatRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(endpoints.ConcatRequest)\n\treturn &pb.ConcatRequest{A: req.A, B: req.B}, nil\n}", "func EncodeGRPCGetContainerKMIRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.GetContainerKMIRequest)\n\treturn &containerPB.GetContainerKMIRequest{\n\t\tContainerID: req.ContainerID,\n\t}, nil\n}", "func EncodeGRPCStopContainerRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.StopContainerRequest)\n\treturn &containerPB.StopContainerRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t}, nil\n}", "func EncodeGRPCRemoveContainerRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.RemoveContainerRequest)\n\treturn &containerPB.RemoveContainerRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t}, nil\n}", "func encodeUserInfoRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treturn &pb.UserInfoRequest{}, nil\n\t// return nil, errors.New(\"'Account' Encoder is not impelemented\")\n}", "func DecodeGRPCLoginRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.LoginRequest)\n\treturn req, nil\n}", "func EncodeBatchGRPCRequest(ctx context.Context, v interface{}, md *metadata.MD) (interface{}, error) {\n\tpayload, ok := v.(*public.TestPayloadBatch)\n\tif !ok {\n\t\treturn nil, goagrpc.ErrInvalidType(\"public\", \"batchGRPC\", \"*public.TestPayloadBatch\", v)\n\t}\n\treturn NewBatchGRPCRequest(payload), nil\n}", "func DecodeGRPCLoginRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.LoginParams)\n\treturn req, nil\n}", "func DecodeGRPCGetUserInfoRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.GetUserInfoParams)\n\treturn req, nil\n}", "func EncodeGrpcReqWorkload(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*Workload)\n\treturn req, nil\n}", "func EncodeGRPCCreateContainerRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.CreateContainerRequest)\n\treturn &containerPB.CreateContainerRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tKmiID: uint32(req.KmiID),\n\t\tName: req.Name,\n\t}, nil\n}", "func EncodeGrpcReqEVPNConfig(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*EVPNConfig)\n\treturn req, nil\n}", "func EncodeCreateUserRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(CreateUserRequest)\n\treturn &pb.CreateUserRequest{\n\t\tUsername: req.Username,\n\t\tEmail: req.Email,\n\t\tPassword: req.Password,\n\t\tPhoneNumber: req.PhoneNumber,\n\t}, nil\n}", "func EncodeGrpcReqLicenseSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*LicenseSpec)\n\treturn req, nil\n}", "func DecodeGRPCUpdateUserInfoRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.UpdateUserInfoParams)\n\treturn req, nil\n}", "func EncodeGrpcReqWorkloadSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadSpec)\n\treturn req, nil\n}", "func encodeGRPCEchoRequest(_ context.Context, request interface{}) (interface{}, error) {\n\n\t\n\tr:=request.(endpoints.EchoRequest);\n\treturn &pb.EchoRequest{Word: r.Word}, nil\n\t\n\t\n}", "func EncodeName(buf []byte, name string) ([]byte, error) {\n\treturn Encode(buf, Names[name])\n}", "func EncodeGrpcReqMetricsQuerySpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQuerySpec)\n\treturn req, nil\n}", "func encodeAuthRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treturn &pb.AuthRequest{}, nil\n}", "func EncodeLoginRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(LoginRequest)\n\treturn &pb.LoginRequest{\n\t\tUsername: req.Username,\n\t\tPassword: req.Password,\n\t}, nil\n}", "func EncodeGRPCSetLinkRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.SetLinkRequest)\n\treturn &containerPB.SetLinkRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tContainerID: req.ContainerID,\n\t\tLinkID: req.LinkID,\n\t\tLinkName: req.LinkName,\n\t\tLinkInterface: req.LinkInterface,\n\t}, nil\n}", "func EncodeGrpcReqFeature(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*Feature)\n\treturn req, nil\n}", "func EncodeGrpcReqDSCProfileSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*DSCProfileSpec)\n\treturn req, nil\n}", "func DecodeGRPCQueryRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.QueryRequest)\n\treturn req, nil\n}", "func EncodeGrpcReqDSCProfile(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*DSCProfile)\n\treturn req, nil\n}", "func EncodeGrpcReqWorkloadStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadStatus)\n\treturn req, nil\n}", "func encodeSigninRequest(_ context.Context, request interface{}) (interface{}, error) {\n\tr := request.(endpoint1.SigninRequest)\n\treturn &pb.SigninRequest{\n\t\tEmail: r.Email,\n\t\tPassword: r.Password}, nil\n}", "func EncodeGRPCRemoveLinkRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.RemoveLinkRequest)\n\treturn &containerPB.RemoveLinkRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tContainerID: req.ContainerID,\n\t\tLinkID: req.LinkID,\n\t\tLinkName: req.LinkName,\n\t\tLinkInterface: req.LinkInterface,\n\t}, nil\n}", "func (c Client) encodeRequest(req messages.PlayerReq) error {\n\treturn c.Encoder.Encode(req)\n}", "func decodeGRPCGreetingRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.GreetingRequest)\n\treturn greeterendpoint.GreetingRequest{Name: req.Name}, nil\n}", "func DecodeGRPCPingRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.PingRequest)\n\treturn req, nil\n}", "func EncodeName(org, bucket platform.ID) [16]byte {\n\tvar nameBytes [16]byte\n\tbinary.BigEndian.PutUint64(nameBytes[0:8], uint64(org))\n\tbinary.BigEndian.PutUint64(nameBytes[8:16], uint64(bucket))\n\treturn nameBytes\n}", "func (e *execPlugin) encodeRequest(request *credentialproviderapi.CredentialProviderRequest) ([]byte, error) {\n\tdata, err := runtime.Encode(e.encoder, request)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error encoding request: %w\", err)\n\t}\n\n\treturn data, nil\n}", "func EncodeQueryRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*resource.QueryPayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"resource\", \"Query\", \"*resource.QueryPayload\", v)\n\t\t}\n\t\tvalues := req.URL.Query()\n\t\tvalues.Add(\"name\", p.Name)\n\t\tfor _, value := range p.Catalogs {\n\t\t\tvalues.Add(\"catalogs\", value)\n\t\t}\n\t\tfor _, value := range p.Categories {\n\t\t\tvalues.Add(\"categories\", value)\n\t\t}\n\t\tfor _, value := range p.Kinds {\n\t\t\tvalues.Add(\"kinds\", value)\n\t\t}\n\t\tfor _, value := range p.Tags {\n\t\t\tvalues.Add(\"tags\", value)\n\t\t}\n\t\tfor _, value := range p.Platforms {\n\t\t\tvalues.Add(\"platforms\", value)\n\t\t}\n\t\tvalues.Add(\"limit\", fmt.Sprintf(\"%v\", p.Limit))\n\t\tvalues.Add(\"match\", p.Match)\n\t\treq.URL.RawQuery = values.Encode()\n\t\treturn nil\n\t}\n}", "func EncodeGrpcReqRDSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*RDSpec)\n\treturn req, nil\n}", "func EncodeGRPCLoremRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(endpoints.LoremRequest)\n\treturn &pb.LoremRequest{\n\t\tRequestType: req.RequestType,\n\t\tMax: req.Max,\n\t\tMin: req.Min,\n\t} , nil\n}", "func decodeGRPCNameResponse(_ context.Context, grpcReply interface{}) (interface{}, error) {\n\treply := grpcReply.(*pb.NameReply)\n\treturn loginendpoint.LoginResponse{V: string(reply.V), Err: str2err(reply.Err)}, nil\n}", "func EncodeLoginByUsernameRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*user.LoginByUsernamePayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"User\", \"LoginByUsername\", \"*user.LoginByUsernamePayload\", v)\n\t\t}\n\t\tbody := NewLoginByUsernameRequestBody(p)\n\t\tif err := encoder(req).Encode(&body); err != nil {\n\t\t\treturn goahttp.ErrEncodingError(\"User\", \"LoginByUsername\", err)\n\t\t}\n\t\treturn nil\n\t}\n}", "func EncodeGRPCLoremRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(LoremRequest)\n\treturn &pb.LoremRequest{\n\t\tRequestType: req.RequestType,\n\t\tMax: req.Max,\n\t\tMin: req.Min,\n\t}, nil\n}", "func EncodeGRPCWebRequestBody(message proto.Message) io.Reader {\n\tdata, _ := proto.Marshal(message)\n\tlengthPrefix := []byte{0, 0, 0, 0, 0}\n\tbinary.BigEndian.PutUint32(lengthPrefix[1:], uint32(len(data)))\n\tvar buf bytes.Buffer\n\tbuf.Write(lengthPrefix)\n\tbuf.Write(data)\n\n\tb := make([]byte, base64.StdEncoding.EncodedLen(buf.Len()))\n\tbase64.StdEncoding.Encode(b, buf.Bytes())\n\treturn bytes.NewReader(b)\n}", "func EncodeGrpcReqWorkloadIntfSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadIntfSpec)\n\treturn req, nil\n}", "func GetPlayername(r *http.Request) string {\n\tparseError := r.ParseForm()\n\tif parseError == nil {\n\t\tusername := r.Form.Get(\"username\")\n\t\tif username != \"\" {\n\t\t\treturn username\n\t\t}\n\t}\n\n\tusernameCookie, noCookieError := r.Cookie(\"username\")\n\tif noCookieError == nil {\n\t\tusername := usernameCookie.Value\n\t\tif username != \"\" {\n\t\t\treturn username\n\t\t}\n\t}\n\n\treturn \"\"\n}", "func encodeCreateRequest(_ context.Context, request interface{}) (interface{}, error) {\n\tr := request.(endpoint1.CreateRequest)\n\treturn &pb.CreateRequest{\n\t\tEmail: r.Email,\n\t\tPassword: r.Password,\n\t\tOrgname: r.OrgName}, nil\n}", "func EncodeGRPCInstancesRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.InstancesRequest)\n\treturn &containerPB.InstancesRequest{\n\t\tRefID: uint32(req.RefID),\n\t}, nil\n}", "func EncodeGrpcReqFeatureStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*FeatureStatus)\n\treturn req, nil\n}", "func EncodeGRPCLoginResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.LoginResponse)\n\treturn resp, nil\n}", "func EncodeGrpcReqMetricsQueryResponse(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQueryResponse)\n\treturn req, nil\n}", "func (g *grpc) Name() string {\n\treturn \"grpc\"\n}", "func EncodeGRPCQueryResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.QueryResponse)\n\treturn resp, nil\n}", "func EncodeGrpcReqRouteDistinguisher(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*RouteDistinguisher)\n\treturn req, nil\n}", "func EncodeGrpcReqBGPAuthStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*BGPAuthStatus)\n\treturn req, nil\n}", "func DecodeGRPCSetUserConfigRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.SetUserConfigParams)\n\treturn req, nil\n}", "func EncodeGrpcReqWorkloadMigrationStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadMigrationStatus)\n\treturn req, nil\n}", "func EncodeGrpcReqLicense(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*License)\n\treturn req, nil\n}", "func DecodeGRPCStatusRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.StatusRequest)\n\treturn req, nil\n}", "func EncodeGrpcReqBGPConfig(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*BGPConfig)\n\treturn req, nil\n}", "func EncodeGrpcReqRoute(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*Route)\n\treturn req, nil\n}", "func EncodeRegisterRequest(ctx context.Context, v interface{}, md *metadata.MD) (interface{}, error) {\n\tpayload, ok := v.(*usermethod.RegisterPayload)\n\tif !ok {\n\t\treturn nil, goagrpc.ErrInvalidType(\"userMethod\", \"register\", \"*usermethod.RegisterPayload\", v)\n\t}\n\treturn NewRegisterRequest(payload), nil\n}", "func DecodeGRPCGetUserConfigRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.GetUserConfigParams)\n\treturn req, nil\n}", "func EncodeNameString(org, bucket platform.ID) string {\n\tname := EncodeName(org, bucket)\n\treturn string(name[:])\n}", "func EncodeGrpcReqLicenseStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*LicenseStatus)\n\treturn req, nil\n}", "func (c *APIGateway) GetDomainNamesRequest(input *GetDomainNamesInput) (req *request.Request, output *GetDomainNamesOutput) {\n\top := &request.Operation{\n\t\tName: opGetDomainNames,\n\t\tHTTPMethod: \"GET\",\n\t\tHTTPPath: \"/domainnames\",\n\t\tPaginator: &request.Paginator{\n\t\t\tInputTokens: []string{\"position\"},\n\t\t\tOutputTokens: []string{\"position\"},\n\t\t\tLimitToken: \"limit\",\n\t\t\tTruncationToken: \"\",\n\t\t},\n\t}\n\n\tif input == nil {\n\t\tinput = &GetDomainNamesInput{}\n\t}\n\n\treq = c.newRequest(op, input, output)\n\toutput = &GetDomainNamesOutput{}\n\treq.Data = output\n\treturn\n}", "func DoGRPCRequest(grpcAction GRPCAction, resultsChannel chan reporter.SampleReqResult, sessionMap map[string]string, _ *config.VUContext, vulog *log.Entry, playbook *config.TestDef) bool {\n\tvar trace_req string\n\tsampleReqResult := buildSampleResult(REPORTER_GRPC, sessionMap[\"UID\"], 0, reporter.NETWORK_ERROR, 0, grpcAction.Title, \"\")\n\tdata := SubstParams(sessionMap, string([]byte(grpcAction.Data)), vulog)\n\n\tif must_trace_request {\n\t\ttrace_req = fmt.Sprintf(\"%s %s\", grpcAction.Call, data)\n\t} else {\n\t\tvulog.Debugf(\"New Request: Call: %s, Data: %s\", grpcAction.Call, data)\n\t}\n\n\t// Try to substitute the server name by an IP address\n\tserver := playbook.DfltValues.Server\n\tif !disable_dns_cache {\n\t\tif addr, status := utils.GetServerAddress(server); status == true {\n\t\t\tserver = addr\n\t\t}\n\t}\n\n\tvar opts []grpc.DialOption\n\topts = append(opts, grpc.WithInsecure())\n\n\t/*\n\t\tsh := &statsHandler{\n\t\t\tid: len(b.handlers),\n\t\t\tresults: b.results,\n\t\t\thasLog: b.config.hasLog,\n\t\t\tlog: b.config.log,\n\t\t}\n\n\t\tb.handlers = append(b.handlers, sh)\n\t\topts = append(opts, grpc.WithStatsHandler(sh))\n\t*/\n\n\topts = append(opts, grpc.WithKeepaliveParams(keepalive.ClientParameters{\n\t\tTime: time.Duration(playbook.Timeout) * time.Second,\n\t\tTimeout: time.Duration(playbook.Timeout) * time.Second,\n\t}))\n\n\t// increase max receive and send message sizes\n\topts = append(opts,\n\t\tgrpc.WithDefaultCallOptions(\n\t\t\tgrpc.MaxCallRecvMsgSize(math.MaxInt32),\n\t\t\tgrpc.MaxCallSendMsgSize(math.MaxInt32),\n\t\t))\n\n\tctx := context.Background()\n\tctx, _ = context.WithTimeout(ctx, time.Duration(playbook.Timeout)*time.Second)\n\t// cancel is ignored here as connection.Close() is used.\n\t// See https://godoc.org/google.golang.org/grpc#DialContext\n\n\tstart := time.Now()\n\n\t// create client connection\n\tconn, err := grpc.DialContext(ctx, server, opts...)\n\tdefer conn.Close()\n\tif err != nil {\n\t\tif must_trace_request {\n\t\t\tvulog.Infof(\"%s: FAILED (%s)\", trace_req, err)\n\t\t}\n\t\tvulog.Errorf(\"GRPC connection failed: %s\", err)\n\t\tbuildGRPCSampleResult(&sampleReqResult, 0, reporter.NETWORK_ERROR, 0, err.Error())\n\t\tresultsChannel <- sampleReqResult\n\t\treturn false\n\t}\n\n\treq := &GRPCRequest{\n\t\tTitle: grpcAction.Title,\n\t\tStub: grpcdynamic.NewStub(conn),\n\t\tCall: grpcAction.Call,\n\t\tData: data,\n\t\tFunc: grpcAction.Func,\n\t}\n\n\t// Unary request\n\tvar inputs []*dynamic.Message\n\tif inputs, err = getMessages(req, data); err != nil {\n\t\tvulog.Error(err)\n\t\treturn false\n\t}\n\tresp, err := makeUnaryRequest(&ctx, req, nil, inputs[0], vulog)\n\n\telapsed := time.Since(start)\n\n\tif err != nil {\n\t\tif must_trace_request {\n\t\t\tvulog.Infof(\"%s: FAILED (%s)\", trace_req, err)\n\t\t}\n\t\tvulog.Printf(\"Reading GRPC response failed: %s\", err)\n\t\tbuildGRPCSampleResult(&sampleReqResult, len(resp.String()), 1, elapsed.Nanoseconds(), req.Call)\n\t\tresultsChannel <- sampleReqResult\n\t\treturn false\n\t}\n\n\tif must_trace_request {\n\t\tvulog.Infof(\"%s; RetCode=%d; RcvdBytes=%d\", trace_req, 0, len(resp.String()))\n\t}\n\tif must_display_srv_resp {\n\t\tvulog.Debugf(\"[GRPC Response=%d] Received data: %s\", 0, resp.String())\n\t}\n\n\tvalid := true\n\n\t// if action specifies response action, parse using regexp/jsonpath\n\tvar empty_http_header http.Header\n\tdynResp := resp.(*dynamic.Message)\n\tjsonData, err := dynResp.MarshalJSON()\n\tif !processResult(grpcAction.ResponseHandlers, sessionMap, vulog, jsonData, empty_http_header) {\n\t\tvalid = false\n\t}\n\n\tbuildGRPCSampleResult(&sampleReqResult, len(resp.String()), 0, elapsed.Nanoseconds(), req.Call)\n\tresultsChannel <- sampleReqResult\n\n\treturn valid\n}", "func EncodeInitStatusRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treturn &pb.InitStatusRequest{}, nil\n}", "func EncodeGrpcReqPaginationSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*PaginationSpec)\n\treturn req, nil\n}", "func EncodeGRPCNewUserResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresp := r.(newUserResponse)\n\treturn &pb.NewUserResponse{\n\t\tId: resp.Id,\n\t\tErr: resp.Err,\n\t}, nil\n}", "func EncodeGrpcReqDSCProfileStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*DSCProfileStatus)\n\treturn req, nil\n}", "func EncodeGrpcReqWorkloadIntfStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadIntfStatus)\n\treturn req, nil\n}", "func EncodeSigninRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*user.Signin)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"user\", \"signin\", \"*user.Signin\", v)\n\t\t}\n\t\tbody := NewSigninRequestBody(p)\n\t\tif err := encoder(req).Encode(&body); err != nil {\n\t\t\treturn goahttp.ErrEncodingError(\"user\", \"signin\", err)\n\t\t}\n\t\treturn nil\n\t}\n}", "func EncodeGRPCGetUserInfoResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.UserInfo)\n\treturn resp, nil\n}", "func DecodeGRPCAllowPortRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.AllowPortRequest)\n\tsrcIP, err := abstraction.NewInet(req.SrcIP)\n\tif err != nil {\n\t\treturn AllowPortRequest{}, err\n\t}\n\tdstIP, err := abstraction.NewInet(req.DstIP)\n\tif err != nil {\n\t\treturn AllowPortRequest{}, err\n\t}\n\treturn AllowPortRequest{\n\t\tSrcIP: srcIP,\n\t\tDstIP: dstIP,\n\t\tSrcNetwork: req.SrcNetwork,\n\t\tDstNetwork: req.DstNetwork,\n\t\tProtocol: req.Protocol,\n\t\tPort: uint16(req.Port),\n\t}, nil\n}", "func decodeGRPCGenerateReportRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.GenerateReportRequest)\n\treturn generateReportRequest{Username: req.Username, Token: req.Token, RepositoryName: req.RepositoryName}, nil\n}", "func EncodeGrpcReqPropagationStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*PropagationStatus)\n\treturn req, nil\n}", "func EncodeOrgName(org platform.ID) [8]byte {\n\tvar orgBytes [8]byte\n\tbinary.BigEndian.PutUint64(orgBytes[0:8], uint64(org))\n\treturn orgBytes\n}", "func (c *DubboTestCodec) EncodeRequest(request *remoting.Request) (*bytes.Buffer, error) {\n\tif request.Event {\n\t\treturn c.encodeHeartbeartReqeust(request)\n\t}\n\n\tinvoc, ok := request.Data.(*invocation.RPCInvocation)\n\tif !ok {\n\t\treturn nil, perrors.Errorf(\"encode request failed for parameter type :%+v\", request)\n\t}\n\ttmpInvocation := invoc\n\n\tsvc := impl.Service{}\n\tsvc.Path = tmpInvocation.GetAttachmentWithDefaultValue(constant.PathKey, \"\")\n\tsvc.Interface = tmpInvocation.GetAttachmentWithDefaultValue(constant.InterfaceKey, \"\")\n\tsvc.Version = tmpInvocation.GetAttachmentWithDefaultValue(constant.VersionKey, \"\")\n\tsvc.Group = tmpInvocation.GetAttachmentWithDefaultValue(constant.GroupKey, \"\")\n\tsvc.Method = tmpInvocation.MethodName()\n\ttimeout, err := strconv.Atoi(tmpInvocation.GetAttachmentWithDefaultValue(constant.TimeoutKey, strconv.Itoa(constant.DefaultRemotingTimeout)))\n\tif err != nil {\n\t\t// it will be wrapped in readwrite.Write .\n\t\treturn nil, perrors.WithStack(err)\n\t}\n\tsvc.Timeout = time.Duration(timeout)\n\n\theader := impl.DubboHeader{}\n\tserialization := tmpInvocation.GetAttachmentWithDefaultValue(constant.SerializationKey, constant.Hessian2Serialization)\n\tif serialization == constant.ProtobufSerialization {\n\t\theader.SerialID = constant.SProto\n\t} else {\n\t\theader.SerialID = constant.SHessian2\n\t}\n\theader.ID = request.ID\n\tif request.TwoWay {\n\t\theader.Type = impl.PackageRequest_TwoWay\n\t} else {\n\t\theader.Type = impl.PackageRequest\n\t}\n\n\tpkg := &impl.DubboPackage{\n\t\tHeader: header,\n\t\tService: svc,\n\t\tBody: impl.NewRequestPayload(tmpInvocation.Arguments(), tmpInvocation.Attachments()),\n\t\tErr: nil,\n\t\tCodec: impl.NewDubboCodec(nil),\n\t}\n\n\tif err := impl.LoadSerializer(pkg); err != nil {\n\t\treturn nil, perrors.WithStack(err)\n\t}\n\n\treturn pkg.Marshal()\n}", "func DecodeGRPCAllowConnectionRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.AllowConnectionRequest)\n\tsrcIP, err := abstraction.NewInet(req.SrcIP)\n\tif err != nil {\n\t\treturn AllowConnectionRequest{}, err\n\t}\n\tdstIP, err := abstraction.NewInet(req.DstIP)\n\tif err != nil {\n\t\treturn AllowConnectionRequest{}, err\n\t}\n\treturn AllowConnectionRequest{\n\t\tSrcIP: srcIP,\n\t\tDstIP: dstIP,\n\t\tSrcNetwork: req.SrcNetwork,\n\t\tDstNetwork: req.DstNetwork,\n\t}, nil\n}", "func DecodeGRPCGetCodeRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.GetCodeReq)\n\treturn req, nil\n}", "func protobufName(f *ast.Field) string {\n\tfor _, attr := range f.Attrs {\n\t\tif strings.HasPrefix(attr.Text, \"@protobuf\") {\n\t\t\tfor _, str := range strings.Split(attr.Text[10:len(attr.Text)-1], \",\") {\n\t\t\t\tif strings.HasPrefix(str, \"name=\") {\n\t\t\t\t\treturn str[5:]\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn \"\"\n}", "func EncodeConfirmUserRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(ConfirmUserRequest)\n\treturn &pb.ConfirmUserRequest{\n\t\tUsername: req.Username,\n\t\tCode: req.Code,\n\t}, nil\n}", "func requestIDFromGRPCContext(ctx context.Context) string {\n\tmd, ok := metadata.FromIncomingContext(ctx)\n\tif ok {\n\t\tif v, ok := md[\"request-id\"]; ok {\n\t\t\treturn v[0]\n\t\t}\n\t}\n\treturn uuid.New().String()\n}", "func EncodeGetAccountRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*restapi.GetAccountPayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"RestAPI\", \"getAccount\", \"*restapi.GetAccountPayload\", v)\n\t\t}\n\t\tvalues := req.URL.Query()\n\t\tvalues.Add(\"address\", p.Address)\n\t\treq.URL.RawQuery = values.Encode()\n\t\treturn nil\n\t}\n}", "func encodeGRPCGreetingResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tres := response.(greeterendpoint.GreetingResponse)\n\treturn &pb.GreetingResponse{Greeting: res.Greeting}, nil\n}", "func EncodeGRPCLoginResponse(_ context.Context, r interface{}) (interface{}, error) {\n\tresp := r.(loginResponse)\n\treturn &pb.LoginResponse{\n\t\tToken: resp.Token,\n\t\tRefreshToken: resp.RefreshToken,\n\t\tErr: resp.Err,\n\t}, nil\n}", "func EncodeGrpcReqMetricsQueryResult(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQueryResult)\n\treturn req, nil\n}" ]
[ "0.72947574", "0.71715724", "0.6907968", "0.673007", "0.6354859", "0.6223447", "0.6130725", "0.60985935", "0.6095571", "0.5966165", "0.5927888", "0.5906733", "0.5886314", "0.5881191", "0.58687496", "0.5832968", "0.5822578", "0.5758928", "0.5684335", "0.5647708", "0.56460917", "0.5644387", "0.56373876", "0.5632121", "0.56066597", "0.555902", "0.54958206", "0.54872614", "0.54871804", "0.54850096", "0.54787135", "0.54750746", "0.54578084", "0.5451321", "0.5448507", "0.54212946", "0.54149103", "0.54064953", "0.5398404", "0.5393427", "0.5392892", "0.5381066", "0.5368873", "0.536075", "0.5352007", "0.53485954", "0.533555", "0.5323577", "0.53151095", "0.52703434", "0.526328", "0.52484804", "0.5248012", "0.5239556", "0.5222083", "0.52104974", "0.52015716", "0.5199986", "0.51948774", "0.5192233", "0.5191731", "0.5190185", "0.5180088", "0.5169672", "0.51688755", "0.5162892", "0.5158915", "0.5154849", "0.51394504", "0.51211345", "0.51156116", "0.50965106", "0.5083538", "0.5061566", "0.5058438", "0.50574577", "0.50524694", "0.505028", "0.505018", "0.505003", "0.50392294", "0.5020406", "0.50095874", "0.5004342", "0.5003883", "0.49941212", "0.49916035", "0.49764246", "0.49705595", "0.49510086", "0.49498206", "0.49471202", "0.49469036", "0.4942867", "0.49129215", "0.49083608", "0.48977247", "0.4888662", "0.48805624", "0.48793158" ]
0.85362494
0
encodeGRPCConcatRequest is a transport/grpc.EncodeRequestFunc that converts a userdomain concat request to a gRPC concat request. Primarily useful in a client.
func encodeGRPCConcatRequest(_ context.Context, request interface{}) (interface{}, error) { req := request.(loginendpoint.LoginRequest) return &pb.NameRequest{N: req.N}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func encodeGRPCConcatRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(endpoints.ConcatRequest)\n\treturn &pb.ConcatRequest{A: req.A, B: req.B}, nil\n}", "func decodeGRPCConcatRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.ConcatRequest)\n\treturn endpoints.ConcatRequest{A: req.A, B: req.B}, nil\n}", "func encodeGRPCSumRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(endpoints.SumRequest)\n\t\n\treturn &pb.SumRequest{A: req.A, B: req.B}, nil\n}", "func encodeGRPCConcatResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\treply := grpcReply.(endpoints.ConcatResponse)\n\treturn &pb.ConcatReply{Rs: reply.Rs}, def.GrpcEncodeError(reply.Err)\n}", "func encodeHTTPConcatRequest(_ context.Context, r *http.Request, request interface{}) (err error) {\n\tvar buf bytes.Buffer\n\tif err := json.NewEncoder(&buf).Encode(request); err != nil {\n\t\treturn err\n\t}\n\tr.Body = ioutil.NopCloser(&buf)\n\treturn nil\n}", "func encodeGRPCGenerateReportRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(generateReportRequest)\n\treturn &pb.GenerateReportRequest{A: int64(req.A), B: int64(req.B)}, nil\n}", "func EncodeGRPCExecuteRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.ExecuteRequest)\n\treturn &containerPB.ExecuteRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t\tCmd: req.CMD,\n\t}, nil\n}", "func encodeGRPCMathOpRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(mathendpoint2.MathOpRequest)\n\treturn &pb.MathOpRequest{A: req.A, B: req.B}, nil\n}", "func EncodeBatchGRPCRequest(ctx context.Context, v interface{}, md *metadata.MD) (interface{}, error) {\n\tpayload, ok := v.(*public.TestPayloadBatch)\n\tif !ok {\n\t\treturn nil, goagrpc.ErrInvalidType(\"public\", \"batchGRPC\", \"*public.TestPayloadBatch\", v)\n\t}\n\treturn NewBatchGRPCRequest(payload), nil\n}", "func encodeGRPCTicRequest(_ context.Context, request interface{}) (interface{}, error) {\n\t_ = request.(endpoints.TicRequest)\n\treturn &pb.TicRequest{}, nil\n}", "func encodeGRPCTacRequest(_ context.Context, request interface{}) (interface{}, error) {\n\t_ = request.(endpoints.TacRequest)\n\treturn &pb.TacRequest{}, nil\n}", "func encodeGRPCNameRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(loginendpoint.LoginRequest)\n\treturn &pb.NameRequest{N: string(req.N)}, nil\n}", "func EncodeGrpcReqWorkload(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*Workload)\n\treturn req, nil\n}", "func EncodeGRPCRemoveContainerRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.RemoveContainerRequest)\n\treturn &containerPB.RemoveContainerRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t}, nil\n}", "func EncodeGrpcReqDSCProfile(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*DSCProfile)\n\treturn req, nil\n}", "func EncodeGRPCGetLinksRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.GetLinksRequest)\n\treturn &containerPB.GetLinksRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tContainerID: req.ContainerID,\n\t}, nil\n}", "func EncodeUsernameTakenRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(UsernameTakenRequest)\n\treturn &pb.UsernameTakenRequest{\n\t\tUsername: req.Username,\n\t}, nil\n}", "func EncodeGRPCGetEnvRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.GetEnvRequest)\n\treturn &containerPB.GetEnvRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t\tKey: req.Key,\n\t}, nil\n}", "func EncodeGrpcReqRDSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*RDSpec)\n\treturn req, nil\n}", "func EncodeGrpcReqDSCProfileSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*DSCProfileSpec)\n\treturn req, nil\n}", "func encodeAuthRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treturn &pb.AuthRequest{}, nil\n}", "func decodeHTTPConcatRequest(_ context.Context, r *http.Request) (interface{}, error) {\n\tvar req endpoints.ConcatRequest\n\terr := json.NewDecoder(r.Body).Decode(&req)\n\treturn req, err\n}", "func EncodeGRPCSetEnvRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.SetEnvRequest)\n\treturn &containerPB.SetEnvRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t\tKey: req.Key,\n\t\tValue: req.Value,\n\t}, nil\n}", "func EncodeGRPCIDForNameRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.IDForNameRequest)\n\treturn &containerPB.IDForNameRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tName: req.Name,\n\t}, nil\n}", "func EncodeGrpcReqWorkloadSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadSpec)\n\treturn req, nil\n}", "func EncodeGRPCRemoveLinkRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.RemoveLinkRequest)\n\treturn &containerPB.RemoveLinkRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tContainerID: req.ContainerID,\n\t\tLinkID: req.LinkID,\n\t\tLinkName: req.LinkName,\n\t\tLinkInterface: req.LinkInterface,\n\t}, nil\n}", "func EncodeGrpcReqMetricsQuerySpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQuerySpec)\n\treturn req, nil\n}", "func EncodeRegisterRequest(ctx context.Context, v interface{}, md *metadata.MD) (interface{}, error) {\n\tpayload, ok := v.(*usermethod.RegisterPayload)\n\tif !ok {\n\t\treturn nil, goagrpc.ErrInvalidType(\"userMethod\", \"register\", \"*usermethod.RegisterPayload\", v)\n\t}\n\treturn NewRegisterRequest(payload), nil\n}", "func decodeGRPCConcatResponse(_ context.Context, grpcReply interface{}) (interface{}, error) {\n\treply := grpcReply.(*pb.ConcatReply)\n\treturn endpoints.ConcatResponse{Rs: reply.Rs}, nil\n}", "func encodeGRPCEchoRequest(_ context.Context, request interface{}) (interface{}, error) {\n\n\t\n\tr:=request.(endpoints.EchoRequest);\n\treturn &pb.EchoRequest{Word: r.Word}, nil\n\t\n\t\n}", "func EncodeGrpcReqRouteDistinguisher(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*RouteDistinguisher)\n\treturn req, nil\n}", "func EncodeGrpcReqEVPNConfig(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*EVPNConfig)\n\treturn req, nil\n}", "func EncodeGrpcReqFeature(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*Feature)\n\treturn req, nil\n}", "func EncodeGrpcReqPaginationSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*PaginationSpec)\n\treturn req, nil\n}", "func EncodeGrpcReqResultSeries(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*ResultSeries)\n\treturn req, nil\n}", "func EncodeResendConfirmationRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(ResendConfirmationRequest)\n\treturn &pb.ResendConfirmationRequest{\n\t\tUsername: req.Username,\n\t}, nil\n}", "func EncodeGrpcReqMetricsQueryResponse(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQueryResponse)\n\treturn req, nil\n}", "func EncodeConfirmUserRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(ConfirmUserRequest)\n\treturn &pb.ConfirmUserRequest{\n\t\tUsername: req.Username,\n\t\tCode: req.Code,\n\t}, nil\n}", "func (e *execPlugin) encodeRequest(request *credentialproviderapi.CredentialProviderRequest) ([]byte, error) {\n\tdata, err := runtime.Encode(e.encoder, request)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error encoding request: %w\", err)\n\t}\n\n\treturn data, nil\n}", "func EncodeQueryRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*resource.QueryPayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"resource\", \"Query\", \"*resource.QueryPayload\", v)\n\t\t}\n\t\tvalues := req.URL.Query()\n\t\tvalues.Add(\"name\", p.Name)\n\t\tfor _, value := range p.Catalogs {\n\t\t\tvalues.Add(\"catalogs\", value)\n\t\t}\n\t\tfor _, value := range p.Categories {\n\t\t\tvalues.Add(\"categories\", value)\n\t\t}\n\t\tfor _, value := range p.Kinds {\n\t\t\tvalues.Add(\"kinds\", value)\n\t\t}\n\t\tfor _, value := range p.Tags {\n\t\t\tvalues.Add(\"tags\", value)\n\t\t}\n\t\tfor _, value := range p.Platforms {\n\t\t\tvalues.Add(\"platforms\", value)\n\t\t}\n\t\tvalues.Add(\"limit\", fmt.Sprintf(\"%v\", p.Limit))\n\t\tvalues.Add(\"match\", p.Match)\n\t\treq.URL.RawQuery = values.Encode()\n\t\treturn nil\n\t}\n}", "func EncodeGrpcReqWorkloadStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadStatus)\n\treturn req, nil\n}", "func EncodeCreateUserRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(CreateUserRequest)\n\treturn &pb.CreateUserRequest{\n\t\tUsername: req.Username,\n\t\tEmail: req.Email,\n\t\tPassword: req.Password,\n\t\tPhoneNumber: req.PhoneNumber,\n\t}, nil\n}", "func EncodeGrpcReqRoute(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*Route)\n\treturn req, nil\n}", "func EncodeGRPCInstancesRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.InstancesRequest)\n\treturn &containerPB.InstancesRequest{\n\t\tRefID: uint32(req.RefID),\n\t}, nil\n}", "func EncodeStreamedBatchGRPCRequest(ctx context.Context, v interface{}, md *metadata.MD) (interface{}, error) {\n\tpayload, ok := v.(*public.StreamMode)\n\tif !ok {\n\t\treturn nil, goagrpc.ErrInvalidType(\"public\", \"streamedBatchGRPC\", \"*public.StreamMode\", v)\n\t}\n\tif payload.Recieveall != nil {\n\t\t(*md).Append(\"recieveall\", fmt.Sprintf(\"%v\", *payload.Recieveall))\n\t}\n\treturn nil, nil\n}", "func EncodeGrpcReqLicenseSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*LicenseSpec)\n\treturn req, nil\n}", "func EncodeGrpcReqBGPAuthStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*BGPAuthStatus)\n\treturn req, nil\n}", "func EncodeGrpcReqBGPConfig(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*BGPConfig)\n\treturn req, nil\n}", "func encodeUserInfoRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treturn &pb.UserInfoRequest{}, nil\n\t// return nil, errors.New(\"'Account' Encoder is not impelemented\")\n}", "func EncodeGRPCWebRequestBody(message proto.Message) io.Reader {\n\tdata, _ := proto.Marshal(message)\n\tlengthPrefix := []byte{0, 0, 0, 0, 0}\n\tbinary.BigEndian.PutUint32(lengthPrefix[1:], uint32(len(data)))\n\tvar buf bytes.Buffer\n\tbuf.Write(lengthPrefix)\n\tbuf.Write(data)\n\n\tb := make([]byte, base64.StdEncoding.EncodedLen(buf.Len()))\n\tbase64.StdEncoding.Encode(b, buf.Bytes())\n\treturn bytes.NewReader(b)\n}", "func QuantizedConcat(scope *Scope, concat_dim tf.Output, values []tf.Output, input_mins []tf.Output, input_maxes []tf.Output) (output tf.Output, output_min tf.Output, output_max tf.Output) {\n\tif scope.Err() != nil {\n\t\treturn\n\t}\n\topspec := tf.OpSpec{\n\t\tType: \"QuantizedConcat\",\n\t\tInput: []tf.Input{\n\t\t\tconcat_dim, tf.OutputList(values), tf.OutputList(input_mins), tf.OutputList(input_maxes),\n\t\t},\n\t}\n\top := scope.AddOperation(opspec)\n\treturn op.Output(0), op.Output(1), op.Output(2)\n}", "func EncodeGRPCStopContainerRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.StopContainerRequest)\n\treturn &containerPB.StopContainerRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tID: req.ID,\n\t}, nil\n}", "func EncodeGrpcReqMetricsQueryList(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQueryList)\n\treturn req, nil\n}", "func EncodeGrpcReqDSCProfileStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*DSCProfileStatus)\n\treturn req, nil\n}", "func EncodeGetOnlineAccountsRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*restapi.GetOnlineAccountsPayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"RestAPI\", \"getOnlineAccounts\", \"*restapi.GetOnlineAccountsPayload\", v)\n\t\t}\n\t\tvalues := req.URL.Query()\n\t\tif p.Version != nil {\n\t\t\tvalues.Add(\"version\", *p.Version)\n\t\t}\n\t\tif p.Threshold != nil {\n\t\t\tvalues.Add(\"threshold\", fmt.Sprintf(\"%v\", *p.Threshold))\n\t\t}\n\t\treq.URL.RawQuery = values.Encode()\n\t\treturn nil\n\t}\n}", "func EncodeGrpcReqMetricsQueryResult(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*MetricsQueryResult)\n\treturn req, nil\n}", "func EncodeGRPCSetLinkRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.SetLinkRequest)\n\treturn &containerPB.SetLinkRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tContainerID: req.ContainerID,\n\t\tLinkID: req.LinkID,\n\t\tLinkName: req.LinkName,\n\t\tLinkInterface: req.LinkInterface,\n\t}, nil\n}", "func EncodeGrpcReqWorkloadIntfSpec(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadIntfSpec)\n\treturn req, nil\n}", "func EncodeGetUserWorkHistoryRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*hyuserworkhistory.GetUserWorkHistoryPayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"hy_userWorkHistory\", \"getUserWorkHistory\", \"*hyuserworkhistory.GetUserWorkHistoryPayload\", v)\n\t\t}\n\t\tif p.Token != nil {\n\t\t\thead := *p.Token\n\t\t\tif !strings.Contains(head, \" \") {\n\t\t\t\treq.Header.Set(\"Authorization\", \"Bearer \"+head)\n\t\t\t} else {\n\t\t\t\treq.Header.Set(\"Authorization\", head)\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n}", "func EncodeGRPCGetContainerKMIRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.GetContainerKMIRequest)\n\treturn &containerPB.GetContainerKMIRequest{\n\t\tContainerID: req.ContainerID,\n\t}, nil\n}", "func (c *Client) Concat(ctx context.Context, p *ConcatPayload) (res string, err error) {\n\tvar ires any\n\tires, err = c.ConcatEndpoint(ctx, p)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn ires.(string), nil\n}", "func EncodeGRPCLoremRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(endpoints.LoremRequest)\n\treturn &pb.LoremRequest{\n\t\tRequestType: req.RequestType,\n\t\tMax: req.Max,\n\t\tMin: req.Min,\n\t} , nil\n}", "func EncodeGRPCQueryResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(*pb.QueryResponse)\n\treturn resp, nil\n}", "func EncodeClientRequest(method string, args interface{}) ([]byte, error) {\n\tc := ClientRequest{\n\t\tJsonRpc: \"2.0\",\n\t\tMethod: method,\n\t\tParams: [1]interface{}{args},\n\t\tId: generateULID(),\n\t}\n\treturn json.Marshal(c)\n}", "func MarshalConcat(va interface{}, vbs ...interface{}) ([]byte, error) {\n\tunique, err := marshalConcat(nil, va, vbs...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn json.Marshal(unique)\n}", "func EncodeRequest(er EncodeRequestFunc) ClientOption {\n\treturn func(c *Client) {\n\t\tc.encode = er\n\t}\n}", "func EncodeAddRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, any) error {\n\treturn func(req *http.Request, v any) error {\n\t\tp, ok := v.(*storage.Bottle)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"storage\", \"add\", \"*storage.Bottle\", v)\n\t\t}\n\t\tbody := NewAddRequestBody(p)\n\t\tif err := encoder(req).Encode(&body); err != nil {\n\t\t\treturn goahttp.ErrEncodingError(\"storage\", \"add\", err)\n\t\t}\n\t\treturn nil\n\t}\n}", "func BuildRegisterFunc(grpccli user_methodpb.UserMethodClient, cliopts ...grpc.CallOption) goagrpc.RemoteFunc {\n\treturn func(ctx context.Context, reqpb interface{}, opts ...grpc.CallOption) (interface{}, error) {\n\t\tfor _, opt := range cliopts {\n\t\t\topts = append(opts, opt)\n\t\t}\n\t\treturn grpccli.Register(ctx, reqpb.(*user_methodpb.RegisterRequest), opts...)\n\t}\n}", "func EncodeGetAccountRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*restapi.GetAccountPayload)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"RestAPI\", \"getAccount\", \"*restapi.GetAccountPayload\", v)\n\t\t}\n\t\tvalues := req.URL.Query()\n\t\tvalues.Add(\"address\", p.Address)\n\t\treq.URL.RawQuery = values.Encode()\n\t\treturn nil\n\t}\n}", "func EncodeGRPCCreateContainerRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*container.CreateContainerRequest)\n\treturn &containerPB.CreateContainerRequest{\n\t\tRefID: uint32(req.RefID),\n\t\tKmiID: uint32(req.KmiID),\n\t\tName: req.Name,\n\t}, nil\n}", "func EncodeGRPCLoremRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(LoremRequest)\n\treturn &pb.LoremRequest{\n\t\tRequestType: req.RequestType,\n\t\tMax: req.Max,\n\t\tMin: req.Min,\n\t}, nil\n}", "func EncodeRequest(format wrp.Format) gokithttp.EncodeRequestFunc {\n\treturn func(ctx context.Context, component *http.Request, v interface{}) error {\n\t\tentity := v.(*Entity)\n\n\t\tif format == entity.Format && len(entity.Contents) > 0 {\n\t\t\t// the entity is already formatted properly, so just write its contents out\n\t\t\tcomponent.Body = ioutil.NopCloser(bytes.NewReader(entity.Contents))\n\t\t\tcomponent.ContentLength = int64(len(entity.Contents))\n\t\t} else {\n\t\t\tvar transcoded []byte\n\t\t\tif err := wrp.NewEncoderBytes(&transcoded, format).Encode(&entity.Message); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tcomponent.Body = ioutil.NopCloser(bytes.NewReader(transcoded))\n\t\t\tcomponent.ContentLength = int64(len(transcoded))\n\t\t}\n\n\t\tcomponent.Header.Set(\"Content-Type\", format.ContentType())\n\t\tcomponent.Header.Set(DestinationHeader, entity.Message.Destination)\n\t\treturn nil\n\t}\n}", "func encodeCreateRequest(_ context.Context, request interface{}) (interface{}, error) {\n\tr := request.(endpoint1.CreateRequest)\n\treturn &pb.CreateRequest{\n\t\tEmail: r.Email,\n\t\tPassword: r.Password,\n\t\tOrgname: r.OrgName}, nil\n}", "func DecodeGRPCQueryRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.QueryRequest)\n\treturn req, nil\n}", "func EncodeGrpcReqFeatureStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*FeatureStatus)\n\treturn req, nil\n}", "func EncodeGrpcReqLicense(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*License)\n\treturn req, nil\n}", "func EncodeGRPCResponse(_ context.Context, response interface{}) (interface{}, error) {\n\treturn response, nil\n}", "func encodeHTTPSumRequest(_ context.Context, r *http.Request, request interface{}) (err error) {\n\tvar buf bytes.Buffer\n\tif err := json.NewEncoder(&buf).Encode(request); err != nil {\n\t\treturn err\n\t}\n\tr.Body = ioutil.NopCloser(&buf)\n\treturn nil\n}", "func EncodeGrpcReqWorkloadMigrationStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadMigrationStatus)\n\treturn req, nil\n}", "func EncodeInitStatusRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treturn &pb.InitStatusRequest{}, nil\n}", "func decodeGRPCSumRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.SumRequest)\n\treturn endpoints.SumRequest{A: req.A, B: req.B}, nil\n}", "func EncodeUserDetailsRequest(_ context.Context, r interface{}) (interface{}, error) {\n\treq := r.(UserDetailsRequest)\n\treturn &pb.UserDetailsRequest{\n\t\tJwt: req.Jwt,\n\t}, nil\n}", "func (c Client) encodeRequest(req messages.PlayerReq) error {\n\treturn c.Encoder.Encode(req)\n}", "func encodeGRPCSumResponse(_ context.Context, grpcReply interface{}) (res interface{}, err error) {\n\treply := grpcReply.(endpoints.SumResponse)\n\treturn &pb.SumReply{Rs: reply.Rs}, def.GrpcEncodeError(reply.Err)\n}", "func EncodeUpdateRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*user.UpdateUser)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"user\", \"update\", \"*user.UpdateUser\", v)\n\t\t}\n\t\treq.Header.Set(\"Authorization\", p.Token)\n\t\tbody := NewUpdateRequestBody(p)\n\t\tif err := encoder(req).Encode(&body); err != nil {\n\t\t\treturn goahttp.ErrEncodingError(\"user\", \"update\", err)\n\t\t}\n\t\treturn nil\n\t}\n}", "func DecodeGRPCUpdateUserInfoRequest(_ context.Context, grpcReq interface{}) (interface{}, error) {\n\treq := grpcReq.(*pb.UpdateUserInfoParams)\n\treturn req, nil\n}", "func (c *DubboTestCodec) EncodeRequest(request *remoting.Request) (*bytes.Buffer, error) {\n\tif request.Event {\n\t\treturn c.encodeHeartbeartReqeust(request)\n\t}\n\n\tinvoc, ok := request.Data.(*invocation.RPCInvocation)\n\tif !ok {\n\t\treturn nil, perrors.Errorf(\"encode request failed for parameter type :%+v\", request)\n\t}\n\ttmpInvocation := invoc\n\n\tsvc := impl.Service{}\n\tsvc.Path = tmpInvocation.GetAttachmentWithDefaultValue(constant.PathKey, \"\")\n\tsvc.Interface = tmpInvocation.GetAttachmentWithDefaultValue(constant.InterfaceKey, \"\")\n\tsvc.Version = tmpInvocation.GetAttachmentWithDefaultValue(constant.VersionKey, \"\")\n\tsvc.Group = tmpInvocation.GetAttachmentWithDefaultValue(constant.GroupKey, \"\")\n\tsvc.Method = tmpInvocation.MethodName()\n\ttimeout, err := strconv.Atoi(tmpInvocation.GetAttachmentWithDefaultValue(constant.TimeoutKey, strconv.Itoa(constant.DefaultRemotingTimeout)))\n\tif err != nil {\n\t\t// it will be wrapped in readwrite.Write .\n\t\treturn nil, perrors.WithStack(err)\n\t}\n\tsvc.Timeout = time.Duration(timeout)\n\n\theader := impl.DubboHeader{}\n\tserialization := tmpInvocation.GetAttachmentWithDefaultValue(constant.SerializationKey, constant.Hessian2Serialization)\n\tif serialization == constant.ProtobufSerialization {\n\t\theader.SerialID = constant.SProto\n\t} else {\n\t\theader.SerialID = constant.SHessian2\n\t}\n\theader.ID = request.ID\n\tif request.TwoWay {\n\t\theader.Type = impl.PackageRequest_TwoWay\n\t} else {\n\t\theader.Type = impl.PackageRequest\n\t}\n\n\tpkg := &impl.DubboPackage{\n\t\tHeader: header,\n\t\tService: svc,\n\t\tBody: impl.NewRequestPayload(tmpInvocation.Arguments(), tmpInvocation.Attachments()),\n\t\tErr: nil,\n\t\tCodec: impl.NewDubboCodec(nil),\n\t}\n\n\tif err := impl.LoadSerializer(pkg); err != nil {\n\t\treturn nil, perrors.WithStack(err)\n\t}\n\n\treturn pkg.Marshal()\n}", "func EncodeGrpcReqWorkloadIntfStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*WorkloadIntfStatus)\n\treturn req, nil\n}", "func encodeGRPCGenerateReportResponse(_ context.Context, response interface{}) (interface{}, error) {\n\tresp := response.(generateReportResponse)\n\treturn &pb.GenerateReportReply{V: int64(resp.V), Err: err2str(resp.Err)}, nil\n}", "func encodeGRPCSayHelloRequest(_ context.Context, request interface{}) (interface{}, error) {\n\treq := request.(endpoints.SayHelloRequest)\n\treturn &pb.SayHelloRequest{Saidword: req.Saidword, Want: req.Want}, nil\n\t\n\t\n}", "func EncodeGrpcReqRoutingConfig(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*RoutingConfig)\n\treturn req, nil\n}", "func marshalEncode(submission protobufproto.Message, cmd Command) (encoded []byte, err error) {\n\traw, err := protobufproto.Marshal(submission)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to marshal request: %w\", err)\n\t}\n\n\tprefixBytes := []byte(uuid.NewV4().String())\n\tcommandInput := append([]byte{byte(cmd)}, raw...)\n\treturn append(prefixBytes, commandInput...), nil\n}", "func EncodeGrpcReqLicenseStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*LicenseStatus)\n\treturn req, nil\n}", "func EncodeAddRequest(encoder func(*http.Request) goahttp.Encoder) func(*http.Request, interface{}) error {\n\treturn func(req *http.Request, v interface{}) error {\n\t\tp, ok := v.(*blog.NewComment)\n\t\tif !ok {\n\t\t\treturn goahttp.ErrInvalidType(\"blog\", \"add\", \"*blog.NewComment\", v)\n\t\t}\n\t\tbody := NewAddRequestBody(p)\n\t\tif err := encoder(req).Encode(&body); err != nil {\n\t\t\treturn goahttp.ErrEncodingError(\"blog\", \"add\", err)\n\t\t}\n\t\treturn nil\n\t}\n}", "func encodeSigninRequest(_ context.Context, request interface{}) (interface{}, error) {\n\tr := request.(endpoint1.SigninRequest)\n\treturn &pb.SigninRequest{\n\t\tEmail: r.Email,\n\t\tPassword: r.Password}, nil\n}", "func EncodeSCEPRequest(ctx context.Context, r *http.Request, request interface{}) error {\n\treq := request.(SCEPRequest)\n\tparams := r.URL.Query()\n\tparams.Set(\"operation\", req.Operation)\n\tswitch r.Method {\n\tcase \"GET\":\n\t\tif len(req.Message) > 0 {\n\t\t\tmsg := base64.URLEncoding.EncodeToString(req.Message)\n\t\t\tparams.Set(\"message\", msg)\n\t\t}\n\t\tr.URL.RawQuery = params.Encode()\n\t\treturn nil\n\tcase \"POST\":\n\t\tbody := bytes.NewReader(req.Message)\n\t\t// recreate the request here because IIS does not support chunked encoding by default\n\t\t// and Go doesn't appear to set Content-Length if we use an io.ReadCloser\n\t\tu := r.URL\n\t\tu.RawQuery = params.Encode()\n\t\trr, err := http.NewRequest(\"POST\", u.String(), body)\n\t\tif err != nil {\n\t\t\treturn errors.Wrapf(err, \"creating new POST request for %s\", req.Operation)\n\t\t}\n\t\t*r = *rr\n\t\treturn nil\n\tdefault:\n\t\treturn fmt.Errorf(\"scep: %s method not supported\", r.Method)\n\t}\n}", "func EncodeGrpcReqPropagationStatus(ctx context.Context, request interface{}) (interface{}, error) {\n\treq := request.(*PropagationStatus)\n\treturn req, nil\n}", "func encodePutRequest(args *PutRequest) []byte {\n\tnum_bytes := uint64(8 + 8 + 8 + 8 + len(args.Value)) // CID + Seq + key + value-len + value\n\te := marshal.NewEnc(num_bytes)\n\te.PutInt(args.CID)\n\te.PutInt(args.Seq)\n\te.PutInt(args.Key)\n\te.PutInt(uint64(len(args.Value)))\n\te.PutBytes(args.Value)\n\n\treturn e.Finish()\n}", "func Concat(in []string) (rv string) {\n\n\tswitch len(in) {\n\n\tcase 0:\n\t\treturn \"\"\n\n\tcase 1:\n\t\treturn in[0]\n\n\tcase 2:\n\t\treturn in[0] + in[1]\n\n\tcase 3:\n\t\treturn in[0] + in[1] + in[2]\n\t}\n\n\tn := 0\n\n\tfor i := 0; i < len(in); i++ {\n\t\tn += len(in[i])\n\t}\n\n\tb := make([]byte, n)\n\n\tbp := copy(b, in[0])\n\n\tfor _, s := range in[1:] {\n\t\tbp += copy(b[bp:], s)\n\t}\n\n\treturn string(b)\n}", "func BuildBatchGRPCFunc(grpccli publicpb.PublicClient, cliopts ...grpc.CallOption) goagrpc.RemoteFunc {\n\treturn func(ctx context.Context, reqpb interface{}, opts ...grpc.CallOption) (interface{}, error) {\n\t\tfor _, opt := range cliopts {\n\t\t\topts = append(opts, opt)\n\t\t}\n\t\tif reqpb != nil {\n\t\t\treturn grpccli.BatchGRPC(ctx, reqpb.(*publicpb.BatchGRPCRequest), opts...)\n\t\t}\n\t\treturn grpccli.BatchGRPC(ctx, &publicpb.BatchGRPCRequest{}, opts...)\n\t}\n}" ]
[ "0.8351589", "0.6935657", "0.68416077", "0.66010845", "0.6436013", "0.635566", "0.6267067", "0.6070977", "0.6024813", "0.60160005", "0.59978575", "0.59751195", "0.58679605", "0.58348876", "0.5766015", "0.57529694", "0.56341875", "0.562872", "0.5587026", "0.55825436", "0.55423826", "0.5502829", "0.54931396", "0.5491988", "0.5487444", "0.5476286", "0.5467192", "0.54594", "0.54292053", "0.54164", "0.54113", "0.54049087", "0.5386624", "0.5381383", "0.5351396", "0.5346515", "0.53319764", "0.5326629", "0.5323823", "0.5280476", "0.52737266", "0.5241024", "0.5240732", "0.52318126", "0.52240276", "0.5217769", "0.52061486", "0.5202692", "0.5202017", "0.5183029", "0.5161814", "0.51540685", "0.51520044", "0.5128133", "0.51215285", "0.5121039", "0.5115194", "0.5113885", "0.5089996", "0.50884753", "0.50832784", "0.50804406", "0.5060873", "0.50576943", "0.505746", "0.5054951", "0.5046725", "0.50447214", "0.50165546", "0.501447", "0.50074816", "0.5001487", "0.49833214", "0.49820387", "0.49682748", "0.49558523", "0.49458346", "0.49412078", "0.4917723", "0.48962843", "0.48887718", "0.48876765", "0.4882612", "0.48755682", "0.48673323", "0.48615712", "0.4861295", "0.48505262", "0.48371577", "0.48311907", "0.4824557", "0.48131523", "0.48114118", "0.4806023", "0.48043203", "0.48014364", "0.47904542", "0.47787377", "0.47654048", "0.4760151" ]
0.7864252
1
These annoying helper functions are required to translate Go error types to and from strings, which is the type we use in our IDLs to represent errors. There is special casing to treat empty strings as nil errors.
func str2err(s string) error { if s == "" { return nil } return errors.New(s) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Str2Err(s string) error {\n\tif s == \"\" {\n\t\treturn nil\n\t}\n\treturn errors.New(s)\n}", "func errstring(err error) string {\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn \"\"\n}", "func errstring(err error) string {\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn \"\"\n}", "func errstring(err error) string {\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn \"\"\n}", "func convertNillableType(str string) (string, error) {\n\tdataType, err := convertType(str)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tnillableDataType, exists := nillableDataTypes[dataType]\n\tif exists {\n\t\treturn nillableDataType, nil \n\t}\n\n\treturn str, fmt.Errorf(\"convertingNillableType '%s'\", dataType)\n}", "func errorString(e *env, err error) string {\n\ttype hasUnderlying interface {\n\t\tHasUnderlying() error\n\t}\n\n\tparseErr := func(err error) error {\n\t\tif apiErr, ok := err.(*parse.Error); ok {\n\t\t\treturn errors.New(apiErr.Message)\n\t\t}\n\t\treturn err\n\t}\n\n\tlastErr := func(err error) error {\n\t\tif serr, ok := err.(*stackerr.Error); ok {\n\t\t\tif errs := stackerr.Underlying(serr); len(errs) != 0 {\n\t\t\t\terr = errs[len(errs)-1]\n\t\t\t}\n\t\t} else {\n\t\t\tif eu, ok := err.(hasUnderlying); ok {\n\t\t\t\terr = eu.HasUnderlying()\n\t\t\t}\n\t\t}\n\n\t\treturn parseErr(err)\n\t}\n\n\tif !e.ErrorStack {\n\t\tif merr, ok := err.(errgroup.MultiError); ok {\n\t\t\tvar multiError []error\n\t\t\tfor _, ierr := range []error(merr) {\n\t\t\t\tmultiError = append(multiError, lastErr(ierr))\n\t\t\t}\n\t\t\terr = errgroup.MultiError(multiError)\n\t\t} else {\n\t\t\terr = lastErr(err)\n\t\t}\n\t\treturn parseErr(err).Error()\n\t}\n\n\treturn err.Error()\n}", "func sanitisedError(err error, objectType string, objectName string, objectVersion string) error {\n\tsanitisedErr := strings.Replace(err.Error(), \"\\\\\", \" \", -1)\n\treturn fmt.Errorf(\"failed to get objectType:%s, objectName:%s, objectVersion:%s %s\", objectType, objectName, objectVersion, sanitisedErr)\n}", "func getErrorMessage(e validator.FieldError) string {\n\ttag := e.Tag()\n\t// StructNamespace returns the namespace for the field error, with the field's actual name.\n\tfieldName := e.StructNamespace()\n\tfieldValue := e.Param()\n\tvar msg string\n\tswitch tag {\n\tcase \"uuid\":\n\t\tmsg = fmt.Sprintf(\"%s field needs a uuid\", fieldName)\n\tcase \"required\":\n\t\tmsg = fmt.Sprintf(\"%s field is required\", fieldName)\n\tcase \"required_without\":\n\t\tmsg = fmt.Sprintf(\"%s field is required if the %s is not present\", fieldName, fieldValue)\n\tcase \"len\":\n\t\tmsg = fmt.Sprintf(\"The length of %s field is not %s\", fieldName, fieldValue)\n\tcase \"oneof\":\n\t\tmsg = fmt.Sprintf(\"%s field should be one of %s\", fieldName, fieldValue)\n\tcase \"gt\":\n\t\tmsg = fmt.Sprintf(\"%s field should greater than %s\", fieldName, fieldValue)\n\tcase dtoDurationTag:\n\t\tmsg = fmt.Sprintf(\"%s field should follows the ISO 8601 Durations format. Eg,100ms, 24h\", fieldName)\n\tcase dtoUuidTag:\n\t\tmsg = fmt.Sprintf(\"%s field needs a uuid\", fieldName)\n\tcase dtoNoneEmptyStringTag:\n\t\tmsg = fmt.Sprintf(\"%s field should not be empty string\", fieldName)\n\tcase dtoRFC3986UnreservedCharTag, emptyOrDtoRFC3986UnreservedCharTag:\n\t\tmsg = fmt.Sprintf(\"%s field only allows unreserved characters which are ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789-_~:;=\", fieldName)\n\tdefault:\n\t\tmsg = fmt.Sprintf(\"%s field validation failed on the %s tag\", fieldName, tag)\n\t}\n\treturn msg\n}", "func errTranslator(err error) (C.int, bool) {\n\tswitch err {\n\tcase nil:\n\t\treturn C.NFS3_OK, true\n\tcase os.ErrPermission:\n\t\treturn C.NFS3ERR_ACCES, true\n\tcase os.ErrNotExist:\n\t\treturn C.NFS3ERR_NOENT, true\n\tcase os.ErrInvalid:\n\t\treturn C.NFS3ERR_INVAL, true\n\tcase os.ErrExist:\n\t\treturn C.NFS3ERR_EXIST, true\n\tdefault:\n\t\tswitch {\n\t\tcase strings.Contains(err.Error(), \"not empty\"):\n\t\t\treturn C.NFS3ERR_NOTEMPTY, true\n\t\tdefault:\n\t\t\treturn C.NFS3ERR_IO, false\n\t\t}\n\t}\n}", "func errStr(err error) string {\n\tif err == nil {\n\t\treturn \"\"\n\t}\n\treturn err.Error()\n}", "func toErrMsg(err error) string {\n\treturn recursErrMsg(\"\", err)\n}", "func CheckValueErrStr(value string, err error) *wrappers.StringValue {\n\tif err == nil && value != \"\" {\n\t\treturn &wrappers.StringValue{Value: value}\n\t}\n\n\treturn nil\n}", "func (ec ErrCode) StringRU() string { return ec.errRu }", "func errorToString(err error) string {\n\tif err == nil {\n\t\treturn \"\"\n\t}\n\treturn err.Error()\n}", "func errToString(err error) string {\n\tif err == nil {\n\t\treturn \"\"\n\t}\n\treturn err.Error()\n}", "func errToString(err error) string {\n\tif err == nil {\n\t\treturn \"\"\n\t}\n\treturn err.Error()\n}", "func translate(err error) error {\n\tswitch errors.Cause(err) {\n\tcase user.ErrNotFound:\n\t\treturn web.ErrNotFound\n\tcase user.ErrInvalidID:\n\t\treturn web.ErrInvalidID\n\tcase user.ErrAuthenticationFailure:\n\t\treturn web.ErrUnauthorized\n\tcase user.ErrForbidden:\n\t\treturn web.ErrForbidden\n\t}\n\treturn err\n}", "func String(prefix interface{}, values map[string]interface{}, wrapped error) string {\n\tmsgs := []string{}\n\n\tmsg := \"\"\n\tif prefix != nil {\n\t\tswitch p := prefix.(type) {\n\t\tcase string:\n\t\t\tmsg = p\n\t\tcase []byte:\n\t\t\tmsg = string(p)\n\t\tdefault:\n\t\t\tmsg = strings.TrimPrefix(reflect.TypeOf(prefix).String(), \"*\")\n\t\t}\n\t}\n\n\tif msg != \"\" {\n\t\tmsgs = append(msgs, msg)\n\t}\n\n\tif len(values) > 0 {\n\t\tkeys := []string{}\n\t\tfor k := range values {\n\t\t\tkeys = append(keys, k)\n\t\t}\n\t\tsort.Strings(keys)\n\n\t\tkv := []string{}\n\t\tfor _, k := range keys {\n\t\t\tkv = append(kv, fmt.Sprintf(\"%s=%v\", k, values[k]))\n\t\t}\n\n\t\tmsgs = append(msgs, \"(\"+strings.Join(kv, \" \")+\")\")\n\n\t}\n\n\tmsg = strings.Join(msgs, \" \")\n\terrMsg := \"\"\n\tif wrapped != nil {\n\t\terrMsg = wrapped.Error()\n\t}\n\n\tswitch {\n\tcase msg == \"\" && errMsg == \"\":\n\t\treturn \"unspecified error\"\n\tcase msg == \"\" && errMsg != \"\":\n\t\treturn errMsg\n\tcase msg != \"\" && errMsg == \"\":\n\t\treturn msg\n\t}\n\n\treturn fmt.Sprintf(\"%s: %s\", msg, errMsg)\n}", "func convertType(str string) (string, error) {\n\tif str == \"\" {\n\t\treturn \"\", fmt.Errorf(\"convertingType '%s' (empty string)\", str)\n\t}\n\n\tdataType, exists := dataTypes[strings.ToLower(strings.TrimSpace(str))]\n\tif exists {\n\t\treturn dataType, nil\n\t}\n\n\treturn str, fmt.Errorf(\"convertingType '%s'\", strings.ToLower(str))\n}", "func convertMgoError(err error) error {\n\tif err == nil {\n\t\treturn nil\n\t} else if err == mgo.ErrNotFound {\n\t\treturn ErrNotFound\n\t}\n\n\treturn ErrUnknown\n}", "func TestTypeName(t *testing.T) {\n\terr1 := &myE{\"woo\"}\n\terr2 := &myE{\"\"}\n\n\ttn1 := errbase.GetTypeMark(err1)\n\ttn2 := errbase.GetTypeMark(err2)\n\n\ttt := testutils.T{T: t}\n\n\ttt.Check(!tn1.Equals(tn2))\n\ttt.CheckEqual(tn1.FamilyName, tn2.FamilyName)\n\ttt.Check(tn1.Extension != tn2.Extension)\n}", "func (e errorString) Error() string {\n\treturn string(e)\n}", "func (e errorString) Error() string {\n\treturn string(e)\n}", "func (i ParseError) String() string { return toString(i) }", "func typeMismatchReason(pValue interface{}, v reflect.Value) string {\n\tentityType := \"empty\"\n\tswitch pValue.(type) {\n\tcase int64:\n\t\tentityType = \"int\"\n\tcase bool:\n\t\tentityType = \"bool\"\n\tcase string:\n\t\tentityType = \"string\"\n\tcase float64:\n\t\tentityType = \"float\"\n\tcase *Key:\n\t\tentityType = \"*datastore.Key\"\n\tcase time.Time:\n\t\tentityType = \"time.Time\"\n\tcase appengine.BlobKey:\n\t\tentityType = \"appengine.BlobKey\"\n\tcase appengine.GeoPoint:\n\t\tentityType = \"appengine.GeoPoint\"\n\tcase ByteString:\n\t\tentityType = \"datastore.ByteString\"\n\tcase []byte:\n\t\tentityType = \"[]byte\"\n\t}\n\treturn fmt.Sprintf(\"type mismatch: %s versus %v\", entityType, v.Type())\n}", "func typeMismatchReason(pValue interface{}, v reflect.Value) string {\n\tentityType := \"empty\"\n\tswitch pValue.(type) {\n\tcase int64:\n\t\tentityType = \"int\"\n\tcase bool:\n\t\tentityType = \"bool\"\n\tcase string:\n\t\tentityType = \"string\"\n\tcase float64:\n\t\tentityType = \"float\"\n\tcase *Key:\n\t\tentityType = \"*datastore.Key\"\n\tcase time.Time:\n\t\tentityType = \"time.Time\"\n\tcase appengine.BlobKey:\n\t\tentityType = \"appengine.BlobKey\"\n\tcase appengine.GeoPoint:\n\t\tentityType = \"appengine.GeoPoint\"\n\tcase ByteString:\n\t\tentityType = \"datastore.ByteString\"\n\tcase []byte:\n\t\tentityType = \"[]byte\"\n\t}\n\treturn fmt.Sprintf(\"type mismatch: %s versus %v\", entityType, v.Type())\n}", "func replaceErrors(a []interface{}) {\n\tfor i, v := range a {\n\t\tif err, ok := v.(error); ok {\n\t\t\ta[i] = errors.String(err)\n\t\t}\n\t}\n}", "func (t OAuthErrorType) String() string {\n\treturn string(t)\n}", "func getErrOfMalformedInput(v interface{}, excludedFields []string) string {\n\terrInfo := make(map[string]string)\n\n\tval := reflect.ValueOf(v).Elem()\n\tfor i := 0; i < val.NumField(); i++ {\n\n\t\tfld := val.Type().Field(i)\n\n\t\tif helpers.StrSliceIndexOf(excludedFields, fld.Name) == -1 {\n\t\t\tjsonFldName := fld.Tag.Get(\"json\")\n\t\t\tif jsonFldName == \"\" {\n\t\t\t\tjsonFldName = fld.Name\n\t\t\t}\n\n\t\t\t// errInfo[jsonFldName] = fld.Type.Kind().String()\n\n\t\t\terrInfo[jsonFldName] = getSimpleType(fld.Type)\n\t\t}\n\t}\n\t/**/\n\tresErrText := \"Malformed payload.\"\n\terrInfoTxt, err := json.Marshal(errInfo)\n\tif err == nil {\n\t\treturn resErrText + \" The payload should look like: \\n\" + string(errInfoTxt)\n\t}\n\n\treturn resErrText\n}", "func (e *TypeError) ToTypeString() *TypeString {\n\treturn String(e.Msg())\n}", "func GetErrString(a, b byte) string {\n\treturn errmap[stringify(a, b)]\n}", "func castType(typed types.Type, value interface{}) string {\n\tvar casted string\n\tswitch typed {\n\tcase types.Double:\n\t\tcasted = Float64Empty(value)\n\tcase types.Int64:\n\t\tcasted = Int64Empty(value)\n\tcase types.Uint64:\n\t\tcasted = Uint64Empty(value)\n\tcase types.Fixed64:\n\t\tcasted = Uint64Empty(value)\n\tcase types.Int32:\n\t\tcasted = Int32Empty(value)\n\tcase types.Uint32:\n\t\tcasted = Uint32Empty(value)\n\tcase types.Fixed32:\n\t\tcasted = Uint32Empty(value)\n\tcase types.Float:\n\t\tcasted = Float32Empty(value)\n\tcase types.String:\n\t\tcasted = StringEmpty(value)\n\tcase types.Enum:\n\t\tcasted = StringEmpty(value)\n\tcase types.Bool:\n\t\tcasted = BoolEmpty(value)\n\tcase types.Bytes:\n\t\tcasted = BytesBase64Empty(value)\n\tcase types.Sfixed32:\n\t\tcasted = Int32Empty(value)\n\tcase types.Sfixed64:\n\t\tcasted = Int64Empty(value)\n\tcase types.Sint32:\n\t\tcasted = Int32Empty(value)\n\tcase types.Sint64:\n\t\tcasted = Int64Empty(value)\n\t}\n\n\treturn casted\n}", "func GetStringErr(key string) (string, error) {\n\treturn c.GetStringErr(key)\n}", "func getValidationMessage(err error) error {\n\n\tfor _, errz := range err.(validator.ValidationErrors) {\n\t\t// Build the custom errors here\n\t\tswitch tag := errz.ActualTag(); tag {\n\t\tcase \"required\":\n\t\t\treturn fmt.Errorf(errz.StructField() + \" is required.\")\n\t\tcase \"email\":\n\t\t\treturn fmt.Errorf(errz.StructField() + \" is an invalid email address.\")\n\t\tcase \"min\":\n\t\t\tif errz.Type().Kind() == reflect.String {\n\t\t\t\treturn fmt.Errorf(errz.StructField() + \" must be more than or equal to \" + errz.Param() + \" character(s).\")\n\t\t\t}\n\t\t\treturn fmt.Errorf(errz.StructField() + \" must be larger than \" + errz.Param() + \".\")\n\n\t\tcase \"max\":\n\t\t\tif errz.Type().Kind() == reflect.String {\n\t\t\t\treturn fmt.Errorf(errz.StructField() + \" must be lesser than or equal to \" + errz.Param() + \" character(s).\")\n\t\t\t}\n\t\t\treturn fmt.Errorf(errz.StructField() + \" must be smaller than \" + errz.Param() + \".\")\n\n\t\tdefault:\n\t\t\treturn fmt.Errorf(errz.StructField() + \" is invalid.\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func castString(input map[string]interface{}, key string) *string {\n\tvar intf interface{}\n\tvar rv string\n\tvar ok bool\n\n\tintf, ok = input[key]\n\tif !ok {\n\t\treturn nil\n\t}\n\n\trv, ok = intf.(string)\n\tif !ok {\n\t\tlog.Print(\"WARNING: type mismatch for field \", key,\n\t\t\t\". Expected string, found \", reflect.TypeOf(intf).Name())\n\t\treturn nil\n\t}\n\n\treturn proto.String(rv)\n}", "func String(uri string) string {\n\tmsk, _ := StringError(uri)\n\treturn msk\n}", "func String(r interface{}, err error) (string, error) {\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tswitch r := r.(type) {\n\tcase []byte:\n\t\treturn string(r), nil\n\tcase string:\n\t\treturn r, nil\n\tcase nil:\n\t\treturn \"\", simplesessions.ErrNil\n\t}\n\treturn \"\", simplesessions.ErrAssertType\n}", "func getErrorText(err error, message string) string {\n\tfor _, err := range err.(validator.ValidationErrors) {\n\t\tmessage += \"Field \" + err.Field() + \" with value \" + fmt.Sprintf(\"%v\", err.Value()) + \" failed \" + err.Tag() + \" tag validation. \"\n\t}\n\treturn message\n}", "func RedisConvErrStr(val string) string {\n\treturn fmt.Sprintf(\"-%s\\r\\n\", val)\n}", "func (e IDEmptyError) Error() string {\n\treturn \"don't know what to do with \\\"\\\"\"\n}", "func errHndlr(errorFrom string, err error) {\n\tif err != nil {\n\t\tfmt.Println(\"error:\", errorFrom, \":: \", err)\n\t}\n}", "func Wrap() {\n\te := errors.New(\"standard error\")\n\n\tfmt.Println(\"Regular Error - \", WrappedError(e))\n\n\tfmt.Println(\"Typed Error - \", WrappedError(ErrorTyped{errors.New(\"typed error\")}))\n\n\tfmt.Println(\"Nil -\", WrappedError(nil))\n\n}", "func handleErr(r interface{}) error {\n\tswitch v := r.(type) {\n\tcase map[string]interface{}:\n\t\tcode, ok := v[\"err_code\"]\n\t\tif !ok {\n\t\t\treturn nil\n\t\t}\n\n\t\terr, ok := v[\"err_msg\"]\n\t\tif !ok {\n\t\t\treturn nil\n\t\t}\n\n\t\treturn fmt.Errorf(\"error %g: %v\", code, err)\n\tdefault:\n\t\treturn fmt.Errorf(\"I don't know about type %T!\\n\", v)\n\t}\n}", "func rErrFmt(err error) string {\n\tif err != nil {\n\t\treturn err.Error()\n\t}\n\treturn \"\"\n}", "func typeString(t []TypeName) (str string) {\n\tswitch len(t) {\n\tcase 0:\n\t\tbreak\n\tcase 1:\n\t\tif t[0].Location.Line == 0 {\n\t\t\t// Use the empty string for undeclared\n\t\t\t// implicit types (such as object).\n\t\t\tbreak\n\t\t}\n\t\tstr = t[0].Str\n\tdefault:\n\t\tstr = \"(either\"\n\t\tfor _, n := range t {\n\t\t\tstr += \" \" + n.Str\n\t\t}\n\t\tstr += \")\"\n\t}\n\treturn\n}", "func TranslateStringToType(tagType TagTypePrimitive, valueString string) (value interface{}, err error) {\n\tdefer func() {\n\t\tif state := recover(); state != nil {\n\t\t\terr = log.Wrap(state.(error))\n\t\t}\n\t}()\n\n\tif tagType == TypeUndefined {\n\t\t// The caller should just call String() on the decoded type.\n\t\tlog.Panicf(\"undefined-type values are not supported\")\n\t}\n\n\tif tagType == TypeByte {\n\t\twide, err := strconv.ParseInt(valueString, 16, 8)\n\t\tlog.PanicIf(err)\n\n\t\treturn byte(wide), nil\n\t} else if tagType == TypeAscii || tagType == TypeAsciiNoNul {\n\t\t// Whether or not we're putting an NUL on the end is only relevant for\n\t\t// byte-level encoding. This function really just supports a user\n\t\t// interface.\n\n\t\treturn valueString, nil\n\t} else if tagType == TypeShort {\n\t\tn, err := strconv.ParseUint(valueString, 10, 16)\n\t\tlog.PanicIf(err)\n\n\t\treturn uint16(n), nil\n\t} else if tagType == TypeLong {\n\t\tn, err := strconv.ParseUint(valueString, 10, 32)\n\t\tlog.PanicIf(err)\n\n\t\treturn uint32(n), nil\n\t} else if tagType == TypeRational {\n\t\tparts := strings.SplitN(valueString, \"/\", 2)\n\n\t\tnumerator, err := strconv.ParseUint(parts[0], 10, 32)\n\t\tlog.PanicIf(err)\n\n\t\tdenominator, err := strconv.ParseUint(parts[1], 10, 32)\n\t\tlog.PanicIf(err)\n\n\t\treturn Rational{\n\t\t\tNumerator: uint32(numerator),\n\t\t\tDenominator: uint32(denominator),\n\t\t}, nil\n\t} else if tagType == TypeSignedLong {\n\t\tn, err := strconv.ParseInt(valueString, 10, 32)\n\t\tlog.PanicIf(err)\n\n\t\treturn int32(n), nil\n\t} else if tagType == TypeFloat {\n\t\tn, err := strconv.ParseFloat(valueString, 32)\n\t\tlog.PanicIf(err)\n\n\t\treturn float32(n), nil\n\t} else if tagType == TypeDouble {\n\t\tn, err := strconv.ParseFloat(valueString, 64)\n\t\tlog.PanicIf(err)\n\n\t\treturn float64(n), nil\n\t} else if tagType == TypeSignedRational {\n\t\tparts := strings.SplitN(valueString, \"/\", 2)\n\n\t\tnumerator, err := strconv.ParseInt(parts[0], 10, 32)\n\t\tlog.PanicIf(err)\n\n\t\tdenominator, err := strconv.ParseInt(parts[1], 10, 32)\n\t\tlog.PanicIf(err)\n\n\t\treturn SignedRational{\n\t\t\tNumerator: int32(numerator),\n\t\t\tDenominator: int32(denominator),\n\t\t}, nil\n\t}\n\n\tlog.Panicf(\"from-string encoding for type not supported; this shouldn't happen: [%s]\", tagType.String())\n\treturn nil, nil\n}", "func recursErrMsg(errMsg string, err error) string {\n\tif errMsg != \"\" {\n\t\terrMsg += \": \"\n\t}\n\n\tswitch tErr := err.(type) {\n\tcase autorest.DetailedError:\n\t\treturn recursErrMsg(errMsg+tErr.Message, tErr.Original)\n\tcase *azure.RequestError:\n\t\tif tErr.DetailedError.Original != nil {\n\t\t\treturn recursErrMsg(errMsg+tErr.DetailedError.Message, tErr.DetailedError.Original)\n\t\t}\n\t\tif tErr.ServiceError != nil {\n\t\t\treturn errMsg + tErr.ServiceError.Message\n\t\t}\n\tcase adal.TokenRefreshError:\n\t\t// This type of error is returned when the OAuth authentication with Azure Active Directory fails, often\n\t\t// due to an invalid or expired secret.\n\t\t//\n\t\t// The associated message is typically opaque and contains elements that are unique to each request\n\t\t// (trace/correlation IDs, timestamps), which causes an infinite loop of reconciliation if propagated to\n\t\t// the object's status conditions.\n\t\t// Instead of resorting to over-engineered error parsing techniques to get around the verbosity of the\n\t\t// message, we simply return a short and generic error description.\n\t\treturn errMsg + \"failed to refresh token: the provided secret is either invalid or expired\"\n\t}\n\n\treturn errMsg + err.Error()\n}", "func nullError(w, text string) *SplitError {\n return &SplitError{w, text, ErrNull}\n}", "func errorsToStrings(errs []error) []string {\n\tvar result []string\n\tfor _, err := range errs {\n\t\tresult = append(result, err.Error())\n\t}\n\n\treturn result\n}", "func formatError(err error, stdErr string) error {\n\tparts := strings.Split(strings.TrimSuffix(stdErr, \"\\n\"), \"\\n\")\n\tif len(parts) > 1 {\n\t\treturn errors.New(strings.Join(parts[1:], \" \"))\n\t}\n\treturn err\n}", "func MessageOf(err error) string {\n\tif err == nil {\n\t\treturn \"\"\n\t}\n\tif e, ok := err.(interface{ Message() string }); ok {\n\t\treturn e.Message()\n\t}\n\treturn err.Error()\n}", "func transformError(resourceType string, resource string, err error) error {\n\tswitch err.(type) {\n\tcase *find.NotFoundError, *find.DefaultNotFoundError:\n\t\treturn k8serrors.NewNotFound(schema.GroupResource{Group: \"vmoperator.vmware.com\", Resource: strings.ToLower(resourceType)}, resource)\n\tcase *find.MultipleFoundError, *find.DefaultMultipleFoundError:\n\t\t// Transform?\n\t\treturn err\n\tdefault:\n\t\treturn err\n\t}\n}", "func ECToString(errCode ecdef.ErrCode) string {\n\tr := \"\"\n\tswitch errCode {\n\tcase Error:\n\t\tr = \"math error\"\n\t\t// Generic floating point error.\n\tcase FloatingPoint:\n\t\tr = \"floating point error\"\n\t// Generic overflow error.\n\tcase Overflow:\n\t\tr = \"overflow\"\n\t// Division by zero.\n\tcase ZeroDivision:\n\t\tr = \"division by zero\"\n\tdefault:\n\t\tr = fmt.Sprintf(\"unknown ecmath error code (%d)\", errCode)\n\t}\n\treturn r\n}", "func parseError(m mxj.Map) string {\n\terrMsg, err := m.ValueForPathString(\"Envelope.Body.Fault.faultstring\")\n\tif err != nil {\n\t\treturn m.StringIndentNoTypeInfo(0)\n\t}\n\n\treturn errMsg\n}", "func ECToString(errCode ecdef.ErrCode) string {\n\tr := \"\"\n\tswitch errCode {\n\tcase Error:\n\t\tr = \"file system error\"\n\tcase NotAFile:\n\t\tr = \"not a file\"\n\tcase NotADir:\n\t\tr = \"not a directory\"\n\tcase NotASymlink:\n\t\tr = \"not a symlink\"\n\tcase NotAHardlink:\n\t\tr = \"not a hardlink\"\n\tcase FileCorrupt:\n\t\tr = \"file is corrupt\"\n\tcase FileTooLarge:\n\t\tr = \"file is too large\"\n\tcase InvalidPath:\n\t\tr = \"invalid path\"\n\tdefault:\n\t\tr = fmt.Sprintf(\"unknown ecfs error code (%d)\", errCode)\n\t}\n\treturn r\n}", "func TestValues_FormatOf_EdgeCases(t *testing.T) {\n\tvar err *errors.Validation\n\n\terr = FormatOf(\"path\", \"in\", \"bugz\", \"\", nil)\n\tif assert.NotNil(t, err) {\n\t\tassert.Equal(t, int(err.Code()), int(errors.InvalidTypeCode))\n\t\tassert.Contains(t, err.Error(), \"bugz is an invalid type name\")\n\t}\n\n\terr = FormatOf(\"path\", \"in\", \"bugz\", \"\", strfmt.Default)\n\tif assert.NotNil(t, err) {\n\t\tassert.Equal(t, int(err.Code()), int(errors.InvalidTypeCode))\n\t\tassert.Contains(t, err.Error(), \"bugz is an invalid type name\")\n\t}\n}", "func Stoerr(s string) error {\n\treturn &errorString{s}\n}", "func catchnil(input string) string {\n\tvar output string\n\n\tparts := strings.Split(strings.Replace(input, `\\.`, `_~_|_~_`, -1), \".\")\n\tfor i := range parts {\n\t\toutput += strings.Join(parts[:i+1], \".\") + \" == nil\"\n\t\tif i+1 < len(parts) {\n\t\t\toutput += \" ||\\n\\t\\t\"\n\t\t}\n\t}\n\toutput = strings.Replace(output, `_~_|_~_`, `.`, -1)\n\n\treturn output\n}", "func unknownIfEmpty(t string) string {\n\tif t == \"\" {\n\t\tt = \"unknown\"\n\t}\n\treturn t\n}", "func internalServerError(resp *ApiResponse, msg string) error {\n if msg == \"\" {\n msg = http.StatusText(http.StatusInternalServerError)\n }\n\n resp.StatusCode = http.StatusInternalServerError\n resp.Message = []byte(msg)\n resp.ErrorMessage = msg\n\n return nil\n}", "func ErrFromString(str string) *Err {\n\tvar e Err\n\tif err := json.Unmarshal([]byte(str), &e); err != nil || e.Code < 10000 {\n\t\treturn &Err{ErrInternalFromString.Code, str}\n\t}\n\treturn &e\n}", "func translateError(err error) (errc int) {\n\tif err == nil {\n\t\treturn 0\n\t}\n\tcause := errors.Cause(err)\n\tif mErr, ok := cause.(mountlib.Error); ok {\n\t\tswitch mErr {\n\t\tcase mountlib.OK:\n\t\t\treturn 0\n\t\tcase mountlib.ENOENT:\n\t\t\treturn -fuse.ENOENT\n\t\tcase mountlib.ENOTEMPTY:\n\t\t\treturn -fuse.ENOTEMPTY\n\t\tcase mountlib.EEXIST:\n\t\t\treturn -fuse.EEXIST\n\t\tcase mountlib.ESPIPE:\n\t\t\treturn -fuse.ESPIPE\n\t\tcase mountlib.EBADF:\n\t\t\treturn -fuse.EBADF\n\t\tcase mountlib.EROFS:\n\t\t\treturn -fuse.EROFS\n\t\t}\n\t}\n\tfs.Errorf(nil, \"IO error: %v\", err)\n\treturn -fuse.EIO\n}", "func errstr(errno int) string {\n\t// ask windows for the remaining errors\n\tvar flags uint32 = syscall.FORMAT_MESSAGE_FROM_SYSTEM | syscall.FORMAT_MESSAGE_ARGUMENT_ARRAY | syscall.FORMAT_MESSAGE_IGNORE_INSERTS\n\tb := make([]uint16, 300)\n\tn, err := syscall.FormatMessage(flags, 0, uint32(errno), 0, b, nil)\n\tif err != nil {\n\t\treturn fmt.Sprintf(\"error %d (FormatMessage failed with: %v)\", errno, err)\n\t}\n\t// trim terminating \\r and \\n\n\tfor ; n > 0 && (b[n-1] == '\\n' || b[n-1] == '\\r'); n-- {\n\t}\n\treturn string(utf16.Decode(b[:n]))\n}", "func makeError(format string, a ...interface{}) error {\n\treturn errors.New(fmt.Sprintf(format, a...))\n}", "func (p statisticsActionProps) tr(in string, err error) string {\n\tvar (\n\t\tpairs = []string{\"{err}\"}\n\t)\n\n\tif err != nil {\n\t\tfor {\n\t\t\t// Unwrap errors\n\t\t\tue := errors.Unwrap(err)\n\t\t\tif ue == nil {\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\terr = ue\n\t\t}\n\n\t\tpairs = append(pairs, err.Error())\n\t} else {\n\t\tpairs = append(pairs, \"nil\")\n\t}\n\treturn strings.NewReplacer(pairs...).Replace(in)\n}", "func (this validate) stringParam(\n\trawValue *model.Data,\n\tparam *model.StringParam,\n) (errs []error) {\n\terrs = []error{}\n\n\t// handle no value passed\n\tif nil == rawValue {\n\t\terrs = append(errs, errors.New(\"String required\"))\n\t\treturn\n\t}\n\n\tvalue := rawValue.String\n\tif \"\" == value && \"\" != param.Default {\n\t\t// apply default if value not set\n\t\tvalue = param.Default\n\t}\n\n\t// guard no constraints\n\tif paramConstraints := param.Constraints; nil != paramConstraints {\n\n\t\t// perform validations supported by gojsonschema\n\t\tconstraintsJsonBytes, err := format.NewJsonFormat().From(paramConstraints)\n\t\tif err != nil {\n\t\t\t// handle syntax errors specially\n\t\t\terrs = append(errs, fmt.Errorf(\"Error interpreting constraints; the op likely has a syntax error.\\n Details: %v\", err.Error()))\n\t\t\treturn\n\t\t}\n\n\t\tvalueJsonBytes, err := format.NewJsonFormat().From(value)\n\t\tif err != nil {\n\t\t\t// handle syntax errors specially\n\t\t\terrs = append(errs, fmt.Errorf(\"Error validating parameter.\\n Details: %v\", err.Error()))\n\t\t\treturn\n\t\t}\n\n\t\tresult, err := gojsonschema.Validate(\n\t\t\tgojsonschema.NewBytesLoader(constraintsJsonBytes),\n\t\t\tgojsonschema.NewBytesLoader(valueJsonBytes),\n\t\t)\n\t\tif err != nil {\n\t\t\t// handle syntax errors specially\n\t\t\terrs = append(errs, fmt.Errorf(\"Error interpreting constraints; the op likely has a syntax error.\\n Details: %v\", err.Error()))\n\t\t\treturn\n\t\t}\n\n\t\tfor _, errString := range result.Errors() {\n\t\t\t// enum validation errors include `(root) ` prefix we don't want\n\t\t\terrs = append(errs, errors.New(strings.TrimPrefix(errString.Description(), \"(root) \")))\n\t\t}\n\n\t}\n\n\treturn\n}", "func validate_string(element String, object any, name string) error {\n\tif getter, ok := object.(StringGetter); ok {\n\t\tif value, ok := getter.GetStringOK(name); ok {\n\t\t\treturn element.Validate(value)\n\t\t}\n\t}\n\n\tif getter, ok := object.(PointerGetter); ok {\n\t\tif value, ok := getter.GetPointer(name); ok {\n\t\t\tif typed, ok := value.(*string); ok {\n\t\t\t\treturn element.Validate(*typed)\n\t\t\t}\n\t\t}\n\t}\n\n\tif element.Required {\n\t\treturn derp.NewValidationError(\"schema.validate_string\", \"Required string property is missing\", element, object, name)\n\t}\n\n\treturn nil\n}", "func (m InvalidType) String() string {\n\treturn \"<invalid>\"\n}", "func (err *error_) String() string {\n\tif err.empty() {\n\t\treturn \"no error\"\n\t}\n\treturn fmt.Sprintf(\"%d: %s\", err.pos(), err.msg(nil, nil))\n}", "func ErrorMessage(error_type string, path string) string {\n\treturn error_type + \" \" + path + \": no such file or directory\"\n}", "func tostr(i interface{}) (string, bool) {\n\tswitch rid := i.(type) {\n\tcase string:\n\t\treturn rid, true\n\tcase []byte:\n\t\treturn string(rid), true\n\tdefault:\n\t\treturn \"\", false\n\t}\n}", "func asRequiredString(key string, target *string) cm.ParseFunc {\n\treturn func(data map[string]string) error {\n\t\tif raw, ok := data[key]; ok {\n\t\t\t*target = raw\n\t\t} else {\n\t\t\treturn fmt.Errorf(\"%q is missing\", key)\n\t\t}\n\t\treturn nil\n\t}\n}", "func (err myError) Type() errorType {\n\treturn err.errType\n}", "func coerceErrawr(err error) errors.Error {\n\terrawr, ok := err.(errors.Error)\n\n\tif ok {\n\t\treturn errawr\n\t}\n\n\treturn errors.NewGeneralUnknownError().WithCause(err)\n}", "func TestValueErrorError(t *testing.T) {\n\te := &ValueError{\"abc\", TypeUnknown}\n\tcases := []struct {\n\t\tinput *ValueError\n\t\texpected string\n\t}{\n\t\t{e, \"data: call of abc on Unknown Type\"},\n\t}\n\tfor i, c := range cases {\n\t\tgot := c.input.Error()\n\t\tif got != c.expected {\n\t\t\tt.Errorf(\"case %d response mismatch: expected: '%s, got: '%s'\", i, c.expected, got)\n\t\t}\n\t}\n}", "func errorString(list []string, got string, want string) string {\n\treturn fmt.Sprintf(\"JoinWithCommas(%#v) = \\\"%s\\\", want \\\"%s\\\"\", list, got, want)\n}", "func resolveType(t radius.Type) string {\n\tswitch t {\n\tcase UserName_Type:\n\t\treturn \"User-Name\"\n\tcase UserPassword_Type:\n\t\treturn \"User-Password\"\n\tcase CHAPPassword_Type:\n\t\treturn \"CHAP-Password\"\n\tcase NASIPAddress_Type:\n\t\treturn \"NAS-IP-Address\"\n\tcase NASPort_Type:\n\t\treturn \"NAS-Port\"\n\tcase ServiceType_Type:\n\t\treturn \"Service-Type\"\n\tcase FramedProtocol_Type:\n\t\treturn \"Framed-Protocol\"\n\tcase FramedIPAddress_Type:\n\t\treturn \"Framed-IP-Address\"\n\tcase FramedIPNetmask_Type:\n\t\treturn \"Framed-IP-Netmask\"\n\tcase FramedRouting_Type:\n\t\treturn \"Framed-Routing\"\n\tcase FilterID_Type:\n\t\treturn \"Filter-ID\"\n\tcase FramedMTU_Type:\n\t\treturn \"Framed-MTU\"\n\tcase FramedCompression_Type:\n\t\treturn \"Framed-Compression\"\n\tcase LoginIPHost_Type:\n\t\treturn \"Login-IP-Host\"\n\tcase LoginService_Type:\n\t\treturn \"Login-Service\"\n\tcase LoginTCPPort_Type:\n\t\treturn \"Login-TCP-Port\"\n\tcase ReplyMessage_Type:\n\t\treturn \"Reply-Message\"\n\tcase CallbackNumber_Type:\n\t\treturn \"Callback-Number\"\n\tcase CallbackID_Type:\n\t\treturn \"Callback-ID\"\n\tcase FramedRoute_Type:\n\t\treturn \"Framed-Route\"\n\tcase FramedIPXNetwork_Type:\n\t\treturn \"Framed-IPX-Network\"\n\tcase State_Type:\n\t\treturn \"State\"\n\tcase Class_Type:\n\t\treturn \"Class\"\n\tcase VendorSpecific_Type:\n\t\treturn \"Vendor-Specific\"\n\tcase SessionTimeout_Type:\n\t\treturn \"Session-Timeout\"\n\tcase IdleTimeout_Type:\n\t\treturn \"Idle-Timeout\"\n\tcase TerminationAction_Type:\n\t\treturn \"Termination-Action\"\n\tcase CalledStationID_Type:\n\t\treturn \"Called-Station-ID\"\n\tcase CallingStationID_Type:\n\t\treturn \"Calling-Station-ID\"\n\tcase NASIdentifier_Type:\n\t\treturn \"NAS-Identifier\"\n\tcase ProxyState_Type:\n\t\treturn \"Proxy-State\"\n\tcase LoginLATService_Type:\n\t\treturn \"Login-LAT-Service\"\n\tcase LoginLATNode_Type:\n\t\treturn \"Login-LAT-Node\"\n\tcase LoginLATGroup_Type:\n\t\treturn \"Login-LAT-Group\"\n\tcase FramedAppleTalkLink_Type:\n\t\treturn \"Framed-Apple-Talk-Link\"\n\tcase FramedAppleTalkNetwork_Type:\n\t\treturn \"Framed-Apple-Talk-Network\"\n\tcase FramedAppleTalkZone_Type:\n\t\treturn \"Framed-Apple-Talk-Zone\"\n\tcase CHAPChallenge_Type:\n\t\treturn \"CHAP-Challenge\"\n\tcase NASPortType_Type:\n\t\treturn \"NAS-Port-Type\"\n\tcase PortLimit_Type:\n\t\treturn \"Port-Limit\"\n\tcase LoginLATPort_Type:\n\t\treturn \"Login-LAT-Port\"\n\t}\n\treturn \"Unknown\"\n}", "func Test_NewErrors(t *testing.T) {\n\terr := service.NewErrors()\n\tassert.Equal(t, err.Error(), []interface{}{})\n\tassert.Equal(t, err.NotEmpty(), false)\n\n\terr = service.NewErrors(\"this is error message\")\n\tassert.Equal(t, err.Error(), []interface{}{\"this is error message\"})\n\tassert.Equal(t, err.NotEmpty(), true)\n\n\terr = service.NewErrors()\n\terr.Add(\"error message 1\")\n\terr.Add(\"error message 2\")\n\tassert.Equal(t, err.Error(), []interface{}{\"error message 1\", \"error message 2\"})\n\tassert.Equal(t, err.NotEmpty(), true)\n}", "func handleError(message string, listOfErrors ...error) error {\n\tvar errSlice []string\n\tfor _, e := range listOfErrors {\n\t\tif e != nil {\n\t\t\terrSlice = append(errSlice, e.Error())\n\t\t}\n\t}\n\treturn fmt.Errorf(\"errors: %s, msg: %s\", strings.Join(errSlice, \", \"), message)\n}", "func String(s string, err error) string {\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\treturn s\n}", "func (e GLenum) ErrorString() string {\n\tif e == GLenum_GL_NO_ERROR {\n\t\treturn \"GL_NO_ERROR\"\n\t}\n\treturn e.String()\n}", "func iStringCollectionToStringArrayErr(disp *ole.IDispatch, err error) ([]string, error) {\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif disp == nil {\n\t\treturn nil, nil\n\t}\n\n\tcount, err := oleconv.ToInt32Err(oleutil.GetProperty(disp, \"Count\"))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tstringCollection := make([]string, count)\n\n\tfor i := 0; i < int(count); i++ {\n\t\tstr, err := oleconv.ToStringErr(oleutil.GetProperty(disp, \"Item\", i))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tstringCollection[i] = str\n\t}\n\treturn stringCollection, nil\n}", "func StringToStatus(name string) (Status, error) {\n\tswitch name {\n\tcase Attach.String():\n\t\treturn Attach, nil\n\tcase AutoUpdate.String():\n\t\treturn AutoUpdate, nil\n\tcase Build.String():\n\t\treturn Build, nil\n\tcase Checkpoint.String():\n\t\treturn Checkpoint, nil\n\tcase Cleanup.String():\n\t\treturn Cleanup, nil\n\tcase Commit.String():\n\t\treturn Commit, nil\n\tcase Create.String():\n\t\treturn Create, nil\n\tcase Exec.String():\n\t\treturn Exec, nil\n\tcase ExecDied.String():\n\t\treturn ExecDied, nil\n\tcase Exited.String():\n\t\treturn Exited, nil\n\tcase Export.String():\n\t\treturn Export, nil\n\tcase HealthStatus.String():\n\t\treturn HealthStatus, nil\n\tcase History.String():\n\t\treturn History, nil\n\tcase Import.String():\n\t\treturn Import, nil\n\tcase Init.String():\n\t\treturn Init, nil\n\tcase Kill.String():\n\t\treturn Kill, nil\n\tcase LoadFromArchive.String():\n\t\treturn LoadFromArchive, nil\n\tcase Mount.String():\n\t\treturn Mount, nil\n\tcase NetworkConnect.String():\n\t\treturn NetworkConnect, nil\n\tcase NetworkDisconnect.String():\n\t\treturn NetworkDisconnect, nil\n\tcase Pause.String():\n\t\treturn Pause, nil\n\tcase Prune.String():\n\t\treturn Prune, nil\n\tcase Pull.String():\n\t\treturn Pull, nil\n\tcase Push.String():\n\t\treturn Push, nil\n\tcase Refresh.String():\n\t\treturn Refresh, nil\n\tcase Remove.String():\n\t\treturn Remove, nil\n\tcase Rename.String():\n\t\treturn Rename, nil\n\tcase Renumber.String():\n\t\treturn Renumber, nil\n\tcase Restart.String():\n\t\treturn Restart, nil\n\tcase Restore.String():\n\t\treturn Restore, nil\n\tcase Rotate.String():\n\t\treturn Rotate, nil\n\tcase Save.String():\n\t\treturn Save, nil\n\tcase Start.String():\n\t\treturn Start, nil\n\tcase Stop.String():\n\t\treturn Stop, nil\n\tcase Sync.String():\n\t\treturn Sync, nil\n\tcase Tag.String():\n\t\treturn Tag, nil\n\tcase Unmount.String():\n\t\treturn Unmount, nil\n\tcase Unpause.String():\n\t\treturn Unpause, nil\n\tcase Untag.String():\n\t\treturn Untag, nil\n\t}\n\treturn \"\", fmt.Errorf(\"unknown event status %q\", name)\n}", "func ErrFromGoErr(err error) *Err {\n\tif sysErr, ok := errorMapping[err.Error()]; ok {\n\t\treturn sysErr\n\t}\n\n\tif e, ok := err.(*Err); ok {\n\t\treturn e\n\t}\n\treturn ErrFromString(err.Error())\n}", "func (in ErrInvalidName) BadRequest() {}", "func checkTableName(s string) error {\n\tswitch s {\n\tcase \"\":\n\t\treturn NewError(TypeError, \"A table name must not be empty.\", map[string]interface{}{\n\t\t\t\"name\": s,\n\t\t})\n\tcase \"Bool\", \"Int8\", \"Int16\", \"Int32\", \"Int64\", \"UInt8\", \"UInt16\", \"UInt32\", \"UInt64\",\n\t\t\"Float\", \"ShortText\", \"Text\", \"LongText\", \"Time\", \"WGS84GeoPoint\", \"TokyoGeoPoint\":\n\t\treturn NewError(TypeError, \"The name specifies a built-in type and not available as a table name.\", map[string]interface{}{\n\t\t\t\"name\": s,\n\t\t})\n\t}\n\tif s[0] == '_' {\n\t\treturn NewError(TypeError, \"A table name must not start with '_'.\", map[string]interface{}{\n\t\t\t\"name\": s,\n\t\t})\n\t}\n\tfor _, c := range s {\n\t\tswitch {\n\t\tcase c >= '0' && c <= '9':\n\t\tcase c >= 'A' && c <= 'Z':\n\t\tcase c >= 'a' && c <= 'z':\n\t\tcase c == '_':\n\t\tdefault:\n\t\t\treturn NewError(TypeError, \"A table name must consist of [0-9A-Za-z_].\", map[string]interface{}{\n\t\t\t\t\"name\": s,\n\t\t\t})\n\t\t}\n\t}\n\treturn nil\n}", "func (e ErrorCode) String() string {\n\tswitch e {\n\tcase \"missing-input-secret\":\n\t\treturn \"The secret parameter was not passed.\"\n\tcase \"invalid-input-secret\":\n\t\treturn \"The secret parameter was invalid or did not exist.\"\n\tcase \"missing-input-response\":\n\t\treturn \"The response parameter was not passed.\"\n\tcase \"invalid-input-response\":\n\t\treturn \"The response parameter is invalid or has expired.\"\n\tcase \"bad-request\":\n\t\treturn \"The request was rejected because it was malformed.\"\n\tcase \"timeout-or-duplicate\":\n\t\treturn \"The response parameter has already been validated before.\"\n\tcase \"internal-error\":\n\t\treturn \"An internal error happened while validating the response. The request can be retried.\"\n\t}\n\treturn string(e)\n}", "func (ec ErrCode) StringEN() string { return ec.errEn }", "func FormatFromType(value interface{}, justFirst bool) (phrase string, err error) {\n\tdefer func() {\n\t\tif state := recover(); state != nil {\n\t\t\terr = log.Wrap(state.(error))\n\t\t}\n\t}()\n\n\t// TODO(dustin): !! Add test\n\n\tswitch t := value.(type) {\n\tcase []byte:\n\t\treturn DumpBytesToString(t), nil\n\tcase string:\n\t\tfor i, c := range t {\n\t\t\tif c == 0 {\n\t\t\t\tt = t[:i]\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif isPrintableText(t) == false {\n\t\t\tphrase = fmt.Sprintf(\"string with binary data (%d bytes)\", len(t))\n\t\t\treturn phrase, nil\n\t\t}\n\n\t\treturn t, nil\n\tcase []uint16, []uint32, []int32, []float64, []float32:\n\t\tval := reflect.ValueOf(t)\n\n\t\tif val.Len() == 0 {\n\t\t\treturn \"\", nil\n\t\t}\n\n\t\tif justFirst == true {\n\t\t\tvar valueSuffix string\n\t\t\tif val.Len() > 1 {\n\t\t\t\tvalueSuffix = \"...\"\n\t\t\t}\n\n\t\t\treturn fmt.Sprintf(\"%v%s\", val.Index(0), valueSuffix), nil\n\t\t}\n\n\t\treturn fmt.Sprintf(\"%v\", val), nil\n\tcase []Rational:\n\t\tif len(t) == 0 {\n\t\t\treturn \"\", nil\n\t\t}\n\n\t\tparts := make([]string, len(t))\n\t\tfor i, r := range t {\n\t\t\tparts[i] = fmt.Sprintf(\"%d/%d\", r.Numerator, r.Denominator)\n\n\t\t\tif justFirst == true {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif justFirst == true {\n\t\t\tvar valueSuffix string\n\t\t\tif len(t) > 1 {\n\t\t\t\tvalueSuffix = \"...\"\n\t\t\t}\n\n\t\t\treturn fmt.Sprintf(\"%v%s\", parts[0], valueSuffix), nil\n\t\t}\n\n\t\treturn fmt.Sprintf(\"%v\", parts), nil\n\tcase []SignedRational:\n\t\tif len(t) == 0 {\n\t\t\treturn \"\", nil\n\t\t}\n\n\t\tparts := make([]string, len(t))\n\t\tfor i, r := range t {\n\t\t\tparts[i] = fmt.Sprintf(\"%d/%d\", r.Numerator, r.Denominator)\n\n\t\t\tif justFirst == true {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif justFirst == true {\n\t\t\tvar valueSuffix string\n\t\t\tif len(t) > 1 {\n\t\t\t\tvalueSuffix = \"...\"\n\t\t\t}\n\n\t\t\treturn fmt.Sprintf(\"%v%s\", parts[0], valueSuffix), nil\n\t\t}\n\n\t\treturn fmt.Sprintf(\"%v\", parts), nil\n\tcase fmt.Stringer:\n\t\ts := t.String()\n\t\tif isPrintableText(s) == false {\n\t\t\tphrase = fmt.Sprintf(\"stringable with binary data (%d bytes)\", len(s))\n\t\t\treturn phrase, nil\n\t\t}\n\n\t\t// An undefined value that is documented (or that we otherwise support).\n\t\treturn s, nil\n\tdefault:\n\t\t// Affects only \"unknown\" values, in general.\n\t\tlog.Panicf(\"type can not be formatted into string: %v\", reflect.TypeOf(value).Name())\n\n\t\t// Never called.\n\t\treturn \"\", nil\n\t}\n}", "func ECToString(errCode ecdef.ErrCode) string {\n\tr := \"\"\n\tswitch errCode {\n\tcase Error:\n\t\tr = \"network error\"\n\t\t// Connection was refused by the remote server.\n\tcase ConnectionRefused:\n\t\tr = \"connection refused\"\n\t// Connection was reset by the remote server.\n\tcase ConnectionReset:\n\t\tr = \"connection reset\"\n\t// Connection was aborted (terminated) by the remote server.\n\tcase ConnectionAborted:\n\t\tr = \"connection aborted\"\n\t// Network operation failed because it was not connected yet.\n\tcase NotConnected:\n\t\tr = \"not connected\"\n\t// Socket address is already in use elsewhere.\n\tcase AddrInUse:\n\t\tr = \"address in use\"\n\t// Nonexistent network interface was requested or the address is not local.\n\tcase AddrNotAvailable:\n\t\tr = \"address not available\"\n\tdefault:\n\t\tr = fmt.Sprintf(\"unknown ecnet error code (%d)\", errCode)\n\t}\n\treturn r\n}", "func TestCheckBinaryExprNilGtrString(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `nil > \"abc\"`, env,\n\t\t`cannot convert nil to type int`,\n\t\t`cannot convert \"abc\" to type int`,\n\t\t`invalid operation: nil > \"abc\" (mismatched types <T> and string)`,\n\t)\n\n}", "func type2name(i int) string {\n\tif i == 0 {\n\t\treturn \"Str\"\n\t}\n\treturn \"Num\"\n}", "func convertType(name string) (string,bool,string) {\n\tconversion := \"string\"\n\tdirectConversion := false\n\tprintfstr := \"%s\"\n\t\n\tswitch name {\n\t\t\tcase \"int\":\n\t\t\t\tconversion = \"int\"\n\t\t\t\tdirectConversion = true\n\t\t\t\tprintfstr = \"%d\"\n\t\t\tcase \"long\":\n\t\t\t\tconversion = \"int64\"\n\t\t\t\tdirectConversion = true\n\t\t\t\tprintfstr = \"%d\"\n\t\t\tcase \"ulong\":\n\t\t\t\tconversion = \"uint64\"\n\t\t\t\tdirectConversion = true\n\t\t\t\tprintfstr = \"%d\"\n\t\t\tcase \"ushort\":\n\t\t\t\tconversion = \"uint16\"\n\t\t\t\tdirectConversion = true\n\t\t\t\tprintfstr = \"%d\"\n\t\t\tcase \"bool\":\n\t\t\t\tconversion = \"bool\"\n\t\t\t\tdirectConversion = true\n\t\t\t\tprintfstr = \"%t\"\n\t\t\tcase \"string\":\n\t\t\t\tconversion = \"string\"\n\t\t\t\tdirectConversion = true\n\t\t\t\tprintfstr = \"%s\"\n\t\t\tcase \"TimeSpan\":\n\t\t\t\tconversion = \"int\"\n\t\t\t\tdirectConversion = true\n\t\t\t\tprintfstr = \"%d\"\n\t\t\tbreak\n\t}\n\treturn conversion,directConversion,printfstr\n}", "func formatErrorMessage(status string, v interface{}) string {\n\tstr := \"\"\n\tmetaValue := reflect.ValueOf(v).Elem()\n\n\tif metaValue.Kind() == reflect.Struct {\n\t\tfield := metaValue.FieldByName(\"Title\")\n\t\tif field != (reflect.Value{}) {\n\t\t\tstr = fmt.Sprintf(\"%s\", field.Interface())\n\t\t}\n\n\t\tfield = metaValue.FieldByName(\"Detail\")\n\t\tif field != (reflect.Value{}) {\n\t\t\tstr = fmt.Sprintf(\"%s (%s)\", str, field.Interface())\n\t\t}\n\t}\n\n\treturn strings.TrimSpace(fmt.Sprintf(\"%s %s\", status, str))\n}", "func Unwrap(err error) error { return errors.Unwrap(err) }", "func (errMap ErrMap) StringMap() map[string]string {\n\tstringMap := make(map[string]string)\n\tfor key, value := range errMap {\n\t\tstringMap[key] = value.Error()\n\t}\n\n\treturn stringMap\n}", "func TestCheckBinaryExprStringQuoNil(t *testing.T) {\n\tenv := MakeSimpleEnv()\n\n\texpectCheckError(t, `\"abc\" / nil`, env,\n\t\t`cannot convert \"abc\" to type int`,\n\t\t`cannot convert nil to type int`,\n\t\t`invalid operation: \"abc\" / nil (mismatched types string and <T>)`,\n\t)\n\n}", "func asStringer(t r.Type, v r.Value) (ret string, okay bool) {\n\ttype stringer interface{ String() string }\n\tstringish := t.AssignableTo(r.TypeOf((*stringer)(nil)).Elem())\n\tif stringish {\n\t\tif !v.IsValid() {\n\t\t\tv = r.Zero(t)\n\t\t}\n\t\tif val, ok := v.Interface().(stringer); ok {\n\t\t\tret = val.String()\n\t\t}\n\t\tokay = true // supports stringer, even if nil\n\t}\n\n\treturn\n}" ]
[ "0.6722059", "0.5995395", "0.5995395", "0.5995395", "0.5967785", "0.57880276", "0.5715489", "0.5708418", "0.56861067", "0.5675173", "0.5628014", "0.5618988", "0.5597196", "0.5537179", "0.55328375", "0.55328375", "0.5530634", "0.54776967", "0.5457534", "0.54475886", "0.544039", "0.5430976", "0.5430976", "0.54263705", "0.54232335", "0.54232335", "0.5407036", "0.53558755", "0.5355731", "0.5333966", "0.533339", "0.53325385", "0.5294331", "0.5290317", "0.52825826", "0.52658707", "0.5254053", "0.52443457", "0.5230572", "0.5229699", "0.52252895", "0.52152514", "0.5215067", "0.5209235", "0.52089614", "0.52083343", "0.5201943", "0.51984686", "0.51972306", "0.5149841", "0.5142551", "0.5138901", "0.51303023", "0.5127399", "0.51065034", "0.51026344", "0.5101823", "0.5095469", "0.5070167", "0.50681615", "0.5064785", "0.50631434", "0.5059546", "0.5058439", "0.5048722", "0.5013601", "0.5012196", "0.5003373", "0.4999912", "0.49973223", "0.4996907", "0.49954456", "0.49882203", "0.49778986", "0.49729523", "0.49684188", "0.49551383", "0.49510956", "0.49483377", "0.49464995", "0.49450657", "0.4933347", "0.49280143", "0.4927707", "0.49261078", "0.49189064", "0.49173036", "0.49165744", "0.49098927", "0.49092308", "0.49054947", "0.48963827", "0.48918805", "0.48874286", "0.48860678", "0.48788637", "0.48780712", "0.4877205" ]
0.68429416
2
Expr builds value expressions for InsertBuilder and UpdateBuilder. Ex: .Values(Expr("FROM_UNIXTIME(?)", t))
func Expr(sql string, args ...interface{}) expr { return expr{sql: sql, args: args} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (stmt *statement) SetExpr(field, expr string, args ...interface{}) Statement {\n\t// TODO How to handle both INSERT ... VALUES and SET in ON DUPLICATE KEY UPDATE?\n\tp := 0\n\tfor _, part := range stmt.parts {\n\t\tif part.position == posInsert || part.position == posUpdate {\n\t\t\tp = part.position\n\t\t\tbreak\n\t\t}\n\t}\n\n\tswitch p {\n\tcase posInsert:\n\t\tstmt.addPart(posInsertFields, \"\", field, nil, \", \")\n\t\tstmt.addPart(posValues, \"\", expr, args, \", \")\n\tcase posUpdate:\n\t\tstmt.addPart(posSet, \"SET\", field+\"=\"+expr, args, \", \")\n\t}\n\treturn stmt\n}", "func NewValue(e expr.Expr, ts time.Time, params expr.Params, metadata goexpr.Params) Sequence {\n\tseq := NewSequence(e.EncodedWidth(), 1)\n\tseq.SetUntil(ts)\n\tseq.UpdateValueAt(0, e, params, metadata)\n\treturn seq\n}", "func Expr(val interface{}) Term {\n\tif val == nil {\n\t\treturn Term{\n\t\t\ttermType: p.Term_DATUM,\n\t\t\tdata: nil,\n\t\t}\n\t}\n\n\tswitch val := val.(type) {\n\tcase Term:\n\t\treturn val\n\tcase []interface{}:\n\t\tvals := make([]Term, len(val))\n\t\tfor i, v := range val {\n\t\t\tvals[i] = Expr(v)\n\t\t}\n\n\t\treturn makeArray(vals)\n\tcase map[string]interface{}:\n\t\tvals := make(map[string]Term, len(val))\n\t\tfor k, v := range val {\n\t\t\tvals[k] = Expr(v)\n\t\t}\n\n\t\treturn makeObject(vals)\n\tcase\n\t\tbool,\n\t\tint,\n\t\tint8,\n\t\tint16,\n\t\tint32,\n\t\tint64,\n\t\tuint,\n\t\tuint8,\n\t\tuint16,\n\t\tuint32,\n\t\tuint64,\n\t\tfloat32,\n\t\tfloat64,\n\t\tuintptr,\n\t\tstring,\n\t\t*bool,\n\t\t*int,\n\t\t*int8,\n\t\t*int16,\n\t\t*int32,\n\t\t*int64,\n\t\t*uint,\n\t\t*uint8,\n\t\t*uint16,\n\t\t*uint32,\n\t\t*uint64,\n\t\t*float32,\n\t\t*float64,\n\t\t*uintptr,\n\t\t*string:\n\t\treturn Term{\n\t\t\ttermType: p.Term_DATUM,\n\t\t\tdata: val,\n\t\t}\n\tdefault:\n\t\t// Use reflection to check for other types\n\t\tvalType := reflect.TypeOf(val)\n\t\tvalValue := reflect.ValueOf(val)\n\n\t\tswitch valType.Kind() {\n\t\tcase reflect.Func:\n\t\t\treturn makeFunc(val)\n\t\tcase reflect.Struct, reflect.Map, reflect.Ptr:\n\t\t\tdata, err := encode(val)\n\n\t\t\tif err != nil || data == nil {\n\t\t\t\treturn Term{\n\t\t\t\t\ttermType: p.Term_DATUM,\n\t\t\t\t\tdata: nil,\n\t\t\t\t\tlastErr: err,\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturn Expr(data)\n\n\t\tcase reflect.Slice, reflect.Array:\n\t\t\t// Check if slice is a byte slice\n\t\t\tif valType.Elem().Kind() == reflect.Uint8 {\n\t\t\t\tdata, err := encode(val)\n\n\t\t\t\tif err != nil || data == nil {\n\t\t\t\t\treturn Term{\n\t\t\t\t\t\ttermType: p.Term_DATUM,\n\t\t\t\t\t\tdata: nil,\n\t\t\t\t\t\tlastErr: err,\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\treturn Expr(data)\n\t\t\t}\n\n\t\t\tvals := make([]Term, valValue.Len())\n\t\t\tfor i := 0; i < valValue.Len(); i++ {\n\t\t\t\tvals[i] = Expr(valValue.Index(i).Interface())\n\t\t\t}\n\n\t\t\treturn makeArray(vals)\n\t\tdefault:\n\t\t\tdata, err := encode(val)\n\n\t\t\tif err != nil || data == nil {\n\t\t\t\treturn Term{\n\t\t\t\t\ttermType: p.Term_DATUM,\n\t\t\t\t\tdata: nil,\n\t\t\t\t\tlastErr: err,\n\t\t\t\t}\n\t\t\t}\n\n\t\t\treturn Term{\n\t\t\t\ttermType: p.Term_DATUM,\n\t\t\t\tdata: data,\n\t\t\t}\n\t\t}\n\t}\n}", "func (s *BasePlSqlParserListener) EnterDatetime_expr(ctx *Datetime_exprContext) {}", "func (c *compiledField) compileExpr(expr influxql.Expr) error {\n\tswitch expr := expr.(type) {\n\tcase *influxql.VarRef:\n\t\t// A bare variable reference will require auxiliary fields.\n\t\tc.global.HasAuxiliaryFields = true\n\t\treturn nil\n\tcase *influxql.Wildcard:\n\t\t// Wildcards use auxiliary fields. We assume there will be at least one\n\t\t// expansion.\n\t\tc.global.HasAuxiliaryFields = true\n\t\tif !c.AllowWildcard {\n\t\t\treturn errors.New(\"unable to use wildcard in a binary expression\")\n\t\t}\n\t\treturn nil\n\tcase *influxql.RegexLiteral:\n\t\tif !c.AllowWildcard {\n\t\t\treturn errors.New(\"unable to use regex in a binary expression\")\n\t\t}\n\t\tc.global.HasAuxiliaryFields = true\n\t\treturn nil\n\tcase *influxql.Call:\n\t\tif isMathFunction(expr) {\n\t\t\treturn c.compileMathFunction(expr)\n\t\t}\n\n\t\t// Register the function call in the list of function calls.\n\t\tc.global.FunctionCalls = append(c.global.FunctionCalls, expr)\n\n\t\tswitch expr.Name {\n\t\tcase \"percentile\":\n\t\t\treturn c.compilePercentile(expr.Args)\n\t\tcase \"sample\":\n\t\t\treturn c.compileSample(expr.Args)\n\t\tcase \"distinct\":\n\t\t\treturn c.compileDistinct(expr.Args, false)\n\t\tcase \"top\", \"bottom\":\n\t\t\treturn c.compileTopBottom(expr)\n\t\tcase \"derivative\", \"non_negative_derivative\":\n\t\t\tisNonNegative := expr.Name == \"non_negative_derivative\"\n\t\t\treturn c.compileDerivative(expr.Args, isNonNegative)\n\t\tcase \"difference\", \"non_negative_difference\":\n\t\t\tisNonNegative := expr.Name == \"non_negative_difference\"\n\t\t\treturn c.compileDifference(expr.Args, isNonNegative)\n\t\tcase \"cumulative_sum\":\n\t\t\treturn c.compileCumulativeSum(expr.Args)\n\t\tcase \"moving_average\":\n\t\t\treturn c.compileMovingAverage(expr.Args)\n\t\tcase \"exponential_moving_average\", \"double_exponential_moving_average\", \"triple_exponential_moving_average\", \"relative_strength_index\", \"triple_exponential_derivative\":\n\t\t\treturn c.compileExponentialMovingAverage(expr.Name, expr.Args)\n\t\tcase \"kaufmans_efficiency_ratio\", \"kaufmans_adaptive_moving_average\":\n\t\t\treturn c.compileKaufmans(expr.Name, expr.Args)\n\t\tcase \"chande_momentum_oscillator\":\n\t\t\treturn c.compileChandeMomentumOscillator(expr.Args)\n\t\tcase \"elapsed\":\n\t\t\treturn c.compileElapsed(expr.Args)\n\t\tcase \"integral\":\n\t\t\treturn c.compileIntegral(expr.Args)\n\t\tcase \"holt_winters\", \"holt_winters_with_fit\":\n\t\t\twithFit := expr.Name == \"holt_winters_with_fit\"\n\t\t\treturn c.compileHoltWinters(expr.Args, withFit)\n\t\tdefault:\n\t\t\treturn c.compileFunction(expr)\n\t\t}\n\tcase *influxql.Distinct:\n\t\tcall := expr.NewCall()\n\t\tc.global.FunctionCalls = append(c.global.FunctionCalls, call)\n\t\treturn c.compileDistinct(call.Args, false)\n\tcase *influxql.BinaryExpr:\n\t\t// Disallow wildcards in binary expressions. RewriteFields, which expands\n\t\t// wildcards, is too complicated if we allow wildcards inside of expressions.\n\t\tc.AllowWildcard = false\n\n\t\t// Check if either side is a literal so we only compile one side if it is.\n\t\tif _, ok := expr.LHS.(influxql.Literal); ok {\n\t\t\tif _, ok := expr.RHS.(influxql.Literal); ok {\n\t\t\t\treturn errors.New(\"cannot perform a binary expression on two literals\")\n\t\t\t}\n\t\t\treturn c.compileExpr(expr.RHS)\n\t\t} else if _, ok := expr.RHS.(influxql.Literal); ok {\n\t\t\treturn c.compileExpr(expr.LHS)\n\t\t} else {\n\t\t\t// Validate both sides of the expression.\n\t\t\tif err := c.compileExpr(expr.LHS); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tif err := c.compileExpr(expr.RHS); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn nil\n\t\t}\n\tcase *influxql.ParenExpr:\n\t\treturn c.compileExpr(expr.Expr)\n\tcase influxql.Literal:\n\t\treturn errors.New(\"field must contain at least one variable\")\n\t}\n\treturn errors.New(\"unimplemented\")\n}", "func (l *Layout) GenInsertStmtValues(colVals []string) string {\n\treturn fmt.Sprintf(\"(null,%s,now())\", strings.Join(colVals, \",\"))\n}", "func (t *WindowedTable) FormatExpr() string {\n\tfrom := t.Window.From.UnixNano() / int64(time.Millisecond)\n\tto := t.Window.To.UnixNano() / int64(time.Millisecond)\n\treturn fmt.Sprintf(\"[%v@%v-%v]\", t.ID, from, to)\n}", "func (b *builtinValuesTimeSig) evalTime(_ chunk.Row) (types.Time, bool, error) {\n\trow := b.ctx.GetSessionVars().CurrInsertValues\n\tif row.IsEmpty() {\n\t\treturn types.ZeroTime, true, nil\n\t}\n\tif b.offset < row.Len() {\n\t\tif row.IsNull(b.offset) {\n\t\t\treturn types.ZeroTime, true, nil\n\t\t}\n\t\treturn row.GetTime(b.offset), false, nil\n\t}\n\treturn types.ZeroTime, true, errors.Errorf(\"Session current insert values len %d and column's offset %v don't match\", row.Len(), b.offset)\n}", "func (e Engine) AddNowTime(cols []string) Engine {\n\tvar buf bytes.Buffer\n\tkey := \"\"\n\tfor _, v := range cols {\n\t\tkey += fmt.Sprintf(\",%s\", v)\n\t}\n\te.s = strings.ReplaceAll(e.s, \") values\", fmt.Sprintf(\"%s) values\", key))\n\n\tvalue := \"\"\n\n\tfor i := 0; i < len(cols); i++ {\n\t\tvalue += fmt.Sprintf(\",%s\", \"now()\")\n\t}\n\n\te.s = strings.ReplaceAll(e.s, \"?)\", fmt.Sprintf(\"?%s)\", value))\n\tbuf.WriteString(e.s)\n\te.s = buf.String()\n\treturn e\n}", "func (e *Evaluator) Eval(expr *tipb.Expr) (types.Datum, error) {\n\tswitch expr.GetTp() {\n\tcase tipb.ExprType_Null:\n\t\treturn types.Datum{}, nil\n\tcase tipb.ExprType_Int64:\n\t\treturn e.evalInt(expr.Val)\n\tcase tipb.ExprType_Uint64:\n\t\treturn e.evalUint(expr.Val)\n\tcase tipb.ExprType_String:\n\t\treturn e.evalString(expr.Val)\n\tcase tipb.ExprType_Bytes:\n\t\treturn types.NewBytesDatum(expr.Val), nil\n\tcase tipb.ExprType_Float32:\n\t\treturn e.evalFloat(expr.Val, true)\n\tcase tipb.ExprType_Float64:\n\t\treturn e.evalFloat(expr.Val, false)\n\tcase tipb.ExprType_MysqlDecimal:\n\t\treturn e.evalDecimal(expr.Val)\n\tcase tipb.ExprType_MysqlDuration:\n\t\treturn e.evalDuration(expr.Val)\n\tcase tipb.ExprType_ColumnRef:\n\t\treturn e.evalColumnRef(expr.Val)\n\tcase tipb.ExprType_LT:\n\t\treturn e.evalLT(expr)\n\tcase tipb.ExprType_LE:\n\t\treturn e.evalLE(expr)\n\tcase tipb.ExprType_EQ:\n\t\treturn e.evalEQ(expr)\n\tcase tipb.ExprType_NE:\n\t\treturn e.evalNE(expr)\n\tcase tipb.ExprType_GE:\n\t\treturn e.evalGE(expr)\n\tcase tipb.ExprType_GT:\n\t\treturn e.evalGT(expr)\n\tcase tipb.ExprType_NullEQ:\n\t\treturn e.evalNullEQ(expr)\n\tcase tipb.ExprType_And:\n\t\treturn e.evalAnd(expr)\n\tcase tipb.ExprType_Or:\n\t\treturn e.evalOr(expr)\n\tcase tipb.ExprType_Like:\n\t\treturn e.evalLike(expr)\n\tcase tipb.ExprType_Not:\n\t\treturn e.evalNot(expr)\n\tcase tipb.ExprType_In:\n\t\treturn e.evalIn(expr)\n\tcase tipb.ExprType_Plus, tipb.ExprType_Div:\n\t\treturn e.evalArithmetic(expr)\n\t}\n\treturn types.Datum{}, nil\n}", "func (tp *TablePlan) generateUpdateValues(sql *sqlparser.TrackedBuffer, before, after []sqltypes.Value) bool {\n\tseparator := \"\"\n\thasSet := false\n\tfor _, cExpr := range tp.ColExprs {\n\t\tif cExpr.IsGrouped {\n\t\t\tcontinue\n\t\t}\n\t\tif len(before) != 0 && len(after) != 0 {\n\t\t\tif cExpr.Operation == OpCount {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif valsEqual(before[cExpr.ColNum], after[cExpr.ColNum]) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tsql.Myprintf(\"%s%v=\", separator, cExpr.ColName)\n\t\tseparator = \", \"\n\t\thasSet = true\n\t\tif cExpr.Operation == OpCount || cExpr.Operation == OpSum {\n\t\t\tsql.Myprintf(\"%v\", cExpr.ColName)\n\t\t}\n\t\tif len(before) != 0 {\n\t\t\tswitch cExpr.Operation {\n\t\t\tcase OpNone:\n\t\t\t\tif len(after) == 0 {\n\t\t\t\t\tsql.WriteString(\"NULL\")\n\t\t\t\t}\n\t\t\tcase OpCount:\n\t\t\t\tsql.WriteString(\"-1\")\n\t\t\tcase OpSum:\n\t\t\t\tif !before[cExpr.ColNum].IsNull() {\n\t\t\t\t\tsql.WriteString(\"-\")\n\t\t\t\t\tencodeValue(sql, before[cExpr.ColNum])\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif len(after) != 0 {\n\t\t\tswitch cExpr.Operation {\n\t\t\tcase OpNone:\n\t\t\t\tencodeValue(sql, after[cExpr.ColNum])\n\t\t\tcase OpCount:\n\t\t\t\tsql.WriteString(\"+1\")\n\t\t\tcase OpSum:\n\t\t\t\tif !after[cExpr.ColNum].IsNull() {\n\t\t\t\t\tsql.WriteString(\"+\")\n\t\t\t\t\tencodeValue(sql, after[cExpr.ColNum])\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn hasSet\n}", "func Expr(sql string, args ...interface{}) Cond {\n\treturn expr{sql, args}\n}", "func (stmt *statement) Expr(expr string, args ...interface{}) Statement {\n\tstmt.addPart(stmt.position, \"\", expr, args, \", \")\n\treturn stmt\n}", "func evalAstExpr(ctx sessionctx.Context, expr ast.ExprNode) (types.Datum, error) {\n\ttrace_util_0.Count(_expression_rewriter_00000, 0)\n\tif val, ok := expr.(*driver.ValueExpr); ok {\n\t\ttrace_util_0.Count(_expression_rewriter_00000, 4)\n\t\treturn val.Datum, nil\n\t}\n\ttrace_util_0.Count(_expression_rewriter_00000, 1)\n\tb := &PlanBuilder{\n\t\tctx: ctx,\n\t\tcolMapper: make(map[*ast.ColumnNameExpr]int),\n\t}\n\tif ctx.GetSessionVars().TxnCtx.InfoSchema != nil {\n\t\ttrace_util_0.Count(_expression_rewriter_00000, 5)\n\t\tb.is = ctx.GetSessionVars().TxnCtx.InfoSchema.(infoschema.InfoSchema)\n\t}\n\ttrace_util_0.Count(_expression_rewriter_00000, 2)\n\tfakePlan := LogicalTableDual{}.Init(ctx)\n\tnewExpr, _, err := b.rewrite(expr, fakePlan, nil, true)\n\tif err != nil {\n\t\ttrace_util_0.Count(_expression_rewriter_00000, 6)\n\t\treturn types.Datum{}, err\n\t}\n\ttrace_util_0.Count(_expression_rewriter_00000, 3)\n\treturn newExpr.Eval(chunk.Row{})\n}", "func (t *Time) Expr() Expr {\n\treturn t.Literal\n}", "func (p *InfluxParser) parseExpr(getExpr influxql.Expr, level int, selectors []string, where [][]*WhereCond) (string, ExprReturn, error) {\n\texprType := Unknown\n\tmc2 := \"\"\n\t// Return the function name or variable name, if available.\n\tswitch expr := getExpr.(type) {\n\tcase *influxql.Call:\n\t\tcall, exprTypeCall, err := p.parseCall(expr, selectors, where)\n\t\tif err != nil {\n\t\t\treturn \"\", Unknown, err\n\t\t}\n\t\texprType = exprTypeCall\n\t\tmc2 += call\n\tcase *influxql.BinaryExpr:\n\t\tleftString, leftType, leftError := p.parseExpr(expr.LHS, level+1, selectors, where)\n\t\tif leftError != nil {\n\t\t\treturn \"\", Unknown, leftError\n\t\t}\n\t\trightString, rightType, rightError := p.parseExpr(expr.RHS, level+1, selectors, where)\n\t\tif rightError != nil {\n\t\t\treturn \"\", Unknown, rightError\n\t\t}\n\n\t\tmc2 += leftString + fmt.Sprintf(\" 'left-%d' STORE\\n\", level)\n\t\tmc2 += rightString + fmt.Sprintf(\" 'right-%d' STORE\\n\", level)\n\n\t\tbinaryType := getBinaryOperationType(leftType, rightType)\n\n\t\topMc2, opErr := p.parseMathOperation(binaryType, expr.Op, level)\n\t\tif opErr != nil {\n\t\t\treturn \"\", Unknown, rightError\n\t\t}\n\t\tmc2 += opMc2\n\t\tif leftType == Scalar && rightType == Scalar {\n\t\t\texprType = Scalar\n\t\t} else {\n\t\t\texprType = SeriesSet\n\t\t}\n\tcase *influxql.ParenExpr:\n\t\treturn p.parseExpr(expr.Expr, level+1, selectors, where)\n\tcase *influxql.VarRef:\n\t\tp.HasWildCard = false\n\t\tfetch, exprTypeCall, err := p.parseFetch(expr.Val, expr.Type, selectors, where)\n\t\tif err != nil {\n\t\t\treturn \"\", Unknown, err\n\t\t}\n\t\texprType = exprTypeCall\n\t\tmc2 += fetch\n\tcase *influxql.IntegerLiteral:\n\t\treturn fmt.Sprintf(\" %d \", expr.Val), Scalar, nil\n\tcase *influxql.NumberLiteral:\n\t\treturn fmt.Sprintf(\" %f \", expr.Val), Scalar, nil\n\tdefault:\n\t\tlog.Warnf(\"Parse expr %s : %T\", expr.String(), expr)\n\t}\n\n\treturn mc2, exprType, nil\n\n}", "func (s InsertStatement) QueryAndValues() (string, []interface{}) {\n\tquery := []string{\"INSERT INTO\", fmt.Sprintf(\"%s.%s\", s.Keyspace(), s.Table())}\n\n\tfieldMap := s.FieldMap()\n\tfieldNames := make([]string, 0, len(fieldMap))\n\tplaceholders := make([]string, 0, len(fieldMap))\n\tvalues := make([]interface{}, 0, len(fieldMap))\n\tfor _, field := range sortedKeys(fieldMap) {\n\t\tfieldNames = append(fieldNames, strings.ToLower(field))\n\t\tplaceholders = append(placeholders, \"?\")\n\t\tvalues = append(values, fieldMap[field])\n\t}\n\n\tquery = append(query, \"(\"+strings.Join(fieldNames, \", \")+\")\")\n\tquery = append(query, \"VALUES (\"+strings.Join(placeholders, \", \")+\")\")\n\n\t// Determine if we need to set a TTL\n\tif s.TTL() > time.Duration(0) {\n\t\tquery = append(query, \"USING TTL ?\")\n\t\tvalues = append(values, int(s.TTL().Seconds()))\n\t}\n\n\treturn strings.Join(query, \" \"), values\n}", "func NewExprEvaluator(arrays []*Array, values [][]byte) *ExprEvaluator {\n\tassert(len(arrays) == len(values), \"array/value count mismatch: %d != %d\", len(arrays), len(values))\n\n\tm := make(map[uint64][]byte)\n\tfor i, array := range arrays {\n\t\t_, ok := m[array.ID]\n\t\tassert(!ok, \"duplicate array: id=%d\", array.ID)\n\t\tm[array.ID] = values[i]\n\t}\n\n\treturn &ExprEvaluator{m: m}\n}", "func Expr(expression string, args ...interface{}) *Query {\n\treturn &Query{expression, args}\n}", "func (sink *influxdbSink) parseRawQueryRow(rawRow influx_models.Row) ([]core.TimestampedMetricValue, error) {\n\tvals := make([]core.TimestampedMetricValue, len(rawRow.Values))\n\twasInt := make(map[string]bool, 1)\n\tfor i, rawVal := range rawRow.Values {\n\t\tval := core.TimestampedMetricValue{}\n\n\t\tif ts, err := time.Parse(time.RFC3339, rawVal[0].(string)); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Unable to parse timestamp %q in series %q\", rawVal[0].(string), rawRow.Name)\n\t\t} else {\n\t\t\tval.Timestamp = ts\n\t\t}\n\n\t\tif err := tryParseMetricValue(\"value\", rawVal, &val.MetricValue, 1, wasInt); err != nil {\n\t\t\tglog.Errorf(\"Unable to parse field \\\"value\\\" in series %q: %v\", rawRow.Name, err)\n\t\t\treturn nil, fmt.Errorf(\"Unable to parse values in series %q\", rawRow.Name)\n\t\t}\n\n\t\tvals[i] = val\n\t}\n\n\tif wasInt[\"value\"] {\n\t\tfor i := range vals {\n\t\t\tvals[i].MetricValue.ValueType = core.ValueInt64\n\t\t}\n\t} else {\n\t\tfor i := range vals {\n\t\t\tvals[i].MetricValue.ValueType = core.ValueFloat\n\t\t}\n\t}\n\n\treturn vals, nil\n}", "func Values(values ...interface{}) string {\n\treturn \" VALUES ( \" + mergeValues(values) + \" )\"\n}", "func BuildInsertColumns(val interface{}, placeholder int) (string, []interface{}) {\n\tt := reflect.ValueOf(val).Elem()\n\ttypeOfT := t.Type()\n\n\tvar columnNamesSql string = \"\"\n\tvar valueFieldsSql string = \"\"\n\n\tvar values []interface{}\n\tvar fieldId int = 1\n\tvar tag reflect.StructTag\n\tvar columnName *string\n\tvar field reflect.Value\n\tvar fieldType reflect.StructField\n\n\tfor i := 0; i < t.NumField(); i++ {\n\t\tfieldType = typeOfT.Field(i)\n\t\ttag = fieldType.Tag\n\n\t\tif columnName = GetColumnNameFromTag(&tag); columnName == nil {\n\t\t\tcontinue\n\t\t}\n\t\tif HasColumnAttributeFromTag(&tag, \"serial\") {\n\t\t\tcontinue\n\t\t}\n\n\t\tfield = t.Field(i)\n\n\t\tvar val interface{} = field.Interface()\n\n\t\t// if time is null or with zero value, just skip it.\n\t\tswitch tv := val.(type) {\n\t\tcase *time.Time:\n\t\t\tif tv == nil || tv.Unix() == -62135596800 {\n\t\t\t\tcontinue\n\t\t\t}\n\t\tcase time.Time:\n\t\t\tif tv.Unix() == -62135596800 {\n\t\t\t\tcontinue\n\t\t\t}\n\t\tcase string:\n\t\t\tif HasColumnAttributeFromTag(&tag, \"date\") && tv == \"\" {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\n\t\tcolumnNamesSql += *columnName + \", \"\n\n\t\tif placeholder == QMARK_HOLDER {\n\t\t\tvalueFieldsSql += \"?, \"\n\t\t} else {\n\t\t\tvalueFieldsSql += \"$\" + strconv.Itoa(fieldId) + \", \"\n\t\t}\n\t\tvalues = append(values, val)\n\t\tfieldId++\n\t}\n\treturn \"( \" + columnNamesSql[:len(columnNamesSql)-2] + \" ) \" +\n\t\t\"VALUES ( \" + valueFieldsSql[:len(valueFieldsSql)-2] + \" )\", values\n}", "func (g *Generator) Expression(genConfig ExprGeneratorConfig) Expr {\n\tvar options []exprF\n\t// this will only be used for tuple expressions, everything else will need genConfig.NumCols = 1\n\tnumCols := genConfig.NumCols\n\tgenConfig = genConfig.SetNumCols(1)\n\n\tswitch genConfig.Type {\n\tcase \"bigint\":\n\t\toptions = append(options, func() Expr { return g.intExpr(genConfig) })\n\tcase \"varchar\":\n\t\toptions = append(options, func() Expr { return g.stringExpr(genConfig) })\n\tcase \"tinyint\":\n\t\toptions = append(options, func() Expr { return g.booleanExpr(genConfig) })\n\tcase \"\":\n\t\toptions = append(options, []exprF{\n\t\t\tfunc() Expr { return g.intExpr(genConfig) },\n\t\t\tfunc() Expr { return g.stringExpr(genConfig) },\n\t\t\tfunc() Expr { return g.booleanExpr(genConfig) },\n\t\t}...)\n\t}\n\n\tfor i := range g.exprGenerators {\n\t\tgenerator := g.exprGenerators[i]\n\t\tif generator == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\t// don't create expressions from the expression exprGenerators if we haven't created an aggregation yet\n\t\tif _, ok := generator.(QueryGenerator); ok || genConfig.AggrRule != IsAggregate {\n\t\t\toptions = append(options, func() Expr {\n\t\t\t\texpr := generator.Generate(g.r, genConfig)\n\t\t\t\tif expr == nil {\n\t\t\t\t\treturn g.randomLiteral()\n\t\t\t\t}\n\t\t\t\treturn expr\n\t\t\t})\n\t\t}\n\t}\n\n\tif genConfig.AggrRule != CannotAggregate {\n\t\toptions = append(options, func() Expr {\n\t\t\tg.isAggregate = true\n\t\t\treturn g.randomAggregate(genConfig.CannotAggregateConfig())\n\t\t})\n\t}\n\n\t// if an arbitrary number of columns may be generated, randomly choose 1-3 columns\n\tif numCols == 0 {\n\t\tnumCols = g.r.Intn(3) + 1\n\t}\n\n\tif numCols == 1 {\n\t\treturn g.makeAggregateIfNecessary(genConfig, g.randomOf(options))\n\t}\n\n\t// with 1/5 probability choose a tuple subquery\n\tif g.randomBool(0.2) {\n\t\treturn g.subqueryExpr(genConfig.SetNumCols(numCols))\n\t}\n\n\ttuple := ValTuple{}\n\tfor i := 0; i < numCols; i++ {\n\t\ttuple = append(tuple, g.makeAggregateIfNecessary(genConfig, g.randomOf(options)))\n\t}\n\n\treturn tuple\n}", "func getExprValue(e tree.Expr, mce *MysqlCmdExecutor, ses *Session) (interface{}, error) {\n\t/*\n\t\tCORNER CASE:\n\t\t\tSET character_set_results = utf8; // e = tree.UnresolvedName{'utf8'}.\n\n\t\t\ttree.UnresolvedName{'utf8'} can not be resolved as the column of some table.\n\t*/\n\tswitch v := e.(type) {\n\tcase *tree.UnresolvedName:\n\t\t// set @a = on, type of a is bool.\n\t\treturn v.Parts[0], nil\n\t}\n\n\tvar err error\n\n\ttable := &tree.TableName{}\n\ttable.ObjectName = \"dual\"\n\n\t//1.composite the 'select (expr) from dual'\n\tcompositedSelect := &tree.Select{\n\t\tSelect: &tree.SelectClause{\n\t\t\tExprs: tree.SelectExprs{\n\t\t\t\ttree.SelectExpr{\n\t\t\t\t\tExpr: e,\n\t\t\t\t},\n\t\t\t},\n\t\t\tFrom: &tree.From{\n\t\t\t\tTables: tree.TableExprs{\n\t\t\t\t\t&tree.JoinTableExpr{\n\t\t\t\t\t\tJoinType: tree.JOIN_TYPE_CROSS,\n\t\t\t\t\t\tLeft: &tree.AliasedTableExpr{\n\t\t\t\t\t\t\tExpr: table,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\t//2.run the select\n\tctx := ses.GetRequestContext()\n\n\t//run the statement in the same session\n\tses.ClearResultBatches()\n\terr = executeStmtInSameSession(ctx, mce, ses, compositedSelect)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbatches := ses.GetResultBatches()\n\tif len(batches) == 0 {\n\t\treturn nil, moerr.NewInternalError(ctx, \"the expr %s does not generate a value\", e.String())\n\t}\n\n\tif batches[0].VectorCount() > 1 {\n\t\treturn nil, moerr.NewInternalError(ctx, \"the expr %s generates multi columns value\", e.String())\n\t}\n\n\t//evaluate the count of rows, the count of columns\n\tcount := 0\n\tvar resultVec *vector.Vector\n\tfor _, b := range batches {\n\t\tif b.RowCount() == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tcount += b.RowCount()\n\t\tif count > 1 {\n\t\t\treturn nil, moerr.NewInternalError(ctx, \"the expr %s generates multi rows value\", e.String())\n\t\t}\n\t\tif resultVec == nil && b.GetVector(0).Length() != 0 {\n\t\t\tresultVec = b.GetVector(0)\n\t\t}\n\t}\n\n\tif resultVec == nil {\n\t\treturn nil, moerr.NewInternalError(ctx, \"the expr %s does not generate a value\", e.String())\n\t}\n\n\t// for the decimal type, we need the type of expr\n\t//!!!NOTE: the type here may be different from the one in the result vector.\n\tvar planExpr *plan.Expr\n\toid := resultVec.GetType().Oid\n\tif oid == types.T_decimal64 || oid == types.T_decimal128 {\n\t\tbuilder := plan2.NewQueryBuilder(plan.Query_SELECT, ses.GetTxnCompileCtx(), false)\n\t\tbindContext := plan2.NewBindContext(builder, nil)\n\t\tbinder := plan2.NewSetVarBinder(builder, bindContext)\n\t\tplanExpr, err = binder.BindExpr(e, 0, false)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn getValueFromVector(resultVec, ses, planExpr)\n}", "func (t Timestamp) Value() (driver.Value, error) {\n\treturn time.Time(t), nil\n}", "func NewValues(col sql.Expression) sql.Expression {\n\treturn &Values{\n\t\tUnaryExpression: expression.UnaryExpression{Child: col},\n\t\tValue: nil,\n\t}\n}", "func NewSetExpr(elements ...Expr) Expr {\n\tvalues := make([]Value, len(elements))\n\tfor i, expr := range elements {\n\t\tvalue, ok := expr.(Value)\n\t\tif !ok {\n\t\t\treturn &SetExpr{elements}\n\t\t}\n\t\tvalues[i] = value\n\t}\n\treturn NewSet(values...)\n}", "func writeValuesAsSQL(sql *sqlparser.TrackedBuffer, tce *tableCacheEntry, rs *mysql.Rows, rowIndex int, getPK bool) (sqltypes.Value, []sqltypes.Value, error) {\n\tvalueIndex := 0\n\tdata := rs.Rows[rowIndex].Data\n\tpos := 0\n\tvar keyspaceIDCell sqltypes.Value\n\tvar pkValues []sqltypes.Value\n\tif getPK {\n\t\tpkValues = make([]sqltypes.Value, len(tce.pkNames))\n\t}\n\n\tif len(tce.ti.Fields) != rs.DataColumns.Count() {\n\t\terr := fmt.Errorf(\"[%v] cached columns count[%d] mismatch binglog row [%d]\", tce.ti.Name, len(tce.ti.Fields), rs.DataColumns.Count())\n\t\treturn sqltypes.Value{}, nil, err\n\t}\n\n\tfor c := 0; c < rs.DataColumns.Count(); c++ {\n\t\tif !rs.DataColumns.Bit(c) {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Print a separator if needed, then print the name.\n\t\tif valueIndex > 0 {\n\t\t\tsql.WriteString(\", \")\n\t\t}\n\t\tsql.Myprintf(\"%v\", sqlparser.NewIdentifierCI(tce.ti.Fields[c].Name))\n\t\tsql.WriteByte('=')\n\n\t\tif rs.Rows[rowIndex].NullColumns.Bit(valueIndex) {\n\t\t\t// This column is represented, but its value is NULL.\n\t\t\tsql.WriteString(\"NULL\")\n\t\t\tvalueIndex++\n\t\t\tcontinue\n\t\t}\n\n\t\t// We have real data.\n\t\tvalue, l, err := binlog.CellValue(data, pos, tce.tm.Types[c], tce.tm.Metadata[c], &querypb.Field{Type: tce.ti.Fields[c].Type})\n\t\tif err != nil {\n\t\t\treturn keyspaceIDCell, nil, err\n\t\t}\n\t\tvBytes, err := value.ToBytes()\n\t\tif err != nil {\n\t\t\treturn sqltypes.Value{}, nil, err\n\t\t}\n\t\tif value.Type() == querypb.Type_TIMESTAMP && !bytes.HasPrefix(vBytes, binlog.ZeroTimestamp) {\n\t\t\t// Values in the binary log are UTC. Let's convert them\n\t\t\t// to whatever timezone the connection is using,\n\t\t\t// so MySQL properly converts them back to UTC.\n\t\t\tsql.WriteString(\"convert_tz(\")\n\t\t\tvalue.EncodeSQL(sql)\n\t\t\tsql.WriteString(\", '+00:00', @@session.time_zone)\")\n\t\t} else {\n\t\t\tvalue.EncodeSQL(sql)\n\t\t}\n\t\tif c == tce.keyspaceIDIndex {\n\t\t\tkeyspaceIDCell = value\n\t\t}\n\t\tif getPK {\n\t\t\tif tce.pkIndexes[c] != -1 {\n\t\t\t\tpkValues[tce.pkIndexes[c]] = value\n\t\t\t}\n\t\t}\n\t\tpos += l\n\t\tvalueIndex++\n\t}\n\n\treturn keyspaceIDCell, pkValues, nil\n}", "func evalAstExpr(sctx sessionctx.Context, expr ast.ExprNode) (types.Datum, error) {\n\tif val, ok := expr.(*driver.ValueExpr); ok {\n\t\treturn val.Datum, nil\n\t}\n\tvar is infoschema.InfoSchema\n\tif sctx.GetSessionVars().TxnCtx.InfoSchema != nil {\n\t\tis = sctx.GetSessionVars().TxnCtx.InfoSchema.(infoschema.InfoSchema)\n\t}\n\tb := NewPlanBuilder(sctx, is)\n\tfakePlan := LogicalTableDual{}.Init(sctx)\n\tnewExpr, _, err := b.rewrite(context.TODO(), expr, fakePlan, nil, true)\n\tif err != nil {\n\t\treturn types.Datum{}, err\n\t}\n\treturn newExpr.Eval(chunk.Row{})\n}", "func Timestamp(layout string) Valuer {\n\treturn func(context.Context) interface{} {\n\t\treturn time.Now().Format(layout)\n\t}\n}", "func (s UpdateStatement) QueryAndValues() (string, []interface{}) {\n\tvalues := make([]interface{}, 0)\n\tquery := []string{\"UPDATE\", fmt.Sprintf(\"%s.%s\", s.Keyspace(), s.Table())}\n\n\t// Determine if we need to set a TTL\n\tif s.TTL() > 0 {\n\t\tquery = append(query, \"USING TTL ?\")\n\t\tvalues = append(values, int(s.TTL().Seconds()))\n\t}\n\n\tsetCQL, setValues := generateUpdateSetCQL(s.FieldMap())\n\tquery = append(query, \"SET\", setCQL)\n\tvalues = append(values, setValues...)\n\n\twhereCQL, whereValues := generateWhereCQL(s.Relations())\n\tif whereCQL != \"\" {\n\t\tquery = append(query, \"WHERE\", whereCQL)\n\t\tvalues = append(values, whereValues...)\n\t}\n\treturn strings.Join(query, \" \"), values\n}", "func prepareValues(field string, values []string) (interface{}, error) {\n\tpreparedValues := make([]interface{}, len(values))\n\tif isTimestampField := timestampFields[field]; isTimestampField {\n\t\tfor idx, value := range values {\n\t\t\ttimestamp, err := time.Parse(time.RFC3339Nano, value)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.NewFlyteAdminErrorf(codes.InvalidArgument,\n\t\t\t\t\t\"Timestamp %s must conform to RFC3339 Nano spec\", value)\n\t\t\t}\n\t\t\tpreparedValues[idx] = timestamp\n\t\t}\n\t} else if isDurationField := durationFields[strings.ToLower(field)]; isDurationField {\n\t\tfor idx, value := range values {\n\t\t\tfloatValue, err := strconv.ParseFloat(value, 64)\n\t\t\tif err == nil {\n\t\t\t\t// The value is an float. By default purely float values are assumed to represent durations in seconds.\n\t\t\t\tvalue = fmt.Sprintf(secondsFormat, floatValue)\n\t\t\t}\n\t\t\tduration, err := time.ParseDuration(value)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.NewFlyteAdminErrorf(codes.InvalidArgument,\n\t\t\t\t\t\"Failed to parse duration [%s]\", value)\n\t\t\t}\n\t\t\tpreparedValues[idx] = duration\n\t\t}\n\t} else {\n\t\tfor idx, value := range values {\n\t\t\tpreparedValues[idx] = value\n\t\t}\n\t}\n\tif len(preparedValues) == 1 {\n\t\treturn preparedValues[0], nil\n\t}\n\treturn preparedValues, nil\n}", "func CreatedAtGTE(v time.Time) predicate.MetaSchema {\n\treturn predicate.MetaSchema(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t},\n\t)\n}", "func NewValue(t int64, v byte) Value {\n\treturn Value{UnixNano: t, Value: v}\n}", "func exprf(format string, a ...interface{}) ast.Expr {\n\tsrc := fmt.Sprintf(format, a...)\n\texpr, err := parser.ParseExpr(src)\n\tif err != nil {\n\t\tlog.Fatalf(\"expr parse error on %q: %v\", src, err)\n\t}\n\treturn expr\n}", "func NewCastExpr(src Expr, width uint, signed bool) Expr {\n\tif signed {\n\t\treturn newSExtExpr(src, width)\n\t}\n\treturn newZExtExpr(src, width)\n}", "func CreatedAtEQ(v time.Time) predicate.MetaSchema {\n\treturn predicate.MetaSchema(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldCreatedAt), v))\n\t},\n\t)\n}", "func (self *TStatement) generate_query(vals map[string]interface{}, includeVersion bool, includeUpdated bool, includeNil bool,\r\n\tincludeAutoIncr bool, allUseBool bool, useAllCols bool, unscoped bool, mustColumnMap map[string]bool) (res_clause string, res_params []interface{}) {\r\n\t//res_domain = utils.NewStringList()\r\n\tlClauses := make([]string, 0)\r\n\tres_params = make([]interface{}, 0)\r\n\r\n\tvar (\r\n\t\t//\t\tfield *TField\r\n\t\tcol IField\r\n\t\t//left, oprator, right string\r\n\r\n\t\tlIsRequiredField bool\r\n\t\tlFieldType reflect.Type\r\n\t\tlFieldVal reflect.Value\r\n\t)\r\n\r\n\tfor name, val := range vals {\r\n\r\n\t\t//field = self.session.model.FieldByName(name)\r\n\t\tcol = self.model.GetFieldByName(name) // field.column\r\n\t\tif col == nil {\r\n\t\t\tcontinue\r\n\t\t}\r\n\r\n\t\tif !includeVersion && col.IsVersion() {\r\n\t\t\tcontinue\r\n\t\t}\r\n\r\n\t\tif !includeUpdated && col.IsUpdated() {\r\n\t\t\tcontinue\r\n\t\t}\r\n\r\n\t\tif !includeAutoIncr && col.IsAutoIncrement() {\r\n\t\t\tcontinue\r\n\t\t}\r\n\r\n\t\tif self.session.orm.dialect.DBType() == MSSQL && col.SQLType().Name == Text {\r\n\t\t\tcontinue\r\n\t\t}\r\n\t\tif col.SQLType().IsJson() {\r\n\t\t\tcontinue\r\n\t\t}\r\n\r\n\t\tif val == nil {\r\n\t\t\tcontinue\r\n\t\t}\r\n\r\n\t\tlFieldType = reflect.TypeOf(val)\r\n\t\tlFieldVal = reflect.ValueOf(val)\r\n\t\tlIsRequiredField = useAllCols\r\n\t\t// 强制过滤已经设定的字段是否作为Query使用\r\n\t\tif b, ok := mustColumnMap[strings.ToLower(col.Name())]; ok {\r\n\t\t\tif b {\r\n\t\t\t\tlIsRequiredField = true\r\n\t\t\t} else {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\t// 处理指针结构\r\n\t\tif lFieldType.Kind() == reflect.Ptr {\r\n\t\t\tif val == nil {\r\n\t\t\t\tif includeNil {\r\n\t\t\t\t\t//args = append(args, nil)\r\n\t\t\t\t\t//colNames = append(colNames, fmt.Sprintf(\"%v %s ?\", colName, engine.dialect.EqStr()))\r\n\t\t\t\t\tlClauses = append(lClauses, fmt.Sprintf(\"%v %s ?\", name, self.session.orm.dialect.EqStr()))\r\n\t\t\t\t\t//res_domain.AddSubList(name, self.session.orm.dialect.EqStr(), \"?\")\r\n\t\t\t\t\tres_params = append(res_params, nil)\r\n\t\t\t\t}\r\n\t\t\t\tcontinue\r\n\r\n\t\t\t} else {\r\n\t\t\t\t// dereference ptr type to instance type\r\n\t\t\t\tlFieldVal = lFieldVal.Elem()\r\n\t\t\t\tlFieldType = reflect.TypeOf(lFieldVal.Interface())\r\n\t\t\t\tlIsRequiredField = true\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\tswitch lFieldType.Kind() {\r\n\t\tcase reflect.Bool:\r\n\t\t\tif !allUseBool || !lIsRequiredField {\r\n\t\t\t\t// if a bool in a struct, it will not be as a condition because it default is false,\r\n\t\t\t\t// please use Where() instead\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\tcase reflect.String:\r\n\t\t\t/*if !requiredField && fieldValue.String() == \"\" {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\t// for MyString, should convert to string or panic\r\n\t\t\tif fieldType.String() != reflect.String.String() {\r\n\t\t\t\tval = fieldValue.String()\r\n\t\t\t} else {\r\n\t\t\t\tval = fieldValue.Interface()\r\n\t\t\t}*/\r\n\t\tcase reflect.Int8, reflect.Int16, reflect.Int, reflect.Int32, reflect.Int64:\r\n\t\t\t/*if !requiredField && fieldValue.Int() == 0 {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\tval = fieldValue.Interface()*/\r\n\t\tcase reflect.Float32, reflect.Float64:\r\n\t\t\t/*if !requiredField && fieldValue.Float() == 0.0 {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\tval = fieldValue.Interface()*/\r\n\t\tcase reflect.Uint8, reflect.Uint16, reflect.Uint, reflect.Uint32, reflect.Uint64:\r\n\t\t\t/*if !requiredField && fieldValue.Uint() == 0 {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\tt := int64(fieldValue.Uint())\r\n\t\t\tval = reflect.ValueOf(&t).Interface()*/\r\n\t\tcase reflect.Struct:\r\n\t\t\tif lFieldType.ConvertibleTo(TimeType) {\r\n\t\t\t\tt := lFieldVal.Convert(TimeType).Interface().(time.Time)\r\n\t\t\t\tif !lIsRequiredField && (t.IsZero() || !lFieldVal.IsValid()) {\r\n\t\t\t\t\tcontinue\r\n\t\t\t\t}\r\n\t\t\t\tval = self.session.orm.FormatTime(col.SQLType().Name, t)\r\n\t\t\t} else if _, ok := reflect.New(lFieldType).Interface().(Conversion); ok {\r\n\t\t\t\tcontinue\r\n\r\n\t\t\t\t/*} else if valNul, ok := fieldValue.Interface().(driver.Valuer); ok {\r\n\t\t\t\tval, _ = valNul.Value()\r\n\t\t\t\tif val == nil {\r\n\t\t\t\t\tcontinue\r\n\t\t\t\t}*/\r\n\t\t\t} else {\r\n\t\t\t\tif col.SQLType().IsJson() {\r\n\t\t\t\t\tif col.SQLType().IsText() {\r\n\t\t\t\t\t\tbytes, err := json.Marshal(val)\r\n\t\t\t\t\t\tif err != nil {\r\n\t\t\t\t\t\t\tlog.Err(\"adas\", err)\r\n\t\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t\t}\r\n\t\t\t\t\t\tval = string(bytes)\r\n\t\t\t\t\t} else if col.SQLType().IsBlob() {\r\n\t\t\t\t\t\tvar bytes []byte\r\n\t\t\t\t\t\tvar err error\r\n\t\t\t\t\t\tbytes, err = json.Marshal(val)\r\n\t\t\t\t\t\tif err != nil {\r\n\t\t\t\t\t\t\tlog.Errf(\"asdf\", err)\r\n\t\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t\t}\r\n\t\t\t\t\t\tval = bytes\r\n\t\t\t\t\t}\r\n\t\t\t\t} else {\r\n\t\t\t\t\t// any other\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\tcase reflect.Array, reflect.Slice, reflect.Map:\r\n\t\t\tif lFieldVal == reflect.Zero(lFieldType) {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\t\tif lFieldVal.IsNil() || !lFieldVal.IsValid() || lFieldVal.Len() == 0 {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\r\n\t\t\tif col.SQLType().IsText() {\r\n\t\t\t\tbytes, err := json.Marshal(lFieldVal.Interface())\r\n\t\t\t\tif err != nil {\r\n\t\t\t\t\tlog.Errf(\"generate_query:\", err)\r\n\t\t\t\t\tcontinue\r\n\t\t\t\t}\r\n\t\t\t\tval = string(bytes)\r\n\t\t\t} else if col.SQLType().IsBlob() {\r\n\t\t\t\tvar bytes []byte\r\n\t\t\t\tvar err error\r\n\t\t\t\tif (lFieldType.Kind() == reflect.Array || lFieldType.Kind() == reflect.Slice) &&\r\n\t\t\t\t\tlFieldType.Elem().Kind() == reflect.Uint8 {\r\n\t\t\t\t\tif lFieldVal.Len() > 0 {\r\n\t\t\t\t\t\tval = lFieldVal.Bytes()\r\n\t\t\t\t\t} else {\r\n\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t}\r\n\t\t\t\t} else {\r\n\t\t\t\t\tbytes, err = json.Marshal(lFieldVal.Interface())\r\n\t\t\t\t\tif err != nil {\r\n\t\t\t\t\t\tlog.Err(\"1\", err)\r\n\t\t\t\t\t\tcontinue\r\n\t\t\t\t\t}\r\n\t\t\t\t\tval = bytes\r\n\t\t\t\t}\r\n\t\t\t} else {\r\n\t\t\t\tcontinue\r\n\t\t\t}\r\n\t\tdefault:\r\n\t\t\t//val = lFieldVal.Interface()\r\n\t\t}\r\n\r\n\t\tvar Clause string\r\n\t\tif col.IsPrimaryKey() && self.session.orm.dialect.DBType() == \"ql\" {\r\n\t\t\t//condi = \"id() == ?\"\r\n\t\t\tClause = \"id() == ?\"\r\n\t\t\t//left = \"id()\"\r\n\t\t\t//oprator = \"=\"\r\n\t\t\t//right = \"?\"\r\n\r\n\t\t} else {\r\n\t\t\t//condi = fmt.Sprintf(\"%v %s ?\", colName, self.session.orm.dialect.EqStr())\r\n\t\t\tClause = fmt.Sprintf(\"%v %s ?\", name, self.session.orm.dialect.EqStr())\r\n\t\t\t//left = name\r\n\t\t\t//oprator = \"=\"\r\n\t\t\t//right = \"?\"\r\n\t\t}\r\n\t\tlClauses = append(lClauses, Clause)\r\n\t\t//res_domain.AddSubList(right, oprator, left)\r\n\t\tres_params = append(res_params, val)\r\n\t}\r\n\r\n\tres_clause = strings.Join(lClauses, \" \"+self.session.orm.dialect.AndStr()+\" \")\r\n\treturn\r\n}", "func evalTypedExpr(expr Expr, t knownType, env Env) (xs []reflect.Value, err error) {\n if expr.IsConst() {\n x := expr.Const()\n if ct, ok := expr.KnownType()[0].(ConstType); ok {\n cx, _ := promoteConstToTyped(ct, constValue(x), t[0], expr)\n xs = []reflect.Value{reflect.Value(cx)}\n } else {\n xs = []reflect.Value{x}\n }\n } else {\n xs, err = EvalExpr(expr, env)\n }\n return xs, err\n}", "func (sink *influxdbSink) composeRawQuery(metricName string, labels map[string]string, metricKeys []core.HistoricalKey, start, end time.Time) string {\n\tseriesName, fieldName := sink.metricToSeriesAndField(metricName)\n\n\tqueries := make([]string, len(metricKeys))\n\tfor i, key := range metricKeys {\n\t\tpred := sink.keyToSelector(key)\n\t\tif labels != nil {\n\t\t\tpred += fmt.Sprintf(\" AND %s\", sink.labelsToPredicate(labels))\n\t\t}\n\t\tif !start.IsZero() {\n\t\t\tpred += fmt.Sprintf(\" AND time > '%s'\", start.Format(time.RFC3339))\n\t\t}\n\t\tif !end.IsZero() {\n\t\t\tpred += fmt.Sprintf(\" AND time < '%s'\", end.Format(time.RFC3339))\n\t\t}\n\t\tqueries[i] = fmt.Sprintf(\"SELECT time, %q FROM %q WHERE %s\", fieldName, seriesName, pred)\n\t}\n\n\treturn strings.Join(queries, \"; \")\n}", "func parseExprs(exprs []string) (ast.Exprs, error) {\n\tstmt, err := ParseOne(fmt.Sprintf(\"SET ROW (%s)\", strings.Join(exprs, \",\")))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tset, ok := stmt.AST.(*ast.SetVar)\n\tif !ok {\n\t\treturn nil, errors.AssertionFailedf(\"expected a SET statement, but found %T\", stmt)\n\t}\n\treturn set.Values, nil\n}", "func Insert(value models.Insert) (string, []interface{}, error) {\n\tif value.TableName == \"\" {\n\t\treturn \"\", nil, errNoTableName\n\t}\n\texpr := \"\"\n\texpr += fmt.Sprintf(\"INSERT INTO %s (\", value.TableName)\n\tnum := len(value.Pairs)\n\tkeys := make([]string, num)\n\tnums := make([]string, num)\n\tdata := make([]interface{}, num)\n\tfor i, v := range value.Pairs {\n\t\tif v.Key == \"\" {\n\t\t\treturn \"\", nil, errNoKey\n\t\t}\n\t\tif v.Value == nil {\n\t\t\treturn \"\", nil, errNoValue\n\t\t}\n\t\tkeys[i] = v.Key\n\t\tdata[i] = v.Value\n\t\tnums[i] = fmt.Sprintf(\"$%d\", i+1)\n\t}\n\tkeysStr := strings.Join(keys, \",\")\n\texpr += keysStr + \") \"\n\tif len(nums) == 0 {\n\t\treturn \"\", nil, errNoValues\n\t}\n\texpr += \"VALUES (\" + strings.Join(nums, \",\") + \")\"\n\treturn expr, data, nil\n}", "func (i *Inserter) Values(v builder.Value) builder.Inserter {\n\tfields := make([]string, 0, len(v))\n\tvalues := make([]string, 0, len(v))\n\n\tfor item := range v {\n\t\tfields = append(fields, item)\n\t}\n\tsort.Strings(fields)\n\tfor _, field := range fields {\n\t\tswitch v[field].(type) {\n\t\tcase string:\n\t\t\tvalues = append(values, fmt.Sprintf(\"'%s'\", v[field].(string)))\n\t\tdefault:\n\t\t\tvalues = append(values, fmt.Sprintf(\"%v\", v[field]))\n\t\t}\n\t}\n\n\tjoinedValues := strings.Join(values, seperator)\n\ti.sql.WriteString(strings.Join([]string{\"(\", joinedValues, \"),\"}, \"\"))\n\treturn i\n}", "func EncodeTableValue(\n\tappendTo []byte, colID descpb.ColumnID, val tree.Datum, scratch []byte,\n) ([]byte, error) {\n\tif val == tree.DNull {\n\t\treturn encoding.EncodeNullValue(appendTo, uint32(colID)), nil\n\t}\n\tswitch t := tree.UnwrapDatum(nil, val).(type) {\n\tcase *tree.DBitArray:\n\t\treturn encoding.EncodeBitArrayValue(appendTo, uint32(colID), t.BitArray), nil\n\tcase *tree.DBool:\n\t\treturn encoding.EncodeBoolValue(appendTo, uint32(colID), bool(*t)), nil\n\tcase *tree.DInt:\n\t\treturn encoding.EncodeIntValue(appendTo, uint32(colID), int64(*t)), nil\n\tcase *tree.DFloat:\n\t\treturn encoding.EncodeFloatValue(appendTo, uint32(colID), float64(*t)), nil\n\tcase *tree.DDecimal:\n\t\treturn encoding.EncodeDecimalValue(appendTo, uint32(colID), &t.Decimal), nil\n\tcase *tree.DString:\n\t\treturn encoding.EncodeBytesValue(appendTo, uint32(colID), []byte(*t)), nil\n\tcase *tree.DBytes:\n\t\treturn encoding.EncodeBytesValue(appendTo, uint32(colID), []byte(*t)), nil\n\tcase *tree.DDate:\n\t\treturn encoding.EncodeIntValue(appendTo, uint32(colID), t.UnixEpochDaysWithOrig()), nil\n\tcase *tree.DBox2D:\n\t\treturn encoding.EncodeBox2DValue(appendTo, uint32(colID), t.CartesianBoundingBox)\n\tcase *tree.DGeography:\n\t\treturn encoding.EncodeGeoValue(appendTo, uint32(colID), t.SpatialObjectRef())\n\tcase *tree.DGeometry:\n\t\treturn encoding.EncodeGeoValue(appendTo, uint32(colID), t.SpatialObjectRef())\n\tcase *tree.DTime:\n\t\treturn encoding.EncodeIntValue(appendTo, uint32(colID), int64(*t)), nil\n\tcase *tree.DTimeTZ:\n\t\treturn encoding.EncodeTimeTZValue(appendTo, uint32(colID), t.TimeTZ), nil\n\tcase *tree.DTimestamp:\n\t\treturn encoding.EncodeTimeValue(appendTo, uint32(colID), t.Time), nil\n\tcase *tree.DTimestampTZ:\n\t\treturn encoding.EncodeTimeValue(appendTo, uint32(colID), t.Time), nil\n\tcase *tree.DInterval:\n\t\treturn encoding.EncodeDurationValue(appendTo, uint32(colID), t.Duration), nil\n\tcase *tree.DUuid:\n\t\treturn encoding.EncodeUUIDValue(appendTo, uint32(colID), t.UUID), nil\n\tcase *tree.DIPAddr:\n\t\treturn encoding.EncodeIPAddrValue(appendTo, uint32(colID), t.IPAddr), nil\n\tcase *tree.DJSON:\n\t\tencoded, err := json.EncodeJSON(scratch, t.JSON)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn encoding.EncodeJSONValue(appendTo, uint32(colID), encoded), nil\n\tcase *tree.DArray:\n\t\ta, err := encodeArray(t, scratch)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn encoding.EncodeArrayValue(appendTo, uint32(colID), a), nil\n\tcase *tree.DTuple:\n\t\treturn encodeTuple(t, appendTo, uint32(colID), scratch)\n\tcase *tree.DCollatedString:\n\t\treturn encoding.EncodeBytesValue(appendTo, uint32(colID), []byte(t.Contents)), nil\n\tcase *tree.DOid:\n\t\treturn encoding.EncodeIntValue(appendTo, uint32(colID), int64(t.DInt)), nil\n\tcase *tree.DEnum:\n\t\treturn encoding.EncodeBytesValue(appendTo, uint32(colID), t.PhysicalRep), nil\n\tdefault:\n\t\treturn nil, errors.Errorf(\"unable to encode table value: %T\", t)\n\t}\n}", "func CreateTimeEQ(v time.Time) predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldCreateTime), v))\n\t},\n\t)\n}", "func (q *Query) Expr(field_name, expr string, args ...interface{}) *Query {\n\treturn q.addCondition(field_name, query.OpExpr, expr, args...)\n}", "func (t *WindowedTable) FormatAbsoluteExpr() string {\n\tfrom := t.Window.From.UnixNano() / int64(time.Millisecond)\n\tto := t.Window.To.UnixNano() / int64(time.Millisecond)\n\tproject := t.ProjectID\n\tif project != \"\" {\n\t\tproject += \":\"\n\t}\n\treturn fmt.Sprintf(\"[%v%v@%v-%v]\", project, t.ID, from, to)\n}", "func CreatedAtGTE(v time.Time) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func CreatedAtGTE(v time.Time) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func DateTimeGTE(v time.Time) predicate.Bulk {\n\treturn predicate.Bulk(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldDateTime), v))\n\t})\n}", "func (s *Serializer) BindAllAsExpr(vals ...interface{}) []AsExpr {\n\tl := make([]AsExpr, len(vals))\n\n\tfor i, val := range vals {\n\t\tl[i] = s.Bind(val)\n\t}\n\n\treturn l\n}", "func NewArrayExpr(elements ...Expr) Expr {\n\tvalues := make([]Value, len(elements))\n\tfor i, expr := range elements {\n\t\tvalue, ok := expr.(Value)\n\t\tif !ok {\n\t\t\ttuples := make([]Expr, len(elements))\n\t\t\tfor i, elt := range elements {\n\t\t\t\tposAttr, err := NewAttrExpr(\"@\", NewNumber(float64(i)))\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t\tvalAttr, err := NewAttrExpr(ArrayItemAttr, elt)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t\ttuples[i] = NewTupleExpr(posAttr, valAttr)\n\t\t\t}\n\t\t\treturn NewSetExpr(tuples...)\n\t\t}\n\t\tvalues[i] = value\n\t}\n\treturn NewArray(values...)\n}", "func (b *BatchInsert) Insert(ctx context.Context, values ...interface{}) error {\n\t//buf := make([]byte, 0, cap(b.query)-b.qStartLen-len(b.qEnd))\n\tb.buf = b.buf[:0]\n\n\tif len(b.query) != b.qStartLen {\n\t\tb.buf = append(b.buf, ',', '(')\n\t} else {\n\t\tb.buf = append(b.buf, '(')\n\t}\n\n\tfor _, arg := range values {\n\t\tmyArg := arg\n\ttryAgain:\n\t\tswitch v := myArg.(type) {\n\t\tcase nil:\n\t\t\tb.buf = append(b.buf, \"NULL\"...)\n\t\tcase SQLField:\n\t\t\tb.buf = append(b.buf, v...)\n\t\tcase int:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase int8:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase int16:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase int32:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase int64:\n\t\t\tb.buf = strconv.AppendInt(b.buf, v, 10)\n\n\t\tcase uint:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase uint8:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase uint16:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase uint32:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10)\n\t\tcase uint64:\n\t\t\tb.buf = strconv.AppendInt(b.buf, int64(v), 10) // Dangerous?\n\n\t\tcase float32:\n\t\t\tb.buf = strconv.AppendFloat(b.buf, float64(v), 'g', -1, 64)\n\t\tcase float64:\n\t\t\tb.buf = strconv.AppendFloat(b.buf, v, 'g', -1, 64)\n\t\tcase bool:\n\t\t\tif b.dbMode == MySQL {\n\t\t\t\tif v {\n\t\t\t\t\tb.buf = append(b.buf, '1')\n\t\t\t\t} else {\n\t\t\t\t\tb.buf = append(b.buf, '0')\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tif v {\n\t\t\t\t\tb.buf = append(b.buf, []byte{'t', 'r', 'u', 'e'}...)\n\t\t\t\t} else {\n\t\t\t\t\tb.buf = append(b.buf, []byte{'f', 'a', 'l', 's', 'e'}...)\n\t\t\t\t}\n\t\t\t}\n\n\t\tcase time.Time:\n\t\t\tb.buf = append(b.buf, []byte(v.Format(\"'2006-01-02 15:04:05.000000'\"))...)\n\n\t\tcase []byte:\n\t\t\tif v == nil {\n\t\t\t\tb.buf = append(b.buf, \"NULL\"...)\n\t\t\t} else {\n\t\t\t\tif b.dbMode == MySQL {\n\t\t\t\t\t// https://dev.mysql.com/doc/refman/5.7/en/mysql-real-escape-string.html\n\t\t\t\t\t// Strictly speaking, MySQL requires only that backslash and the quote character used to quote the string in the query be escaped\n\t\t\t\t\tb.buf = append(b.buf, '\\'')\n\t\t\t\t\tb.buf = append(b.buf, bytes.Replace(bytes.Replace(v, []byte{'\\''}, []byte{'\\'', '\\''}, -1), []byte{'\\''}, []byte{'\\\\', '\\''}, -1)...)\n\t\t\t\t\tb.buf = append(b.buf, '\\'')\n\t\t\t\t} else {\n\t\t\t\t\tdata := make([]byte, hex.EncodedLen(len(v))+6)\n\t\t\t\t\tcopy(data, []byte{'E', '\\'', '\\\\', '\\\\', 'x'})\n\t\t\t\t\thex.Encode(data[5:], v)\n\t\t\t\t\tdata[len(data)-1] = '\\''\n\n\t\t\t\t\tb.buf = append(b.buf, data...)\n\t\t\t\t}\n\t\t\t}\n\t\tcase string:\n\t\t\tb.buf = append(b.buf, '\\'')\n\t\t\tif b.dbMode == MySQL {\n\t\t\t\tb.buf = append(b.buf, bytes.Replace(bytes.Replace([]byte(v), []byte{'\\''}, []byte{'\\'', '\\''}, -1), []byte{'\\''}, []byte{'\\\\', '\\''}, -1)...)\n\t\t\t} else {\n\t\t\t\tb.buf = append(b.buf, []byte(strings.Replace(v, \"'\", \"''\", -1))...)\n\t\t\t}\n\t\t\tb.buf = append(b.buf, '\\'')\n\t\tcase sql.NullBool:\n\t\t\tif v.Valid {\n\t\t\t\tmyArg = v.Bool\n\t\t\t\tgoto tryAgain\n\t\t\t}\n\t\t\tb.buf = append(b.buf, \"NULL\"...)\n\n\t\tcase sql.NullInt64:\n\t\t\tif v.Valid {\n\t\t\t\tmyArg = v.Int64\n\t\t\t\tgoto tryAgain\n\t\t\t}\n\t\t\tb.buf = append(b.buf, \"NULL\"...)\n\n\t\tcase sql.NullString:\n\t\t\tif v.Valid {\n\t\t\t\tmyArg = v.String\n\t\t\t\tgoto tryAgain\n\t\t\t}\n\t\t\tb.buf = append(b.buf, \"NULL\"...)\n\n\t\tcase sql.NullFloat64:\n\t\t\tif v.Valid {\n\t\t\t\tmyArg = v.Float64\n\t\t\t\tgoto tryAgain\n\t\t\t}\n\t\t\tb.buf = append(b.buf, \"NULL\"...)\n\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"%v (%T) not supported by sqlbatchinsert\", v, v)\n\t\t}\n\t\tb.buf = append(b.buf, ',')\n\t}\n\n\tb.buf[len(b.buf)-1] = ')'\n\n\tif len(b.query)+len(b.buf)+len(b.qEnd) > b.qLen {\n\t\tif len(b.buf)+b.qStartLen+len(b.qEnd) > b.qLen {\n\t\t\treturn ErrInsertLargerThanMaxLength\n\t\t}\n\t\tif err := b.flush(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tb.query = append(b.query, b.buf[1:]...) // strip off lead comma\n\t\treturn nil\n\t}\n\n\tb.query = append(b.query, b.buf...)\n\t//log.Infof(ctx, \"Add: %s\", buf)\n\treturn nil\n}", "func (ng *Engine) execEvalStmt(ctx context.Context, query *query, s *EvalStmt) (model.Value, error) {\n\tquerier, err := ng.queryable.Querier()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer querier.Close()\n\n\tprepareTimer := query.stats.GetTimer(stats.QueryPreparationTime).Start()\n\terr = ng.populateIterators(ctx, querier, s)\n\tprepareTimer.Stop()\n\tqueryPrepareTime.Observe(prepareTimer.ElapsedTime().Seconds())\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer ng.closeIterators(s)\n\n\tevalTimer := query.stats.GetTimer(stats.InnerEvalTime).Start()\n\t// Instant evaluation.\n\tif s.Start == s.End && s.Interval == 0 {\n\t\tevaluator := &evaluator{\n\t\t\tTimestamp: s.Start,\n\t\t\tctx: ctx,\n\t\t}\n\t\tval, err := evaluator.Eval(s.Expr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// Turn matrix and vector types with protected metrics into\n\t\t// model.* types.\n\t\tswitch v := val.(type) {\n\t\tcase vector:\n\t\t\tval = v.value()\n\t\tcase matrix:\n\t\t\tval = v.value()\n\t\t}\n\n\t\tevalTimer.Stop()\n\t\tqueryInnerEval.Observe(evalTimer.ElapsedTime().Seconds())\n\n\t\treturn val, nil\n\t}\n\tnumSteps := int(s.End.Sub(s.Start) / s.Interval)\n\n\t// Range evaluation.\n\tsampleStreams := map[model.Fingerprint]*sampleStream{}\n\tfor ts := s.Start; !ts.After(s.End); ts = ts.Add(s.Interval) {\n\n\t\tif err := contextDone(ctx, \"range evaluation\"); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tevaluator := &evaluator{\n\t\t\tTimestamp: ts,\n\t\t\tctx: ctx,\n\t\t}\n\t\tval, err := evaluator.Eval(s.Expr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tswitch v := val.(type) {\n\t\tcase *model.Scalar:\n\t\t\t// As the expression type does not change we can safely default to 0\n\t\t\t// as the fingerprint for scalar expressions.\n\t\t\tss := sampleStreams[0]\n\t\t\tif ss == nil {\n\t\t\t\tss = &sampleStream{Values: make([]model.SamplePair, 0, numSteps)}\n\t\t\t\tsampleStreams[0] = ss\n\t\t\t}\n\t\t\tss.Values = append(ss.Values, model.SamplePair{\n\t\t\t\tValue: v.Value,\n\t\t\t\tTimestamp: v.Timestamp,\n\t\t\t})\n\t\tcase vector:\n\t\t\tfor _, sample := range v {\n\t\t\t\tfp := sample.Metric.Metric.Fingerprint()\n\t\t\t\tss := sampleStreams[fp]\n\t\t\t\tif ss == nil {\n\t\t\t\t\tss = &sampleStream{\n\t\t\t\t\t\tMetric: sample.Metric,\n\t\t\t\t\t\tValues: make([]model.SamplePair, 0, numSteps),\n\t\t\t\t\t}\n\t\t\t\t\tsampleStreams[fp] = ss\n\t\t\t\t}\n\t\t\t\tss.Values = append(ss.Values, model.SamplePair{\n\t\t\t\t\tValue: sample.Value,\n\t\t\t\t\tTimestamp: sample.Timestamp,\n\t\t\t\t})\n\t\t\t}\n\t\tdefault:\n\t\t\tpanic(fmt.Errorf(\"promql.Engine.exec: invalid expression type %q\", val.Type()))\n\t\t}\n\t}\n\tevalTimer.Stop()\n\tqueryInnerEval.Observe(evalTimer.ElapsedTime().Seconds())\n\n\tif err := contextDone(ctx, \"expression evaluation\"); err != nil {\n\t\treturn nil, err\n\t}\n\n\tappendTimer := query.stats.GetTimer(stats.ResultAppendTime).Start()\n\tmat := matrix{}\n\tfor _, ss := range sampleStreams {\n\t\tmat = append(mat, ss)\n\t}\n\tappendTimer.Stop()\n\tqueryResultAppend.Observe(appendTimer.ElapsedTime().Seconds())\n\n\tif err := contextDone(ctx, \"expression evaluation\"); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Turn matrix type with protected metric into model.Matrix.\n\tresMatrix := mat.value()\n\n\tsortTimer := query.stats.GetTimer(stats.ResultSortTime).Start()\n\tsort.Sort(resMatrix)\n\tsortTimer.Stop()\n\tqueryResultSort.Observe(sortTimer.ElapsedTime().Seconds())\n\treturn resMatrix, nil\n}", "func (t Time) Value() (driver.Value, error) {\n\tif t.IsZero() {\n\t\treturn nil, nil\n\t}\n\n\treturn t.In(time.UTC).Format(SQLDateTime), nil\n}", "func (ev *evaluator) eval(expr Expr) model.Value {\n\t// This is the top-level evaluation method.\n\t// Thus, we check for timeout/cancellation here.\n\tif err := contextDone(ev.ctx, \"expression evaluation\"); err != nil {\n\t\tev.error(err)\n\t}\n\n\tswitch e := expr.(type) {\n\tcase *AggregateExpr:\n\t\tvector := ev.evalVector(e.Expr)\n\t\treturn ev.aggregation(e.Op, e.Grouping, e.Without, e.KeepCommonLabels, e.Param, vector)\n\n\tcase *BinaryExpr:\n\t\tlhs := ev.evalOneOf(e.LHS, model.ValScalar, model.ValVector)\n\t\trhs := ev.evalOneOf(e.RHS, model.ValScalar, model.ValVector)\n\n\t\tswitch lt, rt := lhs.Type(), rhs.Type(); {\n\t\tcase lt == model.ValScalar && rt == model.ValScalar:\n\t\t\treturn &model.Scalar{\n\t\t\t\tValue: scalarBinop(e.Op, lhs.(*model.Scalar).Value, rhs.(*model.Scalar).Value),\n\t\t\t\tTimestamp: ev.Timestamp,\n\t\t\t}\n\n\t\tcase lt == model.ValVector && rt == model.ValVector:\n\t\t\tswitch e.Op {\n\t\t\tcase itemLAND:\n\t\t\t\treturn ev.vectorAnd(lhs.(vector), rhs.(vector), e.VectorMatching)\n\t\t\tcase itemLOR:\n\t\t\t\treturn ev.vectorOr(lhs.(vector), rhs.(vector), e.VectorMatching)\n\t\t\tcase itemLUnless:\n\t\t\t\treturn ev.vectorUnless(lhs.(vector), rhs.(vector), e.VectorMatching)\n\t\t\tdefault:\n\t\t\t\treturn ev.vectorBinop(e.Op, lhs.(vector), rhs.(vector), e.VectorMatching, e.ReturnBool)\n\t\t\t}\n\t\tcase lt == model.ValVector && rt == model.ValScalar:\n\t\t\treturn ev.vectorScalarBinop(e.Op, lhs.(vector), rhs.(*model.Scalar), false, e.ReturnBool)\n\n\t\tcase lt == model.ValScalar && rt == model.ValVector:\n\t\t\treturn ev.vectorScalarBinop(e.Op, rhs.(vector), lhs.(*model.Scalar), true, e.ReturnBool)\n\t\t}\n\n\tcase *Call:\n\t\treturn e.Func.Call(ev, e.Args)\n\n\tcase *MatrixSelector:\n\t\treturn ev.matrixSelector(e)\n\n\tcase *NumberLiteral:\n\t\treturn &model.Scalar{Value: e.Val, Timestamp: ev.Timestamp}\n\n\tcase *ParenExpr:\n\t\treturn ev.eval(e.Expr)\n\n\tcase *StringLiteral:\n\t\treturn &model.String{Value: e.Val, Timestamp: ev.Timestamp}\n\n\tcase *UnaryExpr:\n\t\tse := ev.evalOneOf(e.Expr, model.ValScalar, model.ValVector)\n\t\t// Only + and - are possible operators.\n\t\tif e.Op == itemSUB {\n\t\t\tswitch v := se.(type) {\n\t\t\tcase *model.Scalar:\n\t\t\t\tv.Value = -v.Value\n\t\t\tcase vector:\n\t\t\t\tfor i, sv := range v {\n\t\t\t\t\tv[i].Value = -sv.Value\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn se\n\n\tcase *VectorSelector:\n\t\treturn ev.vectorSelector(e)\n\t}\n\tpanic(fmt.Errorf(\"unhandled expression of type: %T\", expr))\n}", "func SerializeExpr(expr Expression, mem memory.Allocator) (*memory.Buffer, error) {\n\tvar (\n\t\tcols []arrow.Array\n\t\tmetaKey []string\n\t\tmetaValue []string\n\t\tvisit func(Expression) error\n\t)\n\n\taddScalar := func(s scalar.Scalar) (string, error) {\n\t\tret := len(cols)\n\t\tarr, err := scalar.MakeArrayFromScalar(s, 1, mem)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tcols = append(cols, arr)\n\t\treturn strconv.Itoa(ret), nil\n\t}\n\n\tvisit = func(e Expression) error {\n\t\tswitch e := e.(type) {\n\t\tcase *Literal:\n\t\t\tif !e.IsScalarExpr() {\n\t\t\t\treturn errors.New(\"not implemented: serialization of non-scalar literals\")\n\t\t\t}\n\t\t\tmetaKey = append(metaKey, \"literal\")\n\t\t\ts, err := addScalar(e.Literal.(*ScalarDatum).Value)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tmetaValue = append(metaValue, s)\n\t\tcase *Parameter:\n\t\t\tif e.ref.Name() == \"\" {\n\t\t\t\treturn errors.New(\"not implemented: serialization of non-name field_ref\")\n\t\t\t}\n\n\t\t\tmetaKey = append(metaKey, \"field_ref\")\n\t\t\tmetaValue = append(metaValue, e.ref.Name())\n\t\tcase *Call:\n\t\t\tmetaKey = append(metaKey, \"call\")\n\t\t\tmetaValue = append(metaValue, e.funcName)\n\n\t\t\tfor _, arg := range e.args {\n\t\t\t\tvisit(arg)\n\t\t\t}\n\n\t\t\tif e.options != nil {\n\t\t\t\tst, err := scalar.ToScalar(e.options, mem)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmetaKey = append(metaKey, \"options\")\n\t\t\t\ts, err := addScalar(st)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tmetaValue = append(metaValue, s)\n\n\t\t\t\tfor _, f := range st.(*scalar.Struct).Value {\n\t\t\t\t\tswitch s := f.(type) {\n\t\t\t\t\tcase releasable:\n\t\t\t\t\t\tdefer s.Release()\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tmetaKey = append(metaKey, \"end\")\n\t\t\tmetaValue = append(metaValue, e.funcName)\n\t\t}\n\t\treturn nil\n\t}\n\n\tif err := visit(expr); err != nil {\n\t\treturn nil, err\n\t}\n\n\tfields := make([]arrow.Field, len(cols))\n\tfor i, c := range cols {\n\t\tfields[i].Type = c.DataType()\n\t\tdefer c.Release()\n\t}\n\n\tmetadata := arrow.NewMetadata(metaKey, metaValue)\n\trec := array.NewRecord(arrow.NewSchema(fields, &metadata), cols, 1)\n\tdefer rec.Release()\n\n\tbuf := &bufferWriteSeeker{mem: mem}\n\twr, err := ipc.NewFileWriter(buf, ipc.WithSchema(rec.Schema()), ipc.WithAllocator(mem))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\twr.Write(rec)\n\twr.Close()\n\treturn buf.buf, nil\n}", "func (s *SelectParser) parseWhereExpr(expr sqlparser.Expr) SqlAny {\n\tswitch expr := expr.(type) {\n\tcase *sqlparser.AndExpr:\n\t\treturn &AndOp{\n\t\t\tLeft: s.parseWhereExpr(expr.Left),\n\t\t\tRight: s.parseWhereExpr(expr.Right),\n\t\t}\n\tcase *sqlparser.OrExpr:\n\t\treturn &OrOp{\n\t\t\tLeft: s.parseWhereExpr(expr.Left),\n\t\t\tRight: s.parseWhereExpr(expr.Right),\n\t\t}\n\tcase *sqlparser.NotExpr:\n\t\treturn &NotOp{\n\t\t\tExpr: s.parseWhereExpr(expr.Expr),\n\t\t}\n\tcase *sqlparser.ParenExpr:\n\t\treturn s.parseWhereExpr(expr.Expr)\n\tcase *sqlparser.ComparisonExpr:\n\t\t//if !supportedCompOps.Contains(expr.Operator) {\n\t\t//\tpanic(fmt.Errorf(\"unsupported operator: %s\", expr.Operator))\n\t\t//}\n\t\treturn &CompOp{\n\t\t\tOperator: expr.Operator,\n\t\t\tLeft: s.parseWhereExpr(expr.Left),\n\t\t\tRight: s.parseWhereExpr(expr.Right),\n\t\t}\n\tcase *sqlparser.RangeCond:\n\t\trangeOp := &RangeOp{\n\t\t\tLeft: s.parseWhereExpr(expr.Left),\n\t\t\tFrom: s.parseWhereExpr(expr.From),\n\t\t\tTo: s.parseWhereExpr(expr.To),\n\t\t}\n\t\tswitch expr.Operator {\n\t\tcase sqlparser.BetweenStr:\n\t\t\treturn rangeOp\n\t\tcase sqlparser.NotBetweenStr:\n\t\t\treturn &NotOp{\n\t\t\t\tExpr: rangeOp,\n\t\t\t}\n\t\tdefault:\n\t\t\tpanic(fmt.Errorf(\"bug: RangeCond.Operator=%s\", expr.Operator))\n\t\t}\n\tcase *sqlparser.SQLVal:\n\t\tswitch expr.Type {\n\t\tcase sqlparser.StrVal:\n\t\t\treturn SqlString(string(expr.Val))\n\t\tcase sqlparser.IntVal:\n\t\t\tval, err := strconv.ParseInt(string(expr.Val), 10, 64)\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\treturn SqlBigInt(val)\n\t\tdefault:\n\t\t\t// TODO\n\t\t\tpanic(\"todo\")\n\t\t}\n\tcase *sqlparser.ColName:\n\t\ttable := s.table.Name\n\t\tif expr.Qualifier.Name.String() != \"\" {\n\t\t\ttable = expr.Qualifier.Name.String()\n\t\t}\n\t\treturn &SqlField{\n\t\t\tField: Field{\n\t\t\t\tTable: table,\n\t\t\t\tName: expr.Name.String(),\n\t\t\t},\n\t\t}\n\tcase *sqlparser.IntervalExpr:\n\t\t// TODO\n\t\tpanic(\"todo\")\n\n\tcase *sqlparser.FuncExpr:\n\t\tif !expr.Qualifier.IsEmpty() {\n\t\t\tpanic(ErrFunctionQualifier)\n\t\t}\n\t\tif expr.Distinct {\n\t\t\tpanic(ErrDistinct)\n\t\t}\n\n\t\tvar found bool\n\t\tvar sqlfunc SqlFunc\n\t\tfor i := range funcs {\n\t\t\tif expr.Name.EqualString(funcs[i].Name) {\n\t\t\t\tfound = true\n\t\t\t\tsqlfunc = funcs[i]\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tpanic(fmt.Errorf(\"not found %s function\", expr.Name.String()))\n\t\t}\n\n\t\t// 関数の引数をパースするは、補完されるテーブル名を SqlFunc で定義されたテーブル名に変更する。\n\t\tparser := s\n\t\tif sqlfunc.Table != \"\" {\n\t\t\ttable, ok := findTableByName(sqlfunc.Table)\n\t\t\tif !ok {\n\t\t\t\tpanic(fmt.Errorf(\"not found %s table\", sqlfunc.Table))\n\t\t\t}\n\t\t\tparser = &SelectParser{\n\t\t\t\ttable: table,\n\t\t\t}\n\t\t}\n\n\t\tvar fnargs []SqlAny\n\t\tfor _, arg := range expr.Exprs {\n\t\t\tfnargs = append(fnargs, parser.parseSelectExpr(arg))\n\t\t}\n\t\treturn sqlfunc.Parse(fnargs...)\n\tdefault:\n\t\tpanic(\"bug\")\n\t}\n\treturn nil\n}", "func TimeStamp(children ...Element) *CompoundElement { return newCE(\"TimeStamp\", children) }", "func CreateTimeGTE(v time.Time) predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreateTime), v))\n\t},\n\t)\n}", "func (d *dbBaseClickHouse) ReadValues(q dbQuerier, qs *querySet, mi *modelInfo, cond *Condition, exprs []string, container interface{}, tz *time.Location) (int64, error) {\n\n\tvar (\n\t\tmaps []Params\n\t\tlists []ParamsList\n\t\tlist ParamsList\n\t)\n\n\ttyp := 0\n\tswitch v := container.(type) {\n\tcase *[]Params:\n\t\td := *v\n\t\tif len(d) == 0 {\n\t\t\tmaps = d\n\t\t}\n\t\ttyp = 1\n\tcase *[]ParamsList:\n\t\td := *v\n\t\tif len(d) == 0 {\n\t\t\tlists = d\n\t\t}\n\t\ttyp = 2\n\tcase *ParamsList:\n\t\td := *v\n\t\tif len(d) == 0 {\n\t\t\tlist = d\n\t\t}\n\t\ttyp = 3\n\tdefault:\n\t\tpanic(fmt.Errorf(\"unsupport read values type `%T`\", container))\n\t}\n\n\ttables := newDbTables(mi, d.ins)\n\n\tvar (\n\t\tcols []string\n\t\tinfos []*fieldInfo\n\t)\n\n\thasExprs := len(exprs) > 0\n\n\tQ := d.ins.TableQuote()\n\n\tif hasExprs {\n\t\tcols = make([]string, 0, len(exprs))\n\t\tinfos = make([]*fieldInfo, 0, len(exprs))\n\t\tfor _, ex := range exprs {\n\t\t\tindex, name, fi, suc := tables.parseExprs(mi, strings.Split(ex, ExprSep))\n\t\t\tif !suc {\n\t\t\t\tpanic(fmt.Errorf(\"unknown field/column name `%s`\", ex))\n\t\t\t}\n\t\t\tcols = append(cols, fmt.Sprintf(\"%s.%s%s%s %s%s%s\", index, Q, fi.column, Q, Q, name, Q))\n\t\t\tinfos = append(infos, fi)\n\t\t}\n\t} else {\n\t\tcols = make([]string, 0, len(mi.fields.dbcols))\n\t\tinfos = make([]*fieldInfo, 0, len(exprs))\n\t\tfor _, fi := range mi.fields.fieldsDB {\n\t\t\tcols = append(cols, fmt.Sprintf(\"T0.%s%s%s %s%s%s\", Q, fi.column, Q, Q, fi.name, Q))\n\t\t\tinfos = append(infos, fi)\n\t\t}\n\t}\n\n\twhere, args := tables.getCondSQL(cond, false, tz)\n\tgroupBy := tables.getGroupSQL(qs.groups)\n\torderBy := tables.getOrderSQL(qs.orders)\n\tlimit := tables.getLimitSQL(mi, qs.offset, qs.limit)\n\tjoin := tables.getJoinSQL()\n\n\tsels := strings.Join(cols, \", \")\n\n\tsqlSelect := \"SELECT\"\n\tif qs.distinct {\n\t\tsqlSelect += \" DISTINCT\"\n\t}\n\tquery := fmt.Sprintf(\"%s %s FROM %s%s%s T0 %s%s%s%s%s\", sqlSelect, sels, Q, mi.table, Q, join, where, groupBy, orderBy, limit)\n\n\td.ins.ReplaceMarks(&query)\n\n\trs, err := q.Query(query, args...)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\trefs := make([]interface{}, len(cols))\n\tfor i := range refs {\n\t\tvar ref interface{}\n\t\trefs[i] = &ref\n\t}\n\n\tdefer rs.Close()\n\n\tvar (\n\t\tcnt int64\n\t\tcolumns []string\n\t)\n\tfor rs.Next() {\n\t\tif cnt == 0 {\n\t\t\tcols, err := rs.Columns()\n\t\t\tif err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t\tcolumns = cols\n\t\t}\n\n\t\tif err := rs.Scan(refs...); err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\tswitch typ {\n\t\tcase 1:\n\t\t\tparams := make(Params, len(cols))\n\t\t\tfor i, ref := range refs {\n\t\t\t\tfi := infos[i]\n\n\t\t\t\tval := reflect.Indirect(reflect.ValueOf(ref)).Interface()\n\n\t\t\t\tvalue, err := d.convertValueFromDB(fi, val, tz)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(fmt.Errorf(\"db value convert failed `%v` %s\", val, err.Error()))\n\t\t\t\t}\n\n\t\t\t\tparams[columns[i]] = value\n\t\t\t}\n\t\t\tmaps = append(maps, params)\n\t\tcase 2:\n\t\t\tparams := make(ParamsList, 0, len(cols))\n\t\t\tfor i, ref := range refs {\n\t\t\t\tfi := infos[i]\n\n\t\t\t\tval := reflect.Indirect(reflect.ValueOf(ref)).Interface()\n\n\t\t\t\tvalue, err := d.convertValueFromDB(fi, val, tz)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(fmt.Errorf(\"db value convert failed `%v` %s\", val, err.Error()))\n\t\t\t\t}\n\n\t\t\t\tparams = append(params, value)\n\t\t\t}\n\t\t\tlists = append(lists, params)\n\t\tcase 3:\n\t\t\tfor i, ref := range refs {\n\t\t\t\tfi := infos[i]\n\n\t\t\t\tval := reflect.Indirect(reflect.ValueOf(ref)).Interface()\n\n\t\t\t\tvalue, err := d.convertValueFromDB(fi, val, tz)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(fmt.Errorf(\"db value convert failed `%v` %s\", val, err.Error()))\n\t\t\t\t}\n\n\t\t\t\tlist = append(list, value)\n\t\t\t}\n\t\t}\n\n\t\tcnt++\n\t}\n\n\tswitch v := container.(type) {\n\tcase *[]Params:\n\t\t*v = maps\n\tcase *[]ParamsList:\n\t\t*v = lists\n\tcase *ParamsList:\n\t\t*v = list\n\t}\n\n\treturn cnt, nil\n}", "func CreatedAtGTE(v time.Time) predicate.Step {\n\treturn predicate.Step(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func CreatedAtGTE(v time.Time) predicate.Job {\n\treturn predicate.Job(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func funcTimestamp(vals []parser.Value, args parser.Expressions, enh *EvalNodeHelper) Vector {\n\tvec := vals[0].(Vector)\n\tfor _, el := range vec {\n\t\tenh.Out = append(enh.Out, Sample{\n\t\t\tMetric: enh.DropMetricName(el.Metric),\n\t\t\tF: float64(el.T) / 1000,\n\t\t})\n\t}\n\treturn enh.Out\n}", "func CreatedAtGTE(v time.Time) predicate.Order {\n\treturn predicate.Order(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func (ts Timestamp) Value() (driver.Value, error) {\n\tif !ts.Valid {\n\t\treturn nil, nil\n\t}\n\n\tif ts.InfinityModifier != Finite {\n\t\treturn ts.InfinityModifier.String(), nil\n\t}\n\treturn ts.Time, nil\n}", "func ValueGTE(v float64) predicate.K8sEvent {\n\treturn predicate.K8sEvent(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldValue), v))\n\t})\n}", "func (d *dbBase) ReadValues(ctx context.Context, q dbQuerier, qs *querySet, mi *modelInfo, cond *Condition, exprs []string, container interface{}, tz *time.Location) (int64, error) {\n\tvar (\n\t\tmaps []Params\n\t\tlists []ParamsList\n\t\tlist ParamsList\n\t)\n\n\ttyp := 0\n\tswitch v := container.(type) {\n\tcase *[]Params:\n\t\td := *v\n\t\tif len(d) == 0 {\n\t\t\tmaps = d\n\t\t}\n\t\ttyp = 1\n\tcase *[]ParamsList:\n\t\td := *v\n\t\tif len(d) == 0 {\n\t\t\tlists = d\n\t\t}\n\t\ttyp = 2\n\tcase *ParamsList:\n\t\td := *v\n\t\tif len(d) == 0 {\n\t\t\tlist = d\n\t\t}\n\t\ttyp = 3\n\tdefault:\n\t\tpanic(fmt.Errorf(\"unsupport read values type `%T`\", container))\n\t}\n\n\ttables := newDbTables(mi, d.ins)\n\n\tvar (\n\t\tcols []string\n\t\tinfos []*fieldInfo\n\t)\n\n\thasExprs := len(exprs) > 0\n\n\tQ := d.ins.TableQuote()\n\n\tif hasExprs {\n\t\tcols = make([]string, 0, len(exprs))\n\t\tinfos = make([]*fieldInfo, 0, len(exprs))\n\t\tfor _, ex := range exprs {\n\t\t\tindex, name, fi, suc := tables.parseExprs(mi, strings.Split(ex, ExprSep))\n\t\t\tif !suc {\n\t\t\t\tpanic(fmt.Errorf(\"unknown field/column name `%s`\", ex))\n\t\t\t}\n\t\t\tcols = append(cols, fmt.Sprintf(\"%s.%s%s%s %s%s%s\", index, Q, fi.column, Q, Q, name, Q))\n\t\t\tinfos = append(infos, fi)\n\t\t}\n\t} else {\n\t\tcols = make([]string, 0, len(mi.fields.dbcols))\n\t\tinfos = make([]*fieldInfo, 0, len(exprs))\n\t\tfor _, fi := range mi.fields.fieldsDB {\n\t\t\tcols = append(cols, fmt.Sprintf(\"T0.%s%s%s %s%s%s\", Q, fi.column, Q, Q, fi.name, Q))\n\t\t\tinfos = append(infos, fi)\n\t\t}\n\t}\n\n\twhere, args := tables.getCondSQL(cond, false, tz)\n\tgroupBy := tables.getGroupSQL(qs.groups)\n\torderBy := tables.getOrderSQL(qs.orders)\n\tlimit := tables.getLimitSQL(mi, qs.offset, qs.limit)\n\tjoin := tables.getJoinSQL()\n\tspecifyIndexes := tables.getIndexSql(mi.table, qs.useIndex, qs.indexes)\n\n\tsels := strings.Join(cols, \", \")\n\n\tsqlSelect := \"SELECT\"\n\tif qs.distinct {\n\t\tsqlSelect += \" DISTINCT\"\n\t}\n\tquery := fmt.Sprintf(\"%s %s FROM %s%s%s T0 %s%s%s%s%s%s\",\n\t\tsqlSelect, sels,\n\t\tQ, mi.table, Q,\n\t\tspecifyIndexes, join, where, groupBy, orderBy, limit)\n\n\td.ins.ReplaceMarks(&query)\n\n\trs, err := q.QueryContext(ctx, query, args...)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\trefs := make([]interface{}, len(cols))\n\tfor i := range refs {\n\t\tvar ref interface{}\n\t\trefs[i] = &ref\n\t}\n\n\tdefer rs.Close()\n\n\tvar (\n\t\tcnt int64\n\t\tcolumns []string\n\t)\n\tfor rs.Next() {\n\t\tif cnt == 0 {\n\t\t\tcols, err := rs.Columns()\n\t\t\tif err != nil {\n\t\t\t\treturn 0, err\n\t\t\t}\n\t\t\tcolumns = cols\n\t\t}\n\n\t\tif err := rs.Scan(refs...); err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\tswitch typ {\n\t\tcase 1:\n\t\t\tparams := make(Params, len(cols))\n\t\t\tfor i, ref := range refs {\n\t\t\t\tfi := infos[i]\n\n\t\t\t\tval := reflect.Indirect(reflect.ValueOf(ref)).Interface()\n\n\t\t\t\tvalue, err := d.convertValueFromDB(fi, val, tz)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(fmt.Errorf(\"db value convert failed `%v` %s\", val, err.Error()))\n\t\t\t\t}\n\n\t\t\t\tparams[columns[i]] = value\n\t\t\t}\n\t\t\tmaps = append(maps, params)\n\t\tcase 2:\n\t\t\tparams := make(ParamsList, 0, len(cols))\n\t\t\tfor i, ref := range refs {\n\t\t\t\tfi := infos[i]\n\n\t\t\t\tval := reflect.Indirect(reflect.ValueOf(ref)).Interface()\n\n\t\t\t\tvalue, err := d.convertValueFromDB(fi, val, tz)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(fmt.Errorf(\"db value convert failed `%v` %s\", val, err.Error()))\n\t\t\t\t}\n\n\t\t\t\tparams = append(params, value)\n\t\t\t}\n\t\t\tlists = append(lists, params)\n\t\tcase 3:\n\t\t\tfor i, ref := range refs {\n\t\t\t\tfi := infos[i]\n\n\t\t\t\tval := reflect.Indirect(reflect.ValueOf(ref)).Interface()\n\n\t\t\t\tvalue, err := d.convertValueFromDB(fi, val, tz)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(fmt.Errorf(\"db value convert failed `%v` %s\", val, err.Error()))\n\t\t\t\t}\n\n\t\t\t\tlist = append(list, value)\n\t\t\t}\n\t\t}\n\n\t\tcnt++\n\t}\n\n\tswitch v := container.(type) {\n\tcase *[]Params:\n\t\t*v = maps\n\tcase *[]ParamsList:\n\t\t*v = lists\n\tcase *ParamsList:\n\t\t*v = list\n\t}\n\n\treturn cnt, nil\n}", "func (v Timestamp) Value() (driver.Value, error) {\n\tif !v.Valid() {\n\t\treturn nil, nil\n\t}\n\treturn v.time, nil\n}", "func CreateTimeEQ(v time.Time) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldCreateTime), v))\n\t})\n}", "func (db *DatabaseIndex) measurementsByExpr(expr influxql.Expr) (Measurements, error) {\n\tswitch e := expr.(type) {\n\tcase *influxql.BinaryExpr:\n\t\tswitch e.Op {\n\t\tcase influxql.EQ, influxql.NEQ, influxql.EQREGEX, influxql.NEQREGEX:\n\t\t\ttag, ok := e.LHS.(*influxql.VarRef)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"left side of '%s' must be a tag name\", e.Op.String())\n\t\t\t}\n\n\t\t\ttf := &TagFilter{\n\t\t\t\tOp: e.Op,\n\t\t\t\tKey: tag.Val,\n\t\t\t}\n\n\t\t\tif influxql.IsRegexOp(e.Op) {\n\t\t\t\tre, ok := e.RHS.(*influxql.RegexLiteral)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn nil, fmt.Errorf(\"right side of '%s' must be a regular expression\", e.Op.String())\n\t\t\t\t}\n\t\t\t\ttf.Regex = re.Val\n\t\t\t} else {\n\t\t\t\ts, ok := e.RHS.(*influxql.StringLiteral)\n\t\t\t\tif !ok {\n\t\t\t\t\treturn nil, fmt.Errorf(\"right side of '%s' must be a tag value string\", e.Op.String())\n\t\t\t\t}\n\t\t\t\ttf.Value = s.Val\n\t\t\t}\n\n\t\t\treturn db.measurementsByTagFilters([]*TagFilter{tf}), nil\n\t\tcase influxql.OR, influxql.AND:\n\t\t\tlhsIDs, err := db.measurementsByExpr(e.LHS)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\trhsIDs, err := db.measurementsByExpr(e.RHS)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tif e.Op == influxql.OR {\n\t\t\t\treturn lhsIDs.union(rhsIDs), nil\n\t\t\t}\n\n\t\t\treturn lhsIDs.intersect(rhsIDs), nil\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"invalid operator\")\n\t\t}\n\tcase *influxql.ParenExpr:\n\t\treturn db.measurementsByExpr(e.Expr)\n\t}\n\treturn nil, fmt.Errorf(\"%#v\", expr)\n}", "func (da *DateArith) Eval(ctx context.Context, args map[interface{}]interface{}) (interface{}, error) {\n\tt, years, months, days, durations, err := da.evalArgs(ctx, args)\n\tif t.IsZero() || err != nil {\n\t\treturn nil, errors.Trace(err)\n\t}\n\n\tif !da.isAdd() {\n\t\tyears, months, days, durations = -years, -months, -days, -durations\n\t}\n\tt.Time = t.Time.Add(durations)\n\tt.Time = t.Time.AddDate(int(years), int(months), int(days))\n\n\t// \"2011-11-11 10:10:20.000000\" outputs \"2011-11-11 10:10:20\".\n\tif t.Time.Nanosecond() == 0 {\n\t\tt.Fsp = 0\n\t}\n\n\treturn t, nil\n}", "func Evaluate(e ast.Node, genCtx *GenCtx) parser_driver.ValueExpr {\n\tswitch t := e.(type) {\n\tcase *ast.ParenthesesExpr:\n\t\treturn Evaluate(t.Expr, genCtx)\n\tcase *ast.BinaryOperationExpr:\n\t\tres, err := operator.BinaryOps.Eval(t.Op.String(), Evaluate(t.L, genCtx), Evaluate(t.R, genCtx))\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"error occurred on eval: %+v\", err))\n\t\t}\n\t\treturn res\n\tcase *ast.UnaryOperationExpr:\n\t\tres, err := operator.UnaryOps.Eval(t.Op.String(), Evaluate(t.V, genCtx))\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"error occurred on eval: %+v\", err))\n\t\t}\n\t\treturn res\n\tcase *ast.IsNullExpr:\n\t\tsubResult := Evaluate(t.Expr, genCtx)\n\t\tc := ConvertToBoolOrNull(subResult)\n\t\tr := parser_driver.ValueExpr{}\n\t\tr.SetInt64(0)\n\t\tif c == -1 {\n\t\t\tr.SetInt64(1)\n\t\t}\n\t\treturn r\n\tcase *ast.ColumnNameExpr:\n\t\tfor key, value := range genCtx.unwrapPivotRows {\n\t\t\toriginTableName := t.Name.Table.L\n\t\t\tfor k, v := range genCtx.TableAlias {\n\t\t\t\tif v == originTableName {\n\t\t\t\t\toriginTableName = k\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\toriginColumnName := t.Name.Name.L\n\t\t\tif key == fmt.Sprintf(\"%s.%s\", originTableName, originColumnName) {\n\t\t\t\tv := parser_driver.ValueExpr{}\n\t\t\t\tv.SetValue(value)\n\t\t\t\tif tmpTable, ok := genCtx.TableAlias[t.Name.Table.L]; ok {\n\t\t\t\t\tt.Name.Table = model.NewCIStr(tmpTable)\n\t\t\t\t}\n\t\t\t\treturn v\n\t\t\t}\n\t\t}\n\t\tpanic(fmt.Sprintf(\"no such col %s in table %s\", t.Name, t.Name.Table))\n\tcase ast.ValueExpr:\n\t\tv := parser_driver.ValueExpr{}\n\t\tv.SetValue(t.GetValue())\n\t\tv.SetType(t.GetType())\n\t\treturn v\n\t}\n\n\t// is useless?\n\t// if e == nil {\n\t// \treturn trueValueExpr()\n\t// }\n\n\tpanic(\"not reachable\")\n\tv := parser_driver.ValueExpr{}\n\tv.SetNull()\n\treturn v\n}", "func dbUnixEpoch(db *gosql.DB) (float64, error) {\n\tvar epoch float64\n\tif err := db.QueryRow(\"SELECT now()::DECIMAL\").Scan(&epoch); err != nil {\n\t\treturn 0, err\n\t}\n\treturn epoch, nil\n}", "func (v TimestampNano) Value() (driver.Value, error) {\n\tif !v.Valid() {\n\t\treturn nil, nil\n\t}\n\treturn v.time.UnixNano(), nil\n}", "func CreatedAtGTE(v time.Time) predicate.GameServer {\n\treturn predicate.GameServer(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func CreatedAtGTE(v time.Time) predicate.Block {\n\treturn predicate.Block(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func updateTimeStampForCreateCallback(db *gorm.DB) {\n\tif db.Error != nil || db.Statement.Schema == nil {\n\t\treturn\n\t}\n\tuserName := username(db)\n\tnow := time.Now()\n\tcols := columns{\n\t\t{Name: \"CreatedOn\", Value: now}, {Name: \"CreatedBy\", Value: userName},\n\t\t{Name: \"ModifiedOn\", Value: now}, {Name: \"ModifiedBy\", Value: userName},\n\t}\n\tfor _, col := range cols {\n\t\tif field := db.Statement.Schema.LookUpField(col.Name); field != nil {\n\t\t\tif _, isZero := field.ValueOf(db.Statement.ReflectValue); isZero {\n\t\t\t\t_ = field.Set(db.Statement.ReflectValue, col.Value)\n\t\t\t}\n\t\t}\n\t}\n}", "func CreateTimeGTE(v time.Time) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreateTime), v))\n\t})\n}", "func MakeBinValExpr(op string, eval valueEval) func(scanner parser.Scanner, a, b Expr) Expr {\n\treturn func(scanner parser.Scanner, a, b Expr) Expr {\n\t\treturn newBinExpr(scanner, a, b, op, \"(%s \"+op+\" %s)\",\n\t\t\tfunc(ctx context.Context, a, b Value, _ Scope) (Value, error) {\n\t\t\t\treturn eval(a, b), nil\n\t\t\t})\n\t}\n}", "func CreatedAtEQ(v time.Time) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldCreatedAt), v))\n\t})\n}", "func CreatedAtEQ(v time.Time) predicate.User {\n\treturn predicate.User(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldCreatedAt), v))\n\t})\n}", "func (esg *expressionSQLGenerator) updateExpressionSQL(b sb.SQLBuilder, update exp.UpdateExpression) {\n\tesg.Generate(b, update.Col())\n\tb.WriteRunes(esg.dialectOptions.SetOperatorRune)\n\tesg.Generate(b, update.Val())\n}", "func (m *Measurement) expandExpr(expr influxql.Expr) []tagSetExpr {\n\t// Retrieve list of unique values for each tag.\n\tvaluesByTagKey := m.uniqueTagValues(expr)\n\n\t// Convert keys to slices.\n\tkeys := make([]string, 0, len(valuesByTagKey))\n\tfor key := range valuesByTagKey {\n\t\tkeys = append(keys, key)\n\t}\n\tsort.Strings(keys)\n\n\t// Order uniques by key.\n\tuniques := make([][]string, len(keys))\n\tfor i, key := range keys {\n\t\tuniques[i] = valuesByTagKey[key]\n\t}\n\n\t// Reduce a condition for each combination of tag values.\n\treturn expandExprWithValues(expr, keys, []tagExpr{}, uniques, 0)\n}", "func BuildNotNullExpr(ctx sessionctx.Context, expr Expression) Expression {\n\ttrace_util_0.Count(_util_00000, 206)\n\tisNull := NewFunctionInternal(ctx, ast.IsNull, types.NewFieldType(mysql.TypeTiny), expr)\n\tnotNull := NewFunctionInternal(ctx, ast.UnaryNot, types.NewFieldType(mysql.TypeTiny), isNull)\n\treturn notNull\n}", "func (b *Builder) buildInsertColumnOrValue(\n\tmodel IModel,\n\tskipFunc func(column string, model IModel) bool,\n\tgenerator func(query string, column string, hasComma bool) (string, bool),\n) string {\n\tvar query string\n\thasComma := true\n\n\tcolumns := model.GetColumns(model)\n\n\tif model.IsTimestamp() {\n\t\tcolumns = append(columns, \"created_at\", \"updated_at\")\n\t}\n\n\tif model.IsSoftDelete() {\n\t\tcolumns = append(columns, \"deleted_at\")\n\t}\n\n\tfor _, col := range columns {\n\t\tif !skipFunc(col, model) {\n\t\t\tquery, hasComma = generator(query, col, hasComma)\n\t\t}\n\t}\n\n\treturn query\n}", "func CreatedAtEQ(v time.Time) predicate.Order {\n\treturn predicate.Order(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldCreatedAt), v))\n\t})\n}", "func CreatedAtGTE(v time.Time) predicate.K8sEvent {\n\treturn predicate.K8sEvent(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func expressionToQuery(qp *model.QueryRangeParamsV2, varToQuery map[string]string, expression *govaluate.EvaluableExpression) (string, error) {\n\tvar formulaQuery string\n\tvars := unique(expression.Vars())\n\tfor idx, var_ := range vars[1:] {\n\t\tx, y := vars[idx], var_\n\t\tif !reflect.DeepEqual(qp.CompositeMetricQuery.BuilderQueries[x].GroupingTags, qp.CompositeMetricQuery.BuilderQueries[y].GroupingTags) {\n\t\t\treturn \"\", fmt.Errorf(\"group by must be same\")\n\t\t}\n\t}\n\tvar modified []govaluate.ExpressionToken\n\ttokens := expression.Tokens()\n\tfor idx := range tokens {\n\t\ttoken := tokens[idx]\n\t\tif token.Kind == govaluate.VARIABLE {\n\t\t\ttoken.Value = fmt.Sprintf(\"%v.value\", token.Value)\n\t\t\ttoken.Meta = fmt.Sprintf(\"%v.value\", token.Meta)\n\t\t}\n\t\tmodified = append(modified, token)\n\t}\n\t// err should be nil here since the expression is already validated\n\tformula, _ := govaluate.NewEvaluableExpressionFromTokens(modified)\n\n\tvar formulaSubQuery string\n\tvar joinUsing string\n\tvar prevVar string\n\tfor idx, var_ := range vars {\n\t\tquery := varToQuery[var_]\n\t\tgroupTags := qp.CompositeMetricQuery.BuilderQueries[var_].GroupingTags\n\t\tgroupTags = append(groupTags, \"ts\")\n\t\tif joinUsing == \"\" {\n\t\t\tfor _, tag := range groupTags {\n\t\t\t\tjoinUsing += fmt.Sprintf(\"%s.%s as %s, \", var_, tag, tag)\n\t\t\t}\n\t\t\tjoinUsing = strings.TrimSuffix(joinUsing, \", \")\n\t\t}\n\t\tformulaSubQuery += fmt.Sprintf(\"(%s) as %s \", query, var_)\n\t\tif idx > 0 {\n\t\t\tformulaSubQuery += \" ON \"\n\t\t\tfor _, tag := range groupTags {\n\t\t\t\tformulaSubQuery += fmt.Sprintf(\"%s.%s = %s.%s AND \", prevVar, tag, var_, tag)\n\t\t\t}\n\t\t\tformulaSubQuery = strings.TrimSuffix(formulaSubQuery, \" AND \")\n\t\t}\n\t\tif idx < len(vars)-1 {\n\t\t\tformulaSubQuery += \" GLOBAL INNER JOIN\"\n\t\t}\n\t\tprevVar = var_\n\t}\n\tformulaQuery = fmt.Sprintf(\"SELECT %s, %s as value FROM \", joinUsing, formula.ExpressionString()) + formulaSubQuery\n\treturn formulaQuery, nil\n}", "func NewExpr(terms interface{}) *Expr {\n\treturn &Expr{\n\t\tNegated: false,\n\t\tTerms: terms,\n\t\tIndex: 0,\n\t}\n}", "func (d *dbBase) convertValueFromDB(fi *fieldInfo, val interface{}, tz *time.Location) (interface{}, error) {\n\tif val == nil {\n\t\treturn nil, nil\n\t}\n\n\tvar value interface{}\n\tvar tErr error\n\n\tvar str *StrTo\n\tswitch v := val.(type) {\n\tcase []byte:\n\t\ts := StrTo(string(v))\n\t\tstr = &s\n\tcase string:\n\t\ts := StrTo(v)\n\t\tstr = &s\n\t}\n\n\tfieldType := fi.fieldType\n\nsetValue:\n\tswitch {\n\tcase fieldType == TypeBooleanField:\n\t\tif str == nil {\n\t\t\tswitch v := val.(type) {\n\t\t\tcase int64:\n\t\t\t\tb := v == 1\n\t\t\t\tvalue = b\n\t\t\tdefault:\n\t\t\t\ts := StrTo(ToStr(v))\n\t\t\t\tstr = &s\n\t\t\t}\n\t\t}\n\t\tif str != nil {\n\t\t\tb, err := str.Bool()\n\t\t\tif err != nil {\n\t\t\t\ttErr = err\n\t\t\t\tgoto end\n\t\t\t}\n\t\t\tvalue = b\n\t\t}\n\tcase fieldType == TypeVarCharField || fieldType == TypeCharField || fieldType == TypeTextField || fieldType == TypeJSONField || fieldType == TypeJsonbField:\n\t\tif str == nil {\n\t\t\tvalue = ToStr(val)\n\t\t} else {\n\t\t\tvalue = str.String()\n\t\t}\n\tcase fieldType == TypeTimeField || fieldType == TypeDateField || fieldType == TypeDateTimeField:\n\t\tif str == nil {\n\t\t\tswitch t := val.(type) {\n\t\t\tcase time.Time:\n\t\t\t\td.ins.TimeFromDB(&t, tz)\n\t\t\t\tvalue = t\n\t\t\tdefault:\n\t\t\t\ts := StrTo(ToStr(t))\n\t\t\t\tstr = &s\n\t\t\t}\n\t\t}\n\t\tif str != nil {\n\t\t\ts := str.String()\n\t\t\tvar (\n\t\t\t\tt time.Time\n\t\t\t\terr error\n\t\t\t)\n\n\t\t\tif fi.timePrecision != nil && len(s) >= (20+*fi.timePrecision) {\n\t\t\t\tlayout := formatDateTime + \".\"\n\t\t\t\tfor i := 0; i < *fi.timePrecision; i++ {\n\t\t\t\t\tlayout += \"0\"\n\t\t\t\t}\n\t\t\t\tt, err = time.ParseInLocation(layout, s[:20+*fi.timePrecision], tz)\n\t\t\t} else if len(s) >= 19 {\n\t\t\t\ts = s[:19]\n\t\t\t\tt, err = time.ParseInLocation(formatDateTime, s, tz)\n\t\t\t} else if len(s) >= 10 {\n\t\t\t\tif len(s) > 10 {\n\t\t\t\t\ts = s[:10]\n\t\t\t\t}\n\t\t\t\tt, err = time.ParseInLocation(formatDate, s, tz)\n\t\t\t} else if len(s) >= 8 {\n\t\t\t\tif len(s) > 8 {\n\t\t\t\t\ts = s[:8]\n\t\t\t\t}\n\t\t\t\tt, err = time.ParseInLocation(formatTime, s, tz)\n\t\t\t}\n\t\t\tt = t.In(DefaultTimeLoc)\n\n\t\t\tif err != nil && s != \"00:00:00\" && s != \"0000-00-00\" && s != \"0000-00-00 00:00:00\" {\n\t\t\t\ttErr = err\n\t\t\t\tgoto end\n\t\t\t}\n\t\t\tvalue = t\n\t\t}\n\tcase fieldType&IsIntegerField > 0:\n\t\tif str == nil {\n\t\t\ts := StrTo(ToStr(val))\n\t\t\tstr = &s\n\t\t}\n\t\tif str != nil {\n\t\t\tvar err error\n\t\t\tswitch fieldType {\n\t\t\tcase TypeBitField:\n\t\t\t\t_, err = str.Int8()\n\t\t\tcase TypeSmallIntegerField:\n\t\t\t\t_, err = str.Int16()\n\t\t\tcase TypeIntegerField:\n\t\t\t\t_, err = str.Int32()\n\t\t\tcase TypeBigIntegerField:\n\t\t\t\t_, err = str.Int64()\n\t\t\tcase TypePositiveBitField:\n\t\t\t\t_, err = str.Uint8()\n\t\t\tcase TypePositiveSmallIntegerField:\n\t\t\t\t_, err = str.Uint16()\n\t\t\tcase TypePositiveIntegerField:\n\t\t\t\t_, err = str.Uint32()\n\t\t\tcase TypePositiveBigIntegerField:\n\t\t\t\t_, err = str.Uint64()\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\ttErr = err\n\t\t\t\tgoto end\n\t\t\t}\n\t\t\tif fieldType&IsPositiveIntegerField > 0 {\n\t\t\t\tv, _ := str.Uint64()\n\t\t\t\tvalue = v\n\t\t\t} else {\n\t\t\t\tv, _ := str.Int64()\n\t\t\t\tvalue = v\n\t\t\t}\n\t\t}\n\tcase fieldType == TypeFloatField || fieldType == TypeDecimalField:\n\t\tif str == nil {\n\t\t\tswitch v := val.(type) {\n\t\t\tcase float64:\n\t\t\t\tvalue = v\n\t\t\tdefault:\n\t\t\t\ts := StrTo(ToStr(v))\n\t\t\t\tstr = &s\n\t\t\t}\n\t\t}\n\t\tif str != nil {\n\t\t\tv, err := str.Float64()\n\t\t\tif err != nil {\n\t\t\t\ttErr = err\n\t\t\t\tgoto end\n\t\t\t}\n\t\t\tvalue = v\n\t\t}\n\tcase fieldType&IsRelField > 0:\n\t\tfi = fi.relModelInfo.fields.pk\n\t\tfieldType = fi.fieldType\n\t\tgoto setValue\n\t}\n\nend:\n\tif tErr != nil {\n\t\terr := fmt.Errorf(\"convert to `%s` failed, field: %s err: %s\", fi.addrValue.Type(), fi.fullName, tErr)\n\t\treturn nil, err\n\t}\n\n\treturn value, nil\n}", "func (df *DateFilter) BuildQuery(t *time.Time) string {\n\tvar bf bytes.Buffer\n\n\tif df.Attribute == \"\" {\n\t\tdf.defaultAttribute()\n\t}\n\n\tif t == nil {\n\t\tn := time.Now()\n\t\tt = &n\n\t}\n\n\tbf.WriteString(string(df.Attribute))\n\tif df.Custom != \"\" {\n\t\tbf.WriteString(df.Custom)\n\t} else {\n\t\tbf.WriteString(\">=\")\n\t\tbf.WriteString(df.getDateAPIFormat(t))\n\t}\n\n\treturn bf.String()\n}", "func CreatedAtEQ(v time.Time) predicate.Job {\n\treturn predicate.Job(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldCreatedAt), v))\n\t})\n}", "func CreateTimeIn(vs ...time.Time) predicate.Location {\n\tv := make([]interface{}, len(vs))\n\tfor i := range v {\n\t\tv[i] = vs[i]\n\t}\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\t// if not arguments were provided, append the FALSE constants,\n\t\t// since we can't apply \"IN ()\". This will make this predicate falsy.\n\t\tif len(vs) == 0 {\n\t\t\ts.Where(sql.False())\n\t\t\treturn\n\t\t}\n\t\ts.Where(sql.In(s.C(FieldCreateTime), v...))\n\t},\n\t)\n}", "func (*ExprTrunc) Immutable() {}", "func (tbl Table) Emit(tw *trap.Writer, values ...interface{}) {\n\tif ncol, nval := len(tbl.schema), len(values); ncol != nval {\n\t\tlog.Fatalf(\"wrong number of values for table %s; expected %d, but got %d\", tbl.name, ncol, nval)\n\t}\n\tfor i, col := range tbl.schema {\n\t\tif !col.columnType.valid(values[i]) {\n\t\t\tpanic(fmt.Sprintf(\"Invalid value for column %d of table %s; expected a %s, but got %s which is a %s\", i, tbl.name, col.columnType.ref(), values[i], reflect.TypeOf(values[i])))\n\t\t}\n\t}\n\ttw.Emit(tbl.name, values)\n}", "func InitCronExpr(ctx context.Context, duration time.Duration) error {\n\tif duration < 0 || duration > 12*time.Hour {\n\t\treturn moerr.NewNotSupported(ctx, \"export cron expr not support cycle: %v\", duration)\n\t}\n\tif duration < 5*time.Minute {\n\t\tMergeTaskCronExpr = fmt.Sprintf(\"@every %.0fs\", duration.Seconds())\n\t} else if duration < time.Hour {\n\t\tconst unit = 5 * time.Minute\n\t\tduration = (duration + unit - 1) / unit * unit\n\t\tswitch duration {\n\t\tcase 5 * time.Minute:\n\t\t\tMergeTaskCronExpr = MergeTaskCronExprEvery05Min\n\t\tcase 15 * time.Minute:\n\t\t\tMergeTaskCronExpr = MergeTaskCronExprEvery15Min\n\t\tdefault:\n\t\t\tMergeTaskCronExpr = fmt.Sprintf(\"@every %.0fm\", duration.Minutes())\n\t\t}\n\t} else {\n\t\tminHour := duration / time.Hour\n\t\tswitch minHour {\n\t\tcase 1:\n\t\t\tMergeTaskCronExpr = MergeTaskCronExprEvery1Hour\n\t\tcase 2:\n\t\t\tMergeTaskCronExpr = MergeTaskCronExprEvery2Hour\n\t\tcase 4:\n\t\t\tMergeTaskCronExpr = MergeTaskCronExprEvery4Hour\n\t\tdefault:\n\t\t\tvar hours = make([]string, 0, 12)\n\t\t\tfor h := minHour; h < 24; h += minHour {\n\t\t\t\thours = append(hours, strconv.Itoa(int(h)))\n\t\t\t}\n\t\t\tMergeTaskCronExpr = fmt.Sprintf(\"0 0 %s * * *\", strings.Join(hours, \",\"))\n\t\t}\n\t}\n\treturn nil\n}", "func ConcatQuery(c *Context, values ...interface{}) string {\n\ts := strings.Builder{}\n\n\tfor _, val := range values {\n\t\tswitch v := val.(type) {\n\t\tcase (Field):\n\t\t\ts.WriteString(v.QueryString(c))\n\t\tcase (Condition):\n\t\t\ts.WriteString(v(c))\n\t\tcase (SelectQuery):\n\t\t\tsql, _ := v.SQL(SQLBuilder{Context: c})\n\t\t\ts.WriteString(getSubQuerySQL(sql))\n\t\tcase (string):\n\t\t\ts.WriteString(v)\n\t\tdefault:\n\t\t\tpanic(`Can only use strings, Fields, Conditions and SelectQueries to build SQL`)\n\t\t}\n\t}\n\treturn s.String()\n}", "func CreatedAtGTE(v time.Time) predicate.Agent {\n\treturn predicate.Agent(func(s *sql.Selector) {\n\t\ts.Where(sql.GTE(s.C(FieldCreatedAt), v))\n\t})\n}", "func ValuePath(column string, opts ...Option) sql.Querier {\n\treturn sql.ExprFunc(func(b *sql.Builder) {\n\t\tvaluePath(b, column, opts...)\n\t})\n}" ]
[ "0.5238076", "0.517202", "0.51271206", "0.50810814", "0.503775", "0.5023392", "0.498908", "0.48464355", "0.4793017", "0.47793496", "0.47737923", "0.4713352", "0.4672723", "0.46356982", "0.45884207", "0.45605707", "0.45489782", "0.45323822", "0.45212737", "0.45011598", "0.44839203", "0.4478501", "0.44739264", "0.44736728", "0.44574115", "0.44564104", "0.44443187", "0.44426635", "0.44414455", "0.4434923", "0.44322744", "0.44314563", "0.4424172", "0.44228107", "0.442109", "0.44162962", "0.4411835", "0.44114336", "0.43985805", "0.43768626", "0.43745354", "0.43729284", "0.43729228", "0.43582177", "0.43525156", "0.43392465", "0.43263456", "0.43259436", "0.43259436", "0.43255126", "0.43251207", "0.43186226", "0.43170163", "0.4315935", "0.43073186", "0.43056718", "0.42967564", "0.4295787", "0.42950606", "0.42948395", "0.42895103", "0.4283023", "0.42824796", "0.4278526", "0.42668355", "0.42656812", "0.42650154", "0.4260609", "0.42591432", "0.425542", "0.42481315", "0.42390957", "0.4232306", "0.42316133", "0.42297158", "0.4226888", "0.4225013", "0.42100316", "0.42051384", "0.41934693", "0.4181433", "0.4181433", "0.4176014", "0.41660348", "0.4165233", "0.4165116", "0.41603023", "0.4158806", "0.4158337", "0.41573292", "0.41564494", "0.41540188", "0.4153603", "0.41529462", "0.41522175", "0.41507167", "0.4146393", "0.41454715", "0.41440758", "0.4142227" ]
0.49611345
7
MarshalJSON amends LogEntry instances with type metadata so that clients do not need to be concerned with the tedium of doing so.
func (l LogEntry) MarshalJSON() ([]byte, error) { type Alias LogEntry return json.Marshal( struct { meta.TypeMeta `json:",inline"` Alias `json:",inline"` }{ TypeMeta: meta.TypeMeta{ APIVersion: meta.APIVersion, Kind: "LogEntry", }, Alias: (Alias)(l), }, ) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o LogContent) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.Attributes != nil {\n\t\ttoSerialize[\"attributes\"] = o.Attributes\n\t}\n\tif o.Host != nil {\n\t\ttoSerialize[\"host\"] = o.Host\n\t}\n\tif o.Message != nil {\n\t\ttoSerialize[\"message\"] = o.Message\n\t}\n\tif o.Service != nil {\n\t\ttoSerialize[\"service\"] = o.Service\n\t}\n\tif o.Tags != nil {\n\t\ttoSerialize[\"tags\"] = o.Tags\n\t}\n\tif o.Timestamp != nil {\n\t\tif o.Timestamp.Nanosecond() == 0 {\n\t\t\ttoSerialize[\"timestamp\"] = o.Timestamp.Format(\"2006-01-02T15:04:05Z07:00\")\n\t\t} else {\n\t\t\ttoSerialize[\"timestamp\"] = o.Timestamp.Format(\"2006-01-02T15:04:05.000Z07:00\")\n\t\t}\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (h HeavyAccountCopyLogDetails) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"accountName\", h.AccountName)\n\tobjectMap[\"copyLogDetailsType\"] = ClassDiscriminatorDataBoxHeavy\n\tpopulate(objectMap, \"copyLogLink\", h.CopyLogLink)\n\tpopulate(objectMap, \"copyVerboseLogLink\", h.CopyVerboseLogLink)\n\treturn json.Marshal(objectMap)\n}", "func (f FlowLog) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"etag\", f.Etag)\n\tpopulate(objectMap, \"id\", f.ID)\n\tpopulate(objectMap, \"location\", f.Location)\n\tpopulate(objectMap, \"name\", f.Name)\n\tpopulate(objectMap, \"properties\", f.Properties)\n\tpopulate(objectMap, \"tags\", f.Tags)\n\tpopulate(objectMap, \"type\", f.Type)\n\treturn json.Marshal(objectMap)\n}", "func (m ConsoleLogEvent) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 1)\n\n\taO0, err := swag.WriteJSON(m.RealtimeBuildEvent)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\t// now for regular properties\n\tvar propsConsoleLogEvent struct {\n\t\tLines []string `json:\"lines\"`\n\n\t\tStepRecordID strfmt.UUID `json:\"stepRecordId,omitempty\"`\n\n\t\tTimelineID strfmt.UUID `json:\"timelineId,omitempty\"`\n\n\t\tTimelineRecordID strfmt.UUID `json:\"timelineRecordId,omitempty\"`\n\t}\n\tpropsConsoleLogEvent.Lines = m.Lines\n\n\tpropsConsoleLogEvent.StepRecordID = m.StepRecordID\n\n\tpropsConsoleLogEvent.TimelineID = m.TimelineID\n\n\tpropsConsoleLogEvent.TimelineRecordID = m.TimelineRecordID\n\n\tjsonDataPropsConsoleLogEvent, errConsoleLogEvent := swag.WriteJSON(propsConsoleLogEvent)\n\tif errConsoleLogEvent != nil {\n\t\treturn nil, errConsoleLogEvent\n\t}\n\t_parts = append(_parts, jsonDataPropsConsoleLogEvent)\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (f FlowLogInformation) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"flowAnalyticsConfiguration\", f.FlowAnalyticsConfiguration)\n\tpopulate(objectMap, \"properties\", f.Properties)\n\tpopulate(objectMap, \"targetResourceId\", f.TargetResourceID)\n\treturn json.Marshal(objectMap)\n}", "func (rec JSONLogRec) MarshalJSONObject(enc *gojay.Encoder) {\n\tif !rec.DisableTimestamp {\n\t\ttimestampFmt := rec.TimestampFormat\n\t\tif timestampFmt == \"\" {\n\t\t\ttimestampFmt = logr.DefTimestampFormat\n\t\t}\n\t\ttime := rec.Time()\n\t\tenc.AddTimeKey(rec.KeyTimestamp, &time, timestampFmt)\n\t}\n\tif !rec.DisableLevel {\n\t\tenc.AddStringKey(rec.KeyLevel, rec.Level().Name)\n\t}\n\tif !rec.DisableMsg {\n\t\tenc.AddStringKey(rec.KeyMsg, rec.Msg())\n\t}\n\tif !rec.DisableContext {\n\t\tctxFields := rec.sorter(rec.Fields())\n\t\tif rec.KeyContextFields != \"\" {\n\t\t\tenc.AddObjectKey(rec.KeyContextFields, jsonFields(ctxFields))\n\t\t} else {\n\t\t\tif len(ctxFields) > 0 {\n\t\t\t\tfor _, cf := range ctxFields {\n\t\t\t\t\tkey := rec.prefixCollision(cf.Key)\n\t\t\t\t\tencodeField(enc, key, cf.Val)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tif rec.stacktrace && !rec.DisableStacktrace {\n\t\tframes := rec.StackFrames()\n\t\tif len(frames) > 0 {\n\t\t\tenc.AddArrayKey(rec.KeyStacktrace, stackFrames(frames))\n\t\t}\n\t}\n\n}", "func (l LogSpecification) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"blobDuration\", l.BlobDuration)\n\tpopulate(objectMap, \"displayName\", l.DisplayName)\n\tpopulate(objectMap, \"name\", l.Name)\n\treturn json.Marshal(objectMap)\n}", "func (l LogSpecification) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"blobDuration\", l.BlobDuration)\n\tpopulate(objectMap, \"displayName\", l.DisplayName)\n\tpopulate(objectMap, \"name\", l.Name)\n\treturn json.Marshal(objectMap)\n}", "func (hacld HeavyAccountCopyLogDetails) MarshalJSON() ([]byte, error) {\n\thacld.CopyLogDetailsType = CopyLogDetailsTypeDataBoxHeavy\n\tobjectMap := make(map[string]interface{})\n\tif hacld.AccountName != nil {\n\t\tobjectMap[\"accountName\"] = hacld.AccountName\n\t}\n\tif hacld.CopyLogLink != nil {\n\t\tobjectMap[\"copyLogLink\"] = hacld.CopyLogLink\n\t}\n\tif hacld.CopyLogDetailsType != \"\" {\n\t\tobjectMap[\"copyLogDetailsType\"] = hacld.CopyLogDetailsType\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (a AccountCopyLogDetails) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"accountName\", a.AccountName)\n\tobjectMap[\"copyLogDetailsType\"] = ClassDiscriminatorDataBox\n\tpopulate(objectMap, \"copyLogLink\", a.CopyLogLink)\n\tpopulate(objectMap, \"copyVerboseLogLink\", a.CopyVerboseLogLink)\n\treturn json.Marshal(objectMap)\n}", "func (e *CloudTrailNotifications) Marshal() ([]byte, error) {\n\treturn json.Marshal(e)\n}", "func (entry Entry) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(map[string]interface{}{\"item\": entry.key, \"count\": entry.value})\n}", "func (m SnoozeLogEntryAllOf2) MarshalJSON() ([]byte, error) {\n\tvar b1, b2, b3 []byte\n\tvar err error\n\tb1, err = json.Marshal(struct {\n\t}{},\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tb2, err = json.Marshal(struct {\n\t\tChangedActions []IncidentAction `json:\"changed_actions,omitempty\"`\n\t}{\n\n\t\tChangedActions: m.changedActionsField,\n\t},\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn swag.ConcatJSON(b1, b2, b3), nil\n}", "func (e *Entry) MarshalJSON() ([]byte, error) {\n\tentry := dbEntry{\n\t\t_Entry: _Entry(*e),\n\t\tIntStatus: int(e.Status),\n\t}\n\treturn json.Marshal(entry)\n}", "func ToJsonableEntry(e iface.IPFSLogEntry) interface{} {\n\tidentity := (*Identity)(nil)\n\tif e.GetIdentity() != nil {\n\t\tidentity = ToJsonableIdentity(e.GetIdentity())\n\t}\n\n\tswitch e.GetV() {\n\tcase 0:\n\t\th := (*string)(nil)\n\t\tif e.GetHash().Defined() {\n\t\t\tval := e.GetHash().String()\n\t\t\th = &val\n\t\t}\n\n\t\tnextValues := make([]string, len(e.GetNext()))\n\t\tfor i, n := range e.GetNext() {\n\t\t\tnextValues[i] = n.String()\n\t\t}\n\n\t\treturn &EntryV0{\n\t\t\tV: e.GetV(),\n\t\t\tID: e.GetLogID(),\n\t\t\tKey: hex.EncodeToString(e.GetKey()),\n\t\t\tSig: hex.EncodeToString(e.GetSig()),\n\t\t\tHash: h,\n\t\t\tNext: nextValues,\n\t\t\tClock: ToJsonableLamportClock(e.GetClock()),\n\t\t\tPayload: string(e.GetPayload()),\n\t\t}\n\tcase 1:\n\t\treturn &EntryV1{\n\t\t\tV: e.GetV(),\n\t\t\tLogID: e.GetLogID(),\n\t\t\tKey: hex.EncodeToString(e.GetKey()),\n\t\t\tSig: hex.EncodeToString(e.GetSig()),\n\t\t\tHash: nil,\n\t\t\tNext: e.GetNext(),\n\t\t\tClock: ToJsonableLamportClock(e.GetClock()),\n\t\t\tPayload: string(e.GetPayload()),\n\t\t\tIdentity: identity,\n\t\t}\n\tdefault:\n\t\tret := &EntryV2{\n\t\t\tV: e.GetV(),\n\t\t\tLogID: e.GetLogID(),\n\t\t\tKey: hex.EncodeToString(e.GetKey()),\n\t\t\tSig: hex.EncodeToString(e.GetSig()),\n\t\t\tHash: nil,\n\t\t\tNext: e.GetNext(),\n\t\t\tRefs: e.GetRefs(),\n\t\t\tClock: ToJsonableLamportClock(e.GetClock()),\n\t\t\tPayload: string(e.GetPayload()),\n\t\t\tIdentity: identity,\n\t\t}\n\n\t\t{\n\t\t\tadd := e.GetAdditionalData()\n\n\t\t\tencryptedLinks, okEncrypted := add[iface.KeyEncryptedLinks]\n\t\t\tencryptedLinksNonce, okEncryptedNonce := add[iface.KeyEncryptedLinksNonce]\n\n\t\t\tif okEncrypted && okEncryptedNonce {\n\t\t\t\tret.EncryptedLinks = encryptedLinks\n\t\t\t\tret.EncryptedLinksNonce = encryptedLinksNonce\n\n\t\t\t\tret.Next = []cid.Cid{}\n\t\t\t\tret.Refs = []cid.Cid{}\n\t\t\t}\n\t\t}\n\n\t\treturn ret\n\t}\n}", "func (l LogAnalytics) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"logType\", l.LogType)\n\tpopulate(objectMap, \"metadata\", l.Metadata)\n\tpopulate(objectMap, \"workspaceId\", l.WorkspaceID)\n\tpopulate(objectMap, \"workspaceKey\", l.WorkspaceKey)\n\tpopulate(objectMap, \"workspaceResourceId\", l.WorkspaceResourceID)\n\treturn json.Marshal(objectMap)\n}", "func (cld CopyLogDetails) MarshalJSON() ([]byte, error) {\n\tcld.CopyLogDetailsType = CopyLogDetailsTypeCopyLogDetails\n\tobjectMap := make(map[string]interface{})\n\tif cld.CopyLogDetailsType != \"\" {\n\t\tobjectMap[\"copyLogDetailsType\"] = cld.CopyLogDetailsType\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (acld AccountCopyLogDetails) MarshalJSON() ([]byte, error) {\n\tacld.CopyLogDetailsType = CopyLogDetailsTypeDataBox\n\tobjectMap := make(map[string]interface{})\n\tif acld.AccountName != nil {\n\t\tobjectMap[\"accountName\"] = acld.AccountName\n\t}\n\tif acld.CopyLogLink != nil {\n\t\tobjectMap[\"copyLogLink\"] = acld.CopyLogLink\n\t}\n\tif acld.CopyLogDetailsType != \"\" {\n\t\tobjectMap[\"copyLogDetailsType\"] = acld.CopyLogDetailsType\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (f FlowLogProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"enabled\", f.Enabled)\n\tpopulate(objectMap, \"format\", f.Format)\n\tpopulate(objectMap, \"retentionPolicy\", f.RetentionPolicy)\n\tpopulate(objectMap, \"storageId\", f.StorageID)\n\treturn json.Marshal(objectMap)\n}", "func (m ServerConfigResultEntry) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 2)\n\n\taO0, err := swag.WriteJSON(m.PolicyAbstractConfigResultEntry)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\tvar dataAO1 struct {\n\t\tConfigResult *ServerConfigResultRef `json:\"ConfigResult,omitempty\"`\n\t}\n\n\tdataAO1.ConfigResult = m.ConfigResult\n\n\tjsonDataAO1, errAO1 := swag.WriteJSON(dataAO1)\n\tif errAO1 != nil {\n\t\treturn nil, errAO1\n\t}\n\t_parts = append(_parts, jsonDataAO1)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (gr gelfRecord) MarshalJSONObject(enc *gojay.Encoder) {\n\tenc.AddStringKey(GelfVersionKey, GelfVersion)\n\tenc.AddStringKey(GelfHostKey, gr.getHostname())\n\tenc.AddStringKey(GelfShortKey, gr.Msg())\n\n\tif gr.level.Stacktrace {\n\t\tframes := gr.StackFrames()\n\t\tif len(frames) != 0 {\n\t\t\tvar sbuf strings.Builder\n\t\t\tfor _, frame := range frames {\n\t\t\t\tfmt.Fprintf(&sbuf, \"%s\\n %s:%d\\n\", frame.Function, frame.File, frame.Line)\n\t\t\t}\n\t\t\tenc.AddStringKey(GelfFullKey, sbuf.String())\n\t\t}\n\t}\n\n\tsecs := float64(gr.Time().UTC().Unix())\n\tmillis := float64(gr.Time().Nanosecond() / 1000000)\n\tts := secs + (millis / 1000)\n\tenc.AddFloat64Key(GelfTimestampKey, ts)\n\n\tenc.AddUint32Key(GelfLevelKey, uint32(gr.level.ID))\n\n\tvar fields []logr.Field\n\tif gr.EnableCaller {\n\t\tcaller := logr.Field{\n\t\t\tKey: \"_caller\",\n\t\t\tType: logr.StringType,\n\t\t\tString: gr.LogRec.Caller(),\n\t\t}\n\t\tfields = append(fields, caller)\n\t}\n\n\tfields = append(fields, gr.Fields()...)\n\tif gr.sorter != nil {\n\t\tfields = gr.sorter(fields)\n\t}\n\n\tif len(fields) > 0 {\n\t\tfor _, field := range fields {\n\t\t\tif !strings.HasPrefix(\"_\", field.Key) {\n\t\t\t\tfield.Key = \"_\" + field.Key\n\t\t\t}\n\t\t\tif err := encodeField(enc, field); err != nil {\n\t\t\t\tenc.AddStringKey(field.Key, fmt.Sprintf(\"<error encoding field: %v>\", err))\n\t\t\t}\n\t\t}\n\t}\n}", "func (f FlowLogPropertiesFormat) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"enabled\", f.Enabled)\n\tpopulate(objectMap, \"flowAnalyticsConfiguration\", f.FlowAnalyticsConfiguration)\n\tpopulate(objectMap, \"format\", f.Format)\n\tpopulate(objectMap, \"provisioningState\", f.ProvisioningState)\n\tpopulate(objectMap, \"retentionPolicy\", f.RetentionPolicy)\n\tpopulate(objectMap, \"storageId\", f.StorageID)\n\tpopulate(objectMap, \"targetResourceGuid\", f.TargetResourceGUID)\n\tpopulate(objectMap, \"targetResourceId\", f.TargetResourceID)\n\treturn json.Marshal(objectMap)\n}", "func (ident DockerConfigEntry) MarshalJSON() ([]byte, error) {\n\ttoEncode := dockerConfigEntryWithAuth{ident.Username, ident.Password, ident.Email, \"\"}\n\ttoEncode.Auth = encodeDockerConfigFieldAuth(ident.Username, ident.Password)\n\n\treturn json.Marshal(toEncode)\n}", "func Encode(entry *Entry) ([]byte, error) {\n\t// marshal the data into protobuf format\n\tdata, err := proto.Marshal(entry)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not marshal log entry\")\n\t}\n\n\t// write the size of the data to a new buffer\n\tsizebytes := make([]byte, 4)\n\tbinary.LittleEndian.PutUint32(sizebytes, uint32(len(data)))\n\tbuf := bytes.NewBuffer(sizebytes)\n\n\t// write the data to the buffer\n\t_, err = buf.Write(data)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not write data to buffer\")\n\t}\n\n\treturn buf.Bytes(), nil\n}", "func (ident RegistryConfigEntry) MarshalJSON() ([]byte, error) {\n\ttoEncode := registryConfigEntryWithAuth{ident.Username, ident.Password, ident.Email, \"\"}\n\ttoEncode.Auth = encodeRegistryConfigFieldAuth(ident.Username, ident.Password)\n\n\treturn json.Marshal(toEncode)\n}", "func (c CustomerDiskCopyLogDetails) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tobjectMap[\"copyLogDetailsType\"] = ClassDiscriminatorDataBoxCustomerDisk\n\tpopulate(objectMap, \"errorLogLink\", c.ErrorLogLink)\n\tpopulate(objectMap, \"serialNumber\", c.SerialNumber)\n\tpopulate(objectMap, \"verboseLogLink\", c.VerboseLogLink)\n\treturn json.Marshal(objectMap)\n}", "func (e Entry) Write() {\n\tif e.o.enc != nil {\n\t\tif e.o.enc.format == FormatJSON {\n\t\t\te.o.enc.closeObject()\n\t\t}\n\t\te.l.write(e)\n\t}\n}", "func (v *TypeReference) MarshalLogObject(enc zapcore.ObjectEncoder) (err error) {\n\tif v == nil {\n\t\treturn nil\n\t}\n\tenc.AddString(\"name\", v.Name)\n\tenc.AddString(\"importPath\", v.ImportPath)\n\tif v.Annotations != nil {\n\t\terr = multierr.Append(err, enc.AddObject(\"annotations\", (_Map_String_String_Zapper)(v.Annotations)))\n\t}\n\treturn err\n}", "func (d DiskCopyLogDetails) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tobjectMap[\"copyLogDetailsType\"] = ClassDiscriminatorDataBoxDisk\n\tpopulate(objectMap, \"diskSerialNumber\", d.DiskSerialNumber)\n\tpopulate(objectMap, \"errorLogLink\", d.ErrorLogLink)\n\tpopulate(objectMap, \"verboseLogLink\", d.VerboseLogLink)\n\treturn json.Marshal(objectMap)\n}", "func (e Entry) Encode() ([]byte, error) {\n\treturn json.Marshal(e)\n}", "func entrySize(e *logging.LogEntry) (int, error) {\n\tb, err := json.Marshal(e)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn len(b), nil\n}", "func (f *CustomJSONFormatter) Format(entry *logrus.Entry) ([]byte, error) {\n\tdata := make(logrus.Fields, len(entry.Data)+3)\n\tfor k, v := range entry.Data {\n\t\tswitch v := v.(type) {\n\t\tcase error:\n\t\t\t// Otherwise errors are ignored by `encoding/json`\n\t\t\t// https://github.com/sirupsen/logrus/issues/137\n\t\t\tdata[k] = v.Error()\n\t\tdefault:\n\t\t\tdata[k] = v\n\t\t}\n\t}\n\n\tprefixFieldClashes(data, f.FieldMap)\n\n\ttimestampFormat := f.TimestampFormat\n\tif timestampFormat == \"\" {\n\t\ttimestampFormat = defaultTimestampFormat\n\t}\n\n\tif !f.DisableTimestamp {\n\t\tdata[f.FieldMap.resolve(FieldKeyTime)] = entry.Time.Format(timestampFormat)\n\t}\n\tdata[f.FieldMap.resolve(FieldKeyMsg)] = entry.Message\n\tdata[f.FieldMap.resolve(FieldKeyLevel)] = entry.Level.String()\n\n\tvar serialized []byte\n\tvar err error\n\n\tif f.PrettyPrint {\n\t\tserialized, err = json.MarshalIndent(data, \"\", \" \")\n\t} else {\n\t\tserialized, err = json.Marshal(data)\n\t}\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to marshal fields to JSON, %v\", err)\n\t}\n\treturn append(serialized, '\\n'), nil\n}", "func (je *JSONEncoder) EncodeEntry(e *Entry) ([]byte, error) {\n\tb, err := json.Marshal(e)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\treturn b, nil\n}", "func (a *NavigateToHistoryEntryArgs) MarshalJSON() ([]byte, error) {\n\ttype Copy NavigateToHistoryEntryArgs\n\tc := &Copy{}\n\t*c = Copy(*a)\n\treturn json.Marshal(&c)\n}", "func (w *Entry) MarshalJSON() ([]byte, error) {\n\t<<!!YOUR_CODE!!>>\n}", "func (t TrackBase) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tobjectMap[\"@odata.type\"] = t.ODataType\n\treturn json.Marshal(objectMap)\n}", "func (sce StartedChaosEvent) MarshalJSON() ([]byte, error) {\n\tsce.Kind = KindStarted\n\tobjectMap := make(map[string]interface{})\n\tif sce.ChaosParameters != nil {\n\t\tobjectMap[\"ChaosParameters\"] = sce.ChaosParameters\n\t}\n\tif sce.TimeStampUtc != nil {\n\t\tobjectMap[\"TimeStampUtc\"] = sce.TimeStampUtc\n\t}\n\tif sce.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = sce.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (ce ChaosEvent) MarshalJSON() ([]byte, error) {\n\tce.Kind = KindChaosEvent\n\tobjectMap := make(map[string]interface{})\n\tif ce.TimeStampUtc != nil {\n\t\tobjectMap[\"TimeStampUtc\"] = ce.TimeStampUtc\n\t}\n\tif ce.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = ce.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (l LogRules) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"filteringTags\", l.FilteringTags)\n\tpopulate(objectMap, \"sendAadLogs\", l.SendAADLogs)\n\tpopulate(objectMap, \"sendActivityLogs\", l.SendActivityLogs)\n\tpopulate(objectMap, \"sendSubscriptionLogs\", l.SendSubscriptionLogs)\n\treturn json.Marshal(objectMap)\n}", "func (e EventTypeInfo) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"inlineEventTypes\", e.InlineEventTypes)\n\tpopulate(objectMap, \"kind\", e.Kind)\n\treturn json.Marshal(objectMap)\n}", "func (v *Type) MarshalLogObject(enc zapcore.ObjectEncoder) (err error) {\n\tif v == nil {\n\t\treturn nil\n\t}\n\tif v.SimpleType != nil {\n\t\terr = multierr.Append(err, enc.AddObject(\"simpleType\", *v.SimpleType))\n\t}\n\tif v.SliceType != nil {\n\t\terr = multierr.Append(err, enc.AddObject(\"sliceType\", v.SliceType))\n\t}\n\tif v.KeyValueSliceType != nil {\n\t\terr = multierr.Append(err, enc.AddObject(\"keyValueSliceType\", v.KeyValueSliceType))\n\t}\n\tif v.MapType != nil {\n\t\terr = multierr.Append(err, enc.AddObject(\"mapType\", v.MapType))\n\t}\n\tif v.ReferenceType != nil {\n\t\terr = multierr.Append(err, enc.AddObject(\"referenceType\", v.ReferenceType))\n\t}\n\tif v.PointerType != nil {\n\t\terr = multierr.Append(err, enc.AddObject(\"pointerType\", v.PointerType))\n\t}\n\treturn err\n}", "func (f FlowLogFormatParameters) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"type\", f.Type)\n\tpopulate(objectMap, \"version\", f.Version)\n\treturn json.Marshal(objectMap)\n}", "func (c *TimeseriesEntry) MarshalJSON() ([]byte, error) {\n\tvalue, err := c.Value.Int64()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn json.Marshal(struct {\n\t\tTime int64 `json:\"time\"`\n\t\tValue int64 `json:\"value\"`\n\t}{\n\t\tTime: c.Time.Unix(),\n\t\tValue: value,\n\t})\n}", "func (o LogQueryDefinition) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.Compute != nil {\n\t\ttoSerialize[\"compute\"] = o.Compute\n\t}\n\tif o.GroupBy != nil {\n\t\ttoSerialize[\"group_by\"] = o.GroupBy\n\t}\n\tif o.Index != nil {\n\t\ttoSerialize[\"index\"] = o.Index\n\t}\n\tif o.MultiCompute != nil {\n\t\ttoSerialize[\"multi_compute\"] = o.MultiCompute\n\t}\n\tif o.Search != nil {\n\t\ttoSerialize[\"search\"] = o.Search\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (d DashboardPartMetadata) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tobjectMap[\"type\"] = d.Type\n\tif d.AdditionalProperties != nil {\n\t\tfor key, val := range d.AdditionalProperties {\n\t\t\tobjectMap[key] = val\n\t\t}\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (v TimeEntryActivity) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson44273644EncodeGithubComSomniSomGoRedmine1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (fe FabricEvent) MarshalJSON() ([]byte, error) {\n\tfe.Kind = KindFabricEvent\n\tobjectMap := make(map[string]interface{})\n\tif fe.EventInstanceID != nil {\n\t\tobjectMap[\"EventInstanceId\"] = fe.EventInstanceID\n\t}\n\tif fe.TimeStamp != nil {\n\t\tobjectMap[\"TimeStamp\"] = fe.TimeStamp\n\t}\n\tif fe.HasCorrelatedEvents != nil {\n\t\tobjectMap[\"HasCorrelatedEvents\"] = fe.HasCorrelatedEvents\n\t}\n\tif fe.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = fe.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (s SetV2LoggingLevelInput) MarshalFields(e protocol.FieldEncoder) error {\n\te.SetValue(protocol.HeaderTarget, \"Content-Type\", protocol.StringValue(\"application/json\"), protocol.Metadata{})\n\n\tif len(s.LogLevel) > 0 {\n\t\tv := s.LogLevel\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"logLevel\", protocol.QuotedValue{ValueMarshaler: v}, metadata)\n\t}\n\tif s.LogTarget != nil {\n\t\tv := s.LogTarget\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"logTarget\", v, metadata)\n\t}\n\treturn nil\n}", "func (el *ZapEventLogger) Event(ctx context.Context, event string, metadata ...Loggable) {\n\n\t// short circuit if theres nothing to write to\n\tif !writer.WriterGroup.Active() {\n\t\treturn\n\t}\n\n\t// Collect loggables for later logging\n\tvar loggables []Loggable\n\n\t// get any existing metadata from the context\n\texisting, err := MetadataFromContext(ctx)\n\tif err != nil {\n\t\texisting = Metadata{}\n\t}\n\tloggables = append(loggables, existing)\n\tloggables = append(loggables, metadata...)\n\n\te := entry{\n\t\tloggables: loggables,\n\t\tsystem: el.system,\n\t\tevent: event,\n\t}\n\n\taccum := Metadata{}\n\tfor _, loggable := range e.loggables {\n\t\taccum = DeepMerge(accum, loggable.Loggable())\n\t}\n\n\t// apply final attributes to reserved keys\n\t// TODO accum[\"level\"] = level\n\taccum[\"event\"] = e.event\n\taccum[\"system\"] = e.system\n\taccum[\"time\"] = FormatRFC3339(time.Now())\n\n\tvar buf bytes.Buffer\n\tencoder := json.NewEncoder(&buf)\n\tencoder.SetEscapeHTML(false)\n\terr = encoder.Encode(accum)\n\tif err != nil {\n\t\tel.Errorf(\"ERROR FORMATTING EVENT ENTRY: %s\", err)\n\t\treturn\n\t}\n\n\t_, _ = writer.WriterGroup.Write(buf.Bytes())\n}", "func (v ActionLog) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson50c2aa5cEncodeGithubComDoraLogsInternalDjson3(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (le *LogEntry) UnmarshalJSON(data []byte) error {\n\tvar err error\n\ttype Alias struct {\n\t\tLevel *string `json:\"logLevel,omitempty\"`\n\t\tArgs []interface{} `json:\"args,omitempty\"`\n\t\tOriginService *string `json:\"originService,omitempty\"`\n\t\tMessage *string `json:\"message,omitempty\"`\n\t\tCreated int64 `json:\"created,omitempty\"`\n\t}\n\ta := Alias{}\n\t// Error with unmarshaling\n\tif err = json.Unmarshal(data, &a); err != nil {\n\t\treturn err\n\t}\n\n\t// Nillable fields\n\tif a.Level != nil {\n\t\tle.Level = *a.Level\n\t}\n\tif a.OriginService != nil {\n\t\tle.OriginService = *a.OriginService\n\t}\n\tif a.Message != nil {\n\t\tle.Message = *a.Message\n\t}\n\tle.Args = a.Args\n\tle.Created = a.Created\n\n\tle.isValidated, err = le.Validate()\n\n\treturn err\n}", "func (m *AuditLogRoot) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.Entity.Serialize(writer)\n if err != nil {\n return err\n }\n if m.GetDirectoryAudits() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetDirectoryAudits())\n err = writer.WriteCollectionOfObjectValues(\"directoryAudits\", cast)\n if err != nil {\n return err\n }\n }\n if m.GetProvisioning() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetProvisioning())\n err = writer.WriteCollectionOfObjectValues(\"provisioning\", cast)\n if err != nil {\n return err\n }\n }\n if m.GetSignIns() != nil {\n cast := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.CollectionCast[i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable](m.GetSignIns())\n err = writer.WriteCollectionOfObjectValues(\"signIns\", cast)\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (l LineRegistration) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"channelAccessToken\", l.ChannelAccessToken)\n\tpopulate(objectMap, \"channelSecret\", l.ChannelSecret)\n\tpopulate(objectMap, \"generatedId\", l.GeneratedID)\n\treturn json.Marshal(objectMap)\n}", "func (o UsageLogsHour) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.BillableIngestedBytes.IsSet() {\n\t\ttoSerialize[\"billable_ingested_bytes\"] = o.BillableIngestedBytes.Get()\n\t}\n\tif o.Hour != nil {\n\t\tif o.Hour.Nanosecond() == 0 {\n\t\t\ttoSerialize[\"hour\"] = o.Hour.Format(\"2006-01-02T15:04:05Z07:00\")\n\t\t} else {\n\t\t\ttoSerialize[\"hour\"] = o.Hour.Format(\"2006-01-02T15:04:05.000Z07:00\")\n\t\t}\n\t}\n\tif o.IndexedEventsCount.IsSet() {\n\t\ttoSerialize[\"indexed_events_count\"] = o.IndexedEventsCount.Get()\n\t}\n\tif o.IngestedEventsBytes.IsSet() {\n\t\ttoSerialize[\"ingested_events_bytes\"] = o.IngestedEventsBytes.Get()\n\t}\n\tif o.LogsForwardingEventsBytes.IsSet() {\n\t\ttoSerialize[\"logs_forwarding_events_bytes\"] = o.LogsForwardingEventsBytes.Get()\n\t}\n\tif o.LogsLiveIndexedCount.IsSet() {\n\t\ttoSerialize[\"logs_live_indexed_count\"] = o.LogsLiveIndexedCount.Get()\n\t}\n\tif o.LogsLiveIngestedBytes.IsSet() {\n\t\ttoSerialize[\"logs_live_ingested_bytes\"] = o.LogsLiveIngestedBytes.Get()\n\t}\n\tif o.LogsRehydratedIndexedCount.IsSet() {\n\t\ttoSerialize[\"logs_rehydrated_indexed_count\"] = o.LogsRehydratedIndexedCount.Get()\n\t}\n\tif o.LogsRehydratedIngestedBytes.IsSet() {\n\t\ttoSerialize[\"logs_rehydrated_ingested_bytes\"] = o.LogsRehydratedIngestedBytes.Get()\n\t}\n\tif o.OrgName != nil {\n\t\ttoSerialize[\"org_name\"] = o.OrgName\n\t}\n\tif o.PublicId != nil {\n\t\ttoSerialize[\"public_id\"] = o.PublicId\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (info *Info) Marshal() ([]byte, error) {\n\tif info.Attributes[\"perf.elapsed.ms\"] == nil && !info.StartTime.IsZero() {\n\t\tinfo.Attributes[\"perf.elapsed.ms\"] = time.Since(info.StartTime).Nanoseconds() / 1e6\n\t}\n\treturn json.Marshal(info.Attributes)\n}", "func (req *AppendEntriesRequest) Encode(w io.Writer) (int, error) {\n\n\tprotoEntries := make([]*protobuf.ProtoAppendEntriesRequest_ProtoLogEntry, len(req.Entries))\n\n\tfor i, entry := range req.Entries {\n\t\tprotoEntries[i] = &protobuf.ProtoAppendEntriesRequest_ProtoLogEntry{\n\t\t\tIndex: proto.Uint64(entry.Index),\n\t\t\tTerm: proto.Uint64(entry.Term),\n\t\t\tCommandName: proto.String(entry.CommandName),\n\t\t\tCommand: entry.Command,\n\t\t}\n\t}\n\n\tpb := &protobuf.ProtoAppendEntriesRequest{\n\t\tTerm: proto.Uint64(req.Term),\n\t\tPrevLogIndex: proto.Uint64(req.PrevLogIndex),\n\t\tPrevLogTerm: proto.Uint64(req.PrevLogTerm),\n\t\tCommitIndex: proto.Uint64(req.CommitIndex),\n\t\tLeaderName: proto.String(req.LeaderName),\n\t\tEntries: protoEntries,\n\t}\n\n\tp, err := proto.Marshal(pb)\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\n\treturn w.Write(p)\n}", "func (l *LogEntry) ToJSON() {\n\tb, _ := json.Marshal(l)\n\tfmt.Fprintf(os.Stderr, string(b)+\"\\n\")\n}", "func (entry *LogEntry) Encode(w io.Writer, parent string) {\n\tentry.assertValid()\n\twriteUInt8(entry.Flags, w)\n\twriteSequence(entry.Sequence, w)\n\twriteString(entry.DocID, w)\n\twriteString(entry.RevID, w)\n\twriteString(parent, w)\n}", "func (t TestLineUpdate) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"tags\", t.Tags)\n\treturn json.Marshal(objectMap)\n}", "func (f FlowLogListResult) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"nextLink\", f.NextLink)\n\tpopulate(objectMap, \"value\", f.Value)\n\treturn json.Marshal(objectMap)\n}", "func (s Logging) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.AccessLog != nil {\n\t\tv := s.AccessLog\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"accessLog\", v, metadata)\n\t}\n\treturn nil\n}", "func (dcld DiskCopyLogDetails) MarshalJSON() ([]byte, error) {\n\tdcld.CopyLogDetailsType = CopyLogDetailsTypeDataBoxDisk\n\tobjectMap := make(map[string]interface{})\n\tif dcld.DiskSerialNumber != nil {\n\t\tobjectMap[\"diskSerialNumber\"] = dcld.DiskSerialNumber\n\t}\n\tif dcld.ErrorLogLink != nil {\n\t\tobjectMap[\"errorLogLink\"] = dcld.ErrorLogLink\n\t}\n\tif dcld.VerboseLogLink != nil {\n\t\tobjectMap[\"verboseLogLink\"] = dcld.VerboseLogLink\n\t}\n\tif dcld.CopyLogDetailsType != \"\" {\n\t\tobjectMap[\"copyLogDetailsType\"] = dcld.CopyLogDetailsType\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (o EventAttributes) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.AggregationKey != nil {\n\t\ttoSerialize[\"aggregation_key\"] = o.AggregationKey\n\t}\n\tif o.DateHappened != nil {\n\t\ttoSerialize[\"date_happened\"] = o.DateHappened\n\t}\n\tif o.DeviceName != nil {\n\t\ttoSerialize[\"device_name\"] = o.DeviceName\n\t}\n\tif o.Duration != nil {\n\t\ttoSerialize[\"duration\"] = o.Duration\n\t}\n\tif o.EventObject != nil {\n\t\ttoSerialize[\"event_object\"] = o.EventObject\n\t}\n\tif o.Evt != nil {\n\t\ttoSerialize[\"evt\"] = o.Evt\n\t}\n\tif o.Hostname != nil {\n\t\ttoSerialize[\"hostname\"] = o.Hostname\n\t}\n\tif o.Monitor.IsSet() {\n\t\ttoSerialize[\"monitor\"] = o.Monitor.Get()\n\t}\n\tif o.MonitorGroups.IsSet() {\n\t\ttoSerialize[\"monitor_groups\"] = o.MonitorGroups.Get()\n\t}\n\tif o.MonitorId.IsSet() {\n\t\ttoSerialize[\"monitor_id\"] = o.MonitorId.Get()\n\t}\n\tif o.Priority.IsSet() {\n\t\ttoSerialize[\"priority\"] = o.Priority.Get()\n\t}\n\tif o.RelatedEventId != nil {\n\t\ttoSerialize[\"related_event_id\"] = o.RelatedEventId\n\t}\n\tif o.Service != nil {\n\t\ttoSerialize[\"service\"] = o.Service\n\t}\n\tif o.SourceTypeName != nil {\n\t\ttoSerialize[\"source_type_name\"] = o.SourceTypeName\n\t}\n\tif o.Sourcecategory != nil {\n\t\ttoSerialize[\"sourcecategory\"] = o.Sourcecategory\n\t}\n\tif o.Status != nil {\n\t\ttoSerialize[\"status\"] = o.Status\n\t}\n\tif o.Tags != nil {\n\t\ttoSerialize[\"tags\"] = o.Tags\n\t}\n\tif o.Timestamp != nil {\n\t\ttoSerialize[\"timestamp\"] = o.Timestamp\n\t}\n\tif o.Title != nil {\n\t\ttoSerialize[\"title\"] = o.Title\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (v UsersLog) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson50c2aa5cEncodeGithubComDoraLogsInternalDjson(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (wl WaterLog) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(waterLog{T: wl.Time, A: wl.Amount})\n}", "func (t TrackDescriptor) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tobjectMap[\"@odata.type\"] = t.ODataType\n\treturn json.Marshal(objectMap)\n}", "func MarshalTLEToJSON(tle *rekor_pb.TransparencyLogEntry) ([]byte, error) {\n\treturn protojson.Marshal(tle)\n}", "func (entry persistedEntry) Marshal() ([]byte, error) {\n\tif entry.DataLen > modules.RegistryDataSize {\n\t\tbuild.Critical(errTooMuchData)\n\t\treturn nil, errTooMuchData\n\t}\n\tb := make([]byte, PersistedEntrySize)\n\tb[0] = entry.Key.Algorithm\n\tcopy(b[1:], entry.Key.Key[:])\n\tcopy(b[33:], entry.Tweak[:])\n\tbinary.LittleEndian.PutUint32(b[65:], uint32(entry.Expiry))\n\tbinary.LittleEndian.PutUint64(b[69:], uint64(entry.Revision))\n\tcopy(b[77:], entry.Signature[:])\n\tb[141] = byte(entry.DataLen)\n\tcopy(b[142:], entry.Data[:])\n\tb[PersistedEntrySize-1] = uint8(entry.Type)\n\treturn b, nil\n}", "func (meta *EventMetaInfo) MarshalJSON() ([]byte, error) {\n\ttype format struct {\n\t\tConnectionID string `json:\"connectionId,omitempty\"`\n\t}\n\tdata := format{}\n\n\tif meta.ConnectionID != uuid.Nil {\n\t\tdata.ConnectionID = meta.ConnectionID.String()\n\t}\n\n\treturn json.Marshal(data)\n}", "func (e Event) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"count\", e.Count)\n\tpopulateTimeRFC3339(objectMap, \"firstTimestamp\", e.FirstTimestamp)\n\tpopulateTimeRFC3339(objectMap, \"lastTimestamp\", e.LastTimestamp)\n\tpopulate(objectMap, \"message\", e.Message)\n\tpopulate(objectMap, \"name\", e.Name)\n\tpopulate(objectMap, \"type\", e.Type)\n\treturn json.Marshal(objectMap)\n}", "func (o IPAllowlistEntryAttributes) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.CidrBlock != nil {\n\t\ttoSerialize[\"cidr_block\"] = o.CidrBlock\n\t}\n\tif o.CreatedAt != nil {\n\t\tif o.CreatedAt.Nanosecond() == 0 {\n\t\t\ttoSerialize[\"created_at\"] = o.CreatedAt.Format(\"2006-01-02T15:04:05Z07:00\")\n\t\t} else {\n\t\t\ttoSerialize[\"created_at\"] = o.CreatedAt.Format(\"2006-01-02T15:04:05.000Z07:00\")\n\t\t}\n\t}\n\tif o.ModifiedAt != nil {\n\t\tif o.ModifiedAt.Nanosecond() == 0 {\n\t\t\ttoSerialize[\"modified_at\"] = o.ModifiedAt.Format(\"2006-01-02T15:04:05Z07:00\")\n\t\t} else {\n\t\t\ttoSerialize[\"modified_at\"] = o.ModifiedAt.Format(\"2006-01-02T15:04:05.000Z07:00\")\n\t\t}\n\t}\n\tif o.Note != nil {\n\t\ttoSerialize[\"note\"] = o.Note\n\t}\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func materializeWithJSON(logFields []model.KeyValue) ([]byte, error) {\n\tfields := make(map[string]string, len(logFields))\n\tfor i := range logFields {\n\t\tfields[logFields[i].Key] = tagValueToString(&logFields[i])\n\t}\n\tif event, ok := fields[\"event\"]; ok && len(fields) == 1 {\n\t\treturn []byte(event), nil\n\t}\n\treturn json.Marshal(fields)\n}", "func (i Identity)MarshalJSON() ([]byte, error){\n objectMap := make(map[string]interface{})\n if(i.Type != \"\") {\n objectMap[\"type\"] = i.Type\n }\n return json.Marshal(objectMap)\n }", "func (mu MachineUpdateIdentity) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tif mu.Type != nil {\n\t\tobjectMap[\"type\"] = mu.Type\n\t}\n\treturn json.Marshal(objectMap)\n}", "func JSONLog(w io.Writer) LogFunc {\n\treturn func(v interface{}) {\n\t\tdata, err := json.Marshal(v)\n\t\tif err != nil {\n\t\t\tdata, err = json.Marshal(struct {\n\t\t\t\tContext string `json:\"context\"`\n\t\t\t\tDebugData string `json:\"debugData\"`\n\t\t\t\tError string `json:\"error\"`\n\t\t\t}{\n\t\t\t\tContext: \"Error marshaling 'debugData' into JSON\",\n\t\t\t\tDebugData: spew.Sdump(v),\n\t\t\t\tError: err.Error(),\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\t// We really REALLY should never get here\n\t\t\t\tlog.Println(\"ERROR MARSHALLING THE MARSHALLING ERROR!:\", err)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tif _, err := fmt.Fprintf(w, \"%s\\n\", data); err != nil {\n\t\t\tlog.Println(\"ERROR WRITING TO LOGGER:\", err)\n\t\t}\n\t}\n}", "func (curse ClusterUpgradeRollbackStartEvent) MarshalJSON() ([]byte, error) {\n\tcurse.Kind = KindClusterUpgradeRollbackStart\n\tobjectMap := make(map[string]interface{})\n\tif curse.TargetClusterVersion != nil {\n\t\tobjectMap[\"TargetClusterVersion\"] = curse.TargetClusterVersion\n\t}\n\tif curse.FailureReason != nil {\n\t\tobjectMap[\"FailureReason\"] = curse.FailureReason\n\t}\n\tif curse.OverallUpgradeElapsedTimeInMs != nil {\n\t\tobjectMap[\"OverallUpgradeElapsedTimeInMs\"] = curse.OverallUpgradeElapsedTimeInMs\n\t}\n\tif curse.EventInstanceID != nil {\n\t\tobjectMap[\"EventInstanceId\"] = curse.EventInstanceID\n\t}\n\tif curse.TimeStamp != nil {\n\t\tobjectMap[\"TimeStamp\"] = curse.TimeStamp\n\t}\n\tif curse.HasCorrelatedEvents != nil {\n\t\tobjectMap[\"HasCorrelatedEvents\"] = curse.HasCorrelatedEvents\n\t}\n\tif curse.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = curse.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (a Event_Payload) MarshalJSON() ([]byte, error) {\n\tvar err error\n\tobject := make(map[string]json.RawMessage)\n\n\tfor fieldName, field := range a.AdditionalProperties {\n\t\tobject[fieldName], err = json.Marshal(field)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, fmt.Sprintf(\"error marshaling '%s'\", fieldName))\n\t\t}\n\t}\n\treturn json.Marshal(object)\n}", "func (enc *jsonEncoder) AddMarshaler(key string, obj LogMarshaler) error {\n\tenc.addKey(key)\n\tenc.bytes = append(enc.bytes, '{')\n\terr := obj.MarshalLog(enc)\n\tenc.bytes = append(enc.bytes, '}')\n\treturn err\n}", "func (m *MobileAppTroubleshootingEvent) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.Entity.Serialize(writer)\n if err != nil {\n return err\n }\n if m.GetAppLogCollectionRequests() != nil {\n cast := make([]i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, len(m.GetAppLogCollectionRequests()))\n for i, v := range m.GetAppLogCollectionRequests() {\n if v != nil {\n cast[i] = v.(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable)\n }\n }\n err = writer.WriteCollectionOfObjectValues(\"appLogCollectionRequests\", cast)\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (lth *LOBSTERTradingHalt) MarshalJSON() (jsonBytes []byte, err error) {\n\treturn json.Marshal(struct {\n\t\tTheMainEvent LOBSTERTradingHalt `json:\"event\"`\n\t\tEventType Event `json:\"eventtype\"`\n\t}{\n\t\tTheMainEvent: LOBSTERTradingHalt(*lth),\n\t\tEventType: TradingHalt,\n\t})\n}", "func (o LogsPipelineProcessor) MarshalJSON() ([]byte, error) {\n\ttoSerialize := map[string]interface{}{}\n\tif o.UnparsedObject != nil {\n\t\treturn json.Marshal(o.UnparsedObject)\n\t}\n\tif o.Filter != nil {\n\t\ttoSerialize[\"filter\"] = o.Filter\n\t}\n\tif o.IsEnabled != nil {\n\t\ttoSerialize[\"is_enabled\"] = o.IsEnabled\n\t}\n\tif o.Name != nil {\n\t\ttoSerialize[\"name\"] = o.Name\n\t}\n\tif o.Processors != nil {\n\t\ttoSerialize[\"processors\"] = o.Processors\n\t}\n\ttoSerialize[\"type\"] = o.Type\n\n\tfor key, value := range o.AdditionalProperties {\n\t\ttoSerialize[key] = value\n\t}\n\treturn json.Marshal(toSerialize)\n}", "func (event FileEvent) MarshalJSON() ([]byte, error) {\n\ttype surrogate FileEvent\n\tpayload, err := json.Marshal(struct {\n\t\tsurrogate\n\t\tType string `json:\"type\"`\n\t}{\n\t\tsurrogate(event),\n\t\tevent.GetType(),\n\t})\n\treturn payload, errors.JSONMarshalError.Wrap(err)\n}", "func (patdb ProvisionApplicationTypeDescriptionBase) MarshalJSON() ([]byte, error) {\n\tpatdb.Kind = KindProvisionApplicationTypeDescriptionBase\n\tobjectMap := make(map[string]interface{})\n\tif patdb.Async != nil {\n\t\tobjectMap[\"Async\"] = patdb.Async\n\t}\n\tif patdb.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = patdb.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (t Task) MarshalLogObject(enc logging.ObjectEncoder) error {\n\tenc.AddString(\"id\", t.ID)\n\tenc.AddString(\"name\", t.Name)\n\tenc.AddString(\"status\", t.StatusDisplay())\n\tenc.AddString(\"payload\", fmt.Sprintf(\"%v\", t.Payload))\n\tenc.AddInt(\"max_retries\", t.MaxRetries)\n\tenc.AddDuration(\"ttl\", t.TTL)\n\tenc.AddDuration(\"timeout\", t.Timeout)\n\tenc.AddTime(\"published_at\", t.PublishedAt)\n\n\tif !t.StartedAt.IsZero() {\n\t\tenc.AddTime(\"started_at\", t.StartedAt)\n\t}\n\n\tif !t.ProcessedAt.IsZero() {\n\t\tenc.AddTime(\"processed_at\", t.ProcessedAt)\n\t\tenc.AddDuration(\"duration\", t.ProcessedAt.Sub(t.StartedAt))\n\t}\n\n\tif !t.ETA.IsZero() {\n\t\tenc.AddTime(\"eta\", t.ETA)\n\t}\n\n\tif t.ExecTime != 0 {\n\t\tenc.AddFloat64(\"exec_time\", t.ExecTime)\n\t}\n\n\tif len(t.RetryIntervals) > 0 {\n\t\tenc.AddString(\"retry_intervals\", t.RetryIntervalsDisplay())\n\t}\n\n\treturn nil\n}", "func TestJSONEncodeEntry(t *testing.T) {\n\ttype bar struct {\n\t\tKey string `json:\"key\"`\n\t\tVal float64 `json:\"val\"`\n\t}\n\n\ttype foo struct {\n\t\tA string `json:\"aee\"`\n\t\tB int `json:\"bee\"`\n\t\tC float64 `json:\"cee\"`\n\t\tD []bar `json:\"dee\"`\n\t}\n\n\ttests := []struct {\n\t\tdesc string\n\t\texpected string\n\t\tent zapcore.Entry\n\t\tfields []zapcore.Field\n\t}{\n\t\t{\n\t\t\tdesc: \"info entry with some fields\",\n\t\t\texpected: `{\n\t\t\t\t\"L\": \"info\",\n\t\t\t\t\"T\": \"2018-06-19T16:33:42.000Z\",\n\t\t\t\t\"N\": \"bob\",\n\t\t\t\t\"M\": \"lob law\",\n\t\t\t\t\"so\": \"passes\",\n\t\t\t\t\"answer\": 42,\n\t\t\t\t\"a_float32\": 2.71,\n\t\t\t\t\"common_pie\": 3.14,\n\t\t\t\t\"complex_value\": \"3.14-2.71i\",\n\t\t\t\t\"null_value\": null,\n\t\t\t\t\"array_with_null_elements\": [{}, null, null, 2],\n\t\t\t\t\"such\": {\n\t\t\t\t\t\"aee\": \"lol\",\n\t\t\t\t\t\"bee\": 123,\n\t\t\t\t\t\"cee\": 0.9999,\n\t\t\t\t\t\"dee\": [\n\t\t\t\t\t\t{\"key\": \"pi\", \"val\": 3.141592653589793},\n\t\t\t\t\t\t{\"key\": \"tau\", \"val\": 6.283185307179586}\n\t\t\t\t\t]\n\t\t\t\t}\n\t\t\t}`,\n\t\t\tent: zapcore.Entry{\n\t\t\t\tLevel: zapcore.InfoLevel,\n\t\t\t\tTime: time.Date(2018, 6, 19, 16, 33, 42, 99, time.UTC),\n\t\t\t\tLoggerName: \"bob\",\n\t\t\t\tMessage: \"lob law\",\n\t\t\t},\n\t\t\tfields: []zapcore.Field{\n\t\t\t\tzap.String(\"so\", \"passes\"),\n\t\t\t\tzap.Int(\"answer\", 42),\n\t\t\t\tzap.Float64(\"common_pie\", 3.14),\n\t\t\t\tzap.Float32(\"a_float32\", 2.71),\n\t\t\t\tzap.Complex128(\"complex_value\", 3.14-2.71i),\n\t\t\t\t// Cover special-cased handling of nil in AddReflect() and\n\t\t\t\t// AppendReflect(). Note that for the latter, we explicitly test\n\t\t\t\t// correct results for both the nil static interface{} value\n\t\t\t\t// (`nil`), as well as the non-nil interface value with a\n\t\t\t\t// dynamic type and nil value (`(*struct{})(nil)`).\n\t\t\t\tzap.Reflect(\"null_value\", nil),\n\t\t\t\tzap.Reflect(\"array_with_null_elements\", []interface{}{&struct{}{}, nil, (*struct{})(nil), 2}),\n\t\t\t\tzap.Reflect(\"such\", foo{\n\t\t\t\t\tA: \"lol\",\n\t\t\t\t\tB: 123,\n\t\t\t\t\tC: 0.9999,\n\t\t\t\t\tD: []bar{\n\t\t\t\t\t\t{\"pi\", 3.141592653589793},\n\t\t\t\t\t\t{\"tau\", 6.283185307179586},\n\t\t\t\t\t},\n\t\t\t\t}),\n\t\t\t},\n\t\t},\n\t}\n\n\tenc := zapcore.NewJSONEncoder(zapcore.EncoderConfig{\n\t\tMessageKey: \"M\",\n\t\tLevelKey: \"L\",\n\t\tTimeKey: \"T\",\n\t\tNameKey: \"N\",\n\t\tCallerKey: \"C\",\n\t\tFunctionKey: \"F\",\n\t\tStacktraceKey: \"S\",\n\t\tEncodeLevel: zapcore.LowercaseLevelEncoder,\n\t\tEncodeTime: zapcore.ISO8601TimeEncoder,\n\t\tEncodeDuration: zapcore.SecondsDurationEncoder,\n\t\tEncodeCaller: zapcore.ShortCallerEncoder,\n\t})\n\n\tfor _, tt := range tests {\n\t\tt.Run(tt.desc, func(t *testing.T) {\n\t\t\tbuf, err := enc.EncodeEntry(tt.ent, tt.fields)\n\t\t\tif assert.NoError(t, err, \"Unexpected JSON encoding error.\") {\n\t\t\t\tassert.JSONEq(t, tt.expected, buf.String(), \"Incorrect encoded JSON entry.\")\n\t\t\t}\n\t\t\tbuf.Free()\n\t\t})\n\t}\n}", "func (s *sender) logToJSON(record plog.LogRecord) (string, error) {\n\tdata := s.filter.filterOut(record.Attributes())\n\trecord.Body().CopyTo(data.orig.PutEmpty(logKey))\n\n\tnextLine, err := json.Marshal(data.orig.AsRaw())\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn bytes.NewBuffer(nextLine).String(), nil\n}", "func (s ThingTypeMetadata) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.CreationDate != nil {\n\t\tv := *s.CreationDate\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"creationDate\",\n\t\t\tprotocol.TimeValue{V: v, Format: protocol.UnixTimeFormatName, QuotedFormatTime: true}, metadata)\n\t}\n\tif s.Deprecated != nil {\n\t\tv := *s.Deprecated\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"deprecated\", protocol.BoolValue(v), metadata)\n\t}\n\tif s.DeprecationDate != nil {\n\t\tv := *s.DeprecationDate\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"deprecationDate\",\n\t\t\tprotocol.TimeValue{V: v, Format: protocol.UnixTimeFormatName, QuotedFormatTime: true}, metadata)\n\t}\n\treturn nil\n}", "func (e EventGrid) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"accessKey1\", e.AccessKey1)\n\tpopulate(objectMap, \"accessKey2\", e.AccessKey2)\n\tpopulate(objectMap, \"authenticationType\", e.AuthenticationType)\n\tpopulateTimeRFC3339(objectMap, \"createdTime\", e.CreatedTime)\n\tpopulate(objectMap, \"deadLetterSecret\", e.DeadLetterSecret)\n\tpopulate(objectMap, \"deadLetterUri\", e.DeadLetterURI)\n\tobjectMap[\"endpointType\"] = EndpointTypeEventGrid\n\tpopulate(objectMap, \"provisioningState\", e.ProvisioningState)\n\tpopulate(objectMap, \"TopicEndpoint\", e.TopicEndpoint)\n\treturn json.Marshal(objectMap)\n}", "func (s SyncIdentityProviderUpdate) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"properties\", s.Properties)\n\tpopulate(objectMap, \"systemData\", s.SystemData)\n\treturn json.Marshal(objectMap)\n}", "func (efce ExecutingFaultsChaosEvent) MarshalJSON() ([]byte, error) {\n\tefce.Kind = KindExecutingFaults\n\tobjectMap := make(map[string]interface{})\n\tif efce.Faults != nil {\n\t\tobjectMap[\"Faults\"] = efce.Faults\n\t}\n\tif efce.TimeStampUtc != nil {\n\t\tobjectMap[\"TimeStampUtc\"] = efce.TimeStampUtc\n\t}\n\tif efce.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = efce.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (i Identifiable) MarshalJSON() ([]byte, error) {\n\ti.Type = TypeIdentifiable\n\tobjectMap := make(map[string]interface{})\n\tif i.ID != nil {\n\t\tobjectMap[\"id\"] = i.ID\n\t}\n\tif i.Type != \"\" {\n\t\tobjectMap[\"_type\"] = i.Type\n\t}\n\treturn json.Marshal(objectMap)\n}", "func getWrappedLogMessagesOfType(t *testing.T, entityName, entityVersion, typ string, logOutput []byte) [][]byte {\n\tlines := bytes.Split(logOutput, []byte(\"\\n\"))\n\tvar logLines [][]byte\n\tfor _, line := range lines {\n\t\tif len(line) == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tvar currEntry map[string]interface{}\n\t\tassert.NoError(t, json.Unmarshal(line, &currEntry), \"failed to parse json line %q\", string(line))\n\t\tassert.Equal(t, \"wrapped.1\", currEntry[\"type\"])\n\t\tassert.Equal(t, entityName, currEntry[\"entityName\"])\n\t\tassert.Equal(t, entityVersion, currEntry[\"entityVersion\"])\n\t\tif payload, ok := currEntry[\"payload\"].(map[string]interface{}); ok {\n\t\t\tif payloadLog, ok := payload[payload[\"type\"].(string)].(map[string]interface{}); ok {\n\t\t\t\tif payloadLogLineType, ok := payloadLog[\"type\"]; ok && payloadLogLineType == typ {\n\t\t\t\t\tif payloadBytes, err := json.Marshal(payloadLog); err == nil {\n\t\t\t\t\t\tlogLines = append(logLines, payloadBytes)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn logLines\n}", "func (t ThirdPartyAuthDetails) MarshalJSON() ([]byte, error) {\n\ttype Alias ThirdPartyAuthDetails\n\treturn json.Marshal(\n\t\tstruct {\n\t\t\tmeta.TypeMeta `json:\",inline\"`\n\t\t\tAlias `json:\",inline\"`\n\t\t}{\n\t\t\tTypeMeta: meta.TypeMeta{\n\t\t\t\tAPIVersion: meta.APIVersion,\n\t\t\t\tKind: \"ThirdPartyAuthDetails\",\n\t\t\t},\n\t\t\tAlias: (Alias)(t),\n\t\t},\n\t)\n}", "func Encode(lr plog.LogRecord, res pcommon.Resource, scope pcommon.InstrumentationScope) (string, error) {\n\tvar logRecord lokiEntry\n\tvar jsonRecord []byte\n\tvar err error\n\tvar body []byte\n\n\tbody, err = serializeBodyJSON(lr.Body())\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tlogRecord = lokiEntry{\n\t\tBody: body,\n\t\tTraceID: traceutil.TraceIDToHexOrEmptyString(lr.TraceID()),\n\t\tSpanID: traceutil.SpanIDToHexOrEmptyString(lr.SpanID()),\n\t\tSeverity: lr.SeverityText(),\n\t\tAttributes: lr.Attributes().AsRaw(),\n\t\tResources: res.Attributes().AsRaw(),\n\t\tFlags: uint32(lr.Flags()),\n\t}\n\n\tscopeName := scope.Name()\n\tif scopeName != \"\" {\n\t\tlogRecord.InstrumentationScope = &instrumentationScope{\n\t\t\tName: scopeName,\n\t\t}\n\t\tlogRecord.InstrumentationScope.Version = scope.Version()\n\t\tlogRecord.InstrumentationScope.Attributes = scope.Attributes().AsRaw()\n\t}\n\n\tjsonRecord, err = json.Marshal(logRecord)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn string(jsonRecord), nil\n}", "func (s EgressFilter) MarshalFields(e protocol.FieldEncoder) error {\n\tif len(s.Type) > 0 {\n\t\tv := s.Type\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"type\", protocol.QuotedValue{ValueMarshaler: v}, metadata)\n\t}\n\treturn nil\n}", "func (e *Entry) Encode(w msg.Writer) error {\n\tmsg.WriteUint(w, e.Timestamp())\n\tmsg.WriteInt(w, e.Level)\n\tmsg.WriteString(w, e.Message)\n\treturn nil\n}", "func (l LineChannelProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"callbackUrl\", l.CallbackURL)\n\tpopulate(objectMap, \"isValidated\", l.IsValidated)\n\tpopulate(objectMap, \"lineRegistrations\", l.LineRegistrations)\n\treturn json.Marshal(objectMap)\n}", "func (pbi PropertyBatchInfo) MarshalJSON() ([]byte, error) {\n\tpbi.Kind = KindPropertyBatchInfo\n\tobjectMap := make(map[string]interface{})\n\tif pbi.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = pbi.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (s VirtualGatewayLogging) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.AccessLog != nil {\n\t\tv := s.AccessLog\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"accessLog\", v, metadata)\n\t}\n\treturn nil\n}" ]
[ "0.5876557", "0.5784184", "0.5697098", "0.56902766", "0.56712556", "0.5645131", "0.56377184", "0.56377184", "0.56044203", "0.5602046", "0.5528011", "0.5516618", "0.5474974", "0.54499847", "0.5445715", "0.5423863", "0.53563595", "0.53111035", "0.5310348", "0.5281776", "0.52635473", "0.5263436", "0.52624196", "0.52536315", "0.52343404", "0.5201463", "0.5168026", "0.51597923", "0.5117289", "0.51009303", "0.5096016", "0.5088021", "0.5084152", "0.5075666", "0.5068229", "0.5058374", "0.50520647", "0.50378287", "0.50362074", "0.50284445", "0.50111234", "0.50095564", "0.500832", "0.4999439", "0.49970138", "0.4996057", "0.49724323", "0.4965821", "0.49550617", "0.4931578", "0.4931122", "0.49232206", "0.49216518", "0.4906307", "0.4893983", "0.48813766", "0.48809448", "0.48799407", "0.4870629", "0.48699185", "0.48656812", "0.48637724", "0.48512557", "0.4848126", "0.48191464", "0.48163408", "0.48127812", "0.48113087", "0.48047578", "0.4804203", "0.4801566", "0.47953242", "0.47944862", "0.4793691", "0.47931826", "0.4792479", "0.4791584", "0.47911587", "0.47898492", "0.47765416", "0.47757947", "0.47748467", "0.47655755", "0.47607887", "0.47555295", "0.47545612", "0.47489664", "0.47279173", "0.47189462", "0.47181952", "0.4714572", "0.47129962", "0.47064674", "0.4706223", "0.47062212", "0.47018752", "0.46989563", "0.46845162", "0.46836597" ]
0.70293987
1
NewLogsService returns a specialized interface for accessing logs.
func NewLogsService( authorize AuthorizeFn, projectAuthorize ProjectAuthorizeFn, projectsStore ProjectsStore, eventsStore EventsStore, warmLogsStore LogsStore, coolLogsStore LogsStore, ) LogsService { return &logsService{ authorize: authorize, projectAuthorize: projectAuthorize, projectsStore: projectsStore, eventsStore: eventsStore, warmLogsStore: warmLogsStore, coolLogsStore: coolLogsStore, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func New(svc role.Service, logger schemago.Logger) *LogService {\n\treturn &LogService{\n\t\tService: svc,\n\t\tlogger: logger,\n\t}\n}", "func New(svc analysis.Service, logger log.Logger) *LogService {\n\treturn &LogService{\n\t\tService: svc,\n\t\tlogger: logger,\n\t}\n}", "func New(svc topic.Service, logger model.Logger) *LogService {\n\treturn &LogService{\n\t\tService: svc,\n\t\tlogger: logger,\n\t}\n}", "func New(svc event.Service, logger takrib.Logger) *LogService {\n\treturn &LogService{\n\t\tService: svc,\n\t\tlogger: logger,\n\t}\n}", "func New(service string) (*Logger, error) {\n\n\tvar logger *log.Logger\n\tvar file *os.File\n\n\tif service != \"\" {\n\t\tfilename := fmt.Sprintf(\"%s_%s.log\", service, time.Now().Format(\"2006_01_02_15_04_05\"))\n\t\tf, err := os.Create(filename)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tlogger = log.New(f, \"\", log.LstdFlags)\n\t\tfile = f\n\t} else {\n\t\tlogger = log.New(os.Stdout, \"\", log.LstdFlags)\n\t}\n\n\tl := new(Logger)\n\tl.logger = logger\n\tl.file = file\n\n\treturn l, nil\n}", "func NewLoggingService(s Service) Service {\n\treturn &loggingService{s, \"cloudsource\"}\n}", "func New(serviceName string, environment string) LoggerWrapper {\n\tlogStore = &loggerWrapper{logrus.New().WithField(\"service\", serviceName).WithField(\"environment\", environment)}\n\tif environment == \"production\" {\n\t\tlogStore.SetFormat(&logrus.JSONFormatter{})\n\t}\n\n\t// fmt.Println(\"Adding hook\")\n\t// hook := logrusly.NewLogglyHook(\"71000042-f956-4c7e-987d-8694a20695a8\", \"https://logs-01.loggly.com/bulk/\", logrus.InfoLevel, serviceName)\n\t// logStore.Logger.Hooks.Add(hook)\n\treturn logStore\n}", "func New(accessToken string, logger *logrus.Logger) Service {\n\treturn &service{accessToken, logger}\n}", "func NewLoggingService(log *log.Logger, s Service) Service {\n\treturn &loggingService{log, s}\n}", "func NewLogsClient(credential azcore.TokenCredential, options *LogsClientOptions) (*LogsClient, error) {\n\tif options == nil {\n\t\toptions = &LogsClientOptions{}\n\t}\n\tif reflect.ValueOf(options.Cloud).IsZero() {\n\t\toptions.Cloud = cloud.AzurePublic\n\t}\n\tc, ok := options.Cloud.Services[ServiceNameLogs]\n\tif !ok || c.Audience == \"\" || c.Endpoint == \"\" {\n\t\treturn nil, errors.New(\"provided Cloud field is missing Azure Monitor Logs configuration\")\n\t}\n\n\tauthPolicy := runtime.NewBearerTokenPolicy(credential, []string{c.Audience + \"/.default\"}, nil)\n\tpl := runtime.NewPipeline(moduleName, version, runtime.PipelineOptions{PerRetry: []policy.Policy{authPolicy}}, &options.ClientOptions)\n\treturn &LogsClient{host: c.Endpoint, pl: pl}, nil\n}", "func NewLoggingService(next Service) Service {\n\treturn &loggingService{\n\t\tnext: next,\n\t}\n}", "func NewLog(svc string) *logrus.Entry {\n\tlog := logrus.New()\n\n\tlog.SetLevel(logrus.DebugLevel)\n\tlog.SetFormatter(&logrus.JSONFormatter{})\n\n\tl := log.WithFields(logrus.Fields{\n\t\t\"service-name\": svc,\n\t})\n\n\treturn l\n}", "func New(logEnable bool, logServiceType int, config *Fluent) ILog {\n\tif !logEnable {\n\t\tlogServiceType = DEFAULT\n\t}\n\n\tswitch logServiceType {\n\tcase FLUENT:\n\t\treturn NewFluent(config)\n\tdefault:\n\t\treturn Default()\n\t}\n\n\treturn nil\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger: logger, Service: s}\n}", "func newLogger() *ServiceLogger {\n\tLogger := log.New()\n\tvar serviceLogger ServiceLogger\n\t// Log as JSON instead of the default ASCII formatter.\n\tLogger.SetFormatter(&log.JSONFormatter{})\n\n\t// Output to stdout instead of the default stderr\n\tLogger.SetOutput(os.Stdout)\n\n\t// Only log the warning severity or above.\n\tLogger.SetLevel(log.InfoLevel)\n\n\tserviceLogger.Logger = Logger\n\n\treturn &serviceLogger\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger: logger, service: s}\n}", "func NewLoggingService(logger log.Logger, service definitions.Service) definitions.Service {\n\treturn logging.Service(logger, service)\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func NewLoggingService(logger log.Logger, s Service) Service {\n\treturn &loggingService{logger, s}\n}", "func (e Engine) newLogs(config dvid.StoreConfig) (*fileLogs, bool, error) {\n\tpath, _, err := parseConfig(config)\n\tif err != nil {\n\t\treturn nil, false, err\n\t}\n\n\tvar created bool\n\tif _, err := os.Stat(path); os.IsNotExist(err) {\n\t\tdvid.Infof(\"Log not already at path (%s). Creating ...\\n\", path)\n\t\tif err := os.MkdirAll(path, 0755); err != nil {\n\t\t\treturn nil, false, err\n\t\t}\n\t\tcreated = true\n\t} else {\n\t\tdvid.Infof(\"Found log at %s (err = %v)\\n\", path, err)\n\t}\n\n\t// opt, err := getOptions(config.Config)\n\t// if err != nil {\n\t// \treturn nil, false, err\n\t// }\n\n\tlog := &fileLogs{\n\t\tpath: path,\n\t\tconfig: config,\n\t\tfiles: make(map[string]*fileLog),\n\t}\n\treturn log, created, nil\n}", "func NewLogger() *LoggerService {\n\tfullFileName := getFullFileName(\"default.log\")\n\n\tfile, err := os.OpenFile(fullFileName, os.O_RDWR|os.O_CREATE, 0755)\n\tif err != nil {\n\t\tfmt.Println(\"Error opening file:\", err)\n\t\tos.Exit(1)\n\t}\n\n\treturn &LoggerService{file}\n}", "func NewLogs(finding *guardduty.Finding) (logs []LogEntry, err error) {\n\tswitch aws.StringValue(finding.Service.Action.ActionType) {\n\t// if this is a port-probe, it can have multiple events in a single record, so we add a new log for each.\n\tcase \"PORT_PROBE\":\n\t\tfor _, probe := range finding.Service.Action.PortProbeAction.PortProbeDetails {\n\t\t\tl := LogEntry{}\n\t\t\tl.addCommon(finding)\n\t\t\tl.addInstance(finding.Resource.InstanceDetails)\n\t\t\tl.addCommonService(finding.Service)\n\t\t\tl.addPortProbe(finding.Service.Action.PortProbeAction, probe)\n\t\t\tlogs = append(logs, l)\n\t\t}\n\tcase \"NETWORK_CONNECTION\":\n\t\tl := LogEntry{}\n\t\tl.addCommon(finding)\n\t\tif finding.Resource != nil && finding.Resource.InstanceDetails != nil {\n\t\t\tl.addInstance(finding.Resource.InstanceDetails)\n\t\t}\n\t\tif finding.Service != nil {\n\t\t\tl.addCommonService(finding.Service)\n\t\t}\n\t\tl.addConnection(finding.Service.Action.NetworkConnectionAction)\n\t\tlogs = append(logs, l)\n\tcase \"DNS_REQUEST\":\n\t\tl := LogEntry{}\n\t\tl.addCommon(finding)\n\t\tif finding.Resource != nil && finding.Resource.InstanceDetails != nil {\n\t\t\tl.addInstance(finding.Resource.InstanceDetails)\n\t\t}\n\t\tif finding.Service != nil {\n\t\t\tl.addCommonService(finding.Service)\n\t\t}\n\t\tl.addDns(finding.Service.Action.DnsRequestAction)\n\t\tlogs = append(logs, l)\n\tcase \"AWS_API_CALL\":\n\t\tl := LogEntry{}\n\t\tl.addCommon(finding)\n\t\tif finding.Service != nil {\n\t\t\tl.addCommonService(finding.Service)\n\t\t}\n\t\tl.addApiCall(finding.Service.Action.AwsApiCallAction)\n\t\tlogs = append(logs, l)\n\t}\n\treturn\n}", "func NewSrvLog() *SrvLog {\n\tobj := &SrvLog{\n\t\tsubs: make(map[string]*LogService),\n\t\tfields: make(map[string]interface{}),\n\t}\n\n\tobj.AddLog(\"debug\", logBuffer)\n\tobj.AddLog(\"error\", logBuffer)\n\tobj.AddLog(\"info\", logBuffer)\n\n\tfileLog := flog.New(\"all\", logDir, flog.BufferedFile)\n\n\tobj.AddLog(\"debug\", fileLog)\n\tobj.AddLog(\"error\", fileLog)\n\tobj.AddLog(\"info\", fileLog)\n\n\treturn obj\n}", "func NewLogsStore(kubeClient kubernetes.Interface) api.LogsStore {\n\treturn &logsStore{\n\t\tkubeClient: kubeClient,\n\t}\n}", "func New(l lg.Logger, s Store) Service {\n\treturn Service{\n\t\tlogger: l,\n\t\tstore: s,\n\t}\n}", "func NewService(store storage.Store) *Service {\n\treturn &Service{\n\t\tstore: store,\n\t\tworkerPool: make(chan LogChannel, maxWorkers),\n\t}\n}", "func NewService(repo Repository, logger log.Logger) Service {\n\treturn service{repo, logger}\n}", "func NewService(repo Repository, logger log.Logger) Service {\n\treturn service{repo, logger}\n}", "func NewService(repo Repository, logger log.Logger) Service {\n\treturn service{repo, logger}\n}", "func NewLoggingService(ctx context.Context, s Service) Service {\n\tlogger := logging.FromContext(ctx)\n\tlogger = log.With(logger, \"component\", \"service\")\n\treturn &loggingService{logger, s}\n}", "func NewService(repo Repo, log logger.ContextLog) *Service {\n\treturn &Service{\n\t\trepo: repo,\n\t\tlog: log,\n\t}\n}", "func NewLogsClient(cc *grpc.ClientConn) LogsClient {\n\treturn &logsClient{rawClient: otlpcollectorlog.NewLogsServiceClient(cc)}\n}", "func LogNew(level, event, msg string) {\n\tvar thelog StructuredLog\n\tthelog.Timestamp = time.Now().Format(time.RFC3339)\n\thostname, _ := os.Hostname()\n\tthelog.Server = hostname\n\tthelog.Level = level\n\tthelog.Event = event\n\tthelog.Message = msg\n\tthelog.Service = \"search-api\"\n\tlogJSON, err := json.Marshal(thelog)\n\tif err != nil {\n\t\tlog.Println(\"Structured logger: Logger JSON Marshal failed !\")\n\t}\n\tlog.Println(string(logJSON))\n}", "func NewService(config *ServiceConfig, configExt ...*ServiceConfigExt) (*Service, error) {\n\tvar serviceLogWriter, accessLogWriter io.Writer\n\tswitch len(configExt) {\n\tcase 0:\n\tcase 1:\n\t\tserviceLogWriter = configExt[0].ServiceLogWriter\n\t\taccessLogWriter = configExt[0].AccessLogWriter\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Only single instance of ServiceConfigExt allowed atmost\")\n\t}\n\n\t// Normalize and validate config\n\tconfig.Normalize()\n\tif err := config.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create the service\n\ts := &Service{\n\t\tconfig: config,\n\t\tdefaultLogger: &log.Logger{Formatter: new(log.JSONFormatter)},\n\t\tapiRouters: make(map[int]*httptreemux.ContextMux, config.Version.Max-config.Version.Min+1),\n\t}\n\ts.globalRouter = s.newRouter()\n\tfor v := config.Version.Min; v <= config.Version.Max; v++ {\n\t\ts.apiRouters[v] = s.newRouter()\n\t}\n\n\t// Configure logging\n\tif serviceLogWriter != nil {\n\t\ts.defaultLogger.Out = serviceLogWriter\n\t} else if config.Log.ServiceLogPath != \"\" {\n\t\t// Service log to file\n\t\topenLogFile(s.defaultLogger, config.Log.ServiceLogPath)\n\t} else {\n\t\t// Service log to stdout\n\t\ts.defaultLogger.Out = os.Stdout\n\t}\n\n\tswitch strings.ToLower(config.Log.ServiceLogLevel) {\n\tcase \"debug\":\n\t\ts.defaultLogger.SetLevel(log.DebugLevel)\n\tdefault:\n\t\tfallthrough\n\tcase \"info\":\n\t\ts.defaultLogger.SetLevel(log.InfoLevel)\n\tcase \"warn\":\n\t\ts.defaultLogger.SetLevel(log.WarnLevel)\n\tcase \"error\":\n\t\ts.defaultLogger.SetLevel(log.ErrorLevel)\n\t}\n\n\ts.accessLogger = &log.Logger{\n\t\tFormatter: new(log.JSONFormatter),\n\t\tLevel: log.InfoLevel,\n\t}\n\tif accessLogWriter != nil {\n\t\ts.accessLogger.Out = accessLogWriter\n\t} else if config.Log.AccessLogPath != \"\" {\n\t\t// Access log to file\n\t\topenLogFile(s.accessLogger, config.Log.AccessLogPath)\n\t} else if config.Log.ServiceLogPath != \"\" {\n\t\t// Access log to stdout\n\t\ts.accessLogger.Out = os.Stdout\n\t} else {\n\t\t// Both service log and access log to stdout (sharing a logger)\n\t\ts.accessLogger = s.defaultLogger\n\t}\n\n\t// Configure tracing\n\ts.tracerKind, s.tracer = s.newTracer()\n\n\t// Dump goroutine stacks on demand\n\tdumpGoroutineStacks()\n\n\t// Configure CORS\n\tif config.CORS.Enabled {\n\t\topts := cors.Options{\n\t\t\tAllowedOrigins: config.CORS.AllowedOrigins,\n\t\t\tAllowedMethods: config.CORS.AllowedMethods,\n\t\t\tAllowedHeaders: config.CORS.AllowedHeaders,\n\t\t\tExposedHeaders: config.CORS.ExposedHeaders,\n\t\t\tAllowCredentials: config.CORS.AllowCredentials,\n\t\t}\n\t\ts.cors = cors.New(opts)\n\t}\n\n\t// Create the default schema filesystem\n\tif config.Schema.Enabled {\n\t\ts.schemas = http.Dir(config.Schema.FilePath)\n\t}\n\n\t// Add default middleware handlers, beginning with \"bottom\"\n\ts.AddHandler(&bottomHandler{\n\t\ts: s,\n\t\tdefaultLogger: s.defaultLogger,\n\t\taccessLogger: s.accessLogger,\n\t\ttracerKind: s.tracerKind,\n\t\ttracer: s.tracer,\n\t\tcors: s.cors,\n\t})\n\n\ts.AddHandler(new(negotiatorHandler))\n\n\ts.AddHandler(&versionHandler{\n\t\tminVersion: s.config.Version.Min,\n\t\tmaxVersion: s.config.Version.Max,\n\t})\n\n\treturn s, nil\n}", "func New(debug bool, errorChannel string) *Logs {\n\tlogrus.SetOutput(os.Stdout)\n\tprimary := logrus.New()\n\n\tif debug {\n\t\tlogrus.SetLevel(logrus.DebugLevel)\n\t\tprimary.SetFormatter(&logrus.TextFormatter{\n\t\t\tForceColors: true,\n\t\t})\n\t} else {\n\t\tlogrus.SetLevel(logrus.InfoLevel)\n\t\tprimary.SetFormatter(&logrus.JSONFormatter{})\n\n\t}\n\n\treturn &Logs{\n\t\tPrimary: primary,\n\t\tCommand: primary.WithField(\"type\", \"command\"),\n\t\tMultiplexer: primary.WithField(\"type\", \"multiplexer\"),\n\t\tdebug: debug,\n\t\terrorChannel: errorChannel,\n\t}\n}", "func Service(logger log.Logger, service definitions.Service) definitions.Service {\n\treturn &loggingService{\n\t\tService: service,\n\t\tLogger: logger,\n\t}\n}", "func newLogsCommand(dockerCli *command.DockerCli) *cobra.Command {\n\tvar opts logsOptions\n\n\tcmd := &cobra.Command{\n\t\tUse: \"logs [OPTIONS] SERVICE|TASK\",\n\t\tShort: \"Fetch the logs of a service or task\",\n\t\tArgs: cli.ExactArgs(1),\n\t\tRunE: func(cmd *cobra.Command, args []string) error {\n\t\t\topts.target = args[0]\n\t\t\treturn runLogs(dockerCli, &opts)\n\t\t},\n\t\tTags: map[string]string{\"version\": \"1.29\"},\n\t}\n\n\tflags := cmd.Flags()\n\t// options specific to service logs\n\tflags.BoolVar(&opts.noResolve, \"no-resolve\", false, \"Do not map IDs to Names in output\")\n\tflags.BoolVar(&opts.noTrunc, \"no-trunc\", false, \"Do not truncate output\")\n\tflags.BoolVar(&opts.noTaskIDs, \"no-task-ids\", false, \"Do not include task IDs in output\")\n\t// options identical to container logs\n\tflags.BoolVarP(&opts.follow, \"follow\", \"f\", false, \"Follow log output\")\n\tflags.StringVar(&opts.since, \"since\", \"\", \"Show logs since timestamp (e.g. 2013-01-02T13:23:37) or relative (e.g. 42m for 42 minutes)\")\n\tflags.BoolVarP(&opts.timestamps, \"timestamps\", \"t\", false, \"Show timestamps\")\n\tflags.StringVar(&opts.tail, \"tail\", \"all\", \"Number of lines to show from the end of the logs\")\n\treturn cmd\n}", "func New(fs fs, repo repo, logger log.FieldLogger, metrics counters, tracer opentracing.Tracer) *Service {\n\treturn &Service{\n\t\tfs,\n\t\trepo,\n\t\tlogger,\n\t\ttracer,\n\t\tmetrics,\n\t}\n}", "func NewService(cfg *config.Config) (Service, error) {\r\n\tlogLevel, err := log.ParseLevel(cfg.Log.Level)\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"parse log level err - %v\", err)\r\n\t}\r\n\tlog.SetLevel(logLevel)\r\n\r\n\tf, err := os.OpenFile(cfg.Log.File, os.O_APPEND|os.O_CREATE|os.O_RDWR, 0666)\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"init log file \\\"%s\\\" err - %v\", cfg.Log.File, err)\r\n\t}\r\n\tlog.SetOutput(io.MultiWriter(os.Stdout, f))\r\n\r\n\tparser, err := parser.NewParser(cfg.Syslog.Templates)\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"init parser err - %v\", err)\r\n\t}\r\n\tlsn, err := syslog.NewListener(cfg.Syslog.Listen, cfg.Syslog.BufSize, parser)\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"init syslog listener err - %v\", err)\r\n\t}\r\n\r\n\tconn, err := net.Listen(\"tcp\", cfg.GRPC.Listen)\r\n\tif err != nil {\r\n\t\treturn nil, fmt.Errorf(\"init grpc conn err - %v\", err)\r\n\t}\r\n\tlog.Debugf(\"listen grpc requests on %s\", cfg.GRPC.Listen)\r\n\r\n\ts := &service{\r\n\t\tserver: grpc.NewServer(),\r\n\t\tconn: conn,\r\n\t\tlistener: lsn,\r\n\t\tsubsMu: sync.Mutex{},\r\n\t\tsubscribers: make(map[string]*subscriber),\r\n\t\tclosed: make(chan struct{}),\r\n\t}\r\n\r\n\tpb.RegisterSyslogCatcherServer(s.server, s)\r\n\r\n\treturn s, nil\r\n}", "func New(s *session.Session) *CloudWatchLogs {\n\treturn &CloudWatchLogs{\n\t\tclient: cloudwatchlogs.New(s),\n\t}\n}", "func New(debug bool) *Logs {\n\tlogrus.SetOutput(os.Stdout)\n\tprimary := logrus.New()\n\n\tif debug {\n\t\tlogrus.SetLevel(logrus.DebugLevel)\n\t\tprimary.SetFormatter(&logrus.TextFormatter{\n\t\t\tForceColors: true,\n\t\t})\n\t} else {\n\t\tlogrus.SetLevel(logrus.InfoLevel)\n\t\tprimary.SetFormatter(&logrus.JSONFormatter{})\n\n\t}\n\n\treturn &Logs{\n\t\tPrimary: primary,\n\t\tCommand: primary.WithField(\"type\", \"command\"),\n\t\tMultiplexer: primary.WithField(\"type\", \"multiplexer\"),\n\t\tdebug: debug,\n\t}\n}", "func NewService(signingKey string, tokenExpiration int, logger log.Logger) Service {\n\treturn service{signingKey, tokenExpiration, logger}\n}", "func (c *ClusterClient) Logs() *LogsClient {\n\treturn NewLogsClient(\n\t\tc.transport,\n\t\tpath.Join(c.path, \"logs\"),\n\t)\n}", "func (a *Accessor) Logs(namespace string, pod string, container string, previousLog bool) (string, error) {\n\treturn a.ctl.logs(namespace, pod, container, previousLog)\n}", "func NewLogs(ctx *Context) (*Logs, error) {\n\tcfg, err := ConfigGetAll(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t//pretty.Println(cfg)\n\tif cfg.PubNub.PubKey != \"\" && cfg.PubNub.SubKey != \"\" {\n\t\tn := messaging.NewPubnub(\n\t\t\tcfg.PubNub.PubKey,\n\t\t\tcfg.PubNub.SubKey, \"\", \"\", false, \"\", nil)\n\t\treturn &Logs{nub: n, ctx: ctx, stop: make(chan struct{})}, nil\n\t}\n\treturn nil, errors.New(\"resingo: no pubnub details found\")\n}", "func newLog() {\n\tlogBuffer = log.NewLogBuffer(DefaultHttpLogBuffers)\n\tsrvLog = NewSrvLog()\n\n\ttrace.DebugLogger = srvLog\n\ttrace.ErrorLogger = srvLog\n}", "func New(nc *nats.Conn, logger log.Logger) (s AddService) {\n\tvar svc AddService\n\t{\n\t\tsvc = &stubAddService{nc: nc, logger: logger}\n\t\tsvc = LoggingMiddleware(logger)(svc)\n\t}\n\treturn svc\n}", "func New(output io.Writer, level LogLevel) *Logger {\n\tvar (\n\t\tsystemdInvocation bool\n\t\tawsLogGroup bool\n\t\tflags int\n\t)\n\t// Detect special compatibility modes\n\t_, systemdInvocation = os.LookupEnv(\"INVOCATION_ID\")\n\t_, awsLogGroup = os.LookupEnv(\"AWS_LAMBDA_LOG_GROUP_NAME\")\n\t// Prepare configuration accordingly\n\tif systemdInvocation && output != os.Stdout && output != os.Stderr {\n\t\t// launched by systemd but logger is not being redirected to std output (may be to a file ?)\n\t\t// disabling journald compat mode\n\t\tsystemdInvocation = false\n\t}\n\tif !systemdInvocation && !awsLogGroup {\n\t\tflags = log.Ltime | log.Ldate\n\t}\n\t// Return the initialized logger\n\treturn &Logger{\n\t\tjournald: systemdInvocation,\n\t\tllevel: level,\n\t\tlogger: log.New(output, \"\", flags),\n\t}\n}", "func newLogger() services.Logger {\n\treturn &services.SystemOutLogger{}\n}", "func GetLogs(client DockerClient, containerId string, since int64, stdout io.Writer, stderr io.Writer) {\n\tgo func() {\n\t\terr := client.Logs(docker.LogsOptions{\n\t\t\tContainer: containerId,\n\t\t\tOutputStream: stdout,\n\t\t\tErrorStream: stderr,\n\t\t\tSince: since,\n\t\t\tStdout: true,\n\t\t\tStderr: true,\n\t\t})\n\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to fetch logs for task: %s\", err.Error())\n\t\t}\n\t}()\n}", "func newLogsProcessor(ctx context.Context, set processor.CreateSettings, nextConsumer consumer.Logs, cfg *Config) (processor.Logs, error) {\n\n\tlsp := &logSamplerProcessor{\n\t\tscaledSamplingRate: uint32(cfg.SamplingPercentage * percentageScaleFactor),\n\t\thashSeed: cfg.HashSeed,\n\t\ttraceIDEnabled: cfg.AttributeSource == traceIDAttributeSource,\n\t\tsamplingPriority: cfg.SamplingPriority,\n\t\tsamplingSource: cfg.FromAttribute,\n\t\tlogger: set.Logger,\n\t}\n\n\treturn processorhelper.NewLogsProcessor(\n\t\tctx,\n\t\tset,\n\t\tcfg,\n\t\tnextConsumer,\n\t\tlsp.processLogs,\n\t\tprocessorhelper.WithCapabilities(consumer.Capabilities{MutatesData: true}))\n}", "func (c *kubernetesDeploymentManifest) Logs(service ServiceRequest) error {\n\tkubectl = cluster.Kubectl().WithNamespace(c.Context, \"default\")\n\t_, err := kubectl.Run(c.Context, \"logs\", \"deployment/\"+service.Name)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"error\": err,\n\t\t\t\"service\": service.Name,\n\t\t}).Error(\"Could not retrieve Elastic Agent logs\")\n\n\t\treturn err\n\t}\n\treturn nil\n}", "func New(logger logging.Logger) Service {\n\ta := &API{\n\t\tlogger: logger,\n\t}\n\n\ta.setupRouting()\n\n\treturn a\n}", "func NewService(dc cache.Service, lggr *logrus.Logger) Service {\n\ts := &service{}\n\ts.watchers = map[string]*job{}\n\tif lggr == nil {\n\t\tlggr = logrus.New()\n\t}\n\ts.lggr = lggr\n\ts.dc = dc\n\n\treturn s\n}", "func New(logsAbsFolder_ string, fileMaxSizeMBytes int, MaxBackupsFiles int, MaxAgeDays int) (*LocalLog, error) {\n\n\tlogger := logrus.New()\n\n\tlogsAbsFolder = filepath.Join(logsAbsFolder_, \"\")\n\tlogsAllAbsFolder = filepath.Join(logsAbsFolder, \"all\")\n\tlogsErrorAbsFolder = filepath.Join(logsAbsFolder, \"error\")\n\n\t//make sure the logs folder exist otherwise create dir\n\tdirError := checkAndMkDir(logsAbsFolder)\n\tif dirError != nil {\n\t\treturn nil, dirError\n\t}\n\tdirError = checkAndMkDir(logsAllAbsFolder)\n\tif dirError != nil {\n\t\treturn nil, dirError\n\t}\n\tdirError = checkAndMkDir(logsErrorAbsFolder)\n\tif dirError != nil {\n\t\treturn nil, dirError\n\t}\n\t///////////////////////\n\t//default info level//\n\tLocalLogPointer := &LocalLog{*logger, logsAllAbsFolder, logsErrorAbsFolder,\n\t\tfileMaxSizeMBytes, MaxBackupsFiles, MaxAgeDays}\n\tLocalLogPointer.ResetLevel(LEVEL_INFO)\n\treturn LocalLogPointer, nil\n}", "func NewAuthLogService(h database.MgoSession) *AuthLogService {\n\treturn &AuthLogService{db: h.DB(\"\")}\n}", "func New(appID, logLevel string) *zapLogger { //nolint\n\tatom := zap.NewAtomicLevel()\n\n\tencoderCfg := zap.NewProductionEncoderConfig()\n\tencoderCfg.EncodeTime = zapcore.ISO8601TimeEncoder\n\n\tlog := zap.New(zapcore.NewCore(\n\t\tzapcore.NewJSONEncoder(encoderCfg),\n\t\tzapcore.Lock(os.Stdout),\n\t\tatom,\n\t))\n\n\tatom.SetLevel(zap.InfoLevel)\n\tif logLevel != \"\" {\n\t\terr := atom.UnmarshalText([]byte(strings.ToLower(logLevel)))\n\t\tif err != nil {\n\t\t\tlog.Error(\"invalid log level\")\n\t\t}\n\t}\n\n\tl := log.With(zap.String(\"svc\", appID))\n\n\treturn &zapLogger{\n\t\tfast: l,\n\t\tsugared: l.Sugar(),\n\t}\n}", "func New(name string, base logrus.Level, level []logrus.Level, dev bool) (*Logger, error) {\n\t// If Logger had been created, return nil.\n\tif Has(name) {\n\t\treturn nil, errors.New(\"Name cannot be duplicated\")\n\t}\n\n\t// Create logger.\n\tlogger := &Logger{Logger: logrus.New()}\n\n\t// Create log file in temp folder.\n\tif logFile, err := ioutil.TempFile(\"\", name+\".*.log\"); err == nil {\n\t\tlogger.Path = logFile.Name()\n\t} else {\n\t\treturn nil, errors.New(\"Cannot create log file\")\n\t}\n\n\t// Enable color logging in Windows console.\n\tlogger.Formatter = &logrus.TextFormatter{ForceColors: true}\n\tlogger.SetOutput(colorable.NewColorableStdout())\n\n\t// Update logger config.\n\tlogger.Config(base, level, dev)\n\n\t// Store logger.\n\tloggers[name] = logger\n\n\treturn logger, nil\n}", "func New(info logger.Info) (logger.Logger, error) {\n\tlogDir := removeLogDirOption(info.Config)\n\tif logDir == \"\" {\n\t\tlogDir = defaultLogDir\n\t}\n\tinfo.LogPath = filepath.Join(logDir, info.ContainerID)\n\n\tif err := os.MkdirAll(filepath.Dir(info.LogPath), 0755); err != nil {\n\t\treturn nil, fmt.Errorf(\"error setting up logger dir: %v\", err)\n\t}\n\n\treturn jsonfilelog.New(info)\n}", "func ServiceLogs(service string) ([]byte, error) {\n\tcmd := DockerComposeCommand(\"logs\", \"--no-color\", service)\n\tout, err := cmd.CombinedOutput()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn out, nil\n}", "func NewUserService(logger logger.Service) Service {\n\treturn newUserService(logger)\n}", "func New(i ILog) *Logger {\n\treturn &Logger{\n\t\tout: i,\n\t}\n}", "func GetLogs(cluster, service, status string) error {\n\tdef, err := GetCurrentTaskDefinition(cluster, service)\n\tif err != nil {\n\t\treturn err\n\t}\n\tallTasks, err := GetAllTasksByDefinition(cluster, def, status)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Problem getting tasks by definition: %v\", err)\n\t}\n\tlogConfig := def.ContainerDefinitions[0].LogConfiguration\n\tif *logConfig.LogDriver != \"awslogs\" {\n\t\treturn fmt.Errorf(\"GetLogs function requires awslogs driver, found %v\", def.ContainerDefinitions[0].LogConfiguration)\n\t}\n\tgroup := logConfig.Options[\"awslogs-group\"]\n\tprefix := logConfig.Options[\"awslogs-stream-prefix\"]\n\tregion := *logConfig.Options[\"awslogs-region\"]\n\tallStreams := make([]*cloudwatchlogs.GetLogEventsInput, len(allTasks))\n\tfor i, task := range allTasks {\n\t\ttaskID := GetTaskIDFromArn(*task.TaskArn)\n\t\tallStreams[i] = &cloudwatchlogs.GetLogEventsInput{\n\t\t\tLogGroupName: group,\n\t\t\tLogStreamName: aws.String(fmt.Sprintf(\"%s/%s/%s\", *prefix, *task.Containers[0].Name, taskID)),\n\t\t}\n\t}\n\twg := &sync.WaitGroup{}\n\twg.Add(len(allStreams))\n\tallLogsStream := make(chan string)\n\tfor _, params := range allStreams {\n\t\tgo func(params *cloudwatchlogs.GetLogEventsInput) {\n\t\t\tdefer wg.Done()\n\t\t\tlogs, err := GetAllLogs(region, params)\n\t\t\tif err == nil {\n\t\t\t\tfor _, log := range logs {\n\t\t\t\t\tallLogsStream <- log\n\t\t\t\t}\n\t\t\t}\n\t\t}(params)\n\t}\n\tallLogs := make([]string, 0)\n\tgo func() {\n\t\tfor curr := range allLogsStream {\n\t\t\tallLogs = append(allLogs, curr)\n\t\t}\n\t}()\n\twg.Wait()\n\tsort.Strings(allLogs)\n\tfmt.Println(strings.Join(allLogs, \"\\n\"))\n\treturn nil\n}", "func New(host string, port uint16, logger logging.Logger) Lister {\n\treturn &service{\n\t\tproxy: http.NewThingProxy(logger, host, port),\n\t}\n}", "func (c *Client) Logs(ctx context.Context, namespace, podName, containerName string, lastContainerLog bool, tail *int64, follow bool) (io.ReadCloser, error) {\n\tretVal := ioutil.NopCloser(strings.NewReader(\"ContainerLogs\"))\n\treturn retVal, nil\n}", "func GetInstance() *zerolog.Logger {\n\tonce.Do(func() {\n\t\tconfig := Config{\n\t\t\tConsoleLoggingEnabled: true,\n\t\t\tEncodeLogsAsJson: true,\n\t\t\tFileLoggingEnabled: true,\n\t\t\tDirectory: \"log\",\n\t\t\tFilename: \"service.log\",\n\t\t\tMaxSize: 10,\n\t\t\tMaxBackups: 5,\n\t\t\tMaxAge: 5,\n\t\t}\n\t\tcustomLogger = configure(config)\n\t})\n\treturn customLogger\n}", "func NewUserlogService(opts ...Option) (*UserlogService, error) {\n\to := &Options{}\n\tfor _, opt := range opts {\n\t\topt(o)\n\t}\n\n\tif o.Stream == nil || o.Store == nil {\n\t\treturn nil, fmt.Errorf(\"need non nil stream (%v) and store (%v) to work properly\", o.Stream, o.Store)\n\t}\n\n\tch, err := events.Consume(o.Stream, \"userlog\", o.RegisteredEvents...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tul := &UserlogService{\n\t\tlog: o.Logger,\n\t\tm: o.Mux,\n\t\tstore: o.Store,\n\t\tcfg: o.Config,\n\t\thistoryClient: o.HistoryClient,\n\t\tgatewaySelector: o.GatewaySelector,\n\t\tvalueClient: o.ValueClient,\n\t\tregisteredEvents: make(map[string]events.Unmarshaller),\n\t\ttp: o.TraceProvider,\n\t\ttracer: o.TraceProvider.Tracer(\"github.com/owncloud/ocis/services/userlog/pkg/service\"),\n\t\tpublisher: o.Stream,\n\t}\n\n\tfor _, e := range o.RegisteredEvents {\n\t\ttyp := reflect.TypeOf(e)\n\t\tul.registeredEvents[typ.String()] = e\n\t}\n\n\tm := roles.NewManager(\n\t\t// TODO: caching?\n\t\troles.Logger(o.Logger),\n\t\troles.RoleService(o.RoleClient),\n\t)\n\n\tul.m.Route(\"/ocs/v2.php/apps/notifications/api/v1/notifications\", func(r chi.Router) {\n\t\tr.Get(\"/\", ul.HandleGetEvents)\n\t\tr.Delete(\"/\", ul.HandleDeleteEvents)\n\t\tr.Post(\"/global\", RequireAdminOrSecret(&m, o.Config.GlobalNotificationsSecret)(ul.HandlePostGlobalEvent))\n\t\tr.Delete(\"/global\", RequireAdminOrSecret(&m, o.Config.GlobalNotificationsSecret)(ul.HandleDeleteGlobalEvent))\n\t})\n\n\tgo ul.MemorizeEvents(ch)\n\n\treturn ul, nil\n}", "func New(crawl crawl.Crawl, maxWorkers int, logger log.Logger) Service {\n\tvar svc Service\n\t{\n\t\tsvc = NewBasicService(crawl, maxWorkers)\n\t\tsvc = LoggingMiddleware(logger)(svc)\n\t}\n\treturn svc\n}", "func (a *LogsApi) ListLogs(ctx _context.Context, o ...ListLogsOptionalParameters) (LogsListResponse, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodPost\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarReturnValue LogsListResponse\n\t\toptionalParams ListLogsOptionalParameters\n\t)\n\n\tif len(o) > 1 {\n\t\treturn localVarReturnValue, nil, datadog.ReportError(\"only one argument of type ListLogsOptionalParameters is allowed\")\n\t}\n\tif len(o) == 1 {\n\t\toptionalParams = o[0]\n\t}\n\n\tlocalBasePath, err := a.Client.Cfg.ServerURLWithContext(ctx, \"v2.LogsApi.ListLogs\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, datadog.GenericOpenAPIError{ErrorMessage: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/api/v2/logs/events/search\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\tlocalVarHeaderParams[\"Content-Type\"] = \"application/json\"\n\tlocalVarHeaderParams[\"Accept\"] = \"application/json\"\n\n\t// body params\n\tif optionalParams.Body != nil {\n\t\tlocalVarPostBody = &optionalParams.Body\n\t}\n\tdatadog.SetAuthKeys(\n\t\tctx,\n\t\t&localVarHeaderParams,\n\t\t[2]string{\"apiKeyAuth\", \"DD-API-KEY\"},\n\t\t[2]string{\"appKeyAuth\", \"DD-APPLICATION-KEY\"},\n\t)\n\treq, err := a.Client.PrepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, nil)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.Client.CallAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := datadog.ReadBody(localVarHTTPResponse)\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := datadog.GenericOpenAPIError{\n\t\t\tErrorBody: localVarBody,\n\t\t\tErrorMessage: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 || localVarHTTPResponse.StatusCode == 403 || localVarHTTPResponse.StatusCode == 429 {\n\t\t\tvar v APIErrorResponse\n\t\t\terr = a.Client.Decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.ErrorModel = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.Client.Decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := datadog.GenericOpenAPIError{\n\t\t\tErrorBody: localVarBody,\n\t\t\tErrorMessage: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func newLog(prefix string) *logging.Logger {\n\tfdFmt := logging.MustStringFormatter(\n\t\t`%{level:.4s}[%{id:03x}]%{time:2006-01-02 15:04:05.000}: %{message}`,\n\t)\n\t// nolint\n\tfd, err := os.OpenFile(logFile, os.O_CREATE|os.O_WRONLY|os.O_APPEND, 0o644)\n\tif err != nil {\n\t\tfmt.Fprint(os.Stderr, err.Error())\n\t}\n\n\tfdLog := logging.NewLogBackend(fd, \"\", 0)\n\tfdFmttr = logging.NewBackendFormatter(fdLog, fdFmt)\n\n\tsysFmttr, err := logging.NewSyslogBackend(prog + \": \")\n\tif err != nil {\n\t\tfmt.Fprint(os.Stderr, err.Error())\n\t}\n\n\tlogging.SetBackend(fdFmttr, sysFmttr)\n\n\treturn logging.MustGetLogger(prog)\n}", "func (k *KubeAPI) Logs(l *LogOpts, out io.Writer) error {\n\treq := k.Client.CoreV1().RESTClient().Get().\n\t\tNamespace(l.Namespace).\n\t\tName(l.Pod).\n\t\tResource(\"pods\").\n\t\tSubResource(\"log\").\n\t\tParam(\"follow\", strconv.FormatBool(l.Follow)).\n\t\tParam(\"container\", l.Container).\n\t\tParam(\"previous\", strconv.FormatBool(l.Previous)).\n\t\tParam(\"timestamps\", strconv.FormatBool(l.Timestamps))\n\n\treadCloser, err := req.Stream()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer readCloser.Close()\n\t_, err = io.Copy(out, readCloser)\n\treturn err\n}", "func New(addr string, files []string, l log.Logger) (*Service, error) {\n\ts, err := server.New(server.WithLogger(l), server.WithSourceFiles(files))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Service{\n\t\taddr: addr,\n\t\tl: l,\n\t\ts: s,\n\t}, nil\n}", "func (m *GraphBaseServiceClient) AuditLogs()(*i20b08d3949f1191430a14a315e0758a1f131dc59bbdc93e654f1dd447a6af14c.AuditLogsRequestBuilder) {\n return i20b08d3949f1191430a14a315e0758a1f131dc59bbdc93e654f1dd447a6af14c.NewAuditLogsRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func (m *GraphBaseServiceClient) AuditLogs()(*i20b08d3949f1191430a14a315e0758a1f131dc59bbdc93e654f1dd447a6af14c.AuditLogsRequestBuilder) {\n return i20b08d3949f1191430a14a315e0758a1f131dc59bbdc93e654f1dd447a6af14c.NewAuditLogsRequestBuilderInternal(m.pathParameters, m.requestAdapter);\n}", "func NewLogger(keptnContext string, eventID string, serviceName string) *Logger {\n\treturn &Logger{\n\t\tKeptnContext: keptnContext,\n\t\tEventID: eventID,\n\t\tServiceName: serviceName,\n\t}\n}", "func New() *Logger {\n\n\tbaseLogrus := logrus.New()\n\n\tvar logger = &Logger{baseLogrus}\n\n\tf, err := os.OpenFile(\"dummy-api.log\", os.O_CREATE|os.O_WRONLY, 0666)\n\tif err != nil {\n\t\tlog.Fatalf(\"unable to interact with log file: %s\", err)\n\t}\n\n\tlogger.SetFormatter(&logrus.JSONFormatter{\n\t\tTimestampFormat: \"02-01-2006 15:04:05\", // DD-MM-YYYY HH:MM:SS\n\n\t})\n\n\toutputs := io.MultiWriter(os.Stderr, f) // Write to both standard error and the log file.\n\tlogger.Out = outputs\n\n\treturn logger\n\n}", "func New(config logging.Config) logging.Logger {\n\treturn rem.New(config, writeWithStd)\n}", "func New(ctx context.Context, params ...Parameter) (*Service, error) {\n\tparameters, err := parseAndCheckParameters(params...)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"problem with parameters\")\n\t}\n\n\t// Set logging.\n\tlog = zerologger.With().Str(\"service\", \"rules\").Str(\"impl\", \"standard\").Logger()\n\tif parameters.logLevel != log.GetLevel() {\n\t\tlog = log.Level(parameters.logLevel)\n\t}\n\n\tstore, err := NewStore(parameters.storagePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Service{\n\t\tstore: store,\n\t\tadminIPs: parameters.adminIPs,\n\t}, nil\n}", "func NewService(m map[string]interface{}) Service {\n\treturn m\n}", "func (a *Client) ListLogService(params *ListLogServiceParams) (*ListLogServiceOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewListLogServiceParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"listLogService\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/Systems/{identifier}/LogServices\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\", \"https\"},\n\t\tParams: params,\n\t\tReader: &ListLogServiceReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*ListLogServiceOK), nil\n\n}", "func New(logFile string) *HistoryLog {\n\tvar customLog *log.Logger\n\tvar file *os.File\n\n\tif logFile == \"\" {\n\t\tcustomLog = log.New(os.Stdout, \"\", log.LstdFlags)\n\t} else {\n\t\tfile, err := os.Create(logFile)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(err)\n\t\t}\n\n\t\t// this file need to be closed handy\n\t\tcustomLog = log.New(file, \"\", log.LstdFlags)\n\t}\n\n\treturn &HistoryLog{\n\t\ttarget: logFile,\n\t\tl: customLog,\n\t\tfile: file,\n\t}\n}", "func LogsFromOtlp(req *otlpcollectorlog.ExportLogsServiceRequest) LogsWrapper {\n\treturn LogsWrapper{req: req}\n}", "func NewDashboardsService(conf config.Config, logger log.LogAdapter) DashboardsService {\n\treturn DashboardsService{\n\t\tconfig: conf,\n\t\tLogger: log.NewSafeAdapter(logger),\n\t}\n}", "func New(l *logrus.Entry) grpclog.Logger {\n\tif l == nil {\n\t\tl = logrus.WithFields(logrus.Fields{\"source\": \"grpc\"})\n\t}\n\treturn &log{l: l}\n}", "func NewService(options ...Option) (Service, error) {\n\ts := &service{}\n\tfor _, option := range options {\n\t\tif err := option(s); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tif s.logger == nil {\n\t\tlogger := logrus.New()\n\t\tlogger.SetOutput(ioutil.Discard)\n\t\ts.logger = logger\n\t}\n\n\treturn s, nil\n}", "func (a *LogsApi) ListLogsGet(ctx _context.Context, o ...ListLogsGetOptionalParameters) (LogsListResponse, *_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarReturnValue LogsListResponse\n\t\toptionalParams ListLogsGetOptionalParameters\n\t)\n\n\tif len(o) > 1 {\n\t\treturn localVarReturnValue, nil, datadog.ReportError(\"only one argument of type ListLogsGetOptionalParameters is allowed\")\n\t}\n\tif len(o) == 1 {\n\t\toptionalParams = o[0]\n\t}\n\n\tlocalBasePath, err := a.Client.Cfg.ServerURLWithContext(ctx, \"v2.LogsApi.ListLogsGet\")\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, datadog.GenericOpenAPIError{ErrorMessage: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/api/v2/logs/events\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\tif optionalParams.FilterQuery != nil {\n\t\tlocalVarQueryParams.Add(\"filter[query]\", datadog.ParameterToString(*optionalParams.FilterQuery, \"\"))\n\t}\n\tif optionalParams.FilterIndexes != nil {\n\t\tlocalVarQueryParams.Add(\"filter[indexes]\", datadog.ParameterToString(*optionalParams.FilterIndexes, \"csv\"))\n\t}\n\tif optionalParams.FilterFrom != nil {\n\t\tlocalVarQueryParams.Add(\"filter[from]\", datadog.ParameterToString(*optionalParams.FilterFrom, \"\"))\n\t}\n\tif optionalParams.FilterTo != nil {\n\t\tlocalVarQueryParams.Add(\"filter[to]\", datadog.ParameterToString(*optionalParams.FilterTo, \"\"))\n\t}\n\tif optionalParams.FilterStorageTier != nil {\n\t\tlocalVarQueryParams.Add(\"filter[storage_tier]\", datadog.ParameterToString(*optionalParams.FilterStorageTier, \"\"))\n\t}\n\tif optionalParams.Sort != nil {\n\t\tlocalVarQueryParams.Add(\"sort\", datadog.ParameterToString(*optionalParams.Sort, \"\"))\n\t}\n\tif optionalParams.PageCursor != nil {\n\t\tlocalVarQueryParams.Add(\"page[cursor]\", datadog.ParameterToString(*optionalParams.PageCursor, \"\"))\n\t}\n\tif optionalParams.PageLimit != nil {\n\t\tlocalVarQueryParams.Add(\"page[limit]\", datadog.ParameterToString(*optionalParams.PageLimit, \"\"))\n\t}\n\tlocalVarHeaderParams[\"Accept\"] = \"application/json\"\n\n\tdatadog.SetAuthKeys(\n\t\tctx,\n\t\t&localVarHeaderParams,\n\t\t[2]string{\"apiKeyAuth\", \"DD-API-KEY\"},\n\t\t[2]string{\"appKeyAuth\", \"DD-APPLICATION-KEY\"},\n\t)\n\treq, err := a.Client.PrepareRequest(ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, nil)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.Client.CallAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := datadog.ReadBody(localVarHTTPResponse)\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := datadog.GenericOpenAPIError{\n\t\t\tErrorBody: localVarBody,\n\t\t\tErrorMessage: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 400 || localVarHTTPResponse.StatusCode == 403 || localVarHTTPResponse.StatusCode == 429 {\n\t\t\tvar v APIErrorResponse\n\t\t\terr = a.Client.Decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.ErrorModel = v\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.Client.Decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := datadog.GenericOpenAPIError{\n\t\t\tErrorBody: localVarBody,\n\t\t\tErrorMessage: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, nil\n}", "func Logs(args ...string) {\n runInstances(\"Logs\", func(i int, id string) error {\n return run(\"logs\", id)\n })\n}", "func NewService(logger log.Logger) OrderService {\n\treturn &service{\n\t\tlogger: logger,\n\t}\n}", "func NewService(c *config.Configs, e *config.ErrorMessage, r Repository, l glog.Logger) *Service {\n\treturn &Service{\n\t\tconf: c,\n\t\tem: e,\n\t\trepo: r,\n\t\tlogger: l,\n\t}\n}", "func (c *Client) GetLogs(ctx context.Context, params logs.Request) (<-chan logs.Message, error) {\n\n\tlogRequest, err := c.newRequest(http.MethodGet, \"/system/logs\", nil)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"cannot connect to OpenFaaS on URL: %s\", c.GatewayURL.String())\n\t}\n\n\tlogRequest.URL.RawQuery = reqAsQueryValues(params).Encode()\n\n\tres, err := c.doRequest(ctx, logRequest)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"cannot connect to OpenFaaS on URL: %s\", c.GatewayURL.String())\n\t}\n\n\tlogStream := make(chan logs.Message, 1000)\n\tswitch res.StatusCode {\n\tcase http.StatusOK:\n\t\tgo func() {\n\t\t\tdefer close(logStream)\n\t\t\tdefer res.Body.Close()\n\n\t\t\tdecoder := json.NewDecoder(res.Body)\n\t\t\tfor decoder.More() {\n\t\t\t\tmsg := logs.Message{}\n\t\t\t\terr := decoder.Decode(&msg)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Printf(\"cannot parse log results: %s\\n\", err.Error())\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tlogStream <- msg\n\t\t\t}\n\t\t}()\n\tcase http.StatusUnauthorized:\n\t\treturn nil, fmt.Errorf(\"unauthorized access, run \\\"faas-cli login\\\" to setup authentication for this server\")\n\tdefault:\n\t\tbytesOut, err := ioutil.ReadAll(res.Body)\n\t\tif err == nil {\n\t\t\treturn nil, fmt.Errorf(\"server returned unexpected status code: %d - %s\", res.StatusCode, string(bytesOut))\n\t\t}\n\t}\n\treturn logStream, nil\n}", "func NewLog(appender Appender, prefix string, tags ...string) (res *Log) {\n\tres = &Log{\n\t\ttags: tags,\n\t\tprefix: prefix,\n\t\tappenderPtr: new(unsafe.Pointer),\n\t\tcallDepth: 2,\n\t}\n\tres.SetAppender(appender)\n\treturn\n}", "func NewLog(dir string, c Config) (*Log, error) {\n\tif c.Segment.MaxStoreBytes == 0 {\n\t\tc.Segment.MaxStoreBytes = 1024\n\t}\n\tif c.Segment.MaxIndexBytes == 0 {\n\t\tc.Segment.MaxIndexBytes = 1024\n\t}\n\tl := &Log{\n\t\tDir: dir,\n\t\tConfig: c,\n\t}\n\treturn l, l.setup()\n}", "func New(ctx ...interface{}) log15.Logger {\n\treturn log15.Root().New(ctx...)\n}" ]
[ "0.7309914", "0.72259873", "0.7207906", "0.7109817", "0.6643609", "0.6547021", "0.64360374", "0.642206", "0.63699096", "0.6364318", "0.63501257", "0.6340149", "0.6262887", "0.62299436", "0.62149596", "0.6210809", "0.6205587", "0.6182399", "0.6182399", "0.6182399", "0.6182399", "0.6182399", "0.6182399", "0.6182399", "0.6182399", "0.6182399", "0.6134504", "0.61096096", "0.60953605", "0.60825276", "0.60753393", "0.60534865", "0.60455495", "0.6011629", "0.6011629", "0.6011629", "0.5955867", "0.5954143", "0.5940303", "0.5914448", "0.5874781", "0.587314", "0.585804", "0.58310664", "0.5824124", "0.58178604", "0.5800628", "0.57936263", "0.57782847", "0.5761629", "0.57481426", "0.57106996", "0.5700052", "0.569286", "0.56827396", "0.5663835", "0.56634843", "0.56512016", "0.56468725", "0.5633806", "0.56324613", "0.56323355", "0.5617043", "0.55791676", "0.5571167", "0.5569853", "0.556028", "0.555496", "0.5539007", "0.5534812", "0.5534164", "0.55337155", "0.5533439", "0.5527111", "0.5524479", "0.55145186", "0.5513612", "0.54960966", "0.5481542", "0.5475406", "0.5475406", "0.5460495", "0.5460431", "0.54207146", "0.5412832", "0.5406513", "0.54060423", "0.5405476", "0.5404079", "0.5403402", "0.5397744", "0.5379541", "0.5378085", "0.53741306", "0.5371285", "0.53630966", "0.5347852", "0.53391755", "0.5333517", "0.53213596" ]
0.75371003
0
test examples from the spec (
func TestCanonicalizationExamples(t *testing.T) { Convey("Given XML Input", t, func() { cases := map[string]exampleXML{ "(Example 3.1 w/o Comments)": exampleXML{input: example31Input, output: example31Output}, "(Example 3.1 w/Comments)": exampleXML{input: example31Input, output: example31OutputWithComments, withComments: true}, "(Example 3.2)": exampleXML{input: example32Input, output: example32Output}, // 3.3 is for Canonical NOT ExclusiveCanonical (one of the exceptions here: http://www.w3.org/TR/xml-exc-c14n/#sec-Specification) //"(Example 3.3)": exampleXML{input: example33Input, output: example33Output}, "(Example 3.4)": exampleXML{input: example34Input, output: example34ModifiedOutput}, //"(Example 3.5)": exampleXML{input: example35Input, output: example35Output}, // 3.6 will work, but requires a change to the etree package first: // http://stackoverflow.com/questions/6002619/unmarshal-an-iso-8859-1-xml-input-in-go //"(Example 3.6)": exampleXML{input: example36Input, output: example36Output}, //"(Example 3.7)": exampleXML{input: example37Input, output: example37Output, expression: example37SubsetExpression}, } for description, test := range cases { Convey(fmt.Sprintf("When transformed %s", description), func() { transform := ExclusiveCanonicalization{WithComments: test.withComments} resultXML, err := transform.Process(test.input, "") Convey("Then the resulting XML match the example output", func() { So(err, ShouldBeNil) So(resultXML, ShouldEqual, test.output) }) }) } }) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func RunExamples(matchString func(pat, str string) (bool, error), examples []InternalExample) (ok bool) {}", "func (t *SpecTest) They(specification string, check func()) { t.Describe(specification, check) }", "func initExamples(obj exampler, attr *expr.AttributeExpr, r *expr.ExampleGenerator) {\n\texamples := attr.ExtractUserExamples()\n\tswitch {\n\tcase len(examples) > 1:\n\t\trefs := make(map[string]*ExampleRef, len(examples))\n\t\tfor _, ex := range examples {\n\t\t\texample := &Example{\n\t\t\t\tSummary: ex.Summary,\n\t\t\t\tDescription: ex.Description,\n\t\t\t\tValue: ex.Value,\n\t\t\t}\n\t\t\trefs[ex.Summary] = &ExampleRef{Value: example}\n\t\t}\n\t\tobj.setExamples(refs)\n\t\treturn\n\tcase len(examples) > 0:\n\t\tobj.setExample(examples[0].Value)\n\tdefault:\n\t\tobj.setExample(attr.Example(r))\n\t}\n}", "func TestExamples(t *testing.T) {\n\ttests := []struct {\n\t\tname string\n\t\targs []string\n\t}{\n\t\t{\n\t\t\tname: \"example/author\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/author/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/author/query.sql\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/complex_params\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/complex_params/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/complex_params/query.sql\",\n\t\t\t\t\"--go-type\", \"int8=int\",\n\t\t\t\t\"--go-type\", \"int4=int\",\n\t\t\t\t\"--go-type\", \"text=string\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/composite\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/composite/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/composite/query.sql\",\n\t\t\t\t\"--go-type\", \"int8=int\",\n\t\t\t\t\"--go-type\", \"int4=int\",\n\t\t\t\t\"--go-type\", \"text=string\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/enums\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/enums/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/enums/query.sql\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"internal/pg\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/author/schema.sql\", // force docker usage\n\t\t\t\t\"--query-glob\", \"internal/pg/query.sql\",\n\t\t\t\t\"--acronym\", \"oid\",\n\t\t\t\t\"--acronym\", \"oids=OIDs\",\n\t\t\t\t\"--go-type\", \"text=string\",\n\t\t\t\t\"--go-type\", \"_int8=[]int\",\n\t\t\t\t\"--go-type\", \"_text=[]string\",\n\t\t\t\t\"--go-type\", \"_float4=[]float32\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/device\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/device/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/device/query.sql\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/erp star glob\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/erp/*.sql\",\n\t\t\t\t\"--query-glob\", \"example/erp/order/*.sql\",\n\t\t\t\t\"--acronym\", \"mrr\",\n\t\t\t\t\"--go-type\", \"tenant_id=int\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/erp question marks\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/erp/??_schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/erp/order/*.sql\",\n\t\t\t\t\"--acronym\", \"mrr\",\n\t\t\t\t\"--go-type\", \"tenant_id=int\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/go_pointer_types\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/go_pointer_types/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/go_pointer_types/query.sql\",\n\t\t\t\t\"--go-type\", \"int8=*int\",\n\t\t\t\t\"--go-type\", \"int4=*int\",\n\t\t\t\t\"--go-type\", \"text=*string\",\n\t\t\t\t\"--go-type\", \"_int8=[]int\",\n\t\t\t\t\"--go-type\", \"_int4=[]int\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/ltree\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/ltree/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/ltree/query.sql\",\n\t\t\t\t\"--go-type\", \"ltree=github.com/jackc/pgtype.Text\",\n\t\t\t\t\"--go-type\", \"_ltree=github.com/jackc/pgtype.TextArray\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/syntax\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/syntax/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/syntax/query.sql\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/custom_types\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/custom_types/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/custom_types/query.sql\",\n\t\t\t\t\"--go-type\", \"text=github.com/jschaf/pggen/example/custom_types/mytype.String\",\n\t\t\t\t\"--go-type\", \"int8=github.com/jschaf/pggen/example/custom_types.CustomInt\",\n\t\t\t\t\"--go-type\", \"my_int=int\",\n\t\t\t\t\"--go-type\", \"_my_int=[]int\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/pgcrypto\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/pgcrypto/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/pgcrypto/query.sql\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/separate_out_dir\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/separate_out_dir/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/separate_out_dir/alpha/query.sql\",\n\t\t\t\t\"--query-glob\", \"example/separate_out_dir/alpha/alpha/query.sql\",\n\t\t\t\t\"--query-glob\", \"example/separate_out_dir/bravo/query.sql\",\n\t\t\t\t\"--output-dir\", \"example/separate_out_dir/out\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/void\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/void/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/void/query.sql\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/nested\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/nested/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/nested/query.sql\",\n\t\t\t\t\"--go-type\", \"int4=int\",\n\t\t\t\t\"--go-type\", \"text=string\",\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tname: \"example/domain\",\n\t\t\targs: []string{\n\t\t\t\t\"--schema-glob\", \"example/domain/schema.sql\",\n\t\t\t\t\"--query-glob\", \"example/domain/query.sql\",\n\t\t\t},\n\t\t},\n\t}\n\tif *update {\n\t\t// update only disables the assertions. Running the tests causes pggen\n\t\t// to overwrite generated code.\n\t\tt.Log(\"updating integration test generated files\")\n\t}\n\tpggen := compilePggen(t)\n\t// Start a single Docker container to use for all tests. Each test will create\n\t// a new database in the Postgres cluster.\n\tctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)\n\tdefer cancel()\n\tdocker, err := pgdocker.Start(ctx, nil, zaptest.NewLogger(t).Sugar())\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdefer errs.CaptureT(t, func() error { return docker.Stop(ctx) }, \"stop docker\")\n\tmainConnStr, err := docker.ConnString()\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tt.Log(\"started dockerized postgres: \" + mainConnStr)\n\tconn, err := pgx.Connect(ctx, mainConnStr)\n\tdefer errs.CaptureT(t, func() error { return conn.Close(ctx) }, \"close conn\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tdbName := \"pggen_example_\" + strconv.FormatInt(int64(rand.Int31()), 36)\n\t\t\tif _, err = conn.Exec(ctx, `CREATE DATABASE `+dbName); err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\tconnStr := mainConnStr + \" dbname=\" + dbName\n\t\t\targs := append(tt.args, \"--postgres-connection\", connStr)\n\t\t\trunPggen(t, pggen, args...)\n\t\t\tif !*update {\n\t\t\t\tassertNoDiff(t)\n\t\t\t}\n\t\t})\n\t}\n}", "func examples(ex interface{}) (*gherkin.Examples, bool) {\n\tt, ok := ex.(*gherkin.Examples)\n\treturn t, ok\n}", "func examples(ex interface{}) (*gherkin.Examples, bool) {\n\tt, ok := ex.(*gherkin.Examples)\n\treturn t, ok\n}", "func execExamples(arity int, p *gop.Context) {\n\targs := p.GetArgs(arity)\n\tconv := func(args []interface{}) []*ast.File {\n\t\tret := make([]*ast.File, len(args))\n\t\tfor i, arg := range args {\n\t\t\tret[i] = arg.(*ast.File)\n\t\t}\n\t\treturn ret\n\t}\n\tret := doc.Examples(conv(args[0:])...)\n\tp.Ret(arity, ret)\n}", "func (suite *PouchAPIHelpSuite) TestExample(c *check.C) {\n}", "func TestExamples(t *testing.T) {\n\tmatches, err := filepath.Glob(\"examples/*.md\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tfor _, match := range matches {\n\t\tm := match\n\t\tt.Run(m, func(t *testing.T) {\n\t\t\tblob, err := ioutil.ReadFile(m)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatal(err)\n\t\t\t}\n\t\t\tvar sql, goc string\n\t\t\tvar captureSQL, captureGo bool\n\t\t\tfor _, line := range strings.Split(string(blob), \"\\n\") {\n\t\t\t\tif strings.HasPrefix(line, \"```sql\") {\n\t\t\t\t\tcaptureSQL = true\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif strings.HasPrefix(line, \"```go\") {\n\t\t\t\t\tcaptureGo = true\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif strings.HasPrefix(line, \"```\") {\n\t\t\t\t\tcaptureSQL = false\n\t\t\t\t\tcaptureGo = false\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif captureSQL {\n\t\t\t\t\tsql += line + \"\\n\"\n\t\t\t\t}\n\t\t\t\tif captureGo {\n\t\t\t\t\tgoc += line + \"\\n\"\n\t\t\t\t}\n\t\t\t}\n\t\t\tif _, err := pg.Parse(sql); err != nil {\n\t\t\t\tt.Errorf(\"could not parse SQL: %s\", err)\n\t\t\t}\n\t\t\tif _, err := format.Source([]byte(goc)); err != nil {\n\t\t\t\tt.Errorf(\"could not format Go: %s\", err)\n\t\t\t}\n\t\t})\n\t}\n}", "func TestSpecs(t *testing.T) {\n\tfor _, tt := range specTests {\n\t\tspec, err := openapi3.ReadFile(tt.filepath, true)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"openapi3.ReadFile('%s', true) Error [%s]\", tt.filepath, err)\n\t\t}\n\t\tif tt.title != spec.Info.Title {\n\t\t\tt.Errorf(\"openapi3.ReadFile('%s', true) Want [%s] Got [%s]\", tt.filepath, tt.title, spec.Info.Title)\n\t\t}\n\t\t_, err = validate.ValidateMore(spec)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"validate.ValidateMore(spec) Error [%s]\", err.Error())\n\t\t} else {\n\t\t\tfmt.Printf(\"VALID [%s]\\n\", tt.filepath)\n\t\t}\n\t\tfmt.Printf(\"TITLE [%s]\\n\", spec.Info.Title)\n\t}\n}", "func init() { //nolint\n\ttestSample := TestCase{\n\t\tName: \"should modify the default preview-lines value [E2E-CLI-014]\",\n\t\tArgs: args{\n\t\t\tArgs: []cmdArgs{\n\t\t\t\t[]string{\"scan\", \"--preview-lines\", \"1\", \"--no-color\", \"--no-progress\",\n\t\t\t\t\t\"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform-single.tf\"},\n\t\t\t},\n\t\t},\n\t\tValidation: func(outputText string) bool {\n\t\t\t// only the match1 must be true\n\t\t\tmatch1, _ := regexp.MatchString(`001\\: resource \\\"aws_redshift_cluster\\\" \\\"default1\\\" \\{`, outputText)\n\t\t\tmatch2, _ := regexp.MatchString(`002\\: publicly_accessible = false`, outputText)\n\t\t\treturn match1 && !match2\n\t\t},\n\t\tWantStatus: []int{40},\n\t}\n\n\tTests = append(Tests, testSample)\n}", "func Test1(t *testing.T) {\n\tlog.Print(\"\\n\\n========== SOLUTION 1 =========\\n\\n\\n\")\n\tfor i, example := range examples {\n\t\tresult, err := parse(example)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\n\t\tx, err := json.MarshalIndent(result, \" \", \" \")\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tlog.Printf(\"Example %d: %s - %s\", i, example, string(x))\n\t}\n}", "func TestParseExamples(t *testing.T) {\n\ts := runtime.NewScheme()\n\n\tif err := scheme.AddToScheme(s); err != nil {\n\t\tt.Fatalf(\"failed to add builtin scheme: %s\", err)\n\t}\n\n\tif err := v1alpha1.AddToScheme(s); err != nil {\n\t\tt.Fatalf(\"failed to add service-api scheme: %s\", err)\n\t}\n\n\tdecoder := serializer.NewCodecFactory(s).UniversalDeserializer()\n\n\tfor _, filename := range findExamples(t, \".\") {\n\t\tt.Run(filename, func(t *testing.T) {\n\t\t\tfor n, buf := range decodeExample(t, filename) {\n\t\t\t\tobj, vers, err := decoder.Decode(buf, nil, nil)\n\t\t\t\tif err != nil {\n\t\t\t\t\tt.Errorf(\"failed to decode YAML object #%d from %q: %s\", n, filename, err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tmetaObj, err := meta.Accessor(obj)\n\t\t\t\tif err != nil {\n\t\t\t\t\tt.Errorf(\"invalid type for decoded object: %s\", err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\tt.Logf(\"decoded YAML object #%d as name=%q group=%q version=%q kind=%q\\n\",\n\t\t\t\t\tn, objectName(metaObj), vers.Group, vers.Version, vers.Kind)\n\t\t\t}\n\t\t})\n\t}\n}", "func doTests(t *testing.T, tests []string) {\n\tdoTestsParam(t, tests, TestParams{\n\t\textensions: parser.CommonExtensions,\n\t})\n}", "func TestPart2(t *testing.T){\n\n\tfor _, test := range tests {\n\n\t\tpart2( test.input )\n\t}\n}", "func init() { //nolint\n\ttestSample := TestCase{\n\t\tName: \"should load libraries from the provided path [E2E-CLI-049]\",\n\t\tArgs: args{\n\t\t\tArgs: []cmdArgs{\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform-single.tf\",\n\t\t\t\t\t\"--libraries-path\", \"fixtures/samples/libraries\"},\n\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/positive.yaml\",\n\t\t\t\t\t\"--libraries-path\", \"fixtures/samples/libraries\"},\n\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/positive.yaml\",\n\t\t\t\t\t\"--libraries-path\", \"fixtures/samples/not-exists-folder\"},\n\t\t\t},\n\t\t},\n\t\tWantStatus: []int{0, 50, 126},\n\t}\n\n\tTests = append(Tests, testSample)\n}", "func TestComplexifyRandom(t *testing.T) {\n\n}", "func init() { //nolint\n\ttestSample := TestCase{\n\t\tName: \"should exclude provided queries [E2E-CLI-020]\",\n\t\tArgs: args{\n\t\t\tArgs: []cmdArgs{\n\t\t\t\t[]string{\"scan\", \"--exclude-queries\", \"15ffbacc-fa42-4f6f-a57d-2feac7365caa,0a494a6a-ebe2-48a0-9d77-cf9d5125e1b3\", \"-s\",\n\t\t\t\t\t\"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform-single.tf\"},\n\t\t\t},\n\t\t},\n\t\tWantStatus: []int{20},\n\t}\n\n\tTests = append(Tests, testSample)\n}", "func init() { //nolint\n\ttestSample := TestCase{\n\t\tName: \"should load custom secrets rules from provided path [E2E-CLI-048]\",\n\t\tArgs: args{\n\t\t\tArgs: []cmdArgs{\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform-secret.tf\",\n\t\t\t\t\t\"--secrets-regexes-path\", \"fixtures/samples/secrets/regex_rules_48_valid.json\"},\n\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform-secret.tf\",\n\t\t\t\t\t\"--secrets-regexes-path\", \"fixtures/samples/secrets/regex_rules_48_valid.json\",\n\t\t\t\t\t\"--exclude-queries\", \"487f4be7-3fd9-4506-a07a-eae252180c08\"},\n\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform-secret.tf\",\n\t\t\t\t\t\"--secrets-regexes-path\", \"fixtures/samples/secrets/regex_rules_48_empty.json\"},\n\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform-secret.tf\",\n\t\t\t\t\t\"--secrets-regexes-path\", \"fixtures/samples/secrets/regex_rules_48_invalid_regex.json\"},\n\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform.tf\",\n\t\t\t\t\t\"--secrets-regexes-path\", \"not-exists-folder\"},\n\n\t\t\t\t[]string{\"scan\", \"--silent\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform.tf\",\n\t\t\t\t\t\"--secrets-regexes-path\", \"samples\"},\n\t\t\t},\n\t\t},\n\t\tWantStatus: []int{50, 40, 40, 126, 126, 126},\n\t}\n\n\tTests = append(Tests, testSample)\n}", "func specs(i interface{}) {\n\tfmt.Println(i)\n}", "func (o QperfSpecOutput) Tests() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v QperfSpec) []string { return v.Tests }).(pulumi.StringArrayOutput)\n}", "func Test_sampe002(t *testing.T) {\n\n}", "func (suite *PouchHelpSuite) TestExample(c *check.C) {\n\tcmd := exec.Command(\"ls\")\n\terr := cmd.Run()\n\tc.Assert(err, check.IsNil)\n}", "func main() {\n\tprintln()\n\tdefer println()\n\n\tdir := flag.String(\"dir\", \"\", \"Directory to test.\")\n\toss := flag.Bool(\"oss\", true, \"Test oss examples.\")\n\tpro := flag.Bool(\"pro\", false, \"Test pro examples.\")\n\tflag.Parse()\n\n\tlog.SetHandler(cli.Default)\n\n\tif *dir != \"\" {\n\t\tif err := testExample(*dir); err != nil {\n\t\t\tlog.Fatalf(\"error: %s\", err)\n\t\t}\n\t\treturn\n\t}\n\n\tif *oss {\n\t\tif err := testExamples(\"oss\"); err != nil {\n\t\t\tlog.Fatalf(\"error: %s\", err)\n\t\t}\n\t}\n\n\tif *pro {\n\t\tif err := testExamples(\"pro\"); err != nil {\n\t\t\tlog.Fatalf(\"error: %s\", err)\n\t\t}\n\t}\n}", "func TestGivenExamples23Part1(t *testing.T) {\n\tcomp := janeMarieComputer{debug: false}\n\tt23ComputerTest(t, &comp, \"./testdata/d23_input_sample.txt\", 0)\n\n\tresultA := (*comp.a)\n\texpectedA := 2\n\tif resultA != expectedA {\n\t\tt.Errorf(\"did not get %v, instead %v\", expectedA, resultA)\n\t}\n}", "func RunExamples(matchString func(pat, str string) (bool, error), examples []InternalExample) (ok bool) {\n\t_, ok = runExamples(matchString, examples)\n\treturn ok\n}", "func TestPrewriteMultiple4A(t *testing.T) {\n}", "func TestDoQuestions(t *testing.T) {\n\tquestions := parseQuestions(\"test\")\n\tdoQuestions(questions, 3)\n}", "func loadRawExamples() {\n\t// load chef_run's for reuse\n\trawruns = make(map[string][]byte)\n\truns := []string{\n\t\t\"../../ingest-service/examples/chef_client_run.json\",\n\t\t\"../../ingest-service/examples/converge-bad-report.json\",\n\t\t\"../../ingest-service/examples/converge-failure-report.json\",\n\t\t\"../../ingest-service/examples/converge-success-report.json\",\n\t}\n\tfor _, r := range runs {\n\t\t// load chef_run json into memory, so that we do not count the json generation\n\t\tcontent, err := ioutil.ReadFile(r)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\trawruns[r] = content\n\t}\n\n\t// load chef_action's for reuse\n\trawactions = make(map[string][]byte)\n\tactions := []string{\n\t\t\"bag_create\",\n\t\t\"bag_create\",\n\t\t\"bag_delete\",\n\t\t\"client_create\",\n\t\t\"cookbookartifactversion_update\",\n\t\t\"environment_create\",\n\t\t\"environment_delete\",\n\t\t\"environment_update\",\n\t\t\"group_create\",\n\t\t\"group_update\",\n\t\t\"item_bag_create\",\n\t\t\"item_bag_update\",\n\t\t\"node_create\",\n\t\t\"node_delete\",\n\t\t\"org_create\",\n\t\t\"permission_update_container\",\n\t\t\"permission_update_cookbook\",\n\t\t\"permission_update_environment\",\n\t\t\"policy_update\",\n\t\t\"user_associate\",\n\t\t\"user_create\",\n\t\t\"user_invite\",\n\t\t\"user_update\",\n\t\t\"version_cookbook_create\",\n\t\t\"version_cookbook_update\",\n\t}\n\tfor _, a := range actions {\n\t\tcontent, err := ioutil.ReadFile(fmt.Sprintf(\"../../ingest-service/examples/actions/%s.json\", a))\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\trawactions[a] = content\n\t}\n\n\t// load liveness's for reuse\n\trawliveness = make(map[string][]byte)\n\tliveness := []string{\n\t\t\"liveness_ping\",\n\t}\n\tfor _, l := range liveness {\n\t\tcontent, err := ioutil.ReadFile(fmt.Sprintf(\"../../ingest-service/examples/%s.json\", l))\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\trawliveness[l] = content\n\t}\n\n\t// load report's for reuse\n\trawreports = make(map[string][]byte)\n\treports := []string{\n\t\t\"compliance-failure-big-report\",\n\t\t\"compliance-success-tiny-report\",\n\t}\n\tfor _, r := range reports {\n\t\tcontent, err := ioutil.ReadFile(fmt.Sprintf(\"../../compliance-service/ingest/examples/%s.json\", r))\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\trawreports[r] = content\n\t}\n}", "func runTests(t *testing.T, tests []test) {\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tresp := executeRequest(tt.method, tt.url, serialize(tt.req), tt.asAdmin)\n\t\t\tif resp.StatusCode != tt.want {\n\t\t\t\tt.Errorf(\"Unexpected status code %d\", resp.StatusCode)\n\t\t\t}\n\n\t\t\tif tt.body != \"\" {\n\t\t\t\tbodyBytes, err := ioutil.ReadAll(resp.Body)\n\t\t\t\tdefer resp.Body.Close()\n\t\t\t\tif err != nil {\n\t\t\t\t\tt.Errorf(\"Error loading body\")\n\t\t\t\t}\n\t\t\t\tif tt.body != string(bodyBytes) {\n\t\t\t\t\tt.Errorf(\"Unexpected body '%s', expected '%s'\", bodyBytes, tt.body)\n\t\t\t\t}\n\t\t\t}\n\t\t})\n\t}\n}", "func genTests() {\n\tanswers := map[int]int64{\n\t\t4: 906609,\n\t\t5: 232792560,\n\t\t6: 25164150,\n\t\t7: 104743,\n\t\t8: 23514624000,\n\t\t9: 31875000,\n\t\t10: 142913828922,\n\t\t11: 70600674,\n\t\t12: 76576500,\n\t\t13: 5537376230,\n\t\t14: 837799,\n\t\t15: 137846528820,\n\t\t16: 1366,\n\t\t17: 21124,\n\t\t18: 1074,\n\t\t19: 171,\n\t\t20: 648,\n\t\t21: 31626,\n\t}\n\n\tfor i := 4; i < len(answers)+4; i++ {\n\t\tfmt.Printf(`func Test%d(t *testing.T) {\n\tif n := Problem%d(); n != %d {\n\t\tt.Errorf(\"#%d got %%d, expected %%d\", n, %d)\n\t}\n}\n`, i, i, answers[i], i, answers[i])\n\t}\n}", "func TestFrameworkExample(t *testing.T) {\n\tScenario(t, \"test example without setup and teardown\", func() {\n\t\ta := 0\n\t\tCase(\"add with verify\", func() {\n\t\t\ta = a + 1\n\t\t\tVerify(a, Equal, 1)\n\t\t})\n\t\tCase(\"add without verify\", func() {\n\t\t\ta = a + 1\n\t\t})\n\t\tCase(\"only verify\", func() {\n\t\t\tVerify(a, Equal, 2)\n\t\t})\n\t})\n}", "func TestExamples(t *testing.T) {\n\tconfigurations, err := ioutil.ReadDir(exampleConfigurationDir)\n\tif err != nil {\n\t\tutil.Die(t, err)\n\t}\n\n\tfor _, configuration := range configurations {\n\t\tname := configuration.Name()\n\n\t\ttest := func(t *testing.T) {\n\t\t\t// Install any prerequisites.\n\t\t\tprerequisitesPath := path.Join(exampleConfigurationDir, name, exampleConfigurationPrerequisites)\n\n\t\t\tif _, err := os.Stat(prerequisitesPath); err == nil {\n\t\t\t\tobjects := util.MustReadYAMLObjects(t, prerequisitesPath)\n\n\t\t\t\tutil.MustCreateResources(t, clients, \"default\", objects)\n\n\t\t\t\tfor i := range objects {\n\t\t\t\t\tobject := objects[i]\n\n\t\t\t\t\tdefer util.DeleteResource(clients, \"default\", object)\n\t\t\t\t}\n\n\t\t\t\t// Hack, we need a better way of synchronizing the readiness of\n\t\t\t\t// resources such as these. For example, a dynamic admission controller\n\t\t\t\t// installed in this phase may not have started e.g. still pulling\n\t\t\t\t// the image, when the service instance is created. The service\n\t\t\t\t// instance then misses all its defaults and collapses in a heap.\n\t\t\t\t// Also we need the clients to sync in the mean time in order to\n\t\t\t\t// see the new CRD types. Make this synchronous.\n\t\t\t\ttime.Sleep(time.Minute)\n\t\t\t}\n\n\t\t\t// Create a clean namespace to test in, we can clean up everything\n\t\t\t// by just deleting it and letting the cascade do its thing.\n\t\t\tnamespace := util.MustCreateResource(t, clients, \"\", util.MustGetNamespace(t))\n\n\t\t\tdefer util.DeleteResource(clients, \"\", namespace)\n\n\t\t\t// Install the service broker configuration for the example.\n\t\t\t// * Tests example passes CRD validation.\n\t\t\tconfigurationPath := path.Join(exampleConfigurationDir, name, exampleConfigurationSpecification)\n\n\t\t\tobjects := util.MustReadYAMLObjects(t, configurationPath)\n\t\t\tserviceBrokerConfiguration := util.MustFindResource(t, objects, \"servicebroker.couchbase.com/v1alpha1\", \"ServiceBrokerConfig\", exampleDefaultResourceName)\n\n\t\t\tutil.MustCreateResources(t, clients, namespace.GetName(), objects)\n\n\t\t\t// Install the service broker, we need to check that the service broker\n\t\t\t// flags the configuration as valid and the deployment is available.\n\t\t\t// As the namespace is ephemeral we need to watch out for any resources\n\t\t\t// that usually refer to \"default\" explicitly.\n\t\t\t// * Tests service broker comes up in Kubernetes.\n\t\t\t// * Tests example passses service broker validation.\n\t\t\tcaCertificate, serverCertificate, serverKey := util.MustGenerateServiceBrokerTLS(t, namespace.GetName())\n\n\t\t\tobjects = util.MustReadYAMLObjects(t, exampleBrokerConfiguration)\n\t\t\tserviceBrokerSecret := util.MustFindResource(t, objects, \"v1\", \"Secret\", exampleDefaultResourceName)\n\t\t\tserviceBrokerRoleBinding := util.MustFindResource(t, objects, \"rbac.authorization.k8s.io/v1\", \"RoleBinding\", exampleDefaultResourceName)\n\t\t\tserviceBrokerDeployment := util.MustFindResource(t, objects, \"apps/v1\", \"Deployment\", exampleDefaultResourceName)\n\n\t\t\t// Override the service broker TLS secret data.\n\t\t\tdata := map[string]interface{}{\n\t\t\t\t\"token\": base64.StdEncoding.EncodeToString([]byte(token)),\n\t\t\t\t\"username\": base64.StdEncoding.EncodeToString([]byte(username)),\n\t\t\t\t\"password\": base64.StdEncoding.EncodeToString([]byte(password)),\n\t\t\t\t\"tls-certificate\": base64.StdEncoding.EncodeToString(serverCertificate),\n\t\t\t\t\"tls-private-key\": base64.StdEncoding.EncodeToString(serverKey),\n\t\t\t}\n\n\t\t\tif err := unstructured.SetNestedField(serviceBrokerSecret.Object, data, \"data\"); err != nil {\n\t\t\t\tutil.Die(t, err)\n\t\t\t}\n\n\t\t\t// Override the service broker role binding namespace.\n\t\t\tsubjects := []interface{}{\n\t\t\t\tmap[string]interface{}{\n\t\t\t\t\t\"kind\": \"ServiceAccount\",\n\t\t\t\t\t\"name\": exampleDefaultResourceName,\n\t\t\t\t\t\"namespace\": namespace.GetName(),\n\t\t\t\t},\n\t\t\t}\n\n\t\t\tif err := unstructured.SetNestedField(serviceBrokerRoleBinding.Object, subjects, \"subjects\"); err != nil {\n\t\t\t\tutil.Die(t, err)\n\t\t\t}\n\n\t\t\t// Set the pull policy so we use the cached version.\n\t\t\tif err := unstructured.SetNestedField(serviceBrokerDeployment.Object, \"IfNotPresent\", \"spec\", \"template\", \"spec\", \"imagePullPolicy\"); err != nil {\n\t\t\t\tutil.Die(t, err)\n\t\t\t}\n\n\t\t\tutil.MustCreateResources(t, clients, namespace.GetName(), objects)\n\n\t\t\tutil.MustWaitFor(t, util.ResourceCondition(clients, namespace.GetName(), serviceBrokerConfiguration, string(v1.ConfigurationValid), string(v1.ConditionTrue)), time.Minute)\n\t\t\tutil.MustWaitFor(t, util.ResourceCondition(clients, namespace.GetName(), serviceBrokerDeployment, string(appsv1.DeploymentAvailable), string(corev1.ConditionTrue)), time.Minute)\n\n\t\t\t// Register the service broker with the service catalog.\n\t\t\t// We replaced the service broker configuration with new TLS due to the\n\t\t\t// namespace change, do the same here.\n\t\t\t// * Tests the service catalog can talk to the service broker.\n\t\t\tobjects = util.MustReadYAMLObjects(t, exampleClusterServiceBroker)\n\t\t\tclusterServiceBroker := util.MustFindResource(t, objects, \"servicecatalog.k8s.io/v1beta1\", \"ClusterServiceBroker\", exampleDefaultResourceName)\n\n\t\t\tif err := unstructured.SetNestedField(clusterServiceBroker.Object, fmt.Sprintf(\"https://%s.%s\", exampleDefaultResourceName, namespace.GetName()), \"spec\", \"url\"); err != nil {\n\t\t\t\tutil.Die(t, err)\n\t\t\t}\n\n\t\t\tif err := unstructured.SetNestedField(clusterServiceBroker.Object, base64.StdEncoding.EncodeToString(caCertificate), \"spec\", \"caBundle\"); err != nil {\n\t\t\t\tutil.Die(t, err)\n\t\t\t}\n\n\t\t\tif err := unstructured.SetNestedField(clusterServiceBroker.Object, namespace.GetName(), \"spec\", \"authInfo\", \"basic\", \"secretRef\", \"namespace\"); err != nil {\n\t\t\t\tutil.Die(t, err)\n\t\t\t}\n\n\t\t\tutil.MustCreateResources(t, clients, namespace.GetName(), objects)\n\n\t\t\tdefer util.DeleteResource(clients, \"\", clusterServiceBroker)\n\n\t\t\tutil.MustWaitFor(t, util.ResourceCondition(clients, namespace.GetName(), clusterServiceBroker, \"Ready\", \"True\"), time.Minute)\n\n\t\t\t// Create the service instance.\n\t\t\t// * Tests the configuration provisions.\n\t\t\tserviceInstancePath := path.Join(exampleConfigurationDir, name, exampleConfigurationServiceInstance)\n\n\t\t\tobjects = util.MustReadYAMLObjects(t, serviceInstancePath)\n\t\t\tserviceInstance := util.MustFindResource(t, objects, \"servicecatalog.k8s.io/v1beta1\", \"ServiceInstance\", exampleDefaultServiceInstanceName)\n\n\t\t\tutil.MustCreateResources(t, clients, namespace.GetName(), objects)\n\n\t\t\tutil.MustWaitFor(t, util.ResourceCondition(clients, namespace.GetName(), serviceInstance, \"Ready\", \"True\"), 10*time.Minute)\n\n\t\t\t// Create the service binding if one exists.\n\t\t\t// * Tests the configuration provisions.\n\t\t\tserviceBindingPath := path.Join(exampleConfigurationDir, name, exampleConfigurationServiceBinding)\n\t\t\tif _, err := os.Stat(serviceBindingPath); err == nil {\n\t\t\t\tobjects = util.MustReadYAMLObjects(t, serviceBindingPath)\n\t\t\t\tserviceBinding := util.MustFindResource(t, objects, \"servicecatalog.k8s.io/v1beta1\", \"ServiceBinding\", exampleDefaultServiceBindingName)\n\n\t\t\t\tutil.MustCreateResources(t, clients, namespace.GetName(), objects)\n\n\t\t\t\tutil.MustWaitFor(t, util.ResourceCondition(clients, namespace.GetName(), serviceBinding, \"Ready\", \"True\"), 5*time.Minute)\n\n\t\t\t\t// Delete the service binding.\n\t\t\t\t// * Tests the service binding is deprovisioned cleanly.\n\t\t\t\tutil.DeleteResource(clients, namespace.GetName(), serviceBinding)\n\t\t\t}\n\n\t\t\t// Delete the service instance.\n\t\t\t// * Tests the service instance is deprovisioned cleanly.\n\t\t\tutil.DeleteResource(clients, namespace.GetName(), serviceInstance)\n\t\t}\n\n\t\tt.Run(\"TestExample-\"+name, test)\n\t}\n}", "func TestComplexTypes(t *testing.T) {\n\n}", "func Test(t *testing.T, p prog.Program, cases ...Case) {\n\tt.Helper()\n\tfor _, c := range cases {\n\t\tt.Run(strings.Join(c.args, \" \"), func(t *testing.T) {\n\t\t\tt.Helper()\n\t\t\tr := run(p, c.args, c.stdin)\n\t\t\tif r.exitCode != c.want.exitCode {\n\t\t\t\tt.Errorf(\"got exit code %v, want %v\", r.exitCode, c.want.exitCode)\n\t\t\t}\n\t\t\tif !matchOutput(r.stdout, c.want.stdout) {\n\t\t\t\tt.Errorf(\"got stdout %v, want %v\", r.stdout, c.want.stdout)\n\t\t\t}\n\t\t\tif !matchOutput(r.stderr, c.want.stderr) {\n\t\t\t\tt.Errorf(\"got stderr %v, want %v\", r.stderr, c.want.stderr)\n\t\t\t}\n\t\t})\n\t}\n}", "func (o QperfSpecPtrOutput) Tests() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *QperfSpec) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Tests\n\t}).(pulumi.StringArrayOutput)\n}", "func TestSolution(t *testing.T) {\n\tfor test, f := range tests {\n\t\tRunTest(t, test, f, SampleTestCases)\n\t}\n}", "func Main(matchString func(pat, str string) (bool, error), tests []InternalTest, benchmarks []InternalBenchmark, examples []InternalExample)", "func (suite *AddCommandTestSuite) TestExecuteWhenMultipleTracksFound() {\n\n}", "func TestSpec_IsInUniverse(t *testing.T) {\n\tcode := `\npackage haha\nimport \"fmt\"\nconst a = 1\nfunc main() {\n var b = 2.0\n type c struct {\n d string\n }\n fmt.Println(a, b)\n}`\n\ts := NewSpec(code)\n\tif s.IsInUniverse(\"bool\") && !s.IsInUniverse(\"a\") && !s.IsInUniverse(\"kaka\") {\n\t} else {\n\t\tt.Error(`test failed`)\n\t}\n}", "func TestAllSpecs(t *testing.T) {\n\tr := gospec.NewRunner()\n\n\tr.Parallel = false\n\n\tr.BeforeEach = func() {\n\t\tConfigure(map[string]string{\n\t\t\t\"server\": \"localhost:7000\",\n\t\t\t\"process\": \"1\",\n\t\t\t\"database\": \"15\",\n\t\t\t\"pool\": \"1\",\n\t\t})\n\n\t\tfor port := 7000; port <= 7002; port++ {\n\t\t\tc, _ := redis.Dial(\"tcp\", \"localhost:\"+strconv.Itoa(port))\n\t\t\tc.Cmd(\"flushdb\")\n\t\t\tc.Close()\n }\n\t}\n\n\t// List all specs here\n\tr.AddSpec(WorkersSpec)\n\tr.AddSpec(ConfigSpec)\n\tr.AddSpec(MsgSpec)\n\tr.AddSpec(FetchSpec)\n\tr.AddSpec(WorkerSpec)\n\tr.AddSpec(ManagerSpec)\n\tr.AddSpec(ScheduledSpec)\n\tr.AddSpec(EnqueueSpec)\n\tr.AddSpec(MiddlewareSpec)\n\tr.AddSpec(MiddlewareRetrySpec)\n\tr.AddSpec(MiddlewareStatsSpec)\n\n\t// Run GoSpec and report any errors to gotest's `testing.T` instance\n\tgospec.MainGoTest(r, t)\n}", "func init() { //nolint\n\ttestSample := TestCase{\n\t\tName: \"should hide the progress bar in the CLI [E2E-CLI-009]\",\n\t\tArgs: args{\n\t\t\tArgs: []cmdArgs{\n\t\t\t\t[]string{\"scan\", \"-q\", \"../assets/queries\", \"-p\", \"fixtures/samples/terraform.tf\", \"--no-progress\"},\n\t\t\t},\n\t\t},\n\t\tWantStatus: []int{50},\n\t\tValidation: func(outputText string) bool {\n\t\t\tgetProgressRegex := \"Executing queries:\"\n\t\t\tmatch, _ := regexp.MatchString(getProgressRegex, outputText)\n\t\t\t// if not found -> the the test was successful\n\t\t\treturn !match\n\t\t},\n\t}\n\n\tTests = append(Tests, testSample)\n}", "func Main(matchString func(pat, str string) (bool, error), tests []InternalTest, benchmarks []InternalBenchmark, examples []InternalExample) {}", "func TestExampleTestSuite(t *testing.T) {\n\tsuite.Run(t, new(ExampleTestSuite))\n}", "func (suite *AddCommandTestSuite) TestExecuteWithMultipleURLs() {\n\n}", "func getTestCases() []testCase {\n\treturn []testCase{\n\t\t{\n\t\t\tinputStart: 1,\n\t\t\tinputEnd: 10,\n\n\t\t\texpectedOutput: \"1 2 Fizz 4 Buzz Fizz 7 8 Fizz Buzz\",\n\t\t\texpectingError: false,\n\t\t},\n\t\t{\n\t\t\tinputStart: 5,\n\t\t\tinputEnd: 8,\n\n\t\t\texpectedOutput: \"Buzz Fizz 7 8\",\n\t\t\texpectingError: false,\n\t\t},\n\t\t{\n\t\t\tinputStart: 5,\n\t\t\tinputEnd: 12,\n\n\t\t\texpectedOutput: \"Buzz Fizz 7 8 Fizz Buzz 11 Fizz\",\n\t\t\texpectingError: false,\n\t\t},\n\t\t{\n\t\t\tinputStart: 5,\n\t\t\tinputEnd: 2,\n\n\t\t\texpectingError: true,\n\t\t},\n\t\t{\n\t\t\tinputStart: 5,\n\t\t\tinputEnd: 10,\n\n\t\t\texpectedOutput: \"Buzz Fizz 7 8 Fizz Buzz\",\n\t\t\texpectingError: false,\n\t\t},\n\t\t{\n\t\t\tinputStart: -12,\n\t\t\tinputEnd: 10,\n\n\t\t\texpectingError: true,\n\t\t},\n\t\t{\n\t\t\tinputStart: 0,\n\t\t\tinputEnd: 10,\n\n\t\t\texpectingError: true,\n\t\t},\n\t\t{\n\t\t\tinputStart: 10,\n\t\t\tinputEnd: 15,\n\n\t\t\texpectedOutput: \"Buzz 11 Fizz 13 14 FizzBuzz\",\n\t\t\texpectingError: false,\n\t\t},\n\t\t{\n\t\t\tinputStart: 10,\n\t\t\tinputEnd: 2,\n\n\t\t\texpectingError: true,\n\t\t},\n\t\t{\n\t\t\tinputStart: 10,\n\t\t\tinputEnd: 10,\n\n\t\t\texpectedOutput: \"Buzz\",\n\t\t\texpectingError: false,\n\t\t},\n\t\t{\n\t\t\tinputStart: 10,\n\t\t\tinputEnd: 9,\n\n\t\t\texpectingError: true,\n\t\t},\n\t\t{\n\t\t\tinputStart: 1,\n\t\t\tinputEnd: 36,\n\n\t\t\texpectedOutput: \"1 2 Fizz 4 Buzz Fizz 7 8 Fizz Buzz 11 Fizz 13 14 FizzBuzz 16 17 Fizz 19 Buzz Fizz 22 23 Fizz Buzz 26 Fizz 28 29 FizzBuzz 31 32 Fizz 34 Buzz Fizz\",\n\t\t\texpectingError: false,\n\t\t},\n\t\t{\n\t\t\tinputStart: 10,\n\t\t\tinputEnd: 20,\n\n\t\t\texpectedOutput: \"Buzz 11 Fizz 13 14 FizzBuzz 16 17 Fizz 19 Buzz\",\n\t\t\texpectingError: false,\n\t\t},\n\t}\n}", "func test(t *testing.T, scene string, f func(*testing.T)) {\n\tif t.Failed() {\n\t\treturn\n\t}\n\tConvey(scene, t, func() {\n\t\tf(t)\n\t})\n}", "func (t *SpecTest) Describe(thing string, does func()) {\n\tt.getSpecRegexp()\n\n\tt.descstack = append(t.descstack, thing)\n\tt.beforestack = append(t.beforestack, nil)\n\tt.deferstack = append(t.deferstack, nil)\n\tt.depth++\n\n\toldrunspec := t.runspec\n\tif specregexp != nil && !specregexp.MatchString(t.String()) {\n\t\tt.runspec = false\n\t} else if specregexp != nil {\n\t\tt.runspec = true\n\t}\n\n\tdefer func() {\n\t\t// Clear the SpecTest when the description's scope is left.\n\t\tt.depth--\n\t\tt.descstack = t.descstack[:t.depth]\n\t\tpopTrigger(&t.beforestack[t.depth])\n\t\tpopTrigger(&t.deferstack[t.depth])\n\t\tt.spec = nil\n\t\tt.passed = true\n\t\tt.ranspec = false\n\t\tt.err = nil\n\t\tt.runspec = oldrunspec\n\t}()\n\n\tafter := t.deferstack\n\tif k := len(after[t.depth-1]); k > 0 {\n\t\tfor j := 0; j < k; j++ {\n\t\t\ttr := after[t.depth-1][j]\n\t\t\tif tr.Quantifier == Last {\n\t\t\t\tdefer tr.fn()\n\t\t\t}\n\t\t}\n\t}\n\tt.deferstack = after\n\n\t// Do the described tests.\n\tdoes()\n\tif t.ranspec {\n\t\t// Compute the result of executed Spec calls.\n ok := t.passed && t.err == nil\n var result string\n\t\tswitch {\n case ok:\n\t\t result = \"PASS\"\n case t.err != nil:\n\t\t\tresult = \"ERROR\"\n\t\tcase !t.passed:\n\t\t\tresult = \"FAIL\"\n default:\n panic(\"unexpected outcome\")\n\t\t}\n\n\t\t// Write a message summarizing Spec calls.\n\t\tmsg := fmt.Sprintf(\"%s: %s\", t.String(), result)\n\t\tif !ok {\n\t\t\tmsg += fmt.Sprintf(\"\\n\\t%s\", specString(t.spec))\n\t\t}\n\t\tif t.err != nil {\n\t\t\tmsg += fmt.Sprintf(\"\\n\\tError: %s\", t.err.Error())\n\t\t}\n\n\t\t// Write the message as an error if there was a problem.\n\t\tif ok {\n\t\t\tt.Log(msg)\n\t\t} else {\n\t\t\tt.Error(msg)\n\t\t}\n\t}\n}", "func TestParseQuestions(t *testing.T) {\n\tquestions := parseQuestions(\"test\")\n\n\tif len(questions) != 2 {\n\t\tt.Fatal(\"Expected a single question in the file\")\n\n\t}\n}", "func TestGetVersions4A(t *testing.T) {\n}", "func TestAllSpecs(t *testing.T) {\n\tr := gospec.NewRunner()\n\n\tr.Parallel = false\n\tredisNodes := 6\n\n\tr.BeforeEach = func() {\n\t\tmrs := []*miniredis.Miniredis{}\n\t\tserverAddrs := []string{}\n\t\tfor i := 0; i < redisNodes; i++ {\n\t\t\tmr, _ := miniredis.Run()\n\t\t\tmrs = append(mrs, mr)\n\t\t\tserverAddrs = append(serverAddrs, mr.Addr())\n\t\t}\n\n\t\tConfigure(map[string]string{\n\t\t\t\"server\": strings.Join(serverAddrs[:], \",\"),\n\t\t\t\"process\": \"1\",\n\t\t})\n\n\t\tfor i := 0; i < redisNodes; i++ {\n\t\t\tmrs[i].FlushAll()\n\t\t}\n\t}\n\n\t// List all specs here\n\tr.AddSpec(WorkersSpec)\n\tr.AddSpec(ConfigSpec)\n\tr.AddSpec(MsgSpec)\n\tr.AddSpec(FetchSpec)\n\tr.AddSpec(WorkerSpec)\n\tr.AddSpec(ManagerSpec)\n\tr.AddSpec(ScheduledSpec)\n\tr.AddSpec(EnqueueSpec)\n\tr.AddSpec(MiddlewareSpec)\n\tr.AddSpec(MiddlewareRetrySpec)\n\tr.AddSpec(MiddlewareStatsSpec)\n\n\t// Run GoSpec and report any errors to gotest's `testing.T` instance\n\tgospec.MainGoTest(r, t)\n}", "func runTestScenarios(t *testing.T, tests []testCase) {\n\trunTestScenariosWithInput(t, tests, nil)\n}", "func TestUnmarshalPathologies(t *testing.T) {\n\n}", "func E2ETests(testEnv *env.TestEnv) func(*testing.T) {\n\tctx := &context{\n\t\ttestEnv,\n\t}\n\n\treturn ctx.testWelcome\n}", "func TestSinglePrewrite4A(t *testing.T) {\n}", "func TestExamplesComplete(t *testing.T) {\n\tt.Parallel()\n\n\tterraformOptions := &terraform.Options{\n\t\t// The path to where our Terraform code is located\n\t\tTerraformDir: \"../../examples/complete\",\n\t\tUpgrade: true,\n\t\t// Variables to pass to our Terraform code using -var-file options\n\t\tVarFiles: []string{\"fixtures.us-east-2.tfvars\"},\n\t}\n\n\t// At the end of the test, run `terraform destroy` to clean up any resources that were created\n\tdefer terraform.Destroy(t, terraformOptions)\n\n\t// This will run `terraform init` and `terraform apply` and fail the test if there are any errors\n\tterraform.InitAndApply(t, terraformOptions)\n\n\t// Run `terraform output` to get the value of an output variable\n\toutput := terraform.Output(t, terraformOptions, \"map\")\n\n\tkey1Result := strings.Contains(output, \"/production/test/master/company\")\n\tvalue1Result := strings.Contains(output, \"Amazon\")\n\tkey2Result := strings.Contains(output, \"/production/test/master/users\")\n\tvalue2Result := strings.Contains(output, \"John,Todd\")\n\n\t// Verify we're getting back the outputs we expect\n\tassert.True(t, key1Result, \"The 'map' output should contain the /production/test/master/company key\")\n\tassert.True(t, value1Result, \"The /production/test/master/company key's value should be 'Amazon'\")\n\n\tassert.True(t, key2Result, \"The 'map' output should contain the /production/test/master/users key\")\n\tassert.True(t, value2Result, \"The /production/test/master/users key's value should be 'John,Todd'\")\n}", "func TestSpecCluster_SpecsReader(t *testing.T) {\n\n\ttests := []struct {\n\t\tname string\n\t\tfile string\n\t\twantErr bool\n\t}{\n\t\t{\n\t\t\tname: \"Read cluster workload spec from yaml\",\n\t\t\tfile: \"/cluster/positive/edge_workload_cluster.yaml\",\n\t\t\twantErr: false,\n\t\t},\n\t\t{\n\t\t\tname: \"Read cluster workload spec from json\",\n\t\t\tfile: \"/cluster/positive/edge_workload_cluster.json\",\n\t\t\twantErr: false,\n\t\t},\n\t\t{\n\t\t\tname: \"Read mgmt cluster spec from yaml\",\n\t\t\tfile: \"/cluster/positive/edge_mgmt_cluster.yaml\",\n\t\t\twantErr: false,\n\t\t},\n\t\t{\n\t\t\tname: \"Read mgmt cluster spec from json\",\n\t\t\tfile: \"/cluster/positive/edge_mgmt_cluster.json\",\n\t\t\twantErr: false,\n\t\t},\n\t}\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\n\t\t\tassetsDir := GetTestAssetsDir()\n\t\t\tfileName := assetsDir + tt.file\n\t\t\tf, err := os.Open(fileName)\n\t\t\tassert.NoError(t, err)\n\n\t\t\tspec, err := SpecCluster{}.SpecsFromReader(f)\n\t\t\tif tt.wantErr && err == nil {\n\t\t\t\tt.Errorf(\"Test failed must not return error\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tif tt.wantErr && err != nil {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif spec == nil {\n\t\t\t\tt.Errorf(\"SpecsFromFile() return nil spec\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tclusterSpec, ok := (*spec).(*SpecCluster)\n\t\t\tif !ok {\n\t\t\t\tt.Errorf(\"Test failed method return wrong type\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\terr = clusterSpec.Validate()\n\t\t\tif err != nil {\n\t\t\t\tt.Errorf(\"SpecsFromFile() Test failed validator \"+\n\t\t\t\t\t\"return error for positive case err %v file %s\", err, fileName)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t})\n\t}\n}", "func TestUsage(t *testing.T) {\n\tinitialize(t)\n\tusageSubscriber(\"testing\", t)\n\tassertUsageSubscriber(\"testing\", true, t)\n\tusageSubscriber(\"witness\", t)\n\tassertUsageSubscriber(\"witness\", true, t)\n\tassertUsageSubscriber(\"badguy\", false, t)\n\tusageWritePayloads(pcnt, t)\n\tusageReadCheck(\"witness\", pcnt, true, t)\n\tusageReadCheck(\"testing\", pcnt, false, t)\n}", "func TestDemo(t *testing.T) {\n\tvar (\n\t\ttn string\n\t\terr error\n\t)\n\trefreshErrHandler := func(err error) {\n\t\tif err != nil {\n\t\t\tt.Error(err)\n\t\t}\n\t}\n\tctx := context.Background()\n\tconfigStorage, err := onlinelab.NewConsulConfigStorage(&api.Config{})\n\tif err != nil {\n\t\tt.Error(\"failed to create config storage\")\n\t}\n\ttreatmentControoler, err := onlinelab.CreateTreatmentController(ctx, configStorage,\n\t\t\"HelloLab\", time.Second*1, refreshErrHandler)\n\tif err != nil {\n\t\tt.Error(\"failed to create treatment controller\")\n\t}\n\n\tcntT1 := 0\n\tcntT2 := 0\n\tfor i := 1; i <= 1000000; i++ {\n\t\ttn, err = treatmentControoler.GetNextTreatment(i)\n\t\tif err != nil {\n\t\t\tt.Error(err)\n\t\t\tbreak\n\t\t}\n\t\tswitch tn {\n\t\tcase \"T1\":\n\t\t\tcntT1++\n\t\tcase \"T2\":\n\t\t\tcntT2++\n\t\t}\n\t}\n\tt.Logf(\"T1 count: %d; T2 count: %d\", cntT1, cntT2)\n}", "func TestParser(t *testing.T) {\n\texamples, err := filepath.Glob(\"example/*.bib\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tfor _, ex := range examples {\n\t\tt.Logf(\"Parsing example: %s\", ex)\n\t\tb, err := ioutil.ReadFile(ex)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Cannot read %s: %v\", ex, err)\n\t\t}\n\t\t_, err = Parse(bytes.NewReader(b))\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Cannot parse valid bibtex file %s: %v\", ex, err)\n\t\t}\n\t}\n}", "func (t *SpecTest) It(specification string, check func()) { t.Describe(specification, check) }", "func UsageExamples() string {\n\treturn os.Args[0] + ` foo foo1 --p 6458532619907689287` + \"\\n\" +\n\t\t\"\"\n}", "func TestExample(t *testing.T) {\n\tinstance := &RestExample{\n\t\tpost: make(map[string]string),\n\t\twatch: make(map[string]chan string),\n\t}\n\n\tinstance.HandleCreateHello(HelloArg{\n\t\tTo: \"rest\",\n\t\tPost: \"rest is powerful\",\n\t})\n\n\tresp, err := rest.SetTest(instance, map[string]string{\"to\": \"rest\"}, nil)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\targ := instance.HandleHello()\n\tif resp.Code != http.StatusOK {\n\t\tt.Error(\"should return ok\")\n\t}\n\tif arg.To != \"rest\" {\n\t\tt.Error(\"arg.To should be rest\")\n\t}\n\tif arg.Post != \"rest is powerful\" {\n\t\tt.Error(\"arg.Post should be 'rest is powerful'\")\n\t}\n\n\tresp, err = rest.SetTest(instance, map[string]string{\"to\": \"123\"}, nil)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\targ = instance.HandleHello()\n\tif resp.Code != http.StatusNotFound {\n\t\tt.Error(\"should return not found\")\n\t}\n}", "func CheckExamples(cmd *cobra.Command) []error {\n\tfmt.Fprint(os.Stdout, \" ↳ checking examples\\n\")\n\tcmdPath := cmd.CommandPath()\n\texamples := cmd.Example\n\terrors := []error{}\n\tif len(examples) > 0 {\n\t\tfor _, line := range strings.Split(examples, \"\\n\") {\n\t\t\tif !strings.HasPrefix(line, templates.Indentation) {\n\t\t\t\terrors = append(errors, fmt.Errorf(`command %q: examples are not normalized, make sure you are calling templates.Examples (from pkg/cmd/templates) before assigning cmd.Example`, cmdPath))\n\t\t\t}\n\t\t\tif trimmed := strings.TrimSpace(line); strings.HasPrefix(trimmed, \"//\") {\n\t\t\t\terrors = append(errors, fmt.Errorf(`command %q: we use # to start comments in examples instead of //`, cmdPath))\n\t\t\t}\n\t\t}\n\t}\n\treturn errors\n}", "func tests() []testcase {\n\titems := [...]testcase{\n\t\t{0, 0},\n\t\t{1, 1},\n\t\t{2, 1},\n\t\t{3, 2},\n\t\t{4, 3},\n\t\t{5, 5},\n\t\t{6, 8},\n\t\t{7, 13},\n\t\t{8, 21},\n\t\t{9, 34},\n\t\t{10, 55},\n\t\t{11, 89},\n\t\t{12, 144},\n\t\t{13, 233},\n\t\t{14, 377},\n\t\t{15, 610},\n\t\t{16, 987},\n\t\t{17, 1597},\n\t\t{18, 2584},\n\t\t{19, 4181},\n\t\t{20, 6765},\n\t}\n\treturn items[:]\n}", "func TestData(t *testing.T) { TestingT(t) }", "func TestCodeDocumentation(t interface {\n\tFatalf(string, ...interface{})\n\tFatal(...interface{})\n}, options *Options) {\n\tif options == nil {\n\t\tt.Fatalf(\"options cannot be nil\")\n\t\treturn\n\t}\n\tif options.Symbols == nil {\n\t\toptions.Symbols = yaegi_template.DefaultSymbols()\n\t}\n\tif options.Options == nil {\n\t\topts := yaegi_template.DefaultOptions()\n\t\toptions.Options = &opts\n\t}\n\n\tset := token.NewFileSet()\n\tpkgs, err := parser.ParseDir(set, options.Path, nil, parser.ParseComments)\n\tif err != nil {\n\t\tt.Fatalf(\"failed to parse package: %w\", err)\n\t\treturn\n\t}\n\n\tpkg, ok := pkgs[options.PkgName]\n\tif !ok {\n\t\tt.Fatalf(\"unable to find %q in %v\", options.PkgName, pkgs)\n\t\treturn\n\t}\n\n\tvar result error\n\tfor fileName, f := range pkg.Files {\n\t\tvar docEntries []docEntry\n\t\tvar errorSlice []error\n\n\t\tif f.Doc != nil && strings.TrimSpace(f.Doc.Text()) != \"\" {\n\t\t\tdocEntries = append(docEntries, docEntry{\n\t\t\t\ttext: f.Doc.Text(),\n\t\t\t\tlineno: set.Position(f.Doc.Pos()).Line,\n\t\t\t})\n\t\t}\n\n\t\tast.Inspect(f, astInspectorFunc(set, &docEntries, &errorSlice, options))\n\n\t\tresult = multierror.Append(result, errorSlice...)\n\n\t\tfor _, doc := range docEntries {\n\t\t\tif err := testDocForFile(fileName, doc.lineno, doc.text, options); err != nil {\n\t\t\t\tresult = multierror.Append(result, err)\n\t\t\t}\n\t\t}\n\t}\n\n\tif result == nil {\n\t\treturn\n\t}\n\tif merr, ok := result.(*multierror.Error); ok {\n\t\tswitch merr.Len() {\n\t\tcase 0:\n\t\t\treturn\n\t\tcase 1:\n\t\t\tt.Fatal(merr.Errors[0])\n\t\t\treturn\n\t\t}\n\t}\n\tt.Fatal(result)\n}", "func Examples(s string) string {\n\tif len(s) == 0 {\n\t\treturn s\n\t}\n\treturn normalizer{s}.Trim().Indent().string\n}", "func loadExamples(path string) []*Example {\n\tfp, err := os.Open(path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer fp.Close()\n\n\tvar examples []*Example\n\tif err := json.NewDecoder(fp).Decode(&examples); err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn examples\n}", "func testSpec1() *v1.EnvironmentSpec {\n\treturn &v1.EnvironmentSpec{\n\t\tInfra: v1.InfraSpec{\n\t\t\tAZ: v1.AZSpec{\n\t\t\t\tSubscription: []v1.AZSubscription{\n\t\t\t\t\t{Name: \"dummy\", ID: \"12345\"},\n\t\t\t\t},\n\t\t\t},\n\t\t\tSource: v1.SourceSpec{\n\t\t\t\tType: \"local\",\n\t\t\t\tURL: \"../e2e/testdata/terraform\", // relative to dir containing this _test.go file.\n\t\t\t},\n\t\t\tMain: \"main.tf\",\n\t\t},\n\t\tDefaults: v1.ClusterSpec{\n\t\t\tInfra: v1.ClusterInfraSpec{\n\t\t\t\tX: map[string]string{\n\t\t\t\t\t\"overridden\": \"default\",\n\t\t\t\t\t\"notOverridden\": \"default\",\n\t\t\t\t},\n\t\t\t},\n\t\t\tAddons: v1.ClusterAddonSpec{\n\t\t\t\tSource: v1.SourceSpec{\n\t\t\t\t\tType: \"local\",\n\t\t\t\t\tURL: \"../e2e/testdata/addons\", // relative to dir containing this _test.go file.\n\t\t\t\t},\n\t\t\t\tJobs: []string{\n\t\t\t\t\t\"cluster/local/minikube/all.yaml\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\tClusters: []v1.ClusterSpec{\n\t\t\t{\n\t\t\t\tName: \"cpe\",\n\t\t\t\tInfra: v1.ClusterInfraSpec{\n\t\t\t\t\tX: map[string]string{\n\t\t\t\t\t\t\"overridden\": \"cpe-cluster\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t}, {\n\t\t\t\tName: \"second\",\n\t\t\t\tInfra: v1.ClusterInfraSpec{\n\t\t\t\t\tX: map[string]string{\n\t\t\t\t\t\t\"overridden\": \"second-cluster\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func TestSDAMSpec(t *testing.T) {\n\tfor _, subdir := range []string{\"single\", \"rs\", \"sharded\"} {\n\t\tfor _, file := range testhelpers.FindJSONFilesInDir(t, path.Join(testsDir, subdir)) {\n\t\t\trunTest(t, subdir, file)\n\t\t}\n\t}\n}", "func (suite *ExampleTestSuite) TestExample() {\n\tgo Handler(suite.listener)\n\t// suite.NoError(err)\n\tsuite.conn.Write([]byte(`<policy-file-request/>\\0`))\n\tvar a []byte = make([]byte, len(policy)/2)\n\tvar b []byte = make([]byte, len(policy)/2)\n\tsuite.conn.SetReadDeadline(time.Now().Add(3 * time.Second))\n\tsuite.conn.Read(a)\n\tsuite.conn.Read(b)\n\tvar buff bytes.Buffer\n\tbuff.Write(a)\n\tbuff.Write(b)\n\tsuite.Equal(policy, buff.Bytes())\n\tsuite.listener.Close()\n\t// suite.Equal(suite.VariableThatShouldStartAtFive, 5)\n}", "func TestCallFunc_arguments(t *testing.T) {\n\n}", "func main() {\n\ttest_plain_background()\n\ttest_cloud()\n\ttest_enemy()\n\ttest_move_background()\n\ttest_display_score()\n}", "func TestHelloWorldTable(t *testing.T) {\n\ttests := []struct{\n\t\tname\t\tstring \t//name of sub-test\n\t\trequest\t\tstring \t//what's ur request?\n\t\texpected\tstring\t//expectation\n\t}{\n\t\t{\n\t\t\tname: \"Aji\",\n\t\t\trequest: \"Aji\",\n\t\t\texpected: \"Hello Aji\",\n\t\t},\n\t\t{\n\t\t\tname: \"Wahidin\",\n\t\t\trequest: \"Wahidin\",\n\t\t\texpected: \"Hello Wahidin\",\n\t\t},\n\t}\n\n\tfor _, test := range tests {\n\t\tt.Run(test.name, func(t *testing.T) {\n\t\t\tresult := HelloWorld(test.request)\n\t\t\trequire.Equal(t, test.expected, result)\n\t\t})\n\t}\n}", "func partChecks(t *testing.T, f func(*parseTest) (Tag, bool)) {\n\tfor i, tt := range parseTests() {\n\t\ttag, skip := f(&tt)\n\t\tif skip {\n\t\t\tcontinue\n\t\t}\n\t\tif l, _ := language.ParseBase(tt.lang); l != tag.Tag().LangID {\n\t\t\tt.Errorf(\"%d: lang was %q; want %q\", i, tag.Tag().LangID, l)\n\t\t}\n\t\tif sc, _ := language.ParseScript(tt.script); sc != tag.Tag().ScriptID {\n\t\t\tt.Errorf(\"%d: script was %q; want %q\", i, tag.Tag().ScriptID, sc)\n\t\t}\n\t\tif r, _ := language.ParseRegion(tt.region); r != tag.Tag().RegionID {\n\t\t\tt.Errorf(\"%d: region was %q; want %q\", i, tag.Tag().RegionID, r)\n\t\t}\n\t\tv := tag.Tag().Variants()\n\t\tif v != \"\" {\n\t\t\tv = v[1:]\n\t\t}\n\t\tif v != tt.variants {\n\t\t\tt.Errorf(\"%d: variants was %q; want %q\", i, v, tt.variants)\n\t\t}\n\t\tif e := strings.Join(tag.Tag().Extensions(), \"-\"); e != tt.ext {\n\t\t\tt.Errorf(\"%d: extensions were %q; want %q\", i, e, tt.ext)\n\t\t}\n\t}\n}", "func runReleaseCases(t *testing.T, tests []releaseCase, rcmd releaseCmd) {\n\tvar buf bytes.Buffer\n\tfor _, tt := range tests {\n\t\tt.Run(tt.name, func(t *testing.T) {\n\t\t\tc := &helm.FakeClient{Rels: tt.rels}\n\t\t\tcmd := rcmd(c, &buf)\n\t\t\tcmd.ParseFlags(tt.flags)\n\t\t\terr := cmd.RunE(cmd, tt.args)\n\t\t\tif (err != nil) != tt.err {\n\t\t\t\tt.Errorf(\"expected error, got '%v'\", err)\n\t\t\t}\n\t\t\tre := regexp.MustCompile(tt.expected)\n\t\t\tif !re.Match(buf.Bytes()) {\n\t\t\t\tt.Errorf(\"expected\\n%q\\ngot\\n%q\", tt.expected, buf.String())\n\t\t\t}\n\t\t\tbuf.Reset()\n\t\t})\n\t}\n}", "func UsageExamples() string {\n\treturn os.Args[0] + ` neat-thing neat-thing-today` + \"\\n\" +\n\t\t\"\"\n}", "func run(t *testing.T, formatter Formatter, suites ...TCatcher) {\n\tvar (\n\t\tbeforeAllFound, afterAllFound bool\n\t\tbeforeAll, afterAll, before, after reflect.Value\n\t\ttotalPassed, totalFailed, totalPending, totalNoAssertions int\n\t)\n\n\tflag.Parse()\n\n\tfor _, s := range suites {\n\t\tbeforeAll, afterAll, before, after = reflect.Value{}, reflect.Value{}, reflect.Value{}, reflect.Value{}\n\t\ts.SetT(t)\n\t\ts.Reset()\n\n\t\tiType := reflect.TypeOf(s)\n\n\t\tformatter.PrintSuiteName(strings.Split(iType.String(), \".\")[1])\n\n\t\t// search for Before and After methods\n\t\tfor i := 0; i < iType.NumMethod(); i++ {\n\t\t\tmethod := iType.Method(i)\n\t\t\tif ok, _ := regexp.MatchString(\"^BeforeAll\", method.Name); ok {\n\t\t\t\tif !beforeAllFound {\n\t\t\t\t\tbeforeAll = method.Func\n\t\t\t\t\tbeforeAllFound = true\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tif ok, _ := regexp.MatchString(\"^AfterAll\", method.Name); ok {\n\t\t\t\tif !afterAllFound {\n\t\t\t\t\tafterAll = method.Func\n\t\t\t\t\tafterAllFound = true\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t}\n\t\t\tif ok, _ := regexp.MatchString(\"^Before\", method.Name); ok {\n\t\t\t\tbefore = method.Func\n\t\t\t}\n\t\t\tif ok, _ := regexp.MatchString(\"^After\", method.Name); ok {\n\t\t\t\tafter = method.Func\n\t\t\t}\n\t\t}\n\n\t\tif beforeAll.IsValid() {\n\t\t\tbeforeAll.Call([]reflect.Value{reflect.ValueOf(s)})\n\t\t}\n\n\t\tfor i := 0; i < iType.NumMethod(); i++ {\n\t\t\tmethod := iType.Method(i)\n\t\t\tif ok, _ := regexp.MatchString(*testToRun, method.Name); ok {\n\t\t\t\tif ok, _ := regexp.MatchString(formatter.AllowedMethodsPattern(), method.Name); ok {\n\n\t\t\t\t\ts.SetStatus(&Status{Code: STATUS_NO_ASSERTIONS})\n\n\t\t\t\t\tif before.IsValid() {\n\t\t\t\t\t\tbefore.Call([]reflect.Value{reflect.ValueOf(s)})\n\t\t\t\t\t}\n\n\t\t\t\t\tmethod.Func.Call([]reflect.Value{reflect.ValueOf(s)})\n\n\t\t\t\t\tif after.IsValid() {\n\t\t\t\t\t\tafter.Call([]reflect.Value{reflect.ValueOf(s)})\n\t\t\t\t\t}\n\n\t\t\t\t\tvar info *suiteInfo\n\t\t\t\t\tstatus := s.GetStatus()\n\n\t\t\t\t\tswitch status.Code {\n\t\t\t\t\tcase STATUS_PASS:\n\t\t\t\t\t\tinfo = s.GetInfo()\n\t\t\t\t\t\ttotalPassed++\n\t\t\t\t\tcase STATUS_FAIL:\n\t\t\t\t\t\tinfo = s.GetInfo()\n\t\t\t\t\t\tt.Error(status.ErrorMessage)\n\t\t\t\t\t\ttotalFailed++\n\t\t\t\t\tcase STATUS_PENDING:\n\t\t\t\t\t\tinfo = s.GetInfo()\n\t\t\t\t\t\tinfo.assertions = 0\n\t\t\t\t\t\ttotalPending++\n\t\t\t\t\tcase STATUS_NO_ASSERTIONS:\n\t\t\t\t\t\tinfo = &suiteInfo{0, method.Name}\n\t\t\t\t\t\ttotalNoAssertions++\n\t\t\t\t\t}\n\t\t\t\t\tformatter.PrintStatus(status, info)\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\n\t\tif afterAll.IsValid() {\n\t\t\tafterAll.Call([]reflect.Value{reflect.ValueOf(s)})\n\t\t}\n\t}\n\n\tformatter.PrintFinalReport(totalPassed, totalFailed, totalPending, totalNoAssertions)\n}", "func GetExamples(w http.ResponseWriter, r *http.Request) *appError {\n rendering.JSON(w, http.StatusOK, config.Examples)\n\n return nil\n}", "func Test_regex(t *testing.T) {\n\n\tfor _, val := range shouldMatch {\n\t\tif !secretKarmaReg.MatchString(val) {\n\t\t\tt.Errorf(\"String %s should have matched but didn't.\", val)\n\t\t}\n\t}\n\n\t// for _, val := range shouldNotMatch {\n\t// \tif secretKarmaReg.MatchString(val) {\n\t// \t\tt.Errorf(\"String %s should not have have matched but did.\", val)\n\t// \t}\n\t// }\n}", "func RunTests(matchString func(pat, str string) (bool, error), tests []InternalTest) (ok bool) {}", "func listTests(ctx context.Context, cfg *config.Config,\n\tdrv *driver.Driver,\n\tdutInfos map[string]*protocol.DUTInfo) ([]*resultsjson.Result, error) {\n\tCompanionFeatures := make(map[string]*frameworkprotocol.DUTFeatures)\n\tfor role, dutInfo := range dutInfos {\n\t\tif role != \"\" {\n\t\t\tCompanionFeatures[role] = dutInfo.GetFeatures()\n\t\t}\n\t}\n\n\tvar dutFeature *frameworkprotocol.DUTFeatures\n\tif _, ok := dutInfos[\"\"]; ok {\n\t\tdutFeature = dutInfos[\"\"].GetFeatures()\n\t}\n\n\ttests, err := drv.ListMatchedTests(ctx, cfg.Features(dutFeature, CompanionFeatures))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar shard *sharding.Shard\n\tif cfg.ShardMethod() == \"hash\" {\n\t\tshard = sharding.ComputeHash(tests, cfg.ShardIndex(), cfg.TotalShards())\n\t} else {\n\t\tshard = sharding.ComputeAlpha(tests, cfg.ShardIndex(), cfg.TotalShards())\n\t}\n\n\tvar testsToPrint []*driver.BundleEntity\n\tif cfg.ExcludeSkipped() {\n\t\ttestsToPrint, _ = removeSkippedTestsFromBundle(shard.Included)\n\t} else {\n\t\ttestsToPrint = shard.Included\n\t}\n\n\t// Convert driver.BundleEntity to resultsjson.Result.\n\tresults := make([]*resultsjson.Result, len(testsToPrint))\n\tfor i, re := range testsToPrint {\n\t\ttest, err := resultsjson.NewTest(re.Resolved.GetEntity())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresults[i] = &resultsjson.Result{\n\t\t\tTest: *test,\n\t\t\tSkipReason: strings.Join(re.Resolved.GetSkip().GetReasons(), \", \"),\n\t\t}\n\t}\n\treturn results, nil\n}", "func UsageExamples() string {\n\treturn os.Args[0] + ` organization list` + \"\\n\" +\n\t\tos.Args[0] + ` step list --id \"Quaerat aut.\"` + \"\\n\" +\n\t\tos.Args[0] + ` walkthrough list --id \"Aut ad doloribus possimus deleniti magni quia.\"` + \"\\n\" +\n\t\t\"\"\n}", "func TestTool(t *testing.T) {\n\ttests := []struct {\n\t\tit string\n\t\tmode Mode\n\t\tjob string\n\t\tsetValues yamlx.Values\n\t\tglobalValues string\n\t\ttemplates map[string]string\n\t\tvault getter\n\t\twant *fakeDoer\n\t}{\n\t\t{\n\t\t\tit: \"should_apply_one_doc_with_tmplt_scoped_values\",\n\t\t\tmode: ModeGenerate,\n\t\t\tjob: `\nsteps:\n- tmplt: tpl/example.txt\n values:\n audience: all\n team:\n lead: pipo`,\n\t\t\ttemplates: map[string]string{\n\t\t\t\t\"tpl/example.txt\": `\n{{ .Values.team.lead }} says hello {{ .Values.audience }}!`,\n\t\t\t},\n\t\t\twant: &fakeDoer{\n\t\t\t\tapply: []string{\"\\npipo says hello all!\"},\n\t\t\t},\n\t\t},\n\n\t\t{\n\t\t\tit: \"should_apply_one_doc_with_global_and_tmplt_scoped_values\",\n\t\t\tmode: ModeGenerate,\n\t\t\tjob: `\nsteps:\n- tmplt: tpl/example.txt\n values:\n team:\n lead: pipo\ndefaults:\n audience: all\n team:\n lead: klukkluk`,\n\t\t\tglobalValues: `\naudience: world\n`,\n\t\t\ttemplates: map[string]string{\n\t\t\t\t\"tpl/example.txt\": `\n{{ .Values.team.lead }} says hello {{ .Values.audience }}!`,\n\t\t\t},\n\t\t\twant: &fakeDoer{\n\t\t\t\tapply: []string{\"\\npipo says hello world!\"},\n\t\t\t},\n\t\t},\n\n\t\t{\n\t\t\tit: \"should_apply_one_doc_with_setvalue_overriding_all_others\",\n\t\t\tmode: ModeGenerate,\n\t\t\tjob: `\nsteps:\n- tmplt: tpl/example.txt\n values:\n name: pipo\ndefaults:\n name: klukkluk`,\n\t\t\tsetValues: yamlx.Values{\"name\": \"dikkedeur\"},\n\t\t\tglobalValues: `\nname: mamaloe\n`,\n\t\t\ttemplates: map[string]string{\n\t\t\t\t\"tpl/example.txt\": `\n{{ .Values.name }}`,\n\t\t\t},\n\t\t\twant: &fakeDoer{\n\t\t\t\tapply: []string{\"\\ndikkedeur\"},\n\t\t\t},\n\t\t},\n\n\t\t{\n\t\t\tit: \"should_wait\",\n\t\t\tjob: `\nsteps:\n- wait: --one 1 --two 2`,\n\t\t\twant: &fakeDoer{\n\t\t\t\twait: []string{\"--one 1 --two 2\"},\n\t\t\t},\n\t\t},\n\n\t\t{\n\t\t\tit: \"should_handle_action_with_portforward_arg\",\n\t\t\tmode: ModeGenerateWithActions,\n\t\t\tjob: `\nsteps:\n- action: action/get.txt\n portForward: --forward-flags\n values:\n type: getSecret`,\n\t\t\ttemplates: map[string]string{\n\t\t\t\t\"action/get.txt\": `\ntype: {{ .Values.type }}`,\n\t\t\t},\n\n\t\t\twant: &fakeDoer{\n\t\t\t\taction: []string{\"\\ntype: getSecret\"},\n\t\t\t\tportForward: []string{\"--forward-flags\"},\n\t\t\t\tpassedValues: yamlx.Values{},\n\t\t\t\tactionTally: 1,\n\t\t\t},\n\t\t},\n\n\t\t{\n\t\t\tit: \"should_handle_action_with_passed_values\",\n\t\t\tmode: ModeGenerateWithActions,\n\t\t\tjob: `\nsteps:\n- action: action/nop.txt\n- action: action/value.txt`,\n\t\t\ttemplates: map[string]string{\n\t\t\t\t\"action/nop.txt\": `\nno operation`,\n\t\t\t\t\"action/value.txt\": `\ntally: {{ .Get.tally }}`,\n\t\t\t},\n\t\t\twant: &fakeDoer{\n\t\t\t\taction: []string{\"\\nno operation\", \"\\ntally: 1\"},\n\t\t\t\tportForward: []string{\"\", \"\"},\n\t\t\t\tpassedValues: yamlx.Values{\"tally\": 1},\n\t\t\t\tactionTally: 2,\n\t\t\t},\n\t\t},\n\n\t\t{\n\t\t\tit: \"should_handle_reads_from_vault\",\n\t\t\tmode: ModeGenerateWithActions,\n\t\t\tjob: `\nsteps:\n- tmplt: tpl/vault.txt`,\n\t\t\ttemplates: map[string]string{\n\t\t\t\t\"tpl/vault.txt\": `\nsecret: {{ vault \"object\" \"field\" }}`,\n\t\t\t},\n\t\t\tvault: &fakeVault{\n\t\t\t\t\"object/field\": \"value\",\n\t\t\t},\n\t\t\twant: &fakeDoer{\n\t\t\t\tapply: []string{\"\\nsecret: value\"},\n\t\t\t},\n\t\t},\n\n\t\t{\n\t\t\tit: \"should_expand_variables_in_job\",\n\t\t\tmode: ModeGenerate,\n\t\t\tjob: `\nsteps:\n- tmplt: tpl/example.txt\n values:\n text: \"{{ .Values.first }}\" # note the quotes to make this valid yaml (arguably)\ndefaults:\n first: \"hello\"\n`,\n\t\t\ttemplates: map[string]string{\n\t\t\t\t\"tpl/example.txt\": `text={{ .Values.text }}`,\n\t\t\t},\n\t\t\twant: &fakeDoer{\n\t\t\t\tapply: []string{\"text=hello\"},\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, tst := range tests {\n\t\tt.Run(tst.it, func(t *testing.T) {\n\t\t\t// create function to read template file content.\n\t\t\treadFile := func(path string) (string, []byte, error) {\n\t\t\t\ts, ok := tst.templates[path]\n\t\t\t\tif !ok {\n\t\t\t\t\treturn \"\", nil, fmt.Errorf(\"not found: %s\", path)\n\t\t\t\t}\n\t\t\t\treturn path, []byte(s), nil\n\t\t\t}\n\n\t\t\tm := &fakeDoer{}\n\n\t\t\ttl := Tool{\n\t\t\t\tMode: tst.mode,\n\t\t\t\tEnviron: []string{},\n\t\t\t\tExecute: m,\n\t\t\t\treadFileFn: readFile,\n\t\t\t\tvault: tst.vault,\n\t\t\t}\n\n\t\t\terr := tl.run(tst.setValues, []byte(tst.globalValues), []byte(tst.job))\n\t\t\tif assert.NoError(t, err) {\n\t\t\t\tassert.Equal(t, tst.want, m)\n\t\t\t}\n\t\t})\n\t}\n}", "func spec(stream []string, uFeat string) []*structs.State {\n if stream == nil || uFeat == \".\" {\n return nil\n }\n uCat, _, _ := defeat(uFeat)\n found := search(stream, \"u\" + uCat, nil)\n return found\n}", "func TestDocs(docPath string, pathToCheck string) error {\n\tbuf := bytes.NewBuffer([]byte{})\n\tdate := time.Now().Format(\"2006-01-02\")\n\ttitle := out.Fmt(title, out.V{\"Command\": \"List of Integration Test Cases\", \"Description\": \"Auto generated list of all minikube integration tests and what they do.\", \"Date\": date})\n\t_, err := buf.Write([]byte(title))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = filepath.Walk(pathToCheck, func(path string, info os.FileInfo, err error) error {\n\t\tif info.IsDir() || !strings.HasSuffix(path, \".go\") {\n\t\t\treturn nil\n\t\t}\n\t\tfset := token.NewFileSet()\n\t\tr, e := os.ReadFile(path)\n\t\tif e != nil {\n\t\t\treturn errors.Wrap(e, fmt.Sprintf(\"error reading file %s\", path))\n\t\t}\n\t\tfile, e := parser.ParseFile(fset, \"\", r, parser.ParseComments)\n\t\tif e != nil {\n\t\t\treturn errors.Wrap(e, fmt.Sprintf(\"error parsing file %s\", path))\n\t\t}\n\n\t\tast.Inspect(file, func(x ast.Node) bool {\n\t\t\tif fd, ok := x.(*ast.FuncDecl); ok {\n\t\t\t\tfnName := fd.Name.Name\n\t\t\t\tif !shouldParse(fnName) {\n\t\t\t\t\treturn true\n\t\t\t\t}\n\t\t\t\ttd := parseFuncDocs(file, fd)\n\t\t\t\t_, e := buf.WriteString(td.toMarkdown())\n\t\t\t\tif e != nil {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn true\n\t\t})\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = os.WriteFile(docPath, buf.Bytes(), 0o644)\n\treturn err\n}", "func TestBasicMaking(t *testing.T) {\n\ttests := []testvector{\n\t\t{\n\t\t\t// Really basic mk operation.\n\t\t\tinput: \"testdata/test1.mk\",\n\t\t\toutput: \"testdata/test1.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// Environment variables are expanded in dependencies\n\t\t\tinput: \"testdata/test2.mk\",\n\t\t\toutput: \"testdata/test2.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// Variables defined in the mkfile are expanded in dependencies\n\t\t\tinput: \"testdata/test3.mk\",\n\t\t\toutput: \"testdata/test3.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// Pair of dependencies in an variable are expanded.\n\t\t\tinput: \"testdata/test4.mk\",\n\t\t\toutput: \"testdata/test4.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// \\ can escape newlines.\n\t\t\tinput: \"testdata/test5.mk\",\n\t\t\toutput: \"testdata/test5.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// variables can be set included from another file\n\t\t\tinput: \"testdata/test6.mk\",\n\t\t\toutput: \"testdata/test6.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// $vars are not expanded in import statements. expected to fail.\n\t\t\tinput: \"testdata/test7.mk\",\n\t\t\toutput: \"testdata/test7.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// Variables expanded in recipes.\n\t\t\tinput: \"testdata/test8.mk\",\n\t\t\toutput: \"testdata/test8.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// EOF can end a variable if no \\n present.\n\t\t\tinput: \"testdata/test9.mk\",\n\t\t\toutput: \"testdata/test9.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// External commands can generate variables\n\t\t\tinput: \"testdata/test10.mk\",\n\t\t\toutput: \"testdata/test10.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// mkfile variables are expanded in backquote substitution\n\t\t\tinput: \"testdata/test11.mk\",\n\t\t\toutput: \"testdata/test11.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// mkfile variables are expanded in backquote substitution\n\t\t\tinput: \"testdata/test13.mk\",\n\t\t\toutput: \"testdata/test13.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: false,\n\t\t},\n\t\t{\n\t\t\t// Rules can be created by pipeing commands\n\t\t\tinput: \"testdata/test14.mk\",\n\t\t\toutput: \"testdata/test14.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// Test alternative recipe shell\n\t\t\tinput: \"testdata/test15.mk\",\n\t\t\toutput: \"testdata/test15.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// Test alternative recipe shell\n\t\t\tinput: \"testdata/test16.mk\",\n\t\t\toutput: \"testdata/test16.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t\t{\n\t\t\t// Test\n\t\t\tinput: \"testdata/test17.mk\",\n\t\t\toutput: \"testdata/test17.mk.expected\",\n\t\t\terrors: \"\",\n\t\t\tpasses: true,\n\t\t},\n\t}\n\n\tfor _, tv := range tests {\n\t\t// TODO(rjk): Validate generated errors.\n\t\tgot, _, err := startMk(\"-n\", \"-f\", tv.input)\n\n\t\tif err != nil {\n\t\t\tif !tv.passes {\n\t\t\t\tt.Logf(\"%s expected failure\", tv.input)\n\t\t\t\tt.Logf(\"%s exec failed: %v\", tv.input, err)\n\t\t\t} else {\n\t\t\t\tt.Errorf(\"%s exec failed: %v\", tv.input, err)\n\t\t\t}\n\t\t}\n\n\t\tefd, err := os.Open(tv.output)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"%s can't open: %v\", tv.input, err)\n\t\t\tcontinue\n\t\t}\n\t\twant, err := ioutil.ReadAll(efd)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"%s can't read: %v\", tv.input, err)\n\t\t\tcontinue\n\t\t}\n\n\t\t// TODO(rjk): Read expected errors if they exist.\n\t\tif diff := cmp.Diff(string(want), string(got)); diff != \"\" {\n\t\t\tif !tv.passes {\n\t\t\t\tt.Logf(\"%s expected failure\", tv.input)\n\t\t\t\tt.Logf(\"%s: mismatch (-want +got):\\n%s\", tv.input, diff)\n\t\t\t} else {\n\t\t\t\tt.Errorf(\"%s: mismatch (-want +got):\\n%s\", tv.input, diff)\n\t\t\t}\n\t\t}\n\t}\n}", "func initIntegrationsTests(id string) {}", "func parseSpec(dir string) (*[]TestSpec, error) {\n\t// Read testspec directory for its files.\n\tfiles, err := ioutil.ReadDir(dir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Insert testspec data into struct.\n\ttests := []TestSpec{}\n\tfor _, file := range files {\n\t\ttest, err := createSpec(path.Join(dir, file.Name()))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif test != nil {\n\t\t\ttests = append(tests, *test)\n\t\t}\n\t}\n\n\treturn &tests, nil\n}", "func TestSBSMultipleResolutions(t *testing.T) {\n\ttest(t,\n\t\tusers(\"alice\", \"bob\", \"charlie\"),\n\t\tinPrivateTlf(\"alice,bob@twitter,charlie@twitter\"),\n\t\tas(alice,\n\t\t\tmkfile(\"alice.txt\", \"hello bob & charlie\"),\n\t\t),\n\t\tas(bob,\n\t\t\texpectError(initRoot(), \"bob does not have read access to directory /keybase/private/alice,bob@twitter,charlie@twitter\"),\n\t\t\tnoSync(),\n\t\t),\n\t\tas(charlie,\n\t\t\texpectError(initRoot(), \"charlie does not have read access to directory /keybase/private/alice,bob@twitter,charlie@twitter\"),\n\t\t\tnoSync(),\n\t\t),\n\n\t\taddNewAssertion(\"bob\", \"bob@twitter\"),\n\t\tas(alice,\n\t\t\t// TODO: Ideally, we wouldn't have to do this,\n\t\t\t// and we'd just wait for a rekey.\n\t\t\trekey(),\n\t\t),\n\n\t\tinPrivateTlfNonCanonical(\"alice,bob@twitter,charlie@twitter\", \"alice,bob,charlie@twitter\"),\n\t\tas(bob,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tmkfile(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t),\n\t\tas(alice,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t),\n\t\tas(charlie,\n\t\t\texpectError(initRoot(), \"charlie does not have read access to directory /keybase/private/alice,bob,charlie@twitter\"),\n\t\t\tnoSync(),\n\t\t),\n\n\t\tinPrivateTlf(\"alice,bob,charlie@twitter\"),\n\t\tas(bob,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tmkfile(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t),\n\t\tas(alice,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t),\n\t\tas(charlie,\n\t\t\texpectError(initRoot(), \"charlie does not have read access to directory /keybase/private/alice,bob,charlie@twitter\"),\n\t\t\tnoSync(),\n\t\t),\n\n\t\taddNewAssertion(\"charlie\", \"charlie@twitter\"),\n\t\tas(bob,\n\t\t\t// TODO: Ideally, we wouldn't have to do this,\n\t\t\t// and we'd just wait for a rekey.\n\t\t\trekey(),\n\t\t),\n\n\t\tinPrivateTlfNonCanonical(\"alice,bob@twitter,charlie@twitter\", \"alice,bob,charlie\"),\n\t\tas(charlie,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t\tmkfile(\"charlie1.txt\", \"hello alice & bob\"),\n\t\t),\n\t\tas(alice,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"charlie1.txt\", \"hello alice & bob\"),\n\t\t),\n\t\tas(bob,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"charlie1.txt\", \"hello alice & bob\"),\n\t\t),\n\n\t\tinPrivateTlf(\"alice,bob,charlie\"),\n\t\tas(charlie,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"charlie1.txt\", \"hello alice & bob\"),\n\t\t\tmkfile(\"charlie2.txt\", \"hello alice & bob\"),\n\t\t),\n\t\tas(alice,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"charlie1.txt\", \"hello alice & bob\"),\n\t\t\tread(\"charlie2.txt\", \"hello alice & bob\"),\n\t\t),\n\t\tas(bob,\n\t\t\tread(\"alice.txt\", \"hello bob & charlie\"),\n\t\t\tread(\"bob1.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"bob2.txt\", \"hello alice & charlie\"),\n\t\t\tread(\"charlie1.txt\", \"hello alice & bob\"),\n\t\t\tread(\"charlie2.txt\", \"hello alice & bob\"),\n\t\t),\n\t)\n}", "func FilterExamples(description string, lang string) string {\n\tif description == \"\" {\n\t\treturn \"\"\n\t}\n\n\tsource := []byte(description)\n\tparsed := schema.ParseDocs(source)\n\tfilterExamples(source, parsed, lang)\n\treturn schema.RenderDocsToString(source, parsed)\n}", "func TestMyCheat(t *testing.T) {\n\ta, b := // YOUR CODE HERE\n\n\n\t//Try to cheat verifier:\n\tproofForth := // YOUR CODE HERE\n\n\tv := gt.VerifyForth(a, b, proofForth)\n\n\tif !v {\n\t\tt.Fatal(\"My cheat does not work\")\n\t}\n}", "func CreateCoreIoTests() []interface{} {\n res := []interface{}{}\n res = addBinIoRoundtripTests( res )\n res = addBinIoSequenceRoundtripTests( res )\n res = addBinIoInvalidDataTests( res )\n return res\n}", "func InitDemo(s interf.Service) error {\n\n\t// first, update file index\n\terr := s.Update()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// create over 1000 small test files for Files() tests (if not exist)\n\tfor i := 1; i < 1011; i++ {\n\t\tname := fmt.Sprintf(\"small-test-file-%d.dat\", i)\n\t\t_, err := s.Files().ByName(name)\n\t\tif err != nil {\n\t\t\t_, err := s.Save(name, strings.NewReader(name), 0)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\t// create big random test files\n\trnd := rand.New(rand.NewSource(1337))\n\n\tname := \"big-test-file-150.dat\"\n\tsize := 150*1024*1024 + 1\n\t_, err = s.Files().ByName(name)\n\tif err != nil {\n\t\t_, err := s.Save(name, rnd, int64(size))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// create type test files (like text, image, ...)\n\tconst fuse = 128 * 1024 // 128 kB\n\tconst buffer = 16 * 1024 * 1024 // 16 MB\n\tconst comp = 1 * 1024 * 1024 // 1 MB\n\tconst bundle = 12 * 1024 * 1024 // 12 MB\n\n\tfor _, size := range []int{0, fuse, fuse - 1, fuse + 1, buffer, buffer - 1, buffer + 1, comp, comp - 1, comp + 1, bundle, bundle - 1, bundle + 1} {\n\t\tfor _, rate := range []float32{0.99, 0.66, 0.33, 0} {\n\t\t\t// data\n\t\t\tdata := make([]byte, size)\n\t\t\trnd.Read(data)\n\t\t\tfor i := 0; i < int(float32(size)*rate); i++ {\n\t\t\t\tdata[i] = 'B'\n\t\t\t}\n\t\t\tif len(data) > 0 {\n\t\t\t\tdata[0] = 'A'\n\t\t\t}\n\n\t\t\t// save\n\t\t\tname := fmt.Sprintf(\"special-file-%d-%f.dat\", size, rate)\n\t\t\t_, err = s.Files().ByName(name)\n\t\t\tif err != nil {\n\t\t\t\t_, err := s.Save(name, bytes.NewReader(data), 0)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\t// final update\n\terr = s.Update()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn nil\n}", "func runTestCases(ctx *gr.FCContext, evt map[string]string) ([]byte, error) {\n\tservingEndpoint := evt[\"servingEndpoint\"]\n\n\thttpClient := &http.Client{\n\t\tCheckRedirect: func(req *http.Request, via []*http.Request) error {\n\t\t\treturn http.ErrUseLastResponse\n\t\t},\n\t\tTransport: &http.Transport{\n\t\t\tMaxIdleConnsPerHost: 100,\n\t\t\tIdleConnTimeout: 50 * time.Second,\n\t\t},\n\t}\n\n\tresp, err := httpClient.Get(servingEndpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer resp.Body.Close()\n\tif resp.StatusCode == 404 || resp.StatusCode == 200 {\n\t\tbody, err := ioutil.ReadAll(resp.Body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbodyStr := strings.Replace(string(body), \"\\n\", \"\", -1)\n\t\treturn []byte(fmt.Sprintf(`{\"httpStatus\": %d, \"servingStatus\": \"succeeded\", \"body\": \"%s\"}`, resp.StatusCode, bodyStr)), nil\n\t}\n\treturn []byte(fmt.Sprintf(`{\"httpStatus\": %d, \"servingStatus\": \"succeeded\"}`, resp.StatusCode)), nil\n}", "func main() {\n\tcwd, err := os.Getwd()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\trootPath := path.Join(cwd, \"test262-parser-tests\")\n\tpassDone := make(chan bool)\n\tpassesDone := 0\n\n\tgo runFiles(path.Join(rootPath, \"pass\"), passDone, func(file string) bool {\n\t\tif shouldSkipPass(file) {\n\t\t\treturn true\n\t\t}\n\n\t\treturn assertFile(file, false)\n\t})\n\n\tgo runFiles(path.Join(rootPath, \"fail\"), passDone, func(file string) bool {\n\t\t//if shouldSkipPass(file) {\n\t\t//\treturn\n\t\t//}\n\n\t\treturn assertFile(file, true)\n\t})\n\n\tselect {\n\tcase <-passDone:\n\t\tpassesDone++\n\n\t\tif passesDone == 2 {\n\t\t\treturn\n\t\t}\n\t}\n}", "func runDefault(f *utils.BaseCustomFactory, logger log.Logger) error {\n\tlogger.Info(\"Run sub test 'default' of test 'enter'\")\n\tlogger.StartWait(\"Run test...\")\n\tdefer logger.StopWait()\n\n\tpods, err := f.Client.KubeClient().CoreV1().Pods(f.Namespace).List(metav1.ListOptions{})\n\tif err != nil {\n\t\treturn errors.Errorf(\"Unable to list the pods: %v\", err)\n\t}\n\n\tpodName := pods.Items[0].Name\n\n\tenterConfigs := []*cmd.EnterCmd{\n\t\t{\n\t\t\tGlobalFlags: &flags.GlobalFlags{\n\t\t\t\tNamespace: f.Namespace,\n\t\t\t\tNoWarn: true,\n\t\t\t\tSilent: true,\n\t\t\t},\n\t\t\tWait: true,\n\t\t\tContainer: \"container-0\",\n\t\t},\n\t\t{\n\t\t\tGlobalFlags: &flags.GlobalFlags{\n\t\t\t\tNamespace: f.Namespace,\n\t\t\t\tNoWarn: true,\n\t\t\t\tSilent: true,\n\t\t\t},\n\t\t\tWait: true,\n\t\t\tPod: podName,\n\t\t},\n\t\t{\n\t\t\tGlobalFlags: &flags.GlobalFlags{\n\t\t\t\tNamespace: f.Namespace,\n\t\t\t\tNoWarn: true,\n\t\t\t\tSilent: true,\n\t\t\t},\n\t\t\tWait: true,\n\t\t\tLabelSelector: \"app=test\",\n\t\t},\n\t\t{\n\t\t\tGlobalFlags: &flags.GlobalFlags{\n\t\t\t\tNamespace: f.Namespace,\n\t\t\t\tNoWarn: true,\n\t\t\t\tSilent: true,\n\t\t\t},\n\t\t\tWait: true,\n\t\t\tPick: true,\n\t\t},\n\t}\n\n\tfor _, c := range enterConfigs {\n\t\tdone := utils.Capture()\n\n\t\toutput := \"My Test Data\"\n\t\terr = c.Run(f, nil, []string{\"echo\", output})\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\ttime.Sleep(time.Second * 5)\n\n\t\tcapturedOutput, err := done()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif !strings.HasPrefix(capturedOutput, output) {\n\t\t\treturn errors.Errorf(\"capturedOutput '%s' is different than output '%s' for the enter cmd\", capturedOutput, output)\n\t\t}\n\t}\n\n\treturn nil\n}", "func assertModuleDocContent(t *testing.T, module context.TestcontainersModule, moduleDocFile string) {\n\tcontent, err := os.ReadFile(moduleDocFile)\n\tassert.Nil(t, err)\n\n\tlower := module.Lower()\n\ttitle := module.Title()\n\n\tdata := sanitiseContent(content)\n\tassert.Equal(t, data[0], \"# \"+title)\n\tassert.Equal(t, data[2], `Not available until the next release of testcontainers-go <a href=\"https://github.com/testcontainers/testcontainers-go\"><span class=\"tc-version\">:material-tag: main</span></a>`)\n\tassert.Equal(t, data[4], \"## Introduction\")\n\tassert.Equal(t, data[6], \"The Testcontainers module for \"+title+\".\")\n\tassert.Equal(t, data[8], \"## Adding this module to your project dependencies\")\n\tassert.Equal(t, data[10], \"Please run the following command to add the \"+title+\" module to your Go dependencies:\")\n\tassert.Equal(t, data[13], \"go get github.com/testcontainers/testcontainers-go/\"+module.ParentDir()+\"/\"+lower)\n\tassert.Equal(t, data[18], \"<!--codeinclude-->\")\n\tassert.Equal(t, data[19], \"[Creating a \"+title+\" container](../../\"+module.ParentDir()+\"/\"+lower+\"/\"+lower+\".go)\")\n\tassert.Equal(t, data[20], \"<!--/codeinclude-->\")\n\tassert.Equal(t, data[22], \"<!--codeinclude-->\")\n\tassert.Equal(t, data[23], \"[Test for a \"+title+\" container](../../\"+module.ParentDir()+\"/\"+lower+\"/\"+lower+\"_test.go)\")\n\tassert.Equal(t, data[24], \"<!--/codeinclude-->\")\n\tassert.Equal(t, data[28], \"The \"+title+\" module exposes one entrypoint function to create the \"+title+\" container, and this function receives two parameters:\")\n\tassert.True(t, strings.HasSuffix(data[31], \"(*\"+title+\"Container, error)\"))\n\tassert.Equal(t, \"for \"+title+\". E.g. `testcontainers.WithImage(\\\"\"+module.Image+\"\\\")`.\", data[44])\n}", "func RunIntruderTests(t *testing.T, ctx context.Context, td testdeck.TestCase, client interface{}, methodName string, req interface{}, data InputValidationTestData) {\n\n\t// get parameters of the sample request using reflection because we do not know the protobuf type\n\tfieldNames := reflect.TypeOf(req).Elem()\n\tfieldValues := reflect.ValueOf(req).Elem()\n\n\t// for each parameter field in this endpoint\n\tfor i := 0; i < fieldValues.NumField(); i++ {\n\t\tfieldName := fieldNames.Field(i).Name\n\n\t\t// skip this parameter if it is an automatically-generated field (field name starts with XXX)\n\t\tif strings.HasPrefix(fieldName, \"XXX\") {\n\t\t\tbreak\n\t\t}\n\n\t\t// run fuzz tests on this field\n\t\tTestThisField(t, ctx, td, client, methodName, req, fieldName, data)\n\t}\n}", "func batchProgrammingsTest(e *httpexpect.Expect, t *testing.T) {\n\ttestCases := []testCase{\n\t\tnotLoggedTestCase,\n\t\t{\n\t\t\tToken: testCtx.User.Token,\n\t\t\tStatus: http.StatusUnauthorized,\n\t\t\tSent: []byte(`{Pend}`),\n\t\t\tBodyContains: []string{\"Droits administrateur requis\"}},\n\t\t{\n\t\t\tToken: testCtx.Admin.Token,\n\t\t\tStatus: http.StatusInternalServerError,\n\t\t\tSent: []byte(`{Pend}`),\n\t\t\tBodyContains: []string{\"Batch programmation, décodage : \"}},\n\t\t//cSpell:disable\n\t\t{\n\t\t\tToken: testCtx.Admin.Token,\n\t\t\tStatus: http.StatusOK,\n\t\t\tSent: []byte(`{\"Programmings\": [\n\t\t\t{\"physical_op_id\":9,\"year\":2018,\"value\":100000000,\n\t\t\t\"commission_id\":7,\"total_value\":null,\"state_ratio\":null},\n\t\t\t{\"physical_op_id\":10,\"year\":2018,\"value\":200000000,\n\t\t\t\"commission_id\":8,\"total_value\":400000000,\"state_ratio\":null},\n\t\t\t{\"physical_op_id\":14 ,\"year\":2018,\"value\":300000000,\n\t\t\t\"commission_id\":3,\"total_value\":600000000,\"state_ratio\":0.35}],\n\t\t\t\"year\":2018}`),\n\t\t\tBodyContains: []string{\"Programmings\", `\"physical_op_id\":9`, `\"physical_op_id\":10`,\n\t\t\t\t`\"physical_op_id\":14`, `\"value\":200000000`, `\"commission_id\":8`, `\"total_value\":400000000`,\n\t\t\t\t`\"total_value\":null`, `\"state_ratio\":null`, `\"state_ratio\":0.35`}},\n\t}\n\t//cSpell:enable\n\tf := func(tc testCase) *httpexpect.Response {\n\t\treturn e.POST(\"/api/programmings/array\").\n\t\t\tWithHeader(\"Authorization\", \"Bearer \"+tc.Token).WithBytes(tc.Sent).Expect()\n\t}\n\tfor _, r := range chkTestCases(testCases, f, \"BatchProgrammings\") {\n\t\tt.Error(r)\n\t}\n}" ]
[ "0.6740999", "0.66714954", "0.65592945", "0.65583795", "0.64752537", "0.64752537", "0.6393368", "0.6335788", "0.6316314", "0.62687653", "0.62477225", "0.6188293", "0.6112569", "0.6095223", "0.5998042", "0.596894", "0.58857256", "0.5860294", "0.5762579", "0.56704915", "0.5660689", "0.5629446", "0.5625085", "0.56138253", "0.5583183", "0.5578782", "0.5574343", "0.5561499", "0.5560602", "0.5549406", "0.55282813", "0.55031526", "0.5502325", "0.54978245", "0.5467684", "0.54413885", "0.54170835", "0.54126585", "0.54123163", "0.5407496", "0.5404678", "0.53992045", "0.5393472", "0.5392698", "0.5383732", "0.5381226", "0.53420913", "0.5326993", "0.53223443", "0.53151584", "0.53035295", "0.53026175", "0.52930397", "0.52912784", "0.5290698", "0.5283299", "0.52722776", "0.5270921", "0.52637553", "0.52634984", "0.52627337", "0.5258015", "0.5256162", "0.52430826", "0.5243016", "0.524128", "0.52373296", "0.5229929", "0.52291644", "0.52251685", "0.5223965", "0.52212685", "0.52114", "0.5209331", "0.52066225", "0.52054894", "0.5202536", "0.51785123", "0.5177627", "0.5176298", "0.51673716", "0.5150198", "0.5147193", "0.51447123", "0.5144087", "0.5140848", "0.51325166", "0.51300067", "0.5125676", "0.51246256", "0.51234233", "0.5111962", "0.5106207", "0.51044065", "0.50953645", "0.50923556", "0.5089164", "0.50879806", "0.5080841", "0.5079382", "0.5079271" ]
0.0
-1
Type returns event type
func (e EventCommonKeyPressure) Type() EventType { return EventTypeCommonKeyPressure }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (e _event) Type() string { return e.Type_ }", "func (e *Event) Type() string {\n\treturn e.eventType\n}", "func (e Event) GetType() int {\n\treturn int(C.caca_get_event_type(e.Ev))\n}", "func (e *Event) Type() EventType {\n\treturn e.eventID\n}", "func (ce *CustomEvent) GetType() string {\n\treturn ce.eventType.String()\n}", "func (event *GGEvent) Type() int {\n\treturn (int)(event._type)\n}", "func (e *basicEvent) Type() Type {\n\treturn e.dtype\n}", "func (e *Event) GetType() string {\n\tif e == nil || e.Type == nil {\n\t\treturn \"\"\n\t}\n\treturn *e.Type\n}", "func (ev binlogEvent) Type() byte {\n\treturn ev.Bytes()[BinlogEventTypeOffset]\n}", "func (e *basicEvent) TypeString() string {\n\treturn e.etype\n}", "func (e *Event) GetType() EventType {\n\treturn EventType(C.CALL_MACRO_GST_EVENT_TYPE(e.g()))\n}", "func (bot) Type(t string) string {\n\tif !contains(knownBotEvents, t) {\n\t\tlog.Printf(\"[WARN] unknown bot event type: %q\", t)\n\t}\n\treturn strings.ToLower(fmt.Sprintf(bot_type_template, t))\n}", "func (bse *BaseEvent) Type(eventType string) {\n\tbse.EventType = eventType\n}", "func (event ParticipantStateChangedEvent) GetType() string {\n\treturn \"participantStateChanged\"\n}", "func (event FileEvent) GetType() string {\n\treturn \"file\"\n}", "func getMessageType(e *pb.Event) pb.EventType {\n\tswitch e.Event.(type) {\n\tcase *pb.Event_Register:\n\t\treturn pb.EventType_REGISTER\n\tcase *pb.Event_Block:\n\t\treturn pb.EventType_BLOCK\n\tcase *pb.Event_ChaincodeEvent:\n\t\treturn pb.EventType_CHAINCODE\n\tcase *pb.Event_Rejection:\n\t\treturn pb.EventType_REJECTION\n\tdefault:\n\t\treturn -1\n\t}\n}", "func (uem UnknownEventMetadata) Type() string {\n\treturn \"unknown\"\n}", "func (QueryEvent) Type() EventType {\n\treturn EventQuery\n}", "func (e *Event) GetTypeName() string {\n\treturn e.GetType().GetName()\n}", "func (MemberEvent) Type() EventType {\n\treturn EventMember\n}", "func (e Timing) Type() int {\n\treturn EventTiming\n}", "func (message *KeyDown) Type() tinygametools.EventName {\n\treturn TypeKeyDown\n}", "func (UserEvent) Type() EventType {\n\treturn EventUser\n}", "func (o FunctionEventTriggerOutput) EventType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v FunctionEventTrigger) *string { return v.EventType }).(pulumi.StringPtrOutput)\n}", "func (e event) EventType() eh.EventType {\n\treturn e.dbEvent.EventType\n}", "func (ErrorEvent) Type() EventType {\n\treturn EventError\n}", "func (l *Listener) Type() string {\n\treturn Type\n}", "func (e EventCommonPitchWheelChange) Type() EventType {\n\treturn EventTypeCommonPitchWheelChange\n}", "func (m *EventMutation) Type() string {\n\treturn m.typ\n}", "func (ge *GollumEvent) EventType() string {\n\treturn \"GollumEvent\"\n}", "func (e BaseEvent) EventType() EventType {\n\treturn e.eventType\n}", "func (e EventCommonSongPositionPointer) Type() EventType {\n\treturn EventTypeCommonSongPositionPointer\n}", "func (tc *Target) EventType(space string, name eventpkg.TypeName) *eventpkg.Type {\n\ttc.eventTypeCache.RLock()\n\tdefer tc.eventTypeCache.RUnlock()\n\treturn tc.eventTypeCache.cache[libkv.EventTypeKey{Space: space, Name: name}]\n}", "func (g *Github) GetEventType(ctx context.Context, r *http.Request) (string, error) {\n\tevent := r.Header.Get(\"X-Github-Event\")\n\tif len(event) == 0 {\n\t\treturn \"\", fmt.Errorf(\"failed to get event type\")\n\t}\n\treturn event, nil\n}", "func (o FunctionEventTriggerPtrOutput) EventType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *FunctionEventTrigger) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.EventType\n\t}).(pulumi.StringPtrOutput)\n}", "func (e *TodoRegistered) GetEventType() string {\n\treturn e.EventType\n}", "func (e event) EventType() eh.EventType {\n\treturn e.AggregateEvent.EventType\n}", "func (l *SpotListener) Type() string {\n\treturn l.listenerType\n}", "func (e WakeupEvent) EventType() string {\n\treturn \"consistency-check\"\n}", "func (e EventRealTimeMetaTempo) Type() EventType {\n\treturn EventTypeRealTimeMetaTempo\n}", "func (e EventCommonChannelPressure) Type() EventType {\n\treturn EventTypeCommonChannelPressure\n}", "func (et EventerType) String() string {\n\tswitch et {\n\tcase LogFile:\n\t\treturn \"file\"\n\tcase Journald:\n\t\treturn \"journald\"\n\tcase Memory:\n\t\treturn \"memory\"\n\tcase Null:\n\t\treturn \"none\"\n\tdefault:\n\t\treturn \"invalid\"\n\t}\n}", "func (t Type) String() string {\n\treturn fmt.Sprintf(`event.%v`, t.Name())\n}", "func (o *AlertingPredefinedEventFilter) GetEventType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.EventType\n}", "func (h EventHandlerFunc) HandlerType() EventHandlerType {\n\treturn EventHandlerType(fmt.Sprintf(\"handler-func-%v\", h))\n}", "func EventTypeString(s string) (EventType, error) {\n\tif val, ok := _EventTypeNameToValueMap[s]; ok {\n\t\treturn val, nil\n\t}\n\treturn 0, fmt.Errorf(\"%s does not belong to EventType values\", s)\n}", "func (s *Structured) GetType() string {\n\treturn s.cloudEvent.EventType\n}", "func (e PrecisionTiming) Type() int {\n\treturn EventPrecisionTiming\n}", "func (e *TodoCompleted) GetEventType() string {\n\treturn e.EventType\n}", "func (f *factory) Type() string {\n\treturn f.receiverType\n}", "func NameOfEvent(t uint32) string {\n\tswitch t {\n\tcase SessionCreate:\n\t\treturn \"SessionCreate\"\n\tcase SessionDestroy:\n\t\treturn \"SessionDestroy\"\n\tcase TopicPublish:\n\t\treturn \"TopicPublish\"\n\tcase TopicSubscribe:\n\t\treturn \"TopicSubscribe\"\n\tcase TopicUnsubscribe:\n\t\treturn \"TopicUnsubscribe\"\n\tcase QutoChange:\n\t\treturn \"QutoChange\"\n\tcase SessionResume:\n\t\treturn \"SessionResume\"\n\tcase AuthChange:\n\t\treturn \"AuthChange\"\n\tdefault:\n\t\treturn \"Unknown\"\n\t}\n}", "func (e EventRealTimeMetaChannelPrefix) Type() EventType {\n\treturn EventTypeRealTimeMetaChannelPrefix\n}", "func (e DriveAddedEvent) EventType() string {\n\treturn \"drive-added\"\n}", "func (e *TodoMessageChanged) GetEventType() string {\n\treturn e.EventType\n}", "func (x *fastReflection_EventReceive) Type() protoreflect.MessageType {\n\treturn _fastReflection_EventReceive_messageType\n}", "func (moduleAfterCreate) EventType() string {\n\treturn \"afterCreate\"\n}", "func (channelOnManual) EventType() string {\n\treturn \"onManual\"\n}", "func (channelAfterCreate) EventType() string {\n\treturn \"afterCreate\"\n}", "func (f *FunctionLike) Type() string {\n\tif f.macro {\n\t\treturn \"macro\"\n\t}\n\treturn \"function\"\n}", "func (*SaveApplicationLogEventRequest) TypeName() string {\n\treturn \"saveApplicationLogEvent\"\n}", "func (moduleOnManual) EventType() string {\n\treturn \"onManual\"\n}", "func (m *EventRSVPMutation) Type() string {\n\treturn m.typ\n}", "func (e EventType) String() string {\n\tswitch e {\n\tcase EventTypeGistCreated:\n\t\treturn \"gistCreated\"\n\tdefault:\n\t\tpanic(fmt.Sprintf(\"unexpected EventType %s, in string()\", string(e)))\n\t}\n\n}", "func (e *Element) Type() string {\n\treturn e.attrs.Get(\"type\")\n}", "func (t EventType) GetName() string {\n\treturn C.GoString((*C.char)(C.gst_event_type_get_name(C.GstEventType(t))))\n}", "func (e DriveReinstatedEvent) EventType() string {\n\treturn \"drive-reinstated\"\n}", "func (moduleBeforeCreate) EventType() string {\n\treturn \"beforeCreate\"\n}", "func (e WasCreated) GetType() string {\n\treturn fmt.Sprintf(\"%T\", e)\n}", "func (f *receiverFactory) Type() string {\n\treturn typeStr\n}", "func (el *EventListener) GetEventSourceType() apicommon.EventSourceType {\n\treturn apicommon.AMQPEvent\n}", "func (o *SyntheticMonitorUpdate) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (e *EventAPI) Get(name string) (*EventType, error) {\n\teventType := &EventType{}\n\terr := e.client.httpGET(e.backOffConf.create(), e.eventURL(name), eventType, \"unable to request event types\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn eventType, nil\n}", "func (channelBeforeCreate) EventType() string {\n\treturn \"beforeCreate\"\n}", "func (e DriveRemovedEvent) EventType() string {\n\treturn \"drive-removed\"\n}", "func (s *SaveApplicationLogEventRequest) GetType() (value string) {\n\tif s == nil {\n\t\treturn\n\t}\n\treturn s.Type\n}", "func (x *fastReflection_EventCreateClass) Type() protoreflect.MessageType {\n\treturn _fastReflection_EventCreateClass_messageType\n}", "func (s *Server) getEventType(w http.ResponseWriter, r *http.Request) {\n\ttmp := s.templates.Lookup(\"event_type_list.html\")\n\tUnableToFindHtmlTemplate(tmp)\n\tet, err := s.store.GetEventType()\n\tUnableToGetData(err)\n\ttempData := eventTypeData{\n\t\tEventType: et,\n\t}\n\terr = tmp.Execute(w, tempData)\n\tExcutionTemplateError(err)\n}", "func (b *EventBus) HandlerType() eh.EventHandlerType {\n\treturn \"eventbus\"\n}", "func (b *EventBus) HandlerType() eh.EventHandlerType {\n\treturn \"eventbus\"\n}", "func (c *InjectEvent) CommandType() eh.CommandType {\n\treturn InjectEventCommand\n}", "func (ce *CustomEvent) GetEventType() model.EventType {\n\treturn ce.eventType\n}", "func (e Created) EventType() string {\n\treturn entityName + \"Created\"\n}", "func (*SaveApplicationLogEventRequest) TypeID() uint32 {\n\treturn SaveApplicationLogEventRequestTypeID\n}", "func (a *AudioSampleEntryBox) Type() string {\n\treturn a.name\n}", "func (moduleAfterUpdate) EventType() string {\n\treturn \"afterUpdate\"\n}", "func eventType(t string) string {\n\t// defines the prefix for event type, organization name or FQDN or more extended possibility,\n\t// use harbor by default.\n\tprefix := \"harbor\"\n\treturn fmt.Sprintf(\"%s.%s\", prefix, t)\n}", "func (channelAfterUpdate) EventType() string {\n\treturn \"afterUpdate\"\n}", "func (channelBeforeUpdate) EventType() string {\n\treturn \"beforeUpdate\"\n}", "func (moduleBeforeUpdate) EventType() string {\n\treturn \"beforeUpdate\"\n}", "func (o SolutionOutput) EventType() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Solution) pulumi.StringPtrOutput { return v.EventType }).(pulumi.StringPtrOutput)\n}", "func (e Button) Type() int {\n\treturn ButtonType\n}", "func (m *TimerMutation) Type() string {\n\treturn m.typ\n}", "func (x *fastReflection_EventRetire) Type() protoreflect.MessageType {\n\treturn _fastReflection_EventRetire_messageType\n}", "func (consumer *Consumer) Type() string {\n\treturn consumer.data.Type\n}", "func (x *fastReflection_EventCancel) Type() protoreflect.MessageType {\n\treturn _fastReflection_EventCancel_messageType\n}", "func (subscription *EventSubscription_Spec_ARM) GetType() string {\n\treturn \"Microsoft.EventGrid/eventSubscriptions\"\n}", "func EventType(val string) zap.Field {\n\treturn zap.String(FieldEventType, val)\n}", "func (c *ContextCollector) GetEventTypes() []string {\n\treturn []string{\"*\"}\n}", "func (r *Rdispatch) Type() int8 {\n\treturn RdispatchTpe\n}", "func (e Event) Is(t EventType) bool {\n\treturn EventType(e.Type) == t\n}" ]
[ "0.8833918", "0.8231284", "0.80096775", "0.7661345", "0.7644077", "0.762825", "0.7621584", "0.75805354", "0.7489805", "0.73479295", "0.73060995", "0.7157061", "0.71027166", "0.7065338", "0.6895248", "0.68624276", "0.685866", "0.6844002", "0.68357396", "0.6811805", "0.6796279", "0.67931056", "0.67340696", "0.66833097", "0.6678585", "0.6673969", "0.6667518", "0.66643673", "0.6641968", "0.6592271", "0.65417063", "0.6501821", "0.64978725", "0.64907557", "0.6485545", "0.6461411", "0.64038676", "0.6392836", "0.63853365", "0.63790685", "0.6364767", "0.63505125", "0.6328191", "0.63116544", "0.6303897", "0.62859386", "0.62819105", "0.62722147", "0.6220366", "0.61696", "0.6167937", "0.61588824", "0.6156126", "0.613011", "0.6113018", "0.6098959", "0.6086891", "0.6083964", "0.60814476", "0.6074693", "0.60360813", "0.6035668", "0.60218954", "0.6002422", "0.5992769", "0.5978624", "0.59777105", "0.59666044", "0.5963192", "0.5949419", "0.5946459", "0.5943445", "0.5931083", "0.59100616", "0.5901375", "0.58754426", "0.58562475", "0.58530116", "0.58530116", "0.5850956", "0.58112687", "0.58051306", "0.5803901", "0.5800516", "0.5797982", "0.57971716", "0.5797069", "0.57880974", "0.5783087", "0.5764731", "0.5762701", "0.57605696", "0.57514894", "0.5750979", "0.5750371", "0.5748154", "0.5740225", "0.57276285", "0.57219833", "0.57183295" ]
0.65992755
29
Marshall marshalls event into writer
func (e EventCommonKeyPressure) Marshall(writer io.Writer) error { if err := marshallVarInt(writer, e.DeltaTime); err != nil { return err } _, err := writer.Write([]byte{ 0xa0 | (e.Channel & 0x0f), e.Key & 0x7f, e.Velocity & 0x7f, }) return err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (t Track) Marshall(writer io.Writer) error {\n\tvar buff bytes.Buffer\n\tfor _, event := range t.Events {\n\t\tif err := event.Marshall(&buff); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// write header\n\tbody := buff.Bytes()\n\tif _, err := writer.Write([]byte(\"MTrk\")); err != nil {\n\t\treturn err\n\t}\n\n\t// write body length\n\tif err := binary.Write(writer, binary.BigEndian, uint32(len(body))); err != nil {\n\t\treturn err\n\t}\n\n\t// write body\n\t_, err := writer.Write(body)\n\treturn err\n}", "func (e EventCommonPitchWheelChange) Marshall(writer io.Writer) error {\n\tif err := marshallVarInt(writer, e.DeltaTime); err != nil {\n\t\treturn err\n\t}\n\n\thh := uint8(e.Change >> 7)\n\tll := uint8(e.Change)\n\t_, err := writer.Write([]byte{\n\t\t0xe0 | (e.Channel & 0x0f),\n\t\tll & 0x7f,\n\t\thh & 0x7f,\n\t})\n\n\treturn err\n}", "func marshalEvents(events []event) ([]byte, error) {\n\tvar b strings.Builder\n\tb.WriteString(\"[\\n\")\n\tfor _, e := range events {\n\t\tb.WriteString(\" \")\n\t\tb.WriteString(formatIso(e.Time))\n\t\tb.WriteString(\": \")\n\t\tb.WriteString(e.Coins.String())\n\t\tb.WriteString(\"\\n\")\n\t}\n\tb.WriteString(\"]\")\n\treturn []byte(b.String()), nil\n}", "func (e EventRealTimeMetaChannelPrefix) Marshall(writer io.Writer) error {\n\tif err := marshallVarInt(writer, e.DeltaTime); err != nil {\n\t\treturn err\n\t}\n\n\t_, err := writer.Write([]byte{0xff, 0x20, 0x01, e.Prefix})\n\treturn err\n}", "func (e EventCommonChannelPressure) Marshall(writer io.Writer) error {\n\tif err := marshallVarInt(writer, e.DeltaTime); err != nil {\n\t\treturn err\n\t}\n\n\t_, err := writer.Write([]byte{\n\t\t0xd0 | (e.Channel & 0x0f),\n\t\te.Pressure & 0x7f,\n\t})\n\treturn err\n}", "func (e EventCommonSongPositionPointer) Marshall(writer io.Writer) error {\n\tif err := marshallVarInt(writer, e.DeltaTime); err != nil {\n\t\treturn err\n\t}\n\n\thh := uint8(e.Pointer >> 7)\n\tll := uint8(e.Pointer)\n\t_, err := writer.Write([]byte{\n\t\t0xf2,\n\t\tll & 0x7f,\n\t\thh & 0x7f,\n\t})\n\n\treturn err\n}", "func TestEventMarshaling(t *testing.T) {\n\tassert := asserts.NewTesting(t, asserts.FailStop)\n\n\tevtIn, err := mesh.NewEvent(\"test\")\n\tassert.NoError(err)\n\tdata, err := json.Marshal(evtIn)\n\tassert.NoError(err)\n\n\tevtOut := mesh.Event{}\n\terr = json.Unmarshal(data, &evtOut)\n\tassert.NoError(err)\n\tassert.Equal(evtOut, evtIn)\n\n\tplEvtA, err := mesh.NewEvent(\"payload-a\")\n\tassert.NoError(err)\n\tplEvtB, err := mesh.NewEvent(\"payload-b\")\n\tassert.NoError(err)\n\tplEvtC, err := mesh.NewEvent(\"payload-c\")\n\tassert.NoError(err)\n\n\tevtIn, err = mesh.NewEvent(\"test\", plEvtA, plEvtB, plEvtC)\n\tassert.NoError(err)\n\tdata, err = json.Marshal(evtIn)\n\tassert.NoError(err)\n\n\tevtOut = mesh.Event{}\n\terr = json.Unmarshal(data, &evtOut)\n\tassert.NoError(err)\n\tassert.Equal(evtOut, evtIn)\n\tpl := []mesh.Event{}\n\terr = evtOut.Payload(&pl)\n\tassert.NoError(err)\n\tassert.Equal(pl[0], plEvtA)\n\tassert.Equal(pl[1], plEvtB)\n\tassert.Equal(pl[2], plEvtC)\n}", "func serializeEvent(event EventInfo) ([]byte, error) {\n\tresult, err := json.Marshal(event)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result, nil\n}", "func (j *Event) MarshalJSONBuf(buf fflib.EncodingBuffer) error {\n\tif j == nil {\n\t\tbuf.WriteString(\"null\")\n\t\treturn nil\n\t}\n\tvar err error\n\tvar obj []byte\n\t_ = obj\n\t_ = err\n\tbuf.WriteString(`{\"type\":`)\n\n\t{\n\n\t\terr = j.TypeField.MarshalJSONBuf(buf)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\tbuf.WriteString(`,\"payload\":`)\n\t/* Falling back. type=types.Payload kind=map */\n\terr = buf.Encode(j.PayloadField)\n\tif err != nil {\n\t\treturn err\n\t}\n\tbuf.WriteByte('}')\n\treturn nil\n}", "func (v WSEvent) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson25363b2dEncodeGithubComDarkfoxs96OpenApiV3SdkOkexGoSdkApi6(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (m *EventMessageDetail) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (s Event) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.Attributes != nil {\n\t\tv := s.Attributes\n\n\t\tmetadata := protocol.Metadata{}\n\t\tms0 := e.Map(protocol.BodyTarget, \"attributes\", metadata)\n\t\tms0.Start()\n\t\tfor k1, v1 := range v {\n\t\t\tms0.MapSetValue(k1, protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v1)})\n\t\t}\n\t\tms0.End()\n\n\t}\n\tif s.EventType != nil {\n\t\tv := *s.EventType\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"eventType\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\tif s.Metrics != nil {\n\t\tv := s.Metrics\n\n\t\tmetadata := protocol.Metadata{}\n\t\tms0 := e.Map(protocol.BodyTarget, \"metrics\", metadata)\n\t\tms0.Start()\n\t\tfor k1, v1 := range v {\n\t\t\tms0.MapSetValue(k1, protocol.Float64Value(v1))\n\t\t}\n\t\tms0.End()\n\n\t}\n\tif s.Session != nil {\n\t\tv := s.Session\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"session\", v, metadata)\n\t}\n\tif s.Timestamp != nil {\n\t\tv := *s.Timestamp\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"timestamp\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\tif s.Version != nil {\n\t\tv := *s.Version\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"version\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\treturn nil\n}", "func (j *Event) MarshalJSON() ([]byte, error) {\n\tvar buf fflib.Buffer\n\tif j == nil {\n\t\tbuf.WriteString(\"null\")\n\t\treturn buf.Bytes(), nil\n\t}\n\terr := j.MarshalJSONBuf(&buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn buf.Bytes(), nil\n}", "func (e *Event) MarshalBinary() ([]byte, error) {\n\treturn e.MarshalJSON()\n}", "func (v Event) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson6601e8cdEncodeGithubComGoParkMailRu2018242GameServerTypes12(w, v)\n}", "func (v Event) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson6601e8cdEncodeGithubComGoParkMailRu2018242GameServerTypes12(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func marshal(event event.ImageUploaded) []byte {\n\tbytes, err := schema.ImageUploadedEvent.Marshal(event)\n\tSo(err, ShouldBeNil)\n\treturn bytes\n}", "func (m *AppliedAuthenticationEventListener) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n if m.GetEventType() != nil {\n cast := (*m.GetEventType()).String()\n err := writer.WriteStringValue(\"eventType\", &cast)\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"executedListenerId\", m.GetExecutedListenerId())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteObjectValue(\"handlerResult\", m.GetHandlerResult())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (j *JSONSerializer) MarshalEvent(v Event) (Record, Error) {\n\teventType, _ := EventType(v)\n\n\tdata, err := json.Marshal(v)\n\tif err != nil {\n\t\treturn Record{}, NewError(err, ErrorInvalidEncoding, \"unable to encode data\")\n\t}\n\n\tdata, err = json.Marshal(jsonEvent{\n\t\tType: eventType,\n\t\tData: json.RawMessage(data),\n\t})\n\tif err != nil {\n\t\treturn Record{}, NewError(err, ErrorInvalidEncoding, \"unable to encode event\")\n\t}\n\n\treturn Record{\n\t\tVersion: v.EventVersion(),\n\t\tData: data,\n\t}, nil\n}", "func (e *Event) packBytes() ([]byte, error) {\n\tdata := make([]interface{}, 2)\n\tdata[0] = e.Header\n\tdata[1] = e.Name\n\n\tfor _, a := range e.Args {\n\t\tdata = append(data, a)\n\t}\n\n\tvar buf []byte\n\n\tenc := codec.NewEncoderBytes(&buf, &mh)\n\tif err := enc.Encode(data); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn buf, nil\n}", "func (e EventRealTimeMetaTempo) Marshall(writer io.Writer) error {\n\tif err := marshallVarInt(writer, e.DeltaTime); err != nil {\n\t\treturn err\n\t}\n\n\thh := uint8(e.MicroSecondPerQuarterNote >> 16)\n\tmm := uint8(e.MicroSecondPerQuarterNote >> 8)\n\tll := uint8(e.MicroSecondPerQuarterNote)\n\t_, err := writer.Write([]byte{0xff, 0x51, 0x03, hh, mm, ll})\n\treturn err\n}", "func (e EventRealTimeTimingClock) Marshall(writer io.Writer) error {\n\tif err := marshallVarInt(writer, e.DeltaTime); err != nil {\n\t\treturn err\n\t}\n\n\t_, err := writer.Write([]byte{0xf8})\n\treturn err\n}", "func MarshalAddOn(object *AddOn, writer io.Writer) error {\n\tstream := helpers.NewStream(writer)\n\twriteAddOn(object, stream)\n\tstream.Flush()\n\treturn stream.Error\n}", "func (v WSEventResponse) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson25363b2dEncodeGithubComDarkfoxs96OpenApiV3SdkOkexGoSdkApi5(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (v WSEvent) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson25363b2dEncodeGithubComDarkfoxs96OpenApiV3SdkOkexGoSdkApi6(w, v)\n}", "func (aee *AuthEconEvent) Marshal() string {\n\tb, _ := json.Marshal(aee)\n\treturn string(b)\n}", "func (d *Device) WriteEvent(i interface{}) error {\n\tif d.file == nil {\n\t\treturn errors.New(\"device has not been initialized\")\n\t}\n\tbuf := new(bytes.Buffer)\n\terr := binary.Write(buf, binary.LittleEndian, i)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = d.file.Write(buf.Bytes())\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (e *EventLog) eventWriter() {\n\tgo func() {\n\t\tdefer func() {\n\t\t\tif r := recover(); r != nil {\n\t\t\t\tfmt.Println(\"Event writer panic: \", r)\n\t\t\t}\n\t\t}()\n\t\tfor event := range e.eventChannel {\n\t\t\theader := SchemaVersionHeader + e.schemaVersion + \"\\n\"\n\t\t\tif createdFlag := e.writeFile(event, header); createdFlag {\n\t\t\t\te.rotateEventLog()\n\t\t\t}\n\t\t}\n\t}()\n}", "func MarshalEvent(e Event) (Annotations, error) {\n\t// Handle event marshalers.\n\tif v, ok := e.(EventMarshaler); ok {\n\t\tas, err := v.MarshalEvent()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tas = append(as, Annotation{Key: SchemaPrefix + e.Schema()})\n\t\treturn as, nil\n\t}\n\n\tvar as Annotations\n\tflattenValue(\"\", reflect.ValueOf(e), func(k, v string) {\n\t\tas = append(as, Annotation{Key: k, Value: []byte(v)})\n\t})\n\tas = append(as, Annotation{Key: SchemaPrefix + e.Schema()})\n\treturn as, nil\n}", "func writeEvent(event smf.Event, prevEvent smf.Event, writer io.Writer) error {\n\n\t//add vlq length\n\twriter.Write(vlq.GetBytes(event.GetDTime()))\n\n\t//get status and event data\n\tstatus := []byte{event.GetStatus()}\n\teventData := event.GetData()\n\n\t//write events\n\tswitch castEvent := event.(type) {\n\tcase *smf.MIDIEvent:\n\t\t{\n\t\t\t//try cast prev event\n\t\t\tprevMidiEvent, ok := prevEvent.(*smf.MIDIEvent)\n\n\t\t\t//check running status\n\t\t\tif !ok || !(prevEvent.GetStatus() == status[0] && castEvent.GetChannel() == prevMidiEvent.GetChannel()) {\n\t\t\t\twriter.Write([]byte{status[0] | castEvent.GetChannel()})\n\t\t\t}\n\t\t}\n\tcase *smf.MetaEvent:\n\t\t{\n\t\t\twriter.Write(status)\n\t\t\twriter.Write([]byte{castEvent.GetMetaType()})\n\t\t\twriter.Write(vlq.GetBytes(uint32(len(eventData))))\n\t\t}\n\tcase *smf.SysexEvent:\n\t\t{\n\t\t\twriter.Write(status)\n\t\t\twriter.Write(vlq.GetBytes(uint32(len(eventData))))\n\t\t}\n\tdefault:\n\t\t{\n\t\t\treturn &WriteError{\"event type: \" + event.String() + \"not supported\"}\n\t\t}\n\t}\n\n\t//write event data\n\twriter.Write(eventData)\n\n\treturn nil\n}", "func (hh Events) ToWire() [][]byte {\n\tres := make([][]byte, len(hh))\n\tfor i, h := range hh {\n\t\tres[i] = h.Bytes()\n\t}\n\n\treturn res\n}", "func encodeMockEvent(e models.Event) ([]byte, error) {\n\tbyteBuffer, err := cbor.Marshal(e)\n\tif err != nil {\n\t\treturn []byte{}, err\n\t}\n\treturn byteBuffer, nil\n}", "func (m *DeviceManagementTroubleshootingEvent) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.Entity.Serialize(writer)\n if err != nil {\n return err\n }\n if m.GetAdditionalInformation() != nil {\n cast := make([]i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, len(m.GetAdditionalInformation()))\n for i, v := range m.GetAdditionalInformation() {\n if v != nil {\n cast[i] = v.(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable)\n }\n }\n err = writer.WriteCollectionOfObjectValues(\"additionalInformation\", cast)\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"correlationId\", m.GetCorrelationId())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteTimeValue(\"eventDateTime\", m.GetEventDateTime())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteStringValue(\"eventName\", m.GetEventName())\n if err != nil {\n return err\n }\n }\n {\n err = writer.WriteObjectValue(\"troubleshootingErrorDetails\", m.GetTroubleshootingErrorDetails())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func writeEvent(filename string, event *Event) (error) {\n\t_, err := os.Stat(filename)\n\tif os.IsNotExist(err) {\n\t\t_, err := os.Create(filename)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t}\n\tfile, err := os.OpenFile(filename, os.O_APPEND|os.O_WRONLY|os.O_CREATE, 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\tdata, _ := json.MarshalIndent(event, \"\", \"\\t\")\n\t_, err = file.Write(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (e Event) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"count\", e.Count)\n\tpopulateTimeRFC3339(objectMap, \"firstTimestamp\", e.FirstTimestamp)\n\tpopulateTimeRFC3339(objectMap, \"lastTimestamp\", e.LastTimestamp)\n\tpopulate(objectMap, \"message\", e.Message)\n\tpopulate(objectMap, \"name\", e.Name)\n\tpopulate(objectMap, \"type\", e.Type)\n\treturn json.Marshal(objectMap)\n}", "func (j *EventMsg) MarshalJSON() ([]byte, error) {\n\tvar buf fflib.Buffer\n\tif j == nil {\n\t\tbuf.WriteString(\"null\")\n\t\treturn buf.Bytes(), nil\n\t}\n\terr := j.MarshalJSONBuf(&buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn buf.Bytes(), nil\n}", "func (e *Event) MarshalJSON() ([]byte, error) {\n\n\tdata := \"\"\n\tif _h := hex.EncodeToString(e.Data); _h != \"\" && _h != strings.Repeat(\"0\", 64) {\n\t\tdata = fmt.Sprintf(\"0x%s\", _h)\n\t}\n\n\treturn []byte(fmt.Sprintf(`{\"origin\":%q,\"index\":%d,\"topics\":%v,\"data\":%q,\"txHash\":%q,\"blockHash\":%q}`,\n\t\te.Origin,\n\t\te.Index,\n\t\tstrings.Join(\n\t\t\tstrings.Fields(\n\t\t\t\tfmt.Sprintf(\"%q\", e.Topics)), \",\"),\n\t\tdata, e.TransactionHash, e.BlockHash)), nil\n\n}", "func (event FileEvent) MarshalJSON() ([]byte, error) {\n\ttype surrogate FileEvent\n\tpayload, err := json.Marshal(struct {\n\t\tsurrogate\n\t\tType string `json:\"type\"`\n\t}{\n\t\tsurrogate(event),\n\t\tevent.GetType(),\n\t})\n\treturn payload, errors.JSONMarshalError.Wrap(err)\n}", "func SerializeEvent(v Event) ([]byte, error) {\n\treturn marshal(v.EventName(), v)\n}", "func (e Event) MarshalJSON() ([]byte, error) {\n\ttype Alias Event // type alias to prevent infinite recursion\n\tev := Alias(e)\n\n\tif ev.Visibility != nil && *ev.Visibility == 1 {\n\t\tcp := *ev.Visibility\n\t\tev.Visibility = &cp\n\t\tev.Visibility = nil // don't serialize default value\n\t}\n\tif ev.Time != nil && ev.Time.Location() != time.UTC {\n\t\tcp := *ev.Time // copy prevents changing the original event\n\t\tev.Time = &cp\n\t\t*ev.Time = ev.Time.UTC() // normalize to UTC\n\t}\n\n\treturn json.Marshal(ev)\n}", "func marshalWriteRequest(samples []*prompb.TimeSeries) ([]byte, error) {\n\treq := &prompb.WriteRequest{\n\t\tTimeseries: samples,\n\t}\n\n\tdata, err := proto.Marshal(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcompressed := snappy.Encode(nil, data)\n\treturn compressed, nil\n}", "func (j *JSONSerializer) MarshalAll(events ...Event) (History, Error) {\n\thistory := make(History, 0, len(events))\n\n\tfor _, event := range events {\n\t\trecord, err := j.MarshalEvent(event)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\thistory = append(history, record)\n\t}\n\n\treturn history, nil\n}", "func (e *Event) WriteTo(w io.Writer) (int64, error) {\n\tsum := int64(0)\n\tvar nint int\n\tn, err := wrapNewlines(w, []byte(\"event: \"), []byte(e.Name))\n\tsum += n\n\tif err != nil {\n\t\treturn sum, err\n\t}\n\n\tif e.Data != nil {\n\t\tvar data []byte\n\t\tswitch v := e.Data.(type) {\n\t\tcase []byte:\n\t\t\tdata = v\n\t\tcase string:\n\t\t\tdata = []byte(v)\n\t\tdefault:\n\t\t\tvar err error\n\t\t\tdata, err = json.Marshal(e.Data)\n\t\t\tif err != nil {\n\t\t\t\treturn sum, err\n\t\t\t}\n\t\t}\n\t\tn, err := wrapNewlines(w, []byte(\"data: \"), data)\n\t\tsum += n\n\t\tif err != nil {\n\t\t\treturn sum, err\n\t\t}\n\t}\n\n\tn, err = wrapNewlines(w, []byte(\"id: \"), []byte(e.ID))\n\tsum += n\n\tif err != nil {\n\t\treturn sum, err\n\t}\n\n\tif e.Retry != 0 {\n\t\tnint, err = fmt.Fprintf(w, \"retry: %d\\n\", int64(e.Retry/time.Millisecond))\n\t\tsum += int64(nint)\n\t\tif err != nil {\n\t\t\treturn sum, err\n\t\t}\n\t}\n\n\tnint, err = w.Write([]byte(\"\\n\"))\n\tsum += int64(nint)\n\n\treturn sum, err\n}", "func (e *Event) WriteTo(w io.Writer) (int64, error) {\n\t// TODO(kevlar): buffer pool? cached encoding?\n\tbuf := new(bytes.Buffer)\n\tfmt.Fprintln(buf, \"id:\", e.ID)\n\tfmt.Fprintln(buf, \"event:\", e.Type)\n\n\tswitch d := e.Data.(type) {\n\tcase string:\n\t\tfor _, line := range strings.Split(d, \"\\n\") {\n\t\t\tfmt.Fprintln(buf, \"data:\", line)\n\t\t}\n\tcase nil:\n\tdefault:\n\t\tfmt.Fprintf(buf, \"data: \")\n\t\tif err := json.NewEncoder(buf).Encode(d); err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\tfmt.Fprintln(buf)\n\treturn buf.WriteTo(w)\n}", "func (v WSEventResponse) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjson25363b2dEncodeGithubComDarkfoxs96OpenApiV3SdkOkexGoSdkApi5(w, v)\n}", "func (j *EventMsg) MarshalJSONBuf(buf fflib.EncodingBuffer) error {\n\tif j == nil {\n\t\tbuf.WriteString(\"null\")\n\t\treturn nil\n\t}\n\tvar err error\n\tvar obj []byte\n\t_ = obj\n\t_ = err\n\tbuf.WriteString(`{\"SenderID\":`)\n\tfflib.WriteJsonString(buf, string(j.SenderID))\n\tbuf.WriteString(`,\"Event\":`)\n\n\t{\n\n\t\terr = j.Event.MarshalJSONBuf(buf)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\tbuf.WriteByte('}')\n\treturn nil\n}", "func (v Event) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonF642ad3eEncodeGithubComFrequencyanalyticsFrequency(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func eventWriter(k kafkametrics.Handler, c chan *kafkametrics.Event) {\n\tfor e := range c {\n\t\terr := k.PostEvent(e)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Error writing event: %s\\n\", err)\n\t\t}\n\t}\n}", "func (v Event) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonF642ad3eEncodeGithubComFrequencyanalyticsFrequency(w, v)\n}", "func (e AggregateEvent) MarshalEvent() (Annotations, error) {\n\t// Encode the entire event as JSON.\n\tdata, err := json.Marshal(e)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn Annotations{\n\t\t{Key: \"JSON\", Value: data},\n\t}, nil\n}", "func (e Event) MarshalJSON() ([]byte, error) {\n\tm := map[string]string{\"module\": e.Module, \"version\": e.Version}\n\treturn json.Marshal(&m)\n}", "func (m *RetentionEventStatus) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteObjectValue(\"error\", m.GetError())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n if m.GetStatus() != nil {\n cast := (*m.GetStatus()).String()\n err := writer.WriteStringValue(\"status\", &cast)\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func Encode(e Event, opt CodecOption) ([]byte, error) {\n\tre := rawEvent{\n\t\tType: e.GetType(),\n\t}\n\tif buf, err := json.Marshal(e); err != nil {\n\t\treturn nil, err\n\t} else {\n\t\t// Only support json format now\n\t\tre.Payload = buf\n\t\treturn json.Marshal(re)\n\t}\n}", "func (s ListViolationEventsOutput) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.NextToken != nil {\n\t\tv := *s.NextToken\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"nextToken\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\tif s.ViolationEvents != nil {\n\t\tv := s.ViolationEvents\n\n\t\tmetadata := protocol.Metadata{}\n\t\tls0 := e.List(protocol.BodyTarget, \"violationEvents\", metadata)\n\t\tls0.Start()\n\t\tfor _, v1 := range v {\n\t\t\tls0.ListAddFields(v1)\n\t\t}\n\t\tls0.End()\n\n\t}\n\treturn nil\n}", "func (s ViolationEvent) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.Behavior != nil {\n\t\tv := s.Behavior\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"behavior\", v, metadata)\n\t}\n\tif s.MetricValue != nil {\n\t\tv := s.MetricValue\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"metricValue\", v, metadata)\n\t}\n\tif s.SecurityProfileName != nil {\n\t\tv := *s.SecurityProfileName\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"securityProfileName\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\tif s.ThingName != nil {\n\t\tv := *s.ThingName\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"thingName\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\tif s.ViolationEventTime != nil {\n\t\tv := *s.ViolationEventTime\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"violationEventTime\",\n\t\t\tprotocol.TimeValue{V: v, Format: protocol.UnixTimeFormatName, QuotedFormatTime: true}, metadata)\n\t}\n\tif len(s.ViolationEventType) > 0 {\n\t\tv := s.ViolationEventType\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"violationEventType\", protocol.QuotedValue{ValueMarshaler: v}, metadata)\n\t}\n\tif s.ViolationId != nil {\n\t\tv := *s.ViolationId\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"violationId\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\treturn nil\n}", "func (e Data) MarshalJSON() ([]byte, error) {\n\teM := make(map[string]interface{})\n\n\teM[\"event\"] = e.event\n\n\tif e.Payload != nil {\n\t\teM[\"payload\"] = e.Payload\n\t} else {\n\t\teM[\"payload\"] = \"\"\n\t}\n\n\teM[\"message\"] = e.Message\n\n\tif e.Error != nil {\n\t\teM[\"error\"] = e.Error.Error()\n\t} else {\n\t\teM[\"error\"] = \"\"\n\t}\n\n\treturn json.Marshal(eM)\n}", "func (a *AvroEventBatchEncoder) EncodeDDLEvent(e *model.DDLEvent) (*MQMessage, error) {\n\treturn nil, nil\n}", "func (m *MobileAppTroubleshootingEvent) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n err := m.Entity.Serialize(writer)\n if err != nil {\n return err\n }\n if m.GetAppLogCollectionRequests() != nil {\n cast := make([]i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable, len(m.GetAppLogCollectionRequests()))\n for i, v := range m.GetAppLogCollectionRequests() {\n if v != nil {\n cast[i] = v.(i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.Parsable)\n }\n }\n err = writer.WriteCollectionOfObjectValues(\"appLogCollectionRequests\", cast)\n if err != nil {\n return err\n }\n }\n return nil\n}", "func marshal(event event.DimensionsInserted, c C) []byte {\n\tbytes, err := schema.DimensionsInsertedEvent.Marshal(event)\n\tc.So(err, ShouldBeNil)\n\treturn bytes\n}", "func (a Event_Payload) MarshalJSON() ([]byte, error) {\n\tvar err error\n\tobject := make(map[string]json.RawMessage)\n\n\tfor fieldName, field := range a.AdditionalProperties {\n\t\tobject[fieldName], err = json.Marshal(field)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, fmt.Sprintf(\"error marshaling '%s'\", fieldName))\n\t\t}\n\t}\n\treturn json.Marshal(object)\n}", "func (v managerListener) MarshalEasyJSON(w *jwriter.Writer) {\n\teasyjsonEd74d837EncodeGithubComKihamoBoggartComponentsBoggartInternalHandlers(w, v)\n}", "func (v EventLoadEventFired) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage70(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func Marshal(w io.Writer, v interface{}) (int, error) {\n if _, ok := v.(xdrType); ok {\n if bm, ok := v.(encoding.BinaryMarshaler); ok {\n b, err := bm.MarshalBinary()\n if err != nil {\n return 0, err\n }\n return w.Write(b)\n }\n }\n // delegate to xdr package's Marshal\n return xdr.Marshal(w, v)\n}", "func (e *Event) Bytes() []byte {\n\tif e.encoded == nil {\n\t\tvar err error\n\t\tprunedData := make(map[string]interface{})\n\t\tfor key, value := range e.data {\n\t\t\tif key == \"@metadata\" {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tprunedData[key] = value\n\t\t}\n\t\tif e.encoded, err = json.Marshal(prunedData); err != nil {\n\t\t\te.encoded = []byte(\"{\\\"tags\\\":[\\\"_encode_failure\\\"]}\")\n\t\t}\n\t}\n\treturn e.encoded\n}", "func (e *Event) Write(b []byte) (int, error) {\n\te.Data = append(e.Data, b...)\n\treturn len(b), nil\n}", "func (e *Encoder) Encode(index string, event *beat.Event) ([]byte, error) {\n\te.buf.Reset()\n\terr := e.folder.Fold(makeEvent(index, e.version, event))\n\tif err != nil {\n\t\te.reset()\n\t\treturn nil, err\n\t}\n\n\tjson := e.buf.Bytes()\n\tif !e.config.Pretty {\n\t\treturn json, nil\n\t}\n\n\tvar buf bytes.Buffer\n\tif err = stdjson.Indent(&buf, json, \"\", \" \"); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn buf.Bytes(), nil\n}", "func Marshal(v interface{}, w io.Writer, order binary.ByteOrder, length LengthType) (err error) {\n defer func() {\n if e := recover(); e != nil {\n switch v := e.(type) {\n case error:\n err = v\n case string:\n err = errors.New(\"marshal error:\" + v)\n default:\n panic(e) //repanic\n }\n }\n }()\n m := &marshaler{w:w, order:order}\n m.marshal(reflect.ValueOf(v), length())\n return nil\n}", "func (vertice *Vertice) Marshall(file *os.File, nbIndentation int) error {\n\tfmt.Fprintf(file, \"&uml.Vertice{\\n\")\n\tindent(file, nbIndentation)\n\tfmt.Fprintf(file, \"\\tX: %f,\\n\", vertice.X)\n\tindent(file, nbIndentation)\n\tfmt.Fprintf(file, \"\\tY: %f,\\n\", vertice.Y)\n\tindent(file, nbIndentation)\n\tfmt.Fprintf(file, \"},\\n\")\n\n\treturn nil\n}", "func (lump *Generic) Marshall() ([]byte, error) {\n\treturn lump.data, nil\n}", "func (w *Writer) encodePayload(e *xml.Encoder) error {\n\tif w.ObjectName == \"\" {\n\t\treturn e.Encode(w.Payload)\n\t}\n\treturn e.EncodeElement(w.Payload, xml.StartElement{Name: xml.Name{Local: w.ObjectName}})\n}", "func WireToEvents(buf [][]byte) Events {\n\tif buf == nil {\n\t\treturn nil\n\t}\n\n\thh := make(Events, len(buf))\n\tfor i, b := range buf {\n\t\thh[i] = BytesToEvent(b)\n\t}\n\n\treturn hh\n}", "func Write(w http.ResponseWriter, e *Event) error {\n\t_, err := e.WriteTo(w)\n\tw.(http.Flusher).Flush()\n\treturn err\n}", "func Marshal(marshal marshalfn) func(*Handler) {\n\treturn func(h *Handler) {\n\t\th.marshal = marshal\n\t}\n}", "func (w *Writer) Marshal(data interface{}) ([]byte, error) {\n\treturn w.encode(data)\n}", "func (v EventDocumentOpened) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage85(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func Marshal(fam byte, e Any) ([]byte, error) {\n\treturn e.marshal(fam)\n}", "func (v EventNavigatedWithinDocument) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage69(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (hm *headerMarshaller) Marshal(r *bytes.Buffer, ev wire.Event) error {\n\tconsensusEv := ev.(*Header)\n\tif err := encoding.WriteVarBytes(r, consensusEv.PubKeyBLS); err != nil {\n\t\treturn err\n\t}\n\n\treturn MarshalSignableVote(r, consensusEv)\n}", "func (c *EventCodec) MarshalEvent(ctx context.Context, event eh.Event) ([]byte, error) {\n\te := evt{\n\t\tEventType: event.EventType(),\n\t\tTimestamp: event.Timestamp(),\n\t\tAggregateType: event.AggregateType(),\n\t\tAggregateID: event.AggregateID().String(),\n\t\tVersion: event.Version(),\n\t\tMetadata: event.Metadata(),\n\t\tContext: eh.MarshalContext(ctx),\n\t}\n\n\t// Marshal event data if there is any.\n\tif event.Data() != nil {\n\t\tvar err error\n\t\tif e.RawData, err = json.Marshal(event.Data()); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not marshal event data: %w\", err)\n\t\t}\n\t}\n\n\t// Marshal the event (using JSON for now).\n\tb, err := json.Marshal(e)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not marshal event: %w\", err)\n\t}\n\n\treturn b, nil\n}", "func (m *UserSimulationEventInfo) Serialize(writer i878a80d2330e89d26896388a3f487eef27b0a0e6c010c493bf80be1452208f91.SerializationWriter)(error) {\n {\n err := writer.WriteStringValue(\"browser\", m.GetBrowser())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteTimeValue(\"eventDateTime\", m.GetEventDateTime())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"eventName\", m.GetEventName())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"ipAddress\", m.GetIpAddress())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"@odata.type\", m.GetOdataType())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteStringValue(\"osPlatformDeviceDetails\", m.GetOsPlatformDeviceDetails())\n if err != nil {\n return err\n }\n }\n {\n err := writer.WriteAdditionalData(m.GetAdditionalData())\n if err != nil {\n return err\n }\n }\n return nil\n}", "func (l *JSONFileWriter) Write(e *client.EventsResponse) error {\n\t// do not log if there is no events\n\tif len(e.Events) == 0 {\n\t\treturn nil\n\t}\n\n\tb, err := json.Marshal(e)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Println(\"un\", string(b))\n\n\tif _, err = l.f.Write(b); err != nil {\n\t\treturn err\n\t}\n\t_, err = l.f.Write([]byte(\"\\n\"))\n\treturn err\n}", "func (c EventOutputContext) MarshalEasyJSON(writer *ej_jwriter.Writer) {\n\tif err := c.Err(); err != nil {\n\t\twriter.Error = err\n\t\treturn\n\t}\n\twrappedWriter := jwriter.NewWriterFromEasyJSONWriter(writer)\n\tContextSerialization.MarshalToJSONWriterEventOutput(&wrappedWriter, &c)\n}", "func (e *encoder) marshal(t reflect.Type, v reflect.Value, n nestedTypeData) error {\n\tswitch t.Kind() {\n\tcase reflect.Ptr:\n\t\treturn e.marshalPointer(t, v, n)\n\tcase reflect.Slice:\n\t\treturn e.marshalVector(t, v, n)\n\tcase reflect.String:\n\t\treturn e.marshalString(v, n)\n\t}\n\tif isHandleType(t) {\n\t\treturn e.marshalHandle(v, n)\n\t}\n\tif isInterfaceType(t) || isInterfaceRequestType(t) {\n\t\t// An interface is represented by a Proxy, whose first field is\n\t\t// a zx.Channel, and we can just marshal that. Same goes for an\n\t\t// interface request, which is just an InterfaceRequest whose\n\t\t// first field is a zx.Channel.\n\t\treturn e.marshalHandle(v.Field(0), n)\n\t}\n\treturn e.marshalInline(t, v, n)\n}", "func (e *Extra) Marshal() ([]byte, error) {\n\tlog.Printf(\"[extra:marshal] Marshaling: %s\", e.name)\n\treturn []byte(e.name), nil\n}", "func (m ConsoleLogEvent) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 1)\n\n\taO0, err := swag.WriteJSON(m.RealtimeBuildEvent)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\t// now for regular properties\n\tvar propsConsoleLogEvent struct {\n\t\tLines []string `json:\"lines\"`\n\n\t\tStepRecordID strfmt.UUID `json:\"stepRecordId,omitempty\"`\n\n\t\tTimelineID strfmt.UUID `json:\"timelineId,omitempty\"`\n\n\t\tTimelineRecordID strfmt.UUID `json:\"timelineRecordId,omitempty\"`\n\t}\n\tpropsConsoleLogEvent.Lines = m.Lines\n\n\tpropsConsoleLogEvent.StepRecordID = m.StepRecordID\n\n\tpropsConsoleLogEvent.TimelineID = m.TimelineID\n\n\tpropsConsoleLogEvent.TimelineRecordID = m.TimelineRecordID\n\n\tjsonDataPropsConsoleLogEvent, errConsoleLogEvent := swag.WriteJSON(propsConsoleLogEvent)\n\tif errConsoleLogEvent != nil {\n\t\treturn nil, errConsoleLogEvent\n\t}\n\t_parts = append(_parts, jsonDataPropsConsoleLogEvent)\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (a ApplicationGatewayHTTPListener) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"etag\", a.Etag)\n\tpopulate(objectMap, \"id\", a.ID)\n\tpopulate(objectMap, \"name\", a.Name)\n\tpopulate(objectMap, \"properties\", a.Properties)\n\tpopulate(objectMap, \"type\", a.Type)\n\treturn json.Marshal(objectMap)\n}", "func (v EventContextChanged) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoWebaudio4(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (s *socket) writeEvent(ev *event, identity string) error {\n\tevBin, err := ev.MarshalBinary()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"event write\")\n\t}\n\n\ts.pipeInLck.Lock()\n\tdefer s.pipeInLck.Unlock()\n\n\tif atomic.LoadUint32(&s.stopping) == 1 {\n\t\treturn zmq4.ErrorSocketClosed\n\t}\n\n\tfor {\n\t\tif identity == \"\" {\n\t\t\t_, err = s.pipeIn.SendMessage(evBin)\n\t\t} else {\n\t\t\t_, err = s.pipeIn.SendMessage(identity, \"\", evBin)\n\t\t}\n\t\tif err != nil {\n\t\t\tif IsIntr(err) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\treturn errors.Wrap(err, \"event write\")\n\t\t}\n\t\tbreak\n\t}\n\n\treturn <-s.pipeErrs\n}", "func (b *SampleFJSONBuilder) Marshal(orig *SampleF) ([]byte, error) {\n\tret, err := b.Convert(orig)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn json.Marshal(ret)\n}", "func (a ApplicationGatewayListener) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"etag\", a.Etag)\n\tpopulate(objectMap, \"id\", a.ID)\n\tpopulate(objectMap, \"name\", a.Name)\n\tpopulate(objectMap, \"properties\", a.Properties)\n\tpopulate(objectMap, \"type\", a.Type)\n\treturn json.Marshal(objectMap)\n}", "func (j *Type) MarshalJSONBuf(buf fflib.EncodingBuffer) error {\n\tif j == nil {\n\t\tbuf.WriteString(\"null\")\n\t\treturn nil\n\t}\n\tvar err error\n\tvar obj []byte\n\t_ = obj\n\t_ = err\n\tbuf.WriteString(`{\"event_version\":`)\n\tfflib.WriteJsonString(buf, string(j.EventVersionField))\n\tbuf.WriteString(`,\"event_name\":`)\n\tfflib.WriteJsonString(buf, string(j.EventNameField))\n\tbuf.WriteByte('}')\n\treturn nil\n}", "func (i EventType) MarshalText() ([]byte, error) {\n\treturn []byte(i.String()), nil\n}", "func (v EventDomContentEventFired) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage84(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (e *eventRecorder) Write(b []byte) (int, error) {\n\tn, err := e.ResponseWriter.Write(b)\n\te.event.Size += int64(n)\n\treturn n, err\n}", "func (client *StdoutClient) Register(subscribeBuilder *SubscribeBuilder) error {\n\tsubscribe := struct {\n\t\tTopic string `json:\"topic\"`\n\t\tEventName string `json:\"name\"`\n\t\tTimestamp time.Time `json:\"timestamp\"`\n\t\tVersion int `json:\"version\"`\n\t}{\n\t\tTopic: subscribeBuilder.topic,\n\t\tEventName: subscribeBuilder.eventName,\n\t\tVersion: defaultVersion,\n\t\tTimestamp: time.Now().UTC(),\n\t}\n\n\teventByte, err := json.Marshal(&subscribe)\n\tif err != nil {\n\t\tlogrus.Errorf(\"unable to marshal event : %s , error : %v\", subscribe.EventName, err)\n\t}\n\n\tfmt.Println(string(eventByte))\n\n\treturn nil\n}", "func (s UpdateFunctionEventInvokeConfigInput) MarshalFields(e protocol.FieldEncoder) error {\n\te.SetValue(protocol.HeaderTarget, \"Content-Type\", protocol.StringValue(\"application/json\"), protocol.Metadata{})\n\n\tif s.DestinationConfig != nil {\n\t\tv := s.DestinationConfig\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"DestinationConfig\", v, metadata)\n\t}\n\tif s.MaximumEventAgeInSeconds != nil {\n\t\tv := *s.MaximumEventAgeInSeconds\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"MaximumEventAgeInSeconds\", protocol.Int64Value(v), metadata)\n\t}\n\tif s.MaximumRetryAttempts != nil {\n\t\tv := *s.MaximumRetryAttempts\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.BodyTarget, \"MaximumRetryAttempts\", protocol.Int64Value(v), metadata)\n\t}\n\tif s.FunctionName != nil {\n\t\tv := *s.FunctionName\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.PathTarget, \"FunctionName\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\tif s.Qualifier != nil {\n\t\tv := *s.Qualifier\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.QueryTarget, \"Qualifier\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\treturn nil\n}", "func (v EventLifecycleEvent) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonC5a4559bEncodeGithubComChromedpCdprotoPage71(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func notify(conn *websocket.Conn, name, event string) error {\n\tdata := [2]string{name, event}\n\treturn conn.WriteJSON(data)\n}", "func (se ServiceEvent) MarshalJSON() ([]byte, error) {\n\tse.Kind = KindServiceEvent\n\tobjectMap := make(map[string]interface{})\n\tif se.ServiceID != nil {\n\t\tobjectMap[\"ServiceId\"] = se.ServiceID\n\t}\n\tif se.EventInstanceID != nil {\n\t\tobjectMap[\"EventInstanceId\"] = se.EventInstanceID\n\t}\n\tif se.TimeStamp != nil {\n\t\tobjectMap[\"TimeStamp\"] = se.TimeStamp\n\t}\n\tif se.HasCorrelatedEvents != nil {\n\t\tobjectMap[\"HasCorrelatedEvents\"] = se.HasCorrelatedEvents\n\t}\n\tif se.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = se.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (w *SQLWriter) Write(event model.Event) error {\n\n\tpayloadText, err := w.Serializer.Serialize(event.Payload)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Storage.Exec(w.Storage.AppendStatement, event.SourceID.String(), event.Created, event.EventType, event.Version, payloadText)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}" ]
[ "0.6963643", "0.6264234", "0.5978546", "0.5833716", "0.58073133", "0.58000696", "0.57313836", "0.57079905", "0.5681556", "0.5660331", "0.5645458", "0.56430364", "0.5630469", "0.56105953", "0.55501556", "0.55430764", "0.55315626", "0.5526544", "0.5514821", "0.5507746", "0.5505744", "0.5495265", "0.54948664", "0.5488168", "0.5447157", "0.5436505", "0.54106784", "0.5374117", "0.5351989", "0.5350321", "0.53358454", "0.5328769", "0.52862537", "0.52860934", "0.52816534", "0.52686507", "0.5245573", "0.5239523", "0.52348125", "0.5230372", "0.52067816", "0.5201576", "0.5198257", "0.5196469", "0.5186949", "0.5170127", "0.5153721", "0.51505774", "0.51497346", "0.5142084", "0.5130531", "0.5124776", "0.5100108", "0.5079816", "0.5079763", "0.5072023", "0.50713015", "0.506785", "0.50608724", "0.5047905", "0.50435877", "0.5042222", "0.50291383", "0.50178903", "0.49963397", "0.49918815", "0.4991876", "0.49822903", "0.49792638", "0.49688792", "0.49562463", "0.49485254", "0.49398693", "0.49324945", "0.49324903", "0.49312395", "0.49301577", "0.49260682", "0.49247628", "0.4922934", "0.49141145", "0.49107355", "0.49101385", "0.48988184", "0.48981625", "0.4897108", "0.48970604", "0.48948434", "0.4891696", "0.48888063", "0.48851112", "0.48846626", "0.4876467", "0.48753572", "0.48624188", "0.4858571", "0.48555958", "0.4848269", "0.48453927", "0.4843801" ]
0.5629498
13
ReadResponse reads a server response into the received o.
func (o *EndpointDeleteReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) { switch response.Code() { case 204: result := NewEndpointDeleteNoContent() if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } return result, nil case 400: result := NewEndpointDeleteBadRequest() if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } return nil, result case 404: result := NewEndpointDeleteNotFound() if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } return nil, result case 500: result := NewEndpointDeleteInternalServerError() if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } return nil, result case 503: result := NewEndpointDeleteServiceUnavailable() if err := result.readResponse(response, consumer, o.formats); err != nil { return nil, err } return nil, result default: return nil, runtime.NewAPIError("unknown error", response, response.Code()) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (d *ResourceHandler) ReadResponse(dataOut unsafe.Pointer, bytesToRead int32, bytesRead *int32, callback *Callback) int32 {\n\treturn lookupResourceHandlerProxy(d.Base()).ReadResponse(d, dataOut, bytesToRead, bytesRead, callback)\n}", "func (o *GetServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *InteractionBindReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewInteractionBindOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewInteractionBindNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 500:\n\t\tresult := NewInteractionBindInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *InteractionUnbindReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewInteractionUnbindOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewInteractionUnbindNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 500:\n\t\tresult := NewInteractionUnbindInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (r *ResponseReader) ReadResponse(req *Request) (res *Response, err error) {\n\tres = CreateEmptyResponse(req)\n\t_, err = readFirstLine(r, res)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = readHeaders(r, res)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t_, err = readBodyContent(r, res)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn res, nil\n}", "func (c *Conn) ReadResponse(rmsg *Response) error {\n\tdata, err := c.ReadDataUnit()\n\tif err != nil {\n\t\treturn err\n\t}\n\tcolor.Printf(\"@{c}<!-- RESPONSE -->\\n%s\\n\\n\", string(data))\n\terr = xml.Unmarshal(data, rmsg)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// color.Fprintf(os.Stderr, \"@{y}%s\\n\", spew.Sprintf(\"%+v\", msg))\n\tif len(rmsg.Results) != 0 {\n\t\tr := rmsg.Results[0]\n\t\tif r.IsError() {\n\t\t\treturn r\n\t\t}\n\t}\n\treturn nil\n}", "func (o *VerifyConnectionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewVerifyConnectionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetAvailableReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetAvailableOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *ClosePositionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewClosePositionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewClosePositionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 401:\n\t\tresult := NewClosePositionUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewClosePositionNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 405:\n\t\tresult := NewClosePositionMethodNotAllowed()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *DescribeServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDescribeServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewDescribeServerBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewDescribeServerNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 409:\n\t\tresult := NewDescribeServerConflict()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewDescribeServerInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetServerSessionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetServerSessionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewGetServerSessionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewGetServerSessionUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetServerSessionNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewGetServerSessionInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tdata, err := ioutil.ReadAll(response.Body())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn nil, fmt.Errorf(\"Requested GET /dsmcontroller/namespaces/{namespace}/servers/{podName}/session returns an error %d: %s\", response.Code(), string(data))\n\t}\n}", "func (o *StartReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewStartOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (resp *PharosResponse) readResponse() {\n\tif !resp.hasBeenRead && resp.Response != nil && resp.Response.Body != nil {\n\t\tresp.data, resp.Error = ioutil.ReadAll(resp.Response.Body)\n\t\tresp.Response.Body.Close()\n\t\tresp.hasBeenRead = true\n\t}\n}", "func (o *HelloWorldReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHelloWorldOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewHelloWorldBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewHelloWorldInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (reader *BasicRpcReader) ReadResponse(r io.Reader, method string, requestID int32, resp proto.Message) error {\n\trrh := &hadoop.RpcResponseHeaderProto{}\n\terr := readRPCPacket(r, rrh, resp)\n\tif err != nil {\n\t\treturn err\n\t} else if int32(rrh.GetCallId()) != requestID {\n\t\treturn errors.New(\"unexpected sequence number\")\n\t} else if rrh.GetStatus() != hadoop.RpcResponseHeaderProto_SUCCESS {\n\t\treturn &NamenodeError{\n\t\t\tmethod: method,\n\t\t\tmessage: rrh.GetErrorMsg(),\n\t\t\tcode: int(rrh.GetErrorDetail()),\n\t\t\texception: rrh.GetExceptionClassName(),\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *UpdateAntivirusServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewUpdateAntivirusServerNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tresult := NewUpdateAntivirusServerDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *HasEventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHasEventsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewHasEventsUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewHasEventsForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetV2Reader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetV2OK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewGetV2InternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SaveReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewSaveNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 500:\n\t\tresult := NewSaveInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *TestWriteReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewTestWriteOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewTestWriteUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *AllConnectionsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewAllConnectionsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewAllConnectionsBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewAllConnectionsNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SendDataToDeviceReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSendDataToDeviceOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewSendDataToDeviceBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewSendDataToDeviceInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *HealthNoopReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHealthNoopOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PutOutOfRotationReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewPutOutOfRotationNoContent()\n\t\tresult.HttpResponse = response\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\terrorResult := kbcommon.NewKillbillError(response.Code())\n\t\tif err := consumer.Consume(response.Body(), &errorResult); err != nil && err != io.EOF {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, errorResult\n\t}\n}", "func (o *GetReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *ReplaceServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewReplaceServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 202:\n\t\tresult := NewReplaceServerAccepted()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewReplaceServerBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewReplaceServerNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewReplaceServerDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *StatusReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewStatusOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewStatusUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewStatusForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func ReadResponse(r *bfe_bufio.Reader, req *Request) (*Response, error) {\n\ttp := textproto.NewReader(r)\n\tresp := &Response{\n\t\tRequest: req,\n\t}\n\n\t// Parse the first line of the response.\n\tline, err := tp.ReadLine()\n\tif err != nil {\n\t\tif err == io.EOF {\n\t\t\terr = io.ErrUnexpectedEOF\n\t\t}\n\t\treturn nil, err\n\t}\n\tf := strings.SplitN(line, \" \", 3)\n\tif len(f) < 2 {\n\t\treturn nil, &badStringError{\"malformed HTTP response\", line}\n\t}\n\treasonPhrase := \"\"\n\tif len(f) > 2 {\n\t\treasonPhrase = f[2]\n\t}\n\tresp.Status = f[1] + \" \" + reasonPhrase\n\tresp.StatusCode, err = strconv.Atoi(f[1])\n\tif err != nil {\n\t\treturn nil, &badStringError{\"malformed HTTP status code\", f[1]}\n\t}\n\n\tresp.Proto = f[0]\n\tvar ok bool\n\tif resp.ProtoMajor, resp.ProtoMinor, ok = ParseHTTPVersion(resp.Proto); !ok {\n\t\treturn nil, &badStringError{\"malformed HTTP version\", resp.Proto}\n\t}\n\n\t// Parse the response headers.\n\tmimeHeader, err := tp.ReadMIMEHeader()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp.Header = Header(mimeHeader)\n\n\tfixPragmaCacheControl(resp.Header)\n\n\terr = readTransfer(resp, r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn resp, nil\n}", "func (o *PostChatroomsChannelHashReadReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewPostChatroomsChannelHashReadOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 403:\n\t\tresult := NewPostChatroomsChannelHashReadForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *TogglePacketGeneratorsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewTogglePacketGeneratorsCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *FrontPutBinaryReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewFrontPutBinaryOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SystemPingReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewSystemPingOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 500:\n\t\tresult := NewSystemPingInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SendDummyAlertReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewSendDummyAlertOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewSendDummyAlertBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewSendDummyAlertNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetViewsConnectionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetViewsConnectionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewGetViewsConnectionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *SyncCopyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSyncCopyOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewSyncCopyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *PostPatientsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewPostPatientsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewPostPatientsBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 405:\n\t\tresult := NewPostPatientsMethodNotAllowed()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (c *Conn) readResponse(res *response_) error {\n\terr := c.readDataUnit()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = IgnoreEOF(scanResponse.Scan(c.decoder, res))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif res.Result.IsError() {\n\t\treturn res.Result\n\t}\n\treturn nil\n}", "func (o *AllConnectionsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n switch response.Code() {\n \n case 200:\n result := NewAllConnectionsOK()\n if err := result.readResponse(response, consumer, o.formats); err != nil {\n return nil, err\n }\n return result, nil\n \n case 400:\n result := NewAllConnectionsBadRequest()\n if err := result.readResponse(response, consumer, o.formats); err != nil {\n return nil, err\n }\n return nil, result\n \n case 404:\n result := NewAllConnectionsNotFound()\n if err := result.readResponse(response, consumer, o.formats); err != nil {\n return nil, err\n }\n return nil, result\n \n default:\n return nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n }\n}", "func (o *GetMsgVpnReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetMsgVpnOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tresult := NewGetMsgVpnDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (r *Response) Read(p []byte) (n int, err error) {\n\n\tif r.Error != nil {\n\t\treturn -1, r.Error\n\t}\n\n\treturn r.RawResponse.Body.Read(p)\n}", "func (o *PostPciLinksMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostPciLinksMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostPciLinksMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *THSRAPIODFare2121Reader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewTHSRAPIODFare2121OK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 299:\n\t\tresult := NewTHSRAPIODFare2121Status299()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 304:\n\t\tresult := NewTHSRAPIODFare2121NotModified()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *PostGatewayConnectNetaddressReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewPostGatewayConnectNetaddressNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostGatewayConnectNetaddressDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *DNSGetReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDNSGetOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewDNSGetDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetGreetStatusReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetGreetStatusOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PostAPIV2EventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostAPIV2EventsNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewPostAPIV2EventsBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 403:\n\t\tresult := NewPostAPIV2EventsForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *CreateAntivirusServerReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewCreateAntivirusServerOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tresult := NewCreateAntivirusServerDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *PostCarsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewPostCarsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 405:\n\t\tresult := NewPostCarsMethodNotAllowed()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *ChatGetConnectedReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewChatGetConnectedOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewChatGetConnectedBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 401:\n\t\tresult := NewChatGetConnectedUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewChatGetConnectedNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *LogReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewLogOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewLogNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *WebModifyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewWebModifyOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 202:\n\t\tresult := NewWebModifyAccepted()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewWebModifyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetHyperflexServerModelsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetHyperflexServerModelsMoidOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetHyperflexServerModelsMoidNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewGetHyperflexServerModelsMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *KillQueryReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewKillQueryNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewKillQueryBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewKillQueryNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 422:\n\t\tresult := NewKillQueryUnprocessableEntity()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetProgressionViewReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetProgressionViewOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewGetProgressionViewBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *UpdateRackTopoReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUpdateRackTopoOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUpdateRackTopoBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetByUIDReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetByUIDOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetByUIDNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *UtilTestReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUtilTestOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetMeReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetMeOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewGetMeDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *Delete1Reader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewDelete1NoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewDelete1NotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *PostGatewayDisconnectNetaddressReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewPostGatewayDisconnectNetaddressNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostGatewayDisconnectNetaddressDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetProtocolsUsingGETReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetProtocolsUsingGETOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *RevokeReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewRevokeOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 401:\n\t\tresult := NewRevokeUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewRevokeNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *DestroySessionUsingPOSTReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDestroySessionUsingPOSTOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *CompleteTransactionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewCompleteTransactionNoContent()\n\t\tresult.HttpResponse = response\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\terrorResult := kbcommon.NewKillbillError(response.Code())\n\t\tif err := consumer.Consume(response.Body(), &errorResult); err != nil && err != io.EOF {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, errorResult\n\t}\n}", "func (o *GetMapNameEventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetMapNameEventsOK(o.writer)\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetMapNameEventsNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *RecoveryReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewRecoveryOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewRecoveryInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetPeersReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetPeersOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 403:\n\t\tresult := NewGetPeersForbidden()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *InstallEventsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewInstallEventsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *UpdateRackTopoReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUpdateRackTopoOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUpdateRackTopoBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 404:\n\t\tresult := NewUpdateRackTopoNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewUpdateRackTopoInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *GetVoicesReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetVoicesOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SetMemoRequiredReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSetMemoRequiredOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewSetMemoRequiredBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewSetMemoRequiredInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PatchHyperflexServerModelsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPatchHyperflexServerModelsMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPatchHyperflexServerModelsMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *BounceReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tresult := NewBounceDefault(response.Code())\n\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\treturn nil, err\n\t}\n\tif response.Code()/100 == 2 {\n\t\treturn result, nil\n\t}\n\treturn nil, result\n}", "func (o *PostHyperflexHxdpVersionsMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostHyperflexHxdpVersionsMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostHyperflexHxdpVersionsMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *DeleteApplianceRestoresMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDeleteApplianceRestoresMoidOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewDeleteApplianceRestoresMoidNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewDeleteApplianceRestoresMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetObmsLibraryIdentifierReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetObmsLibraryIdentifierOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewGetObmsLibraryIdentifierNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\tresult := NewGetObmsLibraryIdentifierDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *UserQuerySessionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUserQuerySessionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUserQuerySessionBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 500:\n\t\tresult := NewUserQuerySessionInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tdata, err := ioutil.ReadAll(response.Body())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn nil, fmt.Errorf(\"Requested GET /sessionbrowser/namespaces/{namespace}/gamesession returns an error %d: %s\", response.Code(), string(data))\n\t}\n}", "func (o *GetDiscoverReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetDiscoverOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *UnclaimTrafficFilterLinkIDReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUnclaimTrafficFilterLinkIDOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewUnclaimTrafficFilterLinkIDBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tcase 500:\n\t\tresult := NewUnclaimTrafficFilterLinkIDInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (r *overwriteConsumerReader) ReadResponse(resp runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tif r.forStatusCode == ForAllStatusCodes || resp.Code() == r.forStatusCode {\n\t\treturn r.requestReader.ReadResponse(resp, r.consumer)\n\t}\n\n\treturn r.requestReader.ReadResponse(resp, consumer)\n}", "func (o *ChangeaspecificSpeedDialReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 204:\n\t\tresult := NewChangeaspecificSpeedDialNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetDebugRequestReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewGetDebugRequestOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewGetDebugRequestNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PostMemoryArraysMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostMemoryArraysMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostMemoryArraysMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (c *Client) readResponse(conn net.Conn) ([]byte, error) {\n\tif c.Timeout > 0 {\n\t\t_ = conn.SetReadDeadline(time.Now().Add(c.Timeout))\n\t}\n\n\tproto := \"udp\"\n\tif _, ok := conn.(*net.TCPConn); ok {\n\t\tproto = \"tcp\"\n\t}\n\n\tif proto == \"udp\" {\n\t\tbufSize := c.UDPSize\n\t\tif bufSize == 0 {\n\t\t\tbufSize = dns.MinMsgSize\n\t\t}\n\t\tresponse := make([]byte, bufSize)\n\t\tn, err := conn.Read(response)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn response[:n], nil\n\t}\n\n\t// If we got here, this is a TCP connection\n\t// so we should read a 2-byte prefix first\n\treturn readPrefixed(conn)\n}", "func (o *PayReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewPayOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 400:\n\t\tresult := NewPayBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewPayNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 409:\n\t\tresult := NewPayConflict()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\tdata, err := ioutil.ReadAll(response.Body())\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\treturn nil, fmt.Errorf(\"Requested POST /platform/public/namespaces/{namespace}/payment/orders/{paymentOrderNo}/pay returns an error %d: %s\", response.Code(), string(data))\n\t}\n}", "func (o *CountReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewCountOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewCountBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *PostNodesIdentifierObmIdentifyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 201:\n\t\tresult := NewPostNodesIdentifierObmIdentifyCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewPostNodesIdentifierObmIdentifyNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\tresult := NewPostNodesIdentifierObmIdentifyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *DeleteEventsEventIDReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 204:\n\t\tresult := NewDeleteEventsEventIDNoContent()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 401:\n\t\tresult := NewDeleteEventsEventIDUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 404:\n\t\tresult := NewDeleteEventsEventIDNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *GetInterpreterReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetInterpreterOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewGetInterpreterNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *UtilityServiceReadyReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewUtilityServiceReadyOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewUtilityServiceReadyDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *SubscriptionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewSubscriptionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *HTTPGetPersistenceItemDataReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewHTTPGetPersistenceItemDataOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewHTTPGetPersistenceItemDataNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *FrontSessionReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewFrontSessionOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (pr *PushedRequest) ReadResponse(ctx context.Context) (*http.Response, error) {\n\tselect {\n\tcase <-ctx.Done():\n\t\tpr.Cancel()\n\t\tpr.pushedStream.bufPipe.CloseWithError(ctx.Err())\n\t\treturn nil, ctx.Err()\n\tcase <-pr.pushedStream.peerReset:\n\t\treturn nil, pr.pushedStream.resetErr\n\tcase resErr := <-pr.pushedStream.resc:\n\t\tif resErr.err != nil {\n\t\t\tfmt.Println(resErr.err.Error())\n\t\t\tpr.Cancel()\n\t\t\tpr.pushedStream.bufPipe.CloseWithError(resErr.err)\n\t\t\treturn nil, resErr.err\n\t\t}\n\t\tresErr.res.Request = pr.Promise\n\t\tresErr.res.TLS = pr.pushedStream.cc.tlsState\n\t\treturn resErr.res, resErr.err\n\t}\n}", "func (o *PostEquipmentIoExpandersMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 201:\n\t\tresult := NewPostEquipmentIoExpandersMoidCreated()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewPostEquipmentIoExpandersMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *DeleteFirmwareUpgradesMoidReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewDeleteFirmwareUpgradesMoidOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tcase 404:\n\t\tresult := NewDeleteFirmwareUpgradesMoidNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\tdefault:\n\t\tresult := NewDeleteFirmwareUpgradesMoidDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GetZippedReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tresult := NewGetZippedDefault(response.Code())\n\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\treturn nil, err\n\t}\n\tif response.Code()/100 == 2 {\n\t\treturn result, nil\n\t}\n\treturn nil, result\n}", "func (o *GetEtherPhysicalPortsReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewGetEtherPhysicalPortsOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\tdefault:\n\t\tresult := NewGetEtherPhysicalPortsDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *ZoneStreamReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\tcase 200:\n\t\tresult := NewZoneStreamOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"response status code does not match any response statuses defined for this endpoint in the swagger spec\", response, response.Code())\n\t}\n}", "func (o *ByNamespaceReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewByNamespaceOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 404:\n\t\tresult := NewByNamespaceNotFound()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}", "func (o *SystemDataUsageReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewSystemDataUsageOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 500:\n\t\tresult := NewSystemDataUsageInternalServerError()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\treturn nil, runtime.NewAPIError(\"unknown error\", response, response.Code())\n\t}\n}" ]
[ "0.76401824", "0.76067877", "0.7520219", "0.7508321", "0.7479227", "0.7471853", "0.7433047", "0.74236375", "0.73745817", "0.73665947", "0.7358356", "0.7354183", "0.73496455", "0.7345725", "0.7345279", "0.7338932", "0.73350984", "0.7322747", "0.73151827", "0.73143595", "0.7309159", "0.7307103", "0.72894984", "0.72862935", "0.72815233", "0.72737324", "0.72730523", "0.7265249", "0.7262986", "0.7261884", "0.7254393", "0.7249128", "0.72484225", "0.7247726", "0.7240097", "0.72237045", "0.72236663", "0.7218388", "0.72152305", "0.72129846", "0.7210294", "0.7208539", "0.7208441", "0.7199809", "0.7196856", "0.7196598", "0.71919525", "0.71768415", "0.7173413", "0.71732086", "0.71650475", "0.7154256", "0.7148943", "0.7148803", "0.7147453", "0.714295", "0.7142482", "0.7140463", "0.7139629", "0.71362376", "0.71358466", "0.71358275", "0.7135605", "0.7134447", "0.71322906", "0.71300906", "0.71237504", "0.7122628", "0.71191347", "0.71190774", "0.7118546", "0.7113784", "0.71044517", "0.7100435", "0.70982236", "0.7097546", "0.7097314", "0.7097285", "0.7095645", "0.70953536", "0.7092934", "0.7092836", "0.7087281", "0.7086462", "0.70831263", "0.7080391", "0.7077702", "0.7076653", "0.70763123", "0.70754945", "0.70695007", "0.70688057", "0.7067524", "0.7066842", "0.7066613", "0.7061669", "0.70610726", "0.7059043", "0.7056359", "0.70529985", "0.7049774" ]
0.0
-1
NewEndpointDeleteNoContent creates a EndpointDeleteNoContent with default headers values
func NewEndpointDeleteNoContent() *EndpointDeleteNoContent { return &EndpointDeleteNoContent{} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewDeleteEndpoint(s Service) goa.Endpoint {\n\treturn func(ctx context.Context, req interface{}) (interface{}, error) {\n\t\tp := req.(*DeletePayload)\n\t\treturn nil, s.Delete(ctx, p)\n\t}\n}", "func (a *DefaultApiService) DeleteEndpoint(ctx _context.Context, id string) ApiDeleteEndpointRequest {\n\treturn ApiDeleteEndpointRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tid: id,\n\t}\n}", "func DeleteEndpoint(w http.ResponseWriter, r *http.Request) {\r\n\tr.Header.Set(\"Content-Type\", \"application/json, charset=UTF-8\")\r\n\tvar dc DeleteConfig\r\n\r\n\tswitch r.Method {\r\n\tcase \"GET\":\r\n\t\tparams := mux.Vars(r)\r\n\t\tdc.Default()\r\n\t\tdc.Endpoint = params[\"endpoint_name\"]\r\n\t\tdc.EndpointList = append(dc.EndpointList, dc.Endpoint)\r\n\tcase \"POST\":\r\n\t\tbody, err := ioutil.ReadAll(r.Body)\r\n\t\tif err != nil {\r\n\t\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint, failed to read request\")\r\n\t\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Failed to read HTTP request\"))\r\n\t\t\treturn\r\n\t\t}\r\n\t\tdc.LoadParams(body)\r\n\t}\r\n\r\n\t// Verify Endpoint is provided in request body\r\n\tif len(dc.EndpointList) == 0 {\r\n\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint - endpoint is required\")\r\n\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Endpoint required\"))\r\n\t\treturn\r\n\t}\r\n\r\n\tvar query elastic.Query\r\n\r\n\t// Convert EndpointList to interface{} slice\r\n\tendpoints := make([]interface{}, len(dc.EndpointList))\r\n\tfor i, v := range dc.EndpointList {\r\n\t\tendpoints[i] = v\r\n\t}\r\n\r\n\tquery = elastic.NewBoolQuery().\r\n\t\tMust(elastic.NewWildcardQuery(\"CaseInfo.CaseName\", dc.CaseName),\r\n\t\t\telastic.NewTermsQuery(\"ComputerName.keyword\", endpoints...))\r\n\r\n\tdeleteEndpointByQuery(w, r, query, \"DeleteEndpoint\")\r\n\r\n}", "func (endpoint *HNSEndpoint) Delete() (*HNSEndpoint, error) {\n\toperation := \"Delete\"\n\ttitle := \"hcsshim::HNSEndpoint::\" + operation\n\tlogrus.Debugf(title+\" id=%s\", endpoint.Id)\n\n\treturn HNSEndpointRequest(\"DELETE\", endpoint.Id, \"\")\n}", "func DeleteEndpoint(serviceAccountProvider provider.ServiceAccountProvider, projectProvider provider.ProjectProvider, userInfoGetter provider.UserInfoGetter) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq, ok := request.(deleteReq)\n\t\tif !ok {\n\t\t\treturn nil, errors.NewBadRequest(\"invalid request\")\n\t\t}\n\t\terr := req.Validate()\n\t\tif err != nil {\n\t\t\treturn nil, errors.NewBadRequest(err.Error())\n\t\t}\n\t\tuserInfo, err := userInfoGetter(ctx, req.ProjectID)\n\t\tif err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\t\t// check if project exist\n\t\tif _, err := projectProvider.Get(userInfo, req.ProjectID, &provider.ProjectGetOptions{}); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\t// check if service account exist before deleting it\n\t\tif _, err := serviceAccountProvider.Get(userInfo, req.ServiceAccountID, nil); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\tif err := serviceAccountProvider.Delete(userInfo, req.ServiceAccountID); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\treturn nil, nil\n\t}\n}", "func (a *DefaultApiService) DeleteEndpointExecute(r ApiDeleteEndpointRequest) (*_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.DeleteEndpoint\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/endpoints/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.PathEscape(parameterToString(r.id, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ErrorNotFound\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, executionError\n}", "func Delete(client *gophercloud.ServiceClient, endpointID string) (r DeleteResult) {\n\t_, r.Err = client.Delete(endpointURL(client, endpointID), nil)\n\treturn\n}", "func (c *Client) Delete(ctx context.Context, p *DeletePayload) (err error) {\n\t_, err = c.DeleteEndpoint(ctx, p)\n\treturn\n}", "func (s *DefaultClient) Del(endpoint string) (resp *http.Response, err error) {\n\t_, resp, err = s.http(http.MethodDelete, endpoint, nil)\n\treturn\n}", "func EndpointDeleteMessage(e notifications.RegenNotificationInfo) AgentNotifyMessage {\n\tnotification := EndpointNotification{\n\t\tEndpointRegenNotification: EndpointRegenNotification{\n\t\t\tID: e.GetID(),\n\t\t\tLabels: e.GetOpLabels(),\n\t\t},\n\t\tPodName: e.GetK8sPodName(),\n\t\tNamespace: e.GetK8sNamespace(),\n\t}\n\n\treturn AgentNotifyMessage{\n\t\tType: AgentNotifyEndpointDeleted,\n\t\tNotification: notification,\n\t}\n}", "func (c *Client) Delete(endpoint string, resource ...RequestResource) error {\n\twrapper := newWrapper(\"delete\", endpoint, resource...)\n\treturn c.do(&wrapper)\n}", "func (c *V2) Delete(endpoint string) (*http.Response, error) {\n\treq, err := c.PrepareRequest(http.MethodDelete, endpoint, nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.Do(req)\n}", "func MakeDeleteEndpoint(s service.DepartmentService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func MakeDeleteEndpoint(s service.TodoService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func (c *APIClient) Delete(endpoint Endpointer, result interface{}, headers ...func(h *http.Header)) error {\n\tepoint, err := endpoint.GetEndpoint()\n\tif err != nil {\n\t\treturn err\n\t}\n\tendUrl := c.CompileEndpointURL(epoint)\n\treq, _ := http.NewRequest(http.MethodDelete, endUrl, nil)\n\n\th := &req.Header\n\tfor _, v := range headers {\n\t\tv(h)\n\t}\n\n\tresp, err := c.Client.Call(req)\n\tc.addResponseCode(resp.StatusCode)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn MakeResponseError(resp)\n\t}\n\n\terr = json.NewDecoder(resp.Body).Decode(result)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func deleteEndpoint(endpointId string) (string, error) {\n\tvar ret string\n\n\turi := fmt.Sprintf(\"%s/%s\", endpointURI, id)\n\turl, err := client.GetURL(uri)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\treq, err := client.BuildReq(nil, url, http.MethodDelete, true)\n\trawResp, err := client.HTTPClient().Do(req)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\tif _, err := client.ParseReq(rawResp); err != nil {\n\t\treturn ret, err\n\t}\n\n\t// Confirm the node was deleted.\n\tif rawResp.StatusCode != 204 {\n\t\treturn ret, errors.New(\"Endpoint was not able to be deleted\")\n\t}\n\tret = \"Endpoint deleted\\n\"\n\n\treturn ret, nil\n}", "func (client *NpmClient) EndpointDeleteReq(epinfo *netproto.Endpoint) (*netproto.Endpoint, error) {\n\treturn nil, nil\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn handleDeleteResponse(bts)\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(ctx context.Context, project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(ctx, path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn checkAPIResponse(bts, nil)\n}", "func DeleteEndpointInfo(ctx iris.Context) {\n\turi := ctx.Request().RequestURI\n\tfabricID := ctx.Params().Get(\"id\")\n\tif _, ok := capdata.FabricDataStore.Data[fabricID]; !ok {\n\t\terrMsg := fmt.Sprintf(\"Fabric data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Fabric\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tendpointData, ok := capdata.EndpointDataStore[uri]\n\tif !ok {\n\t\terrMsg := fmt.Sprintf(\"Endpoint data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Endpoint\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif endpointData.Endpoint.Links != nil && len(endpointData.Endpoint.Links.AddressPools) > 0 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be deleted as there are dependent upon AddressPool\")\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceCannotBeDeleted, errMsg, []interface{}{uri, \"Endpoint\"})\n\t\tctx.StatusCode(http.StatusNotAcceptable)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\t// Todo:Add the validation to verify the links\n\tdelete(capdata.EndpointDataStore, uri)\n\tctx.StatusCode(http.StatusNoContent)\n}", "func NewEndpointDeleteNotFound() *EndpointDeleteNotFound {\n\treturn &EndpointDeleteNotFound{}\n}", "func PostDeleteEndpoint(deleter deleting.Service) func(w http.ResponseWriter, r *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\tid := mux.Vars(r)[\"id\"]\n\n\t\t// Delete a post.\n\t\terr := deleter.PostDelete(id)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error deleting a post:\", err)\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tw.WriteHeader(http.StatusNoContent)\n\t}\n}", "func (*DeleteEndpointRequest) Descriptor() ([]byte, []int) {\n\treturn file_endpoint_api_proto_rawDescGZIP(), []int{11}\n}", "func MakeDeleteNodeEndpoint(s registry.Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteNodeRequest)\n\t\terr := s.DeleteNode(ctx, req.Token, req.Id)\n\t\treturn DeleteNodeResponse{Err: err}, nil\n\t}\n}", "func NewDeleteDataEventEndpoint(s Service, authJWTFn security.AuthJWTFunc) goa.Endpoint {\n\treturn func(ctx context.Context, req interface{}) (interface{}, error) {\n\t\tp := req.(*DeleteDataEventPayload)\n\t\tvar err error\n\t\tsc := security.JWTScheme{\n\t\t\tName: \"jwt\",\n\t\t\tScopes: []string{\"api:access\", \"api:admin\", \"api:ingestion\"},\n\t\t\tRequiredScopes: []string{\"api:access\"},\n\t\t}\n\t\tctx, err = authJWTFn(ctx, p.Auth, &sc)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, s.DeleteDataEvent(ctx, p)\n\t}\n}", "func Delete(uri string) error {\n\treq, err := http.NewRequest(\"DELETE\", Host+uri, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tres, err := Client.Do(req)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer res.Body.Close()\n\n\tif res.StatusCode != 204 {\n\t\treturn fmt.Errorf(\"got %d\", res.StatusCode)\n\t}\n\n\treturn nil\n}", "func DeleteProductEndPoint(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprintln(w, \"not implemented yet !\")\n}", "func Delete(dest interface{}, uri string, data url.Values) error {\n\treturn DefaultClient.Delete(dest, uri, data)\n}", "func (c *EmptyClient) Delete() *EmptyDelete {\n\tmutation := newEmptyMutation(c.config, OpDelete)\n\treturn &EmptyDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (ec *ExtensionClient) DeleteEndpoint(extensionID, serviceID, URL string) error {\n\n\turl := url.QueryEscape(URL)\n\trequest, err := extensionc.BuildDeleteEndpointPayload(extensionID, serviceID, url)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = ec.c.DeleteEndpoint()(context.Background(), request)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *fullEndpoint) delete(expectedOldValue *Endpoint) {\n\tatomic.CompareAndSwapPointer(&p.endpoint, unsafe.Pointer(expectedOldValue), nil)\n}", "func (nb *NetBuilder) DeleteEndpoint(nw *Network, ep *Endpoint) error {\n\t// Generate network name here as endpoint name is dependent upon network name.\n\tnw.Name = nb.generateHNSNetworkName(nw)\n\t// Query the namespace identifier.\n\tnsType, namespaceIdentifier := nb.getNamespaceIdentifier(ep)\n\n\t// Find the HNS endpoint ID.\n\tendpointName := nb.generateHNSEndpointName(nw.Name, namespaceIdentifier)\n\thnsEndpoint, err := hcsshim.GetHNSEndpointByName(endpointName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Detach the HNS endpoint from the container's network namespace.\n\tlog.Infof(\"Detaching HNS endpoint %s from container %s netns.\", hnsEndpoint.Id, ep.ContainerID)\n\tif nsType == hcsNamespace {\n\t\t// Detach the HNS endpoint from the namespace, if we can.\n\t\t// HCN Namespace and HNS Endpoint have a 1-1 relationship, therefore,\n\t\t// even if detachment of endpoint from namespace fails, we can still proceed to delete it.\n\t\terr = hcn.RemoveNamespaceEndpoint(namespaceIdentifier, hnsEndpoint.Id)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to detach endpoint, ignoring: %v\", err)\n\t\t}\n\t} else {\n\t\terr = hcsshim.HotDetachEndpoint(ep.ContainerID, hnsEndpoint.Id)\n\t\tif err != nil && err != hcsshim.ErrComputeSystemDoesNotExist {\n\t\t\treturn err\n\t\t}\n\n\t\t// The rest of the delete logic applies to infrastructure container only.\n\t\tif nsType == nonInfraContainerNS {\n\t\t\t// For non-infra containers, the network must not be deleted.\n\t\t\tnw.UseExisting = true\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// Delete the HNS endpoint.\n\tlog.Infof(\"Deleting HNS endpoint name: %s ID: %s\", endpointName, hnsEndpoint.Id)\n\t_, err = hnsEndpoint.Delete()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to delete HNS endpoint: %v.\", err)\n\t}\n\n\treturn err\n}", "func (c *Client) Delete(endpoint string, data map[string]interface{}) ([]byte, error) {\n\treservation := c.limiter.Reserve()\n\n\tif !reservation.OK() {\n\t\tduration := reservation.DelayFrom(time.Now())\n\t\treservation.Cancel()\n\t\treturn nil, fmt.Errorf(ErrRateLimited, duration.Milliseconds())\n\t}\n\n\tpayload, err := json.Marshal(data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treqUrl := c.client.BaseUrl + endpoint\n\terr = c.SignRequest(http.MethodDelete, endpoint, reqUrl, payload)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = c.client.Submit()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"error submitting request\")\n\t}\n\n\treservation.Cancel()\n\n\treturn c.client.Response.Body(), nil\n}", "func NewWithoutDefaults(endpoint string) BaseClient {\n\treturn BaseClient{\n\t\tClient: autorest.NewClientWithUserAgent(UserAgent()),\n\t\tEndpoint: endpoint,\n\t}\n}", "func (*DeleteEndpointResponse) Descriptor() ([]byte, []int) {\n\treturn file_endpoint_api_proto_rawDescGZIP(), []int{12}\n}", "func MakeDeleteAddressEndpoint(s Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(deleteAddressRequest)\n\t\te := s.DeleteAddress(ctx, req.ProfileID, req.AddressID)\n\t\treturn deleteAddressResponse{Err: e}, nil\n\t}\n}", "func NewEndpoint() *Endpoint {\r\n\t// Create a new Endpoint with an empty list of handler funcs.\r\n\treturn &Endpoint{\r\n\t\thandler: map[string]HandleFunc{},\r\n\t}\r\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsClientDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (*DeleteEndpointApiRequest) Descriptor() ([]byte, []int) {\n\treturn file_endpoint_api_proto_rawDescGZIP(), []int{2}\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func makeDeleteBookEndpoint(svc BookService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\t// convert request into a bookRequest\n\t\treq := request.(deleteBookRequest)\n\n\t\t// call actual service with data from the req\n\t\terr := svc.DeleteBook(req.BookId)\n\t\treturn deleteBookResponse{\n\t\t\tErr: err,\n\t\t}, nil\n\t}\n}", "func DeleteCustomerEndPoint(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprintln(w, \"not implemented yet !\")\n}", "func MakeDeleteSiteEndpoint(svc service.Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (response interface{}, err error) {\n\t\treq := request.(DeleteSiteRequest)\n\t\terr = svc.DeleteSite(ctx, req.SiteID)\n\t\treturn DeleteSiteResponse{Err: err}, nil\n\t}\n}", "func NewEndpoint(dnsName, recordType string, targets ...string) *Endpoint {\n\treturn NewEndpointWithTTL(dnsName, recordType, TTL(0), targets...)\n}", "func (c *ManagedServiceClient) Delete() *ManagedServiceDeleteRequest {\n\treturn &ManagedServiceDeleteRequest{\n\t\ttransport: c.transport,\n\t\tpath: c.path,\n\t}\n}", "func (c *TogglHttpClient) DeleteRequest(endpoint string, body interface{}) (*json.RawMessage, error) {\n\treturn request(c, \"DELETE\", endpoint, body)\n}", "func ExchangeDelete(service string, urlBase string, urlSuffix string, credentials string, goodHttpCodes []int) (httpCode int) {\n\turl := urlBase + \"/\" + urlSuffix\n\tapiMsg := http.MethodDelete + \" \" + url\n\n\t// get message printer\n\tmsgPrinter := i18n.GetMessagePrinter()\n\n\tVerbose(apiMsg)\n\tif IsDryRun() {\n\t\treturn 204\n\t}\n\n\thttpClient := GetHTTPClient(config.HTTPRequestTimeoutS)\n\n\tresp := InvokeRestApi(httpClient, http.MethodDelete, url, credentials, nil, service, apiMsg, make(map[string]string), true)\n\tif resp.Body != nil {\n\t\tdefer resp.Body.Close()\n\t}\n\n\t// delete never returns a body\n\thttpCode = resp.StatusCode\n\tVerbose(msgPrinter.Sprintf(\"HTTP code: %d\", httpCode))\n\tif !isGoodCode(httpCode, goodHttpCodes) {\n\t\tFatal(HTTP_ERROR, msgPrinter.Sprintf(\"bad HTTP code %d from %s\", httpCode, apiMsg))\n\t}\n\treturn\n}", "func NewDataDeleteNoContent() *DataDeleteNoContent {\n\treturn &DataDeleteNoContent{}\n}", "func Delete(h http.Handler) http.Handler {\n\treturn HTTP(h, DELETE)\n}", "func ExamplePrivateEndpointsClient_BeginDelete() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armstreamanalytics.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := clientFactory.NewPrivateEndpointsClient().BeginDelete(ctx, \"sjrg\", \"testcluster\", \"testpe\", nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t_, err = poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n}", "func (handler *BucketWebHandler) Delete(w http.ResponseWriter, r *http.Request) {\n\trespondWithError(w, http.StatusNotImplemented, \"Not implemented\", nil)\n}", "func DeleteHostNCApipaEndpoint(\n\tnetworkContainerID string) error {\n\treturn nil\n}", "func (handler *ObjectWebHandler) Delete(w http.ResponseWriter, r *http.Request) {\n\trespondWithError(w, http.StatusNotImplemented, \"Not implemented\", nil)\n}", "func TestServiceEndpointAzureRM_Delete_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := ResourceServiceEndpointAzureRM()\n\tfor _, resource := range azurermTestServiceEndpointsAzureRM {\n\t\tresourceData := getResourceData(t, resource)\n\t\tflattenServiceEndpointAzureRM(resourceData, &resource, azurermTestServiceEndpointAzureRMProjectID)\n\n\t\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\t\tclients := &client.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\t\texpectedArgs := serviceendpoint.DeleteServiceEndpointArgs{\n\t\t\tEndpointId: resource.Id,\n\t\t\tProjectIds: &[]string{\n\t\t\t\tazurermTestServiceEndpointAzureRMProjectID.String(),\n\t\t\t},\n\t\t}\n\n\t\tbuildClient.\n\t\t\tEXPECT().\n\t\t\tDeleteServiceEndpoint(clients.Ctx, expectedArgs).\n\t\t\tReturn(errors.New(\"DeleteServiceEndpoint() Failed\")).\n\t\t\tTimes(1)\n\n\t\terr := r.Delete(resourceData, clients)\n\t\trequire.Contains(t, err.Error(), \"DeleteServiceEndpoint() Failed\")\n\t}\n}", "func MakeDeleteProfileEndpoint(s Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(deleteProfileRequest)\n\t\te := s.DeleteProfile(ctx, req.ProfileID)\n\t\treturn deleteProfileResponse{Err: e}, nil\n\t}\n}", "func Delete(ctx context.Context, url string, body Body, options ...RequestOption) (*Response, error) {\n\tr, err := newRequest(ctx, http.MethodDelete, url, body, options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr.Header.Set(\"Content-Type\", body.ContentType())\n\treturn doRequest(http.DefaultClient, r)\n}", "func (v *DCHttpClient) Delete(\n\turl string, headers map[string]string) (response *DCHttpResponse, err error) {\n\treturn v.DoWithoutContent(http.MethodDelete, url, headers)\n}", "func (e EndPoint) Delete(container EndPointContainer) {\n\n\tentity := reflect.New(container.GetPrototype()).Interface()\n\tvar id int64\n\t_, err := fmt.Sscanf(container.GetRequest().URL.Query().Get(\":id\"), \"%d\", &id)\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusBadRequest)\n\t\treturn\n\t}\n\trepository := container.GetRepository()\n\n\terr = repository.FindByID(id, entity.(Entity))\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusNotFound)\n\t\treturn\n\t}\n\terr = container.GetSignal().Dispatch(&BeforeResourceDeleteEvent{})\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusInternalServerError)\n\t\treturn\n\t}\n\terr = repository.Delete(entity.(Entity))\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusInternalServerError)\n\t\treturn\n\t}\n\terr = container.GetSignal().Dispatch(&AfterResourceDeleteEvent{})\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusInternalServerError)\n\t\treturn\n\t}\n\tcontainer.GetResponseWriter().WriteHeader(http.StatusOK)\n}", "func (a *DefaultApiService) CreateEndpoint(ctx _context.Context) ApiCreateEndpointRequest {\n\treturn ApiCreateEndpointRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (e *EmailHeaderService) Delete(id int) (*Response, error) {\n\temailHeader := &EmailHeader{ID: id}\n\tendpoint := fmt.Sprintf(\"/assets/email/header/%d\", emailHeader.ID)\n\tresp, err := e.client.deleteRequest(endpoint, emailHeader)\n\treturn resp, err\n}", "func (a *CustomServicesApiService) Delete(ctx context.Context, technology string, id string) ( *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/customServices/{technology}/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"technology\"+\"}\", fmt.Sprintf(\"%v\", technology), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", fmt.Sprintf(\"%v\", id), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{ }\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\n\t\t}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tif ctx != nil {\n\t\t// API Key Authentication\n\t\tif auth, ok := ctx.Value(ContextAPIKey).(APIKey); ok {\n\t\t\tvar key string\n\t\t\tif auth.Prefix != \"\" {\n\t\t\t\tkey = auth.Prefix + \" \" + auth.Key\n\t\t\t} else {\n\t\t\t\tkey = auth.Key\n\t\t\t}\n\t\t\tlocalVarQueryParams.Add(\"Api-Token\", key)\n\t\t}\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\tdefer localVarHttpResponse.Body.Close()\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tbodyBytes, _ := ioutil.ReadAll(localVarHttpResponse.Body)\n\t\treturn localVarHttpResponse, reportError(\"Status: %v, Body: %s\", localVarHttpResponse.Status, bodyBytes)\n\t}\n\n\treturn localVarHttpResponse, err\n}", "func (*DeleteEndpointApiResponse) Descriptor() ([]byte, []int) {\n\treturn file_endpoint_api_proto_rawDescGZIP(), []int{3}\n}", "func (client CloudEndpointsClient) DeleteResponder(resp *http.Response) (result autorest.Response, err error) {\n\terr = autorest.Respond(\n\t\tresp,\n\t\tazure.WithErrorUnlessStatusCode(http.StatusOK, http.StatusAccepted, http.StatusNoContent),\n\t\tautorest.ByClosing())\n\tresult.Response = resp\n\treturn\n}", "func (c *Client) delete(endpoint string, out interface{}, q *WriteOptions) (*WriteMeta, error) {\n\tr, err := c.newRequest(\"DELETE\", endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr.setWriteOptions(q)\n\trtt, resp, err := requireOK(c.doRequest(r))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer resp.Body.Close()\n\n\twm := &WriteMeta{RequestTime: rtt}\n\tparseWriteMeta(resp, wm)\n\n\tif out != nil {\n\t\tif err := decodeBody(resp, &out); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn wm, nil\n}", "func (de *DefaultEndpointHandler) HandleDELETE(w http.ResponseWriter, r *http.Request, resources []string) {\n\thttp.Error(w, \"Method Not Allowed\", http.StatusMethodNotAllowed)\n}", "func Delete(url string) *THttpClient {\r\n\treturn NewHttpClient(url).Delete(\"\")\r\n}", "func RawDelete(restClient *rest.RESTClient, streams genericiooptions.IOStreams, url, filename string) error {\n\treturn raw(restClient, streams, url, filename, \"DELETE\")\n}", "func (c *Client) Delete(url string, headers, queryParams map[string][]string, data interface{}) (response *http.Response, err error) {\n\treturn c.makeRequest(url, http.MethodDelete, headers, queryParams, data)\n}", "func (a *DefaultApiService) DeleteTopicExecute(r ApiDeleteTopicRequest) (*_nethttp.Response, error) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.DeleteTopic\")\n\tif err != nil {\n\t\treturn nil, GenericOpenAPIError{error: err.Error()}\n\t}\n\n\tlocalVarPath := localBasePath + \"/topics/{topicName}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"topicName\"+\"}\", _neturl.PathEscape(parameterToString(r.topicName, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\treturn localVarHTTPResponse, err\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, nil\n}", "func (it *integTestSuite) TestNpmEndpointCreateDelete(c *C) {\n\t// create a network in controller\n\t// if not present create the default tenant\n\tit.CreateTenant(\"default\")\n\terr := it.DeleteAllEndpoints(c)\n\tc.Assert(err, IsNil)\n\terr = it.CreateNetwork(\"default\", \"default\", \"testNetwork\", \"10.1.0.0/22\", \"10.1.1.254\")\n\tc.Assert(err, IsNil)\n\tAssertEventually(c, func() (bool, interface{}) {\n\t\t_, nerr := it.npmCtrler.StateMgr.FindNetwork(\"default\", \"testNetwork\")\n\t\treturn (nerr == nil), nil\n\t}, \"Network not found in statemgr\")\n\n\t// wait till agent has the network\n\tfor _, ag := range it.agents {\n\t\tAssertEventually(c, func() (bool, interface{}) {\n\n\t\t\tnt := netproto.Network{\n\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\t\t\tObjectMeta: api.ObjectMeta{Tenant: \"default\", Namespace: \"default\", Name: \"testNetwork\"},\n\t\t\t}\n\t\t\t_, nerr := ag.dscAgent.PipelineAPI.HandleNetwork(agentTypes.Get, nt)\n\t\t\treturn (nerr == nil), nil\n\t\t}, \"Network not found in agent\")\n\t}\n\n\t// create a wait channel\n\twaitCh := make(chan error, it.numAgents*2)\n\n\t// create one endpoint from each agent\n\tfor i, ag := range it.agents {\n\t\tgo func(i int, ag *Dpagent) {\n\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\t\t\thostName := fmt.Sprintf(\"testHost-%d\", i)\n\n\t\t\t// make the call\n\t\t\tcerr := it.CreateEndpoint(\"default\", \"default\", \"testNetwork\", epname, hostName, \"0101.0101.0101\", hostName, \"20.1.1.1\", map[string]string{\"env\": \"production\", \"app\": \"procurement\"}, 2)\n\t\t\tif cerr != nil {\n\t\t\t\twaitCh <- fmt.Errorf(\"endpoint create failed: %v\", cerr)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(i, ag)\n\t}\n\n\t// wait for all endpoint creates to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Error during endpoint create\")\n\t}\n\n\t// wait for all endpoints to be propagated to other agents\n\tfor _, ag := range it.agents {\n\t\tgo func(ag *Dpagent) {\n\t\t\tfound := CheckEventually(func() (bool, interface{}) {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\treturn len(endpoints) == it.numAgents, nil\n\t\t\t}, \"10ms\", it.pollTimeout())\n\t\t\tif !found {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\tlog.Infof(\"Endpoint count expected [%v] found [%v]\", it.numAgents, len(endpoints))\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint count incorrect in datapath\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor i := range it.agents {\n\t\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\t\t\t\tepmeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t\tObjectMeta: api.ObjectMeta{\n\t\t\t\t\t\tTenant: \"default\",\n\t\t\t\t\t\tNamespace: \"default\",\n\t\t\t\t\t\tName: epname,\n\t\t\t\t\t},\n\t\t\t\t}\n\t\t\t\t_, perr := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.Get, epmeta)\n\t\t\t\tif perr != nil {\n\t\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint not found in datapath\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\twaitCh <- nil\n\t\t}(ag)\n\t}\n\n\t// wait for all goroutines to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint info incorrect in datapath\")\n\t}\n\n\t// now delete the endpoints\n\tfor i, ag := range it.agents {\n\t\tgo func(i int, ag *Dpagent) {\n\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\n\t\t\t// make the call\n\t\t\tcerr := it.DeleteEndpoint(\"default\", \"default\", epname)\n\t\t\tif cerr != nil {\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint delete failed: %v\", cerr)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(i, ag)\n\t}\n\n\t// wait for all endpoint deletes to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint delete failed\")\n\n\t}\n\n\tfor _, ag := range it.agents {\n\t\tgo func(ag *Dpagent) {\n\t\t\tif !CheckEventually(func() (bool, interface{}) {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\treturn len(endpoints) == 0, nil\n\t\t\t}, \"10ms\", it.pollTimeout()) {\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint was not deleted from datapath\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(ag)\n\t}\n\n\t// wait for all goroutines to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint delete error\")\n\t}\n\n\t// delete the network\n\terr = it.DeleteNetwork(\"default\", \"testNetwork\")\n\tc.Assert(err, IsNil)\n\tAssertEventually(c, func() (bool, interface{}) {\n\t\t_, nerr := it.npmCtrler.StateMgr.FindNetwork(\"default\", \"testNetwork\")\n\t\treturn (nerr != nil), nil\n\t}, \"Network still found in statemgr\")\n}", "func (client *WebAppsClient) deleteHostSecretCreateRequest(ctx context.Context, resourceGroupName string, name string, keyType string, keyName string, options *WebAppsDeleteHostSecretOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/host/default/{keyType}/{keyName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif keyType == \"\" {\n\t\treturn nil, errors.New(\"parameter keyType cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{keyType}\", url.PathEscape(keyType))\n\tif keyName == \"\" {\n\t\treturn nil, errors.New(\"parameter keyName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{keyName}\", url.PathEscape(keyName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewDeleteComponentNoContent() *DeleteComponentNoContent {\n\treturn &DeleteComponentNoContent{}\n}", "func NewDelete(appName string) *commander.CommandWrapper {\n\treturn &commander.CommandWrapper{\n\t\tHandler: &Delete{},\n\t\tHelp: &commander.CommandDescriptor{\n\t\t\tName: \"delete\",\n\t\t\tShortDescription: \"Delete a server.\",\n\t\t\tLongDescription: `Delete a server will destroy the world and container and the version file.`,\n\t\t\tArguments: \"name\",\n\t\t\tExamples: []string{\"\", \"my_server\"},\n\t\t},\n\t}\n}", "func (client *WebAppsClient) deletePrivateEndpointConnectionCreateRequest(ctx context.Context, resourceGroupName string, name string, privateEndpointConnectionName string, options *WebAppsBeginDeletePrivateEndpointConnectionOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/privateEndpointConnections/{privateEndpointConnectionName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif privateEndpointConnectionName == \"\" {\n\t\treturn nil, errors.New(\"parameter privateEndpointConnectionName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{privateEndpointConnectionName}\", url.PathEscape(privateEndpointConnectionName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (c *PetruleClient) Delete() *PetruleDelete {\n\tmutation := newPetruleMutation(c.config, OpDelete)\n\treturn &PetruleDelete{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func NewDeleteSpoeAgentDefault(code int) *DeleteSpoeAgentDefault {\n\treturn &DeleteSpoeAgentDefault{\n\t\t_statusCode: code,\n\t\tConfigurationVersion: 0,\n\t}\n}", "func Delete(c *gophercloud.ServiceClient, idOrURL string) (r DeleteResult) {\n\tvar url string\n\tif strings.Contains(idOrURL, \"/\") {\n\t\turl = idOrURL\n\t} else {\n\t\turl = deleteURL(c, idOrURL)\n\t}\n\tresp, err := c.Delete(url, nil)\n\t_, r.Header, r.Err = gophercloud.ParseResponse(resp, err)\n\treturn\n}", "func NewDeleteConsulDefault(code int) *DeleteConsulDefault {\n\tif code <= 0 {\n\t\tcode = 500\n\t}\n\n\treturn &DeleteConsulDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewDeleteOneNoContent() *DeleteOneNoContent {\n\treturn &DeleteOneNoContent{}\n}", "func (nse ErrNoSuchEndpoint) NotFound() {}", "func (client *LiveOutputsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, accountName string, liveEventName string, liveOutputName string, options *LiveOutputsClientBeginDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Media/mediaservices/{accountName}/liveEvents/{liveEventName}/liveOutputs/{liveOutputName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif accountName == \"\" {\n\t\treturn nil, errors.New(\"parameter accountName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{accountName}\", url.PathEscape(accountName))\n\tif liveEventName == \"\" {\n\t\treturn nil, errors.New(\"parameter liveEventName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{liveEventName}\", url.PathEscape(liveEventName))\n\tif liveOutputName == \"\" {\n\t\treturn nil, errors.New(\"parameter liveOutputName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{liveOutputName}\", url.PathEscape(liveOutputName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2022-08-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NewEndpointDeleteServiceUnavailable() *EndpointDeleteServiceUnavailable {\n\treturn &EndpointDeleteServiceUnavailable{}\n}", "func (d *Driver) DeleteEndpoint(r *sdk.DeleteEndpointRequest) error {\n\tlog.Debugf(\"Delete endpoint request: %+v\", &r)\n\t//TODO: null check cidr in case driver restarted and doesn't know the network to avoid panic\n\tlog.Debugf(\"Delete endpoint %s\", r.EndpointID)\n\n\tnid := r.NetworkID\n\teid := r.EndpointID\n\n\t// Get the network handler and make sure it exists\n\td.Lock()\n\tn, ok := d.networks[nid]\n\td.Unlock()\n\n\tif !ok {\n\t\treturn types.InternalMaskableErrorf(\"network %s does not exist\", nid)\n\t}\n\tif n == nil {\n\t\treturn driverapi.ErrNoNetwork(nid)\n\t}\n\n\t// Sanity Check\n\tn.Lock()\n\tif n.id != nid {\n\t\tn.Unlock()\n\t\treturn InvalidNetworkIDError(nid)\n\t}\n\tn.Unlock()\n\n\t// Check endpoint id and if an endpoint is actually there\n\tep, err := n.getEndpoint(eid)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif ep == nil {\n\t\treturn EndpointNotFoundError(eid)\n\t}\n\n\t// Remove it\n\tn.Lock()\n\tdelete(n.endpoints, eid)\n\tn.Unlock()\n\n\t// On failure make sure to set back ep in n.endpoints, but only\n\t// if it hasn't been taken over already by some other thread.\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tn.Lock()\n\t\t\tif _, ok := n.endpoints[eid]; !ok {\n\t\t\t\tn.endpoints[eid] = ep\n\t\t\t}\n\t\t\tn.Unlock()\n\t\t}\n\t}()\n\n\t// Try removal of link. Discard error: it is a best effort.\n\t// Also make sure defer does not see this error either.\n\tif link, err := d.nlh.LinkByName(ep.srcName); err == nil {\n\t\td.nlh.LinkDel(link)\n\t}\n\n\treturn nil\n}", "func (k *xyzProvider) Delete(ctx context.Context, req *pulumirpc.DeleteRequest) (*pbempty.Empty, error) {\n\turn := resource.URN(req.GetUrn())\n\tty := urn.Type()\n\tif ty != \"xyz:index:Random\" {\n\t\treturn nil, fmt.Errorf(\"Unknown resource type '%s'\", ty)\n\t}\n\n\t// Note that for our Random resource, we don't have to do anything on Delete.\n\treturn &pbempty.Empty{}, nil\n}", "func (a *FrinxOpenconfigNetworkInstanceApiService) DeleteFrinxOpenconfigNetworkInstanceNetworkInstancesNetworkInstanceConnectionPointsConnectionPointEndpointsEndpoint(ctx context.Context, name string, connectionPointId string, endpointId string, nodeId string) (*http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = strings.ToUpper(\"Delete\")\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\t\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/config/network-topology:network-topology/network-topology:topology/unified/network-topology:node/{node-id}/yang-ext:mount/frinx-openconfig-network-instance:network-instances/frinx-openconfig-network-instance:network-instance/{name}/frinx-openconfig-network-instance:connection-points/frinx-openconfig-network-instance:connection-point/{connection-point-id}/frinx-openconfig-network-instance:endpoints/frinx-openconfig-network-instance:endpoint/{endpoint-id}/\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"name\"+\"}\", fmt.Sprintf(\"%v\", name), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"connection-point-id\"+\"}\", fmt.Sprintf(\"%v\", connectionPointId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"endpoint-id\"+\"}\", fmt.Sprintf(\"%v\", endpointId), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"node-id\"+\"}\", fmt.Sprintf(\"%v\", nodeId), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{\"application/json\", \"application/xml\"}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\", \"application/xml\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarHttpResponse, err\n\t}\n\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericSwaggerError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\t\n\t\treturn localVarHttpResponse, newErr\n\t}\n\n\treturn localVarHttpResponse, nil\n}", "func (api *Api) Delete(path string, endpoint http.HandlerFunc, queries ...string) {\n\tapi.Router.HandleFunc(path, endpoint).Methods(\"DELETE\").Queries(queries...)\n}", "func Delete(config *HTTPConfig) (*HTTPResult, error) {\n\treturn HandleRequest(\"DELETE\", config)\n}", "func NewDeleteMaybeNoContent() *DeleteMaybeNoContent {\n\treturn &DeleteMaybeNoContent{}\n}", "func (c *Client) delete(rawURL string, authenticate bool, out interface{}) error {\n\terr := c.do(rawURL, \"DELETE\", authenticate, http.StatusOK, nil, out)\n\treturn errio.Error(err)\n}", "func (m *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func DeleteGuestbookNoContent(t goatest.TInterface, ctx context.Context, service *goa.Service, ctrl app.GuestbookController, id string) http.ResponseWriter {\n\t// Setup service\n\tvar (\n\t\tlogBuf bytes.Buffer\n\t\tresp interface{}\n\n\t\trespSetter goatest.ResponseSetterFunc = func(r interface{}) { resp = r }\n\t)\n\tif service == nil {\n\t\tservice = goatest.Service(&logBuf, respSetter)\n\t} else {\n\t\tlogger := log.New(&logBuf, \"\", log.Ltime)\n\t\tservice.WithLogger(goa.NewLogger(logger))\n\t\tnewEncoder := func(io.Writer) goa.Encoder { return respSetter }\n\t\tservice.Encoder = goa.NewHTTPEncoder() // Make sure the code ends up using this decoder\n\t\tservice.Encoder.Register(newEncoder, \"*/*\")\n\t}\n\n\t// Setup request context\n\trw := httptest.NewRecorder()\n\tu := &url.URL{\n\t\tPath: fmt.Sprintf(\"/guestbook/%v\", id),\n\t}\n\treq, err := http.NewRequest(\"DELETE\", u.String(), nil)\n\tif err != nil {\n\t\tpanic(\"invalid test \" + err.Error()) // bug\n\t}\n\tprms := url.Values{}\n\tprms[\"id\"] = []string{fmt.Sprintf(\"%v\", id)}\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\tgoaCtx := goa.NewContext(goa.WithAction(ctx, \"GuestbookTest\"), rw, req, prms)\n\tdeleteCtx, _err := app.NewDeleteGuestbookContext(goaCtx, req, service)\n\tif _err != nil {\n\t\te, ok := _err.(goa.ServiceError)\n\t\tif !ok {\n\t\t\tpanic(\"invalid test data \" + _err.Error()) // bug\n\t\t}\n\t\tt.Errorf(\"unexpected parameter validation error: %+v\", e)\n\t\treturn nil\n\t}\n\n\t// Perform action\n\t_err = ctrl.Delete(deleteCtx)\n\n\t// Validate response\n\tif _err != nil {\n\t\tt.Fatalf(\"controller returned %+v, logs:\\n%s\", _err, logBuf.String())\n\t}\n\tif rw.Code != 204 {\n\t\tt.Errorf(\"invalid response status code: got %+v, expected 204\", rw.Code)\n\t}\n\n\t// Return results\n\treturn rw\n}", "func Delete(url string, data ...interface{}) (*Response, error) {\n\tr := NewRequest()\n\treturn r.Delete(url, data...)\n}", "func deleteEndpointByQuery(w http.ResponseWriter, r *http.Request, query elastic.Query, caller string) {\r\n\r\n\tclient, err := getElasticsearchClient()\r\n\tif err != nil {\r\n\t\tapi.LogError(api.DEBUG, err)\r\n\t\tfmt.Fprintf(w, api.HttpFailureMessage(\"Failed to initialized query. ERROR 0x40000012\"))\r\n\t\treturn\r\n\t}\r\n\r\n\tres, err := client.DeleteByQuery().\r\n\t\tIndex(config.ElasticIndex()).\r\n\t\tType(\"audit_type\").\r\n\t\tQuery(query).\r\n\t\tDo(context.Background())\r\n\r\n\tif err != nil {\r\n\t\tapi.HttpResponseReturn(w, r, \"failed\", err.Error(), nil)\r\n\t\treturn\r\n\t}\r\n\tapi.HttpResponseReturn(w, r, \"success\", \"Case/Endpoint deleted\", res)\r\n}", "func (client CloudEndpointsClient) Delete(ctx context.Context, resourceGroupName string, storageSyncServiceName string, syncGroupName string, cloudEndpointName string) (result CloudEndpointsDeleteFuture, err error) {\n\tif tracing.IsEnabled() {\n\t\tctx = tracing.StartSpan(ctx, fqdn+\"/CloudEndpointsClient.Delete\")\n\t\tdefer func() {\n\t\t\tsc := -1\n\t\t\tif result.FutureAPI != nil && result.FutureAPI.Response() != nil {\n\t\t\t\tsc = result.FutureAPI.Response().StatusCode\n\t\t\t}\n\t\t\ttracing.EndSpan(ctx, sc, err)\n\t\t}()\n\t}\n\tif err := validation.Validate([]validation.Validation{\n\t\t{TargetValue: client.SubscriptionID,\n\t\t\tConstraints: []validation.Constraint{{Target: \"client.SubscriptionID\", Name: validation.MinLength, Rule: 1, Chain: nil}}},\n\t\t{TargetValue: resourceGroupName,\n\t\t\tConstraints: []validation.Constraint{{Target: \"resourceGroupName\", Name: validation.MaxLength, Rule: 90, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.MinLength, Rule: 1, Chain: nil},\n\t\t\t\t{Target: \"resourceGroupName\", Name: validation.Pattern, Rule: `^[-\\w\\._\\(\\)]+$`, Chain: nil}}}}); err != nil {\n\t\treturn result, validation.NewError(\"storagesync.CloudEndpointsClient\", \"Delete\", err.Error())\n\t}\n\n\treq, err := client.DeletePreparer(ctx, resourceGroupName, storageSyncServiceName, syncGroupName, cloudEndpointName)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"storagesync.CloudEndpointsClient\", \"Delete\", nil, \"Failure preparing request\")\n\t\treturn\n\t}\n\n\tresult, err = client.DeleteSender(req)\n\tif err != nil {\n\t\terr = autorest.NewErrorWithError(err, \"storagesync.CloudEndpointsClient\", \"Delete\", result.Response(), \"Failure sending request\")\n\t\treturn\n\t}\n\n\treturn\n}", "func NewEndpointDeleteBadRequest() *EndpointDeleteBadRequest {\n\treturn &EndpointDeleteBadRequest{}\n}", "func Delete(url string, data ...interface{}) (*ClientResponse, error) {\n\treturn DoRequest(\"DELETE\", url, data...)\n}", "func (h *Handler) Delete(w http.ResponseWriter, r *http.Request) {\n\n\tp := mux.Vars(r)\n\tid := p[\"id\"]\n\n\terr := h.Source.Delete(id)\n\n\tif err != nil {\n\t\te := Error{\n\t\t\tMessage: fmt.Sprintf(\"An error occured when trying to delete the delivery : %s\", err),\n\t\t}\n\t\tJSONWithHTTPCode(w, e, 500)\n\t\treturn\n\t}\n\n\tJSONWithHTTPCode(w, nil, 204)\n\n}", "func NewDeleteAPIKeyNoContent() *DeleteAPIKeyNoContent {\n\treturn &DeleteAPIKeyNoContent{}\n}" ]
[ "0.6879929", "0.6832315", "0.6657975", "0.6482036", "0.62831396", "0.624386", "0.6218298", "0.61974895", "0.60938275", "0.6065671", "0.60642606", "0.60351217", "0.59749484", "0.5940223", "0.5927456", "0.587366", "0.587366", "0.57466143", "0.5743995", "0.5703535", "0.56740713", "0.56736565", "0.56595683", "0.5629202", "0.56041336", "0.55935556", "0.5577443", "0.5493569", "0.5468044", "0.54667866", "0.54585826", "0.545063", "0.5435264", "0.54178727", "0.5401154", "0.5388858", "0.538002", "0.53749573", "0.5372519", "0.53546405", "0.5351574", "0.535001", "0.53401804", "0.5322502", "0.53061247", "0.52942544", "0.5285904", "0.52852434", "0.5273139", "0.5244242", "0.5232581", "0.5225796", "0.52250093", "0.52224916", "0.5220996", "0.5202447", "0.5194491", "0.51827335", "0.51735336", "0.5171975", "0.5166028", "0.51658916", "0.5162879", "0.5146594", "0.51464707", "0.51448166", "0.51339155", "0.5129749", "0.512313", "0.51194614", "0.5117747", "0.5109939", "0.51076543", "0.5106669", "0.5105343", "0.5104851", "0.5104709", "0.5101152", "0.51008415", "0.5100477", "0.5099144", "0.5093932", "0.5091823", "0.5091124", "0.5090717", "0.50854385", "0.50839555", "0.5078452", "0.5078106", "0.50769144", "0.50710887", "0.50663966", "0.50619847", "0.5054279", "0.50534254", "0.50477463", "0.50473803", "0.5046114", "0.504076", "0.50382197" ]
0.71023625
0
NewEndpointDeleteBadRequest creates a EndpointDeleteBadRequest with default headers values
func NewEndpointDeleteBadRequest() *EndpointDeleteBadRequest { return &EndpointDeleteBadRequest{} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewDeleteBadRequest(body *DeleteBadRequestResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDataDeleteBadRequest() *DataDeleteBadRequest {\n\treturn &DataDeleteBadRequest{}\n}", "func (ctx *DeleteHostContext) BadRequest() error {\n\tctx.ResponseData.WriteHeader(400)\n\treturn nil\n}", "func (a *DefaultApiService) DeleteEndpoint(ctx _context.Context, id string) ApiDeleteEndpointRequest {\n\treturn ApiDeleteEndpointRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tid: id,\n\t}\n}", "func EndpointDeleteMessage(e notifications.RegenNotificationInfo) AgentNotifyMessage {\n\tnotification := EndpointNotification{\n\t\tEndpointRegenNotification: EndpointRegenNotification{\n\t\t\tID: e.GetID(),\n\t\t\tLabels: e.GetOpLabels(),\n\t\t},\n\t\tPodName: e.GetK8sPodName(),\n\t\tNamespace: e.GetK8sNamespace(),\n\t}\n\n\treturn AgentNotifyMessage{\n\t\tType: AgentNotifyEndpointDeleted,\n\t\tNotification: notification,\n\t}\n}", "func DeleteEndpoint(w http.ResponseWriter, r *http.Request) {\r\n\tr.Header.Set(\"Content-Type\", \"application/json, charset=UTF-8\")\r\n\tvar dc DeleteConfig\r\n\r\n\tswitch r.Method {\r\n\tcase \"GET\":\r\n\t\tparams := mux.Vars(r)\r\n\t\tdc.Default()\r\n\t\tdc.Endpoint = params[\"endpoint_name\"]\r\n\t\tdc.EndpointList = append(dc.EndpointList, dc.Endpoint)\r\n\tcase \"POST\":\r\n\t\tbody, err := ioutil.ReadAll(r.Body)\r\n\t\tif err != nil {\r\n\t\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint, failed to read request\")\r\n\t\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Failed to read HTTP request\"))\r\n\t\t\treturn\r\n\t\t}\r\n\t\tdc.LoadParams(body)\r\n\t}\r\n\r\n\t// Verify Endpoint is provided in request body\r\n\tif len(dc.EndpointList) == 0 {\r\n\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint - endpoint is required\")\r\n\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Endpoint required\"))\r\n\t\treturn\r\n\t}\r\n\r\n\tvar query elastic.Query\r\n\r\n\t// Convert EndpointList to interface{} slice\r\n\tendpoints := make([]interface{}, len(dc.EndpointList))\r\n\tfor i, v := range dc.EndpointList {\r\n\t\tendpoints[i] = v\r\n\t}\r\n\r\n\tquery = elastic.NewBoolQuery().\r\n\t\tMust(elastic.NewWildcardQuery(\"CaseInfo.CaseName\", dc.CaseName),\r\n\t\t\telastic.NewTermsQuery(\"ComputerName.keyword\", endpoints...))\r\n\r\n\tdeleteEndpointByQuery(w, r, query, \"DeleteEndpoint\")\r\n\r\n}", "func NewDeleteEndpoint(s Service) goa.Endpoint {\n\treturn func(ctx context.Context, req interface{}) (interface{}, error) {\n\t\tp := req.(*DeletePayload)\n\t\treturn nil, s.Delete(ctx, p)\n\t}\n}", "func NewDeleteShipmentBadRequest() *DeleteShipmentBadRequest {\n\n\treturn &DeleteShipmentBadRequest{}\n}", "func (m *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewDeleteMediaBadRequest(body *DeleteMediaBadRequestResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func (client *LROSADsClient) delete202RetryInvalidHeaderCreateRequest(ctx context.Context, options *LROSADsClientBeginDelete202RetryInvalidHeaderOptions) (*policy.Request, error) {\n\turlPath := \"/lro/error/delete/202/retry/invalidheader\"\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NewWeaviateSchemaActionsDeleteBadRequest() *WeaviateSchemaActionsDeleteBadRequest {\n\n\treturn &WeaviateSchemaActionsDeleteBadRequest{}\n}", "func NewDeleteAPIKeyBadRequest() *DeleteAPIKeyBadRequest {\n\treturn &DeleteAPIKeyBadRequest{}\n}", "func NewDeleteCustomPropertyForDeviceBadRequest() *DeleteCustomPropertyForDeviceBadRequest {\n\treturn &DeleteCustomPropertyForDeviceBadRequest{}\n}", "func NewDeleteUserBadRequest() *DeleteUserBadRequest {\n\n\treturn &DeleteUserBadRequest{}\n}", "func NewSystemDeleteBadRequest() *SystemDeleteBadRequest {\n\treturn &SystemDeleteBadRequest{}\n}", "func NewDeleteNotFound(body *DeleteNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDeleteNotFound(body *DeleteNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDeleteProductBadRequest() *DeleteProductBadRequest {\n\n\treturn &DeleteProductBadRequest{}\n}", "func NewDeleteLinkBadRequest() *DeleteLinkBadRequest {\n\treturn &DeleteLinkBadRequest{}\n}", "func DeleteEndpoint(serviceAccountProvider provider.ServiceAccountProvider, projectProvider provider.ProjectProvider, userInfoGetter provider.UserInfoGetter) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq, ok := request.(deleteReq)\n\t\tif !ok {\n\t\t\treturn nil, errors.NewBadRequest(\"invalid request\")\n\t\t}\n\t\terr := req.Validate()\n\t\tif err != nil {\n\t\t\treturn nil, errors.NewBadRequest(err.Error())\n\t\t}\n\t\tuserInfo, err := userInfoGetter(ctx, req.ProjectID)\n\t\tif err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\t\t// check if project exist\n\t\tif _, err := projectProvider.Get(userInfo, req.ProjectID, &provider.ProjectGetOptions{}); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\t// check if service account exist before deleting it\n\t\tif _, err := serviceAccountProvider.Get(userInfo, req.ServiceAccountID, nil); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\tif err := serviceAccountProvider.Delete(userInfo, req.ServiceAccountID); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\treturn nil, nil\n\t}\n}", "func NewDeleteConnectorUsingDELETEBadRequest() *DeleteConnectorUsingDELETEBadRequest {\n\treturn &DeleteConnectorUsingDELETEBadRequest{}\n}", "func NewDeleteDatasetBadRequest() *DeleteDatasetBadRequest {\n\treturn &DeleteDatasetBadRequest{}\n}", "func NewDeleteBadRequestResponseBody(res *goa.ServiceError) *DeleteBadRequestResponseBody {\n\tbody := &DeleteBadRequestResponseBody{\n\t\tName: res.Name,\n\t\tID: res.ID,\n\t\tMessage: res.Message,\n\t\tTemporary: res.Temporary,\n\t\tTimeout: res.Timeout,\n\t\tFault: res.Fault,\n\t}\n\treturn body\n}", "func NewDeleteUserBadRequest() *DeleteUserBadRequest {\n\treturn &DeleteUserBadRequest{}\n}", "func (a *DefaultApiService) DeleteEndpointExecute(r ApiDeleteEndpointRequest) (*_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.DeleteEndpoint\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/endpoints/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.PathEscape(parameterToString(r.id, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ErrorNotFound\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, executionError\n}", "func (ctx *DeleteOutputContext) BadRequest(r error) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.goa.error\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 400, r)\n}", "func NewDeleteEnvironmentBadRequest() *DeleteEnvironmentBadRequest {\n\treturn &DeleteEnvironmentBadRequest{}\n}", "func NewDeletePostbyIDBadRequest() *DeletePostbyIDBadRequest {\n\n\treturn &DeletePostbyIDBadRequest{}\n}", "func (client *NpmClient) EndpointDeleteReq(epinfo *netproto.Endpoint) (*netproto.Endpoint, error) {\n\treturn nil, nil\n}", "func (m *GroupPolicyDefinitionsItemNextVersionDefinitionPreviousVersionDefinitionRequestBuilder) Delete(ctx context.Context, requestConfiguration *GroupPolicyDefinitionsItemNextVersionDefinitionPreviousVersionDefinitionRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewDeleteIscsiLunBadRequest() *DeleteIscsiLunBadRequest {\n\treturn &DeleteIscsiLunBadRequest{}\n}", "func NewDeleteIscsiLunBadRequest() *DeleteIscsiLunBadRequest {\n\treturn &DeleteIscsiLunBadRequest{}\n}", "func (m *PayloadResponseItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *PayloadResponseItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (c *Client) Delete(endpoint string, resource ...RequestResource) error {\n\twrapper := newWrapper(\"delete\", endpoint, resource...)\n\treturn c.do(&wrapper)\n}", "func NewPcloudCloudconnectionsDeleteBadRequest() *PcloudCloudconnectionsDeleteBadRequest {\n\treturn &PcloudCloudconnectionsDeleteBadRequest{}\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func (ctx *CreateHostContext) BadRequest() error {\n\tctx.ResponseData.WriteHeader(400)\n\treturn nil\n}", "func NewEndpointDeleteNotFound() *EndpointDeleteNotFound {\n\treturn &EndpointDeleteNotFound{}\n}", "func Delete(client *gophercloud.ServiceClient, endpointID string) (r DeleteResult) {\n\t_, r.Err = client.Delete(endpointURL(client, endpointID), nil)\n\treturn\n}", "func (o *WeaviateSchemaActionsDeleteBadRequest) WithPayload(payload *models.ErrorResponse) *WeaviateSchemaActionsDeleteBadRequest {\n\to.Payload = payload\n\treturn o\n}", "func NewGroupDeletesBadRequest() *GroupDeletesBadRequest {\n\treturn &GroupDeletesBadRequest{}\n}", "func (m *WindowsUpdatesCatalogRequestBuilder) Delete(ctx context.Context, requestConfiguration *WindowsUpdatesCatalogRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func DeleteGuestbookBadRequest(t goatest.TInterface, ctx context.Context, service *goa.Service, ctrl app.GuestbookController, id string) (http.ResponseWriter, *app.GuestbookError) {\n\t// Setup service\n\tvar (\n\t\tlogBuf bytes.Buffer\n\t\tresp interface{}\n\n\t\trespSetter goatest.ResponseSetterFunc = func(r interface{}) { resp = r }\n\t)\n\tif service == nil {\n\t\tservice = goatest.Service(&logBuf, respSetter)\n\t} else {\n\t\tlogger := log.New(&logBuf, \"\", log.Ltime)\n\t\tservice.WithLogger(goa.NewLogger(logger))\n\t\tnewEncoder := func(io.Writer) goa.Encoder { return respSetter }\n\t\tservice.Encoder = goa.NewHTTPEncoder() // Make sure the code ends up using this decoder\n\t\tservice.Encoder.Register(newEncoder, \"*/*\")\n\t}\n\n\t// Setup request context\n\trw := httptest.NewRecorder()\n\tu := &url.URL{\n\t\tPath: fmt.Sprintf(\"/guestbook/%v\", id),\n\t}\n\treq, err := http.NewRequest(\"DELETE\", u.String(), nil)\n\tif err != nil {\n\t\tpanic(\"invalid test \" + err.Error()) // bug\n\t}\n\tprms := url.Values{}\n\tprms[\"id\"] = []string{fmt.Sprintf(\"%v\", id)}\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\tgoaCtx := goa.NewContext(goa.WithAction(ctx, \"GuestbookTest\"), rw, req, prms)\n\tdeleteCtx, _err := app.NewDeleteGuestbookContext(goaCtx, req, service)\n\tif _err != nil {\n\t\te, ok := _err.(goa.ServiceError)\n\t\tif !ok {\n\t\t\tpanic(\"invalid test data \" + _err.Error()) // bug\n\t\t}\n\t\tt.Errorf(\"unexpected parameter validation error: %+v\", e)\n\t\treturn nil, nil\n\t}\n\n\t// Perform action\n\t_err = ctrl.Delete(deleteCtx)\n\n\t// Validate response\n\tif _err != nil {\n\t\tt.Fatalf(\"controller returned %+v, logs:\\n%s\", _err, logBuf.String())\n\t}\n\tif rw.Code != 400 {\n\t\tt.Errorf(\"invalid response status code: got %+v, expected 400\", rw.Code)\n\t}\n\tvar mt *app.GuestbookError\n\tif resp != nil {\n\t\tvar _ok bool\n\t\tmt, _ok = resp.(*app.GuestbookError)\n\t\tif !_ok {\n\t\t\tt.Fatalf(\"invalid response media: got variable of type %T, value %+v, expected instance of app.GuestbookError\", resp, resp)\n\t\t}\n\t\t_err = mt.Validate()\n\t\tif _err != nil {\n\t\t\tt.Errorf(\"invalid response media type: %s\", _err)\n\t\t}\n\t}\n\n\t// Return results\n\treturn rw, mt\n}", "func MakeDeleteEndpoint(s service.DepartmentService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func DeleteEndpointInfo(ctx iris.Context) {\n\turi := ctx.Request().RequestURI\n\tfabricID := ctx.Params().Get(\"id\")\n\tif _, ok := capdata.FabricDataStore.Data[fabricID]; !ok {\n\t\terrMsg := fmt.Sprintf(\"Fabric data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Fabric\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tendpointData, ok := capdata.EndpointDataStore[uri]\n\tif !ok {\n\t\terrMsg := fmt.Sprintf(\"Endpoint data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Endpoint\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif endpointData.Endpoint.Links != nil && len(endpointData.Endpoint.Links.AddressPools) > 0 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be deleted as there are dependent upon AddressPool\")\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceCannotBeDeleted, errMsg, []interface{}{uri, \"Endpoint\"})\n\t\tctx.StatusCode(http.StatusNotAcceptable)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\t// Todo:Add the validation to verify the links\n\tdelete(capdata.EndpointDataStore, uri)\n\tctx.StatusCode(http.StatusNoContent)\n}", "func (s *DefaultClient) Del(endpoint string) (resp *http.Response, err error) {\n\t_, resp, err = s.http(http.MethodDelete, endpoint, nil)\n\treturn\n}", "func (ctx *DeleteFilterContext) BadRequest(r error) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.goa.error\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 400, r)\n}", "func NewDeleteGatewayBundleUsingDELETEBadRequest() *DeleteGatewayBundleUsingDELETEBadRequest {\n\treturn &DeleteGatewayBundleUsingDELETEBadRequest{}\n}", "func (client *LROSADsClient) delete202NonRetry400CreateRequest(ctx context.Context, options *LROSADsClientBeginDelete202NonRetry400Options) (*policy.Request, error) {\n\turlPath := \"/lro/nonretryerror/delete/202/retry/400\"\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\treturn req, nil\n}", "func NewEndpointUpdateBadRequest() *EndpointUpdateBadRequest {\n\treturn &EndpointUpdateBadRequest{}\n}", "func ValidateDeleteBadRequestResponseBody(body *DeleteBadRequestResponseBody) (err error) {\n\tif body.Name == nil {\n\t\terr = goa.MergeErrors(err, goa.MissingFieldError(\"name\", \"body\"))\n\t}\n\tif body.ID == nil {\n\t\terr = goa.MergeErrors(err, goa.MissingFieldError(\"id\", \"body\"))\n\t}\n\tif body.Message == nil {\n\t\terr = goa.MergeErrors(err, goa.MissingFieldError(\"message\", \"body\"))\n\t}\n\tif body.Temporary == nil {\n\t\terr = goa.MergeErrors(err, goa.MissingFieldError(\"temporary\", \"body\"))\n\t}\n\tif body.Timeout == nil {\n\t\terr = goa.MergeErrors(err, goa.MissingFieldError(\"timeout\", \"body\"))\n\t}\n\tif body.Fault == nil {\n\t\terr = goa.MergeErrors(err, goa.MissingFieldError(\"fault\", \"body\"))\n\t}\n\treturn\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsClientDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (ctx *DeleteFeedContext) BadRequest(r error) error {\n\tif ctx.ResponseData.Header().Get(\"Content-Type\") == \"\" {\n\t\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.goa.error\")\n\t}\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 400, r)\n}", "func (m *ManagedDevicesItemSecurityBaselineStatesSecurityBaselineStateItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ManagedDevicesItemSecurityBaselineStatesSecurityBaselineStateItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client HTTPSuccessClient) Delete202Responder(resp *http.Response) (result autorest.Response, err error) { \n err = autorest.Respond(\n resp,\n client.ByInspecting(),\n azure.WithErrorUnlessStatusCode(http.StatusOK,http.StatusAccepted),\n autorest.ByClosing())\n result.Response = resp\n return\n}", "func (m *VirtualEventsWebinarsItemSessionsItemVirtualAppointmentRequestBuilder) Delete(ctx context.Context, requestConfiguration *VirtualEventsWebinarsItemSessionsItemVirtualAppointmentRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (c *APIClient) Delete(endpoint Endpointer, result interface{}, headers ...func(h *http.Header)) error {\n\tepoint, err := endpoint.GetEndpoint()\n\tif err != nil {\n\t\treturn err\n\t}\n\tendUrl := c.CompileEndpointURL(epoint)\n\treq, _ := http.NewRequest(http.MethodDelete, endUrl, nil)\n\n\th := &req.Header\n\tfor _, v := range headers {\n\t\tv(h)\n\t}\n\n\tresp, err := c.Client.Call(req)\n\tc.addResponseCode(resp.StatusCode)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn MakeResponseError(resp)\n\t}\n\n\terr = json.NewDecoder(resp.Body).Decode(result)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (ctx *DeleteDogContext) BadRequest() error {\n\tctx.ResponseData.WriteHeader(400)\n\treturn nil\n}", "func TestServiceEndpointAzureRM_Delete_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := ResourceServiceEndpointAzureRM()\n\tfor _, resource := range azurermTestServiceEndpointsAzureRM {\n\t\tresourceData := getResourceData(t, resource)\n\t\tflattenServiceEndpointAzureRM(resourceData, &resource, azurermTestServiceEndpointAzureRMProjectID)\n\n\t\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\t\tclients := &client.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\t\texpectedArgs := serviceendpoint.DeleteServiceEndpointArgs{\n\t\t\tEndpointId: resource.Id,\n\t\t\tProjectIds: &[]string{\n\t\t\t\tazurermTestServiceEndpointAzureRMProjectID.String(),\n\t\t\t},\n\t\t}\n\n\t\tbuildClient.\n\t\t\tEXPECT().\n\t\t\tDeleteServiceEndpoint(clients.Ctx, expectedArgs).\n\t\t\tReturn(errors.New(\"DeleteServiceEndpoint() Failed\")).\n\t\t\tTimes(1)\n\n\t\terr := r.Delete(resourceData, clients)\n\t\trequire.Contains(t, err.Error(), \"DeleteServiceEndpoint() Failed\")\n\t}\n}", "func (m *ZebraFotaConnectorRequestBuilder) Delete(ctx context.Context, requestConfiguration *ZebraFotaConnectorRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (*DeleteEndpointRequest) Descriptor() ([]byte, []int) {\n\treturn file_endpoint_api_proto_rawDescGZIP(), []int{11}\n}", "func (c *Client) Delete(ctx context.Context, p *DeletePayload) (err error) {\n\t_, err = c.DeleteEndpoint(ctx, p)\n\treturn\n}", "func (client *WebAppsClient) deleteHostSecretCreateRequest(ctx context.Context, resourceGroupName string, name string, keyType string, keyName string, options *WebAppsDeleteHostSecretOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/host/default/{keyType}/{keyName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif keyType == \"\" {\n\t\treturn nil, errors.New(\"parameter keyType cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{keyType}\", url.PathEscape(keyType))\n\tif keyName == \"\" {\n\t\treturn nil, errors.New(\"parameter keyName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{keyName}\", url.PathEscape(keyName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewDeleteDebugRequestBadRequest() *DeleteDebugRequestBadRequest {\n\treturn &DeleteDebugRequestBadRequest{}\n}", "func NewDeleteConfigurationBadRequest() *DeleteConfigurationBadRequest {\n\treturn &DeleteConfigurationBadRequest{}\n}", "func NewDeleteAPIKeyDefault(code int) *DeleteAPIKeyDefault {\n\treturn &DeleteAPIKeyDefault{\n\t\t_statusCode: code,\n\t}\n}", "func MakeDeleteEndpoint(s service.TodoService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func (c *V2) Delete(endpoint string) (*http.Response, error) {\n\treq, err := c.PrepareRequest(http.MethodDelete, endpoint, nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.Do(req)\n}", "func (nse ErrNoSuchEndpoint) NotFound() {}", "func NewDeleteWaitlistEntryBadRequest(body *DeleteWaitlistEntryBadRequestResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func PostDeleteEndpoint(deleter deleting.Service) func(w http.ResponseWriter, r *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\tid := mux.Vars(r)[\"id\"]\n\n\t\t// Delete a post.\n\t\terr := deleter.PostDelete(id)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error deleting a post:\", err)\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tw.WriteHeader(http.StatusNoContent)\n\t}\n}", "func NewDeleteMachinePolicyBadRequest() *DeleteMachinePolicyBadRequest {\n\treturn &DeleteMachinePolicyBadRequest{}\n}", "func (m *VirtualEventsEventsItemSessionsItemVirtualAppointmentRequestBuilder) Delete(ctx context.Context, requestConfiguration *VirtualEventsEventsItemSessionsItemVirtualAppointmentRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (uee *UnknownEndpointError) NotFound() {}", "func (m *ItemManagedDevicesItemSecurityBaselineStatesSecurityBaselineStateItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ItemManagedDevicesItemSecurityBaselineStatesSecurityBaselineStateItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewDeleteVersionControlRequestBadRequest() *DeleteVersionControlRequestBadRequest {\n\treturn &DeleteVersionControlRequestBadRequest{}\n}", "func (o *WeaviateSchemaActionsDeleteBadRequest) WriteResponse(rw http.ResponseWriter, producer runtime.Producer) {\n\n\trw.WriteHeader(400)\n\tif o.Payload != nil {\n\t\tpayload := o.Payload\n\t\tif err := producer.Produce(rw, payload); err != nil {\n\t\t\tpanic(err) // let the recovery middleware deal with this\n\t\t}\n\t}\n}", "func NewDeletePlantBadRequest() *DeletePlantBadRequest {\n\n\treturn &DeletePlantBadRequest{}\n}", "func NewDeleteBundleCustomFieldsBadRequest() *DeleteBundleCustomFieldsBadRequest {\n\treturn &DeleteBundleCustomFieldsBadRequest{}\n}", "func (endpoint *HNSEndpoint) Delete() (*HNSEndpoint, error) {\n\toperation := \"Delete\"\n\ttitle := \"hcsshim::HNSEndpoint::\" + operation\n\tlogrus.Debugf(title+\" id=%s\", endpoint.Id)\n\n\treturn HNSEndpointRequest(\"DELETE\", endpoint.Id, \"\")\n}", "func (ds *DeleteSuite) TestDelete_Failure_on_not_parsable_id() {\n\n\t// Arrange.\n\n\te := echo.New()\n\treq := httptest.NewRequest(http.MethodDelete, \"/:id\", nil)\n\tres := httptest.NewRecorder()\n\tc := e.NewContext(req, res)\n\n\tc.SetPath(\"/:id\")\n\tc.SetParamNames(\"id\")\n\tc.SetParamValues(\"d\")\n\n\t// Act.\n\n\t_ = HandleDelete(c)\n\n\t// Assert.\n\n\tassert.Equal(ds.T(), http.StatusBadRequest, res.Code)\n\tvar httpError echo.HTTPError\n\t_ = json.Unmarshal(res.Body.Bytes(), &httpError)\n\tassert.Equal(ds.T(), \"strconv.Atoi: parsing \\\"d\\\": invalid syntax\", httpError.Message)\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn handleDeleteResponse(bts)\n}", "func NewDeleteBacsIDBadRequest() *DeleteBacsIDBadRequest {\n\treturn &DeleteBacsIDBadRequest{}\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(ctx context.Context, project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(ctx, path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn checkAPIResponse(bts, nil)\n}", "func NewDeleteMessageBadRequestResponseBody(res *goa.ServiceError) *DeleteMessageBadRequestResponseBody {\n\tbody := &DeleteMessageBadRequestResponseBody{\n\t\tName: res.Name,\n\t\tID: res.ID,\n\t\tMessage: res.Message,\n\t\tTemporary: res.Temporary,\n\t\tTimeout: res.Timeout,\n\t\tFault: res.Fault,\n\t}\n\treturn body\n}", "func (m *ItemCalendarViewBookingAppointmentItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ItemCalendarViewBookingAppointmentItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewDeleteInstancesBadRequest() *DeleteInstancesBadRequest {\n\treturn &DeleteInstancesBadRequest{}\n}", "func (client *KeyVaultClient) purgeDeletedSecretCreateRequest(ctx context.Context, vaultBaseURL string, secretName string, options *KeyVaultClientPurgeDeletedSecretOptions) (*policy.Request, error) {\n\thost := \"{vaultBaseUrl}\"\n\thost = strings.ReplaceAll(host, \"{vaultBaseUrl}\", vaultBaseURL)\n\turlPath := \"/deletedsecrets/{secret-name}\"\n\tif secretName == \"\" {\n\t\treturn nil, errors.New(\"parameter secretName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{secret-name}\", url.PathEscape(secretName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"7.2\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func ExchangeDelete(service string, urlBase string, urlSuffix string, credentials string, goodHttpCodes []int) (httpCode int) {\n\turl := urlBase + \"/\" + urlSuffix\n\tapiMsg := http.MethodDelete + \" \" + url\n\n\t// get message printer\n\tmsgPrinter := i18n.GetMessagePrinter()\n\n\tVerbose(apiMsg)\n\tif IsDryRun() {\n\t\treturn 204\n\t}\n\n\thttpClient := GetHTTPClient(config.HTTPRequestTimeoutS)\n\n\tresp := InvokeRestApi(httpClient, http.MethodDelete, url, credentials, nil, service, apiMsg, make(map[string]string), true)\n\tif resp.Body != nil {\n\t\tdefer resp.Body.Close()\n\t}\n\n\t// delete never returns a body\n\thttpCode = resp.StatusCode\n\tVerbose(msgPrinter.Sprintf(\"HTTP code: %d\", httpCode))\n\tif !isGoodCode(httpCode, goodHttpCodes) {\n\t\tFatal(HTTP_ERROR, msgPrinter.Sprintf(\"bad HTTP code %d from %s\", httpCode, apiMsg))\n\t}\n\treturn\n}", "func deleteEndpoint(endpointId string) (string, error) {\n\tvar ret string\n\n\turi := fmt.Sprintf(\"%s/%s\", endpointURI, id)\n\turl, err := client.GetURL(uri)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\treq, err := client.BuildReq(nil, url, http.MethodDelete, true)\n\trawResp, err := client.HTTPClient().Do(req)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\tif _, err := client.ParseReq(rawResp); err != nil {\n\t\treturn ret, err\n\t}\n\n\t// Confirm the node was deleted.\n\tif rawResp.StatusCode != 204 {\n\t\treturn ret, errors.New(\"Endpoint was not able to be deleted\")\n\t}\n\tret = \"Endpoint deleted\\n\"\n\n\treturn ret, nil\n}", "func (ctx *GetByIDHostContext) BadRequest() error {\n\tctx.ResponseData.WriteHeader(400)\n\treturn nil\n}", "func (*DeleteEndpointApiRequest) Descriptor() ([]byte, []int) {\n\treturn file_endpoint_api_proto_rawDescGZIP(), []int{2}\n}", "func NewListBadRequest(body *ListBadRequestResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func (m *TenantStatusRequestBuilder) Delete(ctx context.Context, requestConfiguration *TenantStatusRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (o *DeleteShipmentBadRequest) WithPayload(payload *ghcmessages.Error) *DeleteShipmentBadRequest {\n\to.Payload = payload\n\treturn o\n}", "func (m *ApplicationSignInDetailedSummaryApplicationSignInDetailedSummaryItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ApplicationSignInDetailedSummaryApplicationSignInDetailedSummaryItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func (m *MultiValueLegacyExtendedPropertyItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *MultiValueLegacyExtendedPropertyItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.CreateDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.requestAdapter.SendNoContentAsync(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}" ]
[ "0.6926965", "0.62658376", "0.60880667", "0.6082658", "0.59830004", "0.5965941", "0.58348936", "0.5755714", "0.57305", "0.5724605", "0.5718729", "0.57114536", "0.56896406", "0.5648961", "0.56463647", "0.5641761", "0.5615768", "0.5615768", "0.56145716", "0.5613542", "0.5605007", "0.5603735", "0.5600414", "0.5598333", "0.5581738", "0.5532515", "0.55184096", "0.5491027", "0.54849267", "0.54843515", "0.5479243", "0.5479132", "0.5479132", "0.5473186", "0.5468616", "0.5460562", "0.5457012", "0.5457012", "0.54501766", "0.545008", "0.54465514", "0.5425706", "0.54140633", "0.54107785", "0.5383026", "0.53720737", "0.5370372", "0.536557", "0.53517383", "0.5348267", "0.5338108", "0.5336691", "0.5327431", "0.53267187", "0.532358", "0.530902", "0.52973914", "0.5294016", "0.52789456", "0.5271324", "0.5268483", "0.5261469", "0.5253023", "0.52524805", "0.52465796", "0.52348", "0.523385", "0.5224399", "0.5218181", "0.52181405", "0.5215796", "0.5205816", "0.5204037", "0.51975346", "0.5196129", "0.5194841", "0.5192972", "0.51904273", "0.5182524", "0.51753813", "0.517096", "0.5169783", "0.51601034", "0.5149956", "0.51416254", "0.51360506", "0.5135186", "0.5134336", "0.51240057", "0.5123654", "0.51218283", "0.5118956", "0.5112905", "0.51114404", "0.51106393", "0.51058215", "0.5103215", "0.50954944", "0.50856173", "0.5085028" ]
0.67783815
1
NewEndpointDeleteNotFound creates a EndpointDeleteNotFound with default headers values
func NewEndpointDeleteNotFound() *EndpointDeleteNotFound { return &EndpointDeleteNotFound{} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (nse ErrNoSuchEndpoint) NotFound() {}", "func (uee *UnknownEndpointError) NotFound() {}", "func NewDeleteNotFound(body *DeleteNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDeleteNotFound(body *DeleteNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDeleteEndpoint(s Service) goa.Endpoint {\n\treturn func(ctx context.Context, req interface{}) (interface{}, error) {\n\t\tp := req.(*DeletePayload)\n\t\treturn nil, s.Delete(ctx, p)\n\t}\n}", "func DeleteEndpoint(w http.ResponseWriter, r *http.Request) {\r\n\tr.Header.Set(\"Content-Type\", \"application/json, charset=UTF-8\")\r\n\tvar dc DeleteConfig\r\n\r\n\tswitch r.Method {\r\n\tcase \"GET\":\r\n\t\tparams := mux.Vars(r)\r\n\t\tdc.Default()\r\n\t\tdc.Endpoint = params[\"endpoint_name\"]\r\n\t\tdc.EndpointList = append(dc.EndpointList, dc.Endpoint)\r\n\tcase \"POST\":\r\n\t\tbody, err := ioutil.ReadAll(r.Body)\r\n\t\tif err != nil {\r\n\t\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint, failed to read request\")\r\n\t\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Failed to read HTTP request\"))\r\n\t\t\treturn\r\n\t\t}\r\n\t\tdc.LoadParams(body)\r\n\t}\r\n\r\n\t// Verify Endpoint is provided in request body\r\n\tif len(dc.EndpointList) == 0 {\r\n\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint - endpoint is required\")\r\n\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Endpoint required\"))\r\n\t\treturn\r\n\t}\r\n\r\n\tvar query elastic.Query\r\n\r\n\t// Convert EndpointList to interface{} slice\r\n\tendpoints := make([]interface{}, len(dc.EndpointList))\r\n\tfor i, v := range dc.EndpointList {\r\n\t\tendpoints[i] = v\r\n\t}\r\n\r\n\tquery = elastic.NewBoolQuery().\r\n\t\tMust(elastic.NewWildcardQuery(\"CaseInfo.CaseName\", dc.CaseName),\r\n\t\t\telastic.NewTermsQuery(\"ComputerName.keyword\", endpoints...))\r\n\r\n\tdeleteEndpointByQuery(w, r, query, \"DeleteEndpoint\")\r\n\r\n}", "func (a *DefaultApiService) DeleteEndpoint(ctx _context.Context, id string) ApiDeleteEndpointRequest {\n\treturn ApiDeleteEndpointRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tid: id,\n\t}\n}", "func DeleteEndpoint(serviceAccountProvider provider.ServiceAccountProvider, projectProvider provider.ProjectProvider, userInfoGetter provider.UserInfoGetter) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq, ok := request.(deleteReq)\n\t\tif !ok {\n\t\t\treturn nil, errors.NewBadRequest(\"invalid request\")\n\t\t}\n\t\terr := req.Validate()\n\t\tif err != nil {\n\t\t\treturn nil, errors.NewBadRequest(err.Error())\n\t\t}\n\t\tuserInfo, err := userInfoGetter(ctx, req.ProjectID)\n\t\tif err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\t\t// check if project exist\n\t\tif _, err := projectProvider.Get(userInfo, req.ProjectID, &provider.ProjectGetOptions{}); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\t// check if service account exist before deleting it\n\t\tif _, err := serviceAccountProvider.Get(userInfo, req.ServiceAccountID, nil); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\tif err := serviceAccountProvider.Delete(userInfo, req.ServiceAccountID); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\treturn nil, nil\n\t}\n}", "func EndpointDeleteMessage(e notifications.RegenNotificationInfo) AgentNotifyMessage {\n\tnotification := EndpointNotification{\n\t\tEndpointRegenNotification: EndpointRegenNotification{\n\t\t\tID: e.GetID(),\n\t\t\tLabels: e.GetOpLabels(),\n\t\t},\n\t\tPodName: e.GetK8sPodName(),\n\t\tNamespace: e.GetK8sNamespace(),\n\t}\n\n\treturn AgentNotifyMessage{\n\t\tType: AgentNotifyEndpointDeleted,\n\t\tNotification: notification,\n\t}\n}", "func NotFoundHandler() ServiceHttpHandler { return ServiceHttpHandler{Handler: NotFound} }", "func (endpoint *HNSEndpoint) Delete() (*HNSEndpoint, error) {\n\toperation := \"Delete\"\n\ttitle := \"hcsshim::HNSEndpoint::\" + operation\n\tlogrus.Debugf(title+\" id=%s\", endpoint.Id)\n\n\treturn HNSEndpointRequest(\"DELETE\", endpoint.Id, \"\")\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func (nb *NetBuilder) FindOrCreateEndpoint(nw *Network, ep *Endpoint) error {\n\t// Query the namespace identifier.\n\tnsType, namespaceIdentifier := nb.getNamespaceIdentifier(ep)\n\n\t// Check if the endpoint already exists.\n\tendpointName := nb.generateHNSEndpointName(nw.Name, namespaceIdentifier)\n\thnsEndpoint, err := hcsshim.GetHNSEndpointByName(endpointName)\n\tif err == nil {\n\t\tlog.Infof(\"Found existing HNS endpoint %s.\", endpointName)\n\t\tif nsType == infraContainerNS || nsType == hcsNamespace {\n\t\t\t// This is a benign duplicate create call for an existing endpoint.\n\t\t\t// The endpoint was already attached in a previous call. Ignore and return success.\n\t\t\tlog.Infof(\"HNS endpoint %s is already attached to container ID %s.\",\n\t\t\t\tendpointName, ep.ContainerID)\n\t\t} else {\n\t\t\t// Attach the existing endpoint to the container's network namespace.\n\t\t\t// Attachment of endpoint to each container would occur only when using HNS V1 APIs.\n\t\t\terr = nb.attachEndpointV1(hnsEndpoint, ep.ContainerID)\n\t\t}\n\n\t\tep.MACAddress, ep.IPAddresses, nw.GatewayIPAddresses =\n\t\t\tnb.parseEndpointFieldsFromResponse(hnsEndpoint)\n\t\treturn err\n\t} else {\n\t\tif nsType != infraContainerNS && nsType != hcsNamespace {\n\t\t\t// The endpoint referenced in the container netns does not exist.\n\t\t\tlog.Errorf(\"Failed to find endpoint %s for container %s.\", endpointName, ep.ContainerID)\n\t\t\treturn fmt.Errorf(\"failed to find endpoint %s: %v\", endpointName, err)\n\t\t}\n\t}\n\n\t// Initialize the HNS endpoint.\n\thnsEndpoint = &hcsshim.HNSEndpoint{\n\t\tName: endpointName,\n\t\tVirtualNetworkName: nw.Name,\n\t\tDNSSuffix: strings.Join(nw.DNSSuffixSearchList, \",\"),\n\t\tDNSServerList: strings.Join(nw.DNSServers, \",\"),\n\t}\n\n\tif ep.MACAddress != nil {\n\t\thnsEndpoint.MacAddress = ep.MACAddress.String()\n\t}\n\tif len(ep.IPAddresses) != 0 {\n\t\thnsEndpoint.IPAddress = ep.IPAddresses[0].IP\n\t\tpl, _ := ep.IPAddresses[0].Mask.Size()\n\t\thnsEndpoint.PrefixLength = uint8(pl)\n\t}\n\n\t// Add ACL policies for blocking IMDS access through the endpoint.\n\tif ep.BlockIMDS {\n\t\terr = imds.BlockInstanceMetadataEndpoint(hnsEndpoint)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to block instance metadata endpoint: %v.\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Create the HNS endpoint.\n\tlog.Infof(\"Creating HNS endpoint: %+v\", hnsEndpoint)\n\thnsResponse, err := hnsEndpoint.Create()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create HNS endpoint: %v.\", err)\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Received HNS endpoint response: %+v.\", hnsResponse)\n\n\t// Attach the HNS endpoint to the container's network namespace.\n\tif nsType == infraContainerNS {\n\t\terr = nb.attachEndpointV1(hnsResponse, ep.ContainerID)\n\t}\n\tif nsType == hcsNamespace {\n\t\terr = nb.attachEndpointV2(hnsResponse, namespaceIdentifier)\n\t}\n\tif err != nil {\n\t\t// Cleanup the failed endpoint.\n\t\tlog.Infof(\"Deleting the failed HNS endpoint %s.\", hnsResponse.Id)\n\t\t_, delErr := hnsResponse.Delete()\n\t\tif delErr != nil {\n\t\t\tlog.Errorf(\"Failed to delete HNS endpoint: %v.\", delErr)\n\t\t}\n\n\t\treturn err\n\t}\n\n\t// Return network interface MAC address, IP Address and Gateway.\n\tep.MACAddress, ep.IPAddresses, nw.GatewayIPAddresses =\n\t\tnb.parseEndpointFieldsFromResponse(hnsResponse)\n\treturn nil\n}", "func Delete(client *gophercloud.ServiceClient, endpointID string) (r DeleteResult) {\n\t_, r.Err = client.Delete(endpointURL(client, endpointID), nil)\n\treturn\n}", "func NotFound(w http.ResponseWriter, r *http.Request) {\n\tw.WriteHeader(http.StatusNotFound)\n\tjson.NewEncoder(w).Encode(&ServiceError{\n\t\tMessage: \"Endpoint not found\",\n\t\tSolution: \"See / for possible directives\",\n\t\tErrorCode: http.StatusNotFound,\n\t})\n}", "func (client *NpmClient) EndpointDeleteReq(epinfo *netproto.Endpoint) (*netproto.Endpoint, error) {\n\treturn nil, nil\n}", "func DeleteEndpointInfo(ctx iris.Context) {\n\turi := ctx.Request().RequestURI\n\tfabricID := ctx.Params().Get(\"id\")\n\tif _, ok := capdata.FabricDataStore.Data[fabricID]; !ok {\n\t\terrMsg := fmt.Sprintf(\"Fabric data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Fabric\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tendpointData, ok := capdata.EndpointDataStore[uri]\n\tif !ok {\n\t\terrMsg := fmt.Sprintf(\"Endpoint data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Endpoint\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif endpointData.Endpoint.Links != nil && len(endpointData.Endpoint.Links.AddressPools) > 0 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be deleted as there are dependent upon AddressPool\")\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceCannotBeDeleted, errMsg, []interface{}{uri, \"Endpoint\"})\n\t\tctx.StatusCode(http.StatusNotAcceptable)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\t// Todo:Add the validation to verify the links\n\tdelete(capdata.EndpointDataStore, uri)\n\tctx.StatusCode(http.StatusNoContent)\n}", "func (a *DefaultApiService) DeleteEndpointExecute(r ApiDeleteEndpointRequest) (*_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.DeleteEndpoint\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/endpoints/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.PathEscape(parameterToString(r.id, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ErrorNotFound\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, executionError\n}", "func NewEndpointUpdateNotFound() *EndpointUpdateNotFound {\n\treturn &EndpointUpdateNotFound{}\n}", "func MakeDeleteEndpoint(s service.DepartmentService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func NewEndpointDeleteNoContent() *EndpointDeleteNoContent {\n\treturn &EndpointDeleteNoContent{}\n}", "func (c *Client) Delete(ctx context.Context, p *DeletePayload) (err error) {\n\t_, err = c.DeleteEndpoint(ctx, p)\n\treturn\n}", "func (c *V2) Delete(endpoint string) (*http.Response, error) {\n\treq, err := c.PrepareRequest(http.MethodDelete, endpoint, nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.Do(req)\n}", "func NewDeleteComponentNotFound() *DeleteComponentNotFound {\n\treturn &DeleteComponentNotFound{}\n}", "func (m *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilder) Delete(ctx context.Context, requestConfiguration *ManagedTenantsManagedTenantTicketingEndpointsManagedTenantTicketingEndpointItemRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewNotFound() error {\n\treturn requestError{\n\t\tClientError: ClientError{\n\t\t\tErrors: []clientErrorSubError{{Message: \"status code 404\"}},\n\t\t},\n\t}\n}", "func MakeDeleteEndpoint(s service.TodoService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func (c *Client) Delete(endpoint string, resource ...RequestResource) error {\n\twrapper := newWrapper(\"delete\", endpoint, resource...)\n\treturn c.do(&wrapper)\n}", "func NewDataDeleteNotFound() *DataDeleteNotFound {\n\treturn &DataDeleteNotFound{}\n}", "func (ctx *DeleteHostContext) NotFound() error {\n\tctx.ResponseData.WriteHeader(404)\n\treturn nil\n}", "func deleteEndpoint(endpointId string) (string, error) {\n\tvar ret string\n\n\turi := fmt.Sprintf(\"%s/%s\", endpointURI, id)\n\turl, err := client.GetURL(uri)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\treq, err := client.BuildReq(nil, url, http.MethodDelete, true)\n\trawResp, err := client.HTTPClient().Do(req)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\tif _, err := client.ParseReq(rawResp); err != nil {\n\t\treturn ret, err\n\t}\n\n\t// Confirm the node was deleted.\n\tif rawResp.StatusCode != 204 {\n\t\treturn ret, errors.New(\"Endpoint was not able to be deleted\")\n\t}\n\tret = \"Endpoint deleted\\n\"\n\n\treturn ret, nil\n}", "func (srv *Server) RegisterDelete(ep *Endpoint) (err error) {\n\tif ep == nil || ep.Call == nil {\n\t\treturn nil\n\t}\n\n\tep.Method = RequestMethodDelete\n\tep.RequestType = RequestTypeQuery\n\n\t// Check if the same route already registered.\n\tfor _, rute := range srv.routeDeletes {\n\t\t_, ok := rute.parse(ep.Path)\n\t\tif ok {\n\t\t\treturn ErrEndpointAmbiguous\n\t\t}\n\t}\n\n\trute, err := newRoute(ep)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tsrv.routeDeletes = append(srv.routeDeletes, rute)\n\n\treturn nil\n}", "func (s *DefaultClient) Del(endpoint string) (resp *http.Response, err error) {\n\t_, resp, err = s.http(http.MethodDelete, endpoint, nil)\n\treturn\n}", "func NewDeleteUsingDELETE8NotFound() *DeleteUsingDELETE8NotFound {\n\treturn &DeleteUsingDELETE8NotFound{}\n}", "func NewThingsDeleteNotFound() *ThingsDeleteNotFound {\n\n\treturn &ThingsDeleteNotFound{}\n}", "func (c *APIClient) Delete(endpoint Endpointer, result interface{}, headers ...func(h *http.Header)) error {\n\tepoint, err := endpoint.GetEndpoint()\n\tif err != nil {\n\t\treturn err\n\t}\n\tendUrl := c.CompileEndpointURL(epoint)\n\treq, _ := http.NewRequest(http.MethodDelete, endUrl, nil)\n\n\th := &req.Header\n\tfor _, v := range headers {\n\t\tv(h)\n\t}\n\n\tresp, err := c.Client.Call(req)\n\tc.addResponseCode(resp.StatusCode)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn MakeResponseError(resp)\n\t}\n\n\terr = json.NewDecoder(resp.Body).Decode(result)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *Driver) DeleteEndpoint(r *sdk.DeleteEndpointRequest) error {\n\tlog.Debugf(\"Delete endpoint request: %+v\", &r)\n\t//TODO: null check cidr in case driver restarted and doesn't know the network to avoid panic\n\tlog.Debugf(\"Delete endpoint %s\", r.EndpointID)\n\n\tnid := r.NetworkID\n\teid := r.EndpointID\n\n\t// Get the network handler and make sure it exists\n\td.Lock()\n\tn, ok := d.networks[nid]\n\td.Unlock()\n\n\tif !ok {\n\t\treturn types.InternalMaskableErrorf(\"network %s does not exist\", nid)\n\t}\n\tif n == nil {\n\t\treturn driverapi.ErrNoNetwork(nid)\n\t}\n\n\t// Sanity Check\n\tn.Lock()\n\tif n.id != nid {\n\t\tn.Unlock()\n\t\treturn InvalidNetworkIDError(nid)\n\t}\n\tn.Unlock()\n\n\t// Check endpoint id and if an endpoint is actually there\n\tep, err := n.getEndpoint(eid)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif ep == nil {\n\t\treturn EndpointNotFoundError(eid)\n\t}\n\n\t// Remove it\n\tn.Lock()\n\tdelete(n.endpoints, eid)\n\tn.Unlock()\n\n\t// On failure make sure to set back ep in n.endpoints, but only\n\t// if it hasn't been taken over already by some other thread.\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tn.Lock()\n\t\t\tif _, ok := n.endpoints[eid]; !ok {\n\t\t\t\tn.endpoints[eid] = ep\n\t\t\t}\n\t\t\tn.Unlock()\n\t\t}\n\t}()\n\n\t// Try removal of link. Discard error: it is a best effort.\n\t// Also make sure defer does not see this error either.\n\tif link, err := d.nlh.LinkByName(ep.srcName); err == nil {\n\t\td.nlh.LinkDel(link)\n\t}\n\n\treturn nil\n}", "func NewEndpointDeleteBadRequest() *EndpointDeleteBadRequest {\n\treturn &EndpointDeleteBadRequest{}\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(ctx context.Context, project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(ctx, path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn checkAPIResponse(bts, nil)\n}", "func NewPatchEndpointIDConfigNotFound() *PatchEndpointIDConfigNotFound {\n\treturn &PatchEndpointIDConfigNotFound{}\n}", "func (suite *TenantTestSuite) TestDeleteNotFound() {\n\trequest, _ := http.NewRequest(\"DELETE\", \"/api/v2/admin/tenants/id\", strings.NewReader(\"\"))\n\trequest.Header.Set(\"x-api-key\", suite.clientkey)\n\trequest.Header.Set(\"Accept\", \"application/json\")\n\tresponse := httptest.NewRecorder()\n\n\tsuite.router.ServeHTTP(response, request)\n\n\tcode := response.Code\n\toutput := response.Body.String()\n\n\tsuite.Equal(404, code, \"Internal Server Error\")\n\tsuite.Equal(suite.respTenantNotFound, output, \"Response body mismatch\")\n}", "func TestServiceEndpointAzureRM_Delete_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := ResourceServiceEndpointAzureRM()\n\tfor _, resource := range azurermTestServiceEndpointsAzureRM {\n\t\tresourceData := getResourceData(t, resource)\n\t\tflattenServiceEndpointAzureRM(resourceData, &resource, azurermTestServiceEndpointAzureRMProjectID)\n\n\t\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\t\tclients := &client.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\t\texpectedArgs := serviceendpoint.DeleteServiceEndpointArgs{\n\t\t\tEndpointId: resource.Id,\n\t\t\tProjectIds: &[]string{\n\t\t\t\tazurermTestServiceEndpointAzureRMProjectID.String(),\n\t\t\t},\n\t\t}\n\n\t\tbuildClient.\n\t\t\tEXPECT().\n\t\t\tDeleteServiceEndpoint(clients.Ctx, expectedArgs).\n\t\t\tReturn(errors.New(\"DeleteServiceEndpoint() Failed\")).\n\t\t\tTimes(1)\n\n\t\terr := r.Delete(resourceData, clients)\n\t\trequire.Contains(t, err.Error(), \"DeleteServiceEndpoint() Failed\")\n\t}\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn handleDeleteResponse(bts)\n}", "func NewDeleteMediaNotFound(body *DeleteMediaNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func TestFailedEndpoint0(t *testing.T) {\n\tisTesting = true\n\tvar request = Request{\n\t\tPath: \"/api/devices\",\n\t\tHTTPMethod: \"PUT\",\n\t}\n\tvar response, _ = Handler(request)\n\tif response.StatusCode != 404 {\n\t\tt.Errorf(\"response status code has to be 404 but is %d\", response.StatusCode)\n\t}\n\tif response.Body != `{\"message\":\"requested endpoint not found\"}` {\n\t\tt.Errorf(\"body is: %s\", response.Body)\n\t}\n}", "func notFound(resource string) middleware.Responder {\n\tmessage := fmt.Sprintf(\"404 %s not found\", resource)\n\treturn operations.NewGetChartDefault(http.StatusNotFound).WithPayload(\n\t\t&models.Error{Code: helpers.Int64ToPtr(http.StatusNotFound), Message: &message},\n\t)\n}", "func TestDeleteNotificationNotFound(t *testing.T) {\n\treq, err := http.NewRequest(\"DELETE\", fmt.Sprintf(\"%s/api/v1/notification/hdhdsjh\", server.URL), nil)\n\treq.SetBasicAuth(\"test\", \"test\")\n\n\tif err != nil {\n\t\tt.Fatal(\"Request failed [DELETE] /api/v1/notification\")\n\t}\n\n\tresp, _ := http.DefaultClient.Do(req)\n\n\tassert.Equal(t, 404, resp.StatusCode)\n}", "func NewGetIdentityEndpointsNotFound() *GetIdentityEndpointsNotFound {\n\n\treturn &GetIdentityEndpointsNotFound{}\n}", "func (p *fullEndpoint) delete(expectedOldValue *Endpoint) {\n\tatomic.CompareAndSwapPointer(&p.endpoint, unsafe.Pointer(expectedOldValue), nil)\n}", "func NewDeleteCustomPropertyForDeviceNotFound() *DeleteCustomPropertyForDeviceNotFound {\n\treturn &DeleteCustomPropertyForDeviceNotFound{}\n}", "func NewDeleteCrossConnectNotFound() *DeleteCrossConnectNotFound {\n\treturn &DeleteCrossConnectNotFound{}\n}", "func NewEndpointDeleteServiceUnavailable() *EndpointDeleteServiceUnavailable {\n\treturn &EndpointDeleteServiceUnavailable{}\n}", "func notFound(resp *ApiResponse, msg string) error {\n resp.StatusCode = http.StatusNotFound\n resp.Message = []byte(msg)\n resp.ErrorMessage = http.StatusText(http.StatusNotFound)\n\n return nil\n}", "func NewDeleteAPIKeyNotFound() *DeleteAPIKeyNotFound {\n\treturn &DeleteAPIKeyNotFound{}\n}", "func (m *ZebraFotaConnectorRequestBuilder) Delete(ctx context.Context, requestConfiguration *ZebraFotaConnectorRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewDeleteConfigurationNotFound() *DeleteConfigurationNotFound {\n\treturn &DeleteConfigurationNotFound{}\n}", "func (ds *DeleteSuite) TestDelete_Failure_SensorType_Not_exists() {\n\n\t// Arrange.\n\n\te := echo.New()\n\treq := httptest.NewRequest(http.MethodDelete, \"/:id\", nil)\n\tres := httptest.NewRecorder()\n\tc := e.NewContext(req, res)\n\n\tc.SetPath(\"/:id\")\n\tc.SetParamNames(\"id\")\n\tc.SetParamValues(\"99\")\n\n\t// Act.\n\n\t_ = HandleDelete(c)\n\n\t// Assert.\n\n\tassert.Equal(ds.T(), http.StatusBadRequest, res.Code)\n\tvar httpError echo.HTTPError\n\t_ = json.Unmarshal(res.Body.Bytes(), &httpError)\n\tassert.Equal(ds.T(), \"sensortype not found\", httpError.Message)\n}", "func TestEndpointCase46(t *testing.T) {\n\tvar params = EndpointParameters{\n\t\tRegion: ptr.String(\"us-east-1\"),\n\t\tUseDualStack: ptr.Bool(false),\n\t\tUseFIPS: ptr.Bool(true),\n\t\tEndpoint: ptr.String(\"https://example.com\"),\n\t}\n\n\tresolver := NewDefaultEndpointResolverV2()\n\tresult, err := resolver.ResolveEndpoint(context.Background(), params)\n\t_, _ = result, err\n\n\tif err == nil {\n\t\tt.Fatalf(\"expect error, got none\")\n\t}\n\tif e, a := \"Invalid Configuration: FIPS and custom endpoint are not supported\", err.Error(); !strings.Contains(a, e) {\n\t\tt.Errorf(\"expect %v error in %v\", e, a)\n\t}\n}", "func deleteEndpointByQuery(w http.ResponseWriter, r *http.Request, query elastic.Query, caller string) {\r\n\r\n\tclient, err := getElasticsearchClient()\r\n\tif err != nil {\r\n\t\tapi.LogError(api.DEBUG, err)\r\n\t\tfmt.Fprintf(w, api.HttpFailureMessage(\"Failed to initialized query. ERROR 0x40000012\"))\r\n\t\treturn\r\n\t}\r\n\r\n\tres, err := client.DeleteByQuery().\r\n\t\tIndex(config.ElasticIndex()).\r\n\t\tType(\"audit_type\").\r\n\t\tQuery(query).\r\n\t\tDo(context.Background())\r\n\r\n\tif err != nil {\r\n\t\tapi.HttpResponseReturn(w, r, \"failed\", err.Error(), nil)\r\n\t\treturn\r\n\t}\r\n\tapi.HttpResponseReturn(w, r, \"success\", \"Case/Endpoint deleted\", res)\r\n}", "func NewDeleteProductUsingGETNotFound() *DeleteProductUsingGETNotFound {\n\treturn &DeleteProductUsingGETNotFound{}\n}", "func NewDeleteCompanyNotFound(body *DeleteCompanyNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func (nb *NetBuilder) DeleteEndpoint(nw *Network, ep *Endpoint) error {\n\t// Generate network name here as endpoint name is dependent upon network name.\n\tnw.Name = nb.generateHNSNetworkName(nw)\n\t// Query the namespace identifier.\n\tnsType, namespaceIdentifier := nb.getNamespaceIdentifier(ep)\n\n\t// Find the HNS endpoint ID.\n\tendpointName := nb.generateHNSEndpointName(nw.Name, namespaceIdentifier)\n\thnsEndpoint, err := hcsshim.GetHNSEndpointByName(endpointName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Detach the HNS endpoint from the container's network namespace.\n\tlog.Infof(\"Detaching HNS endpoint %s from container %s netns.\", hnsEndpoint.Id, ep.ContainerID)\n\tif nsType == hcsNamespace {\n\t\t// Detach the HNS endpoint from the namespace, if we can.\n\t\t// HCN Namespace and HNS Endpoint have a 1-1 relationship, therefore,\n\t\t// even if detachment of endpoint from namespace fails, we can still proceed to delete it.\n\t\terr = hcn.RemoveNamespaceEndpoint(namespaceIdentifier, hnsEndpoint.Id)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to detach endpoint, ignoring: %v\", err)\n\t\t}\n\t} else {\n\t\terr = hcsshim.HotDetachEndpoint(ep.ContainerID, hnsEndpoint.Id)\n\t\tif err != nil && err != hcsshim.ErrComputeSystemDoesNotExist {\n\t\t\treturn err\n\t\t}\n\n\t\t// The rest of the delete logic applies to infrastructure container only.\n\t\tif nsType == nonInfraContainerNS {\n\t\t\t// For non-infra containers, the network must not be deleted.\n\t\t\tnw.UseExisting = true\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// Delete the HNS endpoint.\n\tlog.Infof(\"Deleting HNS endpoint name: %s ID: %s\", endpointName, hnsEndpoint.Id)\n\t_, err = hnsEndpoint.Delete()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to delete HNS endpoint: %v.\", err)\n\t}\n\n\treturn err\n}", "func TestAzureDevOpsServiceEndpointDockerRegistry_Delete_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := resourceServiceEndpointDockerRegistry()\n\tresourceData := schema.TestResourceDataRaw(t, r.Schema, nil)\n\tflattenServiceEndpointDockerRegistry(resourceData, &dockerRegistryTestServiceEndpoint, dockerRegistryTestServiceEndpointProjectID)\n\n\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\tclients := &config.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\texpectedArgs := serviceendpoint.DeleteServiceEndpointArgs{EndpointId: dockerRegistryTestServiceEndpoint.Id, Project: dockerRegistryTestServiceEndpointProjectID}\n\tbuildClient.\n\t\tEXPECT().\n\t\tDeleteServiceEndpoint(clients.Ctx, expectedArgs).\n\t\tReturn(errors.New(\"DeleteServiceEndpoint() Failed\")).\n\t\tTimes(1)\n\n\terr := r.Delete(resourceData, clients)\n\trequire.Contains(t, err.Error(), \"DeleteServiceEndpoint() Failed\")\n}", "func NewListByPathNotFound(body *ListByPathNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewListByPathNotFound(body *ListByPathNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewNotFound(msg string) error {\n\treturn &ELBError{\n\t\tmsg: msg,\n\t\tCode: http.StatusNotFound,\n\t}\n}", "func NewPollersDeleteNotFound() *PollersDeleteNotFound {\n\treturn &PollersDeleteNotFound{}\n}", "func (e EndPoint) Delete(container EndPointContainer) {\n\n\tentity := reflect.New(container.GetPrototype()).Interface()\n\tvar id int64\n\t_, err := fmt.Sscanf(container.GetRequest().URL.Query().Get(\":id\"), \"%d\", &id)\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusBadRequest)\n\t\treturn\n\t}\n\trepository := container.GetRepository()\n\n\terr = repository.FindByID(id, entity.(Entity))\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusNotFound)\n\t\treturn\n\t}\n\terr = container.GetSignal().Dispatch(&BeforeResourceDeleteEvent{})\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusInternalServerError)\n\t\treturn\n\t}\n\terr = repository.Delete(entity.(Entity))\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusInternalServerError)\n\t\treturn\n\t}\n\terr = container.GetSignal().Dispatch(&AfterResourceDeleteEvent{})\n\tif err != nil {\n\t\tcontainer.Error(err, http.StatusInternalServerError)\n\t\treturn\n\t}\n\tcontainer.GetResponseWriter().WriteHeader(http.StatusOK)\n}", "func NewDeactivateNotFound(body *DeactivateNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDeactivateNotFound(body *DeactivateNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func ExchangeDelete(service string, urlBase string, urlSuffix string, credentials string, goodHttpCodes []int) (httpCode int) {\n\turl := urlBase + \"/\" + urlSuffix\n\tapiMsg := http.MethodDelete + \" \" + url\n\n\t// get message printer\n\tmsgPrinter := i18n.GetMessagePrinter()\n\n\tVerbose(apiMsg)\n\tif IsDryRun() {\n\t\treturn 204\n\t}\n\n\thttpClient := GetHTTPClient(config.HTTPRequestTimeoutS)\n\n\tresp := InvokeRestApi(httpClient, http.MethodDelete, url, credentials, nil, service, apiMsg, make(map[string]string), true)\n\tif resp.Body != nil {\n\t\tdefer resp.Body.Close()\n\t}\n\n\t// delete never returns a body\n\thttpCode = resp.StatusCode\n\tVerbose(msgPrinter.Sprintf(\"HTTP code: %d\", httpCode))\n\tif !isGoodCode(httpCode, goodHttpCodes) {\n\t\tFatal(HTTP_ERROR, msgPrinter.Sprintf(\"bad HTTP code %d from %s\", httpCode, apiMsg))\n\t}\n\treturn\n}", "func CreateEndpoint(ctx iris.Context) {\n\t// Add logic to check if given ports exits\n\turi := ctx.Request().RequestURI\n\tfabricID := ctx.Params().Get(\"id\")\n\tfabricData, ok := capdata.FabricDataStore.Data[fabricID]\n\tif !ok {\n\t\terrMsg := fmt.Sprintf(\"Fabric data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Fabric\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\n\tvar endpoint model.Endpoint\n\terr := ctx.ReadJSON(&endpoint)\n\tif err != nil {\n\t\terrorMessage := \"error while trying to get JSON body from the request: \" + err.Error()\n\t\tlog.Error(errorMessage)\n\t\tresp := updateErrorResponse(response.MalformedJSON, errorMessage, nil)\n\t\tctx.StatusCode(http.StatusBadRequest)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif len(endpoint.Redundancy) < 1 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be created, Redudancy in the request is missing: \" + err.Error())\n\t\tresp := updateErrorResponse(response.PropertyMissing, errMsg, []interface{}{\"Redundancy\"})\n\t\tctx.StatusCode(http.StatusBadRequest)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif len(endpoint.Redundancy[0].RedundancySet) == 0 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be created, RedudancySet in the request is missing: \" + err.Error())\n\t\tresp := updateErrorResponse(response.PropertyMissing, errMsg, []interface{}{\"RedudancySet\"})\n\t\tctx.StatusCode(http.StatusBadRequest)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\t// get all existing endpoints under fabric check for the name\n\tfor _, endpointData := range capdata.EndpointDataStore {\n\t\tif endpoint.Name == endpointData.Endpoint.Name {\n\t\t\terrMsg := \"Endpoint name is already assigned to other endpoint:\" + endpointData.Endpoint.Name\n\t\t\tresp := updateErrorResponse(response.ResourceAlreadyExists, errMsg, []interface{}{\"Endpoint\", endpointData.Endpoint.Name, endpoint.Name})\n\t\t\tctx.StatusCode(http.StatusConflict)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\t\t}\n\t}\n\tvar switchURI = \"\"\n\tvar portPattern = \"\"\n\tportList := make(map[string]bool)\n\t// check if given ports are present in plugin database\n\tfor i := 0; i < len(endpoint.Redundancy[0].RedundancySet); i++ {\n\t\tportURI := endpoint.Redundancy[0].RedundancySet[i].Oid\n\t\tif _, ok := portList[endpoint.Redundancy[0].RedundancySet[i].Oid]; ok {\n\t\t\terrMsg := \"Duplicate port passed in the request\"\n\t\t\tresp := updateErrorResponse(response.PropertyValueConflict, errMsg, []interface{}{endpoint.Redundancy[0].RedundancySet[i].Oid, endpoint.Redundancy[0].RedundancySet[i].Oid})\n\t\t\tctx.StatusCode(http.StatusBadRequest)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\n\t\t}\n\t\tportList[endpoint.Redundancy[0].RedundancySet[i].Oid] = true\n\n\t\t_, statusCode, resp := getPortData(portURI)\n\t\tif statusCode != http.StatusOK {\n\t\t\tctx.StatusCode(statusCode)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\t\t}\n\t\tstatusCode, resp = checkEndpointPortMapping(endpoint.Redundancy[0].RedundancySet[i].Oid)\n\t\tif statusCode != http.StatusOK {\n\t\t\tctx.StatusCode(statusCode)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\t\t}\n\t\tportURIData := strings.Split(portURI, \"/\")\n\t\tswitchID := portURIData[6]\n\t\tswitchIDData := strings.Split(switchID, \":\")\n\t\tswitchURI = switchURI + \"-\" + switchIDData[1]\n\t\tportIDData := strings.Split(portURIData[8], \":\")\n\t\ttmpPortPattern := strings.Replace(portIDData[1], \"eth\", \"\", -1)\n\t\ttmpPortPattern = strings.Replace(tmpPortPattern, \"-\", \"-ports-\", -1)\n\t\tportPattern = tmpPortPattern\n\t}\n\n\tportPolicyGroupList, err := caputilities.GetPortPolicyGroup(fabricData.PodID, switchURI)\n\tif err != nil || len(portPolicyGroupList) == 0 {\n\t\terrMsg := \"Port policy group not found for given ports\"\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"protpaths\" + switchURI, \"PolicyGroup\"})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\n\t}\n\tpolicyGroupDN := \"\"\n\tfor i := 0; i < len(portPolicyGroupList); i++ {\n\t\tif strings.Contains(portPolicyGroupList[i].BaseAttributes.DistinguishedName, portPattern) {\n\t\t\tpolicyGroupDN = portPolicyGroupList[i].BaseAttributes.DistinguishedName\n\t\t}\n\t}\n\tif policyGroupDN == \"\" {\n\t\terrMsg := \"Port policy group not found for given ports\"\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{portPattern, \"PolicyGroup\"})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tlog.Info(\"Dn of Policy group:\" + policyGroupDN)\n\tsaveEndpointData(uri, fabricID, policyGroupDN, &endpoint)\n\tcommon.SetResponseHeader(ctx, map[string]string{\n\t\t\"Location\": endpoint.ODataID,\n\t})\n\tctx.StatusCode(http.StatusCreated)\n\tctx.JSON(endpoint)\n}", "func NewNotFound(err error, msg ...string) *Errs {\n\tif err == nil {\n\t\terr = ErrNotFound\n\t}\n\treturn &Errs{\n\t\tcodeHTTP: http.StatusNotFound,\n\t\terr: err,\n\t\tkind: trace(2),\n\t\tmessage: msg,\n\t}\n}", "func delEndpoint(domain string, addr string) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\n\tservice, ok := registryMap[domain]\n\tif ok == false {\n\t\treturn ErrServiceNotFound\n\t}\n\tfor k, item := range service.Items {\n\t\tif item.Endpoint == addr {\n\t\t\tendpoints := append(service.Items[:k], service.Items[k+1:]...)\n\t\t\tservice.Items = endpoints\n\t\t\tbreak\n\t\t}\n\t}\n\tregistryMap[domain] = service\n\treturn nil\n}", "func NewNotFound(s string, v ...interface{}) error {\n\treturn asNotFound(fmt.Errorf(s, v...))\n}", "func (v *ModelVerifier) EndpointNotExists(name string) *ModelVerifier {\n\tv.verifiers = append(v.verifiers, &endpointVerifier{\n\t\texists: false,\n\t\tname: name,\n\n\t\tmodel: v.model,\n\t})\n\n\treturn v\n}", "func NewDeleteShipmentNotFound() *DeleteShipmentNotFound {\n\n\treturn &DeleteShipmentNotFound{}\n}", "func NewDeleteTagNotFound() *DeleteTagNotFound {\n\treturn &DeleteTagNotFound{}\n}", "func NewDeleteTagNotFound() *DeleteTagNotFound {\n\treturn &DeleteTagNotFound{}\n}", "func NewDeleteVendorNotFound() *DeleteVendorNotFound {\n\n\treturn &DeleteVendorNotFound{}\n}", "func PostDeleteEndpoint(deleter deleting.Service) func(w http.ResponseWriter, r *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tw.Header().Set(\"Content-Type\", \"application/json\")\n\n\t\tid := mux.Vars(r)[\"id\"]\n\n\t\t// Delete a post.\n\t\terr := deleter.PostDelete(id)\n\t\tif err != nil {\n\t\t\tlog.Println(\"Error deleting a post:\", err)\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tw.WriteHeader(http.StatusNoContent)\n\t}\n}", "func TestAfPacketDeletedLinuxInterfaceHostNotFound(t *testing.T) {\n\tctx, plugin, _ := afPacketTestSetup(t)\n\tdefer afPacketTestTeardown(ctx)\n\n\t// Prepare\n\tplugin.ResolveCreatedLinuxInterface(\"host1\", \"host1\", 1)\n\t// Test un-registered linux interface\n\tplugin.ResolveDeletedLinuxInterface(\"host1\", \"host1\", 1)\n\tExpect(plugin.GetHostInterfacesEntry(\"host1\")).To(BeFalse())\n}", "func NewDeleteConsulNotFound() *DeleteConsulNotFound {\n\n\treturn &DeleteConsulNotFound{}\n}", "func EndpointCreateMessage(e notifications.RegenNotificationInfo) AgentNotifyMessage {\n\tnotification := EndpointNotification{\n\t\tEndpointRegenNotification: EndpointRegenNotification{\n\t\t\tID: e.GetID(),\n\t\t\tLabels: e.GetOpLabels(),\n\t\t},\n\t\tPodName: e.GetK8sPodName(),\n\t\tNamespace: e.GetK8sNamespace(),\n\t}\n\n\treturn AgentNotifyMessage{\n\t\tType: AgentNotifyEndpointCreated,\n\t\tNotification: notification,\n\t}\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func NewDeleteCapabilityNotFound() *DeleteCapabilityNotFound {\n\treturn &DeleteCapabilityNotFound{}\n}", "func (it *integTestSuite) TestNpmEndpointCreateDelete(c *C) {\n\t// create a network in controller\n\t// if not present create the default tenant\n\tit.CreateTenant(\"default\")\n\terr := it.DeleteAllEndpoints(c)\n\tc.Assert(err, IsNil)\n\terr = it.CreateNetwork(\"default\", \"default\", \"testNetwork\", \"10.1.0.0/22\", \"10.1.1.254\")\n\tc.Assert(err, IsNil)\n\tAssertEventually(c, func() (bool, interface{}) {\n\t\t_, nerr := it.npmCtrler.StateMgr.FindNetwork(\"default\", \"testNetwork\")\n\t\treturn (nerr == nil), nil\n\t}, \"Network not found in statemgr\")\n\n\t// wait till agent has the network\n\tfor _, ag := range it.agents {\n\t\tAssertEventually(c, func() (bool, interface{}) {\n\n\t\t\tnt := netproto.Network{\n\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\t\t\tObjectMeta: api.ObjectMeta{Tenant: \"default\", Namespace: \"default\", Name: \"testNetwork\"},\n\t\t\t}\n\t\t\t_, nerr := ag.dscAgent.PipelineAPI.HandleNetwork(agentTypes.Get, nt)\n\t\t\treturn (nerr == nil), nil\n\t\t}, \"Network not found in agent\")\n\t}\n\n\t// create a wait channel\n\twaitCh := make(chan error, it.numAgents*2)\n\n\t// create one endpoint from each agent\n\tfor i, ag := range it.agents {\n\t\tgo func(i int, ag *Dpagent) {\n\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\t\t\thostName := fmt.Sprintf(\"testHost-%d\", i)\n\n\t\t\t// make the call\n\t\t\tcerr := it.CreateEndpoint(\"default\", \"default\", \"testNetwork\", epname, hostName, \"0101.0101.0101\", hostName, \"20.1.1.1\", map[string]string{\"env\": \"production\", \"app\": \"procurement\"}, 2)\n\t\t\tif cerr != nil {\n\t\t\t\twaitCh <- fmt.Errorf(\"endpoint create failed: %v\", cerr)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(i, ag)\n\t}\n\n\t// wait for all endpoint creates to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Error during endpoint create\")\n\t}\n\n\t// wait for all endpoints to be propagated to other agents\n\tfor _, ag := range it.agents {\n\t\tgo func(ag *Dpagent) {\n\t\t\tfound := CheckEventually(func() (bool, interface{}) {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\treturn len(endpoints) == it.numAgents, nil\n\t\t\t}, \"10ms\", it.pollTimeout())\n\t\t\tif !found {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\tlog.Infof(\"Endpoint count expected [%v] found [%v]\", it.numAgents, len(endpoints))\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint count incorrect in datapath\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor i := range it.agents {\n\t\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\t\t\t\tepmeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t\tObjectMeta: api.ObjectMeta{\n\t\t\t\t\t\tTenant: \"default\",\n\t\t\t\t\t\tNamespace: \"default\",\n\t\t\t\t\t\tName: epname,\n\t\t\t\t\t},\n\t\t\t\t}\n\t\t\t\t_, perr := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.Get, epmeta)\n\t\t\t\tif perr != nil {\n\t\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint not found in datapath\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\twaitCh <- nil\n\t\t}(ag)\n\t}\n\n\t// wait for all goroutines to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint info incorrect in datapath\")\n\t}\n\n\t// now delete the endpoints\n\tfor i, ag := range it.agents {\n\t\tgo func(i int, ag *Dpagent) {\n\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\n\t\t\t// make the call\n\t\t\tcerr := it.DeleteEndpoint(\"default\", \"default\", epname)\n\t\t\tif cerr != nil {\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint delete failed: %v\", cerr)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(i, ag)\n\t}\n\n\t// wait for all endpoint deletes to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint delete failed\")\n\n\t}\n\n\tfor _, ag := range it.agents {\n\t\tgo func(ag *Dpagent) {\n\t\t\tif !CheckEventually(func() (bool, interface{}) {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\treturn len(endpoints) == 0, nil\n\t\t\t}, \"10ms\", it.pollTimeout()) {\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint was not deleted from datapath\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(ag)\n\t}\n\n\t// wait for all goroutines to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint delete error\")\n\t}\n\n\t// delete the network\n\terr = it.DeleteNetwork(\"default\", \"testNetwork\")\n\tc.Assert(err, IsNil)\n\tAssertEventually(c, func() (bool, interface{}) {\n\t\t_, nerr := it.npmCtrler.StateMgr.FindNetwork(\"default\", \"testNetwork\")\n\t\treturn (nerr != nil), nil\n\t}, \"Network still found in statemgr\")\n}", "func MakeDeleteNodeEndpoint(s registry.Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteNodeRequest)\n\t\terr := s.DeleteNode(ctx, req.Token, req.Id)\n\t\treturn DeleteNodeResponse{Err: err}, nil\n\t}\n}", "func (m *WindowsUpdatesCatalogRequestBuilder) Delete(ctx context.Context, requestConfiguration *WindowsUpdatesCatalogRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewAPIDeleteUserNotFound() *APIDeleteUserNotFound {\n\treturn &APIDeleteUserNotFound{}\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsClientDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func DeleteHostNCApipaEndpoint(\n\tnetworkContainerID string) error {\n\treturn nil\n}", "func NewDelete1NotFound() *Delete1NotFound {\n\treturn &Delete1NotFound{}\n}", "func (r workloadEndpoints) Delete(ctx context.Context, namespace, name string, opts options.DeleteOptions) (*apiv3.WorkloadEndpoint, error) {\r\n\tout, err := r.client.resources.Delete(ctx, opts, apiv3.KindWorkloadEndpoint, namespace, name)\r\n\tif out != nil {\r\n\t\treturn out.(*apiv3.WorkloadEndpoint), err\r\n\t}\r\n\treturn nil, err\r\n}", "func NewDeleteAPIKeyDefault(code int) *DeleteAPIKeyDefault {\n\treturn &DeleteAPIKeyDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (*DeleteEndpointRequest) Descriptor() ([]byte, []int) {\n\treturn file_endpoint_api_proto_rawDescGZIP(), []int{11}\n}", "func (m *GroupPolicyDefinitionsItemNextVersionDefinitionPreviousVersionDefinitionRequestBuilder) Delete(ctx context.Context, requestConfiguration *GroupPolicyDefinitionsItemNextVersionDefinitionPreviousVersionDefinitionRequestBuilderDeleteRequestConfiguration)(error) {\n requestInfo, err := m.ToDeleteRequestInformation(ctx, requestConfiguration);\n if err != nil {\n return err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": i20a3050780ee0b0cde0a884a4f35429a20d60067e3bcda382ec5400079147459.CreateODataErrorFromDiscriminatorValue,\n }\n err = m.BaseRequestBuilder.RequestAdapter.SendNoContent(ctx, requestInfo, errorMapping)\n if err != nil {\n return err\n }\n return nil\n}", "func NewRevokeNotFound() *RevokeNotFound {\n\treturn &RevokeNotFound{}\n}", "func NewNotFoundError(message string)*RestErr{\n\treturn &RestErr{\n\t\tMessage: message,\n\t\tStatus: http.StatusNotFound,\n\t\tError: \"Not Found\",\n\t}\n}" ]
[ "0.7107122", "0.6872569", "0.6531333", "0.6531333", "0.63523686", "0.63063663", "0.62974745", "0.6083681", "0.60352474", "0.5882893", "0.5882644", "0.5872145", "0.5872145", "0.5846968", "0.58195734", "0.580794", "0.5801622", "0.5795623", "0.5771072", "0.5759165", "0.5699919", "0.56889796", "0.5635919", "0.5635451", "0.5613191", "0.5608332", "0.5589682", "0.55864894", "0.5550732", "0.55206096", "0.55120176", "0.5511406", "0.55040985", "0.55037725", "0.5489709", "0.5472489", "0.54634833", "0.54634774", "0.54369503", "0.543131", "0.54161334", "0.54130715", "0.5408421", "0.54082346", "0.54054993", "0.5402999", "0.53944224", "0.53739816", "0.5364091", "0.5359364", "0.5341378", "0.53312325", "0.5312051", "0.5310487", "0.52985257", "0.52975106", "0.5265858", "0.5253287", "0.52472067", "0.5236633", "0.52292323", "0.5226261", "0.5226175", "0.5216591", "0.5209886", "0.5209886", "0.5205411", "0.5205089", "0.52047676", "0.52009237", "0.52009237", "0.5200738", "0.5195701", "0.51944494", "0.51888674", "0.5186521", "0.5183859", "0.51768667", "0.51756424", "0.51756424", "0.51489", "0.5129286", "0.512609", "0.5123724", "0.51187676", "0.5116403", "0.51147145", "0.51085544", "0.51028645", "0.50946194", "0.50929576", "0.5080714", "0.5076137", "0.5068257", "0.50484335", "0.50447625", "0.5044065", "0.5043926", "0.50437856", "0.5043483" ]
0.6772105
2
NewEndpointDeleteServiceUnavailable creates a EndpointDeleteServiceUnavailable with default headers values
func NewEndpointDeleteServiceUnavailable() *EndpointDeleteServiceUnavailable { return &EndpointDeleteServiceUnavailable{} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewServiceUnavailable(cause error) ServiceUnavailable { return ServiceUnavailable(cause.Error()) }", "func NewDeleteEndpoint(s Service) goa.Endpoint {\n\treturn func(ctx context.Context, req interface{}) (interface{}, error) {\n\t\tp := req.(*DeletePayload)\n\t\treturn nil, s.Delete(ctx, p)\n\t}\n}", "func NewEndpointUpdateServiceUnavailable() *EndpointUpdateServiceUnavailable {\n\treturn &EndpointUpdateServiceUnavailable{}\n}", "func DeleteEndpoint(w http.ResponseWriter, r *http.Request) {\r\n\tr.Header.Set(\"Content-Type\", \"application/json, charset=UTF-8\")\r\n\tvar dc DeleteConfig\r\n\r\n\tswitch r.Method {\r\n\tcase \"GET\":\r\n\t\tparams := mux.Vars(r)\r\n\t\tdc.Default()\r\n\t\tdc.Endpoint = params[\"endpoint_name\"]\r\n\t\tdc.EndpointList = append(dc.EndpointList, dc.Endpoint)\r\n\tcase \"POST\":\r\n\t\tbody, err := ioutil.ReadAll(r.Body)\r\n\t\tif err != nil {\r\n\t\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint, failed to read request\")\r\n\t\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Failed to read HTTP request\"))\r\n\t\t\treturn\r\n\t\t}\r\n\t\tdc.LoadParams(body)\r\n\t}\r\n\r\n\t// Verify Endpoint is provided in request body\r\n\tif len(dc.EndpointList) == 0 {\r\n\t\tapi.LogDebug(api.DEBUG, \"[+] POST /delete/endpoint - endpoint is required\")\r\n\t\tfmt.Fprintln(w, api.HttpFailureMessage(\"Endpoint required\"))\r\n\t\treturn\r\n\t}\r\n\r\n\tvar query elastic.Query\r\n\r\n\t// Convert EndpointList to interface{} slice\r\n\tendpoints := make([]interface{}, len(dc.EndpointList))\r\n\tfor i, v := range dc.EndpointList {\r\n\t\tendpoints[i] = v\r\n\t}\r\n\r\n\tquery = elastic.NewBoolQuery().\r\n\t\tMust(elastic.NewWildcardQuery(\"CaseInfo.CaseName\", dc.CaseName),\r\n\t\t\telastic.NewTermsQuery(\"ComputerName.keyword\", endpoints...))\r\n\r\n\tdeleteEndpointByQuery(w, r, query, \"DeleteEndpoint\")\r\n\r\n}", "func (nse ErrNoSuchEndpoint) NotFound() {}", "func (a *DefaultApiService) DeleteEndpoint(ctx _context.Context, id string) ApiDeleteEndpointRequest {\n\treturn ApiDeleteEndpointRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t\tid: id,\n\t}\n}", "func DeleteEndpoint(serviceAccountProvider provider.ServiceAccountProvider, projectProvider provider.ProjectProvider, userInfoGetter provider.UserInfoGetter) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq, ok := request.(deleteReq)\n\t\tif !ok {\n\t\t\treturn nil, errors.NewBadRequest(\"invalid request\")\n\t\t}\n\t\terr := req.Validate()\n\t\tif err != nil {\n\t\t\treturn nil, errors.NewBadRequest(err.Error())\n\t\t}\n\t\tuserInfo, err := userInfoGetter(ctx, req.ProjectID)\n\t\tif err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\t\t// check if project exist\n\t\tif _, err := projectProvider.Get(userInfo, req.ProjectID, &provider.ProjectGetOptions{}); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\t// check if service account exist before deleting it\n\t\tif _, err := serviceAccountProvider.Get(userInfo, req.ServiceAccountID, nil); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\tif err := serviceAccountProvider.Delete(userInfo, req.ServiceAccountID); err != nil {\n\t\t\treturn nil, common.KubernetesErrorToHTTPError(err)\n\t\t}\n\n\t\treturn nil, nil\n\t}\n}", "func EndpointDeleteMessage(e notifications.RegenNotificationInfo) AgentNotifyMessage {\n\tnotification := EndpointNotification{\n\t\tEndpointRegenNotification: EndpointRegenNotification{\n\t\t\tID: e.GetID(),\n\t\t\tLabels: e.GetOpLabels(),\n\t\t},\n\t\tPodName: e.GetK8sPodName(),\n\t\tNamespace: e.GetK8sNamespace(),\n\t}\n\n\treturn AgentNotifyMessage{\n\t\tType: AgentNotifyEndpointDeleted,\n\t\tNotification: notification,\n\t}\n}", "func (uee *UnknownEndpointError) NotFound() {}", "func Delete(client *gophercloud.ServiceClient, endpointID string) (r DeleteResult) {\n\t_, r.Err = client.Delete(endpointURL(client, endpointID), nil)\n\treturn\n}", "func NewEndpointDeleteNotFound() *EndpointDeleteNotFound {\n\treturn &EndpointDeleteNotFound{}\n}", "func (s *DefaultClient) Del(endpoint string) (resp *http.Response, err error) {\n\t_, resp, err = s.http(http.MethodDelete, endpoint, nil)\n\treturn\n}", "func serviceUnavailable(resp *ApiResponse, msg string) error {\n resp.StatusCode = http.StatusServiceUnavailable\n resp.Message = []byte(msg)\n resp.ErrorMessage = msg\n\n return nil\n}", "func NewEndpointGroupUpdateServiceUnavailable() *EndpointGroupUpdateServiceUnavailable {\n\treturn &EndpointGroupUpdateServiceUnavailable{}\n}", "func (c *Client) Delete(ctx context.Context, p *DeletePayload) (err error) {\n\t_, err = c.DeleteEndpoint(ctx, p)\n\treturn\n}", "func NewServiceAddServiceUnavailable() *ServiceAddServiceUnavailable {\n\n\treturn &ServiceAddServiceUnavailable{}\n}", "func badEndpointsResource(edsServiceName string, host string, ports []uint32) *v3endpointpb.ClusterLoadAssignment {\n\te := e2e.DefaultEndpoint(edsServiceName, host, ports)\n\te.Endpoints[0].LbEndpoints[0].LoadBalancingWeight = &wrapperspb.UInt32Value{Value: 0}\n\treturn e\n}", "func (client *NpmClient) EndpointDeleteReq(epinfo *netproto.Endpoint) (*netproto.Endpoint, error) {\n\treturn nil, nil\n}", "func ServiceUnavailable(message ...interface{}) Err {\n\treturn Boomify(http.StatusServiceUnavailable, message...)\n}", "func (client *WebAppsClient) deletePrivateEndpointConnectionCreateRequest(ctx context.Context, resourceGroupName string, name string, privateEndpointConnectionName string, options *WebAppsBeginDeletePrivateEndpointConnectionOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/privateEndpointConnections/{privateEndpointConnectionName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif privateEndpointConnectionName == \"\" {\n\t\treturn nil, errors.New(\"parameter privateEndpointConnectionName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{privateEndpointConnectionName}\", url.PathEscape(privateEndpointConnectionName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func TestServiceEndpointAzureRM_Delete_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := ResourceServiceEndpointAzureRM()\n\tfor _, resource := range azurermTestServiceEndpointsAzureRM {\n\t\tresourceData := getResourceData(t, resource)\n\t\tflattenServiceEndpointAzureRM(resourceData, &resource, azurermTestServiceEndpointAzureRMProjectID)\n\n\t\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\t\tclients := &client.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\t\texpectedArgs := serviceendpoint.DeleteServiceEndpointArgs{\n\t\t\tEndpointId: resource.Id,\n\t\t\tProjectIds: &[]string{\n\t\t\t\tazurermTestServiceEndpointAzureRMProjectID.String(),\n\t\t\t},\n\t\t}\n\n\t\tbuildClient.\n\t\t\tEXPECT().\n\t\t\tDeleteServiceEndpoint(clients.Ctx, expectedArgs).\n\t\t\tReturn(errors.New(\"DeleteServiceEndpoint() Failed\")).\n\t\t\tTimes(1)\n\n\t\terr := r.Delete(resourceData, clients)\n\t\trequire.Contains(t, err.Error(), \"DeleteServiceEndpoint() Failed\")\n\t}\n}", "func ExampleRDS_DeleteDBClusterEndpoint_shared00() {\n\tsvc := rds.New(session.New())\n\tinput := &rds.DeleteDBClusterEndpointInput{\n\t\tDBClusterEndpointIdentifier: aws.String(\"mycustomendpoint\"),\n\t}\n\n\tresult, err := svc.DeleteDBClusterEndpoint(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeInvalidDBClusterEndpointStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBClusterEndpointStateFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBClusterEndpointNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBClusterEndpointNotFoundFault, aerr.Error())\n\t\t\tcase rds.ErrCodeInvalidDBClusterStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBClusterStateFault, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func NewEndpoint(service health.Service) *Endpoint {\n\treturn &Endpoint{\n\t\tservice: service,\n\t}\n}", "func NewDeleteDatasetServiceUnavailable() *DeleteDatasetServiceUnavailable {\n\treturn &DeleteDatasetServiceUnavailable{}\n}", "func ExamplePrivateEndpointsClient_BeginDelete() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armstreamanalytics.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := clientFactory.NewPrivateEndpointsClient().BeginDelete(ctx, \"sjrg\", \"testcluster\", \"testpe\", nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t_, err = poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n}", "func NewDeleteNotFound(body *DeleteNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDeleteNotFound(body *DeleteNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewServiceCreateServiceUnavailable() *ServiceCreateServiceUnavailable {\n\treturn &ServiceCreateServiceUnavailable{}\n}", "func (c *V2) Delete(endpoint string) (*http.Response, error) {\n\treq, err := c.PrepareRequest(http.MethodDelete, endpoint, nil, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.Do(req)\n}", "func TestFailedEndpoint0(t *testing.T) {\n\tisTesting = true\n\tvar request = Request{\n\t\tPath: \"/api/devices\",\n\t\tHTTPMethod: \"PUT\",\n\t}\n\tvar response, _ = Handler(request)\n\tif response.StatusCode != 404 {\n\t\tt.Errorf(\"response status code has to be 404 but is %d\", response.StatusCode)\n\t}\n\tif response.Body != `{\"message\":\"requested endpoint not found\"}` {\n\t\tt.Errorf(\"body is: %s\", response.Body)\n\t}\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func (e Endpoints) Delete(ctx context.Context, id string) (error error) {\n\trequest := DeleteRequest{Id: id}\n\tresponse, err := e.DeleteEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(DeleteResponse).Error\n}", "func DeleteEndpointInfo(ctx iris.Context) {\n\turi := ctx.Request().RequestURI\n\tfabricID := ctx.Params().Get(\"id\")\n\tif _, ok := capdata.FabricDataStore.Data[fabricID]; !ok {\n\t\terrMsg := fmt.Sprintf(\"Fabric data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Fabric\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tendpointData, ok := capdata.EndpointDataStore[uri]\n\tif !ok {\n\t\terrMsg := fmt.Sprintf(\"Endpoint data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Endpoint\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif endpointData.Endpoint.Links != nil && len(endpointData.Endpoint.Links.AddressPools) > 0 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be deleted as there are dependent upon AddressPool\")\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceCannotBeDeleted, errMsg, []interface{}{uri, \"Endpoint\"})\n\t\tctx.StatusCode(http.StatusNotAcceptable)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\t// Todo:Add the validation to verify the links\n\tdelete(capdata.EndpointDataStore, uri)\n\tctx.StatusCode(http.StatusNoContent)\n}", "func deleteEndpoint(endpointId string) (string, error) {\n\tvar ret string\n\n\turi := fmt.Sprintf(\"%s/%s\", endpointURI, id)\n\turl, err := client.GetURL(uri)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\treq, err := client.BuildReq(nil, url, http.MethodDelete, true)\n\trawResp, err := client.HTTPClient().Do(req)\n\tif err != nil {\n\t\treturn ret, err\n\t}\n\tif _, err := client.ParseReq(rawResp); err != nil {\n\t\treturn ret, err\n\t}\n\n\t// Confirm the node was deleted.\n\tif rawResp.StatusCode != 204 {\n\t\treturn ret, errors.New(\"Endpoint was not able to be deleted\")\n\t}\n\tret = \"Endpoint deleted\\n\"\n\n\treturn ret, nil\n}", "func (c *Client) Delete(endpoint string, resource ...RequestResource) error {\n\twrapper := newWrapper(\"delete\", endpoint, resource...)\n\treturn c.do(&wrapper)\n}", "func (d *Driver) DeleteEndpoint(r *sdk.DeleteEndpointRequest) error {\n\tlog.Debugf(\"Delete endpoint request: %+v\", &r)\n\t//TODO: null check cidr in case driver restarted and doesn't know the network to avoid panic\n\tlog.Debugf(\"Delete endpoint %s\", r.EndpointID)\n\n\tnid := r.NetworkID\n\teid := r.EndpointID\n\n\t// Get the network handler and make sure it exists\n\td.Lock()\n\tn, ok := d.networks[nid]\n\td.Unlock()\n\n\tif !ok {\n\t\treturn types.InternalMaskableErrorf(\"network %s does not exist\", nid)\n\t}\n\tif n == nil {\n\t\treturn driverapi.ErrNoNetwork(nid)\n\t}\n\n\t// Sanity Check\n\tn.Lock()\n\tif n.id != nid {\n\t\tn.Unlock()\n\t\treturn InvalidNetworkIDError(nid)\n\t}\n\tn.Unlock()\n\n\t// Check endpoint id and if an endpoint is actually there\n\tep, err := n.getEndpoint(eid)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif ep == nil {\n\t\treturn EndpointNotFoundError(eid)\n\t}\n\n\t// Remove it\n\tn.Lock()\n\tdelete(n.endpoints, eid)\n\tn.Unlock()\n\n\t// On failure make sure to set back ep in n.endpoints, but only\n\t// if it hasn't been taken over already by some other thread.\n\tdefer func() {\n\t\tif err != nil {\n\t\t\tn.Lock()\n\t\t\tif _, ok := n.endpoints[eid]; !ok {\n\t\t\t\tn.endpoints[eid] = ep\n\t\t\t}\n\t\t\tn.Unlock()\n\t\t}\n\t}()\n\n\t// Try removal of link. Discard error: it is a best effort.\n\t// Also make sure defer does not see this error either.\n\tif link, err := d.nlh.LinkByName(ep.srcName); err == nil {\n\t\td.nlh.LinkDel(link)\n\t}\n\n\treturn nil\n}", "func (p *fullEndpoint) delete(expectedOldValue *Endpoint) {\n\tatomic.CompareAndSwapPointer(&p.endpoint, unsafe.Pointer(expectedOldValue), nil)\n}", "func TestAzureDevOpsServiceEndpointDockerRegistry_Delete_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := resourceServiceEndpointDockerRegistry()\n\tresourceData := schema.TestResourceDataRaw(t, r.Schema, nil)\n\tflattenServiceEndpointDockerRegistry(resourceData, &dockerRegistryTestServiceEndpoint, dockerRegistryTestServiceEndpointProjectID)\n\n\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\tclients := &config.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\texpectedArgs := serviceendpoint.DeleteServiceEndpointArgs{EndpointId: dockerRegistryTestServiceEndpoint.Id, Project: dockerRegistryTestServiceEndpointProjectID}\n\tbuildClient.\n\t\tEXPECT().\n\t\tDeleteServiceEndpoint(clients.Ctx, expectedArgs).\n\t\tReturn(errors.New(\"DeleteServiceEndpoint() Failed\")).\n\t\tTimes(1)\n\n\terr := r.Delete(resourceData, clients)\n\trequire.Contains(t, err.Error(), \"DeleteServiceEndpoint() Failed\")\n}", "func MakeDeleteEndpoint(s service.DepartmentService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func NewEndpoint(ctx *pulumi.Context,\n\tname string, args *EndpointArgs, opts ...pulumi.ResourceOption) (*Endpoint, error) {\n\tif args == nil || args.EndpointId == nil {\n\t\treturn nil, errors.New(\"missing required argument 'EndpointId'\")\n\t}\n\tif args == nil || args.Service == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Service'\")\n\t}\n\tif args == nil {\n\t\targs = &EndpointArgs{}\n\t}\n\tvar resource Endpoint\n\terr := ctx.RegisterResource(\"gcp:servicedirectory/endpoint:Endpoint\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (endpoint *HNSEndpoint) Delete() (*HNSEndpoint, error) {\n\toperation := \"Delete\"\n\ttitle := \"hcsshim::HNSEndpoint::\" + operation\n\tlogrus.Debugf(title+\" id=%s\", endpoint.Id)\n\n\treturn HNSEndpointRequest(\"DELETE\", endpoint.Id, \"\")\n}", "func NewWithoutDefaults(endpoint string) BaseClient {\n\treturn BaseClient{\n\t\tClient: autorest.NewClientWithUserAgent(UserAgent()),\n\t\tEndpoint: endpoint,\n\t}\n}", "func delEndpoint(domain string, addr string) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\n\tservice, ok := registryMap[domain]\n\tif ok == false {\n\t\treturn ErrServiceNotFound\n\t}\n\tfor k, item := range service.Items {\n\t\tif item.Endpoint == addr {\n\t\t\tendpoints := append(service.Items[:k], service.Items[k+1:]...)\n\t\t\tservice.Items = endpoints\n\t\t\tbreak\n\t\t}\n\t}\n\tregistryMap[domain] = service\n\treturn nil\n}", "func NewEndpointDeleteNoContent() *EndpointDeleteNoContent {\n\treturn &EndpointDeleteNoContent{}\n}", "func (r workloadEndpoints) Delete(ctx context.Context, namespace, name string, opts options.DeleteOptions) (*apiv3.WorkloadEndpoint, error) {\r\n\tout, err := r.client.resources.Delete(ctx, opts, apiv3.KindWorkloadEndpoint, namespace, name)\r\n\tif out != nil {\r\n\t\treturn out.(*apiv3.WorkloadEndpoint), err\r\n\t}\r\n\treturn nil, err\r\n}", "func serviceUnavailable(rw http.ResponseWriter, r *http.Request) {\n\n}", "func NewGetPublicIPByUUIDUsingGETServiceUnavailable() *GetPublicIPByUUIDUsingGETServiceUnavailable {\n\treturn &GetPublicIPByUUIDUsingGETServiceUnavailable{}\n}", "func NewDeleteWaitlistEntryServiceUnavailable(body *DeleteWaitlistEntryServiceUnavailableResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func (r *Responder) ServiceUnavailable() { r.write(http.StatusServiceUnavailable) }", "func NewGetInfraEnvPresignedFileURLServiceUnavailable() *GetInfraEnvPresignedFileURLServiceUnavailable {\n\treturn &GetInfraEnvPresignedFileURLServiceUnavailable{}\n}", "func NewEndpointSlice(l log.Logger, eps cache.SharedIndexInformer, svc, pod, node cache.SharedInformer) *EndpointSlice {\n\tif l == nil {\n\t\tl = log.NewNopLogger()\n\t}\n\te := &EndpointSlice{\n\t\tlogger: l,\n\t\tendpointSliceInf: eps,\n\t\tendpointSliceStore: eps.GetStore(),\n\t\tserviceInf: svc,\n\t\tserviceStore: svc.GetStore(),\n\t\tpodInf: pod,\n\t\tpodStore: pod.GetStore(),\n\t\tnodeInf: node,\n\t\twithNodeMetadata: node != nil,\n\t\tqueue: workqueue.NewNamed(\"endpointSlice\"),\n\t}\n\n\t_, err := e.endpointSliceInf.AddEventHandler(cache.ResourceEventHandlerFuncs{\n\t\tAddFunc: func(o interface{}) {\n\t\t\tepslAddCount.Inc()\n\t\t\te.enqueue(o)\n\t\t},\n\t\tUpdateFunc: func(_, o interface{}) {\n\t\t\tepslUpdateCount.Inc()\n\t\t\te.enqueue(o)\n\t\t},\n\t\tDeleteFunc: func(o interface{}) {\n\t\t\tepslDeleteCount.Inc()\n\t\t\te.enqueue(o)\n\t\t},\n\t})\n\tif err != nil {\n\t\tlevel.Error(l).Log(\"msg\", \"Error adding endpoint slices event handler.\", \"err\", err)\n\t}\n\n\tserviceUpdate := func(o interface{}) {\n\t\tsvc, err := convertToService(o)\n\t\tif err != nil {\n\t\t\tlevel.Error(e.logger).Log(\"msg\", \"converting to Service object failed\", \"err\", err)\n\t\t\treturn\n\t\t}\n\n\t\t// TODO(brancz): use cache.Indexer to index endpoints by\n\t\t// disv1beta1.LabelServiceName so this operation doesn't have to\n\t\t// iterate over all endpoint objects.\n\t\tfor _, obj := range e.endpointSliceStore.List() {\n\t\t\tesa, err := e.getEndpointSliceAdaptor(obj)\n\t\t\tif err != nil {\n\t\t\t\tlevel.Error(e.logger).Log(\"msg\", \"converting to EndpointSlice object failed\", \"err\", err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif lv, exists := esa.labels()[esa.labelServiceName()]; exists && lv == svc.Name {\n\t\t\t\te.enqueue(esa.get())\n\t\t\t}\n\t\t}\n\t}\n\t_, err = e.serviceInf.AddEventHandler(cache.ResourceEventHandlerFuncs{\n\t\tAddFunc: func(o interface{}) {\n\t\t\tsvcAddCount.Inc()\n\t\t\tserviceUpdate(o)\n\t\t},\n\t\tUpdateFunc: func(_, o interface{}) {\n\t\t\tsvcUpdateCount.Inc()\n\t\t\tserviceUpdate(o)\n\t\t},\n\t\tDeleteFunc: func(o interface{}) {\n\t\t\tsvcDeleteCount.Inc()\n\t\t\tserviceUpdate(o)\n\t\t},\n\t})\n\tif err != nil {\n\t\tlevel.Error(l).Log(\"msg\", \"Error adding services event handler.\", \"err\", err)\n\t}\n\n\tif e.withNodeMetadata {\n\t\t_, err = e.nodeInf.AddEventHandler(cache.ResourceEventHandlerFuncs{\n\t\t\tAddFunc: func(o interface{}) {\n\t\t\t\tnode := o.(*apiv1.Node)\n\t\t\t\te.enqueueNode(node.Name)\n\t\t\t},\n\t\t\tUpdateFunc: func(_, o interface{}) {\n\t\t\t\tnode := o.(*apiv1.Node)\n\t\t\t\te.enqueueNode(node.Name)\n\t\t\t},\n\t\t\tDeleteFunc: func(o interface{}) {\n\t\t\t\tnode := o.(*apiv1.Node)\n\t\t\t\te.enqueueNode(node.Name)\n\t\t\t},\n\t\t})\n\t\tif err != nil {\n\t\t\tlevel.Error(l).Log(\"msg\", \"Error adding nodes event handler.\", \"err\", err)\n\t\t}\n\t}\n\n\treturn e\n}", "func NewEndpointUpdateNotFound() *EndpointUpdateNotFound {\n\treturn &EndpointUpdateNotFound{}\n}", "func NewPostGocardlessServiceUnavailable() *PostGocardlessServiceUnavailable {\n\treturn &PostGocardlessServiceUnavailable{}\n}", "func (client *WebAppsClient) deletePrivateEndpointConnectionSlotCreateRequest(ctx context.Context, resourceGroupName string, name string, privateEndpointConnectionName string, slot string, options *WebAppsBeginDeletePrivateEndpointConnectionSlotOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/slots/{slot}/privateEndpointConnections/{privateEndpointConnectionName}\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif privateEndpointConnectionName == \"\" {\n\t\treturn nil, errors.New(\"parameter privateEndpointConnectionName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{privateEndpointConnectionName}\", url.PathEscape(privateEndpointConnectionName))\n\tif slot == \"\" {\n\t\treturn nil, errors.New(\"parameter slot cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{slot}\", url.PathEscape(slot))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsClientDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func EndpointCreateMessage(e notifications.RegenNotificationInfo) AgentNotifyMessage {\n\tnotification := EndpointNotification{\n\t\tEndpointRegenNotification: EndpointRegenNotification{\n\t\t\tID: e.GetID(),\n\t\t\tLabels: e.GetOpLabels(),\n\t\t},\n\t\tPodName: e.GetK8sPodName(),\n\t\tNamespace: e.GetK8sNamespace(),\n\t}\n\n\treturn AgentNotifyMessage{\n\t\tType: AgentNotifyEndpointCreated,\n\t\tNotification: notification,\n\t}\n}", "func (a *DefaultApiService) DeleteEndpointExecute(r ApiDeleteEndpointRequest) (*_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodDelete\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"DefaultApiService.DeleteEndpoint\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/endpoints/{id}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"id\"+\"}\", _neturl.PathEscape(parameterToString(r.id, \"\")), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\tif localVarHTTPResponse.StatusCode == 404 {\n\t\t\tvar v ErrorNotFound\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarHTTPResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t}\n\t\treturn localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarHTTPResponse, executionError\n}", "func NewEndpointDeleteBadRequest() *EndpointDeleteBadRequest {\n\treturn &EndpointDeleteBadRequest{}\n}", "func (client *DataCollectionEndpointsClient) deleteCreateRequest(ctx context.Context, resourceGroupName string, dataCollectionEndpointName string, options *DataCollectionEndpointsDeleteOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Insights/dataCollectionEndpoints/{dataCollectionEndpointName}\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif dataCollectionEndpointName == \"\" {\n\t\treturn nil, errors.New(\"parameter dataCollectionEndpointName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{dataCollectionEndpointName}\", url.PathEscape(dataCollectionEndpointName))\n\treq, err := runtime.NewRequest(ctx, http.MethodDelete, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-04-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func newRESTEndpointService(hostPortStr string) endpointService {\n\treturn endpointService(\n\t\tnewRESTDiscoveryService(fmt.Sprintf(edsRestEndpointTemplate, hostPortStr)),\n\t)\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(ctx context.Context, project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(ctx, path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn checkAPIResponse(bts, nil)\n}", "func NewEndpoint(dnsName, recordType string, targets ...string) *Endpoint {\n\treturn NewEndpointWithTTL(dnsName, recordType, TTL(0), targets...)\n}", "func NotFound(w http.ResponseWriter, r *http.Request) {\n\tw.WriteHeader(http.StatusNotFound)\n\tjson.NewEncoder(w).Encode(&ServiceError{\n\t\tMessage: \"Endpoint not found\",\n\t\tSolution: \"See / for possible directives\",\n\t\tErrorCode: http.StatusNotFound,\n\t})\n}", "func NewDeleteConnectorUsingDELETEServiceUnavailable() *DeleteConnectorUsingDELETEServiceUnavailable {\n\treturn &DeleteConnectorUsingDELETEServiceUnavailable{}\n}", "func TestServiceEndpointAzureRM_Create_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := ResourceServiceEndpointAzureRM()\n\tfor _, resource := range azurermTestServiceEndpointsAzureRM {\n\t\tresourceData := getResourceData(t, resource)\n\t\tflattenServiceEndpointAzureRM(resourceData, &resource, azurermTestServiceEndpointAzureRMProjectID)\n\n\t\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\t\tclients := &client.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\t\texpectedArgs := serviceendpoint.CreateServiceEndpointArgs{Endpoint: &resource}\n\n\t\tbuildClient.\n\t\t\tEXPECT().\n\t\t\tCreateServiceEndpoint(clients.Ctx, expectedArgs).\n\t\t\tReturn(nil, errors.New(\"CreateServiceEndpoint() Failed\")).\n\t\t\tTimes(1)\n\n\t\terr := r.Create(resourceData, clients)\n\t\trequire.Contains(t, err.Error(), \"CreateServiceEndpoint() Failed\")\n\n\t}\n}", "func ExampleRDS_CreateDBClusterEndpoint_shared00() {\n\tsvc := rds.New(session.New())\n\tinput := &rds.CreateDBClusterEndpointInput{\n\t\tDBClusterEndpointIdentifier: aws.String(\"mycustomendpoint\"),\n\t\tDBClusterIdentifier: aws.String(\"mydbcluster\"),\n\t\tEndpointType: aws.String(\"reader\"),\n\t\tStaticMembers: []*string{\n\t\t\taws.String(\"dbinstance1\"),\n\t\t\taws.String(\"dbinstance2\"),\n\t\t},\n\t}\n\n\tresult, err := svc.CreateDBClusterEndpoint(input)\n\tif err != nil {\n\t\tif aerr, ok := err.(awserr.Error); ok {\n\t\t\tswitch aerr.Code() {\n\t\t\tcase rds.ErrCodeDBClusterEndpointQuotaExceededFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBClusterEndpointQuotaExceededFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBClusterEndpointAlreadyExistsFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBClusterEndpointAlreadyExistsFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBClusterNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBClusterNotFoundFault, aerr.Error())\n\t\t\tcase rds.ErrCodeInvalidDBClusterStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBClusterStateFault, aerr.Error())\n\t\t\tcase rds.ErrCodeDBInstanceNotFoundFault:\n\t\t\t\tfmt.Println(rds.ErrCodeDBInstanceNotFoundFault, aerr.Error())\n\t\t\tcase rds.ErrCodeInvalidDBInstanceStateFault:\n\t\t\t\tfmt.Println(rds.ErrCodeInvalidDBInstanceStateFault, aerr.Error())\n\t\t\tdefault:\n\t\t\t\tfmt.Println(aerr.Error())\n\t\t\t}\n\t\t} else {\n\t\t\t// Print the error, cast err to awserr.Error to get the Code and\n\t\t\t// Message from an error.\n\t\t\tfmt.Println(err.Error())\n\t\t}\n\t\treturn\n\t}\n\n\tfmt.Println(result)\n}", "func (h *ServiceIntegrationEndpointsHandler) Delete(project, endpointID string) error {\n\tpath := buildPath(\"project\", project, \"integration_endpoint\", endpointID)\n\tbts, err := h.client.doDeleteRequest(path, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn handleDeleteResponse(bts)\n}", "func NewEndpointCreated() filters.Spec {\n\tvar ec endpointCreated\n\treturn ec\n}", "func (it *integTestSuite) TestNpmEndpointCreateDelete(c *C) {\n\t// create a network in controller\n\t// if not present create the default tenant\n\tit.CreateTenant(\"default\")\n\terr := it.DeleteAllEndpoints(c)\n\tc.Assert(err, IsNil)\n\terr = it.CreateNetwork(\"default\", \"default\", \"testNetwork\", \"10.1.0.0/22\", \"10.1.1.254\")\n\tc.Assert(err, IsNil)\n\tAssertEventually(c, func() (bool, interface{}) {\n\t\t_, nerr := it.npmCtrler.StateMgr.FindNetwork(\"default\", \"testNetwork\")\n\t\treturn (nerr == nil), nil\n\t}, \"Network not found in statemgr\")\n\n\t// wait till agent has the network\n\tfor _, ag := range it.agents {\n\t\tAssertEventually(c, func() (bool, interface{}) {\n\n\t\t\tnt := netproto.Network{\n\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Network\"},\n\t\t\t\tObjectMeta: api.ObjectMeta{Tenant: \"default\", Namespace: \"default\", Name: \"testNetwork\"},\n\t\t\t}\n\t\t\t_, nerr := ag.dscAgent.PipelineAPI.HandleNetwork(agentTypes.Get, nt)\n\t\t\treturn (nerr == nil), nil\n\t\t}, \"Network not found in agent\")\n\t}\n\n\t// create a wait channel\n\twaitCh := make(chan error, it.numAgents*2)\n\n\t// create one endpoint from each agent\n\tfor i, ag := range it.agents {\n\t\tgo func(i int, ag *Dpagent) {\n\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\t\t\thostName := fmt.Sprintf(\"testHost-%d\", i)\n\n\t\t\t// make the call\n\t\t\tcerr := it.CreateEndpoint(\"default\", \"default\", \"testNetwork\", epname, hostName, \"0101.0101.0101\", hostName, \"20.1.1.1\", map[string]string{\"env\": \"production\", \"app\": \"procurement\"}, 2)\n\t\t\tif cerr != nil {\n\t\t\t\twaitCh <- fmt.Errorf(\"endpoint create failed: %v\", cerr)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(i, ag)\n\t}\n\n\t// wait for all endpoint creates to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Error during endpoint create\")\n\t}\n\n\t// wait for all endpoints to be propagated to other agents\n\tfor _, ag := range it.agents {\n\t\tgo func(ag *Dpagent) {\n\t\t\tfound := CheckEventually(func() (bool, interface{}) {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\treturn len(endpoints) == it.numAgents, nil\n\t\t\t}, \"10ms\", it.pollTimeout())\n\t\t\tif !found {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\tlog.Infof(\"Endpoint count expected [%v] found [%v]\", it.numAgents, len(endpoints))\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint count incorrect in datapath\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfor i := range it.agents {\n\t\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\t\t\t\tepmeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t\tObjectMeta: api.ObjectMeta{\n\t\t\t\t\t\tTenant: \"default\",\n\t\t\t\t\t\tNamespace: \"default\",\n\t\t\t\t\t\tName: epname,\n\t\t\t\t\t},\n\t\t\t\t}\n\t\t\t\t_, perr := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.Get, epmeta)\n\t\t\t\tif perr != nil {\n\t\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint not found in datapath\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\twaitCh <- nil\n\t\t}(ag)\n\t}\n\n\t// wait for all goroutines to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint info incorrect in datapath\")\n\t}\n\n\t// now delete the endpoints\n\tfor i, ag := range it.agents {\n\t\tgo func(i int, ag *Dpagent) {\n\t\t\tepname := fmt.Sprintf(\"testEndpoint-%d\", i)\n\n\t\t\t// make the call\n\t\t\tcerr := it.DeleteEndpoint(\"default\", \"default\", epname)\n\t\t\tif cerr != nil {\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint delete failed: %v\", cerr)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(i, ag)\n\t}\n\n\t// wait for all endpoint deletes to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint delete failed\")\n\n\t}\n\n\tfor _, ag := range it.agents {\n\t\tgo func(ag *Dpagent) {\n\t\t\tif !CheckEventually(func() (bool, interface{}) {\n\t\t\t\tepMeta := netproto.Endpoint{\n\t\t\t\t\tTypeMeta: api.TypeMeta{Kind: \"Endpoint\"},\n\t\t\t\t}\n\t\t\t\tendpoints, _ := ag.dscAgent.PipelineAPI.HandleEndpoint(agentTypes.List, epMeta)\n\t\t\t\treturn len(endpoints) == 0, nil\n\t\t\t}, \"10ms\", it.pollTimeout()) {\n\t\t\t\twaitCh <- fmt.Errorf(\"Endpoint was not deleted from datapath\")\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\twaitCh <- nil\n\t\t}(ag)\n\t}\n\n\t// wait for all goroutines to complete\n\tfor i := 0; i < it.numAgents; i++ {\n\t\tAssertOk(c, <-waitCh, \"Endpoint delete error\")\n\t}\n\n\t// delete the network\n\terr = it.DeleteNetwork(\"default\", \"testNetwork\")\n\tc.Assert(err, IsNil)\n\tAssertEventually(c, func() (bool, interface{}) {\n\t\t_, nerr := it.npmCtrler.StateMgr.FindNetwork(\"default\", \"testNetwork\")\n\t\treturn (nerr != nil), nil\n\t}, \"Network still found in statemgr\")\n}", "func ErrServiceUnavailablef(format string, arguments ...interface{}) *Status {\n\treturn &Status{Code: http.StatusServiceUnavailable, Text: fmt.Sprintf(format, arguments...)}\n}", "func (r *Response) ServiceUnavailable(v interface{}) {\n\tr.writeResponse(http.StatusServiceUnavailable, v)\n}", "func NewFilePatchServiceUnavailable() *FilePatchServiceUnavailable {\n\treturn &FilePatchServiceUnavailable{}\n}", "func NewDeleteBacsIDServiceUnavailable() *DeleteBacsIDServiceUnavailable {\n\treturn &DeleteBacsIDServiceUnavailable{}\n}", "func ExchangeDelete(service string, urlBase string, urlSuffix string, credentials string, goodHttpCodes []int) (httpCode int) {\n\turl := urlBase + \"/\" + urlSuffix\n\tapiMsg := http.MethodDelete + \" \" + url\n\n\t// get message printer\n\tmsgPrinter := i18n.GetMessagePrinter()\n\n\tVerbose(apiMsg)\n\tif IsDryRun() {\n\t\treturn 204\n\t}\n\n\thttpClient := GetHTTPClient(config.HTTPRequestTimeoutS)\n\n\tresp := InvokeRestApi(httpClient, http.MethodDelete, url, credentials, nil, service, apiMsg, make(map[string]string), true)\n\tif resp.Body != nil {\n\t\tdefer resp.Body.Close()\n\t}\n\n\t// delete never returns a body\n\thttpCode = resp.StatusCode\n\tVerbose(msgPrinter.Sprintf(\"HTTP code: %d\", httpCode))\n\tif !isGoodCode(httpCode, goodHttpCodes) {\n\t\tFatal(HTTP_ERROR, msgPrinter.Sprintf(\"bad HTTP code %d from %s\", httpCode, apiMsg))\n\t}\n\treturn\n}", "func NewDeactivateEndpoint(s Service) goa.Endpoint {\n\treturn func(ctx context.Context, req interface{}) (interface{}, error) {\n\t\tp := req.(*DeactivatePayload)\n\t\treturn nil, s.Deactivate(ctx, p)\n\t}\n}", "func (nb *NetBuilder) FindOrCreateEndpoint(nw *Network, ep *Endpoint) error {\n\t// Query the namespace identifier.\n\tnsType, namespaceIdentifier := nb.getNamespaceIdentifier(ep)\n\n\t// Check if the endpoint already exists.\n\tendpointName := nb.generateHNSEndpointName(nw.Name, namespaceIdentifier)\n\thnsEndpoint, err := hcsshim.GetHNSEndpointByName(endpointName)\n\tif err == nil {\n\t\tlog.Infof(\"Found existing HNS endpoint %s.\", endpointName)\n\t\tif nsType == infraContainerNS || nsType == hcsNamespace {\n\t\t\t// This is a benign duplicate create call for an existing endpoint.\n\t\t\t// The endpoint was already attached in a previous call. Ignore and return success.\n\t\t\tlog.Infof(\"HNS endpoint %s is already attached to container ID %s.\",\n\t\t\t\tendpointName, ep.ContainerID)\n\t\t} else {\n\t\t\t// Attach the existing endpoint to the container's network namespace.\n\t\t\t// Attachment of endpoint to each container would occur only when using HNS V1 APIs.\n\t\t\terr = nb.attachEndpointV1(hnsEndpoint, ep.ContainerID)\n\t\t}\n\n\t\tep.MACAddress, ep.IPAddresses, nw.GatewayIPAddresses =\n\t\t\tnb.parseEndpointFieldsFromResponse(hnsEndpoint)\n\t\treturn err\n\t} else {\n\t\tif nsType != infraContainerNS && nsType != hcsNamespace {\n\t\t\t// The endpoint referenced in the container netns does not exist.\n\t\t\tlog.Errorf(\"Failed to find endpoint %s for container %s.\", endpointName, ep.ContainerID)\n\t\t\treturn fmt.Errorf(\"failed to find endpoint %s: %v\", endpointName, err)\n\t\t}\n\t}\n\n\t// Initialize the HNS endpoint.\n\thnsEndpoint = &hcsshim.HNSEndpoint{\n\t\tName: endpointName,\n\t\tVirtualNetworkName: nw.Name,\n\t\tDNSSuffix: strings.Join(nw.DNSSuffixSearchList, \",\"),\n\t\tDNSServerList: strings.Join(nw.DNSServers, \",\"),\n\t}\n\n\tif ep.MACAddress != nil {\n\t\thnsEndpoint.MacAddress = ep.MACAddress.String()\n\t}\n\tif len(ep.IPAddresses) != 0 {\n\t\thnsEndpoint.IPAddress = ep.IPAddresses[0].IP\n\t\tpl, _ := ep.IPAddresses[0].Mask.Size()\n\t\thnsEndpoint.PrefixLength = uint8(pl)\n\t}\n\n\t// Add ACL policies for blocking IMDS access through the endpoint.\n\tif ep.BlockIMDS {\n\t\terr = imds.BlockInstanceMetadataEndpoint(hnsEndpoint)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to block instance metadata endpoint: %v.\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Create the HNS endpoint.\n\tlog.Infof(\"Creating HNS endpoint: %+v\", hnsEndpoint)\n\thnsResponse, err := hnsEndpoint.Create()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to create HNS endpoint: %v.\", err)\n\t\treturn err\n\t}\n\n\tlog.Infof(\"Received HNS endpoint response: %+v.\", hnsResponse)\n\n\t// Attach the HNS endpoint to the container's network namespace.\n\tif nsType == infraContainerNS {\n\t\terr = nb.attachEndpointV1(hnsResponse, ep.ContainerID)\n\t}\n\tif nsType == hcsNamespace {\n\t\terr = nb.attachEndpointV2(hnsResponse, namespaceIdentifier)\n\t}\n\tif err != nil {\n\t\t// Cleanup the failed endpoint.\n\t\tlog.Infof(\"Deleting the failed HNS endpoint %s.\", hnsResponse.Id)\n\t\t_, delErr := hnsResponse.Delete()\n\t\tif delErr != nil {\n\t\t\tlog.Errorf(\"Failed to delete HNS endpoint: %v.\", delErr)\n\t\t}\n\n\t\treturn err\n\t}\n\n\t// Return network interface MAC address, IP Address and Gateway.\n\tep.MACAddress, ep.IPAddresses, nw.GatewayIPAddresses =\n\t\tnb.parseEndpointFieldsFromResponse(hnsResponse)\n\treturn nil\n}", "func (e ServiceUnavailable) Code() int { return http.StatusServiceUnavailable }", "func NewEndpoint(conn *websocket.Conn, registry *Registry) *Endpoint {\n\tif registry == nil {\n\t\tregistry = dummyRegistry\n\t}\n\te := &Endpoint{}\n\te.conn = conn\n\te.server.registry = registry\n\te.client.pending = make(map[uint64]*rpc.Call)\n\treturn e\n}", "func (nb *NetBuilder) DeleteEndpoint(nw *Network, ep *Endpoint) error {\n\t// Generate network name here as endpoint name is dependent upon network name.\n\tnw.Name = nb.generateHNSNetworkName(nw)\n\t// Query the namespace identifier.\n\tnsType, namespaceIdentifier := nb.getNamespaceIdentifier(ep)\n\n\t// Find the HNS endpoint ID.\n\tendpointName := nb.generateHNSEndpointName(nw.Name, namespaceIdentifier)\n\thnsEndpoint, err := hcsshim.GetHNSEndpointByName(endpointName)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Detach the HNS endpoint from the container's network namespace.\n\tlog.Infof(\"Detaching HNS endpoint %s from container %s netns.\", hnsEndpoint.Id, ep.ContainerID)\n\tif nsType == hcsNamespace {\n\t\t// Detach the HNS endpoint from the namespace, if we can.\n\t\t// HCN Namespace and HNS Endpoint have a 1-1 relationship, therefore,\n\t\t// even if detachment of endpoint from namespace fails, we can still proceed to delete it.\n\t\terr = hcn.RemoveNamespaceEndpoint(namespaceIdentifier, hnsEndpoint.Id)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Failed to detach endpoint, ignoring: %v\", err)\n\t\t}\n\t} else {\n\t\terr = hcsshim.HotDetachEndpoint(ep.ContainerID, hnsEndpoint.Id)\n\t\tif err != nil && err != hcsshim.ErrComputeSystemDoesNotExist {\n\t\t\treturn err\n\t\t}\n\n\t\t// The rest of the delete logic applies to infrastructure container only.\n\t\tif nsType == nonInfraContainerNS {\n\t\t\t// For non-infra containers, the network must not be deleted.\n\t\t\tnw.UseExisting = true\n\t\t\treturn nil\n\t\t}\n\t}\n\n\t// Delete the HNS endpoint.\n\tlog.Infof(\"Deleting HNS endpoint name: %s ID: %s\", endpointName, hnsEndpoint.Id)\n\t_, err = hnsEndpoint.Delete()\n\tif err != nil {\n\t\tlog.Errorf(\"Failed to delete HNS endpoint: %v.\", err)\n\t}\n\n\treturn err\n}", "func createAPIEndpointService(namespace, serviceName string, api *API) error {\n\n\tclientset, err := getClientSet()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar stype v1.ServiceType\n\n\tswitch api.ServiceType {\n\n\tcase \"clusterIP\":\n\t\tstype = v1.ServiceTypeClusterIP\n\n\tcase \"nodePort\":\n\t\tstype = v1.ServiceTypeNodePort\n\n\tcase \"loadBalancer\":\n\t\tstype = v1.ServiceTypeLoadBalancer\n\n\tdefault:\n\t\tstype = v1.ServiceTypeClusterIP\n\n\t}\n\n\t_, err = clientset.CoreV1().Services(namespace).Create(&v1.Service{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: serviceName,\n\t\t\tLabels: map[string]string{\n\t\t\t\t\"kubez-type\": \"api-endpoint\",\n\t\t\t\t\"kubez-name\": api.Name,\n\t\t\t},\n\t\t},\n\t\tSpec: v1.ServiceSpec{\n\t\t\tSelector: map[string]string{\n\t\t\t\t\"kubez-type\": \"api-endpoint\",\n\t\t\t\t\"kubez-name\": api.Name,\n\t\t\t},\n\t\t\tType: stype,\n\t\t\tPorts: []v1.ServicePort{\n\t\t\t\t{\n\t\t\t\t\tName: \"kubez-api\",\n\t\t\t\t\tPort: api.Port,\n\t\t\t\t\tTargetPort: intstr.IntOrString{IntVal: api.Port},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *APIClient) Delete(endpoint Endpointer, result interface{}, headers ...func(h *http.Header)) error {\n\tepoint, err := endpoint.GetEndpoint()\n\tif err != nil {\n\t\treturn err\n\t}\n\tendUrl := c.CompileEndpointURL(epoint)\n\treq, _ := http.NewRequest(http.MethodDelete, endUrl, nil)\n\n\th := &req.Header\n\tfor _, v := range headers {\n\t\tv(h)\n\t}\n\n\tresp, err := c.Client.Call(req)\n\tc.addResponseCode(resp.StatusCode)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer resp.Body.Close()\n\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn MakeResponseError(resp)\n\t}\n\n\terr = json.NewDecoder(resp.Body).Decode(result)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func NewEndpoint() *Endpoint {\r\n\t// Create a new Endpoint with an empty list of handler funcs.\r\n\treturn &Endpoint{\r\n\t\thandler: map[string]HandleFunc{},\r\n\t}\r\n}", "func ServiceUnavailablef(format string, args ...interface{}) ServiceUnavailable {\n\treturn ServiceUnavailable(fmt.Sprintf(format, args...))\n}", "func (r Response) ServiceUnavailable(code string, payload Payload, header ...ResponseHeader) {\n\tr.Response(code, http.ServiceUnavailable, payload, header...)\n}", "func NewPostDeviceURLServiceUnavailable() *PostDeviceURLServiceUnavailable {\n\treturn &PostDeviceURLServiceUnavailable{}\n}", "func (ec *ExtensionClient) DeleteEndpoint(extensionID, serviceID, URL string) error {\n\n\turl := url.QueryEscape(URL)\n\trequest, err := extensionc.BuildDeleteEndpointPayload(extensionID, serviceID, url)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = ec.c.DeleteEndpoint()(context.Background(), request)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func New(endpoint string) BaseClient {\n\treturn NewWithoutDefaults(endpoint)\n}", "func TestAzureDevOpsServiceEndpointDockerRegistry_Create_DoesNotSwallowError(t *testing.T) {\n\tctrl := gomock.NewController(t)\n\tdefer ctrl.Finish()\n\n\tr := resourceServiceEndpointDockerRegistry()\n\tresourceData := schema.TestResourceDataRaw(t, r.Schema, nil)\n\tflattenServiceEndpointDockerRegistry(resourceData, &dockerRegistryTestServiceEndpoint, dockerRegistryTestServiceEndpointProjectID)\n\n\tbuildClient := azdosdkmocks.NewMockServiceendpointClient(ctrl)\n\tclients := &config.AggregatedClient{ServiceEndpointClient: buildClient, Ctx: context.Background()}\n\n\texpectedArgs := serviceendpoint.CreateServiceEndpointArgs{Endpoint: &dockerRegistryTestServiceEndpoint, Project: dockerRegistryTestServiceEndpointProjectID}\n\tbuildClient.\n\t\tEXPECT().\n\t\tCreateServiceEndpoint(clients.Ctx, expectedArgs).\n\t\tReturn(nil, errors.New(\"CreateServiceEndpoint() Failed\")).\n\t\tTimes(1)\n\n\terr := r.Create(resourceData, clients)\n\trequire.Contains(t, err.Error(), \"CreateServiceEndpoint() Failed\")\n}", "func EndpointFactory(args *endpoint.Arg, stats *stats.Stats, workerCount uint) (endpoint.EndPoint, error) {\n\tif FailSetup {\n\t\treturn nil, errors.New(\"Forced Error\")\n\t}\n\treturn &fakeEndpoint{}, nil\n}", "func MakeDeleteEndpoint(s service.TodoService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\treq := request.(DeleteRequest)\n\t\terror := s.Delete(ctx, req.Id)\n\t\treturn DeleteResponse{Error: error}, nil\n\t}\n}", "func NewBindHostServiceUnavailable() *BindHostServiceUnavailable {\n\treturn &BindHostServiceUnavailable{}\n}", "func (manager *Manager) OnDeleteEndpoint(endpoint *k8sTypes.CiliumEndpoint) {\n\tid := types.NamespacedName{\n\t\tName: endpoint.GetName(),\n\t\tNamespace: endpoint.GetNamespace(),\n\t}\n\n\tmanager.pendingEndpointEventsLock.Lock()\n\tdelete(manager.pendingEndpointEvents, id)\n\tmanager.pendingEndpointEventsLock.Unlock()\n\n\tmanager.endpointEventsQueue.Add(id)\n}", "func CreateEndpoint(ctx iris.Context) {\n\t// Add logic to check if given ports exits\n\turi := ctx.Request().RequestURI\n\tfabricID := ctx.Params().Get(\"id\")\n\tfabricData, ok := capdata.FabricDataStore.Data[fabricID]\n\tif !ok {\n\t\terrMsg := fmt.Sprintf(\"Fabric data for uri %s not found\", uri)\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"Fabric\", fabricID})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\n\tvar endpoint model.Endpoint\n\terr := ctx.ReadJSON(&endpoint)\n\tif err != nil {\n\t\terrorMessage := \"error while trying to get JSON body from the request: \" + err.Error()\n\t\tlog.Error(errorMessage)\n\t\tresp := updateErrorResponse(response.MalformedJSON, errorMessage, nil)\n\t\tctx.StatusCode(http.StatusBadRequest)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif len(endpoint.Redundancy) < 1 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be created, Redudancy in the request is missing: \" + err.Error())\n\t\tresp := updateErrorResponse(response.PropertyMissing, errMsg, []interface{}{\"Redundancy\"})\n\t\tctx.StatusCode(http.StatusBadRequest)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tif len(endpoint.Redundancy[0].RedundancySet) == 0 {\n\t\terrMsg := fmt.Sprintf(\"Endpoint cannot be created, RedudancySet in the request is missing: \" + err.Error())\n\t\tresp := updateErrorResponse(response.PropertyMissing, errMsg, []interface{}{\"RedudancySet\"})\n\t\tctx.StatusCode(http.StatusBadRequest)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\t// get all existing endpoints under fabric check for the name\n\tfor _, endpointData := range capdata.EndpointDataStore {\n\t\tif endpoint.Name == endpointData.Endpoint.Name {\n\t\t\terrMsg := \"Endpoint name is already assigned to other endpoint:\" + endpointData.Endpoint.Name\n\t\t\tresp := updateErrorResponse(response.ResourceAlreadyExists, errMsg, []interface{}{\"Endpoint\", endpointData.Endpoint.Name, endpoint.Name})\n\t\t\tctx.StatusCode(http.StatusConflict)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\t\t}\n\t}\n\tvar switchURI = \"\"\n\tvar portPattern = \"\"\n\tportList := make(map[string]bool)\n\t// check if given ports are present in plugin database\n\tfor i := 0; i < len(endpoint.Redundancy[0].RedundancySet); i++ {\n\t\tportURI := endpoint.Redundancy[0].RedundancySet[i].Oid\n\t\tif _, ok := portList[endpoint.Redundancy[0].RedundancySet[i].Oid]; ok {\n\t\t\terrMsg := \"Duplicate port passed in the request\"\n\t\t\tresp := updateErrorResponse(response.PropertyValueConflict, errMsg, []interface{}{endpoint.Redundancy[0].RedundancySet[i].Oid, endpoint.Redundancy[0].RedundancySet[i].Oid})\n\t\t\tctx.StatusCode(http.StatusBadRequest)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\n\t\t}\n\t\tportList[endpoint.Redundancy[0].RedundancySet[i].Oid] = true\n\n\t\t_, statusCode, resp := getPortData(portURI)\n\t\tif statusCode != http.StatusOK {\n\t\t\tctx.StatusCode(statusCode)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\t\t}\n\t\tstatusCode, resp = checkEndpointPortMapping(endpoint.Redundancy[0].RedundancySet[i].Oid)\n\t\tif statusCode != http.StatusOK {\n\t\t\tctx.StatusCode(statusCode)\n\t\t\tctx.JSON(resp)\n\t\t\treturn\n\t\t}\n\t\tportURIData := strings.Split(portURI, \"/\")\n\t\tswitchID := portURIData[6]\n\t\tswitchIDData := strings.Split(switchID, \":\")\n\t\tswitchURI = switchURI + \"-\" + switchIDData[1]\n\t\tportIDData := strings.Split(portURIData[8], \":\")\n\t\ttmpPortPattern := strings.Replace(portIDData[1], \"eth\", \"\", -1)\n\t\ttmpPortPattern = strings.Replace(tmpPortPattern, \"-\", \"-ports-\", -1)\n\t\tportPattern = tmpPortPattern\n\t}\n\n\tportPolicyGroupList, err := caputilities.GetPortPolicyGroup(fabricData.PodID, switchURI)\n\tif err != nil || len(portPolicyGroupList) == 0 {\n\t\terrMsg := \"Port policy group not found for given ports\"\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{\"protpaths\" + switchURI, \"PolicyGroup\"})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\n\t}\n\tpolicyGroupDN := \"\"\n\tfor i := 0; i < len(portPolicyGroupList); i++ {\n\t\tif strings.Contains(portPolicyGroupList[i].BaseAttributes.DistinguishedName, portPattern) {\n\t\t\tpolicyGroupDN = portPolicyGroupList[i].BaseAttributes.DistinguishedName\n\t\t}\n\t}\n\tif policyGroupDN == \"\" {\n\t\terrMsg := \"Port policy group not found for given ports\"\n\t\tlog.Error(errMsg)\n\t\tresp := updateErrorResponse(response.ResourceNotFound, errMsg, []interface{}{portPattern, \"PolicyGroup\"})\n\t\tctx.StatusCode(http.StatusNotFound)\n\t\tctx.JSON(resp)\n\t\treturn\n\t}\n\tlog.Info(\"Dn of Policy group:\" + policyGroupDN)\n\tsaveEndpointData(uri, fabricID, policyGroupDN, &endpoint)\n\tcommon.SetResponseHeader(ctx, map[string]string{\n\t\t\"Location\": endpoint.ODataID,\n\t})\n\tctx.StatusCode(http.StatusCreated)\n\tctx.JSON(endpoint)\n}", "func (r *templateRouter) DeleteEndpoints(id string) {\n\tservice, ok := r.FindServiceUnit(id)\n\tif !ok {\n\t\treturn\n\t}\n\tservice.EndpointTable = make(map[string]Endpoint)\n\n\tr.state[id] = service\n}", "func makeDeleteBookEndpoint(svc BookService) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\t\t// convert request into a bookRequest\n\t\treq := request.(deleteBookRequest)\n\n\t\t// call actual service with data from the req\n\t\terr := svc.DeleteBook(req.BookId)\n\t\treturn deleteBookResponse{\n\t\t\tErr: err,\n\t\t}, nil\n\t}\n}", "func (o *ServiceAddServiceUnavailable) WithPayload(payload *models.Error) *ServiceAddServiceUnavailable {\n\to.Payload = payload\n\treturn o\n}", "func DeleteHostNCApipaEndpoint(\n\tnetworkContainerID string) error {\n\treturn nil\n}", "func NewDeactivateNotFound(body *DeactivateNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func NewDeactivateNotFound(body *DeactivateNotFoundResponseBody) *goa.ServiceError {\n\tv := &goa.ServiceError{\n\t\tName: *body.Name,\n\t\tID: *body.ID,\n\t\tMessage: *body.Message,\n\t\tTemporary: *body.Temporary,\n\t\tTimeout: *body.Timeout,\n\t\tFault: *body.Fault,\n\t}\n\n\treturn v\n}", "func ExampleWebAppsClient_BeginDeletePrivateEndpointConnection() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclient, err := armappservice.NewWebAppsClient(\"34adfa4f-cedf-4dc0-ba29-b6d1a69ab345\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tpoller, err := client.BeginDeletePrivateEndpointConnection(ctx,\n\t\t\"rg\",\n\t\t\"testSite\",\n\t\t\"connection\",\n\t\tnil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\tres, err := poller.PollUntilDone(ctx, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to pull the result: %v\", err)\n\t}\n\t// TODO: use response item\n\t_ = res\n}" ]
[ "0.65823364", "0.6272349", "0.6205341", "0.6025818", "0.5972945", "0.595924", "0.5870526", "0.5839624", "0.57643414", "0.56722456", "0.56486386", "0.56240743", "0.5579028", "0.5546808", "0.54401135", "0.5434158", "0.5398272", "0.5396613", "0.5380882", "0.5380018", "0.5377111", "0.53683066", "0.5339411", "0.53347754", "0.5327948", "0.53225404", "0.53225404", "0.5286684", "0.5267385", "0.52507544", "0.5246859", "0.5246859", "0.5224394", "0.5194632", "0.5187212", "0.5180454", "0.51764786", "0.5159425", "0.5148611", "0.51456267", "0.5124361", "0.51234096", "0.5119787", "0.5087559", "0.50809467", "0.5069893", "0.5065149", "0.5064116", "0.5060476", "0.5059948", "0.5053369", "0.50519055", "0.5040949", "0.50402224", "0.5031785", "0.5019977", "0.5016439", "0.5015174", "0.50148755", "0.5014751", "0.49955478", "0.49936238", "0.49923614", "0.49868354", "0.4971883", "0.49705055", "0.49677563", "0.49667716", "0.49589163", "0.49589005", "0.49576297", "0.49568847", "0.49567202", "0.49561557", "0.49553028", "0.4951734", "0.49432492", "0.49390846", "0.49282414", "0.49209958", "0.49076858", "0.48971164", "0.48903024", "0.48872837", "0.48728412", "0.48699558", "0.4869287", "0.48555943", "0.48473972", "0.48343706", "0.4831087", "0.4830757", "0.48300466", "0.48213086", "0.4816742", "0.48122245", "0.48116758", "0.48101372", "0.48101372", "0.4809626" ]
0.6752515
0
batch is a syntactic helper for registering a migration
func batch(sqls ...string) func(db migrations.DB) error { return func(db migrations.DB) error { for _, sql := range sqls { if _, err := db.Exec(sql); err != nil { return err } } return nil } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func migrate() {\n\n}", "func tableMigrationProcess(newdriver *xorm.Engine) {\n\tfor _, table := range migrationGroup() {\n\t\terr := newdriver.Sync2(table)\n\t\tif err != nil {\n\t\t\ttname := reflect.TypeOf(table).String()\n\t\t\tprintln(\"There is an error while creating: \", tname)\n\t\t}\n\t}\n\n}", "func generateMigration(migrator *dbmigrate.Migrator, engines []string, args ...string) error {\n\tif len(engines) == 1 && engines[0] == enginesNoOptDefVal {\n\t\tengines[0] = migrator.Engine\n\t}\n\n\tif len(engines) == 1 && engines[0] == \"all\" {\n\t\tengines = dbmigrate.Engines()\n\t}\n\n\tfpaths, err := migrator.GenerateMigration(strings.Join(args, \" \"), engines...)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"can't generate migration\")\n\t}\n\n\tfor _, fpath := range fpaths {\n\t\tfmt.Printf(\"created %s\\n\", fpath)\n\t}\n\n\treturn nil\n}", "func autoMigrate() {\n\t// Add auto migrate bellow this line\n\tConn.Exec(`CREATE EXTENSION IF NOT EXISTS \"uuid-ossp\";`)\n\tlog.Println(\"STARTING AUTO MIGRATE \")\n\tConn.AutoMigrate(\n\t\tsa_models.SaUser{},\n\t\tsa_models.SaRole{},\n\t\tsa_models.SaClient{},\n\t\tsa_models.SaBranch{},\n\t\tsa_models.SaCompany{},\n\t\tsa_models.SaUserBranch{},\n\t\tsa_models.SaUserCompany{},\n\t\tsa_models.SaRoleMenu{},\n\t\tsa_models.SaFileUpload{},\n\t\tsa_models.SaMenu{},\n\t)\n\n\tlog.Println(\"FINISHING AUTO MIGRATE \")\n}", "func registerMigration(dir string) {\n\tmigrations.MustRegisterTx(func(db migrations.DB) error {\n\t\tpath := dir + \"/up.sql\"\n\t\tsql, err := loadMigrationFile(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = db.Exec(sql)\n\t\treturn err\n\t}, func(db migrations.DB) error {\n\t\tpath := dir + \"/down.sql\"\n\t\tsql, err := loadMigrationFile(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = db.Exec(sql)\n\t\treturn err\n\t})\n}", "func autoMigrate() {\n\t// Add auto migrate bellow this line\n\tConn.Exec(`CREATE EXTENSION IF NOT EXISTS \"uuid-ossp\";`)\n\tlog.Println(\"STARTING AUTO MIGRATE \")\n\tConn.AutoMigrate(\n\t\tmodels.SsUser{},\n\t\tmodels.Paket{},\n\t\tmodels.CapsterCollection{},\n\t\tmodels.Barber{},\n\t\tmodels.BarberPaket{},\n\t\tmodels.BarberCapster{},\n\t\tmodels.SaFileUpload{},\n\t\tmodels.OrderH{},\n\t\tmodels.OrderD{},\n\t\tmodels.SsSequenceNo{},\n\t)\n\n\tConn.Exec(`\n\tCREATE OR REPLACE FUNCTION update_status_new_job() \n\tRETURNS void AS $$\n\t\t#variable_conflict use_variable\n\t\tDECLARE\n\t\t\tcurtime timestamp := now();\n\t\tBEGIN\n\t\t\tupdate order_h\n\t\t\t\tset status = 'C'\n\t\t\twhere status ='N' \n\t\t\tand order_date < current_timestamp\n\t\t\tand (EXTRACT(EPOCH FROM current_timestamp-order_date)/3600)::int >=2;\n\t\tEND;\n\t$$ LANGUAGE plpgsql;\n\n\tCREATE OR REPLACE FUNCTION public.last_day(date)\n\tRETURNS date AS\n\t$$\n \t\tSELECT (date_trunc('MONTH', $1) + INTERVAL '1 MONTH - 1 day')::date;\n\t$$ LANGUAGE 'sql' IMMUTABLE STRICT;\n\t\t\t\n\tCREATE OR REPLACE FUNCTION public.week_of_month(\n\t\tp_date DATE,\n\t\tp_direction INT -- DEFAULT 1 -- for 8.4 and above\n\t ) RETURNS INT AS\n\t $$\n\t\tSELECT CASE WHEN $2 >= 0 THEN\n\t\t\tcase when CEIL(EXTRACT(DAY FROM $1) / 7)::int > 4 then\n\t\t\t\t4 \n\t\t\telse\n\t\t\t\tCEIL(EXTRACT(DAY FROM $1) / 7)::int\n\t\t\tEND\n\t\tELSE \n\t\t 0 - CEIL(\n\t\t\t(EXTRACT(DAY FROM last_day($1)) - EXTRACT(DAY FROM $1) + 1) / 7\n\t\t )::int\n\t\tEND\n\t $$ LANGUAGE 'sql' IMMUTABLE;\n\t \n\t CREATE OR REPLACE FUNCTION public.fbarber_beranda_s(p_status varchar, p_date varchar)\n\t RETURNS \n\t TABLE(\n\t\t owner_id integer,\n\t\t barber_id integer,\n\t\t barber_name varchar,\n\t\t file_id integer,\n\t\t file_name varchar,\n\t\t file_path varchar,\n\t\t file_type varchar,\n\t\t price numeric\n\t )\n\t LANGUAGE plpgsql\n\t AS $function$\n\t DECLARE v_id INTEGER; \n\t BEGIN \t\n\t\t RETURN QUERY \n\t\t\t select \tbarber.owner_id ,\n\t\t\t\t\t\t barber.barber_id ,\n\t\t\t\t\t\t barber.barber_name ,\n\t\t\t\t\t\t barber.file_id ,\n\t\t\t\t\t\t sa_file_upload.file_name ,\n\t\t\t\t\t\t sa_file_upload.file_path ,\n\t\t\t\t\t\t sa_file_upload.file_type ,\n\t\t\t\t\t\t (\n\t\t\t\t\t\t\t select coalesce(sum(od.price),0) from order_d od join order_h oh\n\t\t\t\t\t\t\t\t on oh.order_id = od.order_id \n\t\t\t\t\t\t\t where oh.barber_id = barber.barber_id \n\t\t\t\t\t\t\t and oh.order_date::date = p_date::date \n\t\t\t\t\t\t\t and oh.status =p_status\n\t\t\t\t\t\t ) as price\n\t\t\t\t\t\t from barber\n\t\t\t\t\t\t left join sa_file_upload on sa_file_upload.file_id = barber.file_id\n\t ;\n\t\t\t\t\n\t END;\n\t $function$\n\t ;\n\n\t \n\t\tCREATE OR replace FUNCTION public.fbarber_beranda_status(p_status varchar, p_date varchar)\n\t\tRETURNS \n\t\tTABLE(\n\t\t\towner_id integer,\n\t\t\tprogress_status integer,\n\t\t\tfinish_status integer,\n\t\t\tcancel_status integer\n\t\t\t,income_price numeric\n\t\t)\n\t\tLANGUAGE plpgsql\n\t\tAS $function$\n\t\tDECLARE v_id INTEGER; \n\t\tBEGIN \t\n\t\t\tRETURN QUERY \n\t\t\tselect \n\t\t\t\tss.owner_id ,\n\t\t\t\tss.progress_status,\n\t\t\t\tss.finish_status,\n\t\t\t\tss.cancel_status\n\t\t\t\t,\n\t\t\t\t(\n\t\t\t\t\t\tselect coalesce(sum(od.price ),0)::numeric \n\t\t\t\t\tfrom order_h oh join order_d od\n\t\t\t\t\ton oh.order_id = od.order_id \n\t\t\t\t\twhere oh.order_date::date= p_date::date \n\t\t\t\t\tand oh.status = p_status\n\t\t\t\t\tand oh.barber_id in(\n\t\t\t\t\t\tselect b2.barber_id from barber b2 \n\t\t\t\t\t\twhere b2.owner_id = ss.owner_id\n\t\t\t\t\t)\n\t\t\t\t)::numeric as income_price\n\t\t\tfrom (\n\t\t\t\tselect b.owner_id,\n\t\t\t\tcount(case a.status when 'P' then 1 else null end)::integer as progress_status,\n\t\t\t\t\t\t\tcount(case a.status when 'F' then 1 else null end)::integer as finish_status,\n\t\t\t\t\t\t\tcount(case a.status when 'C' then 1 else null end)::integer as cancel_status\n\t\t\t\tfrom order_h a join barber b\n\t\t\t\ton a.barber_id = b.barber_id \n\t\t\t\twhere a.order_date::date=p_date::date\n\t\t\t\tgroup by b.owner_id\n\t\t\t) ss\t\t\n\n\t\t;\n\t\t\t\t\n\t\tEND;\n\t\t$function$\n\t\t;\n\n\t \n\n\t\tCREATE OR REPLACE VIEW public.v_order_h\n\t\tAS SELECT barber.owner_id,\n\t\t\tbarber.barber_id,\n\t\t\tbarber.barber_name,\n\t\t\torder_h.order_id,\n\t\t\torder_h.status,\n\t\t\torder_h.from_apps,\n\t\t\torder_h.capster_id,\n\t\t\torder_h.order_date,\n\t\t\tss_user.name AS capster_name,\n\t\t\tss_user.file_id,\n\t\t\tsa_file_upload.file_name,\n\t\t\tsa_file_upload.file_path,\n\t\t\t( SELECT sum(order_d.price) AS sum\n\t\t\t\tFROM order_d\n\t\t\t\tWHERE order_d.order_id = order_h.order_id) AS price,\n\t\t\tweek_of_month(order_h.order_date::date, 1) AS weeks,\n\t\t\tdate_part('year'::text, order_h.order_date) AS years,\n\t\t\tdate_part('month'::text, order_h.order_date) AS months,\n\t\t\torder_h.customer_name,\n\t\t\torder_h.order_no\n\t\tFROM barber\n\t\t\tJOIN order_h ON order_h.barber_id = barber.barber_id\n\t\t\tJOIN ss_user ON ss_user.user_id = order_h.capster_id\n\t\t\tLEFT JOIN sa_file_upload ON sa_file_upload.file_id = ss_user.file_id;\n\n\n\t\tCREATE OR REPLACE VIEW public.v_capster\n\t\tAS\n\t\tSELECT \n\t\t\tss_user.user_id as capster_id,ss_user.user_name,ss_user.name,\n\t\t\tss_user.is_active,sa_file_upload.file_id,sa_file_upload.file_name,\n\t\t\tsa_file_upload.file_path,sa_file_upload.file_type, 0 as rating,\n\t\t\t(case when b.barber_id is not null then true else false end) as in_use,\n\t\t\tss_user.user_type,ss_user.user_input,ss_user.time_edit ,b.barber_id,\n\t\t\tb.barber_name\n\t\t\t\n\t\t FROM \"ss_user\" \n\t\t\tleft join sa_file_upload ON sa_file_upload.file_id = ss_user.file_id\n\t\t\tleft join barber_capster bc on bc.capster_id = ss_user.user_id \n\t\t\tleft join barber b on bc.barber_id =b.barber_id \n\t\t\tand b.owner_id::varchar = ss_user.user_input;\n\t \n\t `)\n\n\tlog.Println(\"FINISHING AUTO MIGRATE \")\n}", "func registerMigrations(ctx context.Context, db dbutil.DB, outOfBandMigrationRunner *oobmigration.Runner) error {\n\tmigrators := map[int]oobmigration.Migrator{\n\t\tmigration.DiagnosticsCountMigrationID: migration.NewDiagnosticsCountMigrator(services.lsifStore, config.DiagnosticsCountMigrationBatchSize),\n\t\tmigration.DefinitionsCountMigrationID: migration.NewLocationsCountMigrator(services.lsifStore, \"lsif_data_definitions\", config.DefinitionsCountMigrationBatchSize),\n\t\tmigration.ReferencesCountMigrationID: migration.NewLocationsCountMigrator(services.lsifStore, \"lsif_data_references\", config.ReferencesCountMigrationBatchSize),\n\t}\n\n\tfor id, migrator := range migrators {\n\t\tif err := outOfBandMigrationRunner.Register(id, migrator, oobmigration.MigratorOptions{Interval: time.Second}); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func ServiceAutoMigration(db *gorm.DB) error {\n\t// Keep a list of migrations here\n\tm := gormigrate.New(db, gormigrate.DefaultOptions, nil)\n\tm.InitSchema(func(db *gorm.DB) error {\n\t\tlog.Info(\"[Migration.InitSchema] Initializing database schema\")\n\t\tdb.Exec(\"create extension \\\"uuid-ossp\\\";\")\n\t\tif err := updateMigration(db); err != nil {\n\t\t\treturn fmt.Errorf(\"[Migration.InitSchema]: %v\", err)\n\t\t}\n\t\t// Add more jobs, etc here\n\t\treturn nil\n\t})\n\tm.Migrate()\n\n\tif err := updateMigration(db); err != nil {\n\t\treturn err\n\t}\n\tm = gormigrate.New(db, gormigrate.DefaultOptions, []*gormigrate.Migration{\n\t\tjobs.SeedUsers,\n\t})\n\treturn m.Migrate()\n}", "func Register(up, down func(DB) error) error {\n\t_, file, _, _ := runtime.Caller(1)\n\tversion, err := extractVersion(file)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tallMigrations = append(allMigrations, Migration{\n\t\tVersion: version,\n\t\tUp: up,\n\t\tDown: down,\n\t})\n\treturn nil\n}", "func AutoMigrate(values ...interface{}) {\n\tfor _, value := range values {\n\t\tdb.AutoMigrate(value)\n\t}\n}", "func migrate(c *sql.DB) error {\n\t// Acquire advisory lock\n\t_, err := c.Exec(\"SELECT pg_advisory_lock($1);\", constants.AdvisoryLockID)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"acquiring advisory lock %d\", constants.AdvisoryLockID)\n\t}\n\n\t_, err = c.Exec(`\n CREATE TABLE IF NOT EXISTS migrations (\n num INTEGER NOT NULL,\n descr TEXT,\n at TIMESTAMPTZ NOT NULL\n );\n`)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"creating migration table\")\n\t}\n\n\ti := 0\n\tdone := false\n\tvar tx *sql.Tx\n\tfor {\n\n\t\ttx, err = c.Begin()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\n\t\t// Within a transaction, perform a single migration.\n\t\tvar num sql.NullInt64\n\t\tn := 0\n\n\t\tif err = tx.QueryRow(`SELECT MAX(num) FROM migrations;`).Scan(&num); err != nil {\n\t\t\terr = errors.Wrap(err, \"select max migration\")\n\t\t\tbreak\n\t\t}\n\t\tif num.Valid { // if not, n will still be 0\n\t\t\tn = int(num.Int64)\n\t\t}\n\n\t\tif n >= len(migrations) {\n\t\t\tdone = true\n\t\t} else {\n\t\t\tmigrationNum := n + 1\n\t\t\tm := migrations[n]\n\t\t\tif _, err = tx.Exec(m.statement); err != nil {\n\t\t\t\terr = errors.Wrapf(err, \"migration '%s' (%d) failed\", m.descr, migrationNum)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tq := `INSERT INTO migrations (num, descr, at) VALUES ($1, $2, now());`\n\t\t\tif _, err = tx.Exec(q, migrationNum, m.descr); err != nil {\n\t\t\t\terr = errors.Wrap(err, \"update migration table\")\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\t// done with one migration, commit transaction\n\t\tif err = tx.Commit(); err != nil {\n\t\t\tbreak\n\t\t}\n\n\t\tif done {\n\t\t\tbreak\n\t\t}\n\t\ti++ // rinse and repeat for next migration\n\t}\n\n\tif err != nil {\n\t\t// This adapter will go away soon. No need to fix this.\n\t\ttx.Rollback() // nolint\n\t}\n\n\t// Release advisory lock\n\t_, releaseErr := c.Exec(\"SELECT pg_advisory_unlock($1);\", constants.AdvisoryLockID)\n\tif releaseErr != nil {\n\t\treleaseErr = errors.Wrapf(releaseErr, \"releasing advisory lock %d\", constants.AdvisoryLockID)\n\t\t// Want to report back on both the migration error and also the advisory release error, if there are both.\n\t\tif err != nil {\n\t\t\treturn errors.New(err.Error() + releaseErr.Error())\n\t\t}\n\t\treturn releaseErr\n\t}\n\n\treturn err\n}", "func init() {\n\tMigrations = append(Migrations, addUserMigration0001)\n}", "func init() {\n\tmigrations.AddPluginMigration(\"inputs.cassandra\", migrate)\n}", "func RunMigrations(db *gorm.DB) {\n\tdb.AutoMigrate(&Todo{})\n\n\tfmt.Println(\"AutoMigrations completed.\")\n}", "func validateAndMigrate(db *gorm.DB, models ...models.Nego) {\n\tfor _, v := range models {\n\t\tdb.AutoMigrate(v)\n\t}\n}", "func Migrate(tx *gorm.DB) error {\n\treturn tx.Exec(`\n\tCREATE OR REPLACE FUNCTION notifyPipelineRunStarted() RETURNS TRIGGER AS $_$\n\tBEGIN\n\t\tIF NEW.finished_at IS NULL THEN\n\t\t\tPERFORM pg_notify('pipeline_run_started', NEW.id::text);\n\t\tEND IF;\n\t\tRETURN NEW;\n\tEND\n\t$_$ LANGUAGE 'plpgsql';\n\t`).Error\n}", "func sqlMigration(db *IndexerDb, state *MigrationState, sqlLines []string) error {\n\tdb.accountingLock.Lock()\n\tdefer db.accountingLock.Unlock()\n\n\tnextState := *state\n\tnextState.NextMigration++\n\n\tf := func(ctx context.Context, tx *sql.Tx) error {\n\t\tdefer tx.Rollback()\n\n\t\tfor _, cmd := range sqlLines {\n\t\t\t_, err := tx.Exec(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\n\t\t\t\t\t\"migration %d exec cmd: \\\"%s\\\" err: %w\", state.NextMigration, cmd, err)\n\t\t\t}\n\t\t}\n\t\tmigrationStateJSON := encoding.EncodeJSON(nextState)\n\t\t_, err := tx.Exec(setMetastateUpsert, migrationMetastateKey, migrationStateJSON)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"migration %d exec metastate err: %w\", state.NextMigration, err)\n\t\t}\n\t\treturn tx.Commit()\n\t}\n\terr := db.txWithRetry(context.Background(), serializable, f)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"migration %d commit err: %w\", state.NextMigration, err)\n\t}\n\n\t*state = nextState\n\treturn nil\n}", "func sqlMigration(db *IndexerDb, state *MigrationState, sqlLines []string) error {\n\tdb.accountingLock.Lock()\n\tdefer db.accountingLock.Unlock()\n\n\tnextState := *state\n\tnextState.NextMigration++\n\n\tf := func(ctx context.Context, tx *sql.Tx) error {\n\t\tdefer tx.Rollback()\n\n\t\tfor _, cmd := range sqlLines {\n\t\t\t_, err := tx.Exec(cmd)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\n\t\t\t\t\t\"migration %d exec cmd: \\\"%s\\\" err: %w\", state.NextMigration, cmd, err)\n\t\t\t}\n\t\t}\n\t\tmigrationStateJSON := encoding.EncodeJSON(nextState)\n\t\t_, err := tx.Exec(setMetastateUpsert, migrationMetastateKey, migrationStateJSON)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"migration %d exec metastate err: %w\", state.NextMigration, err)\n\t\t}\n\t\treturn tx.Commit()\n\t}\n\terr := db.txWithRetry(context.Background(), serializable, f)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"migration %d commit err: %w\", state.NextMigration, err)\n\t}\n\n\t*state = nextState\n\treturn nil\n}", "func migrate(migrator *dbmigrate.Migrator, steps int) (int, error) {\n\tdone := make(chan struct{})\n\tgdone := make(chan struct{})\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase migration := <-migrator.MigrationsCh:\n\t\t\t\tfmt.Printf(\"migration %s has been successfully applied\\n\", migration.FileName())\n\t\t\tcase <-done:\n\t\t\t\tclose(gdone)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\tn, err := migrator.MigrateSteps(steps)\n\tclose(done)\n\n\t<-gdone\n\tif err != nil {\n\t\treturn n, errors.Wrap(err, \"can't migrate\")\n\t}\n\n\tif n == 0 {\n\t\tfmt.Println(\"there are no migrations to apply\")\n\t\treturn n, nil\n\t}\n\tfmt.Printf(\"%d %s successfully applied\\n\", n, pluralize(\"migration\", n))\n\n\treturn n, nil\n}", "func BackfillTable(table db.TableDescriptor, target map[string]interface{}) []Migration {\n var result []Migration\n\n // check for additions\n for key, value := range target {\n var found = false\n var col db.ColumnDescriptor\n for _, col = range table.Columns {\n if (col.Name == key) {\n found = true\n\n // multitask and look for changed type\n // convert types to upper case for ease\n var upperKnown = strings.ToUpper(col.Type)\n var upperTesting = strings.ToUpper(value.(string))\n if (upperKnown != upperTesting) {\n // handle the PRIMARY KEY part a little differently\n if (col.Primary) {\n var trimmedTesting = upperTesting[:len(upperTesting) - len(\" PRIMARY KEY\")]\n if (upperKnown == trimmedTesting) { continue }\n }\n result = append(result, ChangeTypeMigration(table, col.Name, value.(string)))\n }\n\n // escape for loop\n break\n }\n }\n\n if (!found) {\n result = append(result, CreationMigration(table, key, value.(string)))\n }\n }\n\n\n // check for removals\n for _, col := range table.Columns {\n var found = false\n for key, _ := range target {\n if (col.Name == key) {\n found = true\n break\n }\n }\n\n if (!found) {\n result = append(result, RemovalMigration(table, col.Name))\n }\n }\n\n\n return result\n}", "func Migrate(ctx context.Context, s *store.Store, serializer serialization.Serializer) error {\n\tsteps := []MigrationStep{\n\t\tcreateTempTables,\n\t\treencodeDocuments,\n\t\treencodeResultChunks,\n\t\treencodeDefinitions,\n\t\treencodeReferences,\n\t\tswapTables,\n\t}\n\n\t// NOTE: We need to serialize with the JSON serializer, NOT the current serializer. This is\n\t// because future migrations assume that v4 was written with the most current serializer at\n\t// that time. Using the current serializer will cause future migrations to fail to read the\n\t// encoded data.\n\tdeserializer := jsonserializer.New()\n\n\tfor _, step := range steps {\n\t\tif err := step(ctx, s, deserializer, serializer); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func migrate(name string, port uint64, peerListenUrls string, peerAdvertiseUrls string, clientListenUrls string,\n\tbinPath string, dataDirPath string, etcdDataPrefix string, ttlKeysDirectory string,\n\tinitialCluster string, target *EtcdVersionPair, bundledVersions SupportedVersions, etcdServerArgs string) {\n\n\tdataDir, err := OpenOrCreateDataDirectory(dataDirPath)\n\tif err != nil {\n\t\tklog.Fatalf(\"Error opening or creating data directory %s: %v\", dataDirPath, err)\n\t}\n\n\tcfg := &EtcdMigrateCfg{\n\t\tbinPath: binPath,\n\t\tname: name,\n\t\tport: port,\n\t\tpeerListenUrls: peerListenUrls,\n\t\tpeerAdvertiseUrls: peerAdvertiseUrls,\n\t\tclientListenUrls: clientListenUrls,\n\t\tetcdDataPrefix: etcdDataPrefix,\n\t\tttlKeysDirectory: ttlKeysDirectory,\n\t\tinitialCluster: initialCluster,\n\t\tsupportedVersions: bundledVersions,\n\t\tdataDirectory: dataDirPath,\n\t\tetcdServerArgs: etcdServerArgs,\n\t}\n\tclient, err := NewEtcdMigrateClient(cfg)\n\tif err != nil {\n\t\tklog.Fatalf(\"Migration failed: %v\", err)\n\t}\n\tdefer client.Close()\n\n\tmigrator := &Migrator{cfg, dataDir, client}\n\n\terr = migrator.MigrateIfNeeded(target)\n\tif err != nil {\n\t\tklog.Fatalf(\"Migration failed: %v\", err)\n\t}\n}", "func migrate0001(db *gorm.DB) error {\n\tmigrationName := \"oauth_initial\"\n\n\tmigration := new(migrations.Migration)\n\tfound := !db.Where(\"name = ?\", migrationName).First(migration).RecordNotFound()\n\n\tif found {\n\t\tlogger.INFO.Printf(\"Skipping %s migration\", migrationName)\n\t\treturn nil\n\t}\n\n\tlogger.INFO.Printf(\"Running %s migration\", migrationName)\n\n\tvar err error\n\n\t// Create oauth_clients table\n\tif err := db.CreateTable(new(Client)).Error; err != nil {\n\t\treturn fmt.Errorf(\"Error creating oauth_clients table: %s\", err)\n\t}\n\n\t// Create oauth_scopes table\n\tif err := db.CreateTable(new(Scope)).Error; err != nil {\n\t\treturn fmt.Errorf(\"Error creating oauth_scopes table: %s\", err)\n\t}\n\n\t// Create oauth_users table\n\tif err := db.CreateTable(new(User)).Error; err != nil {\n\t\treturn fmt.Errorf(\"Error creating oauth_users table: %s\", err)\n\t}\n\n\t// Create oauth_refresh_tokens table\n\tif err := db.CreateTable(new(RefreshToken)).Error; err != nil {\n\t\treturn fmt.Errorf(\"Error creating oauth_refresh_tokens table: %s\", err)\n\t}\n\n\t// Create oauth_access_tokens table\n\tif err := db.CreateTable(new(AccessToken)).Error; err != nil {\n\t\treturn fmt.Errorf(\"Error creating oauth_access_tokens table: %s\", err)\n\t}\n\n\t// Create oauth_authorization_codes table\n\tif err := db.CreateTable(new(AuthorizationCode)).Error; err != nil {\n\t\treturn fmt.Errorf(\"Error creating oauth_authorization_codes table: %s\", err)\n\t}\n\n\t// Add foreign key on oauth_refresh_tokens.client_id\n\terr = db.Model(new(RefreshToken)).AddForeignKey(\n\t\t\"client_id\",\n\t\t\"oauth_clients(id)\",\n\t\t\"RESTRICT\",\n\t\t\"RESTRICT\",\n\t).Error\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating foreign key on \"+\n\t\t\t\"oauth_refresh_tokens.client_id for oauth_clients(id): %s\", err)\n\t}\n\n\t// Add foreign key on oauth_refresh_tokens.user_id\n\terr = db.Model(new(RefreshToken)).AddForeignKey(\n\t\t\"user_id\",\n\t\t\"oauth_users(id)\",\n\t\t\"RESTRICT\",\n\t\t\"RESTRICT\",\n\t).Error\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating foreign key on \"+\n\t\t\t\"oauth_refresh_tokens.user_id for oauth_users(id): %s\", err)\n\t}\n\n\t// Add foreign key on oauth_access_tokens.client_id\n\terr = db.Model(new(AccessToken)).AddForeignKey(\n\t\t\"client_id\",\n\t\t\"oauth_clients(id)\",\n\t\t\"RESTRICT\",\n\t\t\"RESTRICT\",\n\t).Error\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating foreign key on \"+\n\t\t\t\"oauth_access_tokens.client_id for oauth_clients(id): %s\", err)\n\t}\n\n\t// Add foreign key on oauth_access_tokens.user_id\n\terr = db.Model(new(AccessToken)).AddForeignKey(\n\t\t\"user_id\",\n\t\t\"oauth_users(id)\",\n\t\t\"RESTRICT\",\n\t\t\"RESTRICT\",\n\t).Error\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating foreign key on \"+\n\t\t\t\"oauth_access_tokens.user_id for oauth_users(id): %s\", err)\n\t}\n\n\t// Add foreign key on oauth_authorization_codes.client_id\n\terr = db.Model(new(AuthorizationCode)).AddForeignKey(\n\t\t\"client_id\",\n\t\t\"oauth_clients(id)\",\n\t\t\"RESTRICT\",\n\t\t\"RESTRICT\",\n\t).Error\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating foreign key on \"+\n\t\t\t\"oauth_authorization_codes.client_id for oauth_clients(id): %s\", err)\n\t}\n\n\t// Add foreign key on oauth_authorization_codes.user_id\n\terr = db.Model(new(AuthorizationCode)).AddForeignKey(\n\t\t\"user_id\",\n\t\t\"oauth_users(id)\",\n\t\t\"RESTRICT\",\n\t\t\"RESTRICT\",\n\t).Error\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Error creating foreign key on \"+\n\t\t\t\"oauth_authorization_codes.user_id for oauth_users(id): %s\", err)\n\t}\n\n\t// Save a record to migrations table,\n\t// so we don't rerun this migration again\n\tmigration.Name = migrationName\n\tif err := db.Create(migration).Error; err != nil {\n\t\treturn fmt.Errorf(\"Error saving record to migrations table: %s\", err)\n\t}\n\n\treturn nil\n}", "func RunMigrations(c *datastore.Client, ns string, migratable interface{}) {\n\n\tfmt.Println(\"Running Unison\")\n\n\t// Get the timestamp of the last applied migration\n\tlastAppliedMigrationTS := getLastAppliedMigrationTimeStamp(c, ns)\n\n\ttMigratable := reflect.TypeOf(migratable)\n\n\tmigrationMetaSet, err := generateMigrationMetaSet(tMigratable)\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tsort.Sort(migrationMetaSet)\n\n\tvMigratable := reflect.ValueOf(migratable)\n\n\tfor _, migration := range migrationMetaSet.set {\n\n\t\tif migration.Timestamp <= lastAppliedMigrationTS {\n\t\t\tcontinue\n\t\t}\n\n\t\tfmt.Printf(\"Applying migration %d ... \", migration.Timestamp)\n\n\t\ttx, _ := c.NewTransaction(context.Background())\n\n\t\tmethod := vMigratable.MethodByName(migration.Name)\n\n\t\trtx := reflect.ValueOf(tx)\n\t\trns := reflect.ValueOf(ns)\n\n\t\tvalues := method.Call([]reflect.Value{rtx, rns})\n\n\t\tvalue := values[0]\n\n\t\tif !value.IsNil() {\n\t\t\te, _ := value.Interface().(error)\n\n\t\t\tfmt.Printf(\"Failed. Cause: %s\\n\", e.Error())\n\t\t\treturn\n\t\t}\n\n\t\ttxkey := &datastore.Key{Kind: UnisonMigrationMetaKind, Name: migration.Name, Namespace: ns}\n\t\tmigration.AppliedAt = time.Now()\n\n\t\ttx.Put(txkey, &migration)\n\n\t\t_, err := tx.Commit()\n\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tfmt.Printf(\"Done\\n\")\n\t}\n\n\tfmt.Printf(\"We are done !!\\n\")\n}", "func doDatabaseMigrations(db *WowDB, env *Env, blizzard Blizzard) {\n\n\tif ! db.HasTable(&Race{}) {\n\t\tlog.Debug(\"Migrating race\")\n\t\tdb.AutoMigrate(&Race{})\n\t\terr := UpdateRacesFromBlizzard(env, blizzard)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Could not update races: %v\", err)\n\t\t}\n\t}\n\n\tif ! db.HasTable(&ToonClass{}) {\n\t\tlog.Println(\"Migrating classes\")\n\t\tdb.AutoMigrate(&ToonClass{})\n\t\terr := UpdateClassesFromBlizzard(env, blizzard)\n\t\tif err != nil {\n\t\t\tlog.Errorf(\"Could not update classes: %v\", err)\n\t\t}\n\t}\n\n\tdb.AutoMigrate(&Stat{})\n\tdb.AutoMigrate(&Toon{})\n\n\tif ! db.HasTable(&ClassColor{}) {\n\t\tdb.AutoMigrate(&ClassColor{})\n\t\tdb.Model(&ClassColor{}).AddForeignKey(\"toon_class_id\", \"toon_classes(id)\", \"RESTRICT\", \"RESTRICT\")\n\t\tvar tColor = ClassColor{ToonClassID: 1, Color: \"#C79C63\"}\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 2\n\t\ttColor.Color = \"#F58CBA\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 3\n\t\ttColor.Color = \"#ABD473\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 4\n\t\ttColor.Color = \"#FFF569\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 5\n\t\ttColor.Color = \"#F0EBE0\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 6\n\t\ttColor.Color = \"#C41F3B\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 7\n\t\ttColor.Color = \"#0070DE\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 8\n\t\ttColor.Color = \"#69CCF0\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 9\n\t\ttColor.Color = \"#9482C9\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 10\n\t\ttColor.Color = \"#00FF96\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 11\n\t\ttColor.Color = \"#FF7D0A\"\n\t\tdb.Create(&tColor)\n\t\ttColor.ID = 0\n\t\ttColor.ToonClassID = 12\n\t\ttColor.Color = \"#A330C9\"\n\t\tdb.Create(&tColor)\n\t}\n\n\tdb.Model(&Toon{}).AddForeignKey(\"race_id\", \"races(id)\", \"RESTRICT\", \"RESTRICT\")\n\tdb.Model(&Toon{}).AddForeignKey(\"class_id\", \"toon_classes(id)\", \"RESTRICT\", \"RESTRICT\")\n\tdb.Model(&Stat{}).AddForeignKey(\"toon_id\", \"toons(id)\", \"RESTRICT\", \"RESTRICT\")\n\tdb.Model(&Stat{}).AddUniqueIndex(\"idx_toon_id_create_date\", \"toon_id\", \"insert_date\")\n}", "func migrationGroup() []interface{} {\n\tvar tables []interface{}\n\ttables = append(tables, &Urls{})\n\ttables = append(tables, &Users{})\n\treturn tables\n}", "func main() {\n\tcontainer := server.BuildContainer()\n\n\terr := container.Invoke(func(db *sql.DB) {\n\t\tdefer db.Close()\n\n\t\tmigration := &CreateUsersTable01{db}\n\n\t\tres, err := migration.Up()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tfmt.Println(res)\n\t})\n\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func ExecuteMigrator(config *config.Config, action *string, createConnector func(*config.Config) db.Connector, createLoader func(*config.Config) loader.Loader) int {\n\n\tswitch *action {\n\tcase PrintConfigAction:\n\t\tlog.Printf(\"Configuration file ==>\\n%v\\n\", config)\n\t\treturn 0\n\tcase ListDiskMigrationsAction:\n\t\tdiskMigrations := LoadDiskMigrations(config, createLoader)\n\t\tlog.Printf(\"List of disk migrations ==>\\n%v\", types.MigrationArrayString(diskMigrations))\n\t\treturn 0\n\tcase ListDBMigrationsAction:\n\t\tdbMigrations := LoadDBMigrations(config, createConnector)\n\t\tlog.Printf(\"List of db migrations ==> \\n%v\", types.MigrationDBArrayString(dbMigrations))\n\t\treturn 0\n\tcase ListDBTenantsAction:\n\t\tdbTenants := LoadDBTenants(config, createConnector)\n\t\tlog.Printf(\"List of db tenants ==> \\n%v\", types.TenantArrayString(dbTenants))\n\t\treturn 0\n\tcase ApplyAction:\n\t\tmigrationsApplied := ApplyMigrations(config, createConnector, createLoader)\n\t\tlog.Printf(\"List of migrations applied ==>\\n%v\", types.MigrationArrayString(migrationsApplied))\n\t\treturn 0\n\tdefault:\n\t\tlog.Printf(\"Unknown action to run %q. For usage please run migrator with -h flag.\", *action)\n\t\treturn 1\n\t}\n}", "func (db *Database) RunMigrations(ctx context.Context) error {\n\tlogger := logging.FromContext(ctx)\n\tm := gormigrate.New(db.db, gormigrate.DefaultOptions, []*gormigrate.Migration{\n\t\t{\n\t\t\tID: \"00001-CreateUsers\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: creating users table\")\n\t\t\t\treturn tx.AutoMigrate(&User{}).Error\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\treturn tx.DropTable(\"users\").Error\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00002-CreateVerificationCodes\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: creating verification codes table\")\n\t\t\t\treturn tx.AutoMigrate(&VerificationCode{}).Error\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\treturn tx.DropTable(\"verification_codes\").Error\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00003-CreateAuthorizedApps\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: creating authorized apps table\")\n\t\t\t\treturn tx.AutoMigrate(&AuthorizedApp{}).Error\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\treturn tx.DropTable(\"authorized_apps\").Error\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00004-CreateTokens\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: creating tokens table\")\n\t\t\t\treturn tx.AutoMigrate(&Token{}).Error\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\treturn tx.DropTable(\"tokens\").Error\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00005-CreateCleanups\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: creating cleanup status table\")\n\t\t\t\tif err := tx.AutoMigrate(&CleanupStatus{}).Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\t// Seed database w/ cleanup record.\n\t\t\t\tif err := tx.Create(&CleanupStatus{Type: \"cleanup\", Generation: 1, NotBefore: time.Now()}).Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\treturn tx.DropTable(\"cleanup_statuses\").Error\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00006-AddIndexes\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: add users purge index\")\n\t\t\t\tif err := tx.Model(&User{}).AddIndex(\"users_purge_index\", \"disabled\", \"updated_at\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tlogger.Infof(\"db migrations: add verification code purge index\")\n\t\t\t\tif err := tx.Model(&VerificationCode{}).AddIndex(\"ver_code_purge_index\", \"expires_at\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tlogger.Infof(\"db migrations: add tokens purge index\")\n\t\t\t\tif err := tx.Model(&VerificationCode{}).AddIndex(\"token_purge_index\", \"expires_at\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: drop users purge index\")\n\t\t\t\tif err := tx.Model(&User{}).RemoveIndex(\"users_purge_index\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tlogger.Infof(\"db migrations: drop verification code purge index\")\n\t\t\t\tif err := tx.Model(&VerificationCode{}).RemoveIndex(\"ver_code_purge_index\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tlogger.Infof(\"db migrations: drop tokens purge index\")\n\t\t\t\tif err := tx.Model(&VerificationCode{}).RemoveIndex(\"token_purge_index\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00007-AddSymptomOnset\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Info(\"db migrations: rename test_date to symptom_date\")\n\t\t\t\t// AutoMigrate will add missing fields.\n\t\t\t\tif err := tx.AutoMigrate(&VerificationCode{}).Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\t// If not upgrading from an old version, this column will have never been created.\n\t\t\t\tif tx.NewScope(&VerificationCode{}).HasColumn(\"test_date\") {\n\t\t\t\t\tif err := tx.Model(&VerificationCode{}).DropColumn(\"test_date\").Error; err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tif err := tx.AutoMigrate(&Token{}).Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\t// If not upgrading from an old version, this column will have never been created.\n\t\t\t\tif tx.NewScope(&Token{}).HasColumn(\"test_date\") {\n\t\t\t\t\tif err := tx.Model(&Token{}).DropColumn(\"test_date\").Error; err != nil {\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Info(\"db migrations: rename symptom_date to test_date\")\n\t\t\t\tif err := tx.Model(&VerificationCode{}).DropColumn(\"symptom_date\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tif err := tx.Model(&Token{}).DropColumn(\"symptom_date\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00008-AddKeyTypes\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: upgrading authorized_apps table.\")\n\t\t\t\treturn tx.AutoMigrate(&AuthorizedApp{}).Error\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\tif err := tx.Model(&AuthorizedApp{}).DropColumn(\"admin_key\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00009-AddIssuerColumns\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: adding issuer columns to issued codes\")\n\t\t\t\treturn tx.AutoMigrate(&VerificationCode{}).Error\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\tif err := tx.Model(&AuthorizedApp{}).DropColumn(\"issuing_user\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tif err := tx.Model(&AuthorizedApp{}).DropColumn(\"issuing_app\").Error; err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn nil\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tID: \"00010-AddSMSConfig\",\n\t\t\tMigrate: func(tx *gorm.DB) error {\n\t\t\t\tlogger.Infof(\"db migrations: adding sms_configs table\")\n\t\t\t\treturn tx.AutoMigrate(&SMSConfig{}).Error\n\t\t\t},\n\t\t\tRollback: func(tx *gorm.DB) error {\n\t\t\t\treturn tx.DropTable(\"sms_configs\").Error\n\t\t\t},\n\t\t},\n\t})\n\n\tlogger.Infof(\"database migrations complete\")\n\n\treturn m.Migrate()\n}", "func Migration(table, directory string) Initializer {\n\tif _, err := os.Stat(directory); os.IsNotExist(err) {\n\t\tstartup_base.PanicOnError(err, \"No database migration files found\")\n\t}\n\n\treturn func(db *sqlx.DB) error {\n\t\tmigrate.SetTable(table)\n\n\t\tmigrations := &migrate.FileMigrationSource{Dir: directory}\n\t\tn, err := migrate.Exec(db.DB, \"postgres\", migrations, migrate.Up)\n\t\tif err != nil {\n\t\t\treturn errors.WithMessage(err, \"applying database migration\")\n\t\t}\n\n\t\tlogrus.WithField(\"prefix\", \"database\").\n\t\t\tInfof(\"%d migrations executed\", n)\n\n\t\treturn nil\n\t}\n}", "func Migrate(values ...interface{}) {\n\tdb.AutoMigrate(values...)\n}", "func doMigrations(db *pg.DB) error {\n\tconst op errors.Op = \"repositories.doMigrations\"\n\tfor _, model := range entities {\n\t\terr := db.CreateTable(model, &orm.CreateTableOptions{\n\t\t\tIfNotExists: true,\n\t\t\tFKConstraints: true,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn errors.E(op, errors.Internal, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *subscriptionMigrator) migrateAll() error {\n\tlog.Printf(\"Starting migration of %d Subscription(s)\", len(m.subscriptions))\n\n\tfor _, sub := range m.subscriptions {\n\t\tsubKey := fmt.Sprintf(\"%s/%s\", sub.Namespace, sub.Name)\n\n\t\tif err := m.migrateSubscription(sub); err != nil {\n\t\t\treturn errors.Wrapf(err, \"migrating Subscription %q\", subKey)\n\t\t}\n\n\t\tlog.Printf(\"+ Deleting Subscription %q\", subKey)\n\n\t\tif err := m.deleteSubscriptionWithRetry(sub.Namespace, sub.Name); err != nil {\n\t\t\treturn errors.Wrapf(err, \"deleting Subscription %q\", subKey)\n\t\t}\n\t}\n\n\treturn nil\n}", "func runAllMigrations(db *sql.DB) {\n\tLogInfo(\"RUNNING ALL MIGRATIONS\\n\")\n\n\tcreateMigrationsTable(db)\n\n\tfilenamePattern := fmt.Sprintf(\"./*.sql\")\n\tfiles, err := filepath.Glob(filenamePattern)\n\tif err != nil {\n\t\tLogErr(\"%s\\n\", err)\n\t}\n\n\tfor _, filePath := range files {\n\t\tfilename := filepath.Base(filePath)\n\n\t\tif checkIfMigrated(filename, db) {\n\t\t\tLogNotice(\"SKIPPING %s\\n\", filename)\n\t\t\tcontinue\n\t\t}\n\n\t\tsqlQuery, fileErr := ioutil.ReadFile(fmt.Sprintf(\"./%s\", filename))\n\t\tif fileErr != nil {\n\t\t\tLogErr(\"%s\\n\", fileErr)\n\t\t}\n\n\t\texecQuery(db, string(sqlQuery))\n\t\texecQuery(db, \"INSERT INTO _migrations (migration) VALUES($1)\", filename)\n\t\tLogSuccess(\"%s MIGRATED\\n\", filename)\n\t}\n\n}", "func GetMigrations(configuration MigrationConfiguration) Migrations {\n\tm := Migrations{}\n\n\t// Version 0\n\tm = append(m, steps{ExecuteSQLFile(\"000-bootstrap.sql\")})\n\n\t// Version 1\n\tm = append(m, steps{ExecuteSQLFile(\"001-identities-users.sql\")})\n\n\t// Version 2\n\tm = append(m, steps{ExecuteSQLFile(\"002-oauth-states.sql\")})\n\n\t// Version 3\n\tm = append(m, steps{ExecuteSQLFile(\"003-space-resources.sql\")})\n\n\t// Version 4\n\tm = append(m, steps{ExecuteSQLFile(\"004-unique-resource-space.sql\")})\n\n\t// Version 5\n\tm = append(m, steps{ExecuteSQLFile(\"005-authorization.sql\")})\n\n\t// Version 6\n\tm = append(m, steps{ExecuteSQLFile(\"006-external-provider.sql\")})\n\n\t// Version 7\n\tm = append(m, steps{ExecuteSQLFile(\"007-external-provider-id-index.sql\")})\n\n\t// Version 8\n\tm = append(m, steps{ExecuteSQLFile(\"008-rename-token-table.sql\")})\n\n\t// Version 9\n\tm = append(m, steps{ExecuteSQLFile(\"009-external-token-hard-delete.sql\")})\n\n\t// Version 10\n\tdefaultCluster := configuration.GetOpenShiftClientApiUrl()\n\tm = append(m, steps{ExecuteSQLFile(\"010-add-cluster-to-user.sql\", defaultCluster)})\n\n\t// Version 11\n\tm = append(m, steps{ExecuteSQLFile(\"011-add-username-to-external-token.sql\")})\n\n\t// Version 12\n\tm = append(m, steps{ExecuteSQLFile(\"012-hide-email.sql\")})\n\n\t// Version 13\n\tm = append(m, steps{ExecuteSQLFile(\"013-add-email-verified.sql\")})\n\n\t// Version 14\n\tm = append(m, steps{ExecuteSQLFile(\"014-add-user-feature-level.sql\")})\n\n\t// Version 15\n\tm = append(m, steps{ExecuteSQLFile(\"015-clear-resources-create-resource-types.sql\")})\n\n\t// Version 16\n\tm = append(m, steps{ExecuteSQLFile(\"016-add-state-to-auth-state-reference.sql\")})\n\n\t// Version 17\n\tm = append(m, steps{ExecuteSQLFile(\"017-feature-level-not-null.sql\")})\n\n\t// Version 18\n\tm = append(m, steps{ExecuteSQLFile(\"018-convert-user-feature-level.sql\")})\n\n\t// Version 19\n\tm = append(m, steps{ExecuteSQLFile(\"019-authorization-part-2.sql\")})\n\n\t// Version 20\n\tm = append(m, steps{ExecuteSQLFile(\"020-add-response-mode-to-auth-state-reference.sql\")})\n\n\t// Version 21\n\tm = append(m, steps{ExecuteSQLFile(\"021-organizations-list-create.sql\")})\n\n\t// Version 22\n\tm = append(m, steps{ExecuteSQLFile(\"022-add-deprovisioned-to-user.sql\")})\n\n\t// Version 23\n\tm = append(m, steps{ExecuteSQLFile(\"023-resource-type-index.sql\")})\n\n\t// Version 24\n\tm = append(m, steps{ExecuteSQLFile(\"024-role-mapping-and-team-and-group-identities.sql\")})\n\n\t// Version 25\n\tm = append(m, steps{ExecuteSQLFile(\"025-fix-feature-level.sql\")})\n\n\t// Version 26\n\tm = append(m, steps{ExecuteSQLFile(\"026-identities-users-indexes.sql\")})\n\n\t// Version 27\n\tm = append(m, steps{ExecuteSQLFile(\"027-invitations.sql\")})\n\n\t// Version 28\n\tm = append(m, steps{ExecuteSQLFile(\"028-make-organization-names-unique.sql\")})\n\n\t// Version 29\n\tm = append(m, steps{ExecuteSQLFile(\"029-add-space-resourcetype.sql\")})\n\n\t// Version 30\n\tm = append(m, steps{ExecuteSQLFile(\"030-add-team-admin-role.sql\")})\n\n\t// Version 31\n\tm = append(m, steps{ExecuteSQLFile(\"031-clean-up-roles-scopes.sql\")})\n\n\t// Version 32\n\tm = append(m, steps{ExecuteSQLFile(\"032-invitation-code.sql\")})\n\n\t// Version 33\n\tm = append(m, steps{ExecuteSQLFile(\"033-drop-space-resources.sql\")})\n\n\t// Version 34\n\tm = append(m, steps{ExecuteSQLFile(\"034-rename-token-table.sql\")})\n\n\t// Version 35\n\tm = append(m, steps{ExecuteSQLFile(\"035-unique_constraint_default_role_mapping.sql\")})\n\n\t// Version 36\n\tm = append(m, steps{ExecuteSQLFile(\"036-token-privileges.sql\")})\n\n\t// Version 37\n\tm = append(m, steps{ExecuteSQLFile(\"037-invitation-redirect-url.sql\")})\n\n\t// Version 38\n\tm = append(m, steps{ExecuteSQLFile(\"038-admin-console-resource.sql\")})\n\n\t// Version 39\n\tm = append(m, steps{ExecuteSQLFile(\"039-resource-type-alter.sql\")})\n\n\t// Version 40\n\tm = append(m, steps{ExecuteSQLFile(\"040-deferrable-constraints.sql\")})\n\n\t// Version 41\n\tm = append(m, steps{ExecuteSQLFile(\"041-identity-role-index.sql\")})\n\n\t// Version 42\n\tm = append(m, steps{ExecuteSQLFile(\"042-token-index.sql\")})\n\n\t// Version 43\n\tm = append(m, steps{ExecuteSQLFile(\"043-add-admin-console-resource.sql\")})\n\n\t// Version 44\n\tm = append(m, steps{ExecuteSQLFile(\"044-user-active.sql\")})\n\n\t// Version 45\n\tm = append(m, steps{ExecuteSQLFile(\"045-identity-last-active.sql\")})\n\n\t// Version 46\n\tm = append(m, steps{ExecuteSQLFile(\"046-identity-last-active-default.sql\")})\n\n\t// Version 47\n\tm = append(m, steps{ExecuteSQLFile(\"047-add-user-banned-column.sql\")})\n\n\t// Version 48\n\tm = append(m, steps{ExecuteSQLFile(\"048-identity-deactivation-notification.sql\")})\n\n\t// Version 49\n\tm = append(m, steps{ExecuteSQLFile(\"049-user-banned-index.sql\")})\n\n\t// Version 50\n\tm = append(m, steps{ExecuteSQLFile(\"050-worker-lock.sql\")})\n\n\t// Version 51\n\tm = append(m, steps{ExecuteSQLFile(\"051-identity-deactivation-scheduled.sql\")})\n\n\t// Version 52\n\tm = append(m, steps{ExecuteSQLFile(\"052-deferrable-constraints2.sql\")})\n\n\t// Version 53\n\tm = append(m, steps{ExecuteSQLFile(\"053-deactivation-indexes.sql\")})\n\n\t// Version 54\n\tm = append(m, steps{ExecuteSQLFile(\"054-cleanup-oauth-state-references.sql\")})\n\n\t// Version N\n\t//\n\t// In order to add an upgrade, simply append an array of MigrationFunc to the\n\t// the end of the \"migrations\" slice. The version numbers are determined by\n\t// the index in the array. The following code in comments show how you can\n\t// do a migration in 3 steps. If one of the steps fails, the others are not\n\t// executed.\n\t// If something goes wrong during the migration, all you need to do is return\n\t// an error that is not nil.\n\n\t/*\n\t\tm = append(m, steps{\n\t\t\tfunc(db *sql.Tx) error {\n\t\t\t\t// Execute random go code\n\t\t\t\treturn nil\n\t\t\t},\n\t\t\tExecuteSQLFile(\"YOUR_OWN_FILE.sql\"),\n\t\t\tfunc(db *sql.Tx) error {\n\t\t\t\t// Execute random go code\n\t\t\t\treturn nil\n\t\t\t},\n\t\t})\n\t*/\n\n\treturn m\n}", "func Migrate() {\n\tlog.Info(\"Executing migrations...\")\n\tDb.AutoMigrate(&models.Repository{}, &models.Week{}, &models.Contributor{})\n}", "func Migrate(db *sql.DB, catalog string, configuration MigrationConfiguration) error {\n\n\tvar err error\n\tif db == nil {\n\t\treturn errs.Errorf(\"Database handle is nil\\n\")\n\t}\n\n\tm := GetMigrations(configuration)\n\n\tvar tx *sql.Tx\n\tfor nextVersion := int64(0); nextVersion < int64(len(m)) && err == nil; nextVersion++ {\n\n\t\ttx, err = db.Begin()\n\t\tif err != nil {\n\t\t\treturn errs.Errorf(\"Failed to start transaction: %s\\n\", err)\n\t\t}\n\n\t\terr = MigrateToNextVersion(tx, &nextVersion, m, catalog)\n\n\t\tif err != nil {\n\t\t\toldErr := err\n\t\t\tlog.Info(nil, map[string]interface{}{\n\t\t\t\t\"next_version\": nextVersion,\n\t\t\t\t\"migrations\": m,\n\t\t\t\t\"err\": err,\n\t\t\t}, \"Rolling back transaction due to: %v\", err)\n\n\t\t\tif err = tx.Rollback(); err != nil {\n\t\t\t\tlog.Error(nil, map[string]interface{}{\n\t\t\t\t\t\"next_version\": nextVersion,\n\t\t\t\t\t\"migrations\": m,\n\t\t\t\t\t\"err\": err,\n\t\t\t\t}, \"error while rolling back transaction\")\n\t\t\t\treturn errs.Errorf(\"Error while rolling back transaction: %s\\n\", err)\n\t\t\t}\n\t\t\treturn oldErr\n\t\t}\n\n\t\tif err = tx.Commit(); err != nil {\n\t\t\tlog.Error(nil, map[string]interface{}{\n\t\t\t\t\"migrations\": m,\n\t\t\t\t\"err\": err,\n\t\t\t}, \"error during transaction commit: %v\", err)\n\t\t\treturn errs.Errorf(\"Error during transaction commit: %s\\n\", err)\n\t\t}\n\n\t}\n\n\tif err != nil {\n\t\tlog.Error(nil, map[string]interface{}{\n\t\t\t\"migrations\": m,\n\t\t\t\"err\": err,\n\t\t}, \"migration failed with error: %v\", err)\n\t\treturn errs.Errorf(\"Migration failed with error: %s\\n\", err)\n\t}\n\n\treturn nil\n}", "func TestMigrate(t *testing.T) {\n\tt.Parallel()\n\n\tcount := 100\n\n\tregisterTable := func() string {\n\t\ttype user struct {\n\t\t\tUsername string `ds:\"primary\"`\n\t\t\tEmail string `ds:\"unique\"`\n\t\t\tEnabled bool `ds:\"index\"`\n\t\t\tPassword string\n\t\t}\n\n\t\ttp := path.Join(tmpDir, randomString(12))\n\t\ttable, err := Register(user{}, tp, nil)\n\t\tif err != nil {\n\t\t\tt.Fatalf(\"Error registering table: %s\", err.Error())\n\t\t}\n\n\t\ti := 0\n\t\tfor i < count {\n\t\t\terr = table.Add(user{\n\t\t\t\tUsername: randomString(24),\n\t\t\t\tEmail: randomString(24),\n\t\t\t\tEnabled: true,\n\t\t\t\tPassword: randomString(24),\n\t\t\t})\n\t\t\tif err != nil {\n\t\t\t\tt.Fatalf(\"Error adding value to table: %s\", err.Error())\n\t\t\t}\n\t\t\ti++\n\t\t}\n\n\t\ttable.Close()\n\t\treturn tp\n\t}\n\n\ttablePath := registerTable()\n\n\ttype oldUser struct {\n\t\tUsername string `ds:\"primary\"`\n\t\tEmail string `ds:\"unique\"`\n\t\tEnabled bool `ds:\"index\"`\n\t\tPassword string\n\t}\n\ttype user struct {\n\t\tID string `ds:\"primary\"`\n\t\tUsername string `ds:\"unique\"`\n\t\tEmail string `ds:\"unique\"`\n\t\tEnabled bool `ds:\"index\"`\n\t\tPassword string\n\t}\n\n\tstats := Migrate(MigrateParams{\n\t\tTablePath: tablePath,\n\t\tOldType: oldUser{},\n\t\tNewType: user{},\n\t\tNewPath: tablePath,\n\t\tMigrateObject: func(o interface{}) (interface{}, error) {\n\t\t\told := o.(oldUser)\n\t\t\treturn user{\n\t\t\t\tID: randomString(24),\n\t\t\t\tUsername: old.Username,\n\t\t\t\tEmail: old.Email,\n\t\t\t\tEnabled: old.Enabled,\n\t\t\t\tPassword: old.Password,\n\t\t\t}, nil\n\t\t},\n\t})\n\tif stats.Error != nil {\n\t\tt.Errorf(\"Error migrating table: %s\", stats.Error)\n\t}\n\tif !stats.Success {\n\t\tt.Error(\"Migration not successful but error is nil\")\n\t}\n\tif stats.EntriesMigrated != uint(count) {\n\t\tt.Errorf(\"Not all entries migrated. Expected %d got %d\", count, stats.EntriesMigrated)\n\t}\n}", "func (m Migrator) RunMigrations() error {\n\t// Run everything in a transaction. In case of error, we can roll it back\n\ttx, err := m.Connection.Database.Begin()\n\tif err != nil {\n\t\t// Connection could not be started\n\t\treturn err\n\t}\n\n\t// First check if the database db_migrations exists\n\tres := tx.QueryRow(`SELECT EXISTS(\n\t\tSELECT *\n\t\tFROM information_schema.tables\n\t\tWHERE\n\t\t\ttable_schema = 'public' AND\n\t\t\ttable_name = 'db_migrations'\n\t)`)\n\n\tvar migTablePresent bool\n\terr = res.Scan(&migTablePresent)\n\tif err != nil {\n\t\t// result was invalid\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\n\talreadyRunMigrations := make(map[string]bool)\n\tif !migTablePresent {\n\t\t_, err = tx.Query(`\n\t\t\tCREATE TABLE db_migrations (version VARCHAR(50) NOT NULL, executed_at TIMESTAMP(0) NOT NULL DEFAULT CURRENT_TIMESTAMP, PRIMARY KEY(version))\n\t\t`)\n\t\tif err != nil {\n\t\t\t// could not create db_migration table\n\t\t\ttx.Rollback()\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tversionRows, err := tx.Query(`\n\t\t\tSELECT version FROM db_migrations\n\t\t`)\n\t\tif err != nil {\n\t\t\t// could not fetch the list of executed migrations\n\t\t\ttx.Rollback()\n\t\t\treturn err\n\t\t}\n\t\tfor versionRows.Next() {\n\t\t\tvar version string\n\t\t\terr = versionRows.Scan(&version)\n\t\t\tif err != nil {\n\t\t\t\t// A version number could not be parsed\n\t\t\t\ttx.Rollback()\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\talreadyRunMigrations[version] = true\n\t\t}\n\t}\n\n\tavailableMigrations, err := m.checkAvailableMigrations()\n\tif err != nil {\n\t\ttx.Rollback()\n\t\treturn err\n\t}\n\n\tvar migrationsToRun []string\n\tfor version := range availableMigrations {\n\t\tif _, ok := alreadyRunMigrations[version]; !ok {\n\t\t\tmigrationsToRun = append(migrationsToRun, version)\n\t\t}\n\t}\n\tfor version := range alreadyRunMigrations {\n\t\tif _, ok := availableMigrations[version]; !ok {\n\t\t\t// Warn there is a present migration with no corresponding file\n\t\t}\n\t}\n\n\tfor _, version := range migrationsToRun {\n\t\tmigrationByteContent, err := migration.Asset(fmt.Sprintf(\"%s_up.sql\", version))\n\t\tif err != nil {\n\t\t\ttx.Rollback()\n\t\t\treturn err\n\t\t}\n\t\tmigrationContent := string(migrationByteContent)\n\n\t\t_, err = tx.Query(migrationContent)\n\t\tif err != nil {\n\t\t\t// There was an error running the migration\n\t\t\ttx.Rollback()\n\t\t\treturn err\n\t\t}\n\t\t_, err = tx.Query(`INSERT INTO db_migrations (version) VALUES ($1)`, version)\n\t\tif err != nil {\n\t\t\t// There was an error running the migration\n\t\t\ttx.Rollback()\n\t\t\treturn err\n\t\t}\n\t}\n\n\ttx.Commit()\n\n\treturn nil\n}", "func autoMigrate(db *gorm.DB) {\n\tif os.Getenv(\"DROP_TABLES\") == \"yes\" {\n\t\t_ = db.Migrator().DropTable(&models.User{})\n\t\t_ = db.Migrator().DropTable(&models.Date{})\n\t\t_ = db.Migrator().DropTable(&models.DailyText{})\n\t}\n\n\tif os.Getenv(\"CREATE_TABLE\") == \"yes\" {\n\t\tif err := db.AutoMigrate(&models.User{}); err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t\tif err := db.AutoMigrate(&models.Date{}); err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t\tif err := db.AutoMigrate(&models.DailyText{}); err != nil {\n\t\t\tfmt.Println(err)\n\t\t}\n\t}\n\n}", "func RunMigrations(db *gorm.DB) {\n\tdb.Set(\"gorm:table_options\", \"ENGINE=InnoDB\")\n\tdb.Set(\"gorm:table_options\", \"collation_connection=utf8_general_ci\")\n\tdb.AutoMigrate(&model.User{}, &model.Todo{})\n}", "func (t TicketsDB) migrateDB(sqlFiles ...string) error {\n\tfor _, sqlFile := range sqlFiles {\n\t\tsetupScript, err := ioutil.ReadFile(sqlFile)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = t.DB.Exec(string(setupScript))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func Migrate(x *xorm.Engine) error {\n\tif err := x.Sync2(new(Version)); err != nil {\n\t\treturn fmt.Errorf(\"sync: %v\", err)\n\t}\n\n\tcurrentVersion := &Version{ID: 1}\n\thas, err := x.Get(currentVersion)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"get: %v\", err)\n\t} else if !has {\n\t\t// If the version record does not exist we think\n\t\t// it is a fresh installation and we can skip all migrations.\n\t\tcurrentVersion.ID = 0\n\t\tcurrentVersion.Version = int64(_MIN_DB_VER + len(migrations))\n\n\t\tif _, err = x.InsertOne(currentVersion); err != nil {\n\t\t\treturn fmt.Errorf(\"insert: %v\", err)\n\t\t}\n\t}\n\n\tv := currentVersion.Version\n\tif _MIN_DB_VER > v {\n\t\tlog.Print(`\nQmoon can not auto-migration from your previously installed version.\n`)\n\t\treturn nil\n\t}\n\n\tif int(v-_MIN_DB_VER) > len(migrations) {\n\t\tcurrentVersion.Version = int64(len(migrations) + _MIN_DB_VER)\n\t\t_, err = x.ID(1).Update(currentVersion)\n\t\treturn err\n\t}\n\tfor i, m := range migrations[v-_MIN_DB_VER:] {\n\t\tif err = m.Migrate(x); err != nil {\n\t\t\treturn fmt.Errorf(\"do migrate: %v\", err)\n\t\t}\n\t\tcurrentVersion.Version = v + int64(i) + 1\n\t\tif _, err = x.ID(1).Update(currentVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func Backfill(collection, target string) MigrationCollection {\n if (len(target) <= 0) {\n fmt.Println(\"ERROR: must supply (-f, --file) flag to backfill\")\n os.Exit(1)\n }\n\n if (strings.Index(collection, \".\") < 1) {\n fmt.Println(\"ERROR: backfill can only be used on {keyspace}.{table} items\")\n os.Exit(1)\n }\n\n // read target JSON\n var contents, err = ioutil.ReadFile(target)\n if (err != nil) {\n fmt.Printf(\"ERROR: could not read descriptor JSON:\\n%s\\n\\n\", err)\n os.Exit(1)\n }\n\n // unmarshal target JSON\n var targetJSON map[string]interface{}\n var jsonErr = json.Unmarshal(contents, &targetJSON)\n if (jsonErr != nil) {\n fmt.Printf(\"ERROR: could not parse descriptor JSON:\\n%s\\n\\n\", jsonErr)\n os.Exit(1)\n }\n\n // remove any comments of the suggested form\n delete(targetJSON, \"_\")\n\n // create the placeholder for the result migrations\n var migrations MigrationCollection\n\n // get existing table\n var parts = strings.Split(collection, \".\")\n var table, tblErr = db.Table(parts[0], parts[1])\n if (tblErr != nil) {\n if (tblErr.Error() == \"not found\") {\n migrations = CreateTableMigration(parts[0], parts[1], targetJSON)\n } else {\n os.Exit(1)\n }\n } else {\n migrations = BackfillTable(table, targetJSON)\n }\n\n return migrations\n}", "func HatcheryCmdMigration(store cache.Store, DBFunc func() *gorp.DbMap) {\n\tdb := DBFunc()\n\n\tlog.Info(\"HatcheryCmdMigration> Begin\")\n\n\twms, err := worker.LoadWorkerModels(db)\n\tif err != nil {\n\t\tlog.Warning(\"HatcheryCmdMigration> Cannot load worker models : %v\", err)\n\t\treturn\n\t}\n\n\tfor _, wmTmp := range wms {\n\t\tif wmTmp.ModelDocker.Cmd != \"\" || wmTmp.ModelVirtualMachine.Cmd != \"\" {\n\t\t\tcontinue\n\t\t}\n\t\ttx, errTx := db.Begin()\n\t\tif errTx != nil {\n\t\t\tlog.Warning(\"HatcheryCmdMigration> cannot create a transaction : %v\", errTx)\n\t\t\tcontinue\n\t\t}\n\t\twm, errL := worker.LoadAndLockWorkerModelByID(tx, wmTmp.ID)\n\t\tif errL != nil {\n\t\t\tlog.Warning(\"HatcheryCmdMigration> cannot load and lock a worker model : %v\", errL)\n\t\t\ttx.Rollback()\n\t\t\tcontinue\n\t\t}\n\n\t\tswitch wm.Type {\n\t\tcase sdk.Docker:\n\t\t\tif wm.ModelDocker.Image != \"\" && wm.ModelDocker.Cmd != \"\" {\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\twm.ModelDocker = sdk.ModelDocker{\n\t\t\t\tImage: wm.Image,\n\t\t\t\tShell: \"sh -c\",\n\t\t\t\tCmd: \"curl {{.API}}/download/worker/linux/$(uname -m) -o worker --retry 10 --retry-max-time 120 && chmod +x worker && exec ./worker\",\n\t\t\t}\n\t\tcase sdk.Openstack:\n\t\t\tvar osdata deprecatedOpenstackModelData\n\t\t\tif wm.ModelVirtualMachine.Image != \"\" && wm.ModelVirtualMachine.Cmd != \"\" {\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif wm.Image == \"\" {\n\t\t\t\tlog.Warning(\"HatcheryCmdMigration> worker model image field is empty for %s\", wm.Name)\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err := json.Unmarshal([]byte(wm.Image), &osdata); err != nil {\n\t\t\t\tlog.Warning(\"HatcheryCmdMigration> cannot unmarshal image field is empty for %s : %v\", wm.Name, err)\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tpreCmd := `#!/bin/bash\nset +e\nexport CDS_FROM_WORKER_IMAGE={{.FromWorkerImage}}\nexport CDS_SINGLE_USE=1\nexport CDS_FORCE_EXIT=1\nexport CDS_API={{.API}}\nexport CDS_TOKEN={{.Token}}\nexport CDS_NAME={{.Name}}\nexport CDS_MODEL={{.Model}}\nexport CDS_HATCHERY={{.Hatchery}}\nexport CDS_HATCHERY_NAME={{.HatcheryName}}\nexport CDS_BOOKED_PB_JOB_ID={{.PipelineBuildJobID}}\nexport CDS_BOOKED_WORKFLOW_JOB_ID={{.WorkflowJobID}}\nexport CDS_TTL={{.TTL}}\nexport CDS_GRAYLOG_HOST={{.GraylogHost}}\nexport CDS_GRAYLOG_PORT={{.GraylogPort}}\nexport CDS_GRAYLOG_EXTRA_KEY={{.GraylogExtraKey}}\nexport CDS_GRAYLOG_EXTRA_VALUE={{.GraylogExtraValue}}\n#export CDS_GRPC_API={{.GrpcAPI}}\n#export CDS_GRPC_INSECURE={{.GrpcInsecure}}\nexport CDS_INSECURE={{.HTTPInsecure}}\n`\n\t\t\tuserdata, errD := base64.StdEncoding.DecodeString(osdata.UserData)\n\t\t\tif errD != nil {\n\t\t\t\tlog.Warning(\"HatcheryCmdMigration> cannot decode base64 image field for %s : %v\", wm.Name, errD)\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tpreCmd += string(userdata)\n\n\t\t\tpreCmd += `\n\t\t\tcurl -L \"{{.API}}/download/worker/linux/$(uname -m)\" -o worker --retry 10 --retry-max-time 120\n\t\t\tchmod +x worker\n\t\t\t`\n\t\t\twm.ModelVirtualMachine = sdk.ModelVirtualMachine{\n\t\t\t\tFlavor: osdata.Flavor,\n\t\t\t\tImage: osdata.Image,\n\t\t\t\tPreCmd: preCmd,\n\t\t\t\tCmd: \"./worker\",\n\t\t\t\tPostCmd: \"sudo shutdown -h now\",\n\t\t\t}\n\n\t\tcase sdk.VSphere:\n\t\t\tvar vspheredata deprecatedVSphereModelData\n\t\t\tif wm.ModelVirtualMachine.Image != \"\" && wm.ModelVirtualMachine.Cmd != \"\" {\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif wm.Image == \"\" {\n\t\t\t\tlog.Warning(\"HatcheryCmdMigration> worker model image field is empty for %s\", wm.Name)\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err := json.Unmarshal([]byte(wm.Image), &vspheredata); err != nil {\n\t\t\t\tlog.Warning(\"HatcheryCmdMigration> cannot unmarshal image field is empty for %s : %v\", wm.Name, err)\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tpreCmd := `#!/bin/bash\nset +e\nexport CDS_FROM_WORKER_IMAGE={{.FromWorkerImage}}\nexport CDS_SINGLE_USE=1\nexport CDS_FORCE_EXIT=1\nexport CDS_API={{.API}}\nexport CDS_TOKEN={{.Token}}\nexport CDS_NAME={{.Name}}\nexport CDS_MODEL={{.Model}}\nexport CDS_HATCHERY={{.Hatchery}}\nexport CDS_HATCHERY_NAME={{.HatcheryName}}\nexport CDS_BOOKED_PB_JOB_ID={{.PipelineBuildJobID}}\nexport CDS_BOOKED_WORKFLOW_JOB_ID={{.WorkflowJobID}}\nexport CDS_TTL={{.TTL}}\nexport CDS_INSECURE={{.HTTPInsecure}}\nexport CDS_GRAYLOG_HOST={{.GraylogHost}}\nexport CDS_GRAYLOG_PORT={{.GraylogPort}}\nexport CDS_GRAYLOG_EXTRA_KEY={{.GraylogExtraKey}}\nexport CDS_GRAYLOG_EXTRA_VALUE={{.GraylogExtraValue}}\n#export CDS_GRPC_API={{.GrpcAPI}}\n#export CDS_GRPC_INSECURE={{.GrpcInsecure}}\n\ncurl -L \"{{.API}}/download/worker/linux/$(uname -m)\" -o worker --retry 10 --retry-max-time 120\nchmod +x worker\n`\n\t\t\twm.ModelVirtualMachine = sdk.ModelVirtualMachine{\n\t\t\t\tImage: vspheredata.OS,\n\t\t\t\tPreCmd: preCmd + vspheredata.UserData,\n\t\t\t\tCmd: \"PATH=$PATH ./worker\",\n\t\t\t\tPostCmd: \"shutdown -h now\",\n\t\t\t}\n\t\tcase sdk.HostProcess:\n\t\t\tif wm.ModelVirtualMachine.Image != \"\" && wm.ModelVirtualMachine.Cmd != \"\" {\n\t\t\t\ttx.Rollback()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\twm.ModelVirtualMachine = sdk.ModelVirtualMachine{\n\t\t\t\tImage: wm.Name,\n\t\t\t\tCmd: \"worker --api={{.API}} --token={{.Token}} --basedir={{.BaseDir}} --model={{.Model}} --name={{.Name}} --hatchery={{.Hatchery}} --hatchery-name={{.HatcheryName}} --insecure={{.HTTPInsecure}} --graylog-extra-key={{.GraylogExtraKey}} --graylog-extra-value={{.GraylogExtraValue}} --graylog-host={{.GraylogHost}} --graylog-port={{.GraylogPort}} --single-use --force-exit\",\n\t\t\t}\n\t\t}\n\n\t\tif err := worker.UpdateWorkerModelWithoutRegistration(tx, *wm); err != nil {\n\t\t\tlog.Warning(\"HatcheryCmdMigration> cannot update worker model %s : %v\", wm.Name, err)\n\t\t\ttx.Rollback()\n\t\t\tcontinue\n\t\t}\n\n\t\tif err := tx.Commit(); err != nil {\n\t\t\tlog.Warning(\"HatcheryCmdMigration> cannot commit tx for worker model %s : %v\", wm.Name, err)\n\t\t\ttx.Rollback()\n\t\t}\n\t}\n\n\tlog.Info(\"HatcheryCmdMigration> Done\")\n}", "func (q *Bucket) Batch(fn func(backend.Batch) error) error {\n\treturn q.db.boltDB.Update(func(tx *bolt.Tx) error {\n\t\tbucket, err := tx.CreateBucketIfNotExists([]byte(q.name))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tbatch := &Batch{\n\t\t\tbucket: bucket,\n\t\t}\n\t\treturn fn(batch)\n\t})\n}", "func migrateNew() {\n\tif len(flag.Args()) == 0 {\n\t\tlog.Fatalf(\"ERROR: Migration's name not specified\\n\")\n\t\treturn\n\t}\n\twd, _ := os.Getwd()\n\tname := path.Join(wd, \"migrations\", fmt.Sprintf(\"%d_%s.sql\", time.Now().Unix(), flag.Arg(0)))\n\tfile, err := os.Create(name)\n\tif err != nil {\n\t\tlog.Printf(\"ERROR: %s\\n\", err)\n\t\treturn\n\t}\n\tfmt.Fprintf(file, \"-- +migrate Up\\n\")\n\tfmt.Fprintf(file, \"-- SQL in section 'Up' is executed when this migration is applied\\n\\n\\n\")\n\tfmt.Fprintf(file, \"-- +migrate Down\\n\")\n\tfmt.Fprintf(file, \"-- SQL in section 'Down' is executed when this migration is rolled back\\n\\n\\n\")\n\terr = file.Close()\n\tif err != nil {\n\t\tlog.Printf(\"ERROR: %s\\n\", err)\n\t} else {\n\t\tlog.Printf(\"INFO: File %s has been successfully created\\n\", name)\n\t}\n}", "func (db *MemoryStorage) EnableBatch() {\n}", "func Sync(c *SQLConnection) (done []string, err error) {\n\tcount, err := migrate.ExecMax(c.db, c.dialect, c.source, migrate.Up, 0)\n\tif err != nil {\n\t\treturn\n\t}\n\tif count == 0 {\n\t\treturn\n\t}\n\n\trecords, err := migrate.GetMigrationRecords(c.db, c.dialect)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor i := len(records) - count; i < len(records); i++ {\n\t\tdone = append(done, records[i].Id)\n\t}\n\treturn\n}", "func Migrate(db *gorm.DB) {\n\tdb.AutoMigrate(&User{})\n\tfmt.Println(\"Auto migration has been completed\")\n}", "func RunMigrations(db DB, migrations []Migration, a ...string) (oldVersion, newVersion int64, err error) {\n\tsortMigrations(migrations)\n\n\tvar cmd string\n\tif len(a) > 0 {\n\t\tcmd = a[0]\n\t}\n\n\terr = createTables(db)\n\tif err != nil {\n\t\treturn\n\t}\n\n\toldVersion, err = Version(db)\n\tif err != nil {\n\t\treturn\n\t}\n\tnewVersion = oldVersion\n\n\tswitch cmd {\n\tcase \"create\":\n\t\tif len(a) < 2 {\n\t\t\tfmt.Println(\"Please enter migration description\")\n\t\t\treturn\n\t\t}\n\n\t\tvar version int64\n\t\tif len(migrations) > 0 {\n\t\t\tversion = migrations[len(migrations)-1].Version\n\t\t}\n\n\t\tfilename := fmtMigrationFilename(version+1, strings.Join(a[1:], \"_\"))\n\t\terr = createMigrationFile(filename)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\n\t\tfmt.Println(\"created migration\", filename)\n\t\treturn\n\tcase \"version\":\n\t\treturn\n\tcase \"up\", \"\":\n\t\tfor i := range migrations {\n\t\t\tm := &migrations[i]\n\t\t\tif m.Version <= oldVersion {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\terr = m.Up(db)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tnewVersion = m.Version\n\t\t\terr = SetVersion(db, newVersion)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\treturn\n\tcase \"down\":\n\t\tnewVersion, err = down(db, migrations, oldVersion)\n\t\treturn\n\tcase \"reset\":\n\t\tversion := oldVersion\n\t\tfor {\n\t\t\tnewVersion, err = down(db, migrations, version)\n\t\t\tif err != nil || newVersion == version {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tversion = newVersion\n\t\t}\n\tcase \"set_version\":\n\t\tif len(a) < 2 {\n\t\t\terr = fmt.Errorf(\"set_version requires version as 2nd arg, e.g. set_version 42\")\n\t\t\treturn\n\t\t}\n\n\t\tnewVersion, err = strconv.ParseInt(a[1], 10, 64)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\terr = SetVersion(db, newVersion)\n\t\treturn\n\tdefault:\n\t\terr = fmt.Errorf(\"unsupported command: %q\", cmd)\n\t\treturn\n\t}\n}", "func init() {\n\tmigrations.AddPluginMigration(\"inputs.io\", migrate)\n}", "func Initialize(ctx context.Context, migrationDir string, db Database, queries QueryAdapter) error {\n\t// make sure migrations can persist\n\t_, err := db.ExecContext(ctx, queries.CreateTableSQL())\n\tif err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\t// Get the migration with the highest ID number\n\trow := db.QueryRowContext(ctx, queries.SelectLatestMigrationSQL())\n\tif err := row.Err(); err != nil {\n\t\treturn errors.WithStack(err)\n\t}\n\tlatest := 0\n\n\tif err := row.Scan(&latest); err != nil {\n\t\tif err != sql.ErrNoRows {\n\t\t\treturn errors.WithStack(err)\n\t\t}\n\t}\n\n\t// Load up the migrations and run starting from the last known run migration\n\tmigs := fromDir(migrationDir)\n\tfor _, migration := range migs[:latest] {\n\t\tfmt.Printf(\"Already ran migration: %d\\n\", migration.Order)\n\t}\n\n\tfor _, migration := range migs[latest:] {\n\t\tfmt.Printf(\"Running migration %d\\n\", migration.Order)\n\t\tif _, err := db.ExecContext(ctx, migration.Up); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif _, err := db.ExecContext(ctx, queries.InsertMigrationSQL(), migration.Order); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func AutoMigrate() {\n\tdb := utils.GetDB()\n\n\tdb.AutoMigrate(&CourseModel{})\n\t//db.AutoMigrate(&HomeworkModel{})\n\t//db.AutoMigrate(&CourseUserModel{})\n}", "func (a *ActivityTypeMigrator) Migrate() {\n\tfor _, d := range a.DeprecatedActivityTypes {\n\t\tif a.isDeprecated(d.Domain, d.ActivityType.Name, d.ActivityType.Version) {\n\t\t\tLog.Printf(\"action=migrate at=deprecate-activity domain=%s activity=%s version=%s status=previously-deprecated\", LS(d.Domain), LS(d.ActivityType.Name), LS(d.ActivityType.Version))\n\t\t} else {\n\t\t\ta.deprecate(d)\n\t\t\tLog.Printf(\"action=migrate at=depreacate-activity domain=%s activity=%s version=%s status=deprecated\", LS(d.Domain), LS(d.ActivityType.Name), LS(d.ActivityType.Version))\n\t\t}\n\t}\n\tfor _, r := range a.RegisteredActivityTypes {\n\t\tif a.isRegisteredNotDeprecated(r) {\n\t\t\tLog.Printf(\"action=migrate at=register-activity domain=%s activity=%s version=%s status=previously-registered\", LS(r.Domain), LS(r.Name), LS(r.Version))\n\t\t} else {\n\t\t\ta.register(r)\n\t\t\tLog.Printf(\"action=migrate at=register-activity domain=%s activity=%s version=%s status=registered\", LS(r.Domain), LS(r.Name), LS(r.Version))\n\t\t}\n\t}\n}", "func testMigrationBase(t *testing.T) {\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\n\t// Ensure we work on a clean copy of the provided repo\n\tcopyRepo(t)\n\n\topts := migrate.Options{\n\t\tVerbose: true,\n\t\tFlags: migrate.Flags{\n\t\t\tPath: workingRepo,\n\t\t\tRevert: false,\n\t\t\tVerbose: true,\n\t\t},\n\t}\n\tm := Migration{}\n\terr := m.open(opts)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Print out the blocks before migrating (for debugging)\n\torigBlocks := blocks(t, m.dstore)\n\tt.Logf(\"Blocks pre-migration: %d\", len(origBlocks))\n\tfor _, origb := range origBlocks {\n\t\tk := ds.NewKey(origb)\n\t\tc, err := dsKeyToCid(ds.NewKey(k.BaseNamespace()))\n\t\tif err != nil {\n\t\t\tt.Error(\"block before migration cannot be parsed\")\n\t\t}\n\t\tt.Log(origb, \" -> \", c)\n\t}\n\n\t// Apply the migration\n\terr = m.Apply(opts)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Re-open store, as Apply() closes things.\n\terr = m.open(opts)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Obtain blocks post-migration\n\tpostBlocks := blocks(t, m.dstore)\n\n\tt.Logf(\"Blocks post-migration: %d\", len(postBlocks))\n\n\t// All blocks should be parseable.\n\t// We do not check what there is or not. We will just\n\t// expect that they are all CIDv0 and that revert\n\t// will put everything in its place.\n\tfor _, postb := range postBlocks {\n\t\tk := ds.NewKey(postb)\n\t\tc, err := dsKeyToCid(ds.NewKey(k.BaseNamespace()))\n\t\tif err != nil {\n\t\t\tt.Error(\"key after migration cannot be parsed\")\n\t\t}\n\t\tt.Log(postb, \" -> \", c)\n\n\t\tif c.Version() != 0 {\n\t\t\tt.Error(\"CidV1 keys left after migration\")\n\t\t}\n\t}\n\n\t// Set MFS root to an CIDv1 corresponding to folder b. This means\n\t// this block will need to be created in CIDv1 fashion after\n\t// revert. Note, it is not the same as the one from the B-folder-tree\n\t// added with CidV1 as that one points to a CIDv1 link, while this one\n\t// has a CIDv0 link inside.\n\tcidV1FromB, err := cid.Decode(\"bafybeie4pduk2uwvr5dq36wnbhxspgox7dtqo3fprri4r2wpa7vrej5jqq\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\twriteMFSRoot(t, m.dstore, cidV1FromB)\n\n\t// Pin a CIDv1 corresponding to folder C which was added with CIDv0\n\t// only. This means, we will have to create this block in\n\t// CIDv1-addressed fashion on revert.\n\tcidV1FromC, err := cid.Decode(\"bafybeicgaywc7bcz5dlkbvvjbm7d2epi3wm5yilfuuex6llrxyzr2mlhba\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tpinner, dags, err := getPinner(ctx, m.dstore)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// We need a node for pinning\n\tnd, err := dags.Get(ctx, cidV1FromC)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Pin the node.\n\terr = pinner.Pin(ctx, nd, true)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\terr = pinner.Flush(ctx)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\t// Now revert the migration.\n\topts.Revert = true\n\terr = m.Revert(opts)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\trevertedBlocks := blocks(t, m.dstore)\n\tt.Logf(\"Blocks post-revert: %d\", len(revertedBlocks))\n\n\t// Check all the original blocks are in the reverted version\n\tfor _, origB := range origBlocks {\n\t\tfound := false\n\t\tfor _, revB := range revertedBlocks {\n\t\t\tif origB == revB {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tt.Errorf(\"block %s not found after revert\", origB)\n\t\t}\n\t}\n\n\t// Print the list of reverted blocks for debugging.\n\tfor _, revB := range revertedBlocks {\n\t\tk := ds.NewKey(revB)\n\t\tc, err := dsKeyToCid(ds.NewKey(k.BaseNamespace()))\n\t\tif err != nil {\n\t\t\tt.Error(\"key after revert cannot be parsed\")\n\t\t}\n\t\tt.Log(revB, \" -> \", c)\n\t}\n\n\t// Re-open, as Apply() closes things.\n\terr = m.open(opts)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\tdefer m.dstore.Close()\n\n\t// Check that the CIDv1s that we explicitally pinned or\n\t// added to MFS are now retrievable as CIDv1-addressed nodes.\n\tpostMigrationCids := []cid.Cid{\n\t\tcidV1FromB,\n\t\tcidV1FromC,\n\t}\n\n\tfor _, c := range postMigrationCids {\n\t\tk := blocksPrefix.Child(cidToDsKey(c))\n\t\t_, err := m.dstore.Get(k)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"expected key %s not found\", k)\n\t\t}\n\t}\n}", "func forward(c *cli.Context, name string) error {\n\tdb, err := pomegranate.Connect(c.String(\"dburl\"))\n\tif err != nil {\n\t\treturn cli.NewExitError(err, 1)\n\t}\n\tdir := c.String(\"dir\")\n\tallMigrations, err := pomegranate.ReadMigrationFiles(dir)\n\tif err != nil {\n\t\treturn cli.NewExitError(err, 1)\n\t}\n\terr = pomegranate.MigrateForwardTo(name, db, allMigrations, true)\n\tif err != nil {\n\t\treturn cli.NewExitError(err, 1)\n\t}\n\tfmt.Println(\"Done\")\n\treturn nil\n}", "func RunMigrations(db *gorm.DB) {\n\tdb.AutoMigrate(&models.Article{})\n\tdb.AutoMigrate(&models.Page{})\n}", "func SaveAllMigrations(db gorp.SqlExecutor) error {\n\tfor _, migration := range migrations {\n\t\tmigration.Done = time.Now()\n\t\tmigration.Status = sdk.MigrationStatusDone\n\t\tmigration.Progress = \"Done because it was a fresh installation\"\n\t\tif err := Insert(db, &migration); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (s Service) Apply(migration *Migration, save bool) error {\n\tif migration == nil {\n\t\treturn ErrNilMigration\n\t}\n\n\tfor _, statement := range migration.Up() {\n\t\tif s.dryRun {\n\t\t\tlogger.Info(\"statement: %s\", statement)\n\t\t\tcontinue\n\t\t}\n\n\t\terr := s.conn.Execute(statement)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to apply migration %s: %s\\n%s\", migration.Name, err, statement)\n\t\t}\n\t}\n\n\tif s.dryRun {\n\t\treturn nil\n\t}\n\n\tif !save {\n\t\treturn nil\n\t}\n\n\terr := s.conn.AddApplied(migration.Name)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to mark migration %s as applied: %s\", migration.Name, err)\n\t}\n\n\treturn nil\n}", "func Migrate(db *sql.DB) error {\n\tif err := createTable(db); err != nil {\n\t\treturn err\n\t}\n\tcompleted, err := selectCompleted(db)\n\tif err != nil && err != sql.ErrNoRows {\n\t\treturn err\n\t}\n\tfor _, migration := range migrations {\n\t\tif _, ok := completed[migration.name]; ok {\n\n\t\t\tcontinue\n\t\t}\n\n\t\tif _, err := db.Exec(migration.stmt); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := insertMigration(db, migration.name); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\tif err := initTemplete(db); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (dbWrapper *DB) migrate() (err error) {\n\n\tdb := dbWrapper.gormDB\n\n\tif !db.HasTable(&MamidMetadata{}) {\n\n\t\t// run the populating query\n\n\t\tddlStatements, err := Asset(\"model/sql/mamid_postgresql.sql\")\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"sql DDL data not found: %s\", err)\n\t\t}\n\n\t\terr = db.Exec(string(ddlStatements), []interface{}{}).Error\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error running DDL statements: %s\", err)\n\t\t}\n\n\t\t// persist schema version\n\t\tif err = dbWrapper.setMetadata(\"schema_version\", SCHEMA_VERSION); err != nil {\n\t\t\treturn fmt.Errorf(\"error setting schema version: %s\", err)\n\t\t}\n\n\t} else {\n\n\t\tversion, err := dbWrapper.schemaVersion()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"error determining schema version: %s\", err)\n\t\t}\n\n\t\tif version != SCHEMA_VERSION {\n\t\t\treturn fmt.Errorf(\"the database has already been populated, migrations are not supported\")\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func (pc Controllers) AutoMigration(c *gin.Context) {\n\tmysql.AutoMigration()\n}", "func PostAutoMigrate(db *gorm.DB) error {\n\t// These types don't apply for sqlite -- just mysql.\n\tif db.Dialect().GetName() == mySQLDialect {\n\t\tdb.Model(&Invocation{}).ModifyColumn(\"pattern\", \"text\")\n\t\tdb.Model(&Execution{}).ModifyColumn(\"serialized_operation\", \"text\")\n\t}\n\treturn nil\n}", "func BatchPlacements(ctx iris.Context) {\n\tvar req models.Placements\n\tif err := ctx.ReadJSON(&req); err != nil {\n\t\tctx.StatusCode(http.StatusInternalServerError)\n\t\tctx.JSON(jsonError{\"Batch de stages, décodage : \" + err.Error()})\n\t\treturn\n\t}\n\tdb := ctx.Values().Get(\"db\").(*sql.DB)\n\tif err := req.Save(db); err != nil {\n\t\tctx.StatusCode(http.StatusInternalServerError)\n\t\tctx.JSON(jsonError{\"Batch de stages, requête : \" + err.Error()})\n\t\treturn\n\t}\n\tctx.StatusCode(http.StatusOK)\n\tctx.JSON(jsonMessage{\"Batch de stages importé\"})\n}", "func SetupAndMigrate(\n\tctx context.Context,\n\tlog logrus.FieldLogger,\n\tdb interface {\n\t\tBeginTx(context.Context, pgx.TxOptions) (pgx.Tx, error)\n\t\tExec(context.Context, string, ...any) (pgconn.CommandTag, error)\n\t},\n\ttableName string,\n\tschemas []string,\n) error {\n\ttableName = pgx.Identifier{tableName}.Sanitize()\n\n\tvar version int32\n\tvar migrateErr error\n\n\t// this is split off from the rest because we might not have permissions to\n\t// CREATE TABLE, which is checked even if the table exists\n\tif _, err := RetryIdempotent(ctx, log, func() (struct{}, error) {\n\t\t_, err := db.Exec(ctx,\n\t\t\tfmt.Sprintf(`CREATE TABLE IF NOT EXISTS %v (\n\t\t\t\tversion integer PRIMARY KEY CHECK (version > 0),\n\t\t\t\tcreated timestamptz NOT NULL DEFAULT now()\n\t\t\t)`, tableName), pgx.QueryExecModeExec,\n\t\t)\n\t\treturn struct{}{}, trace.Wrap(err)\n\t}); err != nil {\n\t\t// the very first SELECT in the next transaction will fail, we don't\n\t\t// need anything higher than debug here\n\t\tlog.WithError(err).Debugf(\"Failed to confirm the existence of the %v table.\", tableName)\n\t}\n\n\tconst idempotent = true\n\tif err := RetryTx(ctx, log, db, pgx.TxOptions{\n\t\tIsoLevel: pgx.Serializable,\n\t\tAccessMode: pgx.ReadWrite,\n\t}, idempotent, func(tx pgx.Tx) error {\n\t\tif err := tx.QueryRow(ctx,\n\t\t\tfmt.Sprintf(\"SELECT COALESCE(max(version), 0) FROM %v\", tableName),\n\t\t\tpgx.QueryExecModeExec,\n\t\t).Scan(&version); err != nil {\n\t\t\treturn trace.Wrap(err)\n\t\t}\n\n\t\tif int(version) > len(schemas) {\n\t\t\tmigrateErr = trace.BadParameter(\"unsupported schema version %v\", version)\n\t\t\t// the transaction succeeded, the error is outside of the transaction\n\t\t\treturn nil\n\t\t}\n\n\t\tif int(version) == len(schemas) {\n\t\t\treturn nil\n\t\t}\n\n\t\tfor _, s := range schemas[version:] {\n\t\t\tif _, err := tx.Exec(ctx, s, pgx.QueryExecModeExec); err != nil {\n\t\t\t\treturn trace.Wrap(err)\n\t\t\t}\n\t\t}\n\n\t\tif _, err := tx.Exec(ctx,\n\t\t\tfmt.Sprintf(\"INSERT INTO %v (version) VALUES ($1)\", tableName),\n\t\t\tpgx.QueryExecModeExec, len(schemas),\n\t\t); err != nil {\n\t\t\treturn trace.Wrap(err)\n\t\t}\n\n\t\treturn nil\n\t}); err != nil {\n\t\treturn trace.Wrap(err)\n\t}\n\n\tif migrateErr != nil {\n\t\treturn trace.Wrap(migrateErr)\n\t}\n\n\tif int(version) != len(schemas) {\n\t\tlog.WithFields(logrus.Fields{\n\t\t\t\"previous_version\": version,\n\t\t\t\"current_version\": len(schemas),\n\t\t}).Info(\"Migrated database schema.\")\n\t}\n\n\treturn nil\n}", "func Register(namespace string, migration Migration) {\n\tif _, ok := namespacedMigrations[namespace]; !ok {\n\t\tnamespacedMigrations[namespace] = make(Migrations)\n\t}\n\n\tnamespacedMigrations[namespace][migration.Version] = migration\n}", "func RunMigrations(command *string) {\n\tswitch *command {\n\tcase \"new\":\n\t\tmigrateNew()\n\t\tos.Exit(0)\n\tcase \"up\":\n\t\tmigrateUp(models.GetDB(), 0)\n\t\tos.Exit(0)\n\tcase \"down\":\n\t\tmigrateDown(models.GetDB(), 1)\n\t\tos.Exit(0)\n\tcase \"redo\":\n\t\tmigrateDown(models.GetDB(), 1)\n\t\tmigrateUp(models.GetDB(), 1)\n\t\tos.Exit(0)\n\tcase \"skip\":\n\tdefault:\n\t\tlog.Fatalf(\"Wrong migration flag %q, acceptable values: up, down\\n\", *command)\n\t}\n}", "func Migrate(x *xorm.Engine, db *gorm.DB) error {\n\tif err := x.Sync(new(Version)); err != nil {\n\t\treturn fmt.Errorf(\"sync: %v\", err)\n\t}\n\n\tcurrentVersion := &Version{ID: 1}\n\thas, err := x.Get(currentVersion)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"get: %v\", err)\n\t} else if !has {\n\t\t// If the version record does not exist we think\n\t\t// it is a fresh installation and we can skip all migrations.\n\t\tcurrentVersion.ID = 0\n\t\tcurrentVersion.Version = int64(minDBVersion + len(migrations))\n\n\t\tif _, err = x.InsertOne(currentVersion); err != nil {\n\t\t\treturn fmt.Errorf(\"insert: %v\", err)\n\t\t}\n\t}\n\n\tv := currentVersion.Version\n\tif minDBVersion > v {\n\t\tlog.Fatal(`\nHi there, thank you for using Gogs for so long!\nHowever, Gogs has stopped supporting auto-migration from your previously installed version.\nBut the good news is, it's very easy to fix this problem!\nYou can migrate your older database using a previous release, then you can upgrade to the newest version.\n\nPlease save following instructions to somewhere and start working:\n\n- If you were using below 0.6.0 (e.g. 0.5.x), download last supported archive from following link:\n\thttps://gogs.io/gogs/releases/tag/v0.7.33\n- If you were using below 0.7.0 (e.g. 0.6.x), download last supported archive from following link:\n\thttps://gogs.io/gogs/releases/tag/v0.9.141\n- If you were using below 0.11.55 (e.g. 0.9.141), download last supported archive from following link:\n\thttps://gogs.io/gogs/releases/tag/v0.12.0\n\nOnce finished downloading:\n\n1. Extract the archive and to upgrade steps as usual.\n2. Run it once. To verify, you should see some migration traces.\n3. Once it starts web server successfully, stop it.\n4. Now it's time to put back the release archive you originally intent to upgrade.\n5. Enjoy!\n\nIn case you're stilling getting this notice, go through instructions again until it disappears.`)\n\t\treturn nil\n\t}\n\n\tif int(v-minDBVersion) > len(migrations) {\n\t\t// User downgraded Gogs.\n\t\tcurrentVersion.Version = int64(len(migrations) + minDBVersion)\n\t\t_, err = x.Id(1).Update(currentVersion)\n\t\treturn err\n\t}\n\tfor i, m := range migrations[v-minDBVersion:] {\n\t\tlog.Info(\"Migration: %s\", m.Description())\n\t\tif err = m.Migrate(db); err != nil {\n\t\t\treturn fmt.Errorf(\"do migrate: %v\", err)\n\t\t}\n\t\tcurrentVersion.Version = v + int64(i) + 1\n\t\tif _, err = x.Id(1).Update(currentVersion); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SchemaEditor) Migrate() (err error) {\n\tglog.V(2).Infof(\"patrol: running migrate.\")\n\n\tvar count int\n\tif count, err = s.PendingMigrations(); err != nil {\n\t\treturn\n\t}\n\tif count == 0 {\n\t\treturn fmt.Errorf(\"no pending migrations\")\n\t}\n\n\tif err = s.prepareSchemaMigrations(); err != nil {\n\t\treturn\n\t}\n\n\tresolver := NewSchemaDependenciesResolver(s.pr)\n\titems, err := resolver.Resolve()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, item := range items {\n\t\tisApplied, errApplied := s.IsAppliedMigration(item.migration, item.plugin)\n\t\tif errApplied != nil {\n\t\t\treturn errApplied\n\t\t}\n\n\t\tfmt.Printf(\"migration %s:%s \", yellow(item.plugin.ID()), yellow(item.migration.ID()))\n\n\t\tif isApplied {\n\t\t\tfmt.Println(\"already applied.\")\n\t\t\tcontinue\n\t\t}\n\n\t\t// process migration\n\t\tif err := s.ProcessMigration(item.migration, item.plugin); err != nil {\n\t\t\tfmt.Printf(red(\"... Error %s\\n\"), err.Error())\n\t\t\treturn err\n\t\t}\n\t\tfmt.Println(green(\"... OK\"))\n\n\t\tfmt.Printf(\" post migrate %s:%s \", item.plugin.ID(), item.migration.ID())\n\t\tif err := item.migration.PostMigrate(s.context); err != nil {\n\t\t\tfmt.Printf(red(\"...failed. %s\\n\"), err.Error())\n\t\t}\n\t\tfmt.Println(green(\"... OK\"))\n\t}\n\n\treturn nil\n}", "func (this *Migrate) Run(args []string) (exitCode int) {\n\tvar how bool\n\tctx := base.GetMigrationContext()\n\tcmdFlags := flag.NewFlagSet(\"migrate\", flag.ContinueOnError)\n\tcmdFlags.StringVar(&ctx.InspectorConnectionConfig.Key.Hostname, \"host\", \"127.0.0.1\", \"\")\n\tcmdFlags.IntVar(&ctx.InspectorConnectionConfig.Key.Port, \"port\", 3306, \"\")\n\tcmdFlags.StringVar(&ctx.DatabaseName, \"db\", \"\", \"\")\n\tcmdFlags.StringVar(&ctx.OriginalTableName, \"table\", \"\", \"\")\n\tcmdFlags.StringVar(&ctx.AlterStatement, \"alter\", \"\", \"\")\n\tcmdFlags.BoolVar(&how, \"how\", false, \"\")\n\tcmdFlags.BoolVar(&ctx.Noop, \"dryrun\", false, \"\")\n\tcmdFlags.StringVar(&ctx.CliUser, \"user\", \"root\", \"\")\n\tcmdFlags.StringVar(&ctx.CliPassword, \"pass\", \"\", \"\")\n\tcmdFlags.Usage = func() { this.Ui.Output(this.Help()) }\n\tif err := cmdFlags.Parse(args); err != nil {\n\t\treturn 1\n\t}\n\n\tif how {\n\t\tthis.printHowItWorks()\n\t\treturn\n\t}\n\n\tctx.AllowedRunningOnMaster = true\n\tctx.OkToDropTable = true\n\tctx.InitiallyDropOldTable = true\n\tctx.InitiallyDropGhostTable = true\n\tctx.ServeSocketFile = \"/tmp/dbc.migrate.sock\"\n\tctx.SetChunkSize(1000)\n\tctx.SetDMLBatchSize(100)\n\tctx.SetDefaultNumRetries(3)\n\tctx.ApplyCredentials()\n\n\tlog.SetLevel(log.DEBUG)\n\n\tlog.Infof(\"starting\")\n\n\t// a migrator contains a sql parser(parsing alter statement), an inspector, applier and binlog streamer.\n\t// migrator.onChangelogStateEvent\n\t// migrator.onApplyEventStruct\n\t// 1. change log table\n\t// 2. ghost table\n\t// 3. alter ghost table\n\tmigrator := logic.NewMigrator()\n\tif err := migrator.Migrate(); err != nil {\n\t\tmigrator.ExecOnFailureHook()\n\t\tlog.Fatale(err)\n\t}\n\tlog.Infof(\"OK %d rows copied, %s, %s\", ctx.GetTotalRowsCopied(), ctx.ElapsedRowCopyTime(), ctx.ElapsedTime())\n\n\treturn\n}", "func (c *generateMigrationCommand) Run(args []string) error {\n\tif len(args) < 1 || args[0] == \"\" {\n\t\treturn fmt.Errorf(\"no NAME given\")\n\t}\n\tname := args[0]\n\tif c.option.ORM == \"\" {\n\t\tc.option.ORM = defaultORM\n\t}\n\torm, exists := ORM[c.option.ORM]\n\tif !exists {\n\t\treturn fmt.Errorf(\"unsupported ORM: `%v'\", c.option.ORM)\n\t}\n\tnow := _time.Now().Format(\"20060102150405\")\n\tdata := map[string]interface{}{\n\t\t\"Name\": util.ToCamelCase(name),\n\t\t\"TimeStamp\": now,\n\t\t\"ImportPath\": orm.ImportPath(),\n\t\t\"TxType\": reflect.TypeOf(orm.TransactionType()).String(),\n\t}\n\tif err := util.CopyTemplate(\n\t\tfilepath.Join(skeletonDir(\"migration\"), \"migration.go\"+util.TemplateSuffix),\n\t\tfilepath.Join(\"db\", \"migration\", fmt.Sprintf(\"%v_%v.go\", now, util.ToSnakeCase(name))),\n\t\tdata,\n\t); err != nil {\n\t\treturn err\n\t}\n\tinitPath := filepath.Join(\"db\", \"migration\", \"init.go\")\n\tif _, err := os.Stat(initPath); os.IsNotExist(err) {\n\t\tappDir, err := util.FindAppDir()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := util.CopyTemplate(\n\t\t\tfilepath.Join(skeletonDir(\"migration\"), \"init.go\"+util.TemplateSuffix),\n\t\t\tinitPath, map[string]interface{}{\n\t\t\t\t\"typeName\": c.option.ORM,\n\t\t\t\t\"tx\": strings.TrimSpace(util.GoString(orm)),\n\t\t\t\t\"dbImportPath\": path.Join(appDir, \"db\"),\n\t\t\t},\n\t\t); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (m *migrater) getNewMigrations() []migrationModel {\n\n\tvar names []string\n\tfor k, _ := range pool.migrations {\n\t\tnames = append(names, k)\n\t}\n\n\tsort.Strings(names)\n\n\tstep := 20 // limit\n\tresult := make([]migrationModel, 0)\n\texistMigrations := make(map[string]bool)\n\tfor i := 0; i < len(names); {\n\n\t\ti += step\n\t\tvar chunkNames []string\n\t\tif i <= len(names) {\n\t\t\tchunkNames = names[i-step : i]\n\t\t} else {\n\t\t\tchunkNames = names[i-step:]\n\t\t}\n\n\t\trows := make([]struct{ Name string }, 0)\n\t\tif err := m.db.Model(m.newMigrationModel()).\n\t\t\tWhere(\"name IN (?)\", chunkNames).\n\t\t\tScan(&rows).Error; err != nil {\n\n\t\t\tpanic(err)\n\t\t}\n\t\t\n\t\tfor _, row := range rows {\n\t\t\texistMigrations[row.Name] = true\n\t\t}\n\t}\n\n\tfor _, name := range names {\n\t\tif _, ok := existMigrations[name]; !ok {\n\t\t\tmodel := m.newMigrationModel()\n\t\t\tmodel.Name = name\n\t\t\tresult = append(result, model)\n\t\t}\n\t}\n\n\treturn result\n}", "func AutoMigrate(db *gorm.DB) {\n\tlog.Info(\"Migrating model\")\n\tif err := db.AutoMigrate(&User{}, &Product{}, &ProductProps{}).Error; err != nil {\n\t\tlog.Errorf(\"Can't automigrate schema %v\", err)\n\t}\n}", "func (p *planner) writeDatabaseChangeToBatch(\n\tctx context.Context, desc *dbdesc.Mutable, b *kv.Batch,\n) error {\n\treturn p.Descriptors().WriteDescToBatch(\n\t\tctx,\n\t\tp.extendedEvalCtx.Tracing.KVTracingEnabled(),\n\t\tdesc,\n\t\tb,\n\t)\n}", "func (_Vault *VaultCaller) Migration(opts *bind.CallOpts, arg0 common.Address, arg1 common.Address) (bool, error) {\n\tvar (\n\t\tret0 = new(bool)\n\t)\n\tout := ret0\n\terr := _Vault.contract.Call(opts, out, \"migration\", arg0, arg1)\n\treturn *ret0, err\n}", "func (ob *OrderBook) BatchUpdate() {\n\n}", "func (s *Store) Batch([]newstorage.Operation) error {\n\treturn errors.New(\"batch failure\")\n}", "func TestMigrateFail(t *testing.T) {\n\tt.Parallel()\n\n\ttype oldType struct {\n\t\tPrimary string `ds:\"primary\"`\n\t\tIndex string `ds:\"index\"`\n\t\tUnique string `ds:\"unique\"`\n\t}\n\ttype newType struct {\n\t\tPrimary string `ds:\"primary\"`\n\t\tIndex string `ds:\"index\"`\n\t\tSomethingElse int\n\t}\n\n\ttablePath := path.Join(tmpDir, randomString(12))\n\ttable, err := Register(oldType{}, tablePath, nil)\n\tif err != nil {\n\t\tt.Errorf(\"Error registering table: %s\", err.Error())\n\t}\n\n\ti := 0\n\tcount := 100\n\tindex := randomString(12)\n\tfor i < count {\n\t\terr = table.Add(oldType{\n\t\t\tPrimary: randomString(12),\n\t\t\tIndex: index,\n\t\t\tUnique: randomString(12),\n\t\t})\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Error adding value to table: %s\", err.Error())\n\t\t}\n\t\ti++\n\t}\n\n\ttable.Close()\n\n\ti = 0\n\tstats := Migrate(MigrateParams{\n\t\tTablePath: tablePath,\n\t\tOldType: oldType{},\n\t\tNewType: newType{},\n\t\tNewPath: tablePath,\n\t\tMigrateObject: func(o interface{}) (interface{}, error) {\n\t\t\told := o.(oldType)\n\t\t\tif i == count/2 {\n\t\t\t\treturn nil, fmt.Errorf(\"Fake error\")\n\t\t\t}\n\t\t\ti++\n\t\t\treturn newType{\n\t\t\t\tPrimary: old.Primary,\n\t\t\t\tIndex: old.Index,\n\t\t\t\tSomethingElse: i,\n\t\t\t}, nil\n\t\t},\n\t})\n\tif stats.Error == nil {\n\t\tt.Errorf(\"No error seens for failed migration\")\n\t}\n\tif stats.Success {\n\t\tt.Error(\"Migration successful but migration failed\")\n\t}\n}", "func (prod *InfluxDB) sendBatch() core.AssemblyFunc {\n\tif prod.writer.isConnectionUp() {\n\t\treturn prod.assembly.Write\n\t} else if prod.IsStopping() {\n\t\treturn prod.assembly.Flush\n\t}\n\n\treturn nil\n}", "func Migrate(logger logger.Logger, db *sqlx.DB) {\n\tlogger.Info().Msg(\"Starting database migration\")\n\n\tdbInstance, err := postgres.WithInstance(db.DB, &postgres.Config{})\n\tif err != nil {\n\t\tlogger.Fatal().Err(err).Msg(\"Failed to create migrate database instance\")\n\t}\n\n\tsrcInstance, err := httpfs.New(pkger.Dir(\"/migrate/migrations\"), \"\")\n\tif err != nil {\n\t\tlogger.Fatal().Err(err).Msg(\"Failed to create migrate source instance\")\n\t}\n\n\tms, err := migrate.NewWithInstance(\"httpfs\", srcInstance, \"postgres\", dbInstance)\n\tif err != nil {\n\t\tlogger.Fatal().Err(err).Msg(\"Failed to create migration service\")\n\t}\n\n\tversion, _, err := ms.Version()\n\tif err != nil {\n\t\tif err == migrate.ErrNilVersion {\n\t\t\tlogger.Info().Msg(\"No migrations have been applied yet\")\n\t\t} else {\n\t\t\tlogger.Fatal().Err(err).Msg(\"Failed to get migration version\")\n\t\t}\n\t} else {\n\t\tlogger.Info().Uint(\"migrationVersion\", version).Send()\n\t}\n\n\tif err := ms.Up(); err != nil {\n\t\tif err == migrate.ErrNoChange {\n\t\t\tlogger.Info().Msg(\"No new migrations to apply\")\n\t\t} else {\n\t\t\tlogger.Fatal().Err(err).Msg(\"Failed to apply migrations\")\n\t\t}\n\t}\n\n\tlogger.Info().Msg(\"Database migration complete\")\n}", "func (w *WorkflowTypeMigrator) Migrate() {\n\tfor _, dd := range w.DeprecatedWorkflowTypes {\n\t\tif w.isDeprecated(dd.Domain, dd.WorkflowType.Name, dd.WorkflowType.Version) {\n\t\t\tLog.Printf(\"action=migrate at=deprecate-workflow domain=%s workflow=%s version=%s status=previously-deprecated\", LS(dd.Domain), LS(dd.WorkflowType.Name), LS(dd.WorkflowType.Version))\n\t\t} else {\n\t\t\tw.deprecate(dd)\n\t\t\tLog.Printf(\"action=migrate at=deprecate-workflow domain=%s workflow=%s version=%s status=deprecate\", LS(dd.Domain), LS(dd.WorkflowType.Name), LS(dd.WorkflowType.Version))\n\t\t}\n\t}\n\tfor _, r := range w.RegisteredWorkflowTypes {\n\t\tif w.isRegisteredNotDeprecated(r) {\n\t\t\tLog.Printf(\"action=migrate at=register-workflow domain=%s workflow=%s version=%s status=previously-registered\", LS(r.Domain), LS(r.Name), LS(r.Version))\n\t\t} else {\n\t\t\tw.register(r)\n\t\t\tLog.Printf(\"action=migrate at=register-workflow domain=%s workflow=%s version=%s status=registered\", LS(r.Domain), LS(r.Name), LS(r.Version))\n\t\t}\n\t}\n}", "func Migrate(cfg *config.Config, logger log.Logger, command string) error {\n\tdb, err := sql.Open(\"mysql\", cfg.DB.Dsn)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn err\n\t}\n\tpath, err := os.Getwd()\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn err\n\t}\n\n\tmigrationPath := fmt.Sprintf(\"file://%s/migration\", path)\n\tlogger.Infof(\"migrationPath : %s\", migrationPath)\n\n\tdriver, err := mysql.WithInstance(db, &mysql.Config{})\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn err\n\t}\n\tm, err := migrater.NewWithDatabaseInstance(\n\t\tmigrationPath,\n\t\t\"mysql\",\n\t\tdriver,\n\t)\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn err\n\t}\n\tif command == \"up\" {\n\t\tlogger.Info(\"Migrate up\")\n\t\tif err := m.Up(); err != nil && err != migrater.ErrNoChange {\n\t\t\tlogger.Errorf(\"An error occurred while syncing the database.. %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif command == \"down\" {\n\t\tlogger.Info(\"Migrate down\")\n\t\tif err := m.Down(); err != nil && err != migrater.ErrNoChange {\n\t\t\tlogger.Errorf(\"An error occurred while syncing the database.. %v\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif err != nil {\n\t\tlogger.Error(err)\n\t\treturn err\n\t}\n\n\tlogger.Info(\"Migrate complete\")\n\treturn nil\n}", "func (db *JSONLite) InsertBatch(items []Item) ([]string, error) { // nolint:gocyclo\n\tif len(items) == 0 {\n\t\treturn nil, nil\n\t}\n\tfirstItem := items[0]\n\n\tif _, ok := firstItem[db.Discriminator()]; !ok {\n\t\treturn nil, errors.New(\"missing discriminator in item\")\n\t}\n\n\tif _, ok := firstItem[\"uid\"]; !ok {\n\t\tfirstItem[\"uid\"] = firstItem[db.Discriminator()].(string) + \"--\" + uuid.New().String()\n\t}\n\n\tflatItem, err := goflatten.Flatten(firstItem)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not flatten item\")\n\t}\n\n\tif err := db.ensureTable(flatItem, firstItem); err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not ensure table\")\n\t}\n\n\t// get columnNames\n\tvar columnNames []string\n\tfor k := range flatItem {\n\t\tcolumnNames = append(columnNames, k)\n\t}\n\n\t// get columnValues\n\tvar placeholderGrp []string\n\tvar columnValues []interface{}\n\tvar uids []string\n\tfor _, item := range items {\n\t\tif db.Strict() {\n\t\t\tdb.sqlMutex.Lock()\n\t\t\tvalErr, err := db.validateItemSchema(item)\n\t\t\tdb.sqlMutex.Unlock()\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.Wrap(err, \"validation failed\")\n\t\t\t}\n\t\t\tif len(valErr) > 0 {\n\t\t\t\treturn nil, fmt.Errorf(\"item could not be validated [%s]\", strings.Join(valErr, \",\"))\n\t\t\t}\n\t\t}\n\n\t\tflatItem, err := goflatten.Flatten(item)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"could not flatten item\")\n\t\t}\n\t\tif _, ok := flatItem[\"uid\"]; !ok {\n\t\t\tflatItem[\"uid\"] = flatItem[db.Discriminator()].(string) + \"--\" + uuid.New().String()\n\t\t}\n\t\tfor _, name := range columnNames {\n\t\t\tcolumnValues = append(columnValues, flatItem[name])\n\t\t}\n\t\tplaceholderGrp = append(placeholderGrp, \"(\"+strings.Repeat(\"?,\", len(flatItem)-1)+\"?)\")\n\n\t\tuids = append(uids, flatItem[\"uid\"].(string))\n\t}\n\n\tquery := fmt.Sprintf(\n\t\t\"INSERT INTO \\\"%s\\\"(%s) VALUES %s\",\n\t\tfirstItem[db.Discriminator()].(string),\n\t\t`\"`+strings.Join(columnNames, `\",\"`)+`\"`,\n\t\tstrings.Join(placeholderGrp, \",\"),\n\t) // #nosec\n\tstmt, err := db.cursor.Prepare(query)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, fmt.Sprintf(\"could not prepare statement %s\", query))\n\t}\n\n\tdb.sqlMutex.Lock()\n\tdefer db.sqlMutex.Unlock()\n\t_, err = stmt.Exec(columnValues...)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, fmt.Sprint(\"could not exec statement\", query, columnValues))\n\t}\n\n\treturn uids, nil\n}", "func AutoMigration(conn *gorm.DB) {\n\tconn.AutoMigrate(entity.User{})\n}", "func (*elephant) autoMigrate() error {\n\treturn db.AutoMigrate(&Company{}, &Sentiment{}, &Average{}).Error\n}", "func (ai *AppInteractor) CreateBatch(apps []domain.App) ([]string, error) {\n\treturn ai.AppRepository.CreateBatch(apps)\n}", "func main() {\n\tdsn := newPostgresConf().dsn()\n\tdb, err := gorm.Open(postgres.Open(dsn), &gorm.Config{})\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to connect database %v\", err)\n\t}\n\tif err := db.AutoMigrate(&Country{}); err != nil {\n\t\tlog.Fatalf(\"failed to make migration: %v\", err)\n\t}\n\taddFixtures(db)\n}", "func (m *SqlxMigrate) Migrate() error {\n\terr := m.createMigrationTable()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor _, migration := range m.migrations {\n\t\tvar found string\n\t\terr := m.db.Get(&found, \"SELECT id FROM migrations WHERE id=?\", migration.ID)\n\t\tswitch err {\n\t\tcase sql.ErrNoRows:\n\t\t\tlog.Printf(\"Running migration: %v\\n\", migration.ID)\n\t\t\t// we need to run the migration so we continue to code below\n\t\tcase nil:\n\t\t\tlog.Printf(\"Skipping migration: %v\\n\", migration.ID)\n\t\t\tcontinue\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"looking up migration by id: %w\", err)\n\t\t}\n\t\terr = m.runMigration(migration)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func TestMigrateParams(t *testing.T) {\n\tt.Parallel()\n\n\ttype exampleType struct {\n\t\tPrimary string `ds:\"primary\"`\n\t\tIndex string `ds:\"index\"`\n\t\tUnique string `ds:\"unique\"`\n\t}\n\n\ttablePath := path.Join(tmpDir, randomString(12))\n\ttable, err := Register(exampleType{}, tablePath, nil)\n\tif err != nil {\n\t\tt.Errorf(\"Error registering table: %s\", err.Error())\n\t}\n\n\terr = table.Add(exampleType{\n\t\tPrimary: randomString(12),\n\t\tIndex: randomString(12),\n\t\tUnique: randomString(12),\n\t})\n\tif err != nil {\n\t\tt.Errorf(\"Error adding value to table: %s\", err.Error())\n\t}\n\ttable.Close()\n\n\t// Missing table path\n\tstats := Migrate(MigrateParams{\n\t\tOldType: exampleType{},\n\t\tNewType: exampleType{},\n\t\tNewPath: tablePath,\n\t\tMigrateObject: func(o interface{}) (interface{}, error) {\n\t\t\treturn nil, nil\n\t\t},\n\t})\n\tif stats.Error == nil {\n\t\tt.Errorf(\"No error seen for invalid migration request\")\n\t}\n\n\t// Missing old type\n\tstats = Migrate(MigrateParams{\n\t\tTablePath: tablePath,\n\t\tNewType: exampleType{},\n\t\tNewPath: tablePath,\n\t\tMigrateObject: func(o interface{}) (interface{}, error) {\n\t\t\treturn nil, nil\n\t\t},\n\t})\n\tif stats.Error == nil {\n\t\tt.Errorf(\"No error seen for invalid migration request\")\n\t}\n\n\t// Missing new type\n\tstats = Migrate(MigrateParams{\n\t\tTablePath: tablePath,\n\t\tOldType: exampleType{},\n\t\tNewPath: tablePath,\n\t\tMigrateObject: func(o interface{}) (interface{}, error) {\n\t\t\treturn nil, nil\n\t\t},\n\t})\n\tif stats.Error == nil {\n\t\tt.Errorf(\"No error seen for invalid migration request\")\n\t}\n\n\t// Missing new path\n\tstats = Migrate(MigrateParams{\n\t\tTablePath: tablePath,\n\t\tOldType: exampleType{},\n\t\tNewType: exampleType{},\n\t\tMigrateObject: func(o interface{}) (interface{}, error) {\n\t\t\treturn nil, nil\n\t\t},\n\t})\n\tif stats.Error == nil {\n\t\tt.Errorf(\"No error seen for invalid migration request\")\n\t}\n\n\t// Missing migrate method\n\tstats = Migrate(MigrateParams{\n\t\tTablePath: tablePath,\n\t\tNewPath: tablePath,\n\t\tOldType: exampleType{},\n\t\tNewType: exampleType{},\n\t})\n\tif stats.Error == nil {\n\t\tt.Errorf(\"No error seen for invalid migration request\")\n\t}\n\n\t// Backup already exists\n\tioutil.WriteFile(tablePath+\"_backup\", []byte(\"\"), os.ModePerm)\n\tstats = Migrate(MigrateParams{\n\t\tTablePath: tablePath,\n\t\tOldType: exampleType{},\n\t\tNewType: exampleType{},\n\t\tNewPath: tablePath,\n\t\tMigrateObject: func(o interface{}) (interface{}, error) {\n\t\t\treturn nil, nil\n\t\t},\n\t})\n\tif stats.Error == nil {\n\t\tt.Errorf(\"No error seen for invalid migration request\")\n\t}\n}", "func (m *MigrationService) Migrate(args []string) error {\n\n\tfolder, _ := m.getCurrentFolder()\n\tluaRunner := luascript.NewLGoLua()\n\tluaRunner.RunLua(folder)\n\treturn nil\n}", "func migrateTables() error {\n\n\tif err := Connection.AutoMigrate(&MasterUser{}).Error; err != nil {\n\t\treturn err\n\t}\n\n\tif err := Connection.AutoMigrate(&User{}).Error; err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Perform(db *gorm.DB) {\r\n\r\n\tdb.AutoMigrate(\r\n\t\t&models.WorkItem{})\r\n}", "func RunMigrationsUp(gormdb *gorm.DB) {\n\tgormdb.AutoMigrate(&entities.Tag{})\n\tgormdb.AutoMigrate(&entities.RecordType{})\n\tgormdb.AutoMigrate(&entities.Record{})\n\tgormdb.Model(&entities.Record{}).AddForeignKey(\"record_type_id\", \"record_types(id)\", \"RESTRICT\", \"RESTRICT\")\n}", "func initialMigration() {\n\tdb, err := gorm.Open(\"sqlite3\", \"test.db\")\n\tif err != nil {\n\t\tfmt.Println(err.Error())\n\t\tpanic(\"failed to connect database\")\n\t}\n\tdefer db.Close()\n\n\t// Migrate the schema\n\tdb.AutoMigrate(&User{})\n}", "func (driver *Driver) Migrate(migration *m.PlannedMigration) error {\n\t// Note: Driver does not support DDL statements in a transaction. If DDL statements are\n\t// executed in a transaction, it is an implicit commit.\n\t// See: http://dev.mysql.com/doc/refman/5.7/en/implicit-commit.html\n\tvar migrationStatements *parser.ParsedMigration\n\n\tif migration.Direction == m.Up {\n\t\tmigrationStatements = migration.Up\n\t} else if migration.Direction == m.Down {\n\t\tmigrationStatements = migration.Down\n\t}\n\n\tfor _, sqlStmt := range migrationStatements.Statements {\n\t\tif len(strings.TrimSpace(sqlStmt)) > 0 {\n\t\t\tif _, err := driver.db.Exec(sqlStmt); err != nil {\n\t\t\t\treturn fmt.Errorf(\"Error executing statement: %s\\n%s\", err, sqlStmt)\n\t\t\t}\n\t\t}\n\t}\n\n\tif migration.Direction == m.Up {\n\t\tif _, err := driver.db.Exec(\"INSERT INTO \"+mysqlTableName+\" (version) VALUES (?)\", migration.ID); err != nil {\n\t\t\treturn err\n\t\t}\n\t} else {\n\t\tif _, err := driver.db.Exec(\"DELETE FROM \"+mysqlTableName+\" WHERE version=?\", migration.ID); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func Migrate(tx *gorm.DB) error {\n\treturn tx.Exec(`\n\t ALTER TABLE initiators ADD COLUMN \"polling_interval\" BigInt;\n\t`).Error\n}", "func (b *Botanist) AnnotateExtensionCRsForMigration(ctx context.Context) (err error) {\n\tif err = b.Shoot.Components.Extensions.Network.Migrate(ctx); err != nil {\n\t\treturn err\n\t}\n\n\tvar fns []flow.TaskFn\n\tfns, err = b.applyFuncToAllExtensionCRs(ctx, annotateObjectForMigrationFunc(ctx, b.K8sSeedClient.DirectClient()))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn flow.Parallel(fns...)(ctx)\n}", "func spannerBatchPut(ctx context.Context, db string, m []*spanner.Mutation) error {\n\tclient, err := spanner.NewClient(ctx, db)\n\tif err != nil {\n\t\tlog.Fatalf(\"Failed to create client %v\", err)\n\t\treturn err\n\t}\n\tdefer client.Close()\n\n\tif _, err = client.Apply(ctx, m); err != nil {\n\t\treturn errors.New(\"ResourceNotFoundException: \" + err.Error())\n\t}\n\treturn nil\n}", "func Migrate(migrationPolicy MigrationPolicy) {\n\tcontainer.Make(\n\t\tfunc(gormDB *gorm.DB) {\n\n\t\t\tif migrationPolicy == OrmMigration {\n\t\t\t\terr := gormDB.AutoMigrate(\n\t\t\t\t\t&dbmodel.User{},\n\t\t\t\t\t&dbmodel.Department{},\n\t\t\t\t\t&dbmodel.Province{},\n\t\t\t\t\t&dbmodel.District{},\n\t\t\t\t)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Fatalln(err.Error())\n\t\t\t\t}\n\t\t\t} else if migrationPolicy == SQLScriptsMigration {\n\t\t\t\tsqlDB, err := gormDB.DB()\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Fatalln(err.Error())\n\t\t\t\t}\n\n\t\t\t\terr = sqlDB.Ping()\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Fatalln(err.Error())\n\t\t\t\t}\n\n\t\t\t\tdriver, _ := mysql.WithInstance(\n\t\t\t\t\tsqlDB, &mysql.Config{},\n\t\t\t\t)\n\n\t\t\t\tm, err := migrate.NewWithDatabaseInstance(\n\t\t\t\t\t\"file://internal/database/migrations/mysql\",\n\t\t\t\t\tmySQLDriver,\n\t\t\t\t\tdriver,\n\t\t\t\t)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlog.Fatalln(err.Error())\n\t\t\t\t}\n\n\t\t\t\terr = m.Up()\n\t\t\t\tif err != nil && err != migrate.ErrNoChange {\n\t\t\t\t\tlog.Fatalln(err.Error())\n\t\t\t\t}\n\t\t\t}\n\t\t},\n\t)\n}" ]
[ "0.6200696", "0.6156387", "0.6027621", "0.60257566", "0.59955513", "0.5990754", "0.5964021", "0.595133", "0.59126735", "0.5796585", "0.5782023", "0.5779723", "0.57062995", "0.5700448", "0.56916976", "0.56082976", "0.56056505", "0.56056505", "0.55998796", "0.5583283", "0.55158377", "0.5505171", "0.54951704", "0.5485518", "0.5479103", "0.5461872", "0.54598874", "0.54388297", "0.5431578", "0.541415", "0.5413604", "0.53922886", "0.53916156", "0.5388347", "0.53809386", "0.536937", "0.5302814", "0.5286776", "0.5281757", "0.5276036", "0.52639407", "0.5253076", "0.52046025", "0.51910007", "0.5163371", "0.5162476", "0.5154256", "0.5135911", "0.5124737", "0.51241475", "0.5117955", "0.51083404", "0.51011413", "0.50989336", "0.5094174", "0.5094036", "0.5082706", "0.50733227", "0.5070641", "0.50634146", "0.5059088", "0.50590837", "0.5047906", "0.50242", "0.5023161", "0.5009052", "0.5005393", "0.50047296", "0.50043195", "0.50036186", "0.49919283", "0.4981386", "0.49792767", "0.49780768", "0.49726573", "0.4966295", "0.49659905", "0.49626413", "0.49588895", "0.495852", "0.49583226", "0.4955603", "0.493996", "0.4936772", "0.49363106", "0.49308035", "0.4929044", "0.4928926", "0.49124375", "0.49095666", "0.49063453", "0.4901756", "0.48954508", "0.48919132", "0.48857644", "0.48836693", "0.48821744", "0.48782393", "0.4870996", "0.48582342" ]
0.6020524
4
TimeStrToInt parse time string to unix nano
func TimeStrToInt(ts, layout string) int64 { if ts == "" { return 0 } if layout == "" { layout = "2006-01-02 15:04:05" } t, err := time.ParseInLocation(layout, ts, time.Local) if err != nil { logs.Error(err) return 0 } return t.Unix() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TimeStr2Unix(timeStr string) (int64, error) {\n\tt, err := time.ParseInLocation(\"2006-01-02 15:04:05\", timeStr, time.Local)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn t.Unix(), nil\n}", "func (u *Util) ConvertStrTime2Unix(timeStr string) (int64, error) {\n\tnowT := time.Now()\n\ttm, err := time.Parse(TimeLayoutStr, timeStr)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\ttm.Sub(nowT)\n\treturn tm.UTC().Unix(), nil\n}", "func TimeStr2UnixMilli(timeStr string) (int64, error) {\n\tt, err := time.ParseInLocation(\"2006-01-02 15:04:05\", timeStr, time.Local)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int64(t.UnixNano() / 1000000), nil\n}", "func Unix(sec int64, nsec int64) Time {}", "func stringToUnixTime(s string) int64 {\n\t// Parse YYYY-MM-DD\n\ttimeT, err := time.Parse(\"2006-01-02\", s)\n\tif err != nil {\n\t\tErrorLog(\"cannot convert string: \"+s+\"to millisecond: %s\", err.Error())\n\t\treturn 0\n\t}\n\treturn timeT.Unix()\n}", "func parseTime(t string) (time.Time, error) {\n\tepoch, err := strconv.ParseInt(t, 10, 64)\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\treturn time.Unix(epoch, 0), nil\n}", "func convertTime(timeString string) time.Time {\n\tif timeString == \"\" {\n\t\treturn time.Unix(0, 0)\n\t}\n\n\ti, err := strconv.ParseInt(timeString, 10, 64)\n\tif err != nil {\n\t\treturn time.Unix(63072000, 0)\n\t}\n\n\treturn time.Unix(i, 0)\n}", "func Str2TimeStampS(s string) int64 {\n\tloc, _ := time.LoadLocation(\"UTC\")\n\tt, err := time.ParseInLocation(DATE_FORMAT_SHORT, s, loc)\n\tCheckErr(err, CHECK_FLAG_LOGONLY)\n\treturn t.Unix()\n}", "func parseUnix(str string, def time.Time) (time.Time, error) {\n\tif len(str) == 0 {\n\t\treturn def, nil\n\t}\n\tunix, err := strconv.ParseInt(str, 10, 64)\n\tif err != nil {\n\t\treturn def, err\n\t}\n\treturn time.Unix(unix, 0), nil\n}", "func Time2Int64(t int64) int64 {\n\tdatestring := UnixMilli2TimeStr(t)\n\t_time := datestring[11:]\n\ttimestr := _time + strings.Repeat(\"0\", 12-len(_time))\n\ttimeintstr := strings.Replace(strings.Replace(timestr, \":\", \"\", -1), \".\", \"\", -1)\n\ttimeToInt, _ := strconv.ParseInt(timeintstr, 10, 64)\n\treturn timeToInt\n}", "func (t Time) Unix() int64 {}", "func (u *Util) TimeStr2Seconds(timeStr string) int {\n\tvar (\n\t\ttimeSeconds int\n\t\ttempIntVal int\n\t\ttempFloatVal float64\n\t)\n\n\t//basic check\n\tif timeStr == \"\" {\n\t\treturn timeSeconds\n\t}\n\n\ti := 1\n\ttempSlice := strings.Split(timeStr, \":\")\n\tfor _, info := range tempSlice {\n\t\tswitch i {\n\t\tcase 1://hour\n\t\t\ttempIntVal, _ = strconv.Atoi(info)\n\t\t\ttimeSeconds += tempIntVal * 3600\n\t\tcase 2://minute\n\t\t\ttempIntVal, _ = strconv.Atoi(info)\n\t\t\ttimeSeconds += tempIntVal * 60\n\t\tcase 3:\t//second\n\t\t\ttempFloatVal, _ = strconv.ParseFloat(info, 64)\n\t\t\ttimeSeconds += int(tempFloatVal)\n\t\t}\n\t\ti++\n\t}\n\n\treturn timeSeconds\n}", "func parseTime(value string) (int64, error) {\n\tif len(value) != 0 {\n\t\tt, err := time.Parse(time.RFC3339, value)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t\tif !t.IsZero() {\n\t\t\treturn t.Unix(), nil\n\t\t}\n\t}\n\treturn 0, nil\n}", "func parseTime(msg []byte) *time.Time {\n\t// convert input to integer\n\ti := binary.BigEndian.Uint32(msg)\n\n\t// convert time from 1900 to Unix time\n\tt := int64(i) - timeGap1900\n\n\tres := time.Unix(t, 0)\n\treturn &res\n}", "func TimevalToNsec(tv Timeval) int64 { return tv.Nano() }", "func parseTimeStamp(s string) (int64, error) {\n\tt, err := time.Parse(TimeStampLayout, s)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn t.UnixNano() / int64(time.Millisecond), nil\n}", "func getUnixTime(val []byte) (uint64, error) {\n\tif len(val) < 8 {\n\t\treturn 0, errors.New(\"len(val) < 8, want len(val) => 8\")\n\t}\n\tunixTime := ((uint64)(val[0]) | // 1st\n\t\t((uint64)(val[1]) << 8) | // 2nd\n\t\t((uint64)(val[2]) << 16) | // 3rd\n\t\t((uint64)(val[3]) << 24) | // 4th\n\t\t((uint64)(val[4]) << 32) | // 5th\n\t\t((uint64)(val[5]) << 40) | // 6th\n\t\t((uint64)(val[6]) << 48) | // 7th\n\t\t((uint64)(val[7]) << 56)) // 8th\n\treturn unixTime, nil\n}", "func toUnixMsec(t time.Time) int64 {\n\treturn t.UnixNano() / 1e6\n}", "func Str2TimeStampL(s string) int64 {\n\tloc, _ := time.LoadLocation(\"UTC\")\n\tt, err := time.ParseInLocation(DATE_FORMAT_LONG, s, loc)\n\tCheckErr(err, CHECK_FLAG_LOGONLY)\n\treturn t.Unix()\n}", "func ParseTime(s string, fmt string) (ti time.Time, err error) {\n\tvar i int64\n\tswitch fmt {\n\tcase \"ns\", \"nanosecond\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i), nil\n\t\t}\n\tcase \"us\", \"microsecond\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Microsecond)), nil\n\t\t}\n\tcase \"ms\", \"millisecond\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Millisecond)), nil\n\t\t}\n\tcase \"s\", \"second\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Second)), nil\n\t\t}\n\tcase \"mi\", \"minute\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Minute)), nil\n\t\t}\n\tcase \"h\", \"hour\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Hour)), nil\n\t\t}\n\tcase \"d\", \"day\":\n\t\tif i, err = strconv.ParseInt(s, 10, 64); err != nil {\n\t\t\treturn\n\t\t} else {\n\t\t\treturn time.Unix(0, i*int64(time.Hour)*24), nil\n\t\t}\n\tdefault:\n\t\tif fmt == \"\" {\n\t\t\tfmt = time.RFC3339Nano\n\t\t}\n\t\tif ti, e := time.ParseInLocation(fmt, s, time.Local); e != nil {\n\t\t\treturn ti, errs.New(e)\n\t\t} else {\n\t\t\treturn ti, nil\n\t\t}\n\t}\n}", "func parseClock(s string) (int, error) {\n\tclock := 0\n\n\tfor s != \"\" {\n\t\tcomponent, rest := splitOnce(s, \":\")\n\n\t\tn, err := strconv.Atoi(component)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\n\t\tclock *= 60\n\t\tclock += int(n)\n\n\t\ts = rest\n\t}\n\n\treturn clock, nil\n}", "func convertUnixTime(timeInt int) time.Time {\n\n\ti := int64(timeInt)\n\treturn time.Unix(i, 0)\n\n}", "func parseTimestamp(s string) (time.Time, error) {\n\tsp := strings.Split(s, \".\")\n\tsec, nsec := int64(0), int64(0)\n\tvar err error\n\tif len(sp) > 0 {\n\t\tif sec, err = strconv.ParseInt(sp[0], 10, 64); err != nil {\n\t\t\treturn time.Time{}, fmt.Errorf(\"unable to parse timestamp %s: %s\",\n\t\t\t\ts, err.Error())\n\t\t}\n\t}\n\n\tif len(sp) > 1 {\n\t\tif nsec, err = strconv.ParseInt(sp[1], 10, 64); err != nil {\n\t\t\treturn time.Time{}, fmt.Errorf(\"unable to parse timestamp %s: %s\",\n\t\t\t\ts, err.Error())\n\t\t}\n\n\t\tnsec *= 1000000\n\t}\n\n\treturn time.Unix(sec, nsec), nil\n}", "func parseUnixTimeString(ref *ShapeRef, memName, v string) string {\n\tref.API.AddSDKImport(\"private/protocol\")\n\treturn fmt.Sprintf(\"%s: %s,\\n\", memName, inlineParseModeledTime(protocol.UnixTimeFormatName, v))\n}", "func parseTime(s string) (hour, minute int64, err error) {\n\ttime := strings.Split(s, \":\")\n\thour, err = strconv.ParseInt(time[0], 10, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\tminute, err = strconv.ParseInt(time[1], 10, 64)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func timeToUnix(t time.Time) int64 {\n\tif t.IsZero() {\n\t\treturn 0\n\t}\n\n\treturn t.Unix()\n}", "func ParseStringToTime(tm string) (int64, error) {\n\n\tloc , _ := time.LoadLocation(\"Local\")\n\tresultTime , err := time.ParseInLocation(\"2006-01-02 15:04:05\", tm, loc)\n\n\treturn resultTime.Unix() , err\n\n}", "func toTimeSeconds(value string) (int64, error) {\n\t//is serial format?\n\tserial, err := strconv.ParseFloat(value, 64)\n\tif err != nil {\n\t\treturn -1, err\n\t}\n\treturn int64(serial * 86400), nil\n}", "func timeParse(microsecs string) time.Time {\n\tt := time.Date(1601, time.January, 1, 0, 0, 0, 0, time.UTC)\n\tm, err := strconv.ParseInt(microsecs, 10, 64)\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"%s\\n\", err.Error())\n\t\tos.Exit(1)\n\t}\n\tvar u int64 = 100000000000000\n\tdu := time.Duration(u) * time.Microsecond\n\tf := float64(m)\n\tx := float64(u)\n\tn := f / x\n\tr := int64(n)\n\tremainder := math.Mod(f, x)\n\tiRem := int64(remainder)\n\tvar i int64\n\tfor i = 0; i < r; i++ {\n\t\tt = t.Add(du)\n\t}\n\n\tt = t.Add(time.Duration(iRem) * time.Microsecond)\n\n\t// RFC1123 = \"Mon, 02 Jan 2006 15:04:05 MST\"\n\t// t.Format(time.RFC1123)\n\treturn t\n}", "func timeToUnixMS(t time.Time) int64 {\n\treturn t.UnixNano() / int64(time.Millisecond)\n}", "func ParseTime(s string) (int64, error) {\n\tif t, err := strconv.ParseFloat(s, 64); err == nil {\n\t\ts, ns := math.Modf(t)\n\t\tns = math.Round(ns*1000) / 1000\n\t\ttm := time.Unix(int64(s), int64(ns*float64(time.Second)))\n\t\treturn TimeToMillis(tm), nil\n\t}\n\tif t, err := time.Parse(time.RFC3339Nano, s); err == nil {\n\t\treturn TimeToMillis(t), nil\n\t}\n\treturn 0, httpgrpc.Errorf(http.StatusBadRequest, \"cannot parse %q to a valid timestamp\", s)\n}", "func StrToTime(str string) *Time {\n\tstr = strings.TrimSpace(str)\n\t// Check sign\n\tsign := Time(1)\n\tswitch str[0] {\n\tcase '-':\n\t\tsign = -1\n\t\tfallthrough\n\tcase '+':\n\t\tstr = str[1:]\n\t}\n\tvar (\n\t\tii int\n\t\tok error\n\t\ttt Time\n\t)\n\t// Find houre\n\tif nn := strings.IndexRune(str, ':'); nn != -1 {\n\t\tif ii, ok = strconv.Atoi(str[0:nn]); ok != nil {\n\t\t\treturn nil\n\t\t}\n\t\ttt = Time(ii * 3600)\n\t\tstr = str[nn+1:]\n\t} else {\n\t\treturn nil\n\t}\n\tif len(str) != 5 && len(str) != 15 || str[2] != ':' {\n\t\treturn nil\n\t}\n\tif ii, ok = strconv.Atoi(str[0:2]); ok != nil || ii > 59 {\n\t\treturn nil\n\t}\n\ttt += Time(ii * 60)\n\tif ii, ok = strconv.Atoi(str[3:5]); ok != nil || ii > 59 {\n\t\treturn nil\n\t}\n\ttt += Time(ii)\n\ttt *= 1e9\n\tif len(str) == 15 {\n\t\tif str[5] != '.' {\n\t\t\treturn nil\n\t\t}\n\t\tif ii, ok = strconv.Atoi(str[6:15]); ok != nil {\n\t\t\treturn nil\n\t\t}\n\t\ttt += Time(ii)\n\t}\n\ttt *= sign\n\treturn &tt\n}", "func parsePAXTime(t string) (time.Time, error) {\n\tbuf := []byte(t)\n\tpos := bytes.IndexByte(buf, '.')\n\tvar seconds, nanoseconds int64\n\tvar err error\n\tif pos == -1 {\n\t\tseconds, err = strconv.ParseInt(t, 10, 0)\n\t\tif err != nil {\n\t\t\treturn time.Time{}, err\n\t\t}\n\t} else {\n\t\tseconds, err = strconv.ParseInt(string(buf[:pos]), 10, 0)\n\t\tif err != nil {\n\t\t\treturn time.Time{}, err\n\t\t}\n\t\tnano_buf := string(buf[pos+1:])\n\t\t// Pad as needed before converting to a decimal.\n\t\t// For example .030 -> .030000000 -> 30000000 nanoseconds\n\t\tif len(nano_buf) < maxNanoSecondIntSize {\n\t\t\t// Right pad\n\t\t\tnano_buf += strings.Repeat(\"0\", maxNanoSecondIntSize-len(nano_buf))\n\t\t} else if len(nano_buf) > maxNanoSecondIntSize {\n\t\t\t// Right truncate\n\t\t\tnano_buf = nano_buf[:maxNanoSecondIntSize]\n\t\t}\n\t\tnanoseconds, err = strconv.ParseInt(string(nano_buf), 10, 0)\n\t\tif err != nil {\n\t\t\treturn time.Time{}, err\n\t\t}\n\t}\n\tts := time.Unix(seconds, nanoseconds)\n\treturn ts, nil\n}", "func ntpUnPretty(s string) (int64, error) {\n\tif len(s) < 1 {\n\t\treturn 0, fmt.Errorf(\"Zero length string passed to ntpUnPretty\")\n\t}\n\tvar multiplier int64 = 1\n\tshift := 1\n\tswitch s[len(s)-1] {\n\tcase 'm':\n\t\tmultiplier = 60\n\tcase 'h':\n\t\tmultiplier = 60 * 60\n\tcase 'd':\n\t\tmultiplier = 60 * 60 * 24\n\tdefault:\n\t\tshift = 0\n\t}\n\ti, err := strconv.ParseInt(s[0:len(s)-shift], 10, 64)\n\treturn i * multiplier, err\n}", "func TimeUnix(inputTime time.Time) int64 {\n\treturn (inputTime.UnixNano() / (int64(time.Millisecond) / int64(time.Nanosecond)))\n}", "func parse_time(mtime string) (int64, error) {\n\ttemp, err := time.Parse(\"Mon, 02 Jan 2006 15:04:05 MST\", mtime)\n\tif err == nil {\n\t\treturn temp.Unix(), nil\n\t}\n\n\ttemp, err = time.Parse(\"Mon, 2 Jan 2006 15:04:05 MST\", mtime)\n\tif err == nil {\n\t\treturn temp.Unix(), nil\n\t}\n\n\ttemp, err = time.Parse(\"Mon, 02 Jan 2006 15:04:05 -0700\", mtime)\n\tif err == nil {\n\t\treturn temp.Unix(), nil\n\t}\n\n\ttemp, err = time.Parse(\"Mon, 2 Jan 2006 15:04:05 -0700\", mtime)\n\tif err == nil {\n\t\treturn temp.Unix(), nil\n\t}\n\n\ttemp, err = time.Parse(\"2006-01-02T15:04:05.000Z\", mtime)\n\tif err == nil {\n\t\treturn temp.Unix(), nil\n\t}\n\n\tfmt.Printf(\"There was an error parsing: %s\\n\", mtime)\n\treturn -1, errors.New(\"There was an error parsing the timestamp\")\n}", "func tdParseTimestamp( t string, a string ) int64 {\n n, err := strconv.ParseInt( t, 10, 64 )\n if err == nil {\n // NR feed is in Java time (millis) so convert to Unix time (seconds)\n n := n / int64(1000)\n\n if a != \"\" {\n dt := time.Now().Unix() - n\n // Hide anomaly around local midnight where a blip is recorded going from\n // +- 2400 (14400 when merged into 1m samples)\n if dt > -2400 && dt < 2400 {\n statistics.Set( \"td.\" + a, dt )\n }\n }\n\n return n\n }\n return 0\n}", "func timeConversion(unixTime string) string {\r\n\tuTime, _ := strconv.ParseInt(unixTime, 10, 64)\r\n\tdateTime := time.Unix(uTime, 0)\r\n\tloc, _ := time.LoadLocation(\"America/New_York\")\r\n\tnewTime := dateTime.In(loc).Format(\"2006-01-02 15:04:05\")\r\n\treturn newTime\r\n}", "func ToUsec(t time.Time) int64 {\n\treturn t.UnixNano() / 1e3\n}", "func (c *StringConverter) ToTimeFromNsec() (time.Time, error) {\n\ti, err := c.ToInt64()\n\tif err != nil {\n\t\treturn time.Time{}, err\n\t}\n\treturn time.Unix(0, i), nil\n}", "func convertTimeToTs(ts time.Time) int32 {\n\treturn int32(ts.Unix())\n}", "func humanToNanoTime(value []byte) ([]byte) {\n\tdura, err := time.ParseDuration(string(value))\n\tif err != nil {\n\t\treturn value\n\t}\n\treturn []byte(strconv.FormatInt(dura.Nanoseconds(), 10))\n}", "func parseTime(s string) (time.Time, error) {\n\t// attempt to parse time as RFC3339 string\n\tt, err := time.Parse(time.RFC3339Nano, s)\n\tif err == nil {\n\t\treturn t, nil\n\t}\n\n\t// attempt to parse time as float number of unix seconds\n\tif f, err := strconv.ParseFloat(s, 64); err == nil {\n\t\tsec, dec := math.Modf(f)\n\t\treturn time.Unix(int64(sec), int64(dec*(1e9))), nil\n\t}\n\n\t// attempt to parse time as json marshaled value\n\tif err := json.Unmarshal([]byte(s), &t); err == nil {\n\t\treturn t, nil\n\t}\n\n\treturn time.Time{}, err\n}", "func toNtpTime(t time.Time) (uint32, uint32) {\n\tvar nanoPerSec uint64\n\n\tntpEpoch := time.Date(1900, 1, 1, 0, 0, 0, 0, time.UTC)\n\tnsec := uint64(t.Sub(ntpEpoch))\n\tnanoPerSec = 1000000000\n\tsec := nsec / nanoPerSec\n\t// Round up the fractional component so that repeated conversions\n\t// between time.Time and ntpTime do not yield continually decreasing\n\t// results.\n\tfrac := (((nsec - sec*nanoPerSec) << 32) + nanoPerSec - 1) / nanoPerSec\n\treturn uint32(sec & 0x00000000FFFFFFFF), uint32(frac & 0x00000000FFFFFFFF)\n}", "func seconds(s string) int64 {\n\tt, err := time.Parse(gitime, s)\n\tif err != nil {\n\t\treturn -1\n\t}\n\treturn t.Unix()\n}", "func TimeStampToSeconds(ts string) int {\n\ttotalSec := 0\n\tdata := strings.Split(ts, \":\")\n\ttimeVal := []int{}\n\tfor _, val := range data {\n\t\ttime, err := strconv.Atoi(val)\n\t\tif err != nil {\n\t\t\treturn 0\n\t\t}\n\t\ttimeVal = append(timeVal, time)\n\t}\n\tfor i, j := 0, len(timeVal)-1; i < j; i, j = i+1, j-1 { // reverse\n\t\ttimeVal[i], timeVal[j] = timeVal[j], timeVal[i]\n\t}\n\tfor i := 0; i < len(timeVal); i++ {\n\t\tswitch i {\n\t\tcase 0:\n\t\t\ttotalSec = timeVal[i]\n\t\tcase 1:\n\t\t\ttotalSec = totalSec + (timeVal[i] * 60)\n\t\tcase 2:\n\t\t\ttotalSec = totalSec + (timeVal[i] * 3600)\n\t\tcase 3:\n\t\t\ttotalSec = totalSec + (timeVal[i] * 86400)\n\t\t}\n\t}\n\treturn totalSec\n}", "func poolSyncTimeParseFloat64(e string) float64 {\n\tnum, err := strconv.ParseFloat(e, 64)\n\tif err != nil {\n\t\treturn 0\n\n\t}\n\treturn num\n}", "func NanoTime() int64", "func TimeInSec(period string) int {\n\tif strings.HasSuffix(period, \"sec\") {\n\t\ti, _ := strconv.Atoi(strings.Replace(period, \"sec\", \"\", -1))\n\t\treturn i\n\t} else if strings.HasSuffix(period, \"min\") {\n\t\ti, _ := strconv.Atoi(strings.Replace(period, \"min\", \"\", -1))\n\t\treturn i * 60\n\t} else if strings.HasSuffix(period, \"hours\") {\n\t\ti, _ := strconv.Atoi(strings.Replace(period, \"hours\", \"\", -1))\n\t\treturn i * 60 * 60\n\t} else if strings.HasSuffix(period, \"days\") {\n\t\ti, _ := strconv.Atoi(strings.Replace(period, \"days\", \"\", -1))\n\t\treturn i * 60 * 60 * 24\n\t} else {\n\t\treturn 0\n\t}\n}", "func ParseTimeUnixUTC(in string) time.Time {\n\ti, err := strconv.ParseInt(in, 10, 64)\n\tif err != nil {\n\t\treturn time.Time{}\n\t}\n\tv := time.Unix(i, 0)\n\treturn v\n}", "func (val Time) Unix() (sec, nsec int64) {\n\tsec = int64(val) / 1000000\n\tnsec = (int64(val) % 1000000) * 1000\n\treturn\n}", "func MysqlTimeToUnix(ts string) int64 {\n\tloc, _ := time.LoadLocation(\"Local\")\n\tt, _ := time.ParseInLocation(goMysqlTimeFormat, ts, loc)\n\treturn t.Unix()\n}", "func parseSplitTimestamp(timestamp string) (int64, int, error) {\n\tsplits := strings.Split(timestamp, \".\")\n\tif len(splits) != 2 {\n\t\treturn 0, 0, fmt.Errorf(\"can't parse timestamp %s\", timestamp)\n\t}\n\tnanos, err := strconv.ParseInt(splits[0], 0, 0)\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\tif nanos <= 0 {\n\t\treturn 0, 0, fmt.Errorf(\"nanos must be greater than 0: %d\", nanos)\n\t}\n\tlogical, err := strconv.Atoi(splits[1])\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\treturn nanos, logical, nil\n}", "func TimespecToNsec(ts Timespec) int64 { return ts.Nano() }", "func ParseTS(ts uint64) (time.Time, uint64) {\n\tlogical := ts & logicalBitsMask\n\tphysical := ts >> logicalBits\n\tphysicalTime := time.Unix(int64(physical/1000), int64(physical)%1000*time.Millisecond.Nanoseconds())\n\treturn physicalTime, logical\n}", "func ParseTS(ts uint64) (time.Time, uint64) {\n\tlogical := ts & logicalBitsMask\n\tphysical := ts >> logicalBits\n\tphysicalTime := time.Unix(int64(physical/1000), int64(physical)%1000*time.Millisecond.Nanoseconds())\n\treturn physicalTime, logical\n}", "func (u *Util) DateTime2Unix(dateTime string) (int64, error) {\n\t//remove un useful info\n\tdateTime = strings.Replace(dateTime, \"T\", \" \", -1)\n\tdateTime = strings.Replace(dateTime, \"Z\", \"\", -1)\n\n\t//theTime, err := time.Parse(TimeLayOut, dateTime)\n\ttheTime, err := time.ParseInLocation(TimeLayoutStr, dateTime, time.Local)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn theTime.Unix(), nil\n}", "func Str2TimeL(s string) time.Time {\n\tloc, _ := time.LoadLocation(\"UTC\")\n\tt, err := time.ParseInLocation(DATE_FORMAT_LONG, s, loc)\n\tCheckErr(err, CHECK_FLAG_LOGONLY)\n\treturn t\n}", "func (c *Client) parseTime(lines []string, err error) (time.Time, error) {\n\tvar t time.Time\n\tif err != nil {\n\t\treturn t, err\n\t}\n\n\tif len(lines) != 1 {\n\t\treturn t, NewInvalidResponseError(\"parseTime: unexpected lines\", lines...)\n\t}\n\n\ti, err := strconv.ParseInt(lines[0], 10, 64)\n\tif err != nil {\n\t\treturn t, NewInvalidResponseError(\"parseTime: parse int\", lines[0])\n\t}\n\n\treturn time.Unix(i, 0), nil\n}", "func nanotime() int64", "func nanotime() int64", "func nanotime() int64", "func nanotime() int64", "func (p Packet) TimeUnix() int64 {\n\treturn int64(p.CdTime >> 30)\n}", "func Unix(sec, nsec int64) *Timestamp {\n\tt := time.Unix(sec, nsec).UTC()\n\treturn Time(t)\n}", "func convertTime(time uint64, stream_uot, target_uot UnitOfTime) uint64 {\n\tunitmultiplier := map[UnitOfTime]uint64{\n\t\tUOT_NS: 1000000000,\n\t\tUOT_US: 1000000,\n\t\tUOT_MS: 1000,\n\t\tUOT_S: 1}\n\treturn time / unitmultiplier[stream_uot] * unitmultiplier[target_uot]\n}", "func toMSTimestamp(t time.Time) int64 {\n\treturn t.UnixNano() / 1e6\n}", "func (gdb *Gdb) getUnixTimeStamp(t string, d int) (int64, error) {\n\tif st, err := time.Parse(timeFormatString, t); err != nil {\n\t\treturn -1, err\n\t} else {\n\t\treturn st.Add(time.Duration(d) * time.Second).Unix(), nil\n\t}\n}", "func timeToApigeeInt(t time.Time) int64 {\n\treturn t.UnixNano() / (int64(time.Millisecond) / int64(time.Nanosecond))\n}", "func DecodeUTime(b []byte) time.Time {\n\treturn time.UnixMilli(int64(IntCoder.Uint64(b)))\n}", "func PointerTimeToInt64(i *time.Time) *int64 {\n\tif i == nil {\n\t\treturn nil\n\t}\n\treturn Int64(i.Unix())\n}", "func fromUnixMilli(ms int64) time.Time {\n\treturn time.Unix(ms/int64(millisInSecond), (ms%int64(millisInSecond))*int64(nsInSecond))\n}", "func timeToEpochMS(t time.Time) int64 {\n\treturn t.UnixNano() / 1000000\n}", "func parsePdTime(pdTime int64) *metav1.Time {\n\tt := metav1.Unix(pdTime, 0)\n\treturn &t\n}", "func byteArrToTime(byteArr []byte) int64 {\n\t//This is set to bigendian so that the timestamp is sorted in binary format.\n\ttimeVal := int64(binary.BigEndian.Uint64(byteArr))\n\treturn timeVal\n}", "func CalcTime(s string) (time.Time, error) {\n\tt := time.Time{}\n\tif len(s) != 17 {\n\t\treturn t, fmt.Errorf(\"invalid time string length %d for %s\", len(s), s)\n\t}\n\ty, e := strconv.Atoi(s[:4])\n\tif e != nil {\n\t\treturn t, errors.NewUtil(s, \"reading year\", e)\n\t}\n\tif y < 1900 || y > 3000 {\n\t\treturn t, fmt.Errorf(\"invalid year %d\", y)\n\t}\n\tm, e := strconv.Atoi(s[4:6])\n\tif e != nil {\n\t\treturn t, errors.NewUtil(s, \"reading month\", e)\n\t}\n\tif m < 1 || m > 12 {\n\t\treturn t, fmt.Errorf(\"invalid month %d\", m)\n\t}\n\td, e := strconv.Atoi(s[6:8])\n\tif e != nil {\n\t\treturn t, errors.NewUtil(s, \"reading day\", e)\n\t}\n\tif d < 1 || d > 31 {\n\t\treturn t, fmt.Errorf(\"invalid day %d\", d)\n\t}\n\th, e := strconv.Atoi(s[8:10])\n\tif e != nil {\n\t\treturn t, errors.NewUtil(s, \"reading hour\", e)\n\t}\n\tif h < 0 || h > 24 {\n\t\treturn t, fmt.Errorf(\"invalid hour %d\", h)\n\t}\n\tmi, e := strconv.Atoi(s[10:12])\n\tif e != nil {\n\t\treturn t, errors.NewUtil(s, \"reading minutes\", e)\n\t}\n\tif mi < 0 || mi > 60 {\n\t\treturn t, fmt.Errorf(\"invalid minutes %d\", mi)\n\t}\n\tsc, e := strconv.Atoi(s[12:14])\n\tif e != nil {\n\t\treturn t, errors.NewUtil(s, \"reading seconds\", e)\n\t}\n\tif sc < 0 || sc > 60 {\n\t\treturn t, fmt.Errorf(\"invalid seconds %d\", sc)\n\t}\n\tms, e := strconv.Atoi(s[14:17])\n\tif e != nil {\n\t\treturn t, errors.NewUtil(s, \"reading miliseconds\", e)\n\t}\n\tif ms < 0 || ms > 1000 {\n\t\treturn t, fmt.Errorf(\"invalid miliseconds %d\", ms)\n\t}\n\tl, e := time.LoadLocation(\"Local\")\n\tif e != nil {\n\t\treturn t, errors.NewUtil(\"\", \"loading location\", e)\n\t}\n\treturn time.Date(y, time.Month(m), d, h, mi, sc, ms*1000000, l), nil\n}", "func IsUnixTime(str string) bool {\n\tif _, err := strconv.Atoi(str); err == nil {\n\t\treturn true\n\t}\n\treturn false\n}", "func unixMilli(msec int64) time.Time {\n\treturn time.Unix(msec/1e3, (msec%1e3)*1e6)\n}", "func Parse2(s string) (int64, error) { return parse(s, units2) }", "func to_ms(nano int64) int64 {\n\treturn nano / int64(time.Millisecond)\n}", "func msToTime(ms string) (time.Time, error) {\n\tmsInt, err := strconv.ParseInt(ms, 10, 64)\n\tif err != nil {\n\t\t// todo: support-scheduler will be removed later issue_650a\n\t\tt, err := time.Parse(SCHEDULER_TIMELAYOUT, ms)\n\t\tif err == nil {\n\t\t\treturn t, nil\n\t\t}\n\t\treturn time.Time{}, err\n\t}\n\n\treturn time.Unix(0, msInt*int64(time.Millisecond)), nil\n}", "func SSAtoMS(t string) int {\n\th, m, s, cs := ssatoSplit(t)\n\treturn (h*Hour + m*Minute + s*Second + cs*Centisecond)\n}", "func unpackTime(msg *[]byte) int64 {\n\tresult := int64(0)\n\n\tfor i := 0; i < 8; i++ {\n\t\tresult |= (int64((*msg)[i+4]) << uint(i*8))\n\t}\n\n\treturn result\n}", "func (ts Timespec) Unix() (sec int64, nsec int64) {\n\treturn int64(ts.Sec), int64(ts.Nsec)\n}", "func unixMilli(t time.Time) int64 {\n\treturn t.UnixNano() / int64(time.Millisecond)\n}", "func strToTime(str string) (int, int, float64, bool, bool, formulaArg) {\n\tvar subMatch []string\n\tpattern := \"\"\n\tfor key, tf := range timeFormats {\n\t\tsubMatch = tf.FindStringSubmatch(str)\n\t\tif len(subMatch) > 1 {\n\t\t\tpattern = key\n\t\t\tbreak\n\t\t}\n\t}\n\tif pattern == \"\" {\n\t\treturn 0, 0, 0, false, false, newErrorFormulaArg(formulaErrorVALUE, formulaErrorVALUE)\n\t}\n\tdateIsEmpty := subMatch[1] == \"\"\n\tsubMatch = subMatch[49:]\n\tvar (\n\t\tl = len(subMatch)\n\t\tlast = subMatch[l-1]\n\t\tam = last == \"am\"\n\t\tpm = last == \"pm\"\n\t\thours, minutes int\n\t\tseconds float64\n\t\terr error\n\t)\n\tif handler, ok := map[string]func(match []string) (int, int, float64, error){\n\t\t\"hh\": strToTimePatternHandler1,\n\t\t\"hh:mm\": strToTimePatternHandler2,\n\t\t\"mm:ss\": strToTimePatternHandler3,\n\t\t\"hh:mm:ss\": strToTimePatternHandler4,\n\t}[pattern]; ok {\n\t\tif hours, minutes, seconds, err = handler(subMatch); err != nil {\n\t\t\treturn 0, 0, 0, false, false, newErrorFormulaArg(formulaErrorVALUE, formulaErrorVALUE)\n\t\t}\n\t}\n\tif minutes >= 60 {\n\t\treturn 0, 0, 0, false, false, newErrorFormulaArg(formulaErrorVALUE, formulaErrorVALUE)\n\t}\n\tif am || pm {\n\t\tif hours > 12 || seconds >= 60 {\n\t\t\treturn 0, 0, 0, false, false, newErrorFormulaArg(formulaErrorVALUE, formulaErrorVALUE)\n\t\t} else if hours == 12 {\n\t\t\thours = 0\n\t\t}\n\t} else if hours >= 24 || seconds >= 10000 {\n\t\treturn 0, 0, 0, false, false, newErrorFormulaArg(formulaErrorVALUE, formulaErrorVALUE)\n\t}\n\treturn hours, minutes, seconds, pm, dateIsEmpty, newEmptyFormulaArg()\n}", "func timeFromJournalInt(t int64) time.Time {\n\tsecs := t / 1000000\n\tms := t % 1000000\n\treturn time.Unix(secs, ms).UTC()\n}", "func String2Time(stime string) time.Time {\n\ttimeObj, err := time.Parse(\"2006-01-02 15:04:05\", stime)\n\tif err != nil {\n\t\tpanic(err)\n\t\treturn time.Now()\n\t}\n\treturn timeObj\n}", "func NSToTime(ns int64) time.Time {\n\treturn time.Unix(ns/1e9, ns%1e9)\n}", "func strToInt64Point(s string) (int64, error) {\n\treturn strconv.ParseInt(s, 10, 64)\n}", "func ParseTimestamp(timestamp int64, unit TimeUnit) (time.Time, error) {\n\tswitch unit {\n\tcase UnitSeconds:\n\t\treturn time.Unix(timestamp, 0), nil\n\tcase UnitMilliseconds:\n\t\t// add digits to match nanosecond accuracy\n\t\ttimestamp *= 1000 * 1000\n\t\treturn time.Unix(0, timestamp), nil\n\tcase UnitMicroseconds:\n\t\t// add digits to match nanosecond accuracy\n\t\ttimestamp *= 1000\n\t\treturn time.Unix(0, timestamp), nil\n\tcase UnitNanoseconds:\n\t\treturn time.Unix(0, timestamp), nil\n\tdefault:\n\t\treturn time.Time{}, fmt.Errorf(\"unknown unit '%v'\", unit)\n\t}\n}", "func Str2TimeS(s string) time.Time {\n\tloc, _ := time.LoadLocation(\"UTC\")\n\tt, err := time.ParseInLocation(DATE_FORMAT_SHORT, s, loc)\n\tCheckErr(err, CHECK_FLAG_LOGONLY)\n\treturn t\n}", "func convertTimeStamp(timestamp uint64) int64 {\n\treturn int64(timestamp / uint64(int64(time.Millisecond)/int64(time.Nanosecond)))\n}", "func Unix(sec int64, nsec int64) *Time {\n\treturn &Time{time.Unix(sec, nsec)}\n}", "func Unix(sec int64, nsec int64) *time.Time {\n\tt := time.Unix(sec, nsec)\n\treturn &t\n}", "func parseTimestamp(ts interface{}) (time.Time, bool) {\n\tswitch ts.(type) {\n\tcase float64:\n\t\treturn time.Unix(int64(ts.(float64)), 0), true\n\tcase int64:\n\t\treturn time.Unix(ts.(int64), 0), true\n\tcase string:\n\t\tif tm, err := strconv.ParseInt(ts.(string), 10, 64); err == nil {\n\t\t\treturn time.Unix(tm, 0), true\n\t\t} else {\n\t\t\tlogging.Warn(\"parseTimestamp: %s\", err)\n\t\t}\n\t}\n\treturn time.Now(), false\n}", "func Parse(layout, value string) (Time, error) {}", "func parseIntoDuration(str string) (time.Duration, error) {\n\tvar d time.Duration\n\t/**\n\t * important! When editing this regex, make sure that you specify the \"or\"s as\n\t * whole -> subset instead of subset -> whole, that is \"second|sec|s\" instead of\n\t * \"s|sec|second\". Otherwise, you will find yourself matching \"s\", but with a tailing\n\t * \"econd\"\n\t**/\n\tre := regexp.MustCompile(\"([-+][0-9]+)(hour|hr|h|minute|min|m|second|sec|s|days|day|d)\")\n\tres := re.FindAllStringSubmatch(str, -1)\n\tif len(res) != 1 {\n\t\treturn d, errors.New(\"Invalid timespec: \" + str)\n\t}\n\n\t// handle amount\n\ti, err := strconv.ParseInt(res[0][1], 10, 64)\n\tif err != nil {\n\t\treturn d, err\n\t}\n\td = time.Duration(i)\n\n\t// handle units\n\tswitch res[0][2] {\n\tcase \"h\", \"hr\", \"hour\":\n\t\td *= time.Hour\n\tcase \"m\", \"min\", \"minute\":\n\t\td *= time.Minute\n\tcase \"s\", \"sec\", \"second\":\n\t\td *= time.Second\n\tcase \"d\", \"days\", \"day\":\n\t\td *= 24 * time.Hour\n\tdefault:\n\t\treturn d, errors.New(\"Timespec needs valid units:\" + str)\n\t}\n\n\treturn d, nil\n}", "func run_timeNano() int64", "func ParseEpoch(epoch string) (*time.Time, error) {\n\ti, err := strconv.ParseInt(epoch, 10, 64)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttm := time.Unix(i, 0)\n\treturn &tm, nil\n}" ]
[ "0.7113885", "0.7021678", "0.6913968", "0.6790761", "0.67688674", "0.6602016", "0.65225476", "0.6439093", "0.6437345", "0.6419058", "0.6402529", "0.6263361", "0.6241682", "0.62355626", "0.6205712", "0.6192411", "0.61783", "0.61746967", "0.61625093", "0.61457837", "0.6097658", "0.6092855", "0.60791504", "0.60219014", "0.6020876", "0.59797573", "0.59786093", "0.5948067", "0.59287846", "0.59228057", "0.58929235", "0.5884475", "0.58643126", "0.5855785", "0.5850783", "0.5817403", "0.58147943", "0.5794522", "0.57932043", "0.57688653", "0.57545984", "0.57495964", "0.5731815", "0.57240325", "0.5684751", "0.5679396", "0.56534255", "0.5648858", "0.56438166", "0.56334645", "0.5630788", "0.5610161", "0.5605639", "0.559066", "0.5583086", "0.5583086", "0.5581303", "0.55776215", "0.5570408", "0.5543974", "0.5543974", "0.5543974", "0.5543974", "0.55214113", "0.55150926", "0.5475667", "0.5471455", "0.5466299", "0.5464258", "0.5460122", "0.54559755", "0.543241", "0.5428042", "0.54244524", "0.54144824", "0.54049885", "0.53993857", "0.5387003", "0.5378253", "0.5374258", "0.53686404", "0.53650117", "0.5354804", "0.53434014", "0.5341325", "0.53354424", "0.53323644", "0.52966404", "0.5296025", "0.5291736", "0.52917105", "0.52901083", "0.52897406", "0.5286448", "0.52797896", "0.5275199", "0.5273124", "0.52728873", "0.5270817", "0.5270598" ]
0.7112144
1
Flatten generates a flat map from a nested one. The original may include values of type map, slice and scalar, but not struct. Keys in the flat map will be a compound of descending map keys and slice iterations. The presentation of keys is set by style. A prefix is joined to each key.
func Flatten_(nested map[string]interface{}, prefix string, style SeparatorStyle) (map[string]interface{}, error) { flatmap := make(map[string]interface{}) err := flatten(true, flatmap, nested, prefix, style) if err != nil { return nil, err } return flatmap, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Flatten(nested map[string]interface{}, prefix string, style SeparatorStyle) (map[string]interface{}, error) {\n\tflatmap := make(map[string]interface{})\n\n\terr := flatten(true, flatmap, nested, prefix, style)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn flatmap, nil\n}", "func Flatten(nested map[string]interface{}) (flatmap map[string]interface{}, err error) {\n\treturn flatten(\"\", nested)\n}", "func NestedToFlatMap(input map[string]map[string]interface{}) map[string]interface{} {\n\tret := make(map[string]interface{})\n\tfor k, v := range input {\n\t\tret[k] = interface{}(v)\n\t}\n\treturn ret\n}", "func FlattenMap(data map[string]interface{}, delimiter string) {\n\tfor k, vi := range data {\n\t\tif v2i, ok := vi.(map[string]interface{}); ok {\n\t\t\tFlattenMap(v2i, delimiter)\n\t\t\tfor k3, v3i := range v2i {\n\t\t\t\tdata[k+delimiter+k3] = v3i\n\t\t\t}\n\t\t\tdelete(data, k)\n\t\t}\n\t}\n}", "func FlatToNestedMap(input map[string]interface{}) (map[string]map[string]interface{}, error) {\n\tret := make(map[string]map[string]interface{})\n\tfor k, v := range input {\n\t\tif vMap, ok := v.(map[string]interface{}); ok {\n\t\t\tret[k] = vMap\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"Key '%s' is not a map\", k)\n\t\t}\n\t}\n\treturn ret, nil\n}", "func flatten(data map[string]map[string]specs) map[string]specs {\n\tresult := make(map[string]specs)\n\tfor _, m := range data {\n\t\tfor iType, specs := range m {\n\t\t\tresult[iType] = specs\n\t\t}\n\t}\n\treturn result\n}", "func flattenNestedMap(flatMap map[string]interface{}, nestedMap interface{}) error {\n\tassign := func(newKey string, v interface{}) error {\n\t\tswitch v.(type) {\n\t\tcase map[string]interface{}:\n\t\t\tif err := flattenNestedMap(flatMap, v); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\tdefault:\n\t\t\tflatMap[newKey] = v\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tswitch nestedMap.(type) {\n\tcase map[string]interface{}:\n\t\tfor k, v := range nestedMap.(map[string]interface{}) {\n\t\t\tassign(k, v)\n\t\t}\n\tdefault:\n\t\treturn errors.New(\"Not a valid input, must be a map\")\n\t}\n\n\treturn nil\n}", "func Flatten(m map[string]interface{}, tokenizer func([]string) string) map[string]interface{} {\n\tvar r = make(map[string]interface{})\n\tflattenRecursive(m, []string{}, func(ks []string, v interface{}) {\n\t\tr[tokenizer(ks)] = v\n\t})\n\treturn r\n}", "func Flatten(obj interface{}) map[string]interface{} {\n\treturn flatten(reflect.ValueOf(obj))\n}", "func flattenJSON(in *map[string]interface{}, out *map[string]interface{}, prefix string) {\n\tfor key := range *in {\n\t\tprefixedKey := fmt.Sprintf(\"%s.%s\", prefix, key)\n\n\t\tif s, ok := (*in)[key].(string); ok {\n\t\t\t(*out)[\"JSONFlag\"] = true\n\t\t\t(*out)[prefixedKey] = s\n\t\t} else if m, ok := (*in)[key].(map[string]interface{}); ok {\n\t\t\tflattenJSON(&m, out, prefixedKey)\n\t\t} else if array, ok := (*in)[key].([]interface{}); ok {\n\t\t\tfor i, v := range array {\n\t\t\t\tif s, ok := v.(string); ok {\n\t\t\t\t\t(*out)[fmt.Sprintf(\"%s.%d\", prefixedKey, i)] = s\n\t\t\t\t} else if m, ok := v.(map[string]interface{}); ok {\n\t\t\t\t\tflattenJSON(&m, out, fmt.Sprintf(\"%s.%d\", prefixedKey, i))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func Unflatten(flat map[string]interface{}) (nested map[string]interface{}, err error) {\n\tnested = make(map[string]interface{})\n\n\tfor k, v := range flat {\n\t\ttemp := uf(k, v).(map[string]interface{})\n\t\terr = mergo.Merge(&nested, temp)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\twalk(reflect.ValueOf(nested))\n\n\treturn\n}", "func (f *Flattener) Flatten(input interface{}, flattened *bson.D) error {\n\tf.flattened = flattened\n\tif f.Separator == nil {\n\t\tf.Separator = &defaultSeparator\n\t}\n\treturn f.flatten(input, \"\")\n}", "func flattenKey(c *Client, i interface{}, res *Key) *Key {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\tif len(m) == 0 {\n\t\treturn nil\n\t}\n\n\tresultRes := &Key{}\n\tresultRes.Name = dcl.SelfLinkToName(dcl.FlattenString(m[\"name\"]))\n\tresultRes.DisplayName = dcl.FlattenString(m[\"displayName\"])\n\tresultRes.WebSettings = flattenKeyWebSettings(c, m[\"webSettings\"], res)\n\tresultRes.AndroidSettings = flattenKeyAndroidSettings(c, m[\"androidSettings\"], res)\n\tresultRes.IosSettings = flattenKeyIosSettings(c, m[\"iosSettings\"], res)\n\tresultRes.Labels = dcl.FlattenKeyValuePairs(m[\"labels\"])\n\tresultRes.CreateTime = dcl.FlattenString(m[\"createTime\"])\n\tresultRes.TestingOptions = flattenKeyTestingOptions(c, m[\"testingOptions\"], res)\n\tresultRes.Project = dcl.FlattenString(m[\"project\"])\n\n\treturn resultRes\n}", "func Flatten(toFlatten interface{}) (result interface{}) {\n\ttemp := toFlatten.(map[string]interface{})\n\tif len(temp) > 1 {\n\t\tpanic(\"ndgo.Flatten:: flattened json has more than 1 item, operation not supported\")\n\t}\n\tfor _, item := range temp {\n\t\treturn item\n\t}\n\treturn nil\n}", "func flattenStoredInfoTypeDictionaryMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeDictionary {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeDictionary{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeDictionary{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeDictionary)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeDictionary(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func FlattenString(nestedstr, prefix string, style SeparatorStyle) (string, error) {\n\tif !isJsonMap.MatchString(nestedstr) {\n\t\treturn \"\", NotValidJsonInputError\n\t}\n\n\tvar nested map[string]interface{}\n\terr := json.Unmarshal([]byte(nestedstr), &nested)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tflatmap, err := Flatten(nested, prefix, style)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tflatb, err := json.Marshal(&flatmap)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(flatb), nil\n}", "func Flatten(input interface{}, flattened *bson.D, separator string) error {\n\tf := &Flattener{\n\t\tSeparator: &separator,\n\t}\n\treturn f.Flatten(input, flattened)\n}", "func (p QuantumSuperposition) Flatten() QuantumSuperposition {\n\tdict := Superposition{}\n\tfor _, l := range p.Dict {\n\t\tif _l, ok := l.([]Val); ok {\n\t\t\tfor k, v := range _l[0].(QuantumSuperposition).Dict {\n\t\t\t\tif _, ok := dict[k]; ok {\n\t\t\t\t\tdict[k] = dict[k].(complex128) + mult(v.(complex128), _l[1].(complex128))\n\t\t\t\t} else {\n\t\t\t\t\tif _v, ok := v.([]Val); ok {\n\t\t\t\t\t\tdict[k] = mult(_v[1].(complex128), _l[1].(complex128))\n\t\t\t\t\t} else {\n\t\t\t\t\t\tdict[k] = mult(v.(complex128), _l[1].(complex128))\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\treturn p\n\t\t}\n\t}\n\treturn new(QuantumSuperposition).From(dict)\n}", "func flattenStoredInfoTypeDictionaryCloudStoragePathMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeDictionaryCloudStoragePath {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeDictionaryCloudStoragePath{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeDictionaryCloudStoragePath{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeDictionaryCloudStoragePath)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeDictionaryCloudStoragePath(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenImageShieldedInstanceInitialStateKekMap(c *Client, i interface{}) map[string]ImageShieldedInstanceInitialStateKek {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageShieldedInstanceInitialStateKek{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageShieldedInstanceInitialStateKek{}\n\t}\n\n\titems := make(map[string]ImageShieldedInstanceInitialStateKek)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageShieldedInstanceInitialStateKek(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenKeyTestingOptionsMap(c *Client, i interface{}, res *Key) map[string]KeyTestingOptions {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]KeyTestingOptions{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]KeyTestingOptions{}\n\t}\n\n\titems := make(map[string]KeyTestingOptions)\n\tfor k, item := range a {\n\t\titems[k] = *flattenKeyTestingOptions(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeDictionary(c *Client, i interface{}, res *StoredInfoType) *StoredInfoTypeDictionary {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &StoredInfoTypeDictionary{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyStoredInfoTypeDictionary\n\t}\n\tr.WordList = flattenStoredInfoTypeDictionaryWordList(c, m[\"wordList\"], res)\n\tr.CloudStoragePath = flattenStoredInfoTypeDictionaryCloudStoragePath(c, m[\"cloudStoragePath\"], res)\n\n\treturn r\n}", "func flattenKeyIosSettingsMap(c *Client, i interface{}, res *Key) map[string]KeyIosSettings {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]KeyIosSettings{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]KeyIosSettings{}\n\t}\n\n\titems := make(map[string]KeyIosSettings)\n\tfor k, item := range a {\n\t\titems[k] = *flattenKeyIosSettings(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func nestMap(fMap FieldMap) FieldMap {\n\tres := make(FieldMap)\n\tnested := make(map[string]FieldMap)\n\tfor k, v := range fMap {\n\t\texprs := strings.Split(k, ExprSep)\n\t\tif len(exprs) == 1 {\n\t\t\t// We are in the top map here\n\t\t\tres[k] = v\n\t\t\tcontinue\n\t\t}\n\t\tif _, exists := nested[exprs[0]]; !exists {\n\t\t\tnested[exprs[0]] = make(FieldMap)\n\t\t}\n\t\tnested[exprs[0]][strings.Join(exprs[1:], ExprSep)] = v\n\t}\n\t// Get nested FieldMap and assign to top key\n\tfor k, fm := range nested {\n\t\tres[k] = fm\n\t}\n\treturn res\n}", "func nestMap(fMap FieldMap) FieldMap {\n\tres := make(FieldMap)\n\tnested := make(map[string]FieldMap)\n\tfor k, v := range fMap {\n\t\texprs := strings.Split(k, ExprSep)\n\t\tif len(exprs) == 1 {\n\t\t\t// We are in the top map here\n\t\t\tres[k] = v\n\t\t\tcontinue\n\t\t}\n\t\tif _, exists := nested[exprs[0]]; !exists {\n\t\t\tnested[exprs[0]] = make(FieldMap)\n\t\t}\n\t\tnested[exprs[0]][strings.Join(exprs[1:], ExprSep)] = v\n\t}\n\t// Get nested FieldMap and assign to top key\n\tfor k, fm := range nested {\n\t\tres[k] = fm\n\t}\n\treturn res\n}", "func Flatten(in *corev1.NFSVolumeSource) []interface{} {\n\tout := make([]interface{}, 1)\n\n\trow := map[string]interface{}{}\n\n\tif in.Server != \"\" {\n\t\trow[FieldServer] = in.Server\n\t}\n\n\tif in.Path != \"\" {\n\t\trow[FieldPath] = in.Path\n\t}\n\n\tout[0] = row\n\n\treturn out\n}", "func FlattenString(nestedstr, prefix string, style SeparatorStyle) (string, error) {\n\tvar nested map[string]interface{}\n\terr := json.Unmarshal([]byte(nestedstr), &nested)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tflatmap, err := Flatten_(nested, prefix, style)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tflatb, err := json.Marshal(&flatmap)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(flatb), nil\n}", "func (e Entry) flatten(m map[string]interface{}) {\n\tm[\"message\"] = e.Message\n\tm[\"severity\"] = e.Severity\n\tif e.Trace != \"\" {\n\t\tm[\"logging.googleapis.com/trace\"] = e.Trace\n\t}\n\tif e.Component != \"\" {\n\t\tm[\"component\"] = e.Component\n\t}\n\tif e.Fields != nil {\n\t\tfor k, v := range e.Fields {\n\t\t\tm[k] = v\n\t\t}\n\t}\n}", "func flattenPrivateCloudNsxMap(c *Client, i interface{}, res *PrivateCloud) map[string]PrivateCloudNsx {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]PrivateCloudNsx{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]PrivateCloudNsx{}\n\t}\n\n\titems := make(map[string]PrivateCloudNsx)\n\tfor k, item := range a {\n\t\titems[k] = *flattenPrivateCloudNsx(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeLargeCustomDictionaryMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeLargeCustomDictionary {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionary{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionary{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeLargeCustomDictionary)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeLargeCustomDictionary(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeRegexMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeRegex {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeRegex{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeRegex{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeRegex)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeRegex(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabelMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (sm seqMap) flatten(startAt string) []string {\n\tseen := make(map[string]bool)\n\treturn sm.flattenRecursive(startAt, seen)\n}", "func (a ApplyTo) Flatten() []schema.GroupVersionKind {\n\tvar result []schema.GroupVersionKind\n\tfor _, group := range a.Groups {\n\t\tfor _, version := range a.Versions {\n\t\t\tfor _, kind := range a.Kinds {\n\t\t\t\tgvk := schema.GroupVersionKind{\n\t\t\t\t\tGroup: group,\n\t\t\t\t\tVersion: version,\n\t\t\t\t\tKind: kind,\n\t\t\t\t}\n\t\t\t\tresult = append(result, gvk)\n\t\t\t}\n\t\t}\n\t}\n\treturn result\n}", "func BlankNestedMap(accumulator map[string]interface{}, elements []string) map[string]interface{} {\n\tif len(elements) == 1 {\n\t\taccumulator[elements[0]] = \"\"\n\t\treturn accumulator\n\t} else {\n\t\taccumulator[elements[0]] = BlankNestedMap(map[string]interface{}{}, elements[1:])\n\t\treturn accumulator\n\t}\n}", "func (e *ChefEnvironment) Flatten() map[string]interface{} {\n\treturn util.FlattenObj(e)\n}", "func flattenServiceCustomMap(c *Client, i interface{}, res *Service) map[string]ServiceCustom {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ServiceCustom{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ServiceCustom{}\n\t}\n\n\titems := make(map[string]ServiceCustom)\n\tfor k, item := range a {\n\t\titems[k] = *flattenServiceCustom(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenAttestorUserOwnedGrafeasNotePublicKeysMap(c *Client, i interface{}) map[string]AttestorUserOwnedGrafeasNotePublicKeys {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]AttestorUserOwnedGrafeasNotePublicKeys{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]AttestorUserOwnedGrafeasNotePublicKeys{}\n\t}\n\n\titems := make(map[string]AttestorUserOwnedGrafeasNotePublicKeys)\n\tfor k, item := range a {\n\t\titems[k] = *flattenAttestorUserOwnedGrafeasNotePublicKeys(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeDictionaryWordListMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeDictionaryWordList {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeDictionaryWordList{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeDictionaryWordList{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeDictionaryWordList)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeDictionaryWordList(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenImageShieldedInstanceInitialStatePkMap(c *Client, i interface{}) map[string]ImageShieldedInstanceInitialStatePk {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageShieldedInstanceInitialStatePk{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageShieldedInstanceInitialStatePk{}\n\t}\n\n\titems := make(map[string]ImageShieldedInstanceInitialStatePk)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageShieldedInstanceInitialStatePk(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenProjectParentMap(c *Client, i interface{}) map[string]ProjectParent {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ProjectParent{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ProjectParent{}\n\t}\n\n\titems := make(map[string]ProjectParent)\n\tfor k, item := range a {\n\t\titems[k] = *flattenProjectParent(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func FlatWithFunc(mp map[string]any, fn reflects.FlatFunc) {\n\tif mp == nil || fn == nil {\n\t\treturn\n\t}\n\treflects.FlatMap(reflect.ValueOf(mp), fn)\n}", "func flattenStoredInfoTypeLargeCustomDictionaryOutputPathMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeLargeCustomDictionaryOutputPath {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryOutputPath{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryOutputPath{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeLargeCustomDictionaryOutputPath)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeLargeCustomDictionaryOutputPath(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func Flatten(input interface{}) []interface{} {\n\tflat := make([]interface{}, 0)\n\tfor _, elem := range input.([]interface{}) {\n\t\tif inner, ok := elem.([]interface{}); ok {\n\t\t\tflat = append(flat, Flatten(inner)...)\n\t\t} else if elem != nil {\n\t\t\tflat = append(flat, elem)\n\t\t}\n\t}\n\treturn flat\n}", "func flattenStoredInfoTypeLargeCustomDictionaryBigQueryFieldTableMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldTable {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldTable{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldTable{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldTable)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeLargeCustomDictionaryBigQueryFieldTable(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenKeyWebSettingsMap(c *Client, i interface{}, res *Key) map[string]KeyWebSettings {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]KeyWebSettings{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]KeyWebSettings{}\n\t}\n\n\titems := make(map[string]KeyWebSettings)\n\tfor k, item := range a {\n\t\titems[k] = *flattenKeyWebSettings(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilterMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilter(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenDefaultObjectAccessControlProjectTeamMap(c *Client, i interface{}) map[string]DefaultObjectAccessControlProjectTeam {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]DefaultObjectAccessControlProjectTeam{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]DefaultObjectAccessControlProjectTeam{}\n\t}\n\n\titems := make(map[string]DefaultObjectAccessControlProjectTeam)\n\tfor k, item := range a {\n\t\titems[k] = *flattenDefaultObjectAccessControlProjectTeam(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenImageShieldedInstanceInitialStateMap(c *Client, i interface{}) map[string]ImageShieldedInstanceInitialState {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageShieldedInstanceInitialState{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageShieldedInstanceInitialState{}\n\t}\n\n\titems := make(map[string]ImageShieldedInstanceInitialState)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageShieldedInstanceInitialState(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenImageImageEncryptionKeyMap(c *Client, i interface{}) map[string]ImageImageEncryptionKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageImageEncryptionKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageImageEncryptionKey{}\n\t}\n\n\titems := make(map[string]ImageImageEncryptionKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageImageEncryptionKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeLargeCustomDictionaryBigQueryFieldMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeLargeCustomDictionaryBigQueryField {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryBigQueryField{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryBigQueryField{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeLargeCustomDictionaryBigQueryField)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeLargeCustomDictionaryBigQueryField(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenPrivateCloudHcxMap(c *Client, i interface{}, res *PrivateCloud) map[string]PrivateCloudHcx {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]PrivateCloudHcx{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]PrivateCloudHcx{}\n\t}\n\n\titems := make(map[string]PrivateCloudHcx)\n\tfor k, item := range a {\n\t\titems[k] = *flattenPrivateCloudHcx(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenInterconnectCircuitInfosMap(c *Client, i interface{}) map[string]InterconnectCircuitInfos {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]InterconnectCircuitInfos{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]InterconnectCircuitInfos{}\n\t}\n\n\titems := make(map[string]InterconnectCircuitInfos)\n\tfor k, item := range a {\n\t\titems[k] = *flattenInterconnectCircuitInfos(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenImageSourceSnapshotEncryptionKeyMap(c *Client, i interface{}) map[string]ImageSourceSnapshotEncryptionKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageSourceSnapshotEncryptionKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageSourceSnapshotEncryptionKey{}\n\t}\n\n\titems := make(map[string]ImageSourceSnapshotEncryptionKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageSourceSnapshotEncryptionKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenPrivateCloudNetworkConfigMap(c *Client, i interface{}, res *PrivateCloud) map[string]PrivateCloudNetworkConfig {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]PrivateCloudNetworkConfig{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]PrivateCloudNetworkConfig{}\n\t}\n\n\titems := make(map[string]PrivateCloudNetworkConfig)\n\tfor k, item := range a {\n\t\titems[k] = *flattenPrivateCloudNetworkConfig(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenUrlMapDefaultRouteActionUrlRewriteMap(c *Client, i interface{}) map[string]UrlMapDefaultRouteActionUrlRewrite {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapDefaultRouteActionUrlRewrite{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapDefaultRouteActionUrlRewrite{}\n\t}\n\n\titems := make(map[string]UrlMapDefaultRouteActionUrlRewrite)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapDefaultRouteActionUrlRewrite(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherPathRuleMap(c *Client, i interface{}) map[string]UrlMapPathMatcherPathRule {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherPathRule{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherPathRule{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherPathRule)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherPathRule(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (h *GrayLog) flatten(item map[string]interface{}, fields map[string]interface{}, id string) {\n\tif id != \"\" {\n\t\tid = id + \"_\"\n\t}\n\tfor k, i := range item {\n\t\tswitch i := i.(type) {\n\t\tcase int:\n\t\t\tfields[id+k] = float64(i)\n\t\tcase float64:\n\t\t\tfields[id+k] = i\n\t\tcase map[string]interface{}:\n\t\t\th.flatten(i, fields, id+k)\n\t\tdefault:\n\t\t}\n\t}\n}", "func flattenImageSourceDiskEncryptionKeyMap(c *Client, i interface{}) map[string]ImageSourceDiskEncryptionKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageSourceDiskEncryptionKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageSourceDiskEncryptionKey{}\n\t}\n\n\titems := make(map[string]ImageSourceDiskEncryptionKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageSourceDiskEncryptionKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenImageSourceImageEncryptionKeyMap(c *Client, i interface{}) map[string]ImageSourceImageEncryptionKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageSourceImageEncryptionKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageSourceImageEncryptionKey{}\n\t}\n\n\titems := make(map[string]ImageSourceImageEncryptionKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageSourceImageEncryptionKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenKeyAndroidSettingsMap(c *Client, i interface{}, res *Key) map[string]KeyAndroidSettings {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]KeyAndroidSettings{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]KeyAndroidSettings{}\n\t}\n\n\titems := make(map[string]KeyAndroidSettings)\n\tfor k, item := range a {\n\t\titems[k] = *flattenKeyAndroidSettings(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func normalizeTree(source map[string][]string) map[string][]string {\n\tnormalized := map[string][]string {}\n\n\tfor key, values := range source {\n\t\t// Copy the valid entry from the source map to the normalized map.\n\t\tnormalized[key] = values\n\t\tfor _, node := range values {\n\t\t\tif _, found := source[node]; !found {\n\t\t\t\t// Current node is in the slice, but not as a key in map.\n\t\t\t\t// This means we need to treat it as a leaf-node.\n\t\t\t\tnormalized[node] = []string{}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn normalized\n}", "func Unflatten(m map[string]interface{}, tf TokenizerFunc) map[string]interface{} {\n\ttree := make(map[string]interface{})\n\n\tc := make(chan map[string]interface{})\n\n\tgo mapify(m, c, tf)\n\n\tfor n := range c {\n\t\tmergo.Merge(&tree, n)\n\t}\n\n\treturn tree\n}", "func flattenUrlMapPathMatcherRouteRuleMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRule {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRule{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRule{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRule)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRule(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenPrivateCloudVcenterMap(c *Client, i interface{}, res *PrivateCloud) map[string]PrivateCloudVcenter {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]PrivateCloudVcenter{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]PrivateCloudVcenter{}\n\t}\n\n\titems := make(map[string]PrivateCloudVcenter)\n\tfor k, item := range a {\n\t\titems[k] = *flattenPrivateCloudVcenter(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenUrlMapDefaultUrlRedirectMap(c *Client, i interface{}) map[string]UrlMapDefaultUrlRedirect {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapDefaultUrlRedirect{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapDefaultUrlRedirect{}\n\t}\n\n\titems := make(map[string]UrlMapDefaultUrlRedirect)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapDefaultUrlRedirect(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenKeyTestingOptions(c *Client, i interface{}, res *Key) *KeyTestingOptions {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &KeyTestingOptions{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyKeyTestingOptions\n\t}\n\tr.TestingScore = dcl.FlattenDouble(m[\"testingScore\"])\n\tr.TestingChallenge = flattenKeyTestingOptionsTestingChallengeEnum(m[\"testingChallenge\"])\n\n\treturn r\n}", "func Flatten(given interface{}) []interface{} {\n\tout := []interface{}{}\n\tif slice, ok := given.([]interface{}); ok {\n\t\tfor _, x := range slice {\n\t\t\tout = append(out, Flatten(x)...)\n\t\t}\n\t\treturn out\n\t}\n\tif given != nil {\n\t\tout = []interface{}{given}\n\t}\n\treturn out\n}", "func Flatten(in *corev1.Probe) []interface{} {\n\tout := make([]interface{}, 1)\n\n\trow := map[string]interface{}{}\n\n\tif in.HTTPGet != nil {\n\t\trow[FieldHTTP] = http.Flatten(in.HTTPGet)\n\t}\n\n\tif in.TCPSocket != nil {\n\t\trow[FieldTCP] = tcp.Flatten(in.TCPSocket)\n\t}\n\n\tif in.Exec != nil {\n\t\trow[FieldExec] = exec.Flatten(in.Exec)\n\t}\n\n\tif in.InitialDelaySeconds > 0 {\n\t\trow[FieldInitialDelaySeconds] = in.InitialDelaySeconds\n\t}\n\n\tif in.PeriodSeconds > 0 {\n\t\trow[FieldPeriodSeconds] = in.PeriodSeconds\n\t}\n\n\tif in.TimeoutSeconds > 0 {\n\t\trow[FieldTimeoutSeconds] = in.TimeoutSeconds\n\t}\n\n\tif in.SuccessThreshold > 0 {\n\t\trow[FieldSuccessThreshold] = in.SuccessThreshold\n\t}\n\n\tif in.FailureThreshold > 0 {\n\t\trow[FieldFailureThreshold] = in.FailureThreshold\n\t}\n\n\tout[0] = row\n\n\treturn out\n}", "func flattenUrlMapPathMatcherMap(c *Client, i interface{}) map[string]UrlMapPathMatcher {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcher{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcher{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcher)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcher(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeLargeCustomDictionaryBigQueryFieldFieldMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldField {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldField{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldField{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeLargeCustomDictionaryBigQueryFieldField)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeLargeCustomDictionaryBigQueryFieldField(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenImageRawDiskMap(c *Client, i interface{}) map[string]ImageRawDisk {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageRawDisk{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageRawDisk{}\n\t}\n\n\titems := make(map[string]ImageRawDisk)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageRawDisk(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenAttestorUserOwnedGrafeasNotePublicKeysPkixPublicKeyMap(c *Client, i interface{}) map[string]AttestorUserOwnedGrafeasNotePublicKeysPkixPublicKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]AttestorUserOwnedGrafeasNotePublicKeysPkixPublicKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]AttestorUserOwnedGrafeasNotePublicKeysPkixPublicKey{}\n\t}\n\n\titems := make(map[string]AttestorUserOwnedGrafeasNotePublicKeysPkixPublicKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenAttestorUserOwnedGrafeasNotePublicKeysPkixPublicKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherPathRuleUrlRedirectMap(c *Client, i interface{}) map[string]UrlMapPathMatcherPathRuleUrlRedirect {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherPathRuleUrlRedirect{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherPathRuleUrlRedirect{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherPathRuleUrlRedirect)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherPathRuleUrlRedirect(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherPathRuleRouteActionRequestMirrorPolicyMap(c *Client, i interface{}) map[string]UrlMapPathMatcherPathRuleRouteActionRequestMirrorPolicy {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherPathRuleRouteActionRequestMirrorPolicy{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherPathRuleRouteActionRequestMirrorPolicy{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherPathRuleRouteActionRequestMirrorPolicy)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherPathRuleRouteActionRequestMirrorPolicy(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenAttestorUserOwnedDrydockNotePublicKeysMap(c *Client, i interface{}, res *Attestor) map[string]AttestorUserOwnedDrydockNotePublicKeys {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]AttestorUserOwnedDrydockNotePublicKeys{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]AttestorUserOwnedDrydockNotePublicKeys{}\n\t}\n\n\titems := make(map[string]AttestorUserOwnedDrydockNotePublicKeys)\n\tfor k, item := range a {\n\t\titems[k] = *flattenAttestorUserOwnedDrydockNotePublicKeys(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenKeyIosSettings(c *Client, i interface{}, res *Key) *KeyIosSettings {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &KeyIosSettings{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyKeyIosSettings\n\t}\n\tr.AllowAllBundleIds = dcl.FlattenBool(m[\"allowAllBundleIds\"])\n\tr.AllowedBundleIds = dcl.FlattenStringSlice(m[\"allowedBundleIds\"])\n\n\treturn r\n}", "func flattenAttestorUserOwnedGrafeasNoteMap(c *Client, i interface{}) map[string]AttestorUserOwnedGrafeasNote {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]AttestorUserOwnedGrafeasNote{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]AttestorUserOwnedGrafeasNote{}\n\t}\n\n\titems := make(map[string]AttestorUserOwnedGrafeasNote)\n\tfor k, item := range a {\n\t\titems[k] = *flattenAttestorUserOwnedGrafeasNote(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (f *Flattener) Flattened(input interface{}) (bson.D, error) {\n\tvar flattened bson.D\n\terr := f.Flatten(input, &flattened)\n\treturn flattened, err\n}", "func (b *BaseImpl) Flatten() {\n\n\tnbytes := make([]byte, b.LenBuf())\n\n\tcopy(nbytes, b.bytes)\n\tfor i := range b.Diffs {\n\t\tcopy(nbytes[b.Diffs[i].Offset:], b.Diffs[i].bytes)\n\t}\n\tb.bytes = nbytes\n\tb.Diffs = []Diff{}\n\n}", "func flattenStoredInfoTypeDictionaryCloudStoragePath(c *Client, i interface{}, res *StoredInfoType) *StoredInfoTypeDictionaryCloudStoragePath {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &StoredInfoTypeDictionaryCloudStoragePath{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyStoredInfoTypeDictionaryCloudStoragePath\n\t}\n\tr.Path = dcl.FlattenString(m[\"path\"])\n\n\treturn r\n}", "func flattenUrlMapDefaultRouteActionRequestMirrorPolicyMap(c *Client, i interface{}) map[string]UrlMapDefaultRouteActionRequestMirrorPolicy {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapDefaultRouteActionRequestMirrorPolicy{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapDefaultRouteActionRequestMirrorPolicy{}\n\t}\n\n\titems := make(map[string]UrlMapDefaultRouteActionRequestMirrorPolicy)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapDefaultRouteActionRequestMirrorPolicy(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeLargeCustomDictionary(c *Client, i interface{}, res *StoredInfoType) *StoredInfoTypeLargeCustomDictionary {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &StoredInfoTypeLargeCustomDictionary{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyStoredInfoTypeLargeCustomDictionary\n\t}\n\tr.OutputPath = flattenStoredInfoTypeLargeCustomDictionaryOutputPath(c, m[\"outputPath\"], res)\n\tr.CloudStorageFileSet = flattenStoredInfoTypeLargeCustomDictionaryCloudStorageFileSet(c, m[\"cloudStorageFileSet\"], res)\n\tr.BigQueryField = flattenStoredInfoTypeLargeCustomDictionaryBigQueryField(c, m[\"bigQueryField\"], res)\n\n\treturn r\n}", "func NormalizeMap(zv zcode.Bytes) zcode.Bytes {\n\telements := make([]keyval, 0, 8)\n\tfor it := zv.Iter(); !it.Done(); {\n\t\tkey, _, err := it.NextTagAndBody()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tval, _, err := it.NextTagAndBody()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\telements = append(elements, keyval{key, val})\n\t}\n\tif len(elements) < 2 {\n\t\treturn zv\n\t}\n\tsort.Slice(elements, func(i, j int) bool {\n\t\treturn bytes.Compare(elements[i].key, elements[j].key) == -1\n\t})\n\tnorm := make(zcode.Bytes, 0, len(zv))\n\tnorm = append(norm, elements[0].key...)\n\tnorm = append(norm, elements[0].val...)\n\tfor i := 1; i < len(elements); i++ {\n\t\t// Skip duplicates.\n\t\tif !bytes.Equal(elements[i].key, elements[i-1].key) {\n\t\t\tnorm = append(norm, elements[i].key...)\n\t\t\tnorm = append(norm, elements[i].val...)\n\t\t}\n\t}\n\treturn norm\n}", "func Flatten(nested interface{}) []interface{} {\n\tflattened := []interface{}{}\n\tnestedSlice, ok := nested.([]interface{})\n\tif !ok {\n\t\treturn flattened\n\t}\n\tfor _, e := range nestedSlice {\n\t\tt := reflect.TypeOf(e)\n\t\tif t == nil {\n\t\t\tcontinue\n\t\t}\n\t\tswitch kind := t.Kind(); kind {\n\t\tcase reflect.Slice:\n\t\t\tflattened = append(flattened, Flatten(e)...)\n\t\tdefault:\n\t\t\tflattened = append(flattened, e)\n\t\t}\n\t}\n\treturn flattened\n}", "func (e EdgeMetadatas) Flatten() EdgeMetadata {\n\tresult := EdgeMetadata{}\n\tfor _, v := range e {\n\t\tresult = result.Flatten(v)\n\t}\n\treturn result\n}", "func flattenUrlMapHeaderActionRequestHeadersToAddMap(c *Client, i interface{}) map[string]UrlMapHeaderActionRequestHeadersToAdd {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapHeaderActionRequestHeadersToAdd{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapHeaderActionRequestHeadersToAdd{}\n\t}\n\n\titems := make(map[string]UrlMapHeaderActionRequestHeadersToAdd)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapHeaderActionRequestHeadersToAdd(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherRouteRuleRouteActionRequestMirrorPolicyMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleRouteActionRequestMirrorPolicy {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleRouteActionRequestMirrorPolicy{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleRouteActionRequestMirrorPolicy{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleRouteActionRequestMirrorPolicy)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleRouteActionRequestMirrorPolicy(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherDefaultUrlRedirectMap(c *Client, i interface{}) map[string]UrlMapPathMatcherDefaultUrlRedirect {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherDefaultUrlRedirect{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherDefaultUrlRedirect{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherDefaultUrlRedirect)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherDefaultUrlRedirect(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherPathRuleRouteActionUrlRewriteMap(c *Client, i interface{}) map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherPathRuleRouteActionUrlRewrite(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMap(c *Client, i interface{}) *UrlMap {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\tif len(m) == 0 {\n\t\treturn nil\n\t}\n\n\tr := &UrlMap{}\n\tr.DefaultRouteAction = flattenUrlMapDefaultRouteAction(c, m[\"defaultRouteAction\"])\n\tr.DefaultService = dcl.FlattenString(m[\"defaultService\"])\n\tr.DefaultUrlRedirect = flattenUrlMapDefaultUrlRedirect(c, m[\"defaultUrlRedirect\"])\n\tr.Description = dcl.FlattenString(m[\"description\"])\n\tr.HeaderAction = flattenUrlMapHeaderAction(c, m[\"headerAction\"])\n\tr.HostRule = flattenUrlMapHostRuleSlice(c, m[\"hostRules\"])\n\tr.Name = dcl.FlattenString(m[\"name\"])\n\tr.PathMatcher = flattenUrlMapPathMatcherSlice(c, m[\"pathMatchers\"])\n\tr.Region = dcl.FlattenString(m[\"region\"])\n\tr.Test = flattenUrlMapTestSlice(c, m[\"tests\"])\n\tr.Project = dcl.FlattenString(m[\"project\"])\n\n\treturn r\n}", "func flattenUrlMapDefaultRouteActionRetryPolicyPerTryTimeoutMap(c *Client, i interface{}) map[string]UrlMapDefaultRouteActionRetryPolicyPerTryTimeout {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapDefaultRouteActionRetryPolicyPerTryTimeout{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapDefaultRouteActionRetryPolicyPerTryTimeout{}\n\t}\n\n\titems := make(map[string]UrlMapDefaultRouteActionRetryPolicyPerTryTimeout)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapDefaultRouteActionRetryPolicyPerTryTimeout(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherRouteRuleMatchRuleMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleMatchRule {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRule{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRule{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleMatchRule)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleMatchRule(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenInterconnectInterconnectTypeEnumMap(c *Client, i interface{}) map[string]InterconnectInterconnectTypeEnum {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]InterconnectInterconnectTypeEnum{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]InterconnectInterconnectTypeEnum{}\n\t}\n\n\titems := make(map[string]InterconnectInterconnectTypeEnum)\n\tfor k, item := range a {\n\t\titems[k] = *flattenInterconnectInterconnectTypeEnum(item.(interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherRouteRuleUrlRedirectMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleUrlRedirect {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleUrlRedirect{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleUrlRedirect{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleUrlRedirect)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleUrlRedirect(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (g Gen) FlatMap(f func(interface{}) Gen) Gen {\n\treturn func(genParams *GenParameters) *GenResult {\n\t\tresult := g(genParams)\n\t\tvalue, ok := result.Retrieve()\n\t\tif ok {\n\t\t\treturn f(value)(genParams)\n\t\t}\n\t\tmappedZero := f(reflect.Zero(result.ResultType).Interface())(genParams)\n\t\treturn &GenResult{\n\t\t\tShrinker: NoShrinker,\n\t\t\tresult: nil,\n\t\t\tLabels: result.Labels,\n\t\t\tResultType: mappedZero.ResultType,\n\t\t}\n\t}\n}", "func (d *dataUsageCache) flatten(root dataUsageEntry) dataUsageEntry {\n\tfor id := range root.Children {\n\t\te := d.Cache[id]\n\t\tif len(e.Children) > 0 {\n\t\t\te = d.flatten(e)\n\t\t}\n\t\troot.merge(e)\n\t}\n\troot.Children = nil\n\treturn root\n}", "func flattenInstanceNetworksMap(c *Client, i interface{}) map[string]InstanceNetworks {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]InstanceNetworks{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]InstanceNetworks{}\n\t}\n\n\titems := make(map[string]InstanceNetworks)\n\tfor k, item := range a {\n\t\titems[k] = *flattenInstanceNetworks(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenServiceTelemetryMap(c *Client, i interface{}, res *Service) map[string]ServiceTelemetry {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ServiceTelemetry{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ServiceTelemetry{}\n\t}\n\n\titems := make(map[string]ServiceTelemetry)\n\tfor k, item := range a {\n\t\titems[k] = *flattenServiceTelemetry(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func Flatten(i interface{}) []interface{} {\n\tres := []interface{}{}\n\tswitch t := i.(type) {\n\tcase []interface{}:\n\t\tfor _, item := range t {\n\t\t\tres = append(res, Flatten(item)...)\n\t\t}\n\tcase interface{}:\n\t\tres = append(res, t)\n\t}\n\treturn res\n}" ]
[ "0.7979032", "0.72060883", "0.67485243", "0.6262294", "0.6236859", "0.6156783", "0.60582453", "0.5954483", "0.5859087", "0.5809873", "0.57548505", "0.5574524", "0.55396956", "0.54802305", "0.54689723", "0.53855795", "0.5384726", "0.53739184", "0.53649014", "0.53499806", "0.5341637", "0.5299707", "0.5291604", "0.52659047", "0.52659047", "0.5239705", "0.5238765", "0.5227316", "0.5222273", "0.5217828", "0.5216632", "0.518686", "0.51659477", "0.5143728", "0.5120736", "0.5080306", "0.50673074", "0.5064501", "0.5048545", "0.50480235", "0.5032001", "0.50135267", "0.50096935", "0.49999112", "0.4994429", "0.49852854", "0.49393132", "0.4935146", "0.4928221", "0.4907645", "0.4899542", "0.4891208", "0.48816296", "0.48767847", "0.48730317", "0.48686847", "0.48666146", "0.48593637", "0.48435044", "0.4841692", "0.48389074", "0.48374626", "0.48278108", "0.48274404", "0.48257655", "0.4812468", "0.4799832", "0.47913218", "0.47829056", "0.4776245", "0.47737065", "0.47670516", "0.4761332", "0.47452635", "0.47426197", "0.4733654", "0.47293118", "0.47258604", "0.4719339", "0.47154003", "0.4709039", "0.47061652", "0.47012404", "0.46914583", "0.46859926", "0.46858957", "0.4682685", "0.46783948", "0.46775258", "0.46711156", "0.46695155", "0.46496296", "0.46449953", "0.46403188", "0.4630884", "0.4629196", "0.46269703", "0.46240675", "0.46176618", "0.4606791" ]
0.7336597
1
FlattenString generates a flat JSON map from a nested one. Keys in the flat map will be a compound of descending map keys and slice iterations. The presentation of keys is set by style. A prefix is joined to each key.
func FlattenString(nestedstr, prefix string, style SeparatorStyle) (string, error) { var nested map[string]interface{} err := json.Unmarshal([]byte(nestedstr), &nested) if err != nil { return "", err } flatmap, err := Flatten_(nested, prefix, style) if err != nil { return "", err } flatb, err := json.Marshal(&flatmap) if err != nil { return "", err } return string(flatb), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func FlattenString(nestedstr, prefix string, style SeparatorStyle) (string, error) {\n\tif !isJsonMap.MatchString(nestedstr) {\n\t\treturn \"\", NotValidJsonInputError\n\t}\n\n\tvar nested map[string]interface{}\n\terr := json.Unmarshal([]byte(nestedstr), &nested)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tflatmap, err := Flatten(nested, prefix, style)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tflatb, err := json.Marshal(&flatmap)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(flatb), nil\n}", "func Flatten(nested map[string]interface{}, prefix string, style SeparatorStyle) (map[string]interface{}, error) {\n\tflatmap := make(map[string]interface{})\n\n\terr := flatten(true, flatmap, nested, prefix, style)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn flatmap, nil\n}", "func Flatten_(nested map[string]interface{}, prefix string, style SeparatorStyle) (map[string]interface{}, error) {\n\tflatmap := make(map[string]interface{})\n\n\terr := flatten(true, flatmap, nested, prefix, style)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn flatmap, nil\n}", "func Flatten(nested map[string]interface{}) (flatmap map[string]interface{}, err error) {\n\treturn flatten(\"\", nested)\n}", "func Flatten(m map[string]interface{}, tokenizer func([]string) string) map[string]interface{} {\n\tvar r = make(map[string]interface{})\n\tflattenRecursive(m, []string{}, func(ks []string, v interface{}) {\n\t\tr[tokenizer(ks)] = v\n\t})\n\treturn r\n}", "func flattenJSON(in *map[string]interface{}, out *map[string]interface{}, prefix string) {\n\tfor key := range *in {\n\t\tprefixedKey := fmt.Sprintf(\"%s.%s\", prefix, key)\n\n\t\tif s, ok := (*in)[key].(string); ok {\n\t\t\t(*out)[\"JSONFlag\"] = true\n\t\t\t(*out)[prefixedKey] = s\n\t\t} else if m, ok := (*in)[key].(map[string]interface{}); ok {\n\t\t\tflattenJSON(&m, out, prefixedKey)\n\t\t} else if array, ok := (*in)[key].([]interface{}); ok {\n\t\t\tfor i, v := range array {\n\t\t\t\tif s, ok := v.(string); ok {\n\t\t\t\t\t(*out)[fmt.Sprintf(\"%s.%d\", prefixedKey, i)] = s\n\t\t\t\t} else if m, ok := v.(map[string]interface{}); ok {\n\t\t\t\t\tflattenJSON(&m, out, fmt.Sprintf(\"%s.%d\", prefixedKey, i))\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func FlattenMap(data map[string]interface{}, delimiter string) {\n\tfor k, vi := range data {\n\t\tif v2i, ok := vi.(map[string]interface{}); ok {\n\t\t\tFlattenMap(v2i, delimiter)\n\t\t\tfor k3, v3i := range v2i {\n\t\t\t\tdata[k+delimiter+k3] = v3i\n\t\t\t}\n\t\t\tdelete(data, k)\n\t\t}\n\t}\n}", "func Flatten(records [][]string) string {\n\n\tvar firstRow = false\n\tvar response string\n\tfor _, row := range records {\n\t\tresponse = fmt.Sprintf(\"%s%s%s\", response, addComma(firstRow), strings.Join(row, \",\"))\n\t\tfirstRow = true\n\t}\n\treturn response\n}", "func (f *Flattener) Flatten(input interface{}, flattened *bson.D) error {\n\tf.flattened = flattened\n\tif f.Separator == nil {\n\t\tf.Separator = &defaultSeparator\n\t}\n\treturn f.flatten(input, \"\")\n}", "func flattenKey(c *Client, i interface{}, res *Key) *Key {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\tif len(m) == 0 {\n\t\treturn nil\n\t}\n\n\tresultRes := &Key{}\n\tresultRes.Name = dcl.SelfLinkToName(dcl.FlattenString(m[\"name\"]))\n\tresultRes.DisplayName = dcl.FlattenString(m[\"displayName\"])\n\tresultRes.WebSettings = flattenKeyWebSettings(c, m[\"webSettings\"], res)\n\tresultRes.AndroidSettings = flattenKeyAndroidSettings(c, m[\"androidSettings\"], res)\n\tresultRes.IosSettings = flattenKeyIosSettings(c, m[\"iosSettings\"], res)\n\tresultRes.Labels = dcl.FlattenKeyValuePairs(m[\"labels\"])\n\tresultRes.CreateTime = dcl.FlattenString(m[\"createTime\"])\n\tresultRes.TestingOptions = flattenKeyTestingOptions(c, m[\"testingOptions\"], res)\n\tresultRes.Project = dcl.FlattenString(m[\"project\"])\n\n\treturn resultRes\n}", "func NestedToFlatMap(input map[string]map[string]interface{}) map[string]interface{} {\n\tret := make(map[string]interface{})\n\tfor k, v := range input {\n\t\tret[k] = interface{}(v)\n\t}\n\treturn ret\n}", "func flatten(s string) string {\n\treturn strings.Replace((strings.Replace(s, \"\\n\", \"\", -1)), \"\\t\", \"\", -1)\n}", "func flattenNestedMap(flatMap map[string]interface{}, nestedMap interface{}) error {\n\tassign := func(newKey string, v interface{}) error {\n\t\tswitch v.(type) {\n\t\tcase map[string]interface{}:\n\t\t\tif err := flattenNestedMap(flatMap, v); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\tdefault:\n\t\t\tflatMap[newKey] = v\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tswitch nestedMap.(type) {\n\tcase map[string]interface{}:\n\t\tfor k, v := range nestedMap.(map[string]interface{}) {\n\t\t\tassign(k, v)\n\t\t}\n\tdefault:\n\t\treturn errors.New(\"Not a valid input, must be a map\")\n\t}\n\n\treturn nil\n}", "func Flatten(input interface{}, flattened *bson.D, separator string) error {\n\tf := &Flattener{\n\t\tSeparator: &separator,\n\t}\n\treturn f.Flatten(input, flattened)\n}", "func FlattenMessage(source *map[string]string, message string) {\n\tin, out := map[string]interface{}{}, map[string]interface{}{}\n\n\terr := json.Unmarshal([]byte(message), &in)\n\tif err != nil {\n\t\treturn\n\t}\n\tout[\"JSONFlag\"] = false\n\tflattenJSON(&in, &out, \"json_log_data\")\n\t// remove message if it has already been parsed into json_log_data\n\tif b, ok := out[\"JSONFlag\"].(bool); ok && b {\n\t\tdelete((*source), \"message\")\n\t}\n\tdelete(out, \"JSONFlag\")\n\tfor key := range out {\n\t\tif s, ok := out[key].(string); ok {\n\t\t\t(*source)[key] = s\n\t\t}\n\t}\n}", "func FormatMap(m map[string]string) (fmtStr string) {\n\t// output with keys in sorted order to provide stable output\n\tkeys := sets.NewString()\n\tfor key := range m {\n\t\tkeys.Insert(key)\n\t}\n\tfor _, key := range keys.List() {\n\t\tfmtStr += fmt.Sprintf(\"%v=%q\\n\", key, m[key])\n\t}\n\tfmtStr = strings.TrimSuffix(fmtStr, \"\\n\")\n\n\treturn\n}", "func Flatten(toFlatten interface{}) (result interface{}) {\n\ttemp := toFlatten.(map[string]interface{})\n\tif len(temp) > 1 {\n\t\tpanic(\"ndgo.Flatten:: flattened json has more than 1 item, operation not supported\")\n\t}\n\tfor _, item := range temp {\n\t\treturn item\n\t}\n\treturn nil\n}", "func FlatToNestedMap(input map[string]interface{}) (map[string]map[string]interface{}, error) {\n\tret := make(map[string]map[string]interface{})\n\tfor k, v := range input {\n\t\tif vMap, ok := v.(map[string]interface{}); ok {\n\t\t\tret[k] = vMap\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"Key '%s' is not a map\", k)\n\t\t}\n\t}\n\treturn ret, nil\n}", "func (sm seqMap) flatten(startAt string) []string {\n\tseen := make(map[string]bool)\n\treturn sm.flattenRecursive(startAt, seen)\n}", "func Flatten(obj interface{}) map[string]interface{} {\n\treturn flatten(reflect.ValueOf(obj))\n}", "func flatten(data map[string]map[string]specs) map[string]specs {\n\tresult := make(map[string]specs)\n\tfor _, m := range data {\n\t\tfor iType, specs := range m {\n\t\t\tresult[iType] = specs\n\t\t}\n\t}\n\treturn result\n}", "func TestShouldFlattenList(t *testing.T) {\n\tnested := []*NestedString{\n\t\t&NestedString{[]*NestedString{\n\t\t\t&NestedString{[]*NestedString{}, \"1.1\"},\n\t\t\t&NestedString{[]*NestedString{}, \"1.2\"},\n\t\t}, \"1\"},\n\t\t&NestedString{[]*NestedString{\n\t\t\t&NestedString{[]*NestedString{}, \"2.1\"},\n\t\t\t&NestedString{[]*NestedString{\n\t\t\t\t&NestedString{[]*NestedString{}, \"2.2.1\"},\n\t\t\t}, \"2.2\"},\n\t\t\t&NestedString{[]*NestedString{}, \"2.3\"},\n\t\t}, \"2\"},\n\t\t&NestedString{[]*NestedString{\n\t\t\t&NestedString{[]*NestedString{}, \"3.1\"},\n\t\t}, \"3\"},\n\t}\n\n\tflattened := Flatten(nested)\n\texpected := []string{\"1\", \"1.1\", \"1.2\", \"2\", \"2.1\", \"2.2\", \"2.2.1\", \"2.3\", \"3\", \"3.1\"}\n\n\tassert.Equal(t, expected, flattened, \"Should flatten nested list\")\n}", "func (testStringIdEntity_EntityInfo) Flatten(object interface{}, fbb *flatbuffers.Builder, id uint64) error {\n\n\t// build the FlatBuffers object\n\tfbb.StartObject(1)\n\tfbutils.SetUint64Slot(fbb, 0, id)\n\treturn nil\n}", "func (a *CalendarDate) Flatten() []string {\n\tvar sID string\n\tif a.Service != nil {\n\t\tsID = string(a.Service.ID)\n\t}\n\treturn []string{\n\t\t// service_id\n\t\tsID,\n\t\t// date\n\t\tstring(date.FormatDate(a.Date)),\n\t\t// exception_type\n\t\tstrconv.FormatInt(int64(a.ExceptionType), 10),\n\t}\n}", "func flattenUrlMap(c *Client, i interface{}) *UrlMap {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\tif len(m) == 0 {\n\t\treturn nil\n\t}\n\n\tr := &UrlMap{}\n\tr.DefaultRouteAction = flattenUrlMapDefaultRouteAction(c, m[\"defaultRouteAction\"])\n\tr.DefaultService = dcl.FlattenString(m[\"defaultService\"])\n\tr.DefaultUrlRedirect = flattenUrlMapDefaultUrlRedirect(c, m[\"defaultUrlRedirect\"])\n\tr.Description = dcl.FlattenString(m[\"description\"])\n\tr.HeaderAction = flattenUrlMapHeaderAction(c, m[\"headerAction\"])\n\tr.HostRule = flattenUrlMapHostRuleSlice(c, m[\"hostRules\"])\n\tr.Name = dcl.FlattenString(m[\"name\"])\n\tr.PathMatcher = flattenUrlMapPathMatcherSlice(c, m[\"pathMatchers\"])\n\tr.Region = dcl.FlattenString(m[\"region\"])\n\tr.Test = flattenUrlMapTestSlice(c, m[\"tests\"])\n\tr.Project = dcl.FlattenString(m[\"project\"])\n\n\treturn r\n}", "func resourceLibratoServicesFlatten(settings map[string]string) (string, error) {\n\tbyteArray, err := json.Marshal(settings)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"Error encoding to JSON: %s\", err)\n\t}\n\n\treturn string(byteArray), nil\n}", "func (c *StringValueMap) String() string {\n\tbuilder := \"\"\n\n\t// Todo: User encoder\n\tfor key := range c.value {\n\t\tvalue := c.value[key]\n\n\t\tif len(builder) > 0 {\n\t\t\tbuilder = builder + \";\"\n\t\t}\n\n\t\tif value != \"\" {\n\t\t\tbuilder = builder + fmt.Sprintf(\"%s=%s\", key, value)\n\t\t} else {\n\t\t\tbuilder = builder + key\n\t\t}\n\t}\n\n\treturn builder\n}", "func flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabelMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilterFilterLabel(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (c Context) String() string {\n\tkeys := make([]string, 0, len(c))\n\tfor key, _ := range c {\n\t\tkeys = append(keys, key)\n\t}\n\tsort.Strings(keys)\n\tfor i, key := range keys {\n\t\tkeys[i] = fmt.Sprintf(\"%v:%v\", key, c[key])\n\t}\n\treturn fmt.Sprintf(\"map[%v]\", strings.Join(keys, \" \"))\n}", "func MakeJSON(jsonMap map[string]string, key string) (s string) {\n\tvalue := jsonMap[key]\n\tif strings.HasPrefix(value, \"json:array \") {\n\t\tarray := strings.Split(value, \" \")\n\t\tarraysN := \"\"\n\t\tfor i := 1; i < len(array); i++ {\n\t\t\tnewKey := array[i]\n\t\t\tif key != \"\" {\n\t\t\t\tnewKey = key + \" \" + array[i]\n\t\t\t}\n\t\t\tt := MakeJSON(jsonMap, newKey)\n\t\t\tif arraysN != \"\" {\n\t\t\t\tarraysN = arraysN + \",\" + t\n\t\t\t} else {\n\t\t\t\tarraysN = t\n\t\t\t}\n\t\t}\n\t\ts = \"[\" + arraysN + \"]\"\n\t} else if strings.HasPrefix(value, \"json:object \") {\n\t\tarray := strings.Split(value, \" \")\n\t\tarraysN := \"\"\n\t\tfor i := 1; i < len(array); i++ {\n\t\t\tnewKey := array[i]\n\t\t\tif key != \"\" {\n\t\t\t\tnewKey = key + \" \" + array[i]\n\t\t\t}\n\t\t\tt := MakeJSON(jsonMap, newKey)\n\t\t\tif arraysN != \"\" {\n\t\t\t\tarraysN = arraysN + \",\" + `\"` + array[i] + `\":` + t\n\t\t\t} else {\n\t\t\t\tif array[i] != \"\" {\n\t\t\t\t\tarraysN = `\"` + array[i] + `\":` + t\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\ts = \"{\" + arraysN + \"}\"\n\t} else {\n\t\ts = value\n\t}\n\treturn s\n}", "func Unflatten(flat map[string]interface{}) (nested map[string]interface{}, err error) {\n\tnested = make(map[string]interface{})\n\n\tfor k, v := range flat {\n\t\ttemp := uf(k, v).(map[string]interface{})\n\t\terr = mergo.Merge(&nested, temp)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\twalk(reflect.ValueOf(nested))\n\n\treturn\n}", "func (e Entry) flatten(m map[string]interface{}) {\n\tm[\"message\"] = e.Message\n\tm[\"severity\"] = e.Severity\n\tif e.Trace != \"\" {\n\t\tm[\"logging.googleapis.com/trace\"] = e.Trace\n\t}\n\tif e.Component != \"\" {\n\t\tm[\"component\"] = e.Component\n\t}\n\tif e.Fields != nil {\n\t\tfor k, v := range e.Fields {\n\t\t\tm[k] = v\n\t\t}\n\t}\n}", "func toStringMap(m map[interface{}]interface{}) map[string]interface{} {\n\tresult := make(map[string]interface{})\n\tfor k, v := range m {\n\t\t// convert key to string\n\t\tk := fmt.Sprintf(\"%v\", k)\n\t\tif valMap, ok := v.(map[interface{}]interface{}); ok {\n\t\t\t// convert values recursively\n\t\t\tv = toStringMap(valMap)\n\t\t}\n\t\tresult[k] = v\n\t}\n\n\treturn result\n}", "func expand(s string, format ...string) string {\n\tvar (\n\t\tb bytes.Buffer\n\n\t\tcontainer string\n\t)\n\n\t// Handle the \"literal dot\" in a fieldname\n\ts = strings.Replace(s, `\\.`, `_~_|_~_`, -1)\n\n\tparts := strings.Split(s, \".\")\n\n\tif len(parts) > 0 {\n\t\tif reNumber.MatchString(parts[0]) {\n\t\t\tcontainer = \"slic\"\n\t\t} else {\n\t\t\tcontainer = \"mapi\"\n\t\t}\n\t} else {\n\t\tcontainer = \"mapi\"\n\t}\n\n\tb.WriteString(container)\n\n\tif len(parts) > 0 && parts[0] == \"\" {\n\t\treturn b.String()\n\t}\n\n\tfor i, v := range parts {\n\t\tif reNumber.MatchString(v) {\n\t\t\tif i > 0 {\n\t\t\t\tb.WriteString(`.([]interface{})`)\n\t\t\t}\n\t\t\tb.WriteString(`[`)\n\t\t\tb.WriteString(v)\n\t\t\tb.WriteString(`]`)\n\t\t} else {\n\t\t\tif i > 0 {\n\t\t\t\tif len(format) > 0 && format[0] == \"yaml\" {\n\t\t\t\t\tb.WriteString(`.(map[interface{}]interface{})`)\n\t\t\t\t} else {\n\t\t\t\t\tb.WriteString(`.(map[string]interface{})`)\n\t\t\t\t}\n\t\t\t}\n\t\t\tb.WriteString(`[\"`)\n\t\t\tb.WriteString(strings.Trim(v, `\"`)) // Remove the quotes from keys\n\t\t\tb.WriteString(`\"]`)\n\t\t}\n\n\t}\n\treturn strings.Replace(b.String(), `_~_|_~_`, `\\.`, -1)\n}", "func (e *ChefEnvironment) Flatten() map[string]interface{} {\n\treturn util.FlattenObj(e)\n}", "func flatten(entry logEntry) (string, error) {\n\tvar msgValue string\n\tvar errorValue error\n\tif len(entry.Values)%2 == 1 {\n\t\treturn \"\", errors.New(\"log entry cannot have odd number off keyAndValues\")\n\t}\n\n\tkeys := make([]string, 0, len(entry.Values)/2)\n\tvalues := make(map[string]interface{}, len(entry.Values)/2)\n\tfor i := 0; i < len(entry.Values); i += 2 {\n\t\tk, ok := entry.Values[i].(string)\n\t\tif !ok {\n\t\t\tpanic(fmt.Sprintf(\"key is not a string: %s\", entry.Values[i]))\n\t\t}\n\t\tvar v interface{}\n\t\tif i+1 < len(entry.Values) {\n\t\t\tv = entry.Values[i+1]\n\t\t}\n\t\tswitch k {\n\t\tcase \"msg\":\n\t\t\tmsgValue, ok = v.(string)\n\t\t\tif !ok {\n\t\t\t\tpanic(fmt.Sprintf(\"the msg value is not of type string: %s\", v))\n\t\t\t}\n\t\tcase \"error\":\n\t\t\terrorValue, ok = v.(error)\n\t\t\tif !ok {\n\t\t\t\tpanic(fmt.Sprintf(\"the error value is not of type error: %s\", v))\n\t\t\t}\n\t\tdefault:\n\t\t\tif _, ok := values[k]; !ok {\n\t\t\t\tkeys = append(keys, k)\n\t\t\t}\n\t\t\tvalues[k] = v\n\t\t}\n\t}\n\tstr := \"\"\n\tif entry.Prefix != \"\" {\n\t\tstr += fmt.Sprintf(\"[%s] \", entry.Prefix)\n\t}\n\tstr += msgValue\n\tif errorValue != nil {\n\t\tif msgValue != \"\" {\n\t\t\tstr += \": \"\n\t\t}\n\t\tstr += errorValue.Error()\n\t}\n\tfor _, k := range keys {\n\t\tprettyValue, err := pretty(values[k])\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tstr += fmt.Sprintf(\" %s=%s\", k, prettyValue)\n\t}\n\treturn str, nil\n}", "func flattenKeyTestingOptionsMap(c *Client, i interface{}, res *Key) map[string]KeyTestingOptions {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]KeyTestingOptions{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]KeyTestingOptions{}\n\t}\n\n\titems := make(map[string]KeyTestingOptions)\n\tfor k, item := range a {\n\t\titems[k] = *flattenKeyTestingOptions(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func BlankNestedMap(accumulator map[string]interface{}, elements []string) map[string]interface{} {\n\tif len(elements) == 1 {\n\t\taccumulator[elements[0]] = \"\"\n\t\treturn accumulator\n\t} else {\n\t\taccumulator[elements[0]] = BlankNestedMap(map[string]interface{}{}, elements[1:])\n\t\treturn accumulator\n\t}\n}", "func expandStringMap(v map[string]interface{}) map[string]string {\n\tm := make(map[string]string)\n\tfor key, val := range v {\n\t\tm[key] = val.(string)\n\t}\n\n\treturn m\n}", "func flattenUrlMapPathMatcherPathRuleMap(c *Client, i interface{}) map[string]UrlMapPathMatcherPathRule {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherPathRule{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherPathRule{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherPathRule)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherPathRule(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenImageImageEncryptionKey(c *Client, i interface{}) *ImageImageEncryptionKey {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &ImageImageEncryptionKey{}\n\tr.RawKey = dcl.FlattenString(m[\"rawKey\"])\n\tr.KmsKeyName = dcl.FlattenString(m[\"kmsKeyName\"])\n\tr.Sha256 = dcl.FlattenString(m[\"sha256\"])\n\tr.KmsKeyServiceAccount = dcl.FlattenString(m[\"kmsKeyServiceAccount\"])\n\n\treturn r\n}", "func (schema *Schema) String() (schemaString string) {\n\tlines := []string{} // Set our lines that we'll use to ensure newlines and the like\n\tsort.Strings(schema.Order) // Sort our order\n\n\tfor _, section := range schema.Order { // Use order so our sections are organized alphabetically\n\t\tkv := schema.Map[section] // Get our key/value\n\n\t\tif kv == nil || len(kv.Keys) == 0 { // No keys\n\t\t\tcontinue\n\t\t}\n\n\t\tsectionLabel := fmt.Sprintf(\"[%s]\", section) // Ensure we re-add [ and ]\n\n\t\tlines = append(lines, sectionLabel) // Add our section\n\n\t\tsort.Strings(kv.Order) // Order our SchemaKV\n\n\t\tfor _, orderedKey := range kv.Order { // For each of our ordered keys\n\t\t\tsT := kv.Keys[orderedKey]\n\t\t\tlines = append(lines, orderedKey+\"=\"+sT.String()) // Add in alphabetical order\n\t\t}\n\n\t\tlines = append(lines, \"\", \"\") // Add explicit new line after all our key/values\n\t}\n\n\tschemaString = strings.Join(lines, \"\\n\") // Join our strings, separated by newline\n\tschemaString = strings.ReplaceAll(schemaString, \"\\n\\n\", \"\\n\") // Replace double newlines with just one\n\n\treturn\n}", "func flattenStoredInfoTypeLargeCustomDictionaryOutputPathMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeLargeCustomDictionaryOutputPath {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryOutputPath{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeLargeCustomDictionaryOutputPath{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeLargeCustomDictionaryOutputPath)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeLargeCustomDictionaryOutputPath(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenStoredInfoTypeLargeCustomDictionaryOutputPath(c *Client, i interface{}, res *StoredInfoType) *StoredInfoTypeLargeCustomDictionaryOutputPath {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &StoredInfoTypeLargeCustomDictionaryOutputPath{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyStoredInfoTypeLargeCustomDictionaryOutputPath\n\t}\n\tr.Path = dcl.FlattenString(m[\"path\"])\n\n\treturn r\n}", "func flattenStoredInfoTypeRegexMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeRegex {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeRegex{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeRegex{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeRegex)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeRegex(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func flattenUrlMapDefaultRouteActionUrlRewriteMap(c *Client, i interface{}) map[string]UrlMapDefaultRouteActionUrlRewrite {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapDefaultRouteActionUrlRewrite{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapDefaultRouteActionUrlRewrite{}\n\t}\n\n\titems := make(map[string]UrlMapDefaultRouteActionUrlRewrite)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapDefaultRouteActionUrlRewrite(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (m *TMap) String() string {\n\tvar b strings.Builder\n\tfmt.Fprint(&b, \"{ \")\n\titer := m.Iterator()\n\tfor iter.HasNext() {\n\t\tentry := iter.NextEntry()\n\t\tfmt.Fprintf(&b, \"%s \", entry)\n\t}\n\tfmt.Fprint(&b, \"}\")\n\treturn b.String()\n}", "func flattenStoredInfoTypeDictionaryCloudStoragePathMap(c *Client, i interface{}, res *StoredInfoType) map[string]StoredInfoTypeDictionaryCloudStoragePath {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]StoredInfoTypeDictionaryCloudStoragePath{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]StoredInfoTypeDictionaryCloudStoragePath{}\n\t}\n\n\titems := make(map[string]StoredInfoTypeDictionaryCloudStoragePath)\n\tfor k, item := range a {\n\t\titems[k] = *flattenStoredInfoTypeDictionaryCloudStoragePath(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func BuildJson(unconstructedTypes map[string]UnconstructedTdlTypeIF) map[string]interface{} {\n\tformatted := getFormatted(unconstructedTypes)\n\n\tjson := make(map[string]interface{})\n\tfor path := range formatted {\n\t\tparents, leaf := decompose(path)\n\t\tparentJson := json\n\t\tfor _, parent := range parents {\n\t\t\tif _, ok := parentJson[parent]; !ok {\n\t\t\t\tparentJson[parent] = make(map[string]interface{})\n\t\t\t}\n\t\t\tparentJson = parentJson[parent].(map[string]interface{})\n\t\t}\n\t\tparentJson[leaf] = formatted[path]\n\t}\n\treturn json\n}", "func flattenStoredInfoTypeDictionaryCloudStoragePath(c *Client, i interface{}, res *StoredInfoType) *StoredInfoTypeDictionaryCloudStoragePath {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &StoredInfoTypeDictionaryCloudStoragePath{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyStoredInfoTypeDictionaryCloudStoragePath\n\t}\n\tr.Path = dcl.FlattenString(m[\"path\"])\n\n\treturn r\n}", "func Flatten(flatten bool) SecretOption {\n\treturn func(op *Options) {\n\t\top.Flatten = flatten\n\t}\n}", "func Expand(s string, mapping func(string) string) string", "func ToStringMapStringSlice(i interface{}) (map[string][]string, error) {\n\tvar m = map[string][]string{}\n\n\tswitch v := i.(type) {\n\tcase map[string][]string:\n\t\treturn v, nil\n\tcase map[string][]interface{}:\n\t\tfor k, val := range v {\n\t\t\tm[MustToString(k)] = MustToStringSlice(val)\n\t\t}\n\t\treturn m, nil\n\tcase map[string]string:\n\t\tfor k, val := range v {\n\t\t\tm[MustToString(k)] = []string{val}\n\t\t}\n\tcase map[string]interface{}:\n\t\tfor k, val := range v {\n\t\t\t//m[MustToString(k)] = []string{MustToString(val)}\n\t\t\tm[MustToString(k)] = MustToStringSlice(val)\n\t\t}\n\t\treturn m, nil\n\tcase map[interface{}][]string:\n\t\tfor k, val := range v {\n\t\t\tm[MustToString(k)] = MustToStringSlice(val)\n\t\t}\n\t\treturn m, nil\n\tcase map[interface{}]string:\n\t\tfor k, val := range v {\n\t\t\tm[MustToString(k)] = MustToStringSlice(val)\n\t\t}\n\t\treturn m, nil\n\tcase map[interface{}][]interface{}:\n\t\tfor k, val := range v {\n\t\t\tm[MustToString(k)] = MustToStringSlice(val)\n\t\t}\n\t\treturn m, nil\n\tcase map[interface{}]interface{}:\n\t\tfor k, val := range v {\n\t\t\tkey, err := ToString(k)\n\t\t\tif err != nil {\n\t\t\t\treturn m, fmt.Errorf(\"unable to cast %#v to map[string][]string\", i)\n\t\t\t}\n\t\t\tvalue, err := ToStringSlice(val)\n\t\t\tif err != nil {\n\t\t\t\treturn m, fmt.Errorf(\"unable to Cast %#v to map[string][]string\", i)\n\t\t\t}\n\t\t\tm[key] = value\n\n\t\t}\n\tdefault:\n\t\treturn m, fmt.Errorf(\"unable to Cast %#v to map[string][]string\", i)\n\t}\n\treturn m, nil\n}", "func flattenImageShieldedInstanceInitialStateKekMap(c *Client, i interface{}) map[string]ImageShieldedInstanceInitialStateKek {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageShieldedInstanceInitialStateKek{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageShieldedInstanceInitialStateKek{}\n\t}\n\n\titems := make(map[string]ImageShieldedInstanceInitialStateKek)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageShieldedInstanceInitialStateKek(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenKeyTestingOptions(c *Client, i interface{}, res *Key) *KeyTestingOptions {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &KeyTestingOptions{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyKeyTestingOptions\n\t}\n\tr.TestingScore = dcl.FlattenDouble(m[\"testingScore\"])\n\tr.TestingChallenge = flattenKeyTestingOptionsTestingChallengeEnum(m[\"testingChallenge\"])\n\n\treturn r\n}", "func flattenUrlMapPathMatcherPathRuleRouteActionUrlRewriteMap(c *Client, i interface{}) map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherPathRuleRouteActionUrlRewrite)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherPathRuleRouteActionUrlRewrite(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherRouteRuleMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRule {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRule{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRule{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRule)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRule(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapPathMatcherMap(c *Client, i interface{}) map[string]UrlMapPathMatcher {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcher{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcher{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcher)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcher(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenImageImageEncryptionKeyMap(c *Client, i interface{}) map[string]ImageImageEncryptionKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageImageEncryptionKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageImageEncryptionKey{}\n\t}\n\n\titems := make(map[string]ImageImageEncryptionKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageImageEncryptionKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (m Maps) String() string {\n\tjc, _ := json.Marshal(m)\n\treturn string(jc)\n}", "func (pm *PathMap) String() string {\n\tvar out strings.Builder\n\tout.WriteByte('{')\n\tnames := pm.pathnames()\n\tlastIdx := len(names) - 1\n\tfor idx, name := range names {\n\t\tvalue, _ := pm.get(name)\n\t\tfmt.Fprintf(&out, \"%s: %v\", name, value)\n\t\tif idx != lastIdx {\n\t\t\tout.WriteString(\", \")\n\t\t}\n\t}\n\tout.WriteByte('}')\n\treturn out.String()\n}", "func flattenUrlMapPathMatcherPathRuleUrlRedirectMap(c *Client, i interface{}) map[string]UrlMapPathMatcherPathRuleUrlRedirect {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherPathRuleUrlRedirect{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherPathRuleUrlRedirect{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherPathRuleUrlRedirect)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherPathRuleUrlRedirect(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func FormatString(value string, params map[string]string) string {\n // value should have at least 3 chars to be able to have key.\n if len(value) <= 2 {\n return value\n }\n\n // To avoid mistake when passing Upper or lower case, the params' keys need to be convert to lower case all.\n for k, v := range params {\n delete(params, k)\n params[strings.ToLower(k)] = v\n }\n\n // The result string can be use strings.Builder to process a large value string.\n var result, key string\n index := 0\n size := len(value)\n\n for index < size {\n currentChar := fmt.Sprintf(\"%c\", value[index])\n if currentChar == \"{\" {\n // if key already had value. E.g: {abc{, {{abc{, -> key need to be reset before continue processing\n if key != \"{\" && key != \"{{\" {\n result += key\n key = \"\"\n } else {\n // this covers case that has \"{{{\" -> result need to be added one 1 and continue with key's value \"{{\"\n // this logic mentioned in the last testcase.\n if key == \"{{\" {\n result += \"{\"\n key = \"{\"\n }\n }\n key += \"{\"\n } else if currentChar == \"}\" && strings.Index(key, \"{\") == 0 { // There are only 3 cases.E.g abc}, {{abc}, and {abc}. We only care about 2 last cases.\n keyValue := \"\"\n key += \"}\"\n // As above logic when we detect \"{\". The value of key can only be started with \"{\" or \"{{\".\n // So we only need to check if this key is {{ }} or not.\n if strings.Index(key, \"{{\") == 0 {\n if index < size - 1 && fmt.Sprintf(\"%c\", value[index+1]) == \"}\" { // key has format {{..}}\n index ++\n result += key[1:] //We cut down 1 { } and put value again into result.\n key = \"\"\n } else { // cover key with format {{....}.\n keyValue = key[1: len(key) -1]\n result += \"{\"\n keyValue = key[2: len(key) -1]\n }\n } else {\n keyValue = key[1: len(key) -1]\n }\n if keyValue != \"\" {\n if v, found := params[strings.ToLower(keyValue)]; found { // if the key is in params, we will replace with value in params, if not we will use keyValue.\n result += v\n } else {\n result = fmt.Sprintf(\"%s{%s}\", result, keyValue)\n }\n }\n key = \"\"\n } else {\n key += currentChar\n }\n index++\n }\n // There are some cases that key has value but not completed as a fully pattern in the end of value.\n return result + key\n}", "func marshalMapInOrder(m map[string]interface{}, t interface{}) (string, error) {\n\ts := \"{\"\n\tv := reflect.ValueOf(t)\n\tfor i := 0; i < v.Type().NumField(); i++ {\n\t\tfield := jsonFieldFromTag(v.Type().Field(i).Tag)\n\t\tif field == \"\" {\n\t\t\tcontinue\n\t\t}\n\t\tvalue, err := json.Marshal(m[field])\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\ts += fmt.Sprintf(\"%q:%s,\", field, value)\n\t}\n\ts = s[:len(s)-1]\n\ts += \"}\"\n\tvar buf bytes.Buffer\n\tif err := json.Indent(&buf, []byte(s), \"\", \" \"); err != nil {\n\t\treturn \"\", err\n\t}\n\treturn buf.String(), nil\n}", "func flattenUrlMapPathMatcherRouteRuleRouteActionUrlRewriteMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleRouteActionUrlRewrite {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleRouteActionUrlRewrite{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleRouteActionUrlRewrite{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleRouteActionUrlRewrite)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleRouteActionUrlRewrite(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenUrlMapDefaultRouteActionUrlRewrite(c *Client, i interface{}) *UrlMapDefaultRouteActionUrlRewrite {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &UrlMapDefaultRouteActionUrlRewrite{}\n\tr.PathPrefixRewrite = dcl.FlattenString(m[\"pathPrefixRewrite\"])\n\tr.HostRewrite = dcl.FlattenString(m[\"hostRewrite\"])\n\n\treturn r\n}", "func flattenStoredInfoTypeDictionary(c *Client, i interface{}, res *StoredInfoType) *StoredInfoTypeDictionary {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &StoredInfoTypeDictionary{}\n\n\tif dcl.IsEmptyValueIndirect(i) {\n\t\treturn EmptyStoredInfoTypeDictionary\n\t}\n\tr.WordList = flattenStoredInfoTypeDictionaryWordList(c, m[\"wordList\"], res)\n\tr.CloudStoragePath = flattenStoredInfoTypeDictionaryCloudStoragePath(c, m[\"cloudStoragePath\"], res)\n\n\treturn r\n}", "func nestMap(fMap FieldMap) FieldMap {\n\tres := make(FieldMap)\n\tnested := make(map[string]FieldMap)\n\tfor k, v := range fMap {\n\t\texprs := strings.Split(k, ExprSep)\n\t\tif len(exprs) == 1 {\n\t\t\t// We are in the top map here\n\t\t\tres[k] = v\n\t\t\tcontinue\n\t\t}\n\t\tif _, exists := nested[exprs[0]]; !exists {\n\t\t\tnested[exprs[0]] = make(FieldMap)\n\t\t}\n\t\tnested[exprs[0]][strings.Join(exprs[1:], ExprSep)] = v\n\t}\n\t// Get nested FieldMap and assign to top key\n\tfor k, fm := range nested {\n\t\tres[k] = fm\n\t}\n\treturn res\n}", "func nestMap(fMap FieldMap) FieldMap {\n\tres := make(FieldMap)\n\tnested := make(map[string]FieldMap)\n\tfor k, v := range fMap {\n\t\texprs := strings.Split(k, ExprSep)\n\t\tif len(exprs) == 1 {\n\t\t\t// We are in the top map here\n\t\t\tres[k] = v\n\t\t\tcontinue\n\t\t}\n\t\tif _, exists := nested[exprs[0]]; !exists {\n\t\t\tnested[exprs[0]] = make(FieldMap)\n\t\t}\n\t\tnested[exprs[0]][strings.Join(exprs[1:], ExprSep)] = v\n\t}\n\t// Get nested FieldMap and assign to top key\n\tfor k, fm := range nested {\n\t\tres[k] = fm\n\t}\n\treturn res\n}", "func setPrefix(prefix string, m map[string]string) map[string]string {\n\tn := make(map[string]string)\n\tfor k, v := range m {\n\t\tn[prefix+k] = prefix + v\n\t}\n\n\treturn n\n}", "func lookupJSONStringFields(genericData map[string]interface{}) map[string]string {\n\tvar result = make(map[string]string)\n\tfor key, genericValue := range genericData {\n\t\tswitch castValue := genericValue.(type) {\n\t\tcase map[string]interface{}:\n\t\t\tnestedFields := lookupJSONStringFields(castValue)\n\t\t\tfor nestedKey, value := range nestedFields {\n\t\t\t\tresult[fmt.Sprintf(\"%s.%s\", key, nestedKey)] = value\n\t\t\t}\n\t\tcase string:\n\t\t\tresult[key] = castValue\n\t\t}\n\t}\n\treturn result\n}", "func MarshalIndentCompact(v interface{}, prefix, indent string, lineLen int) ([]byte, error) {\n\tb, err := marshalUnescaped(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Return early if document is small enough.\n\tif len(b) <= lineLen {\n\t\treturn b, nil\n\t}\n\n\tm := orderedmap.New()\n\n\t// Create a temporary JSON object to make sure it can be unmarshaled into a map.\n\ttmpMap := append([]byte(`{\"t\":`), b...)\n\ttmpMap = append(tmpMap, '}')\n\n\t// Unmarshal JSON payload into ordered map to recursively walk the document.\n\terr = json.Unmarshal(tmpMap, m)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ti, ok := m.Get(\"t\")\n\tif !ok {\n\t\treturn nil, errors.New(\"no value for this key\")\n\t}\n\n\t// Create first level padding.\n\tpad := append([]byte(prefix), []byte(indent)...)\n\n\t// Call recursive function to walk the document.\n\treturn marshalIndentCompact(i, indent, pad, lineLen)\n}", "func (m *Map) String() string {\n\tvar out bytes.Buffer\n\n\tout.WriteString(\"{\\n\")\n\tfor idx, val := range m.Elements {\n\t\tout.WriteString(fmt.Sprintf(\"%s : %s,\\n\", idx.String(), val.String()))\n\t}\n\tout.WriteString(\"}\")\n\n\treturn out.String()\n}", "func Mask(jsonString string, config ...*MaskConfig) (string, error) {\n\n\tbefore := make(map[string]interface{})\n\n\tif err := json.Unmarshal([]byte(jsonString), &before); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tcfg := defaultMaskConfig\n\tif len(config) > 0 {\n\t\tcfg = config[0]\n\t}\n\tif cfg.Callback == nil {\n\t\tcfg.Callback = defaultMaskFunc\n\t}\n\n\tskipFieldMap := make(map[string]bool, len(cfg.SkipFields))\n\tfor _, skipField := range cfg.SkipFields {\n\t\tskipFieldMap[skipField] = true\n\t}\n\n\tafter := make(map[string]interface{})\n\tfor k, v := range before {\n\t\tmask(k, v, after, cfg.Callback, skipFieldMap)\n\t}\n\n\tb, err := json.Marshal(after)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(b), nil\n}", "func flattenImageSourceSnapshotEncryptionKeyMap(c *Client, i interface{}) map[string]ImageSourceSnapshotEncryptionKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageSourceSnapshotEncryptionKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageSourceSnapshotEncryptionKey{}\n\t}\n\n\titems := make(map[string]ImageSourceSnapshotEncryptionKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageSourceSnapshotEncryptionKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (ctx *TemplateContext) formatMap() (out string) {\n\talphaSortMap(ctx.substitutionsMap,\n\t\tfunc(s string) {\n\t\t\tv := ctx.substitutionsMap[s]\n\t\t\tconst TRIM = 80\n\t\t\tif len(v) > TRIM {\n\t\t\t\tv = v[:TRIM] + \"...\"\n\t\t\t}\n\t\t\tout += fmt.Sprintf(\" % 20s '%v'\\n\\n\", s, v)\n\t\t})\n\treturn\n}", "func StringMap(m Map) (str string) {\n\tsb := poolStringBuilder.Get().(*StringBuilder)\n\tsb.Grow(ceil32(ByteSizeJSONMap(m)))\n\tEncodeJSONMap(sb, m)\n\tstr = sb.String()\n\tsb.Reset()\n\tpoolStringBuilder.Put(sb)\n\treturn\n}", "func flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilterMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleMatchRuleMetadataFilter)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleMatchRuleMetadataFilter(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (m Map) String() string {\n\tjc, _ := json.Marshal(m)\n\treturn string(jc)\n}", "func (p QuantumSuperposition) Flatten() QuantumSuperposition {\n\tdict := Superposition{}\n\tfor _, l := range p.Dict {\n\t\tif _l, ok := l.([]Val); ok {\n\t\t\tfor k, v := range _l[0].(QuantumSuperposition).Dict {\n\t\t\t\tif _, ok := dict[k]; ok {\n\t\t\t\t\tdict[k] = dict[k].(complex128) + mult(v.(complex128), _l[1].(complex128))\n\t\t\t\t} else {\n\t\t\t\t\tif _v, ok := v.([]Val); ok {\n\t\t\t\t\t\tdict[k] = mult(_v[1].(complex128), _l[1].(complex128))\n\t\t\t\t\t} else {\n\t\t\t\t\t\tdict[k] = mult(v.(complex128), _l[1].(complex128))\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\treturn p\n\t\t}\n\t}\n\treturn new(QuantumSuperposition).From(dict)\n}", "func NormalizeMap(zv zcode.Bytes) zcode.Bytes {\n\telements := make([]keyval, 0, 8)\n\tfor it := zv.Iter(); !it.Done(); {\n\t\tkey, _, err := it.NextTagAndBody()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tval, _, err := it.NextTagAndBody()\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\telements = append(elements, keyval{key, val})\n\t}\n\tif len(elements) < 2 {\n\t\treturn zv\n\t}\n\tsort.Slice(elements, func(i, j int) bool {\n\t\treturn bytes.Compare(elements[i].key, elements[j].key) == -1\n\t})\n\tnorm := make(zcode.Bytes, 0, len(zv))\n\tnorm = append(norm, elements[0].key...)\n\tnorm = append(norm, elements[0].val...)\n\tfor i := 1; i < len(elements); i++ {\n\t\t// Skip duplicates.\n\t\tif !bytes.Equal(elements[i].key, elements[i-1].key) {\n\t\t\tnorm = append(norm, elements[i].key...)\n\t\t\tnorm = append(norm, elements[i].val...)\n\t\t}\n\t}\n\treturn norm\n}", "func DumpMapStrings(data map[string]string, leftPad int) {\n\tlongest := LongestString(StringsMapKeys(data), 0) + leftPad\n\n\tIterateStringsMap(data, func(k, v string) {\n\t\tfmt.Printf(\"%s: %s\\n\", strings.Repeat(\" \", longest-len(k))+k, v)\n\t})\n}", "func (m *Map) ReduceString(\n\treduce func(map[interface{}]interface{}) string,\n\tjoin func(x, y string) string,\n) string {\n\tsplits := m.splits\n\t// NewMap ensures that len(splits) > 0\n\tresult := splits[0].reduceString(reduce)\n\tfor i := 1; i < len(splits); i++ {\n\t\tresult = join(result, splits[i].reduceString(reduce))\n\t}\n\treturn result\n}", "func FlattenVersion(version string) string {\n\treturn flattener.Replace(version)\n}", "func toJSONDot(prefix string, keys []string) string {\n\tif len(keys) == 0 {\n\t\treturn prefix\n\t}\n\n\tcontainsDots := false\n\tfor _, key := range keys {\n\t\tif strings.Contains(key, \".\") {\n\t\t\tcontainsDots = true\n\t\t}\n\t}\n\n\tvar b strings.Builder\n\tb.WriteString(prefix)\n\tif containsDots {\n\t\tfor _, key := range keys {\n\t\t\tb.WriteString(`['`)\n\t\t\tb.WriteString(key)\n\t\t\tb.WriteString(`']`)\n\t\t}\n\t} else {\n\t\tb.WriteString(\".\")\n\t\tfor i, key := range keys {\n\t\t\tif i != 0 {\n\t\t\t\tb.WriteString(\".\")\n\t\t\t}\n\t\t\tb.WriteString(key)\n\t\t}\n\t}\n\n\treturn b.String()\n}", "func flattenAttestorUserOwnedGrafeasNotePublicKeysMap(c *Client, i interface{}) map[string]AttestorUserOwnedGrafeasNotePublicKeys {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]AttestorUserOwnedGrafeasNotePublicKeys{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]AttestorUserOwnedGrafeasNotePublicKeys{}\n\t}\n\n\titems := make(map[string]AttestorUserOwnedGrafeasNotePublicKeys)\n\tfor k, item := range a {\n\t\titems[k] = *flattenAttestorUserOwnedGrafeasNotePublicKeys(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func (m *Map) String() string {\n\tkeys := \"\"\n\ti := 0\n\tfor k, _ := range *m {\n\t\tkeys = keys + k\n\t\tif i++; i < len(*m) {\n\t\t\tkeys = keys + \",\"\n\t\t}\n\t}\n\treturn keys\n}", "func (fpm *RegionalDistribution_FieldPathMap) JSONString() string {\n\treturn strcase.ToLowerCamel(fpm.selector.String()) + \".\" + fpm.key\n}", "func (v *Vertex) flatten() []string {\n\n\tlineage := []string{}\n\n\tp := v\n\tfor p != nil {\n\t\t// Prepend the lineage\n\t\tlineage = append([]string{p.Identifier}, lineage...)\n\t\tp = p.Parent\n\t}\n\n\treturn lineage\n}", "func flattenUrlMapPathMatcherRouteRuleUrlRedirectMap(c *Client, i interface{}) map[string]UrlMapPathMatcherRouteRuleUrlRedirect {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleUrlRedirect{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]UrlMapPathMatcherRouteRuleUrlRedirect{}\n\t}\n\n\titems := make(map[string]UrlMapPathMatcherRouteRuleUrlRedirect)\n\tfor k, item := range a {\n\t\titems[k] = *flattenUrlMapPathMatcherRouteRuleUrlRedirect(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func flattenKeyIosSettingsMap(c *Client, i interface{}, res *Key) map[string]KeyIosSettings {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]KeyIosSettings{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]KeyIosSettings{}\n\t}\n\n\titems := make(map[string]KeyIosSettings)\n\tfor k, item := range a {\n\t\titems[k] = *flattenKeyIosSettings(c, item.(map[string]interface{}), res)\n\t}\n\n\treturn items\n}", "func (s *MapSchema) String() string {\n\tbytes, err := json.MarshalIndent(s, \"\", \" \")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn string(bytes)\n}", "func (testEntityRelated_EntityInfo) Flatten(object interface{}, fbb *flatbuffers.Builder, id uint64) error {\n\tobj := object.(*TestEntityRelated)\n\tvar offsetName = fbutils.CreateStringOffset(fbb, obj.Name)\n\n\tvar rIdNext uint64\n\tif rel := obj.Next; rel != nil {\n\t\tif rId, err := EntityByValueBinding.GetId(rel); err != nil {\n\t\t\treturn err\n\t\t} else {\n\t\t\trIdNext = rId\n\t\t}\n\t}\n\n\t// build the FlatBuffers object\n\tfbb.StartObject(3)\n\tfbutils.SetUint64Slot(fbb, 0, id)\n\tfbutils.SetUOffsetTSlot(fbb, 1, offsetName)\n\tfbutils.SetUint64Slot(fbb, 2, rIdNext)\n\treturn nil\n}", "func FlattenPaths(paths, flatten []string) []string {\n\tres := make([]string, 0, len(paths))\n\tflattened := make(map[string]bool)\nnextPath:\n\tfor _, path := range paths {\n\t\tvar found bool\n\t\tfor _, flatten := range flatten {\n\t\t\tif flatten == path || strings.HasPrefix(path, flatten+\".\") {\n\t\t\t\tif !flattened[flatten] {\n\t\t\t\t\tres = append(res, flatten)\n\t\t\t\t\tflattened[flatten] = true\n\t\t\t\t}\n\t\t\t\tcontinue nextPath\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tres = append(res, path)\n\t\t}\n\t}\n\treturn res\n}", "func flattenImageSourceImageEncryptionKey(c *Client, i interface{}) *ImageSourceImageEncryptionKey {\n\tm, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn nil\n\t}\n\n\tr := &ImageSourceImageEncryptionKey{}\n\tr.RawKey = dcl.FlattenString(m[\"rawKey\"])\n\tr.KmsKeyName = dcl.FlattenString(m[\"kmsKeyName\"])\n\tr.Sha256 = dcl.FlattenString(m[\"sha256\"])\n\tr.KmsKeyServiceAccount = dcl.FlattenString(m[\"kmsKeyServiceAccount\"])\n\n\treturn r\n}", "func MakeStringDict(m map[string]interface{}) (skylark.StringDict, error) {\n\tdict := make(skylark.StringDict, len(m))\n\tfor k, v := range m {\n\t\tval, err := ToValue(v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdict[k] = val\n\t}\n\treturn dict, nil\n}", "func (p *SliceOfMap) String() string {\n\tvar builder strings.Builder\n\tbuilder.WriteString(\"[\")\n\tif p != nil {\n\t\tfor i := range *p {\n\t\t\tbuilder.WriteString(ToString((*p)[i]))\n\t\t\tif i+1 < len(*p) {\n\t\t\t\tbuilder.WriteString(\" \")\n\t\t\t}\n\t\t}\n\t}\n\tbuilder.WriteString(\"]\")\n\treturn builder.String()\n}", "func flattenImageSourceImageEncryptionKeyMap(c *Client, i interface{}) map[string]ImageSourceImageEncryptionKey {\n\ta, ok := i.(map[string]interface{})\n\tif !ok {\n\t\treturn map[string]ImageSourceImageEncryptionKey{}\n\t}\n\n\tif len(a) == 0 {\n\t\treturn map[string]ImageSourceImageEncryptionKey{}\n\t}\n\n\titems := make(map[string]ImageSourceImageEncryptionKey)\n\tfor k, item := range a {\n\t\titems[k] = *flattenImageSourceImageEncryptionKey(c, item.(map[string]interface{}))\n\t}\n\n\treturn items\n}", "func Unflatten(m map[string]interface{}, tf TokenizerFunc) map[string]interface{} {\n\ttree := make(map[string]interface{})\n\n\tc := make(chan map[string]interface{})\n\n\tgo mapify(m, c, tf)\n\n\tfor n := range c {\n\t\tmergo.Merge(&tree, n)\n\t}\n\n\treturn tree\n}", "func Flatten(in *corev1.NFSVolumeSource) []interface{} {\n\tout := make([]interface{}, 1)\n\n\trow := map[string]interface{}{}\n\n\tif in.Server != \"\" {\n\t\trow[FieldServer] = in.Server\n\t}\n\n\tif in.Path != \"\" {\n\t\trow[FieldPath] = in.Path\n\t}\n\n\tout[0] = row\n\n\treturn out\n}" ]
[ "0.75922185", "0.6829765", "0.6402381", "0.63900155", "0.57017297", "0.5587964", "0.5310287", "0.5248318", "0.5099017", "0.50407", "0.5025752", "0.4986255", "0.48482332", "0.48161975", "0.48159418", "0.4791885", "0.47653565", "0.47326958", "0.47076184", "0.470675", "0.4607495", "0.46065253", "0.4605455", "0.45825297", "0.45594317", "0.45515278", "0.4526397", "0.44960192", "0.44948253", "0.4468376", "0.44675678", "0.44462398", "0.4438139", "0.44175923", "0.44174528", "0.44111282", "0.43821648", "0.43720928", "0.43569022", "0.4352756", "0.4336617", "0.4333434", "0.43180516", "0.43160295", "0.43076006", "0.4300736", "0.42954907", "0.42939103", "0.42893323", "0.42890218", "0.4285958", "0.427994", "0.427283", "0.4264865", "0.42587596", "0.425408", "0.42494193", "0.42210925", "0.42184317", "0.42172837", "0.4208604", "0.41841483", "0.41839504", "0.41800755", "0.4179559", "0.41772446", "0.41708416", "0.41701305", "0.41701305", "0.41673544", "0.41665995", "0.41637155", "0.416111", "0.41573063", "0.41556314", "0.41529617", "0.41423896", "0.41330066", "0.41314045", "0.4129861", "0.41229358", "0.41142106", "0.41137204", "0.41116312", "0.41101047", "0.41083163", "0.41065127", "0.41026914", "0.4101583", "0.40994528", "0.40950143", "0.40936136", "0.40918237", "0.40862632", "0.40859053", "0.40824205", "0.40824163", "0.4075016", "0.40743688", "0.40701082" ]
0.7555226
1
New creates a new quote shipping method management v1 API client.
func New(transport runtime.ClientTransport, formats strfmt.Registry) *Client { return &Client{transport: transport, formats: formats} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *PurchaseOrderLinesEndpoint) New() *PurchaseOrderLines {\n\treturn &PurchaseOrderLines{}\n}", "func (c *ChargeClient) New(ctx context.Context, idempotencyKey string, paymentID string, params *ChargeParams, clientInfo *ClientInfo) (*Charge, error) {\n\tcharge := &Charge{}\n\n\theaders := map[string]string{headerIdempotencyKey: idempotencyKey}\n\n\tif clientInfo != nil {\n\t\theaders[headerClientIPAddress] = clientInfo.IPAddress\n\t\theaders[headerClientUserAgent] = clientInfo.UserAgent\n\t}\n\n\tif err := c.Caller.Call(ctx, \"POST\", c.chargesPath(paymentID), headers, params, charge); err != nil {\n\t\treturn nil, err\n\t}\n\treturn charge, nil\n}", "func New(client *http.Client) (*Service, error) {\n\tif client == nil {\n\t\treturn nil, errors.New(\"client is nil\")\n\t}\n\ts := &Service{client: client, BasePath: basePath}\n\ts.Purchases = NewPurchasesService(s)\n\treturn s, nil\n}", "func (s *PurchaseOrdersEndpoint) New() *PurchaseOrders {\n\treturn &PurchaseOrders{}\n}", "func New(g *godo.Client) Client {\n\tc := &client{\n\t\tg: g,\n\t}\n\treturn c\n}", "func New(country, city string, client customer.Customer, items []invoiceitem.Item) Invoice {\n\treturn Invoice{\n\t\tcountry: country,\n\t\tcity: city,\n\t\tclient: client,\n\t\titems: items,\n\t}\n}", "func New(apiKey string) *Client {\n\tclient := new(Client)\n\tclient.apiKey = apiKey\n\treturn client\n}", "func New(key, secret, symbol, currency string, priority int, fee, availShort, availFunds float64) *Client {\n\treturn &Client{\n\t\tkey: key,\n\t\tsecret: secret,\n\t\tsymbol: symbol,\n\t\tcurrency: currency,\n\t\tpriority: priority,\n\t\tfee: fee,\n\t\tavailShort: availShort,\n\t\tavailFunds: availFunds,\n\t\tcurrencyCode: 0,\n\t\tname: fmt.Sprintf(\"Bitfinex(%s)\", currency),\n\t\tbaseURL: \"https://api.bitfinex.com\",\n\t\tdone: make(chan bool, 1),\n\t}\n}", "func New(pg *postgres.Client, consumer mb.Service, cfgNS config.NotificationService, cfgQuote config.QuoteService) Service {\n\tonce.Do(func() {\n\t\tsrv = service{\n\t\t\tpg: pg,\n\t\t\tconsumer: consumer,\n\t\t}\n\t\tsrv.msgChan = make(chan []byte)\n\t\tsrv.urlNS = cfgNS.URL + \"/mail/send\"\n\t\tsrv.urlQuote = cfgQuote.URL + \"/quotes\"\n\t\tgo srv.consumer.Consume([]string{topicCurrencies}, srv.msgChan)\n\t\tgo srv.QuoteConsumer()\n\t})\n\n\treturn srv\n}", "func New(options Options, optFns ...func(*Options)) *Client {\n\toptions = options.Copy()\n\n\tresolveRetryer(&options)\n\n\tresolveHTTPClient(&options)\n\n\tresolveHTTPSignerV4(&options)\n\n\tresolveDefaultEndpointConfiguration(&options)\n\n\tresolveIdempotencyTokenProvider(&options)\n\n\tfor _, fn := range optFns {\n\t\tfn(&options)\n\t}\n\n\tclient := &Client{\n\t\toptions: options,\n\t}\n\n\treturn client\n}", "func (o *OrderService) New(domains []string) (acme.ExtendedOrder, error) {\n\treturn o.NewWithOptions(domains, nil)\n}", "func (cs *checkoutService) quoteShipping(ctx context.Context, address *pb.Address, items []*pb.CartItem) (*pb.Money, error) {\n\tshippingQuote, err := pb.NewShippingServiceClient(cs.shippingSvcConn).\n\t\tGetQuote(ctx, &pb.GetQuoteRequest{\n\t\t\tAddress: address,\n\t\t\tItems: items})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get shipping quote: %+v\", err)\n\t}\n\treturn shippingQuote.GetCostUsd(), nil\n}", "func New(\n\tcountry,\n\tcity string,\n\tclient customer.Customer,\n\t// items []invoiceitem.Item,\n\titems invoiceitem.Items,\n) Invoice {\n\treturn Invoice{\n\t\tcountry: country,\n\t\tcity: city,\n\t\tclient: client,\n\t\titems: items,\n\t}\n}", "func New(transport runtime.ClientTransport, formats strfmt.Registry) *OpenbankingPaymentsClient {\n\t// ensure nullable parameters have default\n\tif formats == nil {\n\t\tformats = strfmt.Default\n\t}\n\n\tcli := new(OpenbankingPaymentsClient)\n\tcli.Transport = transport\n\tcli.DomesticPayments = domestic_payments.New(transport, formats)\n\tcli.DomesticScheduledPayments = domestic_scheduled_payments.New(transport, formats)\n\tcli.DomesticStandingOrders = domestic_standing_orders.New(transport, formats)\n\tcli.FilePayments = file_payments.New(transport, formats)\n\tcli.InternationalPayments = international_payments.New(transport, formats)\n\tcli.InternationalScheduledPayments = international_scheduled_payments.New(transport, formats)\n\tcli.InternationalStandingOrders = international_standing_orders.New(transport, formats)\n\tcli.PaymentDetails = payment_details.New(transport, formats)\n\treturn cli\n}", "func New(client *client.Client, properties ClientProperties) *Client {\n\treturn &Client{\n\t\tclient: client,\n\n\t\taccountSid: properties.AccountSid,\n\t\tsid: properties.Sid,\n\n\t\tMember: func(sid string) *member.Client {\n\t\t\treturn member.New(client, member.ClientProperties{\n\t\t\t\tAccountSid: properties.AccountSid,\n\t\t\t\tQueueSid: properties.Sid,\n\t\t\t\tSid: sid,\n\t\t\t})\n\t\t},\n\t\tMembers: members.New(client, members.ClientProperties{\n\t\t\tAccountSid: properties.AccountSid,\n\t\t\tQueueSid: properties.Sid,\n\t\t}),\n\t}\n}", "func New(sid string, token string) *Client {\n\treturn &Client{\n\t\taccountSid: sid,\n\t\tauthToken: token,\n\t\tmessageFrom: \"Genesis\",\n\t}\n}", "func (c *Client) New() goa.Endpoint {\n\tvar (\n\t\tdecodeResponse = DecodeNewResponse(c.decoder, c.RestoreResponseBody)\n\t)\n\treturn func(ctx context.Context, v interface{}) (interface{}, error) {\n\t\treq, err := c.BuildNewRequest(ctx, v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresp, err := c.NewDoer.Do(req)\n\t\tif err != nil {\n\t\t\treturn nil, goahttp.ErrRequestError(\"spin-broker\", \"new\", err)\n\t\t}\n\t\treturn decodeResponse(resp)\n\t}\n}", "func New(client *client.Client, properties ClientProperties) *Client {\n\treturn &Client{\n\t\tclient: client,\n\n\t\taccountSid: properties.AccountSid,\n\t\tcountryCode: properties.CountryCode,\n\t}\n}", "func New(opt Option) *Client {\n\treturn &Client{\n\t\ttoken: opt.Token,\n\t\taddr: opt.Addr,\n\t}\n}", "func New(c service.GapidClient) service.Service {\n\treturn &client{c, func() error { return nil }}\n}", "func New(client *client.Client, properties ClientProperties) *Client {\n\treturn &Client{\n\t\tclient: client,\n\n\t\tsid: properties.Sid,\n\n\t\tAccessTokens: access_tokens.New(client, access_tokens.ClientProperties{\n\t\t\tServiceSid: properties.Sid,\n\t\t}),\n\t\tEntities: entities.New(client, entities.ClientProperties{\n\t\t\tServiceSid: properties.Sid,\n\t\t}),\n\t\tEntity: func(identity string) *entity.Client {\n\t\t\treturn entity.New(client, entity.ClientProperties{\n\t\t\t\tIdentity: identity,\n\t\t\t\tServiceSid: properties.Sid,\n\t\t\t})\n\t\t},\n\t\tMessagingConfiguration: func(countryCode string) *messaging_configuration.Client {\n\t\t\treturn messaging_configuration.New(client, messaging_configuration.ClientProperties{\n\t\t\t\tCountryCode: countryCode,\n\t\t\t\tServiceSid: properties.Sid,\n\t\t\t})\n\t\t},\n\t\tMessagingConfigurations: messaging_configurations.New(client, messaging_configurations.ClientProperties{\n\t\t\tServiceSid: properties.Sid,\n\t\t}),\n\t\tRateLimit: func(rateLimitSid string) *rate_limit.Client {\n\t\t\treturn rate_limit.New(client, rate_limit.ClientProperties{\n\t\t\t\tServiceSid: properties.Sid,\n\t\t\t\tSid: rateLimitSid,\n\t\t\t})\n\t\t},\n\t\tRateLimits: rate_limits.New(client, rate_limits.ClientProperties{\n\t\t\tServiceSid: properties.Sid,\n\t\t}),\n\t\tVerification: func(verificationSid string) *verification.Client {\n\t\t\treturn verification.New(client, verification.ClientProperties{\n\t\t\t\tServiceSid: properties.Sid,\n\t\t\t\tSid: verificationSid,\n\t\t\t})\n\t\t},\n\t\tVerificationCheck: verification_check.New(client, verification_check.ClientProperties{\n\t\t\tServiceSid: properties.Sid,\n\t\t}),\n\t\tVerifications: verifications.New(client, verifications.ClientProperties{\n\t\t\tServiceSid: properties.Sid,\n\t\t}),\n\t\tWebhook: func(webhookSid string) *webhook.Client {\n\t\t\treturn webhook.New(client, webhook.ClientProperties{\n\t\t\t\tServiceSid: properties.Sid,\n\t\t\t\tSid: webhookSid,\n\t\t\t})\n\t\t},\n\t\tWebhooks: webhooks.New(client, webhooks.ClientProperties{\n\t\t\tServiceSid: properties.Sid,\n\t\t}),\n\t}\n}", "func New(instance string, options ...httptransport.ClientOption) (pb.CustomerServer, error) {\n\n\tif !strings.HasPrefix(instance, \"http\") {\n\t\tinstance = \"http://\" + instance\n\t}\n\tu, err := url.Parse(instance)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_ = u\n\n\tpanic(\"No HTTP Endpoints, this client will not work, define bindings in your proto definition\")\n\n\treturn svc.Endpoints{}, nil\n}", "func New(l hclog.Logger, qs QuoteStore, a Auth) *QuoteServer {\n\tx := new(QuoteServer)\n\tx.log = l.Named(\"http\")\n\tx.Echo = echo.New()\n\tx.db = qs\n\tx.auth = a\n\n\tx.rndr = NewRenderer(x.log)\n\tx.rndr.Reload()\n\n\tx.Echo.Renderer = x.rndr\n\tx.Echo.IPExtractor = echo.ExtractIPFromXFFHeader()\n\n\tx.GET(\"/\", x.home)\n\tx.GET(\"/quote/:id\", x.showQuote)\n\tx.GET(\"/search/:query/:page/:count\", x.searchQuotes)\n\tx.POST(\"/dosearch\", x.searchReflect)\n\n\tx.GET(\"/add\", x.addQuoteForm)\n\tx.POST(\"/add\", x.addQuote)\n\n\tx.GET(\"/login\", x.loginForm)\n\tx.POST(\"/login\", x.loginHandler)\n\tx.GET(\"/logout\", x.logoutHandler)\n\n\tx.GET(\"/reload\", x.reload)\n\n\tadm := x.Group(\"/admin\")\n\tadm.Use(middleware.JWTWithConfig(middleware.JWTConfig{\n\t\tSigningKey: []byte(os.Getenv(\"NF_TOKEN_STRING\")),\n\t\tTokenLookup: \"cookie:auth\",\n\t}))\n\tadm.GET(\"/\", x.adminLanding)\n\tadm.POST(\"/quote/:id/approve\", x.approveQuote)\n\tadm.POST(\"/quote/:id/remove\", x.removeQuote)\n\n\tx.Static(\"/static\", \"web/static\")\n\n\treturn x\n}", "func New(target string, opts ...grpc.DialOption) (*Client, error) {\n\tconn, err := grpc.Dial(target, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Client{conn, quotaservice.NewQuotaServiceClient(conn)}, nil\n}", "func New(\n\tquoteid *field.QuoteIDField,\n\tquotecanceltype *field.QuoteCancelTypeField,\n\tnoquoteentries *field.NoQuoteEntriesField) Message {\n\tbuilder := Message{Message: quickfix.NewMessage()}\n\tbuilder.Header.Set(field.NewBeginString(enum.BeginStringFIX42))\n\tbuilder.Header.Set(field.NewMsgType(\"Z\"))\n\tbuilder.Body.Set(quoteid)\n\tbuilder.Body.Set(quotecanceltype)\n\tbuilder.Body.Set(noquoteentries)\n\treturn builder\n}", "func New(transport client.Transport, formats strfmt.Registry) *Client {\n\treturn &Client{transport: transport, formats: formats}\n}", "func New(opts ...ClientOption) *Client {\n\tc := &Client{\n\t\tName: \"retriable\",\n\t\thttpClient: &http.Client{\n\t\t\tTimeout: time.Second * 30,\n\t\t},\n\t\tPolicy: NewDefaultPolicy(),\n\t}\n\n\tfor _, opt := range opts {\n\t\topt.applyOption(c)\n\t}\n\treturn c\n}", "func NewClient(env common.Environment, apiKey string, processingChannelId *string) *CheckoutComClient {\n\treturn NewWithHTTPClient(env, apiKey, processingChannelId, common.DefaultHttpClient())\n}" ]
[ "0.5735634", "0.5582585", "0.5393141", "0.52955633", "0.5239658", "0.5238384", "0.5111851", "0.5105488", "0.51014143", "0.5082375", "0.5059564", "0.5036858", "0.5021886", "0.50124085", "0.50055", "0.49916774", "0.4942524", "0.4880882", "0.48651394", "0.48411164", "0.4839529", "0.48343754", "0.48322204", "0.4829109", "0.48204097", "0.4810704", "0.4807365", "0.47912708" ]
0.0
-1
/ QuoteShippingMethodManagementV1EstimateByAddressIDPost Estimate shipping
func (a *Client) QuoteShippingMethodManagementV1EstimateByAddressIDPost(params *QuoteShippingMethodManagementV1EstimateByAddressIDPostParams) (*QuoteShippingMethodManagementV1EstimateByAddressIDPostOK, error) { // TODO: Validate the params before sending if params == nil { params = NewQuoteShippingMethodManagementV1EstimateByAddressIDPostParams() } result, err := a.transport.Submit(&runtime.ClientOperation{ ID: "quoteShippingMethodManagementV1EstimateByAddressIdPost", Method: "POST", PathPattern: "/V1/carts/{cartId}/estimate-shipping-methods-by-address-id", ProducesMediaTypes: []string{""}, ConsumesMediaTypes: []string{""}, Schemes: []string{"http"}, Params: params, Reader: &QuoteShippingMethodManagementV1EstimateByAddressIDPostReader{formats: a.formats}, Context: params.Context, Client: params.HTTPClient, }) if err != nil { return nil, err } return result.(*QuoteShippingMethodManagementV1EstimateByAddressIDPostOK), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (a *Client) QuoteShippingMethodManagementV1EstimateByAddressIDPostMine(params *QuoteShippingMethodManagementV1EstimateByAddressIDPostMineParams) (*QuoteShippingMethodManagementV1EstimateByAddressIDPostMineOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewQuoteShippingMethodManagementV1EstimateByAddressIDPostMineParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"quoteShippingMethodManagementV1EstimateByAddressIdPostMine\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/V1/carts/mine/estimate-shipping-methods-by-address-id\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &QuoteShippingMethodManagementV1EstimateByAddressIDPostMineReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*QuoteShippingMethodManagementV1EstimateByAddressIDPostMineOK), nil\n\n}", "func (o *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostBody) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := o.validateAddressID(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK() *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK {\n\treturn &NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK{}\n}", "func NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault(code int) *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault {\n\treturn &NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (o *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault) Code() int {\n\treturn o._statusCode\n}", "func (e Endpoints) PostAddress(ctx context.Context, profileID string, a Address) error {\n\n\t// TODO: Create detailed ref spec\n\trequest := postAddressRequest{ProfileID: profileID, Address: a}\n\n\tresponse, err := e.PostAddressEndpoint(ctx, request)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tresp := response.(postAddressResponse)\n\n\treturn resp.Err\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) SetGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineBody) {\n\to.GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody = giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineBody) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody)\n\treturn o\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody)\n\treturn o\n}", "func NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized() *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized {\n\treturn &NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized{}\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) SetGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody) {\n\to.GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody = giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody\n}", "func (svc *inmemService) PostAddress(ctx context.Context, profileID string, address Address) error {\n\n\t// Get a Lock on the svc for atomic access to the datastore\n\tsvc.mtx.Lock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.Unlock()\n\n\t// Check the datastore for the specified profile and set it to a profile variable when found\n\tprofile, ok := svc.profiles[profileID]\n\n\t// If the profile was not found in the datastore\n\tif !ok {\n\n\t\t// Return an error to the caller informing them that the specified profile could not be found\n\t\treturn ErrNotFound\n\t}\n\n\t// Iterate over the profile's existent addresses\n\tfor _, existingAddress := range profile.Addresses {\n\n\t\t// Check to see if the current existent address has an ID that matches the ID of the address passed in\n\t\tif existingAddress.ID == address.ID {\n\n\t\t\t// Return an error informing the caller that address they wanted to add is already present\n\t\t\treturn ErrAlreadyExists\n\t\t}\n\t}\n\n\t// Add the new address to the profile Address collection\n\tprofile.Addresses = append(profile.Addresses, address)\n\n\t// Replace the profile with the one that has the new Address added to the Address collection\n\tsvc.profiles[profileID] = profile\n\n\t// Return a nil error value\n\treturn nil\n}", "func (c *Client) Addresses(ctx context.Context, foreignID, currency string) (Address, error) {\n\treqBody := map[string]string{\n\t\t\"foreign_id\": foreignID,\n\t\t\"currency\": currency,\n\t}\n\n\treqJSON, err := json.Marshal(reqBody)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request body marshaling error: %w\", err)\n\t}\n\n\taddressesURL, err := joinURL(c.api, addressesEndpoint)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request url creating error: %w\", err)\n\t}\n\n\treq, err := http.NewRequestWithContext(ctx, http.MethodPost, addressesURL.String(), bytes.NewBuffer(reqJSON))\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request creating error: %w\", err)\n\t}\n\n\tsig, err := createHmac(c.secret, reqJSON)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"hmac signature creationg error: %w\", err)\n\t}\n\n\treq.Header.Set(contentTypeHeader, jsonContentType)\n\treq.Header.Set(keyHeader, c.apiKey)\n\treq.Header.Set(signatureHeader, sig)\n\n\tresp, err := c.client.Do(req)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request error: %w\", err)\n\t}\n\tdefer resp.Body.Close()\n\n\terr = ensureSuccessResponse(resp)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request failed: %w\", err)\n\t}\n\n\trespBody := struct {\n\t\tData Address `json:\"data\"`\n\t}{}\n\n\terr = json.NewDecoder(resp.Body).Decode(&respBody)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"response unmarshaling error: %w\", err)\n\t}\n\n\treturn respBody.Data, nil\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithTimeout(timeout time.Duration) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithContext(ctx context.Context) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithTimeout(timeout time.Duration) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func MakePostAddressEndpoint(s Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\n\t\t// Assert the request interface passed in as a postProfileRequest value\n\t\t// req will be a value of type postAddressRequest with the values properly mapped from the dynamic value passed in\n\t\treq := request.(postAddressRequest)\n\n\t\t// Create an error based on the return from the Service's PostAddress function\n\t\te := s.PostAddress(ctx, req.ProfileID, req.Address)\n\n\t\treturn postAddressResponse{Err: e}, nil\n\n\t}\n}", "func (s *OrderShippingAddressService) Show(ctx context.Context, orderID int, id int) (*OrderShippingAddress, *http.Response, error) {\n\tosa := new(OrderShippingAddress)\n\tapiError := new(APIError)\n\n\tresp, err := performRequest(ctx, s.sling.New().Get(fmt.Sprintf(\"%d/shipping_addresses/%d\", orderID, id)), s.httpClient, osa, apiError)\n\treturn osa, resp, relevantError(err, *apiError)\n}", "func main() {\n\tmoveIDString := flag.String(\"moveID\", \"\", \"The ID of the move where shipments are found\")\n\tenv := flag.String(\"env\", \"development\", \"The environment to run in, which configures the database.\")\n\tflag.Parse()\n\n\tif *moveIDString == \"\" {\n\t\tlog.Fatal(\"Usage: generate_shipment_edi -moveID <29cb984e-c70d-46f0-926d-cd89e07a6ec3>\")\n\t}\n\n\tdb, err := pop.Connect(*env)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tmoveID := uuid.Must(uuid.FromString(*moveIDString))\n\tvar shipments models.Shipments\n\n\terr = db.Eager(\n\t\t\"Move.Orders\",\n\t\t\"PickupAddress\",\n\t\t\"DeliveryAddress\",\n\t\t\"ServiceMember\",\n\t\t\"ShipmentOffers.TransportationServiceProviderPerformance\",\n\t).Where(\"shipment_offers.accepted=true\").\n\t\tWhere(\"move_id = $1\", &moveID).\n\t\tJoin(\"shipment_offers\", \"shipment_offers.shipment_id = shipments.id\").\n\t\tAll(&shipments)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif len(shipments) == 0 {\n\t\tlog.Fatal(\"No accepted shipments found\")\n\t}\n\tvar logger = zap.NewNop()\n\n\tvar costsByShipments []rateengine.CostByShipment\n\n\tengine := rateengine.NewRateEngine(db, logger, route.NewTestingPlanner(362)) //TODO: create the proper route/planner\n\tfor _, shipment := range shipments {\n\t\tcostByShipment, err := engine.HandleRunOnShipment(shipment)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tcostsByShipments = append(costsByShipments, costByShipment)\n\t}\n\tedi, err := ediinvoice.Generate858C(costsByShipments, db)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfmt.Println(edi)\n}", "func (o *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 401:\n\t\tresult := NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\tresult := NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithContext(ctx context.Context) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithHTTPClient(client *http.Client) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (a API) EstimateFee(cmd *btcjson.EstimateFeeCmd) (e error) {\n\tRPCHandlers[\"estimatefee\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}", "func (a *Client) PostReturnAddressesByReturnAddressIDGet(params *PostReturnAddressesByReturnAddressIDGetParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesByReturnAddressIDGetOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesByReturnAddressIDGetParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesByReturnAddressIdGet\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/post/return-addresses/{return_address_id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesByReturnAddressIDGetReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesByReturnAddressIDGetOK), nil\n\n}", "func (e Endpoints) GetAddress(ctx context.Context, profileID string, addressID string) (Address, error) {\n\n\t// TODO: Create detailed ref spec\n\trequest := getAddressRequest{ProfileID: profileID, AddressID: addressID}\n\n\tresponse, err := e.GetAddressEndpoint(ctx, request)\n\n\tif err != nil {\n\t\treturn Address{}, err\n\t}\n\n\tresp := response.(getAddressResponse)\n\n\treturn resp.Address, resp.Err\n}", "func (a *Api) AddressAtHeight(address string, height int) (*AddressResult, error) {\n\n\tparams := make(map[string]string)\n\tparams[\"address\"] = address\n\tif height > 0 {\n\t\tparams[\"height\"] = strconv.Itoa(height)\n\t}\n\n\tres, err := a.client.R().SetQueryParams(params).Get(\"/address\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif res.IsError() {\n\t\treturn nil, NewResponseError(res)\n\t}\n\n\tresponse := new(AddressResponse)\n\terr = json.Unmarshal(res.Body(), response)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.Error != nil {\n\t\treturn nil, response.Error\n\t}\n\n\treturn response.Result, nil\n}", "func (a *Client) PostReturnAddressesGet(params *PostReturnAddressesGetParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesGetOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesGetParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesGet\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/post/return-addresses\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesGetReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesGetOK), nil\n\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams() *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (s *OrderShippingAddressService) Count(ctx context.Context, orderID int, params *OrderShippingAddressListParams) (int, *http.Response, error) {\n\tvar cnt count\n\tapiError := new(APIError)\n\n\tresp, err := performRequest(ctx, s.sling.New().Get(fmt.Sprintf(\"%d/shipping_addresses/count\", orderID)).QueryStruct(params), s.httpClient, &cnt, apiError)\n\treturn cnt.Count, resp, relevantError(err, *apiError)\n}", "func HandleShippingDetails(authToken string, cb func(orderId string, req ShippingCostRequest) (ShippingCostResponse, error)) http.HandlerFunc {\n\tfn := func(w http.ResponseWriter, r *http.Request) {\n\t\tif r.Method != http.MethodPost {\n\t\t\tw.Header().Set(\"ALLOW\", \"POST\")\n\t\t\thttp.Error(w, \"Unsupported method\", http.StatusMethodNotAllowed)\n\t\t\treturn\n\t\t}\n\t\tif authToken != \"\" && r.Header.Get(\"Authorization\") != authToken {\n\t\t\thttp.Error(w, \"Unauthorized\", http.StatusUnauthorized)\n\t\t\treturn\n\t\t}\n\t\tpathBySegments := strings.Split(r.URL.Path, \"/\")\n\t\torderID := pathBySegments[len(pathBySegments)-2]\n\n\t\tbodyDec := json.NewDecoder(r.Body)\n\t\tdefer r.Body.Close()\n\n\t\treq := ShippingCostRequest{}\n\t\terr := bodyDec.Decode(&req)\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tsh, err := cb(orderID, req)\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tj, err := json.Marshal(sh)\n\t\tif err != nil {\n\t\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\n\t\tw.Write(j)\n\t}\n\treturn fn\n}", "func EstimateFee(tx StdTx) StdFee {\n\treturn NewStdFee(txparam.DefaultMsgGas*uint64(len(tx.Msgs)), tx.Fee.GasPrice)\n}", "func withAddressID(id int64) addressOption {\n\treturn func(m *AddressMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Address\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Address, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Address.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func GetAddress(latlng []float64, key string, useTier bool, tier common.AlertTier) (string, error) {\n\t// Retrieve location from MapQuest Geocoding API\n\tgeocoder.SetAPIKey(key)\n\ta, err := geocoder.ReverseGeocode(latlng[0], latlng[1])\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\t// Filter data returned depending on provided AlertTier\n\tvar address string\n\tif a.Street != \"\" && tier == common.FIRST {\n\t\taddress = fmt.Sprintf(\"%v, \", a.Street)\n\t}\n\n\tif a.City != \"\" && (tier == common.FIRST || tier == common.SECOND) {\n\t\taddress = address + fmt.Sprintf(\"%v, \", a.City)\n\t}\n\tif a.State != \"\" {\n\t\taddress = address + fmt.Sprintf(\"%v, \", a.State)\n\t}\n\tif a.PostalCode != \"\" {\n\t\taddress = address + fmt.Sprintf(\"%v, \", a.PostalCode)\n\t}\n\tif a.CountryCode != \"\" {\n\t\taddress = address + fmt.Sprintf(\"%v\", a.CountryCode)\n\t}\n\n\treturn address, nil\n}", "func (a *Client) PostReturnAddressesPost(params *PostReturnAddressesPostParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesPostOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesPostParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesPost\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/post/return-addresses\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesPostReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesPostOK), nil\n\n}", "func (cs *checkoutService) quoteShipping(ctx context.Context, address *pb.Address, items []*pb.CartItem) (*pb.Money, error) {\n\tshippingQuote, err := pb.NewShippingServiceClient(cs.shippingSvcConn).\n\t\tGetQuote(ctx, &pb.GetQuoteRequest{\n\t\t\tAddress: address,\n\t\t\tItems: items})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get shipping quote: %+v\", err)\n\t}\n\treturn shippingQuote.GetCostUsd(), nil\n}", "func (a *Client) PostReturnAddressesByReturnAddressIDPut(params *PostReturnAddressesByReturnAddressIDPutParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesByReturnAddressIDPutOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesByReturnAddressIDPutParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesByReturnAddressIdPut\",\n\t\tMethod: \"PUT\",\n\t\tPathPattern: \"/post/return-addresses/{return_address_id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesByReturnAddressIDPutReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesByReturnAddressIDPutOK), nil\n\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithHTTPClient(client *http.Client) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func NewGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams() *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\tvar ()\n\treturn &GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (a *Api) Address(address string) (*AddressResult, error) {\n\treturn a.AddressAtHeight(address, LatestBlockHeight)\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithCartID(cartID string) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetCartID(cartID)\n\treturn o\n}", "func (m *MemberService) UpsertAddress(ctx context.Context, op *WithdrawAddressView) (*WithdrawAddressView, error) {\n\tdata, err := m.POST(\"/address\").\n\t\tBody(op).\n\t\tAuth(m.Presign(time.Minute)).\n\t\tDo(ctx).Bytes()\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp struct {\n\t\tErr\n\t\tAddress *WithdrawAddressView `json:\"address\"`\n\t}\n\n\tif err := jsoniter.Unmarshal(data, &resp); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif resp.Code > 0 {\n\t\treturn nil, resp.Err\n\t}\n\n\treturn resp.Address, nil\n}", "func Explode(address string) (*Address, error) {\n\tif client == nil {\n\t\terr := initClient()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Build the API request.\n\treq := &maps.GeocodingRequest{\n\t\tAddress: address,\n\t}\n\n\t// Execute the request.\n\tresp, err := client.Geocode(context.Background(), req)\n\tif len(resp) < 1 {\n\t\treturn nil, err\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Using the first/closest match in our response, grab the values we need.\n\tcomponents := resp[0].AddressComponents\n\tformattedAddress := resp[0].FormattedAddress\n\tlat := resp[0].Geometry.Location.Lat\n\tlng := resp[0].Geometry.Location.Lng\n\n\t// Construct the return *Address{}\n\tresponse := &Address{\n\t\tAddressLine1: compose(addressLine1Composition, \"\", components, false),\n\t\tAddressLine2: compose(addressLine2Composition, addressLineDelimeter, components, false),\n\t\tAddressCity: compose(addressCityComposition, addressLineDelimeter, components, false),\n\t\tAddressState: compose(addressStateComposition, addressLineDelimeter, components, false),\n\t\tAddressCountry: compose(addressCountryComposition, addressLineDelimeter, components, false),\n\t\tAddressCountryCode: compose(addressCountryCodeComposition, addressLineDelimeter, components, true),\n\t\tAddressZip: compose(addressPostalCodeComposition, addressLineDelimeter, components, false),\n\t\tAddressLat: &lat,\n\t\tAddressLng: &lng,\n\t\tFormattedAddress: &formattedAddress,\n\t}\n\n\treturn response, err\n}", "func (mr *Repo) UpdateAddress(addr models.Address) (models.Address, error) {\n\tdbAddr := &dbAddress{}\n\n\tid, err := primitive.ObjectIDFromHex(addr.ID)\n\tif err != nil {\n\t\terr = &repo.InvalidIDError{ID: addr.ID}\n\t\tlog.Println(err)\n\t\treturn addr, err\n\t}\n\tfilter := bson.M{\n\t\t\"_id\": id,\n\t}\n\n\treturnOption := options.After\n\tcollection := mr.client.Database(mr.dbInfo.DBName).Collection(mr.dbInfo.CollectionName)\n\terr = collection.FindOneAndUpdate(mr.ctx, filter, addr, &options.FindOneAndUpdateOptions{\n\t\tReturnDocument: &returnOption,\n\t}).Decode(&dbAddr)\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\taddr = dbAddr.toServiceAddress()\n\treturn addr, err\n}", "func (_Mapping *MappingCaller) AddressTest(opts *bind.CallOpts, arg0 common.Address) (struct {\n\tNumber *big.Int\n\tPhrase string\n}, error) {\n\tret := new(struct {\n\t\tNumber *big.Int\n\t\tPhrase string\n\t})\n\tout := ret\n\terr := _Mapping.contract.Call(opts, out, \"addressTest\", arg0)\n\treturn *ret, err\n}", "func (*OrderDetail_ShippingAddress) Descriptor() ([]byte, []int) {\n\treturn file_order_proto_rawDescGZIP(), []int{1, 17}\n}", "func (mt *EasypostShipment) Validate() (err error) {\n\tif mt.ID == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"id\"))\n\t}\n\tif mt.Object == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"object\"))\n\t}\n\n\tif mt.BuyerAddress != nil {\n\t\tif err2 := mt.BuyerAddress.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif mt.CustomsInfo != nil {\n\t\tif err2 := mt.CustomsInfo.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tfor _, e := range mt.Fees {\n\t\tif ok := goa.ValidatePattern(`^Fee$`, e.Object); !ok {\n\t\t\terr = goa.MergeErrors(err, goa.InvalidPatternError(`response.fees[*].object`, e.Object, `^Fee$`))\n\t\t}\n\t\tif e.Type != nil {\n\t\t\tif !(*e.Type == \"LabelFee\" || *e.Type == \"PostageFee\" || *e.Type == \"InsuranceFee\" || *e.Type == \"TrackerFee\") {\n\t\t\t\terr = goa.MergeErrors(err, goa.InvalidEnumValueError(`response.fees[*].type`, *e.Type, []interface{}{\"LabelFee\", \"PostageFee\", \"InsuranceFee\", \"TrackerFee\"}))\n\t\t\t}\n\t\t}\n\t}\n\tif mt.FromAddress != nil {\n\t\tif err2 := mt.FromAddress.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif ok := goa.ValidatePattern(`^shp_`, mt.ID); !ok {\n\t\terr = goa.MergeErrors(err, goa.InvalidPatternError(`response.id`, mt.ID, `^shp_`))\n\t}\n\tif mt.Insurance != nil {\n\t\tif err2 := mt.Insurance.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif !(mt.Mode == \"test\" || mt.Mode == \"production\") {\n\t\terr = goa.MergeErrors(err, goa.InvalidEnumValueError(`response.mode`, mt.Mode, []interface{}{\"test\", \"production\"}))\n\t}\n\tif ok := goa.ValidatePattern(`^Shipment$`, mt.Object); !ok {\n\t\terr = goa.MergeErrors(err, goa.InvalidPatternError(`response.object`, mt.Object, `^Shipment$`))\n\t}\n\tif mt.Options != nil {\n\t\tif err2 := mt.Options.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif mt.PostageLabel != nil {\n\t\tif err2 := mt.PostageLabel.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tfor _, e := range mt.Rates {\n\t\tif e.ID != nil {\n\t\t\tif ok := goa.ValidatePattern(`^rate_`, *e.ID); !ok {\n\t\t\t\terr = goa.MergeErrors(err, goa.InvalidPatternError(`response.rates[*].id`, *e.ID, `^rate_`))\n\t\t\t}\n\t\t}\n\t\tif !(e.Mode == \"test\" || e.Mode == \"production\") {\n\t\t\terr = goa.MergeErrors(err, goa.InvalidEnumValueError(`response.rates[*].mode`, e.Mode, []interface{}{\"test\", \"production\"}))\n\t\t}\n\t\tif ok := goa.ValidatePattern(`^Rate$`, e.Object); !ok {\n\t\t\terr = goa.MergeErrors(err, goa.InvalidPatternError(`response.rates[*].object`, e.Object, `^Rate$`))\n\t\t}\n\t}\n\tif mt.RefundStatus != nil {\n\t\tif !(*mt.RefundStatus == \"submitted\" || *mt.RefundStatus == \"rejected\" || *mt.RefundStatus == \"refunded\") {\n\t\t\terr = goa.MergeErrors(err, goa.InvalidEnumValueError(`response.refund_status`, *mt.RefundStatus, []interface{}{\"submitted\", \"rejected\", \"refunded\"}))\n\t\t}\n\t}\n\tif mt.ReturnAddress != nil {\n\t\tif err2 := mt.ReturnAddress.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif mt.ScanForm != nil {\n\t\tif err2 := mt.ScanForm.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif mt.SelectedRate != nil {\n\t\tif err2 := mt.SelectedRate.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif mt.ToAddress != nil {\n\t\tif err2 := mt.ToAddress.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\tif mt.Tracker != nil {\n\t\tif err2 := mt.Tracker.Validate(); err2 != nil {\n\t\t\terr = goa.MergeErrors(err, err2)\n\t\t}\n\t}\n\treturn\n}", "func (b *Builder) Post(addr value.Pointer, size uint64, p Postback) {\n\tif !addr.IsValid() {\n\t\tpanic(fmt.Errorf(\"Pointer address %v is not valid\", addr))\n\t}\n\tb.instructions = append(b.instructions, asm.Post{\n\t\tSource: b.remap(addr),\n\t\tSize: size,\n\t})\n\tb.decoders = append(b.decoders, postBackDecoder{\n\t\texpectedSize: int(size),\n\t\tdecode: p,\n\t})\n}", "func UpdateAddress(a *AddressDAL) (*AddressDAL, error) {\n\t_, err := GetAddress(a.AddressID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresult := db.DB().Save(a)\n\tif result.Error != nil {\n\t\treturn nil, result.Error\n\t}\n\treturn a, nil\n}", "func (svc *inmemService) GetAddress(ctx context.Context, profileID string, addressID string) (Address, error) {\n\n\t// Get a Read Lock on the svc for atomic read access to the datastore\n\tsvc.mtx.RLock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.RUnlock()\n\n\t// Check the data store to make sure the requested profile exists and set\n\tprofile, ok := svc.profiles[profileID]\n\n\t// If no entry for the profile was fund in the datastore\n\tif !ok {\n\n\t\t// Return an empty valued Address and an error informing the caller that no profile was found with the provided ID.\n\t\treturn Address{}, ErrNotFound\n\t}\n\n\t// Loop through each address attached to the found profile\n\tfor _, address := range profile.Addresses {\n\n\t\t// Check to see if the current address's ID matches the addressID passed in\n\t\tif address.ID == addressID {\n\n\t\t\t// Return that address and a nil error for a value\n\t\t\treturn address, nil\n\t\t}\n\t}\n\n\t// Return an empty Address value and a not found error since we were unable to find the specified address.\n\treturn Address{}, ErrNotFound\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParamsWithTimeout(timeout time.Duration) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func CreateAddress(address models.RequestAddress) (bool, models.ResponseAddress, error) {\n\n\t//Create request\n\trequest := models.Request{}\n\trequest.AddBody(address)\n\trequest.SetUri(\"https://api.easypost.com/v2/addresses\")\n\trequest.SetMethod(\"POST\")\n\n\t//Send request\n\tresponseBody, err := SendRequest(request)\n\n\t//Initialize response address\n\tresponseAddress := models.ResponseAddress{}\n\n\terr = json.Unmarshal(responseBody, &responseAddress)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn false, responseAddress, fmt.Errorf(\"Unrecognized response from easypost %s\", err.Error())\n\t}\n\n\treturn responseAddress.Verifications.Delivery.Success, responseAddress, err\n}", "func (c *Client) Address(assetCode ...string) *AddressService {\n\tif len(assetCode) == 0 {\n\t\tassetCode[0] = \"\"\n\t}\n\n\treturn &AddressService{service{c}, assetCode[0]}\n}", "func (addressManager *AddressManager) Address(addressIndex uint64) address.Address {\n\t// update lastUnspentAddressIndex if necessary\n\taddressManager.spentAddressIndexes(addressIndex)\n\n\treturn addressManager.seed.Address(addressIndex)\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParamsWithHTTPClient(client *http.Client) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\t\tHTTPClient: client,\n\t}\n}", "func GetAddrBalance(ee engine.Exchange) sknet.HandlerFunc {\n\treturn func(c *sknet.Context) error {\n\t\tvar rlt *pp.EmptyRes\n\t\tfor {\n\t\t\treq := pp.GetAddrBalanceReq{}\n\t\t\tif err := c.BindJSON(&req); err != nil {\n\t\t\t\tlogger.Error(err.Error())\n\t\t\t\trlt = pp.MakeErrResWithCode(pp.ErrCode_WrongRequest)\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tcoin, err := ee.GetCoin(req.GetCoinType())\n\t\t\tif err != nil {\n\t\t\t\trlt = pp.MakeErrRes(err)\n\t\t\t\tlogger.Error(err.Error())\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\taddrs := strings.Split(req.GetAddrs(), \",\")\n\t\t\tb, err := coin.GetBalance(addrs)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Error(err.Error())\n\t\t\t\trlt = pp.MakeErrRes(err)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tres := pp.GetAddrBalanceRes{\n\t\t\t\tResult: pp.MakeResultWithCode(pp.ErrCode_Success),\n\t\t\t\tBalance: &b,\n\t\t\t}\n\n\t\t\treturn c.SendJSON(&res)\n\t\t}\n\t\treturn c.Error(rlt)\n\t}\n}", "func (rpcServer HttpServer) estimateFeeWithEstimator(defaultFee int64, shardID byte, numBlock uint64, tokenId *common.Hash) uint64 {\n\testimateFeeCoinPerKb := uint64(0)\n\tif defaultFee == -1 {\n\t\tif _, ok := rpcServer.config.FeeEstimator[shardID]; ok {\n\t\t\ttemp, _ := rpcServer.config.FeeEstimator[shardID].EstimateFee(numBlock, tokenId)\n\t\t\testimateFeeCoinPerKb = uint64(temp)\n\t\t}\n\t\tif estimateFeeCoinPerKb == 0 {\n\t\t\tif feeEstimator, ok := rpcServer.config.FeeEstimator[shardID]; ok {\n\t\t\t\testimateFeeCoinPerKb = feeEstimator.GetLimitFee()\n\t\t\t}\n\t\t}\n\t} else {\n\t\testimateFeeCoinPerKb = uint64(defaultFee)\n\t}\n\treturn estimateFeeCoinPerKb\n}", "func (httpServer *HttpServer) handleEstimateFeeWithEstimator(params interface{}, closeChan <-chan struct{}) (interface{}, *rpcservice.RPCError) {\n\tLogger.log.Debugf(\"handleEstimateFeeWithEstimator params: %+v\", params)\n\t// all params\n\tarrayParams := common.InterfaceSlice(params)\n\tif arrayParams == nil || len(arrayParams) < 2 {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"Not enough params\"))\n\t}\n\t// param #1: estimation fee coin per kb from client\n\tdefaultFeeCoinPerKbTemp, ok := arrayParams[0].(float64)\n\tif !ok {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"defaultFeeCoinPerKbTemp is invalid\"))\n\t}\n\tdefaultFeeCoinPerKb := int64(defaultFeeCoinPerKbTemp)\n\n\t// param #2: payment address\n\tpaymentAddressParam, ok := arrayParams[1].(string)\n\tif !ok {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"sender key param is invalid\"))\n\t}\n\t_, shardIDSender, err := rpcservice.GetKeySetFromPaymentAddressParam(paymentAddressParam)\n\tif err != nil {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.InvalidSenderPrivateKeyError, err)\n\t}\n\n\t// param #2: numbloc\n\tnumblock := uint64(8)\n\tif len(arrayParams) >= 3 {\n\t\tnumBlockParam, ok := arrayParams[2].(float64)\n\t\tif !ok {\n\t\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"num block param is invalid\"))\n\t\t}\n\t\tnumblock = uint64(numBlockParam)\n\t}\n\n\t// param #3: tokenId\n\t// if tokenID != nil, return fee for privacy token\n\t// if tokenID != nil, return fee for native token\n\tvar tokenId *common.Hash\n\tif len(arrayParams) >= 4 && arrayParams[3] != nil {\n\t\ttokenIdParam, ok := arrayParams[3].(string)\n\t\tif !ok {\n\t\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"token id param is invalid\"))\n\t\t}\n\t\ttokenId, err = common.Hash{}.NewHashFromStr(tokenIdParam)\n\t\tif err != nil {\n\t\t\treturn nil, rpcservice.NewRPCError(rpcservice.UnexpectedError, err)\n\t\t}\n\t}\n\n\tbeaconState, err := httpServer.blockService.BlockChain.BestState.GetClonedBeaconBestState()\n\tbeaconHeight := beaconState.BeaconHeight\n\n\testimateFeeCoinPerKb, err := httpServer.txService.EstimateFeeWithEstimator(defaultFeeCoinPerKb, shardIDSender, numblock, tokenId, int64(beaconHeight), *httpServer.config.Database)\n\tif err != nil{\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.UnexpectedError, err)\n\t}\n\n\tresult := jsonresult.NewEstimateFeeResult(estimateFeeCoinPerKb, 0)\n\tLogger.log.Debugf(\"handleEstimateFeeWithEstimator result: %+v\", result)\n\treturn result, nil\n}", "func (rpcServer RpcServer) handleEstimateFeeWithEstimator(params interface{}, closeChan <-chan struct{}) (interface{}, *RPCError) {\n\tLogger.log.Infof(\"handleEstimateFeeWithEstimator params: %+v\", params)\n\t// all params\n\tarrayParams := common.InterfaceSlice(params)\n\tif len(arrayParams) < 2 {\n\t\treturn nil, NewRPCError(ErrRPCInvalidParams, errors.New(\"Not enough params\"))\n\t}\n\t// param #1: estimation fee coin per kb from client\n\tdefaultFeeCoinPerKbTemp, ok := arrayParams[0].(float64)\n\tif !ok {\n\t\treturn nil, NewRPCError(ErrRPCInvalidParams, errors.New(\"defaultFeeCoinPerKbTemp is invalid\"))\n\t}\n\tdefaultFeeCoinPerKb := int64(defaultFeeCoinPerKbTemp)\n\n\t// param #2: payment address\n\tsenderKeyParam := arrayParams[1]\n\tsenderKeySet, err := rpcServer.GetKeySetFromKeyParams(senderKeyParam.(string))\n\tif err != nil {\n\t\treturn nil, NewRPCError(ErrInvalidSenderPrivateKey, err)\n\t}\n\tlastByte := senderKeySet.PaymentAddress.Pk[len(senderKeySet.PaymentAddress.Pk)-1]\n\tshardIDSender := common.GetShardIDFromLastByte(lastByte)\n\n\t// param #2: numblocl\n\testimateFeeCoinPerKb := rpcServer.estimateFeeWithEstimator(defaultFeeCoinPerKb, shardIDSender, 8)\n\n\tresult := jsonresult.EstimateFeeResult{\n\t\tEstimateFeeCoinPerKb: estimateFeeCoinPerKb,\n\t}\n\tLogger.log.Infof(\"handleEstimateFeeWithEstimator result: %+v\", result)\n\treturn result, nil\n}", "func GetAddress(addr string) (*model.Address, error) {\n\n\turl := fmt.Sprintf(bchapi.AddressUrl, addr)\n\tresult, err := bchapi.HttpGet(url, bchapi.ConnTimeoutMS, bchapi.ServeTimeoutMS)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\taddress, err := model.StringToAddress(result)\n\treturn address, err\n}", "func (e Endpoints) GetAddresses(ctx context.Context, profileID string) ([]Address, error) {\n\n\t// TODO: Create detailed ref spec\n\trequest := getAddressesRequest{ProfileID: profileID}\n\n\tresponse, err := e.GetAddressesEndpoint(ctx, request)\n\n\tif err != nil {\n\t\treturn []Address{}, nil\n\t}\n\n\tresp := response.(getAddressesResponse)\n\n\treturn resp.Addresses, resp.Err\n\n}", "func EstimateTxFeeRequestHandlerFn(cliCtx context.CLIContext) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tvar req utils.EstimateFeeReq\n\n\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusBadRequest, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\terr = cliCtx.Codec.UnmarshalJSON(body, &req)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusBadRequest, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tgasAdjustment, err := utils.ParseFloat64(req.GasAdjustment, client.DefaultGasAdjustment)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusBadRequest, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tfees, gas, err := utils.ComputeFeesWithStdTx(cliCtx, req.Tx, gasAdjustment, req.GasPrices)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusInternalServerError, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tresponse := utils.EstimateFeeResp{Fees: fees, Gas: gas}\n\t\trest.PostProcessResponse(w, cliCtx, response)\n\t}\n}", "func (pg *PGStorage) Update(a *Address) error {\n\tvar err error\n\t_, err = pg.con.Exec(`\n\t\t\tUPDATE address set ballance = $1, income = $2, outcome = $3\n\t\t\tWHERE id=$4`,\n\t\ta.Ballance,\n\t\ta.Income,\n\t\ta.Outcome,\n\t\ta.ID,\n\t)\n\treturn err\n}", "func (o *moveTaskOrderUpdater) UpdatePostCounselingInfo(appCtx appcontext.AppContext, moveTaskOrderID uuid.UUID, eTag string) (*models.Move, error) {\n\t// Fetch the move and associations.\n\tsearchParams := services.MoveTaskOrderFetcherParams{\n\t\tIncludeHidden: false,\n\t\tMoveTaskOrderID: moveTaskOrderID,\n\t\tExcludeExternalShipments: true,\n\t}\n\tmoveTaskOrder, fetchErr := o.FetchMoveTaskOrder(appCtx, &searchParams)\n\tif fetchErr != nil {\n\t\treturn &models.Move{}, fetchErr\n\t}\n\n\tapprovedForPrimeCounseling := false\n\tfor _, serviceItem := range moveTaskOrder.MTOServiceItems {\n\t\tif serviceItem.ReService.Code == models.ReServiceCodeCS && serviceItem.Status == models.MTOServiceItemStatusApproved {\n\t\t\tapprovedForPrimeCounseling = true\n\t\t\tbreak\n\t\t}\n\t}\n\tif !approvedForPrimeCounseling {\n\t\treturn &models.Move{}, apperror.NewConflictError(moveTaskOrderID, \"Counseling is not an approved service item\")\n\t}\n\n\ttransactionError := appCtx.NewTransaction(func(txnAppCtx appcontext.AppContext) error {\n\t\t// Check the If-Match header against existing eTag before updating.\n\t\tencodedUpdatedAt := etag.GenerateEtag(moveTaskOrder.UpdatedAt)\n\t\tif encodedUpdatedAt != eTag {\n\t\t\treturn apperror.NewPreconditionFailedError(moveTaskOrderID, nil)\n\t\t}\n\n\t\tnow := time.Now()\n\t\tmoveTaskOrder.PrimeCounselingCompletedAt = &now\n\n\t\tverrs, err := appCtx.DB().ValidateAndSave(moveTaskOrder)\n\t\tif verrs != nil && verrs.HasAny() {\n\t\t\treturn apperror.NewInvalidInputError(moveTaskOrderID, nil, verrs, \"\")\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Note: Avoiding the copy of the element in the range so we can preserve the changes to the\n\t\t// statuses when we return the entire move tree.\n\t\tfor i := range moveTaskOrder.MTOShipments {\n\t\t\tif moveTaskOrder.MTOShipments[i].PPMShipment != nil {\n\t\t\t\tmoveTaskOrder.MTOShipments[i].PPMShipment.Status = models.PPMShipmentStatusWaitingOnCustomer\n\t\t\t\tmoveTaskOrder.MTOShipments[i].PPMShipment.ApprovedAt = &now\n\n\t\t\t\tverrs, err = appCtx.DB().ValidateAndSave(moveTaskOrder.MTOShipments[i].PPMShipment)\n\t\t\t\tif verrs != nil && verrs.HasAny() {\n\t\t\t\t\treturn apperror.NewInvalidInputError(moveTaskOrder.MTOShipments[i].PPMShipment.ID, nil, verrs, \"\")\n\t\t\t\t}\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n\n\tif transactionError != nil {\n\t\treturn &models.Move{}, transactionError\n\t}\n\n\treturn moveTaskOrder, nil\n}", "func NewGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParamsWithTimeout(timeout time.Duration) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\tvar ()\n\treturn &GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func NewGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParamsWithHTTPClient(client *http.Client) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\tvar ()\n\treturn &GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (s *Service) GetAddresses(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tsizeStr := r.FormValue(\"size\")\n\tprefix := r.FormValue(\"prefix\")\n\tif sizeStr == \"\" {\n\t\tsizeStr = defaultPageSize\n\t}\n\tdata := &Data{}\n\tdefer func() {\n\t\tif err := json.NewEncoder(w).Encode(data.Addresses); err != nil {\n\t\t\tutils.Logger().Warn().Err(err).Msg(\"cannot JSON-encode addresses\")\n\t\t}\n\t}()\n\n\tsize, err := strconv.Atoi(sizeStr)\n\tif err != nil || size > maxAddresses {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\tdata.Addresses, err = s.Storage.GetAddresses(size, prefix)\n\tif err != nil {\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\tutils.Logger().Warn().Err(err).Msg(\"wasn't able to fetch addresses from storage\")\n\t\treturn\n\t}\n}", "func (mt *EasypostAddress) Validate() (err error) {\n\tif mt.ID == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"id\"))\n\t}\n\tif mt.Object == \"\" {\n\t\terr = goa.MergeErrors(err, goa.MissingAttributeError(`response`, \"object\"))\n\t}\n\n\tif ok := goa.ValidatePattern(`^adr_`, mt.ID); !ok {\n\t\terr = goa.MergeErrors(err, goa.InvalidPatternError(`response.id`, mt.ID, `^adr_`))\n\t}\n\tif !(mt.Mode == \"test\" || mt.Mode == \"production\") {\n\t\terr = goa.MergeErrors(err, goa.InvalidEnumValueError(`response.mode`, mt.Mode, []interface{}{\"test\", \"production\"}))\n\t}\n\tif ok := goa.ValidatePattern(`^Address$`, mt.Object); !ok {\n\t\terr = goa.MergeErrors(err, goa.InvalidPatternError(`response.object`, mt.Object, `^Address$`))\n\t}\n\treturn\n}", "func (os *OutboundShipping) Update() *OutboundShippingUpdateOne {\n\treturn (&OutboundShippingClient{config: os.config}).UpdateOne(os)\n}", "func (a *Transactions) Address(ctx context.Context, address proto.WavesAddress, limit uint) ([]proto.Transaction, *Response, error) {\n\turl, err := joinUrl(a.options.BaseUrl, fmt.Sprintf(\"/transactions/address/%s/limit/%d\", address.String(), limit))\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", url.String(), nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar out []TransactionsField\n\tresponse, err := doHttp(ctx, a.options, req, &out)\n\tif err != nil {\n\t\treturn nil, response, err\n\t}\n\tif len(out) == 0 {\n\t\treturn nil, response, nil\n\t}\n\treturn out[0], response, nil\n}", "func MakeMTOShipment(db *pop.Connection, assertions Assertions) models.MTOShipment {\n\tmoveTaskOrder := assertions.MoveTaskOrder\n\tif isZeroUUID(moveTaskOrder.ID) {\n\t\tmoveTaskOrder = MakeMoveTaskOrder(db, assertions)\n\t}\n\n\tpickupAddress := MakeAddress(db, assertions)\n\tdestinationAddress := MakeAddress2(db, assertions)\n\tsecondaryPickupAddress := MakeAddress(db, assertions)\n\tsecondaryDeliveryAddress := MakeAddress(db, assertions)\n\tshipmentType := models.MTOShipmentTypeHHG\n\n\tif assertions.MTOShipment.ShipmentType != \"\" {\n\t\tshipmentType = assertions.MTOShipment.ShipmentType\n\t}\n\n\t// mock remarks\n\tremarks := \"please treat gently\"\n\trejectionReason := \"shipment not good enough\"\n\n\t// mock weights\n\tactualWeight := unit.Pound(980)\n\n\t// mock dates\n\tscheduledPickupDate := time.Date(TestYear, time.March, 16, 0, 0, 0, 0, time.UTC)\n\trequestedPickupDate := time.Date(TestYear, time.March, 15, 0, 0, 0, 0, time.UTC)\n\n\tMTOShipment := models.MTOShipment{\n\t\tMoveTaskOrder: moveTaskOrder,\n\t\tMoveTaskOrderID: moveTaskOrder.ID,\n\t\tScheduledPickupDate: &scheduledPickupDate,\n\t\tRequestedPickupDate: &requestedPickupDate,\n\t\tCustomerRemarks: &remarks,\n\t\tPickupAddress: &pickupAddress,\n\t\tPickupAddressID: &pickupAddress.ID,\n\t\tDestinationAddress: &destinationAddress,\n\t\tDestinationAddressID: &destinationAddress.ID,\n\t\tPrimeActualWeight: &actualWeight,\n\t\tSecondaryPickupAddress: &secondaryPickupAddress,\n\t\tSecondaryDeliveryAddress: &secondaryDeliveryAddress,\n\t\tShipmentType: shipmentType,\n\t\tStatus: \"SUBMITTED\",\n\t\tRejectionReason: &rejectionReason,\n\t}\n\n\tif assertions.MTOShipment.Status == models.MTOShipmentStatusApproved {\n\t\tapprovedDate := time.Date(TestYear, time.March, 20, 0, 0, 0, 0, time.UTC)\n\t\tMTOShipment.ApprovedDate = &approvedDate\n\t}\n\n\tif assertions.MTOShipment.ScheduledPickupDate != nil {\n\t\trequiredDeliveryDate := time.Date(TestYear, time.April, 15, 0, 0, 0, 0, time.UTC)\n\t\tMTOShipment.RequiredDeliveryDate = &requiredDeliveryDate\n\t}\n\n\t// Overwrite values with those from assertions\n\tmergeModels(&MTOShipment, assertions.MTOShipment)\n\n\tmustCreate(db, &MTOShipment)\n\n\treturn MTOShipment\n}", "func (auo *AddressUpdateOne) Save(ctx context.Context) (*Address, error) {\n\tvar (\n\t\terr error\n\t\tnode *Address\n\t)\n\tif len(auo.hooks) == 0 {\n\t\tnode, err = auo.sqlSave(ctx)\n\t} else {\n\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\tmutation, ok := m.(*AddressMutation)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t}\n\t\t\tauo.mutation = mutation\n\t\t\tnode, err = auo.sqlSave(ctx)\n\t\t\tmutation.done = true\n\t\t\treturn node, err\n\t\t})\n\t\tfor i := len(auo.hooks) - 1; i >= 0; i-- {\n\t\t\tmut = auo.hooks[i](mut)\n\t\t}\n\t\tif _, err := mut.Mutate(ctx, auo.mutation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn node, err\n}", "func TestEstimateFor(t *testing.T) {\n\tt.Parallel()\n\n\ttype fields struct {\n\t\tpathSVC pathUsecase.Service\n\t}\n\n\ttype args struct {\n\t\tride *rideDomain.Ride\n\t}\n\n\tpathSVCMock := new(pathUsecaseMocks.Service)\n\n\ttestCases := []struct {\n\t\tname string\n\t\tfields fields\n\t\targs args\n\t\texpectedDistance float64\n\t\texpectedFare *fareDomain.Fare\n\t}{\n\t\t{\n\t\t\tname: \"Happy path\",\n\t\t\tfields: fields{\n\t\t\t\tpathSVC: pathSVCMock,\n\t\t\t},\n\t\t\targs: args{\n\t\t\t\tride: &rideDomain.Ride{\n\t\t\t\t\tID: 1,\n\t\t\t\t\tPaths: []pathDomain.Path{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tLatitude: 37.966660,\n\t\t\t\t\t\t\tLongitude: 23.728308,\n\t\t\t\t\t\t\tTimestamp: 1405594957,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tLatitude: 37.935490,\n\t\t\t\t\t\t\tLongitude: 23.625655,\n\t\t\t\t\t\t\tTimestamp: 1405596220,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\texpectedDistance: 9.12,\n\t\t\texpectedFare: &fareDomain.Fare{\n\t\t\t\tRideID: 1,\n\t\t\t\tEstimatedAmount: 8.0488,\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, tc := range testCases {\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\tsvc := &serviceImpl{\n\t\t\t\tpathSVC: tc.fields.pathSVC,\n\t\t\t}\n\n\t\t\tstartPath := tc.args.ride.Paths[0]\n\t\t\tendPath := tc.args.ride.Paths[1]\n\n\t\t\t// set expectations\n\t\t\tpathSVCMock.On(\"CalculateDistance\", startPath, endPath).Return(tc.expectedDistance)\n\n\t\t\tif got := svc.estimateFor(tc.args.ride); !reflect.DeepEqual(got, tc.expectedFare) {\n\t\t\t\tt.Errorf(\"estimateFor(): %v, want: %v\", got, tc.expectedFare)\n\t\t\t}\n\t\t})\n\t}\n}", "func (ta TouristAttraction) AsPostalAddress() (*PostalAddress, bool) {\n\treturn nil, false\n}", "func (o *CheckoutShippingInformationManagementV1SaveAddressInformationPostDefault) Code() int {\n\treturn o._statusCode\n}", "func (fe FoodEstablishment) AsPostalAddress() (*PostalAddress, bool) {\n\treturn nil, false\n}", "func (bq *BRQuerier) GetAddress() string {\n\treturn bq.Address\n}", "func (as *AddressService) Create(name string) (*Address, error) {\n\tif isEmptyStr(as.assetCode) {\n\t\treturn nil, errAssetCode\n\t}\n\n\tvar (\n\t\taddresses []*Address\n\t\tbody struct {\n\t\t\tAddress struct {\n\t\t\t\tName string `json:\"name\"`\n\t\t\t} `json:\"address\"`\n\t\t}\n\t)\n\tbody.Address.Name = name\n\n\tif err := as.client.Post(buildString(\"address/deposit/\", as.assetCode), &body, &addresses); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn addresses[0], nil\n}", "func (s *Script) ExtractAddress() (types.Address, error) {\n\n\tswitch {\n\tcase s.IsPayToPubKeyHash():\n\t\tfallthrough\n\tcase s.IsTokenTransfer():\n\t\tfallthrough\n\tcase s.IsTokenIssue():\n\t\t_, pubKeyHash, _, err := s.getNthOp(2, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn types.NewAddressPubKeyHash(pubKeyHash)\n\tcase s.IsContractPubkey():\n\t\treturn s.ParseContractAddr()\n\tcase s.IsSplitAddrScript():\n\t\t_, pubKeyHash, _, err := s.getNthOp(1, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn types.NewSplitAddressFromHash(pubKeyHash)\n\tcase s.IsPayToPubKeyHashCLTVScript():\n\t\tl := len(*s)\n\t\t_, pubKeyHash, _, err := s.getNthOp(l-23, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn types.NewAddressPubKeyHash(pubKeyHash)\n\tcase s.IsPayToScriptHash():\n\t\t_, pubKeyHash, _, err := s.getNthOp(1, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn types.NewAddressPubKeyHash(pubKeyHash)\n\tdefault:\n\t\treturn nil, ErrAddressNotApplicable\n\t}\n}", "func (_Mapping *MappingCallerSession) AddressTest(arg0 common.Address) (struct {\n\tNumber *big.Int\n\tPhrase string\n}, error) {\n\treturn _Mapping.Contract.AddressTest(&_Mapping.CallOpts, arg0)\n}", "func (a *Client) PostReturnAddressesByReturnAddressIDDelete(params *PostReturnAddressesByReturnAddressIDDeleteParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesByReturnAddressIDDeleteOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesByReturnAddressIDDeleteParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesByReturnAddressIdDelete\",\n\t\tMethod: \"DELETE\",\n\t\tPathPattern: \"/post/return-addresses/{return_address_id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesByReturnAddressIDDeleteReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesByReturnAddressIDDeleteOK), nil\n\n}", "func (c *Client) BalanceForAddress(address string) (int, error) {\n\turl := fmt.Sprintf(\"%s/addr/%s/balance?token=%s\", c.URL, address, c.Token)\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tresp, err := c.HC.Do(req)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tbodyBytes, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tresponseString := string(bodyBytes)\n\tduffs, err := strconv.ParseInt(responseString, 10, 64)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn int(duffs), nil\n}", "func (_Mapping *MappingSession) AddressTest(arg0 common.Address) (struct {\n\tNumber *big.Int\n\tPhrase string\n}, error) {\n\treturn _Mapping.Contract.AddressTest(&_Mapping.CallOpts, arg0)\n}", "func GetAddress(id int32) (*AddressDAL, error) {\n\ta := &AddressDAL{}\n\tresult := db.DB().First(a, id)\n\tif result.Error != nil {\n\t\treturn nil, result.Error\n\t}\n\treturn a, nil\n}", "func (r *NucypherAccountRepository) UpdateAddress(address string, updatedBy string, accountID int, now time.Time) error {\n\n\t_, err := r.store.db.NamedExec(`UPDATE nucypher_accounts \n\tSET address=:address, updated_by=:updated_by, updated_at=:updated_at\n\tWHERE (created_by=:updated_by AND account_id=:account_id)`,\n\t\tmap[string]interface{}{\n\t\t\t\"address\": address,\n\t\t\t\"updated_by\": updatedBy,\n\t\t\t\"account_id\": accountID,\n\t\t\t\"updated_at\": now,\n\t\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n\n}", "func (d *AddressDeriver) Derive(change uint32, addressIndex uint32) *Address {\n\tif d.singleAddress != \"\" {\n\t\treturn &Address{\n\t\t\tpath: \"n/a\",\n\t\t\taddr: d.singleAddress,\n\t\t\tnet: d.network,\n\t\t\tchange: 0,\n\t\t\taddrIndex: 0,\n\t\t}\n\t}\n\n\tpath := fmt.Sprintf(\"m/.../%d/%d\", change, addressIndex)\n\taddr := &Address{path: path, net: d.network, change: change, addrIndex: addressIndex}\n\tif len(d.xpubs) == 1 {\n\t\taddr.addr = d.singleDerive(change, addressIndex)\n\t\treturn addr\n\t}\n\taddr.addr = d.multiSigSegwitDerive(change, addressIndex)\n\treturn addr\n}", "func (a Answer) AsPostalAddress() (*PostalAddress, bool) {\n\treturn nil, false\n}", "func (chaincode *Chaincode) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n\t// Extract the function and args from the transaction proposal\n\tfn, args := stub.GetFunctionAndParameters()\n\n\tswitch fn {\n\tcase \"exchange\": // echange addr currency coins\n\t\tif len(args) != 3 {\n\t\t\treturn shim.Error(fmt.Sprintf(\"expected 3 args (addr currency quantity), got %d\", len(args)))\n\t\t}\n\n\t\tvar addr, currencyName, quatityStr = args[0], args[1], args[2]\n\t\tvar quantity, errParseQuantity = strconv.ParseUint(quatityStr, 10, 64)\n\n\t\tif errParseQuantity == nil {\n\t\t\treturn shim.Error(errParseQuantity.Error())\n\t\t}\n\n\t\tvar price, errGetPrice = chaincode.GetPrice()\n\t\tif errGetPrice != nil {\n\t\t\treturn shim.Error(errGetPrice.Error())\n\t\t}\n\n\t\tvar marshalledWallet, errGetWallet = stub.GetState(addr)\n\t\tif errGetWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\tvar wallet, errParseWallet = strconv.ParseUint(string(marshalledWallet), 10, 64)\n\t\tif errParseWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\tvar currency, exists = price.Bpi[currencyName]\n\t\tif !exists {\n\t\t\treturn shim.Error(fmt.Sprintf(\"currency %q not found\", currencyName))\n\t\t}\n\n\t\twallet = uint64(currency.RateFloat) * quantity\n\n\t\tvar errPutState = stub.PutState(addr, []byte(strconv.FormatUint(wallet, 10)))\n\t\tif errPutState != nil {\n\t\t\treturn shim.Error(errPutState.Error())\n\t\t}\n\t\treturn shim.Success([]byte(fmt.Sprintf(\"wallet: %d\", wallet)))\n\tcase \"get\":\n\n\t\tif len(args) != 1 {\n\t\t\treturn shim.Error(fmt.Sprintf(\"expected 1 args (addr), got %d\", len(args)))\n\t\t}\n\n\t\tvar addr = args[0]\n\n\t\tvar marshalledWallet, errGetWallet = stub.GetState(addr)\n\t\tif errGetWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\tvar wallet, errParseWallet = strconv.ParseUint(string(marshalledWallet), 10, 64)\n\t\tif errParseWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\treturn shim.Success([]byte(fmt.Sprintf(\"wallet: %d\", wallet)))\n\tdefault:\n\t\treturn shim.Error(fmt.Sprintf(\"unknown command %q\", fn))\n\t}\n\t// Return the result as success payload\n\n}", "func (_IUniswapV2Factory *IUniswapV2FactoryCaller) FeeToSetter(opts *bind.CallOpts) (common.Address, error) {\r\n\tvar out []interface{}\r\n\terr := _IUniswapV2Factory.contract.Call(opts, &out, \"feeToSetter\")\r\n\r\n\tif err != nil {\r\n\t\treturn *new(common.Address), err\r\n\t}\r\n\r\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\r\n\r\n\treturn out0, err\r\n\r\n}", "func (as *ApiService) CreateDepositAddress(currency string) (*ApiResponse, error) {\n\treq := NewRequest(http.MethodPost, \"/api/v1/deposit-addresses\", map[string]string{\"currency\": currency})\n\treturn as.Call(req)\n}", "func (ioVar ImageObject) AsPostalAddress() (*PostalAddress, bool) {\n\treturn nil, false\n}", "func (o *CheckoutShippingInformationManagementV1SaveAddressInformationPostBody) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := o.validateAddressInformation(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (m *PlacesRequestBuilder) Post(ctx context.Context, body iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.Placeable, requestConfiguration *PlacesRequestBuilderPostRequestConfiguration)(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.Placeable, error) {\n requestInfo, err := m.CreatePostRequestInformation(ctx, body, requestConfiguration);\n if err != nil {\n return nil, err\n }\n errorMapping := i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.ErrorMappings {\n \"4XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n \"5XX\": ia572726a95efa92ddd544552cd950653dc691023836923576b2f4bf716cf204a.CreateODataErrorFromDiscriminatorValue,\n }\n res, err := m.requestAdapter.SendAsync(ctx, requestInfo, iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.CreatePlaceFromDiscriminatorValue, errorMapping)\n if err != nil {\n return nil, err\n }\n if res == nil {\n return nil, nil\n }\n return res.(iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.Placeable), nil\n}", "func (as *AddrServer) HandleAddrBalance(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\taddr := mux.Vars(r)[\"addr\"]\n\n\t// paginate through transactions\n\ttxns, err := as.GetAddressBalance([]string{addr})\n\tif err != nil {\n\t\tw.WriteHeader(400)\n\t\tw.Write(NewPostError(\"error fetching all transactions for address\", err))\n\t\treturn\n\t}\n\tout, _ := json.Marshal(txns.Result.Balance)\n\tw.Write(out)\n}", "func (a *Address) ZipCodeByState(stateName string) string {\n\treturn a.faker.Bothify(a.faker.MustParse(\"address.postcode_by_state.\" + stateName))\n}", "func GetAddressOnly(addID int) structs.Address {\n\tvar addr structs.Address\n\n\tdb := mysql.InitializeMySQL()\n\tsqlQueryAddr := \"select id, province_id, province_name, city_id, city_name, kecamatan_id, kecamatan_name, kelurahan_id, kelurahan_name, zipcode from address_map where id = ?\"\n\tres, err := db.Query(sqlQueryAddr, addID)\n\tdefer mysql.CloseRows(res)\n\n\tif err != nil {\n\t\treturn addr\n\t}\n\n\tfor res.Next() {\n\t\tres.Scan(&addr.ID, &addr.ProvinceID, &addr.ProvinceName, &addr.CityID, &addr.CityName, &addr.KecamatanID, &addr.KecamatanName, &addr.KelurahanID, &addr.KelurahanName, &addr.ZipCode)\n\t}\n\treturn addr\n}", "func (s *Service) GetExplorerAddress(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tid := r.FormValue(\"id\")\n\tif strings.HasPrefix(id, \"0x\") {\n\t\tparsedAddr := common2.ParseAddr(id)\n\t\toneAddr, err := common2.AddressToBech32(parsedAddr)\n\t\tif err != nil {\n\t\t\tutils.Logger().Warn().Err(err).Msg(\"unrecognized address format\")\n\t\t\tw.WriteHeader(http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t\tid = oneAddr\n\t}\n\tkey := GetAddressKey(id)\n\ttxViewParam := r.FormValue(\"tx_view\")\n\tpageParam := r.FormValue(\"page\")\n\toffsetParam := r.FormValue(\"offset\")\n\torder := r.FormValue(\"order\")\n\ttxView := txViewNone\n\tif txViewParam != \"\" {\n\t\ttxView = txViewParam\n\t}\n\tutils.Logger().Info().Str(\"Address\", id).Msg(\"Querying address\")\n\tdata := &Data{}\n\tdefer func() {\n\t\tif err := json.NewEncoder(w).Encode(data.Address); err != nil {\n\t\t\tutils.Logger().Warn().Err(err).Msg(\"cannot JSON-encode Address\")\n\t\t}\n\t}()\n\tif id == \"\" {\n\t\tutils.Logger().Warn().Msg(\"missing address id param\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\tvar err error\n\tvar offset int\n\tif offsetParam != \"\" {\n\t\toffset, err = strconv.Atoi(offsetParam)\n\t\tif err != nil || offset < 1 {\n\t\t\tutils.Logger().Warn().Msg(\"invalid offset parameter\")\n\t\t\tw.WriteHeader(http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t} else {\n\t\toffset = paginationOffset\n\t}\n\tvar page int\n\tif pageParam != \"\" {\n\t\tpage, err = strconv.Atoi(pageParam)\n\t\tif err != nil {\n\t\t\tutils.Logger().Warn().Err(err).Msg(\"invalid page parameter\")\n\t\t\tw.WriteHeader(http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\t} else {\n\t\tpage = 0\n\t}\n\n\tdata.Address.ID = id\n\t// Try to populate the banace by directly calling get balance.\n\t// Check the balance from blockchain rather than local DB dump\n\tbalanceAddr := big.NewInt(0)\n\tif s.GetAccountBalance != nil {\n\t\taddress := common2.ParseAddr(id)\n\t\tbalance, err := s.GetAccountBalance(address)\n\t\tif err == nil {\n\t\t\tbalanceAddr = balance\n\t\t}\n\t}\n\n\tdb := s.Storage.GetDB()\n\tbytes, err := db.Get([]byte(key))\n\tif err != nil {\n\t\tutils.Logger().Warn().Err(err).Str(\"id\", id).Msg(\"cannot fetch address from db\")\n\t\tdata.Address.Balance = balanceAddr\n\t\treturn\n\t}\n\n\tif err = rlp.DecodeBytes(bytes, &data.Address); err != nil {\n\t\tutils.Logger().Warn().Str(\"id\", id).Msg(\"cannot convert address data\")\n\t\tdata.Address.Balance = balanceAddr\n\t\treturn\n\t}\n\n\tdata.Address.Balance = balanceAddr\n\n\tswitch txView {\n\tcase txViewNone:\n\t\tdata.Address.TXs = nil\n\tcase Received:\n\t\treceivedTXs := make([]*Transaction, 0)\n\t\tfor _, tx := range data.Address.TXs {\n\t\t\tif tx.Type == Received {\n\t\t\t\treceivedTXs = append(receivedTXs, tx)\n\t\t\t}\n\t\t}\n\t\tdata.Address.TXs = receivedTXs\n\tcase Sent:\n\t\tsentTXs := make([]*Transaction, 0)\n\t\tfor _, tx := range data.Address.TXs {\n\t\t\tif tx.Type == Sent {\n\t\t\t\tsentTXs = append(sentTXs, tx)\n\t\t\t}\n\t\t}\n\t\tdata.Address.TXs = sentTXs\n\t}\n\tif offset*page >= len(data.Address.TXs) {\n\t\tdata.Address.TXs = []*Transaction{}\n\t} else if offset*page+offset > len(data.Address.TXs) {\n\t\tdata.Address.TXs = data.Address.TXs[offset*page:]\n\t} else {\n\t\tdata.Address.TXs = data.Address.TXs[offset*page : offset*page+offset]\n\t}\n\tif order == \"DESC\" {\n\t\tsort.Slice(data.Address.TXs[:], func(i, j int) bool {\n\t\t\treturn data.Address.TXs[i].Timestamp > data.Address.TXs[j].Timestamp\n\t\t})\n\t} else {\n\t\tsort.Slice(data.Address.TXs[:], func(i, j int) bool {\n\t\t\treturn data.Address.TXs[i].Timestamp < data.Address.TXs[j].Timestamp\n\t\t})\n\t}\n}", "func (r *Restaurant) QueryAddress() *AddressQuery {\n\treturn (&RestaurantClient{config: r.config}).QueryAddress(r)\n}", "func (o *DKSharesInfo) GetAddress() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Address\n}", "func (*BgpNeighbor) PostPath() string {\n\treturn \"/api/objects/bgp/neighbor/\"\n}", "func NewPostGenerateAddressesParams() *PostGenerateAddressesParams {\n\tvar ()\n\treturn &PostGenerateAddressesParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}" ]
[ "0.7513478", "0.6583214", "0.62728506", "0.6064372", "0.5696682", "0.5493261", "0.5445068", "0.54150504", "0.5304678", "0.524681", "0.521757", "0.50714326", "0.5026679", "0.4987674", "0.49597186", "0.4869086", "0.48330227", "0.4767647", "0.47553527", "0.46762457", "0.46656176", "0.46553367", "0.46468893", "0.45908692", "0.4582162", "0.4581187", "0.45372558", "0.44778305", "0.44746318", "0.44424048", "0.4433992", "0.44228247", "0.4418164", "0.44139922", "0.4413724", "0.44017908", "0.4394037", "0.43662027", "0.43657154", "0.43628034", "0.43484682", "0.42996758", "0.42812034", "0.42743394", "0.42638457", "0.4261464", "0.42574087", "0.42365187", "0.4232754", "0.42320633", "0.42053163", "0.4194893", "0.41939682", "0.41826463", "0.41813397", "0.41724223", "0.41511425", "0.4144907", "0.41337344", "0.41292194", "0.4122372", "0.41220173", "0.4118262", "0.41169953", "0.4114712", "0.41104597", "0.4087867", "0.4087102", "0.40803242", "0.40778384", "0.40732017", "0.4072699", "0.4064873", "0.40606382", "0.40531686", "0.4044013", "0.40439656", "0.40372235", "0.40367246", "0.403528", "0.40295476", "0.40191963", "0.3996605", "0.39965644", "0.39930725", "0.3990333", "0.3989755", "0.39876154", "0.3974598", "0.3972348", "0.39711016", "0.39698052", "0.39689308", "0.3955295", "0.39522249", "0.39518076", "0.39505324", "0.39403957", "0.39269274", "0.3924645" ]
0.79218256
0
/ QuoteShippingMethodManagementV1EstimateByAddressIDPostMine Estimate shipping
func (a *Client) QuoteShippingMethodManagementV1EstimateByAddressIDPostMine(params *QuoteShippingMethodManagementV1EstimateByAddressIDPostMineParams) (*QuoteShippingMethodManagementV1EstimateByAddressIDPostMineOK, error) { // TODO: Validate the params before sending if params == nil { params = NewQuoteShippingMethodManagementV1EstimateByAddressIDPostMineParams() } result, err := a.transport.Submit(&runtime.ClientOperation{ ID: "quoteShippingMethodManagementV1EstimateByAddressIdPostMine", Method: "POST", PathPattern: "/V1/carts/mine/estimate-shipping-methods-by-address-id", ProducesMediaTypes: []string{""}, ConsumesMediaTypes: []string{""}, Schemes: []string{"http"}, Params: params, Reader: &QuoteShippingMethodManagementV1EstimateByAddressIDPostMineReader{formats: a.formats}, Context: params.Context, Client: params.HTTPClient, }) if err != nil { return nil, err } return result.(*QuoteShippingMethodManagementV1EstimateByAddressIDPostMineOK), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (a *Client) QuoteShippingMethodManagementV1EstimateByAddressIDPost(params *QuoteShippingMethodManagementV1EstimateByAddressIDPostParams) (*QuoteShippingMethodManagementV1EstimateByAddressIDPostOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewQuoteShippingMethodManagementV1EstimateByAddressIDPostParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"quoteShippingMethodManagementV1EstimateByAddressIdPost\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/V1/carts/{cartId}/estimate-shipping-methods-by-address-id\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &QuoteShippingMethodManagementV1EstimateByAddressIDPostReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*QuoteShippingMethodManagementV1EstimateByAddressIDPostOK), nil\n\n}", "func NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK() *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK {\n\treturn &NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK{}\n}", "func NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault(code int) *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault {\n\treturn &NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (o *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostBody) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := o.validateAddressID(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineBody) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody)\n\treturn o\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) SetGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineBody) {\n\to.GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody = giftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody\n}", "func (o *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault) Code() int {\n\treturn o._statusCode\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithTimeout(timeout time.Duration) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithContext(ctx context.Context) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody)\n\treturn o\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) SetGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody(giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody) {\n\to.GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody = giftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody\n}", "func NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized() *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized {\n\treturn &NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized{}\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams() *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithHTTPClient(client *http.Client) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParamsWithHTTPClient(client *http.Client) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\t\tHTTPClient: client,\n\t}\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParamsWithTimeout(timeout time.Duration) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithContext(ctx context.Context) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithTimeout(timeout time.Duration) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (a API) EstimateFee(cmd *btcjson.EstimateFeeCmd) (e error) {\n\tRPCHandlers[\"estimatefee\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithHTTPClient(client *http.Client) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (o *NegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 401:\n\t\tresult := NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\tresult := NewNegotiableQuoteShippingMethodManagementV1EstimateByAddressIDPostDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (e Endpoints) PostAddress(ctx context.Context, profileID string, a Address) error {\n\n\t// TODO: Create detailed ref spec\n\trequest := postAddressRequest{ProfileID: profileID, Address: a}\n\n\tresponse, err := e.PostAddressEndpoint(ctx, request)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tresp := response.(postAddressResponse)\n\n\treturn resp.Err\n}", "func main() {\n\tmoveIDString := flag.String(\"moveID\", \"\", \"The ID of the move where shipments are found\")\n\tenv := flag.String(\"env\", \"development\", \"The environment to run in, which configures the database.\")\n\tflag.Parse()\n\n\tif *moveIDString == \"\" {\n\t\tlog.Fatal(\"Usage: generate_shipment_edi -moveID <29cb984e-c70d-46f0-926d-cd89e07a6ec3>\")\n\t}\n\n\tdb, err := pop.Connect(*env)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tmoveID := uuid.Must(uuid.FromString(*moveIDString))\n\tvar shipments models.Shipments\n\n\terr = db.Eager(\n\t\t\"Move.Orders\",\n\t\t\"PickupAddress\",\n\t\t\"DeliveryAddress\",\n\t\t\"ServiceMember\",\n\t\t\"ShipmentOffers.TransportationServiceProviderPerformance\",\n\t).Where(\"shipment_offers.accepted=true\").\n\t\tWhere(\"move_id = $1\", &moveID).\n\t\tJoin(\"shipment_offers\", \"shipment_offers.shipment_id = shipments.id\").\n\t\tAll(&shipments)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif len(shipments) == 0 {\n\t\tlog.Fatal(\"No accepted shipments found\")\n\t}\n\tvar logger = zap.NewNop()\n\n\tvar costsByShipments []rateengine.CostByShipment\n\n\tengine := rateengine.NewRateEngine(db, logger, route.NewTestingPlanner(362)) //TODO: create the proper route/planner\n\tfor _, shipment := range shipments {\n\t\tcostByShipment, err := engine.HandleRunOnShipment(shipment)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tcostsByShipments = append(costsByShipments, costByShipment)\n\t}\n\tedi, err := ediinvoice.Generate858C(costsByShipments, db)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tfmt.Println(edi)\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParamsWithContext(ctx context.Context) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\n\t\tContext: ctx,\n\t}\n}", "func EstimateFee(tx StdTx) StdFee {\n\treturn NewStdFee(txparam.DefaultMsgGas*uint64(len(tx.Msgs)), tx.Fee.GasPrice)\n}", "func (svc *inmemService) PostAddress(ctx context.Context, profileID string, address Address) error {\n\n\t// Get a Lock on the svc for atomic access to the datastore\n\tsvc.mtx.Lock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.Unlock()\n\n\t// Check the datastore for the specified profile and set it to a profile variable when found\n\tprofile, ok := svc.profiles[profileID]\n\n\t// If the profile was not found in the datastore\n\tif !ok {\n\n\t\t// Return an error to the caller informing them that the specified profile could not be found\n\t\treturn ErrNotFound\n\t}\n\n\t// Iterate over the profile's existent addresses\n\tfor _, existingAddress := range profile.Addresses {\n\n\t\t// Check to see if the current existent address has an ID that matches the ID of the address passed in\n\t\tif existingAddress.ID == address.ID {\n\n\t\t\t// Return an error informing the caller that address they wanted to add is already present\n\t\t\treturn ErrAlreadyExists\n\t\t}\n\t}\n\n\t// Add the new address to the profile Address collection\n\tprofile.Addresses = append(profile.Addresses, address)\n\n\t// Replace the profile with the one that has the new Address added to the Address collection\n\tsvc.profiles[profileID] = profile\n\n\t// Return a nil error value\n\treturn nil\n}", "func NewGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams() *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\tvar ()\n\treturn &GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func (cs *checkoutService) quoteShipping(ctx context.Context, address *pb.Address, items []*pb.CartItem) (*pb.Money, error) {\n\tshippingQuote, err := pb.NewShippingServiceClient(cs.shippingSvcConn).\n\t\tGetQuote(ctx, &pb.GetQuoteRequest{\n\t\t\tAddress: address,\n\t\t\tItems: items})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get shipping quote: %+v\", err)\n\t}\n\treturn shippingQuote.GetCostUsd(), nil\n}", "func (rpcServer HttpServer) estimateFeeWithEstimator(defaultFee int64, shardID byte, numBlock uint64, tokenId *common.Hash) uint64 {\n\testimateFeeCoinPerKb := uint64(0)\n\tif defaultFee == -1 {\n\t\tif _, ok := rpcServer.config.FeeEstimator[shardID]; ok {\n\t\t\ttemp, _ := rpcServer.config.FeeEstimator[shardID].EstimateFee(numBlock, tokenId)\n\t\t\testimateFeeCoinPerKb = uint64(temp)\n\t\t}\n\t\tif estimateFeeCoinPerKb == 0 {\n\t\t\tif feeEstimator, ok := rpcServer.config.FeeEstimator[shardID]; ok {\n\t\t\t\testimateFeeCoinPerKb = feeEstimator.GetLimitFee()\n\t\t\t}\n\t\t}\n\t} else {\n\t\testimateFeeCoinPerKb = uint64(defaultFee)\n\t}\n\treturn estimateFeeCoinPerKb\n}", "func (c *Client) Addresses(ctx context.Context, foreignID, currency string) (Address, error) {\n\treqBody := map[string]string{\n\t\t\"foreign_id\": foreignID,\n\t\t\"currency\": currency,\n\t}\n\n\treqJSON, err := json.Marshal(reqBody)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request body marshaling error: %w\", err)\n\t}\n\n\taddressesURL, err := joinURL(c.api, addressesEndpoint)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request url creating error: %w\", err)\n\t}\n\n\treq, err := http.NewRequestWithContext(ctx, http.MethodPost, addressesURL.String(), bytes.NewBuffer(reqJSON))\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request creating error: %w\", err)\n\t}\n\n\tsig, err := createHmac(c.secret, reqJSON)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"hmac signature creationg error: %w\", err)\n\t}\n\n\treq.Header.Set(contentTypeHeader, jsonContentType)\n\treq.Header.Set(keyHeader, c.apiKey)\n\treq.Header.Set(signatureHeader, sig)\n\n\tresp, err := c.client.Do(req)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request error: %w\", err)\n\t}\n\tdefer resp.Body.Close()\n\n\terr = ensureSuccessResponse(resp)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"request failed: %w\", err)\n\t}\n\n\trespBody := struct {\n\t\tData Address `json:\"data\"`\n\t}{}\n\n\terr = json.NewDecoder(resp.Body).Decode(&respBody)\n\tif err != nil {\n\t\treturn Address{}, fmt.Errorf(\"response unmarshaling error: %w\", err)\n\t}\n\n\treturn respBody.Data, nil\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostBody); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WithCartID(cartID string) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\to.SetCartID(cartID)\n\treturn o\n}", "func (s *OrderShippingAddressService) Show(ctx context.Context, orderID int, id int) (*OrderShippingAddress, *http.Response, error) {\n\tosa := new(OrderShippingAddress)\n\tapiError := new(APIError)\n\n\tresp, err := performRequest(ctx, s.sling.New().Get(fmt.Sprintf(\"%d/shipping_addresses/%d\", orderID, id)), s.httpClient, osa, apiError)\n\treturn osa, resp, relevantError(err, *apiError)\n}", "func TestEstimateFor(t *testing.T) {\n\tt.Parallel()\n\n\ttype fields struct {\n\t\tpathSVC pathUsecase.Service\n\t}\n\n\ttype args struct {\n\t\tride *rideDomain.Ride\n\t}\n\n\tpathSVCMock := new(pathUsecaseMocks.Service)\n\n\ttestCases := []struct {\n\t\tname string\n\t\tfields fields\n\t\targs args\n\t\texpectedDistance float64\n\t\texpectedFare *fareDomain.Fare\n\t}{\n\t\t{\n\t\t\tname: \"Happy path\",\n\t\t\tfields: fields{\n\t\t\t\tpathSVC: pathSVCMock,\n\t\t\t},\n\t\t\targs: args{\n\t\t\t\tride: &rideDomain.Ride{\n\t\t\t\t\tID: 1,\n\t\t\t\t\tPaths: []pathDomain.Path{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tLatitude: 37.966660,\n\t\t\t\t\t\t\tLongitude: 23.728308,\n\t\t\t\t\t\t\tTimestamp: 1405594957,\n\t\t\t\t\t\t},\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tLatitude: 37.935490,\n\t\t\t\t\t\t\tLongitude: 23.625655,\n\t\t\t\t\t\t\tTimestamp: 1405596220,\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t\texpectedDistance: 9.12,\n\t\t\texpectedFare: &fareDomain.Fare{\n\t\t\t\tRideID: 1,\n\t\t\t\tEstimatedAmount: 8.0488,\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, tc := range testCases {\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\tsvc := &serviceImpl{\n\t\t\t\tpathSVC: tc.fields.pathSVC,\n\t\t\t}\n\n\t\t\tstartPath := tc.args.ride.Paths[0]\n\t\t\tendPath := tc.args.ride.Paths[1]\n\n\t\t\t// set expectations\n\t\t\tpathSVCMock.On(\"CalculateDistance\", startPath, endPath).Return(tc.expectedDistance)\n\n\t\t\tif got := svc.estimateFor(tc.args.ride); !reflect.DeepEqual(got, tc.expectedFare) {\n\t\t\t\tt.Errorf(\"estimateFor(): %v, want: %v\", got, tc.expectedFare)\n\t\t\t}\n\t\t})\n\t}\n}", "func GetAddress(latlng []float64, key string, useTier bool, tier common.AlertTier) (string, error) {\n\t// Retrieve location from MapQuest Geocoding API\n\tgeocoder.SetAPIKey(key)\n\ta, err := geocoder.ReverseGeocode(latlng[0], latlng[1])\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\t// Filter data returned depending on provided AlertTier\n\tvar address string\n\tif a.Street != \"\" && tier == common.FIRST {\n\t\taddress = fmt.Sprintf(\"%v, \", a.Street)\n\t}\n\n\tif a.City != \"\" && (tier == common.FIRST || tier == common.SECOND) {\n\t\taddress = address + fmt.Sprintf(\"%v, \", a.City)\n\t}\n\tif a.State != \"\" {\n\t\taddress = address + fmt.Sprintf(\"%v, \", a.State)\n\t}\n\tif a.PostalCode != \"\" {\n\t\taddress = address + fmt.Sprintf(\"%v, \", a.PostalCode)\n\t}\n\tif a.CountryCode != \"\" {\n\t\taddress = address + fmt.Sprintf(\"%v\", a.CountryCode)\n\t}\n\n\treturn address, nil\n}", "func (a *Api) AddressAtHeight(address string, height int) (*AddressResult, error) {\n\n\tparams := make(map[string]string)\n\tparams[\"address\"] = address\n\tif height > 0 {\n\t\tparams[\"height\"] = strconv.Itoa(height)\n\t}\n\n\tres, err := a.client.R().SetQueryParams(params).Get(\"/address\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif res.IsError() {\n\t\treturn nil, NewResponseError(res)\n\t}\n\n\tresponse := new(AddressResponse)\n\terr = json.Unmarshal(res.Body(), response)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif response.Error != nil {\n\t\treturn nil, response.Error\n\t}\n\n\treturn response.Result, nil\n}", "func (rpcServer RpcServer) handleEstimateFeeWithEstimator(params interface{}, closeChan <-chan struct{}) (interface{}, *RPCError) {\n\tLogger.log.Infof(\"handleEstimateFeeWithEstimator params: %+v\", params)\n\t// all params\n\tarrayParams := common.InterfaceSlice(params)\n\tif len(arrayParams) < 2 {\n\t\treturn nil, NewRPCError(ErrRPCInvalidParams, errors.New(\"Not enough params\"))\n\t}\n\t// param #1: estimation fee coin per kb from client\n\tdefaultFeeCoinPerKbTemp, ok := arrayParams[0].(float64)\n\tif !ok {\n\t\treturn nil, NewRPCError(ErrRPCInvalidParams, errors.New(\"defaultFeeCoinPerKbTemp is invalid\"))\n\t}\n\tdefaultFeeCoinPerKb := int64(defaultFeeCoinPerKbTemp)\n\n\t// param #2: payment address\n\tsenderKeyParam := arrayParams[1]\n\tsenderKeySet, err := rpcServer.GetKeySetFromKeyParams(senderKeyParam.(string))\n\tif err != nil {\n\t\treturn nil, NewRPCError(ErrInvalidSenderPrivateKey, err)\n\t}\n\tlastByte := senderKeySet.PaymentAddress.Pk[len(senderKeySet.PaymentAddress.Pk)-1]\n\tshardIDSender := common.GetShardIDFromLastByte(lastByte)\n\n\t// param #2: numblocl\n\testimateFeeCoinPerKb := rpcServer.estimateFeeWithEstimator(defaultFeeCoinPerKb, shardIDSender, 8)\n\n\tresult := jsonresult.EstimateFeeResult{\n\t\tEstimateFeeCoinPerKb: estimateFeeCoinPerKb,\n\t}\n\tLogger.log.Infof(\"handleEstimateFeeWithEstimator result: %+v\", result)\n\treturn result, nil\n}", "func (httpServer *HttpServer) handleEstimateFeeWithEstimator(params interface{}, closeChan <-chan struct{}) (interface{}, *rpcservice.RPCError) {\n\tLogger.log.Debugf(\"handleEstimateFeeWithEstimator params: %+v\", params)\n\t// all params\n\tarrayParams := common.InterfaceSlice(params)\n\tif arrayParams == nil || len(arrayParams) < 2 {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"Not enough params\"))\n\t}\n\t// param #1: estimation fee coin per kb from client\n\tdefaultFeeCoinPerKbTemp, ok := arrayParams[0].(float64)\n\tif !ok {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"defaultFeeCoinPerKbTemp is invalid\"))\n\t}\n\tdefaultFeeCoinPerKb := int64(defaultFeeCoinPerKbTemp)\n\n\t// param #2: payment address\n\tpaymentAddressParam, ok := arrayParams[1].(string)\n\tif !ok {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"sender key param is invalid\"))\n\t}\n\t_, shardIDSender, err := rpcservice.GetKeySetFromPaymentAddressParam(paymentAddressParam)\n\tif err != nil {\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.InvalidSenderPrivateKeyError, err)\n\t}\n\n\t// param #2: numbloc\n\tnumblock := uint64(8)\n\tif len(arrayParams) >= 3 {\n\t\tnumBlockParam, ok := arrayParams[2].(float64)\n\t\tif !ok {\n\t\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"num block param is invalid\"))\n\t\t}\n\t\tnumblock = uint64(numBlockParam)\n\t}\n\n\t// param #3: tokenId\n\t// if tokenID != nil, return fee for privacy token\n\t// if tokenID != nil, return fee for native token\n\tvar tokenId *common.Hash\n\tif len(arrayParams) >= 4 && arrayParams[3] != nil {\n\t\ttokenIdParam, ok := arrayParams[3].(string)\n\t\tif !ok {\n\t\t\treturn nil, rpcservice.NewRPCError(rpcservice.RPCInvalidParamsError, errors.New(\"token id param is invalid\"))\n\t\t}\n\t\ttokenId, err = common.Hash{}.NewHashFromStr(tokenIdParam)\n\t\tif err != nil {\n\t\t\treturn nil, rpcservice.NewRPCError(rpcservice.UnexpectedError, err)\n\t\t}\n\t}\n\n\tbeaconState, err := httpServer.blockService.BlockChain.BestState.GetClonedBeaconBestState()\n\tbeaconHeight := beaconState.BeaconHeight\n\n\testimateFeeCoinPerKb, err := httpServer.txService.EstimateFeeWithEstimator(defaultFeeCoinPerKb, shardIDSender, numblock, tokenId, int64(beaconHeight), *httpServer.config.Database)\n\tif err != nil{\n\t\treturn nil, rpcservice.NewRPCError(rpcservice.UnexpectedError, err)\n\t}\n\n\tresult := jsonresult.NewEstimateFeeResult(estimateFeeCoinPerKb, 0)\n\tLogger.log.Debugf(\"handleEstimateFeeWithEstimator result: %+v\", result)\n\treturn result, nil\n}", "func (e StaticFeeEstimator) EstimateFeePerByte(numBlocks uint32) (btcutil.Amount, error) {\n\treturn e.FeeRate, nil\n}", "func NewGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParamsWithHTTPClient(client *http.Client) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\tvar ()\n\treturn &GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (a *Client) QuoteShippingMethodManagementV1GetListGetMine(params *QuoteShippingMethodManagementV1GetListGetMineParams) (*QuoteShippingMethodManagementV1GetListGetMineOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewQuoteShippingMethodManagementV1GetListGetMineParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"quoteShippingMethodManagementV1GetListGetMine\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/V1/carts/mine/shipping-methods\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &QuoteShippingMethodManagementV1GetListGetMineReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*QuoteShippingMethodManagementV1GetListGetMineOK), nil\n\n}", "func (*OrderDetail_ShippingAddress) Descriptor() ([]byte, []int) {\n\treturn file_order_proto_rawDescGZIP(), []int{1, 17}\n}", "func NewQuoteBillingAddressManagementV1AssignPostMineDefault(code int) *QuoteBillingAddressManagementV1AssignPostMineDefault {\n\treturn &QuoteBillingAddressManagementV1AssignPostMineDefault{\n\t\t_statusCode: code,\n\t}\n}", "func EstimateTxFeeRequestHandlerFn(cliCtx context.CLIContext) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tvar req utils.EstimateFeeReq\n\n\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusBadRequest, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\terr = cliCtx.Codec.UnmarshalJSON(body, &req)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusBadRequest, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tgasAdjustment, err := utils.ParseFloat64(req.GasAdjustment, client.DefaultGasAdjustment)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusBadRequest, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tfees, gas, err := utils.ComputeFeesWithStdTx(cliCtx, req.Tx, gasAdjustment, req.GasPrices)\n\t\tif err != nil {\n\t\t\trest.WriteErrorResponse(w, http.StatusInternalServerError, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tresponse := utils.EstimateFeeResp{Fees: fees, Gas: gas}\n\t\trest.PostProcessResponse(w, cliCtx, response)\n\t}\n}", "func (e Endpoints) GetAddress(ctx context.Context, profileID string, addressID string) (Address, error) {\n\n\t// TODO: Create detailed ref spec\n\trequest := getAddressRequest{ProfileID: profileID, AddressID: addressID}\n\n\tresponse, err := e.GetAddressEndpoint(ctx, request)\n\n\tif err != nil {\n\t\treturn Address{}, err\n\t}\n\n\tresp := response.(getAddressResponse)\n\n\treturn resp.Address, resp.Err\n}", "func (rpcServer HttpServer) estimateFee(\n\tdefaultFee int64,\n\tcandidateOutputCoins []*privacy.OutputCoin,\n\tpaymentInfos []*privacy.PaymentInfo, shardID byte,\n\tnumBlock uint64, hasPrivacy bool,\n\tmetadata metadata.Metadata,\n\tcustomTokenParams *transaction.CustomTokenParamTx,\n\tprivacyCustomTokenParams *transaction.CustomTokenPrivacyParamTx) (uint64, uint64, uint64) {\n\tif numBlock == 0 {\n\t\tnumBlock = 1000\n\t}\n\t// check real fee(nano PRV) per tx\n\tvar realFee uint64\n\testimateFeeCoinPerKb := uint64(0)\n\testimateTxSizeInKb := uint64(0)\n\n\ttokenId := &common.Hash{}\n\tif privacyCustomTokenParams != nil {\n\t\ttokenId, _ = common.Hash{}.NewHashFromStr(privacyCustomTokenParams.PropertyID)\n\t}\n\n\testimateFeeCoinPerKb = rpcServer.estimateFeeWithEstimator(defaultFee, shardID, numBlock, tokenId)\n\n\tif rpcServer.config.Wallet != nil {\n\t\testimateFeeCoinPerKb += uint64(rpcServer.config.Wallet.GetConfig().IncrementalFee)\n\t}\n\n\tlimitFee := uint64(0)\n\tif feeEstimator, ok := rpcServer.config.FeeEstimator[shardID]; ok {\n\t\tlimitFee = feeEstimator.GetLimitFee()\n\t}\n\testimateTxSizeInKb = transaction.EstimateTxSize(transaction.NewEstimateTxSizeParam(candidateOutputCoins, paymentInfos, hasPrivacy, metadata, customTokenParams, privacyCustomTokenParams, limitFee))\n\n\trealFee = uint64(estimateFeeCoinPerKb) * uint64(estimateTxSizeInKb)\n\treturn realFee, estimateFeeCoinPerKb, estimateTxSizeInKb\n}", "func (a *Client) PostReturnAddressesGet(params *PostReturnAddressesGetParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesGetOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesGetParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesGet\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/post/return-addresses\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesGetReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesGetOK), nil\n\n}", "func (_IUniswapV2Router02 *IUniswapV2Router02Caller) Quote(opts *bind.CallOpts, amountA *big.Int, reserveA *big.Int, reserveB *big.Int) (*big.Int, error) {\r\n\tvar out []interface{}\r\n\terr := _IUniswapV2Router02.contract.Call(opts, &out, \"quote\", amountA, reserveA, reserveB)\r\n\r\n\tif err != nil {\r\n\t\treturn *new(*big.Int), err\r\n\t}\r\n\r\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\r\n\r\n\treturn out0, err\r\n\r\n}", "func (addressManager *AddressManager) Address(addressIndex uint64) address.Address {\n\t// update lastUnspentAddressIndex if necessary\n\taddressManager.spentAddressIndexes(addressIndex)\n\n\treturn addressManager.seed.Address(addressIndex)\n}", "func NewGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParamsWithTimeout(timeout time.Duration) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\tvar ()\n\treturn &GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams{\n\n\t\ttimeout: timeout,\n\t}\n}", "func (a *Client) PostReturnAddressesByReturnAddressIDGet(params *PostReturnAddressesByReturnAddressIDGetParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesByReturnAddressIDGetOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesByReturnAddressIDGetParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesByReturnAddressIdGet\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/post/return-addresses/{return_address_id}\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesByReturnAddressIDGetReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesByReturnAddressIDGetOK), nil\n\n}", "func (_UniswapV2Router02 *UniswapV2Router02Caller) Quote(opts *bind.CallOpts, amountA *big.Int, reserveA *big.Int, reserveB *big.Int) (*big.Int, error) {\r\n\tvar out []interface{}\r\n\terr := _UniswapV2Router02.contract.Call(opts, &out, \"quote\", amountA, reserveA, reserveB)\r\n\r\n\tif err != nil {\r\n\t\treturn *new(*big.Int), err\r\n\t}\r\n\r\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\r\n\r\n\treturn out0, err\r\n\r\n}", "func (_Mapping *MappingCaller) AddressTest(opts *bind.CallOpts, arg0 common.Address) (struct {\n\tNumber *big.Int\n\tPhrase string\n}, error) {\n\tret := new(struct {\n\t\tNumber *big.Int\n\t\tPhrase string\n\t})\n\tout := ret\n\terr := _Mapping.contract.Call(opts, out, \"addressTest\", arg0)\n\treturn *ret, err\n}", "func (i *Invoice) GetShippingAddressRequested() (value bool) {\n\tif i == nil {\n\t\treturn\n\t}\n\treturn i.Flags.Has(4)\n}", "func (svc *inmemService) GetAddress(ctx context.Context, profileID string, addressID string) (Address, error) {\n\n\t// Get a Read Lock on the svc for atomic read access to the datastore\n\tsvc.mtx.RLock()\n\n\t// Immediately set up a lock release to occur when the function finishes\n\tdefer svc.mtx.RUnlock()\n\n\t// Check the data store to make sure the requested profile exists and set\n\tprofile, ok := svc.profiles[profileID]\n\n\t// If no entry for the profile was fund in the datastore\n\tif !ok {\n\n\t\t// Return an empty valued Address and an error informing the caller that no profile was found with the provided ID.\n\t\treturn Address{}, ErrNotFound\n\t}\n\n\t// Loop through each address attached to the found profile\n\tfor _, address := range profile.Addresses {\n\n\t\t// Check to see if the current address's ID matches the addressID passed in\n\t\tif address.ID == addressID {\n\n\t\t\t// Return that address and a nil error for a value\n\t\t\treturn address, nil\n\t\t}\n\t}\n\n\t// Return an empty Address value and a not found error since we were unable to find the specified address.\n\treturn Address{}, ErrNotFound\n}", "func (a *Api) Address(address string) (*AddressResult, error) {\n\treturn a.AddressAtHeight(address, LatestBlockHeight)\n}", "func (_IUniswapV2Router01 *IUniswapV2Router01Caller) Quote(opts *bind.CallOpts, amountA *big.Int, reserveA *big.Int, reserveB *big.Int) (*big.Int, error) {\r\n\tvar out []interface{}\r\n\terr := _IUniswapV2Router01.contract.Call(opts, &out, \"quote\", amountA, reserveA, reserveB)\r\n\r\n\tif err != nil {\r\n\t\treturn *new(*big.Int), err\r\n\t}\r\n\r\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\r\n\r\n\treturn out0, err\r\n\r\n}", "func (e StaticFeeEstimator) EstimateFeePerWeight(numBlocks uint32) (btcutil.Amount, error) {\n\treturn e.FeeRate / blockchain.WitnessScaleFactor, nil\n}", "func (o *CheckoutPaymentInformationManagementV1SavePaymentInformationAndPlaceOrderPostMineParams) WithTimeout(timeout time.Duration) *CheckoutPaymentInformationManagementV1SavePaymentInformationAndPlaceOrderPostMineParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (n *Node) Estimate(e graph2.EstimateNode) float64 {\n\treturn n.Data.(graph2.Estimator).Estimate(e)\n}", "func MakeMTOShipment(db *pop.Connection, assertions Assertions) models.MTOShipment {\n\tmoveTaskOrder := assertions.MoveTaskOrder\n\tif isZeroUUID(moveTaskOrder.ID) {\n\t\tmoveTaskOrder = MakeMoveTaskOrder(db, assertions)\n\t}\n\n\tpickupAddress := MakeAddress(db, assertions)\n\tdestinationAddress := MakeAddress2(db, assertions)\n\tsecondaryPickupAddress := MakeAddress(db, assertions)\n\tsecondaryDeliveryAddress := MakeAddress(db, assertions)\n\tshipmentType := models.MTOShipmentTypeHHG\n\n\tif assertions.MTOShipment.ShipmentType != \"\" {\n\t\tshipmentType = assertions.MTOShipment.ShipmentType\n\t}\n\n\t// mock remarks\n\tremarks := \"please treat gently\"\n\trejectionReason := \"shipment not good enough\"\n\n\t// mock weights\n\tactualWeight := unit.Pound(980)\n\n\t// mock dates\n\tscheduledPickupDate := time.Date(TestYear, time.March, 16, 0, 0, 0, 0, time.UTC)\n\trequestedPickupDate := time.Date(TestYear, time.March, 15, 0, 0, 0, 0, time.UTC)\n\n\tMTOShipment := models.MTOShipment{\n\t\tMoveTaskOrder: moveTaskOrder,\n\t\tMoveTaskOrderID: moveTaskOrder.ID,\n\t\tScheduledPickupDate: &scheduledPickupDate,\n\t\tRequestedPickupDate: &requestedPickupDate,\n\t\tCustomerRemarks: &remarks,\n\t\tPickupAddress: &pickupAddress,\n\t\tPickupAddressID: &pickupAddress.ID,\n\t\tDestinationAddress: &destinationAddress,\n\t\tDestinationAddressID: &destinationAddress.ID,\n\t\tPrimeActualWeight: &actualWeight,\n\t\tSecondaryPickupAddress: &secondaryPickupAddress,\n\t\tSecondaryDeliveryAddress: &secondaryDeliveryAddress,\n\t\tShipmentType: shipmentType,\n\t\tStatus: \"SUBMITTED\",\n\t\tRejectionReason: &rejectionReason,\n\t}\n\n\tif assertions.MTOShipment.Status == models.MTOShipmentStatusApproved {\n\t\tapprovedDate := time.Date(TestYear, time.March, 20, 0, 0, 0, 0, time.UTC)\n\t\tMTOShipment.ApprovedDate = &approvedDate\n\t}\n\n\tif assertions.MTOShipment.ScheduledPickupDate != nil {\n\t\trequiredDeliveryDate := time.Date(TestYear, time.April, 15, 0, 0, 0, 0, time.UTC)\n\t\tMTOShipment.RequiredDeliveryDate = &requiredDeliveryDate\n\t}\n\n\t// Overwrite values with those from assertions\n\tmergeModels(&MTOShipment, assertions.MTOShipment)\n\n\tmustCreate(db, &MTOShipment)\n\n\treturn MTOShipment\n}", "func (o *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\t// path param cartId\n\tif err := r.SetPathParam(\"cartId\", o.CartID); err != nil {\n\t\treturn err\n\t}\n\n\tif err := r.SetBodyParam(o.GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostBody); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (wlt *WhitelistableToken) egMint(address, amount string) (uint64, error) {\n\t// method\n\ttransferFnSignature := []byte(\"mint(address,uint256)\")\n\thash := sha3.NewLegacyKeccak256()\n\thash.Write(transferFnSignature)\n\tmethodID := hash.Sum(nil)[:4]\n\n\t// address\n\taddr := common.HexToAddress(address)\n\tpaddedAddress := common.LeftPadBytes(addr.Bytes(), 32)\n\n\t// amount\n\tamountBN := new(big.Int)\n\tamountBN.SetString(amount, 10)\n\tpaddedAmount := common.LeftPadBytes(amountBN.Bytes(), 32)\n\n\tvar data []byte\n\tdata = append(data, methodID...)\n\tdata = append(data, paddedAddress...)\n\tdata = append(data, paddedAmount...)\n\n\tgasLimit, err := wlt.EthClient.EstimateGas(context.Background(), ethereum.CallMsg{\n\t\tFrom: *wlt.CallerAddres,\n\t\tTo: wlt.ContractAddress,\n\t\tData: data,\n\t})\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn gasLimit, nil\n}", "func (os *OutboundShipping) Update() *OutboundShippingUpdateOne {\n\treturn (&OutboundShippingClient{config: os.config}).UpdateOne(os)\n}", "func Explode(address string) (*Address, error) {\n\tif client == nil {\n\t\terr := initClient()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Build the API request.\n\treq := &maps.GeocodingRequest{\n\t\tAddress: address,\n\t}\n\n\t// Execute the request.\n\tresp, err := client.Geocode(context.Background(), req)\n\tif len(resp) < 1 {\n\t\treturn nil, err\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Using the first/closest match in our response, grab the values we need.\n\tcomponents := resp[0].AddressComponents\n\tformattedAddress := resp[0].FormattedAddress\n\tlat := resp[0].Geometry.Location.Lat\n\tlng := resp[0].Geometry.Location.Lng\n\n\t// Construct the return *Address{}\n\tresponse := &Address{\n\t\tAddressLine1: compose(addressLine1Composition, \"\", components, false),\n\t\tAddressLine2: compose(addressLine2Composition, addressLineDelimeter, components, false),\n\t\tAddressCity: compose(addressCityComposition, addressLineDelimeter, components, false),\n\t\tAddressState: compose(addressStateComposition, addressLineDelimeter, components, false),\n\t\tAddressCountry: compose(addressCountryComposition, addressLineDelimeter, components, false),\n\t\tAddressCountryCode: compose(addressCountryCodeComposition, addressLineDelimeter, components, true),\n\t\tAddressZip: compose(addressPostalCodeComposition, addressLineDelimeter, components, false),\n\t\tAddressLat: &lat,\n\t\tAddressLng: &lng,\n\t\tFormattedAddress: &formattedAddress,\n\t}\n\n\treturn response, err\n}", "func (chaincode *Chaincode) Invoke(stub shim.ChaincodeStubInterface) peer.Response {\n\t// Extract the function and args from the transaction proposal\n\tfn, args := stub.GetFunctionAndParameters()\n\n\tswitch fn {\n\tcase \"exchange\": // echange addr currency coins\n\t\tif len(args) != 3 {\n\t\t\treturn shim.Error(fmt.Sprintf(\"expected 3 args (addr currency quantity), got %d\", len(args)))\n\t\t}\n\n\t\tvar addr, currencyName, quatityStr = args[0], args[1], args[2]\n\t\tvar quantity, errParseQuantity = strconv.ParseUint(quatityStr, 10, 64)\n\n\t\tif errParseQuantity == nil {\n\t\t\treturn shim.Error(errParseQuantity.Error())\n\t\t}\n\n\t\tvar price, errGetPrice = chaincode.GetPrice()\n\t\tif errGetPrice != nil {\n\t\t\treturn shim.Error(errGetPrice.Error())\n\t\t}\n\n\t\tvar marshalledWallet, errGetWallet = stub.GetState(addr)\n\t\tif errGetWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\tvar wallet, errParseWallet = strconv.ParseUint(string(marshalledWallet), 10, 64)\n\t\tif errParseWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\tvar currency, exists = price.Bpi[currencyName]\n\t\tif !exists {\n\t\t\treturn shim.Error(fmt.Sprintf(\"currency %q not found\", currencyName))\n\t\t}\n\n\t\twallet = uint64(currency.RateFloat) * quantity\n\n\t\tvar errPutState = stub.PutState(addr, []byte(strconv.FormatUint(wallet, 10)))\n\t\tif errPutState != nil {\n\t\t\treturn shim.Error(errPutState.Error())\n\t\t}\n\t\treturn shim.Success([]byte(fmt.Sprintf(\"wallet: %d\", wallet)))\n\tcase \"get\":\n\n\t\tif len(args) != 1 {\n\t\t\treturn shim.Error(fmt.Sprintf(\"expected 1 args (addr), got %d\", len(args)))\n\t\t}\n\n\t\tvar addr = args[0]\n\n\t\tvar marshalledWallet, errGetWallet = stub.GetState(addr)\n\t\tif errGetWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\tvar wallet, errParseWallet = strconv.ParseUint(string(marshalledWallet), 10, 64)\n\t\tif errParseWallet != nil {\n\t\t\treturn shim.Error(errGetWallet.Error())\n\t\t}\n\n\t\treturn shim.Success([]byte(fmt.Sprintf(\"wallet: %d\", wallet)))\n\tdefault:\n\t\treturn shim.Error(fmt.Sprintf(\"unknown command %q\", fn))\n\t}\n\t// Return the result as success payload\n\n}", "func (_m *MTOShipmentUpdater) UpdateMTOShipmentPrime(appCtx appcontext.AppContext, mtoShipment *models.MTOShipment, eTag string) (*models.MTOShipment, error) {\n\tret := _m.Called(appCtx, mtoShipment, eTag)\n\n\tvar r0 *models.MTOShipment\n\tif rf, ok := ret.Get(0).(func(appcontext.AppContext, *models.MTOShipment, string) *models.MTOShipment); ok {\n\t\tr0 = rf(appCtx, mtoShipment, eTag)\n\t} else {\n\t\tif ret.Get(0) != nil {\n\t\t\tr0 = ret.Get(0).(*models.MTOShipment)\n\t\t}\n\t}\n\n\tvar r1 error\n\tif rf, ok := ret.Get(1).(func(appcontext.AppContext, *models.MTOShipment, string) error); ok {\n\t\tr1 = rf(appCtx, mtoShipment, eTag)\n\t} else {\n\t\tr1 = ret.Error(1)\n\t}\n\n\treturn r0, r1\n}", "func (s *Service) UpdateAddressPOI(address common.Address, senderTotalFee *big.Int, poiPeriod uint64) {\n\t/*if senderTotalFee.Sign() == 0 {\n\t\ts.store.SetAddressPOI(address, common.Big0)\n\t\treturn // avoid division by 0\n\t}\n\tpoi := new(big.Int).Mul(senderTotalFee, lachesis.PercentUnit)\n\tpoi.Div(poi, s.store.GetPoiFee(poiPeriod)) // rebase user's PoI as <= 1.0 ratio\n\ts.store.SetAddressPOI(address, poi)*/\n}", "func (_IUniswapV2Factory *IUniswapV2FactoryCaller) FeeToSetter(opts *bind.CallOpts) (common.Address, error) {\r\n\tvar out []interface{}\r\n\terr := _IUniswapV2Factory.contract.Call(opts, &out, \"feeToSetter\")\r\n\r\n\tif err != nil {\r\n\t\treturn *new(common.Address), err\r\n\t}\r\n\r\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\r\n\r\n\treturn out0, err\r\n\r\n}", "func (o *CheckoutPaymentInformationManagementV1SavePaymentInformationAndPlaceOrderPostMineParams) WithHTTPClient(client *http.Client) *CheckoutPaymentInformationManagementV1SavePaymentInformationAndPlaceOrderPostMineParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func withAddressID(id int64) addressOption {\n\treturn func(m *AddressMutation) {\n\t\tvar (\n\t\t\terr error\n\t\t\tonce sync.Once\n\t\t\tvalue *Address\n\t\t)\n\t\tm.oldValue = func(ctx context.Context) (*Address, error) {\n\t\t\tonce.Do(func() {\n\t\t\t\tif m.done {\n\t\t\t\t\terr = fmt.Errorf(\"querying old values post mutation is not allowed\")\n\t\t\t\t} else {\n\t\t\t\t\tvalue, err = m.Client().Address.Get(ctx, id)\n\t\t\t\t}\n\t\t\t})\n\t\t\treturn value, err\n\t\t}\n\t\tm.id = &id\n\t}\n}", "func EstimateFeeWei(callbackGasLimit uint32, maxGasPriceWei *big.Int) (*big.Int, error) {\n\tmaxGasUsed := big.NewInt(int64(callbackGasLimit + GasProofVerification))\n\tcostWei := maxGasUsed.Mul(maxGasUsed, maxGasPriceWei)\n\treturn costWei, nil\n}", "func ShipperShipped(orderID uint) error {\n\tctx := context.Background()\n\t_, err := zbClient.NewPublishMessageCommand().MessageName(\"ShipperShipped\").CorrelationKey(fmt.Sprint(orderID)).TimeToLive(1 * time.Minute).Send(ctx)\n\tif err != nil {\n\t\t// failed to set the updated variables\n\t\treturn err\n\t}\n\treturn nil\n}", "func (trpn *TestRetrievalProviderNode) GetMinerWorkerAddress(ctx context.Context, addr address.Address, tok shared.TipSetToken) (address.Address, error) {\n\treturn addr, nil\n}", "func (b *BitcoindFeeEstimator) EstimateFeePerByte(numBlocks uint32) (btcutil.Amount, error) {\n\tfeeEstimate, err := b.fetchEstimatePerByte(numBlocks)\n\tswitch {\n\t// If the estimator doesn't have enough data, or returns an error, then\n\t// to return a proper value, then we'll return the default fall back\n\t// fee rate.\n\tcase err != nil:\n\t\twalletLog.Errorf(\"unable to query estimator: %v\", err)\n\t\tfallthrough\n\n\tcase feeEstimate == 0:\n\t\treturn b.fallBackFeeRate, nil\n\t}\n\n\treturn feeEstimate, nil\n}", "func (d *AddressDeriver) Derive(change uint32, addressIndex uint32) *Address {\n\tif d.singleAddress != \"\" {\n\t\treturn &Address{\n\t\t\tpath: \"n/a\",\n\t\t\taddr: d.singleAddress,\n\t\t\tnet: d.network,\n\t\t\tchange: 0,\n\t\t\taddrIndex: 0,\n\t\t}\n\t}\n\n\tpath := fmt.Sprintf(\"m/.../%d/%d\", change, addressIndex)\n\taddr := &Address{path: path, net: d.network, change: change, addrIndex: addressIndex}\n\tif len(d.xpubs) == 1 {\n\t\taddr.addr = d.singleDerive(change, addressIndex)\n\t\treturn addr\n\t}\n\taddr.addr = d.multiSigSegwitDerive(change, addressIndex)\n\treturn addr\n}", "func NewQuoteBillingAddressManagementV1AssignPostMineOK() *QuoteBillingAddressManagementV1AssignPostMineOK {\n\treturn &QuoteBillingAddressManagementV1AssignPostMineOK{}\n}", "func (b *CustomerRequestBuilder) ShipmentMethod() *ShipmentMethodRequestBuilder {\n\tbb := &ShipmentMethodRequestBuilder{BaseRequestBuilder: b.BaseRequestBuilder}\n\tbb.baseURL += \"/shipmentMethod\"\n\treturn bb\n}", "func MakePostAddressEndpoint(s Service) endpoint.Endpoint {\n\treturn func(ctx context.Context, request interface{}) (interface{}, error) {\n\n\t\t// Assert the request interface passed in as a postProfileRequest value\n\t\t// req will be a value of type postAddressRequest with the values properly mapped from the dynamic value passed in\n\t\treq := request.(postAddressRequest)\n\n\t\t// Create an error based on the return from the Service's PostAddress function\n\t\te := s.PostAddress(ctx, req.ProfileID, req.Address)\n\n\t\treturn postAddressResponse{Err: e}, nil\n\n\t}\n}", "func (auo *AddressUpdateOne) Save(ctx context.Context) (*Address, error) {\n\tvar (\n\t\terr error\n\t\tnode *Address\n\t)\n\tif len(auo.hooks) == 0 {\n\t\tnode, err = auo.sqlSave(ctx)\n\t} else {\n\t\tvar mut Mutator = MutateFunc(func(ctx context.Context, m Mutation) (Value, error) {\n\t\t\tmutation, ok := m.(*AddressMutation)\n\t\t\tif !ok {\n\t\t\t\treturn nil, fmt.Errorf(\"unexpected mutation type %T\", m)\n\t\t\t}\n\t\t\tauo.mutation = mutation\n\t\t\tnode, err = auo.sqlSave(ctx)\n\t\t\tmutation.done = true\n\t\t\treturn node, err\n\t\t})\n\t\tfor i := len(auo.hooks) - 1; i >= 0; i-- {\n\t\t\tmut = auo.hooks[i](mut)\n\t\t}\n\t\tif _, err := mut.Mutate(ctx, auo.mutation); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn node, err\n}", "func ShipperConfirmed(orderID uint) error {\n\tctx := context.Background()\n\t_, err := zbClient.NewPublishMessageCommand().MessageName(\"ShipperConfirmed\").CorrelationKey(fmt.Sprint(orderID)).TimeToLive(1 * time.Minute).Send(ctx)\n\tif err != nil {\n\t\t// failed to set the updated variables\n\t\treturn err\n\t}\n\treturn nil\n}", "func (dcr *ExchangeWallet) estimateSwap(lots, lotSize, feeSuggestion uint64, utxos []*compositeUTXO,\n\tnfo *dex.Asset, trySplit bool, feeBump float64) (*asset.SwapEstimate, bool /*split used*/, uint64 /* locked */, error) {\n\n\tvar avail uint64\n\tfor _, utxo := range utxos {\n\t\tavail += toAtoms(utxo.rpc.Amount)\n\t}\n\n\t// If there is a fee bump, the networkFeeRate can be higher than the\n\t// MaxFeeRate\n\tbumpedMaxRate := nfo.MaxFeeRate\n\tbumpedNetRate := feeSuggestion\n\tif feeBump > 1 {\n\t\tbumpedMaxRate = uint64(math.Round(float64(bumpedMaxRate) * feeBump))\n\t\tbumpedNetRate = uint64(math.Round(float64(bumpedNetRate) * feeBump))\n\t}\n\n\tval := lots * lotSize\n\tsum, inputsSize, _, _, _, err := dcr.tryFund(utxos, orderEnough(val, lots, bumpedMaxRate, nfo))\n\tif err != nil {\n\t\treturn nil, false, 0, err\n\t}\n\n\treqFunds := calc.RequiredOrderFundsAlt(val, uint64(inputsSize), lots, nfo.SwapSizeBase, nfo.SwapSize, bumpedMaxRate)\n\tmaxFees := reqFunds - val\n\n\testHighFunds := calc.RequiredOrderFundsAlt(val, uint64(inputsSize), lots, nfo.SwapSizeBase, nfo.SwapSize, bumpedNetRate)\n\testHighFees := estHighFunds - val\n\n\testLowFunds := calc.RequiredOrderFundsAlt(val, uint64(inputsSize), 1, nfo.SwapSizeBase, nfo.SwapSize, bumpedNetRate)\n\testLowFunds += dexdcr.P2SHOutputSize * (lots - 1) * bumpedNetRate\n\testLowFees := estLowFunds - val\n\n\t// Math for split transactions is a little different.\n\tif trySplit {\n\t\textraFees := splitTxBaggage * bumpedMaxRate\n\t\tsplitFees := splitTxBaggage * bumpedNetRate\n\t\tif avail >= reqFunds+extraFees {\n\t\t\tlocked := val + maxFees + extraFees\n\t\t\treturn &asset.SwapEstimate{\n\t\t\t\tLots: lots,\n\t\t\t\tValue: val,\n\t\t\t\tMaxFees: maxFees + extraFees,\n\t\t\t\tRealisticBestCase: estLowFees + splitFees,\n\t\t\t\tRealisticWorstCase: estHighFees + splitFees,\n\t\t\t}, true, locked, nil\n\t\t}\n\t}\n\n\t// No split transaction.\n\treturn &asset.SwapEstimate{\n\t\tLots: lots,\n\t\tValue: val,\n\t\tMaxFees: maxFees,\n\t\tRealisticBestCase: estLowFees,\n\t\tRealisticWorstCase: estHighFees,\n\t}, false, sum, nil\n}", "func (a API) EstimateFeeGetRes() (out *float64, e error) {\n\tout, _ = a.Result.(*float64)\n\te, _ = a.Result.(error)\n\treturn \n}", "func ShipperReceivedMoney(orderID uint) error {\n\tctx := context.Background()\n\t_, err := zbClient.NewPublishMessageCommand().MessageName(\"ShipperReceivedMoney\").CorrelationKey(fmt.Sprint(orderID)).TimeToLive(1 * time.Minute).Send(ctx)\n\tif err != nil {\n\t\t// failed to set the updated variables\n\t\treturn err\n\t}\n\treturn nil\n}", "func (_IUniswapV2Router02 *IUniswapV2Router02CallerSession) Quote(amountA *big.Int, reserveA *big.Int, reserveB *big.Int) (*big.Int, error) {\r\n\treturn _IUniswapV2Router02.Contract.Quote(&_IUniswapV2Router02.CallOpts, amountA, reserveA, reserveB)\r\n}", "func (_Mapping *MappingCallerSession) AddressTest(arg0 common.Address) (struct {\n\tNumber *big.Int\n\tPhrase string\n}, error) {\n\treturn _Mapping.Contract.AddressTest(&_Mapping.CallOpts, arg0)\n}", "func (b *Bitcoind) GetReceivedByAddress(address string, minconf uint32) (amount float64, err error) {\n\tr, err := b.client.call(\"getreceivedbyaddress\", []interface{}{address, minconf})\n\tif err = handleError(err, &r); err != nil {\n\t\treturn\n\t}\n\n\terr = json.Unmarshal(r.Result, &amount)\n\treturn\n}", "func GetAddressOnly(addID int) structs.Address {\n\tvar addr structs.Address\n\n\tdb := mysql.InitializeMySQL()\n\tsqlQueryAddr := \"select id, province_id, province_name, city_id, city_name, kecamatan_id, kecamatan_name, kelurahan_id, kelurahan_name, zipcode from address_map where id = ?\"\n\tres, err := db.Query(sqlQueryAddr, addID)\n\tdefer mysql.CloseRows(res)\n\n\tif err != nil {\n\t\treturn addr\n\t}\n\n\tfor res.Next() {\n\t\tres.Scan(&addr.ID, &addr.ProvinceID, &addr.ProvinceName, &addr.CityID, &addr.CityName, &addr.KecamatanID, &addr.KecamatanName, &addr.KelurahanID, &addr.KelurahanName, &addr.ZipCode)\n\t}\n\treturn addr\n}", "func (o *QuoteBillingAddressManagementV1AssignPostMineBody) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := o.validateAddress(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (_BaseAccessWallet *BaseAccessWalletCaller) AddressKMS(opts *bind.CallOpts) (common.Address, error) {\n\tvar out []interface{}\n\terr := _BaseAccessWallet.contract.Call(opts, &out, \"addressKMS\")\n\n\tif err != nil {\n\t\treturn *new(common.Address), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(common.Address)).(*common.Address)\n\n\treturn out0, err\n\n}", "func (bq *BRQuerier) GetAddress() string {\n\treturn bq.Address\n}", "func NewGiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParamsWithContext(ctx context.Context) *GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams {\n\tvar ()\n\treturn &GiftRegistryGuestCartShippingMethodManagementV1EstimateByRegistryIDPostParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (_Contract *ContractCaller) ProposalBurntFee(opts *bind.CallOpts) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _Contract.contract.Call(opts, &out, \"proposalBurntFee\")\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func (_Bindings *BindingsCaller) InterestRateModel(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Bindings.contract.Call(opts, out, \"interestRateModel\")\n\treturn *ret0, err\n}", "func main() {\n\te1 := NewQuote(\"EURUSD\", 1.23450)\n\te2 := NewQuote(\"EURUSD\", 1.23470)\n\n\tj1 := NewQuote(\"USDJPY\", 123.450)\n\tj2 := NewQuote(\"USDJPY\", 123.470)\n\n\tfmt.Printf(\"EURUSD profit: %.2f pips\\n\", e1.ProfitInPipsAt(e2)) // # => 2.0\n\tfmt.Printf(\"USDJPY profit: %.2f pips\\n\", j1.ProfitInPipsAt(j2)) // # => 2.0\n\n\te1Plus := NewQuote(\"EURUSD\", 1.23450)\n\te1Plus.AddPips(pips.Pip(1.0))\n\tfmt.Printf(\"%.5f + 1.0 pips = %.5f\\n\", e1.ToFloat64(), e1Plus.ToFloat64())\n\n\tj1Plus := NewQuote(\"USDJPY\", 123.450)\n\tj1Plus.AddPips(pips.Pip(1.0))\n\tfmt.Printf(\"%.3f + 1.0 pips = %.3f\\n\", j1.ToFloat64(), j1Plus.ToFloat64())\n\n\te1Minus := NewQuote(\"EURUSD\", 1.23450)\n\te1Minus.SubtractPips(pips.Pip(1.0))\n\tfmt.Printf(\"%.5f - 1.0 pips = %.5f\\n\", e1.ToFloat64(), e1Minus.ToFloat64())\n\n\tj1Minus := NewQuote(\"USDJPY\", 123.450)\n\tj1Minus.SubtractPips(pips.Pip(1.0))\n\tfmt.Printf(\"%.3f - 1.0 pips = %.3f\\n\", j1.ToFloat64(), j1Minus.ToFloat64())\n}", "func (a *Client) PostReturnAddressesPost(params *PostReturnAddressesPostParams, authInfo runtime.ClientAuthInfoWriter) (*PostReturnAddressesPostOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewPostReturnAddressesPostParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"PostReturnAddressesPost\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/post/return-addresses\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"https\"},\n\t\tParams: params,\n\t\tReader: &PostReturnAddressesPostReader{formats: a.formats},\n\t\tAuthInfo: authInfo,\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*PostReturnAddressesPostOK), nil\n\n}", "func (s *OrderShippingAddressService) Count(ctx context.Context, orderID int, params *OrderShippingAddressListParams) (int, *http.Response, error) {\n\tvar cnt count\n\tapiError := new(APIError)\n\n\tresp, err := performRequest(ctx, s.sling.New().Get(fmt.Sprintf(\"%d/shipping_addresses/count\", orderID)).QueryStruct(params), s.httpClient, &cnt, apiError)\n\treturn cnt.Count, resp, relevantError(err, *apiError)\n}", "func (_Contract *ContractCaller) ProposalFee(opts *bind.CallOpts) (*big.Int, error) {\n\tvar out []interface{}\n\terr := _Contract.contract.Call(opts, &out, \"proposalFee\")\n\n\tif err != nil {\n\t\treturn *new(*big.Int), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(*big.Int)).(**big.Int)\n\n\treturn out0, err\n\n}", "func (w *rpcWallet) AddressInfo(ctx context.Context, address string) (*AddressInfo, error) {\n\ta, err := stdaddr.DecodeAddress(address, w.chainParams)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tres, err := w.client().ValidateAddress(ctx, a)\n\tif err != nil {\n\t\treturn nil, translateRPCCancelErr(err)\n\t}\n\tif !res.IsValid {\n\t\treturn nil, fmt.Errorf(\"address is invalid\")\n\t}\n\tif !res.IsMine {\n\t\treturn nil, fmt.Errorf(\"address does not belong to this wallet\")\n\t}\n\tif res.Branch == nil {\n\t\treturn nil, fmt.Errorf(\"no account branch info for address\")\n\t}\n\treturn &AddressInfo{Account: res.Account, Branch: *res.Branch}, nil\n}", "func (o *CheckoutShippingInformationManagementV1SaveAddressInformationPostDefault) Code() int {\n\treturn o._statusCode\n}" ]
[ "0.7514819", "0.64364034", "0.6253201", "0.62505233", "0.60103106", "0.5877299", "0.5747413", "0.5707778", "0.5600922", "0.55864745", "0.5458676", "0.5409771", "0.53614163", "0.5342406", "0.5076149", "0.50638795", "0.50095284", "0.49321228", "0.48394236", "0.48188487", "0.46886334", "0.46340492", "0.46200007", "0.4606888", "0.459103", "0.45492187", "0.45278397", "0.44432402", "0.444064", "0.44333047", "0.4430384", "0.4403816", "0.4391249", "0.43870038", "0.43536365", "0.43473035", "0.43447417", "0.4304283", "0.42906052", "0.42597714", "0.42539066", "0.42345956", "0.42313987", "0.42069036", "0.42065677", "0.4196615", "0.418464", "0.4175517", "0.41705385", "0.41691887", "0.41669074", "0.41559595", "0.41497958", "0.41422457", "0.414022", "0.41314268", "0.41239277", "0.41238812", "0.4116518", "0.4107366", "0.410583", "0.40963763", "0.4086447", "0.40783736", "0.40771767", "0.40735334", "0.40708497", "0.40627596", "0.4059017", "0.40435275", "0.4030263", "0.40278172", "0.40211937", "0.4016413", "0.40095899", "0.40059194", "0.40018386", "0.3998516", "0.39865366", "0.3984351", "0.39792958", "0.39778042", "0.39757916", "0.39675424", "0.39623824", "0.39566433", "0.3952483", "0.39435232", "0.3930078", "0.3929943", "0.39241183", "0.3922672", "0.39150098", "0.39149162", "0.3913378", "0.39128268", "0.39109296", "0.39095217", "0.3904979", "0.39038357" ]
0.8197049
0
/ QuoteShippingMethodManagementV1GetListGet Lists applicable shipping methods for a specified quote.
func (a *Client) QuoteShippingMethodManagementV1GetListGet(params *QuoteShippingMethodManagementV1GetListGetParams) (*QuoteShippingMethodManagementV1GetListGetOK, error) { // TODO: Validate the params before sending if params == nil { params = NewQuoteShippingMethodManagementV1GetListGetParams() } result, err := a.transport.Submit(&runtime.ClientOperation{ ID: "quoteShippingMethodManagementV1GetListGet", Method: "GET", PathPattern: "/V1/carts/{cartId}/shipping-methods", ProducesMediaTypes: []string{""}, ConsumesMediaTypes: []string{""}, Schemes: []string{"http"}, Params: params, Reader: &QuoteShippingMethodManagementV1GetListGetReader{formats: a.formats}, Context: params.Context, Client: params.HTTPClient, }) if err != nil { return nil, err } return result.(*QuoteShippingMethodManagementV1GetListGetOK), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *ShippingMethodsEndpoint) List(ctx context.Context, division int, all bool, o *api.ListOptions) ([]*ShippingMethods, error) {\n\tvar entities []*ShippingMethods\n\tu, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/sales/ShippingMethods\", division) // #nosec\n\tapi.AddListOptionsToURL(u, o)\n\n\tif all {\n\t\terr := s.client.ListRequestAndDoAll(ctx, u.String(), &entities)\n\t\treturn entities, err\n\t}\n\t_, _, err := s.client.NewRequestAndDo(ctx, \"GET\", u.String(), nil, &entities)\n\treturn entities, err\n}", "func (a *Client) QuoteShippingMethodManagementV1GetListGetMine(params *QuoteShippingMethodManagementV1GetListGetMineParams) (*QuoteShippingMethodManagementV1GetListGetMineOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewQuoteShippingMethodManagementV1GetListGetMineParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"quoteShippingMethodManagementV1GetListGetMine\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/V1/carts/mine/shipping-methods\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &QuoteShippingMethodManagementV1GetListGetMineReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*QuoteShippingMethodManagementV1GetListGetMineOK), nil\n\n}", "func GetShippingMethods(w http.ResponseWriter, r *http.Request) []CommerceShippingResponse {\n\n\tvar shippingMethods []CommerceShippingResponse\n\n\t//httpClient := getHttpClient()\n\turl := \"https://api.sphere.io/flexy-commerce/shipping-methods?country=US\"\n\n\treq, err := http.NewRequest(\"GET\", url, bytes.NewBuffer([]byte{}))\n\tif err != nil {\n\t\tlog.Fatal(\"Error reading request. \", err)\n\t}\n\tservices.SetAuthToken(w, r, req)\n\n\tresponse, err := httpClient.Do(req)\n\tif err != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", err)\n\t}\n\n\tif response.StatusCode > 299 {\n\t\tb, _ := ioutil.ReadAll(response.Body)\n\t\tfmt.Println(string(b))\n\t} else {\n\t\tjson.NewDecoder(response.Body).Decode(&shippingMethods)\n\t}\n\n\treturn shippingMethods\n}", "func (s *OrderShippingAddressService) List(ctx context.Context, orderID int, params *OrderShippingAddressListParams) ([]OrderShippingAddress, *http.Response, error) {\n\tvar osa []OrderShippingAddress\n\tapiError := new(APIError)\n\n\tresp, err := performRequest(ctx, s.sling.New().Get(fmt.Sprintf(\"%d/shipping_addresses\", orderID)).QueryStruct(params), s.httpClient, &osa, apiError)\n\treturn osa, resp, relevantError(err, *apiError)\n}", "func (s *SystemService) ListMethods() ([]string, error) {\r\n\tvar reply []string\r\n\tif err := s.rpc.Call(system_list_methods, nil, &reply); err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\treturn reply, nil\r\n}", "func (r *CompanyShipmentMethodsCollectionRequest) Get(ctx context.Context) ([]ShipmentMethod, error) {\n\treturn r.GetN(ctx, 0)\n}", "func (o *OfflineWebCartPaymentGateway) Methods() []domain.Method {\n\treturn []domain.Method{\n\t\t{\n\t\t\tTitle: \"cash on delivery\",\n\t\t\tCode: \"offlinepayment_cashondelivery\",\n\t\t},\n\t\t{\n\t\t\tTitle: \"cash in advance\",\n\t\t\tCode: \"offlinepayment_cashinadvance\",\n\t\t},\n\t}\n}", "func ListMethods(source DescriptorSource, serviceName string) ([]string, error) {\n\tdsc, err := source.FindSymbol(serviceName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif sd, ok := dsc.(*desc.ServiceDescriptor); !ok {\n\t\treturn nil, notFound(\"Service\", serviceName)\n\t} else {\n\t\tmethods := make([]string, 0, len(sd.GetMethods()))\n\t\tfor _, method := range sd.GetMethods() {\n\t\t\tmethods = append(methods, method.GetName())\n\t\t}\n\t\tsort.Strings(methods)\n\t\treturn methods, nil\n\t}\n}", "func (c *FakeApiGatewayMethodSettingses) List(opts v1.ListOptions) (result *v1alpha1.ApiGatewayMethodSettingsList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(apigatewaymethodsettingsesResource, apigatewaymethodsettingsesKind, c.ns, opts), &v1alpha1.ApiGatewayMethodSettingsList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &v1alpha1.ApiGatewayMethodSettingsList{ListMeta: obj.(*v1alpha1.ApiGatewayMethodSettingsList).ListMeta}\n\tfor _, item := range obj.(*v1alpha1.ApiGatewayMethodSettingsList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (m *RetrieveShippingMethodsResponse) Validate() error {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\tfor key, val := range m.GetShippingMethods() {\n\t\t_ = val\n\n\t\t// no validation rules for ShippingMethods[key]\n\n\t\tif v, ok := interface{}(val).(interface{ Validate() error }); ok {\n\t\t\tif err := v.Validate(); err != nil {\n\t\t\t\treturn RetrieveShippingMethodsResponseValidationError{\n\t\t\t\t\tfield: fmt.Sprintf(\"ShippingMethods[%v]\", key),\n\t\t\t\t\treason: \"embedded message failed validation\",\n\t\t\t\t\tcause: err,\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func NewQuotePaymentMethodManagementV1GetListGetMineDefault(code int) *QuotePaymentMethodManagementV1GetListGetMineDefault {\n\treturn &QuotePaymentMethodManagementV1GetListGetMineDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (m *memory) List() ([]Quote, error) {\n\tquotes := make([]Quote, len(m.quotes))\n\tfor _, v := range m.quotes {\n\t\tquotes = append(quotes, v)\n\t}\n\treturn quotes, nil\n}", "func (s *Handler) ListGateways(ctx context.Context, req *api.ListGatewaysRequest) (*api.ListGatewaysResponse, error) {\n\tgws := s.DaemonService.ListGateways()\n\n\tapiGws := make([]*api.Gateway, 0, len(gws))\n\tfor _, gw := range gws {\n\t\tapiGateway, err := s.newAPIGateway(gw)\n\t\tif err != nil {\n\t\t\treturn nil, trace.Wrap(err)\n\t\t}\n\n\t\tapiGws = append(apiGws, apiGateway)\n\t}\n\n\treturn &api.ListGatewaysResponse{\n\t\tGateways: apiGws,\n\t}, nil\n}", "func (r *CompanyPaymentMethodsCollectionRequest) Get(ctx context.Context) ([]PaymentMethod, error) {\n\treturn r.GetN(ctx, 0)\n}", "func ListPaymentMethods(orgID string) error {\n\tclient, err := NewPacketClient()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tpaymentMethods, _, err := client.Organizations.ListPaymentMethods(orgID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = MarshallAndPrint(paymentMethods)\n\treturn err\n}", "func (s *CouponsService) List(ctx context.Context, productID int) ([]*Coupon, *http.Response, error) {\n\tu := \"2.0/product/list_coupons\"\n\n\toptions := &CouponsOptions{\n\t\tProductID: productID,\n\t}\n\n\treq, err := s.client.NewRequest(\"POST\", u, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tcouponsResponse := new(CouponsResponse)\n\tresponse, err := s.client.Do(ctx, req, couponsResponse)\n\tif err != nil {\n\t\treturn nil, response, err\n\t}\n\n\treturn couponsResponse.Response, response, nil\n}", "func (o CorsPolicyOutput) AllowMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v CorsPolicy) []string { return v.AllowMethods }).(pulumi.StringArrayOutput)\n}", "func (s *CheckoutAPIOp) GetShippingRates(token string) ([]ShippingRate, error) {\n\tpath := fmt.Sprintf(\"%s/%s/shipping_rates.json\", checkoutsBasePath, token)\n\tresource := new(ShippingRatesResource)\n\terr := s.client.Get(path, resource, nil)\n\treturn resource.ShippingRates, err\n}", "func (c *ChargeClient) GetList(ctx context.Context, paymentID string) ([]Charge, error) {\n\tvar charges []Charge\n\tif err := c.Caller.Call(ctx, \"GET\", c.chargesPath(paymentID), nil, nil, &charges); err != nil {\n\t\treturn nil, err\n\t}\n\treturn charges, nil\n}", "func (s *ShippingMethodsEndpoint) Get(ctx context.Context, division int, id *types.GUID) (*ShippingMethods, error) {\n\tb, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/sales/ShippingMethods\", division) // #nosec\n\tu, err := api.AddOdataKeyToURL(b, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\te := &ShippingMethods{}\n\t_, _, requestError := s.client.NewRequestAndDo(ctx, \"GET\", u.String(), nil, e)\n\treturn e, requestError\n}", "func (o BucketCorsConfigurationV2CorsRuleOutput) AllowedMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketCorsConfigurationV2CorsRule) []string { return v.AllowedMethods }).(pulumi.StringArrayOutput)\n}", "func NewQuotePaymentMethodManagementV1GetListGetMineOK() *QuotePaymentMethodManagementV1GetListGetMineOK {\n\treturn &QuotePaymentMethodManagementV1GetListGetMineOK{}\n}", "func (s *ServiceDescriptor) GetMethods() []*MethodDescriptor { return s.Methods }", "func (o BucketV2CorsRuleOutput) AllowedMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketV2CorsRule) []string { return v.AllowedMethods }).(pulumi.StringArrayOutput)\n}", "func (in *MethodList) DeepCopy() *MethodList {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(MethodList)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (o CorsPolicyResponseOutput) AllowMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v CorsPolicyResponse) []string { return v.AllowMethods }).(pulumi.StringArrayOutput)\n}", "func (r *bitroute) allowedMethods(path string) []string {\n\tvar allowed []string\n\tfor method, parser := range r.handlers {\n\t\tif _, _, ok := parser.get(path); ok {\n\t\t\tallowed = append(allowed, method)\n\t\t}\n\t}\n\n\treturn allowed\n}", "func RealmGateway_Methods(methods []server.Method, s RealmGateway_Server) []server.Method {\n\tif cap(methods) == 0 {\n\t\tmethods = make([]server.Method, 0, 2)\n\t}\n\n\tmethods = append(methods, server.Method{\n\t\tMethod: capnp.Method{\n\t\t\tInterfaceID: 0x84ff286cd00a3ed4,\n\t\t\tMethodID: 0,\n\t\t\tInterfaceName: \"persistent.capnp:RealmGateway\",\n\t\t\tMethodName: \"import\",\n\t\t},\n\t\tImpl: func(ctx context.Context, call *server.Call) error {\n\t\t\treturn s.Import(ctx, RealmGateway_import{call})\n\t\t},\n\t})\n\n\tmethods = append(methods, server.Method{\n\t\tMethod: capnp.Method{\n\t\t\tInterfaceID: 0x84ff286cd00a3ed4,\n\t\t\tMethodID: 1,\n\t\t\tInterfaceName: \"persistent.capnp:RealmGateway\",\n\t\t\tMethodName: \"export\",\n\t\t},\n\t\tImpl: func(ctx context.Context, call *server.Call) error {\n\t\t\treturn s.Export(ctx, RealmGateway_export{call})\n\t\t},\n\t})\n\n\treturn methods\n}", "func handleQuotesList(w http.ResponseWriter, r *http.Request) {\n\tif r.Method != \"GET\" {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\tio.WriteString(w, \"Bad Request: \"+r.Method+\" not allowed for \"+r.URL.Path)\n\t\treturn\n\t}\n\tio.WriteString(w, \"List: \"+r.URL.Path)\n}", "func (sd *ServiceDescriptor) GetMethods() []*MethodDescriptor {\n\treturn sd.methods\n}", "func (o *CheckoutResponse) GetAllowedPaymentMethods() []string {\n\tif o == nil || IsNil(o.AllowedPaymentMethods) {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\treturn o.AllowedPaymentMethods\n}", "func (c *FakeProxyRoutes) List(ctx context.Context, opts v1.ListOptions) (result *v1alpha1.ProxyRouteList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(proxyroutesResource, proxyroutesKind, c.ns, opts), &v1alpha1.ProxyRouteList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &v1alpha1.ProxyRouteList{ListMeta: obj.(*v1alpha1.ProxyRouteList).ListMeta}\n\tfor _, item := range obj.(*v1alpha1.ProxyRouteList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (o CorsPolicyPtrOutput) AllowMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *CorsPolicy) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.AllowMethods\n\t}).(pulumi.StringArrayOutput)\n}", "func (w *ServerInterfaceWrapper) ListGateways(ctx echo.Context) error {\n\tvar err error\n\n\t// Invoke the callback with all the unmarshalled arguments\n\terr = w.Handler.ListGateways(ctx)\n\treturn err\n}", "func (s *awsCustomerGatewayLister) List(selector labels.Selector) (ret []*v1.AwsCustomerGateway, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1.AwsCustomerGateway))\n\t})\n\treturn ret, err\n}", "func (m *RetrieveShippingMethodsRequest) Validate() error {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\t// no validation rules for AccessToken\n\n\tif v, ok := interface{}(m.GetTime()).(interface{ Validate() error }); ok {\n\t\tif err := v.Validate(); err != nil {\n\t\t\treturn RetrieveShippingMethodsRequestValidationError{\n\t\t\t\tfield: \"Time\",\n\t\t\t\treason: \"embedded message failed validation\",\n\t\t\t\tcause: err,\n\t\t\t}\n\t\t}\n\t}\n\n\tif _, ok := _RetrieveShippingMethodsRequest_Dow_InLookup[m.GetDow()]; !ok {\n\t\treturn RetrieveShippingMethodsRequestValidationError{\n\t\t\tfield: \"Dow\",\n\t\t\treason: \"value must be in list [monday tuesday wednesday thursday friday saturday sunday]\",\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r Notification_User_Subscriber_Billing) GetDeliveryMethods() (resp []datatypes.Notification_Delivery_Method, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Notification_User_Subscriber_Billing\", \"getDeliveryMethods\", nil, &r.Options, &resp)\n\treturn\n}", "func (o BucketCorsRuleOutput) AllowedMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketCorsRule) []string { return v.AllowedMethods }).(pulumi.StringArrayOutput)\n}", "func (directLinkProvider *DirectLinkProviderV2) ListProviderGateways(listProviderGatewaysOptions *ListProviderGatewaysOptions) (result *ProviderGatewayCollection, response *core.DetailedResponse, err error) {\n\treturn directLinkProvider.ListProviderGatewaysWithContext(context.Background(), listProviderGatewaysOptions)\n}", "func (m *CompaniesCompanyItemRequestBuilder) ShipmentMethods()(*CompaniesItemShipmentMethodsRequestBuilder) {\n return NewCompaniesItemShipmentMethodsRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func (s *FastDNSv2Service) ListZones(ctx context.Context, opt *ZoneListOptions) (*ZoneList, *Response, error) {\n\tu := fmt.Sprintf(\"config-dns/v2/zones\")\n\tu, err := addOptions(u, opt)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treq, err := s.client.NewRequest(\"GET\", u, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tvar zones *ZoneList\n\tresp, err := s.client.Do(ctx, req, &zones)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn zones, resp, nil\n}", "func listRegisteredMethods() string {\n\tmethods := make([]string, 0, len(hbaAuthMethods))\n\tfor method := range hbaAuthMethods {\n\t\tmethods = append(methods, method)\n\t}\n\tsort.Strings(methods)\n\treturn strings.Join(methods, \", \")\n}", "func (o BucketCorsItemOutput) Method() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketCorsItem) []string { return v.Method }).(pulumi.StringArrayOutput)\n}", "func (o CorsRuleOutput) AllowedMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v CorsRule) []string { return v.AllowedMethods }).(pulumi.StringArrayOutput)\n}", "func (tr *TaxRateMaster) List() []string {\n\tlist := make([]string, 0)\n\tfor _, v := range tr.taxRate {\n\t\ts := fmt.Sprint(v)\n\t\tlist = append(list, s)\n\t}\n\treturn list\n}", "func (s *sqsQueuePolicyLister) List(selector labels.Selector) (ret []*v1alpha1.SqsQueuePolicy, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.SqsQueuePolicy))\n\t})\n\treturn ret, err\n}", "func NewMethodsList(receiverName, entityNameWithPackage string) Appender {\n\treturn &methodsList{\n\t\treceiverName: receiverName,\n\t\tentityNameWithPackage: entityNameWithPackage,\n\t\tmethods: []method{\n\t\t\tfindAllMethod(receiverName, entityNameWithPackage),\n\t\t\tfindByIdMethod(receiverName, entityNameWithPackage),\n\t\t\tsaveMethod(receiverName, entityNameWithPackage),\n\t\t\tupdateMethod(receiverName, entityNameWithPackage),\n\t\t\tdeleteMethod(receiverName, entityNameWithPackage),\n\t\t\tcountMethod(receiverName, entityNameWithPackage),\n\t\t},\n\t}\n}", "func (m *ShippingMethod) Validate() error {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\t// no validation rules for Amount\n\n\t// no validation rules for DurationInMinutes\n\n\t// no validation rules for UserId\n\n\t// no validation rules for Title\n\n\treturn nil\n}", "func List(params *stripe.CouponListParams) *Iter {\n\treturn getC().List(params)\n}", "func (b *CompanyRequestBuilder) ShipmentMethods() *CompanyShipmentMethodsCollectionRequestBuilder {\n\tbb := &CompanyShipmentMethodsCollectionRequestBuilder{BaseRequestBuilder: b.BaseRequestBuilder}\n\tbb.baseURL += \"/shipmentMethods\"\n\treturn bb\n}", "func (o CorsPolicyResponsePtrOutput) AllowMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *CorsPolicyResponse) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.AllowMethods\n\t}).(pulumi.StringArrayOutput)\n}", "func (c *FwRouter) GetList() ([]string, error) {\n\tresult, _ := c.versioning()\n\treturn c.ns.Listing(util.Get, c.xpath(nil), result)\n}", "func (cano *Canopus) GetAvailableMethod(amount float64) ([]PaymentMethod, error) {\n\turl := fmt.Sprintf(\"%v/api/v1/merchants/%v/method\", BaseURL, cano.MerchantID)\n\tbody := make(map[string]interface{})\n\tbody[\"MerchantID\"] = cano.MerchantID\n\tbody[\"amount\"] = amount\n\tbodyJson, _ := json.Marshal(body)\n\tsignature, err := cano.GenerateSignature(bodyJson)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", url, bytes.NewBuffer(bodyJson))\n\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"X-Signature\", signature)\n\n\tresp, err := cano.Client.Do(req)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\tresponse, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\tdefer resp.Body.Close()\n\n\trespResult, err := ValidateResponse(response)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\t_, err = VerifySignature(cano.MerchantPem, response, respResult.Signature)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\tvar result []PaymentMethod\n\tfor _, payment := range respResult.Response.Data[\"method\"].([]interface{}) {\n\t\tvar tmp PaymentMethod\n\t\tpaymentJson, err := json.Marshal(payment)\n\t\tif err != nil {\n\t\t\treturn []PaymentMethod{}, err\n\t\t}\n\t\terr = json.Unmarshal(paymentJson, &tmp)\n\t\tif err != nil {\n\t\t\treturn []PaymentMethod{}, err\n\t\t}\n\t\tresult = append(result, tmp)\n\t}\n\n\treturn result, nil\n}", "func (cs *checkoutService) quoteShipping(ctx context.Context, address *pb.Address, items []*pb.CartItem) (*pb.Money, error) {\n\tshippingQuote, err := pb.NewShippingServiceClient(cs.shippingSvcConn).\n\t\tGetQuote(ctx, &pb.GetQuoteRequest{\n\t\t\tAddress: address,\n\t\t\tItems: items})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get shipping quote: %+v\", err)\n\t}\n\treturn shippingQuote.GetCostUsd(), nil\n}", "func (s *SalesPriceListDetailsEndpoint) List(ctx context.Context, division int, all bool, o *api.ListOptions) ([]*SalesPriceListDetails, error) {\n\tvar entities []*SalesPriceListDetails\n\tu, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/sales/SalesPriceListDetails\", division) // #nosec\n\tapi.AddListOptionsToURL(u, o)\n\n\tif all {\n\t\terr := s.client.ListRequestAndDoAll(ctx, u.String(), &entities)\n\t\treturn entities, err\n\t}\n\t_, _, err := s.client.NewRequestAndDo(ctx, \"GET\", u.String(), nil, &entities)\n\treturn entities, err\n}", "func (o CorsRuleResponseOutput) AllowedMethods() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v CorsRuleResponse) []string { return v.AllowedMethods }).(pulumi.StringArrayOutput)\n}", "func (n VpcNatGateways) List(region account.Region, account account.Account, force bool) ([]cloud.Resource, error) {\n\tclient, err := vpc.NewClientWithAccessKey(string(region), account.AccessKeyID, account.AccessKeySecret)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trequest := vpc.CreateDescribeNatGatewaysRequest()\n\trequest.PageSize = \"50\"\n\tresponse, err := client.DescribeNatGateways(request)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tnatGateways := make([]cloud.Resource, 0)\n\tfor _, natGatewayItem := range response.NatGateways.NatGateway {\n\t\tsnatTables, err := fetchSnatTables(client, natGatewayItem.SnatTableIds)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tnatGateways = append(natGateways, VpcNatGateway{NatGateway: natGatewayItem, SnatTables: snatTables})\n\t}\n\n\treturn natGateways, nil\n}", "func (s *PlanServiceOp) List() ([]Plan, *Response, error) {\n\troot := new(planRoot)\n\n\tresp, err := s.client.DoRequest(\"GET\", planBasePath, nil, root)\n\tif err != nil {\n\t\treturn nil, resp, err\n\t}\n\n\treturn root.Plans, resp, err\n}", "func (c *Firewall) GetList() ([]string, error) {\n\tans := c.container()\n\treturn c.ns.Listing(util.Get, c.pather(), ans)\n}", "func (s *HistoricalRatesService) List(date time.Time) (*RateResponse, error) {\n\t// If we have cached results, use them.\n\tif results, ok := s.client.Cache.Get(s.baseCurrency, date); ok {\n\t\treturn results, nil\n\t}\n\n\t// No cached results, go and fetch them.\n\tif err := s.fetch(date); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn s.List(date)\n}", "func (r *OffersService) List() *OffersListCall {\n\tc := &OffersListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\treturn c\n}", "func (api *powerdnsProvider) ListZones() ([]string, error) {\n\tvar result []string\n\tmyZones, err := api.client.Zones().ListZones(context.Background(), api.ServerName)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tfor _, zone := range myZones {\n\t\tresult = append(result, zone.Name)\n\t}\n\treturn result, nil\n}", "func (service *Service) GetMethods() []*Method {\n\tresult := make([]*Method, len(service.Methods))\n\n\tindex := 0\n\tfor key, method := range service.Methods {\n\t\tresult[index] = NewMethod(key, method)\n\t\tindex++\n\t}\n\n\treturn result\n}", "func (s *apiGatewayModelLister) List(selector labels.Selector) (ret []*v1alpha1.ApiGatewayModel, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.ApiGatewayModel))\n\t})\n\treturn ret, err\n}", "func (s *PriorityService) GetList(ctx context.Context) ([]Priority, *Response, error) {\n\tapiEndpoint := \"rest/api/2/priority\"\n\treq, err := s.client.NewRequest(ctx, http.MethodGet, apiEndpoint, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tpriorityList := []Priority{}\n\tresp, err := s.client.Do(req, &priorityList)\n\tif err != nil {\n\t\treturn nil, resp, NewJiraError(resp, err)\n\t}\n\treturn priorityList, resp, nil\n}", "func List(ctx context.Context, params *razorpay.OrderListParams) (*razorpay.OrderList, error) {\n\treturn getDefaultClient().List(ctx, params)\n}", "func NewQuotePaymentMethodManagementV1GetListGetMineUnauthorized() *QuotePaymentMethodManagementV1GetListGetMineUnauthorized {\n\treturn &QuotePaymentMethodManagementV1GetListGetMineUnauthorized{}\n}", "func List(c *drycc.Client, appID string, results int) (api.Gateways, int, error) {\n\tu := fmt.Sprintf(\"/v2/apps/%s/gateways/\", appID)\n\tbody, count, reqErr := c.LimitedRequest(u, results)\n\n\tif reqErr != nil && !drycc.IsErrAPIMismatch(reqErr) {\n\t\treturn []api.Gateway{}, -1, reqErr\n\t}\n\n\tvar gateways []api.Gateway\n\tif err := json.Unmarshal([]byte(body), &gateways); err != nil {\n\t\treturn []api.Gateway{}, -1, err\n\t}\n\n\treturn gateways, count, reqErr\n}", "func (*DirectLinkProviderV2) NewListProviderGatewaysOptions() *ListProviderGatewaysOptions {\n\treturn &ListProviderGatewaysOptions{}\n}", "func GetAddressList() []common.Address {\n\treturn allSkrAddress\n}", "func List(params *CouponListParams) *CouponIter {\n\treturn getC().List(params)\n}", "func (s *proxyRouteLister) List(selector labels.Selector) (ret []*v1alpha1.ProxyRoute, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1alpha1.ProxyRoute))\n\t})\n\treturn ret, err\n}", "func Flistx() []string {\n\tmethods := []string{}\n\tfor k, _ := range fmethods {\n\t\tmethods = append(methods, k)\n\t}\n\treturn methods\n}", "func (m OrderRepository) List(context.Context) ([]*model.Order, error) {\n\treturn m.ListResponse, m.Err\n}", "func (c *Client) List(ctx context.Context, params *razorpay.OrderListParams) (*razorpay.OrderList, error) {\n\tif params == nil {\n\t\tparams = &razorpay.OrderListParams{}\n\t}\n\n\torderList := &razorpay.OrderList{}\n\terr := c.Call(ctx, http.MethodGet, \"/orders\", params, orderList)\n\treturn orderList, err\n}", "func (o *ConditionRequestRateCondition) GetHttpMethods() []WafHttpMethod {\n\tif o == nil || o.HttpMethods == nil {\n\t\tvar ret []WafHttpMethod\n\t\treturn ret\n\t}\n\treturn *o.HttpMethods\n}", "func (v *VaultAccessor) List() (PathList, error) {\n\treturn v.list(v.basePath)\n}", "func GetSupportedHTTPMethodTypes() []HTTPMethodType {\n\tvar result []HTTPMethodType\n\tfor index := HTTPMethodType(1); index < limit; index++ {\n\t\tresult = append(result, index)\n\t}\n\treturn result\n}", "func List(paths ...PathRetriever) PathsRetriever {\n\tvar list []string\n\tvar err error\n\tfor _, path := range paths {\n\t\tvar resolved string\n\t\tresolved, err = path()\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tlist = append(list, resolved)\n\t}\n\n\treturn func() ([]string, error) {\n\t\treturn list, err\n\t}\n}", "func (p *listDiscoveryPlugin) Method() string {\n\treturn \"list\"\n}", "func (client StorageGatewayClient) ListStorageGateways(ctx context.Context, request ListStorageGatewaysRequest) (response ListStorageGatewaysResponse, err error) {\n\tvar ociResponse common.OCIResponse\n\tpolicy := common.NoRetryPolicy()\n\tif client.RetryPolicy() != nil {\n\t\tpolicy = *client.RetryPolicy()\n\t}\n\tif request.RetryPolicy() != nil {\n\t\tpolicy = *request.RetryPolicy()\n\t}\n\tociResponse, err = common.Retry(ctx, request, client.listStorageGateways, policy)\n\tif err != nil {\n\t\tif ociResponse != nil {\n\t\t\tif httpResponse := ociResponse.HTTPResponse(); httpResponse != nil {\n\t\t\t\topcRequestId := httpResponse.Header.Get(\"opc-request-id\")\n\t\t\t\tresponse = ListStorageGatewaysResponse{RawResponse: httpResponse, OpcRequestId: &opcRequestId}\n\t\t\t} else {\n\t\t\t\tresponse = ListStorageGatewaysResponse{}\n\t\t\t}\n\t\t}\n\t\treturn\n\t}\n\tif convertedResponse, ok := ociResponse.(ListStorageGatewaysResponse); ok {\n\t\tresponse = convertedResponse\n\t} else {\n\t\terr = fmt.Errorf(\"failed to convert OCIResponse into ListStorageGatewaysResponse\")\n\t}\n\treturn\n}", "func (o BucketCorsItemResponseOutput) Method() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketCorsItemResponse) []string { return v.Method }).(pulumi.StringArrayOutput)\n}", "func (r *SitemapsService) List(siteUrl string) *SitemapsListCall {\n\tc := &SitemapsListCall{s: r.s, opt_: make(map[string]interface{})}\n\tc.siteUrl = siteUrl\n\treturn c\n}", "func (o *RuleMatch) GetMethods() []string {\n\tif o == nil || o.Methods == nil {\n\t\tvar ret []string\n\t\treturn ret\n\t}\n\treturn o.Methods\n}", "func (r *Replenish) List(args *Replenish) (*ReplenishList, error) {\n\taction := \"sales.replenish.list.get\"\n\tdata, err := json.Marshal(args)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbody, err := mafengwo.NewDeals().Fetch(action, data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresult := ReplenishList{}\n\terr = json.Unmarshal(body, &result)\n\treturn &result, err\n}", "func List(repo name.Repository, options ...Option) ([]string, error) {\n\to, err := makeOptions(options...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn newPuller(o).List(o.context, repo)\n}", "func (l *MethodList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (e *OptionListService) List(opts *ListOptions) ([]OptionList, *Response, error) {\n\tendpoint := \"/assets/optionLists\"\n\toptionLists := new([]OptionList)\n\tresp, err := e.client.getRequestListDecode(endpoint, optionLists, opts)\n\treturn *optionLists, resp, err\n}", "func (s *SalesItemPricesEndpoint) List(ctx context.Context, division int, all bool, o *api.ListOptions) ([]*SalesItemPrices, error) {\n\tvar entities []*SalesItemPrices\n\tu, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/logistics/SalesItemPrices\", division) // #nosec\n\tapi.AddListOptionsToURL(u, o)\n\n\tif all {\n\t\terr := s.client.ListRequestAndDoAll(ctx, u.String(), &entities)\n\t\treturn entities, err\n\t}\n\t_, _, err := s.client.NewRequestAndDo(ctx, \"GET\", u.String(), nil, &entities)\n\treturn entities, err\n}", "func (list AddressList) List() []string {\r\n\tvar ss []string\r\n\tfor _, addr := range list {\r\n\t\tss = append(ss, addr.Address)\r\n\t}\r\n\treturn ss\r\n}", "func (b *ListBuilder) List() *CSPList {\n\treturn b.CspList\n}", "func (ShippingOption) Values() []ShippingOption {\n\treturn []ShippingOption{\n\t\t\"SECOND_DAY\",\n\t\t\"NEXT_DAY\",\n\t\t\"EXPRESS\",\n\t\t\"STANDARD\",\n\t}\n}", "func (BillingMethod) Values() []BillingMethod {\n\treturn []BillingMethod{\n\t\t\"METERED\",\n\t\t\"UNMETERED\",\n\t}\n}", "func (e Endpoints) List(ctx context.Context) (products []io.Product, err error) {\n\trequest := ListRequest{}\n\tresponse, err := e.ListEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(ListResponse).Products, response.(ListResponse).Err\n}", "func (r Plugin) ByMethod(method string) []RouteInfo {\n\tvar routesByMethod []RouteInfo\n\trlen := len(r.routes)\n\tmethod = strings.ToUpper(method)\n\tfor i := 0; i < rlen; i++ {\n\t\tif r.routes[i].Method == method {\n\t\t\troutesByMethod = append(routesByMethod, r.routes[i])\n\t\t}\n\t}\n\treturn routesByMethod\n}", "func (sp *SymbolsProvider) Get() (symbols []schemas.Symbol, err error) {\n\tsmbls, err := sp.getSymbols()\n\tif err != nil {\n\t\treturn\n\t}\n\tcoins, err := sp.getCoins()\n\tif err != nil {\n\t\treturn\n\t}\n\n\tfor _, smb := range smbls {\n\t\tvar basePrec, quotePrec int\n\n\t\ts := smb.Map()\n\t\tif p, ok := coins[smb.CoinType]; ok {\n\t\t\tbasePrec = int(p.TradePrecision)\n\t\t}\n\t\tif p, ok := coins[smb.CoinTypePair]; ok {\n\t\t\tquotePrec = int(p.TradePrecision)\n\t\t}\n\n\t\tif basePrec != 0 && quotePrec != 0 {\n\t\t\tif basePrec > quotePrec {\n\t\t\t\ts.PricePrecision = quotePrec\n\t\t\t} else if quotePrec > basePrec {\n\t\t\t\ts.PricePrecision = basePrec\n\t\t\t} else {\n\t\t\t\ts.PricePrecision = basePrec\n\t\t\t}\n\t\t} else {\n\t\t\ts.PricePrecision = defaultPrecision\n\t\t}\n\n\t\tsymbols = append(symbols, s)\n\t}\n\n\treturn\n}", "func (b *Base) GetMechanismList(req *GetMechanismListReq) (*GetMechanismListResp, error) {\n\treturn nil, ErrFunctionNotSupported\n}", "func (c *FakeQuobyteServices) List(opts v1.ListOptions) (result *quobyte_com_v1.QuobyteServiceList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(quobyteservicesResource, quobyteservicesKind, c.ns, opts), &quobyte_com_v1.QuobyteServiceList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &quobyte_com_v1.QuobyteServiceList{ListMeta: obj.(*quobyte_com_v1.QuobyteServiceList).ListMeta}\n\tfor _, item := range obj.(*quobyte_com_v1.QuobyteServiceList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func (r Notification_User_Subscriber_Mobile) GetDeliveryMethods() (resp []datatypes.Notification_Delivery_Method, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Notification_User_Subscriber_Mobile\", \"getDeliveryMethods\", nil, &r.Options, &resp)\n\treturn\n}", "func (c *Client) GetList(dir string) ([]string, error) {\n\tvar list []string\n\n\t// Prepare the URL\n\tURL := fmt.Sprintf(wpListURL, dir)\n\n\t// Make the Request\n\tresp, err := c.getRequest(URL)\n\tif err != nil {\n\t\treturn list, err\n\t}\n\n\t// Drain body and check Close error\n\tdefer drainAndClose(resp.Body, &err)\n\tbytes, err := ioutil.ReadAll(resp.Body)\n\tmatches := regexList.FindAllStringSubmatch(string(bytes), -1)\n\n\t// Add all matches to extension list\n\tfor _, match := range matches {\n\t\tlist = append(list, match[1])\n\t}\n\n\treturn list, err\n}" ]
[ "0.6360206", "0.5676647", "0.5576999", "0.53600466", "0.5120917", "0.50071275", "0.5005693", "0.49260736", "0.49173674", "0.48397222", "0.4784371", "0.46715364", "0.46630764", "0.4655576", "0.46332535", "0.4600606", "0.4593155", "0.45691648", "0.45391914", "0.45183414", "0.44881973", "0.447122", "0.44656286", "0.44384417", "0.44214466", "0.44160658", "0.4392001", "0.43905482", "0.43878505", "0.43863043", "0.43845603", "0.4363968", "0.43156108", "0.43092632", "0.4300388", "0.42929634", "0.42919812", "0.42804167", "0.42738047", "0.42699462", "0.42668146", "0.42664346", "0.4259232", "0.42388484", "0.42320117", "0.4223973", "0.42153233", "0.42124507", "0.4208985", "0.4207857", "0.42033356", "0.42010576", "0.41966516", "0.41788745", "0.41776222", "0.41698053", "0.415686", "0.41534987", "0.41512764", "0.41503233", "0.41460592", "0.41432714", "0.41370234", "0.41362125", "0.41331065", "0.41229787", "0.41224936", "0.41189536", "0.4111637", "0.4110704", "0.4101767", "0.41011804", "0.4101119", "0.409771", "0.40889958", "0.40834072", "0.40802994", "0.40621194", "0.40577295", "0.40480366", "0.40468603", "0.40396678", "0.40377617", "0.40369537", "0.40222573", "0.40159595", "0.40121374", "0.40099448", "0.4008488", "0.40016544", "0.39926666", "0.39897242", "0.39880398", "0.39819956", "0.3981264", "0.3980982", "0.39762834", "0.39681187", "0.39667645", "0.3962617" ]
0.6407668
0
/ QuoteShippingMethodManagementV1GetListGetMine Lists applicable shipping methods for a specified quote.
func (a *Client) QuoteShippingMethodManagementV1GetListGetMine(params *QuoteShippingMethodManagementV1GetListGetMineParams) (*QuoteShippingMethodManagementV1GetListGetMineOK, error) { // TODO: Validate the params before sending if params == nil { params = NewQuoteShippingMethodManagementV1GetListGetMineParams() } result, err := a.transport.Submit(&runtime.ClientOperation{ ID: "quoteShippingMethodManagementV1GetListGetMine", Method: "GET", PathPattern: "/V1/carts/mine/shipping-methods", ProducesMediaTypes: []string{""}, ConsumesMediaTypes: []string{""}, Schemes: []string{"http"}, Params: params, Reader: &QuoteShippingMethodManagementV1GetListGetMineReader{formats: a.formats}, Context: params.Context, Client: params.HTTPClient, }) if err != nil { return nil, err } return result.(*QuoteShippingMethodManagementV1GetListGetMineOK), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewQuotePaymentMethodManagementV1GetListGetMineDefault(code int) *QuotePaymentMethodManagementV1GetListGetMineDefault {\n\treturn &QuotePaymentMethodManagementV1GetListGetMineDefault{\n\t\t_statusCode: code,\n\t}\n}", "func NewQuotePaymentMethodManagementV1GetListGetMineOK() *QuotePaymentMethodManagementV1GetListGetMineOK {\n\treturn &QuotePaymentMethodManagementV1GetListGetMineOK{}\n}", "func (c *Client) ListMine(ctx context.Context, p *ListMinePayload) (res *StationsFull, err error) {\n\tvar ires interface{}\n\tires, err = c.ListMineEndpoint(ctx, p)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn ires.(*StationsFull), nil\n}", "func NewQuotePaymentMethodManagementV1GetListGetMineBadRequest() *QuotePaymentMethodManagementV1GetListGetMineBadRequest {\n\treturn &QuotePaymentMethodManagementV1GetListGetMineBadRequest{}\n}", "func NewQuotePaymentMethodManagementV1GetListGetMineUnauthorized() *QuotePaymentMethodManagementV1GetListGetMineUnauthorized {\n\treturn &QuotePaymentMethodManagementV1GetListGetMineUnauthorized{}\n}", "func (a *Client) QuoteShippingMethodManagementV1GetListGet(params *QuoteShippingMethodManagementV1GetListGetParams) (*QuoteShippingMethodManagementV1GetListGetOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewQuoteShippingMethodManagementV1GetListGetParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"quoteShippingMethodManagementV1GetListGet\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/V1/carts/{cartId}/shipping-methods\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &QuoteShippingMethodManagementV1GetListGetReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*QuoteShippingMethodManagementV1GetListGetOK), nil\n\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WithContext(ctx context.Context) *QuotePaymentMethodManagementV1SetPutMineParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *QuotePaymentMethodManagementV1GetListGetMineDefault) Code() int {\n\treturn o._statusCode\n}", "func NewListMineHandler(\n\tendpoint goa.Endpoint,\n\tmux goahttp.Muxer,\n\tdecoder func(*http.Request) goahttp.Decoder,\n\tencoder func(context.Context, http.ResponseWriter) goahttp.Encoder,\n\terrhandler func(context.Context, http.ResponseWriter, error),\n\tformatter func(err error) goahttp.Statuser,\n) http.Handler {\n\tvar (\n\t\tdecodeRequest = DecodeListMineRequest(mux, decoder)\n\t\tencodeResponse = EncodeListMineResponse(encoder)\n\t\tencodeError = EncodeListMineError(encoder, formatter)\n\t)\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tctx := context.WithValue(r.Context(), goahttp.AcceptTypeKey, r.Header.Get(\"Accept\"))\n\t\tctx = context.WithValue(ctx, goa.MethodKey, \"list mine\")\n\t\tctx = context.WithValue(ctx, goa.ServiceKey, \"station\")\n\t\tpayload, err := decodeRequest(r)\n\t\tif err != nil {\n\t\t\tif err := encodeError(ctx, w, err); err != nil {\n\t\t\t\terrhandler(ctx, w, err)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tres, err := endpoint(ctx, payload)\n\t\tif err != nil {\n\t\t\tif err := encodeError(ctx, w, err); err != nil {\n\t\t\t\terrhandler(ctx, w, err)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tif err := encodeResponse(ctx, w, res); err != nil {\n\t\t\terrhandler(ctx, w, err)\n\t\t}\n\t})\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WithHTTPClient(client *http.Client) *QuotePaymentMethodManagementV1SetPutMineParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (s *ShippingMethodsEndpoint) List(ctx context.Context, division int, all bool, o *api.ListOptions) ([]*ShippingMethods, error) {\n\tvar entities []*ShippingMethods\n\tu, _ := s.client.ResolvePathWithDivision(\"/api/v1/{division}/sales/ShippingMethods\", division) // #nosec\n\tapi.AddListOptionsToURL(u, o)\n\n\tif all {\n\t\terr := s.client.ListRequestAndDoAll(ctx, u.String(), &entities)\n\t\treturn entities, err\n\t}\n\t_, _, err := s.client.NewRequestAndDo(ctx, \"GET\", u.String(), nil, &entities)\n\treturn entities, err\n}", "func (a *Client) QuoteShippingMethodManagementV1EstimateByAddressIDPostMine(params *QuoteShippingMethodManagementV1EstimateByAddressIDPostMineParams) (*QuoteShippingMethodManagementV1EstimateByAddressIDPostMineOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewQuoteShippingMethodManagementV1EstimateByAddressIDPostMineParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"quoteShippingMethodManagementV1EstimateByAddressIdPostMine\",\n\t\tMethod: \"POST\",\n\t\tPathPattern: \"/V1/carts/mine/estimate-shipping-methods-by-address-id\",\n\t\tProducesMediaTypes: []string{\"\"},\n\t\tConsumesMediaTypes: []string{\"\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &QuoteShippingMethodManagementV1EstimateByAddressIDPostMineReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result.(*QuoteShippingMethodManagementV1EstimateByAddressIDPostMineOK), nil\n\n}", "func (d *POW) Mine(tracer *tracing.Tracer, nonce []uint8, numTrailingZeros uint) error {\n\ttracer.RecordAction(PowlibMiningBegin{Nonce: nonce, NumTrailingZeros: numTrailingZeros})\n\n\tgo func(nonce []uint8, numTrailingZeros uint) {\n\t\tvar secret []uint8\n\t\targs := PowlibMine{Nonce: nonce, NumTrailingZeros: numTrailingZeros}\n\t\ttracer.RecordAction(args)\n\t\td.client.Call(\"Coordinator.Mine\", args, &secret)\n\t\t//if err != nil {\n\t\t//\treturn err\n\t\t//}\n\t\ttracer.RecordAction(PowlibSuccess{\n\t\t\tNonce: nonce,\n\t\t\tNumTrailingZeros: numTrailingZeros,\n\t\t\tSecret: secret,\n\t\t})\n\t\ttracer.RecordAction(PowlibMiningComplete{\n\t\t\tNonce: nonce,\n\t\t\tNumTrailingZeros: numTrailingZeros,\n\t\t\tSecret: secret,\n\t\t})\n\t\td.nc <- MineResult{\n\t\t\tNonce: nonce,\n\t\t\tNumTrailingZeros: numTrailingZeros,\n\t\t\tSecret: secret,\n\t\t}\n\t}(nonce, numTrailingZeros)\n\treturn nil\n}", "func GetShippingMethods(w http.ResponseWriter, r *http.Request) []CommerceShippingResponse {\n\n\tvar shippingMethods []CommerceShippingResponse\n\n\t//httpClient := getHttpClient()\n\turl := \"https://api.sphere.io/flexy-commerce/shipping-methods?country=US\"\n\n\treq, err := http.NewRequest(\"GET\", url, bytes.NewBuffer([]byte{}))\n\tif err != nil {\n\t\tlog.Fatal(\"Error reading request. \", err)\n\t}\n\tservices.SetAuthToken(w, r, req)\n\n\tresponse, err := httpClient.Do(req)\n\tif err != nil {\n\t\tfmt.Printf(\"The HTTP request failed with error %s\\n\", err)\n\t}\n\n\tif response.StatusCode > 299 {\n\t\tb, _ := ioutil.ReadAll(response.Body)\n\t\tfmt.Println(string(b))\n\t} else {\n\t\tjson.NewDecoder(response.Body).Decode(&shippingMethods)\n\t}\n\n\treturn shippingMethods\n}", "func (r *CompanyShipmentMethodsCollectionRequest) Get(ctx context.Context) ([]ShipmentMethod, error) {\n\treturn r.GetN(ctx, 0)\n}", "func (m *Miner) Mine() {\n\tm.miningWorkChannel = make(chan *miningWork, len(m.ClDevices))\n\tgo m.createWork()\n\tfor minerID, device := range m.ClDevices {\n\t\tsdm := &singleDeviceMiner{\n\t\t\tClDevice: device,\n\t\t\tMinerID: minerID,\n\t\t\tHashRateReports: m.HashRateReports,\n\t\t\tminingWorkChannel: m.miningWorkChannel,\n\t\t\tGlobalItemSize: m.GlobalItemSize,\n\t\t\tClient: m.Client,\n\t\t}\n\n\t\tgo sdm.mine()\n\t}\n}", "func (client WorkloadNetworksClient) ListGatewaysSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func (o *ViewReactionsForObject) SetMine(v string) {\n\to.Mine = &v\n}", "func (l *MethodList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (o *OfflineWebCartPaymentGateway) Methods() []domain.Method {\n\treturn []domain.Method{\n\t\t{\n\t\t\tTitle: \"cash on delivery\",\n\t\t\tCode: \"offlinepayment_cashondelivery\",\n\t\t},\n\t\t{\n\t\t\tTitle: \"cash in advance\",\n\t\t\tCode: \"offlinepayment_cashinadvance\",\n\t\t},\n\t}\n}", "func (m *CompaniesCompanyItemRequestBuilder) ShipmentMethods()(*CompaniesItemShipmentMethodsRequestBuilder) {\n return NewCompaniesItemShipmentMethodsRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WithTimeout(timeout time.Duration) *QuotePaymentMethodManagementV1SetPutMineParams {\n\to.SetTimeout(timeout)\n\treturn o\n}", "func (o *QuoteCartManagementV1PlaceOrderPutMineBody) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := o.validatePaymentMethod(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (o *QuotePaymentMethodManagementV1GetListGetMineReader) ReadResponse(response runtime.ClientResponse, consumer runtime.Consumer) (interface{}, error) {\n\tswitch response.Code() {\n\n\tcase 200:\n\t\tresult := NewQuotePaymentMethodManagementV1GetListGetMineOK()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn result, nil\n\n\tcase 400:\n\t\tresult := NewQuotePaymentMethodManagementV1GetListGetMineBadRequest()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tcase 401:\n\t\tresult := NewQuotePaymentMethodManagementV1GetListGetMineUnauthorized()\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn nil, result\n\n\tdefault:\n\t\tresult := NewQuotePaymentMethodManagementV1GetListGetMineDefault(response.Code())\n\t\tif err := result.readResponse(response, consumer, o.formats); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif response.Code()/100 == 2 {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, result\n\t}\n}", "func (o *ViewReactionsForObject) GetMine() string {\n\tif o == nil || o.Mine == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Mine\n}", "func (o *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams) WithContext(ctx context.Context) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func (o *GiftCardAccountGiftCardAccountManagementV1DeleteByQuoteIDDeleteMineParams) WithContext(ctx context.Context) *GiftCardAccountGiftCardAccountManagementV1DeleteByQuoteIDDeleteMineParams {\n\to.SetContext(ctx)\n\treturn o\n}", "func NewQuotePaymentMethodManagementV1SetPutMineParamsWithContext(ctx context.Context) *QuotePaymentMethodManagementV1SetPutMineParams {\n\tvar ()\n\treturn &QuotePaymentMethodManagementV1SetPutMineParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (cano *Canopus) GetAvailableMethod(amount float64) ([]PaymentMethod, error) {\n\turl := fmt.Sprintf(\"%v/api/v1/merchants/%v/method\", BaseURL, cano.MerchantID)\n\tbody := make(map[string]interface{})\n\tbody[\"MerchantID\"] = cano.MerchantID\n\tbody[\"amount\"] = amount\n\tbodyJson, _ := json.Marshal(body)\n\tsignature, err := cano.GenerateSignature(bodyJson)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\treq, err := http.NewRequest(\"GET\", url, bytes.NewBuffer(bodyJson))\n\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"X-Signature\", signature)\n\n\tresp, err := cano.Client.Do(req)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\tresponse, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\tdefer resp.Body.Close()\n\n\trespResult, err := ValidateResponse(response)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\t_, err = VerifySignature(cano.MerchantPem, response, respResult.Signature)\n\tif err != nil {\n\t\treturn []PaymentMethod{}, err\n\t}\n\n\tvar result []PaymentMethod\n\tfor _, payment := range respResult.Response.Data[\"method\"].([]interface{}) {\n\t\tvar tmp PaymentMethod\n\t\tpaymentJson, err := json.Marshal(payment)\n\t\tif err != nil {\n\t\t\treturn []PaymentMethod{}, err\n\t\t}\n\t\terr = json.Unmarshal(paymentJson, &tmp)\n\t\tif err != nil {\n\t\t\treturn []PaymentMethod{}, err\n\t\t}\n\t\tresult = append(result, tmp)\n\t}\n\n\treturn result, nil\n}", "func (_Crowdsale *CrowdsaleCaller) BonusOffList(opts *bind.CallOpts) (*big.Int, error) {\n\tvar (\n\t\tret0 = new(*big.Int)\n\t)\n\tout := ret0\n\terr := _Crowdsale.contract.Call(opts, out, \"bonusOffList\")\n\treturn *ret0, err\n}", "func (p *BoxPeer) Miners() []*types.PeerInfo {\n\tminers, _ := p.minerreader.Miners()\n\tre, _ := regexp.Compile(ipRegex)\n\n\tinfos := []*types.PeerInfo{}\n\tfor _, m := range miners {\n\t\tpid, err := peer.IDB58Decode(m)\n\t\tif err != nil {\n\t\t\tlogger.Errorf(\"IDFromString failed, Err: %v, m: %s\", err, m)\n\t\t\tcontinue\n\t\t}\n\t\tinfo := p.table.peerStore.PeerInfo(pid)\n\t\tiplist := []string{}\n\t\tfor _, ip := range info.Addrs {\n\t\t\tipp := re.Find([]byte(ip.String()))\n\t\t\tif len(ipp) == 0 {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tiplist = append(iplist, string(ipp))\n\t\t}\n\n\t\tpinfo := &types.PeerInfo{\n\t\t\tID: m,\n\t\t\tIplist: iplist,\n\t\t}\n\t\tinfos = append(infos, pinfo)\n\t}\n\treturn infos\n}", "func (self *BlockChain) Mine() {\n\t//Get the last transactions proof\n\tprintln(\"Starting to mine...\")\n\tlast := self.LastBlock()\n\tlastProof := last.Proof\n\n\t//Work out the proof\n\tfmt.Printf(\"Last proof = %d \\n\", lastProof)\n\tnewProof := self.ProofOfWork(lastProof)\n\tself.NewTransaction(\"0\", \"dest\", 1)\n\n\t//Add to blockchain with Proof + HASH\n\t//TODO get hash of previous\n\tfmt.Printf(\"new proof = %d \\n\", newProof)\n\tself.NewBlock(newProof, \"xxx\")\n\n}", "func (gw2 *GW2Api) Minis() (res []int, err error) {\n\tver := \"v2\"\n\ttag := \"minis\"\n\terr = gw2.fetchEndpoint(ver, tag, nil, &res)\n\treturn\n}", "func (m *MarketHoursMutation) MarketInfoIDs() (ids []int) {\n\tif id := m.market_info; id != nil {\n\t\tids = append(ids, *id)\n\t}\n\treturn\n}", "func (serv *ExchangeServer) GetSupportCoins() []string {\n\tsymbols := make([]string, len(serv.coins))\n\ti := 0\n\tfor _, coin := range serv.coins {\n\t\tsymbols[i] = coin.Symbol()\n\t\ti++\n\t}\n\treturn symbols\n}", "func (p PricedItems) ShippingItems() map[string]domain.Price {\n\treturn p.shippingItems\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) SetContext(ctx context.Context) {\n\to.Context = ctx\n}", "func (r *CompanyPaymentMethodsCollectionRequest) Get(ctx context.Context) ([]PaymentMethod, error) {\n\treturn r.GetN(ctx, 0)\n}", "func (o *GiftCardAccountGiftCardAccountManagementV1DeleteByQuoteIDDeleteMineParams) WithHTTPClient(client *http.Client) *GiftCardAccountGiftCardAccountManagementV1DeleteByQuoteIDDeleteMineParams {\n\to.SetHTTPClient(client)\n\treturn o\n}", "func (client ModelClient) ListIntentsSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func NewQuotePaymentMethodManagementV1SetPutMineParams() *QuotePaymentMethodManagementV1SetPutMineParams {\n\tvar ()\n\treturn &QuotePaymentMethodManagementV1SetPutMineParams{\n\n\t\ttimeout: cr.DefaultTimeout,\n\t}\n}", "func NewQuotePaymentMethodManagementV1SetPutMineParamsWithHTTPClient(client *http.Client) *QuotePaymentMethodManagementV1SetPutMineParams {\n\tvar ()\n\treturn &QuotePaymentMethodManagementV1SetPutMineParams{\n\t\tHTTPClient: client,\n\t}\n}", "func (m *memory) List() ([]Quote, error) {\n\tquotes := make([]Quote, len(m.quotes))\n\tfor _, v := range m.quotes {\n\t\tquotes = append(quotes, v)\n\t}\n\treturn quotes, nil\n}", "func (o BucketCorsItemOutput) Method() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v BucketCorsItem) []string { return v.Method }).(pulumi.StringArrayOutput)\n}", "func (s *OrderShippingAddressService) List(ctx context.Context, orderID int, params *OrderShippingAddressListParams) ([]OrderShippingAddress, *http.Response, error) {\n\tvar osa []OrderShippingAddress\n\tapiError := new(APIError)\n\n\tresp, err := performRequest(ctx, s.sling.New().Get(fmt.Sprintf(\"%d/shipping_addresses\", orderID)).QueryStruct(params), s.httpClient, &osa, apiError)\n\treturn osa, resp, relevantError(err, *apiError)\n}", "func (s *SystemService) ListMethods() ([]string, error) {\r\n\tvar reply []string\r\n\tif err := s.rpc.Call(system_list_methods, nil, &reply); err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\treturn reply, nil\r\n}", "func (client PatternClient) GetIntentPatternsSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tautorest.DoRetryForStatusCodes(client.RetryAttempts, client.RetryDuration, autorest.StatusCodesForRetry...))\n}", "func (client FavoritesClient) ListSender(req *http.Request) (*http.Response, error) {\n\treturn autorest.SendWithSender(client, req,\n\t\tazure.DoRetryWithRegistration(client.Client))\n}", "func (dm *FloatingIPService) GetFloatingIPList(serverID string) (floatingIPs []*types.FloatingIP, err error) {\n\tlog.Debug(\"GetFloatingIPList\")\n\n\tpath := \"/network/floating_ips\"\n\tif serverID != \"\" {\n\t\tpath = fmt.Sprintf(\"/cloud/servers/%s/floating_ips\", serverID)\n\n\t}\n\tdata, status, err := dm.concertoService.Get(path)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err = utils.CheckStandardStatus(status, data); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err = json.Unmarshal(data, &floatingIPs); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn floatingIPs, nil\n}", "func (tracker *PeerTracker) List() []*types.ChainInfo {\n\ttracker.mu.Lock()\n\tdefer tracker.mu.Unlock()\n\n\tvar tracked []*types.ChainInfo\n\tfor _, ci := range tracker.peers {\n\t\ttracked = append(tracked, ci)\n\t}\n\tout := make([]*types.ChainInfo, len(tracked))\n\tcopy(out, tracked)\n\treturn out\n}", "func (x *Gateway) GetAllowedIPs() []string {\n\treturn append(x.GetRoutes(), x.GetIp())\n}", "func (bi *Blockchainidentifier) Mine(w http.ResponseWriter, r *http.Request) {\n\tlog.Println(r.RemoteAddr + \" GET /mine\")\n\n\tlastblock := bi.lastBlock()\n\tlastblockhash := blockHasher(lastblock)\n\tlastproof := lastblock.Proof\n\tnewproof := bi.proofOfWork(lastproof)\n\tnewblockindex := bi.newTransaction(transaction{\n\t\tAmount: 1,\n\t\tRecipient: bi.Nodeidentifier,\n\t\tSender: \"0\",\n\t\tTimestamp: time.Now().UTC().Format(\"2006-01-02 15:04:05\"),\n\t})\n\n\tblockforged := bi.newBlock(newblockindex, newproof, lastblockhash)\n\n\tresponseMessage := map[string]interface{}{\n\t\t\"message\": \"New Block Forged\",\n\t\t\"index\": blockforged.Index,\n\t\t\"previous_block_hash\": blockforged.Previousblockhash,\n\t\t\"proof\": blockforged.Proof,\n\t\t\"transactions\": blockforged.Transactions,\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\tw.WriteHeader(http.StatusCreated)\n\tjson.NewEncoder(w).Encode(responseMessage)\n}", "func (tr *TaxRateMaster) List() []string {\n\tlist := make([]string, 0)\n\tfor _, v := range tr.taxRate {\n\t\ts := fmt.Sprint(v)\n\t\tlist = append(list, s)\n\t}\n\treturn list\n}", "func (_FeeCurrencyWhitelist *FeeCurrencyWhitelistCaller) GetWhitelist(opts *bind.CallOpts) ([]common.Address, error) {\n\tvar (\n\t\tret0 = new([]common.Address)\n\t)\n\tout := ret0\n\terr := _FeeCurrencyWhitelist.contract.Call(opts, out, \"getWhitelist\")\n\treturn *ret0, err\n}", "func (_Crowdsale *CrowdsaleCallerSession) BonusOffList() (*big.Int, error) {\n\treturn _Crowdsale.Contract.BonusOffList(&_Crowdsale.CallOpts)\n}", "func (b *CompanyRequestBuilder) ShipmentMethods() *CompanyShipmentMethodsCollectionRequestBuilder {\n\tbb := &CompanyShipmentMethodsCollectionRequestBuilder{BaseRequestBuilder: b.BaseRequestBuilder}\n\tbb.baseURL += \"/shipmentMethods\"\n\treturn bb\n}", "func (l *MethodResponseList) GetItems() []resource.Managed {\n\titems := make([]resource.Managed, len(l.Items))\n\tfor i := range l.Items {\n\t\titems[i] = &l.Items[i]\n\t}\n\treturn items\n}", "func (c *FakeApiGatewayMethodSettingses) List(opts v1.ListOptions) (result *v1alpha1.ApiGatewayMethodSettingsList, err error) {\n\tobj, err := c.Fake.\n\t\tInvokes(testing.NewListAction(apigatewaymethodsettingsesResource, apigatewaymethodsettingsesKind, c.ns, opts), &v1alpha1.ApiGatewayMethodSettingsList{})\n\n\tif obj == nil {\n\t\treturn nil, err\n\t}\n\n\tlabel, _, _ := testing.ExtractFromListOptions(opts)\n\tif label == nil {\n\t\tlabel = labels.Everything()\n\t}\n\tlist := &v1alpha1.ApiGatewayMethodSettingsList{ListMeta: obj.(*v1alpha1.ApiGatewayMethodSettingsList).ListMeta}\n\tfor _, item := range obj.(*v1alpha1.ApiGatewayMethodSettingsList).Items {\n\t\tif label.Matches(labels.Set(item.Labels)) {\n\t\t\tlist.Items = append(list.Items, item)\n\t\t}\n\t}\n\treturn list, err\n}", "func filterMethods(args []sexp) (Filter, error) {\n\tmethods := make([]layers.SIPMethod, len(args))\n\n\t// first pull out all our methods. If we catch any that aren't known sip\n\t// methods, that's an error.\n\tfor i, a := range args {\n\t\ts := \"\"\n\t\tswitch v := a.i.(type) {\n\t\tcase qString:\n\t\t\ts = string(v)\n\t\tcase string:\n\t\t\ts = v\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"arg type %v: %w\", i, ErrMethodsType)\n\t\t}\n\t\tmethod, err := layers.GetSIPMethod(s)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"bad argument (#%v), %v %w: %v\", i, s, ErrMethodsType, err)\n\t\t}\n\t\tmethods[i] = method\n\t}\n\n\treturn func(msg *layers.SIP) bool {\n\t\tfor _, m := range methods {\n\t\t\tif msg.Method == m {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t\treturn false\n\t}, nil\n}", "func (m *User) GetImAddresses()([]string) {\n return m.imAddresses\n}", "func List(client *gophercloud.ServiceClient) pagination.Pager {\n\treturn pagination.NewPager(client, listURL(client), func(r pagination.PageResult) pagination.Page {\n\t\treturn FloatingIPsPage{pagination.SinglePageBase(r)}\n\t})\n}", "func (s *CheckoutAPIOp) GetShippingRates(token string) ([]ShippingRate, error) {\n\tpath := fmt.Sprintf(\"%s/%s/shipping_rates.json\", checkoutsBasePath, token)\n\tresource := new(ShippingRatesResource)\n\terr := s.client.Get(path, resource, nil)\n\treturn resource.ShippingRates, err\n}", "func GiveMyMine(path string) Mine {\n\tpanic(\"non implementato\")\n}", "func (r *OffersService) List() *OffersListCall {\n\tc := &OffersListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\treturn c\n}", "func (r *SitemapsService) List(siteUrl string) *SitemapsListCall {\n\tc := &SitemapsListCall{s: r.s, opt_: make(map[string]interface{})}\n\tc.siteUrl = siteUrl\n\treturn c\n}", "func (fn *formulaFuncs) IPMT(argsList *list.List) formulaArg {\n\treturn fn.ipmt(\"IPMT\", argsList)\n}", "func GetBlacklistedIPs() ([]IP, error) {\n\tvar result []IP\n\trows, err := db.Query(\n\t\t`SELECT web_ips.*\n\t\t\t FROM blacklist\n \t\t\t JOIN web_users ON blacklist.discord_id = web_users.id\n \t\t\t JOIN web_ips ON web_users.id = web_ips.user_id`)\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\n\tfor rows.Next() {\n\t\tnextIP := IP{}\n\t\terr = rows.Scan(&nextIP.UserID, &nextIP.IP, &nextIP.LastUsed)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresult = append(result, nextIP)\n\t}\n\terr = rows.Err()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn result, nil\n}", "func (cs *checkoutService) quoteShipping(ctx context.Context, address *pb.Address, items []*pb.CartItem) (*pb.Money, error) {\n\tshippingQuote, err := pb.NewShippingServiceClient(cs.shippingSvcConn).\n\t\tGetQuote(ctx, &pb.GetQuoteRequest{\n\t\t\tAddress: address,\n\t\t\tItems: items})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get shipping quote: %+v\", err)\n\t}\n\treturn shippingQuote.GetCostUsd(), nil\n}", "func (a API) GetMiningInfo(cmd *None) (e error) {\n\tRPCHandlers[\"getmininginfo\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}", "func (r *SitemapsService) List(siteUrl string) *SitemapsListCall {\n\tc := &SitemapsListCall{s: r.s, urlParams_: make(gensupport.URLParams)}\n\tc.siteUrl = siteUrl\n\treturn c\n}", "func NewGiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParamsWithContext(ctx context.Context) *GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams {\n\tvar ()\n\treturn &GiftRegistryShippingMethodManagementV1EstimateByRegistryIDPostMineParams{\n\n\t\tContext: ctx,\n\t}\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WithQuotePaymentMethodManagementV1SetPutBody(quotePaymentMethodManagementV1SetPutBody QuotePaymentMethodManagementV1SetPutMineBody) *QuotePaymentMethodManagementV1SetPutMineParams {\n\to.SetQuotePaymentMethodManagementV1SetPutBody(quotePaymentMethodManagementV1SetPutBody)\n\treturn o\n}", "func (m *SiteItemRequestBuilder) Lists()(*ItemListsRequestBuilder) {\n return NewItemListsRequestBuilderInternal(m.pathParameters, m.requestAdapter)\n}", "func (ps *PgStore) GetGatewaysLoc(ctx context.Context) ([]GatewayLocation, error) {\n\tvar gwsLoc []GatewayLocation\n\n\terr := sqlx.SelectContext(ctx, ps.db, &gwsLoc, `\n\t\tSELECT latitude, longitude, altitude\n\t\tFROM gateway\n\t\tWHERE latitude != 0 OR longitude != 0`,\n\t)\n\treturn gwsLoc, err\n}", "func (client IotHubResourceClient) ListJobsSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func (in *MethodList) DeepCopy() *MethodList {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(MethodList)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func GetIPList(ip string) []string {\n\tif len(ip) == 0 {\n\t\treturn make([]string, 0)\n\t}\n\n\treturn strings.Split(ip, \",\")\n}", "func (r *friendRepository) RequestList(id string) (*[]model.FriendRequest, error) {\n\tvar requests []model.FriendRequest\n\n\tresult := r.DB.\n\t\tRaw(`\n\t\t select u.id, u.username, u.image, 1 as \"type\" from users u\n\t\t join friend_requests fr on u.id = fr.\"sender_id\"\n\t\t where fr.\"receiver_id\" = @id\n\t\t UNION\n\t\t select u.id, u.username, u.image, 0 as \"type\" from users u\n\t\t join friend_requests fr on u.id = fr.\"receiver_id\"\n\t\t where fr.\"sender_id\" = @id\n\t\t order by username;\n\t\t`, sql.Named(\"id\", id)).\n\t\tFind(&requests)\n\n\treturn &requests, result.Error\n}", "func (r *bitroute) allowedMethods(path string) []string {\n\tvar allowed []string\n\tfor method, parser := range r.handlers {\n\t\tif _, _, ok := parser.get(path); ok {\n\t\t\tallowed = append(allowed, method)\n\t\t}\n\t}\n\n\treturn allowed\n}", "func (b Block) CalculateMinerFees() Currency {\n\tfees := NewCurrency64(0)\n\tfor _, txn := range b.Transactions {\n\t\tfor _, fee := range txn.MinerFees {\n\t\t\tfees = fees.Add(fee)\n\t\t}\n\t}\n\treturn fees\n}", "func (_FeeCurrencyWhitelist *FeeCurrencyWhitelistCaller) Whitelist(opts *bind.CallOpts, arg0 *big.Int) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _FeeCurrencyWhitelist.contract.Call(opts, out, \"whitelist\", arg0)\n\treturn *ret0, err\n}", "func (p *listDiscoveryPlugin) Method() string {\n\treturn \"list\"\n}", "func (s *CouponsService) List(ctx context.Context, productID int) ([]*Coupon, *http.Response, error) {\n\tu := \"2.0/product/list_coupons\"\n\n\toptions := &CouponsOptions{\n\t\tProductID: productID,\n\t}\n\n\treq, err := s.client.NewRequest(\"POST\", u, options)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tcouponsResponse := new(CouponsResponse)\n\tresponse, err := s.client.Do(ctx, req, couponsResponse)\n\tif err != nil {\n\t\treturn nil, response, err\n\t}\n\n\treturn couponsResponse.Response, response, nil\n}", "func (m *Market) Symbols() (MarketResponse, error) {\n\tsymbolURL := URL(\"/v1/symbols\")\n\n\tvar result MarketResponse\n\tresp, err := method.Get(symbolURL, nil, nil)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tdefer resp.Body.Close()\n\n\treturn bodyToMarketResponse(resp.Body, &result)\n}", "func (o *QuotePaymentMethodManagementV1SetPutMineParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif err := r.SetBodyParam(o.QuotePaymentMethodManagementV1SetPutBody); err != nil {\n\t\treturn err\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (m *RetrieveShippingMethodsRequest) Validate() error {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\t// no validation rules for AccessToken\n\n\tif v, ok := interface{}(m.GetTime()).(interface{ Validate() error }); ok {\n\t\tif err := v.Validate(); err != nil {\n\t\t\treturn RetrieveShippingMethodsRequestValidationError{\n\t\t\t\tfield: \"Time\",\n\t\t\t\treason: \"embedded message failed validation\",\n\t\t\t\tcause: err,\n\t\t\t}\n\t\t}\n\t}\n\n\tif _, ok := _RetrieveShippingMethodsRequest_Dow_InLookup[m.GetDow()]; !ok {\n\t\treturn RetrieveShippingMethodsRequestValidationError{\n\t\t\tfield: \"Dow\",\n\t\t\treason: \"value must be in list [monday tuesday wednesday thursday friday saturday sunday]\",\n\t\t}\n\t}\n\n\treturn nil\n}", "func NewQuoteCouponManagementV1SetPutMineDefault(code int) *QuoteCouponManagementV1SetPutMineDefault {\n\treturn &QuoteCouponManagementV1SetPutMineDefault{\n\t\t_statusCode: code,\n\t}\n}", "func (s *awsCustomerGatewayLister) List(selector labels.Selector) (ret []*v1.AwsCustomerGateway, err error) {\n\terr = cache.ListAll(s.indexer, selector, func(m interface{}) {\n\t\tret = append(ret, m.(*v1.AwsCustomerGateway))\n\t})\n\treturn ret, err\n}", "func (s *serfNet) MembersIP() (addr []net.IP) {\n\n\tmembers := s._serf.Members()\n\tfor i := 0; i < len(members); i++ {\n\t\t//fmt.Println(\"localMember \", []byte(s.serf.LocalMember().Name), \"members[i].Name \", []byte(members[i].Name), \" status \", members[i].Status, \" addr \", members[i].Addr)\n\t\tif members[i].Name != s._serf.LocalMember().Name {\n\t\t\taddr = append(addr, members[i].Addr)\n\t\t}\n\t}\n\treturn\n}", "func (s *Server) List(ctx context.Context, in *proto.GetBlockRequest) (*proto.GetBlockResponse, error) {\n\ti, err := metrics.Gauge(\"List\", func() (interface{}, error) {\n\t\tresp := new(proto.GetBlockResponse)\n\n\t\tfor _, b := range s.Blockchain.Blocks {\n\t\t\tresp.Blocks = append(resp.Blocks, &proto.Block{\n\t\t\t\tPrevBlockHash: b.PrevBlockHash,\n\t\t\t\tData: b.Data,\n\t\t\t\tHash: b.Hash,\n\t\t\t})\n\t\t}\n\n\t\treturn resp, nil\n\t})\n\treturn i.(*proto.GetBlockResponse), err\n}", "func (r ListProcessingJobsRequest) Send(ctx context.Context) (*ListProcessingJobsResponse, error) {\n\tr.Request.SetContext(ctx)\n\terr := r.Request.Send()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp := &ListProcessingJobsResponse{\n\t\tListProcessingJobsOutput: r.Request.Data.(*ListProcessingJobsOutput),\n\t\tresponse: &aws.Response{Request: r.Request},\n\t}\n\n\treturn resp, nil\n}", "func (s *SQLStorage) GetWhiteList(ctx context.Context) ([]*IPNet, error) {\n\treturn s.getSubnetList(ctx, \"SELECT ip, mask FROM whitelist\")\n}", "func (m *HotfixMessage) GetFixes() []*Hotfix {\n\treturn m.fixes\n}", "func (_Crowdsale *CrowdsaleCaller) WhiteList(opts *bind.CallOpts) (common.Address, error) {\n\tvar (\n\t\tret0 = new(common.Address)\n\t)\n\tout := ret0\n\terr := _Crowdsale.contract.Call(opts, out, \"whiteList\")\n\treturn *ret0, err\n}", "func (client MSIXPackagesClient) ListSender(req *http.Request) (*http.Response, error) {\n\treturn client.Send(req, azure.DoRetryWithRegistration(client.Client))\n}", "func (m *RetrieveShippingMethodsResponse) Validate() error {\n\tif m == nil {\n\t\treturn nil\n\t}\n\n\tfor key, val := range m.GetShippingMethods() {\n\t\t_ = val\n\n\t\t// no validation rules for ShippingMethods[key]\n\n\t\tif v, ok := interface{}(val).(interface{ Validate() error }); ok {\n\t\t\tif err := v.Validate(); err != nil {\n\t\t\t\treturn RetrieveShippingMethodsResponseValidationError{\n\t\t\t\t\tfield: fmt.Sprintf(\"ShippingMethods[%v]\", key),\n\t\t\t\t\treason: \"embedded message failed validation\",\n\t\t\t\t\tcause: err,\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func (db2 *DB2) GetMarketSymbols(ctx context.Context, market string) ([]string, error) {\n\tif market == \"\" || tradingdb2utils.IndexOfStringSlice(db2.cfg.DB2Markets, market, 0) < 0 {\n\t\treturn nil, ErrInvalidMarket\n\t}\n\n\tsymbols := []string{}\n\terr := db2.AnkaDB.ForEachWithPrefix(ctx, market, makeSymbolDB2KeyPrefix(market), func(key string, buf []byte) error {\n\t\tsi := &tradingpb.SymbolInfo{}\n\n\t\terr := proto.Unmarshal(buf, si)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tsymbols = append(symbols, si.Symbol)\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn symbols, nil\n}", "func (m *ItemSitesSiteItemRequestBuilder) Lists()(*ItemSitesItemListsRequestBuilder) {\n return NewItemSitesItemListsRequestBuilderInternal(m.BaseRequestBuilder.PathParameters, m.BaseRequestBuilder.RequestAdapter)\n}", "func New(ctx context.Context, minionAddresses []string, opts ...grpc.DialOption) (*Server, error) {\n\tserver := &Server{\n\t\tminions: make(map[string]mpb.MinionClient),\n\t\tinitialInterests: nil,\n\t}\n\n\tlog.Println(\"Reaching out to all minions.\")\n\t// Build map of minions.\n\tfor _, addr := range minionAddresses {\n\t\tlog.Printf(\"Contacting %s\\n\", addr)\n\t\tc, err := grpc.Dial(addr, opts...)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tlog.Println(\"Ok, minion connected\")\n\t\tserver.minions[addr] = mpb.NewMinionClient(c)\n\t}\n\n\tlog.Println(\"Retrieving initial interests\")\n\tfor _, m := range server.minions {\n\t\t// TODO(paradoxengine): most likely, a deadline here?\n\t\tintResp, err := m.ListInitialInterests(ctx, &mpb.ListInitialInterestsRequest{})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tfor _, i := range intResp.GetInterests() {\n\t\t\tlog.Printf(\"Got interest: %s\", i)\n\t\t\tserver.initialInterests = append(server.initialInterests, i)\n\t\t}\n\t}\n\tlog.Printf(\"Minimizing interests, now %d\", len(server.initialInterests))\n\tserver.initialInterests = interests.Minify(server.initialInterests)\n\tlog.Printf(\"Minimized interests, now %d\", len(server.initialInterests))\n\treturn server, nil\n}", "func (miner *Miner) Mine(team *Team, index int) {\n\tticker := time.NewTicker(time.Second * time.Duration(miner.interval))\n\tdefer ticker.Stop()\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\tif !team.DiscoveredNodes[index].Active || !miner.Process.Alive {\n\t\t\t\treturn\n\t\t\t}\n\t\t\t//team.mutex.Unlock()\n\t\t\tteam.mutex.Lock()\n\t\t\tdefer team.mutex.Unlock()\n\t\t\tswitch miner.minerType {\n\t\t\tcase \"Bandwidth\":\n\t\t\t\tteam.Bandwidth = team.Bandwidth + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\tcase \"IO\":\n\t\t\t\tteam.Io = team.Io + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\tcase \"Entropy\":\n\t\t\t\tteam.Entropy = team.Entropy + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\tcase \"CPU\":\n\t\t\t\tteam.Cpu = team.Cpu + miner.amount\n\t\t\t\tteam.mutex.Unlock()\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t//team.mutex.Unlock()\n\t\t}\n\t}\n}" ]
[ "0.6247422", "0.6232551", "0.5750774", "0.56089175", "0.5156768", "0.5034006", "0.48612842", "0.47912076", "0.4702766", "0.4579636", "0.44830504", "0.44254896", "0.43943653", "0.42248517", "0.422466", "0.42198053", "0.41508842", "0.41488403", "0.40638202", "0.40435565", "0.39952555", "0.39860296", "0.39765477", "0.3930736", "0.39237684", "0.3917167", "0.39022204", "0.38867593", "0.387692", "0.38449684", "0.3814913", "0.3811153", "0.37816307", "0.3773806", "0.37608498", "0.37437394", "0.37391454", "0.3737431", "0.37344772", "0.37300545", "0.37281635", "0.37156767", "0.37039557", "0.37020686", "0.36859316", "0.36743423", "0.36701766", "0.36699247", "0.36534795", "0.36532444", "0.36488122", "0.3643858", "0.3628113", "0.36240584", "0.3602763", "0.36023572", "0.36002365", "0.35999623", "0.35986525", "0.35873505", "0.3587036", "0.3578396", "0.35699573", "0.3568151", "0.35614383", "0.35610983", "0.35510805", "0.3540908", "0.3529289", "0.35250145", "0.35216525", "0.35215843", "0.3516068", "0.35138264", "0.35109428", "0.35058817", "0.34974802", "0.34912914", "0.34899583", "0.34873426", "0.34850848", "0.34820327", "0.3480262", "0.34776214", "0.34744298", "0.3468909", "0.34650066", "0.34639168", "0.34632838", "0.34605852", "0.3456211", "0.3456099", "0.34527785", "0.34502545", "0.34415415", "0.34358835", "0.3433193", "0.3431332", "0.3430394", "0.34289882" ]
0.72267836
0
SetTransport changes the transport on the client
func (a *Client) SetTransport(transport runtime.ClientTransport) { a.transport = transport }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *JusticeLobbyService) SetTransport(transport runtime.ClientTransport) {\n\tc.Transport = transport\n\tc.LobbyOperations.SetTransport(transport)\n\tc.Admin.SetTransport(transport)\n\tc.Config.SetTransport(transport)\n\tc.Friends.SetTransport(transport)\n\tc.Notification.SetTransport(transport)\n\tc.Party.SetTransport(transport)\n\tc.Player.SetTransport(transport)\n\tc.Presence.SetTransport(transport)\n\tc.Profanity.SetTransport(transport)\n\tc.ThirdParty.SetTransport(transport)\n}", "func (c *Client) SetTransport(tripper http.RoundTripper) {\n\tc.resty.SetTransport(tripper)\n}", "func (c *Jasperserver) SetTransport(transport runtime.ClientTransport) {\n\tc.Transport = transport\n\n\tc.Operations.SetTransport(transport)\n\n}", "func (c *OpenbankingPaymentsClient) SetTransport(transport runtime.ClientTransport) {\n\tc.Transport = transport\n\tc.DomesticPayments.SetTransport(transport)\n\tc.DomesticScheduledPayments.SetTransport(transport)\n\tc.DomesticStandingOrders.SetTransport(transport)\n\tc.FilePayments.SetTransport(transport)\n\tc.InternationalPayments.SetTransport(transport)\n\tc.InternationalScheduledPayments.SetTransport(transport)\n\tc.InternationalStandingOrders.SetTransport(transport)\n\tc.PaymentDetails.SetTransport(transport)\n}", "func (c *Jiskefet) SetTransport(transport runtime.ClientTransport) {\n\tc.Transport = transport\n\n\tc.Attachments.SetTransport(transport)\n\n\tc.Authentication.SetTransport(transport)\n\n\tc.Flp.SetTransport(transport)\n\n\tc.Logs.SetTransport(transport)\n\n\tc.Overview.SetTransport(transport)\n\n\tc.Runs.SetTransport(transport)\n\n\tc.Setting.SetTransport(transport)\n\n\tc.Subsystems.SetTransport(transport)\n\n\tc.Tags.SetTransport(transport)\n\n\tc.Users.SetTransport(transport)\n\n}" ]
[ "0.8350255", "0.81898487", "0.79902893", "0.79455817", "0.78605556" ]
0.0
-1
NewIdentifierFactory creates an IdentifierFactory ready for use
func NewIdentifierFactory() IdentifierFactory { return &identifierFactory{ renames: createRenames(), reservedWords: createReservedWords(), idCache: make(idCache), } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewIdentifierFactory() IdentifierFactory {\n\treturn &identifierFactory{\n\t\trenames: createRenames(),\n\t}\n}", "func NewIdentifierFactory() IdentifierFactory {\n\treturn &identifierFactory{\n\t\trenames: createRenames(),\n\t\treservedWords: createReservedWords(),\n\t\tidCache: make(idCache),\n\t\tinternCache: make(internCache),\n\t\treceiverCache: make(map[string]string),\n\t\tforbiddenReceiverSuffixes: createForbiddenReceiverSuffixes(),\n\t}\n}", "func NewIdentifier(login string, password string) *Identifier {\n\treturn &Identifier{\n\t\tlogin: login,\n\t\tpassword: password,\n\t\tauthEndpoint: authEndpoint,\n\t\tHTTPClient: &http.Client{Timeout: 10 * time.Second},\n\t}\n}", "func NewIdentifier() string {\n\tif !seededrng {\n\t\trand.Seed(time.Now().Unix())\n\t\tseededrng = true\n\t}\n\tr := make([]byte, 9) // 9 bytes * (4/3 encoding) = 12 characters\n\trand.Read(r)\n\ts := base64.URLEncoding.EncodeToString(r)\n\treturn s\n}", "func NewIdentifier(opts ...Option) Profile {\n\treturn Profile{\n\t\toptions: getOpts(opts...),\n\t\tclass: identifier,\n\t}\n}", "func newIdentifier(req messages.Alias, c *models.Customer) (m *models.Identifier, err error) {\n\tm = models.NewIdentifier(c)\n\tm.AliasName = req.AliasName\n\tm.IP, err = newIp(req.Ip)\n\tif err != nil {\n\t\treturn\n\t}\n\tm.Prefix, err = newPrefix(req.Prefix)\n\tif err != nil {\n\t\treturn\n\t}\n\tm.PortRange, err = newPortRange(req.PortRange)\n\tif err != nil {\n\t\treturn\n\t}\n\tm.TrafficProtocol.AddList(req.TrafficProtocol)\n\tm.FQDN.AddList(req.FQDN)\n\tm.URI.AddList(req.URI)\n\n\treturn\n}", "func (factory *identifierFactory) createIdentifierImpl(name string, visibility Visibility, reservedWords reservedWordConsideration) string {\n\tcacheKey := idCacheKey{name, visibility, reservedWords}\n\tfactory.rwLock.RLock()\n\tcached, ok := factory.idCache[cacheKey]\n\tfactory.rwLock.RUnlock()\n\tif ok {\n\t\treturn cached\n\t}\n\n\tresult := factory.createIdentifierUncached(name, visibility, reservedWords)\n\tfactory.rwLock.Lock()\n\tresult = factory.internCache.intern(result)\n\tfactory.idCache[cacheKey] = result\n\tfactory.rwLock.Unlock()\n\treturn result\n}", "func NewIdentifier(opts ...Option) *Profile {\n\treturn &Profile{\n\t\toptions: getOpts(opts...),\n\t\tclass: identifier,\n\t}\n}", "func NewIdentifier(t token.Token) *Identifier {\n\treturn &Identifier{\n\t\tToken: t,\n\t\tValue: t.Literal,\n\t}\n}", "func Factory(loader func(v interface{}) error) (uniqueid.Driver, error) {\n\ti := NewUUID()\n\tconf := &UUIDConfig{}\n\n\terr := loader(conf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tswitch conf.Version {\n\tcase 4:\n\t\ti.creator = uuid.NewV4\n\tdefault:\n\t\ti.creator = uuid.NewV1\n\t}\n\treturn i, nil\n\n}", "func (factory *identifierFactory) CreateIdentifier(name string, visibility Visibility) string {\n\tcacheKey := idCacheKey{name, visibility}\n\tfactory.rwLock.RLock()\n\tcached, ok := factory.idCache[cacheKey]\n\tfactory.rwLock.RUnlock()\n\tif ok {\n\t\treturn cached\n\t}\n\n\tresult := factory.createIdentifierUncached(name, visibility)\n\tfactory.rwLock.Lock()\n\tfactory.idCache[cacheKey] = result\n\tfactory.rwLock.Unlock()\n\treturn result\n}", "func NewID() string {\n\treturn idFunc()\n}", "func (factory *identifierFactory) CreateIdentifier(name string, visibility Visibility) string {\n\treturn factory.createIdentifierImpl(name, visibility, avoidReservedWords)\n}", "func (f *testIDService) New() (string, error) {\n\treturn \"new-ID\", nil\n}", "func New(cacheNodeidentityInfo bool) (nodeidentity.Identifier, error) {\n\tclient, err := newClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &nodeIdentifier{\n\t\tvmssGetter: client,\n\t\tcache: expirationcache.NewTTLStore(stringKeyFunc, cacheTTL),\n\t\tcacheEnabled: cacheNodeidentityInfo,\n\t}, nil\n}", "func NewIdent(name string) *Ident { return &Ident{token.NoPos, name} }", "func New() ID {\n\treturn dgen.New()\n}", "func New(d *Dispatcher) *Identity {\n\td.Lock()\n\tdefer d.Unlock()\n\tstate := randToken()\n\ti := &Identity{\n\t\tstate: state,\n\t}\n\td.identities[state] = i\n\treturn i\n}", "func (s *session) newID() (interface{}, error) {\n\tvar b [32]byte\n\t_, err := rand.Read(b[:])\n\treturn hex.EncodeToString(b[:]), err\n}", "func NewID(typ string) string {\n\treturn typ + \"_\" + NewBareID(16)\n}", "func NewID() string {\n\treturn ksuid.New().String()\n}", "func (*dynamicUUIDProvider) New() string {\n\treturn uuid.New()\n}", "func NewID(length int) string {\n\tb := make([]byte, length*6/8)\n\t_, err := rand.Read(b)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn base64.RawURLEncoding.EncodeToString(b)\n}", "func newID() string {\n\tvar b [8]byte\n\t_, err := rand.Read(b[:])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn fmt.Sprintf(\"%x\", b[:])\n}", "func newID() string {\n\treturn \"_\" + uuid.New().String()\n}", "func NewID() string {\n\treturn replacer.Replace(NewV4().String())\n}", "func MakeIdentifier() string {\n\tb := make([]byte, 12)\n\t_, err := io.ReadFull(rand.Reader, b)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn fmt.Sprintf(\"%x\", b)\n}", "func (this *IdBuilder) NewId() (rv string) {\n\tarr := [24]byte{}\n\ts := arr[:]\n\n\t//\n\t// base32 encode all 64 bits of counter\n\t// - encode 5 bits at a time allows us to use base62 routine to get base32\n\t// - we stop at 10 because that's where date/time begins\n\t//\n\tu := atomic.AddUint64(&this.counter, 1)\n\tvar shift uint = 0\n\tfor i := len(arr) - 1; i > 10; i-- {\n\t\tb := byte(u >> shift)\n\t\tshift += 5\n\t\ts[i] = base62(b & 0x1f)\n\t}\n\n\t//\n\t// encode YMMDDHHMMSS\n\t//\n\t// we need the year as these can get databased\n\t//\n\t// we want the MMDDHHMMSS to be easily human readable\n\t//\n\tnow := time.Now().UTC()\n\tyear, month, day := now.Date()\n\thour, minute, second := now.Clock()\n\ts[0] = base62(byte(year - 2016))\n\ts[1], s[2] = base10(byte(month))\n\ts[3], s[4] = base10(byte(day))\n\ts[5], s[6] = base10(byte(hour))\n\ts[7], s[8] = base10(byte(minute))\n\ts[9], s[10] = base10(byte(second))\n\n\treturn string(s)\n}", "func newid() string {\n\treturn make_key(6)\n}", "func newHashID() (*hashids.HashID, error) {\n\t// Defaults\n\tsalt := \"Best salt\"\n\tminLength := 8\n\n\t// Initiliazing HashID\n\thd := hashids.NewData()\n\thd.Salt = salt\n\thd.MinLength = minLength\n\th, err := hashids.NewWithData(hd)\n\treturn h, err\n}", "func (f JwtFactory) NewIDToken(username, host, clientID, nonce string) (string, error) {\r\n\tt := jwt.New(jwt.GetSigningMethod(\"RS256\"))\r\n\r\n\tt.Claims = &struct {\r\n\t\tUsername string `json:\"username\"`\r\n\t\tNonce string `json:\"nonce,omitempty\"`\r\n\r\n\t\t// azp is the authorized party - the party to which the ID Token was\r\n\t\t// issued. Same as Audience.\r\n\t\tAzp string `json:\"azp\"`\r\n\r\n\t\t// Purpose defines what this JWT is for, either access_token or\r\n\t\t// id_token.\r\n\t\tPurpose string `json:\"purpose\"`\r\n\r\n\t\tjwt.StandardClaims\r\n\t}{\r\n\t\tusername,\r\n\t\tnonce,\r\n\t\tclientID,\r\n\t\t\"id_token\",\r\n\t\tgetStandardClaims(host, username, clientID),\r\n\t}\r\n\r\n\treturn f.sign(t)\r\n}", "func TestIdentifier(t *testing.T) {\n\t// Type as identifier.\n\tvar kvlf KeyValueLessFunc\n\n\tidp := TypeAsIdentifierPart(kvlf)\n\n\tif idp != \"key-value-less-func\" {\n\t\tt.Errorf(\"Identifier part for KeyValueLessFunc is wrong, returned '%v'!\", idp)\n\t}\n\n\tidp = TypeAsIdentifierPart(NewUUID())\n\n\tif idp != \"u-u-i-d\" {\n\t\tt.Errorf(\"Identifier part for UUID is wrong, returned '%v'!\", idp)\n\t}\n\n\t// Identifier.\n\tid := Identifier(\"One\", 2, \"three four\")\n\n\tif id != \"one:2:three-four\" {\n\t\tt.Errorf(\"First identifier is wrong! Id: %v\", id)\n\t}\n\n\tid = Identifier(2011, 6, 22, \"One, two, or three things.\")\n\n\tif id != \"2011:6:22:one-two-or-three-things\" {\n\t\tt.Errorf(\"Second identifier is wrong! Id: %v\", id)\n\t}\n\n\tid = SepIdentifier(\"+\", 1, \"oNe\", 2, \"TWO\", \"3\", \"ÄÖÜ\")\n\n\tif id != \"1+one+2+two+3+äöü\" {\n\t\tt.Errorf(\"Third identifier is wrong! Id: %v\", id)\n\t}\n\n\tid = LimitedSepIdentifier(\"+\", true, \" \", 1, \"oNe\", 2, \"TWO\", \"3\", \"ÄÖÜ\", \"Four\", \"+#-:,\")\n\n\tif id != \"1+one+2+two+3+four\" {\n\t\tt.Errorf(\"Fourth identifier is wrong! Id: %v\", id)\n\t}\n}", "func (factory *identifierFactory) CreateStringIdentifier(name string, visibility Visibility) string {\n\treturn factory.createIdentifierImpl(name, visibility, allowReservedWords)\n}", "func NewID(v string) ID {\n\treturn ID{v, true}\n}", "func NewID(b []byte) (*SHA1, error) {\n\tif len(b) != 20 {\n\t\treturn nil, errors.New(\"length must be 20\")\n\t}\n\treturn MustID(b), nil\n}", "func (s *staticUUIDProvider) New() string {\n\treturn s.staticID\n}", "func (e *exprHelper) NewIdent(name string) ast.Expr {\n\treturn e.exprFactory.NewIdent(e.nextMacroID(), name)\n}", "func NewID(body []byte) ID {\n\treturn ID(xxhash.Sum64(body))\n}", "func NewId() string {\n\t// generate 128 random bits (6 more than standard UUID)\n\tbytes := make([]byte, 16)\n\t_, err := rand.Read(bytes)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// convert them to base 32 encoding\n\ts := base32.StdEncoding.EncodeToString(bytes)\n\treturn strings.ToLower(strings.TrimRight(s, \"=\"))\n}", "func newID() int64 {\n\treturn time.Now().UnixNano()\n}", "func New() (UUID, error) {\n\treturn globalSource.new()\n}", "func newIdentities(c *Client) *identities {\n\treturn &identities{\n\t\tr: c,\n\t}\n}", "func (factory *identifierFactory) CreateIdentifier(name string, visibility Visibility) string {\n\tif identifier, ok := factory.renames[name]; ok {\n\t\t// Just lowercase the first character according to visibility\n\t\tr := []rune(identifier)\n\t\tif visibility == NotExported {\n\t\t\tr[0] = unicode.ToLower(r[0])\n\t\t} else {\n\t\t\tr[0] = unicode.ToUpper(r[0])\n\t\t}\n\t\treturn string(r)\n\t}\n\n\t// replace with spaces so titlecasing works nicely\n\tclean := filterRegex.ReplaceAllLiteralString(name, \" \")\n\n\tcleanWords := sliceIntoWords(clean)\n\tvar caseCorrectedWords []string\n\tfor i, word := range cleanWords {\n\t\tif visibility == NotExported && i == 0 {\n\t\t\tcaseCorrectedWords = append(caseCorrectedWords, strings.ToLower(word))\n\t\t} else {\n\t\t\tcaseCorrectedWords = append(caseCorrectedWords, strings.Title(word))\n\t\t}\n\t}\n\n\tresult := strings.Join(caseCorrectedWords, \"\")\n\treturn result\n}", "func New() string {\n\treturn shortuuid.New()\n}", "func createMyId(localip string) id {\n\tcurTime := time.Now().String()\n\treturn id{localip, curTime}\n}", "func NewID(service Service, modelIndex string, uuid string) ID {\n\treturn &id{service, modelIndex, uuid}\n}", "func NewFactory(config io.Reader) (admission.Interface, error) {\n\treturn New()\n}", "func newGoIdent(f *File, d protoreflect.Descriptor) GoIdent {\n\tname := strings.TrimPrefix(string(d.FullName()), string(f.Desc.Package())+\".\")\n\treturn GoIdent{\n\t\tGoName: strs.GoCamelCase(name),\n\t\tGoImportPath: f.GoImportPath,\n\t}\n}", "func NewID(size int) string {\n\tb := make([]byte, size)\n\tfor i := range b {\n\t\tb[i] = ABC[rand.Intn(len(ABC))]\n\t}\n\n\treturn string(b)\n}", "func newJobID(tm Time) (Job, error) {\n\tk, err := ksuid.NewRandomWithTime(tm)\n\tif err != nil {\n\t\treturn Job{}, err\n\t}\n\treturn Job(k), nil\n}", "func newUID() ([]byte, error) {\n\t// uuid := make([]byte, 16)\n\t// n, err := io.ReadFull(rand.Reader, uuid)\n\t// if n != len(uuid) || err != nil {\n\t// \treturn nil, err\n\t// }\n\t// // variant bits; see section 4.1.1\n\t// uuid[8] = uuid[8]&^0xc0 | 0x80\n\t// // version 4 (pseudo-random); see section 4.1.3\n\t// uuid[6] = uuid[6]&^0xf0 | 0x40\n\t// return []byte(fmt.Sprintf(\"%x-%x-%x-%x-%x\", uuid[0:4], uuid[4:6], uuid[6:8], uuid[8:10], uuid[10:])), nil\n\treturn []byte(uniuri.New()), nil\n}", "func NewObjectIdentifier(oid string) (oi asn1.ObjectIdentifier, err error) {\n\tif len(oid) == 0 {\n\t\treturn nil, errors.Errorf(\"zero length OBJECT IDENTIFIER\")\n\t}\n\n\tif oid[0] == '{' {\n\t\t// ASN.1 notation. (eg {iso(1) member-body(2) us(840) rsadsi(113549) pkcs(1) pkcs-9(9) messageDigest(4)})\n\t\tparts := strings.Split(oid[1:len(oid)-1], \" \")\n\t\toi = make(asn1.ObjectIdentifier, len(parts), len(parts))\n\t\tfor i, part := range parts {\n\t\t\tidx := strings.IndexRune(part, '(')\n\t\t\toi[i], err = strconv.Atoi(part[idx+1 : len(part)-1])\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t} else {\n\t\t// Dot notation. (eg 1.2.840.113549.1.9.4)\n\t\tparts := strings.Split(oid, \".\")\n\t\toi = make(asn1.ObjectIdentifier, len(parts), len(parts))\n\t\tfor i, part := range parts {\n\t\t\toi[i], err = strconv.Atoi(part)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\treturn oi, nil\n}", "func NewIDFromString(body string) ID {\n\treturn ID(xxhash.Sum64String(body))\n}", "func genSID() string {\n\treturn fmt.Sprintf(\"s6a_proxy;%d_%X\", uint(time.Now().Unix()), rand.Uint32())\n}", "func NewStringID(v string) ID { return ID{name: v} }", "func NewFactory(buckets int) xkit.Factory {\n\treturn factory{\n\t\tbuckets: buckets,\n\t}\n}", "func newIDList(p *idElementPool) *idList {\n\tl := &idList{Pool: p}\n\treturn l.Init()\n}", "func NewID() string {\n\tvar p [randomIDEntropyBytes]byte\n\n\tif _, err := io.ReadFull(idReader, p[:]); err != nil {\n\t\tpanic(fmt.Errorf(\"failed to read random bytes: %v\", err))\n\t}\n\n\tvar nn big.Int\n\tnn.SetBytes(p[:])\n\treturn fmt.Sprintf(\"%0[1]*s\", maxRandomIDLength, nn.Text(randomIDBase))\n}", "func NewIdentifierBuilder() IdentifierBuilder {\n\treturn createIdentifierBuilder()\n}", "func newClientID() string {\n\thostname, err := os.Hostname()\n\tif err != nil {\n\t\tip, err := getIP()\n\t\tif err != nil {\n\t\t\tbuffer := make([]byte, 8)\n\t\t\t_, _ = rand.Read(buffer)\n\t\t\thostname = fmt.Sprintf(\"%X\", buffer)\n\n\t\t} else {\n\t\t\thostname = ip.String()\n\t\t}\n\t}\n\ttimestamp := time.Now().UTC().Format(time.RFC3339)\n\t// sarama validation regexp for the client ID doesn't allow ':' characters\n\ttimestamp = strings.Replace(timestamp, \":\", \".\", -1)\n\treturn fmt.Sprintf(\"pixy_%s_%s_%d\", hostname, timestamp, os.Getpid())\n}", "func newULID() ulid.ULID {\n\tt := time.Now()\n\tentropy := rand.New(rand.NewSource(t.UnixNano()))\n\treturn ulid.MustNew(ulid.Timestamp(t), entropy)\n}", "func NewIdentSaver(key string) *CFGoReadLexUnit {\n\treturn NewIdentReader(nil, func(reader snreader.StateNodeReader, stateNode *snreader.StateNode, lex *lex_pgl.LexProduct) error {\n\t\tstateNode.Datas[key] = lex.Value\n\t\treturn nil\n\t}).SetName(\"IdentSaver\")\n}", "func NewID(zone, node int) ID {\n\tif zone < 0 {\n\t\tzone = -zone\n\t}\n\tif node < 0 {\n\t\tnode = -node\n\t}\n\t// return ID(fmt.Sprintf(\"%d.%d\", zone, node))\n\treturn ID(strconv.Itoa(zone) + \".\" + strconv.Itoa(node))\n}", "func NewId(jobType, jobName, jobId string) Id {\n\treturn Id{\n\t\tType: jobType,\n\t\tName: jobName,\n\t\tId: jobId,\n\t}\n}", "func NewId() string {\n\treturn uuid.NewV4().String()\n}", "func newRandomID(n int) string {\n\trand.Seed(UTCNow().UnixNano())\n\tsid := make([]rune, n)\n\tfor i := range sid {\n\t\tsid[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(sid)\n}", "func NewIdentifierSet(identifiers ...ModuleIdentifier) (set ModuleIdentifierSet, err error) {\n\tif len(identifiers) == 0 {\n\t\treturn ModuleIdentifierSet{}, errors.New(\"no identifiers given to create a set from\")\n\t}\n\tfor _, id := range identifiers {\n\t\terr = set.Append(id)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func NewFactory(_ context.Context) (*Factory, error) {\n\tf := &Factory{}\n\treturn f, nil\n}", "func (correlation) New() string {\n\treturn utils.NewUUID()\n}", "func make_id_generator() (func() int) {\n base_id := -1\n return func() int {\n base_id += 1\n return base_id\n }\n}", "func newRandomID(n int) string {\n\trand.Seed(time.Now().UTC().UnixNano())\n\tsid := make([]rune, n)\n\tfor i := range sid {\n\t\tsid[i] = letters[rand.Intn(len(letters))]\n\t}\n\treturn string(sid)\n}", "func NewGameID(str string) *GameID {\n\tgID := GameID{\n\t\tPrefix: PreUnknown,\n\t\tSource: SrcUnknown,\n\t\tID: DefGameID,\n\t}\n\tif len(str) == 0 {\n\t\tfmt.Printf(\"[NewGameID] Invalid input string. \" +\n\t\t\t\"Input string shouldn't be empty.\\n\")\n\t\treturn &gID\n\t}\n\tparts := strings.Split(str, \":\")\n\tif len(parts) != 3 {\n\t\tfmt.Printf(\"[NewGameID] Invalid input string. \"+\n\t\t\t\"There should be exactly 3 parts seperated by ':'. \"+\n\t\t\t\"Instead, found %d parts in %s\\n\", len(parts), str)\n\t\treturn &gID\n\t}\n\tif parts[0] != PreGame {\n\t\tfmt.Printf(\"[NewGameID] Invalid prefix. \"+\n\t\t\t\"Found %s, expected %s\\n\", parts[0], PreGame)\n\t}\n\tgID.Prefix = parts[0]\n\tif !regexp.MustCompile(`^[a-z]{3}$`).MatchString(parts[1]) {\n\t\tfmt.Printf(\"[NewGameID] Invalid source. \"+\n\t\t\t\"Found %s, expected [ABC]\\n\", parts[1])\n\t}\n\tgID.Source = parts[1]\n\tif !regexp.MustCompile(`^[A-Za-z0-9]+$`).MatchString(parts[2]) {\n\t\tfmt.Printf(\"[NewGameID] Invalid ID. \"+\n\t\t\t\"Found %s, expected alphanumeric letters\\n\", parts[2])\n\t}\n\tgID.ID = parts[2]\n\n\treturn &gID\n}", "func newSelector() map[string]string {\n\treturn map[string]string{selectorKey: string(uuid.NewUUID())}\n}", "func NewFactory() extension.Factory {\n\treturn extension.NewFactory(\n\t\tmetadata.Type,\n\t\tcreateDefaultConfig,\n\t\tcreateExtension,\n\t\tmetadata.ExtensionStability,\n\t)\n}", "func NewFactory() extension.Factory {\n\treturn extension.NewFactory(\n\t\tmetadata.Type,\n\t\tcreateDefaultConfig,\n\t\tcreateExtension,\n\t\tmetadata.ExtensionStability,\n\t)\n}", "func NewFactory(config ...FactoryConfig) *Factory {\n\tconf := defaultFactoryConfig()\n\tif len(config) == 1 {\n\t\tconf = config[0]\n\t}\n\n\treturn &Factory{\n\t\tregistry: make(map[string]registryEntry, conf.RegistryCapacity),\n\t}\n}", "func NewID() string {\n\tid := ulid.MustNew(ulid.Now(), entropy)\n\treturn id.String()\n}", "func LangIdNew(filename string) *LangId {\n\tlId := new(LangId)\n\tlId.filename = filename\n\tlId.varNames = varNamesList\n\treturn lId\n}", "func genID(article Article) string {\n\tflake := sonyflake.NewSonyflake(sonyflake.Settings{})\n\tid, err := flake.NextID()\n\tif err != nil {\n\t\tlog.Fatalf(\"NextID() failed with %s\\n\", err)\n\t}\n\tnewID := strconv.FormatUint(id, 10)\n\treturn newID\n}", "func NewId(id string) mgobson.ObjectId { return mgobson.ObjectIdHex(id) }", "func NewFactory() *Factory {\n\treturn &Factory{}\n}", "func NewFactory() *Factory {\n\treturn &Factory{}\n}", "func NewFactory() *Factory {\n\treturn &Factory{}\n}", "func NewFactory() *Factory {\n\treturn &Factory{}\n}", "func NewNumberID(v int64) ID { return ID{number: v} }", "func NewUID(param ...int) (string, error) {\n\tvar size int\n\tif len(param) == 0 {\n\t\tsize = defaultSize\n\t} else {\n\t\tsize = param[0]\n\t}\n\tbytes := make([]byte, size)\n\t_, err := BytesGenerator(bytes)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tid := make([]byte, size)\n\tfor i := 0; i < size; i++ {\n\t\tid[i] = defaultAlphabet[bytes[i]&63]\n\t}\n\treturn string(id[:size]), nil\n}", "func (g *generator) NewId() int {\n\tg.mutex.Lock()\n\tdefer g.mutex.Unlock()\n\tvar id = 0\n\tvar now = time.Now().UnixNano()\n\tid |= int(now) & zeroLowShort\n\tid |= int(g.uniqueKey) << 8\n\tid |= int(g.seqNum)\n\tg.seqNum += 1\n\treturn id\n}", "func (dup *codeDup) newIdent(from string) string {\n\tident := fmt.Sprintf(\"%s$%d\", from, dup.genID())\n\tdup.replacedIdents[from] = ident\n\treturn ident\n}", "func NewID() string {\n\tif i, err := uuid.NewV4(); err == nil {\n\t\treturn i.String()\n\t}\n\treturn \"\"\n}", "func New() string {\n\treturn uuid.NewV4().String()\n}", "func NewFactory(cfg *config.Database, log *logging.Logger) Factory {\n\treturn Factory{\n\t\tcfg: cfg,\n\t\tlog: log,\n\t\tmx: &sync.Mutex{},\n\t}\n}", "func NewID(infix Infix) ID {\n\treturn ID(idgen.Generate(uint16(infix)).String())\n}", "func newContext(wrapped context.Context, identifier int) Context {\n\treturn &contextImpl{wrapped, identifier, false, sync.Mutex{}}\n}", "func NewId() string {\n\tid := uuid.New()\n\tbytes, err := id.MarshalBinary()\n\tif err != nil {\n\t\tpanic(\"UUID encoding error.\")\n\t}\n\n\treturn strings.TrimRight(base64.URLEncoding.EncodeToString(bytes), \"=\")\n}", "func New(sec int64, nsec int, random []byte) UUID {\n\tvar id UUID\n\n\ts := uint32(sec - idEpochBase)\n\tms := uint16(nsec / int(time.Millisecond))\n\n\t// second part\n\tid[0] = byte(s >> 24)\n\tid[1] = byte(s >> 16)\n\tid[2] = byte(s >> 8)\n\tid[3] = byte(s)\n\n\t// millisecond part\n\tid[4] = byte(ms >> 8)\n\tid[5] = byte(ms)\n\n\tcopy(id[6:], random)\n\n\treturn id\n}", "func (ids *IDServiceImpl) NewUUID(clientID string) string {\n\tresult := atomic.AddInt64(&lastID, 1)\n\treturn fmt.Sprintf(\"%v:%v\", clientID, result)\n}", "func New() *Tkeyid {\n\treturn &Tkeyid{keytoid: make(map[string]uint), idtokey: make(map[uint]string)}\n}", "func NewFactory(config FactoryConfig) (*Factory, error) {\n\tf := &Factory{FactoryConfig: config}\n\tuniqueTypes := map[string]struct{}{\n\t\tf.RcdReaderTypes: {},\n\t}\n\n\tfor _, storageType := range f.RcdWriterTypes {\n\t\tuniqueTypes[storageType] = struct{}{}\n\t}\n\n\tf.factories = make(map[string]storage.Factory)\n\n\tfor t := range uniqueTypes {\n\t\tff, err := f.getFactoryOfType(t)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tf.factories[t] = ff\n\t}\n\treturn f, nil\n}", "func NewRemoteIdentifierService(config *config.Config) Service {\n\thost := config.IdentifierServiceHost\n\treturn &remoteIdentifierService{\n\t\tTransportConfig: client.DefaultTransportConfig().WithHost(host),\n\t}\n}", "func New(privateKeyFile string, infoFile string) (*Identity, error) {\n\tidentity := Identity{}\n\n\t// this uses crypto/rand, which is cryptographically secure\n\tpublic, private, err := ed25519.GenerateKey(nil)\n\tif err != nil {\n\t\tmessage := fmt.Sprintf(\"could not generate new Nyz identity: %s\", err.Error())\n\t\treturn nil, errors.New(message)\n\t}\n\tidentity.PrivateKey = private\n\tidentity.PublicKey = public\n\tidentity.addConvenienceDerivatives()\n\n\t// write the pk file\n\terr = utilities.StringToFile(identity.PrivateHex, privateKeyFile)\n\tif err != nil {\n\t\tmessage := fmt.Sprintf(\"error writing new Nyzo identity to file: %s, %s\", privateKeyFile, err.Error())\n\t\treturn nil, errors.New(message)\n\t}\n\n\t// write the info file\n\terr = utilities.StringToFile(identity.NyzoStringPrivate+\"\\n\"+identity.NyzoStringPublic, infoFile)\n\tif err != nil {\n\t\tmessage := fmt.Sprintf(\"error writing new Nyzo identity info to file: %s, %s\", infoFile, err.Error())\n\t\treturn nil, errors.New(message)\n\t}\n\n\treturn &identity, nil\n}" ]
[ "0.795896", "0.78330094", "0.71756345", "0.6960698", "0.690218", "0.68929267", "0.65562814", "0.65105563", "0.64056695", "0.63346016", "0.6318249", "0.6313877", "0.6290458", "0.6251852", "0.62352794", "0.6064076", "0.60508585", "0.60328346", "0.60248464", "0.60030055", "0.5910699", "0.58923143", "0.5885702", "0.58509105", "0.58437896", "0.5712743", "0.5699028", "0.56568754", "0.5647427", "0.56426555", "0.563415", "0.560554", "0.55831265", "0.55419326", "0.5528048", "0.5525521", "0.5522022", "0.55214465", "0.55161935", "0.55089176", "0.55001104", "0.54956365", "0.5486808", "0.5485123", "0.5484764", "0.54806715", "0.5479555", "0.5471678", "0.54613477", "0.5460681", "0.54523337", "0.54492563", "0.5413995", "0.5404086", "0.5400231", "0.53998435", "0.53993434", "0.5398592", "0.5395248", "0.538834", "0.5364885", "0.5338764", "0.533686", "0.53324205", "0.5328347", "0.53259015", "0.53217804", "0.53211427", "0.53210694", "0.53208405", "0.5314832", "0.5311111", "0.53008604", "0.5295572", "0.5295572", "0.52867854", "0.5278055", "0.52693117", "0.5261109", "0.5256039", "0.5253081", "0.5253081", "0.5253081", "0.5253081", "0.52474695", "0.5244638", "0.5243271", "0.5236317", "0.52316415", "0.52002573", "0.5191469", "0.51912075", "0.5187035", "0.51846313", "0.51822865", "0.51784956", "0.51706535", "0.5169953", "0.51674", "0.5157366" ]
0.7809945
2
CreateIdentifier returns a valid Go public identifier
func (factory *identifierFactory) CreateIdentifier(name string, visibility Visibility) string { cacheKey := idCacheKey{name, visibility} factory.rwLock.RLock() cached, ok := factory.idCache[cacheKey] factory.rwLock.RUnlock() if ok { return cached } result := factory.createIdentifierUncached(name, visibility) factory.rwLock.Lock() factory.idCache[cacheKey] = result factory.rwLock.Unlock() return result }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewIdentifier() string {\n\tif !seededrng {\n\t\trand.Seed(time.Now().Unix())\n\t\tseededrng = true\n\t}\n\tr := make([]byte, 9) // 9 bytes * (4/3 encoding) = 12 characters\n\trand.Read(r)\n\ts := base64.URLEncoding.EncodeToString(r)\n\treturn s\n}", "func (factory *identifierFactory) CreateIdentifier(name string, visibility Visibility) string {\n\treturn factory.createIdentifierImpl(name, visibility, avoidReservedWords)\n}", "func NewIdentifier(login string, password string) *Identifier {\n\treturn &Identifier{\n\t\tlogin: login,\n\t\tpassword: password,\n\t\tauthEndpoint: authEndpoint,\n\t\tHTTPClient: &http.Client{Timeout: 10 * time.Second},\n\t}\n}", "func (factory *identifierFactory) createIdentifierImpl(name string, visibility Visibility, reservedWords reservedWordConsideration) string {\n\tcacheKey := idCacheKey{name, visibility, reservedWords}\n\tfactory.rwLock.RLock()\n\tcached, ok := factory.idCache[cacheKey]\n\tfactory.rwLock.RUnlock()\n\tif ok {\n\t\treturn cached\n\t}\n\n\tresult := factory.createIdentifierUncached(name, visibility, reservedWords)\n\tfactory.rwLock.Lock()\n\tresult = factory.internCache.intern(result)\n\tfactory.idCache[cacheKey] = result\n\tfactory.rwLock.Unlock()\n\treturn result\n}", "func (factory *identifierFactory) CreateIdentifier(name string, visibility Visibility) string {\n\tif identifier, ok := factory.renames[name]; ok {\n\t\t// Just lowercase the first character according to visibility\n\t\tr := []rune(identifier)\n\t\tif visibility == NotExported {\n\t\t\tr[0] = unicode.ToLower(r[0])\n\t\t} else {\n\t\t\tr[0] = unicode.ToUpper(r[0])\n\t\t}\n\t\treturn string(r)\n\t}\n\n\t// replace with spaces so titlecasing works nicely\n\tclean := filterRegex.ReplaceAllLiteralString(name, \" \")\n\n\tcleanWords := sliceIntoWords(clean)\n\tvar caseCorrectedWords []string\n\tfor i, word := range cleanWords {\n\t\tif visibility == NotExported && i == 0 {\n\t\t\tcaseCorrectedWords = append(caseCorrectedWords, strings.ToLower(word))\n\t\t} else {\n\t\t\tcaseCorrectedWords = append(caseCorrectedWords, strings.Title(word))\n\t\t}\n\t}\n\n\tresult := strings.Join(caseCorrectedWords, \"\")\n\treturn result\n}", "func (factory *identifierFactory) CreateStringIdentifier(name string, visibility Visibility) string {\n\treturn factory.createIdentifierImpl(name, visibility, allowReservedWords)\n}", "func MakeIdentifier() string {\n\tb := make([]byte, 12)\n\t_, err := io.ReadFull(rand.Reader, b)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn fmt.Sprintf(\"%x\", b)\n}", "func New() string {\n\treturn shortuuid.New()\n}", "func NewID() string {\n\treturn ksuid.New().String()\n}", "func newIdentifier(req messages.Alias, c *models.Customer) (m *models.Identifier, err error) {\n\tm = models.NewIdentifier(c)\n\tm.AliasName = req.AliasName\n\tm.IP, err = newIp(req.Ip)\n\tif err != nil {\n\t\treturn\n\t}\n\tm.Prefix, err = newPrefix(req.Prefix)\n\tif err != nil {\n\t\treturn\n\t}\n\tm.PortRange, err = newPortRange(req.PortRange)\n\tif err != nil {\n\t\treturn\n\t}\n\tm.TrafficProtocol.AddList(req.TrafficProtocol)\n\tm.FQDN.AddList(req.FQDN)\n\tm.URI.AddList(req.URI)\n\n\treturn\n}", "func CreateID(address string, publicKey []byte) ID {\n\treturn ID{PublicKey: publicKey, Address: address}\n}", "func createBundleID(projectName string) string {\n\treturn \"com.elko.\" + projectName\n}", "func GenOpenId() string {\n\t// static $guid = '';\n\t// \t\t$uid = uniqid(\"\", true);\n\t// var data string\n\t// \t\t$data .= $_SERVER['REQUEST_TIME'];\n\t// data = $_SERVER['HTTP_USER_AGENT'];\n\t// data += $_SERVER['SERVER_ADDR'];\n\t// \t\t$data .= $_SERVER['SERVER_PORT'];\n\t// \t\t$data .= $_SERVER['REMOTE_ADDR'];\n\t// \t\t$data .= $_SERVER['REMOTE_PORT'];\n\t// \t\t$hash = strtoupper(hash('ripemd128', $uid . $guid . md5($data)));\n\t// \t\t$guid =\n\t// \t\t\tsubstr($hash, 0, 8) .\n\t// \t\t\tsubstr($hash, 8, 4) .\n\t// \t\t\tsubstr($hash, 12, 4) .\n\t// \t\t\tsubstr($hash, 16, 4) .\n\t// \t\t\tsubstr($hash, 20, 12);\n\n\t// \t\treturn $guid;\n\tvar err error\n\tvar u string = uuid.Must(uuid.NewV4(), err).String()\n\tu = strings.ToUpper(strings.Replace(u, \"-\", \"\", -1))\n\n\treturn u\n}", "func makeCertID(leaf, issuer *x509.Certificate, hashName string) (string, error) {\n\tif leaf == nil {\n\t\treturn \"\", fmt.Errorf(\"leaf certificate is nil\")\n\t}\n\tif issuer == nil {\n\t\treturn \"\", fmt.Errorf(\"issuer certificate is nil\")\n\t}\n\n\tvar hashFunc crypto.Hash\n\tvar oid asn1.ObjectIdentifier\n\n\tswitch hashName {\n\t// The following correlation of hashFunc to OID is copied from a private mapping in golang.org/x/crypto/ocsp:\n\t// https://cs.opensource.google/go/x/crypto/+/refs/tags/v0.8.0:ocsp/ocsp.go;l=156\n\tcase crypto.SHA1.String():\n\t\thashFunc = crypto.SHA1\n\t\toid = asn1.ObjectIdentifier([]int{1, 3, 14, 3, 2, 26})\n\n\tcase crypto.SHA256.String():\n\t\thashFunc = crypto.SHA256\n\t\toid = asn1.ObjectIdentifier([]int{2, 16, 840, 1, 101, 3, 4, 2, 1})\n\n\tcase crypto.SHA384.String():\n\t\thashFunc = crypto.SHA384\n\t\toid = asn1.ObjectIdentifier([]int{2, 16, 840, 1, 101, 3, 4, 2, 2})\n\n\tcase crypto.SHA512.String():\n\t\thashFunc = crypto.SHA512\n\t\toid = asn1.ObjectIdentifier([]int{2, 16, 840, 1, 101, 3, 4, 2, 3})\n\n\tdefault:\n\t\treturn \"\", fmt.Errorf(\"hashName %q is not supported by this package\", hashName)\n\t}\n\n\tif !hashFunc.Available() {\n\t\t// This should never happen.\n\t\treturn \"\", fmt.Errorf(\"hash function %q is not available on your platform\", hashFunc)\n\t}\n\n\tvar spki struct {\n\t\tAlgorithm pkix.AlgorithmIdentifier\n\t\tPublicKey asn1.BitString\n\t}\n\n\t_, err := asn1.Unmarshal(issuer.RawSubjectPublicKeyInfo, &spki)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\th := hashFunc.New()\n\th.Write(spki.PublicKey.RightAlign())\n\tissuerKeyHash := h.Sum(nil)\n\n\th.Reset()\n\th.Write(issuer.RawSubject)\n\tissuerNameHash := h.Sum(nil)\n\n\ttype certID struct {\n\t\tHashAlgorithm pkix.AlgorithmIdentifier\n\t\tIssuerNameHash []byte\n\t\tIssuerKeyHash []byte\n\t\tSerialNumber *big.Int\n\t}\n\n\t// DER-encode the CertID ASN.1 sequence [RFC6960].\n\tcertIDBytes, err := asn1.Marshal(certID{\n\t\tHashAlgorithm: pkix.AlgorithmIdentifier{\n\t\t\tAlgorithm: oid,\n\t\t},\n\t\tIssuerNameHash: issuerNameHash,\n\t\tIssuerKeyHash: issuerKeyHash,\n\t\tSerialNumber: leaf.SerialNumber,\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// base64url-encode [RFC4648] the bytes of the DER-encoded CertID ASN.1 sequence [RFC6960].\n\tencodedBytes := base64.URLEncoding.EncodeToString(certIDBytes)\n\n\t// Any trailing '=' characters MUST be stripped.\n\treturn strings.TrimRight(encodedBytes, \"=\"), nil\n}", "func newClientID() string {\n\thostname, err := os.Hostname()\n\tif err != nil {\n\t\tip, err := getIP()\n\t\tif err != nil {\n\t\t\tbuffer := make([]byte, 8)\n\t\t\t_, _ = rand.Read(buffer)\n\t\t\thostname = fmt.Sprintf(\"%X\", buffer)\n\n\t\t} else {\n\t\t\thostname = ip.String()\n\t\t}\n\t}\n\ttimestamp := time.Now().UTC().Format(time.RFC3339)\n\t// sarama validation regexp for the client ID doesn't allow ':' characters\n\ttimestamp = strings.Replace(timestamp, \":\", \".\", -1)\n\treturn fmt.Sprintf(\"pixy_%s_%s_%d\", hostname, timestamp, os.Getpid())\n}", "func createMyId(localip string) id {\n\tcurTime := time.Now().String()\n\treturn id{localip, curTime}\n}", "func createID(rec *OutputRecord) string {\n\tstr := rec.Type + rec.Path + rec.Datetime + rec.IPAddress + rec.UserID\n\tsum := sha1.Sum([]byte(str))\n\treturn hex.EncodeToString(sum[:])\n}", "func NewID(typ string) string {\n\treturn typ + \"_\" + NewBareID(16)\n}", "func TestIdentifier(t *testing.T) {\n\t// Type as identifier.\n\tvar kvlf KeyValueLessFunc\n\n\tidp := TypeAsIdentifierPart(kvlf)\n\n\tif idp != \"key-value-less-func\" {\n\t\tt.Errorf(\"Identifier part for KeyValueLessFunc is wrong, returned '%v'!\", idp)\n\t}\n\n\tidp = TypeAsIdentifierPart(NewUUID())\n\n\tif idp != \"u-u-i-d\" {\n\t\tt.Errorf(\"Identifier part for UUID is wrong, returned '%v'!\", idp)\n\t}\n\n\t// Identifier.\n\tid := Identifier(\"One\", 2, \"three four\")\n\n\tif id != \"one:2:three-four\" {\n\t\tt.Errorf(\"First identifier is wrong! Id: %v\", id)\n\t}\n\n\tid = Identifier(2011, 6, 22, \"One, two, or three things.\")\n\n\tif id != \"2011:6:22:one-two-or-three-things\" {\n\t\tt.Errorf(\"Second identifier is wrong! Id: %v\", id)\n\t}\n\n\tid = SepIdentifier(\"+\", 1, \"oNe\", 2, \"TWO\", \"3\", \"ÄÖÜ\")\n\n\tif id != \"1+one+2+two+3+äöü\" {\n\t\tt.Errorf(\"Third identifier is wrong! Id: %v\", id)\n\t}\n\n\tid = LimitedSepIdentifier(\"+\", true, \" \", 1, \"oNe\", 2, \"TWO\", \"3\", \"ÄÖÜ\", \"Four\", \"+#-:,\")\n\n\tif id != \"1+one+2+two+3+four\" {\n\t\tt.Errorf(\"Fourth identifier is wrong! Id: %v\", id)\n\t}\n}", "func GetUniqueIdentifier(namespace string, name string, kind string) string {\n\treturn fmt.Sprintf(\"%s/%s/%s\", namespace, kind, name)\n}", "func CreateDIDKey(pubKey []byte) (string, string) {\n\tmethodID := KeyFingerprint(ed25519pub, pubKey)\n\tdidKey := fmt.Sprintf(\"did:key:%s\", methodID)\n\tkeyID := fmt.Sprintf(\"%s#%s\", didKey, methodID)\n\n\treturn didKey, keyID\n}", "func CreateId() string {\n\tuuid.New()\n\n\treturn uuid.New().String()\n}", "func NewId() string {\n\t// generate 128 random bits (6 more than standard UUID)\n\tbytes := make([]byte, 16)\n\t_, err := rand.Read(bytes)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t// convert them to base 32 encoding\n\ts := base32.StdEncoding.EncodeToString(bytes)\n\treturn strings.ToLower(strings.TrimRight(s, \"=\"))\n}", "func NewIdentifier(opts ...Option) Profile {\n\treturn Profile{\n\t\toptions: getOpts(opts...),\n\t\tclass: identifier,\n\t}\n}", "func generateUuidForIdent() string {\n\tkey, _ := uuid.NewUUID()\n\tkeyString := strings.ReplaceAll(key.String(), \"-\", \"_\")\n\treturn \"generated_ident__\" + keyString\n}", "func (factory *identifierFactory) CreateLocal(name string) string {\n\treturn factory.CreateIdentifier(name, NotExported)\n}", "func makeID() (string, error) {\n\tdata := make([]byte, 32)\n\t_, err := rand.Read(data)\n\tx := sha256.Sum256(data)\n\treturn hex.EncodeToString(x[:]), err\n}", "func New() string {\n\tvar result string\n\tbackoff.RetryNotify(func() error { //nolint:errcheck\n\t\tuuid, err := uuid.NewV4()\n\t\tif err != nil {\n\t\t\treturn errors.EnsureStack(err)\n\t\t}\n\t\tresult = uuid.String()\n\t\treturn nil\n\t}, backoff.NewInfiniteBackOff(), func(err error, d time.Duration) error {\n\t\tfmt.Printf(\"error from uuid.NewV4: %v\", err)\n\t\treturn nil\n\t})\n\treturn result\n}", "func newID() string {\n\treturn \"_\" + uuid.New().String()\n}", "func VolumeIdentifierCreate(service, vid string) string {\n\treturn fmt.Sprintf(\"%s:%s\", service, vid)\n}", "func generateID(port int) string {\n\tportInByte := new([]byte)\n\thash := sha256.New()\n\thash.Write(append(*portInByte, byte(port)))\n\treturn hex.EncodeToString(hash.Sum(*portInByte)[:20])\n}", "func (sd *PrivateDescriptor) IdentifierString() string {\n\tb := make([]byte, 4)\n\tbinary.BigEndian.PutUint32(b, sd.Identifier)\n\treturn string(b)\n}", "func NewID(length int) string {\n\tb := make([]byte, length*6/8)\n\t_, err := rand.Read(b)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn base64.RawURLEncoding.EncodeToString(b)\n}", "func CreateUserIDToken(key []byte, userID string) (string, error) {\n\tnow := time.Now()\n\texp := now.Add(90 * 24 * time.Hour)\n\n\tjsonToken := paseto.JSONToken{\n\t\tAudience: \"recruitment.empirica.app\",\n\t\tIssuer: \"recruitment.empirica.app\",\n\t\tJti: xid.New().String(),\n\t\tSubject: userID,\n\t\tIssuedAt: now,\n\t\tExpiration: exp,\n\t\tNotBefore: now,\n\t}\n\n\ttoken, err := paseto.Encrypt(key, jsonToken, \"\")\n\t// token = \"v2.local.E42A2iMY9SaZVzt-WkCi45_aebky4vbSUJsfG45OcanamwXwieieMjSjUkgsyZzlbYt82miN1xD-X0zEIhLK_RhWUPLZc9nC0shmkkkHS5Exj2zTpdNWhrC5KJRyUrI0cupc5qrctuREFLAvdCgwZBjh1QSgBX74V631fzl1IErGBgnt2LV1aij5W3hw9cXv4gtm_jSwsfee9HZcCE0sgUgAvklJCDO__8v_fTY7i_Regp5ZPa7h0X0m3yf0n4OXY9PRplunUpD9uEsXJ_MTF5gSFR3qE29eCHbJtRt0FFl81x-GCsQ9H9701TzEjGehCC6Bhw.c29tZSBmb290ZXI\"\n\n\treturn token, err\n}", "func NewStringID(v string) ID { return ID{name: v} }", "func createNSName(prefix string) string {\n\tsuffix := make([]byte, 10)\n\trand.Read(suffix)\n\treturn fmt.Sprintf(\"%s-%x\", prefix, suffix)\n}", "func (this *IdBuilder) NewId() (rv string) {\n\tarr := [24]byte{}\n\ts := arr[:]\n\n\t//\n\t// base32 encode all 64 bits of counter\n\t// - encode 5 bits at a time allows us to use base62 routine to get base32\n\t// - we stop at 10 because that's where date/time begins\n\t//\n\tu := atomic.AddUint64(&this.counter, 1)\n\tvar shift uint = 0\n\tfor i := len(arr) - 1; i > 10; i-- {\n\t\tb := byte(u >> shift)\n\t\tshift += 5\n\t\ts[i] = base62(b & 0x1f)\n\t}\n\n\t//\n\t// encode YMMDDHHMMSS\n\t//\n\t// we need the year as these can get databased\n\t//\n\t// we want the MMDDHHMMSS to be easily human readable\n\t//\n\tnow := time.Now().UTC()\n\tyear, month, day := now.Date()\n\thour, minute, second := now.Clock()\n\ts[0] = base62(byte(year - 2016))\n\ts[1], s[2] = base10(byte(month))\n\ts[3], s[4] = base10(byte(day))\n\ts[5], s[6] = base10(byte(hour))\n\ts[7], s[8] = base10(byte(minute))\n\ts[9], s[10] = base10(byte(second))\n\n\treturn string(s)\n}", "func genSID() string {\n\treturn fmt.Sprintf(\"s6a_proxy;%d_%X\", uint(time.Now().Unix()), rand.Uint32())\n}", "func newID() string {\n\tvar b [8]byte\n\t_, err := rand.Read(b[:])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn fmt.Sprintf(\"%x\", b[:])\n}", "func NewID(b []byte) (*SHA1, error) {\n\tif len(b) != 20 {\n\t\treturn nil, errors.New(\"length must be 20\")\n\t}\n\treturn MustID(b), nil\n}", "func NewObjectIdentifier(oid string) (oi asn1.ObjectIdentifier, err error) {\n\tif len(oid) == 0 {\n\t\treturn nil, errors.Errorf(\"zero length OBJECT IDENTIFIER\")\n\t}\n\n\tif oid[0] == '{' {\n\t\t// ASN.1 notation. (eg {iso(1) member-body(2) us(840) rsadsi(113549) pkcs(1) pkcs-9(9) messageDigest(4)})\n\t\tparts := strings.Split(oid[1:len(oid)-1], \" \")\n\t\toi = make(asn1.ObjectIdentifier, len(parts), len(parts))\n\t\tfor i, part := range parts {\n\t\t\tidx := strings.IndexRune(part, '(')\n\t\t\toi[i], err = strconv.Atoi(part[idx+1 : len(part)-1])\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t} else {\n\t\t// Dot notation. (eg 1.2.840.113549.1.9.4)\n\t\tparts := strings.Split(oid, \".\")\n\t\toi = make(asn1.ObjectIdentifier, len(parts), len(parts))\n\t\tfor i, part := range parts {\n\t\t\toi[i], err = strconv.Atoi(part)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\treturn oi, nil\n}", "func NewIdentifier(t token.Token) *Identifier {\n\treturn &Identifier{\n\t\tToken: t,\n\t\tValue: t.Literal,\n\t}\n}", "func NewIDFromString(s string) (*SHA1, error) {\n\ts = strings.TrimSpace(s)\n\tif len(s) != 40 {\n\t\treturn nil, errors.New(\"length must be 40\")\n\t}\n\tb, err := hex.DecodeString(s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn NewID(b)\n}", "func NewID() string {\n\treturn idFunc()\n}", "func NewUID(param ...int) (string, error) {\n\tvar size int\n\tif len(param) == 0 {\n\t\tsize = defaultSize\n\t} else {\n\t\tsize = param[0]\n\t}\n\tbytes := make([]byte, size)\n\t_, err := BytesGenerator(bytes)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tid := make([]byte, size)\n\tfor i := 0; i < size; i++ {\n\t\tid[i] = defaultAlphabet[bytes[i]&63]\n\t}\n\treturn string(id[:size]), nil\n}", "func TestNewId(t *testing.T) {\n\thexChars := \"0123456789abcdefABCDEF\"\n\tid := string(NewID())\n\n\t// if the generated ID is not started with 0x, which is the prefix added to the ID\n\tif !strings.HasPrefix(id, \"0x\") {\n\t\tt.Fatalf(\"Invalid id prefix, got id: %s, expected id started with 0x\",\n\t\t\tid)\n\t}\n\n\tid = id[2:]\n\n\t// check the id length\n\tif len(id) == 0 || len(id) > 32 {\n\t\tt.Fatalf(\"Invalid id length, got: %d, expected 32\", len(id))\n\t}\n\n\t// check id's characters\n\tfor i := 0; i < len(id); i++ {\n\t\t// check if the byte can be found within hexChars\n\t\tif strings.IndexByte(hexChars, id[i]) == -1 {\n\t\t\tt.Fatalf(\"unexpected byte, got %c, expected valid hex character\", id[i])\n\t\t}\n\t}\n}", "func NewIdentifier(opts ...Option) *Profile {\n\treturn &Profile{\n\t\toptions: getOpts(opts...),\n\t\tclass: identifier,\n\t}\n}", "func generateUniqueName() (string, error) {\n\tid, err := getUniqueID()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn intToMultiBaseString(id, digitSpec), nil\n}", "func New(cacheNodeidentityInfo bool) (nodeidentity.Identifier, error) {\n\tclient, err := newClient()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &nodeIdentifier{\n\t\tvmssGetter: client,\n\t\tcache: expirationcache.NewTTLStore(stringKeyFunc, cacheTTL),\n\t\tcacheEnabled: cacheNodeidentityInfo,\n\t}, nil\n}", "func RewriteAsIdentifier(\n\tk8sVersion string, rawID fmt.Stringer,\n) Identifier {\n\tvar id = rawID.String()\n\n\tif len(id) == 0 {\n\t\tlog.Fatalf(\"Can't lowercase first letter of 0-rune string\")\n\t}\n\tkindString := kubeversion.MapIdentifier(k8sVersion, id)\n\n\tupper := strings.ToLower(kindString[:1])\n\treturn Identifier(upper + kindString[1:])\n}", "func NewObjectId() string {\n\tvar b [12]byte\n\t// Timestamp, 4 bytes, big endian\n\tbinary.BigEndian.PutUint32(b[:], uint32(time.Now().Unix()))\n\t// Machine, first 3 bytes of md5(hostname)\n\tb[4] = machineId[0]\n\tb[5] = machineId[1]\n\tb[6] = machineId[2]\n\t// Pid, 2 bytes, specs don't specify endianness, but we use big endian.\n\tb[7] = byte(processId >> 8)\n\tb[8] = byte(processId)\n\t// Increment, 3 bytes, big endian\n\ti := atomic.AddUint32(&objectIdCounter, 1)\n\tb[9] = byte(i >> 16)\n\tb[10] = byte(i >> 8)\n\tb[11] = byte(i)\n\treturn hex.EncodeToString(b[:])\n}", "func NewID() string {\n\tif i, err := uuid.NewV4(); err == nil {\n\t\treturn i.String()\n\t}\n\treturn \"\"\n}", "func New() string {\n\treturn uuid.NewV4().String()\n}", "func New() GID {\n\tvar id [12]byte\n\n\t// Timestamp, 4 bytes, big endian\n\tbinary.BigEndian.PutUint32(id[:], uint32(time.Now().Unix()))\n\n\t// Machine, first 3 bytes of md5(hostname)\n\tid[4] = macID[0]\n\tid[5] = macID[1]\n\tid[6] = macID[2]\n\n\t// Pid, 2 bytes, specs don't specify endianness, but we use big endian.\n\tpid := os.Getpid()\n\tid[7] = byte(pid >> 8)\n\tid[8] = byte(pid)\n\n\t// Increment, 3 bytes, big endian\n\ti := atomic.AddUint32(&gidCounter, 1)\n\tid[9] = byte(i >> 16)\n\tid[10] = byte(i >> 8)\n\tid[11] = byte(i)\n\n\treturn GID(id[:])\n}", "func (i *InstanceIdentity) GetIdentifier() string {\n\tident := strings.Join([]string{i.AccountID, i.InstanceID}, \"-\")\n\th := sha1.New()\n\th.Write([]byte(ident))\n\tbid := h.Sum(nil)\n\treturn fmt.Sprintf(\"%x\", bid)\n}", "func generateID() string {\n\treturn stringid.GenerateNonCryptoID()\n}", "func id(s string) string {\n\treturn uuid.NewV5(namespaceUUID, s).String()\n}", "func makeId(allowedTo []common.Endpoint, name string) string {\n\tvar data string\n\tdata = name\n\n\tfor _, e := range allowedTo {\n\t\tif data == \"\" {\n\t\t\tdata = fmt.Sprintf(\"%s\", e)\n\t\t} else {\n\t\t\tdata = fmt.Sprintf(\"%s\\n%s\", data, e)\n\t\t}\n\t}\n\n\thasher := sha1.New()\n\thasher.Write([]byte(data))\n\tsum := hasher.Sum(nil)\n\n\t// Taking 6 bytes of a hash which is 12 chars length\n\treturn fmt.Sprint(hex.EncodeToString(sum[:6]))\n}", "func newUID() ([]byte, error) {\n\t// uuid := make([]byte, 16)\n\t// n, err := io.ReadFull(rand.Reader, uuid)\n\t// if n != len(uuid) || err != nil {\n\t// \treturn nil, err\n\t// }\n\t// // variant bits; see section 4.1.1\n\t// uuid[8] = uuid[8]&^0xc0 | 0x80\n\t// // version 4 (pseudo-random); see section 4.1.3\n\t// uuid[6] = uuid[6]&^0xf0 | 0x40\n\t// return []byte(fmt.Sprintf(\"%x-%x-%x-%x-%x\", uuid[0:4], uuid[4:6], uuid[6:8], uuid[8:10], uuid[10:])), nil\n\treturn []byte(uniuri.New()), nil\n}", "func New() ID {\n\tvar id ID\n\t// Timestamp, 4 bytes, big endian\n\tbinary.BigEndian.PutUint32(id[:], uint32(time.Now().Unix()))\n\t// Machine, first 3 bytes of md5(hostname)\n\tid[4] = machineID[0]\n\tid[5] = machineID[1]\n\tid[6] = machineID[2]\n\t// Pid, 2 bytes, specs don't specify endianness, but we use big endian.\n\tpid := os.Getpid()\n\tid[7] = byte(pid >> 8)\n\tid[8] = byte(pid)\n\t// Increment, 3 bytes, big endian\n\ti := atomic.AddUint32(&objectIDCounter, 1)\n\tid[9] = byte(i >> 16)\n\tid[10] = byte(i >> 8)\n\tid[11] = byte(i)\n\treturn id\n}", "func makeValidIdentifier(name string) string {\n\tvar builder strings.Builder\n\tfor i, c := range name {\n\t\tif i == 0 && c == '@' {\n\t\t\tbuilder.WriteRune(c)\n\t\t\tcontinue\n\t\t}\n\t\tif !isLegalIdentifierPart(c) {\n\t\t\tbuilder.WriteRune('_')\n\t\t} else {\n\t\t\tif i == 0 && !isLegalIdentifierStart(c) {\n\t\t\t\tbuilder.WriteRune('_')\n\t\t\t}\n\t\t\tbuilder.WriteRune(c)\n\t\t}\n\t}\n\tname = builder.String()\n\tif isReservedWord(name) {\n\t\treturn \"@\" + name\n\t}\n\treturn name\n}", "func (o *SparseOAUTHKey) Identifier() string {\n\n\treturn \"\"\n}", "func CreateSessionID(sessionKey string) (string, error) {\n\tif len(sessionKey) < 1 {\n\t\treturn \"\", errors.New(\"signing key must not be empty\")\n\t}\n\n\ttoken := make([]byte, 0)\n\n\tsalt := make([]byte, 32)\n\t_, err := rand.Read(salt)\n\tif err != nil {\n\t\tfmt.Printf(\"error generating salt: %v\\n\", err)\n\t\tos.Exit(1)\n\t}\n\n\thash := hmac.New(sha256.New, []byte(sessionKey))\n\thash.Write(salt)\n\tsignature := hash.Sum(nil)\n\n\tfor _, byte := range salt {\n\t\ttoken = append(token, byte)\n\t}\n\n\tfor _, byte := range signature {\n\t\ttoken = append(token, byte)\n\t}\n\n\tres := base64.URLEncoding.EncodeToString(token)\n\n\treturn res, nil\n}", "func New(localpart, domainpart, resourcepart string) (JID, error) {\n\t// Ensure that parts are valid UTF-8 (and short circuit the rest of the\n\t// process if they're not).\n\t// The domainpart is checked in normalizeDomainpart.\n\tif !utf8.ValidString(localpart) || !utf8.ValidString(resourcepart) {\n\t\treturn JID{}, errInvalidUTF8\n\t}\n\n\tvar err error\n\tdomainpart, err = normalizeDomainpart(domainpart)\n\tif err != nil {\n\t\treturn JID{}, err\n\t}\n\n\tvar lenlocal int\n\tdata := make([]byte, 0, len(localpart)+len(domainpart)+len(resourcepart))\n\n\tif localpart != \"\" {\n\t\tdata, err = precis.UsernameCaseMapped.Append(data, []byte(localpart))\n\t\tif err != nil {\n\t\t\treturn JID{}, err\n\t\t}\n\t\tlenlocal = len(data)\n\t}\n\n\tdata = append(data, []byte(domainpart)...)\n\n\tif resourcepart != \"\" {\n\t\tdata, err = precis.OpaqueString.Append(data, []byte(resourcepart))\n\t\tif err != nil {\n\t\t\treturn JID{}, err\n\t\t}\n\t}\n\n\terr = localChecks(data[:lenlocal])\n\tif err != nil {\n\t\treturn JID{}, err\n\t}\n\n\terr = resourceChecks(data[lenlocal+len(domainpart):])\n\tif err != nil {\n\t\treturn JID{}, err\n\t}\n\n\treturn JID{\n\t\tlocallen: lenlocal,\n\t\tdomainlen: len(domainpart),\n\t\tdata: data,\n\t}, nil\n}", "func createHeadingID(headingName string) (id string, err error) {\n\tif headingName == \"\" {\n\t\treturn \"\", fmt.Errorf(\"need heading name\")\n\t}\n\n\t// Munge the original heading into an id by:\n\t//\n\t// - removing invalid characters.\n\t// - lower-casing.\n\t// - replace spaces\n\tid = strings.Map(validHeadingIDChar, headingName)\n\n\tid = strings.ToLower(id)\n\tid = strings.Replace(id, \" \", \"-\", -1)\n\n\treturn id, nil\n}", "func Generate() (string, error) {\n\treturn shortid.Generate()\n}", "func newid() string {\n\treturn make_key(6)\n}", "func MakeOID(hexStr []byte) (OID, error) {\n\tvar oid OID\n\t// this is twice as much length as we need,\n\t// because hexStr in string encoding takes twice as much space as raw bytes\n\t// but let it be this way for readability\n\tdecoded := make([]byte, len(hexStr))\n\tn, err := hex.Decode(decoded, hexStr)\n\tif err != nil {\n\t\treturn ZeroOID, fmt.Errorf(\"makeOID: %w\", err)\n\t}\n\tif n != sha1.Size {\n\t\treturn ZeroOID, fmt.Errorf(\"makeOID: invalid length (%d), expected %d, oid: %s\", n, sha1.Size, hexStr)\n\t}\n\tcopy(oid[:], decoded)\n\treturn oid, nil\n}", "func NewID(v string) ID {\n\treturn ID{v, true}\n}", "func CreateResourceID(bucket, expectedBucketOwner string) string {\n\tif expectedBucketOwner == \"\" {\n\t\treturn bucket\n\t}\n\n\tparts := []string{bucket, expectedBucketOwner}\n\tid := strings.Join(parts, resourceIDSeparator)\n\n\treturn id\n}", "func NewId() string {\n\tid := uuid.New()\n\tbytes, err := id.MarshalBinary()\n\tif err != nil {\n\t\tpanic(\"UUID encoding error.\")\n\t}\n\n\treturn strings.TrimRight(base64.URLEncoding.EncodeToString(bytes), \"=\")\n}", "func newGoIdent(f *File, d protoreflect.Descriptor) GoIdent {\n\tname := strings.TrimPrefix(string(d.FullName()), string(f.Desc.Package())+\".\")\n\treturn GoIdent{\n\t\tGoName: strs.GoCamelCase(name),\n\t\tGoImportPath: f.GoImportPath,\n\t}\n}", "func GetGoID() int64", "func (s *session) newID() (interface{}, error) {\n\tvar b [32]byte\n\t_, err := rand.Read(b[:])\n\treturn hex.EncodeToString(b[:]), err\n}", "func NewGameID(str string) *GameID {\n\tgID := GameID{\n\t\tPrefix: PreUnknown,\n\t\tSource: SrcUnknown,\n\t\tID: DefGameID,\n\t}\n\tif len(str) == 0 {\n\t\tfmt.Printf(\"[NewGameID] Invalid input string. \" +\n\t\t\t\"Input string shouldn't be empty.\\n\")\n\t\treturn &gID\n\t}\n\tparts := strings.Split(str, \":\")\n\tif len(parts) != 3 {\n\t\tfmt.Printf(\"[NewGameID] Invalid input string. \"+\n\t\t\t\"There should be exactly 3 parts seperated by ':'. \"+\n\t\t\t\"Instead, found %d parts in %s\\n\", len(parts), str)\n\t\treturn &gID\n\t}\n\tif parts[0] != PreGame {\n\t\tfmt.Printf(\"[NewGameID] Invalid prefix. \"+\n\t\t\t\"Found %s, expected %s\\n\", parts[0], PreGame)\n\t}\n\tgID.Prefix = parts[0]\n\tif !regexp.MustCompile(`^[a-z]{3}$`).MatchString(parts[1]) {\n\t\tfmt.Printf(\"[NewGameID] Invalid source. \"+\n\t\t\t\"Found %s, expected [ABC]\\n\", parts[1])\n\t}\n\tgID.Source = parts[1]\n\tif !regexp.MustCompile(`^[A-Za-z0-9]+$`).MatchString(parts[2]) {\n\t\tfmt.Printf(\"[NewGameID] Invalid ID. \"+\n\t\t\t\"Found %s, expected alphanumeric letters\\n\", parts[2])\n\t}\n\tgID.ID = parts[2]\n\n\treturn &gID\n}", "func createKey(a, b string) string {\n\treturn a + \":\" + b\n}", "func isValidIdentifier(name string) bool {\n\treturn identifierRE.MatchString(name)\n}", "func generateVolumeID(volName string) string {\n\t//uuid := uuid.New()\n\t//return fmt.Sprintf(\"volid-%s\", uuid.String())\n\treturn volName\n}", "func NewID(size int) string {\n\tb := make([]byte, size)\n\tfor i := range b {\n\t\tb[i] = ABC[rand.Intn(len(ABC))]\n\t}\n\n\treturn string(b)\n}", "func (us *UserStorage) IDByName(name string) (string, error) {\n\treturn randomdata.StringNumber(2, \"-\"), nil\n}", "func NewInstanceID() string {\n\tg := uuid.Must(uuid.NewV4())\n\tp := strings.Split(g.String(), \"-\")\n\treturn p[len(p)-1]\n}", "func IdFromName(nm string) uint8 {\n\tswitch nm {\n\tcase \"gob\":\n\t\treturn GOB\n\tcase \"zipgob\":\n\t\treturn ZIPGOB\n\tcase \"json\":\n\t\treturn JSON\n\tcase \"protobuf\":\n\t\treturn PROTOBUF\n\tcase \"gorilla\":\n\t\treturn GORILLA\n\tcase \"msgpack\":\n\t\treturn MSGPACK\n\tcase \"binc\":\n\t\treturn BINC\n\tcase \"cbor\":\n\t\treturn CBOR\n\tcase \"repr\":\n\t\treturn REPR\n\tdefault:\n\t\treturn 0\n\t}\n}", "func (o *OAUTHKey) Identifier() string {\n\n\treturn \"\"\n}", "func NewID() string {\n\tid := ulid.MustNew(ulid.Now(), entropy)\n\treturn id.String()\n}", "func NameID(name string) (nm string, err error) {\n\ts, err := binary.Blake2bHash([]byte(name))\n\tif err != nil {\n\t\treturn\n\t}\n\treturn binary.Encode(binary.PrefixName, s), nil\n}", "func (m *CreateIdentifier) String() (result string) {\n\tresult = \"\\n\"\n\tfor key, alias := range m.Identifier.Alias {\n\t\tresult += fmt.Sprintf(\" \\\"%s[%d]\\\":\\n\", \"alias\", key+1)\n\t\tresult += fmt.Sprintf(\" \\\"%s\\\": %s\\n\", \"alias-name\", alias.AliasName)\n\t\tif alias.Ip != nil {\n\t\t\tfor k, v := range alias.Ip {\n\t\t\t\tresult += fmt.Sprintf(\" \\\"%s[%d]\\\": %s\\n\", \"ip\", k+1, v)\n\t\t\t}\n\t\t}\n\t\tif alias.Prefix != nil {\n\t\t\tfor k, v := range alias.Prefix {\n\t\t\t\tresult += fmt.Sprintf(\" \\\"%s[%d]\\\": %s\\n\", \"prefix\", k+1, v)\n\t\t\t}\n\t\t}\n\t\tif alias.PortRange != nil {\n\t\t\tfor k, v := range alias.PortRange {\n\t\t\t\tresult += fmt.Sprintf(\" \\\"%s[%d]\\\":\\n\", \"port-range\", k+1)\n\t\t\t\tresult += fmt.Sprintf(\" \\\"%s\\\": %d\\n\", \"lower-port\", v.LowerPort)\n\t\t\t\tresult += fmt.Sprintf(\" \\\"%s\\\": %d\\n\", \"upper-port\", v.UpperPort)\n\t\t\t}\n\t\t}\n\t\tif alias.FQDN != nil {\n\t\t\tfor k, v := range alias.FQDN {\n\t\t\t\tresult += fmt.Sprintf(\" \\\"%s[%d]\\\": %s\\n\", \"FQDN\", k+1, v)\n\t\t\t}\n\t\t}\n\t\tif alias.URI != nil {\n\t\t\tfor k, v := range alias.URI {\n\t\t\t\tresult += fmt.Sprintf(\" \\\"%s[%d]\\\": %s\\n\", \"URI\", k+1, v)\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func CreateDIDKey(pubKey []byte) (string, string) {\n\treturn fingerprint.CreateDIDKey(pubKey)\n}", "func GenerateID() string {\n return uuid.New().String()\n}", "func CreateSessionID() string {\n\tsess := strconv.FormatInt(time.Now().UTC().UnixNano(), 36)\n\tif len(sess) == 12 {\n\t\treturn sess\n\t} else if len(sess) < 12 {\n\t\treturn (sess + \"000\")[:12]\n\t} else {\n\t\treturn sess[:12]\n\t}\n}", "func (f *testIDService) New() (string, error) {\n\treturn \"new-ID\", nil\n}", "func New(worker uint8, alphabet string, seed uint64) (*Shortid, error) {\n\tif worker > 31 {\n\t\treturn nil, errors.New(\"expected worker in the range [0,31]\")\n\t}\n\tabc, err := NewAbc(alphabet, seed)\n\tif err == nil {\n\t\tsid := &Shortid{\n\t\t\tabc: abc,\n\t\t\tworker: uint(worker),\n\t\t\tepoch: time.Date(2016, time.January, 1, 0, 0, 0, 0, time.UTC),\n\t\t\tms: 0,\n\t\t\tcount: 0,\n\t\t}\n\t\treturn sid, nil\n\t}\n\treturn nil, err\n}", "func Create(userID string) (string, error) {\n\tt := jwt.NewWithClaims(jwt.SigningMethodHS256, claims(userID))\n\treturn t.SignedString(secret)\n}", "func NewID() string {\n\treturn replacer.Replace(NewV4().String())\n}", "func NewId() string {\n\treturn uuid.NewV4().String()\n}", "func (*dynamicUUIDProvider) New() string {\n\treturn uuid.New()\n}", "func createJWTID(id uint64) string {\n\tstr := fmt.Sprintf(\"%v-%v\", id, time.Now().UnixNano())\n\tbytes, _ := bcrypt.GenerateFromPassword([]byte(str), 12)\n\treturn string(bytes)\n}", "func (o *SparseSSHAuthorizationPolicy) Identifier() string {\n\n\tif o.ID == nil {\n\t\treturn \"\"\n\t}\n\treturn *o.ID\n}", "func NewID() string {\n\tid, err := uuid.NewUUID()\n\tif err != nil {\n\t\tlog.Fatalf(\"Error while getting id: %v\\n\", err)\n\t}\n\treturn fmt.Sprintf(\"id-%s\", id.String())\n}", "func NewId(id string) mgobson.ObjectId { return mgobson.ObjectIdHex(id) }", "func NewID() string {\n\tu2 := uuid.NewV4()\n\treturn u2.String()\n}" ]
[ "0.72283757", "0.69594693", "0.67079806", "0.6524352", "0.6468972", "0.63878125", "0.63809896", "0.63100255", "0.6242566", "0.6235022", "0.616394", "0.6156025", "0.6136713", "0.6058159", "0.6051575", "0.60508806", "0.60451144", "0.60405385", "0.60357136", "0.6015317", "0.59919757", "0.5936397", "0.5903236", "0.5899814", "0.58986676", "0.5861901", "0.58525306", "0.58439815", "0.5823754", "0.5811652", "0.5773183", "0.5763636", "0.57565075", "0.57524854", "0.57468814", "0.57349235", "0.57316315", "0.5731377", "0.57255924", "0.5722045", "0.5720517", "0.57175815", "0.5695172", "0.56825286", "0.568224", "0.5663173", "0.56313944", "0.5629499", "0.562799", "0.5624363", "0.5612611", "0.5603164", "0.55960506", "0.55940336", "0.5591846", "0.55801153", "0.5578306", "0.5576951", "0.556523", "0.55627435", "0.5562604", "0.5558326", "0.5555124", "0.5538486", "0.5520739", "0.5511781", "0.5507358", "0.5506454", "0.550308", "0.55018383", "0.54922366", "0.5474362", "0.5462485", "0.5447823", "0.54400027", "0.543749", "0.5428906", "0.5424846", "0.5399486", "0.5395977", "0.539307", "0.53750575", "0.5373492", "0.5358211", "0.53548926", "0.5352635", "0.5341585", "0.5339372", "0.5334351", "0.5332505", "0.5329921", "0.53116524", "0.53035617", "0.5300641", "0.5300568", "0.530017", "0.52984583", "0.52925545", "0.52923423", "0.52914184" ]
0.6858647
2
These are words reserved by go, along with our chosen substitutes
func createReservedWords() map[string]string { return map[string]string{ "break": "brk", "case": "c", "chan": "chn", "const": "cnst", "continue": "cont", "default": "def", "defer": "deferVar", "else": "els", "fallthrough": "fallthrgh", "for": "f", "func": "funcVar", "go": "g", "goto": "gotoVar", "if": "ifVar", "import": "imp", "interface": "iface", "map": "m", "package": "pkg", "range": "rng", "return": "ret", "select": "sel", "struct": "strct", "switch": "sw", "type": "typeVar", "var": "v", } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func isReservedWord(s string) bool {\n\tswitch s {\n\tcase \"abstract\", \"as\", \"base\", \"bool\", \"break\", \"byte\", \"case\", \"catch\", \"char\", \"checked\", \"class\", \"const\",\n\t\t\"continue\", \"decimal\", \"default\", \"delegate\", \"do\", \"double\", \"else\", \"enum\", \"event\", \"explicit\", \"extern\",\n\t\t\"false\", \"finally\", \"fixed\", \"float\", \"for\", \"foreach\", \"goto\", \"if\", \"implicit\", \"in\", \"int\", \"interface\",\n\t\t\"internal\", \"is\", \"lock\", \"long\", \"namespace\", \"new\", \"null\", \"object\", \"operator\", \"out\", \"override\",\n\t\t\"params\", \"private\", \"protected\", \"public\", \"readonly\", \"ref\", \"return\", \"sbyte\", \"sealed\", \"short\",\n\t\t\"sizeof\", \"stackalloc\", \"static\", \"string\", \"struct\", \"switch\", \"this\", \"throw\", \"true\", \"try\", \"typeof\",\n\t\t\"uint\", \"ulong\", \"unchecked\", \"unsafe\", \"ushort\", \"using\", \"virtual\", \"void\", \"volatile\", \"while\":\n\t\treturn true\n\t// Treat contextual keywords as keywords, as we don't validate the context around them.\n\tcase \"add\", \"alias\", \"ascending\", \"async\", \"await\", \"by\", \"descending\", \"dynamic\", \"equals\", \"from\", \"get\",\n\t\t\"global\", \"group\", \"into\", \"join\", \"let\", \"nameof\", \"on\", \"orderby\", \"partial\", \"remove\", \"select\", \"set\",\n\t\t\"unmanaged\", \"value\", \"var\", \"when\", \"where\", \"yield\":\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func reservedWord(ident string, mode reservedMode) bool {\n\treturn reservedWordJava(ident, mode) ||\n\t\treservedWordJavascript(ident, mode) ||\n\t\treservedWordGo(ident)\n\t// TODO(bprosnitz) Other identifiers? (set, assert, raise, with, etc)\n}", "func checkReservedWords(field string) bool {\n\treservedWordsSet := []string{\"abstract\", \"and\", \"arguments\", \"as\", \"assert\", \"async\", \"await\", \"boolean\", \"break\", \"byte\",\n\t\t\"case\", \"catch\", \"char\", \"class\", \"const\", \"continue\", \"debugger\", \"def\", \"default\", \"del\", \"delete\", \"do\", \"double\", \"elif\",\n\t\t\"else\", \"enum\", \"eval\", \"except\", \"export\", \"extends\", \"false\", \"final\", \"finally\", \"float\", \"for\", \"from\", \"function\", \"global\",\n\t\t\"goto\", \"if\", \"implements\", \"import\", \"in\", \"instanceof\", \"int\", \"interface\", \"is\", \"lambda\", \"let\", \"long\", \"native\", \"new\", \"nonlocal\",\n\t\t\"not\", \"null\", \"or\", \"package\", \"pass\", \"private\", \"protected\", \"public\", \"raise\", \"return\", \"short\", \"static\", \"strictfp\",\n\t\t\"super\", \"switch\", \"synchronized\", \"this\", \"throw\", \"throws\", \"transient\", \"true\", \"try\", \"typeof\", \"var\", \"void\", \"volatile\",\n\t\t\"while\", \"with\", \"yield\"}\n\n\tfor _, segment := range strings.Split(field, \"_\") {\n\t\tresult := sort.SearchStrings(reservedWordsSet, segment)\n\t\tif result < len(reservedWordsSet) && reservedWordsSet[result] == segment {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s *BasePlSqlParserListener) EnterNon_reserved_keywords_in_12c(ctx *Non_reserved_keywords_in_12cContext) {\n}", "func nameAllowed(name string) error {\n\tfor _, reservedName := range reservedWikiNames {\n\t\tif name == reservedName {\n\t\t\treturn ErrWikiReservedName{name}\n\t\t}\n\t}\n\treturn nil\n}", "func caseAndRegister(wlName string) string {\n\tvar orj string = wlName\n\twlName = CamelCase(wlName)\n\twlNames[orj] = wlName\n\treturn wlName\n}", "func sanitizeKeywords(name string) string {\n\tfor _, k := range keywords {\n\t\tif name == k {\n\t\t\treturn name + \"Arg\"\n\t\t}\n\t}\n\treturn name\n}", "func sanitizeKeywords(name string) string {\n\tfor _, k := range keywords {\n\t\tif name == k {\n\t\t\treturn name + \"Arg\"\n\t\t}\n\t}\n\treturn name\n}", "func (s *BasePlSqlParserListener) EnterNon_reserved_keywords_pre12c(ctx *Non_reserved_keywords_pre12cContext) {\n}", "func reservedWords(s string) bool {\n\tlc := strings.ToLower(s)\n\tif len(s) == 1 {\n\t\tfor _, b := range \"abcxyzij\" {\n\t\t\tif lc[0] == byte(b) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\n\tfor _, res := range keywords {\n\t\tif lc == res {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func pySafeName(nm string) string {\n\tif _, bad := pyKeywords[nm]; bad {\n\t\treturn \"my\" + nm\n\t}\n\treturn nm\n}", "func key(name string, locale string) string {\n\tkey := name\n\tvar shortened bool\n\tfor {\n\t\tkey, shortened = data.ShortenName(key)\n\t\tif !shortened {\n\t\t\tbreak\n\t\t}\n\t}\n\n\tvar location []string\n\tfor _, word := range strings.Split(locale, \" \") {\n\t\tif word == strings.Title(word) {\n\t\t\tif _, exists := data.ExtraWords[strings.ToUpper(word)]; !exists {\n\t\t\t\tlocation = append(location, word)\n\t\t\t}\n\t\t} else if len(location) > 1 {\n\t\t\tlocation = []string{}\n\t\t}\n\t}\n\n\tif len(location) > 2 {\n\t\tlocation = location[0:2]\n\t}\n\treturn key + \" - \" + strings.Join(location, \" \")\n}", "func Underscorize(term string) string {\n\tterm = acronymRegex.ReplaceAllStringFunc(term, func(match string) string {\n\t\treturn \"_\" + strings.ToLower(match)\n\t})\n\tif strings.HasPrefix(term, \"_\") {\n\t\tterm = term[1:]\n\t}\n\treplacement := \"${1}_${2}\"\n\tterm = upperWordsRegex.ReplaceAllString(term, replacement)\n\tterm = lowerWordsRegex.ReplaceAllString(term, replacement)\n\tterm = strings.Replace(term, \"-\", \"_\", -1)\n\treturn strings.ToLower(term)\n}", "func init() {\n\treplaceSet = []string{\n\t\t\" \", \"-\",\n\t\t\"'\", \"\",\n\t\t\"ı\", \"i\",\n\t\t\",\", \"\",\n\t\t\".\", \"\",\n\t\t\"#\", \"\",\n\t\t\"!\", \"\",\n\t}\n}", "func (s *BasePlSqlParserListener) ExitNon_reserved_keywords_in_12c(ctx *Non_reserved_keywords_in_12cContext) {\n}", "func newServiceNameReplacer() *strings.Replacer {\n\tvar mapping [256]byte\n\t// we start with everything being replaces with underscore, and later fix some safe characters\n\tfor i := range mapping {\n\t\tmapping[i] = '_'\n\t}\n\t// digits are safe\n\tfor i := '0'; i <= '9'; i++ {\n\t\tmapping[i] = byte(i)\n\t}\n\t// lower case letters are safe\n\tfor i := 'a'; i <= 'z'; i++ {\n\t\tmapping[i] = byte(i)\n\t}\n\t// upper case letters are safe, but convert them to lower case\n\tfor i := 'A'; i <= 'Z'; i++ {\n\t\tmapping[i] = byte(i - 'A' + 'a')\n\t}\n\t// dash and dot are safe\n\tmapping['-'] = '-'\n\tmapping['.'] = '.'\n\n\t// prepare array of pairs of bad/good characters\n\toldnew := make([]string, 0, 2*(256-2-10-int('z'-'a'+1)))\n\tfor i := range mapping {\n\t\tif mapping[i] != byte(i) {\n\t\t\toldnew = append(oldnew, string(rune(i)), string(rune(mapping[i])))\n\t\t}\n\t}\n\n\treturn strings.NewReplacer(oldnew...)\n}", "func (nt *applyNameTable) disambiguate(name string) string {\n\tif name == \"\" {\n\t\tname = \"arg\"\n\t} else if isReservedWord(name) {\n\t\tname = \"_\" + name\n\t}\n\n\tif !nt.assigned[name] {\n\t\treturn name\n\t}\n\n\troot := name\n\tfor i := 1; nt.nameCounts[name] != 0; i++ {\n\t\tname = fmt.Sprintf(\"%s%d\", root, i)\n\t}\n\treturn name\n}", "func main(){\n\t// NOTE 1: keywords should not be used as identifiers\n\t//break:= \"one\" // .\\e_keywords.go:21:7: syntax error: unexpected := at end of statement\n\n\t// NOTE 2: CamelCase is recommended for identifiers\n\tbreak_one := \"one\"\n\tfmt.Println(\"break_one:\", break_one)\n\n\tbreakOne := \"one\"\n\tfmt.Println(\"breakOne :\", breakOne)\n}", "func gcTranslate(gcname string, m map[string]string) string {\n\tswitch gcname {\n\tcase \"Langara\",\"L\", \"1\":\n\t\treturn m[\"Langara\"]\n\tcase \"Fraserview\", \"F\", \"2\":\n\t\treturn m[\"Fraserview\"]\n\tcase \"McCleery\"\t, \"M\", \"3\":\n\t\treturn m[\"McCleery\"]\n\tcase \"All\":\n\t\treturn \"1,2,3\"\n\t}\n\treturn \"no courses options recognized\"\n}", "func init() {\n\tprotoErrorVerbose = true\n\n\t// fix up the generated \"token name\" array so that error messages are nicer\n\tsetTokenName(_STRING_LIT, \"string literal\")\n\tsetTokenName(_INT_LIT, \"int literal\")\n\tsetTokenName(_FLOAT_LIT, \"float literal\")\n\tsetTokenName(_NAME, \"identifier\")\n\tsetTokenName(_ERROR, \"error\")\n\t// for keywords, just show the keyword itself wrapped in quotes\n\tfor str, i := range keywords {\n\t\tsetTokenName(i, fmt.Sprintf(`\"%s\"`, str))\n\t}\n}", "func init() {\n\ttag.Register(\"el-CY\", \"Greek (Cyprus)\")\n\ttag.Register(\"el-GR\", \"Greek (Greece)\")\n\ttag.Register(\"grc-GR\", \"Ancient Greek (Greece)\")\n\ttag.Register(\"gre-GR\", \"Modern Greek (Greece)\")\n}", "func (s *BasePlSqlParserListener) ExitNon_reserved_keywords_pre12c(ctx *Non_reserved_keywords_pre12cContext) {\n}", "func DeclareVariables() {\r\n // OK\r\n word := 1\r\n multipleWords := 2\r\n // Not OK\r\n multiplewords\r\n multiple_words\r\n}", "func ToGoPrivate(name string) string {\n\trunes := make([]rune, 0, len(name))\n\n\tfirst := true\n\twordWalker(name, func(info *wordInfo) {\n\t\tword := info.Word\n\t\tswitch {\n\t\tcase first:\n\t\t\tif strings.ToUpper(word) == word || strings.ToLower(word) == word {\n\t\t\t\t// ID → id, CAMEL → camel\n\t\t\t\tword = strings.ToLower(info.Word)\n\t\t\t} else {\n\t\t\t\t// ITicket → iTicket\n\t\t\t\tword = LcFirst(info.Word)\n\t\t\t}\n\t\t\tfirst = false\n\t\tcase info.MatchCommonInitial:\n\t\t\tword = strings.ToUpper(word)\n\t\tcase !info.HasCommonInitial:\n\t\t\tword = UcFirst(strings.ToLower(word))\n\t\t}\n\t\trunes = append(runes, []rune(word)...)\n\t})\n\n\treturn sanitizeKeywords(string(runes))\n}", "func Sanitize(s string) string {\n\ts = CamelCase(s)\n\tif slices.Contains(ReservedNames, s) {\n\t\treturn s + \"_\"\n\t}\n\n\treturn s\n}", "func goNameFromStrings(parts ...string) string {\n\tname := \"\"\n\tre := regexp.MustCompile(\"{|}|-|_\")\n\tfor _, p := range parts {\n\t\tc := re.ReplaceAllString(p, \"\")\n\t\tswitch c {\n\t\tcase \"id\":\n\t\t\tname += \"ID\"\n\t\tcase \"url\":\n\t\t\tname += \"URL\"\n\t\tcase \"api\":\n\t\t\tname += \"API\"\n\t\tdefault:\n\t\t\tname += strings.Title(c)\n\t\t}\n\t}\n\treturn name\n}", "func (lId *LangId) getVarNames() {\n\n\tvar getVarNames = func(toCamel bool) []string {\n\n\t\tvar (\n\t\t\ttmpMap = make(map[string]string)\n\t\t\tout []string\n\t\t\ttmpName string\n\t\t)\n\t\tfor _, val := range lId.lines {\n\t\t\ttmpSl := strings.Split(val, \":\")\n\t\t\tif len(tmpSl) > 1 {\n\t\t\t\ttmpMap[tmpSl[0]] = \"\"\n\t\t\t}\n\t\t}\n\t\tfor name, _ := range tmpMap {\n\t\t\tif toCamel {\n\t\t\t\ttmpName = ToCamel(strings.TrimSpace(name))\n\t\t\t} else {\n\t\t\t\ttmpName = strings.TrimSpace(name)\n\t\t\t}\n\t\t\tif len(tmpName) > 25 || len(tmpName) == 0 {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tout = append(out, tmpName)\n\t\t}\n\t\t// Sort string preserving order ascendant\n\t\tsort.SliceStable(out, func(i, j int) bool {\n\t\t\treturn out[i] < out[j]\n\t\t})\n\t\treturn out\n\t}\n\n\tfmt.Println(\"To replace 'langIdEntry' structure\")\n\tfmt.Println(\"type langIdEntry struct {\")\n\tvarNames := getVarNames(true)\n\tfor idx, name := range varNames {\n\t\tfmt.Printf(\"\\t%s\", name)\n\t\tif idx < len(varNames)-1 {\n\t\t\tfmt.Println(\",\")\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tfmt.Println(` string\n}\n`)\n\n\tfmt.Println(\"\\nTo replace 'varNamesList'\")\n\tfmt.Println(\"var varNamesList = []string{\")\n\tfor _, name := range getVarNames(false) {\n\t\tfmt.Printf(\"\\t\\\"%s:\\\",\\n\", name)\n\t}\n\tfmt.Println(\"}\")\n\n\tfmt.Println(\"\\nTo replace 'storeEntry' method\")\n\tvar cases string\n\tfor _, name := range getVarNames(false) {\n\t\tcases += buildCaseStatement(name)\n\t}\n\tfmt.Println(buildFuncContent(cases))\n}", "func WhyAreYouUsingCapitalLetters_InACFunctionName() {}", "func NameDefinitions(nameDefs map[string][]string, match string) string {\n\tw := io.NewBufferWriter()\n\tif len(nameDefs) == 0 {\n\t\tio.ReplyNL(w, io.Grey+\"nothing to show\")\n\t\treturn w.String()\n\t}\n\tkeys := make([]string, 0)\n\tfor k := range nameDefs {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Sort(sort.StringSlice(keys))\n\tfor _, k := range keys {\n\t\tv := nameDefs[k]\n\t\tcmd := s.Join(v, \" \")\n\t\tif match == \"\" || s.Contains(k, match) || s.Contains(cmd, match) {\n\t\t\tio.ReplyNL(w, io.Magenta+k+io.Grey+\" \"+cmd)\n\t\t}\n\t}\n\treturn w.String()\n}", "func init() {\n\ttag.Register(\"tn-BW\", \"Setswana (Botswana)\")\n\ttag.Register(\"tn-ZA\", \"Setswana (South Africa)\")\n}", "func TestMisspellNameNegative(t *testing.T) {\n\ttest := linttest.NewSuite(t)\n\ttest.Config().TypoFixer = misspell.New()\n\ttest.AddFile(`<?php\nfunction includ() {\n}\n\nclass impelments {}\n\nclass PostRedirect {}\n`)\n\ttest.RunAndMatch()\n}", "func CheckCompareReservedWordsWithMySQL(t *testing.T, db *dbsql.DB, dir string) {\n\tdata, err := ioutil.ReadFile(filepath.Join(dir, \"parser.y\"))\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tcontent := string(data)\n\treservedKeywords := iextract.KeywordsFromTokens(content, iextract.KeywordReserved)\n\tunreservedKeywords := iextract.KeywordsFromTokens(content, iextract.KeywordUnreserved)\n\tnotKeywordTokens := iextract.KeywordsFromTokens(content, iextract.KeywordNot)\n\ttidbKeywords := iextract.KeywordsFromTokens(content, iextract.KeywordTiDB)\n\n\tp := New()\n\tfor _, kw := range reservedKeywords {\n\t\tswitch kw {\n\t\tcase \"CURRENT_ROLE\":\n\t\t\t// special case: we do reserve CURRENT_ROLE but MySQL didn't,\n\t\t\t// and unreservering it causes legit parser conflict.\n\t\t\tcontinue\n\t\t}\n\n\t\tquery := \"do (select 1 as \" + kw + \")\"\n\n\t\tvar err error\n\n\t\tif _, ok := windowFuncTokenMap[kw]; !ok {\n\t\t\t// for some reason the query does parse even then the keyword is reserved in TiDB.\n\t\t\t_, _, err = p.Parse(query, \"\", \"\")\n\t\t\tif !strings.Contains(err.Error(), kw) {\n\t\t\t\tt.Errorf(\"error should contain '%s': %s\", kw, err)\n\t\t\t}\n\t\t}\n\n\t\t_, err = db.Exec(query)\n\t\tif !strings.Contains(err.Error(), kw) {\n\t\t\tt.Errorf(\"MySQL suggests that '%s' should *not* be reserved!\", kw)\n\t\t}\n\t}\n\n\tfor _, kws := range [][]string{unreservedKeywords, notKeywordTokens, tidbKeywords} {\n\t\tfor _, kw := range kws {\n\t\t\tswitch kw {\n\t\t\tcase \"FUNCTION\", // reserved in 8.0.1\n\t\t\t\t\"SEPARATOR\": // ?\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tquery := \"do (select 1 as \" + kw + \")\"\n\n\t\t\tstmts, _, err := p.Parse(query, \"\", \"\")\n\t\t\tif err != nil {\n\t\t\t\tt.Errorf(\"%s: %s\", kw, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif len(stmts) != 1 {\n\t\t\t\tt.Errorf(\"%s should have one statement; has %d\", kw, len(stmts))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// c.Assert(stmts[0], FitsTypeOf, &ast.DoStmt{})\n\t\t\t_, err = db.Exec(query)\n\t\t\tif err != nil {\n\t\t\t\tt.Errorf(\"MySQL suggests that '%s' should be reserved!: %s\", kw, err)\n\t\t\t}\n\t\t}\n\t}\n}", "func TestMisspellNamePositive(t *testing.T) {\n\ttest := linttest.NewSuite(t)\n\ttest.Config().TypoFixer = misspell.New()\n\ttest.AddFile(`<?php\nfunction unconditionnally_rollback() {}\n\nfunction f($notificaton) {\n}\n\nclass c {\n private function m($flag_normallized) {}\n}\n\nclass Mocrotransactions {\n}\n\nfunction f_overpoweing() {\n}\n\nclass c {\n private function set_persistance() {}\n}\n`)\n\ttest.Expect = []string{\n\t\t`\"unconditionnally\" is a misspelling of \"unconditionally\"`,\n\t\t`\"notificaton\" is a misspelling of \"notification\"`,\n\t\t`\"normallized\" is a misspelling of \"normalized\"`,\n\t\t`\"Mocrotransactions\" is a misspelling of \"Microtransactions\"`,\n\t\t`\"overpoweing\" is a misspelling of \"overpowering\"`,\n\t\t`\"persistance\" is a misspelling of \"persistence\"`,\n\t}\n\ttest.RunAndMatch()\n}", "func addUse(name string){\r\n\tnewName = name;\r\n\tfileName = newName\r\n}", "func sanitizeName(name string) string {\n\toutput := strings.ToLower(illegalChars.ReplaceAllString(name, \"_\"))\n\n\tif legalLabel.MatchString(output) {\n\t\treturn output\n\t}\n\t// Prefix name with _ if it begins with a number\n\treturn \"_\" + output\n}", "func TestHellosName(t *testing.T) {\n\tname := \"Gladys\"\n\twant := regexp.MustCompile(`\\b` + name + `\\b`)\n\tmsg, err := Hellos([]string{name})\n\tif !want.MatchString(msg[name]) || err != nil {\n\t\tt.Fatalf(`Hello(\"%v\") = %q, %v, want match for %#q, nil`, name, msg, err, want)\n\t}\n}", "func tagEnvName(name string) string {\n\t// Clean up disallowed characters.\n\tname = regexp.MustCompile(`[^a-zA-Z0-9]`).ReplaceAllString(name, \"_\")\n\tname = regexp.MustCompile(`_+`).ReplaceAllString(name, \"_\")\n\n\t// Convert to standard case format.\n\tif strings.Contains(name, \"_\") {\n\t\tname = strings.ToUpper(name)\n\t} else {\n\t\tname = toSnake(name)\n\t}\n\treturn name\n}", "func substitution(answer string) string {\n\treflections := map[string]string{\n\t\t\"am\": \"are\",\n\t\t\"was\": \"were\",\n\t\t\"i\": \"you\",\n\t\t\"i'd\": \"you would\",\n\t\t\"i've\": \"you have\",\n\t\t\"i'll\": \"you will\",\n\t\t\"my\": \"your\",\n\t\t\"are\": \"am\",\n\t\t\"you've\": \"I have\",\n\t\t\"you'll\": \"I will\",\n\t\t\"your\": \"my\",\n\t\t\"yours\": \"mine\",\n\t\t\"you\": \"me\",\n\t\t\"me\": \"you\",\n\t\t\"myself\": \"yourself\",\n\t\t\"yourself\": \"myself\",\n\t\t\"i'm\": \"you are\",\n\t}\n\n\twords := strings.Split(answer, \" \") // get slices of the words\n\n\tfor i, word := range words {\t// loop through whole sentence\n\t\tif val, ok := reflections[word]; ok {\t// check for the word in reflection\n\t\t\twords[i] = val // substitite the value\n\t\t}//if\n\t}//for\n\n\t// Return substituted string\n\treturn strings.Join(words, \" \") // join back into sentence\n}", "func stripReservedKeywords(str string) string {\n\tif sanitizedValue, exists := keywords[str]; exists {\n\t\treturn sanitizedValue\n\t}\n\n\treturn str\n}", "func escapeBadChars(target string) string {\n\ts := strings.Replace(target, \"*\", \"__ASTERISK__\", -1)\n\ts = strings.Replace(s, \"=\", \"__ASSIGN__\", -1)\n\treturn strings.Replace(s, \"-\", \"__DASH__\", -1)\n}", "func init() {\n\ttag.Register(\"ro-MD\", \"Romanian (Moldova)\")\n\ttag.Register(\"ro-RO\", \"Romanian (Romania)\")\n}", "func isUnreserved(c byte) bool {\n\treturn (c >= 'a' && c <= 'z') ||\n\t\t(c >= 'A' && c <= 'Z') ||\n\t\t(c >= '0' && c <= '9') ||\n\t\tc == '-' ||\n\t\tc == '_' ||\n\t\tc == '.' ||\n\t\tc == '!' ||\n\t\tc == '~' ||\n\t\tc == '*' ||\n\t\tc == '\\'' ||\n\t\tc == '(' ||\n\t\tc == ')'\n}", "func (s *BasePlSqlParserListener) EnterSynonym_name(ctx *Synonym_nameContext) {}", "func DefNameIsAllowed(name string) error {\n\tnames := []string{\"\", \"C\", \"ErrorCallback\", \"Container\", \"NewContainer\"}\n\n\tformatted := FormatDefName(name)\n\n\tfor _, n := range names {\n\t\tif n == formatted {\n\t\t\treturn errors.New(\"DefName '\" + name + \"' is not allowed (reserved key word)\")\n\t\t}\n\t}\n\n\tif bytes.ContainsRune(digits, rune(formatted[0])) {\n\t\treturn errors.New(\"DefName '\" + name + \"' is not allowed (first char is a digit)\")\n\t}\n\n\treturn nil\n}", "func reservedWorkloadNames() map[string]bool {\n\treturn map[string]bool{\n\t\t\"pipelines\": true, // reserved to avoid directory conflict with copilot pipelines\n\t\t\"environments\": true, // reserved to avoid directory conflict with copilot environments\n\t}\n}", "func (*unifinames) Name() string { return \"unifi-names\" }", "func varSlugged(s string) (vslug string) {\n\tvslug = strings.Trim(slugger.ReplaceAllString(s, \"_\"), \"_\")\n\tvslug = strings.ToUpper(string(vslug[0])) + vslug[1:]\n\treturn\n}", "func handleAllowedKeywordsToString(allowed []string) string {\n\tbldr := \"'\" + strings.Join(allowed, \"', '\") + \"'\"\n\treturn \"[\" + bldr + \"]\"\n}", "func quoteIdentifier(s string) string {\n\tquote := false\n\tif reserved[strings.ToUpper(s)] {\n\t\tquote = true\n\t}\n\tif !ident.MatchString(s) {\n\t\tquote = true\n\t}\n\tif quote {\n\t\treturn fmt.Sprintf(\"\\\"%s\\\"\", strings.ReplaceAll(s, \"\\\"\", \"\\\"\\\"\"))\n\t} else {\n\t\treturn s\n\t}\n}", "func f() {\n\t_ = 12 // nolint // want `\\Qremove a space between // and \"nolint\" directive`\n\n\t_ = 30 // nolint2 foo bar // want `\\Qsuggestion: //nolint2 foo bar`\n\n\t/*\n\t\tnolint // want `\\Qdon't put \"nolint\" inside a multi-line comment`\n\t*/\n\n\t//go:baddirective // want `\\Qdon't use baddirective go directive`\n\t//go:noinline\n\t//go:generate foo bar\n\n\t//nolint:gocritic // want `\\Qhey, this is kinda upsetting`\n\n\t// This is a begining // want `\\Q\"begining\" may contain a typo`\n\t// Of a bizzare text with typos. // want `\\Q\"bizzare\" may contain a typo`\n\n\t// I can't give you a buisness advice. // want `\\Q\"buisness advice\" may contain a typo`\n\n\t// calender // want `\\Qfirst=calender`\n\t// cemetary // want `\\Qsecond=cemetary`\n\n\t// collegue // want `\\Qx=\"collegue\"`\n\t// commitee // want `\\Qx=\"\"`\n}", "func SafeIdentifier(s string) string {\n\tvar (\n\t\tisL, isD, last bool\n\t\tbuf bytes.Buffer\n\t)\n\tfor i, r := range s {\n\t\tif isL, isD = unicode.IsLetter(r), unicode.IsDigit(r); isL || isD || ((r == '_') && (i == 0)) {\n\t\t\tif (i > 0) && (isL != last) {\n\t\t\t\tbuf.WriteRune(' ')\n\t\t\t}\n\t\t\tbuf.WriteRune(r)\n\t\t} else {\n\t\t\tbuf.WriteRune(' ')\n\t\t}\n\t\tlast = isL\n\t}\n\twords := Split(strings.Title(buf.String()), \" \")\n\tfor i, w := range words {\n\t\tif (len(w) > 1) && IsUpper(w) {\n\t\t\twords[i] = strings.Title(strings.ToLower(w))\n\t\t}\n\t}\n\treturn strings.Join(words, \"\")\n}", "func checkForbiddenMessageField(name string) error {\n\tif name == \"creator\" {\n\t\treturn fmt.Errorf(\"%s is used by the message scaffolder\", name)\n\t}\n\n\treturn checkGoReservedWord(name)\n}", "func replaceKeptnPlaceholders(input string, keptnEvent BaseKeptnEvent) string {\n\tresult := input\n\n\t// first we do the regular keptn values\n\tresult = strings.Replace(result, \"$TIMESTRING\", keptnEvent.time, -1)\n\tresult = strings.Replace(result, \"$TIMEUTCSTRING\", keptnEvent.timeutc, -1)\n\tresult = strings.Replace(result, \"$TIMEUTCMS\", keptnEvent.timeutcms, -1)\n\n\tresult = strings.Replace(result, \"$CONTEXT\", keptnEvent.context, -1)\n\tresult = strings.Replace(result, \"$EVENT\", keptnEvent.event, -1)\n\tresult = strings.Replace(result, \"$SOURCE\", keptnEvent.source, -1)\n\n\tresult = strings.Replace(result, \"$PROJECT\", keptnEvent.project, -1)\n\tresult = strings.Replace(result, \"$STAGE\", keptnEvent.stage, -1)\n\tresult = strings.Replace(result, \"$SERVICE\", keptnEvent.service, -1)\n\tresult = strings.Replace(result, \"$DEPLOYMENT\", keptnEvent.deployment, -1)\n\tresult = strings.Replace(result, \"$TESTSTRATEGY\", keptnEvent.testStrategy, -1)\n\n\tresult = strings.Replace(result, \"$DEPLOYMENTURILOCAL\", keptnEvent.deploymentURILocal, -1)\n\tresult = strings.Replace(result, \"$DEPLOYMENTURIPUBLIC\", keptnEvent.deploymentURIPublic, -1)\n\n\tresult = strings.Replace(result, \"$ACTION\", keptnEvent.action, -1)\n\n\tresult = strings.Replace(result, \"$PROBLEMID\", keptnEvent.problemID, -1)\n\tresult = strings.Replace(result, \"$PROBLEMSTATE\", keptnEvent.problemState, -1)\n\tresult = strings.Replace(result, \"$PID\", keptnEvent.pid, -1)\n\tresult = strings.Replace(result, \"$PROBLEMTITLE\", keptnEvent.problemTitle, -1)\n\tresult = strings.Replace(result, \"$PROBLEMURL\", keptnEvent.problemURL, -1)\n\n\t// now we do the labels\n\tfor key, value := range keptnEvent.labels {\n\t\tresult = strings.Replace(result, \"$LABEL_\"+strings.ToUpper(key), value, -1)\n\t}\n\n\t// now we do the remediation values\n\tfor remediationKey, remediationValue := range keptnEvent.remediationValues {\n\t\tresult = strings.Replace(result, \"$VALUE_\"+strings.ToUpper(remediationKey), remediationValue, -1)\n\t}\n\n\t// now we do all environment variables\n\tfor _, env := range os.Environ() {\n\t\tpair := strings.SplitN(env, \"=\", 2)\n\t\tresult = strings.Replace(result, \"$ENV_\"+strings.ToUpper(pair[0]), pair[1], -1)\n\t}\n\n\t// TODO: iterate through k8s secrets!\n\n\treturn result\n}", "func makeValidIdentifier(name string) string {\n\tvar builder strings.Builder\n\tfor i, c := range name {\n\t\tif i == 0 && c == '@' {\n\t\t\tbuilder.WriteRune(c)\n\t\t\tcontinue\n\t\t}\n\t\tif !isLegalIdentifierPart(c) {\n\t\t\tbuilder.WriteRune('_')\n\t\t} else {\n\t\t\tif i == 0 && !isLegalIdentifierStart(c) {\n\t\t\t\tbuilder.WriteRune('_')\n\t\t\t}\n\t\t\tbuilder.WriteRune(c)\n\t\t}\n\t}\n\tname = builder.String()\n\tif isReservedWord(name) {\n\t\treturn \"@\" + name\n\t}\n\treturn name\n}", "func abbreviations(abv string) string {\n\tswitch strings.ToLower(abv) {\n\tcase \"id\", \"ppid\", \"pid\", \"mac\", \"ip\", \"iana\", \"uid\", \"ecs\", \"url\", \"os\", \"http\":\n\t\treturn strings.ToUpper(abv)\n\tdefault:\n\t\treturn abv\n\t}\n}", "func (g *Generator) declareNameVars(runs [][]Value, typeName string, suffix string) {\n\tg.Printf(\"const _%s_name%s = \\\"\", typeName, suffix)\n\tfor _, run := range runs {\n\t\tfor i := range run {\n\t\t\tg.Printf(\"%s\", run[i].typeInfo.originalName)\n\t\t}\n\t}\n\tg.Printf(\"\\\"\\n\")\n}", "func (this *WordDictionary) AddWord(word string) {\n \n}", "func escapeKey(originKey string) string {\n\tfinalKey := originKey\n\tfor _, str := range originKey {\n\t\tswitch {\n\t\tcase unicode.IsLetter(str):\n\t\tcase unicode.IsNumber(str):\n\t\tcase byte(str) == '-':\n\t\tcase byte(str) == '_':\n\t\tcase byte(str) == '.':\n\t\tdefault:\n\t\t\tfinalKey = strings.Replace(finalKey, string(str), \"_\", -1)\n\t\t}\n\t}\n\treturn finalKey\n}", "func (t *trs80) writeWord(w *scottpb.Word) {\n\tif w.Synonym {\n\t\tfmt.Fprintf(t.out, \"\\\"*%s\\\"\\n\", w.Word)\n\t} else {\n\t\tfmt.Fprintf(t.out, \"\\\"%s\\\"\\n\", w.Word)\n\t}\n}", "func explainCipher(d description) description {\n\tkexAndCipher := strings.Split(d.Slug, \"_WITH_\")\n\tif len(kexAndCipher) == 2 {\n\t\td.Name = fmt.Sprintf(\"%s key exchange, %s cipher\", kexAndCipher[0][len(\"TLS_\"):], kexAndCipher[1])\n\t} else {\n\t\td.Name = fmt.Sprintf(\"%s cipher\", d.Slug[len(\"TLS_\"):])\n\t}\n\treturn d\n}", "func replaceKeptnPlaceholders(input string, event adapter.EventContentAdapter) string {\n\tresult := input\n\n\t// first we do the regular keptn values\n\tresult = strings.Replace(result, \"$CONTEXT\", event.GetShKeptnContext(), -1)\n\tresult = strings.Replace(result, \"$EVENT\", event.GetEvent(), -1)\n\tresult = strings.Replace(result, \"$SOURCE\", event.GetSource(), -1)\n\tresult = strings.Replace(result, \"$PROJECT\", event.GetProject(), -1)\n\tresult = strings.Replace(result, \"$STAGE\", event.GetStage(), -1)\n\tresult = strings.Replace(result, \"$SERVICE\", event.GetService(), -1)\n\tresult = strings.Replace(result, \"$DEPLOYMENT\", event.GetDeployment(), -1)\n\tresult = strings.Replace(result, \"$TESTSTRATEGY\", event.GetTestStrategy(), -1)\n\n\t// now we do the labels\n\tfor key, value := range event.GetLabels() {\n\t\tresult = strings.Replace(result, \"$LABEL.\"+key, value, -1)\n\t}\n\n\t// now we do all environment variables\n\tfor _, env := range os.Environ() {\n\t\tpair := strings.SplitN(env, \"=\", 2)\n\t\tresult = strings.Replace(result, \"$ENV.\"+pair[0], pair[1], -1)\n\t}\n\n\t// TODO: iterate through k8s secrets!\n\n\treturn result\n}", "func typeAdjustName(purlType, name string) string {\n\tswitch purlType {\n\tcase TypeBitbucket, TypeDebian, TypeGithub, TypeGolang, TypeNPM:\n\t\treturn strings.ToLower(name)\n\tcase TypePyPi:\n\t\treturn strings.ToLower(strings.ReplaceAll(name, \"_\", \"-\"))\n\t}\n\treturn name\n}", "func getSecretPhrase() (secret string) {\n\tsecret = \"LORAXSNUGGLEGEORGEICE\"\n\treturn\n}", "func cleanName(name string) string {\n\tvar builder strings.Builder\n\tfor i, c := range name {\n\t\tif !isLegalIdentifierPart(c) {\n\t\t\tbuilder.WriteRune('_')\n\t\t} else {\n\t\t\tif i == 0 && !isLegalIdentifierStart(c) {\n\t\t\t\tbuilder.WriteRune('_')\n\t\t\t}\n\t\t\tbuilder.WriteRune(c)\n\t\t}\n\t}\n\treturn builder.String()\n}", "func setupSubst(objType string, search string, replace string) {\n\tif !globalType[objType] {\n\t\tabort.Msg(\"Unknown type %s\", objType)\n\t}\n\taddSubst := func(objType, search, replace string) {\n\t\tsubMap, ok := subst[objType]\n\t\tif !ok {\n\t\t\tsubMap = make(map[string]string)\n\t\t\tsubst[objType] = subMap\n\t\t}\n\t\tsubMap[search] = replace\n\t}\n\n\taddSubst(objType, search, replace)\n\n\tfor _, other := range aliases[objType] {\n\t\taddSubst(other, search, replace)\n\t}\n}", "func (e Environment) Overridef(name string, format string, a ...interface{}) {\n\te[fmt.Sprintf(\"%s.override\", name)] = fmt.Sprintf(format, a...)\n}", "func (f *Forth) compileWord(token string) error {\n\tw := word{func(f *Forth) error { return nil }, \"\"}\n\tf.dict[strings.ToLower(token)] = w\n\treturn nil\n}", "func ExpandAbbreviations(template, abbreviations string) {\n\n}", "func buildSynonyms(a *Parser) map[string]string {\n\tsynonyms := make(map[string]string)\n\tfor _, n := range a.seq {\n\t\tp := a.params[n]\n\t\tif n == p.name {\n\t\t\tif len(n) == 0 {\n\t\t\t\tsynonyms[n] = \"(nameless)\"\n\t\t\t} else {\n\t\t\t\tsynonyms[n] = n\n\t\t\t}\n\t\t} else {\n\t\t\tsynonyms[p.name] += \", \" + n\n\t\t}\n\t}\n\treturn synonyms\n}", "func (g *Generator) cName(d desc.Descriptor) string {\n\tname := d.GetName()\n\tswitch name {\n\tcase \"and\", \"and_eq\", \"asm\", \"auto\", \"bitand\", \"bitor\", \"bool\", \"break\",\n\t\t\"case\", \"catch\", \"char\", \"class\", \"compl\", \"const\", \"const_cast\", \"continue\",\n\t\t\"default\", \"delete\", \"do\", \"double\", \"dynamic_cast\", \"else\", \"enum\",\n\t\t\"explicit\", \"extern\", \"false\", \"float\", \"for\", \"friend\", \"goto\", \"if\",\n\t\t\"inline\", \"int\", \"long\", \"mutable\", \"namespace\", \"new\", \"not\", \"not_eq\",\n\t\t\"operator\", \"or\", \"or_eq\", \"private\", \"protected\", \"public\", \"register\",\n\t\t\"reinterpret_cast\", \"return\", \"short\", \"signed\", \"sizeof\", \"static\",\n\t\t\"static_cast\", \"struct\", \"switch\", \"template\", \"this\", \"throw\", \"true\", \"try\",\n\t\t\"typedef\", \"typeid\", \"typename\", \"union\", \"unsigned\", \"using\", \"virtual\",\n\t\t\"void\", \"volatile\", \"wchar_t\", \"while\", \"xor\", \"xor_eq\":\n\t\tname += \"_\"\n\t}\n\treturn name\n}", "func sanitiseName(key string) string {\n\treturn strings.ReplaceAll(key, \".\", \"_\")\n}", "func (f *Forth) addBuiltins() {\n\tf.dict[\"+\"] = word{bAdd, \"\"}\n\tf.dict[\"-\"] = word{bSub, \"\"}\n\tf.dict[\"*\"] = word{bMul, \"\"}\n\tf.dict[\"/\"] = word{bDiv, \"\"}\n\tf.dict[\"1+\"] = word{bAdd1, \"\"}\n\tf.dict[\"1-\"] = word{bSub1, \"\"}\n\tf.dict[\"2*\"] = word{bMul2, \"\"}\n\tf.dict[\"2/\"] = word{bDiv2, \"\"}\n\tf.dict[\"/mod\"] = word{bDivMod, \"\"}\n\tf.dict[\"mod\"] = word{bMod, \"\"}\n\tf.dict[\"=\"] = word{bEq, \"\"}\n\tf.dict[\"<>\"] = word{bNe, \"\"}\n\tf.dict[\"<\"] = word{bLt, \"\"}\n\tf.dict[\">\"] = word{bGt, \"\"}\n\tf.dict[\"<=\"] = word{bLe, \"\"}\n\tf.dict[\">=\"] = word{bGe, \"\"}\n\tf.dict[\"true\"] = word{bTrue, \"\"}\n\tf.dict[\"false\"] = word{bFalse, \"\"}\n\tf.dict[\"dup\"] = word{bDup, \"\"}\n\tf.dict[\"2dup\"] = word{b2Dup, \"\"}\n\tf.dict[\"drop\"] = word{bDrop, \"\"}\n\tf.dict[\"2drop\"] = word{b2Drop, \"\"}\n\tf.dict[\"swap\"] = word{bSwap, \"\"}\n\tf.dict[\"over\"] = word{bOver, \"\"}\n\tf.dict[\"rot\"] = word{bRot, \"\"}\n\tf.dict[\".\"] = word{bDot, \"\"}\n\tf.dict[\".s\"] = word{bShow, \"\"}\n\tf.dict[\".r\"] = word{bShowR, \"\"}\n\tf.dict[\"cr\"] = word{bCr, \"\"}\n\tf.dict[\"emit\"] = word{bEmit, \"\"}\n}", "func fn3() { // want fn3:`Deprecated: Don't use this\\.`\n}", "func initTagConversionMap() {\n\n\tTagStrToInt[\"bos\"] = 0\n\tTagStrToInt[\"$\"] = 1\n\tTagStrToInt[\"\\\"\"] = 2\n\tTagStrToInt[\"(\"] = 3\n\tTagStrToInt[\")\"] = 4\n\tTagStrToInt[\",\"] = 5\n\tTagStrToInt[\"--\"] = 6\n\tTagStrToInt[\".\"] = 7\n\tTagStrToInt[\":\"] = 8\n\tTagStrToInt[\"cc\"] = 9\n\tTagStrToInt[\"cd\"] = 10\n\tTagStrToInt[\"dt\"] = 11\n\tTagStrToInt[\"fw\"] = 12\n\tTagStrToInt[\"jj\"] = 13\n\tTagStrToInt[\"ls\"] = 14\n\tTagStrToInt[\"nn\"] = 15\n\tTagStrToInt[\"np\"] = 16\n\tTagStrToInt[\"pos\"] = 17\n\tTagStrToInt[\"pr\"] = 18\n\tTagStrToInt[\"rb\"] = 19\n\tTagStrToInt[\"sym\"] = 20\n\tTagStrToInt[\"to\"] = 21\n\tTagStrToInt[\"uh\"] = 22\n\tTagStrToInt[\"vb\"] = 23\n\tTagStrToInt[\"md\"] = 24\n\tTagStrToInt[\"in\"] = 25\n\n\tTagIntToStr[0] = \"bos\"\n\tTagIntToStr[1] = \"$\"\n\tTagIntToStr[2] = \"\\\"\"\n\tTagIntToStr[3] = \"(\"\n\tTagIntToStr[4] = \")\"\n\tTagIntToStr[5] = \",\"\n\tTagIntToStr[6] = \"--\"\n\tTagIntToStr[7] = \".\"\n\tTagIntToStr[8] = \":\"\n\tTagIntToStr[9] = \"cc\"\n\tTagIntToStr[10] = \"cd\"\n\tTagIntToStr[11] = \"dt\"\n\tTagIntToStr[12] = \"fw\"\n\tTagIntToStr[13] = \"jj\"\n\tTagIntToStr[14] = \"ls\"\n\tTagIntToStr[15] = \"nn\"\n\tTagIntToStr[16] = \"np\"\n\tTagIntToStr[17] = \"pos\"\n\tTagIntToStr[18] = \"pr\"\n\tTagIntToStr[19] = \"rb\"\n\tTagIntToStr[20] = \"sym\"\n\tTagIntToStr[21] = \"to\"\n\tTagIntToStr[22] = \"uh\"\n\tTagIntToStr[23] = \"vb\"\n\tTagIntToStr[24] = \"md\"\n\tTagIntToStr[25] = \"in\"\n}", "func Translate(input string) string {\n\n\tt, err := tokenizer.New(ipa.Dict(), tokenizer.OmitBosEos())\n\tif err != nil {\n\t\treturn fmt.Sprintln(\"error in initializing tokenizer\", err)\n\t}\n\n\t// split into word list\n\ttokens := t.Analyze(input, tokenizer.Search)\n\n\t// replace 'translatable' words\n\tdatabaseURL := os.Getenv(\"DATABASE_URL\")\n\tdb, err := gorm.Open(\"postgres\", databaseURL)\n\tif err != nil {\n\t\treturn fmt.Sprintln(\"error in openning database,\", err)\n\t}\n\tdefer db.Close()\n\n\tret := \"\"\n\tprecedingPos := \"\"\n\tfor i, token := range tokens {\n\t\tif token.Class == tokenizer.DUMMY || token.Surface == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\t// prefix addition\n\t\t// ! these process should be refactored\n\t\t// to have more generality\n\t\t// 連続する名詞の頭に「お」\n\t\tpos := token.POS()\n\t\tif pos[0] == \"名詞\" &&\n\t\t\t(pos[1] == \"一般\" || pos[1] == \"サ変接続\" || pos[1] == \"数\" || pos[1] == \"形容動詞語幹\") {\n\t\t\t// 先頭にあるか,一つ前が名詞,接頭詞でない\n\t\t\tif i == 0 || (precedingPos != \"名詞\" && precedingPos != \"接頭詞\") {\n\t\t\t\tret += \"お\"\n\t\t\t}\n\t\t}\n\t\tprecedingPos = pos[0]\n\n\t\t// look up database\n\t\tcand := []RegisteredWord{}\n\t\tposStr := strings.Join(token.POS(), \",\")\n\t\tresult := db.Where(\"(source_surface=? OR source_surface IS NULL) AND (? LIKE source_pos || '%' OR source_pos IS NULL)\", token.Surface, posStr).Find(&cand)\n\t\tif result.Error != nil {\n\t\t\treturn fmt.Sprintln(\"error in db query,\", result.Error)\n\t\t}\n\n\t\t// translate\n\t\tif len(cand) > 0 {\n\t\t\t// [TODO] consider better replacement logic\n\t\t\t// such as maximizing `digree of fun'\n\n\t\t\t// if the word has multiple candidates, choose one of them at random\n\t\t\trand.Seed(time.Now().UnixNano())\n\t\t\tp := rand.Intn(len(cand))\n\t\t\tret += cand[p].TargetSurface\n\t\t} else {\n\t\t\t// not registered word\n\t\t\tret += token.Surface\n\t\t}\n\n\t\t// suffix addition\n\t\t// 丁寧語変換\n\t\tif token.POS()[0] == \"動詞\" {\n\t\t\t// collect required info about the next token\n\t\t\tvar nextPos string\n\t\t\tvar nextBase string\n\t\t\tvar nextSurface string\n\t\t\tif i+1 < len(tokens) {\n\t\t\t\tnextPos = tokens[i+1].POS()[0]\n\t\t\t\tnextBase, _ = tokens[i+1].BaseForm()\n\t\t\t\tnextSurface = tokens[i+1].Surface\n\t\t\t}\n\n\t\t\t// 動詞で終わる or 動詞のすぐ後に「ます」「て」以外の助詞助動詞あるいは句点が続く\n\t\t\t// => 丁寧語でないとみなす\n\t\t\tif i == len(tokens)-1 || nextPos == \"句点\" ||\n\t\t\t\t(nextPos == \"助詞\" && nextBase != \"て\") ||\n\t\t\t\t(nextPos == \"助動詞\" && nextBase != \"ます\") {\n\t\t\t\t// 動詞を連用形に活用する\n\t\t\t\tconj := ConjugateVerb(token, renyo)\n\t\t\t\t// remove overlapping\n\t\t\t\truneret := []rune(ret)\n\t\t\t\tsurflen := len([]rune(token.Surface))\n\t\t\t\tret = string(runeret[:len(runeret)-surflen])\n\t\t\t\t// concat conjugated verb\n\t\t\t\tret += conj\n\t\t\t\t// 「ます」を適切な活用の上追加する\n\t\t\t\tret += Conjugate(\"ます\", nextBase, nextPos)\n\t\t\t\t// [TBC] しない -> しません\n\t\t\t\tif nextPos == \"助動詞\" && nextSurface == \"ない\" {\n\t\t\t\t\ttokens[i+1].Surface = \"ん\"\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\n\t\t// explicit EOS\n\t\t// e.g., ました。 -> ましたわ。\n\t\tif token.POS()[0] == \"句点\" && i > 0 &&\n\t\t\ttokens[i-1].POS()[0] != \"助詞\" &&\n\t\t\ttokens[i-1].POS()[0] != \"記号\" &&\n\t\t\ttokens[i-1].POS()[0] != \"感動詞\" {\n\t\t\t// at random (at 50% probability)\n\t\t\trand.Seed(time.Now().UnixNano())\n\t\t\tp := rand.Float32()\n\t\t\tif p < 0.5 {\n\t\t\t\tret += \"わ\"\n\t\t\t} else {\n\t\t\t\tret += \"の\"\n\t\t\t}\n\t\t}\n\t\t// implicit EOS\n\t\t// e.g., した -> したの。\n\t\tif i == len(tokens)-1 &&\n\t\t\t(token.POS()[0] != \"助詞\" &&\n\t\t\t\ttoken.POS()[0] != \"記号\" &&\n\t\t\t\ttoken.POS()[0] != \"名詞\" &&\n\t\t\t\ttoken.POS()[0] != \"感動詞\") {\n\n\t\t\trand.Seed(time.Now().UnixNano())\n\t\t\tp := rand.Float32()\n\t\t\tif p < 0.5 {\n\t\t\t\tret += \"わ\"\n\t\t\t} else {\n\t\t\t\tret += \"の\"\n\t\t\t}\n\t\t}\n\t}\n\n\treturn ret\n}", "func redefinition(short rune, long string) error {\n\tfor _, opt := range flags {\n\t\tif long != \"\" && long == opt.Long() {\n\t\t\treturn fmt.Errorf(\"cannot add option that duplicates long flag: %q\", long)\n\t\t}\n\t\tif short != utf8.RuneError && short == opt.Short() {\n\t\t\treturn fmt.Errorf(\"cannot add option that duplicates short flag: %q\", short)\n\t\t}\n\t}\n\treturn nil\n}", "func GenNaiveSearchIndex(item models.Item) string {\n\twords := make(map[string]struct{})\n\n\t// Extract name.\n\tfor _, v := range extractWords(item.Name) {\n\t\twords[v] = struct{}{}\n\t}\n\n\t// Extract type of item.\n\tfor _, v := range extractWords(item.Type) {\n\t\twords[v] = struct{}{}\n\t}\n\n\t// Extract properties.\n\tfor _, mod := range item.ExplicitMods {\n\t\tfor _, v := range extractWords(mod) {\n\t\t\twords[v] = struct{}{}\n\t\t}\n\t}\n\tfor _, mod := range item.ImplicitMods {\n\t\tfor _, v := range extractWords(mod) {\n\t\t\twords[v] = struct{}{}\n\t\t}\n\t}\n\tfor _, mod := range item.UtilityMods {\n\t\tfor _, v := range extractWords(mod) {\n\t\t\twords[v] = struct{}{}\n\t\t}\n\t}\n\tfor _, mod := range item.EnchantMods {\n\t\tfor _, v := range extractWords(mod) {\n\t\t\twords[v] = struct{}{}\n\t\t}\n\t}\n\tfor _, mod := range item.CraftedMods {\n\t\tfor _, v := range extractWords(mod) {\n\t\t\twords[v] = struct{}{}\n\t\t}\n\t}\n\n\t// Construct final string with sorted keywords.\n\tkeys := make([]string, 0, len(words))\n\tfor key := range words {\n\t\tkeys = append(keys, key)\n\t}\n\tsort.Strings(keys)\n\treturn strings.Join(keys, \" \")\n}", "func mangleName(name string) string {\n\tr, _ := regexp.Compile(\"[^0-9a-zA-Z]+\")\n\treturn r.ReplaceAllString(name, \"\")\n}", "func main() {\n\tvar name string\n\ts := \"Gopher นำแน่!!!\"\n\ti := 5\n\tf := 3.7\n\tb := true\n\tr := '界'\n\n\tfmt.Printf(\"name: %q\\n\", name)\n\tfmt.Printf(\"type: %T\\n\", name)\n\n\tfmt.Printf(\"int: %v\\n\", i)\n\tfmt.Printf(\"float64: %v\\n\", f)\n\tfmt.Printf(\"bool: %v\\n\", b)\n\tfmt.Printf(\"string: %v\\n\", s)\n\tfmt.Printf(\"rune: %v\\n\", r)\n\n}", "func SecretVariableName(name string) string {\n\tsecretType := bdv1.DeploymentSecretTypeVariable\n\tif name == \"\" {\n\t\tname = secretType.String()\n\t} else {\n\t\tname = fmt.Sprintf(\"%s-%s\", secretType, name)\n\t}\n\treturn names.SanitizeSubdomain(name)\n}", "func Define(usr string, fw io.FileWriter, reserved, cmd []string, nameDefs map[string][]string) (string, map[string][]string) {\n\tvar err error\n\tout := \"ok\"\n\tm := strcoll.Copy(nameDefs)\n\tw := io.NewBufferWriter()\n\tleft, right := strcoll.Nth(0, cmd), strcoll.Rest(1, cmd)\n\tif left == \"rm\" {\n\t\t// this might leave dangling names, todo delete them\n\t\tdelete(m, strcoll.Nth(0, right))\n\t\terr = io.StoreDefs(usr, fw, m)\n\t} else {\n\t\tif strcoll.Contains(left, reserved) {\n\t\t\terr = errors.New(left + \" is a reserved word\")\n\t\t} else if strcoll.Contains(left, right) {\n\t\t\terr = errors.New(left + \" can't appear in the right side\")\n\t\t} else {\n\t\t\tif v, ok := m[left]; ok {\n\t\t\t\tout = \"updated old value: \" + s.Join(v, \" \")\n\t\t\t}\n\t\t\tm[left] = right\n\t\t\terr = io.StoreDefs(usr, fw, m)\n\t\t}\n\t}\n\tio.ReplyEither(w, err, io.Grey+out)\n\treturn w.String(), m\n}", "func keyWords() []defCookie {\n\treturn []defCookie{\n\t\t{\"if\", \"TOKEN_IF\"},\n\t\t{\"then\", \"TOKEN_THEN\"},\n\t\t{\"else\", \"TOKEN_ELSE\"},\n\t\t{\"while\", \"TOKEN_WHILE\"},\n\t\t{\"elif\", \"TOKEN_ELIF\"},\n\t\t{\"endif\", \"TOKEN_ENDIF\"},\n\t\t{\"while\", \"TOKEN_WHILE\"},\n\t\t{\"for\", \"TOKEN_FOR\"},\n\t\t{\"print\", \"IDENTIFIER_PRINT\"},\n\t\t{\"return\", \"TOKEN_RETURN\"},\n\t\t{\"exit\", \"TOKEN_EXIT\"},\n\t\t{\"define\", \"TOKEN_DEFINE\"},\n\t\t{\"fn\", \"TOKEN_FUNCTION\"},\n\t\t{\"memes\", \"TOKEN_MEMES\"},\n\t\t{\"int\", \"IDENTIFIER_INT\"},\n\t\t{\"uint\", \"IDENTFIER_INT_U\"},\n\t\t{\"string\", \"IDENTIFIER_STRING\"},\n\t\t{\"char\", \"IDENTIFIER_CHAR\"},\n\t\t{\"bool\", \"IDENTIFIER_BOOL\"},\n\t\t{\"double\", \"IDENTIFIER_DOUBLE\"},\n\t\t{\"float\", \"IDENTIFIER_FLOAT\"},\n\t\t{\"pingu\", \"TOKEN_PINGU\"},\n\t}\n}", "func treatKey(key string) string {\n\tspaces := regexp.MustCompile(`\\s+`)\n\tkey = spaces.ReplaceAllString(key, \"_\")\n\treturn strings.ToLower(key)\n}", "func HashtagDisambiguation(ctx *context.Context) {\n\n\thashtag := ctx.Params(\":hashtag\")\n\tprint(hashtag)\n}", "func Words(prefix, suffix string, words ...string) string {\n\tfor i, word := range words {\n\t\twords[i] = regexp.QuoteMeta(word)\n\t}\n\treturn prefix + `(` + strings.Join(words, `|`) + `)` + suffix\n}", "func (inp inputT) IsReserved() bool {\n\tif inp.Name == \"page\" {\n\t\treturn true\n\t}\n\tif inp.Name == \"lang_code\" {\n\t\treturn true\n\t}\n\treturn false\n}", "func PlaceholderExtension() gval.Language {\n\treturn placeholderExtension\n}", "func isLegalIdentifierPart(c rune) bool {\n\treturn c == '_' ||\n\t\tunicode.In(c, unicode.Lu, unicode.Ll, unicode.Lt, unicode.Lm, unicode.Lo, unicode.Nl, unicode.Mn, unicode.Mc,\n\t\t\tunicode.Nd, unicode.Pc, unicode.Cf)\n}", "func isValidSpecialHeaderLikeVariable(value string) []string {\n\t// underscores in a header-like variable represent '-'.\n\terrMsgs := validation.IsHTTPHeaderName(strings.Replace(value, \"_\", \"-\", -1))\n\tif len(errMsgs) >= 1 || strings.Contains(value, \"-\") {\n\t\treturn []string{\"a valid special variable must consists of alphanumeric characters or '_'\"}\n\t}\n\treturn nil\n}", "func (s *BaselimboListener) EnterString_constant(ctx *String_constantContext) {}", "func AltNames(altNames ...string) OptFunc {\n\treturn func(p *ByName) error {\n\t\tfor _, altName := range altNames {\n\t\t\taltName = strings.TrimSpace(altName)\n\n\t\t\tif err := p.ps.nameCheck(altName, p.whereAdded); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tp.ps.nameToParam[altName] = p\n\t\t\tp.altNames = append(p.altNames, altName)\n\t\t}\n\t\treturn nil\n\t}\n}", "func Test_specialWords(t *testing.T) {\n\n\t// Test true\n\tknownTrueSpecialwords := [...]string{\n\t\t\"exceeding\",\n\t\t\"early\",\n\t\t\"outing\",\n\t}\n\tfor _, word := range knownTrueSpecialwords {\n\t\tif stemmed := stemSpecialWord(word); stemmed == \"\" {\n\t\t\tt.Errorf(\"Expected %v, to be in specialWords\", word)\n\t\t}\n\t}\n\n\t// Test false\n\tknownFalseSpecialwords := [...]string{\n\t\t\"truck\",\n\t\t\"deoxyribonucleic\",\n\t\t\"farse\",\n\t\t\"bullschnizzle\",\n\t}\n\tfor _, word := range knownFalseSpecialwords {\n\t\tif stemmed := stemSpecialWord(word); stemmed != \"\" {\n\t\t\tt.Errorf(\"Expected %v, to NOT be in specialWords\", word)\n\t\t}\n\t}\n}", "func initFormatPlaceholders() {\n\tphfs = map[string]string{\n\t\t\"%{file_with_line}\": \"(%[1]s)\",\n\t\t\"%{duration}\": \"[%.2[2]fms]\",\n\t\t\"%{sql}\": \"%[3]s\",\n\t\t\"%{rows}\": \"[%[4]d rows affected or returned]\",\n\t\t\"%{rows_simple}\": \"[%[5]d]\",\n\t\t\"%{error_msg}\": \"%[6]s\",\n\t}\n}", "func PlaceHolderize(thing []string) string {\n\treturn fmt.Sprintf(\"(%s)\", strings.Join(thing, \"|\"))\n}", "func _(n int) {\n\tre, _ := regexp.Compile(\"const pattern\")\n\tre.FindAllString(\"sdosdos\", -1)\n}", "func (app *generateReference) userNames(name string) []string {\n\tname = strings.ToLower(name)\n\treturn []string{name}\n}", "func (r *REST) ShortNames() []string {\n\treturn []string{\"pvc\"}\n}", "func sanitize(ident string) string {\n\t// TODO: use regex to replace all other characters other than [A-Za-z0-9_]\n\treturn strings.Replace(strings.Trim(ident, \" \"), \" \", \"_\", -1)\n}", "func isStopWord(word string) bool {\n\tswitch word {\n\tcase \"de\", \"la\", \"que\", \"el\", \"en\", \"y\", \"a\", \"los\", \"del\", \"se\", \"las\",\n\t\t\"por\", \"un\", \"para\", \"con\", \"no\", \"una\", \"su\", \"al\", \"lo\", \"como\",\n\t\t\"más\", \"pero\", \"sus\", \"le\", \"ya\", \"o\", \"este\", \"sí\", \"porque\", \"esta\",\n\t\t\"entre\", \"cuando\", \"muy\", \"sin\", \"sobre\", \"también\", \"me\", \"hasta\",\n\t\t\"hay\", \"donde\", \"quien\", \"desde\", \"todo\", \"nos\", \"durante\", \"todos\",\n\t\t\"uno\", \"les\", \"ni\", \"contra\", \"otros\", \"ese\", \"eso\", \"ante\", \"ellos\",\n\t\t\"e\", \"esto\", \"mí\", \"antes\", \"algunos\", \"qué\", \"unos\", \"yo\", \"otro\",\n\t\t\"otras\", \"otra\", \"él\", \"tanto\", \"esa\", \"estos\", \"mucho\", \"quienes\",\n\t\t\"nada\", \"muchos\", \"cual\", \"poco\", \"ella\", \"estar\", \"estas\", \"algunas\",\n\t\t\"algo\", \"nosotros\", \"mi\", \"mis\", \"tú\", \"te\", \"ti\", \"tu\", \"tus\", \"ellas\",\n\t\t\"nosotras\", \"vosostros\", \"vosostras\", \"os\", \"mío\", \"mía\", \"míos\", \"mías\",\n\t\t\"tuyo\", \"tuya\", \"tuyos\", \"tuyas\", \"suyo\", \"suya\", \"suyos\", \"suyas\",\n\t\t\"nuestro\", \"nuestra\", \"nuestros\", \"nuestras\", \"vuestro\", \"vuestra\",\n\t\t\"vuestros\", \"vuestras\", \"esos\", \"esas\", \"estoy\", \"estás\", \"está\", \"estamos\",\n\t\t\"estáis\", \"están\", \"esté\", \"estés\", \"estemos\", \"estéis\", \"estén\", \"estaré\",\n\t\t\"estarás\", \"estará\", \"estaremos\", \"estaréis\", \"estarán\", \"estaría\",\n\t\t\"estarías\", \"estaríamos\", \"estaríais\", \"estarían\", \"estaba\", \"estabas\",\n\t\t\"estábamos\", \"estabais\", \"estaban\", \"estuve\", \"estuviste\", \"estuvo\",\n\t\t\"estuvimos\", \"estuvisteis\", \"estuvieron\", \"estuviera\", \"estuvieras\",\n\t\t\"estuviéramos\", \"estuvierais\", \"estuvieran\", \"estuviese\", \"estuvieses\",\n\t\t\"estuviésemos\", \"estuvieseis\", \"estuviesen\", \"estando\", \"estado\",\n\t\t\"estada\", \"estados\", \"estadas\", \"estad\", \"he\", \"has\", \"ha\", \"hemos\",\n\t\t\"habéis\", \"han\", \"haya\", \"hayas\", \"hayamos\", \"hayáis\", \"hayan\",\n\t\t\"habré\", \"habrás\", \"habrá\", \"habremos\", \"habréis\", \"habrán\", \"habría\",\n\t\t\"habrías\", \"habríamos\", \"habríais\", \"habrían\", \"había\", \"habías\",\n\t\t\"habíamos\", \"habíais\", \"habían\", \"hube\", \"hubiste\", \"hubo\", \"hubimos\",\n\t\t\"hubisteis\", \"hubieron\", \"hubiera\", \"hubieras\", \"hubiéramos\", \"hubierais\",\n\t\t\"hubieran\", \"hubiese\", \"hubieses\", \"hubiésemos\", \"hubieseis\", \"hubiesen\",\n\t\t\"habiendo\", \"habido\", \"habida\", \"habidos\", \"habidas\", \"soy\", \"eres\",\n\t\t\"es\", \"somos\", \"sois\", \"son\", \"sea\", \"seas\", \"seamos\", \"seáis\", \"sean\",\n\t\t\"seré\", \"serás\", \"será\", \"seremos\", \"seréis\", \"serán\", \"sería\", \"serías\",\n\t\t\"seríamos\", \"seríais\", \"serían\", \"era\", \"eras\", \"éramos\", \"erais\",\n\t\t\"eran\", \"fui\", \"fuiste\", \"fue\", \"fuimos\", \"fuisteis\", \"fueron\", \"fuera\",\n\t\t\"fueras\", \"fuéramos\", \"fuerais\", \"fueran\", \"fuese\", \"fueses\", \"fuésemos\",\n\t\t\"fueseis\", \"fuesen\", \"sintiendo\", \"sentido\", \"sentida\", \"sentidos\",\n\t\t\"sentidas\", \"siente\", \"sentid\", \"tengo\", \"tienes\", \"tiene\", \"tenemos\",\n\t\t\"tenéis\", \"tienen\", \"tenga\", \"tengas\", \"tengamos\", \"tengáis\", \"tengan\",\n\t\t\"tendré\", \"tendrás\", \"tendrá\", \"tendremos\", \"tendréis\", \"tendrán\",\n\t\t\"tendría\", \"tendrías\", \"tendríamos\", \"tendríais\", \"tendrían\", \"tenía\",\n\t\t\"tenías\", \"teníamos\", \"teníais\", \"tenían\", \"tuve\", \"tuviste\", \"tuvo\",\n\t\t\"tuvimos\", \"tuvisteis\", \"tuvieron\", \"tuviera\", \"tuvieras\", \"tuviéramos\",\n\t\t\"tuvierais\", \"tuvieran\", \"tuviese\", \"tuvieses\", \"tuviésemos\", \"tuvieseis\",\n\t\t\"tuviesen\", \"teniendo\", \"tenido\", \"tenida\", \"tenidos\", \"tenidas\", \"tened\":\n\t\treturn true\n\t}\n\treturn false\n}" ]
[ "0.59122366", "0.5840971", "0.57974476", "0.57077396", "0.5615138", "0.55619144", "0.5538959", "0.5538959", "0.5526737", "0.5453809", "0.54248405", "0.54105705", "0.53264964", "0.5263973", "0.5154795", "0.51538616", "0.511369", "0.5073781", "0.50569284", "0.50515044", "0.5032897", "0.5018442", "0.5007439", "0.49963832", "0.49944574", "0.49732348", "0.49452198", "0.49435723", "0.49415877", "0.49403954", "0.4938293", "0.49301535", "0.49230728", "0.49106312", "0.4909711", "0.4897054", "0.4872757", "0.4870261", "0.4859515", "0.4856846", "0.48549792", "0.48508033", "0.4844038", "0.48408484", "0.48239812", "0.48180807", "0.48035344", "0.48034003", "0.47909245", "0.4786391", "0.4784299", "0.47814912", "0.47731307", "0.47385582", "0.47289127", "0.47182348", "0.47114277", "0.46841258", "0.46797207", "0.4679234", "0.46781754", "0.46709013", "0.46623245", "0.46566993", "0.4653067", "0.46502146", "0.46491188", "0.46476114", "0.46351552", "0.4630987", "0.4630498", "0.4626283", "0.46260905", "0.46258876", "0.46233836", "0.4622669", "0.4612608", "0.46118054", "0.4602718", "0.46009737", "0.4600845", "0.45999217", "0.4591661", "0.45788452", "0.4578201", "0.45751482", "0.4571503", "0.45664513", "0.456276", "0.45577", "0.45562315", "0.45556477", "0.45538953", "0.45525303", "0.45521465", "0.4551726", "0.4549962", "0.45483625", "0.454267" ]
0.7108631
1
transformToSnakeCase transforms a string LikeThis to a snakecase string like_this
func transformToSnakeCase(input string) string { words := sliceIntoWords(input) // my kingdom for LINQ var lowerWords []string for _, word := range words { lowerWords = append(lowerWords, strings.ToLower(word)) } return strings.Join(lowerWords, "_") }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func transformToSnakeCase(input string) string {\n\twords := sliceIntoWords(input)\n\n\t// my kingdom for LINQ\n\tlowerWords := make([]string, 0, len(words))\n\tfor _, word := range words {\n\t\tif len(word) > 0 {\n\t\t\tlowerWords = append(lowerWords, strings.ToLower(word))\n\t\t}\n\t}\n\n\treturn strings.Join(lowerWords, \"_\")\n}", "func toSnakeCase(str string) string {\n\tsnake := matchAllCap.ReplaceAllString(str, \"${1}_${2}\")\n\treturn strings.ToLower(snake)\n}", "func ToSnakeCase(str string) string {\n\tsnake := matchFirstCap.ReplaceAllString(str, \"${1}_${2}\")\n\tsnake = matchAllCap.ReplaceAllString(snake, \"${1}_${2}\")\n\treturn strings.ToLower(snake)\n}", "func ToSnakeCase(str string) string {\n\tsnake := matchFirstCap.ReplaceAllString(str, \"${1}_${2}\")\n\tsnake = matchAllCap.ReplaceAllString(snake, \"${1}_${2}\")\n\treturn strings.ToLower(snake)\n}", "func ToSnakeCase(str string) string {\n\tsnake := matchFirstCap.ReplaceAllString(str, \"${1}_${2}\")\n\tsnake = matchAllCap.ReplaceAllString(snake, \"${1}_${2}\")\n\treturn strings.ToLower(snake)\n}", "func ToSnakeCase(str string) string {\n\tsnakeCase := matchFirstCap.ReplaceAllString(str, \"${1}_${2}\")\n\tsnakeCase = matchAllCap.ReplaceAllString(snakeCase, \"${1}_${2}\")\n\treturn strings.ToLower(snakeCase)\n}", "func ToSnakeCase(str string) string {\n\tsnake := matchFirstCap.ReplaceAllString(str, \"${1}_${2}\")\n\tsnake = matchAllCap.ReplaceAllString(snake, \"${1}_${2}\")\n\tsnake = strings.Replace(snake, \"__\", \"_\", -1)\n\n\treturn strings.ToLower(snake)\n}", "func ToSnakeCase(str string) string {\n\tmatchFirstCap := regexp.MustCompile(\"(.)([A-Z][a-z]+)\")\n\tmatchAllCap := regexp.MustCompile(\"([a-z0-9])([A-Z])\")\n\tmatchAllSpaces := regexp.MustCompile(\"(\\\\s)\")\n\tcleanUpHack := regexp.MustCompile(\"i_ds\")\n\n\tsnake := matchFirstCap.ReplaceAllString(str, \"${1}_${2}\")\n\tsnake = matchAllCap.ReplaceAllString(snake, \"${1}_${2}\")\n\tsnake = matchAllSpaces.ReplaceAllString(snake, \"_\")\n\tsnake = strings.ToLower(snake)\n\tsnake = cleanUpHack.ReplaceAllString(snake, \"ids\")\n\n\treturn snake\n}", "func ToSnakeCase(s string) string {\n\toutput := \"\"\n\tfor i, c := range s {\n\t\tif i > 0 && isUppercaseRune(c) && output[len(output)-1] != '_' && i < len(s)-1 && !isUppercaseRune(rune(s[i+1])) {\n\t\t\toutput += \"_\" + string(c)\n\t\t} else {\n\t\t\toutput += string(c)\n\t\t}\n\t}\n\treturn output\n}", "func ToSnakeCase(in string) string {\n\trunes := []rune(in)\n\tout := make([]rune, 0, 1+len(runes))\n\tfor i := 0; i < len(runes); i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) && ((i+1 < len(runes) && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t}\n\treturn string(out)\n}", "func CamelToSnake(str string) string {\n\treturn strings.ToLower(match.ReplaceAllString(str, \"${1}_${2}\"))\n}", "func ToSnakeCase(in string) string {\n\trunes := []rune(in)\n\tlength := len(runes)\n\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) &&\n\t\t\t((i+1 < length && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t}\n\n\treturn string(out)\n}", "func ToSnakeCase(s string) string {\n\tvar b strings.Builder\n\tb.Grow(len(s) + 2)\n\tlastWasUpper := true\n\tfor _, r := range s {\n\t\tlr := unicode.ToLower(r)\n\t\tisUpper := lr != r\n\t\tif isUpper && !lastWasUpper {\n\t\t\tb.WriteByte('_')\n\t\t}\n\t\tb.WriteRune(lr)\n\t\tlastWasUpper = isUpper\n\t}\n\treturn b.String()\n}", "func ToSnakeCase(s string) string {\n\trs := []rune(s)\n\tb := strings.Builder{}\n\tb.Grow(len(rs))\n\n\tfor i := 0; i < len(rs); i++ {\n\t\tif IsUpper(rs[i]) { // find first upper char\n\t\t\tbegin, end := i, i\n\t\t\tfor ; i < len(rs); i++ { // find upper chars after first upper char\n\t\t\t\tif IsUpper(rs[i]) {\n\t\t\t\t\tend = i\n\t\t\t\t} else {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif begin != 0 {\n\t\t\t\tb.WriteRune('_')\n\t\t\t}\n\n\t\t\tb.WriteString(strings.ToLower(string(rs[begin : end+1])))\n\t\t\ti--\n\t\t\tcontinue\n\t\t}\n\t\tb.WriteRune(rs[i])\n\t}\n\n\treturn b.String()\n}", "func ToSnakeCase(s string) string {\n\ts = ToCamelCase(s)\n\trunes := []rune(s)\n\tlength := len(runes)\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t\tif i+1 < length && (unicode.IsUpper(runes[i+1]) && unicode.IsLower(runes[i])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t}\n\n\treturn string(out)\n}", "func ToSnakeCase(s string) string {\n\ts = ToCamelCase(s)\n\trunes := []rune(s)\n\tlength := len(runes)\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t\tif i+1 < length && (unicode.IsUpper(runes[i+1]) && unicode.IsLower(runes[i])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t}\n\n\treturn string(out)\n}", "func SnakeCase(text string) string {\n\tresult := snakeRegexp.ReplaceAllStringFunc(text, func(match string) string {\n\t\treturn \"_\" + match\n\t})\n\n\treturn ParameterizeString(result)\n}", "func ToSnake(s string) string {\n\tdel := uint8('_')\n\trunLen := 0\n\ts = strings.Trim(s, \" \")\n\tn := \"\"\n\tfor i, v := range s {\n\t\t// treat acronyms as words, eg for JSONData -> JSON is a whole word\n\t\tnextCaseIsChanged := false\n\t\tif i+1 < len(s) {\n\t\t\tnext := s[i+1]\n\t\t\tif (v >= 'A' && v <= 'Z' && next >= 'a' && next <= 'z') || (v >= 'a' && v <= 'z' && next >= 'A' && next <= 'Z') {\n\t\t\t\tnextCaseIsChanged = true\n\t\t\t}\n\t\t}\n\n\t\tif i > 0 && n[len(n)-1] != del && nextCaseIsChanged && runLen > 2 {\n\t\t\t// add underscore if next letter case type is changed\n\t\t\tif v >= 'A' && v <= 'Z' {\n\t\t\t\tn += string(del) + string(v)\n\t\t\t} else if v >= 'a' && v <= 'z' {\n\t\t\t\tn += string(v) + string(del)\n\t\t\t}\n\t\t\trunLen = 0\n\t\t} else if v == ' ' || v == '_' || v == '.' || v == ':' || v == '-' || v == '$' {\n\t\t\t// replace spaces/underscores with delimiters\n\t\t\tn += string(del)\n\t\t\trunLen = 0\n\t\t} else {\n\t\t\tn = n + string(v)\n\t\t\trunLen++\n\t\t}\n\t}\n\n\tn = strings.ToLower(n)\n\n\t// avoid conflict with Terraform keywords\n\tif IsKeyword(n) {\n\t\tn = \"_\" + n\n\t}\n\n\treturn n\n}", "func ToSnakeCase(str string) string {\n\tvar output []rune\n\tvar segment []rune\n\tfor _, r := range str {\n\t\tif !unicode.IsLower(r) {\n\t\t\toutput = addSegment(output, segment)\n\t\t\tsegment = nil\n\t\t}\n\t\tsegment = append(segment, unicode.ToLower(r))\n\t}\n\toutput = addSegment(output, segment)\n\treturn string(output)\n}", "func ToSnakeCase(s string) string {\n\n\t// Convert.\n\tresult := strings.Map(\n\t\tfunc(r rune) rune {\n\n\t\t\tif whitespace.IsWhitespace(r) || '-' == r {\n\t\t\t\treturn '_'\n\t\t\t} else {\n\t\t\t\treturn unicode.ToLower(r)\n\t\t\t}\n\t\t},\n\t\ts)\n\n\t// Return\n\treturn result\n}", "func ToSnakeCase(camelCased string, lowerCaseResult bool) string {\n\n\tvar result string\n\tvar currentUpperCase bool = false\n\tcamelCased = adjustCamelCase(camelCased)\n\tcurrentChar := \"\"\nloop:\n\tfor i, char := range camelCased {\n\t\tcharString := string(char)\n\t\tif charString == \"_\" {\n\t\t\tresult += charString\n\t\t\tcurrentChar = charString\n\t\t\tcontinue loop\n\t\t}\n\n\t\tif i > 0 && isUpperCase(charString) && currentUpperCase == false {\n\t\t\tcurrentUpperCase = true\n\t\t\tif i > 1 && currentChar != \"_\" {\n\t\t\t\tresult += \"_\"\n\t\t\t}\n\n\t\t} else {\n\t\t\tcurrentUpperCase = false\n\t\t}\n\n\t\tcharLower := strings.ToLower(charString)\n\t\tif 0 == i && lowerCaseResult {\n\t\t\tresult += strings.ToLower(charString)\n\t\t} else {\n\t\t\tif 0 == i {\n\t\t\t\tresult += charString\n\t\t\t} else {\n\t\t\t\tresult += (charLower)\n\t\t\t}\n\n\t\t}\n\t\tcurrentChar = charString\n\n\t}\n\n\tif lowerCaseResult {\n\t\tresult = strings.ToLower(result)\n\t}\n\n\tlog.Println(\"SNAKE CASED from: \", camelCased, \"-->\", result, \";lowerCaseResult(\", lowerCaseResult, \")\")\n\n\treturn result\n\n}", "func ToSnakeCase(name string) string {\n\tparts := nameParts(name)\n\tfor i := range parts {\n\t\tparts[i] = strings.ToLower(parts[i])\n\t}\n\treturn strings.Join(parts, \"_\")\n}", "func ToSnakeCase(in string) string {\n\trunes := []rune(in)\n\tvar out []rune\n\tfor i := 0; i < len(runes); i++ {\n\t\tif i > 0 && (unicode.IsUpper(runes[i]) || unicode.IsNumber(runes[i])) && ((i+1 < len(runes) && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t}\n\treturn string(out)\n}", "func ToLowerSnakeCase(s string) string {\n\treturn strings.ToLower(ToSnakeCase(s))\n}", "func camelToSnake(s string) string {\n\ts = strings.TrimPrefix(upperLower.ReplaceAllString(s, `_${0}`), \"_\")\n\ts = lowerUpper.ReplaceAllString(s, `${1}_${2}`)\n\ts = strings.ToLower(s)\n\treturn s\n}", "func SnakeToLowerCamel(original string) string {\n\treturn toCamel(original, snakeDelimiter, false)\n}", "func toSnake(in string) string {\n\trunes := []rune(in)\n\tlength := len(runes)\n\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) && ((i+1 < length && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t}\n\n\treturn string(out)\n}", "func toSnake(in string) string {\n\trunes := []rune(in)\n\tlength := len(runes)\n\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) && ((i+1 < length && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t}\n\n\treturn string(out)\n}", "func SnakeCase(str string) string {\n\treturn xstrings.ToSnakeCase(str)\n}", "func FromSnakeCase(s string) string {\n\treturn strings.Replace(s, \"_\", \" \", -1)\n}", "func (c *Config) CamelToSnake(camel string) string {\n snake := matchAllCap.ReplaceAllString(camel, \"${1}_${2}\")\n return strings.ToLower(snake)\n}", "func toLowerCamelCase(s string) string {\n\treturn toCamelInitCase(s, false)\n}", "func SnakeCase(in string) string {\n\trunes := []rune(in)\n\tlength := len(runes)\n\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) && ((i+1 < length && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t}\n\n\treturn string(out)\n}", "func camelToSnake(str string) string {\n\tvar snakeName bytes.Buffer\n\tfor i, char := range str {\n\t\tif unicode.IsUpper(char) {\n\t\t\tif i > 0 {\n\t\t\t\tsnakeName.WriteRune('_')\n\t\t\t}\n\t\t\tsnakeName.WriteRune(unicode.ToLower(char))\n\t\t} else {\n\t\t\tsnakeName.WriteRune(char)\n\t\t}\n\t}\n\treturn snakeName.String()\n}", "func ToLowerSnake(s string) (string, bool) {\n\tsnake := strcase.ToSnake(strings.ToLower(s))\n\treturn snake, s == snake\n}", "func toSnake(name string) string {\n\tbuf := bytes.Buffer{}\n\tfor i := 0; i < len(name); i++ {\n\t\tc := rune(name[i])\n\t\tif unicode.IsUpper(c) {\n\t\t\tif i > 0 && (i+1 >= len(name) || unicode.IsLower(rune(name[i+1]))) {\n\t\t\t\tbuf.WriteRune('_')\n\t\t\t}\n\t\t\tbuf.WriteRune(c)\n\t\t} else {\n\t\t\tbuf.WriteRune(unicode.ToUpper(c))\n\t\t}\n\t}\n\treturn buf.String()\n}", "func ToSnake(in string) string {\n\trunes := []rune(in)\n\tlength := len(runes)\n\n\tvar out []rune\n\tfor i := 0; i < length; i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) && ((i+1 < length && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tout = append(out, '_')\n\t\t}\n\t\tout = append(out, unicode.ToLower(runes[i]))\n\t}\n\treturn string(out)\n}", "func ToSnake(s string) string {\n\treturn snaker(s, '_', unicode.ToLower, unicode.ToLower, unicode.ToLower)\n}", "func SnakeCase(s string) string {\n\tvar buf bytes.Buffer\n\tfor i, r := range s {\n\t\tif unicode.IsUpper(r) && i > 0 && s[i-1] != '_' {\n\t\t\tfmt.Fprintf(&buf, \"_\")\n\t\t}\n\t\tr = unicode.ToLower(r)\n\t\tfmt.Fprintf(&buf, \"%c\", r)\n\t}\n\treturn buf.String()\n}", "func SnakeCase(s string) string {\n\treturn Delimit(s, '_')\n}", "func LowerSnakeCase(s string) string {\n\tparts := []string{}\n\tfor _, part := range SplitSymbol(s) {\n\t\tif part == \"\" {\n\t\t\tparts = append(parts, \"_\")\n\t\t\tcontinue\n\t\t}\n\t\tparts = append(parts, strings.ToLower(part))\n\t}\n\treturn strings.Join(parts, \"_\")\n}", "func CamelToSnakeUnderscore(camelCase string) string {\n\tmatchFirstCap := regexp.MustCompile(\"(.)([A-Z][a-z]+)\")\n\tmatchAllCap := regexp.MustCompile(\"([a-z0-9])([A-Z])\")\n\n\tsnakeUnderscore := matchFirstCap.ReplaceAllString(camelCase, \"${1}_${2}\")\n\tsnakeUnderscore = matchAllCap.ReplaceAllString(snakeUnderscore, \"${1}_${2}\")\n\treturn strings.ToUpper(snakeUnderscore)\n}", "func Snake2Camel(s string, exported bool) (retVal string) {\n\tnextUpper := exported\n\tfor i, v := range s {\n\t\tswitch {\n\t\tcase unicode.IsNumber(v):\n\t\t\tretVal += string(v)\n\t\tcase unicode.IsUpper(v):\n\t\t\tif i == 0 && !nextUpper {\n\t\t\t\tretVal += strings.ToLower(string(v))\n\t\t\t} else {\n\t\t\t\tretVal += string(v)\n\t\t\t}\n\t\tcase unicode.IsLower(v):\n\t\t\tif nextUpper {\n\t\t\t\tretVal += strings.ToUpper(string(v))\n\t\t\t} else {\n\t\t\t\tretVal += string(v)\n\t\t\t}\n\t\tcase v == '_':\n\t\t\tnextUpper = true\n\t\t\tcontinue\n\t\tdefault:\n\t\t\tretVal += string(v)\n\t\t}\n\t\tnextUpper = false\n\t}\n\treturn\n}", "func ScreamingSnakeRenamer() Renamer {\n\treturn strcase.ToScreamingSnake\n}", "func ToLowerCamel(s string) string {\n\treturn snaker(s, rune(0), unicode.ToLower, unicode.ToUpper, noop)\n}", "func snakeToCamel(s string) string {\n\tvar result string\n\n\twords := strings.Split(s, \"_\")\n\n\tfor _, word := range words {\n\t\tif upper := strings.ToUpper(word); commonInitialisms[upper] {\n\t\t\tresult += upper\n\t\t\tcontinue\n\t\t}\n\n\t\tif len(word) > 0 {\n\t\t\tw := []rune(word)\n\t\t\tw[0] = unicode.ToUpper(w[0])\n\t\t\tresult += string(w)\n\t\t}\n\t}\n\n\treturn result\n}", "func ConstNameToAllCapsSnake(name string) string {\n\tparts := nameParts(RemoveLeadingK(name))\n\tfor i := range parts {\n\t\tparts[i] = strings.ToUpper(parts[i])\n\t}\n\treturn strings.Join(parts, \"_\")\n}", "func SnakeString(s string) string {\n\tdata := make([]byte, 0, len(s)*2)\n\tj := false\n\tfor _, d := range StringToBytes(s) {\n\t\tif d >= 'A' && d <= 'Z' {\n\t\t\tif j {\n\t\t\t\tdata = append(data, '_')\n\t\t\t\tj = false\n\t\t\t}\n\t\t} else if d != '_' {\n\t\t\tj = true\n\t\t}\n\t\tdata = append(data, d)\n\t}\n\treturn strings.ToLower(BytesToString(data))\n}", "func (sc SnakeCaseConvention) Convert(name string) string {\n\trunes := []rune(name)\n\tn := len(runes)\n\tvar buf bytes.Buffer\n\n\tfor i := 0; i < n; i++ {\n\t\tif i > 0 && unicode.IsUpper(runes[i]) && ((i+1 < n && unicode.IsLower(runes[i+1])) || unicode.IsLower(runes[i-1])) {\n\t\t\tbuf.WriteRune('_')\n\t\t}\n\t\tbuf.WriteRune(unicode.ToLower(runes[i]))\n\t}\n\n\treturn buf.String()\n}", "func CamelToSnake(original string) string {\n\treturn fromCamel(original, snakeDelimiter)\n}", "func snakeCasedNameOld(name string) string {\n\tnewstr := make([]rune, 0)\n\tfor idx, chr := range name {\n\t\tif isUpper := 'A' <= chr && chr <= 'Z'; isUpper {\n\t\t\tif idx > 0 {\n\t\t\t\tnewstr = append(newstr, '_')\n\t\t\t}\n\t\t\tchr -= ('A' - 'a')\n\t\t}\n\t\tnewstr = append(newstr, chr)\n\t}\n\n\treturn string(newstr)\n}", "func ToSnake(s string) string {\n\treturn toSeparated(s, '_', false)\n}", "func SnakeString(s string) string {\n\tdata := make([]byte, 0, len(s)*2)\n\tj := false\n\tnum := len(s)\n\tfor i := 0; i < num; i++ {\n\t\td := s[i]\n\t\tif i > 0 && d >= 'A' && d <= 'Z' && j {\n\t\t\tdata = append(data, '_')\n\t\t}\n\t\tif d != '_' {\n\t\t\tj = true\n\t\t}\n\t\tdata = append(data, d)\n\t}\n\treturn strings.ToLower(string(data[:]))\n}", "func SnakeString(s string) string {\n\tdata := make([]byte, 0, len(s)*2)\n\tj := false\n\tnum := len(s)\n\tfor i := 0; i < num; i++ {\n\t\td := s[i]\n\t\tif i > 0 && d >= 'A' && d <= 'Z' && j {\n\t\t\tdata = append(data, '_')\n\t\t}\n\t\tif d != '_' {\n\t\t\tj = true\n\t\t}\n\t\tdata = append(data, d)\n\t}\n\treturn strings.ToLower(string(data[:]))\n}", "func SnakeString(s string) string {\n\tdata := make([]byte, 0, len(s)*2)\n\tj := false\n\tnum := len(s)\n\tfor i := 0; i < num; i++ {\n\t\td := s[i]\n\t\tif i > 0 && d >= 'A' && d <= 'Z' && j {\n\t\t\tdata = append(data, '_')\n\t\t}\n\t\tif d != '_' {\n\t\t\tj = true\n\t\t}\n\t\tdata = append(data, d)\n\t}\n\treturn strings.ToLower(string(data[:]))\n}", "func IsLowerSnakeCase(s string, extraRunes ...rune) bool {\n\tif s == \"\" {\n\t\treturn false\n\t}\n\tif s[0] == '_' {\n\t\treturn false\n\t}\n\tif s[len(s)-1] == '_' {\n\t\treturn false\n\t}\n\tfor _, c := range s {\n\t\tif !((c >= 'a' && c <= 'z') || (c >= '0' && c <= '9') || c == '_') && !containsRune(c, extraRunes) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func snakeString(s string) string {\n\tdata := make([]byte, 0, len(s)*2)\n\tj := false\n\tnum := len(s)\n\tfor i := 0; i < num; i++ {\n\t\td := s[i]\n\t\tif i > 0 && d >= 'A' && d <= 'Z' && j {\n\t\t\tdata = append(data, '_')\n\t\t}\n\t\tif d != '_' {\n\t\t\tj = true\n\t\t}\n\t\tdata = append(data, d)\n\t}\n\treturn strings.ToUpper(string(data[:]))\n}", "func TrainCase(t string) string {\n\treturn KebabCase(t, true, false)\n}", "func GetSnakeCaseName(s string) string {\n\t// order of operations matters here\n\t// PickupLocation -> pickup_location\n\treturn strings.ToLower(strcase.ToSnake(s))\n}", "func toCamelCase(s string) string {\n\treturn toCamelInitCase(s, true)\n}", "func ToLowerCamel(s string) string {\n\tif s == \"\" {\n\t\treturn s\n\t}\n\tif uppercaseAcronym[s] {\n\t\ts = strings.ToLower(s)\n\t}\n\tif r := rune(s[0]); r == '_' {\n\t\ts = s[1:]\n\t}\n\tif r := rune(s[0]); r >= 'A' && r <= 'Z' {\n\t\ts = strings.ToLower(string(r)) + s[1:]\n\t}\n\treturn toCamelInitCase(s, false)\n}", "func ToLowerCamel(s string) string {\n\treturn ToLowerCamelWithInitialisms(s, CommonInitialisms)\n}", "func SnakeCasedName(name string) string {\n\tnewstr := make([]byte, 0, len(name)+1)\n\tfor i := 0; i < len(name); i++ {\n\t\tc := name[i]\n\t\tif isUpper := 'A' <= c && c <= 'Z'; isUpper {\n\t\t\tif i > 0 {\n\t\t\t\tnewstr = append(newstr, '_')\n\t\t\t}\n\t\t\tc += 'a' - 'A'\n\t\t}\n\t\tnewstr = append(newstr, c)\n\t}\n\n\treturn bytesconv.BytesToStr(newstr)\n}", "func Snake(name string) string {\n\tvar ret bytes.Buffer\n\n\tmultipleUpper := false\n\tvar lastUpper rune\n\tvar beforeUpper rune\n\n\tfor _, c := range name {\n\t\t// Non-lowercase character after uppercase is considered to be uppercase too.\n\t\tisUpper := (unicode.IsUpper(c) || (lastUpper != 0 && !unicode.IsLower(c)))\n\n\t\tif lastUpper != 0 {\n\t\t\t// Output a delimiter if last character was either the\n\t\t\t// first uppercase character in a row, or the last one\n\t\t\t// in a row (e.g. 'S' in \"HTTPServer\"). Do not output\n\t\t\t// a delimiter at the beginning of the name.\n\t\t\tfirstInRow := !multipleUpper\n\t\t\tlastInRow := !isUpper\n\n\t\t\tif ret.Len() > 0 && (firstInRow || lastInRow) && beforeUpper != '_' {\n\t\t\t\tret.WriteByte('_')\n\t\t\t}\n\n\t\t\tret.WriteRune(unicode.ToLower(lastUpper))\n\t\t}\n\n\t\t// Buffer uppercase char, do not output it yet as a delimiter\n\t\t// may be required if the next character is lowercase.\n\t\tif isUpper {\n\t\t\tmultipleUpper = (lastUpper != 0)\n\t\t\tlastUpper = c\n\t\t\tcontinue\n\t\t}\n\n\t\tret.WriteRune(c)\n\t\tlastUpper = 0\n\t\tbeforeUpper = c\n\t\tmultipleUpper = false\n\t}\n\n\tif lastUpper != 0 {\n\t\tret.WriteRune(unicode.ToLower(lastUpper))\n\t}\n\n\treturn ret.String()\n}", "func SnakeToCamel(s string) string {\n\tif !strings.Contains(s, \"_\") {\n\t\treturn s\n\t}\n\n\tvar result string\n\tparts := strings.Split(s, \"_\")\n\tcapitalize := false\n\tfor _, p := range parts {\n\t\tif p == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tif !capitalize {\n\t\t\tresult += p\n\t\t\tcapitalize = true\n\t\t} else {\n\t\t\tresult += strings.Title(p)\n\t\t}\n\t}\n\treturn result\n}", "func CamelToSnake(s string) string {\n\tb := buffer{\n\t\tr: make([]byte, 0, len(s)),\n\t}\n\tvar m rune\n\tvar w bool\n\tfor _, ch := range s {\n\t\tif unicode.IsUpper(ch) {\n\t\t\tif m != 0 {\n\t\t\t\tif !w {\n\t\t\t\t\tb.indent()\n\t\t\t\t\tw = true\n\t\t\t\t}\n\t\t\t\tb.write(m)\n\t\t\t}\n\t\t\tm = unicode.ToLower(ch)\n\t\t} else {\n\t\t\tif m != 0 {\n\t\t\t\tb.indent()\n\t\t\t\tb.write(m)\n\t\t\t\tm = 0\n\t\t\t\tw = false\n\t\t\t}\n\t\t\tb.write(ch)\n\t\t}\n\t}\n\tif m != 0 {\n\t\tif !w {\n\t\t\tb.indent()\n\t\t}\n\t\tb.write(m)\n\t}\n\treturn string(b.r)\n}", "func SnakeString2(s string) string {\n\tdata := make([]byte, 0, len(s)*2)\n\tj := false\n\tfor _, d := range StringToBytes(s) {\n\t\tif d >= 'A' && d <= 'Z' {\n\t\t\tif j {\n\t\t\t\tdata = append(data, '_')\n\t\t\t\tj = false\n\t\t\t}\n\t\t} else if d != '_' {\n\t\t\tj = true\n\t\t}\n\t\tdata = append(data, d)\n\t}\n\treturn strings.ToLower(BytesToString(data))\n}", "func ToLowerCamelWithInitialisms(s string, initialisms map[string]bool) string {\n\tif s == \"\" {\n\t\treturn s\n\t}\n\tif initialisms == nil {\n\t\tinitialisms = map[string]bool{}\n\t}\n\tss := SplitIntoWordsWithInitialisms(s, initialisms)\n\tfor i, s := range ss {\n\t\tif i == 0 {\n\t\t\tss[i] = strings.ToLower(s)\n\t\t\tcontinue\n\t\t}\n\t\tif initialisms[strings.ToLower(s)] {\n\t\t\tss[i] = strings.ToUpper(s)\n\t\t\tcontinue\n\t\t}\n\t\tif strings.ToLower(s[len(s)-1:]) == \"s\" && initialisms[strings.ToLower(s[:len(s)-1])] {\n\t\t\tss[i] = strings.ToUpper(s[:len(s)-1]) + \"s\"\n\t\t\tcontinue\n\t\t}\n\t\tss[i] = strings.ToUpper(s[:1]) + strings.ToLower(s[1:])\n\t}\n\treturn strings.Join(ss, \"\")\n}", "func ToUpperSnake(s string) string {\n\treturn snaker(s, '_', unicode.ToUpper, unicode.ToUpper, unicode.ToUpper)\n}", "func (b *Base) UseSnakeCase() Serializer {\n\treturn b.ConvertKeys(xstrings.ToSnakeCase)\n}", "func Camelize(text string) string {\n\tnewText := camelRegex.ReplaceAllStringFunc(text, func(s string) string {\n\t\treturn upper(s[1])\n\t})\n\n\t// ensure first letter is lower case\n\treturn strings.ToLower(string(newText[0])) + newText[1:]\n}", "func fixForInitialismCase(s string) string {\n\tfor _, initialism := range initialismList {\n\t\tparts := strings.Split(s, \"_\")\n\t\tfor i, part := range parts {\n\t\t\tif part == initialism {\n\t\t\t\tparts[i] = strings.ToUpper(initialism)\n\t\t\t}\n\t\t}\n\t\ts = strings.Join(parts, \"_\")\n\t}\n\treturn s\n}", "func ToLowerCamelCase(in string) string {\n\tout := toCamelCase([]rune(in))\n\tlength := len(out)\n\tfor i := 0; i < length; i++ {\n\t\tisUpper := unicode.IsUpper(out[i])\n\t\tif isUpper && (i == 0 || i+1 == length || unicode.IsUpper(out[i+1])) {\n\t\t\tout[i] -= 'A' - 'a'\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\treturn string(out)\n}", "func (tokens Tokens) Snake() string {\n\treturn strings.Join(tokens.Map(strings.ToLower), \"_\")\n}", "func CamelCase(in string) string {\n\ttokens := strings.Split(in, \"_\")\n\tfor i := range tokens {\n\t\ttokens[i] = strings.Title(strings.Trim(tokens[i], \" \"))\n\t}\n\treturn strings.Join(tokens, \"\")\n}", "func SnakeToUpperCamel(original string) string {\n\treturn toCamel(original, snakeDelimiter, true)\n}", "func lowCamelName(s string) string {\n\ts = gogen.CamelCase(s)\n\tnew := []rune(s)\n\tif len(new) < 1 {\n\t\treturn s\n\t}\n\trv := []rune{}\n\trv = append(rv, unicode.ToLower(new[0]))\n\trv = append(rv, new[1:]...)\n\treturn string(rv)\n}", "func ToCamelCase(in string) string {\n\trunes := []rune(in)\n\tout := make([]rune, 0, len(runes))\n\tup := true\n\tfor i := 0; i < len(runes); i++ {\n\t\tr := runes[i]\n\t\tif r == '_' {\n\t\t\tup = true\n\t\t} else {\n\t\t\tif up {\n\t\t\t\tr = unicode.ToUpper(r)\n\t\t\t\tup = false\n\t\t\t}\n\t\t\tout = append(out, r)\n\t\t}\n\t}\n\treturn string(out)\n}", "func KebabToLowerCamel(original string) string {\n\treturn toCamel(original, kebabDelimiter, false)\n}", "func toCamelInitCase(s string, initCase bool) string {\n\ts = addWordBoundariesToNumbers(s)\n\ts = strings.Trim(s, \" \")\n\tn := \"\"\n\tcapNext := initCase\n\tfor _, v := range s {\n\t\tif v >= 'A' && v <= 'Z' {\n\t\t\tn += string(v)\n\t\t}\n\t\tif v >= '0' && v <= '9' {\n\t\t\tn += string(v)\n\t\t}\n\t\tif v >= 'a' && v <= 'z' {\n\t\t\tif capNext {\n\t\t\t\tn += strings.ToUpper(string(v))\n\t\t\t} else {\n\t\t\t\tn += string(v)\n\t\t\t}\n\t\t}\n\t\tif v == '_' || v == ' ' || v == '-' {\n\t\t\tcapNext = true\n\t\t} else {\n\t\t\tcapNext = false\n\t\t}\n\t}\n\treturn n\n}", "func toUnderscore(s string) string {\n\tif s == \"\" {\n\t\treturn \"\"\n\t}\n\tresult := make([]rune, 0, len(s))\n\n\tresult = append(result, unicode.ToLower(rune(s[0])))\n\tfor _, r := range s[1:] {\n\t\tif unicode.ToUpper(r) == r {\n\t\t\tresult = append(result, '_', unicode.ToLower(r))\n\t\t\tcontinue\n\t\t}\n\t\tresult = append(result, r)\n\t}\n\treturn string(result)\n}", "func ToLowerCamelCase(name string) string {\n\tparts := nameParts(name)\n\tfor i := range parts {\n\t\tif i == 0 {\n\t\t\tparts[i] = strings.ToLower(parts[i])\n\t\t} else {\n\t\t\tparts[i] = strings.Title(strings.ToLower(parts[i]))\n\t\t}\n\t\tif parts[i] == \"\" {\n\t\t\tparts[i] = \"_\"\n\t\t}\n\t}\n\treturn strings.Join(parts, \"\")\n}", "func toLower(s string) string {\n\treturn strings.ToLower(s)\n}", "func ToFriendlyCase(name string) string {\n\tparts := nameParts(name)\n\tfor i := range parts {\n\t\tparts[i] = strings.ToLower(parts[i])\n\t}\n\treturn strings.Join(parts, \" \")\n}", "func toLowerAndFormat(args []string) string {\n\tvar endStr []string\n\tfor _, word := range args {\n\t\tendStr = append(endStr, strings.ToLower(word))\n\t}\n\tif len(args) > 1 {\n\t\treturn strings.Join(endStr, \"_\")\n\t}\n\treturn endStr[0]\n}", "func transformFirstName(input string) string {\n\treturn strings.Replace(toLowerAndTrim(input), \" \", \"-\", -1)\n}", "func ActiveSnakeCase() {\n\tmarshaler = func(v interface{}) ([]byte, error) {\n\t\tmarshaler := conjson.NewMarshaler(v, transform.ConventionalKeys())\n\t\treturn json.MarshalIndent(marshaler, \"\", \" \")\n\t}\n}", "func LowerCamelCase(s string) string {\n\tfirst := true\n\tparts := []string{}\n\tfor _, part := range SplitSymbol(s) {\n\t\tif part == \"\" {\n\t\t\tparts = append(parts, \"_\")\n\t\t\tcontinue\n\t\t}\n\t\tif first {\n\t\t\tparts = append(parts, strings.ToLower(part))\n\t\t\tfirst = false\n\t\t} else {\n\t\t\t// Merge trailing s\n\t\t\tif part == \"s\" && len(parts) > 0 {\n\t\t\t\tparts[len(parts)-1] += part\n\t\t\t} else {\n\t\t\t\tif commonInitialisms[strings.ToUpper(part)] {\n\t\t\t\t\tpart = strings.ToUpper(part)\n\t\t\t\t} else {\n\t\t\t\t\tpart = title(part)\n\t\t\t\t}\n\t\t\t\tparts = append(parts, part)\n\t\t\t}\n\t\t}\n\t}\n\treturn strings.Join(parts, \"\")\n}", "func ToLowerCamel(s string) (string, bool) {\n\tcamel := strcase.ToLowerCamel(s)\n\tsnake := strcase.ToSnake(camel)\n\treturn camel, s == snake\n}", "func ToTrain(s string) string {\n\treturn snaker(s, '-', unicode.ToUpper, unicode.ToUpper, noop)\n}", "func ToLowerFirstCamelCase(s string) string {\n\tif s == \"\" {\n\t\treturn s\n\t}\n\tif len(s) == 1 {\n\t\treturn strings.ToLower(string(s[0]))\n\t}\n\treturn strings.ToLower(string(s[0])) + ToCamelCase(s)[1:]\n}", "func toLower(tk token.Token) token.Token {\n\ts := strings.ToLower(tk.Text())\n\treturn token.UpdateText(tk, s)\n}", "func ToUnderScore(name string) string {\n\tl := len(name)\n\tss := strings.Split(name, \"\")\n\n\t// we just care about the key of idx map,\n\t// the value of map is meaningless\n\tidx := make(map[int]int, 1)\n\n\tvar rs []rune\n\tfor _, s := range name {\n\t\trs = append(rs, []rune(string(s))...)\n\t}\n\n\tfor i := l - 1; i >= 0; {\n\t\tif unicode.IsUpper(rs[i]) {\n\t\t\tvar start, end int\n\t\t\tend = i\n\t\t\tj := i - 1\n\t\t\tfor ; j >= 0; j-- {\n\t\t\t\tif unicode.IsLower(rs[j]) {\n\t\t\t\t\tstart = j + 1\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\t// handle the case: \"BBC\" or \"AaBBB\" case\n\t\t\tif end == l-1 {\n\t\t\t\tidx[start] = 1\n\t\t\t} else {\n\t\t\t\tif start == end {\n\t\t\t\t\t// value=1 is meaningless\n\t\t\t\t\tidx[start] = 1\n\t\t\t\t} else {\n\t\t\t\t\tidx[start] = 1\n\t\t\t\t\tidx[end] = 1\n\t\t\t\t}\n\t\t\t}\n\t\t\ti = j - 1\n\t\t} else {\n\t\t\ti--\n\t\t}\n\t}\n\n\tfor i := l - 1; i >= 0; i-- {\n\t\tss[i] = strings.ToLower(ss[i])\n\t\tif _, ok := idx[i]; ok && i != 0 {\n\t\t\tss = append(ss[0:i], append([]string{\"_\"}, ss[i:]...)...)\n\t\t}\n\t}\n\n\treturn strings.Join(ss, \"\")\n}", "func doLintName(name string, initialisms *names.Initials) (should string) {\n\t// Fast path for simple cases: \"_\" and all lowercase.\n\tif name == \"_\" {\n\t\treturn name\n\t}\n\tallLower := true\n\tfor _, r := range name {\n\t\tif !unicode.IsLower(r) {\n\t\t\tallLower = false\n\t\t\tbreak\n\t\t}\n\t}\n\tif allLower {\n\t\treturn name\n\t}\n\n\t// Split camelCase at any lower->upper transition, and split on underscores.\n\t// Check each word for common initialisms.\n\trunes := []rune(name)\n\tw, i := 0, 0 // index of start of word, scan\n\tfor i+1 <= len(runes) {\n\t\teow := false // whether we hit the end of a word\n\t\tif i+1 == len(runes) {\n\t\t\teow = true\n\t\t} else if runes[i+1] == '_' {\n\t\t\t// underscore; shift the remainder forward over any run of underscores\n\t\t\teow = true\n\t\t\tn := 1\n\t\t\tfor i+n+1 < len(runes) && runes[i+n+1] == '_' {\n\t\t\t\tn++\n\t\t\t}\n\n\t\t\t// Leave at most one underscore if the underscore is between two digits\n\t\t\tif i+n+1 < len(runes) && unicode.IsDigit(runes[i]) && unicode.IsDigit(runes[i+n+1]) {\n\t\t\t\tn--\n\t\t\t}\n\n\t\t\tcopy(runes[i+1:], runes[i+n+1:])\n\t\t\trunes = runes[:len(runes)-n]\n\t\t} else if unicode.IsLower(runes[i]) && !unicode.IsLower(runes[i+1]) {\n\t\t\t// lower->non-lower\n\t\t\teow = true\n\t\t}\n\t\ti++\n\t\tif !eow {\n\t\t\tcontinue\n\t\t}\n\n\t\t// [w,i) is a word.\n\t\tword := string(runes[w:i])\n\t\tif u := strings.ToUpper(word); initialisms.Has(u) {\n\t\t\t// Keep consistent case, which is lowercase only at the start.\n\t\t\tif w == 0 && unicode.IsLower(runes[w]) {\n\t\t\t\tu = strings.ToLower(u)\n\t\t\t}\n\t\t\t// All the common initialisms are ASCII,\n\t\t\t// so we can replace the bytes exactly.\n\t\t\tcopy(runes[w:], []rune(u))\n\t\t} else if w > 0 && strings.ToLower(word) == word {\n\t\t\t// already all lowercase, and not the first word, so uppercase the first character.\n\t\t\trunes[w] = unicode.ToUpper(runes[w])\n\t\t}\n\t\tw = i\n\t}\n\treturn string(runes)\n}", "func CamelcaseToUnderscore(str string) string {\n\treturn CamelcaseToLower(str, \"_\")\n}", "func ToCamelCase(s string) string {\n\tbyteSrc := []byte(s)\n\tchunks := rxCameling.FindAll(byteSrc, -1)\n\tfor idx, val := range chunks {\n\t\tchunks[idx] = bytes.Title(val)\n\t}\n\treturn string(bytes.Join(chunks, nil))\n}", "func ToCamelCase(s string) string {\n\tbyteSrc := []byte(s)\n\tchunks := rxCameling.FindAll(byteSrc, -1)\n\tfor idx, val := range chunks {\n\t\tchunks[idx] = bytes.Title(val)\n\t}\n\treturn string(bytes.Join(chunks, nil))\n}", "func LowerCamelCase(s string) string {\n\tif s == \"\" {\n\t\treturn \"\"\n\t}\n\trunes := []rune(s)\n\treturn string(append([]rune{unicode.ToLower(runes[0])}, runes[1:]...))\n}", "func CamelCase(s string) string {\n\tif s == \"\" {\n\t\treturn \"\"\n\t}\n\tt := make([]byte, 0, 32)\n\ti := 0\n\tif s[0] == '_' {\n\t\t// Need a capital letter; drop the '_'.\n\t\tt = append(t, 'X')\n\t\ti++\n\t}\n\t// Invariant: if the next letter is lower case, it must be converted\n\t// to upper case.\n\t// That is, we process a word at a time, where words are marked by _ or\n\t// upper case letter. Digits are treated as words.\n\tfor ; i < len(s); i++ {\n\t\tc := s[i]\n\t\tif c == '_' && i+1 < len(s) && isASCIILower(s[i+1]) {\n\t\t\tcontinue // Skip the underscore in s.\n\t\t}\n\t\tif isASCIIDigit(c) {\n\t\t\tt = append(t, c)\n\t\t\tcontinue\n\t\t}\n\t\t// Assume we have a letter now - if not, it's a bogus identifier.\n\t\t// The next word is a sequence of characters that must start upper case.\n\t\tif isASCIILower(c) {\n\t\t\tc ^= ' ' // Make it a capital letter.\n\t\t}\n\t\tt = append(t, c) // Guaranteed not lower case.\n\t\t// Accept lower case sequence that follows.\n\t\tfor i+1 < len(s) && isASCIILower(s[i+1]) {\n\t\t\ti++\n\t\t\tt = append(t, s[i])\n\t\t}\n\t}\n\treturn string(t)\n}" ]
[ "0.80201656", "0.780781", "0.7694911", "0.7694911", "0.7694911", "0.76857805", "0.7677266", "0.76719576", "0.744376", "0.74331975", "0.74155873", "0.7400728", "0.739891", "0.73898923", "0.73286057", "0.73286057", "0.7324523", "0.73189044", "0.72823215", "0.7234862", "0.719377", "0.71800953", "0.7177759", "0.7163344", "0.7118351", "0.70770186", "0.707009", "0.707009", "0.7053678", "0.70393413", "0.7030684", "0.7014624", "0.69898015", "0.6945334", "0.69217837", "0.6903463", "0.6889612", "0.6877201", "0.68176514", "0.6754094", "0.675146", "0.66908765", "0.66867656", "0.66447645", "0.66219854", "0.6589442", "0.6546844", "0.65374184", "0.6522769", "0.65203625", "0.6493627", "0.64759684", "0.64753765", "0.64753765", "0.64753765", "0.64703244", "0.6469527", "0.6448201", "0.64266145", "0.6406633", "0.63604534", "0.63361055", "0.6333807", "0.63267845", "0.63164926", "0.6312704", "0.62971085", "0.6251793", "0.6247916", "0.62350273", "0.6223487", "0.6208648", "0.6196269", "0.6145262", "0.61427397", "0.61264104", "0.61179316", "0.610419", "0.60967666", "0.6083538", "0.6072483", "0.6050729", "0.60420996", "0.6039236", "0.60181034", "0.6006178", "0.5997789", "0.59828335", "0.5973596", "0.5966084", "0.59625554", "0.5959998", "0.59326303", "0.59289885", "0.5907915", "0.5904829", "0.5904829", "0.5903311", "0.59008884" ]
0.814207
1
sliceIntoWords splits the provided identifier into a slice of individual words. A word is defined by one of the following: 1. A space ("a test" becomes "a" and "test") 2. A transition between lowercase and uppercase ("aWord" becomes "a" and "Word") 3. A transition between multiple uppercase letters and a lowercase letter ("XMLDocument" becomes "XML" and "Document") 4. A transition between a letter and a digit ("book12" becomes "book" and "12") 5. A transition between a digit and a letter ("12monkeys" becomes "12" and "monkeys")
func sliceIntoWords(identifier string) []string { // Trim any leading and trailing spaces to make our life easier later identifier = strings.Trim(identifier, " ") var result []string chars := []rune(identifier) lastStart := 0 for i := range chars { preceedingLower := i > 0 && unicode.IsLower(chars[i-1]) preceedingDigit := i > 0 && unicode.IsDigit(chars[i-1]) succeedingLower := i+1 < len(chars) && unicode.IsLower(chars[i+1]) // This case is for handling acronyms like XMLDocument isSpace := unicode.IsSpace(chars[i]) foundUpper := unicode.IsUpper(chars[i]) foundDigit := unicode.IsDigit(chars[i]) caseTransition := foundUpper && (preceedingLower || succeedingLower) digitTransition := (foundDigit && !preceedingDigit) || (!foundDigit && preceedingDigit) if isSpace { r := string(chars[lastStart:i]) r = strings.Trim(r, " ") // If r is entirely spaces... just don't append anything if len(r) != 0 { result = append(result, r) } lastStart = i + 1 // skip the space } else if i > lastStart && (caseTransition || digitTransition) { result = append(result, string(chars[lastStart:i])) lastStart = i } } if lastStart < len(chars) { result = append(result, string(chars[lastStart:])) } return result }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func sliceIntoWords(identifier string) []string {\n\t// Trim any leading and trailing spaces to make our life easier later\n\tidentifier = strings.Trim(identifier, \" \")\n\n\tvar result []string\n\tchars := []rune(identifier)\n\tlastStart := 0\n\tfor i := range chars {\n\t\tpreceedingLower := i > 0 && unicode.IsLower(chars[i-1])\n\t\tpreceedingDigit := i > 0 && unicode.IsDigit(chars[i-1])\n\t\tsucceedingLower := i+1 < len(chars) && unicode.IsLower(chars[i+1]) // This case is for handling acronyms like XMLDocument\n\t\tisSeparator := unicode.IsSpace(chars[i]) || chars[i] == '_'\n\t\tfoundUpper := unicode.IsUpper(chars[i])\n\t\tfoundDigit := unicode.IsDigit(chars[i])\n\t\tcaseTransition := foundUpper && (preceedingLower || succeedingLower)\n\t\tdigitTransition := (foundDigit && !preceedingDigit) || (!foundDigit && preceedingDigit)\n\t\tif isSeparator {\n\t\t\tr := string(chars[lastStart:i])\n\t\t\tr = strings.Trim(r, \" \")\n\t\t\t// If r is entirely spaces... just don't append anything\n\t\t\tif len(r) != 0 {\n\t\t\t\tresult = append(result, r)\n\t\t\t}\n\t\t\tlastStart = i + 1 // skip the space\n\t\t} else if i > lastStart && (caseTransition || digitTransition) {\n\t\t\tresult = append(result, string(chars[lastStart:i]))\n\t\t\tlastStart = i\n\t\t}\n\t}\n\n\tif lastStart < len(chars) {\n\t\tresult = append(result, string(chars[lastStart:]))\n\t}\n\n\treturn result\n}", "func Wordize(t string) []string {\n\treturn wordizeRe.Split(t, -1)\n}", "func splitWidthWords(s string, w int) []string {\n\tvar (\n\t\tsl []string\n\t\tline string\n\t)\n\tfor _, word := range splitWords(s) {\n\t\tif uLen(line)+uLen(word) < w {\n\t\t\tline += word\n\t\t} else {\n\t\t\ttrimmedLine := strings.TrimSpace(line)\n\t\t\tif strings.HasSuffix(trimmedLine, \"--\") {\n\t\t\t\t// Move the double dash to the beginning of the next line\n\t\t\t\ttrimmedLine = trimmedLine[:uLen(trimmedLine)-2]\n\t\t\t\tsl = append(sl, trimmedLine)\n\t\t\t\tline = \"-- \" + word\n\t\t\t} else {\n\t\t\t\tsl = append(sl, trimmedLine)\n\t\t\t\tline = word\n\t\t\t}\n\t\t}\n\t}\n\tif uLen(line) == 0 {\n\t\treturn sl\n\t}\n\treturn append(sl, strings.TrimSpace(line))\n}", "func Words(bytes []byte) (boundaries []Boundary) {\n\tboundaries = make([]Boundary, 0, len(bytes)) // TODO memory efficient\n\tfor pos := 0; pos < len(bytes); {\n\t\tlength := FirstWord(bytes[pos:])\n\t\tboundaries = append(boundaries, Boundary{pos, pos + length})\n\t\tpos += length\n\t}\n\treturn\n}", "func WordsInString(s string) (boundaries []Boundary) {\n\tboundaries = make([]Boundary, 0, len(s)) // TODO memory efficient\n\tfor pos := 0; pos < len(s); {\n\t\tlength := FirstWordInString(s[pos:])\n\t\tboundaries = append(boundaries, Boundary{pos, pos + length})\n\t\tpos += length\n\t}\n\treturn\n}", "func TestExtractWords(t *testing.T) {\n\ttests := []struct {\n\t\tinput string\n\t\texpected []string\n\t}{\n\t\t{\n\t\t\tinput: \"An Item To LowerCase\",\n\t\t\texpected: []string{\"an\", \"item\", \"to\", \"lowercase\"},\n\t\t},\n\t\t{\n\t\t\tinput: \" ignore multiple whitespaces \",\n\t\t\texpected: []string{\"ignore\", \"multiple\", \"whitespaces\"},\n\t\t},\n\t\t{\n\t\t\tinput: \"remove some useless char like ' or :\",\n\t\t\texpected: []string{\"remove\", \"some\", \"useless\", \"char\", \"like\", \"or\"},\n\t\t},\n\t}\n\n\tfor _, current := range tests {\n\t\tres := extractWords(current.input)\n\t\tif !reflect.DeepEqual(res, current.expected) {\n\t\t\tt.Errorf(\"\\n\\texpected:\\n%v\\n\\tbut got:\\n%v\\n\", current.expected, res)\n\t\t}\n\t}\n}", "func wordWalker(str string, f func(*wordInfo)) {\n\trunes := []rune(strings.TrimFunc(str, isDelimiter))\n\tw, i := 0, 0 // index of start of word, scan\n\thasCommonInitial := false\n\tfor i+1 <= len(runes) {\n\t\teow := false // whether we hit the end of a word\n\t\tswitch {\n\t\tcase i+1 == len(runes):\n\t\t\teow = true\n\t\tcase isDelimiter(runes[i+1]):\n\t\t\t// underscore; shift the remainder forward over any run of underscores\n\t\t\teow = true\n\t\t\tn := 1\n\t\t\tfor i+n+1 < len(runes) && isDelimiter(runes[i+n+1]) {\n\t\t\t\tn++\n\t\t\t}\n\n\t\t\t// Leave at most one underscore if the underscore is between two digits\n\t\t\tif i+n+1 < len(runes) && unicode.IsDigit(runes[i]) && unicode.IsDigit(runes[i+n+1]) {\n\t\t\t\tn--\n\t\t\t}\n\n\t\t\tcopy(runes[i+1:], runes[i+n+1:])\n\t\t\trunes = runes[:len(runes)-n]\n\t\tcase unicode.IsLower(runes[i]) && !unicode.IsLower(runes[i+1]):\n\t\t\t// lower->non-lower\n\t\t\teow = true\n\t\t}\n\t\ti++\n\n\t\t// [w,i) is a word.\n\t\tword := string(runes[w:i])\n\t\tif !eow && commonInitialisms[word] && !unicode.IsLower(runes[i]) {\n\t\t\t// through\n\t\t\t// split IDFoo → ID, Foo\n\t\t\t// but URLs → URLs\n\t\t} else if !eow {\n\t\t\tif commonInitialisms[word] {\n\t\t\t\thasCommonInitial = true\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tmatchCommonInitial := false\n\t\tif commonInitialisms[strings.ToUpper(word)] {\n\t\t\thasCommonInitial = true\n\t\t\tmatchCommonInitial = true\n\t\t}\n\n\t\tf(&wordInfo{\n\t\t\tWord: word,\n\t\t\tMatchCommonInitial: matchCommonInitial,\n\t\t\tHasCommonInitial: hasCommonInitial,\n\t\t})\n\t\thasCommonInitial = false\n\t\tw = i\n\t}\n}", "func wordWalker(str string, f func(*wordInfo)) {\n\trunes := []rune(str)\n\tw, i := 0, 0 // index of start of word, scan\n\thasCommonInitial := false\n\tfor i+1 <= len(runes) {\n\t\teow := false // whether we hit the end of a word\n\t\tswitch {\n\t\tcase i+1 == len(runes):\n\t\t\teow = true\n\t\tcase isDelimiter(runes[i+1]):\n\t\t\t// underscore; shift the remainder forward over any run of underscores\n\t\t\teow = true\n\t\t\tn := 1\n\t\t\tfor i+n+1 < len(runes) && isDelimiter(runes[i+n+1]) {\n\t\t\t\tn++\n\t\t\t}\n\n\t\t\t// Leave at most one underscore if the underscore is between two digits\n\t\t\tif i+n+1 < len(runes) && unicode.IsDigit(runes[i]) && unicode.IsDigit(runes[i+n+1]) {\n\t\t\t\tn--\n\t\t\t}\n\n\t\t\tcopy(runes[i+1:], runes[i+n+1:])\n\t\t\trunes = runes[:len(runes)-n]\n\t\tcase unicode.IsLower(runes[i]) && !unicode.IsLower(runes[i+1]):\n\t\t\t// lower->non-lower\n\t\t\teow = true\n\t\t}\n\t\ti++\n\n\t\t// [w,i) is a word.\n\t\tword := string(runes[w:i])\n\t\tif !eow && commonInitialisms[word] && !unicode.IsLower(runes[i]) {\n\t\t\t// through\n\t\t\t// split IDFoo → ID, Foo\n\t\t\t// but URLs → URLs\n\t\t} else if !eow {\n\t\t\tif commonInitialisms[word] {\n\t\t\t\thasCommonInitial = true\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tmatchCommonInitial := false\n\t\tif commonInitialisms[strings.ToUpper(word)] {\n\t\t\thasCommonInitial = true\n\t\t\tmatchCommonInitial = true\n\t\t}\n\n\t\tf(&wordInfo{\n\t\t\tWord: word,\n\t\t\tMatchCommonInitial: matchCommonInitial,\n\t\t\tHasCommonInitial: hasCommonInitial,\n\t\t})\n\t\thasCommonInitial = false\n\t\tw = i\n\t}\n}", "func WordBreaksInString(s string) (breaks []int) {\n\tl := len(s)\n\tif l == 0 {\n\t\treturn\n\t}\n\tbreaks = make([]int, 1, len(s)) // TODO memory efficient\n\tbreaks[0] = 0\n\tfor pos := 0; pos < l; {\n\t\tlength := FirstWordInString(s[pos:])\n\t\tpos += length\n\t\tbreaks = append(breaks, pos)\n\t}\n\treturn\n}", "func (impl *Impl) ExtractIndexableWords() []string {\n\tw := make([]string, 0, 20)\n\tw = append(w, fmt.Sprintf(\"%d\", impl.Id))\n\tw = append(w, strings.ToLower(impl.LanguageName))\n\tw = append(w, SplitForIndexing(impl.ImportsBlock, true)...)\n\tw = append(w, SplitForIndexing(impl.CodeBlock, true)...)\n\tif len(impl.AuthorComment) >= 3 {\n\t\tw = append(w, SplitForIndexing(impl.AuthorComment, true)...)\n\t}\n\tif langExtras, ok := langsExtraKeywords[impl.LanguageName]; ok {\n\t\tw = append(w, langExtras...)\n\t}\n\t// Note: we don't index external URLs.\n\treturn w\n}", "func BytesToWords(bank string, data []byte) (words []string) {\n\twordsAll := GetWords(\"english\")\n\n\t// 2048 words per bank, which is 2^11.\n\tnumWords := (8*len(data) + 10) / 11\n\n\tn2048 := big.NewInt(2048)\n\tnData := big.NewInt(0).SetBytes(data)\n\tnRem := big.NewInt(0)\n\t// Alternative, use condition \"nData.BitLen() > 0\"\n\t// to allow for shorter words when data has leading 0's\n\tfor i := 0; i < numWords; i++ {\n\t\tnData.DivMod(nData, n2048, nRem)\n\t\trem := nRem.Int64()\n\t\twords = append(words, wordsAll[rem])\n\t}\n\treturn words\n}", "func BreakdownInto(container []string, sliceSize int) [][]string {\n\n\tvar ret [][]string\n\tvar acc []string\n\tvar tracker int = 0\n\tfor _, content := range container {\n\t\tacc = append(acc, content)\n\t\tif tracker == sliceSize {\n\t\t\tret = append(ret, acc)\n\t\t\ttracker = 0\n\t\t\tacc = []string{}\n\t\t}\n\n\t\ttracker++\n\t}\n\n\treturn ret\n}", "func (t *Ticket) Words() []string {\n\tre := regexp.MustCompile(\"[^a-zA-Z ]\")\n\tphrase := re.ReplaceAllString(t.Section, \" \")\n\n\twords := strings.Split(phrase, \" \")\n\tfor i := len(words) - 1; i >= 0; i-- {\n\t\tif words[i] == \"\" {\n\t\t\twords = append(words[:i], words[i+1:]...)\n\t\t}\n\t}\n\n\treturn words\n}", "func wordBreak(s string, wordDict []string) bool {\n\t// Accepted\n\t// 36/36 cases passed (0 ms)\n\t// Your runtime beats 100 % of golang submissions\n\t// Your memory usage beats 95.19 % of golang submissions (2.1 MB)\n\tif wordDict == nil {\n\t\tif s != \"\" {\n\t\t\treturn false\n\t\t} else {\n\t\t\treturn true\n\t\t}\n\t}\n\n\tl := len(s)\n\tbl := make([]bool, l+1)\n\tbl[0] = true\n\tfor i := 1; i <= l; i++ {\n\t\tbl[i] = false\n\t\tfor j := i - 1; j > -1; j-- {\n\t\t\tif bl[j] && StringsContains(wordDict, s[j:i]) != -1 {\n\t\t\t\tbl[i] = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn bl[l]\n}", "func NewWordSegmenter(r io.Reader) *Segmenter {\n\treturn &Segmenter{\n\t\tr: r,\n\t\tsegment: SegmentWords,\n\t\tmaxTokenSize: MaxScanTokenSize,\n\t\tbuf: make([]byte, 4096), // Plausible starting size; needn't be large.\n\t}\n}", "func wordsMessage(str string) (buf []string, result string) {\n buf = make([]string, 5);\n var start, end, pos int;\n result = \"\";\n leng := len(str);\n for {\n for start < leng && str[start] == ' ' { start++; }\n if start == leng { break; }\n if str[start] == ':' { result = str[start+1:len(str)]; break; }\n for end = start+1; end < leng && str[end] != ' '; { end++; }\n if pos >= len(buf) {\n buf2 := make([]string, len(buf)*2);\n for i, x := range buf { buf2[i] = x; }\n buf = buf2;\n }\n buf[pos] = str[start:end];\n pos++;\n start = end;\n }\n buf = buf[0:pos];\n return;\n}", "func wordBreak(s string, wordDict []string) bool {\n\tlen2rowMap := make(map[int]int)\n\twordsByLen := make([][]string, 0)\n\tbaseLen, maxLen := 0, 0\n\tfor _, v := range wordDict {\n\t\tif len(v) > len(s) {\n\t\t\tcontinue\n\t\t}\n\t\ti, ok := len2rowMap[len(v)]\n\t\tif !ok {\n\t\t\twordsByLen = append(wordsByLen, make([]string, 0))\n\t\t\ti = baseLen\n\t\t\tlen2rowMap[len(v)] = i\n\t\t\tbaseLen++\n\t\t\tif len(v) > maxLen {\n\t\t\t\tmaxLen = len(v)\n\t\t\t}\n\t\t}\n\t\twordsByLen[i] = append(wordsByLen[i], v)\n\t}\n\n\tbs := []byte(s)\n\tdp := make([]bool, len(s)+1)\n\tdp[0] = true\n\tbase := 0\n\tfor base = 0; base <= len(s); base++ {\n\t\tif !dp[base] {\n\t\t\tcontinue\n\t\t}\n\t\tfor l := 1; l <= maxLen; l++ {\n\t\t\tif i, ok := len2rowMap[l]; ok {\n\t\t\t\tpred := base + l\n\t\t\t\tif pred > len(s) || dp[pred] {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\twords := wordsByLen[i]\n\t\t\t\tpartS := string(bs[base:pred])\n\t\t\t\tfor _, w := range words {\n\t\t\t\t\tif w == partS {\n\t\t\t\t\t\tdp[pred] = true\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn dp[len(s)]\n}", "func WordBreak(s string, wordDict []string) bool {\n\tdp, hash_map := make([]bool, len(s)+1), make(map[string]bool)\n\tfor _, word := range wordDict {\n\t\thash_map[word] = true\n\t}\n\tdp[0] = true\n\n\tfor i := 1; i <= len(s); i++ {\n\t\tfor j := 0; j < i; j++ {\n\t\t\tif dp[j] {\n\t\t\t\tif _, ok := hash_map[s[j:i]]; ok {\n\t\t\t\t\tdp[i] = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn dp[len(s)]\n}", "func SplitBy(s string, decider func(r rune) bool) []string {\n\n\t// split by caps\n\tvar segments []string\n\tvar currentSeg []rune\n\n\tfor rIndex, r := range s {\n\n\t\tif decider(r) {\n\t\t\t// new word\n\t\t\tif len(currentSeg) > 0 {\n\t\t\t\tsegments = append(segments, string(currentSeg))\n\t\t\t\tcurrentSeg = nil\n\t\t\t}\n\t\t}\n\n\t\tcurrentSeg = append(currentSeg, r)\n\n\t\t// is this the last one?\n\t\tif rIndex == len(s)-1 {\n\t\t\tsegments = append(segments, string(currentSeg))\n\t\t}\n\n\t}\n\n\treturn segments\n}", "func wordBreak(s string, wordDict []string) bool {\n\twordMap := make(map[string]struct{})\n\tfor i := 0; i < len(wordDict); i++ {\n\t\twordMap[wordDict[i]] = struct{}{}\n\t}\n\n\tdp := make([]bool, len(s)+1)\n\tdp[0] = true\n\n\tfor i := 0; i < len(s); i++ {\n\t\tfor _, word := range wordDict {\n\t\t\tif dp[i] {\n\t\t\t\tif i+len(word) <= len(s) && s[i:i+len(word)] == word {\n\t\t\t\t\tdp[i+len(word)] = true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn dp[len(s)]\n}", "func WordBreaks(bytes []byte) (breaks []int) {\n\tl := len(bytes)\n\tif l == 0 {\n\t\treturn\n\t}\n\tbreaks = make([]int, 1, len(bytes)) // TODO memory efficient\n\tbreaks[0] = 0\n\tfor pos := 0; pos < l; {\n\t\tlength := FirstWord(bytes[pos:])\n\t\tpos += length\n\t\tbreaks = append(breaks, pos)\n\t}\n\treturn\n}", "func (factory *identifierFactory) cleanPart(part string, visibility Visibility) string {\n\tclean := filterRegex.ReplaceAllLiteralString(part, \" \")\n\tcleanWords := sliceIntoWords(clean)\n\tcaseCorrectedWords := make([]string, 0, len(cleanWords))\n\tfor ix, word := range cleanWords {\n\t\tvar w string\n\t\tif ix == 0 && visibility == NotExported {\n\t\t\tw = strings.ToLower(word)\n\t\t} else {\n\t\t\t// Disable lint: the suggested \"replacement\" for this in /x/cases has fundamental\n\t\t\t// differences in how it works (e.g. 'JSON' becomes 'Json'; we don’t want that).\n\t\t\t// Furthermore, the cases (ha) that it \"fixes\" are not relevant to us\n\t\t\t// (something about better handling of various punctuation characters;\n\t\t\t// our words are punctuation-free).\n\t\t\t//nolint:staticcheck\n\t\t\tw = strings.Title(word)\n\t\t}\n\n\t\tcaseCorrectedWords = append(caseCorrectedWords, w)\n\t}\n\n\treturn strings.Join(caseCorrectedWords, \"\")\n}", "func WordAt(i int) string {\n\treturn words[i]\n}", "func splitCamel(s string, sep byte) string {\n\tvar splStr []byte\n\tsLen := len(s)\n\t// easier to handle short strings here:\n\tif sLen < 2 {\n\t\treturn strings.ToLower(s)\n\t} else if sLen == 2 {\n\t\tif isLower(s[0]) && isUpper(s[1]) {\n\t\t\treturn fmt.Sprintf(\"%c-%c\", s[0], toLower(s[1]))\n\t\t}\n\t\treturn strings.ToLower(s)\n\t}\n\tvar appendLastCharacter []byte\n\t// the final letter doesn't follow the same rules (to support plurals like URLs)\n\t// TestURLs --> test-urls (Ending Multiword Caps + one lowercase)\n\t// Also catches TestT --> test-t\n\tif (isLower(s[sLen-1]) && isUpper(s[sLen-2]) && isUpper(s[sLen-3])) || (isLower(s[sLen-2]) && isUpper(s[sLen-1])) {\n\t\tif isUpper(s[sLen-1]) {\n\t\t\tappendLastCharacter = append(appendLastCharacter, sep)\n\t\t}\n\t\tappendLastCharacter = append(appendLastCharacter, s[sLen-1])\n\t\ts = s[0 : sLen-1]\n\t\tsLen--\n\t}\n\tinWord := false\n\twordStart := 0\n\tfor i := 0; i < sLen-1; i++ {\n\t\tl, m := s[i], s[i+1]\n\t\tif !inWord && isLower(l) && isUpper(m) { // aU -> a-u...\n\t\t\twordStart = i - 1\n\t\t\tif wordStart < 0 {\n\t\t\t\twordStart = 0\n\t\t\t}\n\t\t\t// do not continue (keep going).\n\t\t} else if !inWord {\n\t\t\tinWord = true\n\t\t\twordStart = i\n\t\t\tcontinue\n\t\t}\n\t\t// now in word\n\t\tif sameCase(l, m) {\n\t\t\tcontinue\n\t\t}\n\t\t// URLStuff -> url-stuff -- use l as boundry instead of m (case of multi caps)\n\t\tend := i + 1\n\t\tif i-1 >= 0 && isLower(m) && isUpper(l) && isUpper(s[i-1]) {\n\t\t\tend--\n\t\t}\n\t\t//FOOFoo -> foo-foo\n\t\tsplStr = append(splStr, s[wordStart:end]...)\n\t\tsplStr = append(splStr, sep)\n\t\twordStart = end\n\t\tinWord = (end == i) // if we moved back above, then we're still in a word.\n\t}\n\tif inWord {\n\t\tsplStr = append(splStr, s[wordStart:]...)\n\t}\n\tif len(appendLastCharacter) > 0 {\n\t\tsplStr = append(splStr, appendLastCharacter...)\n\t}\n\treturn strings.ToLower(string(splStr))\n}", "func wordBreak(s string, wordDict []string) bool {\n\t// 动态规划 d[i] 表示以i结尾的字符串是否可以被拆分\n\t// d[i] = 判定\n\twordMap := make(map[string]bool)\n\tfor _, val := range wordDict {\n\t\twordMap[val] = true\n\t}\n\n\td := make([]bool, len(s) + 1)\n\td[0] = true // 表示“”一定可以拆分\n\tfor i := 1; i < len(s) + 1; i++ {\n\t\t// 判定字符串s[j:i+1]是否在wordMap中\n\t\tfor j := 0; j < i; j++ {\n\t\t\tif d[j] && wordMap[s[j:i]] {\n\t\t\t\td[i] = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn d[len(s)]\n}", "func (gn *Gen) GenPartWordWavs() {\n\t// next the part words\n\tfor i := 0; i < 4; i++ {\n\t\tfor j := 0; j < 4; j++ {\n\t\t\tk := j % 4\n\t\t\tif k == i {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tgn.GenTriCVWavs(gn.Syl3[i], gn.Syl1[k], gn.Syl2[k], 3)\n\t\t}\n\t}\n}", "func SpinWords(sentence string) string {\n\twords := strings.Split(sentence, \" \")\n\n\tfor i := 0; i < len(words); i++ {\n\t\tif len(words[i]) >= 5 {\n\t\t\twords[i] = reverse(words[i])\n\t\t}\n\t}\n\n\treturn strings.Join(words, \" \")\n}", "func wordBreak(s string, wordDict []string) bool {\n\tlookup := toSet(wordDict)\n\tbreakable := []int{0}\n\tfor i := 1; i <= len(s); i++ {\n\t\tbroken := false\n\t\tfor _, j := range breakable {\n\t\t\tif _, ok := lookup[s[j:i]]; ok {\n\t\t\t\tbroken = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif broken {\n\t\t\tbreakable = append(breakable, i)\n\t\t}\n\t}\n\treturn breakable[len(breakable)-1] == len(s)\n}", "func WordAtInString(s string, pos int) Boundary {\n\treturn Boundary{WordBeginInString(s, pos), WordEndInString(s, pos)}\n}", "func WordAt(bytes []byte, pos int) Boundary {\n\treturn Boundary{WordBegin(bytes, pos), WordEnd(bytes, pos)}\n}", "func splitWords(s string) []string {\n\tvar (\n\t\tsplitpoint bool\n\t\twords []string\n\t\tletters strings.Builder\n\t\ttmp string\n\t)\n\tlenS := uLen(s)\n\tfor i, r := range s {\n\t\tsplitpoint = false\n\t\tswitch r {\n\t\tcase '.', '!', ',', ':', '-', ' ', '?', ';', '\\n':\n\t\t\t// Check if the next character is not an end quote\n\t\t\tif i+1 < lenS && s[i+1] != '\"' && s[i+1] != '\\'' {\n\t\t\t\tsplitpoint = true\n\t\t\t}\n\t\t}\n\t\t// Combine repeated dashes\n\t\tif r == '-' && i+1 < lenS && s[i+1] == '-' {\n\t\t\tsplitpoint = false\n\t\t}\n\t\t// Combine repeated dots\n\t\tif r == '.' && i+1 < lenS && s[i+1] == '.' {\n\t\t\tsplitpoint = false\n\t\t}\n\t\tif splitpoint || i == lenS {\n\t\t\tletters.WriteRune(r)\n\t\t\ttmp = letters.String()\n\t\t\tif uLen(tmp) > 0 {\n\t\t\t\twords = append(words, tmp)\n\t\t\t}\n\t\t\tletters.Reset()\n\t\t} else {\n\t\t\tletters.WriteRune(r)\n\t\t}\n\t}\n\ttmp = strings.TrimSpace(letters.String())\n\tif uLen(tmp) > 0 {\n\t\twords = append(words, tmp)\n\t}\n\treturn words\n}", "func (wp Wordpiece) Tokenize(text string) []string {\n\t// TODO: determine if utf8 conversion is necessary, per python impl\n\t// text = convert_to_unicode(text)\n\tvar toks []string\n\tfor _, tok := range tokenizeWhitespace(text) {\n\t\tif len(tok) > wp.maxWordChars {\n\t\t\ttoks = append(toks, wp.unknownToken)\n\t\t\tcontinue\n\t\t}\n\t\tfor len(tok) > 0 && tok != \"##\" {\n\t\t\tsub := wp.vocab.LongestSubstring(tok)\n\t\t\tif sub == \"\" {\n\t\t\t\ttoks = append(toks, wp.unknownToken)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\ttoks = append(toks, sub)\n\t\t\ttok = fmt.Sprintf(\"##%s\", tok[len(sub):])\n\t\t}\n\t}\n\treturn toks\n}", "func (idiom *Idiom) ExtractIndexableWords() (w []string, wTitle []string, wLead []string) {\n\tw = SplitForIndexing(idiom.Title, true)\n\tw = append(w, fmt.Sprintf(\"%d\", idiom.Id))\n\twTitle = w\n\n\twLead = SplitForIndexing(idiom.LeadParagraph, true)\n\tw = append(w, wLead...)\n\t// ExtraKeywords as not as important as Title, rather as important as Lead\n\twKeywords := SplitForIndexing(idiom.ExtraKeywords, true)\n\twLead = append(wLead, wKeywords...)\n\tw = append(w, wKeywords...)\n\n\tfor i := range idiom.Implementations {\n\t\timpl := &idiom.Implementations[i]\n\t\twImpl := impl.ExtractIndexableWords()\n\t\tw = append(w, wImpl...)\n\t}\n\n\treturn w, wTitle, wLead\n}", "func doubleDigitToWords(digit int) string {\n\tword := \"\"\n\tif digit >= 10 && digit < 20 {\n\t\tword = []string{\"Ten\", \"Eleven\", \"Twelve\", \"Thirteen\", \"Fourteen\",\n\t\t\t\"Fifteen\", \"Sixteen\", \"Seventeen\", \"Eighteen\", \"Nineteen\"}[digit-10]\n\t}\n\treturn word\n}", "func words(data []string) <-chan string {\n\tout := make(chan string)\n\n\tgo func() {\n\t\tdefer close(out)\n\n\t\tfor _, line := range data {\n\t\t\twords := strings.Split(line, \" \")\n\t\t\tfor _, word := range words {\n\t\t\t\tword = strings.ToLower(word)\n\n\t\t\t\tout <- word\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn out\n}", "func digitToWords(digit int) string {\n\tword := \"\"\n\tif digit >= 0 && digit < 10 {\n\t\tword = []string{\"Zero\", \"One\", \"Two\", \"Three\", \"Four\",\n\t\t\t\"Five\", \"Six\", \"Seven\", \"Eight\", \"Nine\"}[digit]\n\t}\n\treturn word\n}", "func BetweenWords(sentence string, emoji string) string {\n\treturn strings.ReplaceAll(sentence, \" \", emoji)\n}", "func WordCount(s string) Messages {\n\tstrs := strings.Fields(s)\n\tres := make(map[string]int)\n\n\tfor _, str := range strs {\n\t\tif len(str) > 2 {\n\t\t\tres[strings.ToLower(str)]++\n\t\t}\n\t}\n\n\tmessages := []Msg{}\n\n\tfor k, v := range res {\n\t\tm := Msg{\n\t\t\tText: k,\n\t\t\tCount: v,\n\t\t}\n\t\tmessages = append(messages, m)\n\t}\n\n\treturn messages\n}", "func NewWordpiece(voc vocab.Dict) Wordpiece {\n\treturn Wordpiece{\n\t\tvocab: voc,\n\t\tmaxWordChars: DefaultMaxWordChars,\n\t\tunknownToken: DefaultUnknownToken,\n\t}\n}", "func splitData(fileName string, chunkSize int) (numMapFiles int, err error) {\n\tvar(\n\t\tfileBuffer []byte //Buffer that stores everything from the file\n\t\tfile *os.File //File that we create to store the pieces\n\t\tfileSize int //We keep track of the file size in bytes\n\t)\n\n\t//Attempts to open the input, if not successful, \"throws\" the erro\n\tif fileBuffer, err = ioutil.ReadFile(fileName); err != nil {\n\t\treturn 0, err\n\t}\n\n\n\tnumMapFiles = 0\n\tword := make([]byte, 0)\n\n\t//Creates the first split file, \"throws\" an error if any\n\tif file, err = os.Create(mapFileName(numMapFiles)); err != nil {\n\t\treturn 0, err\n\t}\n\tnumMapFiles++\n\tfileSize = 0\n\n\t//Iterates through all the runes in the input\n\tfor len(fileBuffer) > 0 {\n\t\tr, size := utf8.DecodeRune(fileBuffer)\n\n\t\t//If the character is somehow larger than the chunk size, return an error\n\t\tif size > chunkSize {\n\t\t\treturn 0, errors.New(\"Character size larger than chunk size\")\n\t\t}\n\n\t\t/*We attempt to assemble an entire word before adding it to the file as not to\n\t\tcut a word in half during the process. If another character is found, adds it \n\t\tdirectly to the file */\n\n\t\tif(unicode.IsLetter(r) || unicode.IsNumber(r)) {\n\t\t\tword = append(word, fileBuffer[:size]...)\n\t\t} else {\n\t\t\t//If another character is encountered, first attempts do add the finished\n\t\t\t//word to the file\n\t\t\tif len(word) > 0 {\n\t\t\t\t//If a word that is bigger than the chunk size is found, we cant handle it\n\t\t\t\tif len(word) > chunkSize {\n\t\t\t\t\treturn 0, errors.New(\"Word larger than chunk size\")\n\t\t\t\t} \n\t\t\t\t//If adding the word exceed the chunk size, starts a new file\n\t\t\t\tif fileSize + len(word) > chunkSize {\n\t\t\t\t\tfile.Close()\n\t\t\t\t\tif file, err = os.Create(mapFileName(numMapFiles)); err != nil {\n\t\t\t\t\t\treturn numMapFiles, err\n\t\t\t\t\t}\n\t\t\t\t\tnumMapFiles++\n\t\t\t\t\tfileSize = 0\n\t\t\t\t}\n\t\t\t\tif _, err = file.Write(word); err != nil {\n\t\t\t\t\treturn numMapFiles, err\n\t\t\t\t}\n\t\t\t\tfileSize += len(word)\n\t\t\t\t//Cleans the word buffer\n\t\t\t\tword = word[:0]\n\t\t\t}\n\t\t\t//Now attempts to add the next character to the file\n\t\t\t//If it passes the cuhunk size, starts a new file\n\t\t\tif fileSize + size > chunkSize {\n\t\t\t\tfile.Close()\n\t\t\t\tif file, err = os.Create(mapFileName(numMapFiles)); err != nil {\n\t\t\t\t\treturn numMapFiles, err\n\t\t\t\t}\n\t\t\t\tnumMapFiles++\n\t\t\t\tfileSize = 0\n\t\t\t}\n\t\t\tif _, err = file.Write(fileBuffer[:size]); err != nil {\n\t\t\t\treturn numMapFiles, err\n\t\t\t}\n\t\t\tfileSize += size\n\t\t}\n\t\t//Continue to the next rune\n\t\tfileBuffer = fileBuffer[size:]\n\t}\n\n\t//If there's still a word in the buffer, add the word to the last file\n\t//Or creates a new file if there's no space\n\tif len(word) > 0 {\n\t\tif len(word) > chunkSize {\n\t\t\treturn 0, errors.New(\"Word larger than chunk size\")\n\t\t}\n\t\tif fileSize + len(word) > chunkSize {\n\t\t\tfile.Close()\n\t\t\tif file, err = os.Create(mapFileName(numMapFiles)); err != nil {\n\t\t\t\treturn numMapFiles, err\n\t\t\t}\n\t\t\tnumMapFiles++\n\t\t}\n\t\tif _, err = file.Write(word); err != nil {\n\t\t\treturn numMapFiles, err\n\t\t}\n\t}\n\n\tfile.Close()\n\n\treturn numMapFiles, nil\n}", "func wordWrap(s string, width int) []string {\n\ts = colorPattern.ReplaceAllString(s, \"\")\n\tif len(s) <= width {\n\t\treturn []string{s}\n\t}\n\n\tvar wrapped []string\n\tfor len(s) > 0 {\n\t\textract := runewidth.Truncate(s, width, \"\")\n\t\tif len(extract) == 0 {\n\t\t\tgr := uniseg.NewGraphemes(s)\n\t\t\tgr.Next()\n\t\t\t_, to := gr.Positions()\n\t\t\textract = s[:to]\n\t\t}\n\t\tif len(extract) < len(s) {\n\t\t\t// Add any spaces from the next line.\n\t\t\tif spaces := spacePattern.FindStringIndex(s[len(extract):]); spaces != nil && spaces[0] == 0 {\n\t\t\t\textract = s[:len(extract)+spaces[1]]\n\t\t\t}\n\n\t\t\t// Can we split before the mandatory end?\n\t\t\tmatches := boundaryPattern.FindAllStringIndex(extract, -1)\n\t\t\tif len(matches) > 0 {\n\t\t\t\t// Yes. Let's split there.\n\t\t\t\textract = extract[:matches[len(matches)-1][1]]\n\t\t\t}\n\t\t}\n\t\twrapped = append(wrapped, extract)\n\t\ts = s[len(extract):]\n\t}\n\treturn wrapped\n}", "func (r *Replacer) Subwords() []Subword {\n\tr.commit()\n\n\tvar subwords []Subword\n\n\tif len(r.levels) == 0 {\n\t\treturn subwords\n\t}\n\n\tlevel := r.levels[0]\n\n\tvar buf bytes.Buffer\n\n\tfor i, ch := range r.word {\n\t\tif r.levels[i] != level {\n\t\t\tsubwords = append(subwords, New(buf.String(), level))\n\t\t\tlevel = r.levels[i]\n\t\t\tbuf.Reset()\n\t\t}\n\t\tbuf.WriteRune(ch)\n\t}\n\tsubwords = append(subwords, New(buf.String(), level))\n\n\treturn subwords\n}", "func SplitAndTrimIndex(str []byte, isSplitRune, isTrimRune IsRuneFunc) (indices [][]int) {\n\tinWord := false\n\tinTrimmedWord := false\n\tlastWasTrim := false\n\tonlyTrimsSince := -1\n\twordStart := -1\n\tr, n := utf8.DecodeRune(str)\n\n\tfor i := 0; r != utf8.RuneError; {\n\t\tisSplit := isSplitRune(r)\n\t\tisTrim := isTrimRune(r)\n\n\t\tif isTrim && !lastWasTrim {\n\t\t\tonlyTrimsSince = i\n\t\t}\n\n\t\tif inWord {\n\t\t\tif inTrimmedWord {\n\t\t\t\tif isSplit {\n\t\t\t\t\twordEnd := i\n\t\t\t\t\tif onlyTrimsSince != -1 {\n\t\t\t\t\t\t// if there have been only trim rune since the last word rune\n\t\t\t\t\t\t// backtrack wordEnd to first trim character\n\t\t\t\t\t\twordEnd = onlyTrimsSince\n\t\t\t\t\t}\n\t\t\t\t\tif wordEnd > wordStart {\n\t\t\t\t\t\t// panicIfNot(wordStart >= 0)\n\t\t\t\t\t\tindices = append(indices, []int{wordStart, wordEnd})\n\t\t\t\t\t}\n\t\t\t\t\tinTrimmedWord = false\n\t\t\t\t\tinWord = false\n\t\t\t\t\twordStart = -2\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tif isSplit {\n\t\t\t\t\tinTrimmedWord = false\n\t\t\t\t\tinWord = false\n\t\t\t\t\twordStart = -3\n\t\t\t\t} else if !isTrim {\n\t\t\t\t\tinWord = true\n\t\t\t\t\tinTrimmedWord = true\n\t\t\t\t\twordStart = i\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\t// panicIfNot(inTrimmedWord == false)\n\t\t\tif !isSplit {\n\t\t\t\tinWord = true\n\t\t\t\tif !isTrim {\n\t\t\t\t\tinTrimmedWord = true\n\t\t\t\t\twordStart = i\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\n\t\tlastWasTrim = isTrim\n\t\tif !isTrim {\n\t\t\tonlyTrimsSince = -1\n\t\t}\n\n\t\ti += n\n\t\tr, n = utf8.DecodeRune(str[i:])\n\t}\n\tif inTrimmedWord {\n\t\twordEnd := len(str)\n\t\tif onlyTrimsSince != -1 {\n\t\t\twordEnd = onlyTrimsSince\n\t\t}\n\t\tif wordEnd > wordStart {\n\t\t\tpanicIfNot(wordStart >= 0)\n\t\t\tindices = append(indices, []int{wordStart, wordEnd})\n\t\t}\n\t}\n\treturn indices\n}", "func WordIndexes(s []byte, word []byte) (idxs []int) {\n\ttmp := Indexes(s, word)\n\tif len(tmp) == 0 {\n\t\treturn nil\n\t}\n\n\tfor _, idx := range tmp {\n\t\tx := idx - 1\n\t\tif x >= 0 {\n\t\t\tif !unicode.IsSpace(rune(s[x])) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tx = idx + len(word)\n\t\tif x >= len(s) {\n\t\t\tidxs = append(idxs, idx)\n\t\t\tcontinue\n\t\t}\n\t\tif !unicode.IsSpace(rune(s[x])) {\n\t\t\tcontinue\n\t\t}\n\t\tidxs = append(idxs, idx)\n\t}\n\n\treturn idxs\n}", "func FlattenToWordsFunction(env *Zlisp, name string, args []Sexp) (Sexp, error) {\n\tif len(args) == 0 {\n\t\treturn SexpNull, WrongNargs\n\t}\n\tstringArgs, err := flattenToWordsHelper(args)\n\tif err != nil {\n\t\treturn SexpNull, err\n\t}\n\n\t// Now convert to []Sexp{SexpStr}\n\tres := make([]Sexp, len(stringArgs))\n\tfor i := range stringArgs {\n\t\tres[i] = &SexpStr{S: stringArgs[i]}\n\t}\n\treturn env.NewSexpArray(res), nil\n}", "func NewWords(words []*Word) *Words {\n\treturn &Words{words: words, offset: -1}\n}", "func arrangeWords(text string) string {\n\tss := strings.Split(text, \" \")\n\tsort.SliceStable(ss, func(i, j int) bool {\n\t\treturn len(ss[i]) < len(ss[j])\n\t})\n\n\tvar sb strings.Builder\n\n\tc := ss[0][0]\n\tif 'a' <= c && c <= 'z' {\n\t\tsb.WriteByte(c + 'A' - 'a')\n\t} else {\n\t\tsb.WriteByte(c)\n\t}\n\n\tsb.WriteString(ss[0][1:])\n\n\tfor _, s := range ss[1:] {\n\t\tsb.WriteByte(' ')\n\t\tsb.WriteString(strings.ToLower(s))\n\t}\n\n\treturn sb.String()\n}", "func SplitByRanges(s string, ranges map[string][]RuneRange) []Word {\n\tvar sentence []Word\n\tvar currentWord = new(Word)\n\n\tgr := uniseg.NewGraphemes(s)\n\n\tfor gr.Next() {\n\t\tmatched := false\n\t\trunes := gr.Runes()\n\n\t\tfor name, rng := range ranges {\n\t\t\tif IsPartOfRange(runes[0], rng) {\n\t\t\t\tif currentWord.Type != name && currentWord.Type != \"\" && len(currentWord.Text) > 0 {\n\t\t\t\t\tsentence = append(sentence, *currentWord)\n\t\t\t\t\tcurrentWord = new(Word)\n\t\t\t\t}\n\t\t\t\tif currentWord.Type == \"\" {\n\t\t\t\t\tcurrentWord.Type = name\n\t\t\t\t}\n\t\t\t\tmatched = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\n\t\tif !matched {\n\t\t\tif (currentWord.Type != \"unmatched\" || currentWord.Type == \"\") && len(currentWord.Text) > 0 {\n\t\t\t\tsentence = append(sentence, *currentWord)\n\t\t\t\tcurrentWord = new(Word)\n\t\t\t}\n\t\t\tif currentWord.Type == \"\" {\n\t\t\t\tcurrentWord.Type = \"unmatched\"\n\t\t\t}\n\t\t}\n\n\t\tcurrentWord.Text += string(runes)\n\t}\n\n\treturn append(sentence, *currentWord)\n}", "func extractWords(line string) []string {\n\tline = strings.ToLower(line)\n\tline = strings.ReplaceAll(line, \"'\", \"\")\n\tline = strings.ReplaceAll(line, \":\", \"\")\n\n\tsplitted := strings.Split(line, \" \")\n\tres := make([]string, 0, len(splitted))\n\tfor _, v := range splitted {\n\t\tif strings.Trim(v, \" \") != \"\" {\n\t\t\tres = append(res, v)\n\t\t}\n\t}\n\treturn res\n}", "func NewSeedFromWords(input string) (Seed, error) {\n\treturn walletseed.DecodeUserInput(input)\n\n}", "func getWord(begin, end int, t string) string {\n for end >= len(t) {\n return \"\"\n }\n d := make([]uint8, end-begin+1)\n for j, i := 0, begin; i <= end; i, j = i+1, j+1 {\n d[j] = t[i]\n }\n return string(d)\n}", "func (c *Client) Words() goa.Endpoint {\n\tvar (\n\t\tdecodeResponse = DecodeWordsResponse(c.decoder, c.RestoreResponseBody)\n\t)\n\treturn func(ctx context.Context, v interface{}) (interface{}, error) {\n\t\treq, err := c.BuildWordsRequest(ctx, v)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tresp, err := c.WordsDoer.Do(req)\n\t\tif err != nil {\n\t\t\treturn nil, goahttp.ErrRequestError(\"shiritori\", \"words\", err)\n\t\t}\n\t\treturn decodeResponse(resp)\n\t}\n}", "func FirstWord(b []byte, state int) (word, rest []byte, newState int) {\n\t// An empty byte slice returns nothing.\n\tif len(b) == 0 {\n\t\treturn\n\t}\n\n\t// Extract the first rune.\n\tr, length := utf8.DecodeRune(b)\n\tif len(b) <= length { // If we're already past the end, there is nothing else to parse.\n\t\treturn b, nil, wbAny\n\t}\n\n\t// If we don't know the state, determine it now.\n\tif state < 0 {\n\t\tstate, _ = transitionWordBreakState(state, r, b[length:], \"\")\n\t}\n\n\t// Transition until we find a boundary.\n\tvar boundary bool\n\tfor {\n\t\tr, l := utf8.DecodeRune(b[length:])\n\t\tstate, boundary = transitionWordBreakState(state, r, b[length+l:], \"\")\n\n\t\tif boundary {\n\t\t\treturn b[:length], b[length:], state\n\t\t}\n\n\t\tlength += l\n\t\tif len(b) <= length {\n\t\t\treturn b, nil, wbAny\n\t\t}\n\t}\n}", "func (p *Entry) Words() [][]byte {\n\treturn regexp.MustCompile(wordRegex).FindAll(p.Body, -1)\n}", "func (b *GroupsEditBuilder) ObsceneWords(v []string) *GroupsEditBuilder {\n\tb.Params[\"obscene_words\"] = v\n\treturn b\n}", "func findLadders(beginWord string, endWord string, wordList []string) [][]string {\n\n}", "func ToWord() Tokeniser {\n\treturn createToWordTokeniser()\n}", "func SplitStringSliceByLimit(in []string, limit uint) [][]string {\n\tvar stringSlice []string\n\tsliceSlice := make([][]string, 0, len(in)/int(limit)+1)\n\tfor len(in) >= int(limit) {\n\t\tstringSlice, in = in[:limit], in[limit:]\n\t\tsliceSlice = append(sliceSlice, stringSlice)\n\t}\n\tif len(in) > 0 {\n\t\tsliceSlice = append(sliceSlice, in)\n\t}\n\treturn sliceSlice\n}", "func collectWords(filename string) ([]string, []string) {\n\tf, err := ioutil.ReadFile(filename)\n\tif err != nil {\n\t\tfmt.Println(\"Cannot access file.\")\n\t\tprintUsage()\n\t}\n\tsize, err := os.Stat(filename)\n\tif err != nil {\n\t\terrHandler(err)\n\t}\n\tfmt.Println((\"Using file \" + filename + \", \" +\n\t\tstrconv.FormatInt(size.Size()/1000, 10) + \"kb\"))\n\tdoc, err := prose.NewDocument(string(f))\n\tif err != nil {\n\t\terrHandler(err)\n\t}\n\ttok := doc.Tokens()\n\tvar nouns, adjs []string\n\trepunc, err := regexp.Compile(\"[.,;!/\\\\'\\\"\\\\p{Pd}]+\")\n\tif err != nil {\n\t\terrHandler(err)\n\t}\n\tfor _, word := range tok {\n\t\ttword := strings.Title(strings.ToLower(word.Text))\n\t\ttword = repunc.ReplaceAllString(tword, \"\")\n\t\tif len(tword) > 1 {\n\t\t\tif word.Tag == \"NN\" || word.Tag == \"NNS\" {\n\t\t\t\tmember, _ := isMember(tword, nouns)\n\t\t\t\tif !member {\n\t\t\t\t\tnouns = append(nouns, tword)\n\t\t\t\t}\n\t\t\t} else if word.Tag == \"JJ\" {\n\t\t\t\tmember, _ := isMember(tword, adjs)\n\t\t\t\tif !member {\n\t\t\t\t\tadjs = append(adjs, tword)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn nouns, adjs\n}", "func (w *W3W) Words(words What3Words, opts *Options) (*Position, error) {\n\tvals := url.Values{}\n\n\tvals.Set(\"key\", w.apikey)\n\tvals.Set(\"string\", strings.Join(words[:], \".\"))\n\n\tpos, err := w.exec(endpoint+\"/w3w\", &vals, opts, &Position{})\n\n\treturn pos.(*Position), err\n}", "func findWords(words []string) []string {\n \n}", "func text2words(raw string) []string {\n\tparts := strings.Fields(raw)\n\tout := make([]string, 0, len(parts))\n\tfor _, word := range parts {\n\t\tword = strings.Trim(word, `'\".,:;-()`)\n\t\tword = strings.ToLower(word)\n\t\tif isAZ(word) {\n\t\t\tout = append(out, word)\n\t\t}\n\t}\n\treturn out\n}", "func words(s string) int {\n\tb, n := true, 0\n\tfor _, r := range []rune(s) {\n\t\tswitch {\n\t\tcase unicode.IsSpace(r):\n\t\t\tb = true\n\t\tcase b:\n\t\t\tb = false\n\t\t\tn++\n\t\t}\n\t}\n\treturn n\n}", "func ControlWordOf(count int, controlType ControlType, param uint32) ControlWord {\n\treturn (ControlWord(count) << 20) | (ControlWord(controlType&0x7) << 17) | ControlWord(param&ControlWordParamLimit)\n}", "func SplitWord(data uint16) (uint8, uint8) {\n\treturn uint8((data & 0xFF00) >> 8), uint8(data & 0xFF)\n}", "func CreateInputSlice(input string) {\n\tindicesNeedingTrailSpace := regexp.MustCompile(`[/\\{+/\\(+/\\)+/\\}+/\\[+/\\]+]`).FindAllStringIndex(input, -1)\n\tfor i := len(indicesNeedingTrailSpace) - 1; i >= 0; i-- {\n\t\tsubstr := \" \" + input[indicesNeedingTrailSpace[i][0]:indicesNeedingTrailSpace[i][1]] + \" \"\n\t\tinput = input[:indicesNeedingTrailSpace[i][0]] + substr + input[indicesNeedingTrailSpace[i][1]:]\n\t}\n\n\tinput = strings.TrimSpace(input)\n\tsplitString := strings.Split(input, \" \")\n\n\tfor i := 0; i < len(splitString); i++ {\n\t\tif splitString[i] != \"\" && splitString[i] != \" \" {\n\t\t\tinputTokens = append(inputTokens, token{splitIndexIncr, splitString[i]})\n\t\t\tsplitIndexIncr++\n\t\t}\n\t}\n\n\ttotalTokens = len(inputTokens)\n}", "func IndexizeWord(w string) string {\n\treturn strings.TrimSpace(strings.ToLower(w))\n}", "func SplitForIndexing(s string, normalize bool) []string {\n\tif normalize {\n\t\ts = NormalizeRunes(s)\n\t}\n\tchunks := regexpWhiteSpace.Split(s, -1)\n\trealChunks := make([]string, 0, len(chunks))\n\n\tfor _, chunk := range chunks {\n\t\t// Accepted :\n\t\t// All words having at least 3 characters\n\t\t// All 1-digits words and 2-digits words\n\t\tif len(chunk) >= 3 || RegexpDigitsOnly.MatchString(chunk) {\n\t\t\trealChunks = append(realChunks, NormalizeRunes(chunk))\n\t\t}\n\t}\n\n\t// Stategy for dash-compound words: all bits get indexed (in addition to the full compound)\n\tfor _, chunk := range chunks {\n\t\tif strings.Contains(chunk, \"-\") {\n\t\t\tfor _, bit := range strings.Split(chunk, \"-\") {\n\t\t\t\tif bit != \"\" {\n\t\t\t\t\trealChunks = append(realChunks, bit)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\n\treturn realChunks\n}", "func tokenizeSubjectIntoSlice(tts []string, subject string) []string {\n\tstart := 0\n\tfor i := 0; i < len(subject); i++ {\n\t\tif subject[i] == btsep {\n\t\t\ttts = append(tts, subject[start:i])\n\t\t\tstart = i + 1\n\t\t}\n\t}\n\ttts = append(tts, subject[start:])\n\treturn tts\n}", "func wordBreak(s string, wordDict []string) bool {\n\twordMap := make(map[string]struct{})\n\tfor i := 0; i < len(wordDict); i++ {\n\t\twordMap[wordDict[i]] = struct{}{}\n\t}\n\n\tmemoize := make(map[string]bool)\n\treturn wordBreakRecurse(s, wordMap, 0, memoize)\n}", "func findSubstring(s string, words []string) []int {\n\tres := []int{}\n\tls, lw := len(s), len(words)\n\tif lw == 0 || ls == 0 {\n\t\treturn res\n\t}\n\n\tmp := make(map[string]int)\n\tfor _, v := range words {\n\t\tmp[v]++\n\t}\n\n\twLen := len(words[0])\n\twindow := wLen * lw\n\tfor i := 0; i < wLen; i++ {\n\t\tfor j := i; j+window <= ls; j = j + wLen {\n\t\t\ttmp := s[j : j+window]\n\t\t\ttp := make(map[string]int)\n\t\t\tfor k := lw - 1; k >= 0; k-- {\n\t\t\t\t// get the word from tmp\n\t\t\t\tword := tmp[k*wLen : (k+1)*wLen]\n\t\t\t\tcount := tp[word] + 1\n\t\t\t\tif count > mp[word] {\n\t\t\t\t\tj = j + k*wLen\n\t\t\t\t\tbreak\n\t\t\t\t} else if k == 0 {\n\t\t\t\t\tres = append(res, j)\n\t\t\t\t} else {\n\t\t\t\t\ttp[word] = count\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn res\n}", "func countWordsInString(sInput string) map[string]int {\n\t//change the input into a rune array\n\tsTemp := []rune(sInput)\n\treturn (countWordsInRunes(sTemp))\n}", "func TextToWords(text string) []string {\n\twords := []string{}\n\tfor _, s := range SentenceTokenizer.Tokenize(text) {\n\t\twords = append(words, strings.Fields(s)...)\n\t}\n\treturn words\n}", "func WordsToMap(s string) map[string]int {\n\twords := strings.Fields(s)\n\twordmap := map[string]int{}\n\tfor _, word := range words {\n\t\twordmap[word]++\n\t}\n\treturn wordmap\n}", "func WordEdits(word string) []string {\n\tcloseWords := []string{}\n\n\t// Perhaps they mistyped a letter try mutating one\n\tfor i := 0; i < len(word); i++ {\n\t\tfor letter := 'a'; letter <= 'z'; letter++ {\n\t\t\tcloseWords = append(closeWords, word[:i]+string(letter)+word[i+1:])\n\t\t}\n\t}\n\n\t// Perhaps they missed a letter try adding one in\n\tfor i := 0; i < len(word)+1; i++ {\n\t\tfor letter := 'a'; letter <= 'z'; letter++ {\n\t\t\tcloseWords = append(closeWords, word[:i]+string(letter)+word[i:])\n\t\t}\n\t}\n\n\t// Perhaps they added an extra letter try removing it\n\tfor i := 0; i < len(word); i++ {\n\t\tcloseWords = append(closeWords, word[:i]+word[i+1:])\n\t}\n\n\treturn closeWords\n}", "func mkWrdArray(rawBytes []byte) []TaggedWord {\n\n\tcurrByte := 0\n\twordStart := currByte\n\tvar taggedWords []TaggedWord = make([]TaggedWord, 0)\n\n\tfor currByte < len(rawBytes) {\n\t\tif isSpace(rawBytes[currByte]) {\n\t\t\tif wordStart != currByte { // add the word if I can\n\t\t\t\ttaggedWords = append(taggedWords, TaggedWord{word: string(rawBytes[wordStart:currByte]), tag: \"\", byteStart: wordStart})\n\t\t\t}\n\t\t\tcurrByte++\n\t\t\twordStart = currByte\n\t\t} else if isSymbol(rawBytes[currByte]) {\n\t\t\tif wordStart != currByte { // add the word if I can\n\t\t\t\ttaggedWords = append(taggedWords, TaggedWord{word: string(rawBytes[wordStart:currByte]), tag: \"\", byteStart: wordStart})\n\t\t\t}\n\t\t\twordStart = currByte\n\t\t\tcurrByte++\n\t\t\ttaggedWords = append(taggedWords, TaggedWord{word: string(rawBytes[wordStart:currByte]), tag: \"\", byteStart: wordStart})\n\t\t\twordStart = currByte\n\t\t} else {\n\t\t\tcurrByte++\n\t\t}\n\t}\n\ttaggedWords = append(taggedWords, TaggedWord{word: string(rawBytes[wordStart:currByte]), tag: \"\", byteStart: wordStart})\n\treturn taggedWords\n}", "func CountWords(bytes *[]byte) int {\n\tnumWords := 0\n\tisWord := false\n\n\tfor _, c := range *bytes {\n\t\tr := rune(c)\n\t\tif unicode.IsLetter(r) {\n\t\t\tisWord = true\n\t\t} else if isWord && !unicode.IsLetter(r) {\n\t\t\tnumWords++\n\t\t\tisWord = false\n\t\t}\n\t}\n\n\treturn numWords\n}", "func stringToSlice(s, d string) []string {\n\tss := make([]string, 0)\n\tif strings.Contains(s, d) {\n\t\tss = strings.Split(s, d)\n\t} else {\n\t\tss = append(ss, s)\n\t}\n\treturn ss\n}", "func splitNumbersString(pinyin string) []word {\n words := strings.Fields(pinyin)\n\n splitPinyin := make([]word, len(words))\n\n for i, v := range words {\n lastChar := v[len(v)-1:]\n\n if unicode.IsNumber(rune(lastChar[0])) {\n n, _ := strconv.Atoi(lastChar)\n\n // convert 5's to 0's for neutral tones\n if n == 5 {\n n = 0\n }\n splitPinyin[i] = word{syllable: v[0 : len(v)-1], tone: n}\n } else {\n splitPinyin[i] = word{syllable: v, tone: -1}\n }\n }\n return splitPinyin\n}", "func NewWords(log *logrus.Logger, textfile string) *Words {\n\tw := &Words{\n\t\tlog: log,\n\t}\n\tw.loadData(textfile)\n\treturn w\n}", "func (gn *Gen) GenWholeWordWavs() {\n\t// first the whole words\n\tfor i := 0; i < 4; i++ {\n\t\tgn.GenTriCVWavs(gn.Syl1[i], gn.Syl2[i], gn.Syl3[i], 3)\n\t}\n}", "func getWord(id int64) string {\n\treturn diceware8k[id&8191]\n}", "func foldString(maxLength int, prefix, s string) string {\n\tvar foldedBuffer bytes.Buffer\n\tlineBuffer := bytes.NewBufferString(prefix)\n\tlineLength := lineBuffer.Len()\n\n\tfor _, word := range strings.Split(s, \" \") {\n\t\twordLength := len(word)\n\t\tif wordLength+lineLength+1 <= maxLength {\n\t\t\tlineBuffer.WriteString(word)\n\t\t\tlineBuffer.WriteString(\" \")\n\t\t\tlineLength += wordLength + 1\n\t\t} else {\n\t\t\tfoldedBuffer.Write(lineBuffer.Bytes())\n\t\t\tfoldedBuffer.WriteString(\"\\r\\n \")\n\t\t\tlineBuffer.Reset()\n\t\t\tlineBuffer.WriteString(word)\n\t\t\tlineBuffer.WriteString(\" \")\n\t\t\tlineLength = wordLength + 1\n\t\t}\n\t}\n\tfoldedBuffer.Write(lineBuffer.Bytes())\n\tfoldedBuffer.WriteString(\"\\r\\n\")\n\treturn foldedBuffer.String()\n}", "func bigramWordByCopy(str string) []string {\n\tss := strings.Split(str, \" \")\n\n\t// Handle unexpected string input:\n\t// ss = \"\" => []string{\"\"}\n\t// ss = \"foobar\" => []string{\"foobar\"}\n\tif len(ss) <= 1 {\n\t\treturn ss\n\t}\n\n\tbigram := make([]string, len(ss)-1)\n\tfor i := 0; i < len(ss)-1; i++ {\n\t\t// Counts the length of primary and secondary words and whitespace\n\t\t// and copy it to the element of slice.\n\t\tbigram[i] = str[:(len(ss[i])+1)+len(ss[i+1])]\n\t\t// Drop the primary word and whitespace.\n\t\tstr = str[len(ss[i])+1:]\n\t}\n\treturn bigram\n}", "func loadWords() []string {\n\twords, err := readFileByLine()\n\tif err != nil {\n\t\tpanic(\"load english cut word failed,\" + err.Error())\n\t}\n\treturn words\n}", "func countWordLenByCounter(str string) []int {\n\tif len(str) == 0 {\n\t\treturn []int{0}\n\t}\n\tstr = strings.Replace(str, \",\", \"\", -1)\n\n\tcount := []int{}\n\tcounter := 0\n\tfor _, s := range str {\n\t\tif (s != ' ') && (s != '.') {\n\t\t\tcounter++\n\t\t\tcontinue\n\t\t}\n\t\tcount = append(count, counter)\n\t\tcounter = 0\n\t}\n\treturn count\n}", "func WordEol(s string, wordIdx int) string {\n\tsplit := strings.Split(s, \" \")\n\tif wordIdx > -1 && len(split) >= wordIdx {\n\t\treturn strings.Join(split[wordIdx:], \" \")\n\t}\n\n\treturn \"\"\n}", "func Tokenize(s Sentence) []string {\n\treturn strings.Split(s.Input, \" \")\n}", "func WordWrap(s string) template.HTML {\n\tmaxTextSize := 53\n\tparts := strings.SplitAfter(s, \" \")\n\tres := \"\"\n\tnb := 0\n\tfor _, part := range parts {\n\t\tnb += len(part)\n\t\tres += part\n\t\tif nb > maxTextSize {\n\t\t\tnb = 0\n\t\t\tres += \"<br />\"\n\t\t}\n\t}\n\treturn template.HTML(res)\n}", "func wordBeginInString(s string, r0Pos int) int {\n\tr0, r1Delta := wFirstSequenceInString(s[r0Pos:])\n\tr1, _ := wFirstSequenceInString(s[r0Pos+r1Delta:])\n\tl0, l0Pos := wLastSequenceInString(s[:r0Pos])\n\n\tfor r0Pos >= 0 {\n\t\tl1, l1Pos := wLastSequenceInString(s[:l0Pos])\n\t\tlOddRI := wIsOpenRIInString(s[:l1Pos], l1, l0)\n\t\tif wDecision(l1, l0, lOddRI, r0, r1) {\n\t\t\treturn r0Pos\n\t\t}\n\t\tr1 = r0\n\t\tr0 = l0\n\t\tl0 = l1\n\t\tr0Pos = l0Pos\n\t\tl0Pos = l1Pos\n\t}\n\n\treturn 0\n}", "func WordCount(s string) map[string]int {\n\twcount := make(map[string]int)\n\tfor _, v := range strings.Fields(s) {\n\t\twcount[v] = wcount[v] + 1\n\t}\n\treturn wcount\n}", "func EvaluateWord(text string, matter string) (word model.Word) {\n\ttext = strings.ToLower(strings.Replace(text, \" \", \"\", -1))\n\tword.Length = len(text)\n\tword.Text = text\n\tword.Matter.Name = matter\n\treturn\n}", "func generateCutWordMap(words []string) {\n\twordCost = make(map[string]float64)\n\tvar wordLen int\n\tlogLen := math.Log(float64(len(words)))\n\tfor idx, word := range words {\n\t\twordLen = len(word)\n\t\tif wordLen > maxLenWord {\n\t\t\tmaxLenWord = wordLen\n\t\t}\n\t\twordCost[word] = math.Log(logLen * float64(idx+1))\n\t}\n}", "func FindWords(txt, pat string, T []int) (offsets []int) {\n\treturn FindYourOwnWay(txt, pat, T, -1, isWord)\n}", "func (w KyTeaWord) Word(util StringUtil) Word {\n\tsurface := w.Surface(util)\n\ttagsLen := w.TagsLen()\n\ttags := make([][]Tag, tagsLen)\n\tfor i := 0; i < tagsLen; i++ {\n\t\tcandidateTagsLen := w.CandidateTagsLen(i)\n\t\ttmp := make([]Tag, candidateTagsLen)\n\t\tfor j := 0; j < candidateTagsLen; j++ {\n\t\t\ttmp[j].Feature, tmp[j].Score = w.Tag(i, j, util)\n\t\t}\n\t\ttags[i] = tmp\n\t}\n\treturn Word{\n\t\tSurface: surface,\n\t\tTags: tags,\n\t}\n}", "func compose(dict map[string]bool, word []byte) []string {\n\n\t// The approach is to find all words in the dictionary\n\t// that are prefixes of the argument word and then recurse\n\t// to perform the rest of the word's parsing.\n\n\t// Search for prefixes by prefix length.\n\t//\n\tword_len := len(word)\n\tfor prefix_len := 1; prefix_len <= word_len; prefix_len++ {\n\n\t\tprefix := string(word[0:prefix_len])\n\n\t\t// Lookup the prefix in the dictionary.\n\t\t// If the prefix does not match, try the next length of prefix.\n\t\tif !dict[prefix] {\n\t\t\tcontinue\n\t\t}\n\n\t\tprefix_word_list := []string{prefix}\n\n\t\t// If the prefix consumed the entire string, then the parse succeeded.\n\t\tif prefix_len == word_len {\n\t\t\treturn prefix_word_list\n\t\t}\n\n\t\t// If the prefix did not consume the entire string,\n\t\t// attempt to parse the suffix by recursing.\n\t\tsuffix_word_list := compose(dict, []byte(word[prefix_len:word_len]))\n\n\t\t// If the suffix parsed, then the parse succeeded.\n\t\tif len(suffix_word_list) > 0 {\n\t\t\treturn append(prefix_word_list, suffix_word_list...)\n\t\t}\n\n\t\t// If the suffix parse failed, this parse has failed, but we\n\t\t// still need to go around to try the remaining prefix lengths.\n\t}\n\n\treturn []string{}\n}", "func FirstWordInString(str string, state int) (word, rest string, newState int) {\n\t// An empty byte slice returns nothing.\n\tif len(str) == 0 {\n\t\treturn\n\t}\n\n\t// Extract the first rune.\n\tr, length := utf8.DecodeRuneInString(str)\n\tif len(str) <= length { // If we're already past the end, there is nothing else to parse.\n\t\treturn str, \"\", wbAny\n\t}\n\n\t// If we don't know the state, determine it now.\n\tif state < 0 {\n\t\tstate, _ = transitionWordBreakState(state, r, nil, str[length:])\n\t}\n\n\t// Transition until we find a boundary.\n\tvar boundary bool\n\tfor {\n\t\tr, l := utf8.DecodeRuneInString(str[length:])\n\t\tstate, boundary = transitionWordBreakState(state, r, nil, str[length+l:])\n\n\t\tif boundary {\n\t\t\treturn str[:length], str[length:], state\n\t\t}\n\n\t\tlength += l\n\t\tif len(str) <= length {\n\t\t\treturn str, \"\", wbAny\n\t\t}\n\t}\n}", "func wordToInt(s string) int {\n\tsum := 0\n\tfor _, v := range s {\n\t\tsum += int(v) - 'A' + 1\n\t}\n\treturn sum\n\n}", "func (fn *TruncateFn) SplitRestriction(_ []byte, rest offsetrange.Restriction) []offsetrange.Restriction {\n\treturn rest.EvenSplits(2)\n}", "func Generate(words int) ([]string, error) {\n\tlist := make([]string, 0, words)\n\tseen := make(map[string]struct{}, words)\n\n\tfor i := 0; i < words; i++ {\n\t\tn, err := RollWord(digits)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tword := WordAt(n)\n\t\tif _, ok := seen[word]; ok {\n\t\t\ti--\n\t\t\tcontinue\n\t\t}\n\n\t\tlist = append(list, word)\n\t\tseen[word] = struct{}{}\n\t}\n\n\treturn list, nil\n}" ]
[ "0.7728493", "0.5840942", "0.54964876", "0.5313557", "0.53101194", "0.5250707", "0.523979", "0.5197165", "0.5141871", "0.5132436", "0.50922287", "0.49990085", "0.49727967", "0.49727017", "0.49687827", "0.49514958", "0.49503365", "0.48901954", "0.48864898", "0.48817423", "0.4837451", "0.48054814", "0.4803117", "0.47909352", "0.47832704", "0.47827432", "0.47752172", "0.4763622", "0.4735842", "0.47287557", "0.46869493", "0.4683388", "0.46702686", "0.46622157", "0.46338448", "0.46051678", "0.4603205", "0.4598943", "0.4584575", "0.45798582", "0.45784885", "0.45566058", "0.45448616", "0.4544024", "0.4540316", "0.4537294", "0.4524631", "0.4498609", "0.44926262", "0.4486093", "0.4469519", "0.44691962", "0.4465725", "0.44538498", "0.44460398", "0.44395667", "0.44350338", "0.44256732", "0.440733", "0.44056728", "0.4400321", "0.4392671", "0.4385313", "0.43843004", "0.4367121", "0.43663707", "0.43654513", "0.43610933", "0.43566433", "0.4338007", "0.4330621", "0.43278047", "0.43212092", "0.43149632", "0.43132952", "0.43048304", "0.4300665", "0.4298827", "0.4297219", "0.4294277", "0.42832932", "0.4269795", "0.42639023", "0.42591956", "0.4250484", "0.42392522", "0.42284435", "0.42103463", "0.42100793", "0.42044604", "0.42032886", "0.4203009", "0.4202392", "0.418814", "0.41868752", "0.41788155", "0.41741782", "0.41713023", "0.41687644", "0.41655597" ]
0.77382016
0
Finish return data is written here
func (b *BaseController) Finish() { so := pack(b.Data["json"], b.GetObjectName()) if so.ErrorCode != 0 { b.Ctx.ResponseWriter.WriteHeader(500) } b.Data["json"] = &so b.ServeJSON() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (op operation) Finish(res interface{}) { dyngo.FinishOperation(op, res) }", "func (p *ArtAddressPacket) finish() {\n\tp.Header.finish()\n}", "func (c *Controller) Finish() {}", "func (res *Results) Finalize() {\n res.Replies = len(res.Took)\n res.min()\n res.max()\n res.avg()\n res.med()\n res.pct()\n\n // Code counts\n for _, code := range res.Code {\n if code < 100 { // ignore\n } else if code < 200 {\n res.Code1xx++\n } else if code < 300 {\n res.Code2xx++\n } else if code < 400 {\n res.Code3xx++\n } else if code < 500 {\n res.Code4xx++\n } else if code < 600 {\n res.Code5xx++\n }\n }\n\n // Error counts\n res.ErrorsTotal = len(res.Errors)\n\n for _, err := range res.Errors {\n e := err.(*url.Error).Err.(*net.OpError).Error()\n if strings.Contains(e, \"connection refused\") {\n res.ErrorsConnRefused++\n } else if strings.Contains(e, \"connection reset\") {\n res.ErrorsConnReset++\n } else if strings.Contains(e, \"connection timed out\") {\n res.ErrorsConnTimeout++\n } else if strings.Contains(e, \"no free file descriptors\") {\n res.ErrorsFdUnavail++\n } else if strings.Contains(e, \"no such host\") {\n res.ErrorsAddrUnavail++\n } else {\n res.ErrorsOther++\n }\n }\n}", "func Finish() {\n\tsyscall.Syscall(gpFinish, 0, 0, 0, 0)\n}", "func (p *Header) finish() {\n\tp.ID = ArtNet\n\tp.Version = version.Bytes()\n\tp.swapOpCode()\n}", "func (b *Builder) Finish() []byte {\n\tbd := b.Done()\n\tbuf := make([]byte, bd.Size)\n\twritten := bd.Copy(buf)\n\ty.AssertTrue(written == len(buf))\n\treturn buf\n}", "func (p *Header) finish() {\n\tp.ID = ArtNet\n\tp.OpCode = code.OpCode(uint16(p.OpCode&0xff) + uint16(p.OpCode>>8))\n\tp.Version = version.Bytes()\n}", "func (r *SnpReportReqABI) Finish(_ BinaryConvertible) error { return nil }", "func (c *Command) Finish() {}", "func (data KeepAliveData) Complete() {\n}", "func (c FinalOutput) Close() {}", "func (f *Finisher) Finish(result interface{}) {\n\tif f.isFinished {\n\t\treturn\n\t}\n\tf.isFinished = true\n\tf.callback(result)\n}", "func Finalize(w http.ResponseWriter, object interface{}) {\n\tfinalData := data{\n\t\tItem: object,\n\t}\n\tfinal := response{\n\t\tData: finalData,\n\t}\n\tjs, err := json.Marshal(final)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tresponseStatus := status{\n\t\tStatusCode: http.StatusOK,\n\t\tStatusText: http.StatusText(http.StatusOK),\n\t}\n\tsendOutJs(w, responseStatus, js)\n}", "func (p *Preparer) Finish() {\n\tp.Writer.Finish()\n}", "func (buf *RespBuffer) Finished() {\n\tbuf.finLock <- struct{}{}\n\tbuf.finished <- struct{}{}\n\t<-buf.finLock\n}", "func (s *PostgresSlurper) Finish() {\n}", "func (operation *Operation) Finish() {\n\toperation.Action.Finish()\n\toperation.updateInDatabase()\n}", "func (l Log) finish(path string, file *os.File) (Log, error) {\n\tl.EndTime = time.Now()\n\tl.Status = \"Finished\"\n\treturn l, l.saveAndWriteToLog(path, file, \"Finished\")\n}", "func (t *Trace) Finish() {\n\tt.t.Finish()\n}", "func (t *Trace) Finish() {\n\tt.t.Finish()\n}", "func (m *MockDoer) Finish() interface{} {\n\targs := m.MethodCalled(\"Finish\")\n\n\treturn args.Get(0)\n}", "func (res *Result) End() (err os.Error) {\n for err == nil && res.db.unreaded_rows {\n _, err = res.GetRow()\n }\n return\n}", "func (s *Stream) objDone(obj *Obj, err error) {\n\tvar rc int64\n\tif obj.prc != nil {\n\t\trc = obj.prc.Dec()\n\t\tcmn.Assert(rc >= 0) // remove\n\t}\n\t// SCQ completion callback\n\tif rc == 0 {\n\t\tif obj.Callback != nil {\n\t\t\tobj.Callback(obj.Hdr, obj.Reader, obj.CmplPtr, err)\n\t\t} else if s.callback != nil {\n\t\t\ts.callback(obj.Hdr, obj.Reader, obj.CmplPtr, err)\n\t\t}\n\t}\n\tif obj.Reader != nil {\n\t\tobj.Reader.Close() // NOTE: always closing\n\t}\n}", "func (m *Master) Done() bool {\n\n\t// Your code here.\n\n\treturn m.end\n}", "func (b *Writer) Finish() []byte {\n\tif b.closed {\n\t\tpanic(\"opeartion on closed block writer\")\n\t}\n\n\tb.closed = true\n\n\t// C++ leveldb need atleast 1 restart entry\n\tif b.restarts == nil {\n\t\tb.restarts = make([]uint32, 1)\n\t}\n\n\tfor _, restart := range b.restarts {\n\t\tbinary.Write(b.buf, binary.LittleEndian, restart)\n\t}\n\tbinary.Write(b.buf, binary.LittleEndian, uint32(len(b.restarts)))\n\n\treturn b.buf.Bytes()\n}", "func FinishFunc(pipe *pipe.Pipe) error {\n\tfor _, handler := range pipe.Handlers {\n\t\tif handler == \"yacht\" {\n\t\t\tgoto export\n\t\t}\n\t}\n\treturn nil\n\nexport:\n\tdataset := pipe.Get(\"yacht\").(*Dataset)\n\n\tpath := pipe.GetDistPath(\"templates\", \"yacht\")\n\tos.MkdirAll(path, os.ModePerm)\n\n\tres, err := json.MarshalIndent(dataset.Templates, \"\", \" \")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"marshal templates error: %s\", err.Error())\n\t}\n\n\tfilename := path + \"/yacht.json\"\n\tioutil.WriteFile(filename, res, os.ModePerm)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"write template file [%s] error: %s\", filename, err.Error())\n\t}\n\n\treturn nil\n}", "func (m *OutboundMock) Finish() {\n\tm.MinimockFinish()\n}", "func (w *Writer) Finalize() []byte {\n\t// update size\n\tvar length = w.index\n\tw.buffer[2] = byte(length)\n\tw.buffer[3] = byte(length >> 8)\n\n\treturn w.buffer[:length]\n}", "func (localOptimizer) finish(operation chan<- Task, result <-chan Task) {\n\t// Guarantee that result is closed before operation is closed.\n\tfor range result {\n\t}\n}", "func (a *Activity) Finish() {\n\tC.ANativeActivity_finish(a.cptr())\n}", "func (r *Response) End() {\n\tr.apiTest.run()\n}", "func (l LcvTrans) End(args ...interface{}) error {\n\tlog.INFO.Printf(\"LCV transfer perf-test done.\")\n\treturn nil\n}", "func Done(Pack) interface{} {\n\treturn nil\n}", "func (w *writable) Finish() error {\n\tw.g.flush()\n\treturn w.w.Finish()\n}", "func (_GameJam *GameJamTransactor) Finish(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _GameJam.contract.Transact(opts, \"finish\")\n}", "func Finish() {\n\tC.glowFinish(gpFinish)\n}", "func Finish() {\n\tC.glowFinish(gpFinish)\n}", "func (t *Transport) finish(stream *Stream, out []byte) (n int) {\n\tatomic.AddUint64(&t.nTxfin, 1)\n\tvar scratch [16]byte\n\tn = tag2cbor(tagCborPrefix, out) // prefix\n\tout[n] = 0xc8 // 0xc8 (end stream, 0b110_01000 <tag,8>)\n\tn++ //\n\tm := tag2cbor(stream.opaque, scratch[:]) // tag-opaque\n\tscratch[m] = 0x40 // zero-len byte-string\n\tm++\n\tn += valbytes2cbor(scratch[:m], out[n:]) // packet\n\tout[n] = 0xff // 0xff CBOR indefinite end.\n\tn++\n\treturn n\n}", "func (j *Job) finish() {\n\tst := <-j.Signal\n\tj.mu.Lock()\n\tj.st = monitor.READY\n\tj.comp = append(j.comp, monitor.CompletedReport{\n\t\tStatus: st.String(),\n\t\tOutput: j.buffer.String(),\n\t\tId: int16(len(j.comp) + 1),\n\t})\n\tj.mu.Unlock()\n}", "func finish(_ js.Value, args []js.Value) interface{} {\n\tmsgB, err := base64.URLEncoding.DecodeString(args[0].String())\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\tmk, err := state.Finish(msgB)\n\tif err != nil {\n\t\treturn nil\n\t}\n\thkdf := hkdf.New(sha256.New, mk, nil, nil)\n\tkey := [32]byte{}\n\t_, err = io.ReadFull(hkdf, key[:])\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\tdst := js.Global().Get(\"Uint8Array\").New(32)\n\tjs.CopyBytesToJS(dst, key[:])\n\n\treturn dst\n}", "func (process *Process) Finish() {\n\tprocess.Action.Finish()\n\tprocess.updateInDatabase()\n}", "func (w *tWriter) finish() (f *tFile, err error) {\n\tdefer w.close()\n\terr = w.tw.Close()\n\tif err != nil {\n\t\treturn\n\t}\n\tif !w.t.noSync {\n\t\terr = w.w.Sync()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t}\n\tf = newTableFile(w.fd, int64(w.tw.BytesLen()), internalKey(w.first), internalKey(w.last))\n\treturn\n}", "func (streamPair *StreamPair) maybeFinish() {\n\tswitch {\n\tcase streamPair.request == nil:\n\t\tlog.Fatalf(\"[%v] request should always be non-nil, since it's set when bidis are created\", streamPair.key)\n\tcase !streamPair.request.done:\n\t\t//log.Printf(\"[%v] still waiting on first stream\", streamPair.key)\n\tcase streamPair.response == nil:\n\t\t//log.Printf(\"[%v] no second stream yet\", streamPair.key)\n\tcase !streamPair.response.done:\n\t\t//log.Printf(\"[%v] still waiting on second stream\", streamPair.key)\n\tdefault:\n\t\t//log.Printf(\"[%v] FINISHED, bytes: %d tx, %d rx\", streamPair.key, streamPair.request.bytes, streamPair.response.bytes)\n\t\trecord(streamPair)\n\t}\n}", "func (eb *ElectrumBackend) Finish() {\n\tclose(eb.doneCh)\n\teb.removeAllNodes()\n\t// TODO: we could gracefully disconnect from all the nodes. We currently don't, because the\n\t// program is going to terminate soon anyways.\n}", "func (worker *Worker) finalize(\n\tsignature *TaskSignature, result reflect.Value, err error,\n) {\n\tif err != nil {\n\t\tlog.Printf(\"Failed processing %s\", signature.Name)\n\t\tlog.Printf(\"Error = %v\", err)\n\n\t\tfor _, errorTask := range signature.OnError {\n\t\t\t// Pass error as a first argument to error callbacks\n\t\t\targs := append([]TaskArg{TaskArg{\n\t\t\t\tType: reflect.TypeOf(err).String(),\n\t\t\t\tValue: reflect.ValueOf(err).Interface(),\n\t\t\t}}, errorTask.Args...)\n\t\t\terrorTask.Args = args\n\t\t\tworker.server.SendTask(errorTask)\n\t\t}\n\t\treturn\n\t}\n\n\tlog.Printf(\"Finished processing %s\", signature.Name)\n\tlog.Printf(\"Result = %v\", result.Interface())\n\n\tfor _, successTask := range signature.OnSuccess {\n\t\tif signature.Immutable == false {\n\t\t\t// Pass results of the task to success callbacks\n\t\t\targs := append([]TaskArg{TaskArg{\n\t\t\t\tType: result.Type().String(),\n\t\t\t\tValue: result.Interface(),\n\t\t\t}}, successTask.Args...)\n\t\t\tsuccessTask.Args = args\n\t\t}\n\t\tworker.server.SendTask(successTask)\n\t}\n}", "func (o *Output) cleanup() {\r\n\to.writer.Flush()\r\n\to.file.Close()\r\n}", "func (w *Writer) End()", "func (l *Logger) Finished(url string, code int, latency time.Duration, urls, errors int) {\n\tl.m.Lock()\n\tdefer l.m.Unlock()\n\tl.Log = append(l.Log, fmt.Sprintf(\"finish %s: %d, %d, %d\", url, code, urls, errors))\n}", "func (builder *OnDiskBuilder) finish() {\n\t// Write the docID to path map\n\tgo builder.index.WriteDocIDToFilePath(\"./saved/meta/idToPath\")\n\t// Find out which terms are in memory, on disk or both\n\tonDiskOnly, inMemoryOnly, onDiskAndInMemory := builder.index.CategorizeTerms()\n\t// Get from frequency scores to tf-idf\n\tvar wg sync.WaitGroup\n\twg.Add(3)\n\tgo builder.writeTfIdfInMemoryTerms(inMemoryOnly, &wg)\n\tgo builder.tfIdfOnDiskTerms(onDiskOnly, &wg)\n\tgo builder.mergeDiskMemoryThenTfIdfTerms(onDiskAndInMemory, &wg)\n\twg.Wait()\n\tclose(builder.writingChannel)\n}", "func (b *bar) finish() {\n\tb.mutex.Lock()\n\tdefer b.mutex.Unlock()\n\n\tif b.finished {\n\t\treturn\n\t}\n\n\tclose(b.finishChan)\n\tb.finished = true\n\tb.clear()\n}", "func (b *builder) finish() {\n\tfmt.Fprintln(b, \"}\")\n}", "func (Sm *ServiceMaster) ReceiveFinish(filenames interface{}, response *bool) error {\n\tfnames, _ := filenames.([]string)\n\tSm.Mr.ReceiveFinish(fnames)\n\t*response = true\n\treturn nil\n}", "func (udc *unfinishedDownloadChunk) returnMemory() {\n\t// The maximum amount of memory is the pieces completed plus the number of\n\t// workers remaining.\n\tmaxMemory := uint64(udc.workersRemaining+udc.piecesCompleted) * udc.staticPieceSize\n\t// If enough pieces have completed, max memory is the number of registered\n\t// pieces plus the number of completed pieces.\n\tif udc.piecesCompleted >= udc.erasureCode.MinPieces() {\n\t\t// udc.piecesRegistered is guaranteed to be at most equal to the number\n\t\t// of overdrive pieces, meaning it will be equal to or less than\n\t\t// initialMemory.\n\t\tmaxMemory = uint64(udc.piecesCompleted+udc.piecesRegistered) * udc.staticPieceSize\n\t}\n\t// If the chunk recovery has completed, the maximum number of pieces is the\n\t// number of registered.\n\tif udc.recoveryComplete {\n\t\tmaxMemory = uint64(udc.piecesRegistered) * udc.staticPieceSize\n\t}\n\t// Return any memory we don't need.\n\tif uint64(udc.memoryAllocated) > maxMemory {\n\t\tudc.download.memoryManager.Return(udc.memoryAllocated - maxMemory)\n\t\tudc.memoryAllocated = maxMemory\n\t}\n}", "func (this *BaseController) Finish() {\r\n}", "func (this *BaseController) Finish() {\r\n}", "func (cr *callResult) NextResultSet() error { return io.EOF }", "func (cr *callResult) Close() error { return nil }", "func (c DBStoreDoneFuncCall) Results() []interface{} {\n\treturn []interface{}{c.Result0}\n}", "func (c DBStoreDoneFuncCall) Results() []interface{} {\n\treturn []interface{}{c.Result0}\n}", "func (ret *OpRet) Done(err ...error) {\n\tif ret.delayed == nil {\n\t\treturn\n\t}\n\n\tif len(err) > 0 {\n\t\tret.error = err[0]\n\t}\n\tclose(ret.delayed)\n}", "func IMPL_FINISH(...interface{}) {\n\tpanic(\"Not yet implemented in the spec\")\n}", "func (s *BasedifListener) ExitData(ctx *DataContext) {}", "func (gp *GoPool) Done() {\n\t<-gp.buffer // read a value from the buffer\n}", "func (clt *Client) Exit() {\n\tdefer lib.Debugf(\"SQS client %d: Exit, %d records lost\", clt.ID, len(clt.batch))\n\n\tclt.done <- true\n\t<-clt.finish\n}", "func retWrite(w http.ResponseWriter, r *http.Request, res map[string]interface{}, start time.Time) {\n\tdata, err := json.Marshal(res)\n\tif err != nil {\n\t\tfmt.Printf(\"json.Marshal(\\\"%v\\\") error(%v)\\n\", res, err)\n\t\treturn\n\t}\n\tdataStr := string(data)\n\tif _, err := w.Write([]byte(dataStr)); err != nil {\n\t\tfmt.Printf(\"w.Write(\\\"%s\\\") error(%v)\\n\", dataStr, err)\n\t}\n\t//fmt.Printf(\"req: \\\"%s\\\", get: res:\\\"%s\\\", ip:\\\"%s\\\", time:\\\"%fs\\\"\\n\", r.URL.String(), dataStr, r.RemoteAddr, time.Now().Sub(start).Seconds())\n}", "func (dpos *DummyDpos) Finalize(*types.Block) error { return nil }", "func (m *Master) Done() bool {\n\n\t// Your code here.\n\treturn m.ReduceFinish\n}", "func (c *TestCase) FinishCase() {\n\tc.Finish <- true\n\t<-c.Finish\n}", "func (cfg *config) end() {\n\tcfg.checkTimeout()\n\tif cfg.t.Failed() == false {\n\t\tcfg.mu.Lock()\n\t\tt := time.Since(cfg.t0).Seconds() // real time\n\t\tnpeers := cfg.n // number of Raft peers\n\t\tnrpc := cfg.rpcTotal() - cfg.rpcs0 // number of RPC sends\n\t\tnbytes := cfg.bytesTotal() - cfg.bytes0 // number of bytes\n\t\tncmds := cfg.maxIndex - cfg.maxIndex0 // number of Raft agreements reported\n\t\tcfg.mu.Unlock()\n\n\t\tfmt.Printf(\" ... Passed --\")\n\t\tfmt.Printf(\" %4.1f %d %4d %7d %4d\\n\", t, npeers, nrpc, nbytes, ncmds)\n\t}\n}", "func (e *EventLog) Finish() {\n\te.el.Finish()\n}", "func (p *ProgressMeter) Finish() {\n\tclose(p.finished)\n\tp.update()\n\tp.logger.Close()\n\tif !p.dryRun && p.estimatedBytes > 0 {\n\t\tfmt.Fprintf(os.Stdout, \"\\n\")\n\t}\n}", "func (p *ProgressMeter) Finish() {\n\tclose(p.finished)\n\tp.update()\n\tp.logger.Close()\n\tif !p.dryRun && p.estimatedBytes > 0 {\n\t\tfmt.Fprintf(os.Stdout, \"\\n\")\n\t}\n}", "func (s *scanner) finishString(buf []byte) string {\n\tstr := *(*string)(unsafe.Pointer(&buf))\n\ts.returnBuffer(buf)\n\treturn str\n}", "func (w *Watcher) finish() {\n\tclose(w.Fork)\n\tclose(w.Exec)\n\tclose(w.Exit)\n\tclose(w.Error)\n}", "func (f *FileWriter) Finish() error {\n\tif f.Status == PAUSED {\n\t\treturn errors.New(\"File Writing is paused\")\n\t}\n\tf.DataFile.Close()\n\tf = nil\n\treturn nil\n}", "func (r *SnpDerivedKeyReqABI) Finish(BinaryConvertible) error { return nil }", "func (s *SubmissionHandler) End(e error) {\n\tdefer s.conn.Close()\n\tmsg := OK\n\tif e != nil {\n\t\tmsg = \"ERROR: \" + e.Error()\n\t\tutil.Log(e, LOG_RECEIVER)\n\t}\n\ts.write(msg)\n}", "func (r *SnpDerivedKeyRespABI) Finish(BinaryConvertible) error {\n\tswitch r.Status {\n\tcase 0:\n\t\treturn nil\n\tcase 0x16:\n\t\treturn errors.New(\"msg_key_req error: invalid parameters\")\n\tdefault:\n\t\treturn fmt.Errorf(\"msg_key_req unknown status code: 0x%x\", r.Status)\n\t}\n}", "func (wkr *worker) done(ctx context.Context) (task.Result, string) {\n\t// close\n\tfor _, rdr := range wkr.stsRdrs {\n\t\trdr.r.Close()\n\t}\n\terr := wkr.w.CloseWithContext(ctx)\n\tif err != nil {\n\t\treturn task.ErrResult, fmt.Sprint(err.Error())\n\t}\n\n\t// publish files stats\n\tallSts := wkr.w.Stats()\n\tfor _, sts := range allSts {\n\t\tif sts.Size > 0 { // only successful files\n\t\t\tproducer.Send(appOpt.FileTopic, sts.JSONBytes())\n\t\t}\n\t}\n\n\t// msg\n\tvar msg string\n\tif wkr.iOpt.Discard {\n\t\tmsg = fmt.Sprintf(\"wrote %v lines over %v files (%v discarded)\", wkr.w.LineCnt(), len(allSts), wkr.discardedCnt)\n\t} else {\n\t\tmsg = fmt.Sprintf(\"wrote %v lines over %v files\", wkr.w.LineCnt(), len(allSts))\n\t}\n\n\treturn task.CompleteResult, msg\n}", "func (a *Aggregator) Done() {\n\tclose(a.done)\n\ta.w.Wait()\n}", "func (l localOptimizer) finishMethodDone(operation chan<- Task, result <-chan Task, task Task) {\n\ttask.Op = MethodDone\n\toperation <- task\n\ttask = <-result\n\tif task.Op != PostIteration {\n\t\tpanic(\"optimize: task should have returned post iteration\")\n\t}\n\tl.finish(operation, result)\n}", "func (z *Stream) End() {\n\tC.lzma_end(z.C())\n}", "func (m *Master) Done() bool {\n\t// Your code here.\n\n\treturn m.done\n}", "func (m *Message) Finish() {\n\tif !atomic.CompareAndSwapInt32(&m.responded, 0, 1) {\n\t\treturn\n\t}\n\tm.Delegate.OnFinish(m)\n}", "func (m *Message) Finish() {\n\tif !atomic.CompareAndSwapInt32(&m.responded, 0, 1) {\n\t\treturn\n\t}\n\tm.Delegate.OnFinish(m)\n}", "func (bf *BlockIndexFilterManager) Quit() {\n\n\tif bf.ioCloseChan != nil {\n\t\tclose(bf.ioChan)\n\t\tlog.Info(\"wait for ioTraceRoutine\")\n\t\t<-bf.ioCloseChan\n\t}\n\n\tif bf.outputFilename == \"\" {\n\t\treturn\n\t}\n\n\tjsonMap := make(map[string]interface{})\n\tfor i, generationStats := range bf.StatsList {\n\t\tjsonMap[\"generation \"+strconv.Itoa(i)] = generationStats\n\t}\n\n\tif encodedStats, err := json.MarshalIndent(jsonMap, \"\", \" \"); err != nil {\n\t\tlog.Error(\"Couldn't marshal results: \", err)\n\t} else if f, err := os.Create(bf.outputFilename); err != nil {\n\t\tlog.Error(\"Couldn't create results file: \", err)\n\t} else if n, err := f.Write(encodedStats); (n != len(encodedStats)) || err != nil {\n\t\tlog.Error(\"Couldn't write to results file: \", err)\n\t} else {\n\t\tlog.Info(bytes.NewBuffer(encodedStats).String())\n\t\tf.Close()\n\t}\n}", "func (m *TesterMock) Finish() {\n\tm.MinimockFinish()\n}", "func (bar *Progress) Finish() {\n\tfmt.Println()\n}", "func (a *Agent) Finish() *Sketch {\n\ta.flush()\n\n\tif a.IsEmpty() {\n\t\treturn nil\n\t}\n\n\treturn a.Sketch.Copy()\n}", "func (call *Call) done() {\n\tdelete(call.Owner.calls, call.ID)\n\tselect {\n\tcase call.Done <- call:\n\t\t// ok\n\tdefault:\n\t\t// We don't want to block here. It is the caller's responsibility to make\n\t\t// sure the channel has enough buffer space. See comment in Go().\n\t\tcall.l.Debug(\"rpc: discarding Call reply due to insufficient Done chan capacity\")\n\t}\n}", "func (m *ParcelMock) Finish() {\n\tm.MinimockFinish()\n}", "func (data TeleportConfirmData) Complete() {\n}", "func (r *SnpReportRespABI) Finish(_ BinaryConvertible) error {\n\tif r.Status != 0 {\n\t\tswitch r.Status {\n\t\tcase 0x16: // Value from MSG_REPORT_RSP specification in SNP API.\n\t\t\treturn errors.New(\"get_report had invalid parameters\")\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"unknown status: 0x%x\", r.Status)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Client) done(cmd *Command, rsp *Response) {\n\tif cmd.result != nil {\n\t\treturn\n\t}\n\tcmd.result = rsp\n\tif tag := cmd.tag; c.cmds[tag] != nil {\n\t\tdelete(c.cmds, tag)\n\t\tif c.tags[0] == tag {\n\t\t\tc.tags = c.tags[1:]\n\t\t} else if n := len(c.tags); c.tags[n-1] == tag {\n\t\t\tc.tags = c.tags[:n-1]\n\t\t} else {\n\t\t\tfor i, v := range c.tags {\n\t\t\t\tif v == tag {\n\t\t\t\t\tc.tags = append(c.tags[:i], c.tags[i+1:]...)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tif rsp == abort {\n\t\tc.Logln(LogCmd, \"<<<\", cmd.tag, \"(Abort)\")\n\t} else {\n\t\tc.Logln(LogCmd, \"<<<\", rsp)\n\t}\n}", "func (j *Job) done() { j.isDone <- true }", "func (s *Basememcached_protocolListener) ExitEnd(ctx *EndContext) {}", "func (dc *Decompressor) Finish() error {\n\tvar err error\n\tselect {\n\tcase <-dc.ctx.Done():\n\t\terr = dc.ctx.Err()\n\tdefault:\n\t}\n\t// NOTE, that the the assemble method must read all of the output\n\t// produced by the workers, even in the event of an error. Otherwise\n\t// a deadlock will occur with the workers trying to write blocks to\n\t// the channel that the assemble method is no longer reading from.\n\tclose(dc.workCh)\n\tdc.workWg.Wait()\n\tclose(dc.doneCh)\n\tdc.doneWg.Wait()\n\treturn err\n}", "func (m *CryptographyServiceMock) Finish() {\n\tm.MinimockFinish()\n}", "func (b *Builder) FinishBlock() {\n\tb.append(y.U32SliceToBytes(b.entryOffsets))\n\tb.append(y.U32ToBytes(uint32(len(b.entryOffsets))))\n\tchecksum := utils.NewCRC(b.currentBlock.Data[:b.sz]).Value()\n\tb.append(y.U32ToBytes(checksum))\n\n\txlog.Logger.Debugf(\"real block size is %d, len of entries is %d\\n\", b.sz, len(b.entryOffsets))\n\t//truncate block to b.sz\n\tb.currentBlock.Data = b.currentBlock.Data[:b.sz]\n\tb.writeCh <- writeBlock{\n\t\tbaseKey: y.Copy(b.baseKey),\n\t\tb: b.currentBlock,\n\t}\n\treturn\n}" ]
[ "0.6376619", "0.6357808", "0.6102739", "0.60721344", "0.6030876", "0.6011511", "0.5989714", "0.59651756", "0.59304285", "0.5905428", "0.5894113", "0.58744043", "0.5866252", "0.57594055", "0.57539856", "0.57482886", "0.571417", "0.5702825", "0.5674391", "0.5595613", "0.5595613", "0.5580288", "0.5572597", "0.55687886", "0.55640996", "0.5563451", "0.5559736", "0.5545928", "0.5533751", "0.55235416", "0.55015963", "0.54977405", "0.5469346", "0.5467941", "0.5461695", "0.5456543", "0.54558724", "0.54558724", "0.5455727", "0.54467523", "0.5425058", "0.54234177", "0.54192454", "0.54105663", "0.5409943", "0.5403597", "0.54010797", "0.5394108", "0.53926504", "0.5383736", "0.53815573", "0.5380703", "0.5352168", "0.532549", "0.532192", "0.532192", "0.5314674", "0.53057206", "0.5302042", "0.5302042", "0.53006023", "0.5297771", "0.529756", "0.52953386", "0.5289709", "0.52763706", "0.527512", "0.5274543", "0.52626723", "0.5261316", "0.5255197", "0.525018", "0.525018", "0.5248704", "0.5234118", "0.52316344", "0.5230497", "0.52290654", "0.52276057", "0.5202949", "0.5196978", "0.5196108", "0.5192939", "0.5190133", "0.5190129", "0.5190129", "0.5189157", "0.5187985", "0.51872975", "0.51828045", "0.5182227", "0.51799905", "0.5178636", "0.51767546", "0.5174505", "0.5161734", "0.5161524", "0.5160131", "0.51540077", "0.51439124" ]
0.62393755
2
GetType return type of request
func (b *BaseController) GetObjectName() string { u := strings.Split(b.Ctx.Request.URL.Path, "/") if len(u) >= 2 { return u[2] } else { return "" } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (req *GetRequest) Type() iproto.Type {\n\treturn req.impl.Type()\n}", "func (f *AccessReq) GetType() string {\n\treturn f.Type\n}", "func (s *SaveApplicationLogEventRequest) GetType() (value string) {\n\tif s == nil {\n\t\treturn\n\t}\n\treturn s.Type\n}", "func (fr *fakeRequest) Type() graphsync.RequestType {\n\treturn fr.requestType\n}", "func (r GetCodeRequest) Type() RequestType {\n\treturn GetCode\n}", "func (a *AddProxyRequest) GetType() (value ProxyTypeClass) {\n\tif a == nil {\n\t\treturn\n\t}\n\treturn a.Type\n}", "func (req *SetRequest) Type() iproto.Type {\n\treturn req.impl.Type()\n}", "func (o *PaymentMethodCardRequest) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (r GetPublicKeyRequest) Type() RequestType {\n\treturn GetPublicKey\n}", "func (*GetWebAppURLRequest) TypeName() string {\n\treturn \"getWebAppUrl\"\n}", "func (g *GetWebAppURLRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getWebAppUrl\",\n\t\tID: GetWebAppURLRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"BotUserID\",\n\t\t\tSchemaName: \"bot_user_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"URL\",\n\t\t\tSchemaName: \"url\",\n\t\t},\n\t\t{\n\t\t\tName: \"Theme\",\n\t\t\tSchemaName: \"theme\",\n\t\t},\n\t\t{\n\t\t\tName: \"ApplicationName\",\n\t\t\tSchemaName: \"application_name\",\n\t\t},\n\t}\n\treturn typ\n}", "func (o *ApiResponse) GetType() string {\n\tif o == nil || IsNil(o.Type) {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (r *Request) ContentType() string {\n\treturn r.contentType\n}", "func (o *ApiResponse) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (r *request) Kind() Kind {\n\treturn Request\n}", "func (f *Predicter) GetType() string {\n\treturn f.GetTypeReactor()\n}", "func (o *ControllerServiceAPI) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (ctx *HijackRequest) Type() proto.NetworkResourceType {\n\treturn ctx.event.ResourceType\n}", "func (r GetExpiryRequest) Type() RequestType {\n\treturn GetExpiry\n}", "func (o *GroupReplaceRequest) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func getRequestEntityType(r *http.Request) string {\n\treturn r.Header.Get(X_GOME_TYPE)\n}", "func (r DeployServiceRequest) Type() RequestType {\n\treturn Deploy\n}", "func (o *Wireless) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (r ExecuteServiceRequest) Type() RequestType {\n\treturn Execute\n}", "func (*GetWebAppURLRequest) TypeID() uint32 {\n\treturn GetWebAppURLRequestTypeID\n}", "func (*GetProxyLinkRequest) TypeName() string {\n\treturn \"getProxyLink\"\n}", "func (data *RequestData) BaseType() string {\n\treturn \"RequestData\"\n}", "func (*GetStoryRequest) TypeName() string {\n\treturn \"getStory\"\n}", "func (o *CheckoutResponse) GetType() string {\n\tif o == nil || IsNil(o.Type) {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (*GetLoginURLInfoRequest) TypeName() string {\n\treturn \"getLoginUrlInfo\"\n}", "func (g *GetLogTagsRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getLogTags\",\n\t\tID: GetLogTagsRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{}\n\treturn typ\n}", "func (o *View) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (s *Service) Type() interface{} {\n\treturn (*s)[jsonldType]\n}", "func (*MessagesGetBotAppRequest) TypeName() string {\n\treturn \"messages.getBotApp\"\n}", "func (p *Predict) GetType() string {\n\treturn Type\n}", "func (p *Predict) GetType() string {\n\treturn Type\n}", "func (o *WorkflowCliCommandAllOf) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (execution *Execution) GetType() (execType string) {\n\tswitch strings.ToLower(execution.Type) {\n\tcase \"\":\n\t\tfallthrough\n\tcase \"local\":\n\t\texecType = \"local\"\n\tcase \"remote\":\n\t\texecType = \"remote\"\n\tdefault:\n\t\tpanic(execution)\n\t}\n\n\treturn\n}", "func (g *GetBackgroundsRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getBackgrounds\",\n\t\tID: GetBackgroundsRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"ForDarkTheme\",\n\t\t\tSchemaName: \"for_dark_theme\",\n\t\t},\n\t}\n\treturn typ\n}", "func (*MessagesRequestEncryptionRequest) TypeName() string {\n\treturn \"messages.requestEncryption\"\n}", "func (*SendPaymentFormRequest) TypeName() string {\n\treturn \"sendPaymentForm\"\n}", "func (g *MessagesGetBotAppRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"messages.getBotApp\",\n\t\tID: MessagesGetBotAppRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"App\",\n\t\t\tSchemaName: \"app\",\n\t\t},\n\t\t{\n\t\t\tName: \"Hash\",\n\t\t\tSchemaName: \"hash\",\n\t\t},\n\t}\n\treturn typ\n}", "func (*InvokeWithTraceRequest) TypeName() string {\n\treturn \"invokeWithTrace\"\n}", "func (c *HTTPChecker) Type() string {\n\treturn \"http\"\n}", "func (*AddProxyRequest) TypeName() string {\n\treturn \"addProxy\"\n}", "func (g *GetConnectedWebsitesRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getConnectedWebsites\",\n\t\tID: GetConnectedWebsitesRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{}\n\treturn typ\n}", "func (o *GetMessagesAllOf) GetType() interface{} {\n\tif o == nil {\n\t\tvar ret interface{}\n\t\treturn ret\n\t}\n\treturn o.Type\n}", "func (o ApiOperationRequestQueryParameterOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApiOperationRequestQueryParameter) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o *Application) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (o ApiOperationRequestRepresentationFormParameterOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApiOperationRequestRepresentationFormParameter) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o ApiOperationRequestHeaderOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApiOperationRequestHeader) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o *VersionedControllerService) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (g *GetLanguagePackStringRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getLanguagePackString\",\n\t\tID: GetLanguagePackStringRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"LanguagePackDatabasePath\",\n\t\t\tSchemaName: \"language_pack_database_path\",\n\t\t},\n\t\t{\n\t\t\tName: \"LocalizationTarget\",\n\t\t\tSchemaName: \"localization_target\",\n\t\t},\n\t\t{\n\t\t\tName: \"LanguagePackID\",\n\t\t\tSchemaName: \"language_pack_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"Key\",\n\t\t\tSchemaName: \"key\",\n\t\t},\n\t}\n\treturn typ\n}", "func (o *Service) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (r *Riddler) Type() string {\n\treturn r.SourceType\n}", "func (o *DataPlaneClusterUpdateStatusRequestConditions) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (g *GetStoryRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getStory\",\n\t\tID: GetStoryRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"StorySenderChatID\",\n\t\t\tSchemaName: \"story_sender_chat_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"StoryID\",\n\t\t\tSchemaName: \"story_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"OnlyLocal\",\n\t\t\tSchemaName: \"only_local\",\n\t\t},\n\t}\n\treturn typ\n}", "func (o *CredentialsResponseElement) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (this *Self) Type() value.Type { return value.JSON }", "func (req *DronaRequest) GetContentType() string {\n\treq.Lock()\n\tdefer req.Unlock()\n\treturn req.contentType\n}", "func (*GetLoginURLInfoRequest) TypeID() uint32 {\n\treturn GetLoginURLInfoRequestTypeID\n}", "func (msg *RegisterRMRequest) Type() uint16 {\n\treturn TypeRegRM\n}", "func (req *Request) ContentType() string {\n\tif req == nil {\n\t\treturn \"\"\n\t}\n\n\tif req._contentType == \"\" {\n\t\tcontentType := req.Request.Header.Get(\"Content-Type\")\n\n\t\tif contentType == \"\" {\n\t\t\treq._contentType = \"text/html\"\n\t\t} else {\n\t\t\treq._contentType = strings.ToLower(strings.TrimSpace(strings.Split(contentType, \";\")[0]))\n\t\t}\n\t}\n\n\treturn req._contentType\n}", "func (*MessagesRequestSimpleWebViewRequest) TypeName() string {\n\treturn \"messages.requestSimpleWebView\"\n}", "func getReqType(path string) string {\n\tif len(path) < 9 {\n\t\treturn \"None\"\n\t}\n\n\tif path[len(path)-9:] == \"info.json\" {\n\t\treturn \"Info\"\n\t}\n\n\tvar parts = strings.Split(path, \"/\")\n\tif len(parts) < 5 {\n\t\treturn \"None\"\n\t}\n\n\tvar iiifPath = strings.Join(parts[len(parts)-5:], \"/\")\n\tvar u, err = iiif.NewURL(iiifPath)\n\tif err != nil || !u.Valid() {\n\t\treturn \"None\"\n\t}\n\n\tif err == nil {\n\t\tif u.Region.Type == iiif.RTFull || u.Region.Type == iiif.RTSquare {\n\t\t\treturn \"Resize\"\n\t\t} else if u.Size.W <= 1024 && u.Size.H <= 1024 {\n\t\t\treturn \"Tile\"\n\t\t}\n\t}\n\n\treturn \"Unknown\"\n}", "func (g *GetProxyLinkRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getProxyLink\",\n\t\tID: GetProxyLinkRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"ProxyID\",\n\t\t\tSchemaName: \"proxy_id\",\n\t\t},\n\t}\n\treturn typ\n}", "func (j *Jsonnet) Type() string {\n\treturn \"jsonnet\"\n}", "func (x *fastReflection_QueryParamsRequest) Type() protoreflect.MessageType {\n\treturn _fastReflection_QueryParamsRequest_messageType\n}", "func (c CredentialService) GetType() string {\n\treturn c.Type\n}", "func (o *Ga4ghChemotherapy) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (r *ResendEmailAddressVerificationCodeRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"resendEmailAddressVerificationCode\",\n\t\tID: ResendEmailAddressVerificationCodeRequestTypeID,\n\t}\n\tif r == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{}\n\treturn typ\n}", "func (rtspService *RTSPService) GetType() string {\n\treturn wssapi.OBJRTSPServer\n}", "func (o *UiNodeInputAttributes) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (r *Request) contentType() int {\n\tif r.Method == \"HEAD\" || r.Method == \"OPTIONS\" {\n\t\treturn contentNone\n\t}\n\n\tct := r.Header.Get(\"content-type\")\n\tif strings.Contains(ct, \"application/x-www-form-urlencoded\") {\n\t\treturn contentFormData\n\t}\n\n\tif strings.Contains(ct, \"multipart/form-data\") {\n\t\treturn contentMultipart\n\t}\n\n\treturn contentStream\n}", "func (g *AccountGetWallPapersRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"account.getWallPapers\",\n\t\tID: AccountGetWallPapersRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"Hash\",\n\t\t\tSchemaName: \"hash\",\n\t\t},\n\t}\n\treturn typ\n}", "func (msg *GlobalBeginRequest) Type() uint16 {\n\treturn TypeGlobalBegin\n}", "func (msg *RegisterTMRequest) Type() uint16 {\n\treturn TypeRegClt\n}", "func (rc *RequiredCapability) GetType() string {\n\treturn \"deliveryservice.RequiredCapability\"\n}", "func (m *kubeGenericRuntimeManager) Type() string {\n\treturn m.runtimeName\n}", "func (con *Controller) GetType() string {\n\treturn con.EntryType\n}", "func (r *MessagesRequestEncryptionRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"messages.requestEncryption\",\n\t\tID: MessagesRequestEncryptionRequestTypeID,\n\t}\n\tif r == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"UserID\",\n\t\t\tSchemaName: \"user_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"RandomID\",\n\t\t\tSchemaName: \"random_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"GA\",\n\t\t\tSchemaName: \"g_a\",\n\t\t},\n\t}\n\treturn typ\n}", "func (r PollServiceRequest) Type() RequestType {\n\treturn Poll\n}", "func (*ResendEmailAddressVerificationCodeRequest) TypeName() string {\n\treturn \"resendEmailAddressVerificationCode\"\n}", "func (s *SaveApplicationLogEventRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"saveApplicationLogEvent\",\n\t\tID: SaveApplicationLogEventRequestTypeID,\n\t}\n\tif s == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"Type\",\n\t\t\tSchemaName: \"type\",\n\t\t},\n\t\t{\n\t\t\tName: \"ChatID\",\n\t\t\tSchemaName: \"chat_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"Data\",\n\t\t\tSchemaName: \"data\",\n\t\t},\n\t}\n\treturn typ\n}", "func (msg *GlobalLockQueryRequest) Type() uint16 {\n\treturn TypeGlobalLockQuery\n}", "func (o JsonSerializationResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v JsonSerializationResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (o CSharpFunctionBindingResponseOutput) Type() pulumi.StringOutput {\n\treturn o.ApplyT(func(v CSharpFunctionBindingResponse) string { return v.Type }).(pulumi.StringOutput)\n}", "func (ns NodeSolver) GetType() string {\n\treturn ns.MType\n}", "func (o *UserActionNamingPlaceholderProcessingStep) GetType() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.Type\n}", "func (g *GetLoginURLInfoRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"getLoginUrlInfo\",\n\t\tID: GetLoginURLInfoRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"ChatID\",\n\t\t\tSchemaName: \"chat_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"MessageID\",\n\t\t\tSchemaName: \"message_id\",\n\t\t},\n\t\t{\n\t\t\tName: \"ButtonID\",\n\t\t\tSchemaName: \"button_id\",\n\t\t},\n\t}\n\treturn typ\n}", "func (pp *Pingreq) Type() Type {\n\treturn PINGREQ\n}", "func (a *ChangeRequest) Type() AttributeType {\n\treturn CHANGEREQUEST\n}", "func (p *Provider) GetType() string {\n\treturn strings.ToLower(p.Type)\n}", "func (i *InvokeWithTraceRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"invokeWithTrace\",\n\t\tID: InvokeWithTraceRequestTypeID,\n\t}\n\tif i == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"TraceParent\",\n\t\t\tSchemaName: \"traceParent\",\n\t\t\tNull: !i.Flags.Has(1),\n\t\t},\n\t\t{\n\t\t\tName: \"TraceState\",\n\t\t\tSchemaName: \"traceState\",\n\t\t\tNull: !i.Flags.Has(2),\n\t\t},\n\t\t{\n\t\t\tName: \"Baggage\",\n\t\t\tSchemaName: \"baggage\",\n\t\t\tNull: !i.Flags.Has(3),\n\t\t},\n\t\t{\n\t\t\tName: \"Query\",\n\t\t\tSchemaName: \"query\",\n\t\t},\n\t}\n\treturn typ\n}", "func (o *PaymentMethodCardRequest) GetTypeOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.Type, true\n}", "func typeOf(obj interface{}) string { return reflect.TypeOf(obj).Name() }", "func (_CraftingI *CraftingICallerSession) GetType(_type_id *big.Int) (string, error) {\n\treturn _CraftingI.Contract.GetType(&_CraftingI.CallOpts, _type_id)\n}", "func (g *UsersGetFullUserRequest) TypeInfo() tdp.Type {\n\ttyp := tdp.Type{\n\t\tName: \"users.getFullUser\",\n\t\tID: UsersGetFullUserRequestTypeID,\n\t}\n\tif g == nil {\n\t\ttyp.Null = true\n\t\treturn typ\n\t}\n\ttyp.Fields = []tdp.Field{\n\t\t{\n\t\t\tName: \"ID\",\n\t\t\tSchemaName: \"id\",\n\t\t},\n\t}\n\treturn typ\n}", "func (m *Message) GetType(simpleMode bool) int {\n\tif simpleMode {\n\t\treturn SIMPLE_MESSAGE\n\t} else if m.File != nil && m.Destination == nil && len(*m.Request) == 0 {\n\t\treturn FILE_INDEXING\n\t} else if m.File != nil && m.Request != nil {\n\t\treturn DATA_REQUEST\n\t} else if m.KeyWords != nil {\n\t\treturn SEARCH_REQUEST\n\t} else if m.MasterKey != nil && m.NewPassword != nil {\n\t\treturn PASSWORD_INSERT\n\t} else if m.MasterKey != nil && m.NewPassword == nil && m.UserName != nil {\n\t\treturn PASSWORD_RETRIEVE\n\t} else if m.MasterKey != nil && m.NewPassword == nil && m.DeleteUser != nil {\n\t\treturn PASSWORD_DELETE\n\t} else if m.Destination != nil {\n\t\treturn PRIVATE_MESSAGE\n\t} else {\n\t\treturn RUMOUR_MESSAGE\n\t}\n}", "func (o *KubernetesEthernetMatcher) GetType() string {\n\tif o == nil || o.Type == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Type\n}", "func (e WasCreated) GetType() string {\n\treturn fmt.Sprintf(\"%T\", e)\n}" ]
[ "0.75369483", "0.74166656", "0.7192883", "0.71340615", "0.70874155", "0.70625603", "0.69063324", "0.6881202", "0.66364485", "0.6562451", "0.65017396", "0.64967084", "0.64939296", "0.64904207", "0.64837843", "0.6481509", "0.644261", "0.64332646", "0.639813", "0.63534534", "0.6337974", "0.62902004", "0.6289433", "0.6282866", "0.628207", "0.6274976", "0.6263763", "0.62175745", "0.6194873", "0.61842597", "0.6171819", "0.6171735", "0.6164019", "0.61483335", "0.61450744", "0.61450744", "0.612952", "0.61178184", "0.6117174", "0.6111678", "0.6105939", "0.6100119", "0.6096842", "0.609195", "0.60868704", "0.60761946", "0.60691726", "0.6058747", "0.60489005", "0.60469246", "0.60456944", "0.6024615", "0.60178936", "0.60151243", "0.60136944", "0.60055435", "0.6004504", "0.59931856", "0.59840447", "0.5978819", "0.59699243", "0.5950949", "0.59507793", "0.59490275", "0.59481215", "0.5943963", "0.5930276", "0.5927764", "0.59272933", "0.592293", "0.5920184", "0.5915358", "0.5914014", "0.59132546", "0.5912463", "0.5911322", "0.59103966", "0.59008396", "0.58976793", "0.58973134", "0.589415", "0.58833766", "0.5882001", "0.58811194", "0.5880829", "0.5864702", "0.5862606", "0.58614683", "0.5859305", "0.5856246", "0.5855434", "0.58513147", "0.5837165", "0.58332485", "0.583213", "0.5819815", "0.58173555", "0.58173054", "0.5808653", "0.58070266", "0.58040595" ]
0.0
-1
GetAttributes get object's Attribute
func GetReflectType(v reflect.Value) reflect.Type { for i := 0; i < 10; i++ { if v.Kind() == reflect.Ptr { v = v.Elem() } else { return v.Type() } } return v.Type() }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *ObjectData) GetAttr(name string) (value Object, present bool) {\n value, present = o.Attrs[name]\n return \n}", "func (e *FakeElement) GetAttribute(attr string) Object {\n\treturn MakeFakeObject(e.Attributes[attr])\n}", "func (c EntityObject) Attributes() map[string]interface{} {\n\treturn c.attributes\n}", "func (a Attributes) Get(key interface{}) interface{} {\n\treturn a[key]\n}", "func (o *Giveaway) GetAttributes() map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func (b *Bucket) Attributes(_ context.Context, name string) (objstore.ObjectAttributes, error) {\n\tobjMeta, err := b.client.GetObjectMeta(b.name, name)\n\tif err != nil {\n\t\treturn objstore.ObjectAttributes{}, errors.Wrapf(err, \"gettting objectmeta of %s\", name)\n\t}\n\n\tlastModified, err := time.Parse(time.RFC1123, objMeta.LastModified)\n\tif err != nil {\n\t\treturn objstore.ObjectAttributes{}, err\n\t}\n\n\treturn objstore.ObjectAttributes{\n\t\tSize: objMeta.ContentLength,\n\t\tLastModified: lastModified,\n\t}, nil\n}", "func (o *Object) Attribute(name string) *AttributeExpr {\n\tfor _, nat := range *o {\n\t\tif nat.Name == name {\n\t\t\treturn nat.Attribute\n\t\t}\n\t}\n\treturn nil\n}", "func (obj *Edge) GetAttributes() ([]types.TGAttribute, types.TGError) {\n\treturn obj.getAttributes()\n}", "func (attestedClaim *AttestedClaim) getAttributes() ([]*Attribute, error) {\n\tbInts := attestedClaim.getRawAttributes()\n\tattributes, err := BigIntsToAttributes(bInts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsorted := sort.SliceIsSorted(attributes, func(p, q int) bool {\n\t\treturn strings.Compare(attributes[p].Name, attributes[q].Name) < 0\n\t})\n\tif !sorted {\n\t\treturn nil, errors.New(\"expected attributes inside credential to be sorted\")\n\t}\n\treturn attributes, nil\n}", "func (m *Device) GetAttributes() (val map[string]interface{}, set bool) {\n\tif m.Attributes == nil {\n\t\treturn\n\t}\n\n\treturn *m.Attributes, true\n}", "func (c EntityObject) Attribute(key string) (interface{}, error) {\n\tif key == \"\" {\n\t\treturn nil, errors.New(\"Attribute key must be non-empty\")\n\t}\n\n\tif c.attributes == nil {\n\t\treturn nil, nil\n\t}\n\n\tvalue, found := c.attributes[key]\n\tif !found {\n\t\treturn nil, nil\n\t}\n\n\treturn value, nil\n}", "func (b *Bucket) Attributes(ctx context.Context, name string) (objstore.ObjectAttributes, error) {\n\tattrs, err := b.bkt.Object(name).Attrs(ctx)\n\tif err != nil {\n\t\treturn objstore.ObjectAttributes{}, err\n\t}\n\n\treturn objstore.ObjectAttributes{\n\t\tSize: attrs.Size,\n\t\tLastModified: attrs.Updated,\n\t}, nil\n}", "func (c CredentialService) GetAttributes() string {\n\treturn c.Attributes\n}", "func (o *Service) GetAttributes() map[string]string {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]string\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func (mouse *Mouse) GetAttr(attr string) (interface{}, error) {\n\treturn 0, nil\n}", "func (O *Object) GetAttribute(data *Data, name string) error {\n\tif O == nil || O.dpiObject == nil {\n\t\tpanic(\"nil dpiObject\")\n\t}\n\tattr, ok := O.Attributes[name]\n\tif !ok {\n\t\treturn errors.Errorf(\"%s: %w\", name, ErrNoSuchKey)\n\t}\n\n\tdata.reset()\n\tdata.NativeTypeNum = attr.NativeTypeNum\n\tdata.ObjectType = attr.ObjectType\n\tdata.implicitObj = true\n\t// the maximum length of that buffer must be supplied\n\t// in the value.asBytes.length attribute before calling this function.\n\tif attr.NativeTypeNum == C.DPI_NATIVE_TYPE_BYTES && attr.OracleTypeNum == C.DPI_ORACLE_TYPE_NUMBER {\n\t\tvar a [39]byte\n\t\tC.dpiData_setBytes(&data.dpiData, (*C.char)(unsafe.Pointer(&a[0])), C.uint32_t(len(a)))\n\t}\n\n\t//fmt.Printf(\"getAttributeValue(%p, %p, %d, %+v)\\n\", O.dpiObject, attr.dpiObjectAttr, data.NativeTypeNum, data.dpiData)\n\tif C.dpiObject_getAttributeValue(O.dpiObject, attr.dpiObjectAttr, data.NativeTypeNum, &data.dpiData) == C.DPI_FAILURE {\n\t\treturn errors.Errorf(\"getAttributeValue(%q, obj=%+v, attr=%+v, typ=%d): %w\", name, O, attr.dpiObjectAttr, data.NativeTypeNum, O.getError())\n\t}\n\t//fmt.Printf(\"getAttributeValue(%p, %q=%p, %d, %+v)\\n\", O.dpiObject, attr.Name, attr.dpiObjectAttr, data.NativeTypeNum, data.dpiData)\n\treturn nil\n}", "func (o *LDAPIdentityProvider) GetAttributes() (value *LDAPAttributes, ok bool) {\n\tok = o != nil && o.bitmap_&4 != 0\n\tif ok {\n\t\tvalue = o.attributes\n\t}\n\treturn\n}", "func (w *wrapper) Getattr(path string, stat *fuse.Stat_t, fd uint64) int {\n\tfi, err := w.underlying.Stat(path)\n\tif err != nil {\n\t\treturn convertError(err)\n\t}\n\tfileInfoToStat(fi, stat)\n\treturn 0\n}", "func GetAttributes(t reflect.Type) (attrs []AttributeDesc) {\n\tif t == nil {\n\t\treturn\n\t}\n\tfor i := 0; i < t.NumField(); i++ {\n\t\tdfield := t.Field(i)\n\t\tattr := AttributeDesc{\n\t\t\tName: dfield.Name,\n\t\t\tIndex: i + 1,\n\t\t\tDesc: dfield.Tag.Get(\"description\"),\n\t\t\tKind: dfield.Tag.Get(\"kind\"),\n\t\t\tRelation: dfield.Tag.Get(\"ref\"),\n\t\t}\n\t\tattrs = append(attrs, attr)\n\t}\n\treturn\n}", "func (c *jsiiProxy_CfnFilter) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (a *ManagementApiService) GetAttributes(ctx _context.Context) apiGetAttributesRequest {\n\treturn apiGetAttributesRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (c *jsiiProxy_CfnLayer) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (graph *Graph) GetAttr(x, y int) byte {\n\treturn graph.Tiles[y][x].Attr\n}", "func (repo *Repository) GetAttr(path, name string, flags ...AttrFlag) (string, error) {\n\tvar value [1]int8\n\tcvalue := (*C.char)(&value[0])\n\tcpath := C.CString(path)\n\tdefer C.free(unsafe.Pointer(cpath))\n\tcname := C.CString(name)\n\tdefer C.free(unsafe.Pointer(cname))\n\n\tvar cflags C.uint32_t\n\tfor _, flag := range flags {\n\t\tcflags |= C.uint32_t(flag)\n\t}\n\n\tecode := C.git_attr_get(&cvalue, repo.git_repository, cflags, cpath, cname)\n\tif ecode != git_SUCCESS {\n\t\treturn \"\", gitError()\n\t}\n\treturn C.GoString(cvalue), nil\n}", "func (o *LogContent) GetAttributes() map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret\n\t}\n\treturn o.Attributes\n}", "func (c *jsiiProxy_CfnInstance) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (e *Environment) Attr(environmentName, attr string) ([]Attr, error) {\n\n\targkeys := []string{\"attr\"}\n\targvalues := []interface{}{attr}\n\tbaseCommand := fmt.Sprintf(\"list environment attr %s\", environmentName)\n\n\tc, err := cmd.ArgsExpander(baseCommand, argkeys, argvalues)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tb, err := cmd.RunCommand(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tattrs := []Attr{}\n\terr = json.Unmarshal(b, &attrs)\n\tif err != nil {\n\t\t// it may have been just an empty output from the Frontend\n\t\tnullOutput := NullOutput{}\n\t\terr = json.Unmarshal(b, &nullOutput)\n\t\tif err != nil {\n\t\t\t// if we still can't recognize the output, return an error\n\t\t\treturn nil, err\n\t\t}\n\t\treturn attrs, err\n\t}\n\treturn attrs, err\n}", "func (r *Thing) Attributes() pulumi.MapOutput {\n\treturn (pulumi.MapOutput)(r.s.State[\"attributes\"])\n}", "func Attr(attrs ...a.Attribute) []a.Attribute {\n return attrs\n}", "func (o *NewCoupons) GetAttributes() map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func GetCustomAttribute(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *CustomAttributeState, opts ...pulumi.ResourceOption) (*CustomAttribute, error) {\n\tvar resource CustomAttribute\n\terr := ctx.ReadResource(\"vsphere:index/customAttribute:CustomAttribute\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (obj *Edge) GetAttribute(attrName string) types.TGAttribute {\n\treturn obj.getAttribute(attrName)\n}", "func (obj *SObject) AttributesField() *SObjectAttributes {\n\tattributes := obj.InterfaceField(sobjectAttributesKey)\n\n\tswitch attributes.(type) {\n\tcase SObjectAttributes:\n\t\t// Use a temporary variable to copy the value of attributes and return the address of the temp value.\n\t\tattrs := (attributes).(SObjectAttributes)\n\t\treturn &attrs\n\tcase map[string]interface{}:\n\t\t// Can't convert attributes to concrete type; decode interface.\n\t\tmapper := attributes.(map[string]interface{})\n\t\tattrs := &SObjectAttributes{}\n\t\tif mapper[\"type\"] != nil {\n\t\t\tattrs.Type = mapper[\"type\"].(string)\n\t\t}\n\t\tif mapper[\"url\"] != nil {\n\t\t\tattrs.URL = mapper[\"url\"].(string)\n\t\t}\n\t\treturn attrs\n\tdefault:\n\t\treturn nil\n\t}\n}", "func (f *File) getObjectAttrs() (*storage.ObjectAttrs, error) {\n\thandle, err := f.getObjectHandle()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn handle.Attrs(f.fileSystem.ctx)\n}", "func (m *IdentityUserFlowAttributeAssignment) GetUserAttribute()(IdentityUserFlowAttributeable) {\n return m.userAttribute\n}", "func (c *jsiiProxy_CfnStack) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnStack) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (fsys *FS) Getattr(path string, stat *fuse.Stat_t, fh uint64) (errc int) {\n\tdefer fs.Trace(path, \"fh=0x%X\", fh)(\"errc=%v\", &errc)\n\tnode, errc := fsys.getNode(path, fh)\n\tif errc == 0 {\n\t\terrc = fsys.stat(node, stat)\n\t}\n\treturn\n}", "func (a *ManagementApiService) GetAttribute(ctx _context.Context, attributeId int32) apiGetAttributeRequest {\n\treturn apiGetAttributeRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t\tattributeId: attributeId,\n\t}\n}", "func (elem *remoteWE) GetAttribute(name string) (string, error) {\n\ttemplate := \"/session/%%s/element/%s/attribute/%s\"\n\turlTemplate := fmt.Sprintf(template, elem.id, name)\n\n\treturn elem.parent.stringCommand(urlTemplate)\n}", "func (it isType) GetAttr(name string) Value {\n\treturn NilValue\n}", "func (c *jsiiProxy_CfnDetector) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (o *IamLdapBasePropertiesAllOf) GetAttribute() string {\n\tif o == nil || o.Attribute == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Attribute\n}", "func (o *IamLdapBasePropertiesAllOf) GetAttributeOk() (*string, bool) {\n\tif o == nil || o.Attribute == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Attribute, true\n}", "func (c *jsiiProxy_CfnDetectorModel) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnCustomResource) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnStudio) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (t *Transaction) GetAttributes(typ AttrType) []Attribute {\n\tvar result []Attribute\n\tfor _, attr := range t.Attributes {\n\t\tif attr.Type == typ {\n\t\t\tresult = append(result, attr)\n\t\t}\n\t}\n\treturn result\n}", "func (ac *AttrCache) GetAttr(options internal.GetAttrOptions) (*internal.ObjAttr, error) {\n\tlog.Trace(\"AttrCache::GetAttr : %s\", options.Name)\n\ttruncatedPath := internal.TruncateDirName(options.Name)\n\n\tac.cacheLock.RLock()\n\tvalue, found := ac.cacheMap[truncatedPath]\n\tac.cacheLock.RUnlock()\n\n\t// Try to serve the request from the attribute cache\n\tif found && value.valid() && time.Since(value.cachedAt).Seconds() < float64(ac.cacheTimeout) {\n\t\tif value.isDeleted() {\n\t\t\tlog.Debug(\"AttrCache::GetAttr : %s served from cache\", options.Name)\n\t\t\t// no entry if path does not exist\n\t\t\treturn &internal.ObjAttr{}, syscall.ENOENT\n\t\t} else {\n\t\t\t// IsMetadataRetrieved is false in the case of ADLS List since the API does not support metadata.\n\t\t\t// Once migration of ADLS list to blob endpoint is done (in future service versions), we can remove this.\n\t\t\t// options.RetrieveMetadata is set by CopyFromFile and WriteFile which need metadata to ensure it is preserved.\n\t\t\tif value.getAttr().IsMetadataRetrieved() || (ac.noSymlinks && !options.RetrieveMetadata) {\n\t\t\t\t// path exists and we have all the metadata required or we do not care about metadata\n\t\t\t\tlog.Debug(\"AttrCache::GetAttr : %s served from cache\", options.Name)\n\t\t\t\treturn value.getAttr(), nil\n\t\t\t}\n\t\t}\n\t}\n\n\t// Get the attributes from next component and cache them\n\tpathAttr, err := ac.NextComponent().GetAttr(options)\n\n\tac.cacheLock.Lock()\n\tdefer ac.cacheLock.Unlock()\n\n\tif err == nil {\n\t\t// Retrieved attributes so cache them\n\t\tif len(ac.cacheMap) < maxTotalFiles {\n\t\t\tac.cacheMap[truncatedPath] = newAttrCacheItem(pathAttr, true, time.Now())\n\t\t}\n\t} else if err == syscall.ENOENT {\n\t\t// Path does not exist so cache a no-entry item\n\t\tac.cacheMap[truncatedPath] = newAttrCacheItem(&internal.ObjAttr{}, false, time.Now())\n\t}\n\n\treturn pathAttr, err\n}", "func (attestedClaim *AttestedClaim) getRawAttributes() []*big.Int {\n\treturn attestedClaim.Credential.Attributes[1:]\n}", "func (a *AliyunInstanceAttribute) fetchAttributes(targetReader TargetReader, nodeName string) {\n\ta.ShootName = GetFromTargetInfo(targetReader, \"shootTechnicalID\")\n\tvar err error\n\tif a.FlagProviderID != \"\" {\n\t\ta.InstanceID = a.FlagProviderID\n\t} else {\n\t\ta.InstanceID, err = fetchAlicloudInstanceIDByNodeName(nodeName)\n\t\tcheckError(err)\n\t}\n\n\tres, err := ExecCmdReturnOutput(\"bash\", \"-c\", \"aliyun ecs DescribeInstanceAttribute --InstanceId=\"+a.InstanceID)\n\tcheckError(err)\n\tdecodedQuery := decodeAndQueryFromJSONString(res)\n\n\ta.RegionID, err = decodedQuery.String(\"RegionId\")\n\tcheckError(err)\n\ta.ZoneID, err = decodedQuery.String(\"ZoneId\")\n\tcheckError(err)\n\ta.VpcID, err = decodedQuery.String(\"VpcAttributes\", \"VpcId\")\n\tcheckError(err)\n\ta.VSwitchID, err = decodedQuery.String(\"VpcAttributes\", \"VSwitchId\")\n\tcheckError(err)\n\ta.ImageID, err = decodedQuery.String(\"ImageId\")\n\tcheckError(err)\n\tips, err := decodedQuery.ArrayOfStrings(\"VpcAttributes\", \"PrivateIpAddress\", \"IpAddress\")\n\tcheckError(err)\n\ta.PrivateIP = ips[0]\n\ta.BastionSecurityGroupName = a.ShootName + \"-bsg\"\n\ta.BastionInstanceName = a.ShootName + \"-bastion\"\n\n\ta.InstanceChargeType = \"PostPaid\"\n\ta.InternetChargeType = \"PayByTraffic\"\n\ta.InternetMaxBandwidthIn = \"10\"\n\ta.InternetMaxBandwidthOut = \"100\"\n\ta.IoOptimized = \"optimized\"\n\ta.InstanceType = a.getMinimumInstanceSpec()\n}", "func (c *jsiiProxy_CfnConnection) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnResourceVersion) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (o *object) Attr(i uint8) bool {\n\treturn o.Attributes[i/8]&(1<<(7-i%8)) != 0\n}", "func (o *LDAPIdentityProvider) Attributes() *LDAPAttributes {\n\tif o != nil && o.bitmap_&4 != 0 {\n\t\treturn o.attributes\n\t}\n\treturn nil\n}", "func (self Channel) GetAttribute(attrib int) (float64, error) {\n\tvar cvalue C.float\n\tresult:=C.BASS_ChannelGetAttribute(self.cint(), C.DWORD(attrib), &cvalue)\n\treturn float64(cvalue), boolToError(result)\n}", "func (c *jsiiProxy_CfnMember) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func GetAttrs(obj runtime.Object) (labels.Set, fields.Set, error) {\n\tsecret, ok := obj.(*api.Secret)\n\tif !ok {\n\t\treturn nil, nil, fmt.Errorf(\"not a secret\")\n\t}\n\treturn labels.Set(secret.Labels), SelectableFields(secret), nil\n}", "func (a *Client) GetAttributes(params *GetAttributesParams) (*GetAttributesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetAttributesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"get_attributes\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/rest/v1/attributes\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &GetAttributesReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetAttributesOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for get_attributes: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (o *NodeUpdate) GetAttributes() []map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret []map[string]interface{}\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func (a adapter) Attrs(key string) []string {\n\treturn a.entry.GetAttributeValues(key)\n}", "func (c *jsiiProxy_CfnStudioComponent) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (g *gcs) Attributes(ctx context.Context, remotePath string) (attributes interface{}, err error) {\n\tfile := g.bucket.Object(remotePath)\n\tattributes, err = file.Attrs(g.context)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (i *Index) Attr(name string) (starlark.Value, error) {\n\tswitch name {\n\tcase \"name\":\n\t\treturn starlark.String(i.name), nil\n\tcase \"str\":\n\t\treturn &stringMethods{subject: i}, nil\n\t}\n\treturn nil, starlark.NoSuchAttrError(name)\n}", "func (c *jsiiProxy_CfnClassifier) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnMacro) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnInput) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnRepository) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnMLTransform) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (a adapter) Attr(key string) string {\n\treturn a.entry.GetAttributeValue(key)\n}", "func (e *EntryBase) Attributes() *EntryAttributes {\n\treturn &e.attr\n}", "func (a *Attribute) GetName() string {\n\treturn a.Name\n}", "func (c *jsiiProxy_CfnDatabase) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnEnvironment) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (o *Service) GetAttributesOk() (map[string]string, bool) {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]string\n\t\treturn ret, false\n\t}\n\treturn *o.Attributes, true\n}", "func (h *Handler) getDeveloperAttributes(c *gin.Context) handlerResponse {\n\n\tdeveloper, err := h.service.Developer.Get(c.Param(developerParameter))\n\tif err != nil {\n\t\treturn handleError(err)\n\t}\n\treturn handleOKAttributes(developer.Attributes)\n}", "func (o *Giveaway) GetAttributesOk() (map[string]interface{}, bool) {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret, false\n\t}\n\treturn *o.Attributes, true\n}", "func (o *ObjectNode) getObjectXAttr(w http.ResponseWriter, r *http.Request) {\n\t// TODO: implement 'getObjectXAttr'\n}", "func (c *jsiiProxy_CfnFileSystem) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnStreamingImage) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnDeliveryStream) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnApplication) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnApplication) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *DOM) GetAttributesWithParams(v *DOMGetAttributesParams) ([]string, error) {\n\tresp, err := gcdmessage.SendCustomReturn(c.target, c.target.GetSendCh(), &gcdmessage.ParamRequest{Id: c.target.GetId(), Method: \"DOM.getAttributes\", Params: v})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar chromeData struct {\n\t\tResult struct {\n\t\t\tAttributes []string\n\t\t}\n\t}\n\n\tif resp == nil {\n\t\treturn nil, &gcdmessage.ChromeEmptyResponseErr{}\n\t}\n\n\t// test if error first\n\tcerr := &gcdmessage.ChromeErrorResponse{}\n\tjson.Unmarshal(resp.Data, cerr)\n\tif cerr != nil && cerr.Error != nil {\n\t\treturn nil, &gcdmessage.ChromeRequestErr{Resp: cerr}\n\t}\n\n\tif err := json.Unmarshal(resp.Data, &chromeData); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn chromeData.Result.Attributes, nil\n}", "func (dir HgmDir) Attr(ctx context.Context, a *fuse.Attr) error {\n\tresp, err := httpClient.Get(dir.getStatEndpoint(dir.localDir, false))\n\tif err != nil {\n\t\treturn fuse.EIO\n\t}\n\n\tfuseErr := stattool.HttpStatusToFuseErr(resp.StatusCode)\n\tif fuseErr == nil {\n\t\tbodyBytes, err := ioutil.ReadAll(resp.Body)\n\t\tif err == nil {\n\t\t\tattr := stattool.HgmStatAttr{}\n\t\t\terr = json.Unmarshal(bodyBytes, &attr)\n\t\t\tif err == nil {\n\t\t\t\tstattool.AttrFromHgmStat(attr, a)\n\t\t\t}\n\t\t}\n\t\tif err != nil {\n\t\t\tfuseErr = fuse.EIO\n\t\t}\n\t}\n\n\treturn fuseErr\n}", "func (c *jsiiProxy_CfnRegistry) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func getAttribute(name string,attrs []xml.Attr) (string) {\n\tval := \"\"\n\tfor _,attr := range attrs { if strings.EqualFold(attr.Name.Local,name) {val = attr.Value } }\n\treturn val\n}", "func (c *jsiiProxy_CfnPublicRepository) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (repo *Repository) GetAttributes(id string) (attributes []Attribute, err error) {\n\trows, err := repo.db.Query(\"SELECT attribute_id FROM product_attributes WHERE product_id LIKE $1\", id)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer func() { _ = rows.Close() }()\n\tvar attributeIds []int\n\tvar attr int\n\tfor rows.Next() {\n\t\terr = rows.Scan(&attr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tattributeIds = append(attributeIds, attr)\n\t}\n\n\tattributes = make([]Attribute, len(attributeIds))\n\tfor i := range attributeIds {\n\t\tattributes[i], err = repo.attributeName(attributeIds[i])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn attributes, rows.Err()\n}", "func (eClass *eClassImpl) GetEAttributes() EList {\n\teClass.getInitializers().initEAttributes()\n\treturn eClass.eAttributes\n}", "func (c *jsiiProxy_CfnProfilingGroup) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnModuleVersion) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnSecurityConfiguration) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnWorkflow) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnPreset) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnVolume) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func getPiAttribute(w http.ResponseWriter, r *http.Request) {\n\t// Get pi name and property/attribute from request\n\tvars := mux.Vars(r)\n\tpiname := vars[\"piname\"]\n\tpiproperty := vars[\"piattribute\"]\n\n\t// Get pi entry from data store\n\tc := appengine.NewContext(r)\n\tq := datastore.NewQuery(piListKind).Filter(\"name =\", piname)\n\tt := q.Run(c)\n\tvar pi Pi\n\t_, err := t.Next(&pi)\n\tif err == datastore.Done {\n\t\thttp.Error(w, \"404 Not found\", http.StatusNotFound)\n\t\treturn\n\t}\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t// Print attribute value in plain text\n\tw.Header().Set(\"Content-Type\", \"text/plain\")\n\tswitch piproperty {\n\tcase \"name\":\n\t\tfmt.Fprint(w, pi.Name)\n\tcase \"ip\":\n\t\tfmt.Fprint(w, pi.Ip)\n\tcase \"lastSeen\":\n\t\tfmt.Fprint(w, pi.LastSeen)\n\tcase \"pingCount\":\n\t\tfmt.Fprint(w, pi.PingCount)\n\tdefault:\n\t\thttp.Error(w, \"404 Not found\", http.StatusNotFound)\n\t}\n}", "func (n *Use) Attributes() map[string]interface{} {\n\treturn nil\n}", "func (obj *BlobAttribute) GetAttributeDescriptor() types.TGAttributeDescriptor {\n\treturn obj.getAttributeDescriptor()\n}", "func (d *cudaDevice) GetAttributes() (map[string]interface{}, error) {\n\treturn nil, fmt.Errorf(\"GetAttributes is not supported for CUDA devices\")\n}", "func (bo BinaryOperator) Attributes() []string {\n\tset := make(map[string]struct{})\n\tfor _, x := range bo.Left.Attributes() {\n\t\tset[x] = struct{}{}\n\t}\n\tfor _, x := range bo.Right.Attributes() {\n\t\tset[x] = struct{}{}\n\t}\n\tattrs := make([]string, 0, len(set))\n\tfor x := range set {\n\t\tattrs = append(attrs, x)\n\t}\n\treturn attrs\n}" ]
[ "0.7258507", "0.6508018", "0.6349299", "0.6330042", "0.63010705", "0.6297056", "0.62293595", "0.6208222", "0.6193886", "0.6189464", "0.6178482", "0.61777824", "0.6163363", "0.6150132", "0.61387676", "0.61117226", "0.60997057", "0.6091487", "0.60762525", "0.60761094", "0.60756046", "0.6065097", "0.60635644", "0.6063549", "0.6058266", "0.60493886", "0.6046884", "0.60419756", "0.6020591", "0.60147256", "0.5999066", "0.5996933", "0.59855324", "0.59753865", "0.59596235", "0.5945901", "0.5945901", "0.5936748", "0.5926221", "0.59108084", "0.5909558", "0.58552235", "0.5850155", "0.5849978", "0.5834076", "0.5816888", "0.5810684", "0.5810122", "0.58062667", "0.58028805", "0.57971567", "0.5779583", "0.5764129", "0.57582283", "0.5755919", "0.574963", "0.5747703", "0.57471675", "0.5740872", "0.5739216", "0.57374865", "0.5732717", "0.5727843", "0.57217765", "0.5721542", "0.57164365", "0.5710569", "0.57085043", "0.5701868", "0.5700806", "0.5693142", "0.56840515", "0.5683008", "0.5681846", "0.567936", "0.56780845", "0.56657696", "0.5660581", "0.5655827", "0.5649793", "0.56488335", "0.56447744", "0.56447744", "0.56414974", "0.56378824", "0.56365347", "0.56277174", "0.56245774", "0.5615122", "0.5611439", "0.5611109", "0.5605331", "0.5602537", "0.56000984", "0.55964714", "0.5594738", "0.55916214", "0.55882907", "0.5584701", "0.5580629", "0.5576676" ]
0.0
-1
GetAttributes get object's Attribute
func GetAttributes(t reflect.Type) (attrs []AttributeDesc) { if t == nil { return } for i := 0; i < t.NumField(); i++ { dfield := t.Field(i) attr := AttributeDesc{ Name: dfield.Name, Index: i + 1, Desc: dfield.Tag.Get("description"), Kind: dfield.Tag.Get("kind"), Relation: dfield.Tag.Get("ref"), } attrs = append(attrs, attr) } return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o *ObjectData) GetAttr(name string) (value Object, present bool) {\n value, present = o.Attrs[name]\n return \n}", "func (e *FakeElement) GetAttribute(attr string) Object {\n\treturn MakeFakeObject(e.Attributes[attr])\n}", "func (c EntityObject) Attributes() map[string]interface{} {\n\treturn c.attributes\n}", "func (a Attributes) Get(key interface{}) interface{} {\n\treturn a[key]\n}", "func (o *Giveaway) GetAttributes() map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func (b *Bucket) Attributes(_ context.Context, name string) (objstore.ObjectAttributes, error) {\n\tobjMeta, err := b.client.GetObjectMeta(b.name, name)\n\tif err != nil {\n\t\treturn objstore.ObjectAttributes{}, errors.Wrapf(err, \"gettting objectmeta of %s\", name)\n\t}\n\n\tlastModified, err := time.Parse(time.RFC1123, objMeta.LastModified)\n\tif err != nil {\n\t\treturn objstore.ObjectAttributes{}, err\n\t}\n\n\treturn objstore.ObjectAttributes{\n\t\tSize: objMeta.ContentLength,\n\t\tLastModified: lastModified,\n\t}, nil\n}", "func (o *Object) Attribute(name string) *AttributeExpr {\n\tfor _, nat := range *o {\n\t\tif nat.Name == name {\n\t\t\treturn nat.Attribute\n\t\t}\n\t}\n\treturn nil\n}", "func (obj *Edge) GetAttributes() ([]types.TGAttribute, types.TGError) {\n\treturn obj.getAttributes()\n}", "func (attestedClaim *AttestedClaim) getAttributes() ([]*Attribute, error) {\n\tbInts := attestedClaim.getRawAttributes()\n\tattributes, err := BigIntsToAttributes(bInts)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsorted := sort.SliceIsSorted(attributes, func(p, q int) bool {\n\t\treturn strings.Compare(attributes[p].Name, attributes[q].Name) < 0\n\t})\n\tif !sorted {\n\t\treturn nil, errors.New(\"expected attributes inside credential to be sorted\")\n\t}\n\treturn attributes, nil\n}", "func (m *Device) GetAttributes() (val map[string]interface{}, set bool) {\n\tif m.Attributes == nil {\n\t\treturn\n\t}\n\n\treturn *m.Attributes, true\n}", "func (c EntityObject) Attribute(key string) (interface{}, error) {\n\tif key == \"\" {\n\t\treturn nil, errors.New(\"Attribute key must be non-empty\")\n\t}\n\n\tif c.attributes == nil {\n\t\treturn nil, nil\n\t}\n\n\tvalue, found := c.attributes[key]\n\tif !found {\n\t\treturn nil, nil\n\t}\n\n\treturn value, nil\n}", "func (b *Bucket) Attributes(ctx context.Context, name string) (objstore.ObjectAttributes, error) {\n\tattrs, err := b.bkt.Object(name).Attrs(ctx)\n\tif err != nil {\n\t\treturn objstore.ObjectAttributes{}, err\n\t}\n\n\treturn objstore.ObjectAttributes{\n\t\tSize: attrs.Size,\n\t\tLastModified: attrs.Updated,\n\t}, nil\n}", "func (c CredentialService) GetAttributes() string {\n\treturn c.Attributes\n}", "func (o *Service) GetAttributes() map[string]string {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]string\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func (mouse *Mouse) GetAttr(attr string) (interface{}, error) {\n\treturn 0, nil\n}", "func (O *Object) GetAttribute(data *Data, name string) error {\n\tif O == nil || O.dpiObject == nil {\n\t\tpanic(\"nil dpiObject\")\n\t}\n\tattr, ok := O.Attributes[name]\n\tif !ok {\n\t\treturn errors.Errorf(\"%s: %w\", name, ErrNoSuchKey)\n\t}\n\n\tdata.reset()\n\tdata.NativeTypeNum = attr.NativeTypeNum\n\tdata.ObjectType = attr.ObjectType\n\tdata.implicitObj = true\n\t// the maximum length of that buffer must be supplied\n\t// in the value.asBytes.length attribute before calling this function.\n\tif attr.NativeTypeNum == C.DPI_NATIVE_TYPE_BYTES && attr.OracleTypeNum == C.DPI_ORACLE_TYPE_NUMBER {\n\t\tvar a [39]byte\n\t\tC.dpiData_setBytes(&data.dpiData, (*C.char)(unsafe.Pointer(&a[0])), C.uint32_t(len(a)))\n\t}\n\n\t//fmt.Printf(\"getAttributeValue(%p, %p, %d, %+v)\\n\", O.dpiObject, attr.dpiObjectAttr, data.NativeTypeNum, data.dpiData)\n\tif C.dpiObject_getAttributeValue(O.dpiObject, attr.dpiObjectAttr, data.NativeTypeNum, &data.dpiData) == C.DPI_FAILURE {\n\t\treturn errors.Errorf(\"getAttributeValue(%q, obj=%+v, attr=%+v, typ=%d): %w\", name, O, attr.dpiObjectAttr, data.NativeTypeNum, O.getError())\n\t}\n\t//fmt.Printf(\"getAttributeValue(%p, %q=%p, %d, %+v)\\n\", O.dpiObject, attr.Name, attr.dpiObjectAttr, data.NativeTypeNum, data.dpiData)\n\treturn nil\n}", "func (o *LDAPIdentityProvider) GetAttributes() (value *LDAPAttributes, ok bool) {\n\tok = o != nil && o.bitmap_&4 != 0\n\tif ok {\n\t\tvalue = o.attributes\n\t}\n\treturn\n}", "func (w *wrapper) Getattr(path string, stat *fuse.Stat_t, fd uint64) int {\n\tfi, err := w.underlying.Stat(path)\n\tif err != nil {\n\t\treturn convertError(err)\n\t}\n\tfileInfoToStat(fi, stat)\n\treturn 0\n}", "func (c *jsiiProxy_CfnFilter) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (a *ManagementApiService) GetAttributes(ctx _context.Context) apiGetAttributesRequest {\n\treturn apiGetAttributesRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (c *jsiiProxy_CfnLayer) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (graph *Graph) GetAttr(x, y int) byte {\n\treturn graph.Tiles[y][x].Attr\n}", "func (repo *Repository) GetAttr(path, name string, flags ...AttrFlag) (string, error) {\n\tvar value [1]int8\n\tcvalue := (*C.char)(&value[0])\n\tcpath := C.CString(path)\n\tdefer C.free(unsafe.Pointer(cpath))\n\tcname := C.CString(name)\n\tdefer C.free(unsafe.Pointer(cname))\n\n\tvar cflags C.uint32_t\n\tfor _, flag := range flags {\n\t\tcflags |= C.uint32_t(flag)\n\t}\n\n\tecode := C.git_attr_get(&cvalue, repo.git_repository, cflags, cpath, cname)\n\tif ecode != git_SUCCESS {\n\t\treturn \"\", gitError()\n\t}\n\treturn C.GoString(cvalue), nil\n}", "func (o *LogContent) GetAttributes() map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret\n\t}\n\treturn o.Attributes\n}", "func (c *jsiiProxy_CfnInstance) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (e *Environment) Attr(environmentName, attr string) ([]Attr, error) {\n\n\targkeys := []string{\"attr\"}\n\targvalues := []interface{}{attr}\n\tbaseCommand := fmt.Sprintf(\"list environment attr %s\", environmentName)\n\n\tc, err := cmd.ArgsExpander(baseCommand, argkeys, argvalues)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tb, err := cmd.RunCommand(c)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tattrs := []Attr{}\n\terr = json.Unmarshal(b, &attrs)\n\tif err != nil {\n\t\t// it may have been just an empty output from the Frontend\n\t\tnullOutput := NullOutput{}\n\t\terr = json.Unmarshal(b, &nullOutput)\n\t\tif err != nil {\n\t\t\t// if we still can't recognize the output, return an error\n\t\t\treturn nil, err\n\t\t}\n\t\treturn attrs, err\n\t}\n\treturn attrs, err\n}", "func (r *Thing) Attributes() pulumi.MapOutput {\n\treturn (pulumi.MapOutput)(r.s.State[\"attributes\"])\n}", "func Attr(attrs ...a.Attribute) []a.Attribute {\n return attrs\n}", "func (o *NewCoupons) GetAttributes() map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func GetCustomAttribute(ctx *pulumi.Context,\n\tname string, id pulumi.IDInput, state *CustomAttributeState, opts ...pulumi.ResourceOption) (*CustomAttribute, error) {\n\tvar resource CustomAttribute\n\terr := ctx.ReadResource(\"vsphere:index/customAttribute:CustomAttribute\", name, id, state, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func (obj *Edge) GetAttribute(attrName string) types.TGAttribute {\n\treturn obj.getAttribute(attrName)\n}", "func (obj *SObject) AttributesField() *SObjectAttributes {\n\tattributes := obj.InterfaceField(sobjectAttributesKey)\n\n\tswitch attributes.(type) {\n\tcase SObjectAttributes:\n\t\t// Use a temporary variable to copy the value of attributes and return the address of the temp value.\n\t\tattrs := (attributes).(SObjectAttributes)\n\t\treturn &attrs\n\tcase map[string]interface{}:\n\t\t// Can't convert attributes to concrete type; decode interface.\n\t\tmapper := attributes.(map[string]interface{})\n\t\tattrs := &SObjectAttributes{}\n\t\tif mapper[\"type\"] != nil {\n\t\t\tattrs.Type = mapper[\"type\"].(string)\n\t\t}\n\t\tif mapper[\"url\"] != nil {\n\t\t\tattrs.URL = mapper[\"url\"].(string)\n\t\t}\n\t\treturn attrs\n\tdefault:\n\t\treturn nil\n\t}\n}", "func (f *File) getObjectAttrs() (*storage.ObjectAttrs, error) {\n\thandle, err := f.getObjectHandle()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn handle.Attrs(f.fileSystem.ctx)\n}", "func (m *IdentityUserFlowAttributeAssignment) GetUserAttribute()(IdentityUserFlowAttributeable) {\n return m.userAttribute\n}", "func (c *jsiiProxy_CfnStack) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnStack) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (fsys *FS) Getattr(path string, stat *fuse.Stat_t, fh uint64) (errc int) {\n\tdefer fs.Trace(path, \"fh=0x%X\", fh)(\"errc=%v\", &errc)\n\tnode, errc := fsys.getNode(path, fh)\n\tif errc == 0 {\n\t\terrc = fsys.stat(node, stat)\n\t}\n\treturn\n}", "func (a *ManagementApiService) GetAttribute(ctx _context.Context, attributeId int32) apiGetAttributeRequest {\n\treturn apiGetAttributeRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t\tattributeId: attributeId,\n\t}\n}", "func (elem *remoteWE) GetAttribute(name string) (string, error) {\n\ttemplate := \"/session/%%s/element/%s/attribute/%s\"\n\turlTemplate := fmt.Sprintf(template, elem.id, name)\n\n\treturn elem.parent.stringCommand(urlTemplate)\n}", "func (it isType) GetAttr(name string) Value {\n\treturn NilValue\n}", "func (c *jsiiProxy_CfnDetector) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (o *IamLdapBasePropertiesAllOf) GetAttribute() string {\n\tif o == nil || o.Attribute == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Attribute\n}", "func (o *IamLdapBasePropertiesAllOf) GetAttributeOk() (*string, bool) {\n\tif o == nil || o.Attribute == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Attribute, true\n}", "func (c *jsiiProxy_CfnDetectorModel) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnCustomResource) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnStudio) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (t *Transaction) GetAttributes(typ AttrType) []Attribute {\n\tvar result []Attribute\n\tfor _, attr := range t.Attributes {\n\t\tif attr.Type == typ {\n\t\t\tresult = append(result, attr)\n\t\t}\n\t}\n\treturn result\n}", "func (ac *AttrCache) GetAttr(options internal.GetAttrOptions) (*internal.ObjAttr, error) {\n\tlog.Trace(\"AttrCache::GetAttr : %s\", options.Name)\n\ttruncatedPath := internal.TruncateDirName(options.Name)\n\n\tac.cacheLock.RLock()\n\tvalue, found := ac.cacheMap[truncatedPath]\n\tac.cacheLock.RUnlock()\n\n\t// Try to serve the request from the attribute cache\n\tif found && value.valid() && time.Since(value.cachedAt).Seconds() < float64(ac.cacheTimeout) {\n\t\tif value.isDeleted() {\n\t\t\tlog.Debug(\"AttrCache::GetAttr : %s served from cache\", options.Name)\n\t\t\t// no entry if path does not exist\n\t\t\treturn &internal.ObjAttr{}, syscall.ENOENT\n\t\t} else {\n\t\t\t// IsMetadataRetrieved is false in the case of ADLS List since the API does not support metadata.\n\t\t\t// Once migration of ADLS list to blob endpoint is done (in future service versions), we can remove this.\n\t\t\t// options.RetrieveMetadata is set by CopyFromFile and WriteFile which need metadata to ensure it is preserved.\n\t\t\tif value.getAttr().IsMetadataRetrieved() || (ac.noSymlinks && !options.RetrieveMetadata) {\n\t\t\t\t// path exists and we have all the metadata required or we do not care about metadata\n\t\t\t\tlog.Debug(\"AttrCache::GetAttr : %s served from cache\", options.Name)\n\t\t\t\treturn value.getAttr(), nil\n\t\t\t}\n\t\t}\n\t}\n\n\t// Get the attributes from next component and cache them\n\tpathAttr, err := ac.NextComponent().GetAttr(options)\n\n\tac.cacheLock.Lock()\n\tdefer ac.cacheLock.Unlock()\n\n\tif err == nil {\n\t\t// Retrieved attributes so cache them\n\t\tif len(ac.cacheMap) < maxTotalFiles {\n\t\t\tac.cacheMap[truncatedPath] = newAttrCacheItem(pathAttr, true, time.Now())\n\t\t}\n\t} else if err == syscall.ENOENT {\n\t\t// Path does not exist so cache a no-entry item\n\t\tac.cacheMap[truncatedPath] = newAttrCacheItem(&internal.ObjAttr{}, false, time.Now())\n\t}\n\n\treturn pathAttr, err\n}", "func (attestedClaim *AttestedClaim) getRawAttributes() []*big.Int {\n\treturn attestedClaim.Credential.Attributes[1:]\n}", "func (a *AliyunInstanceAttribute) fetchAttributes(targetReader TargetReader, nodeName string) {\n\ta.ShootName = GetFromTargetInfo(targetReader, \"shootTechnicalID\")\n\tvar err error\n\tif a.FlagProviderID != \"\" {\n\t\ta.InstanceID = a.FlagProviderID\n\t} else {\n\t\ta.InstanceID, err = fetchAlicloudInstanceIDByNodeName(nodeName)\n\t\tcheckError(err)\n\t}\n\n\tres, err := ExecCmdReturnOutput(\"bash\", \"-c\", \"aliyun ecs DescribeInstanceAttribute --InstanceId=\"+a.InstanceID)\n\tcheckError(err)\n\tdecodedQuery := decodeAndQueryFromJSONString(res)\n\n\ta.RegionID, err = decodedQuery.String(\"RegionId\")\n\tcheckError(err)\n\ta.ZoneID, err = decodedQuery.String(\"ZoneId\")\n\tcheckError(err)\n\ta.VpcID, err = decodedQuery.String(\"VpcAttributes\", \"VpcId\")\n\tcheckError(err)\n\ta.VSwitchID, err = decodedQuery.String(\"VpcAttributes\", \"VSwitchId\")\n\tcheckError(err)\n\ta.ImageID, err = decodedQuery.String(\"ImageId\")\n\tcheckError(err)\n\tips, err := decodedQuery.ArrayOfStrings(\"VpcAttributes\", \"PrivateIpAddress\", \"IpAddress\")\n\tcheckError(err)\n\ta.PrivateIP = ips[0]\n\ta.BastionSecurityGroupName = a.ShootName + \"-bsg\"\n\ta.BastionInstanceName = a.ShootName + \"-bastion\"\n\n\ta.InstanceChargeType = \"PostPaid\"\n\ta.InternetChargeType = \"PayByTraffic\"\n\ta.InternetMaxBandwidthIn = \"10\"\n\ta.InternetMaxBandwidthOut = \"100\"\n\ta.IoOptimized = \"optimized\"\n\ta.InstanceType = a.getMinimumInstanceSpec()\n}", "func (c *jsiiProxy_CfnConnection) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnResourceVersion) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (o *object) Attr(i uint8) bool {\n\treturn o.Attributes[i/8]&(1<<(7-i%8)) != 0\n}", "func (o *LDAPIdentityProvider) Attributes() *LDAPAttributes {\n\tif o != nil && o.bitmap_&4 != 0 {\n\t\treturn o.attributes\n\t}\n\treturn nil\n}", "func (self Channel) GetAttribute(attrib int) (float64, error) {\n\tvar cvalue C.float\n\tresult:=C.BASS_ChannelGetAttribute(self.cint(), C.DWORD(attrib), &cvalue)\n\treturn float64(cvalue), boolToError(result)\n}", "func (c *jsiiProxy_CfnMember) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func GetAttrs(obj runtime.Object) (labels.Set, fields.Set, error) {\n\tsecret, ok := obj.(*api.Secret)\n\tif !ok {\n\t\treturn nil, nil, fmt.Errorf(\"not a secret\")\n\t}\n\treturn labels.Set(secret.Labels), SelectableFields(secret), nil\n}", "func (a *Client) GetAttributes(params *GetAttributesParams) (*GetAttributesOK, error) {\n\t// TODO: Validate the params before sending\n\tif params == nil {\n\t\tparams = NewGetAttributesParams()\n\t}\n\n\tresult, err := a.transport.Submit(&runtime.ClientOperation{\n\t\tID: \"get_attributes\",\n\t\tMethod: \"GET\",\n\t\tPathPattern: \"/api/rest/v1/attributes\",\n\t\tProducesMediaTypes: []string{\"application/json\"},\n\t\tConsumesMediaTypes: []string{\"application/json\"},\n\t\tSchemes: []string{\"http\"},\n\t\tParams: params,\n\t\tReader: &GetAttributesReader{formats: a.formats},\n\t\tContext: params.Context,\n\t\tClient: params.HTTPClient,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tsuccess, ok := result.(*GetAttributesOK)\n\tif ok {\n\t\treturn success, nil\n\t}\n\t// unexpected success response\n\t// safeguard: normally, absent a default response, unknown success responses return an error above: so this is a codegen issue\n\tmsg := fmt.Sprintf(\"unexpected success response for get_attributes: API contract not enforced by server. Client expected to get an error, but got: %T\", result)\n\tpanic(msg)\n}", "func (o *NodeUpdate) GetAttributes() []map[string]interface{} {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret []map[string]interface{}\n\t\treturn ret\n\t}\n\treturn *o.Attributes\n}", "func (a adapter) Attrs(key string) []string {\n\treturn a.entry.GetAttributeValues(key)\n}", "func (c *jsiiProxy_CfnStudioComponent) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (g *gcs) Attributes(ctx context.Context, remotePath string) (attributes interface{}, err error) {\n\tfile := g.bucket.Object(remotePath)\n\tattributes, err = file.Attrs(g.context)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (i *Index) Attr(name string) (starlark.Value, error) {\n\tswitch name {\n\tcase \"name\":\n\t\treturn starlark.String(i.name), nil\n\tcase \"str\":\n\t\treturn &stringMethods{subject: i}, nil\n\t}\n\treturn nil, starlark.NoSuchAttrError(name)\n}", "func (c *jsiiProxy_CfnClassifier) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnMacro) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnInput) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnRepository) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnMLTransform) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (a adapter) Attr(key string) string {\n\treturn a.entry.GetAttributeValue(key)\n}", "func (e *EntryBase) Attributes() *EntryAttributes {\n\treturn &e.attr\n}", "func (a *Attribute) GetName() string {\n\treturn a.Name\n}", "func (c *jsiiProxy_CfnDatabase) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnEnvironment) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (o *Service) GetAttributesOk() (map[string]string, bool) {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]string\n\t\treturn ret, false\n\t}\n\treturn *o.Attributes, true\n}", "func (h *Handler) getDeveloperAttributes(c *gin.Context) handlerResponse {\n\n\tdeveloper, err := h.service.Developer.Get(c.Param(developerParameter))\n\tif err != nil {\n\t\treturn handleError(err)\n\t}\n\treturn handleOKAttributes(developer.Attributes)\n}", "func (o *Giveaway) GetAttributesOk() (map[string]interface{}, bool) {\n\tif o == nil || o.Attributes == nil {\n\t\tvar ret map[string]interface{}\n\t\treturn ret, false\n\t}\n\treturn *o.Attributes, true\n}", "func (o *ObjectNode) getObjectXAttr(w http.ResponseWriter, r *http.Request) {\n\t// TODO: implement 'getObjectXAttr'\n}", "func (c *jsiiProxy_CfnFileSystem) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnStreamingImage) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnDeliveryStream) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnApplication) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnApplication) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *DOM) GetAttributesWithParams(v *DOMGetAttributesParams) ([]string, error) {\n\tresp, err := gcdmessage.SendCustomReturn(c.target, c.target.GetSendCh(), &gcdmessage.ParamRequest{Id: c.target.GetId(), Method: \"DOM.getAttributes\", Params: v})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar chromeData struct {\n\t\tResult struct {\n\t\t\tAttributes []string\n\t\t}\n\t}\n\n\tif resp == nil {\n\t\treturn nil, &gcdmessage.ChromeEmptyResponseErr{}\n\t}\n\n\t// test if error first\n\tcerr := &gcdmessage.ChromeErrorResponse{}\n\tjson.Unmarshal(resp.Data, cerr)\n\tif cerr != nil && cerr.Error != nil {\n\t\treturn nil, &gcdmessage.ChromeRequestErr{Resp: cerr}\n\t}\n\n\tif err := json.Unmarshal(resp.Data, &chromeData); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn chromeData.Result.Attributes, nil\n}", "func (dir HgmDir) Attr(ctx context.Context, a *fuse.Attr) error {\n\tresp, err := httpClient.Get(dir.getStatEndpoint(dir.localDir, false))\n\tif err != nil {\n\t\treturn fuse.EIO\n\t}\n\n\tfuseErr := stattool.HttpStatusToFuseErr(resp.StatusCode)\n\tif fuseErr == nil {\n\t\tbodyBytes, err := ioutil.ReadAll(resp.Body)\n\t\tif err == nil {\n\t\t\tattr := stattool.HgmStatAttr{}\n\t\t\terr = json.Unmarshal(bodyBytes, &attr)\n\t\t\tif err == nil {\n\t\t\t\tstattool.AttrFromHgmStat(attr, a)\n\t\t\t}\n\t\t}\n\t\tif err != nil {\n\t\t\tfuseErr = fuse.EIO\n\t\t}\n\t}\n\n\treturn fuseErr\n}", "func (c *jsiiProxy_CfnRegistry) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func getAttribute(name string,attrs []xml.Attr) (string) {\n\tval := \"\"\n\tfor _,attr := range attrs { if strings.EqualFold(attr.Name.Local,name) {val = attr.Value } }\n\treturn val\n}", "func (c *jsiiProxy_CfnPublicRepository) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (repo *Repository) GetAttributes(id string) (attributes []Attribute, err error) {\n\trows, err := repo.db.Query(\"SELECT attribute_id FROM product_attributes WHERE product_id LIKE $1\", id)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer func() { _ = rows.Close() }()\n\tvar attributeIds []int\n\tvar attr int\n\tfor rows.Next() {\n\t\terr = rows.Scan(&attr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tattributeIds = append(attributeIds, attr)\n\t}\n\n\tattributes = make([]Attribute, len(attributeIds))\n\tfor i := range attributeIds {\n\t\tattributes[i], err = repo.attributeName(attributeIds[i])\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn attributes, rows.Err()\n}", "func (eClass *eClassImpl) GetEAttributes() EList {\n\teClass.getInitializers().initEAttributes()\n\treturn eClass.eAttributes\n}", "func (c *jsiiProxy_CfnProfilingGroup) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnModuleVersion) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnSecurityConfiguration) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnWorkflow) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnPreset) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func (c *jsiiProxy_CfnVolume) GetAtt(attributeName *string) awscdk.Reference {\n\tvar returns awscdk.Reference\n\n\t_jsii_.Invoke(\n\t\tc,\n\t\t\"getAtt\",\n\t\t[]interface{}{attributeName},\n\t\t&returns,\n\t)\n\n\treturn returns\n}", "func getPiAttribute(w http.ResponseWriter, r *http.Request) {\n\t// Get pi name and property/attribute from request\n\tvars := mux.Vars(r)\n\tpiname := vars[\"piname\"]\n\tpiproperty := vars[\"piattribute\"]\n\n\t// Get pi entry from data store\n\tc := appengine.NewContext(r)\n\tq := datastore.NewQuery(piListKind).Filter(\"name =\", piname)\n\tt := q.Run(c)\n\tvar pi Pi\n\t_, err := t.Next(&pi)\n\tif err == datastore.Done {\n\t\thttp.Error(w, \"404 Not found\", http.StatusNotFound)\n\t\treturn\n\t}\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t// Print attribute value in plain text\n\tw.Header().Set(\"Content-Type\", \"text/plain\")\n\tswitch piproperty {\n\tcase \"name\":\n\t\tfmt.Fprint(w, pi.Name)\n\tcase \"ip\":\n\t\tfmt.Fprint(w, pi.Ip)\n\tcase \"lastSeen\":\n\t\tfmt.Fprint(w, pi.LastSeen)\n\tcase \"pingCount\":\n\t\tfmt.Fprint(w, pi.PingCount)\n\tdefault:\n\t\thttp.Error(w, \"404 Not found\", http.StatusNotFound)\n\t}\n}", "func (n *Use) Attributes() map[string]interface{} {\n\treturn nil\n}", "func (obj *BlobAttribute) GetAttributeDescriptor() types.TGAttributeDescriptor {\n\treturn obj.getAttributeDescriptor()\n}", "func (d *cudaDevice) GetAttributes() (map[string]interface{}, error) {\n\treturn nil, fmt.Errorf(\"GetAttributes is not supported for CUDA devices\")\n}", "func (bo BinaryOperator) Attributes() []string {\n\tset := make(map[string]struct{})\n\tfor _, x := range bo.Left.Attributes() {\n\t\tset[x] = struct{}{}\n\t}\n\tfor _, x := range bo.Right.Attributes() {\n\t\tset[x] = struct{}{}\n\t}\n\tattrs := make([]string, 0, len(set))\n\tfor x := range set {\n\t\tattrs = append(attrs, x)\n\t}\n\treturn attrs\n}" ]
[ "0.7258507", "0.6508018", "0.6349299", "0.6330042", "0.63010705", "0.6297056", "0.62293595", "0.6208222", "0.6193886", "0.6189464", "0.6178482", "0.61777824", "0.6163363", "0.6150132", "0.61387676", "0.61117226", "0.60997057", "0.6091487", "0.60761094", "0.60756046", "0.6065097", "0.60635644", "0.6063549", "0.6058266", "0.60493886", "0.6046884", "0.60419756", "0.6020591", "0.60147256", "0.5999066", "0.5996933", "0.59855324", "0.59753865", "0.59596235", "0.5945901", "0.5945901", "0.5936748", "0.5926221", "0.59108084", "0.5909558", "0.58552235", "0.5850155", "0.5849978", "0.5834076", "0.5816888", "0.5810684", "0.5810122", "0.58062667", "0.58028805", "0.57971567", "0.5779583", "0.5764129", "0.57582283", "0.5755919", "0.574963", "0.5747703", "0.57471675", "0.5740872", "0.5739216", "0.57374865", "0.5732717", "0.5727843", "0.57217765", "0.5721542", "0.57164365", "0.5710569", "0.57085043", "0.5701868", "0.5700806", "0.5693142", "0.56840515", "0.5683008", "0.5681846", "0.567936", "0.56780845", "0.56657696", "0.5660581", "0.5655827", "0.5649793", "0.56488335", "0.56447744", "0.56447744", "0.56414974", "0.56378824", "0.56365347", "0.56277174", "0.56245774", "0.5615122", "0.5611439", "0.5611109", "0.5605331", "0.5602537", "0.56000984", "0.55964714", "0.5594738", "0.55916214", "0.55882907", "0.5584701", "0.5580629", "0.5576676" ]
0.60762525
18
getActions get object's actions
func GetActions(obj_name string) (acts []Action) { apiversion, _ := beego.AppConfig.String("apiversion") if _, ok := ActionsMap[obj_name]; !ok { return } acts = ActionsMap[obj_name] for i := 0; i < len(acts); i++ { acts[i].URI = "/" + apiversion + "/" + obj_name if acts[i].Primary != "" { acts[i].URI += "/:" + acts[i].Primary } } return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (m *MessageRule) GetActions()(MessageRuleActionsable) {\n return m.actions\n}", "func (e *CachedEnforcer) GetAllActions() []string {\n\treturn e.api.GetAllActions()\n}", "func (e *Enforcer) GetAllActions(ctx context.Context) ([]string, error) {\n\tres, err := e.client.remoteClient.GetAllActions(ctx, &pb.EmptyRequest{Handler: e.handler})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn res.Array, nil\n}", "func (g *Game) getActions() *Actions {\n\n\tactions := make([]Action, 0)\n\n\tif g.Phase == Development {\n\t\tfor _, loco := range g.Locos {\n\t\t\tif g.isLocoAvailableForDevelopment(loco) {\n\t\t\t\tabbr := fmt.Sprintf(\"D:%s\", loco.Key)\n\t\t\t\tactions = append(actions, Action{\n\t\t\t\t\tAbbr: abbr,\n\t\t\t\t\tVerb: \"Develop\",\n\t\t\t\t\tNoun: loco.Name,\n\t\t\t\t\tCost: loco.DevelopmentCost,\n\t\t\t\t\tLoco: loco})\n\t\t\t}\n\t\t}\n\t}\n\n\t// I think you can always pass.\n\tactions = append(actions, Action{Abbr: \"P\", Verb: \"Pass\"})\n\tphase := Phases[g.Phase-1]\n\treturn &Actions{Phase: phase, Actions: actions}\n}", "func (h *Handler) Actions() []*Action {\n\treturn h.actions\n}", "func (e *Enforcer) GetAllActions() []string {\n\treturn e.GetAllNamedActions(\"p\")\n}", "func (e *SyncedEnforcer) GetAllActions() []string {\n\te.m.RLock()\n\tdefer e.m.RUnlock()\n\treturn e.Enforcer.GetAllActions()\n}", "func (b *Bill) GetActions() []LegistarAction {\n\tvar actions []LegistarAction\n\n\terr := json.Unmarshal([]byte(b.Data), &actions)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn actions\n}", "func (g *Game) getActionsJson() []byte {\n\tactions := g.getActions()\n\tb, err := json.Marshal(actions)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn b\n}", "func Actions() []Action {\n\n\tvar curActions = actions\n\n\tlist := make([]Action, 0, len(curActions))\n\n\tfor _, value := range curActions {\n\t\tlist = append(list, value)\n\t}\n\n\treturn list\n}", "func (c *Client) Actions() ([]introduce.Action, error) {\n\treturn c.service.Actions()\n}", "func (o *TransactionResult) GetActions() []LabeledActionResult {\n\tif o == nil {\n\t\tvar ret []LabeledActionResult\n\t\treturn ret\n\t}\n\treturn o.Actions\n}", "func Actions(w http.ResponseWriter, r *http.Request, _ httprouter.Params) {\n\tallActions, err := database.GetActions()\n\tcheck(err)\n\n\tactionJson, err := json.Marshal(allActions)\n\tcheck(err)\n\n\tfmt.Fprintf(w, \"%s\", string(actionJson))\n}", "func (actionService *ActionService) GetActions(\n\tdeviceUuid,\n\tUserUuid string,\n\tctx context.Context) (models.ResourcesActionResponseV1, error) {\n\n\tactions, err := actionService.actionRepository.GetActions(deviceUuid, UserUuid, ctx)\n\n\tif err != nil {\n\t\treturn models.ResourcesActionResponseV1{}, err\n\t}\n\n\tresourceActionModels := actionEntityToModel(actions)\n\n\tresourceActions := models.ResourcesActionResponseV1{\n\t\tActions: resourceActionModels,\n\t}\n\n\treturn resourceActions, nil\n}", "func (core *coreService) Actions(start uint64, count uint64) ([]*iotexapi.ActionInfo, error) {\n\tif err := core.checkActionIndex(); err != nil {\n\t\treturn nil, err\n\t}\n\tif count == 0 {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"count must be greater than zero\")\n\t}\n\tif count > core.cfg.RangeQueryLimit {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"range exceeds the limit\")\n\t}\n\n\ttotalActions, err := core.indexer.GetTotalActions()\n\tif err != nil {\n\t\treturn nil, status.Error(codes.Internal, err.Error())\n\t}\n\tif start >= totalActions {\n\t\treturn nil, status.Error(codes.InvalidArgument, \"start exceeds the total actions in the block\")\n\t}\n\tif totalActions == uint64(0) || count == 0 {\n\t\treturn []*iotexapi.ActionInfo{}, nil\n\t}\n\tif start+count > totalActions {\n\t\tcount = totalActions - start\n\t}\n\tif core.indexer != nil {\n\t\treturn core.getActionsFromIndex(totalActions, start, count)\n\t}\n\t// Finding actions in reverse order saves time for querying most recent actions\n\treverseStart := totalActions - (start + count)\n\tif totalActions < start+count {\n\t\treverseStart = uint64(0)\n\t\tcount = totalActions - start\n\t}\n\n\tvar res []*iotexapi.ActionInfo\n\tvar actsList [][]*iotexapi.ActionInfo\n\tvar hit bool\n\tfor height := core.bc.TipHeight(); height >= 1 && count > 0; height-- {\n\t\tblk, err := core.dao.GetBlockByHeight(height)\n\t\tif err != nil {\n\t\t\treturn nil, status.Error(codes.NotFound, err.Error())\n\t\t}\n\t\tif !hit && reverseStart >= uint64(len(blk.Actions)) {\n\t\t\treverseStart -= uint64(len(blk.Actions))\n\t\t\tcontinue\n\t\t}\n\t\t// now reverseStart < len(blk.Actions), we are going to fetch actions from this block\n\t\thit = true\n\t\tacts := core.reverseActionsInBlock(blk, reverseStart, count)\n\t\tactsList = append(actsList, acts)\n\t\tcount -= uint64(len(acts))\n\t\treverseStart = 0\n\t}\n\tfor i := len(actsList) - 1; i >= 0; i-- {\n\t\tres = append(res, actsList[i]...)\n\t}\n\treturn res, nil\n}", "func (o ObfuscationRuleOutput) Actions() ObfuscationRuleActionArrayOutput {\n\treturn o.ApplyT(func(v *ObfuscationRule) ObfuscationRuleActionArrayOutput { return v.Actions }).(ObfuscationRuleActionArrayOutput)\n}", "func (c *Collection) Actions() *ActionList {\n\treturn &ActionList{coll: c}\n}", "func (ref *UIElement) Actions() []string {\n\ta := []string{}\n\tvar actions C.CFArrayRef\n\tC.AXUIElementCopyActionNames(ref.obj, &actions)\n\tif actions == nil {\n\t\treturn a\n\t}\n\tdefer C.CFRelease((C.CFTypeRef)(actions))\n\tnum := int(C.CFArrayGetCount(actions))\n\tfor i := 0; i < num; i++ {\n\t\titem := C.CFArrayGetValueAtIndex(actions, C.CFIndex(i))\n\t\ta = append(a, stringFromCFString((C.CFStringRef)(item)))\n\t}\n\treturn a\n}", "func (m *Meow) Actions() ui.KeyActions {\n\treturn m.actions\n}", "func (e *CachedEnforcer) GetAllNamedActions(ptype string) []string {\n\treturn e.api.GetAllNamedActions(ptype)\n}", "func (ati *actionTrackerImpl) getSortedActions() []string {\n\tactions := make([]string, len(ati.actions))\n\tindex := 0\n\tfor action := range ati.actions {\n\t\tactions[index] = action\n\t\tindex++\n\t}\n\tsort.Strings(actions)\n\treturn actions\n}", "func (e *Enforcer) GetAllNamedActions(ctx context.Context, ptype string) ([]string, error) {\n\tres, err := e.client.remoteClient.GetAllNamedActions(ctx, &pb.SimpleGetRequest{\n\t\tEnforcerHandler: e.handler,\n\t\tPType: ptype,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn res.Array, nil\n}", "func (s modelService) ReadActions(ctx context.Context, in *pb.User) (*pb.ActionsResponse, error) {\n\t// TODO: Input validation\n\tvar resp pb.ActionsResponse\n\tresp = pb.ActionsResponse{\n\t// Actions:\n\t}\n\treturn &resp, nil\n}", "func (s *CertificatesService) GetActions(id string) ([]CertificateAction, error) {\n\tvar res []CertificateAction\n\n\tdata, err := s.client.Get(\"/v1/certificates/\" + id + \"/actions\")\n\tif err != nil {\n\t\treturn res, err\n\t}\n\n\terr = json.Unmarshal(data, &res)\n\n\treturn res, err\n}", "func (p *DefaultPolicy) GetActions() []string {\n\treturn p.Actions\n}", "func (e *Enforcer) GetAllNamedActions(ptype string) []string {\n\treturn e.model.GetValuesForFieldInPolicy(\"p\", ptype, 2)\n}", "func KnownActions() []string {\n\tvar res []string\n\tfor k := range actions {\n\t\tres = append(res, k)\n\t}\n\tsort.Strings(res)\n\treturn res\n}", "func (o Actions) Get(gvk string) ActionInfoList {\n\treturn o[gvk]\n}", "func (*HttpCffActions) GetPath() string { return \"/api/objects/http/cff_action/\" }", "func (o ListenerRuleOutput) Actions() ListenerRuleActionArrayOutput {\n\treturn o.ApplyT(func(v *ListenerRule) ListenerRuleActionArrayOutput { return v.Actions }).(ListenerRuleActionArrayOutput)\n}", "func (s *TestSource) Actions() <-chan ActionRequest {\n\treturn s.outgoing\n}", "func (hh *HookHandler) getHookActions(payload *Payload, payloadPath string) []HookAction {\n\tvar hookActions []HookAction\n\n\t// hook action gets triggered first before deploy\n\tif hh.Hook.Command != nil && len(hh.Hook.Command) > 0 {\n\t\tif hh.Josuke.LogEnabled(TraceLevel) {\n\t\t\tlog.Println(\"[TRAC] hook action\")\n\t\t}\n\t\thookActions = append(\n\t\t\thookActions,\n\t\t\tHookAction{\n\t\t\t\tAction: &Action{\n\t\t\t\t\tAction: \"hook\",\n\t\t\t\t\tCommands: [][]string{hh.Hook.Command},\n\t\t\t\t},\n\t\t\t\tInfo: &Info{\n\t\t\t\t\tBaseDir: \"\",\n\t\t\t\t\tProjDir: \"\",\n\t\t\t\t\tHtmlUrl: \"\",\n\t\t\t\t\tPayloadHook: hh.Hook.Name,\n\t\t\t\t\tPayloadPath: payloadPath,\n\t\t\t\t\tPayloadEvent: payload.Action,\n\t\t\t\t},\n\t\t\t})\n\t}\n\tif hh.Josuke.LogEnabled(TraceLevel) {\n\t\tlog.Println(\"[TRAC] hook action from deployment\")\n\t}\n\tif hh.Josuke.Deployment == nil {\n\t\treturn hookActions\n\t}\n\taction, info := payload.getDeployAction(hh.Josuke.Deployment, payloadPath, hh.Hook.Name, payload.Action)\n\n\t// No deployment found\n\tif action == nil {\n\t\treturn hookActions\n\t}\n\n\t// deployment action gets triggered second after hook action\n\treturn append(\n\t\thookActions,\n\t\tHookAction{\n\t\t\tAction: action,\n\t\t\tInfo: info,\n\t\t})\n}", "func (p *processor) Actions() []control.Action {\n\treturn nil\n}", "func (o DataSourceResourcePermissionOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DataSourceResourcePermission) []string { return v.Actions }).(pulumi.StringArrayOutput)\n}", "func listActions(actions []Actionable) error {\n\tlog.Printf(\"query=%s\", query)\n\n\tfor _, a := range actions {\n\t\tit := wf.NewItem(a.Title()).\n\t\t\tArg(a.Title()).\n\t\t\tIcon(a.Icon()).\n\t\t\tCopytext(a.Title()).\n\t\t\tUID(a.Title()).\n\t\t\tValid(true).\n\t\t\tVar(\"ALSF_ACTION\", a.Title())\n\n\t\tit.NewModifier(\"cmd\").\n\t\t\tSubtitle(\"Blacklist action\").\n\t\t\tArg(a.Title()).\n\t\t\tValid(true).\n\t\t\tIcon(IconBlacklist).\n\t\t\tVar(\"action\", \"blacklist\")\n\n\t\tif _, ok := a.(TabActionable); ok {\n\t\t\tit.Var(\"ALSF_ACTION_TYPE\", \"tab\").\n\t\t\t\tVar(\"action\", \"tab-action\")\n\t\t} else if _, ok := a.(URLActionable); ok {\n\t\t\tit.Var(\"ALSF_ACTION_TYPE\", \"url\").\n\t\t\t\tVar(\"action\", \"url-action\")\n\t\t}\n\t}\n\n\tif query != \"\" {\n\t\tres := wf.Filter(query)\n\t\tlog.Printf(\"%d action(s) for %q\", len(res), query)\n\t}\n\twf.WarnEmpty(\"No actions found\", \"Try a different query?\")\n\twf.SendFeedback()\n\treturn nil\n}", "func (m *VirtualEndpoint) GetBulkActions()([]CloudPcBulkActionable) {\n val, err := m.GetBackingStore().Get(\"bulkActions\")\n if err != nil {\n panic(err)\n }\n if val != nil {\n return val.([]CloudPcBulkActionable)\n }\n return nil\n}", "func (c *CommerceController) Actions(ctx *app.ActionsCommerceContext) error {\n\tactions, err := c.CommerceStorage.Actions(ctx.Category)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ctx.OK(actions)\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyStandardPredeployPtrOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *DeliveryPipelineSerialPipelineStageStrategyStandardPredeploy) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func (e *EndComponent) Actions() []control.Action {\n\treturn nil\n}", "func (o DataSetResourcePermissionOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DataSetResourcePermission) []string { return v.Actions }).(pulumi.StringArrayOutput)\n}", "func (this *AVTransport) GetCurrentTransportActions(instanceId uint32) ([]string, error) {\n\ttype Response struct {\n\t\tXMLName xml.Name\n\t\tActions string\n\t\tErrorResponse\n\t}\n\targs := []Arg{\n\t\t{\"InstanceID\", instanceId},\n\t}\n\tresponse := this.Svc.Call(\"GetCurrentTransportActions\", args)\n\tdoc := Response{}\n\txml.Unmarshal([]byte(response), &doc)\n\treturn strings.Split(doc.Actions, \", \"), doc.Error()\n}", "func GetActionAction(c *gin.Context) {\n\tresult := render.NewResult()\n\tdefer c.JSON(http.StatusOK, result)\n\n\tidArg := c.Param(\"id\")\n\tid, err := strconv.ParseUint(idArg, 10, 64)\n\tif nil != err {\n\t\tresult.Error(err)\n\n\t\treturn\n\t}\n\n\tsrv := service.FromContext(c)\n\tdata, err := srv.Actions.Find(c, id)\n\tif nil == data {\n\t\tresult.Error(err)\n\n\t\treturn\n\t}\n\n\tresult.Result = data\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyStandardPredeployOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DeliveryPipelineSerialPipelineStageStrategyStandardPredeploy) []string { return v.Actions }).(pulumi.StringArrayOutput)\n}", "func (c MethodsCollection) ActionGet() pActionGet {\n\treturn pActionGet{\n\t\tMethod: c.MustGet(\"ActionGet\"),\n\t}\n}", "func (m *MockIDistributedEnforcer) GetAllActions() []string {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"GetAllActions\")\n\tret0, _ := ret[0].([]string)\n\treturn ret0\n}", "func (m *MockOobService) Actions() ([]outofband.Action, error) {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"Actions\")\n\tret0, _ := ret[0].([]outofband.Action)\n\tret1, _ := ret[1].(error)\n\treturn ret0, ret1\n}", "func List() []string {\n\tkeys := []string{}\n\tfor k := range actions {\n\t\tkeys = append(keys, k)\n\t}\n\treturn keys\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyStandardPostdeployPtrOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *DeliveryPipelineSerialPipelineStageStrategyStandardPostdeploy) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func (a *RequestServiceApiService) GetActions(ctx _context.Context) ApiGetActionsRequest {\n\treturn ApiGetActionsRequest{\n\t\tApiService: a,\n\t\tctx: ctx,\n\t}\n}", "func (a *RequestServiceApiService) GetActionsExecute(r ApiGetActionsRequest) (*os.File, *_nethttp.Response, GenericOpenAPIError) {\n\tvar (\n\t\tlocalVarHTTPMethod = _nethttp.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\texecutionError GenericOpenAPIError\n\t\tlocalVarReturnValue *os.File\n\t)\n\n\tlocalBasePath, err := a.client.cfg.ServerURLWithContext(r.ctx, \"RequestServiceApiService.GetActions\")\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarPath := localBasePath + \"/requests/actions\"\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := _neturl.Values{}\n\tlocalVarFormParams := _neturl.Values{}\n\n\tif r.limit != nil {\n\t\tlocalVarQueryParams.Add(\"limit\", parameterToString(*r.limit, \"\"))\n\t}\n\tif r.offset != nil {\n\t\tlocalVarQueryParams.Add(\"offset\", parameterToString(*r.offset, \"\"))\n\t}\n\t// to determine the Content-Type header\n\tlocalVarHTTPContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHTTPContentType := selectHeaderContentType(localVarHTTPContentTypes)\n\tif localVarHTTPContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHTTPContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHTTPHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHTTPHeaderAccept := selectHeaderAccept(localVarHTTPHeaderAccepts)\n\tif localVarHTTPHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHTTPHeaderAccept\n\t}\n\treq, err := a.client.prepareRequest(r.ctx, localVarPath, localVarHTTPMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, nil, executionError\n\t}\n\n\tlocalVarHTTPResponse, err := a.client.callAPI(req)\n\tif err != nil || localVarHTTPResponse == nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tlocalVarBody, err := _ioutil.ReadAll(localVarHTTPResponse.Body)\n\tlocalVarHTTPResponse.Body.Close()\n\tlocalVarHTTPResponse.Body = _ioutil.NopCloser(bytes.NewBuffer(localVarBody))\n\tif err != nil {\n\t\texecutionError.error = err.Error()\n\t\treturn localVarReturnValue, localVarHTTPResponse, executionError\n\t}\n\n\tif localVarHTTPResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHTTPResponse.Status,\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHTTPResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHTTPResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHTTPResponse, executionError\n}", "func (d *SQLike) Actions() map[string]destination.Action {\n\treturn map[string]destination.Action{\n\t\t\"run-statements\": RunStatements{\n\t\t\tenv: d.env,\n\t\t},\n\t\t\"run-operation\": RunOperation{\n\t\t\tenv: d.env,\n\t\t\twh: d.wh,\n\t\t},\n\t}\n}", "func (factory *SevenBeeFactory) Actions() []bees.ActionDescriptor {\n\tactions := []bees.ActionDescriptor{\n\t\t{\n\t\t\tNamespace: factory.Name(),\n\t\t\tName: \"sms\",\n\t\t\tDescription: \"Sends SMS message(s)\",\n\t\t\tOptions: []bees.PlaceholderDescriptor{\n\t\t\t\t{\n\t\t\t\t\tDescription: \"Message text\",\n\t\t\t\t\tMandatory: true,\n\t\t\t\t\tName: \"text\",\n\t\t\t\t\tType: \"string\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\treturn actions\n}", "func (p *Parser) GetDefaultActions() []string {\n\treturn p.defaultActions\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPredeployPtrOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPredeploy) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func (h *Handler) Actions(a ...HandlerAction) *Handler {\n\th.actionMutex.Lock()\n\tdefer h.actionMutex.Unlock()\n\tfor _, action := range a {\n\t\th.actions = append(h.actions, action)\n\t}\n\treturn h\n}", "func (core *coreService) ActionsInActPool(actHashes []string) ([]action.SealedEnvelope, error) {\n\tvar ret []action.SealedEnvelope\n\tif len(actHashes) == 0 {\n\t\tfor _, sealeds := range core.ap.PendingActionMap() {\n\t\t\tret = append(ret, sealeds...)\n\t\t}\n\t\treturn ret, nil\n\t}\n\n\tfor _, hashStr := range actHashes {\n\t\ths, err := hash.HexStringToHash256(hashStr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tsealed, err := core.ap.GetActionByHash(hs)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tret = append(ret, sealed)\n\t}\n\treturn ret, nil\n}", "func LoadActionsView(ctx context.Context, list *ListWidget) error {\n\tlist.statusView.Status(\"Getting available Actions\", true)\n\n\tcurrentItem := list.CurrentItem()\n\tspan, ctx := tracing.StartSpanFromContext(ctx, \"actions:\"+currentItem.Name, tracing.SetTag(\"item\", currentItem))\n\tdefer span.Finish()\n\n\tdata, err := armclient.DoRequest(ctx, \"GET\", \"/providers/Microsoft.Authorization/providerOperations/\"+list.CurrentItem().Namespace+\"?api-version=2018-01-01-preview&$expand=resourceTypes\")\n\tif err != nil {\n\t\tlist.statusView.Status(\"Failed to get actions: \"+err.Error(), false)\n\t}\n\tvar opsRequest armclient.OperationsRequest\n\terr = json.Unmarshal([]byte(data), &opsRequest)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\titems := []*handlers.TreeNode{}\n\tfor _, resOps := range opsRequest.ResourceTypes {\n\t\tif resOps.Name == strings.Split(list.CurrentItem().ArmType, \"/\")[1] {\n\t\t\tfor _, op := range resOps.Operations {\n\t\t\t\tresourceAPIVersion, err := armclient.GetAPIVersion(currentItem.ArmType)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlist.statusView.Status(\"Failed to find an api version: \"+err.Error(), false)\n\t\t\t\t}\n\t\t\t\tstripArmType := strings.Replace(op.Name, currentItem.ArmType, \"\", -1)\n\t\t\t\tactionURL := strings.Replace(stripArmType, \"/action\", \"\", -1) + \"?api-version=\" + resourceAPIVersion\n\t\t\t\titems = append(items, &handlers.TreeNode{\n\t\t\t\t\tName: op.DisplayName,\n\t\t\t\t\tDisplay: op.DisplayName,\n\t\t\t\t\tExpandURL: currentItem.ID + \"/\" + actionURL,\n\t\t\t\t\tExpandReturnType: handlers.ActionType,\n\t\t\t\t\tItemType: \"action\",\n\t\t\t\t\tID: currentItem.ID,\n\t\t\t\t})\n\t\t\t}\n\t\t}\n\t}\n\tif len(items) > 1 {\n\t\tlist.SetNodes(items)\n\t}\n\tlist.statusView.Status(\"Fetched available Actions\", false)\n\n\treturn nil\n}", "func (_AuthContract *AuthContractSession) Actions(arg0 [32]byte) (struct {\n\tTarget common.Address\n\tSender common.Address\n\tReceiver common.Address\n\tAmt *big.Int\n\tConfirmations *big.Int\n\tTriggered bool\n}, error) {\n\treturn _AuthContract.Contract.Actions(&_AuthContract.CallOpts, arg0)\n}", "func _listUserAction(query *db.QueryParser) ([]*model.UserAction, error) {\n\tmodels := make([]*model.UserAction, 0)\n\tif err := query.Query(\n\t\tfunc(rows *sql.Rows) (bool, error) {\n\t\t\tm := &model.UserAction{}\n\t\t\terr := rows.Scan(\n\t\t\t\t&m.Id, &m.UserId, &m.Action, &m.Context, &m.CreateTime,\n\t\t\t)\n\t\t\tmodels = append(models, m)\n\t\t\treturn true, err\n\t\t},\n\t); err != nil {\n\t\treturn nil, err\n\t}\n\treturn models, nil\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyStandardPostdeployOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DeliveryPipelineSerialPipelineStageStrategyStandardPostdeploy) []string { return v.Actions }).(pulumi.StringArrayOutput)\n}", "func (c *TaskAction) ActionInfo() []base.ActionInfo {\n\treturn actionInfo\n}", "func FindActions(c *gin.Context) {\n\tvar action []models.Action\n\tresult := models.DB.Find(&action)\n\n\tif result.RowsAffected == 0 {\n\t\tc.JSON(http.StatusOK, helpers.NoResults())\n\t\treturn\n\t}\n\n\tc.JSON(http.StatusOK, gin.H{\"data\": helpers.Results{\n\t\tCount: len(action),\n\t\tResults: action,\n\t}})\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPredeployPtrOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPredeploy) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func (_AuthContract *AuthContractCallerSession) Actions(arg0 [32]byte) (struct {\n\tTarget common.Address\n\tSender common.Address\n\tReceiver common.Address\n\tAmt *big.Int\n\tConfirmations *big.Int\n\tTriggered bool\n}, error) {\n\treturn _AuthContract.Contract.Actions(&_AuthContract.CallOpts, arg0)\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPredeployOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPredeploy) []string {\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func (_AuthContract *AuthContractCaller) Actions(opts *bind.CallOpts, arg0 [32]byte) (struct {\n\tTarget common.Address\n\tSender common.Address\n\tReceiver common.Address\n\tAmt *big.Int\n\tConfirmations *big.Int\n\tTriggered bool\n}, error) {\n\tret := new(struct {\n\t\tTarget common.Address\n\t\tSender common.Address\n\t\tReceiver common.Address\n\t\tAmt *big.Int\n\t\tConfirmations *big.Int\n\t\tTriggered bool\n\t})\n\tout := ret\n\terr := _AuthContract.contract.Call(opts, out, \"actions\", arg0)\n\treturn *ret, err\n}", "func (a *Action) Get(i int) string {\n if a==nil || i<0 || i>= len(a.actions) {\n return \"\"\n }\n return a.actions[i]\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPostdeployPtrOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPostdeploy) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func (a *IAMApiService) GetResourceGroupActions(ctx context.Context, rid string, gid string) (IamActions, *http.Response, error) {\n\tvar (\n\t\tlocalVarHttpMethod = http.MethodGet\n\t\tlocalVarPostBody interface{}\n\t\tlocalVarFormFileName string\n\t\tlocalVarFileName string\n\t\tlocalVarFileBytes []byte\n\t\tlocalVarReturnValue IamActions\n\t)\n\n\t// create path and map variables\n\tlocalVarPath := a.client.cfg.BasePath + \"/acs/api/v1/acls/{rid}/groups/{gid}\"\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"rid\"+\"}\", fmt.Sprintf(\"%v\", rid), -1)\n\tlocalVarPath = strings.Replace(localVarPath, \"{\"+\"gid\"+\"}\", fmt.Sprintf(\"%v\", gid), -1)\n\n\tlocalVarHeaderParams := make(map[string]string)\n\tlocalVarQueryParams := url.Values{}\n\tlocalVarFormParams := url.Values{}\n\n\t// to determine the Content-Type header\n\tlocalVarHttpContentTypes := []string{}\n\n\t// set Content-Type header\n\tlocalVarHttpContentType := selectHeaderContentType(localVarHttpContentTypes)\n\tif localVarHttpContentType != \"\" {\n\t\tlocalVarHeaderParams[\"Content-Type\"] = localVarHttpContentType\n\t}\n\n\t// to determine the Accept header\n\tlocalVarHttpHeaderAccepts := []string{\"application/json\"}\n\n\t// set Accept header\n\tlocalVarHttpHeaderAccept := selectHeaderAccept(localVarHttpHeaderAccepts)\n\tif localVarHttpHeaderAccept != \"\" {\n\t\tlocalVarHeaderParams[\"Accept\"] = localVarHttpHeaderAccept\n\t}\n\tr, err := a.client.prepareRequest(ctx, localVarPath, localVarHttpMethod, localVarPostBody, localVarHeaderParams, localVarQueryParams, localVarFormParams, localVarFormFileName, localVarFileName, localVarFileBytes)\n\tif err != nil {\n\t\treturn localVarReturnValue, nil, err\n\t}\n\n\tlocalVarHttpResponse, err := a.client.callAPI(r)\n\tif err != nil || localVarHttpResponse == nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tlocalVarBody, err := ioutil.ReadAll(localVarHttpResponse.Body)\n\tlocalVarHttpResponse.Body.Close()\n\tif err != nil {\n\t\treturn localVarReturnValue, localVarHttpResponse, err\n\t}\n\n\tif localVarHttpResponse.StatusCode >= 300 {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: localVarHttpResponse.Status,\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 200 {\n\t\t\tvar v IamActions\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\tif localVarHttpResponse.StatusCode == 0 {\n\t\t\tvar v IamError\n\t\t\terr = a.client.decode(&v, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\t\t\tif err != nil {\n\t\t\t\tnewErr.error = err.Error()\n\t\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t\t}\n\t\t\tnewErr.model = v\n\t\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\terr = a.client.decode(&localVarReturnValue, localVarBody, localVarHttpResponse.Header.Get(\"Content-Type\"))\n\tif err != nil {\n\t\tnewErr := GenericOpenAPIError{\n\t\t\tbody: localVarBody,\n\t\t\terror: err.Error(),\n\t\t}\n\t\treturn localVarReturnValue, localVarHttpResponse, newErr\n\t}\n\n\treturn localVarReturnValue, localVarHttpResponse, nil\n}", "func GetAction(name string) (Action, error) {\n\tvar a Action\n\n\tpath := fmt.Sprintf(\"/action/%s\", name)\n\tdata, _, err := Request(\"GET\", path, nil)\n\tif err != nil {\n\t\treturn a, err\n\t}\n\n\tif err := json.Unmarshal(data, &a); err != nil {\n\t\treturn a, err\n\t}\n\n\treturn a, nil\n}", "func GetActionCards(cards []Card) []Card {\n\tactions := []Card{}\n\n\tfor _, card := range cards {\n\t\tif strings.ToLower(card.CardType) == \"action\" {\n\t\t\tactions = append(actions, card)\n\t\t}\n\t}\n\n\treturn actions\n}", "func (d TinkDB) GetWorkflowActions(ctx context.Context, wfID string) (*pb.WorkflowActionList, error) {\n\tquery := `\n\tSELECT action_list\n\tFROM workflow_state\n\tWHERE\n\t\tworkflow_id = $1;\n\t`\n\trow := d.instance.QueryRowContext(ctx, query, wfID)\n\tvar actionList string\n\terr := row.Scan(&actionList)\n\tif err == nil {\n\t\tactions := []*pb.WorkflowAction{}\n\t\tif err := json.Unmarshal([]byte(actionList), &actions); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn &pb.WorkflowActionList{\n\t\t\tActionList: actions}, nil\n\t}\n\tif err != sql.ErrNoRows {\n\t\terr = errors.Wrap(err, \"SELECT from worflow_state\")\n\t\td.logger.Error(err)\n\t}\n\treturn &pb.WorkflowActionList{}, nil\n}", "func (t *trs80) writeActions() {\n\tfor i := 0; i < int(t.pb.Header.NumActions); i++ {\n\t\ta := t.pb.Actions[i]\n\t\tt.writeIntLn(a.VerbIndex*150 + a.NounIndex)\n\t\tfor j := 0; j < 5; j++ {\n\t\t\tt.writeIntLn(int32(a.Conditions[j].Type) + 20*a.Conditions[j].Value)\n\t\t}\n\t\tt.writeIntLn(int32(a.Actions[0])*150 + int32(a.Actions[1]))\n\t\tt.writeIntLn(int32(a.Actions[2])*150 + int32(a.Actions[3]))\n\t}\n}", "func (o *ReservationModel) GetActions() []ActionModelReservationActionNotAllowedReservationActionReason {\n\tif o == nil || o.Actions == nil {\n\t\tvar ret []ActionModelReservationActionNotAllowedReservationActionReason\n\t\treturn ret\n\t}\n\treturn *o.Actions\n}", "func Get(actionType string) Action {\n\treturn actions[actionType]\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPostdeployPtrOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v *DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPostdeploy) []string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func (s *ActionService) List(packageName string, options *ActionListOptions) ([]Action, *http.Response, error) {\n\tvar route string\n\tvar actions []Action\n\n\tif len(packageName) > 0 {\n\t\t// Encode resource name as a path (with no query params) before inserting it into the URI\n\t\t// This way any '?' chars in the name won't be treated as the beginning of the query params\n\t\tpackageName = (&url.URL{Path: packageName}).String()\n\t\troute = fmt.Sprintf(\"actions/%s/\", packageName)\n\t} else {\n\t\troute = fmt.Sprintf(\"actions\")\n\t}\n\n\trouteUrl, err := addRouteOptions(route, options)\n\tif err != nil {\n\t\tDebug(DbgError, \"addRouteOptions(%s, %#v) error: '%s'\\n\", route, options, err)\n\t\terrMsg := wski18n.T(\"Unable to add route options '{{.options}}'\",\n\t\t\tmap[string]interface{}{\"options\": options})\n\t\twhiskErr := MakeWskErrorFromWskError(errors.New(errMsg), err, EXIT_CODE_ERR_GENERAL, DISPLAY_MSG,\n\t\t\tNO_DISPLAY_USAGE)\n\t\treturn nil, nil, whiskErr\n\t}\n\tDebug(DbgError, \"Action list route with options: %s\\n\", route)\n\n\treq, err := s.client.NewRequestUrl(\"GET\", routeUrl, nil, IncludeNamespaceInUrl, AppendOpenWhiskPathPrefix, EncodeBodyAsJson, AuthRequired)\n\tif err != nil {\n\t\tDebug(DbgError, \"http.NewRequestUrl(GET, %s, nil, IncludeNamespaceInUrl, AppendOpenWhiskPathPrefix, EncodeBodyAsJson, AuthRequired) error: '%s'\\n\", routeUrl, err)\n\t\terrMsg := wski18n.T(\"Unable to create HTTP request for GET '{{.route}}': {{.err}}\",\n\t\t\tmap[string]interface{}{\"route\": routeUrl, \"err\": err})\n\t\twhiskErr := MakeWskErrorFromWskError(errors.New(errMsg), err, EXIT_CODE_ERR_NETWORK, DISPLAY_MSG,\n\t\t\tNO_DISPLAY_USAGE)\n\t\treturn nil, nil, whiskErr\n\t}\n\n\tresp, err := s.client.Do(req, &actions, ExitWithSuccessOnTimeout)\n\tif err != nil {\n\t\tDebug(DbgError, \"s.client.Do() error - HTTP req %s; error '%s'\\n\", req.URL.String(), err)\n\t\treturn nil, resp, err\n\t}\n\n\treturn actions, resp, err\n}", "func (o ForwardingRuleOutput) RuleActions() ForwardingRuleRuleActionArrayOutput {\n\treturn o.ApplyT(func(v *ForwardingRule) ForwardingRuleRuleActionArrayOutput { return v.RuleActions }).(ForwardingRuleRuleActionArrayOutput)\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPredeployOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPredeploy) []string {\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func ListAction(req *http.Request) (interface{}, error) {\n\treturn listActionProcess(req)\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPostdeployOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DeliveryPipelineSerialPipelineStageStrategyCanaryCanaryDeploymentPostdeploy) []string {\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func NewActions() Actions {\n\treturn &actions{metrics: map[string]*metrics{}}\n}", "func (c Client) Actions() actions.Client {\n\treturn actions.NewClient(c...)\n}", "func (h *History) At(i int) []*Action {\n\treturn h.actions[i]\n}", "func (s UserSet) ActionGet() *actions.Action {\n\tres := s.Collection().Call(\"ActionGet\")\n\tresTyped, _ := res.(*actions.Action)\n\treturn resTyped\n}", "func filterInformerActions(actions []k8stesting.Action) []k8stesting.Action {\n\tret := []k8stesting.Action{}\n\tfor _, action := range actions {\n\t\tif action.Matches(\"get\", \"workflows\") ||\n\t\t\taction.Matches(\"list\", \"workflows\") ||\n\t\t\taction.Matches(\"watch\", \"workflows\") {\n\t\t\tcontinue\n\t\t}\n\t\tret = append(ret, action)\n\t}\n\n\treturn ret\n}", "func (l *ActionList) Get(doc Document, fps ...FieldPath) *ActionList {\n\treturn l.add(&Action{\n\t\tkind: driver.Get,\n\t\tdoc: doc,\n\t\tfieldpaths: fps,\n\t})\n}", "func (o DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPostdeployOutput) Actions() pulumi.StringArrayOutput {\n\treturn o.ApplyT(func(v DeliveryPipelineSerialPipelineStageStrategyCanaryCustomCanaryDeploymentPhaseConfigPostdeploy) []string {\n\t\treturn v.Actions\n\t}).(pulumi.StringArrayOutput)\n}", "func Get(name string) (Action, error) {\n\tif initFunc, exists := actions[name]; exists {\n\t\treturn initFunc()\n\t}\n\n\treturn nil, fmt.Errorf(\"action %q does not exist as a supported action\", name)\n}", "func Actions() *Command {\n\tcmd := &Command{\n\t\tCommand: &cobra.Command{\n\t\t\tUse: \"action\",\n\t\t\tShort: \"Display commands for retrieving resource action history\",\n\t\t\tLong: `The sub-commands of ` + \"`\" + `doctl compute action` + \"`\" + ` retrieve the history of actions taken on your resources.\n\nThis can be filtered to a specific action. For example, while ` + \"`\" + `doctl compute action list` + \"`\" + ` will list all of the actions taken on all of the resources in your account, ` + \"`\" + `doctl compute action get <action-id>` + \"`\" + ` will retrieve details for a specific action.`,\n\t\t},\n\t}\n\n\tactionDetails := `\n\n- The action ID\n- The action status (` + \"`\" + `pending` + \"`\" + `, ` + \"`\" + `completed` + \"`\" + `, etc)\n- The action type (` + \"`\" + `create` + \"`\" + `, ` + \"`\" + `destroy` + \"`\" + `, ` + \"`\" + `power_cycle` + \"`\" + `, ` + \"`\" + `power_off` + \"`\" + `, ` + \"`\" + `power_on` + \"`\" + `, ` + \"`\" + `backup` + \"`\" + `, ` + \"`\" + `migrate` + \"`\" + `, ` + \"`\" + `attach_volume` + \"`\" + `, etc)\n- The Date/Time when the action started, in RFC3339 format\n- The Date/Time when the action completed, in RFC3339 format\n- The resource ID of the resource upon which the action was taken\n- The resource type (Droplet, backend)\n- The region in which the action took place (nyc3, sfo2, etc)`\n\n\tCmdBuilder(cmd, RunCmdActionGet, \"get <action-id>\", \"Retrieve details about a specific action\", `This command retrieves the following details about a specific action taken on one of your resources:`+actionDetails, Writer,\n\t\taliasOpt(\"g\"), displayerType(&displayers.Action{}))\n\n\tcmdActionList := CmdBuilder(cmd, RunCmdActionList, \"list\", \"Retrieve a list of all recent actions taken on your resources\", `This command retrieves a list of all actions taken on your resources. The following details are provided:`+actionDetails, Writer,\n\t\taliasOpt(\"ls\"), displayerType(&displayers.Action{}))\n\tAddStringFlag(cmdActionList, doctl.ArgActionResourceType, \"\", \"\", \"Action resource type\")\n\tAddStringFlag(cmdActionList, doctl.ArgActionRegion, \"\", \"\", \"Action region\")\n\tAddStringFlag(cmdActionList, doctl.ArgActionAfter, \"\", \"\", \"Action completed after in RFC3339 format\")\n\tAddStringFlag(cmdActionList, doctl.ArgActionBefore, \"\", \"\", \"Action completed before in RFC3339 format\")\n\tAddStringFlag(cmdActionList, doctl.ArgActionStatus, \"\", \"\", \"Action status\")\n\tAddStringFlag(cmdActionList, doctl.ArgActionType, \"\", \"\", \"Action type\")\n\n\tcmdActionWait := CmdBuilder(cmd, RunCmdActionWait, \"wait <action-id>\", \"Block thread until an action completes\", `The command blocks the current thread, returning when an action completes.\n\nFor example, if you find an action when calling `+\"`\"+`doctl compute action list`+\"`\"+` that has a status of `+\"`\"+`in-progress`+\"`\"+`, you can note the action ID and call `+\"`\"+`doctl compute action wait <action-id>`+\"`\"+`, and doctl will appear to \"hang\" until the action has completed. This can be useful for scripting purposes.`, Writer,\n\t\taliasOpt(\"w\"), displayerType(&displayers.Action{}))\n\tAddIntFlag(cmdActionWait, doctl.ArgPollTime, \"\", 5, \"Re-poll time in seconds\")\n\n\treturn cmd\n}", "func (p *PdfiumImplementation) GetJavaScriptActions(request *requests.GetJavaScriptActions) (*responses.GetJavaScriptActions, error) {\n\tp.Lock()\n\tdefer p.Unlock()\n\n\tdocumentHandle, err := p.getDocumentHandle(request.Document)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := p.Module.ExportedFunction(\"FPDFDoc_GetJavaScriptActionCount\").Call(p.Context, *documentHandle.handle)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcJavaScriptActionCount := *(*int32)(unsafe.Pointer(&res[0]))\n\tjavaScriptActionCount := int(cJavaScriptActionCount)\n\tif int(javaScriptActionCount) == -1 {\n\t\treturn nil, errors.New(\"could not get JavaScript Action count\")\n\t}\n\n\tjavaScriptActions := []responses.JavaScriptAction{}\n\tfor i := 0; i < javaScriptActionCount; i++ {\n\t\tres, err = p.Module.ExportedFunction(\"FPDFDoc_GetJavaScriptAction\").Call(p.Context, *documentHandle.handle, uint64(i))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tjavaScriptAction := res[0]\n\t\tif javaScriptAction == 0 {\n\t\t\tcontinue\n\t\t}\n\t\tdefer p.Module.ExportedFunction(\"FPDFDoc_CloseJavaScriptAction\").Call(p.Context, javaScriptAction)\n\n\t\t// First get the name value length.\n\t\tres, err = p.Module.ExportedFunction(\"FPDFJavaScriptAction_GetName\").Call(p.Context, javaScriptAction, 0, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tnameSize := *(*int32)(unsafe.Pointer(&res[0]))\n\t\tif nameSize == 0 {\n\t\t\treturn nil, errors.New(\"Could not get name\")\n\t\t}\n\n\t\tcharDataPointer, err := p.ByteArrayPointer(uint64(nameSize), nil)\n\t\tdefer charDataPointer.Free()\n\n\t\t_, err = p.Module.ExportedFunction(\"FPDFJavaScriptAction_GetName\").Call(p.Context, javaScriptAction, charDataPointer.Pointer, uint64(nameSize))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tcharData, err := charDataPointer.Value(false)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\ttransformedName, err := p.transformUTF16LEToUTF8(charData)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// First get the script value length.\n\t\tres, err = p.Module.ExportedFunction(\"FPDFJavaScriptAction_GetScript\").Call(p.Context, javaScriptAction, 0, 0)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tscriptSize := *(*int32)(unsafe.Pointer(&res[0]))\n\t\tif scriptSize == 0 {\n\t\t\treturn nil, errors.New(\"Could not get script\")\n\t\t}\n\n\t\tcharDataPointer, err = p.ByteArrayPointer(uint64(scriptSize), nil)\n\t\tdefer charDataPointer.Free()\n\n\t\t_, err = p.Module.ExportedFunction(\"FPDFJavaScriptAction_GetScript\").Call(p.Context, javaScriptAction, charDataPointer.Pointer, uint64(scriptSize))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tcharData, err = charDataPointer.Value(false)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\ttransformedScript, err := p.transformUTF16LEToUTF8(charData)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tjavaScriptActions = append(javaScriptActions, responses.JavaScriptAction{\n\t\t\tName: transformedName,\n\t\t\tScript: transformedScript,\n\t\t})\n\t}\n\n\treturn &responses.GetJavaScriptActions{\n\t\tJavaScriptActions: javaScriptActions,\n\t}, nil\n}", "func GetAction(method string) Action {\n\tswitch strings.ToUpper(method) {\n\tcase http.MethodGet:\n\t\treturn ReadAction\n\tcase http.MethodHead:\n\t\treturn ReadAction\n\tcase http.MethodPost:\n\t\treturn CreateAction\n\tcase http.MethodPut:\n\t\treturn UpdateAction\n\tcase http.MethodDelete:\n\t\treturn DeleteAction\n\tcase http.MethodPatch:\n\t\treturn UpdateAction\n\tcase http.MethodOptions:\n\t\treturn ReadAction\n\tdefault:\n\t\treturn UnknownAction\n\t}\n}", "func GetPlayerActions(playerID uint, matchID uint, data common.IDatasource) (*PlayerMatchStats, error) {\n\n\t// On récupère le joueur\n\tplayer, err := data.GetPlayer(playerID)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// On récupère le match\n\tmatch, err := data.GetMatch(matchID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tplayerActions := []action{}\n\tfor i := 0; i < len(match.Actions); i++ {\n\t\tif uint(match.Actions[i].JoueurID) == playerID {\n\t\t\tplayerAction := action{\n\t\t\t\tID: match.Actions[i].ID,\n\t\t\t\tTypeAction: typeAction{Name: match.Actions[i].TypeAction.Nom},\n\t\t\t\tIsValid: match.Actions[i].ActionPositive,\n\t\t\t\tX1: match.Actions[i].X1,\n\t\t\t\tY1: match.Actions[i].Y1,\n\t\t\t\tX2: match.Actions[i].X2,\n\t\t\t\tY2: match.Actions[i].Y2,\n\t\t\t\tHomeScore: match.Actions[i].PointageMaison,\n\t\t\t\tAdvScore: match.Actions[i].PointageAdverse,\n\t\t\t\tTime: match.Actions[i].Temps,\n\t\t\t}\n\t\t\tplayerActions = append(playerActions, playerAction)\n\t\t}\n\t}\n\n\tposition, err := data.GetMatchPosition(player.ID, match.ID)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tstats := PlayerMatchStats{\n\t\tID: match.ID,\n\t\tNumber: player.Numero,\n\t\tFirstName: player.Prenom,\n\t\tLastName: player.Nom,\n\t\tPosition: position.Nom,\n\t\tActions: playerActions,\n\t}\n\n\treturn &stats, err\n}", "func (m *MessageRule) SetActions(value MessageRuleActionsable)() {\n m.actions = value\n}", "func getActionInfo(action *driveactivity.ActionDetail) string {\n\treturn getOneOf(*action)\n}", "func (schematics *SchematicsV1) ListActions(listActionsOptions *ListActionsOptions) (result *ActionList, response *core.DetailedResponse, err error) {\n\treturn schematics.ListActionsWithContext(context.Background(), listActionsOptions)\n}", "func NewActions() times.Actions {\n\ta := new(actions)\n\ta.data = make(map[string]times.Action)\n\ta.alwaysFunction = EmptyAction\n\treturn a\n}", "func (b *Block) RunnableActions() RunnableActions {\n\treturn RunnableActions{actions: b.Actions, txHash: b.txRoot}\n}", "func getActions(userData []byte) ([]provisioning.Cmd, error) {\n\tvar commands []provisioning.Cmd\n\n\tvar ignition ignitionTypes.Config\n\tif err := json.Unmarshal(userData, &ignition); err != nil {\n\t\treturn nil, fmt.Errorf(\"unmarshalling Ignition JSON: %w\", err)\n\t}\n\n\t// Generate commands for files.\n\tfor _, f := range ignition.Storage.Files {\n\t\traw := strings.TrimSpace(f.Contents.Source)\n\t\tcontents, err := decodeFileContents(raw)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"decoding file contents: %w\", err)\n\t\t}\n\n\t\tmode := strconv.FormatInt(int64(*f.Mode), 8)\n\t\tif len(mode) == 3 {\n\t\t\t// Sticky bit isn't specified - pad with a zero.\n\t\t\tmode = \"0\" + mode\n\t\t}\n\n\t\tif f.Path == \"/etc/kubeadm.sh\" {\n\t\t\tcontents = hackKubeadmIgnoreErrors(contents)\n\t\t}\n\n\t\tcommands = append(commands, []provisioning.Cmd{\n\t\t\t// Idempotently create the directory.\n\t\t\t{Cmd: \"mkdir\", Args: []string{\"-p\", filepath.Dir(f.Path)}},\n\t\t\t// Write the file.\n\t\t\t{Cmd: \"/bin/sh\", Args: []string{\"-c\", fmt.Sprintf(\"cat > %s /dev/stdin\", f.Path)}, Stdin: contents},\n\t\t\t// Set file permissions.\n\t\t\t{Cmd: \"chmod\", Args: []string{mode, f.Path}},\n\t\t}...)\n\t}\n\n\tfor _, u := range ignition.Systemd.Units {\n\t\tcontents := strings.TrimSpace(u.Contents)\n\t\tpath := fmt.Sprintf(\"/etc/systemd/system/%s\", u.Name)\n\n\t\tcommands = append(commands, []provisioning.Cmd{\n\t\t\t{Cmd: \"/bin/sh\", Args: []string{\"-c\", fmt.Sprintf(\"cat > %s /dev/stdin\", path)}, Stdin: contents},\n\t\t\t{Cmd: \"systemctl\", Args: []string{\"daemon-reload\"}},\n\t\t}...)\n\n\t\tif u.Enable || (u.Enabled != nil && *u.Enabled) {\n\t\t\tcommands = append(commands, provisioning.Cmd{Cmd: \"systemctl\", Args: []string{\"enable\", \"--now\", u.Name}})\n\t\t}\n\t}\n\n\treturn commands, nil\n}", "func (m *UnifiedRoleAssignmentScheduleRequest) GetAction()(*UnifiedRoleScheduleRequestActions) {\n return m.action\n}" ]
[ "0.7176492", "0.6920503", "0.690926", "0.6880014", "0.6809491", "0.68027526", "0.6601725", "0.6530065", "0.64685965", "0.63547814", "0.63450646", "0.6299845", "0.6253166", "0.6206097", "0.61842597", "0.6139663", "0.6119756", "0.6104244", "0.61000055", "0.60412616", "0.6028857", "0.6003722", "0.60034454", "0.60022604", "0.5992077", "0.59838057", "0.59726334", "0.5946072", "0.5940963", "0.59389246", "0.5926338", "0.59199077", "0.5873255", "0.58554137", "0.5848232", "0.57807153", "0.57756025", "0.5768139", "0.57225984", "0.5672082", "0.56672126", "0.56387705", "0.5634501", "0.56187737", "0.56173736", "0.561679", "0.561297", "0.55899495", "0.5574501", "0.5569226", "0.5550858", "0.5550546", "0.5542542", "0.55327827", "0.5524711", "0.5514844", "0.54987574", "0.54837674", "0.5472392", "0.5455788", "0.5436262", "0.5429807", "0.5427687", "0.5413783", "0.5405845", "0.53925544", "0.5375063", "0.5370594", "0.53656346", "0.53386176", "0.53383815", "0.5336541", "0.5330773", "0.5320305", "0.5319046", "0.53154254", "0.5306729", "0.5298746", "0.52901053", "0.5275129", "0.52698", "0.52414054", "0.5241096", "0.5237207", "0.52210176", "0.5210196", "0.52094936", "0.51812696", "0.51802105", "0.51725507", "0.5167698", "0.5159417", "0.51485246", "0.51429445", "0.5111262", "0.5097949", "0.5038582", "0.50348824", "0.50192106", "0.5013235" ]
0.7839278
0
New constructs a new instance of SQLStore.
func New(cfg Config) (*SQLStore, error) { db, err := sqlx.Open(string(cfg.Driver), cfg.DataSource) if err != nil { return nil, errors.Wrap(err, "unable to open db connection") } builder := sq.StatementBuilder.PlaceholderFormat(sq.Question) switch cfg.Driver { case DBDriverMySQL: // mysql db.MapperFunc(func(s string) string { return s }) //case DBDriverPostgres: // postgres // builder = builder.PlaceholderFormat(sq.Dollar) } return &SQLStore{ cfg: cfg, db: db, builder: builder, }, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func New(c *sqlstore.Config) Datastore {\n\treturn sqlstore.New(c)\n}", "func New(c *Config) *sqlstore {\n\tdsn := c.DSN()\n\tdbx, err := sqlx.Connect(c.Driver, dsn)\n\tif err != nil {\n\t\tlog.Fatalf(\"Error connecting to database: %s\", err)\n\t}\n\tdb := &sqlstore{dbx, c.Driver, c}\n\n\tif c.Host == \"\" || c.Driver == SQLITE {\n\t\tlog.Printf(\"Connected to %s database: %s\", c.Driver, c.Dbname())\n\t} else {\n\t\tlog.Printf(\"Connected to %s database: %s on %s\", c.Driver, c.Dbname(), c.Host)\n\t}\n\n\t// apply database migrations (if any)\n\tdb.Migrate()\n\n\treturn db\n}", "func New(opts ...StoreOption) (*Store, error) {\n\tso := storeOptions{}\n\n\tfor _, opt := range opts {\n\t\topt.apply(&so)\n\t}\n\n\tdbOptions := url.Values{}\n\tdbOptions.Set(\"_foreign_keys\", \"true\")\n\n\tdbURL := url.URL{\n\t\tScheme: \"file\",\n\t\tPath: so.path,\n\t\tRawQuery: dbOptions.Encode(),\n\t}\n\n\tdb, err := sqlx.Connect(\"sqlite3\", dbURL.String())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Mapper = reflectx.NewMapperFunc(\"json\", strings.ToLower)\n\n\treturn &Store{\n\t\tdb: db,\n\t}, nil\n}", "func New(ctx context.Context, db *sql.DB, m map[string]string) (*Store, error) {\n\tstore := &Store{db: db}\n\terr := store.InitTable(ctx, m)\n\treturn store, err\n}", "func New(options Options) *Store {\n\tif options.Codec == nil {\n\t\toptions.Codec = DefaultOptions.Codec\n\t}\n\n\tif options.Interval == 0 {\n\t\toptions.Interval = DefaultOptions.Interval\n\t}\n\n\tif options.TableName == \"\" {\n\t\toptions.TableName = DefaultOptions.TableName\n\t}\n\n\tsql := newSqlSvr(options.User, options.Pwd, options.Host, options.Db, options.TableName, options.Split)\n\tif sql == nil {\n\t\treturn nil\n\t}\n\n\ts := &Store{\n\t\tSql: sql,\n\t\tCodec: options.Codec,\n\t}\n\n\t//go s.autoGC(options.Interval)\n\n\treturn s\n}", "func NewStore(p string) (Store, error) {\n\tp = path.Join(p, SQLiteDBName)\n\tdb, err := sql.Open(\"sqlite3\", fmt.Sprintf(\"file:%s?mode=ro\", p))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tstmt, err := db.Prepare(\"select value from entries where key = ?\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcache, err := lru.New(DirCacheSize)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\taclCache, err := lru.New(AccessCacheSize)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &sqlStore{\n\t\tdb: db,\n\t\tstmt: stmt,\n\t\tcache: cache,\n\t\tacl: aclCache,\n\n\t\tusers: make(map[string]int),\n\t\tgroups: make(map[string]int),\n\t}, nil\n}", "func NewStore(db *sql.DB) Store {\n\treturn &SQLStore{\n\t\tdb:\t\tdb, \n\t\tQueries: New(db),\n\t}\n}", "func New() *Store {\n\treturn &Store{}\n}", "func New() *Store {\n\treturn &Store{}\n}", "func New() (Store, error) {\n dbUsername := os.Getenv(\"DB_USERNAME\")\n dbPassword := os.Getenv(\"DB_PASSWORD\")\n dbHost := os.Getenv(\"DB_HOST\")\n dbTable := os.Getenv(\"DB_TABLE\")\n dbPort := os.Getenv(\"DB_PORT\")\n dbSSLMode := os.Getenv(\"DB_SSL_MODE\")\n\n connectionString := fmt.Sprintf(\n \"host=%s port=%s user=%s dbname=%s password=%s sslmode=%s\",\n dbHost, dbPort, dbUsername, dbTable, dbPassword, dbSSLMode,\n )\n\n db, err := sqlx.Connect(\"postgres\", connectionString)\n if err != nil {\n return Store{}, err\n }\n\n return Store{\n db: db,\n }, nil\n}", "func New(ds datastore.Datastore) *Store {\n\treturn &Store{\n\t\tds: ds,\n\t}\n}", "func New(db *sqlite_db.Db) qfxdb.Store {\n\treturn sqliteStore{db}\n}", "func New(ctx context.Context, log logger.Logger, db *db.Store, cache *cache.Cache) (*Store, error) {\n\ts := &Store{\n\t\tlog: log,\n\t\tcache: cache,\n\t}\n\n\t// Set configuration\n\ts.setConfig()\n\n\tvar err error\n\n\tswitch s.typeStore {\n\tcase \"postgres\":\n\t\tfallthrough\n\tdefault:\n\t\ts.store, err = postgres.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tlog.Info(\"init queryStore\", field.Fields{\n\t\t\"db\": s.typeStore,\n\t})\n\n\treturn s, nil\n}", "func NewStore(db *sql.DB) *Store {\n\n\treturn &Store{\n\t\tdb: db,\n\t\tQueries: New(db), //defined in db.go by sqlc\n\t}\n}", "func New() *Store {\n\treturn &Store{\n\t\tmu: sync.Mutex{},\n\t\tsess: make(map[string]*entities.Session),\n\t}\n}", "func New(c Config) (*Store, error) {\n\tstore := &Store{config: c}\n\tdb, err := sqlx.Connect(\"postgres\", fmt.Sprintf(\"user=%s password=%s dbname=%s host=%s port=%d sslmode=disable\", c.User, c.Password, c.Database, c.Host, c.Port))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdb.DB.SetMaxOpenConns(c.MaxConnections)\n\tstore.DB = db\n\treturn store, nil\n}", "func NewStore(db *sql.DB) (*Store, error) {\n\tunprepared := map[string]string{\n\t\tQueryCreateItem: `\n\t\t\tINSERT INTO todo.items (title, description)\n\t\t\tVALUES(?, ?);\n\t\t`,\n\t\tQueryDeleteItem: `\n\t\t\tDELETE FROM todo.items\n\t\t\tWHERE id = ?;\n\t\t`,\n\t\tQueryFindItemByID: `\n\t\t\tSELECT i.id, i.title, i.description, i.completed, i.created_at, i.updated_at\n\t\t\tFROM todo.items i\n\t\t\tWHERE id = ?;\n\t\t`,\n\t\tQueryUpdateItemByID: `\n\t\t\tUPDATE todo.items i\n\t\t\tSET \n\t\t\t\ti.title = ?,\n\t\t\t\ti.description = ?,\n\t\t\t\ti.completed = ?\n\t\t\tWHERE i.id = ?;\n\t\t`,\n\t}\n\n\t// prepare all statements to verify syntax\n\tstmts, err := prepareStmts(db, unprepared)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ts := Store{\n\t\tdb: db,\n\t\tstmts: stmts,\n\t}\n\n\treturn &s, nil\n}", "func New(tableName string, accessor Accessor) (*Store, error) {\n\tstore := &Store{\n\t\ttableName: tableName,\n\t\taccessor: accessor,\n\t}\n\n\treturn store, nil\n}", "func New() *Store {\n\treturn &Store{\n\t\tm: make(map[interface{}]entry),\n\t}\n}", "func NewStore(db *sql.DB) *Store {\n\treturn &Store{\n\t\tdb: db,\n\t\tQueries: New(db), // New creates and returns a queries object\n\t}\n}", "func New(db *pgxpool.Pool, pools *pools.Pools) (*Store, error) {\n\t// Confirm the database has the right schema.\n\texpectedSchema, err := expectedschema.Load()\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\n\tactual, err := schema.GetDescription(db, Tables{})\n\tif err != nil {\n\t\treturn nil, skerr.Wrap(err)\n\t}\n\tif diff := assertdeep.Diff(expectedSchema, *actual); diff != \"\" {\n\t\treturn nil, skerr.Fmt(\"Schema needs to be updated: %s.\", diff)\n\t}\n\n\treturn &Store{\n\t\tdb: db,\n\t\tpools: pools,\n\t}, nil\n}", "func New(addr, password string) *Store {\n\treturn &Store{\n\t\tpool: newPool(addr, password),\n\t}\n}", "func New() *Store {\n\treturn &Store{\n\t\tingredient.NewStore(),\n\t\trecipe.NewStore(),\n\t}\n}", "func NewStore() (StoreType, error) {\n\tconnection, err := config.DBConnectionString()\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tdb, err := sqlx.Connect(driver, connection)\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn nil, err\n\t}\n\ts := &Store{\n\t\tdb: db,\n\t}\n\n\treturn s, nil\n}", "func New(dir string) *Store {\n\treturn NewWithDB(\"\", badger.New(dir))\n}", "func New(ctx context.Context, cfg *config.Config, logger *logrus.Logger) (*Store, error) {\n\tvar store Store\n\n\tstore.config = cfg\n\tstore.logger = logger\n\n\t// connect to postgres\n\tpgConn, err := postgres.NewConnect(cfg.Postgres)\n\tif err != nil {\n\t\treturn &store, errpath.Err(err)\n\t}\n\tstore.Pg = pgConn\n\n\tif pgConn != nil {\n\t\tgo store.keepAlivePg()\n\t}\n\n\treturn &store, nil\n}", "func NewStore(yogo *sql.DB) Store {\n\treturn &SQLStore{\n\t\tyogo: yogo,\n\t\tQueries: New(yogo),\n\t}\n}", "func NewStore()(*Store) {\n m := &Store{\n Entity: *iadcd81124412c61e647227ecfc4449d8bba17de0380ddda76f641a29edf2b242.NewEntity(),\n }\n return m\n}", "func New(dburl string) *Store {\n\treturn &Store{\n\t\tDatabaseURL: dburl,\n\t}\n}", "func New() CStore {\n\treturn CStore{\n\t\tstore: make(map[string]string),\n\t}\n}", "func New(provider storage.Provider) (*Store, error) {\n\tstore, err := provider.OpenStore(nameSpace)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to open did anchor store: %w\", err)\n\t}\n\n\treturn &Store{\n\t\tstore: store,\n\t}, nil\n}", "func NewSQLStore(db *gorm.DB) *sqlStore {\n\treturn &sqlStore{db: db}\n}", "func New(storeConfig config.Store) (*Store, error) {\n\tdb, err := bolt.Open(storeConfig.DBPath, 0644, &bolt.Options{Timeout: 1 * time.Second})\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not open bolt DB database\")\n\t}\n\tbucketName := []byte(\"shorted\")\n\terr = db.Update(func(tx *bolt.Tx) error {\n\t\t_, err := tx.CreateBucketIfNotExists(bucketName)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Store{\n\t\tdb: db,\n\t\tidLength: storeConfig.ShortedIDLength,\n\t\tbucketName: bucketName,\n\t}, nil\n}", "func New(db *db.DB) core.StepStore {\n\treturn &stepStore{db}\n}", "func New() base.Store {\n\tvar s store\n\ts.data = make(map[string]entry)\n\treturn &s\n}", "func NewStore(opts ...store.Option) store.Store {\n\toptions := store.Options{\n\t\tDatabase: defaultDatabase,\n\t\tTable: defaultTable,\n\t}\n\n\tfor _, o := range opts {\n\t\to(&options)\n\t}\n\n\t// new store\n\ts := new(sqlStore)\n\ts.options = options\n\ts.databases = make(map[string]DB)\n\ts.re = regexp.MustCompile(\"[^a-zA-Z0-9]+\")\n\n\tgo s.expiryLoop()\n\t// return store\n\treturn s\n}", "func NewStore() *Store {\n\treturn &Store{\n\t\tES: MustOpenConnection(),\n\t}\n}", "func New(opt StoreOptions) *Store {\n\tstore := &Store{}\n\n\tfmt.Println(opt.toString())\n\tsession, err := mgo.Dial(opt.toString())\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tstore.cli = session\n\tstore.database = opt.Database\n\treturn store\n}", "func New(opts ...Opt) (*Store, error) {\n\tbadgerOptions := getBadgerOptions(opts...)\n\n\tdb, err := badger.Open(badgerOptions)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not open database: %w\", err)\n\t}\n\n\tstore := &Store{db, newChangelog()}\n\tif err = store.setup(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn store, nil\n}", "func New(cfg *config.Config, ssmSvc ssmiface.SSMAPI) *Store {\n\treturn &Store{\n\t\tssmSvc: ssmSvc,\n\t\tcfg: cfg,\n\t}\n}", "func New(client *ifirestore.Client, crsID string) *StoreImpl {\n\treturn &StoreImpl{\n\t\tclient: client,\n\t\tcrsID: crsID,\n\t}\n}", "func New(ctx context.Context, log logger.Logger, db *db.Store, cache *cache.Cache) (*Store, error) { // nolint:gocognit\n\ts := &Store{\n\t\tlog: log,\n\t\tcache: cache,\n\t}\n\n\t// Set configuration\n\ts.setConfig()\n\n\tvar err error\n\n\tswitch s.typeStore {\n\tcase \"postgres\":\n\t\ts.store, err = postgres.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"mongo\":\n\t\ts.store, err = mongo.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"redis\":\n\t\ts.store, err = redis.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"dgraph\":\n\t\ts.store, err = dgraph.New(ctx, db, log)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"leveldb\":\n\t\ts.store, err = leveldb.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"badger\":\n\t\ts.store, err = badger.New(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"ram\":\n\t\tfallthrough\n\tdefault:\n\t\ts.store, err = ram.New(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tlog.Info(\"init linkStore\", field.Fields{\n\t\t\"db\": s.typeStore,\n\t})\n\n\treturn s, nil\n}", "func NewSQLStore(db *sql.DB, searchIndex *indexes.TrieNode) *SQLStore {\n\tmySQLStore := SQLStore{\n\t\tdb: db,\n\t\tsearchIndex: searchIndex,\n\t}\n\treturn &mySQLStore\n}", "func New(config *Config) Store {\n\treturn newStore(config)\n}", "func New(tableName string, opts ...Option) (*Store, error) {\n\tstore := &Store{\n\t\tregion: DefaultRegion,\n\t\ttableName: tableName,\n\t\thashKey: HashKey,\n\t\trangeKey: RangeKey,\n\t\teventsPerItem: 100,\n\t}\n\n\tfor _, opt := range opts {\n\t\topt(store)\n\t}\n\n\tif store.api == nil {\n\t\tcfg := &aws.Config{Region: aws.String(store.region)}\n\t\ts, err := session.NewSession(cfg)\n\t\tif err != nil {\n\t\t\tif v, ok := err.(awserr.Error); ok {\n\t\t\t\treturn nil, errors.Wrapf(err, \"Unable to create AWS Session - %v [%v]\", v.Message(), v.Code())\n\t\t\t}\n\t\t\treturn nil, err\n\t\t}\n\t\tstore.api = dynamodb.New(s)\n\t}\n\n\treturn store, nil\n}", "func New(_ string) (s *Store, err error) {\n\treturn &Store{xz.NewMap()}, nil\n}", "func New(storeType string) (*Store, error) {\n\tswitch storeType {\n\tcase \"memory\":\n\t\treturn &Store{storeType: storeType, engine: makeMemoryStore()}, nil\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported store type: %s\", storeType)\n\t}\n}", "func NewStore(db *cockroach.DB, logger *zap.Logger) *Store {\n\treturn &Store{\n\t\tdb: db,\n\t\tlogger: logger,\n\t\tclock: DefaultClock,\n\t}\n}", "func New(dbConn *sql.DB, logger log.Factory, tracer opentracing.Tracer) Store {\n\treturn &store{\n\t\tdb: dbConn,\n\t\tlogger: logger.With(zap.String(\"package\", \"healthrepo\")),\n\t\ttracer: tracer,\n\t}\n}", "func NewStore() *Store {\n\tvar st Store\n\tst.Records = make(map[string]HostSet)\n\tst.Netviews = make(map[string]string)\n\tst.Cidrs = make(map[string]string)\n\treturn &st\n}", "func NewStore() *Store {\n\treturn &Store{\n\t\tls: make(map[string]InitFunc),\n\t}\n}", "func New() kv.Store {\n\treturn newStore(newMapStore())\n}", "func New(metainfo *Client, encStore *encryption.Store) *DB {\n\treturn &DB{\n\t\tmetainfo: metainfo,\n\t\tencStore: encStore,\n\t}\n}", "func New() *thingsStore {\n\tdb, err := gorm.Open(\"postgres\", \"host=localhost port=5432 user=postgres dbname=gorming sslmode=disable\")\n\tif err != nil {\n\t\treturn nil\n\t}\n\treturn &thingsStore{\n\t\tdb: db,\n\t}\n}", "func New(connectionString string, log *log.Logger) (*Store, error) {\n\tdb, err := gorm.Open(postgres.Open(connectionString), &gorm.Config{})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := db.AutoMigrate(&gh.Repository{}, &gh.Commit{}); err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Println(\"db init successful\")\n\treturn &Store{\n\t\tdb: db,\n\t\tlog: log,\n\t}, nil\n}", "func NewStore(l log.Logger, dsn string) (hydrocarbon.PrimitiveStore, error) {\n\tdb, err := sqlx.Connect(\"postgres\", dsn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Mapper = reflectx.NewMapperFunc(\"json\", strings.ToLower)\n\n\terr = Migrate(l, db)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Store{\n\t\tdb: db,\n\t}, nil\n}", "func New(db *leveldb.DB) (Storage, error) {\n\ttx, err := db.OpenTransaction()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error opening leveldb transaction: %v\", err)\n\t}\n\n\treturn &storage{\n\t\tstore: tx,\n\t\tdb: db,\n\t\ttx: tx,\n\t}, nil\n}", "func NewStore(d *db.DB) *Store {\n\treturn &Store{\n\t\tdb: d,\n\t}\n}", "func New(client *ifirestore.Client) *Store {\n\treturn &Store{\n\t\tclient: client,\n\t}\n}", "func NewSQLStore(c config.Database) (*SQLStore, error) {\n\tvar driver string\n\tvar dsn string\n\tswitch c[\"type\"] {\n\tcase \"mysql\":\n\t\tdriver = \"mysql\"\n\t\taddress := c[\"address\"]\n\t\t_, _, err := net.SplitHostPort(address)\n\t\tif err != nil {\n\t\t\taddress = address + \":3306\"\n\t\t}\n\t\tm := &mysql.Config{\n\t\t\tUser: c[\"username\"],\n\t\t\tPasswd: c[\"password\"],\n\t\t\tNet: \"tcp\",\n\t\t\tAddr: address,\n\t\t\tDBName: \"certs\",\n\t\t\tParseTime: true,\n\t\t}\n\t\tdsn = m.FormatDSN()\n\tcase \"sqlite\":\n\t\tdriver = \"sqlite3\"\n\t\tdsn = c[\"filename\"]\n\t}\n\tconn, err := sqlx.Open(driver, dsn)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"SQLStore: could not get a connection: %v\", err)\n\t}\n\tif err := conn.Ping(); err != nil {\n\t\tconn.Close()\n\t\treturn nil, fmt.Errorf(\"SQLStore: could not establish a good connection: %v\", err)\n\t}\n\n\tdb := &SQLStore{\n\t\tconn: conn,\n\t}\n\n\tif db.set, err = conn.Preparex(\"INSERT INTO issued_certs (key_id, principals, created_at, expires_at, raw_key) VALUES (?, ?, ?, ?, ?)\"); err != nil {\n\t\treturn nil, fmt.Errorf(\"SQLStore: prepare set: %v\", err)\n\t}\n\tif db.get, err = conn.Preparex(\"SELECT * FROM issued_certs WHERE key_id = ?\"); err != nil {\n\t\treturn nil, fmt.Errorf(\"SQLStore: prepare get: %v\", err)\n\t}\n\tif db.listAll, err = conn.Preparex(\"SELECT * FROM issued_certs\"); err != nil {\n\t\treturn nil, fmt.Errorf(\"SQLStore: prepare listAll: %v\", err)\n\t}\n\tif db.listCurrent, err = conn.Preparex(\"SELECT * FROM issued_certs WHERE ? <= expires_at\"); err != nil {\n\t\treturn nil, fmt.Errorf(\"SQLStore: prepare listCurrent: %v\", err)\n\t}\n\tif db.revoked, err = conn.Preparex(\"SELECT * FROM issued_certs WHERE revoked = 1 AND ? <= expires_at\"); err != nil {\n\t\treturn nil, fmt.Errorf(\"SQLStore: prepare revoked: %v\", err)\n\t}\n\treturn db, nil\n}", "func NewStore(c *cli.Context) Store {\n\treturn &datastore{\n\t\tDB: open(c.String(\"database-config\")),\n\t}\n}", "func New(ctx context.Context, c Config) (*Conn, error) {\n\tfiles, err := ioutil.ReadDir(\"./store/sql\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdots := make([]*dotsql.DotSql, 0, len(files))\n\tfor _, f := range files {\n\t\tdot, err := dotsql.LoadFromFile(filepath.Join(\"./store/sql\", f.Name()))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdots = append(dots, dot)\n\t}\n\tqueries := dotsql.Merge(dots[0])\n\tdb, err := sql.Open(\"postgres\", fmt.Sprintf(\"host=%s port=%s user=%s password=%s, dbname=%s sslmode=disable\", \"localhost\", \"5432\", c.Username, c.Password, c.Name))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres := &Conn{Queries: queries, db: db}\n\n\treturn res, db.PingContext(ctx)\n}", "func New(config string, w io.Writer, wErr io.Writer) (db *Storage, err error) {\n\tif w == nil {\n\t\tw = os.Stdout\n\t}\n\tif wErr == nil {\n\t\twErr = os.Stderr\n\t}\n\tdb = &Storage{\n\t\tlog: alog.New(w, \"SQL: \", 0),\n\t\tlogErr: alog.New(w, \"SQLErr: \", 0),\n\t}\n\n\tif config == \"\" {\n\t\terr = fmt.Errorf(\"Invalid configuration passed (empty)\")\n\t\treturn\n\t}\n\tif db.db, err = sqlx.Open(\"mysql\", config); err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func NewStore() *Store {\n\treturn &Store{\n\t\topaStore: inmem.New(),\n\t}\n}", "func NewStore() *Store {\n\treturn &Store{commands: make(map[string]*Config, 0)}\n}", "func New(pool *redis.Pool) *Store {\n\treturn &Store{\n\t\tpool: pool,\n\t\tidPrefix: defaultIDPrefix,\n\t\turlPrefix: defaultURLPrefix,\n\t}\n}", "func NewStore() (s Store) {\n\ts = make(Store, 0)\n\treturn s\n}", "func NewStore(client dbClient, validSec int) *Store {\n\treturn &Store{\n\t\tclient: client,\n\t\tvalid: time.Duration(validSec) * time.Second,\n\t}\n}", "func New(provider storage.Provider) (*Store, error) {\n\tstore, err := provider.OpenStore(nameSpace)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to open unpublished operation store: %w\", err)\n\t}\n\n\treturn &Store{\n\t\tstore: store,\n\t}, nil\n}", "func New(path string, opts *opt.Options) (*Store, error) {\n\tvar (\n\t\terr error\n\t\tdb *leveldb.DB\n\t)\n\n\tif path == \"\" {\n\t\tdb, err = leveldb.Open(ldbStorage.NewMemStorage(), opts)\n\t} else {\n\t\tdb, err = leveldb.OpenFile(path, opts)\n\t\tif ldbErrors.IsCorrupted(err) && !opts.GetReadOnly() {\n\t\t\tdb, err = leveldb.RecoverFile(path, opts)\n\t\t}\n\t}\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Store{\n\t\tdb: db,\n\t\tpath: path,\n\t}, nil\n}", "func New(options Options) *Store {\n\t// Set default options\n\tif options.Codec == nil {\n\t\toptions.Codec = DefaultOptions.Codec\n\t}\n\n\ts := Store{\n\t\tm: make(map[string]*Item),\n\t\tlock: new(sync.RWMutex),\n\t\tcodec: options.Codec,\n\t}\n\n\tgo s.autoGC(options.Interval)\n\n\treturn &s\n}", "func NewStore(dbName string) *Store {\n\tdb, err := bolt.Open(dbName, 0600, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn &Store{db: db}\n}", "func NewStore(c *Config) *Store {\n\t// create a new store\n\tstore := Store{}\n\tstore.config = c\n\tstore.DBname = c.DBname\n\t// check if the file exists\n\tvar build bool\n\t_, err := os.Stat(c.DBname)\n\tif err != nil {\n\t\tlogger.Critical(\"error on stat , %s\", err)\n\t\tbuild = true\n\t}\n\t// if it is a new file build some tables\n\tif build {\n\t\tstore.Build(c)\n\t}\n\tstore.leases = Load(c.DBname)\n\treturn &store\n}", "func NewStore() *Store {\n\treturn &Store{}\n}", "func NewStore() *Store {\n\treturn &Store{}\n}", "func NewStore() *Store {\n\treturn &Store{}\n}", "func NewStore(schema Schema, options ...CeousOption) *BaseStore {\n\tstore := &BaseStore{\n\t\tschema: schema,\n\t}\n\tfor _, option := range options {\n\t\toption(store)\n\t}\n\tif !store.disableCache {\n\t\tstore.runner = store._runner\n\t} else {\n\t\tstore.runner = sq.NewStmtCacher(store._runner)\n\t}\n\treturn store\n}", "func NewStore(ctx context.Context, l log.Logger, db *sqlx.DB, beaconName string) (*Store, error) {\n\tp := Store{\n\t\tlog: l,\n\t\tdb: db,\n\n\t\trequiresPrevious: chain.PreviousRequiredFromContext(ctx),\n\t}\n\n\tid, err := p.AddBeaconID(ctx, beaconName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tp.beaconID = id\n\n\treturn &p, nil\n}", "func New() *LocalStore {\n\treturn &LocalStore{}\n}", "func New(observationCtx *observation.Context, db database.DB) Store {\n\treturn &store{\n\t\tlogger: logger.Scoped(\"uploads.store\", \"\"),\n\t\tdb: basestore.NewWithHandle(db.Handle()),\n\t\toperations: newOperations(observationCtx),\n\t}\n}", "func New(client *ifirestore.Client) *StoreImpl {\n\treturn &StoreImpl{\n\t\tclient: client,\n\t}\n}", "func NewSQLStore(dsn string) (*MySQLStore, error) {\n\tdb, err := sql.Open(\"mysql\", dsn)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error opening database: %v\", err)\n\t}\n\tif err := db.Ping(); err != nil {\n\t\treturn nil, fmt.Errorf(\"error pinging database: %v\", err)\n\t}\n\tstore := &MySQLStore{db: db}\n\treturn store, nil\n}", "func NewStore() *Store {\n\treturn &Store{\n\t\tRaftDir: \"\",\n\t\tRaftBindAddr: \"\",\n\t\tdata: make(map[string]string),\n\t}\n}", "func New(db *db.DB) core.StageStore {\n\treturn &stageStore{db}\n}", "func NewStore() (*gorm.DB, sqlmock.Sqlmock, error) {\n\tsqldb, mock, err := sqlmock.New()\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tdb, err := gorm.Open(postgres.New(postgres.Config{\n\t\tConn: sqldb,\n\t}), &gorm.Config{})\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn db, mock, nil\n}", "func New(ctx context.Context, alias, path string) (*Store, error) {\n\tdebug.Log(\"Instantiating %q at %q\", alias, path)\n\n\ts := &Store{\n\t\talias: alias,\n\t\tpath: path,\n\t}\n\n\t// init storage and rcs backend\n\tif err := s.initStorageBackend(ctx); err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to init storage backend: %w\", err)\n\t}\n\n\tdebug.Log(\"Storage for %s => %s initialized as %v\", alias, path, s.storage)\n\n\t// init crypto backend\n\tif err := s.initCryptoBackend(ctx); err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to init crypto backend: %w\", err)\n\t}\n\n\tdebug.Log(\"Crypto for %s => %s initialized as %v\", alias, path, s.crypto)\n\n\treturn s, nil\n}", "func New(dsn string) (*SQLStorage, error) {\n\tdb, err := sqlx.Open(\"pgx\", dsn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &SQLStorage{db: db}, nil\n}", "func New(t *testing.T) *sql.DB {\n\tt.Helper()\n\n\tif testing.Short() {\n\t\tt.Skip(\"skip store test because short mode\")\n\t}\n\n\tdb, err := sql.Open(\"txdb\", t.Name())\n\tif err != nil {\n\t\tt.Fatalf(\"can't open db: %s\", err)\n\t}\n\n\tt.Cleanup(func() {\n\t\tdb.Close()\n\t})\n\n\treturn db\n}", "func NewStore(db *database.DB, collection string) Store {\n\treturn &store{db, collection}\n}", "func New(user, pass, dbName, host, port string) (*GoPGStore, error) {\n\tdb := pg.Connect(&pg.Options{\n\t\tUser: user,\n\t\tPassword: pass,\n\t\tDatabase: dbName,\n\t\tAddr: fmt.Sprintf(\"%s:%s\", host, port),\n\t})\n\n\treturn &GoPGStore{\n\t\tdb: db,\n\t}, nil\n}", "func New(db *db.DB) core.ProjectStore {\n\treturn &projectStore{db}\n}", "func New(db *sql.DB, tName string, d time.Duration) (*PgStore, error) {\n\tp := &PgStore{db: db, tName: tName, errChan: make(chan error)}\n\t_, err := p.db.Exec(fmt.Sprintf(table, p.tName))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif d > 0 {\n\t\tgo p.startCleanup(d)\n\t}\n\treturn p, nil\n}", "func New(db *bolt.DB, cleanupInterval time.Duration) *BoltStore {\n\tdb.Update(func(tx *bolt.Tx) error {\n\t\t_, err := tx.CreateBucketIfNotExists(dataBucketName)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = tx.CreateBucketIfNotExists(expiryBucketName)\n\t\treturn err\n\t})\n\tbs := &BoltStore{\n\t\tdb: db,\n\t}\n\tif cleanupInterval > 0 {\n\t\tgo bs.startCleanup(cleanupInterval)\n\t}\n\treturn bs\n}", "func New(path string, config *Config) (*Store, error) {\n\t// open the store\n\tf, err := os.OpenFile(path, os.O_RDWR|os.O_CREATE, 0755)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar dS = &Store{\n\t\tConfig: config,\n\t\tf: f,\n\t\trecords: map[string]interface{}{},\n\t}\n\n\tswitch config.StoreType {\n\tcase JSONStore:\n\t\tdS.Storer = JSONDataStore{}\n\tcase YAMLStore:\n\t\tdS.Storer = YAMLDataStore{}\n\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported store type: %v\", config.StoreType)\n\t}\n\n\t// read records into store\n\tif err := dS.Open(f, dS.records); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// configure name\n\tif config.StoreName == \"\" {\n\t\tconfig.StoreName = dS.f.Name()\n\t}\n\n\treturn dS, nil\n}", "func NewStore() *Store {\n\treturn &Store{\n\t\tstore: make(map[workloadmeta.Kind]map[string]workloadmeta.Entity),\n\t}\n}", "func New(channelID string, transientStore transientStore, collDataStore storeapi.Store) *pvtdatastore.Store {\n\treturn pvtdatastore.New(channelID, transientStore, collDataStore)\n}", "func NewStore(init StoreInit) (s *Store, err error) {\r\n\tvar initialCapacity int\r\n\tif utils.IsSet(init.InitCapacity) {\r\n\t\tinitialCapacity = init.InitCapacity\r\n\t} else {\r\n\t\tinitialCapacity = len(init.Args)\r\n\t}\r\n\r\n\tlocalStore := Store{\r\n\t\tstore: make(map[string]string, initialCapacity),\r\n\t\tparent: init.Parent,\r\n\t}\r\n\r\n\tfor _, arg := range init.Args {\r\n\t\tkey, value, err := splitArgument(arg)\r\n\t\tif err != nil {\r\n\t\t\treturn nil, err\r\n\t\t}\r\n\r\n\t\tif value, err = encode.ConvertStringToUtf8(value); err != nil {\r\n\t\t\treturn nil, fmt.Errorf(\"Error converting value for key '%v' to UTF-8: %v\", key, err)\r\n\t\t}\r\n\r\n\t\tif !localStore.hasKey(key) {\r\n\t\t\tlocalStore.Set(key, value)\r\n\t\t} else {\r\n\t\t\treturn nil, fmt.Errorf(\"Duplicate key '%v' found\", key)\r\n\t\t}\r\n\t}\r\n\r\n\treturn &localStore, nil\r\n}", "func NewStore() (*Store, error) {\n\tpgDB, err := db.Dial()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"store.Dial\")\n\t}\n\n\tlog.Println(\"Running PostgreSQL migrations...\")\n\tif err := db.RunPgMigrations(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"store.runPgMigrations\")\n\t}\n\n\tstore := &Store{\n\t\tDB: pgDB,\n\t\tArticles: repositories.NewArticlesRepository(pgDB),\n\t\tFeeds: repositories.NewFeedsRepository(pgDB),\n\t\tCategories: repositories.NewCategoryRepository(pgDB),\n\t\tSubcategories: repositories.NewSubcategoryRepository(pgDB),\n\t}\n\n\treturn store, nil\n}", "func newDBStore(db *leveldbhelper.DBHandle, dbName string) *store {\n\treturn &store{db, dbName}\n}", "func New(client dynamodbiface.ClientAPI) app.ConfigStore {\n\treturn &db{\n\t\tlog: logrus.StandardLogger().WithField(\"type\", \"app/dynamodb\"),\n\t\tdb: client,\n\t}\n}" ]
[ "0.8235283", "0.7783136", "0.76942456", "0.76699954", "0.7669162", "0.76145947", "0.7557565", "0.7553222", "0.7553222", "0.75039524", "0.7495782", "0.74797624", "0.7476449", "0.7475631", "0.74340194", "0.73144305", "0.7242601", "0.722981", "0.72063226", "0.7171092", "0.7145273", "0.7107777", "0.70779335", "0.70424294", "0.70413613", "0.7036974", "0.70087206", "0.6995882", "0.6984907", "0.69387084", "0.693718", "0.6926064", "0.69178283", "0.69165456", "0.6913564", "0.6897761", "0.6891296", "0.6871945", "0.68708384", "0.68696356", "0.6867638", "0.68673617", "0.68575084", "0.68559617", "0.68433625", "0.6842414", "0.6835606", "0.6834324", "0.68295157", "0.6826507", "0.6823833", "0.6822119", "0.68187857", "0.6812809", "0.6811637", "0.68063617", "0.67885697", "0.6766539", "0.6763366", "0.6746503", "0.67332655", "0.6731472", "0.6720115", "0.67018336", "0.6699403", "0.6692906", "0.66921675", "0.6688294", "0.66847146", "0.66754246", "0.66748774", "0.66723025", "0.66494805", "0.66466284", "0.66466284", "0.66466284", "0.66249186", "0.6623821", "0.66172993", "0.6592006", "0.65839034", "0.6556406", "0.65552694", "0.6550413", "0.6544546", "0.65211886", "0.6504514", "0.6503076", "0.6500932", "0.64846456", "0.64840615", "0.6480484", "0.6458152", "0.64540166", "0.6431707", "0.6430514", "0.64271086", "0.6420062", "0.64107174", "0.6410325" ]
0.7877457
1
get queries for a single row, building the sql, and writing the result into dest. Use this to simplify querying for a single row or column. Dest may be a pointer to a simple type, or a struct with fields to be populated from the returned columns.
func (s *SQLStore) getBuilder(q sqlx.Queryer, dest interface{}, b builder) error { sqlString, args, err := b.ToSql() if err != nil { return errors.Wrap(err, "failed to build sql") } sqlString = s.db.Rebind(sqlString) return sqlx.Get(q, dest, sqlString, args...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (db *DB) Get(ctx context.Context, dest interface{}, query string, args ...interface{}) error {\n\tr := db.QueryRow(ctx, query, args...)\n\treturn r.scanAny(dest, false)\n}", "func queryRow(ctx context.Context, conn driver.Conn, query string, dest ...*string) error {\n\tvar err error\n\n\t// prepare\n\tvar stmt driver.Stmt\n\tif connCtx, ok := conn.(driver.ConnPrepareContext); ok {\n\t\tstmt, err = connCtx.PrepareContext(ctx, query)\n\t} else {\n\t\tstmt, err = conn.Prepare(query)\n\t\tif err == nil {\n\t\t\tselect {\n\t\t\tdefault:\n\t\t\tcase <-ctx.Done():\n\t\t\t\tstmt.Close()\n\t\t\t\treturn ctx.Err()\n\t\t\t}\n\t\t}\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\t// execute query\n\tvar rows driver.Rows\n\tif queryCtx, ok := stmt.(driver.StmtQueryContext); ok {\n\t\trows, err = queryCtx.QueryContext(ctx, []driver.NamedValue{})\n\t} else {\n\t\tselect {\n\t\tdefault:\n\t\tcase <-ctx.Done():\n\t\t\treturn ctx.Err()\n\t\t}\n\t\trows, err = stmt.Query([]driver.Value{})\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer rows.Close()\n\n\t// scan\n\tif len(dest) != len(rows.Columns()) {\n\t\treturn fmt.Errorf(\"xraysql: expected %d destination arguments in Scan, not %d\", len(dest), len(rows.Columns()))\n\t}\n\tcols := make([]driver.Value, len(rows.Columns()))\n\tif err := rows.Next(cols); err != nil {\n\t\treturn err\n\t}\n\tfor i, src := range cols {\n\t\td := dest[i]\n\t\tswitch s := src.(type) {\n\t\tcase string:\n\t\t\t*d = s\n\t\tcase []byte:\n\t\t\t*d = string(s)\n\t\tcase time.Time:\n\t\t\t*d = s.Format(time.RFC3339Nano)\n\t\tcase int64:\n\t\t\t*d = strconv.FormatInt(s, 10)\n\t\tcase float64:\n\t\t\t*d = strconv.FormatFloat(s, 'g', -1, 64)\n\t\tcase bool:\n\t\t\t*d = strconv.FormatBool(s)\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"sql: Scan error on column index %d, name %q: type missmatch\", i, rows.Columns()[i])\n\t\t}\n\t}\n\n\treturn nil\n}", "func (d *Database) QueryRow(db DB, dst interface{}, query string, args ...interface{}) error {\n\t// perform the query\n\trows, err := db.Query(query, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// gather the result\n\treturn d.ScanRow(rows, dst)\n}", "func QueryRow(db DB, dst interface{}, query string, args ...interface{}) error {\n\treturn Default.QueryRow(db, dst, query, args...)\n}", "func (stmt *statement) QueryRow(ctx context.Context, db Executor) error {\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\trow := db.QueryRowContext(ctx, stmt.String(), stmt.args...)\n\n\treturn row.Scan(stmt.dest...)\n}", "func (db *DB) Select(dest interface{}, query string, args ...interface{}) error {\n return Select(db, dest, query, args...)\n}", "func (d *Database) QueryRow(db DB, dst interface{}, query string, args ...interface{}) error {\n\treturn d.QueryRowContext(context.Background(), db, dst, query, args...)\n}", "func findOne(m Model, wh Where, args ...interface{}) error {\n\tvar err error\n\t// trim spaces just in case (injection)\n\tfld, crit, val := wh.trimSpace()\n\tq := \"SELECT * FROM \" + m.TableName() + \" WHERE \" + fld + crit + \"? LIMIT 1\"\n\t// write data to dest\n\terr = DB.QueryRow(q, val).Scan(args...)\n\treturn err\n}", "func QueryRow(db DB, dst interface{}, query string, args ...interface{}) error {\n\treturn QueryRowContext(context.Background(), db, dst, query, args...)\n}", "func (this *Database) SingleQuery(sql string, args ...interface{}) *sql.Row {\n\treturn this.db.QueryRow(sql, args...)\n}", "func (this *Database) SingleQuery(sql string, args ...interface{}) *sql.Row {\n\treturn this.db.QueryRow(sql, args...)\n}", "func SelectStruct(dest interface{}, query string, args ...interface{}) (err error) {\n\tvar stmt *sql.Stmt\n\tvar rows *sql.Rows\n\tvar numRows int\n\n\tdefer func() {\n\t\tstmt.Close()\n\t\trows.Close()\n\t}()\n\n\tstmt = database.prepare(query)\n\n\tif rows, err = stmt.Query(args...); err != nil {\n\t\treturn\n\t}\n\tfor rows.Next() {\n\t\t// this is a hacky way to limit results to one row. Currently, the\n\t\t// sqlstruct library does not let you use the sql.QueryRow function,\n\t\t// which would be a better choice here. It likely won't ever support it.\n\t\t// See issue #5 in that repo for some details.\n\t\tif numRows > 1 {\n\t\t\treturn\n\t\t}\n\t\tif err = sqlstruct.Scan(dest, rows); err != nil {\n\t\t\treturn\n\t\t}\n\t\tnumRows++\n\t}\n\tif err = rows.Err(); err != nil {\n\t\treturn\n\t} else if numRows < 1 {\n\t\terr = errNoRows{sql.ErrNoRows}\n\t\treturn\n\t}\n\treturn\n}", "func (r *Result) One(arg interface{}) error {\n\t// Confirm that there is at least one row to return\n\tif ok := r.rows.Next(); !ok {\n\t\treturn ErrNoResult\n\t}\n\n\tcolumns, err := r.rows.Columns()\n\tif err != nil {\n\t\treturn fmt.Errorf(\n\t\t\t\"aspect: error returning columns from result: %s\",\n\t\t\terr,\n\t\t)\n\t}\n\n\tvalue := reflect.ValueOf(arg)\n\tif value.Kind() == reflect.Map {\n\t\tvalues, ok := arg.(Values)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"aspect: maps as destinations are only allowed if they are of type aspect.Values\")\n\t\t}\n\n\t\t// TODO scan directly into values?\n\t\taddr := make([]interface{}, len(columns))\n\t\tdest := make([]interface{}, len(columns))\n\t\tfor i, _ := range addr {\n\t\t\tdest[i] = &addr[i]\n\t\t}\n\n\t\tif err := r.rows.Scan(dest...); err != nil {\n\t\t\treturn fmt.Errorf(\"aspect: error while scanning map: %s\", err)\n\t\t}\n\n\t\tfor i, name := range columns {\n\t\t\tvalues[name] = addr[i]\n\t\t}\n\t\treturn r.rows.Err()\n\n\t} else if value.Kind() != reflect.Ptr {\n\t\treturn fmt.Errorf(\n\t\t\t\"aspect: received a non-pointer destination for result.One\",\n\t\t)\n\t}\n\n\t// Get the value of the given interface\n\telem := reflect.Indirect(value)\n\n\tswitch elem.Kind() {\n\tcase reflect.Struct:\n\t\t// Build the fields of the given struct\n\t\t// TODO this operation could be cached\n\t\tfields := SelectFields(arg)\n\n\t\t// Align the fields to the selected columns\n\t\t// This will discard unmatched fields\n\t\t// TODO strict mode? error if not all columns were matched?\n\t\taligned := AlignColumns(columns, fields)\n\n\t\t// If the aligned struct is empty, fallback to matching the fields in\n\t\t// order, but only if the length of the columns equals the fields\n\t\tif aligned.Empty() && len(columns) == len(fields) {\n\t\t\taligned = fields\n\t\t}\n\n\t\t// Get an interface for each field and save a pointer to it\n\t\tdest := make([]interface{}, len(aligned))\n\t\tfor i, field := range aligned {\n\t\t\t// If the field does not exist, the value will be discarded\n\t\t\tif !field.Exists() {\n\t\t\t\tdest[i] = &dest[i]\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\t// Recursively get an interface to the elem's fields\n\t\t\tvar fieldElem reflect.Value = elem\n\t\t\tfor _, index := range field.index {\n\t\t\t\tfieldElem = fieldElem.Field(index)\n\t\t\t}\n\t\t\tdest[i] = fieldElem.Addr().Interface()\n\t\t}\n\n\t\tif err := r.rows.Scan(dest...); err != nil {\n\t\t\treturn fmt.Errorf(\"aspect: error while scanning struct: %s\", err)\n\t\t}\n\n\tcase reflect.Slice:\n\t\treturn fmt.Errorf(\"aspect: cannot scan single results into slices\")\n\n\tdefault:\n\t\tif len(columns) != 1 {\n\t\t\treturn fmt.Errorf(\n\t\t\t\t\"aspect: unsupported destination for multi-column result: %s\",\n\t\t\t\telem.Kind(),\n\t\t\t)\n\t\t}\n\t\t// Attempt to scan directly into the elem\n\t\treturn r.rows.Scan(elem.Addr().Interface())\n\t}\n\treturn r.rows.Err()\n}", "func (this *dataStore) Read(queryName string, dest, arg interface{}) (error) {\r\n\r\n\tif stmt, err := this.namedStmt(queryName, dest); err != nil {\r\n\t\treturn err\r\n\t} else if stmt.query.Command != `select` {\r\n\t\treturn fmt.Errorf(`invalid SQL command for Read: %s`, stmt.query.Command)\r\n\t} else if stmt.query.MultiRow == true {\r\n\t\treturn stmt.Select(dest, arg)\r\n\t} else {\r\n\t\treturn stmt.Get(dest, arg)\r\n\t}\r\n}", "func (c *Cache) QueryRow(q string, args ...interface{}) (ret RowScanner) {\n\tif stmt, e := c.prep(q); e != nil {\n\t\tret = &prepError{e}\n\t} else {\n\t\tret = stmt.QueryRow(args...)\n\t}\n\treturn\n}", "func (c *Conn) QueryRow(sql string, args ...interface{}) *Row {\n\trows, _ := c.Query(sql, args...)\n\treturn (*Row)(rows)\n}", "func (db *DB) Get(dest interface{}, query string, args ...interface{}) error {\n return Get(db, dest, query, args...)\n}", "func (w *Wrapper) queryRow(query string, args ...interface{}) *sql.Row {\n\tw.connLock.RLock()\n\tdefer w.connLock.RUnlock()\n\n\treturn w.connection.QueryRow(w.prepare(query), args...)\n}", "func (r *rows) Next(dest []driver.Value) (err error) {\n\tif trace {\n\t\tdefer func() {\n\t\t\ttracer(r, \"Next(%v): %v\", dest, err)\n\t\t}()\n\t}\n\trc := r.rc0\n\tif r.doStep {\n\t\tif rc, err = r.step(r.pstmt); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tr.doStep = true\n\n\tswitch rc {\n\tcase bin.XSQLITE_ROW:\n\t\tif g, e := len(dest), len(r.columns); g != e {\n\t\t\treturn fmt.Errorf(\"Next(): have %v destination values, expected %v\", g, e)\n\t\t}\n\n\t\tfor i := range dest {\n\t\t\tct, err := r.columnType(i)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tswitch ct {\n\t\t\tcase bin.XSQLITE_INTEGER:\n\t\t\t\tv, err := r.columnInt64(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_FLOAT:\n\t\t\t\tv, err := r.columnDouble(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_TEXT:\n\t\t\t\tv, err := r.columnText(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_BLOB:\n\t\t\t\tv, err := r.columnBlob(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_NULL:\n\t\t\t\tdest[i] = nil\n\t\t\tdefault:\n\t\t\t\tpanic(\"internal error\")\n\t\t\t}\n\t\t}\n\t\treturn nil\n\tcase bin.XSQLITE_DONE:\n\t\treturn io.EOF\n\tdefault:\n\t\treturn r.errstr(int32(rc))\n\t}\n}", "func QueryOne(ctx context.Context, queryable Queryable, mapper RowMapper, query string, args ...interface{}) (interface{}, error) {\n\tvar err error\n\tstart := time.Now()\n\tdefer func(e error) {\n\t\tlatency := time.Since(start)\n\t\tzap.L().Info(\"queryOne\", zap.Int(\"latency\", int(latency.Seconds()*1000)), zap.Bool(\"success\", e == sql.ErrNoRows || e == nil), zap.String(\"activityId\", GetTraceID(ctx)))\n\t}(err)\n\trows, err := queryable.QueryContext(ctx, query, args...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer rows.Close()\n\tif !rows.Next() {\n\t\treturn nil, nil\n\t}\n\n\tobj, err := mapper.Map(rows)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn obj, nil\n}", "func QueryRow(db *sql.DB, sql string, obj interface{}) error {\n\tcns, err := columnGetter(sql)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// create reflect.Value\n\tv := reflect.Indirect(reflect.ValueOf(obj))\n\n\t// get tag mapping list\n\ttm, err := tagMappingGetter(cns, v)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// call scan\n\tcolumns := make([]interface{}, len(cns))\n\tfor i := 0; i < len(cns); i++ {\n\t\tvar str string\n\t\tcolumns[i] = &str\n\t}\n\n\terr = db.QueryRow(sql).Scan(columns...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor i, column := range columns {\n\t\tsubv := v.Field(tm[i])\n\t\tstr, ok := column.(*string)\n\t\tif !ok {\n\t\t\treturn errors.New(\"could not cast interface{} to *string type\")\n\t\t}\n\t\tsubv.SetString(*str)\n\t}\n\n\treturn nil\n}", "func (r rowsRes) Next(dest []driver.Value) error {\n\terr := r.my.ScanRow(r.row)\n\tif err != nil {\n\t\treturn errFilter(err)\n\t}\n\tfor i, col := range r.row {\n\t\tif col == nil {\n\t\t\tdest[i] = nil\n\t\t\tcontinue\n\t\t}\n\t\tswitch c := col.(type) {\n\t\tcase time.Time:\n\t\t\tdest[i] = c\n\t\t\tcontinue\n\t\tcase mysql.Timestamp:\n\t\t\tdest[i] = c.Time\n\t\t\tcontinue\n\t\tcase mysql.Date:\n\t\t\tdest[i] = c.Localtime()\n\t\t\tcontinue\n\t\t}\n\t\tv := reflect.ValueOf(col)\n\t\tswitch v.Kind() {\n\t\tcase reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\t\t// this contains time.Duration to\n\t\t\tdest[i] = v.Int()\n\t\tcase reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64:\n\t\t\tu := v.Uint()\n\t\t\tif u > math.MaxInt64 {\n\t\t\t\tpanic(\"Value to large for int64 type\")\n\t\t\t}\n\t\t\tdest[i] = int64(u)\n\t\tcase reflect.Float32, reflect.Float64:\n\t\t\tdest[i] = v.Float()\n\t\tcase reflect.Slice:\n\t\t\tif v.Type().Elem().Kind() == reflect.Uint8 {\n\t\t\t\tdest[i] = v.Interface().([]byte)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tdefault:\n\t\t\tpanic(fmt.Sprint(\"Unknown type of column: \", v.Type()))\n\t\t}\n\t}\n\treturn nil\n}", "func (db *DB) Select(ctx context.Context, dest interface{}, query string, args ...interface{}) error {\n\trows, err := db.Query(ctx, query, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// if something happens here, we want to make sure the rows are Closed\n\tdefer rows.Close()\n\treturn scanAll(rows, dest, false)\n}", "func (d *Database) QueryRow(query string, args ...interface{}) *sql.Row {\n\treturn d.Conn.QueryRow(query, args...)\n}", "func (s SQLite) QueryRow(query string, args ...interface{}) (*sql.Row, error) {\n\tif s.DB == nil {\n\t\treturn nil, fmt.Errorf(\"db is not created\")\n\t}\n\treturn s.DB.QueryRow(query, args...), nil\n}", "func (h StmtHandle) QueryRow(ctx context.Context, args ...interface{}) *pgx.Row {\n\th.check()\n\tp := h.s.sr.mcp.Get()\n\tswitch h.s.sr.method {\n\tcase prepare:\n\t\treturn p.QueryRowEx(ctx, h.s.prepared.Name, nil /* options */, args...)\n\n\tcase noprepare:\n\t\treturn p.QueryRowEx(ctx, h.s.sql, nil /* options */, args...)\n\n\tcase simple:\n\t\treturn p.QueryRowEx(ctx, h.s.sql, simpleProtocolOpt, args...)\n\n\tdefault:\n\t\tpanic(\"invalid method\")\n\t}\n}", "func (m *PgSQL) QueryRow(query string, args ...interface{}) *sql.Row {\n\treturn m.Connection.QueryRow(query, args...)\n}", "func (res *Result) GetRow() (row *Row, err os.Error) {\n if res.FieldCount == 0 {\n // There is no fields in result (OK result)\n return\n }\n row, err = res.getRow()\n if err != nil {\n // Unlock if error\n res.db.unlock()\n } else if row == nil && res.Status & _SERVER_MORE_RESULTS_EXISTS == 0 {\n // Unlock if no more rows to read\n res.db.unreaded_rows = false\n res.db.unlock()\n }\n return\n}", "func (stmt *statement) Query(ctx context.Context, db Executor, handler func(rows *sql.Rows)) error {\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\n\t// Fetch rows\n\trows, err := db.QueryContext(ctx, stmt.String(), stmt.args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Iterate through rows of returned dataset\n\tfor rows.Next() {\n\t\tif len(stmt.dest) > 0 {\n\t\t\terr = rows.Scan(stmt.dest...)\n\t\t\tif err != nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\t// Call a callback function\n\t\thandler(rows)\n\t}\n\t// Check for errors during rows \"Close\".\n\t// This may be more important if multiple statements are executed\n\t// in a single batch and rows were written as well as read.\n\tif closeErr := rows.Close(); closeErr != nil {\n\t\treturn closeErr\n\t}\n\n\t// Check for row scan errors.\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Check for errors during row iteration.\n\treturn rows.Err()\n}", "func (c *Conn) QueryRow(ctx context.Context, sql string, args ...any) Row {\n\trows, _ := c.Query(ctx, sql, args...)\n\treturn (*connRow)(rows.(*baseRows))\n}", "func RowToQueryResult(row *sql.Row, colDefines []database.Column) (QueryResult, error) {\n\tcols := database.Columns(colDefines).Names()\n\tcolumns := make([]interface{}, len(cols))\n\tcolumnPointers := make([]interface{}, len(cols))\n\tfor i := range columns {\n\t\tcolumnPointers[i] = &columns[i]\n\t}\n\t// Scan the result into the column pointers...\n\tif err := row.Scan(columnPointers...); err != nil {\n\t\treturn nil, err\n\t}\n\n\trowData := makeRowDataSet(colDefines)\n\tfor i, colName := range cols {\n\t\tval := columnPointers[i].(*interface{})\n\t\trowData[colName] = ColData{Data: val, DataType: rowData[colName].DataType}\n\t}\n\n\treturn QueryResult(rowData), nil\n}", "func (r *result) Next(dest []driver.Value) error {\n\tif r.data == nil {\n\t\tif err := r.readNext(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tfor i := 0; i < len(r.columns); i++ {\n\t\tkey := r.columns[i]\n\t\tval := r.data[key]\n\t\tdest[i] = val\n\t}\n\tr.data = nil\n\tr.offset++\n\treturn nil\n}", "func duplicateRowQuery(\n\tsrcTbl catalog.TableDescriptor, columnIDs []descpb.ColumnID, pred string, limitResults bool,\n) (sql string, colNames []string, _ error) {\n\tcolNames, err := srcTbl.NamesForColumnIDs(columnIDs)\n\tif err != nil {\n\t\treturn \"\", nil, err\n\t}\n\n\tsrcCols := make([]string, len(colNames))\n\tfor i, n := range colNames {\n\t\tsrcCols[i] = tree.NameString(n)\n\t}\n\n\t// There will be an expression in the WHERE clause for each of the columns,\n\t// and possibly one for pred.\n\tsrcWhere := make([]string, 0, len(srcCols)+1)\n\tfor i := range srcCols {\n\t\tsrcWhere = append(srcWhere, fmt.Sprintf(\"%s IS NOT NULL\", srcCols[i]))\n\t}\n\n\t// Wrap the predicate in parentheses.\n\tif pred != \"\" {\n\t\tsrcWhere = append(srcWhere, fmt.Sprintf(\"(%s)\", pred))\n\t}\n\n\tlimit := \"\"\n\tif limitResults {\n\t\tlimit = \" LIMIT 1\"\n\t}\n\treturn fmt.Sprintf(\n\t\t`SELECT %[1]s FROM [%[2]d AS tbl] WHERE %[3]s GROUP BY %[1]s HAVING count(*) > 1 %[4]s`,\n\t\tstrings.Join(srcCols, \", \"), // 1\n\t\tsrcTbl.GetID(), // 2\n\t\tstrings.Join(srcWhere, \" AND \"), // 3\n\t\tlimit, // 4\n\t), colNames, nil\n}", "func (db *DB) FetchOne(sql string, v ...interface{}) (Row, error) {\n\trows, err := db.query(sql, 1, v...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif len(rows) > 0 {\n\t\treturn rows[0], nil\n\t}\n\treturn nil, nil\n}", "func (db *DB) QueryRowx(query string, args ...interface{}) *Row {\n rows, err := db.DB.Query(query, args...)\n return &Row{rows: rows, err: err, unsafe: db.unsafe, Mapper: db.Mapper}\n}", "func (b *bot) queryRow(query string, args ...interface{}) *sql.Row {\n\treturn b.DB.client.QueryRow(query, args...)\n}", "func (f *FieldValues) Row(idx int, dest []driver.Value) {\n\tcopy(dest, f.values[idx*f.cols:(idx+1)*f.cols])\n\n\tif f.lobCols == 0 {\n\t\treturn\n\t}\n\n\tfor i, descr := range f.descrs {\n\t\tcol := descr.col\n\t\twriter := dest[col].(lobWriter)\n\t\tf.writers[i] = writer\n\t\tdescr.w = writer\n\t\tdest[col] = lobReadDescrToPointer(descr)\n\t}\n\n\t// last descriptor triggers lob read\n\tf.descrs[f.lobCols-1].fn = func() error {\n\t\treturn f.s.readLobStream(f.writers)\n\t}\n}", "func FetchRow(rows *sql.Rows, dst interface{}) error {\n\tvar columns []string\n\tvar err error\n\n\tdstv := reflect.ValueOf(dst)\n\n\tif dstv.IsNil() || dstv.Kind() != reflect.Ptr {\n\t\treturn db.ErrExpectingPointer\n\t}\n\n\titem_v := dstv.Elem()\n\n\tif columns, err = rows.Columns(); err != nil {\n\t\treturn err\n\t}\n\n\treset(dst)\n\n\tnext := rows.Next()\n\n\tif next == false {\n\t\tif err = rows.Err(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn db.ErrNoMoreRows\n\t}\n\n\titem, err := fetchResult(item_v.Type(), rows, columns)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\titem_v.Set(reflect.Indirect(item))\n\n\treturn nil\n}", "func (r Result) One(obj interface{}) error {\n\t// There must be at least one row to return\n\tif ok := r.Next(); !ok {\n\t\treturn sql.ErrNoRows\n\t}\n\n\tcolumns, err := r.Columns()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"sol: error returning columns from result: %s\", err)\n\t}\n\n\t// Since maps are already pointers, they can be used as destinations\n\t// no matter what - as long as they are of type Values\n\tvalue := reflect.ValueOf(obj)\n\tswitch value.Kind() {\n\tcase reflect.Map:\n\t\treturn r.oneMap(columns, value, obj)\n\tcase reflect.Ptr:\n\t\t// Other types must be given as pointers to be valid destinations\n\t\telem := reflect.Indirect(value)\n\t\tswitch elem.Kind() {\n\t\tcase reflect.Struct:\n\t\t\treturn r.oneStruct(columns, elem, obj)\n\t\tdefault: // TODO enumerate types?\n\t\t\treturn r.oneNative(columns, elem)\n\t\t}\n\t}\n\n\treturn fmt.Errorf(\n\t\t\"sol: unsupported destination type %T for Result.One\", obj,\n\t)\n}", "func (s *Session) QueryRow(query string, args ...interface{}) *sql.Row {\n\tif s.tx != nil {\n\t\treturn s.tx.QueryRow(query, args...)\n\t}\n\tif s.useMaster {\n\t\treturn s.db.Master().QueryRow(query, args...)\n\t}\n\treturn s.db.Slave().QueryRow(query, args...)\n}", "func (extDb *Database) QueryRow(query string, args ...interface{}) (*sql.Rows, error) {\n\trows, err := extDb.Query(query, args...)\n\tif err == nil && rows.Next() {\n\t\treturn rows, nil\n\t}\n\t// close rows if some error happen and rows exist\n\tif rows != nil {\n\t\trows.Close()\n\t}\n\tif err == nil {\n\t\terr = sql.ErrNoRows\n\t}\n\treturn nil, err\n}", "func (s *Session) FindOne(dest interface{}) error {\n\ts.initStatemnt()\n\ts.Limit(1)\n\tscanner, err := NewScanner(dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif scanner.entityPointer.Kind() != reflect.Struct {\n\t\treturn FindOneExpectStruct\n\t}\n\tdefer scanner.Close()\n\tif s.statement.table == \"\" {\n\t\ts.statement.From(scanner.GetTableName())\n\t}\n\n\tif s.explainModel {\n\t\ts.explain()\n\t}\n\n\ts.initCtx()\n\tsql, args, err := s.statement.ToSQL()\n\tif err != nil {\n\t\treturn err\n\t}\n\ts.logger.Debugf(\"[Session FindOne] sql: %s, args: %v\", sql, args)\n\trows, err := s.QueryContext(s.ctx, sql, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tscanner.SetRows(rows)\n\treturn scanner.Convert()\n}", "func (c Client) QueryForRow(query string, mapper RowMapper, args ...interface{}) (interface{}, error) {\n\trow := c.db.QueryRow(query, args...)\n\tresult, err := mapper(row)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result, nil\n}", "func (rc *SQLiteRows) Next(dest []driver.Value) error {\n\tif rc.s.closed {\n\t\treturn io.EOF\n\t}\n\trc.s.mu.Lock()\n\tdefer rc.s.mu.Unlock()\n\trv := C.sqlite3_step(rc.s.s)\n\tif rv == C.SQLITE_DONE {\n\t\treturn io.EOF\n\t}\n\tif rv != C.SQLITE_ROW {\n\t\trv = C.sqlite3_reset(rc.s.s)\n\t\tif rv != C.SQLITE_OK {\n\t\t\treturn rc.s.c.lastError()\n\t\t}\n\t\treturn nil\n\t}\n\n\trc.declTypes()\n\n\tfor i := range dest {\n\t\tswitch C.sqlite3_column_type(rc.s.s, C.int(i)) {\n\t\tcase C.SQLITE_INTEGER:\n\t\t\tval := int64(C.sqlite3_column_int64(rc.s.s, C.int(i)))\n\t\t\tswitch rc.decltype[i] {\n\t\t\tcase columnTimestamp, columnDatetime, columnDate:\n\t\t\t\tvar t time.Time\n\t\t\t\t// Assume a millisecond unix timestamp if it's 13 digits -- too\n\t\t\t\t// large to be a reasonable timestamp in seconds.\n\t\t\t\tif val > 1e12 || val < -1e12 {\n\t\t\t\t\tval *= int64(time.Millisecond) // convert ms to nsec\n\t\t\t\t\tt = time.Unix(0, val)\n\t\t\t\t} else {\n\t\t\t\t\tt = time.Unix(val, 0)\n\t\t\t\t}\n\t\t\t\tt = t.UTC()\n\t\t\t\tif rc.s.c.tz != nil {\n\t\t\t\t\tt = t.In(rc.s.c.tz)\n\t\t\t\t}\n\t\t\t\tdest[i] = t\n\t\t\tcase \"boolean\":\n\t\t\t\tdest[i] = val > 0\n\t\t\tdefault:\n\t\t\t\tdest[i] = val\n\t\t\t}\n\t\tcase C.SQLITE_FLOAT:\n\t\t\tdest[i] = float64(C.sqlite3_column_double(rc.s.s, C.int(i)))\n\t\tcase C.SQLITE_BLOB:\n\t\t\tp := C.sqlite3_column_blob(rc.s.s, C.int(i))\n\t\t\tif p == nil {\n\t\t\t\tdest[i] = nil\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tn := int(C.sqlite3_column_bytes(rc.s.s, C.int(i)))\n\t\t\tswitch dest[i].(type) {\n\t\t\tdefault:\n\t\t\t\tslice := make([]byte, n)\n\t\t\t\tcopy(slice[:], (*[1 << 30]byte)(p)[0:n])\n\t\t\t\tdest[i] = slice\n\t\t\t}\n\t\tcase C.SQLITE_NULL:\n\t\t\tdest[i] = nil\n\t\tcase C.SQLITE_TEXT:\n\t\t\tvar err error\n\t\t\tvar timeVal time.Time\n\n\t\t\tn := int(C.sqlite3_column_bytes(rc.s.s, C.int(i)))\n\t\t\ts := C.GoStringN((*C.char)(unsafe.Pointer(C.sqlite3_column_text(rc.s.s, C.int(i)))), C.int(n))\n\n\t\t\tswitch rc.decltype[i] {\n\t\t\tcase columnTimestamp, columnDatetime, columnDate:\n\t\t\t\tvar t time.Time\n\t\t\t\ts = strings.TrimSuffix(s, \"Z\")\n\t\t\t\tfor _, format := range SQLiteTimestampFormats {\n\t\t\t\t\tif timeVal, err = time.ParseInLocation(format, s, time.UTC); err == nil {\n\t\t\t\t\t\tt = timeVal\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif err != nil {\n\t\t\t\t\t// The column is a time value, so return the zero time on parse failure.\n\t\t\t\t\tt = time.Time{}\n\t\t\t\t}\n\t\t\t\tif rc.s.c.tz != nil {\n\t\t\t\t\tt = t.In(rc.s.c.tz)\n\t\t\t\t}\n\t\t\t\tdest[i] = t\n\t\t\tdefault:\n\t\t\t\tdest[i] = []byte(s)\n\t\t\t}\n\n\t\t}\n\t}\n\treturn nil\n}", "func (r RowsImpl) Scan(dest ...interface{}) error {\n\treturn r.R.Scan(dest...)\n}", "func (db TestDB) QueryRow(query string, args ...interface{}) *sql.Row {\n\treturn db.testTx.QueryRow(query, args...)\n}", "func (q sourceQuery) One(exec boil.Executor) (*Source, error) {\n\to := &Source{}\n\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Bind(nil, exec, o)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"mdbmodels: failed to execute a one query for sources\")\n\t}\n\n\treturn o, nil\n}", "func QueryRow(query string, args ...interface{}) (*sql.Row, error) {\n\tdb, err := sql.Open(driver, conninfo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer db.Close()\n\n\trow := db.QueryRow(query, args...)\n\n\treturn row, nil\n}", "func (q sourceQuery) One() (*Source, error) {\n\to := &Source{}\n\n\tqueries.SetLimit(q.Query, 1)\n\n\terr := q.Bind(o)\n\tif err != nil {\n\t\tif errors.Cause(err) == sql.ErrNoRows {\n\t\t\treturn nil, sql.ErrNoRows\n\t\t}\n\t\treturn nil, errors.Wrap(err, \"mdbmdbmodels: failed to execute a one query for sources\")\n\t}\n\n\treturn o, nil\n}", "func (db *DB) QueryRow(name string, args ...interface{}) (*pgx.Row, error) {\n\tq, err := db.qm.getQuery(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn db.Pool.QueryRow(q.getSQL(), args...), nil\n}", "func (t JsonType) SQL(ctx *sql.Context, dest []byte, v interface{}) (sqltypes.Value, error) {\n\tif v == nil {\n\t\treturn sqltypes.NULL, nil\n\t}\n\n\t// Convert to jsonType\n\tjsVal, _, err := t.Convert(v)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\tjs := jsVal.(JSONValue)\n\n\ts, err := js.ToString(ctx)\n\tif err != nil {\n\t\treturn sqltypes.NULL, err\n\t}\n\n\tval := AppendAndSliceString(dest, s)\n\n\treturn sqltypes.MakeTrusted(sqltypes.TypeJSON, val), nil\n}", "func ScanRow(typeMap *pgtype.Map, fieldDescriptions []pgconn.FieldDescription, values [][]byte, dest ...any) error {\n\tif len(fieldDescriptions) != len(values) {\n\t\treturn fmt.Errorf(\"number of field descriptions must equal number of values, got %d and %d\", len(fieldDescriptions), len(values))\n\t}\n\tif len(fieldDescriptions) != len(dest) {\n\t\treturn fmt.Errorf(\"number of field descriptions must equal number of destinations, got %d and %d\", len(fieldDescriptions), len(dest))\n\t}\n\n\tfor i, d := range dest {\n\t\tif d == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\terr := typeMap.Scan(fieldDescriptions[i].DataTypeOID, fieldDescriptions[i].Format, values[i], d)\n\t\tif err != nil {\n\t\t\treturn ScanArgError{ColumnIndex: i, Err: err}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (ex *Execer) queryScalarFn(destinations []interface{}) error {\n\tfullSQL, args, blob, err := ex.cacheOrSQL()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif blob != nil {\n\t\terr = json.Unmarshal(blob, &destinations)\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\t\t// log it and fallthrough to let the query continue\n\t\tlogger.Warn(\"queryScalarFn.10: Could not unmarshal cache data. Continuing with query\")\n\t}\n\n\tdefer logExecutionTime(time.Now(), fullSQL, args)\n\t// Run the query:\n\tvar rows *sqlx.Rows\n\trows, err = ex.database.Queryx(fullSQL, args...)\n\tif err != nil {\n\t\treturn logSQLError(err, \"queryScalarFn.12: querying database\", fullSQL, args)\n\t}\n\n\tdefer rows.Close()\n\tif rows.Next() {\n\t\terr = rows.Scan(destinations...)\n\t\tif err != nil {\n\t\t\treturn logSQLError(err, \"queryScalarFn.14: scanning to destination\", fullSQL, args)\n\t\t}\n\t\tex.setCache(destinations, dtStruct)\n\t\treturn nil\n\t}\n\tif err := rows.Err(); err != nil {\n\t\treturn logSQLError(err, \"queryScalarFn.20: iterating through rows\", fullSQL, args)\n\t}\n\n\treturn dat.ErrNotFound\n}", "func (m *RowMock) Scan(dest ...interface{}) error {\n\tfor i, e := range m.data {\n\t\tif e != nil {\n\t\t\tswitch v := dest[i].(type) {\n\t\t\tcase *[]byte:\n\t\t\t\t*v = e.([]byte)\n\t\t\tcase *string:\n\t\t\t\t*v = e.(string)\n\t\t\tcase *bool:\n\t\t\t\t*v = e.(bool)\n\t\t\tcase *int64:\n\t\t\t\t*v = e.(int64)\n\t\t\t}\n\t\t}\n\t}\n\treturn m.err\n}", "func (dbi *DB) SelectRow(sqls string, arg ...string) (map[string]string, error) {\r\n\tif dbi.status == false {\r\n\t\treturn nil, errors.New(\"database was not opened.\")\r\n\t}\r\n\r\n\t// start select\r\n\tdbi.createOperation(\"DB_SELECT\")\r\n\tdbi.data.reqSql = sqls\r\n\tdbi.data.inVar = arg\r\n\t// data\r\n\tdbi.data.commPrepare()\r\n\t// communicate\r\n\tif dbi.data.comm() == false {\r\n\t\tmylog.Println(dbi.data.sqlCode, \":\", dbi.data.sqlErrm)\r\n\t\tdbi.Close()\r\n\t\treturn nil, errors.New(dbi.Sqlerrm)\r\n\t}\r\n\t// parse\r\n\tdbi.data.commParse()\r\n\tdbi.parseError()\r\n\tif dbi.Sqlcode != 0 {\r\n\t\treturn nil, errors.New(dbi.Sqlerrm)\r\n\t}\r\n\r\n\t// get column name list\r\n\tdbi.createOperation(\"DB_COLUMNS\")\r\n\t// data\r\n\tdbi.data.curId = -1\r\n\tdbi.data.commPrepare()\r\n\t// communicate\r\n\tif dbi.data.comm() == false {\r\n\t\tdbi.Close()\r\n\t\treturn nil, errors.New(dbi.Sqlerrm)\r\n\t}\r\n\t// parse\r\n\tdbi.data.commParse()\r\n\t// no need to change dbi.sqlcode\r\n\r\n\tcolvar := make(map[string]string)\r\n\tfor i := 0; i < len(dbi.data.colName); i++ {\r\n\t\tcolvar[strings.ToLower(dbi.data.colName[i])] = dbi.data.outVar[i]\r\n\t}\r\n\treturn colvar, nil\r\n}", "func (r *Row) Scan(dest ...interface{}) (err error) {\n\trows := (*Rows)(r)\n\n\tif rows.Err() != nil {\n\t\treturn rows.Err()\n\t}\n\n\tif !rows.Next() {\n\t\tif rows.Err() == nil {\n\t\t\treturn ErrNoRows\n\t\t}\n\t\treturn rows.Err()\n\t}\n\n\trows.Scan(dest...)\n\trows.Close()\n\treturn rows.Err()\n}", "func Query(col *tiedot.Col, q map[string]interface{}) (resp []byte, err error) {\n\tvar query interface{}\n\n\tquery = interface{}(q)\n\n\tqueryResult := make(map[int]struct{})\n\t// Do the query\n\ttiedot.EvalQuery(query, col, &queryResult)\n\treturn FlatResult(col, &queryResult)\n\n}", "func (db *DB) Select(funcName string, dest interface{}, query string, args ...interface{}) error {\n\n\terr := db.DB.Select(dest, query, args...)\n\treturn err\n}", "func JSONGetOne(table database.Table, w http.ResponseWriter, r *http.Request, sb *sqrl.SelectBuilder) {\n\ttable.Clear()\n\n\texSb := *sb\n\n\tvars := mux.Vars(r)\n\tfmt.Println(\"query\")\n\n\tvar val string\n\n\tif value, ok := vars[\"id\"]; ok {\n\t\tval = value\n\t} else if value, ok := vars[\"login\"]; ok {\n\t\tval = value\n\t}\n\n\ttype Item struct {\n\t\tData interface{} `json:\"data\"`\n\t}\n\n\tfmt.Println(val)\n\t//fmt.Println(table.GetPrimaryKey())\n\n\tquery, params, _ := exSb.Where(sqrl.Eq{table.GetPrimaryKey(): val}).ToSql()\n\tfmt.Println(query)\n\n\tif params != nil {\n\t\terrs := db.DB.Select(table.GetItems(), query, params[0])\n\t\tfmt.Println(errs)\n\t} else {\n\t\terrs := db.DB.Select(table.GetItems(), query)\n\t\tfmt.Println(errs)\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tjson.NewEncoder(w).Encode(Item{Data: table.GetItem()})\n}", "func (r *Rows) Next(dest []driver.Value) error {\n\ttypes, err := r.columnTypes(false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor i := range types {\n\t\tswitch types[i] {\n\t\tcase Integer:\n\t\t\tdest[i] = r.message.getInt64()\n\t\tcase Float:\n\t\t\tdest[i] = r.message.getFloat64()\n\t\tcase Blob:\n\t\t\tdest[i] = r.message.getBlob()\n\t\tcase Text:\n\t\t\tdest[i] = r.message.getString()\n\t\tcase Null:\n\t\t\tr.message.getUint64()\n\t\t\tdest[i] = nil\n\t\tcase UnixTime:\n\t\t\ttimestamp := time.Unix(r.message.getInt64(), 0)\n\t\t\tdest[i] = timestamp\n\t\tcase ISO8601:\n\t\t\tvalue := r.message.getString()\n\t\t\tif value == \"\" {\n\t\t\t\tdest[i] = nil\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tvar t time.Time\n\t\t\tvar timeVal time.Time\n\t\t\tvar err error\n\t\t\tvalue = strings.TrimSuffix(value, \"Z\")\n\t\t\tfor _, format := range iso8601Formats {\n\t\t\t\tif timeVal, err = time.ParseInLocation(format, value, time.UTC); err == nil {\n\t\t\t\t\tt = timeVal\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tdest[i] = t\n\t\tcase Boolean:\n\t\t\tdest[i] = r.message.getInt64() != 0\n\t\tdefault:\n\t\t\tpanic(\"unknown data type\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (db *DB) query(sql string, num int, v ...interface{}) (Rows, error) {\n\tvar st interface{}\n\tif len(v) > 0 {\n\t\tstValue := reflect.ValueOf(v[0])\n\t\tif reflect.TypeOf(stValue).Kind() == reflect.Struct {\n\t\t\tst = v[0]\n\t\t\tv = v[1:]\n\t\t}\n\t}\n\n\trows, err := db.Queryx(sql, v...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\n\tvar result Rows\n\tif st != nil {\n\t\tfor rows.Next() {\n\t\t\terr := rows.StructScan(st)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tresult = append(result, reflect.ValueOf(st).Elem().Interface())\n\t\t\tif num == 1 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t} else {\n\t\t// get column's name\n\t\tcolumns, err := rows.Columns()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tvar cLen = len(columns)\n\t\tvaluePtrs := make([]interface{}, cLen)\n\t\tvalues := make([][]byte, cLen)\n\t\tfor i := 0; i < cLen; i++ {\n\t\t\tvaluePtrs[i] = &values[i]\n\t\t}\n\t\tfor rows.Next() {\n\t\t\terr := rows.Scan(valuePtrs...)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\trow := make(MapRow, cLen)\n\t\t\tfor i, v := range columns {\n\t\t\t\trow[v] = string(values[i])\n\t\t\t}\n\n\t\t\tresult = append(result, row)\n\t\t\tif num == 1 {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn result, nil\n}", "func (w *Wrapper) Get(columns ...string) (err error) {\n\tw.query = w.buildSelect(columns...)\n\t_, err = w.runQuery()\n\tif err != nil {\n\t\treturn\n\t}\n\treturn\n}", "func (c *jsonTableCol) Next(obj interface{}, pass bool, ord int) (sql.Row, error) {\n\t// nested column should recurse\n\tif len(c.cols) != 0 {\n\t\tif c.data == nil {\n\t\t\tc.LoadData(obj)\n\t\t}\n\n\t\tvar innerObj interface{}\n\t\tif !c.finished {\n\t\t\tinnerObj = c.data[c.pos]\n\t\t}\n\n\t\tvar row sql.Row\n\t\tfor i, col := range c.cols {\n\t\t\tinnerPass := len(col.cols) != 0 && i != c.currSib\n\t\t\trowPart, err := col.Next(innerObj, pass || innerPass, c.pos+1)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\trow = append(row, rowPart...)\n\t\t}\n\n\t\tif pass {\n\t\t\treturn row, nil\n\t\t}\n\n\t\tif c.NextSibling() {\n\t\t\tfor _, col := range c.cols {\n\t\t\t\tcol.Reset()\n\t\t\t}\n\t\t\tc.pos++\n\t\t}\n\n\t\tif c.pos >= len(c.data) {\n\t\t\tc.finished = true\n\t\t}\n\n\t\treturn row, nil\n\t}\n\n\t// this should only apply to nested columns, maybe...\n\tif pass {\n\t\treturn sql.Row{nil}, nil\n\t}\n\n\t// FOR ORDINAL is a special case\n\tif c.opts != nil && c.opts.forOrd {\n\t\treturn sql.Row{ord}, nil\n\t}\n\n\t// TODO: cache this?\n\tval, err := jsonpath.JsonPathLookup(obj, c.path)\n\tif c.opts.exists {\n\t\tif err != nil {\n\t\t\treturn sql.Row{0}, nil\n\t\t} else {\n\t\t\treturn sql.Row{1}, nil\n\t\t}\n\t}\n\n\t// key error means empty\n\tif err != nil {\n\t\tif c.opts.errOnEmp {\n\t\t\treturn nil, fmt.Errorf(\"missing value for JSON_TABLE column '%s'\", c.opts.name)\n\t\t}\n\t\tval = c.opts.defEmpVal\n\t}\n\n\tval, _, err = c.opts.typ.Convert(val)\n\tif err != nil {\n\t\tif c.opts.errOnErr {\n\t\t\treturn nil, err\n\t\t}\n\t\tval, _, err = c.opts.typ.Convert(c.opts.defErrVal)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Base columns are always finished\n\tc.finished = true\n\treturn sql.Row{val}, nil\n}", "func (ex *Execer) queryObject(dest interface{}) error {\n\tblob, err := ex.queryJSON()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif blob != nil {\n\t\treturn json.Unmarshal(blob, dest)\n\t}\n\treturn nil\n}", "func (r *QueryResult) Scan(ctx context.Context, ptr interface{}) error {\n\tif r.Rows == nil {\n\t\treturn errors.New(\"result rows is nil\")\n\t}\n\tdefer r.Close()\n\n\tif !r.Rows.Next() {\n\t\tif r.Rows.Err() != nil {\n\t\t\treturn r.Rows.Err()\n\t\t}\n\n\t\treturn ErrNoRows\n\t}\n\n\tptrType := reflect.TypeOf(ptr)\n\n\t// make sure this is a pointer\n\tif err := verifyPtr(ptrType); err != nil {\n\t\treturn err\n\t}\n\n\tif !scanAsStruct(ptrType) {\n\t\treturn r.Rows.Scan(ptr)\n\t}\n\n\tcols, err := r.Rows.Columns()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbaseType := ptrType.Elem()\n\tfieldIdxs := indexes(baseType)\n\n\tv := reflect.ValueOf(ptr).Elem()\n\n\tvar vals []interface{}\n\tif ptrType.Implements(modelInterfaceType) {\n\t\tvals = modelVals(v, fieldIdxs, cols)\n\t} else {\n\t\tvals = structVals(v, cols)\n\t}\n\n\treturn r.Rows.Scan(vals...)\n}", "func (m *DBMock) QueryRow(ctx context.Context, query string, params ...interface{}) pgx.Row {\n\targs := m.Called(append([]interface{}{query}, params...)...)\n\trowMock := &RowMock{\n\t\terr: args.Error(1),\n\t}\n\tswitch v := args.Get(0).(type) {\n\tcase []interface{}:\n\t\trowMock.data = v\n\tcase interface{}:\n\t\trowMock.data = []interface{}{args.Get(0)}\n\t}\n\treturn rowMock\n}", "func scanRow(scanner db.Scanner, dest *core.Connection) error {\n\treturn scanner.Scan(\n\t\t&dest.ID,\n\t\t&dest.Name,\n\t\t&dest.PID,\n\t\t&dest.DataBase,\n\t\t&dest.Host,\n\t\t&dest.Port,\n\t\t&dest.User,\n\t\t&dest.Password,\n\t\t&dest.Description,\n\t\t&dest.Created,\n\t\t&dest.Updated,\n\t)\n}", "func SelectExists(dest interface{}, query string, args ...interface{}) (exists bool, err error) {\n\tstmt := database.prepare(query)\n\tdefer stmt.Close()\n\terr = stmt.QueryRow(args...).Scan(dest)\n\tif err == sql.ErrNoRows {\n\t\terr = nil // consider a non-error, means the row does not exist.\n\t\treturn\n\t}\n\tif err == nil {\n\t\texists = true\n\t}\n\treturn\n}", "func (r *QueryResult) Next(dest []driver.Value) error {\n\tif len(r.page.Columns) == 0 {\n\t\tr.close()\n\t\treturn io.EOF\n\t}\n\trowCount := int32(len(r.page.Columns[0]))\n\tif r.index >= rowCount {\n\t\tif r.page.Last {\n\t\t\tr.close()\n\t\t\treturn io.EOF\n\t\t}\n\t\tctx, cancel := r.contextWithCancel()\n\t\tdefer cancel()\n\t\tif err := r.fetchNextPage(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tfor i := 0; i < len(r.page.Columns); i++ {\n\t\tcol := r.page.Columns[i]\n\t\t// TODO: find out the reason for requiring the following fix\n\t\tif len(col) <= int(r.index) {\n\t\t\tr.close()\n\t\t\treturn io.EOF\n\t\t}\n\t\tdest[i] = col[r.index]\n\t}\n\tr.index++\n\treturn nil\n}", "func buildResultRow(e *base.Definition, columns []string) []interface{} {\n\n\tresults := make([]interface{}, len(columns))\n\ttimeType := reflect.ValueOf(time.Now())\n\tgocqlUUIDType := reflect.ValueOf(gocql.UUIDFromTime(time.Now()))\n\n\tfor i, column := range columns {\n\t\t// get the type of the field from the ColumnToType mapping for object\n\t\t// That we we can allocate appropriate memory for this field\n\t\ttyp := e.ColumnToType[column]\n\n\t\tswitch typ.Kind() {\n\t\tcase reflect.String:\n\t\t\tvar value *string\n\t\t\tresults[i] = &value\n\t\tcase reflect.Int32, reflect.Uint32, reflect.Int:\n\t\t\t// C* internally uses int and int64\n\t\t\tvar value *int\n\t\t\tresults[i] = &value\n\t\tcase reflect.Int64, reflect.Uint64:\n\t\t\t// C* internally uses int and int64\n\t\t\tvar value *int64\n\t\t\tresults[i] = &value\n\t\tcase reflect.Bool:\n\t\t\tvar value *bool\n\t\t\tresults[i] = &value\n\t\tcase reflect.Slice:\n\t\t\tvar value *[]byte\n\t\t\tresults[i] = &value\n\t\tcase timeType.Kind():\n\t\t\tvar value *time.Time\n\t\t\tresults[i] = &value\n\t\tcase gocqlUUIDType.Kind():\n\t\t\tvar value *gocql.UUID\n\t\t\tresults[i] = &value\n\t\tcase reflect.Ptr:\n\t\t\t// Special case for custom optional string type:\n\t\t\t// string type used in Cassandra\n\t\t\t// converted to/from custom type in ORM layer\n\t\t\tif typ == reflect.TypeOf(&base.OptionalString{}) {\n\t\t\t\tvar value *string\n\t\t\t\tresults[i] = &value\n\t\t\t\tbreak\n\t\t\t}\n\t\t\t// Special case for custom optional int type:\n\t\t\t// int64 type used in Cassandra\n\t\t\t// converted to/from custom type in ORM layer\n\t\t\tif typ == reflect.TypeOf(&base.OptionalUInt64{}) {\n\t\t\t\tvar value *int64\n\t\t\t\tresults[i] = &value\n\t\t\t\tbreak\n\t\t\t}\n\t\t\t// for unrecognized pointer types, fall back to default logging\n\t\t\tfallthrough\n\t\tdefault:\n\t\t\t// This should only happen if we start using a new cassandra type\n\t\t\t// without adding to the translation layer\n\t\t\tlog.WithFields(log.Fields{\"type\": typ.Kind(), \"column\": column}).\n\t\t\t\tInfof(\"type not found\")\n\t\t}\n\t}\n\n\treturn results\n}", "func (access ObjectAccess) Get(row int) interface{} {\n return access.rawData[access.indices[row]]\n}", "func SelectOne(db *sql.DB, table interface{}, args ...interface{}) (int, error) {\n\ttypes := reflect.TypeOf(table)\n\tif types.Kind() != reflect.Ptr || types.Elem().Kind() != reflect.Struct {\n\t\treturn 0, fmt.Errorf(\"you should give a ptr of a struct\")\n\t}\n\trltype := types.Elem()\n\tsqlcmd := \"select * from \" + rltype.Name()\n\tif len(args) > 0 {\n\t\tif reflect.TypeOf(args[0]).Kind() != reflect.String {\n\t\t\treturn 0, fmt.Errorf(\"args[0] should is a string\")\n\t\t}\n\t\tsqlcmd += \" \" + args[0].(string)\n\t}\n\tsqlcmd += \" limit 1\"\n\n\tvar (\n\t\trows *sql.Rows\n\t\terr error\n\t)\n\tif len(args) > 1 {\n\t\trows, err = db.Query(sqlcmd, args[1:]...)\n\t} else {\n\t\trows, err = db.Query(sqlcmd)\n\t}\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer rows.Close()\n\n\tcols, er := readRows(rows, rltype)\n\tif len(cols) > 0 {\n\t\treflect.ValueOf(table).Elem().Set(reflect.ValueOf(cols[0]).Elem())\n\t}\n\n\treturn len(cols), er\n}", "func (q Query) Single(ctx Context) (Record, bool, error) {\n\tnext := q.Iterate()\n\titem, err := next(ctx)\n\tif err != nil {\n\t\tif IsNoRows(err) {\n\t\t\terr = nil\n\t\t}\n\t\treturn Record{}, false, err\n\t}\n\n\t_, err = next(ctx)\n\tif err == nil {\n\t\treturn Record{}, false, nil\n\t}\n\tif !IsNoRows(err) {\n\t\treturn Record{}, false, err\n\t}\n\n\treturn item, true, nil\n}", "func (c *Conn) OneValue(query string, value interface{}, args ...interface{}) error {\n\ts, err := c.Prepare(query, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer s.Finalize()\n\tb, err := s.Next()\n\tif err != nil {\n\t\treturn err\n\t} else if !b {\n\t\tif s.ColumnCount() == 0 {\n\t\t\treturn s.specificError(\"don't use OneValue with query that returns no data such as %q\", query)\n\t\t}\n\t\treturn io.EOF\n\t}\n\treturn s.Scan(value)\n}", "func (m *Manager) Query(stmt sol.Executable, dest ...interface{}) error {\n\t// We only care about events that change things\n\t// If the statement has a conditional, apply additional selections\n\t// TODO switches can't handle new statement types\n\tswitch t := stmt.(type) {\n\tcase sol.DeleteStmt:\n\t\tfor _, clause := range m.onDelete {\n\t\t\tt.AddConditional(clause)\n\t\t}\n\t\tstmt = t\n\tcase sol.InsertStmt, postgres.InsertStmt:\n\t\t// Do nothing\n\tcase sol.UpdateStmt:\n\t\tfor _, clause := range m.onUpdate {\n\t\t\tt.AddConditional(clause)\n\t\t}\n\t\tstmt = t\n\tcase sol.SelectStmt:\n\t\tfor _, clause := range m.onSelect {\n\t\t\tt.AddConditional(clause)\n\t\t}\n\t\tstmt = t\n\t}\n\treturn m.conn.Query(stmt, dest...)\n}", "func (db *DB) QueryRow(query string, args ...interface{}) (rows *sql.Row) {\n\t_, readReplica := db.readReplicaRR()\n\treturn readReplica.QueryRow(query, args...)\n}", "func (dev) Query(w io.Writer) {\n\tl.destMu.Lock()\n\t{\n\t\tl.dest[DevQuery] = w\n\t}\n\tl.destMu.Unlock()\n}", "func (ex *Execer) queryJSONStruct(dest interface{}) error {\n\tblob, err := ex.queryJSONBlob(true)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif blob != nil {\n\t\treturn json.Unmarshal(blob, dest)\n\t}\n\treturn nil\n}", "func (rows *Rows) Scan(dest ...interface{}) (err error) {\n\tif len(rows.fields) != len(dest) {\n\t\terr = fmt.Errorf(\"Scan received wrong number of arguments, got %d but expected %d\", len(dest), len(rows.fields))\n\t\trows.Fatal(err)\n\t\treturn err\n\t}\n\n\tfor i, d := range dest {\n\t\tvr, _ := rows.nextColumn()\n\n\t\tif d == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Check for []byte first as we allow sidestepping the decoding process and retrieving the raw bytes\n\t\tif b, ok := d.(*[]byte); ok {\n\t\t\t// If it actually is a bytea then pass it through decodeBytea (so it can be decoded if it is in text format)\n\t\t\t// Otherwise read the bytes directly regardless of what the actual type is.\n\t\t\tif vr.Type().DataType == ByteaOid {\n\t\t\t\t*b = decodeBytea(vr)\n\t\t\t} else {\n\t\t\t\tif vr.Len() != -1 {\n\t\t\t\t\t*b = vr.ReadBytes(vr.Len())\n\t\t\t\t} else {\n\t\t\t\t\t*b = nil\n\t\t\t\t}\n\t\t\t}\n\t\t} else if s, ok := d.(Scanner); ok {\n\t\t\terr = s.Scan(vr)\n\t\t\tif err != nil {\n\t\t\t\trows.Fatal(scanArgError{col: i, err: err})\n\t\t\t}\n\t\t} else if s, ok := d.(sql.Scanner); ok {\n\t\t\tvar val interface{}\n\t\t\tif 0 <= vr.Len() {\n\t\t\t\tswitch vr.Type().DataType {\n\t\t\t\tcase BoolOid:\n\t\t\t\t\tval = decodeBool(vr)\n\t\t\t\tcase Int8Oid:\n\t\t\t\t\tval = int64(decodeInt8(vr))\n\t\t\t\tcase Int2Oid:\n\t\t\t\t\tval = int64(decodeInt2(vr))\n\t\t\t\tcase Int4Oid:\n\t\t\t\t\tval = int64(decodeInt4(vr))\n\t\t\t\tcase TextOid, VarcharOid:\n\t\t\t\t\tval = decodeText(vr)\n\t\t\t\tcase OidOid:\n\t\t\t\t\tval = int64(decodeOid(vr))\n\t\t\t\tcase Float4Oid:\n\t\t\t\t\tval = float64(decodeFloat4(vr))\n\t\t\t\tcase Float8Oid:\n\t\t\t\t\tval = decodeFloat8(vr)\n\t\t\t\tcase DateOid:\n\t\t\t\t\tval = decodeDate(vr)\n\t\t\t\tcase TimestampOid:\n\t\t\t\t\tval = decodeTimestamp(vr)\n\t\t\t\tcase TimestampTzOid:\n\t\t\t\t\tval = decodeTimestampTz(vr)\n\t\t\t\tdefault:\n\t\t\t\t\tval = vr.ReadBytes(vr.Len())\n\t\t\t\t}\n\t\t\t}\n\t\t\terr = s.Scan(val)\n\t\t\tif err != nil {\n\t\t\t\trows.Fatal(scanArgError{col: i, err: err})\n\t\t\t}\n\t\t} else if vr.Type().DataType == JsonOid || vr.Type().DataType == JsonbOid {\n\t\t\t// Because the argument passed to decodeJSON will escape the heap.\n\t\t\t// This allows d to be stack allocated and only copied to the heap when\n\t\t\t// we actually are decoding JSON. This saves one memory allocation per\n\t\t\t// row.\n\t\t\td2 := d\n\t\t\tdecodeJSON(vr, &d2)\n\t\t} else {\n\t\t\tif err := Decode(vr, d); err != nil {\n\t\t\t\trows.Fatal(scanArgError{col: i, err: err})\n\t\t\t}\n\t\t}\n\t\tif vr.Err() != nil {\n\t\t\trows.Fatal(scanArgError{col: i, err: vr.Err()})\n\t\t}\n\n\t\tif rows.Err() != nil {\n\t\t\treturn rows.Err()\n\t\t}\n\t}\n\n\treturn nil\n}", "func QueryReturn(query string, db *sql.DB, arg ...interface{}) (bool, string) {\n\trow := db.QueryRow(query, arg...)\n\n\tvar rowReturned string\n\terr := row.Scan(&rowReturned)\n\tif err == sql.ErrNoRows {\n\t\treturn false, \"\"\n\t} else {\n\t\tCheck(err)\n\t\treturn true, rowReturned\n\t}\n}", "func (db *dberWithContext) QueryRow(query string, args ...interface{}) *sql.Row {\n\treturn db.MySQLExec.QueryRowContext(db.ctx, query, args...)\n}", "func (ex *Execer) queryStructFn(dest interface{}) error {\n\tfullSQL, args, blob, err := ex.cacheOrSQL()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif blob != nil {\n\t\terr = json.Unmarshal(blob, &dest)\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\t\t// log it and fallthrough to let the query continue\n\t\tlogger.Warn(\"queryStruct.2: Could not unmarshal queryStruct cache data. Continuing with query\")\n\t}\n\n\tdefer logExecutionTime(time.Now(), fullSQL, args)\n\terr = ex.database.Get(dest, fullSQL, args...)\n\tif err != nil {\n\t\treturn logSQLError(err, \"queryStruct.3\", fullSQL, args)\n\t}\n\n\tex.setCache(dest, dtStruct)\n\treturn nil\n}", "func (qr *queryResult) Next(dest []driver.Value) error {\n\tif qr.pos >= qr.numRow() {\n\t\tif qr.attributes.LastPacket() {\n\t\t\treturn io.EOF\n\t\t}\n\t\tif err := qr.conn._fetchNext(context.Background(), qr); err != nil {\n\t\t\tqr.lastErr = err //fieldValues and attrs are nil\n\t\t\treturn err\n\t\t}\n\t\tif qr.numRow() == 0 {\n\t\t\treturn io.EOF\n\t\t}\n\t\tqr.pos = 0\n\t}\n\n\tqr.copyRow(qr.pos, dest)\n\terr := qr.decodeErrors.RowError(qr.pos)\n\tqr.pos++\n\n\tfor _, v := range dest {\n\t\tif v, ok := v.(p.LobDecoderSetter); ok {\n\t\t\tv.SetDecoder(qr.conn.decodeLob)\n\t\t}\n\t}\n\treturn err\n}", "func (r *Rows) Next(dest []driver.Value) error {\nagain:\n\trow, err := r.rows.Next(r.ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(row) == 0 {\n\t\treturn nil\n\t}\n\n\tif _, ok := row[0].(types.OkResult); ok {\n\t\t// skip OK results\n\t\tgoto again\n\t}\n\n\tfor i := range row {\n\t\tdest[i] = r.convert(i, row[i])\n\t}\n\treturn nil\n}", "func (q *OperationsQ) Select(dest interface{}) error {\n\tif q.Err != nil {\n\t\treturn q.Err\n\t}\n\n\tq.Err = q.parent.Select(dest, q.sql)\n\treturn q.Err\n}", "func (qr *queryResult) Next(dest []driver.Value) error {\n\tif qr.pos >= qr.numRow() {\n\t\tif qr.attributes.LastPacket() {\n\t\t\treturn io.EOF\n\t\t}\n\t\tif err := qr.conn._fetchNext(qr); err != nil {\n\t\t\tqr.lastErr = err //fieldValues and attrs are nil\n\t\t\treturn err\n\t\t}\n\t\tif qr.numRow() == 0 {\n\t\t\treturn io.EOF\n\t\t}\n\t\tqr.pos = 0\n\t}\n\n\tqr.copyRow(qr.pos, dest)\n\terr := qr.decodeErrors.RowError(qr.pos)\n\tqr.pos++\n\n\tfor _, v := range dest {\n\t\tif v, ok := v.(p.LobDecoderSetter); ok {\n\t\t\tv.SetDecoder(qr.conn.decodeLob)\n\t\t}\n\t}\n\treturn err\n}", "func (db *DB) Select(dest interface{}, query string, args ...interface{}) error {\n\treturn db.follower.Select(dest, query, args...)\n}", "func (db *DB) doSelectOrWithReturning(query string, arguments []interface{}, recordDescription *recordDescription, pointersGetter pointersGetter) (int64, error) {\n\trows, columns, err := db.executeQuery(query, arguments, false, false)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer rows.Close()\n\n\t// If the given slice is empty, the slice grows as the rows are read.\n\t// If the given slice isn't empty it's filled with rows, and both rows and\n\t// slice length have to be equals.\n\t// If it's a single instance, it's juste filled, and the result must have\n\t// only one row.\n\tvar rowsCount int\n\tif recordDescription.len() > 0 {\n\t\trowsCount, err = db.fillWithValues(recordDescription, pointersGetter, columns, rows)\n\t} else {\n\t\trowsCount, err = db.growAndFillWithValues(recordDescription, pointersGetter, columns, rows)\n\t}\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn 0, err\n\t}\n\n\terr = rows.Err()\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t}\n\treturn int64(rowsCount), err\n}", "func (r *Rows) one(a interface{}) error {\n\tdefer r.Close()\n\n\tif !r.Next() {\n\t\tif err := r.Err(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn sql.ErrNoRows\n\t}\n\n\tvar err error\n\n\trt := reflect.TypeOf(a)\n\tif rt.Kind() == reflect.Ptr && rt.Elem().Kind() == reflect.Map {\n\t\t// pointer to map\n\t\tv := indirect(reflect.ValueOf(a))\n\t\tif v.IsNil() {\n\t\t\tv.Set(reflect.MakeMap(v.Type()))\n\t\t}\n\t\ta = v.Interface()\n\t\trt = reflect.TypeOf(a)\n\t}\n\n\tif rt.Kind() == reflect.Map {\n\t\tv, ok := a.(NullStringMap)\n\t\tif !ok {\n\t\t\treturn VarTypeError(\"must be a NullStringMap\")\n\t\t}\n\t\tif v == nil {\n\t\t\treturn VarTypeError(\"NullStringMap is nil\")\n\t\t}\n\t\terr = r.ScanMap(v)\n\t} else {\n\t\terr = r.ScanStruct(a)\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn r.Close()\n}", "func (my *MySQL) Query(sql string, params ...interface{}) (\n rows []*Row, res *Result, err os.Error) {\n\n res, err = my.Start(sql, params...)\n if err != nil {\n return\n }\n // Read rows\n var row *Row\n for {\n row, err = res.GetRow()\n if err != nil || row == nil {\n break\n }\n rows = append(rows, row)\n }\n return\n}", "func executeQueryWithRowResponse(db *sql.DB, query string, args ...interface{}) *sql.Row {\n\treturn db.QueryRow(query, args...)\n}", "func executeQueryWithRowResponse(db *sql.DB, query string, args ...interface{}) *sql.Row {\n\treturn db.QueryRow(query, args...)\n}", "func RowTo[T any](row CollectableRow) (T, error) {\n\tvar value T\n\terr := row.Scan(&value)\n\treturn value, err\n}", "func (t CustomerTable) Get(\n\tdb DBi,\n\tCustomerID string,\n) (\n\trow Customer,\n\terr error,\n) {\n\tsrc := db.QueryRow(getQuery_Customer,\n\t\tCustomerID)\n\n\treturn t.Scan(src)\n}", "func (driver *SQLDriver) Get(id string) (*pastes.Paste, error) {\n\t// Execute a SELECT query to retrieve the paste\n\trow := driver.database.QueryRow(\"SELECT * FROM ? WHERE id = ?\", driver.table, id)\n\terr := row.Err()\n\tif err != nil {\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn nil, nil\n\t\t}\n\t\treturn nil, err\n\t}\n\n\t// Scan the row into a paste and return it\n\tpaste := new(pastes.Paste)\n\terr = row.Scan(&paste)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn paste, nil\n}", "func (conn *db) QueryMapRow(query string, mapper MapMapper, args ...any) (rowsReturned int, err error) {\n\treturn conn.runMapRow(builder.NewQuery(query, args), mapper)\n}", "func (r *Resultset) GetValue(row, column int) (interface{}, error) {\n\tif row >= len(r.Values) || row < 0 {\n\t\treturn nil, fmt.Errorf(\"invalid row index %d\", row)\n\t}\n\n\tif column >= len(r.Fields) || column < 0 {\n\t\treturn nil, fmt.Errorf(\"invalid column index %d\", column)\n\t}\n\n\treturn r.Values[row][column], nil\n}", "func (r *rows) Next(dest []driver.Value) error {\n\tif r.index >= len(r.rows) {\n\t\tif r.parent.offset < r.parent.total {\n\t\t\tif err := fetchNextPage(r.parent.conn, r.parent.jobid, r.parent.offset, r.parent.total, r.parent); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tr.index = r.parent.rows.index\n\t\t\tr.rows = r.parent.rows.rows\n\t\t} else {\n\t\t\tr.rows = nil\n\t\t}\n\t}\n\tif len(dest) != len(r.parent.columns) {\n\t\treturn fmt.Errorf(\"invalid scan, expected %d arguments and received %d\", len(r.parent.columns), len(dest))\n\t}\n\tif r.rows == nil || len(r.rows) == 0 {\n\t\treturn sql.ErrNoRows\n\t}\n\ttherow := r.rows[r.index]\n\tfor i := 0; i < len(r.parent.columns); i++ {\n\t\tkey := r.parent.columns[i]\n\t\tval := therow[key]\n\t\tdest[i] = val\n\t}\n\tr.index++\n\treturn nil\n}", "func Do(query string, result interface{}, src Obj) error {\n\tfmt.Println(query)\n\ttree, err := sqlparser.Parse(query)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbase.Debug(pretty.Sprint(tree))\n\n\tswitch tree.(type) {\n\tcase sqlparser.SelectStatement:\n\t\terr := sel.Do(tree.(sqlparser.SelectStatement), result, base.Obj(src))\n\t\treturn err\n\t//case *sqlparser.Union:\n\t//\ttree.(*sqlparser.Union).\n\tdefault:\n\t\treturn fmt.Errorf(\"Query type not supported\")\n\t}\n}", "func (a ArticleDB) QueryOne(sess *sqlabble.Session, st stmt.Statement) (Article, error) {\n\tquery, values := sess.Builder.Build(st)\n\trows, err := sess.Query(query, values...)\n\tif err != nil {\n\t\treturn Article{}, err\n\t}\n\tms, err := a.Map(rows)\n\tif err != nil {\n\t\treturn Article{}, err\n\t}\n\tif len(ms) == 0 {\n\t\treturn Article{}, sqlabble.NewErrRecordNotFound(a.Table.Name)\n\t}\n\tif len(ms) > 1 {\n\t\treturn Article{}, sqlabble.NewErrFoundMultipleRecords(a.Table.Name)\n\t}\n\treturn ms[0], nil\n}", "func (r *Rows) row(a ...interface{}) error {\n\tdefer r.Close()\n\n\tfor _, dp := range a {\n\t\tif _, ok := dp.(*sql.RawBytes); ok {\n\t\t\treturn VarTypeError(\"RawBytes isn't allowed on Row()\")\n\t\t}\n\t}\n\n\tif !r.Next() {\n\t\tif err := r.Err(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn sql.ErrNoRows\n\t}\n\tif err := r.Scan(a...); err != nil {\n\t\treturn err\n\t}\n\n\treturn r.Close()\n}" ]
[ "0.66265637", "0.64222854", "0.61475146", "0.606679", "0.60028976", "0.59463704", "0.59340054", "0.59254646", "0.5910851", "0.5894946", "0.5894946", "0.57269907", "0.57042986", "0.5691884", "0.56900215", "0.56087774", "0.5585383", "0.5564866", "0.55224884", "0.55074376", "0.55050516", "0.54759324", "0.5473581", "0.54224366", "0.54160583", "0.5410885", "0.5396572", "0.53556997", "0.5332152", "0.53295463", "0.53260314", "0.5288303", "0.5246949", "0.52433383", "0.5233847", "0.5219939", "0.52187455", "0.5218344", "0.52174", "0.5200534", "0.51962143", "0.5195952", "0.5173656", "0.5132519", "0.51316583", "0.5116557", "0.5109946", "0.50983447", "0.50947046", "0.50614554", "0.5060173", "0.5056654", "0.50532037", "0.5026695", "0.50200343", "0.50186485", "0.49847555", "0.49691737", "0.4967628", "0.4963812", "0.49498117", "0.4947513", "0.49459624", "0.49457818", "0.49279764", "0.492093", "0.49146977", "0.49122974", "0.4890281", "0.48872662", "0.48838383", "0.4883549", "0.4877875", "0.48744643", "0.4873294", "0.48694918", "0.4846507", "0.48462185", "0.48432902", "0.48363972", "0.48268333", "0.48227704", "0.48211628", "0.479403", "0.47849256", "0.4781433", "0.47661084", "0.47591856", "0.47558483", "0.47469708", "0.47392783", "0.47392783", "0.47306943", "0.47274944", "0.47245732", "0.47236225", "0.47193074", "0.47084093", "0.46983907", "0.4690802", "0.46709692" ]
0.0
-1
selectBuilder queries for one or more rows, building the sql, and writing the result into dest. Use this to simplify querying for multiple rows (and possibly columns). Dest may be a slice of a simple, or a slice of a struct with fields to be populated from the returned columns.
func (s *SQLStore) selectBuilder(q sqlx.Queryer, dest interface{}, b builder) error { sqlString, args, err := b.ToSql() if err != nil { return errors.Wrap(err, "failed to build sql") } sqlString = s.db.Rebind(sqlString) return sqlx.Select(q, dest, sqlString, args...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (db *DB) Select(dest interface{}, query string, args ...interface{}) error {\n return Select(db, dest, query, args...)\n}", "func (db *DB) Select(ctx context.Context, dest interface{}, query string, args ...interface{}) error {\n\trows, err := db.Query(ctx, query, args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// if something happens here, we want to make sure the rows are Closed\n\tdefer rows.Close()\n\treturn scanAll(rows, dest, false)\n}", "func (sb *SQLBuilder) Select(cols ...string) *SQLBuilder {\n\tvar buf strings.Builder\n\n\tfor k, col := range cols {\n\n\t\tbuf.WriteString(col)\n\n\t\tif k != len(cols)-1 {\n\t\t\tbuf.WriteString(\",\")\n\t\t}\n\t}\n\n\tsb._select = buf.String()\n\n\treturn sb\n}", "func SelectStruct(dest interface{}, query string, args ...interface{}) (err error) {\n\tvar stmt *sql.Stmt\n\tvar rows *sql.Rows\n\tvar numRows int\n\n\tdefer func() {\n\t\tstmt.Close()\n\t\trows.Close()\n\t}()\n\n\tstmt = database.prepare(query)\n\n\tif rows, err = stmt.Query(args...); err != nil {\n\t\treturn\n\t}\n\tfor rows.Next() {\n\t\t// this is a hacky way to limit results to one row. Currently, the\n\t\t// sqlstruct library does not let you use the sql.QueryRow function,\n\t\t// which would be a better choice here. It likely won't ever support it.\n\t\t// See issue #5 in that repo for some details.\n\t\tif numRows > 1 {\n\t\t\treturn\n\t\t}\n\t\tif err = sqlstruct.Scan(dest, rows); err != nil {\n\t\t\treturn\n\t\t}\n\t\tnumRows++\n\t}\n\tif err = rows.Err(); err != nil {\n\t\treturn\n\t} else if numRows < 1 {\n\t\terr = errNoRows{sql.ErrNoRows}\n\t\treturn\n\t}\n\treturn\n}", "func Select(cols ...string) *Builder {\r\n\tbuilder := &Builder{cond: NewCond()}\r\n\treturn builder.Select(cols...)\r\n}", "func (b *SqliteBuilder) Select(cols ...string) *SelectQuery {\n\treturn NewSelectQuery(b, b.db).Select(cols...)\n}", "func Select(columns ...string) *SelectBuilder {\n\treturn NewSelectBuilder(columns...)\n}", "func (w *Wrapper) buildSelect(columns ...string) (query string) {\n\tbeforeOptions, _ := w.buildQueryOptions()\n\n\tif len(columns) == 0 {\n\t\tquery = fmt.Sprintf(\"SELECT %s* FROM %s \", beforeOptions, w.tableName[0])\n\t} else {\n\t\tquery = fmt.Sprintf(\"SELECT %s%s FROM %s \", beforeOptions, strings.Join(columns, \", \"), w.tableName[0])\n\t}\n\treturn\n}", "func (q *OperationsQ) Select(dest interface{}) error {\n\tif q.Err != nil {\n\t\treturn q.Err\n\t}\n\n\tq.Err = q.parent.Select(dest, q.sql)\n\treturn q.Err\n}", "func (db *DB) Select(funcName string, dest interface{}, query string, args ...interface{}) error {\n\n\terr := db.DB.Select(dest, query, args...)\n\treturn err\n}", "func (this *DbAdapter) Select(dest interface{}, query string, args ...interface{}) error {\n\tstarted := time.Now()\n\tdbMonitoring := this.Slave()\n\terr := dbMonitoring.Db.Select(dest, query, args...)\n\tdbMonitoring.addRequest(err, started)\n\tmysqlmon.ResponseTime.WithLabelValues(dbMonitoring.GetHost(), dbMonitoring.Conf.DBName, metrics.IsError(err), \"SELECT\").Observe(metrics.SinceMs(started))\n\treturn err\n\n}", "func (mySQL) BuildSelect(columns ...string) string {\n\treturn buildSelect(columns...)\n}", "func Select(mock sqlmock.Sqlmock, table string, columns []string, err error, values ...[]driver.Value) {\n\tsql := fmt.Sprintf(\"SELECT (.+) FROM %s\", table)\n\n\tif err != nil {\n\t\tmock.ExpectQuery(sql).WillReturnError(err)\n\t\treturn\n\t}\n\n\tif values == nil || len(values) == 0 {\n\t\tmock.ExpectQuery(sql).WillReturnRows(&sqlmock.Rows{})\n\t\treturn\n\t}\n\n\trows := sqlmock.NewRows(columns)\n\tfor _, value := range values {\n\t\trows.AddRow(value...)\n\t}\n\n\tmock.ExpectQuery(sql).WillReturnRows(rows)\n}", "func (db *DB) Select(dest interface{}, query string, args ...interface{}) error {\n\treturn db.follower.Select(dest, query, args...)\n}", "func Select(columns ...string) QueryMod {\n\treturn func(q *queries.Query) {\n\t\tqueries.AppendSelect(q, columns...)\n\t}\n}", "func (db *DB) Select(query interface{}, args ...interface{}) (tx *DB) {\n\ttx = db.getInstance()\n\n\tswitch v := query.(type) {\n\tcase []string:\n\t\ttx.Statement.Selects = v\n\n\t\tfor _, arg := range args {\n\t\t\tswitch arg := arg.(type) {\n\t\t\tcase string:\n\t\t\t\ttx.Statement.Selects = append(tx.Statement.Selects, arg)\n\t\t\tcase []string:\n\t\t\t\ttx.Statement.Selects = append(tx.Statement.Selects, arg...)\n\t\t\tdefault:\n\t\t\t\ttx.AddError(fmt.Errorf(\"unsupported select args %v %v\", query, args))\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\tdelete(tx.Statement.Clauses, \"SELECT\")\n\tcase string:\n\t\tif strings.Count(v, \"?\") >= len(args) && len(args) > 0 {\n\t\t\ttx.Statement.AddClause(clause.Select{\n\t\t\t\tDistinct: db.Statement.Distinct,\n\t\t\t\tExpression: clause.Expr{SQL: v, Vars: args},\n\t\t\t})\n\t\t} else if strings.Count(v, \"@\") > 0 && len(args) > 0 {\n\t\t\ttx.Statement.AddClause(clause.Select{\n\t\t\t\tDistinct: db.Statement.Distinct,\n\t\t\t\tExpression: clause.NamedExpr{SQL: v, Vars: args},\n\t\t\t})\n\t\t} else {\n\t\t\ttx.Statement.Selects = []string{v}\n\n\t\t\tfor _, arg := range args {\n\t\t\t\tswitch arg := arg.(type) {\n\t\t\t\tcase string:\n\t\t\t\t\ttx.Statement.Selects = append(tx.Statement.Selects, arg)\n\t\t\t\tcase []string:\n\t\t\t\t\ttx.Statement.Selects = append(tx.Statement.Selects, arg...)\n\t\t\t\tdefault:\n\t\t\t\t\ttx.Statement.AddClause(clause.Select{\n\t\t\t\t\t\tDistinct: db.Statement.Distinct,\n\t\t\t\t\t\tExpression: clause.Expr{SQL: v, Vars: args},\n\t\t\t\t\t})\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tdelete(tx.Statement.Clauses, \"SELECT\")\n\t\t}\n\tdefault:\n\t\ttx.AddError(fmt.Errorf(\"unsupported select args %v %v\", query, args))\n\t}\n\n\treturn\n}", "func Select(top uint64, distinct bool, columns []string, table string, where string, groupby []string, orderby []string, terminate bool) string {\n\n\tif table == \"\" {\n\t\treturn \"\"\n\t}\n\n\tvar stmt bytes.Buffer\n\n\tstmt.WriteString(`SELECT`)\n\n\t// top\n\tif top > 0 {\n\t\tstmt.WriteString(` TOP `)\n\t\tstmt.WriteString(strconv.FormatUint(top, 10))\n\t}\n\n\tif distinct {\n\t\tstmt.WriteString(` DISTINCT`)\n\t}\n\n\t// columns\n\tif len(columns) > 0 {\n\t\tstmt.WriteString(\" \" + strings.Join(columns, `, `))\n\t} else {\n\t\tstmt.WriteString(\" *\")\n\t}\n\n\t// table\n\tstmt.WriteString(` FROM ` + table)\n\n\t// where\n\tif where != \"\" {\n\t\tstmt.WriteString(` WHERE ` + where)\n\t}\n\n\t// Group by\n\tif len(groupby) > 0 {\n\t\tstmt.WriteString(\" GROUP BY \" + strings.Join(groupby, \", \") )\n\t}\n\n\t// order by\n\tif len(orderby) > 0 {\n\t\tstmt.WriteString(\" ORDER BY \" + strings.Join(orderby, \", \") )\n\t}\n\n\t// terminate statement\n\tif terminate {\n\t\tstmt.WriteString(\";\")\n\t}\n\n\treturn stmt.String()\n}", "func (t *Transaction) Select(dest interface{}, query string, args ...interface{}) error {\n\treturn hookedselect(t.dbmap, t, dest, query, args...)\n}", "func (db *Database) Select(cols ...string) *Database {\n\tdb.Query.Action = \"SELECT\"\n\tdb.Query.Columns = columns(cols)\n\n\treturn db\n}", "func (sd *SelectDataset) Select(selects ...interface{}) *SelectDataset {\n\tif len(selects) == 0 {\n\t\treturn sd.ClearSelect()\n\t}\n\treturn sd.copy(sd.clauses.SetSelect(exp.NewColumnListExpression(selects...)))\n}", "func Select(columns ...string) QueryOption {\n\treturn newFuncQueryOption(func(wrapper *QueryWrapper) {\n\t\twrapper.columns = columns\n\t\twrapper.queryLen += 2\n\t})\n}", "func (pb *primitiveBuilder) pushSelectRoutes(selectExprs sqlparser.SelectExprs, reservedVars *sqlparser.ReservedVars) ([]*resultColumn, error) {\n\tresultColumns := make([]*resultColumn, 0, len(selectExprs))\n\tfor _, node := range selectExprs {\n\t\tswitch node := node.(type) {\n\t\tcase *sqlparser.AliasedExpr:\n\t\t\tpullouts, origin, expr, err := pb.findOrigin(node.Expr, reservedVars)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tnode.Expr = expr\n\t\t\tnewBuilder, rc, _, err := planProjection(pb, pb.plan, node, origin)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tpb.plan = newBuilder\n\t\t\tresultColumns = append(resultColumns, rc)\n\t\t\tpb.addPullouts(pullouts)\n\t\tcase *sqlparser.StarExpr:\n\t\t\tvar expanded bool\n\t\t\tvar err error\n\t\t\tresultColumns, expanded, err = pb.expandStar(resultColumns, node)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tif expanded {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// We'll allow select * for simple routes.\n\t\t\trb, ok := pb.plan.(*route)\n\t\t\tif !ok {\n\t\t\t\treturn nil, errors.New(\"unsupported: '*' expression in cross-shard query\")\n\t\t\t}\n\t\t\t// Validate keyspace reference if any.\n\t\t\tif !node.TableName.IsEmpty() {\n\t\t\t\tif _, err := pb.st.FindTable(node.TableName); err != nil {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t}\n\t\t\tresultColumns = append(resultColumns, rb.PushAnonymous(node))\n\t\tcase *sqlparser.Nextval:\n\t\t\trb, ok := pb.plan.(*route)\n\t\t\tif !ok {\n\t\t\t\t// This code is unreachable because the parser doesn't allow joins for next val statements.\n\t\t\t\treturn nil, errors.New(\"unsupported: SELECT NEXT query in cross-shard query\")\n\t\t\t}\n\t\t\tif rb.eroute.Opcode != engine.Next {\n\t\t\t\treturn nil, errors.New(\"NEXT used on a non-sequence table\")\n\t\t\t}\n\t\t\trb.eroute.Opcode = engine.Next\n\t\t\tresultColumns = append(resultColumns, rb.PushAnonymous(node))\n\t\tdefault:\n\t\t\treturn nil, fmt.Errorf(\"BUG: unexpected select expression type: %T\", node)\n\t\t}\n\t}\n\treturn resultColumns, nil\n}", "func (b *Builder) Select(cols ...string) *Builder {\r\n\tb.selects = cols\r\n\tb.optype = selectType\r\n\treturn b\r\n}", "func Select(db *sql.DB, q Queryer, filters string) ([]Queryer, error) {\n\tquery := fmt.Sprintf(\"SELECT * FROM %s %s\", q.TableName(), filters)\n\trows, err := db.Query(query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tqs, err := receive(q, rows)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn qs, nil\n}", "func ExampleZSelectBuilder_ToPrepared() {\n\tsql, gen := q.Select().From(q.T(\"user\")).Where(\n\t\tq.Lte(q.C(\"id\"), 100),\n\t\tq.Lte(q.C(\"age\"), q.V(18, \"findAge\")),\n\t).ToPrepared()\n\n\t// // You can use by performing the following steps.\n\t// stmt, err := db.Prepare(sql)\n\t// if err != nil {\n\t// return err\n\t// }\n\t// defer stmt.Close()\n\t//\n\t// ab := gen()\n\t// ab.Set(\"findAge\", 24)\n\t// stmt.Query(ab.Args...)\n\n\tfmt.Println(\"SQL:\", sql)\n\n\t// build arguments\n\tab := gen()\n\tfmt.Println(\"Default Args:\", ab.Args)\n\tab.Set(\"findAge\", 24)\n\tfmt.Println(\"Modified Args:\", ab.Args)\n\t// You can also rewrite other values by using an index,\n\t// but there is a problem readability and weak to SQL change,\n\t// so it isn't recommended.\n\t// ab.Args[0] = 123\n\t// ab.Args[1] = 24\n\n\t// Output:\n\t// SQL: SELECT * FROM \"user\" WHERE (\"id\" <= ?)AND(\"age\" <= ?)\n\t// Default Args: [100 18]\n\t// Modified Args: [100 24]\n}", "func (q ReleaseQuery) Select(b sq.SelectBuilder) ReleaseQuery {\n\tq.builder = b\n\treturn q\n}", "func (b *JoinBuilder) Select(exprs ...interface{}) *SelectBuilder {\n\treturn makeSelectBuilder(b, exprs...)\n}", "func (t *ColumnsType) Select(f ...qb.Field) *qb.SelectBuilder {\n\treturn t.table.Select(f)\n}", "func pushSelectExprs(sel *sqlparser.Select, bldr builder) (builder, error) {\n\tresultColumns, err := pushSelectRoutes(sel.SelectExprs, bldr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbldr.Symtab().ResultColumns = resultColumns\n\n\tif err := pushGroupBy(sel, bldr); err != nil {\n\t\treturn nil, err\n\t}\n\treturn bldr, nil\n}", "func queryRow(ctx context.Context, conn driver.Conn, query string, dest ...*string) error {\n\tvar err error\n\n\t// prepare\n\tvar stmt driver.Stmt\n\tif connCtx, ok := conn.(driver.ConnPrepareContext); ok {\n\t\tstmt, err = connCtx.PrepareContext(ctx, query)\n\t} else {\n\t\tstmt, err = conn.Prepare(query)\n\t\tif err == nil {\n\t\t\tselect {\n\t\t\tdefault:\n\t\t\tcase <-ctx.Done():\n\t\t\t\tstmt.Close()\n\t\t\t\treturn ctx.Err()\n\t\t\t}\n\t\t}\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\t// execute query\n\tvar rows driver.Rows\n\tif queryCtx, ok := stmt.(driver.StmtQueryContext); ok {\n\t\trows, err = queryCtx.QueryContext(ctx, []driver.NamedValue{})\n\t} else {\n\t\tselect {\n\t\tdefault:\n\t\tcase <-ctx.Done():\n\t\t\treturn ctx.Err()\n\t\t}\n\t\trows, err = stmt.Query([]driver.Value{})\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer rows.Close()\n\n\t// scan\n\tif len(dest) != len(rows.Columns()) {\n\t\treturn fmt.Errorf(\"xraysql: expected %d destination arguments in Scan, not %d\", len(dest), len(rows.Columns()))\n\t}\n\tcols := make([]driver.Value, len(rows.Columns()))\n\tif err := rows.Next(cols); err != nil {\n\t\treturn err\n\t}\n\tfor i, src := range cols {\n\t\td := dest[i]\n\t\tswitch s := src.(type) {\n\t\tcase string:\n\t\t\t*d = s\n\t\tcase []byte:\n\t\t\t*d = string(s)\n\t\tcase time.Time:\n\t\t\t*d = s.Format(time.RFC3339Nano)\n\t\tcase int64:\n\t\t\t*d = strconv.FormatInt(s, 10)\n\t\tcase float64:\n\t\t\t*d = strconv.FormatFloat(s, 'g', -1, 64)\n\t\tcase bool:\n\t\t\t*d = strconv.FormatBool(s)\n\t\tdefault:\n\t\t\treturn fmt.Errorf(\"sql: Scan error on column index %d, name %q: type missmatch\", i, rows.Columns()[i])\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r *rows) Next(dest []driver.Value) (err error) {\n\tif trace {\n\t\tdefer func() {\n\t\t\ttracer(r, \"Next(%v): %v\", dest, err)\n\t\t}()\n\t}\n\trc := r.rc0\n\tif r.doStep {\n\t\tif rc, err = r.step(r.pstmt); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tr.doStep = true\n\n\tswitch rc {\n\tcase bin.XSQLITE_ROW:\n\t\tif g, e := len(dest), len(r.columns); g != e {\n\t\t\treturn fmt.Errorf(\"Next(): have %v destination values, expected %v\", g, e)\n\t\t}\n\n\t\tfor i := range dest {\n\t\t\tct, err := r.columnType(i)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tswitch ct {\n\t\t\tcase bin.XSQLITE_INTEGER:\n\t\t\t\tv, err := r.columnInt64(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_FLOAT:\n\t\t\t\tv, err := r.columnDouble(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_TEXT:\n\t\t\t\tv, err := r.columnText(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_BLOB:\n\t\t\t\tv, err := r.columnBlob(i)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\n\t\t\t\tdest[i] = v\n\t\t\tcase bin.XSQLITE_NULL:\n\t\t\t\tdest[i] = nil\n\t\t\tdefault:\n\t\t\t\tpanic(\"internal error\")\n\t\t\t}\n\t\t}\n\t\treturn nil\n\tcase bin.XSQLITE_DONE:\n\t\treturn io.EOF\n\tdefault:\n\t\treturn r.errstr(int32(rc))\n\t}\n}", "func (b *SelectBuilder) Build(opts ...interface{}) (string, []interface{}) {\n\tif len(b.Tables) > 1 {\n\t\topts = removeOptTableAlias(opts...)\n\t}\n\tcosmosdbTblName := strings.TrimSpace(b.Tables[0])\n\tcosmosdbTblAlias := \"c\"\n\toptTableAlias := extractOptTableAlias(opts...)\n\ttablesClause := strings.Join(b.Tables, \",\")\n\tif optTableAlias != \"\" {\n\t\ttblAlias := optTableAlias[:len(optTableAlias)-1]\n\t\tif !reTblNameWithAlias.MatchString(tablesClause) {\n\t\t\ttablesClause += fmt.Sprintf(\" %s\", tblAlias)\n\t\t}\n\t\tcosmosdbTblAlias = tblAlias\n\t}\n\n\tif b.Flavor == sql.FlavorCosmosDb {\n\t\t/* START: special case for gocosmos */\n\t\tif tokens := reTblNameWithAlias.FindStringSubmatch(cosmosdbTblName); tokens != nil {\n\t\t\tcosmosdbTblName = tokens[1]\n\t\t\tcosmosdbTblAlias = strings.TrimSpace(tokens[3])\n\t\t}\n\t\ttablesClause = fmt.Sprintf(\"%s %s\", cosmosdbTblName, cosmosdbTblAlias)\n\n\t\tif optTableAlias == \"\" {\n\t\t\toptTableAlias = cosmosdbTblAlias + \".\"\n\t\t\topts = append([]interface{}{OptTableAlias{TableAlias: cosmosdbTblAlias}}, opts...)\n\t\t}\n\t\t/* END: special case for gocosmos */\n\t}\n\n\tcolsClause := optTableAlias + \"*\"\n\tif b.Flavor == sql.FlavorCosmosDb {\n\t\tcolsClause = \"*\"\n\t}\n\tif len(b.Columns) > 0 {\n\t\tcols := make([]string, len(b.Columns))\n\t\tcopy(cols, b.Columns)\n\t\tfor i, col := range cols {\n\t\t\tif b.Flavor == sql.FlavorCosmosDb && col == \"*\" {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif !reColnamePrefixedTblname.MatchString(col) {\n\t\t\t\tcols[i] = optTableAlias + col\n\t\t\t}\n\t\t}\n\t\tcolsClause = strings.Join(cols, \",\")\n\t}\n\n\tsqlStm := fmt.Sprintf(\"SELECT %s FROM %s\", colsClause, tablesClause)\n\tvalues := make([]interface{}, 0)\n\tvar tempValues []interface{}\n\n\twhereClause := \"\"\n\tif b.Filter != nil {\n\t\tnewOpts := append([]interface{}{OptDbFlavor{b.Flavor}}, opts...)\n\t\twhereClause, tempValues = b.Filter.Build(b.PlaceholderGenerator, newOpts...)\n\t\tvalues = append(values, tempValues...)\n\t}\n\tif whereClause != \"\" {\n\t\tsqlStm += fmt.Sprintf(\" WHERE %s\", whereClause)\n\t}\n\n\tgroupClause := \"\"\n\tif len(b.GroupBy) > 0 {\n\t\tgroupByList := make([]string, len(b.GroupBy))\n\t\tcopy(groupByList, b.GroupBy)\n\t\tfor i, col := range groupByList {\n\t\t\tgroupByList[i] = col\n\t\t\tif !reColnamePrefixedTblname.MatchString(col) {\n\t\t\t\tgroupByList[i] = optTableAlias + col\n\t\t\t}\n\t\t}\n\t\tgroupClause = strings.Join(groupByList, \",\")\n\t}\n\tif groupClause != \"\" {\n\t\tsqlStm += fmt.Sprintf(\" GROUP BY %s\", groupClause)\n\t}\n\n\thavingClause := \"\"\n\tif b.Having != nil {\n\t\tnewOpts := append([]interface{}{OptDbFlavor{b.Flavor}}, opts...)\n\t\thavingClause, tempValues = b.Having.Build(b.PlaceholderGenerator, newOpts...)\n\t\tvalues = append(values, tempValues...)\n\t}\n\tif havingClause != \"\" {\n\t\tsqlStm += fmt.Sprintf(\" HAVING %s\", havingClause)\n\t}\n\n\torderClause := \"\"\n\tif b.Sorting != nil {\n\t\torderClause = b.Sorting.Build(opts...)\n\t}\n\tif orderClause != \"\" {\n\t\tsqlStm += fmt.Sprintf(\" ORDER BY %s\", orderClause)\n\t}\n\n\tif b.LimitNumRows != 0 || b.LimitOffset != 0 {\n\t\tswitch b.Flavor {\n\t\tcase sql.FlavorMySql, sql.FlavorSqlite:\n\t\t\tsqlStm += fmt.Sprintf(\" LIMIT %d,%d\", b.LimitOffset, b.LimitNumRows)\n\t\tcase sql.FlavorPgSql:\n\t\t\tsqlStm += fmt.Sprintf(\" LIMIT %d OFFSET %d\", b.LimitNumRows, b.LimitOffset)\n\t\tcase sql.FlavorMsSql:\n\t\t\tif orderClause != \"\" {\n\t\t\t\t// available since SQL Server 2012 && Azure SQL Database\n\t\t\t\tsqlStm += fmt.Sprintf(\" OFFSET %d ROWS FETCH NEXT %d ROWS ONLY\", b.LimitOffset, b.LimitNumRows)\n\t\t\t}\n\t\tcase sql.FlavorOracle:\n\t\t\tsqlStm += fmt.Sprintf(\" OFFSET %d ROWS FETCH NEXT %d ROWS ONLY\", b.LimitOffset, b.LimitNumRows)\n\t\tcase sql.FlavorCosmosDb:\n\t\t\tsqlStm += fmt.Sprintf(\" OFFSET %d LIMIT %d\", b.LimitOffset, b.LimitNumRows)\n\t\t}\n\t}\n\n\tif b.Flavor == sql.FlavorCosmosDb {\n\t\t/* START: special case for gocosmos */\n\t\t// sqlStm += \" WITH collection=\" + cosmosdbTblName\n\t\tsqlStm += \" WITH cross_partition=true\"\n\t\t/* END: special case for gocosmos */\n\t}\n\n\treturn sqlStm, values\n}", "func (pb *primitiveBuilder) pushSelectExprs(sel *sqlparser.Select, reservedVars *sqlparser.ReservedVars) error {\n\tresultColumns, err := pb.pushSelectRoutes(sel.SelectExprs, reservedVars)\n\tif err != nil {\n\t\treturn err\n\t}\n\tpb.st.SetResultColumns(resultColumns)\n\treturn pb.pushGroupBy(sel)\n}", "func (s *TableQuery) Select(args ...interface{}) drivers.ITableQuery {\n\ts.SelectSQL = NewColumn().GetSelectColumns(args...)\n\treturn s\n}", "func (ses *Ses) Sel(sqlFrom string, columnPairs ...interface{}) (rset *Rset, err error) {\n\tses.log(_drv.Cfg().Log.Ses.Sel)\n\terr = ses.checkClosed()\n\tif err != nil {\n\t\treturn nil, errE(err)\n\t}\n\tif len(columnPairs) == 0 {\n\t\treturn nil, errF(\"No column name-type pairs specified.\")\n\t}\n\tif len(columnPairs)%2 != 0 {\n\t\treturn nil, errF(\"Variadic parameter 'columnPairs' received an odd number of elements. Parameter 'columnPairs' expects an even number of elements.\")\n\t}\n\t// build select statement, gcts\n\tgcts := make([]GoColumnType, len(columnPairs)/2)\n\tbuf := new(bytes.Buffer)\n\tbuf.WriteString(\"SELECT \")\n\tfor n := 0; n < len(columnPairs); n += 2 {\n\t\tcolumnName, ok := columnPairs[n].(string)\n\t\tif !ok {\n\t\t\treturn nil, errF(\"Variadic parameter 'columnPairs' expected an element at index %v to be of type string\", n)\n\t\t}\n\t\tgct, ok := columnPairs[n+1].(GoColumnType)\n\t\tif !ok {\n\t\t\treturn nil, errF(\"Variadic parameter 'columnPairs' expected an element at index %v to be of type ora.GoColumnType\", n+1)\n\t\t}\n\t\tbuf.WriteString(columnName)\n\t\tif n != len(columnPairs)-2 {\n\t\t\tbuf.WriteRune(',')\n\t\t}\n\t\tbuf.WriteRune(' ')\n\t\tgcts[n/2] = gct\n\t}\n\t// add FROM keyword?\n\tfromIndex := strings.Index(strings.ToUpper(sqlFrom), \"FROM\")\n\tif fromIndex < 0 {\n\t\tbuf.WriteString(\"FROM \")\n\t}\n\tbuf.WriteString(sqlFrom)\n\t// prep\n\tstmt, err := ses.Prep(buf.String(), gcts...)\n\tif err != nil {\n\t\tdefer stmt.Close()\n\t\treturn nil, errE(err)\n\t}\n\t// qry\n\trset, err = stmt.Qry()\n\tif err != nil {\n\t\tdefer stmt.Close()\n\t\treturn nil, errE(err)\n\t}\n\trset.autoClose = true\n\treturn rset, nil\n}", "func (u *__Room_Selector) _stoSql() (string, []interface{}) {\n\tsqlWherrs, whereArgs := whereClusesToSql(u.wheres, u.whereSep)\n\n\tsqlstr := \"SELECT \" + u.selectCol + \" FROM ms.room\"\n\n\tif len(strings.Trim(sqlWherrs, \" \")) > 0 { //2 for safty\n\t\tsqlstr += \" WHERE \" + sqlWherrs\n\t}\n\n\tif u.orderBy != \"\" {\n\t\tsqlstr += u.orderBy\n\t}\n\n\tif u.limit != 0 {\n\t\tsqlstr += \" LIMIT \" + strconv.Itoa(u.limit)\n\t}\n\n\tif u.offset != 0 {\n\t\tsqlstr += \" OFFSET \" + strconv.Itoa(u.offset)\n\t}\n\treturn sqlstr, whereArgs\n}", "func (u *__FileStorage_Selector) _toSql() (string, []interface{}) {\n\n\tsqlWheres, whereArgs := whereClusesToSql(u.wheres, \"\")\n\tselectCols := \"*\"\n\tif len(u.selectCol) > 0 {\n\t\tselectCols = strings.Join(u.selectCol, \", \")\n\t}\n\tsqlstr := \"SELECT \" + selectCols + \" FROM sunc_file.file_storage\"\n\n\tif len(strings.Trim(sqlWheres, \" \")) > 0 { //2 for safty\n\t\tsqlstr += \" WHERE \" + sqlWheres\n\t}\n\n\tif len(u.orderBy) > 0 {\n\t\torders := strings.Join(u.orderBy, \", \")\n\t\tsqlstr += \" ORDER BY \" + orders\n\t}\n\n\tif u.limit != 0 {\n\t\tsqlstr += \" LIMIT \" + strconv.Itoa(u.limit)\n\t}\n\tif u.allowFilter {\n\t\tsqlstr += \" ALLOW FILTERING\"\n\t}\n\n\treturn sqlstr, whereArgs\n}", "func ExecuteSelect(queryer db.Queryer, query string) ([]Row, error) {\n\trows, err := queryer.Query(query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\tcols, err := rows.Columns()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar results []Row\n\tfor rows.Next() {\n\t\t// Create a slice of interface{}'s to represent each column,\n\t\t// and a second slice to contain pointers to each item in the columns slice.\n\t\tcolumns := make([]interface{}, len(cols))\n\t\tcolumnPointers := make([]interface{}, len(cols))\n\t\tfor i := range columns {\n\t\t\tcolumnPointers[i] = &columns[i]\n\t\t}\n\n\t\t// Scan the result into the column pointers...\n\t\tif err := rows.Scan(columnPointers...); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// Create our map, and retrieve the value for each column from the pointers slice,\n\t\t// storing it in the map with the name of the column as the key.\n\t\tm := make(map[string]interface{})\n\t\tfor i, colName := range cols {\n\t\t\tval := columnPointers[i].(*interface{})\n\t\t\tm[colName] = *val\n\t\t}\n\t\tresults = append(results, Row(m))\n\t}\n\tif err := rows.Err(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn results, nil\n}", "func Select(\n\tdb *sql.DB, params map[string]interface{},\n\tfieldMap map[string]string, table string, orderBy string,\n) ([]map[string]interface{}, error) {\n\tvar (\n\t\trows *sql.Rows\n\t\tcondition string\n\t\tlimit string\n\t)\n\tvalues := []interface{}{}\n\tresult := []map[string]interface{}{}\n\tcolumns, keys, processColumn, err := prepareSelectVariables(fieldMap)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif PK, ok := params[\"PK\"]; ok {\n\t\tif pk, ok := PK.(map[string]interface{}); ok {\n\t\t\tif condition, err = WherePK(pk, fieldMap, keys, &values, 0); err != nil {\n\t\t\t\tif err != ErrorZeroParamsInPK {\n\t\t\t\t\treturn nil, err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tcondition = fmt.Sprintf(\"where %s\", condition)\n\t\t\t}\n\t\t}\n\t}\n\n\tif Limit, ok := params[\"Limit\"]; ok {\n\t\tif l, ok := Limit.(float64); ok {\n\t\t\tli := int64(l)\n\t\t\tlimit = fmt.Sprintf(\"limit %d\", li)\n\t\t} else {\n\t\t\treturn nil, fmt.Errorf(\"Cannot convert limit param %v into integer\", Limit)\n\t\t}\n\t}\n\n\tob := \"\"\n\tif orderBy != \"\" {\n\t\tob = fmt.Sprintf(`order by %s`, orderBy)\n\t}\n\tquery := fmt.Sprintf(`select %s from \"%s\" %s %s %s`,\n\t\tstrings.Join(columns, \",\"), table, condition, ob, limit)\n\trows, err = db.Query(query, values...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\n\tcount := len(columns)\n\tslots := make([]interface{}, count)\n\tslotsPtrs := make([]interface{}, count)\n\tfor i := range slots {\n\t\tslotsPtrs[i] = &slots[i]\n\t}\n\tfor rows.Next() {\n\t\tif err := rows.Scan(slotsPtrs...); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\trow := map[string]interface{}{}\n\t\tfor i, key := range keys {\n\t\t\tif err := processColumn[i](slots[i], row, key); err != nil {\n\t\t\t\tlog.Println(err, \"at\", table)\n\t\t\t}\n\t\t}\n\t\tresult = append(result, row)\n\t}\n\n\tif err := rows.Err(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn result, nil\n}", "func (r rowsRes) Next(dest []driver.Value) error {\n\terr := r.my.ScanRow(r.row)\n\tif err != nil {\n\t\treturn errFilter(err)\n\t}\n\tfor i, col := range r.row {\n\t\tif col == nil {\n\t\t\tdest[i] = nil\n\t\t\tcontinue\n\t\t}\n\t\tswitch c := col.(type) {\n\t\tcase time.Time:\n\t\t\tdest[i] = c\n\t\t\tcontinue\n\t\tcase mysql.Timestamp:\n\t\t\tdest[i] = c.Time\n\t\t\tcontinue\n\t\tcase mysql.Date:\n\t\t\tdest[i] = c.Localtime()\n\t\t\tcontinue\n\t\t}\n\t\tv := reflect.ValueOf(col)\n\t\tswitch v.Kind() {\n\t\tcase reflect.Int8, reflect.Int16, reflect.Int32, reflect.Int64:\n\t\t\t// this contains time.Duration to\n\t\t\tdest[i] = v.Int()\n\t\tcase reflect.Uint8, reflect.Uint16, reflect.Uint32, reflect.Uint64:\n\t\t\tu := v.Uint()\n\t\t\tif u > math.MaxInt64 {\n\t\t\t\tpanic(\"Value to large for int64 type\")\n\t\t\t}\n\t\t\tdest[i] = int64(u)\n\t\tcase reflect.Float32, reflect.Float64:\n\t\t\tdest[i] = v.Float()\n\t\tcase reflect.Slice:\n\t\t\tif v.Type().Elem().Kind() == reflect.Uint8 {\n\t\t\t\tdest[i] = v.Interface().([]byte)\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tfallthrough\n\t\tdefault:\n\t\t\tpanic(fmt.Sprint(\"Unknown type of column: \", v.Type()))\n\t\t}\n\t}\n\treturn nil\n}", "func (s *Session) Select(columns ...string) *Session {\n\ts.initStatemnt()\n\ts.statement.Select(columns...)\n\treturn s\n}", "func (rdq *ResultsDefinitionQuery) Select(fields ...string) *ResultsDefinitionSelect {\n\trdq.fields = append(rdq.fields, fields...)\n\tselbuild := &ResultsDefinitionSelect{ResultsDefinitionQuery: rdq}\n\tselbuild.label = resultsdefinition.Label\n\tselbuild.flds, selbuild.scan = &rdq.fields, selbuild.Scan\n\treturn selbuild\n}", "func (db *DB) Select(table string, cb RowCB, columns ...string) error {\n\tinnerCB := func(row Row) bool {\n\t\tcb(row)\n\t\treturn false // not done\n\t}\n\treturn db.SelectDone(table, innerCB, columns...)\n}", "func (u *__DirectToMessage_Selector) _stoSql() (string, []interface{}) {\n\tsqlWherrs, whereArgs := whereClusesToSql(u.wheres, u.whereSep)\n\n\tsqlstr := \"SELECT \" + u.selectCol + \" FROM ms.direct_to_message\"\n\n\tif len(strings.Trim(sqlWherrs, \" \")) > 0 { //2 for safty\n\t\tsqlstr += \" WHERE \" + sqlWherrs\n\t}\n\n\tif u.orderBy != \"\" {\n\t\tsqlstr += u.orderBy\n\t}\n\n\tif u.limit != 0 {\n\t\tsqlstr += \" LIMIT \" + strconv.Itoa(u.limit)\n\t}\n\n\tif u.offset != 0 {\n\t\tsqlstr += \" OFFSET \" + strconv.Itoa(u.offset)\n\t}\n\treturn sqlstr, whereArgs\n}", "func (api *MysqlAPI) Select(option QueryOption) (rs []map[string]interface{}, err error) {\n\tvar sql string\n\tfor _, f := range option.Fields {\n\t\tif !api.databaseMetadata.TableHaveField(option.Table, f) {\n\t\t\terr = fmt.Errorf(\"Table '%s' not have '%s' field !/n\", option.Table, f)\n\t\t\treturn\n\t\t}\n\t}\n\tif option.Id != \"\" {\n\t\tsql, err = api.sql.GetByTableAndID(option)\n\t} else {\n\t\tsql, err = api.sql.GetByTable(option)\n\t}\n\tif err != nil {\n\t\treturn\n\t}\n\treturn api.query(sql)\n}", "func (conn *db) SelectSlice(stmt Stmt, mapper SliceMapper) (rowsReturned int, err error) {\n\treturn conn.runSlice(stmt, mapper)\n}", "func (sd *SelectDataset) copy(clauses exp.SelectClauses) *SelectDataset {\n\treturn &SelectDataset{\n\t\tdialect: sd.dialect,\n\t\tclauses: clauses,\n\t\tisPrepared: sd.isPrepared,\n\t\tqueryFactory: sd.queryFactory,\n\t\terr: sd.err,\n\t}\n}", "func (op *OpQuery) Select(fields ...string) *OpQuery {\n\tif op.selectFileds == nil {\n\t\top.selectFileds = make([]string, 0)\n\t}\n\top.selectFileds = append(op.selectFileds, fields...)\n\treturn op\n}", "func (q *PersonQuery) Select(columns ...kallax.SchemaField) *PersonQuery {\n\tif len(columns) == 0 {\n\t\treturn q\n\t}\n\tq.BaseQuery.Select(columns...)\n\treturn q\n}", "func (s Select) Build() string {\n\tvar buf strings.Builder\n\n\tbuf.WriteString(\"SELECT \")\n\tbuf.WriteString(buildColumns(s.columns))\n\n\tbuf.WriteByte(' ')\n\n\tbuf.WriteString(s.from.Build())\n\n\tif s.where != \"\" {\n\t\tbuf.WriteByte(' ')\n\t\tbuf.WriteString(\"WHERE \")\n\t\tbuf.WriteString(s.where)\n\t}\n\n\tif len(s.orderBy.cols) != 0 {\n\t\tbuf.WriteByte(' ')\n\t\tbuf.WriteString(s.orderBy.Build())\n\t}\n\n\tif s.limit > 0 {\n\t\tbuf.WriteByte(' ')\n\t\tbuf.WriteString(\"LIMIT \")\n\t\tbuf.WriteString(strconv.Itoa(s.limit))\n\n\t} else if s.paged {\n\t\tbuf.WriteByte(' ')\n\t\tbuf.WriteString(\"LIMIT ? OFFSET ?\")\n\t}\n\n\tif s.lock {\n\t\tbuf.WriteByte(' ')\n\t\tbuf.WriteString(\"FOR UPDATE\")\n\t}\n\n\treturn buf.String()\n}", "func (b *QueryBuilder) Select(nodes ...NodeI) {\n\tif b.GroupBys != nil {\n\t\tpanic(\"You cannot have Select and GroupBy statements in the same query. The GroupBy columns will automatically be selected.\")\n\t}\n\tfor _, n := range nodes {\n\t\tif NodeGetType(n) != ColumnNodeType {\n\t\t\tpanic(\"you can only select column nodes\")\n\t\t}\n\t}\n\tb.Selects = append(b.Selects, nodes...)\n}", "func (q *Query) selectClause() (columns []string, parser func(*spanner.Row) (*pb.TestResult, error)) {\n\tcolumns = []string{\n\t\t\"InvocationId\",\n\t\t\"TestId\",\n\t\t\"ResultId\",\n\t\t\"IsUnexpected\",\n\t\t\"Status\",\n\t\t\"StartTime\",\n\t\t\"RunDurationUsec\",\n\t}\n\n\t// Select extra columns depending on the mask.\n\tvar extraColumns []string\n\treadMask := q.Mask\n\tif readMask.IsEmpty() {\n\t\treadMask = defaultListMask\n\t}\n\tselectIfIncluded := func(column, field string) {\n\t\tswitch inc, err := readMask.Includes(field); {\n\t\tcase err != nil:\n\t\t\tpanic(err)\n\t\tcase inc != mask.Exclude:\n\t\t\textraColumns = append(extraColumns, column)\n\t\t\tcolumns = append(columns, column)\n\t\t}\n\t}\n\tselectIfIncluded(\"SummaryHtml\", \"summary_html\")\n\tselectIfIncluded(\"Tags\", \"tags\")\n\tselectIfIncluded(\"TestMetadata\", \"test_metadata\")\n\tselectIfIncluded(\"Variant\", \"variant\")\n\tselectIfIncluded(\"VariantHash\", \"variant_hash\")\n\tselectIfIncluded(\"FailureReason\", \"failure_reason\")\n\tselectIfIncluded(\"Properties\", \"properties\")\n\n\t// Build a parser function.\n\tvar b spanutil.Buffer\n\tvar summaryHTML spanutil.Compressed\n\tvar tmd spanutil.Compressed\n\tvar fr spanutil.Compressed\n\tvar properties spanutil.Compressed\n\tparser = func(row *spanner.Row) (*pb.TestResult, error) {\n\t\tvar invID invocations.ID\n\t\tvar maybeUnexpected spanner.NullBool\n\t\tvar micros spanner.NullInt64\n\t\ttr := &pb.TestResult{}\n\n\t\tptrs := []any{\n\t\t\t&invID,\n\t\t\t&tr.TestId,\n\t\t\t&tr.ResultId,\n\t\t\t&maybeUnexpected,\n\t\t\t&tr.Status,\n\t\t\t&tr.StartTime,\n\t\t\t&micros,\n\t\t}\n\n\t\tfor _, v := range extraColumns {\n\t\t\tswitch v {\n\t\t\tcase \"SummaryHtml\":\n\t\t\t\tptrs = append(ptrs, &summaryHTML)\n\t\t\tcase \"Tags\":\n\t\t\t\tptrs = append(ptrs, &tr.Tags)\n\t\t\tcase \"TestMetadata\":\n\t\t\t\tptrs = append(ptrs, &tmd)\n\t\t\tcase \"Variant\":\n\t\t\t\tptrs = append(ptrs, &tr.Variant)\n\t\t\tcase \"VariantHash\":\n\t\t\t\tptrs = append(ptrs, &tr.VariantHash)\n\t\t\tcase \"FailureReason\":\n\t\t\t\tptrs = append(ptrs, &fr)\n\t\t\tcase \"Properties\":\n\t\t\t\tptrs = append(ptrs, &properties)\n\t\t\tdefault:\n\t\t\t\tpanic(\"impossible\")\n\t\t\t}\n\t\t}\n\n\t\terr := b.FromSpanner(row, ptrs...)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\t// Generate test result name now in case tr.TestId and tr.ResultId become\n\t\t// empty after q.Mask.Trim(tr).\n\t\ttrName := pbutil.TestResultName(string(invID), tr.TestId, tr.ResultId)\n\t\ttr.SummaryHtml = string(summaryHTML)\n\t\tPopulateExpectedField(tr, maybeUnexpected)\n\t\tPopulateDurationField(tr, micros)\n\t\tif err := populateTestMetadata(tr, tmd); err != nil {\n\t\t\treturn nil, errors.Annotate(err, \"error unmarshalling test_metadata for %s\", trName).Err()\n\t\t}\n\t\tif err := populateFailureReason(tr, fr); err != nil {\n\t\t\treturn nil, errors.Annotate(err, \"error unmarshalling failure_reason for %s\", trName).Err()\n\t\t}\n\t\tif err := populateProperties(tr, properties); err != nil {\n\t\t\treturn nil, errors.Annotate(err, \"failed to unmarshal properties\").Err()\n\t\t}\n\t\tif err := q.Mask.Trim(tr); err != nil {\n\t\t\treturn nil, errors.Annotate(err, \"error trimming fields for %s\", trName).Err()\n\t\t}\n\t\t// Always include name in tr because name is needed to calculate\n\t\t// page token.\n\t\ttr.Name = trName\n\t\treturn tr, nil\n\t}\n\treturn\n}", "func (sd *SelectDataset) SelectAppend(selects ...interface{}) *SelectDataset {\n\treturn sd.copy(sd.clauses.SelectAppend(exp.NewColumnListExpression(selects...)))\n}", "func processSelect(sel *sqlparser.Select, vschema VSchema, outer builder) (builder, error) {\n\tbldr, err := processTableExprs(sel.From, vschema)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif outer != nil {\n\t\tbldr.Symtab().Outer = outer.Symtab()\n\t}\n\tif sel.Where != nil {\n\t\terr = pushFilter(sel.Where.Expr, bldr, sqlparser.WhereStr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tbldr, err = checkAggregates(sel, bldr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbldr, err = pushSelectExprs(sel, bldr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif sel.Having != nil {\n\t\terr = pushFilter(sel.Having.Expr, bldr, sqlparser.HavingStr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\terr = pushOrderBy(sel.OrderBy, bldr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbldr, err = pushLimit(sel.Limit, bldr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbldr.PushMisc(sel)\n\treturn bldr, nil\n}", "func (q *SessionQuery) Select(columns ...kallax.SchemaField) *SessionQuery {\n\tif len(columns) == 0 {\n\t\treturn q\n\t}\n\tq.BaseQuery.Select(columns...)\n\treturn q\n}", "func (u *__StorageUsage_Selector) _toSql() (string, []interface{}) {\n\n\tsqlWheres, whereArgs := whereClusesToSql(u.wheres, \"\")\n\tselectCols := \"*\"\n\tif len(u.selectCol) > 0 {\n\t\tselectCols = strings.Join(u.selectCol, \", \")\n\t}\n\tsqlstr := \"SELECT \" + selectCols + \" FROM sunc_file.storage_usage\"\n\n\tif len(strings.Trim(sqlWheres, \" \")) > 0 { //2 for safty\n\t\tsqlstr += \" WHERE \" + sqlWheres\n\t}\n\n\tif len(u.orderBy) > 0 {\n\t\torders := strings.Join(u.orderBy, \", \")\n\t\tsqlstr += \" ORDER BY \" + orders\n\t}\n\n\tif u.limit != 0 {\n\t\tsqlstr += \" LIMIT \" + strconv.Itoa(u.limit)\n\t}\n\tif u.allowFilter {\n\t\tsqlstr += \" ALLOW FILTERING\"\n\t}\n\n\treturn sqlstr, whereArgs\n}", "func (r *result) Next(dest []driver.Value) error {\n\tif r.data == nil {\n\t\tif err := r.readNext(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tfor i := 0; i < len(r.columns); i++ {\n\t\tkey := r.columns[i]\n\t\tval := r.data[key]\n\t\tdest[i] = val\n\t}\n\tr.data = nil\n\tr.offset++\n\treturn nil\n}", "func (stmt *statement) Dest() []interface{} {\n\treturn stmt.dest\n}", "func (ex *Execer) querySliceFn(dest interface{}) error {\n\t// Validate the dest and reflection values we need\n\n\t// This must be a pointer to a slice\n\tvalueOfDest := reflect.ValueOf(dest)\n\tkindOfDest := valueOfDest.Kind()\n\n\tif kindOfDest != reflect.Ptr {\n\t\tpanic(\"invalid type passed to LoadValues. Need a pointer to a slice\")\n\t}\n\n\t// This must a slice\n\tvalueOfDest = reflect.Indirect(valueOfDest)\n\tkindOfDest = valueOfDest.Kind()\n\n\tif kindOfDest != reflect.Slice {\n\t\tpanic(\"invalid type passed to LoadValues. Need a pointer to a slice\")\n\t}\n\n\trecordType := valueOfDest.Type().Elem()\n\n\trecordTypeIsPtr := recordType.Kind() == reflect.Ptr\n\tif recordTypeIsPtr {\n\t\treflect.ValueOf(dest)\n\t}\n\n\tfullSQL, args, blob, err := ex.cacheOrSQL()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif blob != nil {\n\t\terr = json.Unmarshal(blob, &dest)\n\t\tif err == nil {\n\t\t\treturn nil\n\t\t}\n\t\t// log it and fallthrough to let the query continue\n\t\tlogger.Warn(\"querySlice.2: Could not unmarshal cache data. Continuing with query\")\n\t}\n\n\tdefer logExecutionTime(time.Now(), fullSQL, args)\n\trows, err := ex.database.Queryx(fullSQL, args...)\n\tif err != nil {\n\t\treturn logSQLError(err, \"querySlice.load_all_values.query\", fullSQL, args)\n\t}\n\n\tsliceValue := valueOfDest\n\tdefer rows.Close()\n\tfor rows.Next() {\n\t\t// Create a new value to store our row:\n\t\tpointerToNewValue := reflect.New(recordType)\n\t\tnewValue := reflect.Indirect(pointerToNewValue)\n\n\t\terr = rows.Scan(pointerToNewValue.Interface())\n\t\tif err != nil {\n\t\t\treturn logSQLError(err, \"querySlice.load_all_values.scan\", fullSQL, args)\n\t\t}\n\n\t\t// Append our new value to the slice:\n\t\tsliceValue = reflect.Append(sliceValue, newValue)\n\t}\n\tvalueOfDest.Set(sliceValue)\n\n\tif err := rows.Err(); err != nil {\n\t\treturn logSQLError(err, \"querySlice.load_all_values.rows_err\", fullSQL, args)\n\t}\n\n\tex.setCache(dest, dtStruct)\n\n\treturn nil\n}", "func pushSelectRoutes(selectExprs sqlparser.SelectExprs, bldr builder) ([]*resultColumn, error) {\n\tresultColumns := make([]*resultColumn, len(selectExprs))\n\tfor i, node := range selectExprs {\n\t\tswitch node := node.(type) {\n\t\tcase *sqlparser.AliasedExpr:\n\t\t\torigin, err := findOrigin(node.Expr, bldr)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tresultColumns[i], _, err = bldr.PushSelect(node, origin)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\tcase *sqlparser.StarExpr:\n\t\t\t// We'll allow select * for simple routes.\n\t\t\trb, ok := bldr.(*route)\n\t\t\tif !ok {\n\t\t\t\treturn nil, errors.New(\"unsupported: '*' expression in cross-shard query\")\n\t\t\t}\n\t\t\t// Validate keyspace reference if any.\n\t\t\tif !node.TableName.IsEmpty() {\n\t\t\t\tif qual := node.TableName.Qualifier; !qual.IsEmpty() {\n\t\t\t\t\tif qual.String() != rb.ERoute.Keyspace.Name {\n\t\t\t\t\t\treturn nil, fmt.Errorf(\"cannot resolve %s to keyspace %s\", sqlparser.String(node), rb.ERoute.Keyspace.Name)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t\tresultColumns[i] = rb.PushAnonymous(node)\n\t\tcase sqlparser.Nextval:\n\t\t\trb, ok := bldr.(*route)\n\t\t\tif !ok {\n\t\t\t\t// This code is unreachable because the parser doesn't allow joins for next val statements.\n\t\t\t\treturn nil, errors.New(\"unsupported: SELECT NEXT query in cross-shard query\")\n\t\t\t}\n\t\t\tif err := rb.SetOpcode(engine.SelectNext); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tresultColumns[i] = rb.PushAnonymous(node)\n\t\tdefault:\n\t\t\tpanic(fmt.Sprintf(\"BUG: unexpceted select expression type: %T\", node))\n\t\t}\n\t}\n\treturn resultColumns, nil\n}", "func (c *commentsQueryBuilder) Select(fields ...string) *commentsQueryBuilder {\n\tif c.err != nil {\n\t\treturn c\n\t}\n\tvar fieldSet []*mapping.StructField\n\tfor _, field := range fields {\n\t\tstructField, ok := NRN_Comments.Model.FieldByName(field)\n\t\tif !ok {\n\t\t\tc.err = errors.Wrapf(mapping.ErrInvalidModelField, \"field: '%s' is not valid for model: '_Comments'\", field)\n\t\t\treturn c\n\t\t}\n\t\tfieldSet = append(fieldSet, structField)\n\t}\n\tc.builder.Select(fieldSet...)\n\treturn c\n}", "func (dbi *DB) SelectRow(sqls string, arg ...string) (map[string]string, error) {\r\n\tif dbi.status == false {\r\n\t\treturn nil, errors.New(\"database was not opened.\")\r\n\t}\r\n\r\n\t// start select\r\n\tdbi.createOperation(\"DB_SELECT\")\r\n\tdbi.data.reqSql = sqls\r\n\tdbi.data.inVar = arg\r\n\t// data\r\n\tdbi.data.commPrepare()\r\n\t// communicate\r\n\tif dbi.data.comm() == false {\r\n\t\tmylog.Println(dbi.data.sqlCode, \":\", dbi.data.sqlErrm)\r\n\t\tdbi.Close()\r\n\t\treturn nil, errors.New(dbi.Sqlerrm)\r\n\t}\r\n\t// parse\r\n\tdbi.data.commParse()\r\n\tdbi.parseError()\r\n\tif dbi.Sqlcode != 0 {\r\n\t\treturn nil, errors.New(dbi.Sqlerrm)\r\n\t}\r\n\r\n\t// get column name list\r\n\tdbi.createOperation(\"DB_COLUMNS\")\r\n\t// data\r\n\tdbi.data.curId = -1\r\n\tdbi.data.commPrepare()\r\n\t// communicate\r\n\tif dbi.data.comm() == false {\r\n\t\tdbi.Close()\r\n\t\treturn nil, errors.New(dbi.Sqlerrm)\r\n\t}\r\n\t// parse\r\n\tdbi.data.commParse()\r\n\t// no need to change dbi.sqlcode\r\n\r\n\tcolvar := make(map[string]string)\r\n\tfor i := 0; i < len(dbi.data.colName); i++ {\r\n\t\tcolvar[strings.ToLower(dbi.data.colName[i])] = dbi.data.outVar[i]\r\n\t}\r\n\treturn colvar, nil\r\n}", "func (duo *DocUpdateOne) Select(field string, fields ...string) *DocUpdateOne {\n\tduo.fields = append([]string{field}, fields...)\n\treturn duo\n}", "func (c *Conn) Select(query string, rowCallbackHandler func(s *Stmt) error, args ...interface{}) error {\n\ts, err := c.Prepare(query)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer s.Finalize()\n\treturn s.Select(rowCallbackHandler, args...)\n}", "func (dbi *DB) Select(sqls string, arg ...string) ([]map[string]string, error) {\r\n\tif dbi.status == false {\r\n\t\treturn nil, errors.New(\"database was not opened.\")\r\n\t}\r\n\r\n\trows, err := dbi.Cursor(sqls, arg...)\r\n\tif err != nil {\r\n\t\treturn nil, errors.New(\"Cannot open cursor for sql: \" + sqls)\r\n\t}\r\n\r\n\tvar resvar []map[string]string\r\n\tfor {\r\n\t\tcolvar, rowerr := rows.Next()\r\n\t\tif rowerr != nil {\r\n\t\t\trows.Close()\r\n\t\t\treturn nil, errors.New(\"Cannot fetch records for sql: \" + sqls)\r\n\t\t} else if colvar == nil && rowerr == nil {\r\n\t\t\t// no records found\r\n\t\t\trows.status = false\r\n\t\t\tbreak\r\n\t\t}\r\n\r\n\t\tline := make(map[string]string)\r\n\t\tfor i := 0; i < len(rows.Cols); i++ {\r\n\t\t\tline[rows.Cols[i]] = colvar[i]\r\n\t\t}\r\n\t\tresvar = append(resvar, line)\r\n\t}\r\n\treturn resvar, nil\r\n}", "func (r *rows) Next(dest []driver.Value) error {\n\tif r.index >= len(r.rows) {\n\t\tif r.parent.offset < r.parent.total {\n\t\t\tif err := fetchNextPage(r.parent.conn, r.parent.jobid, r.parent.offset, r.parent.total, r.parent); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tr.index = r.parent.rows.index\n\t\t\tr.rows = r.parent.rows.rows\n\t\t} else {\n\t\t\tr.rows = nil\n\t\t}\n\t}\n\tif len(dest) != len(r.parent.columns) {\n\t\treturn fmt.Errorf(\"invalid scan, expected %d arguments and received %d\", len(r.parent.columns), len(dest))\n\t}\n\tif r.rows == nil || len(r.rows) == 0 {\n\t\treturn sql.ErrNoRows\n\t}\n\ttherow := r.rows[r.index]\n\tfor i := 0; i < len(r.parent.columns); i++ {\n\t\tkey := r.parent.columns[i]\n\t\tval := therow[key]\n\t\tdest[i] = val\n\t}\n\tr.index++\n\treturn nil\n}", "func (p *postsQueryBuilder) Select(fields ...string) *postsQueryBuilder {\n\tif p.err != nil {\n\t\treturn p\n\t}\n\tvar fieldSet []*mapping.StructField\n\tfor _, field := range fields {\n\t\tstructField, ok := NRN_Posts.Model.FieldByName(field)\n\t\tif !ok {\n\t\t\tp.err = errors.Wrapf(mapping.ErrInvalidModelField, \"field: '%s' is not valid for model: '_Posts'\", field)\n\t\t\treturn p\n\t\t}\n\t\tfieldSet = append(fieldSet, structField)\n\t}\n\tp.builder.Select(fieldSet...)\n\treturn p\n}", "func (u *__Comment_Selector) _stoSql() (string, []interface{}) {\n\tsqlWherrs, whereArgs := whereClusesToSql(u.wheres, u.whereSep)\n\n\tsqlstr := \"SELECT \" + u.selectCol + \" FROM sun.comment\"\n\n\tif len(strings.Trim(sqlWherrs, \" \")) > 0 { //2 for safty\n\t\tsqlstr += \" WHERE \" + sqlWherrs\n\t}\n\n\tif u.orderBy != \"\" {\n\t\tsqlstr += u.orderBy\n\t}\n\n\tif u.limit != 0 {\n\t\tsqlstr += \" LIMIT \" + strconv.Itoa(u.limit)\n\t}\n\n\tif u.offset != 0 {\n\t\tsqlstr += \" OFFSET \" + strconv.Itoa(u.offset)\n\t}\n\treturn sqlstr, whereArgs\n}", "func (s Select) ToSQL() (string, []interface{}) {\n\tvalues := []interface{}{}\n\n\tquery := fmt.Sprintf(\n\t\t`SELECT %s FROM %s`,\n\t\tstrings.Join(s.columns, \", \"),\n\t\ts.BaseQuery.table,\n\t)\n\n\tif s.BaseQuery.hasExpression {\n\t\texpression, prepared, _ := s.BaseQuery.expression.ToQuery(1)\n\t\tquery += \" WHERE \" + expression\n\t\tvalues = prepared\n\t}\n\n\tif s.BaseQuery.hasSort {\n\t\tquery += fmt.Sprintf(\" ORDER BY %s\", s.BaseQuery.sort.ToString())\n\t}\n\n\tif s.BaseQuery.limit > 0 {\n\t\tquery += fmt.Sprintf(\" LIMIT %d\", s.BaseQuery.limit)\n\t}\n\n\tif s.BaseQuery.offset > 0 {\n\t\tquery += fmt.Sprintf(\" OFFSET %d\", s.BaseQuery.offset)\n\t}\n\n\treturn query, values\n}", "func (q *UserQuery) Select(columns ...kallax.SchemaField) *UserQuery {\n\tif len(columns) == 0 {\n\t\treturn q\n\t}\n\tq.BaseQuery.Select(columns...)\n\treturn q\n}", "func (ruo *RevisionUpdateOne) Select(field string, fields ...string) *RevisionUpdateOne {\n\truo.fields = append([]string{field}, fields...)\n\treturn ruo\n}", "func (ouo *OtherUpdateOne) Select(field string, fields ...string) *OtherUpdateOne {\n\touo.fields = append([]string{field}, fields...)\n\treturn ouo\n}", "func (gsuo *GameServerUpdateOne) Select(field string, fields ...string) *GameServerUpdateOne {\n\tgsuo.fields = append([]string{field}, fields...)\n\treturn gsuo\n}", "func (auo *AdministratorUpdateOne) Select(field string, fields ...string) *AdministratorUpdateOne {\n\tauo.fields = append([]string{field}, fields...)\n\treturn auo\n}", "func (u *__Notify_Selector) _stoSql() (string, []interface{}) {\n\tsqlWherrs, whereArgs := whereClusesToSql(u.wheres, u.whereSep)\n\n\tsqlstr := \"SELECT \" + u.selectCol + \" FROM sun.notify\"\n\n\tif len(strings.Trim(sqlWherrs, \" \")) > 0 { //2 for safty\n\t\tsqlstr += \" WHERE \" + sqlWherrs\n\t}\n\n\tif u.orderBy != \"\" {\n\t\tsqlstr += u.orderBy\n\t}\n\n\tif u.limit != 0 {\n\t\tsqlstr += \" LIMIT \" + strconv.Itoa(u.limit)\n\t}\n\n\tif u.offset != 0 {\n\t\tsqlstr += \" OFFSET \" + strconv.Itoa(u.offset)\n\t}\n\treturn sqlstr, whereArgs\n}", "func (q *PetQuery) Select(columns ...kallax.SchemaField) *PetQuery {\n\tif len(columns) == 0 {\n\t\treturn q\n\t}\n\tq.BaseQuery.Select(columns...)\n\treturn q\n}", "func Select(fields ...string) *Query {\n\tquery := newQuery()\n\tquery.action = action_select\n\n\treturn query.Select(fields...)\n}", "func (b *InsertBuilder) Build(opts ...interface{}) (string, []interface{}) {\n\tif b.Flavor == sql.FlavorCosmosDb {\n\t\topts = removeOptTableAlias(opts...)\n\t}\n\tcols := make([]string, 0, len(b.Values))\n\tplaceholders := make([]string, 0, len(b.Values))\n\tvalues := make([]interface{}, 0, len(b.Values))\n\n\tfor k := range b.Values {\n\t\tcols = append(cols, k)\n\t}\n\tsort.Strings(cols)\n\tfor _, col := range cols {\n\t\tvalues = append(values, b.Values[col])\n\t\tplaceholders = append(placeholders, b.PlaceholderGenerator(col))\n\t}\n\n\ttableAliasForField := extractOptTableAlias(opts...)\n\ttableAlias := \"\"\n\tif tableAliasForField != \"\" {\n\t\tif !reTblNameWithAlias.MatchString(b.Table) {\n\t\t\ttableAlias = \" \" + tableAliasForField[:len(tableAliasForField)-1]\n\t\t}\n\t\tfor i, _ := range cols {\n\t\t\tif !reColnamePrefixedTblname.MatchString(cols[i]) {\n\t\t\t\tcols[i] = tableAliasForField + cols[i]\n\t\t\t}\n\t\t}\n\t}\n\tsql := fmt.Sprintf(\"INSERT INTO %s%s (%s) VALUES (%s)\", b.Table, tableAlias, strings.Join(cols, \",\"), strings.Join(placeholders, \",\"))\n\treturn sql, values\n}", "func SelectExists(dest interface{}, query string, args ...interface{}) (exists bool, err error) {\n\tstmt := database.prepare(query)\n\tdefer stmt.Close()\n\terr = stmt.QueryRow(args...).Scan(dest)\n\tif err == sql.ErrNoRows {\n\t\terr = nil // consider a non-error, means the row does not exist.\n\t\treturn\n\t}\n\tif err == nil {\n\t\texists = true\n\t}\n\treturn\n}", "func MakeSelectBuilder(db Queryer, cols ...string) *SelectBuilder {\n\treturn &SelectBuilder{\n\t\tdb: db,\n\t\tcols: cols,\n\t}\n}", "func (c *cassandraConnector) buildSelectQuery(\n\tctx context.Context,\n\te *base.Definition,\n\tkeyCols []base.Column,\n\tcolNamesToRead []string,\n) (*gocql.Query, error) {\n\n\t// split keyCols into a list of names and values to compose query stmt using\n\t// names and use values in the session query call, so the order needs to be\n\t// maintained.\n\tkeyColNames, keyColValues := splitColumnNameValue(keyCols)\n\n\t// Prepare select statement\n\tstmt, err := SelectStmt(\n\t\tTable(e.Name),\n\t\tColumns(colNamesToRead),\n\t\tConditions(keyColNames),\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.Session.Query(stmt, keyColValues...).WithContext(ctx), nil\n}", "func FetchRows(rows *sql.Rows, dst interface{}) error {\n\tvar columns []string\n\tvar err error\n\n\t// Destination.\n\tdstv := reflect.ValueOf(dst)\n\n\tif dstv.IsNil() || dstv.Kind() != reflect.Ptr {\n\t\treturn db.ErrExpectingPointer\n\t}\n\n\tif dstv.Elem().Kind() != reflect.Slice {\n\t\treturn db.ErrExpectingSlicePointer\n\t}\n\n\tif dstv.Kind() != reflect.Ptr || dstv.Elem().Kind() != reflect.Slice || dstv.IsNil() {\n\t\treturn db.ErrExpectingSliceMapStruct\n\t}\n\n\tif columns, err = rows.Columns(); err != nil {\n\t\treturn err\n\t}\n\n\tslicev := dstv.Elem()\n\titem_t := slicev.Type().Elem()\n\n\treset(dst)\n\n\tfor rows.Next() {\n\n\t\titem, err := fetchResult(item_t, rows, columns)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tslicev = reflect.Append(slicev, reflect.Indirect(item))\n\t}\n\n\trows.Close()\n\n\tdstv.Elem().Set(slicev)\n\n\treturn nil\n}", "func (q *PollQuery) Select(columns ...kallax.SchemaField) *PollQuery {\n\tif len(columns) == 0 {\n\t\treturn q\n\t}\n\tq.BaseQuery.Select(columns...)\n\treturn q\n}", "func (b *blogsQueryBuilder) Select(fields ...string) *blogsQueryBuilder {\n\tif b.err != nil {\n\t\treturn b\n\t}\n\tvar fieldSet []*mapping.StructField\n\tfor _, field := range fields {\n\t\tstructField, ok := NRN_Blogs.Model.FieldByName(field)\n\t\tif !ok {\n\t\t\tb.err = errors.Wrapf(mapping.ErrInvalidModelField, \"field: '%s' is not valid for model: '_Blogs'\", field)\n\t\t\treturn b\n\t\t}\n\t\tfieldSet = append(fieldSet, structField)\n\t}\n\tb.builder.Select(fieldSet...)\n\treturn b\n}", "func Select(fields ...interface{}) QueryOptions {\n\treturn func(query *Query) (*Query, error) {\n\t\tif len(fields) == 0 {\n\t\t\treturn query, nil\n\t\t}\n\t\tvar fieldPaths []interface{}\n\t\tfor _, field := range fields {\n\t\t\tif _, ok := field.(string); !ok {\n\t\t\t\treturn nil, errors.New(\"field path type can be only string\")\n\t\t\t}\n\t\t\tfieldPaths = append(fieldPaths, field)\n\t\t}\n\t\tresMap, err := mergeQueryMaps(query.content,\n\t\t\tmap[string]interface{}{operations[SELECT]: fieldPaths})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tquery.content = resMap.(map[string]interface{})\n\t\treturn query, nil\n\t}\n}", "func (duo *DealUpdateOne) Select(field string, fields ...string) *DealUpdateOne {\n\tduo.fields = append([]string{field}, fields...)\n\treturn duo\n}", "func (tuo *TransactionUpdateOne) Select(field string, fields ...string) *TransactionUpdateOne {\n\ttuo.fields = append([]string{field}, fields...)\n\treturn tuo\n}", "func (suo *SettingUpdateOne) Select(field string, fields ...string) *SettingUpdateOne {\n\tsuo.fields = append([]string{field}, fields...)\n\treturn suo\n}", "func (me *NamedExecHelper) SelectQuery(condition *ConditionBuilder, fields ...string) string {\r\n\tvar (\r\n\t\tsField = \"*\"\r\n\t\tsCondition string\r\n\t)\r\n\tif len(fields) > 0 {\r\n\t\t//use provided fields\r\n\t\tsField = strings.Join(fields, \",\")\r\n\t}\r\n\tif condition != nil {\r\n\t\tsCondition = condition.UnnamedConditionString()\r\n\t}\r\n\r\n\treturn fmt.Sprintf(\"SELECT %s FROM %s %s;\", sField, me.tablename, sCondition)\r\n}", "func (dr *DataRow) selectByIndices(columnIndices []int) DataRow {\n\toutItems := make([]DataItem, 0, len(columnIndices))\n\n\tfor _, index := range columnIndices {\n\t\toutItems = append(outItems, dr.Items[index])\n\t}\n\n\treturn DataRow{\n\t\tItems: outItems,\n\t}\n}", "func (isrtuo *IPStaticRoutingTableUpdateOne) Select(field string, fields ...string) *IPStaticRoutingTableUpdateOne {\n\tisrtuo.fields = append([]string{field}, fields...)\n\treturn isrtuo\n}", "func (druo *DeviceRequestUpdateOne) Select(field string, fields ...string) *DeviceRequestUpdateOne {\n\tdruo.fields = append([]string{field}, fields...)\n\treturn druo\n}", "func buildResultRow(e *base.Definition, columns []string) []interface{} {\n\n\tresults := make([]interface{}, len(columns))\n\ttimeType := reflect.ValueOf(time.Now())\n\tgocqlUUIDType := reflect.ValueOf(gocql.UUIDFromTime(time.Now()))\n\n\tfor i, column := range columns {\n\t\t// get the type of the field from the ColumnToType mapping for object\n\t\t// That we we can allocate appropriate memory for this field\n\t\ttyp := e.ColumnToType[column]\n\n\t\tswitch typ.Kind() {\n\t\tcase reflect.String:\n\t\t\tvar value *string\n\t\t\tresults[i] = &value\n\t\tcase reflect.Int32, reflect.Uint32, reflect.Int:\n\t\t\t// C* internally uses int and int64\n\t\t\tvar value *int\n\t\t\tresults[i] = &value\n\t\tcase reflect.Int64, reflect.Uint64:\n\t\t\t// C* internally uses int and int64\n\t\t\tvar value *int64\n\t\t\tresults[i] = &value\n\t\tcase reflect.Bool:\n\t\t\tvar value *bool\n\t\t\tresults[i] = &value\n\t\tcase reflect.Slice:\n\t\t\tvar value *[]byte\n\t\t\tresults[i] = &value\n\t\tcase timeType.Kind():\n\t\t\tvar value *time.Time\n\t\t\tresults[i] = &value\n\t\tcase gocqlUUIDType.Kind():\n\t\t\tvar value *gocql.UUID\n\t\t\tresults[i] = &value\n\t\tcase reflect.Ptr:\n\t\t\t// Special case for custom optional string type:\n\t\t\t// string type used in Cassandra\n\t\t\t// converted to/from custom type in ORM layer\n\t\t\tif typ == reflect.TypeOf(&base.OptionalString{}) {\n\t\t\t\tvar value *string\n\t\t\t\tresults[i] = &value\n\t\t\t\tbreak\n\t\t\t}\n\t\t\t// Special case for custom optional int type:\n\t\t\t// int64 type used in Cassandra\n\t\t\t// converted to/from custom type in ORM layer\n\t\t\tif typ == reflect.TypeOf(&base.OptionalUInt64{}) {\n\t\t\t\tvar value *int64\n\t\t\t\tresults[i] = &value\n\t\t\t\tbreak\n\t\t\t}\n\t\t\t// for unrecognized pointer types, fall back to default logging\n\t\t\tfallthrough\n\t\tdefault:\n\t\t\t// This should only happen if we start using a new cassandra type\n\t\t\t// without adding to the translation layer\n\t\t\tlog.WithFields(log.Fields{\"type\": typ.Kind(), \"column\": column}).\n\t\t\t\tInfof(\"type not found\")\n\t\t}\n\t}\n\n\treturn results\n}", "func (q *Query) buildSlice(sliceResult []int, t *Table, rt reflect.Type, sliceRv reflect.Value) (e error) {\n\tif len(sliceResult) < 1 {\n\t\treturn\n\t}\n\tfor _, pk := range sliceResult {\n\t\trv := reflect.New(rt)\n\t\tif pk < 1 {\n\t\t\tcontinue\n\t\t}\n\t\t_, iftValue, e := t.Get(pk)\n\t\tif e != nil {\n\t\t\treturn e\n\t\t}\n\t\tif iftValue == nil {\n\t\t\tcontinue\n\t\t}\n\t\te = q.s.saver.ToStruct(iftValue, rv.Interface())\n\t\tif e != nil {\n\t\t\treturn e\n\t\t}\n\t\tsliceRv.Elem().Set(reflect.Append(sliceRv.Elem(), rv))\n\t}\n\treturn\n}", "func buildSelectPlan(sel *sqlparser.Select, vschema VSchema) (primitive engine.Primitive, err error) {\n\tbindvars := sqlparser.GetBindvars(sel)\n\tbuilder, err := processSelect(sel, vschema, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tjt := newJointab(bindvars)\n\terr = builder.Wireup(builder, jt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn builder.Primitive(), nil\n}", "func SelectAll(db *sql.DB, table interface{}, args ...interface{}) ([]interface{}, error) {\n\ttypes := reflect.TypeOf(table)\n\tif types.Kind() != reflect.Ptr && types.Kind() != reflect.Struct {\n\t\treturn nil, fmt.Errorf(\"you should give a struct or a ptr of a struct\")\n\t}\n\tif types.Kind() == reflect.Ptr && types.Elem().Kind() != reflect.Struct {\n\t\treturn nil, fmt.Errorf(\"you should give a struct or a ptr of a struct\")\n\t}\n\tvar rltype reflect.Type\n\tif types.Kind() == reflect.Struct {\n\t\trltype = types\n\t} else {\n\t\trltype = types.Elem()\n\t}\n\tsqlcmd := \"select * from \" + rltype.Name()\n\tif len(args) > 0 {\n\t\tif reflect.TypeOf(args[0]).Kind() != reflect.String {\n\t\t\treturn nil, fmt.Errorf(\"args[0] should is a slice\")\n\t\t}\n\t\tsqlcmd += \" \" + args[0].(string)\n\t}\n\n\tvar (\n\t\trows *sql.Rows\n\t\terr error\n\t)\n\tif len(args) > 1 {\n\t\trows, err = db.Query(sqlcmd, args[1:]...)\n\t} else {\n\t\trows, err = db.Query(sqlcmd)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer rows.Close()\n\n\treturn readRows(rows, rltype)\n}", "func (q ReleaseGroupQuery) Select(b sq.SelectBuilder) ReleaseGroupQuery {\n\tq.builder = b\n\treturn q\n}", "func (db *DB) doSelectOrWithReturning(query string, arguments []interface{}, recordDescription *recordDescription, pointersGetter pointersGetter) (int64, error) {\n\trows, columns, err := db.executeQuery(query, arguments, false, false)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tdefer rows.Close()\n\n\t// If the given slice is empty, the slice grows as the rows are read.\n\t// If the given slice isn't empty it's filled with rows, and both rows and\n\t// slice length have to be equals.\n\t// If it's a single instance, it's juste filled, and the result must have\n\t// only one row.\n\tvar rowsCount int\n\tif recordDescription.len() > 0 {\n\t\trowsCount, err = db.fillWithValues(recordDescription, pointersGetter, columns, rows)\n\t} else {\n\t\trowsCount, err = db.growAndFillWithValues(recordDescription, pointersGetter, columns, rows)\n\t}\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn 0, err\n\t}\n\n\terr = rows.Err()\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t}\n\treturn int64(rowsCount), err\n}", "func (r *QueryResult) Next(dest []driver.Value) error {\n\tif len(r.page.Columns) == 0 {\n\t\tr.close()\n\t\treturn io.EOF\n\t}\n\trowCount := int32(len(r.page.Columns[0]))\n\tif r.index >= rowCount {\n\t\tif r.page.Last {\n\t\t\tr.close()\n\t\t\treturn io.EOF\n\t\t}\n\t\tctx, cancel := r.contextWithCancel()\n\t\tdefer cancel()\n\t\tif err := r.fetchNextPage(ctx); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tfor i := 0; i < len(r.page.Columns); i++ {\n\t\tcol := r.page.Columns[i]\n\t\t// TODO: find out the reason for requiring the following fix\n\t\tif len(col) <= int(r.index) {\n\t\t\tr.close()\n\t\t\treturn io.EOF\n\t\t}\n\t\tdest[i] = col[r.index]\n\t}\n\tr.index++\n\treturn nil\n}", "func (e *SqlExecutor) Select(i interface{}, query string, args ...interface{}) ([]interface{}, error) {\n\thook := e.db.ExecutorHook()\n\thook.BeforeSelect(e.ctx, query, args...)\n\tv, err := e.SqlExecutor.Select(i, query, args...)\n\thook.AfterSelect(e.ctx, query, args...)\n\treturn v, err\n}" ]
[ "0.6446001", "0.63682157", "0.620896", "0.59297305", "0.5906448", "0.58667505", "0.58301604", "0.5821693", "0.5815763", "0.5595056", "0.5586066", "0.55711603", "0.5556469", "0.55466646", "0.5514916", "0.5478851", "0.5471615", "0.5434563", "0.542008", "0.5377258", "0.5376172", "0.53716177", "0.53464466", "0.5343583", "0.5332242", "0.5325894", "0.53001714", "0.52908975", "0.5290586", "0.5262423", "0.5221622", "0.5219287", "0.5190563", "0.51872075", "0.5184345", "0.51509154", "0.51231426", "0.51225364", "0.51209235", "0.51192987", "0.51040065", "0.50932574", "0.50900644", "0.5034819", "0.5020297", "0.5016551", "0.50126034", "0.5011983", "0.50116044", "0.4993576", "0.49920574", "0.49868748", "0.49812916", "0.4975316", "0.4962711", "0.4954644", "0.4952589", "0.49299827", "0.49296722", "0.49200976", "0.49166438", "0.491394", "0.49084538", "0.49075848", "0.48951104", "0.489162", "0.48838973", "0.48773366", "0.48747185", "0.4865345", "0.4862792", "0.48626703", "0.48442802", "0.48435903", "0.4829771", "0.48288167", "0.48186764", "0.48173782", "0.48173434", "0.48146746", "0.48144042", "0.48123533", "0.48096266", "0.48045865", "0.47954974", "0.4785195", "0.4782688", "0.47768652", "0.47733125", "0.4773084", "0.47697595", "0.4767179", "0.4759803", "0.4759395", "0.47588798", "0.47563478", "0.4745628", "0.47452962", "0.47442776", "0.4744074" ]
0.6437408
1
exec executes the given query using positional arguments, automatically rebinding for the db.
func (s *SQLStore) exec(e execer, sqlString string, args ...interface{}) (sql.Result, error) { sqlString = s.db.Rebind(sqlString) return e.Exec(sqlString, args...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (w *Wrapper) exec(query string, args ...interface{}) (sql.Result, error) {\n\tw.connLock.RLock()\n\tdefer w.connLock.RUnlock()\n\n\treturn w.connection.Exec(w.prepare(query), args...)\n}", "func execQuery(db *sql.DB, sqlQuery string, args ...interface{}) {\n\t_, dbErr := db.Exec(sqlQuery, args...)\n\tif dbErr != nil {\n\t\tLogErr(\"%s\\n\", dbErr)\n\t}\n}", "func (this *Database) Exec(sql string, args ...interface{}) {\n\t//fmt.Printf(\"About to EXEC\\n\")\n\t_, err := this.db.Exec(sql, args...)\n\t//fmt.Printf(\"Done EXEC\\n\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (c *conn) Exec(query string, args []driver.Value) (driver.Result, error) {\n\treturn c.exec(context.Background(), query, toNamedValues(args))\n}", "func (c *connImpl) Exec(query string, args []driver.Value) (driver.Result, error) {\n\t// http://code.google.com/p/go-wiki/wiki/InterfaceSlice\n\ttmp := make([]interface{}, len(args))\n\tfor i, arg := range args {\n\t\ttmp[i] = arg\n\t}\n\tif err := c.c.Exec(query, tmp...); err != nil {\n\t\treturn nil, err\n\t}\n\treturn c, nil // FIXME RowAffected/noRows\n}", "func (database *Database) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn database.ExecContext(context.Background(), query, args...)\n}", "func (s *Schema) Exec(query string, args ...interface{}) (sql.Result, error) {\n\n\tvar e error\n\tvar stmt *sql.Stmt\n\tstmt, e = s.Dal.Connection.Prepare(query)\n\t// fmt.Printf(\"Stmt: %v\\n\", stmt)\n\tif e != nil {\n\t\tfmt.Printf(\"Error: %s\", e.Error())\n\t\treturn nil, e\n\t}\n\tdefer stmt.Close()\n\treturn stmt.Exec(args...)\n}", "func Exec(query string, args ...interface{}) (sql.Result, error) {\n\tdb, err := sql.Open(driver, conninfo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer db.Close()\n\n\tres, err := db.Exec(query, args...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n}", "func Execute(query string, args ...interface{}) (sql.Result, error){\n result, err := db.Exec(query, args...)\n if err != nil && !debug { \n log.Println(err)\n }\n return result, err\n}", "func (d *Database) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn d.Db.Exec(query, args...)\n}", "func (c *conn) Exec(query string, args []driver.Value) (driver.Result, error) {\n\treturn c.exec(context.Background(), query, args)\n}", "func (q *Querier) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tq.logBefore(query, args)\n\tstart := time.Now()\n\tres, err := q.dbtxCtx.ExecContext(q.ctx, query, args...)\n\tq.logAfter(query, args, time.Since(start), err)\n\treturn res, err\n}", "func (s *dbase) Exec(query string, params ...interface{}) error {\n\t_, err := s.connection.Exec(query, params...)\n\n\treturn err\n}", "func (q *Querier) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tq.logBefore(query, args)\n\tstart := time.Now()\n\n\tdbtxCtx := q.selectDBTXContext(query)\n\tres, err := dbtxCtx.ExecContext(q.ctx, query, args...)\n\tq.logAfter(query, args, time.Since(start), err)\n\treturn res, err\n}", "func (db *DB) Exec(name string, args ...interface{}) (pgx.CommandTag, error) {\n\tq, err := db.qm.getQuery(name)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn db.Pool.Exec(q.getSQL(), args...)\n}", "func (d *sqlDB) Exec(query string, args ...interface{}) (result sql.Result, err error) {\n\terr = d.retry(func() error {\n\t\tresult, err = d.conn.Exec(query, args...)\n\t\treturn err\n\t})\n\treturn\n}", "func (conn *dbconn) exec(ctx context.Context, name string, params []interface{}) (result sql.Result, err error) {\n\n\t// TODO: check Parameters for injection here\n\t// TODO: update the params parameter to use an interface type? So that Parameters to the sproc can have the types validated?\n\n\t// Build the command\n\tvar command string\n\tif command, err = conn.buildSprocCall(name, params); err == nil {\n\n\t\t// Verify that a valid command was created\n\t\tif len(command) > 0 {\n\t\t\tresult, err = conn.db.ExecContext(ctx, command, params...)\n\t\t} else {\n\t\t\terr = errors.Errorf(\"Invalid command creation for stored Procedure %s\", name)\n\t\t}\n\t} else {\n\t\terr = errors.Errorf(\"Error while building execution call for stored Procedure %s: %v\", name, err)\n\t}\n\treturn result, err\n}", "func (conn *n1qlConn) Exec(query string, args ...interface{}) (godbc.Result, error) {\n\n\tif len(args) > 0 {\n\t\tvar argCount int\n\t\tquery, argCount = prepareQuery(query)\n\t\tif argCount != len(args) {\n\t\t\treturn nil, fmt.Errorf(\"Argument count mismatch %d != %d\", argCount, len(args))\n\t\t}\n\t\tquery, args = preparePositionalArgs(query, argCount, args)\n\t}\n\n\treturn conn.performExec(query, nil)\n}", "func (this *Database) Exec(sql string, args ...interface{}) sql.Result {\n\t//fmt.Printf(\"About to EXEC\\n\")\n\tresult, err := this.db.Exec(sql, args...)\n\t//fmt.Printf(\"Done EXEC\\n\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn result\n}", "func (s *stmt) Exec(args []driver.Value) (driver.Result, error) {\n\treturn s.c.Exec(s.query, args)\n}", "func (s *Shard) Exec(query string, args ...interface{}) (sql.Result, errs.Err) {\n\tfixArgs(args)\n\tres, stdErr := s.sqlConn.Exec(query, args...)\n\tif stdErr != nil {\n\t\treturn nil, errs.Wrap(stdErr, errInfo(\"Exec sqlConn.Exec() error\", query, args))\n\t}\n\treturn res, nil\n}", "func (q *QueryGoPg) Exec(sql string, args ...interface{}) (err error) {\n\t_, err = goPgConnection.Exec(sql, args...)\n\treturn\n}", "func (s SQLite) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tif s.DB == nil {\n\t\treturn nil, fmt.Errorf(\"db is not created\")\n\t}\n\treturn s.DB.Exec(query, args...)\n}", "func (extDb *Database) Exec(query string, args ...interface{}) (sql.Result, error) {\n\terr := extDb.checkStatus()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdb := extDb.getDb()\n\tif db == nil {\n\t\treturn nil, ErrNotInitialized\n\t}\n\tresult, err := db.Exec(query, args...)\n\tif err != nil {\n\t\tif !extDb.i.isConnectionError(err) {\n\t\t\treturn nil, err\n\t\t}\n\t\tlog.Info(\"Database exec error \", err)\n\t\terrConn := extDb.Reconnect()\n\t\tif errConn == nil {\n\t\t\tdb := extDb.getDb()\n\t\t\tif db == nil {\n\t\t\t\treturn nil, ErrNotInitialized\n\t\t\t}\n\t\t\tresult, err = db.Exec(query, args...)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, errConn\n\t}\n\treturn result, nil\n}", "func exec(stmt *sql.Stmt, args ...interface{}) error {\n\t_, err := stmt.Exec(args...)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"mysql: could not execute statement: %v\", err)\n\t}\n\treturn nil\n}", "func (c *sqlmock) Exec(query string, args []driver.Value) (driver.Result, error) {\n\tnamedArgs := make([]driver.NamedValue, len(args))\n\tfor i, v := range args {\n\t\tnamedArgs[i] = driver.NamedValue{\n\t\t\tOrdinal: i + 1,\n\t\t\tValue: v,\n\t\t}\n\t}\n\n\tex, err := c.exec(query, namedArgs)\n\tif ex != nil {\n\t\ttime.Sleep(ex.delay)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ex.result, nil\n}", "func (lq LoggingQueryable) Exec(query string, args ...interface{}) (result sql.Result, err error) {\n\tresult, err = lq.Q.Exec(query, args...)\n\tlog.Printf(\"SQL: Exec(%v, %v) -> %v\\n\", query, args, err)\n\treturn result, err\n}", "func (db *DB) Exec(query interface{}, params ...interface{}) (orm.Result, error) {\n\tvar res orm.Result\n\tvar err error\n\ttracing.Trace(db.inner.Context(), query, func() error {\n\t\tif db.tx != nil {\n\t\t\tres, err = db.tx.Exec(query, params...)\n\t\t} else {\n\t\t\tres, err = db.inner.Exec(query, params...)\n\t\t}\n\t\treturn err\n\t})\n\treturn res, err\n}", "func (s *Stmt) Exec(args []driver.Value) (driver.Result, error) {\n\t// Binds all arguments.\n\tif err := s.Bind(args); err != nil {\n\t\treturn nil, err\n\t}\n\t// Executes query.\n\tswitch s.p.(type) {\n\tcase parser.CreateViewStmt:\n\t\treturn NewCreateViewStmt(s).Exec()\n\t}\n\treturn s.si.Exec(args)\n}", "func (ng *Engine) exec(ctx context.Context, q *query) (model.Value, error) {\n\tcurrentQueries.Inc()\n\tdefer currentQueries.Dec()\n\tctx, cancel := context.WithTimeout(ctx, ng.options.Timeout)\n\tq.cancel = cancel\n\n\tqueueTimer := q.stats.GetTimer(stats.ExecQueueTime).Start()\n\n\tif err := ng.gate.Start(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\tdefer ng.gate.Done()\n\n\tqueueTimer.Stop()\n\n\t// Cancel when execution is done or an error was raised.\n\tdefer q.cancel()\n\n\tconst env = \"query execution\"\n\n\tevalTimer := q.stats.GetTimer(stats.TotalEvalTime).Start()\n\tdefer evalTimer.Stop()\n\n\t// The base context might already be canceled on the first iteration (e.g. during shutdown).\n\tif err := contextDone(ctx, env); err != nil {\n\t\treturn nil, err\n\t}\n\n\tswitch s := q.Statement().(type) {\n\tcase *EvalStmt:\n\t\treturn ng.execEvalStmt(ctx, q, s)\n\tcase testStmt:\n\t\treturn nil, s(ctx)\n\t}\n\n\tpanic(fmt.Errorf(\"promql.Engine.exec: unhandled statement of type %T\", q.Statement()))\n}", "func ExecDB(query string, args ...interface{}) sql.Result {\n\trwlock.Lock()\n\tdefer rwlock.Unlock()\n\tvar res sql.Result\n\tstmt, err := db.Prepare(query)\n\tif err != nil {\n\t\tfmt.Println(\"Error creating prepared statement (my_server.go: ExecDB)\")\n\t\tfmt.Println(err)\n\t\treturn res\n\t}\n\n\tres, err = stmt.Exec(args...)\n\tif err != nil {\n\t\tfmt.Println(\"Error executing prepared statement (my_server.go: ExecDB)\")\n\t\tfmt.Println(err)\n\t}\n\treturn res\n}", "func Exec(x Execer, query string, args ...interface{}) (res sql.Result, err error) {\n\tres, err = x.Exec(query, args...)\n\terr = interpretError(err)\n\treturn\n}", "func (s *stmt) Exec(args []driver.Value) (driver.Result, error) {\n\treturn s.exec(context.Background(), toNamedValues(args))\n}", "func (db *DB) NamedExec(query string, arg interface{}) (sql.Result, error) {\n return NamedExec(db, query, arg)\n}", "func (c Conn) Exec(ctx context.Context, query string, args, v any) error {\n\targv, ok := args.([]any)\n\tif !ok {\n\t\treturn fmt.Errorf(\"dialect/sql: invalid type %T. expect []any for args\", v)\n\t}\n\tswitch v := v.(type) {\n\tcase nil:\n\t\tif _, err := c.ExecContext(ctx, query, argv...); err != nil {\n\t\t\treturn err\n\t\t}\n\tcase *sql.Result:\n\t\tres, err := c.ExecContext(ctx, query, argv...)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t*v = res\n\tdefault:\n\t\treturn fmt.Errorf(\"dialect/sql: invalid type %T. expect *sql.Result\", v)\n\t}\n\treturn nil\n}", "func (db *DB) do(query string, arguments []interface{}) (sql.Result, error) {\n\tquery = db.replacePlaceholders(query)\n\n\t// Execute the statement\n\tstartTime := time.Now()\n\tqueryable, err := db.getQueryable(query)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn nil, err\n\t}\n\tresult, err := queryable.Exec(arguments...)\n\tconsumedTime := timeElapsedSince(startTime)\n\tdb.addConsumedTime(consumedTime)\n\tdb.logExecution(consumedTime, query, arguments)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\tif db.useErrorParser {\n\t\t\treturn nil, db.adapter.ParseError(err)\n\t\t}\n\t\treturn nil, err\n\t}\n\n\treturn result, err\n}", "func (e *SqlExecutor) Exec(query string, args ...interface{}) (sql.Result, error) {\n\thook := e.db.ExecutorHook()\n\thook.BeforeExec(e.ctx, query, args...)\n\tv, err := e.SqlExecutor.Exec(query, args...)\n\thook.AfterExec(e.ctx, query, args...)\n\treturn v, err\n}", "func (p PgxWrapper) Exec(ctx context.Context, queryName, sql string, args ...interface{}) error {\n\tspan, _ := opentracing.StartSpanFromContext(ctx, queryName)\n\tspan.Finish()\n\n\tif _, err := p.pool.Exec(ctx, sql, args...); err != nil {\n\t\treturn fmt.Errorf(\"could not execute query: %w\", err)\n\t}\n\n\treturn nil\n}", "func (t *Transaction) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tt.dbmap.trace(query, args)\n\treturn t.tx.Exec(query, args...)\n}", "func (tx *Transaction) Exec(query string, args ...interface{}) error {\n\treturn exec(tx.tx, query, args...)\n}", "func (db *DB) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn db.master.Exec(query, args...)\n}", "func (db *DB) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn db.Master().Exec(query, args...)\n}", "func (s *stmt) Exec(args []driver.Value) (driver.Result, error) {\n\tlist := driverValueToNamedValue(args)\n\treturn s.exec(context.Background(), list)\n}", "func (ex *Execer) execSQL(fullSQL string, args []interface{}) (sql.Result, error) {\n\tdefer logExecutionTime(time.Now(), fullSQL, args)\n\n\tvar result sql.Result\n\tvar err error\n\tresult, err = ex.database.Exec(fullSQL, args...)\n\tif err != nil {\n\t\treturn nil, logSQLError(err, \"execSQL.30\", fullSQL, args)\n\t}\n\n\treturn result, nil\n}", "func (m *SchemaData) Exec(query string) (sql.Result, error) {\n\tlog.Println(\"[SQL]\", \"[\"+m.dbType+\"]\", query)\n\treturn nil, nil\n}", "func Execute(db *sql.DB, query string, args ...interface{}) error {\n\treturn crdb.Execute(func() error {\n\t\t_, err := db.Exec(query, args...)\n\t\treturn err\n\t})\n}", "func (stmt *statement) Exec(ctx context.Context, db Executor) (sql.Result, error) {\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\treturn db.ExecContext(ctx, stmt.String(), stmt.args...)\n}", "func (q *Query) Exec() error {\n\treturn q.query.Exec()\n}", "func (s *Server) Exec(c *Conn, statement string, args Args) (interface{}, error) {\n\treq := c.acquireRequest(statement, args)\n\tdefer c.releaseRequest(req)\n\treturn s.engine.handlers.exec(req)\n}", "func (q *query) Exec(ctx context.Context) *Result {\n\tif span := opentracing.SpanFromContext(ctx); span != nil {\n\t\tspan.SetTag(queryTag, q.stmt.String())\n\t}\n\n\tres, err := q.ng.exec(ctx, q)\n\treturn &Result{Err: err, Value: res}\n}", "func (db TestDB) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn db.testTx.Exec(query, args...)\n}", "func (m *DBMock) Exec(ctx context.Context, query string, params ...interface{}) (pgconn.CommandTag, error) {\n\targs := m.Called(append([]interface{}{query}, params...)...)\n\treturn nil, args.Error(0)\n}", "func ExecQuery(ctx context.Context, sql string) {\n\t_, span := opentelemetry.StartSpan(ctx, \"database-call\")\n\tdefer span.Finish(0)\n\n\t// Do the query execution\n}", "func (stmt *statement) ExecContext(ctx context.Context, args []driver.NamedValue) (driver.Result, error) {\n\treturn stmt.conn.ExecContext(ctx, stmt.query, args)\n}", "func (st *ogonoriStmt) Exec(args []driver.Value) (driver.Result, error) {\n\togl.Debugln(\"** ogonoriStmt.Exec\")\n\tif st.conn == nil || st.conn.dbc == nil {\n\t\treturn nil, oerror.ErrInvalidConn{Msg: \"obinary.DBClient not initialized in ogonoriStmt#Exec\"}\n\t}\n\n\treturn doExec(st.conn.dbc, st.query, args)\n}", "func (p *Pool) ExecuteWithBindings(query string, bindings, rebindings map[string]string) (resp []Response, err error) {\n\tpc, err := p.Get()\n\tif err != nil {\n\t\tfmt.Printf(\"Error aquiring connection from pool: %s\", err)\n\t\treturn nil, err\n\t}\n\tdefer pc.Close()\n\treturn pc.Client.ExecuteWithBindings(query, bindings, rebindings)\n}", "func (c *ConnWrapper) Exec(query string, args []driver.Value) (res driver.Result, err error) {\n\tspan, _ := opentracing.StartSpanFromContext(\n\t\temptyCtx,\n\t\tc.integration.getOperationName(c.dsn),\n\t)\n\n\tdefer span.Finish()\n\n\trawSpan, ok := tracer.GetRaw(span)\n\tif ok {\n\t\tc.integration.beforeCall(query, rawSpan, c.dsn)\n\t}\n\ttracer.OnSpanStarted(span)\n\n\tif execer, ok := c.Conn.(driver.Execer); ok {\n\t\tres, err = execer.Exec(query, args)\n\t\tif err != nil {\n\t\t\tutils.SetSpanError(span, err)\n\t\t}\n\t\treturn\n\t}\n\n\treturn nil, driver.ErrSkip\n}", "func (ar *ActiveRecord) Exec() (sql.Result, error) {\n\treturn ar.DB.Exec(ar.ExecString(), ar.Args...)\n}", "func (c *PostgreSQLConnection) Execute(query string, arguments []interface{}) sql.Result {\n\tif arguments == nil {\n\t\tresult, err := c.db.Exec(query)\n\n\t\tif err != nil {\n\t\t\tlog.Printf(\"[!] Couldn't execute query. Reason %v\", err)\n\t\t\treturn nil\n\t\t}\n\n\t\treturn result\n\t}\n\n\tstmt, err := c.db.Prepare(query)\n\n\tif err != nil {\n\t\tlog.Printf(\"[!] Couldn't prepare statement. Reason %v\", err)\n\t\treturn nil\n\t}\n\n\tresult, err := stmt.Exec(arguments...)\n\n\tif err != nil {\n\t\tlog.Printf(\"[!] Couldn't execute query. Reason %v\", err)\n\t\treturn nil\n\t}\n\n\treturn result\n}", "func (db *DB) executeQuery(query string, arguments []interface{}, noTx, noStmtCache bool) (*sql.Rows, []string, error) {\n\tquery = db.replacePlaceholders(query)\n\n\tstartTime := time.Now()\n\tqueryable, err := db.getQueryableWithOptions(query, noTx, noStmtCache)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn nil, nil, err\n\t}\n\trows, err := queryable.Query(arguments...)\n\tconsumedTime := timeElapsedSince(startTime)\n\tdb.addConsumedTime(consumedTime)\n\tdb.logExecution(consumedTime, query, arguments)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn nil, nil, err\n\t}\n\n\tcolumns, err := rows.Columns()\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\trows.Close()\n\t\treturn nil, nil, err\n\t}\n\n\treturn rows, columns, nil\n}", "func (w *Wrapper) query(query string, args ...interface{}) (*sql.Rows, error) {\n\tw.connLock.RLock()\n\tdefer w.connLock.RUnlock()\n\n\treturn w.connection.Query(w.prepare(query), args...)\n}", "func (this *dataStore) Exec(queryName string, arg interface{}) (int64, error) {\r\n\r\n\tif stmt, err := this.namedStmt(queryName, arg); err != nil {\r\n\t\treturn 0, err\r\n\t} else if res, err := stmt.NamedStmt.Exec(arg); err != nil {\r\n\t\treturn 0, err\r\n\t} else if stmt.query.Command == `insert` {\r\n\t\treturn res.LastInsertId()\r\n\t} else {\r\n\t\treturn res.RowsAffected()\r\n\t}\r\n}", "func (m *MSSQLDatastore) Exec(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\tif m == nil {\n\t\treturn nil, ErrEmptyObject\n\t}\n\n\tif _, ok := ctx.Deadline(); !ok {\n\t\tvar cancel context.CancelFunc\n\t\tctx, cancel = context.WithTimeout(ctx, QueryLimit)\n\t\tdefer cancel()\n\t}\n\n\treturn m.db.ExecContext(ctx, query, args...)\n}", "func Exec(sql string) (sql.Result, error) {\n\treturn database.Exec(sql)\n}", "func (c *Conn) handleExec(query string, args []driver.Value) (driver.Result, error) {\n\tvar (\n\t\tb []byte\n\t\terr error\n\t)\n\n\t// reset the protocol packet sequence number\n\tc.resetSeqno()\n\n\tif b, err = c.createComQuery(replacePlaceholders(query, args)); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// send COM_QUERY to the server\n\tif err := c.writePacket(b); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.handleExecResponse()\n}", "func (mc *MysqlConn) exec(query string) error {\n\t// Send command\n\terr := mc.writeCommandPacketStr(COM_QUERY, query)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Read Result\n\tresLen, err := mc.readResultSetHeaderPacket()\n\tif err == nil && resLen > 0 {\n\t\tif err = mc.readUntilEOF(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = mc.readUntilEOF()\n\t}\n\n\treturn err\n}", "func (a ArticleDB) Exec(sess *sqlabble.Session, st stmt.Statement) (sql.Result, error) {\n\tquery, values := sess.Builder.Build(st)\n\tresult, err := sess.Exec(query, values...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result, nil\n}", "func Exec(a config.AppContext, q interpreter.Query) ([]map[string]interface{}, error) {\n\t/*\n\t * Based on the number of tables avaiable in the query, we will execute the same\n\t */\n\n\tif len(q.Tables) == 0 {\n\t\treturn nil, errors.New(\"Couldn't find the table to be queried from\")\n\t}\n\n\tif len(q.Tables) == 1 {\n\t\treturn SingleTableMode(a, q)\n\t}\n\n\treturn nil, errors.New(\"multiple table join query not supported yet\")\n}", "func (mc *mysqlConn) exec(query string) (e error) {\n\t// Send command\n\te = mc.writeCommandPacket(COM_QUERY, query)\n\tif e != nil {\n\t\treturn\n\t}\n\n\t// Read Result\n\tresLen, e := mc.readResultSetHeaderPacket()\n\tif e != nil {\n\t\treturn\n\t}\n\n\tmc.affectedRows = 0\n\tmc.insertId = 0\n\n\tif resLen > 0 {\n\t\t_, e = mc.readUntilEOF()\n\t\tif e != nil {\n\t\t\treturn\n\t\t}\n\n\t\tmc.affectedRows, e = mc.readUntilEOF()\n\t\tif e != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\treturn\n}", "func (s *stmt) ExecContext(ctx context.Context, args []driver.NamedValue) (driver.Result, error) {\n\treturn s.c.ExecContext(ctx, s.query, args)\n}", "func (c *Conn) ExecContext(ctx context.Context, query string, args []driver.NamedValue) (driver.Result, error) {\n\trawStmt, err := c.Prepare(query)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"prepare: %w\", err)\n\t}\n\n\tif stmt, ok := rawStmt.(*Stmt); ok {\n\t\treturn stmt.ExecContext(ctx, args)\n\t}\n\n\treturn nil, fmt.Errorf(\"cannot execute Stmt of type %T, expected %T\", rawStmt, &Stmt{})\n}", "func (p PostgreSQL) Exec(sql string, args ...interface{}) (pgconn.CommandTag, error) {\n\tif p.Connection == nil {\n\t\treturn nil, fmt.Errorf(\"connection is not created\")\n\t}\n\n\treturn p.Connection.Exec(context.TODO(), sql, args...)\n}", "func (d *Database) Execute(query string, args ...interface{}) (sql.Result, error) {\n\tvar result sql.Result\n\n\tstmtIns, err := d.Conn.Prepare(query)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tdefer stmtIns.Close()\n\n\tresult, err = stmtIns.Exec(args...)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\n\treturn result, nil\n}", "func (p *PgSQL) Execute(query string, args ...interface{}) (sql.Result, error) {\n\tvar result sql.Result\n\n\tstmtIns, err := p.Connection.Prepare(query)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\tdefer stmtIns.Close()\n\n\tresult, err = stmtIns.Exec(args...)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\n\treturn result, nil\n}", "func (s *Stmt) Exec(args []driver.Value) (driver.Result, error) {\n\t// in transaction, need run Executor\n\texecutor, err := exec.BuildExecutor(s.res.dbType, s.txCtx.TransactionMode, s.query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\texecCtx := &types.ExecContext{\n\t\tTxCtx: s.txCtx,\n\t\tQuery: s.query,\n\t\tValues: args,\n\t}\n\n\tret, err := executor.ExecWithValue(context.Background(), execCtx,\n\t\tfunc(ctx context.Context, query string, args []driver.NamedValue) (types.ExecResult, error) {\n\t\t\tret, err := s.stmt.Exec(util.NamedValueToValue(args))\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\treturn types.NewResult(types.WithResult(ret)), nil\n\t\t})\n\n\treturn ret.GetResult(), err\n}", "func (c *Conn) Exec(cmd string, args ...interface{}) error {\n\tfor len(cmd) > 0 {\n\t\ts, err := c.prepare(cmd)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t} else if s.stmt == nil {\n\t\t\t// this happens for a comment or white-space\n\t\t\tcmd = s.tail\n\t\t\tcontinue\n\t\t}\n\t\tvar subargs []interface{}\n\t\tcount := s.BindParameterCount()\n\t\tif len(s.tail) > 0 && len(args) >= count {\n\t\t\tsubargs = args[:count]\n\t\t\targs = args[count:]\n\t\t} else {\n\t\t\tsubargs = args\n\t\t}\n\t\terr = s.Exec(subargs...)\n\t\tif err != nil {\n\t\t\t_ = s.finalize()\n\t\t\treturn err\n\t\t}\n\t\tif err = s.finalize(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tcmd = s.tail\n\t}\n\treturn nil\n}", "func (database *Database) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\treturn database.db.ExecContext(ctx, query, args...)\n}", "func (t *Tx) Exec(query string, args ...interface{}) (int64, error) {\n\treturn rowsChanged(t.Tx.Exec(query, args...))\n}", "func PythonODBCExec(\n\tcfg Config,\n\tquery string,\n) (string, error) {\n\tapplicationintent := \"readwrite\"\n\tif cfg.ReadOnly {\n\t\tapplicationintent = \"readonly\"\n\t}\n\n\targs := []string{\n\t\t\"--server\", fmt.Sprintf(\"%s,%s\", cfg.Host, cfg.Port),\n\t\t\"--username\", cfg.Username,\n\t\t\"--password\", cfg.Password,\n\t\t\"--query\", query,\n\t\t\"--application-intent\", applicationintent,\n\t}\n\n\tif db := cfg.Database; db != \"\" {\n\t\targs = append(args, \"--database\", db)\n\t}\n\n\tout, err := exec.Command(\n\t\t\"./client/odbc_client.py\",\n\t\targs...,\n\t).Output()\n\n\tif err != nil {\n\t\tif exitErrr, ok := err.(*exec.ExitError); ok {\n\t\t\treturn \"\", errors.New(string(exitErrr.Stderr))\n\t\t}\n\n\t\treturn \"\", err\n\t}\n\n\treturn string(out), nil\n}", "func (db *MockDB) Exec(ctx context.Context, q string, args ...interface{}) (sql.Result, error) {\n\tdb.ExecCount++\n\n\tif !assert.True(db.t, len(db.ExecExpected) > 0, \"No ExecExpected Defined\") {\n\t\tdb.t.FailNow()\n\t}\n\n\t// We repeat the final registered Exec mock if we run out.\n\tif db.ExecPointer >= len(db.ExecExpected) {\n\t\tfmt.Print(\"\\nMore Exec Calls than Expected\\n\\n\")\n\t\tdb.t.FailNow()\n\t}\n\n\texec := db.ExecExpected[db.ExecPointer]\n\tdb.ExecPointer++\n\n\tif !assert.Equal(db.t, exec.Query, q) {\n\t\tdb.t.FailNow()\n\t}\n\n\tassertDeepEqual(db.t, exec.Args, args)\n\n\tif exec.Error != nil {\n\t\treturn nil, exec.Error\n\t}\n\n\treturn &exec.Result, nil\n}", "func (db *DatabaseConnection) query(stmt mysql.Stmt, args ...interface{}) mysql.Result {\n\treturn db.exec(stmt, args...)\n}", "func (db *DB) Exec(ctx context.Context, query SQLMutation, args ...interface{}) (sql.Result, error) {\n\tif err := query.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn db.master.ExecContext(ctx, query.String(), args...)\n}", "func (p *Param) Exec() (sql.Result, error) {\n\treturn p.T().Exec(p)\n}", "func (c *Conn) ExecContext(ctx context.Context, q string, args []driver.NamedValue) (driver.Result, error) {\n\tsession, err := c.OpenSession(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\ttmpl := template(q)\n\tstmt := statement(tmpl, args)\n\treturn exec(ctx, session, stmt)\n}", "func (s *Session) Exec() (result sql.Result, err error) {\n\tdefer s.Clear()\n\tlog.Info(s.sql.String(), s.sqlVars)\n\tif result, err = s.DB().Exec(s.sql.String(), s.sqlVars...); err != nil {\n\t\tlog.Error(err)\n\t}\n\treturn\n}", "func (q *Querier) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\treturn q.WithContext(ctx).Exec(query, args...)\n}", "func (q *Querier) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\treturn q.WithContext(ctx).Exec(query, args...)\n}", "func (w *Wrapper) executeQuery() (res sql.Result, err error) {\n\tw.cleanBefore()\n\tw.buildQuery()\n\tw.LastQuery = w.query\n\tw.LastParams = w.params\n\t// Calculate the execution time.\n\tvar start time.Time\n\tif w.tracing {\n\t\tstart = time.Now()\n\t}\n\t// Execute the query if the wrapper is executable.\n\tif w.executable {\n\t\tvar stmt *sql.Stmt\n\t\tvar count int64\n\t\tstmt, err = w.db.Prepare(w.query)\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tres, err = stmt.Exec(w.params...)\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tw.LastResult = res\n\t\tcount, err = res.RowsAffected()\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tw.count = int(count)\n\t\terr = stmt.Close()\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t}\n\tif w.tracing {\n\t\tw.saveTrace(err, w.query, start)\n\t}\n\tw.cleanAfter()\n\treturn\n}", "func (stmt *Statement) Run(params ...interface{}) (res *Result, err os.Error) {\n defer stmt.db.unlockIfError(&err)\n defer catchOsError(&err)\n stmt.db.lock()\n\n if stmt.db.conn == nil {\n return nil, NOT_CONN_ERROR\n }\n if stmt.db.unreaded_rows {\n return nil, UNREADED_ROWS_ERROR\n }\n\n // Bind parameters if any\n if len(params) != 0 {\n stmt.BindParams(params...)\n }\n\n // Send EXEC command with binded parameters\n stmt.sendCmdExec()\n // Get response\n res = stmt.db.getResponse(true)\n res.binary = true\n return\n}", "func (s *SqlDb) Exec(ctx context.Context, sql string, args ...interface{}) (r sql.Result, err error) {\n\tvar c *SqlContext\n\ti := ctx.Value(goradd.SqlContext)\n\tif i != nil {\n\t\tc = i.(*SqlContext)\n\t}\n\tlog.FrameworkDebug(\"Exec: \", sql, args)\n\n\tvar beginTime = time.Now()\n\n\tif c != nil && c.tx != nil {\n\t\tr, err = c.tx.ExecContext(ctx, sql, args...)\n\t} else {\n\t\tr, err = s.db.ExecContext(ctx, sql, args...)\n\t}\n\n\tvar endTime = time.Now()\n\n\tif c != nil && s.profiling {\n\t\tif args != nil {\n\t\t\tfor _, arg := range args {\n\t\t\t\tsql = strings.TrimSpace(sql)\n\t\t\t\tsql += fmt.Sprintf(\",\\n%#v\", arg)\n\t\t\t}\n\t\t}\n\t\tc.profiles = append(c.profiles, ProfileEntry{DbKey: s.dbKey, BeginTime: beginTime, EndTime: endTime, Typ: \"Exec\", Sql: sql})\n\t}\n\n\treturn\n}", "func (q *Query) Exec(ctx context.Context) (*QueryResult, error) {\n\tvar r QueryResult\n\n\tif q.client == nil || !q.client.Started() {\n\t\treturn &r, fmt.Errorf(\"client or db is nil\")\n\t}\n\n\tswitch q.action {\n\tcase \"select\":\n\t\trows, err := q.execSelect(ctx)\n\t\tr.Rows = rows\n\t\treturn &r, err\n\tcase \"insert\":\n\t\trows, err := q.execInsert(ctx)\n\t\tr.Rows = rows\n\t\treturn &r, err\n\tcase \"update\":\n\t\tvar err error\n\t\tif len(q.returning) == 0 {\n\t\t\tr.RowsAffected, err = q.execUpdate(ctx)\n\t\t} else {\n\t\t\tr.Rows, err = q.execUpdateR(ctx)\n\t\t}\n\t\treturn &r, err\n\tcase \"delete\":\n\t\tvar err error\n\t\tif len(q.returning) == 0 {\n\t\t\tr.RowsAffected, err = q.execDelete(ctx)\n\t\t} else {\n\t\t\tr.Rows, err = q.execDeleteR(ctx)\n\t\t}\n\t\treturn &r, err\n\tdefault:\n\t\treturn &r, fmt.Errorf(\"unsupported action %v\", q.action)\n\t}\n}", "func (db *DB) NamedExec(query string, arg interface{}) (sql.Result, error) {\n\treturn db.leader.NamedExec(query, arg)\n}", "func (ex *Execer) execFn() (sql.Result, error) {\n\tfullSQL, args, err := ex.Interpolate()\n\tif err != nil {\n\t\treturn nil, logger.Error(\"execFn.10\", \"err\", err, \"sql\", fullSQL)\n\t}\n\tdefer logExecutionTime(time.Now(), fullSQL, args)\n\n\tvar result sql.Result\n\tresult, err = ex.database.Exec(fullSQL, args...)\n\tif err != nil {\n\t\treturn nil, logSQLError(err, \"execFn.30:\"+fmt.Sprintf(\"%T\", err), fullSQL, args)\n\t}\n\n\treturn result, nil\n}", "func (s StmtWrapper) Exec(args []driver.Value) (res driver.Result, err error) {\n\tspan, _ := opentracing.StartSpanFromContext(\n\t\ts.ctx,\n\t\ts.integration.getOperationName(s.dsn),\n\t)\n\n\tdefer span.Finish()\n\n\trawSpan, ok := tracer.GetRaw(span)\n\tif ok {\n\t\ts.integration.beforeCall(s.query, rawSpan, s.dsn)\n\t}\n\ttracer.OnSpanStarted(span)\n\n\tres, err = s.Stmt.Exec(args)\n\tif err != nil {\n\t\tutils.SetSpanError(span, err)\n\t}\n\treturn\n\n}", "func (tx *Tx) Exec(ctx context.Context, query SQLMutation, args ...interface{}) (sql.Result, error) {\n\tif err := query.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn tx.parent.ExecContext(ctx, query.String(), args...)\n}", "func (s *Stmt) Query(args []driver.Value) (driver.Rows, error) {\n\texecutor, err := exec.BuildExecutor(s.res.dbType, s.txCtx.TransactionMode, s.query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\texecCtx := &types.ExecContext{\n\t\tTxCtx: s.txCtx,\n\t\tQuery: s.query,\n\t\tValues: args,\n\t}\n\n\tret, err := executor.ExecWithValue(context.Background(), execCtx,\n\t\tfunc(ctx context.Context, query string, args []driver.NamedValue) (types.ExecResult, error) {\n\t\t\tret, err := s.stmt.Query(util.NamedValueToValue(args))\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\treturn types.NewResult(types.WithRows(ret)), nil\n\t\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ret.GetRows(), nil\n}", "func Query(query string, args ...interface{}) error {\n\tstmt := database.prepare(query)\n\tdefer stmt.Close()\n\ttx := database.begin()\n\tif _, err := tx.Stmt(stmt).Exec(args...); err != nil {\n\t\tlog.Println(\"Query error: \", err)\n\t\ttx.Rollback()\n\t}\n\terr := tx.Commit()\n\treturn err\n}", "func (stmt *Statement) Exec(params ...interface{}) (\n rows []*Row, res *Result, err os.Error) {\n\n res, err = stmt.Run(params...)\n if err != nil {\n return\n }\n // Read rows\n var row *Row\n for {\n row, err = res.GetRow()\n if err != nil || row == nil {\n break\n }\n rows = append(rows, row)\n }\n return\n}", "func (c *Conn) Exec(ctx context.Context, sql string, arguments ...any) (pgconn.CommandTag, error) {\n\tif c.queryTracer != nil {\n\t\tctx = c.queryTracer.TraceQueryStart(ctx, c, TraceQueryStartData{SQL: sql, Args: arguments})\n\t}\n\n\tif err := c.deallocateInvalidatedCachedStatements(ctx); err != nil {\n\t\treturn pgconn.CommandTag{}, err\n\t}\n\n\tcommandTag, err := c.exec(ctx, sql, arguments...)\n\n\tif c.queryTracer != nil {\n\t\tc.queryTracer.TraceQueryEnd(ctx, c, TraceQueryEndData{CommandTag: commandTag, Err: err})\n\t}\n\n\treturn commandTag, err\n}", "func (ctl *ControllerWebsite) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn ctl.ExecContext(ctl.Context.GetContext(), query, args...)\n}" ]
[ "0.75578105", "0.68279743", "0.6808188", "0.67874706", "0.67764795", "0.6770237", "0.6751396", "0.6749864", "0.67377067", "0.671777", "0.67130256", "0.6711118", "0.67096317", "0.6698934", "0.66943854", "0.6653425", "0.6593627", "0.6591759", "0.65818924", "0.657622", "0.65748", "0.6567073", "0.6540341", "0.6539644", "0.65001744", "0.6487167", "0.6441829", "0.64402694", "0.64181143", "0.6417234", "0.64153594", "0.6400041", "0.6379492", "0.6367704", "0.6359998", "0.63559943", "0.6339042", "0.6328794", "0.6317089", "0.63118666", "0.6284516", "0.6260724", "0.62554646", "0.6218225", "0.6191741", "0.617191", "0.6159733", "0.6152384", "0.61474913", "0.6127481", "0.61256737", "0.6100942", "0.6094956", "0.6071949", "0.6068864", "0.6061755", "0.6050191", "0.6050131", "0.6039243", "0.6038775", "0.6036114", "0.60313076", "0.60294515", "0.60190916", "0.60091716", "0.6006364", "0.6001551", "0.5981775", "0.59722877", "0.59533006", "0.5950843", "0.5941519", "0.5934857", "0.5933446", "0.59194887", "0.59167725", "0.590884", "0.5905244", "0.59031236", "0.5899297", "0.58896893", "0.5883911", "0.5878558", "0.58761215", "0.58711463", "0.5862885", "0.5862885", "0.58478963", "0.58469623", "0.584366", "0.5821458", "0.5815658", "0.58015746", "0.57843095", "0.5778135", "0.57761246", "0.5772535", "0.577096", "0.5763145", "0.5762656" ]
0.75762707
0
exec executes the given query, building the necessary sql.
func (s *SQLStore) execBuilder(e execer, b builder) (sql.Result, error) { sqlString, args, err := b.ToSql() if err != nil { return nil, errors.Wrap(err, "failed to build sql") } return s.exec(e, sqlString, args...) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (w *Wrapper) exec(query string, args ...interface{}) (sql.Result, error) {\n\tw.connLock.RLock()\n\tdefer w.connLock.RUnlock()\n\n\treturn w.connection.Exec(w.prepare(query), args...)\n}", "func execQuery(db *sql.DB, sqlQuery string, args ...interface{}) {\n\t_, dbErr := db.Exec(sqlQuery, args...)\n\tif dbErr != nil {\n\t\tLogErr(\"%s\\n\", dbErr)\n\t}\n}", "func (s *Schema) Exec(query string, args ...interface{}) (sql.Result, error) {\n\n\tvar e error\n\tvar stmt *sql.Stmt\n\tstmt, e = s.Dal.Connection.Prepare(query)\n\t// fmt.Printf(\"Stmt: %v\\n\", stmt)\n\tif e != nil {\n\t\tfmt.Printf(\"Error: %s\", e.Error())\n\t\treturn nil, e\n\t}\n\tdefer stmt.Close()\n\treturn stmt.Exec(args...)\n}", "func exec(stmt *sql.Stmt, args ...interface{}) error {\n\t_, err := stmt.Exec(args...)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"mysql: could not execute statement: %v\", err)\n\t}\n\treturn nil\n}", "func (q *Querier) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tq.logBefore(query, args)\n\tstart := time.Now()\n\tres, err := q.dbtxCtx.ExecContext(q.ctx, query, args...)\n\tq.logAfter(query, args, time.Since(start), err)\n\treturn res, err\n}", "func (m *SchemaData) Exec(query string) (sql.Result, error) {\n\tlog.Println(\"[SQL]\", \"[\"+m.dbType+\"]\", query)\n\treturn nil, nil\n}", "func (q *Querier) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tq.logBefore(query, args)\n\tstart := time.Now()\n\n\tdbtxCtx := q.selectDBTXContext(query)\n\tres, err := dbtxCtx.ExecContext(q.ctx, query, args...)\n\tq.logAfter(query, args, time.Since(start), err)\n\treturn res, err\n}", "func ExecQuery(ctx context.Context, sql string) {\n\t_, span := opentelemetry.StartSpan(ctx, \"database-call\")\n\tdefer span.Finish(0)\n\n\t// Do the query execution\n}", "func (q *Query) Exec() error {\n\treturn q.query.Exec()\n}", "func Exec(sql string) (sql.Result, error) {\n\treturn database.Exec(sql)\n}", "func Exec(query string, args ...interface{}) (sql.Result, error) {\n\tdb, err := sql.Open(driver, conninfo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer db.Close()\n\n\tres, err := db.Exec(query, args...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n}", "func (t *Transaction) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tt.dbmap.trace(query, args)\n\treturn t.tx.Exec(query, args...)\n}", "func (s *SQLStore) exec(e execer, sqlString string, args ...interface{}) (sql.Result, error) {\n\tsqlString = s.db.Rebind(sqlString)\n\treturn e.Exec(sqlString, args...)\n}", "func (ex *Execer) execSQL(fullSQL string, args []interface{}) (sql.Result, error) {\n\tdefer logExecutionTime(time.Now(), fullSQL, args)\n\n\tvar result sql.Result\n\tvar err error\n\tresult, err = ex.database.Exec(fullSQL, args...)\n\tif err != nil {\n\t\treturn nil, logSQLError(err, \"execSQL.30\", fullSQL, args)\n\t}\n\n\treturn result, nil\n}", "func (q *QueryGoPg) Exec(sql string, args ...interface{}) (err error) {\n\t_, err = goPgConnection.Exec(sql, args...)\n\treturn\n}", "func (this *Database) Exec(sql string, args ...interface{}) {\n\t//fmt.Printf(\"About to EXEC\\n\")\n\t_, err := this.db.Exec(sql, args...)\n\t//fmt.Printf(\"Done EXEC\\n\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (lq LoggingQueryable) Exec(query string, args ...interface{}) (result sql.Result, err error) {\n\tresult, err = lq.Q.Exec(query, args...)\n\tlog.Printf(\"SQL: Exec(%v, %v) -> %v\\n\", query, args, err)\n\treturn result, err\n}", "func (database *Database) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn database.ExecContext(context.Background(), query, args...)\n}", "func (s *dbase) Exec(query string, params ...interface{}) error {\n\t_, err := s.connection.Exec(query, params...)\n\n\treturn err\n}", "func (d *Database) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn d.Db.Exec(query, args...)\n}", "func (d *sqlDB) Exec(query string, args ...interface{}) (result sql.Result, err error) {\n\terr = d.retry(func() error {\n\t\tresult, err = d.conn.Exec(query, args...)\n\t\treturn err\n\t})\n\treturn\n}", "func (this *Database) Exec(sql string, args ...interface{}) sql.Result {\n\t//fmt.Printf(\"About to EXEC\\n\")\n\tresult, err := this.db.Exec(sql, args...)\n\t//fmt.Printf(\"Done EXEC\\n\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn result\n}", "func (conn *dbconn) exec(ctx context.Context, name string, params []interface{}) (result sql.Result, err error) {\n\n\t// TODO: check Parameters for injection here\n\t// TODO: update the params parameter to use an interface type? So that Parameters to the sproc can have the types validated?\n\n\t// Build the command\n\tvar command string\n\tif command, err = conn.buildSprocCall(name, params); err == nil {\n\n\t\t// Verify that a valid command was created\n\t\tif len(command) > 0 {\n\t\t\tresult, err = conn.db.ExecContext(ctx, command, params...)\n\t\t} else {\n\t\t\terr = errors.Errorf(\"Invalid command creation for stored Procedure %s\", name)\n\t\t}\n\t} else {\n\t\terr = errors.Errorf(\"Error while building execution call for stored Procedure %s: %v\", name, err)\n\t}\n\treturn result, err\n}", "func (db *DB) Exec(query interface{}, params ...interface{}) (orm.Result, error) {\n\tvar res orm.Result\n\tvar err error\n\ttracing.Trace(db.inner.Context(), query, func() error {\n\t\tif db.tx != nil {\n\t\t\tres, err = db.tx.Exec(query, params...)\n\t\t} else {\n\t\t\tres, err = db.inner.Exec(query, params...)\n\t\t}\n\t\treturn err\n\t})\n\treturn res, err\n}", "func (ng *Engine) exec(ctx context.Context, q *query) (model.Value, error) {\n\tcurrentQueries.Inc()\n\tdefer currentQueries.Dec()\n\tctx, cancel := context.WithTimeout(ctx, ng.options.Timeout)\n\tq.cancel = cancel\n\n\tqueueTimer := q.stats.GetTimer(stats.ExecQueueTime).Start()\n\n\tif err := ng.gate.Start(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\tdefer ng.gate.Done()\n\n\tqueueTimer.Stop()\n\n\t// Cancel when execution is done or an error was raised.\n\tdefer q.cancel()\n\n\tconst env = \"query execution\"\n\n\tevalTimer := q.stats.GetTimer(stats.TotalEvalTime).Start()\n\tdefer evalTimer.Stop()\n\n\t// The base context might already be canceled on the first iteration (e.g. during shutdown).\n\tif err := contextDone(ctx, env); err != nil {\n\t\treturn nil, err\n\t}\n\n\tswitch s := q.Statement().(type) {\n\tcase *EvalStmt:\n\t\treturn ng.execEvalStmt(ctx, q, s)\n\tcase testStmt:\n\t\treturn nil, s(ctx)\n\t}\n\n\tpanic(fmt.Errorf(\"promql.Engine.exec: unhandled statement of type %T\", q.Statement()))\n}", "func (stmt *statement) Exec(ctx context.Context, db Executor) (sql.Result, error) {\n\tif ctx == nil {\n\t\tctx = context.Background()\n\t}\n\treturn db.ExecContext(ctx, stmt.String(), stmt.args...)\n}", "func (s *stmt) Exec(args []driver.Value) (driver.Result, error) {\n\treturn s.c.Exec(s.query, args)\n}", "func ExecQuery(db *sql.DB, query string) (string) {\n\tDCL, err := db.Exec(\"Begin Transaction;\")\n\tDCL, err = db.Exec(query)\n\n\tif err != nil {\n\t\t\t\t\tlog.Fatal(err)\n\t\t\t\t\treturn \"Fail\"\n\t\t\t\t}\n\n\tdb.Exec(\"Commit;\")\n\tfmt.Println(\" Function Executed : \",DCL)\t\n\treturn \"Success\"\n\n}", "func (mc *MysqlConn) exec(query string) error {\n\t// Send command\n\terr := mc.writeCommandPacketStr(COM_QUERY, query)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Read Result\n\tresLen, err := mc.readResultSetHeaderPacket()\n\tif err == nil && resLen > 0 {\n\t\tif err = mc.readUntilEOF(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = mc.readUntilEOF()\n\t}\n\n\treturn err\n}", "func (e *SqlExecutor) Exec(query string, args ...interface{}) (sql.Result, error) {\n\thook := e.db.ExecutorHook()\n\thook.BeforeExec(e.ctx, query, args...)\n\tv, err := e.SqlExecutor.Exec(query, args...)\n\thook.AfterExec(e.ctx, query, args...)\n\treturn v, err\n}", "func (tx *Transaction) Exec(query string, args ...interface{}) error {\n\treturn exec(tx.tx, query, args...)\n}", "func (s *Session) Exec() (result sql.Result, err error) {\n\tdefer s.Clear()\n\tlog.Info(s.sql.String(), s.sqlVars)\n\tif result, err = s.DB().Exec(s.sql.String(), s.sqlVars...); err != nil {\n\t\tlog.Error(err)\n\t}\n\treturn\n}", "func (mc *mysqlConn) exec(query string) (e error) {\n\t// Send command\n\te = mc.writeCommandPacket(COM_QUERY, query)\n\tif e != nil {\n\t\treturn\n\t}\n\n\t// Read Result\n\tresLen, e := mc.readResultSetHeaderPacket()\n\tif e != nil {\n\t\treturn\n\t}\n\n\tmc.affectedRows = 0\n\tmc.insertId = 0\n\n\tif resLen > 0 {\n\t\t_, e = mc.readUntilEOF()\n\t\tif e != nil {\n\t\t\treturn\n\t\t}\n\n\t\tmc.affectedRows, e = mc.readUntilEOF()\n\t\tif e != nil {\n\t\t\treturn\n\t\t}\n\t}\n\n\treturn\n}", "func (s *Shard) Exec(query string, args ...interface{}) (sql.Result, errs.Err) {\n\tfixArgs(args)\n\tres, stdErr := s.sqlConn.Exec(query, args...)\n\tif stdErr != nil {\n\t\treturn nil, errs.Wrap(stdErr, errInfo(\"Exec sqlConn.Exec() error\", query, args))\n\t}\n\treturn res, nil\n}", "func (s SQLite) Exec(query string, args ...interface{}) (sql.Result, error) {\n\tif s.DB == nil {\n\t\treturn nil, fmt.Errorf(\"db is not created\")\n\t}\n\treturn s.DB.Exec(query, args...)\n}", "func (s *SqlDb) Exec(ctx context.Context, sql string, args ...interface{}) (r sql.Result, err error) {\n\tvar c *SqlContext\n\ti := ctx.Value(goradd.SqlContext)\n\tif i != nil {\n\t\tc = i.(*SqlContext)\n\t}\n\tlog.FrameworkDebug(\"Exec: \", sql, args)\n\n\tvar beginTime = time.Now()\n\n\tif c != nil && c.tx != nil {\n\t\tr, err = c.tx.ExecContext(ctx, sql, args...)\n\t} else {\n\t\tr, err = s.db.ExecContext(ctx, sql, args...)\n\t}\n\n\tvar endTime = time.Now()\n\n\tif c != nil && s.profiling {\n\t\tif args != nil {\n\t\t\tfor _, arg := range args {\n\t\t\t\tsql = strings.TrimSpace(sql)\n\t\t\t\tsql += fmt.Sprintf(\",\\n%#v\", arg)\n\t\t\t}\n\t\t}\n\t\tc.profiles = append(c.profiles, ProfileEntry{DbKey: s.dbKey, BeginTime: beginTime, EndTime: endTime, Typ: \"Exec\", Sql: sql})\n\t}\n\n\treturn\n}", "func Exec(x Execer, query string, args ...interface{}) (res sql.Result, err error) {\n\tres, err = x.Exec(query, args...)\n\terr = interpretError(err)\n\treturn\n}", "func (q *query) Exec(ctx context.Context) *Result {\n\tif span := opentracing.SpanFromContext(ctx); span != nil {\n\t\tspan.SetTag(queryTag, q.stmt.String())\n\t}\n\n\tres, err := q.ng.exec(ctx, q)\n\treturn &Result{Err: err, Value: res}\n}", "func (c *Conn) Exec(ctx context.Context, sql string, arguments ...any) (pgconn.CommandTag, error) {\n\tif c.queryTracer != nil {\n\t\tctx = c.queryTracer.TraceQueryStart(ctx, c, TraceQueryStartData{SQL: sql, Args: arguments})\n\t}\n\n\tif err := c.deallocateInvalidatedCachedStatements(ctx); err != nil {\n\t\treturn pgconn.CommandTag{}, err\n\t}\n\n\tcommandTag, err := c.exec(ctx, sql, arguments...)\n\n\tif c.queryTracer != nil {\n\t\tc.queryTracer.TraceQueryEnd(ctx, c, TraceQueryEndData{CommandTag: commandTag, Err: err})\n\t}\n\n\treturn commandTag, err\n}", "func (c *conn) Exec(query string, args []driver.Value) (driver.Result, error) {\n\treturn c.exec(context.Background(), query, toNamedValues(args))\n}", "func (c *conn) Exec(query string, args []driver.Value) (driver.Result, error) {\n\treturn c.exec(context.Background(), query, args)\n}", "func (extDb *Database) Exec(query string, args ...interface{}) (sql.Result, error) {\n\terr := extDb.checkStatus()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdb := extDb.getDb()\n\tif db == nil {\n\t\treturn nil, ErrNotInitialized\n\t}\n\tresult, err := db.Exec(query, args...)\n\tif err != nil {\n\t\tif !extDb.i.isConnectionError(err) {\n\t\t\treturn nil, err\n\t\t}\n\t\tlog.Info(\"Database exec error \", err)\n\t\terrConn := extDb.Reconnect()\n\t\tif errConn == nil {\n\t\t\tdb := extDb.getDb()\n\t\t\tif db == nil {\n\t\t\t\treturn nil, ErrNotInitialized\n\t\t\t}\n\t\t\tresult, err = db.Exec(query, args...)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\treturn result, nil\n\t\t}\n\t\treturn nil, errConn\n\t}\n\treturn result, nil\n}", "func (db TestDB) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn db.testTx.Exec(query, args...)\n}", "func (a ArticleDB) Exec(sess *sqlabble.Session, st stmt.Statement) (sql.Result, error) {\n\tquery, values := sess.Builder.Build(st)\n\tresult, err := sess.Exec(query, values...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn result, nil\n}", "func (tx *Tx) Exec(ctx context.Context, query SQLMutation, args ...interface{}) (sql.Result, error) {\n\tif err := query.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn tx.parent.ExecContext(ctx, query.String(), args...)\n}", "func (conn *n1qlConn) Exec(query string, args ...interface{}) (godbc.Result, error) {\n\n\tif len(args) > 0 {\n\t\tvar argCount int\n\t\tquery, argCount = prepareQuery(query)\n\t\tif argCount != len(args) {\n\t\t\treturn nil, fmt.Errorf(\"Argument count mismatch %d != %d\", argCount, len(args))\n\t\t}\n\t\tquery, args = preparePositionalArgs(query, argCount, args)\n\t}\n\n\treturn conn.performExec(query, nil)\n}", "func (c *connImpl) Exec(query string, args []driver.Value) (driver.Result, error) {\n\t// http://code.google.com/p/go-wiki/wiki/InterfaceSlice\n\ttmp := make([]interface{}, len(args))\n\tfor i, arg := range args {\n\t\ttmp[i] = arg\n\t}\n\tif err := c.c.Exec(query, tmp...); err != nil {\n\t\treturn nil, err\n\t}\n\treturn c, nil // FIXME RowAffected/noRows\n}", "func (mc *MysqlConn) Exec(query string) error {\n\treturn mc.exec(query)\n}", "func Exec(a config.AppContext, q interpreter.Query) ([]map[string]interface{}, error) {\n\t/*\n\t * Based on the number of tables avaiable in the query, we will execute the same\n\t */\n\n\tif len(q.Tables) == 0 {\n\t\treturn nil, errors.New(\"Couldn't find the table to be queried from\")\n\t}\n\n\tif len(q.Tables) == 1 {\n\t\treturn SingleTableMode(a, q)\n\t}\n\n\treturn nil, errors.New(\"multiple table join query not supported yet\")\n}", "func (w *Wrapper) executeQuery() (res sql.Result, err error) {\n\tw.cleanBefore()\n\tw.buildQuery()\n\tw.LastQuery = w.query\n\tw.LastParams = w.params\n\t// Calculate the execution time.\n\tvar start time.Time\n\tif w.tracing {\n\t\tstart = time.Now()\n\t}\n\t// Execute the query if the wrapper is executable.\n\tif w.executable {\n\t\tvar stmt *sql.Stmt\n\t\tvar count int64\n\t\tstmt, err = w.db.Prepare(w.query)\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tres, err = stmt.Exec(w.params...)\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tw.LastResult = res\n\t\tcount, err = res.RowsAffected()\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t\tw.count = int(count)\n\t\terr = stmt.Close()\n\t\tif err != nil {\n\t\t\tif w.tracing {\n\t\t\t\tw.saveTrace(err, w.query, start)\n\t\t\t}\n\t\t\treturn\n\t\t}\n\t}\n\tif w.tracing {\n\t\tw.saveTrace(err, w.query, start)\n\t}\n\tw.cleanAfter()\n\treturn\n}", "func (st *ogonoriStmt) Exec(args []driver.Value) (driver.Result, error) {\n\togl.Debugln(\"** ogonoriStmt.Exec\")\n\tif st.conn == nil || st.conn.dbc == nil {\n\t\treturn nil, oerror.ErrInvalidConn{Msg: \"obinary.DBClient not initialized in ogonoriStmt#Exec\"}\n\t}\n\n\treturn doExec(st.conn.dbc, st.query, args)\n}", "func (conn *db) Exec(stmt Stmt) (res sql.Result, err error) {\n\tif err := conn.Connect(); err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar t time.Time\n\tif conn.hasProfiling() {\n\t\tt = time.Now()\n\t}\n\n\tif conn.tx != nil {\n\t\tres, err = conn.tx.Exec(stmt.String(), stmt.Args()...)\n\t} else {\n\t\tres, err = (*conn.dbx).Exec(stmt.String(), stmt.Args()...)\n\t}\n\n\tconn.profilingStmt(stmt, err, t)\n\treturn\n}", "func (db *DB) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn db.master.Exec(query, args...)\n}", "func (s *Stmt) Exec(args []driver.Value) (driver.Result, error) {\n\t// Binds all arguments.\n\tif err := s.Bind(args); err != nil {\n\t\treturn nil, err\n\t}\n\t// Executes query.\n\tswitch s.p.(type) {\n\tcase parser.CreateViewStmt:\n\t\treturn NewCreateViewStmt(s).Exec()\n\t}\n\treturn s.si.Exec(args)\n}", "func (db *DB) Exec(ctx context.Context, query SQLMutation, args ...interface{}) (sql.Result, error) {\n\tif err := query.Validate(); err != nil {\n\t\treturn nil, err\n\t}\n\treturn db.master.ExecContext(ctx, query.String(), args...)\n}", "func (ar *ActiveRecord) Exec() (sql.Result, error) {\n\treturn ar.DB.Exec(ar.ExecString(), ar.Args...)\n}", "func (tx *PgxTx) Exec(ctx context.Context, sql string, args ...interface{}) error {\n\t_, err := (*pgx.Tx)(tx).ExecEx(ctx, sql, nil /* QueryExOptions */, args...)\n\treturn err\n}", "func (q *Query) Exec(ctx context.Context) (*QueryResult, error) {\n\tvar r QueryResult\n\n\tif q.client == nil || !q.client.Started() {\n\t\treturn &r, fmt.Errorf(\"client or db is nil\")\n\t}\n\n\tswitch q.action {\n\tcase \"select\":\n\t\trows, err := q.execSelect(ctx)\n\t\tr.Rows = rows\n\t\treturn &r, err\n\tcase \"insert\":\n\t\trows, err := q.execInsert(ctx)\n\t\tr.Rows = rows\n\t\treturn &r, err\n\tcase \"update\":\n\t\tvar err error\n\t\tif len(q.returning) == 0 {\n\t\t\tr.RowsAffected, err = q.execUpdate(ctx)\n\t\t} else {\n\t\t\tr.Rows, err = q.execUpdateR(ctx)\n\t\t}\n\t\treturn &r, err\n\tcase \"delete\":\n\t\tvar err error\n\t\tif len(q.returning) == 0 {\n\t\t\tr.RowsAffected, err = q.execDelete(ctx)\n\t\t} else {\n\t\t\tr.Rows, err = q.execDeleteR(ctx)\n\t\t}\n\t\treturn &r, err\n\tdefault:\n\t\treturn &r, fmt.Errorf(\"unsupported action %v\", q.action)\n\t}\n}", "func (ub *UpdateBuilder) Exec(ctx context.Context) (sql.Result, error) {\n\tr, err := ub.Table.Session.Exec(ctx, ub.sql)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"select failed\")\n\t}\n\n\treturn r, nil\n}", "func (s *stmt) Exec(args []driver.Value) (driver.Result, error) {\n\treturn s.exec(context.Background(), toNamedValues(args))\n}", "func (db *DB) Exec(query string, args ...interface{}) (sql.Result, error) {\n\treturn db.Master().Exec(query, args...)\n}", "func (s *Stmt) Exec(args []driver.Value) (driver.Result, error) {\n\t// in transaction, need run Executor\n\texecutor, err := exec.BuildExecutor(s.res.dbType, s.txCtx.TransactionMode, s.query)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\texecCtx := &types.ExecContext{\n\t\tTxCtx: s.txCtx,\n\t\tQuery: s.query,\n\t\tValues: args,\n\t}\n\n\tret, err := executor.ExecWithValue(context.Background(), execCtx,\n\t\tfunc(ctx context.Context, query string, args []driver.NamedValue) (types.ExecResult, error) {\n\t\t\tret, err := s.stmt.Exec(util.NamedValueToValue(args))\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\treturn types.NewResult(types.WithResult(ret)), nil\n\t\t})\n\n\treturn ret.GetResult(), err\n}", "func (stmt *Statement) Exec(params ...interface{}) (\n rows []*Row, res *Result, err os.Error) {\n\n res, err = stmt.Run(params...)\n if err != nil {\n return\n }\n // Read rows\n var row *Row\n for {\n row, err = res.GetRow()\n if err != nil || row == nil {\n break\n }\n rows = append(rows, row)\n }\n return\n}", "func (p PostgreSQL) Exec(sql string, args ...interface{}) (pgconn.CommandTag, error) {\n\tif p.Connection == nil {\n\t\treturn nil, fmt.Errorf(\"connection is not created\")\n\t}\n\n\treturn p.Connection.Exec(context.TODO(), sql, args...)\n}", "func (db *DB) Exec(name string, args ...interface{}) (pgx.CommandTag, error) {\n\tq, err := db.qm.getQuery(name)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn db.Pool.Exec(q.getSQL(), args...)\n}", "func Execute(query string, args ...interface{}) (sql.Result, error){\n result, err := db.Exec(query, args...)\n if err != nil && !debug { \n log.Println(err)\n }\n return result, err\n}", "func Exec(db *sqlx.DB, statement string) error {\n\tif db != nil {\n\t\tr, err := db.Exec(statement)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\trows, _ := r.RowsAffected()\n\t\tlogrus.Debugf(\"Affected Rows %d\", rows)\n\t} else {\n\t\treturn fmt.Errorf(\"Database connection is nil\")\n\t}\n\n\treturn nil\n}", "func (c *sqlmock) Exec(query string, args []driver.Value) (driver.Result, error) {\n\tnamedArgs := make([]driver.NamedValue, len(args))\n\tfor i, v := range args {\n\t\tnamedArgs[i] = driver.NamedValue{\n\t\t\tOrdinal: i + 1,\n\t\t\tValue: v,\n\t\t}\n\t}\n\n\tex, err := c.exec(query, namedArgs)\n\tif ex != nil {\n\t\ttime.Sleep(ex.delay)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn ex.result, nil\n}", "func execSQL(stmt *sql.Stmt, args ...interface{}) (sql.Result, error) {\n\tr, err := stmt.Exec(args...)\n\tif err != nil {\n\t\treturn r, fmt.Errorf(\"mysql: could not execute statement: %v\", err)\n\t}\n\trowsAffected, err := r.RowsAffected()\n\tif err != nil {\n\t\treturn r, fmt.Errorf(\"mysql: could not get rows affected: %v\", err)\n\t} else if rowsAffected != 1 {\n\t\treturn r, fmt.Errorf(\"mysql: expected 1 row affected, got %d\", rowsAffected)\n\t}\n\treturn r, nil\n}", "func (c Conn) Exec(ctx context.Context, query string, args, v any) error {\n\targv, ok := args.([]any)\n\tif !ok {\n\t\treturn fmt.Errorf(\"dialect/sql: invalid type %T. expect []any for args\", v)\n\t}\n\tswitch v := v.(type) {\n\tcase nil:\n\t\tif _, err := c.ExecContext(ctx, query, argv...); err != nil {\n\t\t\treturn err\n\t\t}\n\tcase *sql.Result:\n\t\tres, err := c.ExecContext(ctx, query, argv...)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t*v = res\n\tdefault:\n\t\treturn fmt.Errorf(\"dialect/sql: invalid type %T. expect *sql.Result\", v)\n\t}\n\treturn nil\n}", "func (ex *Execer) execFn() (sql.Result, error) {\n\tfullSQL, args, err := ex.Interpolate()\n\tif err != nil {\n\t\treturn nil, logger.Error(\"execFn.10\", \"err\", err, \"sql\", fullSQL)\n\t}\n\tdefer logExecutionTime(time.Now(), fullSQL, args)\n\n\tvar result sql.Result\n\tresult, err = ex.database.Exec(fullSQL, args...)\n\tif err != nil {\n\t\treturn nil, logSQLError(err, \"execFn.30:\"+fmt.Sprintf(\"%T\", err), fullSQL, args)\n\t}\n\n\treturn result, nil\n}", "func SqlcmdExec(\n\tcfg Config,\n\tquery string,\n) (string, error) {\n\targs := []string{\n\t\t\"-S\", fmt.Sprintf(\"%s,%s\", cfg.Host, cfg.Port),\n\t\t\"-U\", cfg.Username,\n\t\t\"-P\", cfg.Password,\n\t\t\"-Q\", query,\n\t}\n\n\tif cfg.ReadOnly == true {\n\t\targs = append(args, \"-K\", \"ReadOnly\")\n\t}\n\n\tif db := cfg.Database; db != \"\" {\n\t\targs = append(args, \"-d\", db)\n\t}\n\n\tout, err := exec.Command(\n\t\t\"sqlcmd\",\n\t\targs...,\n\t).Output()\n\n\tif err != nil {\n\t\tif exitErrr, ok := err.(*exec.ExitError); ok {\n\t\t\treturn \"\", errors.New(string(exitErrr.Stderr))\n\t\t}\n\n\t\treturn \"\", err\n\t}\n\n\treturn string(out), nil\n}", "func (e *Executor) ExecSQL(sql *types.SQL) error {\n\te.Lock()\n\tdefer e.Unlock()\n\t// e.parseOnlineTable(sql)\n\te.SQLCh <- sql\n\treturn errors.Trace(<-e.ErrCh)\n}", "func (db *DB) do(query string, arguments []interface{}) (sql.Result, error) {\n\tquery = db.replacePlaceholders(query)\n\n\t// Execute the statement\n\tstartTime := time.Now()\n\tqueryable, err := db.getQueryable(query)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn nil, err\n\t}\n\tresult, err := queryable.Exec(arguments...)\n\tconsumedTime := timeElapsedSince(startTime)\n\tdb.addConsumedTime(consumedTime)\n\tdb.logExecution(consumedTime, query, arguments)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\tif db.useErrorParser {\n\t\t\treturn nil, db.adapter.ParseError(err)\n\t\t}\n\t\treturn nil, err\n\t}\n\n\treturn result, err\n}", "func (ar *ActiveRecord) ExecSQL(sql string, args ...interface{}) (sql.Result, error) {\n\treturn ar.DB.Exec(sql, args...)\n}", "func (m *MSSQLTx) Exec(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\treturn m.ExecWithMetrics(ctx, &metrics.NoOp{}, query, args...)\n}", "func (tq *Query) Exec() error {\n\treturn tq.Iter().Close()\n}", "func (wd *WithDDL) Exec(ctx context.Context, query string, fQuery any, fDDL any) (*sqltypes.Result, error) {\n\texecQuery, err := wd.unify(ctx, fQuery)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\texecDDL, err := wd.unify(ctx, fDDL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tqr, err := execQuery(query)\n\tif err == nil {\n\t\treturn qr, nil\n\t}\n\tif !wd.isSchemaError(err) {\n\t\treturn nil, err\n\t}\n\n\tlog.Infof(\"Updating schema for %v and retrying: %v\", sqlparser.TruncateForUI(err.Error()), err)\n\tfor _, applyQuery := range wd.ddls {\n\t\t_, merr := execDDL(applyQuery)\n\t\tif merr == nil {\n\t\t\tcontinue\n\t\t}\n\t\tif mysql.IsSchemaApplyError(merr) {\n\t\t\tcontinue\n\t\t}\n\t\tlog.Warningf(\"DDL apply %v failed: %v\", applyQuery, merr)\n\t\t// Return the original error.\n\t\treturn nil, err\n\t}\n\treturn execQuery(query)\n}", "func (stmt *Stmt) Exec(params ...interface{}) (Result, error) {\n\treturn stmt.exec(context.TODO(), params...)\n}", "func (dbi *DBInstance) ExecTXQuery(q string, args ...interface{}) error {\n\ttx, err := dbi.SQLSession.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer tx.Rollback()\n\n\tstmt, err := tx.Prepare(q)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\t_, err = stmt.Exec(args...)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn tx.Commit()\n}", "func (db *DatabaseConnection) query(stmt mysql.Stmt, args ...interface{}) mysql.Result {\n\treturn db.exec(stmt, args...)\n}", "func executeSQL(db *sql.DB, sql string) error {\n\tsql = strings.TrimSpace(sql)\n\tif db == nil {\n\t\tfmt.Printf(\"%s;\\n\", sql)\n\t} else {\n\t\tlog.Info(fmt.Printf(\"executeSQL: %s\", sql))\n\t\tif _, err := db.Exec(sql); err != nil {\n\t\t\treturn fmt.Errorf(\"Error executing SQL: %v: %v\", sql, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *DBMock) Exec(ctx context.Context, query string, params ...interface{}) (pgconn.CommandTag, error) {\n\targs := m.Called(append([]interface{}{query}, params...)...)\n\treturn nil, args.Error(0)\n}", "func (p PgxWrapper) Exec(ctx context.Context, queryName, sql string, args ...interface{}) error {\n\tspan, _ := opentracing.StartSpanFromContext(ctx, queryName)\n\tspan.Finish()\n\n\tif _, err := p.pool.Exec(ctx, sql, args...); err != nil {\n\t\treturn fmt.Errorf(\"could not execute query: %w\", err)\n\t}\n\n\treturn nil\n}", "func (*Stmt) Exec(args []driver.Value) (driver.Result, error) {\n\tpanic(\"not supported\")\n}", "func (tcdb *Teocdb) execStmt(s *gocql.Session, stmt string) error {\n\tq := s.Query(stmt).RetryPolicy(nil)\n\tdefer q.Release()\n\treturn q.Exec()\n}", "func (q *Querier) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\treturn q.WithContext(ctx).Exec(query, args...)\n}", "func (q *Querier) ExecContext(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\treturn q.WithContext(ctx).Exec(query, args...)\n}", "func (m *MSSQLDatastore) Exec(ctx context.Context, query string, args ...interface{}) (sql.Result, error) {\n\tif m == nil {\n\t\treturn nil, ErrEmptyObject\n\t}\n\n\tif _, ok := ctx.Deadline(); !ok {\n\t\tvar cancel context.CancelFunc\n\t\tctx, cancel = context.WithTimeout(ctx, QueryLimit)\n\t\tdefer cancel()\n\t}\n\n\treturn m.db.ExecContext(ctx, query, args...)\n}", "func (sqd *SurveyQuestionDelete) Exec(ctx context.Context) (int, error) {\n\treturn sqd.sqlExec(ctx)\n}", "func (stmt *statement) ExecContext(ctx context.Context, args []driver.NamedValue) (driver.Result, error) {\n\treturn stmt.conn.ExecContext(ctx, stmt.query, args)\n}", "func (p *Param) Exec() (sql.Result, error) {\n\treturn p.T().Exec(p)\n}", "func (m *MysqlProxy) Exec(args []string, user *client.Client) (interface{}, error) {\n execPlan, err := planbuilder.GetExecPlan(args, user)\n if err != nil { return nil, err }\n\n result, err := execPlan.Do()\n if err != nil { return nil, err }\n\n switch result.(type) {\n case *schema.MysqlTable:\n err = m.AddTable(result.(*schema.MysqlTable))\n if err == nil { result = \"create succesfully!\" }\n }\n\n return result, err\n}", "func GomssqlExec(\n\tcfg Config,\n\tquery string,\n) (string, error) {\n\tapplicationIntent := \"ReadWrite\"\n\tif cfg.ReadOnly {\n\t\tapplicationIntent = \"ReadOnly\"\n\t}\n\n\tdsnString := fmt.Sprintf(\n\t\t\"user id=%s;password=%s;server=%s;port=%s;encrypt=%s;applicationintent=%s\",\n\t\tcfg.Username,\n\t\tcfg.Password,\n\t\tcfg.Host,\n\t\tcfg.Port,\n\t\t\"disable\",\n\t\tapplicationIntent,\n\t)\n\n\tif db := cfg.Database; db != \"\" {\n\t\tdsnString += fmt.Sprintf(\";database=%s\", db)\n\t}\n\n\t// Open the connection\n\tconn, err := sql.Open(\n\t\t\"mssql\",\n\t\tdsnString,\n\t)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer conn.Close()\n\n\tctx, _ := context.WithDeadline(\n\t\tcontext.Background(),\n\t\ttime.Now().Add(1*time.Second),\n\t)\n\n\tif query == \"\" {\n\t\t_, err := conn.Conn(ctx)\n\t\treturn \"\", err\n\t}\n\n\t// Execute the query\n\trows, err := conn.QueryContext(ctx, query)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer rows.Close()\n\n\t// Execute the query\n\tcols, err := rows.Columns()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\trawResult := make([][]byte, len(cols))\n\n\tdest := make([]interface{}, len(cols)) // A temporary interface{} slice\n\tfor i := range rawResult {\n\t\tdest[i] = &rawResult[i] // Put pointers to each string in the interface slice\n\t}\n\n\tw := new(tabwriter.Writer)\n\tbuf := &bytes.Buffer{}\n\tw.Init(buf, 0, 0, 0, ' ', tabwriter.Debug|tabwriter.AlignRight)\n\n\tfor rows.Next() {\n\t\terr = rows.Scan(dest...)\n\t\tif err != nil {\n\t\t\treturn \"\", nil\n\t\t}\n\n\t\trowString := \"\"\n\t\tfor _, raw := range rawResult {\n\t\t\tif raw == nil {\n\t\t\t\trowString += \"\\\\N\"\n\t\t\t} else {\n\t\t\t\trowString += string(raw)\n\t\t\t}\n\n\t\t\trowString += \"\\t\"\n\t\t}\n\n\t\tfmt.Fprintln(w, rowString)\n\t}\n\n\tw.Flush()\n\n\treturn string(buf.Bytes()), err\n}", "func (erd *EpidemiologicResultsDelete) Exec(ctx context.Context) (int, error) {\n\treturn erd.sqlExec(ctx)\n}", "func (db *DB) executeQuery(query string, arguments []interface{}, noTx, noStmtCache bool) (*sql.Rows, []string, error) {\n\tquery = db.replacePlaceholders(query)\n\n\tstartTime := time.Now()\n\tqueryable, err := db.getQueryableWithOptions(query, noTx, noStmtCache)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn nil, nil, err\n\t}\n\trows, err := queryable.Query(arguments...)\n\tconsumedTime := timeElapsedSince(startTime)\n\tdb.addConsumedTime(consumedTime)\n\tdb.logExecution(consumedTime, query, arguments)\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\treturn nil, nil, err\n\t}\n\n\tcolumns, err := rows.Columns()\n\tif err != nil {\n\t\tdb.logExecutionErr(err, query, arguments)\n\t\trows.Close()\n\t\treturn nil, nil, err\n\t}\n\n\treturn rows, columns, nil\n}", "func (s StmtWrapper) Exec(args []driver.Value) (res driver.Result, err error) {\n\tspan, _ := opentracing.StartSpanFromContext(\n\t\ts.ctx,\n\t\ts.integration.getOperationName(s.dsn),\n\t)\n\n\tdefer span.Finish()\n\n\trawSpan, ok := tracer.GetRaw(span)\n\tif ok {\n\t\ts.integration.beforeCall(s.query, rawSpan, s.dsn)\n\t}\n\ttracer.OnSpanStarted(span)\n\n\tres, err = s.Stmt.Exec(args)\n\tif err != nil {\n\t\tutils.SetSpanError(span, err)\n\t}\n\treturn\n\n}", "func ExecDB(query string, args ...interface{}) sql.Result {\n\trwlock.Lock()\n\tdefer rwlock.Unlock()\n\tvar res sql.Result\n\tstmt, err := db.Prepare(query)\n\tif err != nil {\n\t\tfmt.Println(\"Error creating prepared statement (my_server.go: ExecDB)\")\n\t\tfmt.Println(err)\n\t\treturn res\n\t}\n\n\tres, err = stmt.Exec(args...)\n\tif err != nil {\n\t\tfmt.Println(\"Error executing prepared statement (my_server.go: ExecDB)\")\n\t\tfmt.Println(err)\n\t}\n\treturn res\n}", "func (rd *RecipeDelete) Exec(ctx context.Context) (int, error) {\n\treturn rd.sqlExec(ctx)\n}", "func (q queryManager) processQuery(sql string, pubKey []byte, executeifallowed bool) (uint, []byte, []byte, *structures.Transaction, error) {\n\tlocalError := func(err error) (uint, []byte, []byte, *structures.Transaction, error) {\n\t\treturn SQLProcessingResultError, nil, nil, nil, err\n\t}\n\tqp := q.getQueryParser()\n\t// this will get sql type and data from comments. data can be pubkey, txBytes, signature\n\tqparsed, err := qp.ParseQuery(sql)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\t// maybe this query contains signature and txData from previous calls\n\tif len(qparsed.Signature) > 0 && len(qparsed.TransactionBytes) > 0 {\n\t\t// this is a case when signature and txdata were part of SQL comments.\n\t\ttx, err := q.processQueryWithSignature(qparsed.TransactionBytes, qparsed.Signature, executeifallowed)\n\n\t\tif err != nil {\n\t\t\treturn localError(err)\n\t\t}\n\n\t\treturn SQLProcessingResultTranactionComplete, nil, nil, tx, nil\n\t}\n\n\tneedsTX, err := q.checkQueryNeedsTransaction(qparsed)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\tif !needsTX {\n\t\tif !executeifallowed {\n\t\t\t// no need to execute query. just return\n\t\t\treturn SQLProcessingResultExecuted, nil, nil, nil, nil\n\t\t}\n\t\t// no need to have TX\n\t\tif qparsed.IsUpdate() {\n\t\t\t_, err := qp.ExecuteQuery(qparsed.SQL)\n\t\t\tif err != nil {\n\t\t\t\treturn localError(err)\n\t\t\t}\n\t\t}\n\t\treturn SQLProcessingResultExecuted, nil, nil, nil, nil\n\t}\n\t// decide which pubkey to use.\n\n\t// first priority for a key posted as argument, next is the key in SQL comment (parsed) and final is the key\n\t// provided to thi module\n\tif len(pubKey) == 0 {\n\t\tif len(qparsed.PubKey) > 0 {\n\t\t\tpubKey = qparsed.PubKey\n\t\t} else if len(q.pubKey) > 0 {\n\t\t\tpubKey = q.pubKey\n\t\t} else {\n\t\t\t// no pubkey to use. return notice about pubkey required\n\t\t\treturn SQLProcessingResultPubKeyRequired, nil, nil, nil, nil\n\t\t}\n\t}\n\n\t// check if the key has permissions to execute this query\n\thasPerm, err := q.checkExecutePermissions(qparsed, pubKey)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\tif !hasPerm {\n\t\treturn localError(errors.New(\"No permissions to execute this query\"))\n\t}\n\n\tamount, err := q.checkQueryNeedsPayment(qparsed)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\t// prepare SQL part of a TX\n\t// this builds RefID for a TX update\n\tsqlUpdate, err := qp.MakeSQLUpdateStructure(qparsed)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\t// prepare curency TX and add SQL part\n\n\ttxBytes, datatosign, err := q.getTransactionsManager().PrepareNewSQLTransaction(pubKey, sqlUpdate, amount, \"MINTER\")\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\ttx, err := structures.DeserializeTransaction(txBytes)\n\n\tif err != nil {\n\t\treturn localError(err)\n\t}\n\n\tif len(q.pubKey) > 0 && bytes.Compare(q.pubKey, pubKey) == 0 {\n\t\t// transaction was created by internal pubkey. we have private key for it\n\t\tsignature, err := utils.SignDataByPubKey(q.pubKey, q.privKey, datatosign)\n\t\tif err != nil {\n\t\t\treturn localError(err)\n\t\t}\n\n\t\ttx, err = q.processQueryWithSignature(txBytes, signature, executeifallowed)\n\n\t\tif err != nil {\n\t\t\treturn localError(err)\n\t\t}\n\n\t\treturn SQLProcessingResultTranactionCompleteInternally, nil, nil, tx, nil\n\t}\n\treturn SQLProcessingResultSignatureRequired, txBytes, datatosign, nil, nil\n}", "func (db *DB) Exec(query jdh.Query, table jdh.Table, param interface{}) (string, error) {\n\treturn \"\", errors.New(\"ncbi is a read only database\")\n}" ]
[ "0.764619", "0.75909364", "0.7349952", "0.7330214", "0.7247736", "0.72415066", "0.7208735", "0.71455485", "0.7144922", "0.71196324", "0.7115943", "0.7083062", "0.70793813", "0.7000444", "0.69897205", "0.6981616", "0.69492847", "0.6945496", "0.6930081", "0.6914955", "0.69060594", "0.6899077", "0.68980277", "0.6894907", "0.68643", "0.68485904", "0.68371713", "0.6831422", "0.6795959", "0.67919743", "0.6787915", "0.6782243", "0.6781331", "0.6771644", "0.6724049", "0.6717044", "0.66977245", "0.66536117", "0.66470844", "0.6638979", "0.66178554", "0.6599706", "0.65494525", "0.6518457", "0.65177226", "0.649768", "0.64849895", "0.64829284", "0.6460553", "0.64542913", "0.64382166", "0.64363885", "0.6436009", "0.6432001", "0.6424604", "0.6399757", "0.63932955", "0.63896406", "0.63611156", "0.6352559", "0.6350988", "0.6348753", "0.6346307", "0.63415617", "0.6338572", "0.63333285", "0.63081586", "0.6300645", "0.62882835", "0.62702096", "0.62681985", "0.6266116", "0.62613016", "0.6250434", "0.62212616", "0.6220473", "0.6214065", "0.6194196", "0.6183024", "0.61826503", "0.6177584", "0.617308", "0.6129056", "0.61277187", "0.60982484", "0.6078556", "0.6076337", "0.6076337", "0.606492", "0.60644513", "0.6053965", "0.6046674", "0.6042225", "0.60417676", "0.6035972", "0.6020345", "0.6005136", "0.59788775", "0.59730417", "0.596811", "0.59664184" ]
0.0
-1
finalizeTransaction ensures a transaction is closed after use, rolling back if not already committed.
func (s *SQLStore) finalizeTransaction(tx *sqlx.Tx) { // Rollback returns sql.ErrTxDone if the transaction was already closed. if err := tx.Rollback(); err != nil && err != sql.ErrTxDone { logrus.Errorf("Failed to rollback transaction; err: %v", err) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func finishTransaction(ctx context.Context, exec sqlexec.SQLExecutor, err error) error {\n\tif err == nil {\n\t\t_, err = exec.ExecuteInternal(ctx, \"commit\")\n\t} else {\n\t\t_, err1 := exec.ExecuteInternal(ctx, \"rollback\")\n\t\tterror.Log(errors.Trace(err1))\n\t}\n\treturn errors.Trace(err)\n}", "func (db *database) cleanupTransaction(tx *sqlx.Tx, err error) error {\n\t// Handle panicking goroutine by rolling back the transaction, then\n\t// continue panicking.\n\tif r := recover(); r != nil {\n\t\tdb.tryRollback(tx)\n\t\tpanic(r)\n\t}\n\n\t// Handle errors from the transaction handler by rolling back the\n\t// transaction, then returning the wrapped error.\n\tif err != nil {\n\t\tdb.tryRollback(tx)\n\t\treturn errors.Wrap(err, \"transaction failed\")\n\t}\n\n\t// Transaction handler was successful!\n\t// Attempt to commit the transaction, reporting an error on failure.\n\tif err = tx.Commit(); err != nil {\n\t\treturn errors.Wrap(err, \"failed to commit transaction\")\n\t}\n\n\treturn nil\n}", "func (_RefundableCrowdsale *RefundableCrowdsaleTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _RefundableCrowdsale.contract.Transact(opts, \"finalize\")\n}", "func (t *Transaction) cleanup() {\n\tt.mu.Lock()\n\tdefer t.mu.Unlock()\n\tt.alive = false\n\tt.sendError(status.ErrShutdown)\n\tt.streamer.Close()\n\t// signal closed to Close() caller\n\tclose(t.signalClosed)\n}", "func (_RefundableCrowdsale *RefundableCrowdsaleTransactorSession) Finalize() (*types.Transaction, error) {\n\treturn _RefundableCrowdsale.Contract.Finalize(&_RefundableCrowdsale.TransactOpts)\n}", "func (_RefundableCrowdsale *RefundableCrowdsaleSession) Finalize() (*types.Transaction, error) {\n\treturn _RefundableCrowdsale.Contract.Finalize(&_RefundableCrowdsale.TransactOpts)\n}", "func (_CRLv0 *CRLv0Transactor) Finalize(opts *bind.TransactOpts, fin bool) (*types.Transaction, error) {\n\treturn _CRLv0.contract.Transact(opts, \"finalize\", fin)\n}", "func (_FinalizableCrowdsaleImpl *FinalizableCrowdsaleImplTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _FinalizableCrowdsaleImpl.contract.Transact(opts, \"finalize\")\n}", "func (tx *transaction) Rollback() error {\n\t// Prevent rollbacks on managed transactions.\n\tif tx.managed {\n\t\ttx.close()\n\t\tpanic(\"managed transaction rollback not allowed\")\n\t}\n\n\t// Ensure transaction state is valid.\n\tif err := tx.checkClosed(); err != nil {\n\t\treturn err\n\t}\n\n\ttx.close()\n\treturn nil\n}", "func (_FinalizableCrowdsaleImpl *FinalizableCrowdsaleImplTransactorSession) Finalize() (*types.Transaction, error) {\n\treturn _FinalizableCrowdsaleImpl.Contract.Finalize(&_FinalizableCrowdsaleImpl.TransactOpts)\n}", "func (_Finalizable *FinalizableTransactor) Finalize(opts *bind.TransactOpts, fin bool) (*types.Transaction, error) {\n\treturn _Finalizable.contract.Transact(opts, \"finalize\", fin)\n}", "func (_FinalizableCrowdsaleImpl *FinalizableCrowdsaleImplSession) Finalize() (*types.Transaction, error) {\n\treturn _FinalizableCrowdsaleImpl.Contract.Finalize(&_FinalizableCrowdsaleImpl.TransactOpts)\n}", "func (_CRLv0 *CRLv0TransactorSession) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _CRLv0.Contract.Finalize(&_CRLv0.TransactOpts, fin)\n}", "func (_CRLv0 *CRLv0Session) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _CRLv0.Contract.Finalize(&_CRLv0.TransactOpts, fin)\n}", "func (_Finalizable *FinalizableTransactorSession) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _Finalizable.Contract.Finalize(&_Finalizable.TransactOpts, fin)\n}", "func (_Crowdsale *CrowdsaleTransactorSession) Finalise() (*types.Transaction, error) {\n\treturn _Crowdsale.Contract.Finalise(&_Crowdsale.TransactOpts)\n}", "func (_Crowdsale *CrowdsaleTransactor) Finalise(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Crowdsale.contract.Transact(opts, \"finalise\")\n}", "func (tx *transaction) close() {\n\ttx.closed = true\n\n\t// Clear pending blocks that would have been written on commit.\n\ttx.pendingBlocks = nil\n\ttx.pendingBlockData = nil\n\n\t// Clear pending keys that would have been written or deleted on commit.\n\ttx.pendingKeys = nil\n\ttx.pendingRemove = nil\n\n\t// Release the snapshot.\n\tif tx.snapshot != nil {\n\t\ttx.snapshot.Release()\n\t\ttx.snapshot = nil\n\t}\n\n\ttx.db.closeLock.RUnlock()\n\n\t// Release the writer lock for writable transactions to unblock any\n\t// other write transaction which are possibly waiting.\n\tif tx.writable {\n\t\ttx.db.writeLock.Unlock()\n\t}\n}", "func (s *Statement) Finalize() (e error) {\n\treturn SQLiteError(C.sqlite3_finalize(s.cptr))\n}", "func (_Crowdsale *CrowdsaleSession) Finalise() (*types.Transaction, error) {\n\treturn _Crowdsale.Contract.Finalise(&_Crowdsale.TransactOpts)\n}", "func (a *models.App) Finalize() {\n\tvar err error\n\n\ta.DB, err = sql.Close(\"postgres\", connectionString)\n\tutils.ErrCheck(err)\n}", "func (c *Client) FinalizeTransaction(dst *network.ServerIdentity,\n\ttx []byte, signature []byte) (*proto.TransactionResponse, error) {\n\trequest := &proto.TransactionFinalizationRequest{\n\t\tTransaction: tx,\n\t\tSignature: signature,\n\t}\n\tresponse := &proto.TransactionResponse{}\n\n\terr := c.SendProtobuf(dst, request, response)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn response, err\n}", "func (_TransferProxyRegistry *TransferProxyRegistryTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _TransferProxyRegistry.contract.Transact(opts, \"finalize\")\n}", "func (_Finalizable *FinalizableSession) Finalize(fin bool) (*types.Transaction, error) {\n\treturn _Finalizable.Contract.Finalize(&_Finalizable.TransactOpts, fin)\n}", "func (l *EventLogger) CommitTransaction() (lastErr error) {\n\tdefer l.Unlock()\n\tl.endTransaction()\n\treturn l.close()\n}", "func (e *engineImpl) Finalize(\n\tchain engine.ChainReader, header *block.Header,\n\tstate *state.DB, txs []*types.Transaction,\n\treceipts []*types.Receipt, outcxs []*types.CXReceipt,\n\tincxs []*types.CXReceiptsProof, stks []*staking.StakingTransaction,\n) (*types.Block, *big.Int, error) {\n\n\t// Accumulate any block and uncle rewards and commit the final state root\n\t// Header seems complete, assemble into a block and return\n\tpayout, err := AccumulateRewards(\n\t\tchain, state, header, e.Rewarder(), e.Slasher(), e.Beaconchain(),\n\t)\n\tif err != nil {\n\t\treturn nil, nil, ctxerror.New(\"cannot pay block reward\").WithCause(err)\n\t}\n\n\t// Withdraw unlocked tokens to the delegators' accounts\n\t// Only do such at the last block of an epoch\n\tif header.ShardID() == shard.BeaconChainShardID && len(header.ShardState()) > 0 {\n\t\tvalidators, err := chain.ReadValidatorList()\n\t\tif err != nil {\n\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed to read active validators\").WithCause(err)\n\t\t}\n\t\t// Payout undelegated/unlocked tokens\n\t\tfor _, validator := range validators {\n\t\t\twrapper := state.GetStakingInfo(validator)\n\t\t\tif wrapper != nil {\n\t\t\t\tfor i := range wrapper.Delegations {\n\t\t\t\t\tdelegation := &wrapper.Delegations[i]\n\t\t\t\t\ttotalWithdraw := delegation.RemoveUnlockedUndelegations(header.Epoch(), wrapper.LastEpochInCommittee)\n\t\t\t\t\tstate.AddBalance(delegation.DelegatorAddress, totalWithdraw)\n\t\t\t\t}\n\t\t\t\tif err := state.UpdateStakingInfo(validator, wrapper); err != nil {\n\t\t\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed update validator info\").WithCause(err)\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\terr = errors.New(\"[Finalize] validator came back empty \" + common2.MustAddressToBech32(validator))\n\t\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed getting validator info\").WithCause(err)\n\t\t\t}\n\t\t}\n\n\t\t// Set the LastEpochInCommittee field for all external validators in the upcoming epoch.\n\t\tnewShardState, err := header.GetShardState()\n\t\tif err != nil {\n\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed to read shard state\").WithCause(err)\n\t\t}\n\t\tprocessed := make(map[common.Address]struct{})\n\t\tfor i := range newShardState.Shards {\n\t\t\tshard := newShardState.Shards[i]\n\t\t\tfor j := range shard.Slots {\n\t\t\t\tslot := shard.Slots[j]\n\t\t\t\tif slot.EffectiveStake != nil { // For external validator\n\t\t\t\t\t_, ok := processed[slot.EcdsaAddress]\n\t\t\t\t\tif !ok {\n\t\t\t\t\t\tprocessed[slot.EcdsaAddress] = struct{}{}\n\t\t\t\t\t\twrapper := state.GetStakingInfo(slot.EcdsaAddress)\n\t\t\t\t\t\twrapper.LastEpochInCommittee = newShardState.Epoch\n\n\t\t\t\t\t\tif err := state.UpdateStakingInfo(slot.EcdsaAddress, wrapper); err != nil {\n\t\t\t\t\t\t\treturn nil, nil, ctxerror.New(\"[Finalize] failed update validator info\").WithCause(err)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\theader.SetRoot(state.IntermediateRoot(chain.Config().IsS3(header.Epoch())))\n\treturn types.NewBlock(header, txs, receipts, outcxs, incxs, stks), payout, nil\n}", "func (_OwnerProxyRegistry *OwnerProxyRegistryTransactor) Finalize(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _OwnerProxyRegistry.contract.Transact(opts, \"finalize\")\n}", "func (tx *Transaction) Rollback() {\n\tif tx.finished {\n\t\treturn\n\t}\n\n\td, lock := driver.Get()\n\tdefer lock.Unlock()\n\n\td.Rollback(tx.tx)\n\ttx.finished = true\n}", "func (r *dsState) CommitTransaction(c context.Context) error {\n\treturn r.run(c, func() error { return nil })\n}", "func (connection *Connection) Close() {\n\tif connection.adabasToData != nil {\n\t\t_ = connection.adabasToData.BackoutTransaction()\n\t\tconnection.adabasToData.Close()\n\t}\n\tif connection.adabasToMap != nil {\n\t\t_ = connection.adabasToMap.BackoutTransaction()\n\t\tconnection.adabasToMap.Close()\n\t}\n}", "func (c *conn) Rollback() error {\n\tif atomic.LoadInt32(&c.closed) != 0 {\n\t\treturn driver.ErrBadConn\n\t}\n\tif c.txCtx == nil {\n\t\treturn sql.ErrTxDone\n\t}\n\tc.txCtx = nil\n\tstmts := c.stmts\n\tc.stmts = stmts[:0]\n\n\tif len(stmts) == 0 {\n\t\t// there is no statements, so nothing to rollback\n\t\treturn sql.ErrTxDone\n\t}\n\t// the statements will be closed by sql.Tx\n\treturn nil\n}", "func (_Content *ContentTransactor) RunFinalize(opts *bind.TransactOpts, arg0 *big.Int, arg1 *big.Int) (*types.Transaction, error) {\n\treturn _Content.contract.Transact(opts, \"runFinalize\", arg0, arg1)\n}", "func (service *Stainless) FinalizeTransaction(\n\treq *proto.TransactionFinalizationRequest) (network.Message, error) {\n\tsigner := types.HomesteadSigner{}\n\n\tvar tx types.Transaction\n\terr := tx.UnmarshalJSON(req.Transaction)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsignedTx, err := tx.WithSignature(signer, req.Signature)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tsignedBuffer, err := signedTx.MarshalJSON()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tlog.Lvl4(\"Returning\", signedBuffer)\n\n\treturn &proto.TransactionResponse{\n\t\tTransaction: signedBuffer,\n\t}, nil\n}", "func (n *Notary) finalize(acc *wallet.Account, tx *transaction.Transaction, h util.Uint256) error {\n\tnotaryWitness := transaction.Witness{\n\t\tInvocationScript: append([]byte{byte(opcode.PUSHDATA1), 64}, acc.PrivateKey().SignHashable(uint32(n.Network), tx)...),\n\t\tVerificationScript: []byte{},\n\t}\n\tfor i, signer := range tx.Signers {\n\t\tif signer.Account == n.Config.Chain.GetNotaryContractScriptHash() {\n\t\t\ttx.Scripts[i] = notaryWitness\n\t\t\tbreak\n\t\t}\n\t}\n\tnewTx, err := updateTxSize(tx)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to update completed transaction's size: %w\", err)\n\t}\n\n\tn.pushNewTx(newTx, h)\n\n\treturn nil\n}", "func (tangle *Tangle) setTransactionFinalized(transactionID transaction.ID, eventSource EventSource) (modified bool, err error) {\n\t// retrieve metadata and consume\n\tcachedTransactionMetadata := tangle.TransactionMetadata(transactionID)\n\tcachedTransactionMetadata.Consume(func(metadata *TransactionMetadata) {\n\t\t// update the finalized flag of the transaction\n\t\tmodified = metadata.SetFinalized(true)\n\n\t\t// only propagate the changes if the flag was modified\n\t\tif modified {\n\t\t\t// retrieve transaction from the database (for the events)\n\t\t\tcachedTransaction := tangle.Transaction(transactionID)\n\t\t\tdefer cachedTransaction.Release()\n\t\t\tif !cachedTransaction.Exists() {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// trigger the corresponding event\n\t\t\ttangle.Events.TransactionFinalized.Trigger(cachedTransaction, cachedTransactionMetadata)\n\n\t\t\t// propagate changes to value tangle and branch DAG if we were called from the tangle\n\t\t\t// Note: if the update was triggered by a change in the branch DAG then we do not propagate the confirmed\n\t\t\t// and rejected changes yet as those require the branch to be liked before (we instead do it in the\n\t\t\t// BranchLiked event)\n\t\t\tif eventSource == EventSourceTangle {\n\t\t\t\t// propagate changes to the branches (UTXO DAG)\n\t\t\t\tif metadata.Conflicting() {\n\t\t\t\t\t_, err = tangle.branchManager.SetBranchFinalized(metadata.BranchID())\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\ttangle.Events.Error.Trigger(err)\n\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// propagate changes to future cone of transaction (value tangle)\n\t\t\t\ttangle.propagateValuePayloadConfirmedRejectedUpdates(transactionID)\n\t\t\t}\n\t\t}\n\t})\n\n\treturn\n}", "func (connection *Connection) EndTransaction() error {\n\tif connection.adabasToData != nil {\n\t\terr := connection.adabasToData.EndTransaction()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif connection.adabasToMap != nil {\n\t\terr := connection.adabasToMap.EndTransaction()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (*txDriver) Rollback() error { return nil }", "func (*txDriver) Rollback() error { return nil }", "func (ct *ConnectionTransportTLS) Finalize() {\n\tct.mutex.Lock()\n\tdefer ct.mutex.Unlock()\n\tif ct.transport != nil {\n\t\tct.transport.Close()\n\t}\n\tct.transport = ct.stagedTransport\n\tct.stagedTransport = nil\n\tct.srvRemote.Reset()\n}", "func (tx *transaction) Rollback() error {\n\tif tx.terminated {\n\t\treturn errors.Wrap(engine.ErrTransactionDiscarded)\n\t}\n\n\ttx.terminated = true\n\n\tif tx.writable {\n\t\tfor _, undo := range tx.onRollback {\n\t\t\tundo()\n\t\t}\n\t}\n\n\tselect {\n\tcase <-tx.ctx.Done():\n\t\treturn tx.ctx.Err()\n\tdefault:\n\t}\n\n\treturn nil\n}", "func (consensus *Consensus) Finalize(chain ChainReader, header *types.Header, state *state.DB, txs []*types.Transaction, receipts []*types.Receipt) (*types.Block, error) {\n\t// Accumulate any block and uncle rewards and commit the final state root\n\t// Header seems complete, assemble into a block and return\n\taccumulateRewards(chain.Config(), state, header)\n\theader.Root = state.IntermediateRoot(false)\n\treturn types.NewBlock(header, txs, receipts), nil\n}", "func (cs *ConsensusState) tryFinalizeCommit(height int64) {\n\tif cs.Height != height {\n\t\tpanic(fmt.Sprintf(\"tryFinalizeCommit() cs.Height: %v vs height: %v\", cs.Height, height))\n\t}\n\n\tvar blockID tmtypes.QbftBlockID\n\tvar ok bool\n\tif cs.state.Sequence == 0 {\n\t\tblockID, ok = cs.Votes.Precommits(cs.CommitRound).TwoThirdsMajority()\n\t} else {\n\t\tblockID, ok = cs.Votes.Prevotes(cs.CommitRound).TwoThirdsMajority()\n\t}\n\tif !ok || len(blockID.Hash) == 0 {\n\t\tqbftlog.Error(\"Attempt to finalize failed. There was no +2/3 majority, or +2/3 was for <nil>.\")\n\t\tqbftlog.Info(fmt.Sprintf(\"Continue consensus. Current: %v/%v/%v\", cs.Height, cs.CommitRound, cs.Step), \"cost\", types.Since(cs.begCons))\n\t\treturn\n\t}\n\tif !cs.ProposalBlock.HashesTo(blockID.Hash) {\n\t\t// TODO: this happens every time if we're not a validator (ugly logs)\n\t\t// TODO: ^^ wait, why does it matter that we're a validator?\n\t\tqbftlog.Error(\"Attempt to finalize failed. We don't have the commit block.\",\n\t\t\t\"ProposalBlock-hash\", fmt.Sprintf(\"%X\", cs.ProposalBlock.Hash()), \"CommitBlock-hash\", fmt.Sprintf(\"%X\", blockID.Hash))\n\t\tqbftlog.Info(fmt.Sprintf(\"Continue consensus. Current: %v/%v/%v\", cs.Height, cs.CommitRound, cs.Step), \"cost\", types.Since(cs.begCons))\n\t\treturn\n\t}\n\n\t// go\n\tcs.finalizeCommit(height)\n}", "func (_LvRecordableStream *LvRecordableStreamTransactor) RunFinalize(opts *bind.TransactOpts, arg0 *big.Int, arg1 *big.Int) (*types.Transaction, error) {\n\treturn _LvRecordableStream.contract.Transact(opts, \"runFinalize\", arg0, arg1)\n}", "func (hc *LegacyHealthCheckImpl) finalizeConn(hcc *legacyHealthCheckConn) {\n\thcc.tabletStats.Up = false\n\thcc.setServingState(false, \"finalizeConn closing connection\")\n\t// Note: checkConn() exits only when hcc.ctx.Done() is closed. Thus it's\n\t// safe to simply get Err() value here and assign to LastError.\n\thcc.tabletStats.LastError = hcc.ctx.Err()\n\thc.updateHealth(hcc.tabletStats.Copy(), nil)\n\tif hcc.conn != nil {\n\t\t// Don't use hcc.ctx because it's already closed.\n\t\t// Use a separate context, and add a timeout to prevent unbounded waits.\n\t\tctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)\n\t\tdefer cancel()\n\t\thcc.conn.Close(ctx)\n\t\thcc.conn = nil\n\t}\n}", "func (cd *ConnectionDetails) Finalize() error {\n\tcd.Dialect = normalizeSynonyms(cd.Dialect)\n\n\tif cd.Options == nil { // for safety\n\t\tcd.Options = make(map[string]string)\n\t}\n\n\tif cd.URL != \"\" {\n\t\tif err := cd.withURL(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif fin, ok := finalizer[cd.Dialect]; ok {\n\t\tfin(cd)\n\t}\n\n\tif DialectSupported(cd.Dialect) {\n\t\tif cd.Database != \"\" || cd.URL != \"\" {\n\t\t\treturn nil\n\t\t}\n\t\treturn errors.New(\"no database or URL specified\")\n\t}\n\treturn errors.Errorf(\"unsupported dialect '%v'\", cd.Dialect)\n}", "func (c *Conn) Close() error {\n\tif c == nil {\n\t\treturn errors.New(\"nil sqlite database\")\n\t}\n\tif c.db == nil {\n\t\treturn nil\n\t}\n\n\tc.stmtCache.flush()\n\n\trv := C.sqlite3_close(c.db)\n\n\tif rv&0xFF == C.SQLITE_BUSY {\n\t\t// Dangling statements\n\t\tstmt := C.sqlite3_next_stmt(c.db, nil)\n\t\tfor stmt != nil {\n\t\t\tif C.sqlite3_stmt_busy(stmt) != 0 {\n\t\t\t\tLog(C.SQLITE_MISUSE, \"Dangling statement (not reset): \\\"\"+C.GoString(C.sqlite3_sql(stmt))+\"\\\"\")\n\t\t\t} else {\n\t\t\t\tLog(C.SQLITE_MISUSE, \"Dangling statement (not finalized): \\\"\"+C.GoString(C.sqlite3_sql(stmt))+\"\\\"\")\n\t\t\t}\n\t\t\tC.sqlite3_finalize(stmt)\n\t\t\tstmt = C.sqlite3_next_stmt(c.db, nil)\n\t\t}\n\t\trv = C.sqlite3_close(c.db)\n\t}\n\n\tif rv != C.SQLITE_OK {\n\t\tLog(int32(rv), \"error while closing Conn\")\n\t\treturn c.error(rv, \"Conn.Close\")\n\t}\n\tc.db = nil\n\treturn nil\n}", "func (t *TransactionContext) Finish() error {\n\tif t.PfsPropagater != nil {\n\t\tif err := t.PfsPropagater.Run(); err != nil {\n\t\t\treturn errors.EnsureStack(err)\n\t\t}\n\t}\n\tif t.PpsPropagater != nil {\n\t\tif err := t.PpsPropagater.Run(); err != nil {\n\t\t\treturn errors.EnsureStack(err)\n\t\t}\n\t}\n\tif t.PpsJobStopper != nil {\n\t\tif err := t.PpsJobStopper.Run(); err != nil {\n\t\t\treturn errors.EnsureStack(err)\n\t\t}\n\t}\n\tif t.PpsJobFinisher != nil {\n\t\tif err := t.PpsJobFinisher.Run(); err != nil {\n\t\t\treturn errors.EnsureStack(err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *RawMapper) Close() {\n\tif m != nil && m.tx != nil {\n\t\tm.tx.Rollback()\n\t}\n}", "func (_m *Repository) RollbackTransaction(ctx context.Context) {\n\t_m.Called(ctx)\n}", "func (c *Conn) Commit() error {\n\t// Although there are situations when it is possible to recover and continue a transaction,\n\t// it is considered a best practice to always issue a ROLLBACK if an error is encountered.\n\t// In situations when SQLite was already forced to roll back the transaction and has returned to autocommit mode,\n\t// the ROLLBACK will do nothing but return an error that can be safely ignored.\n\terr := c.FastExec(\"COMMIT\")\n\tif err != nil && !c.GetAutocommit() {\n\t\t_ = c.Rollback()\n\t}\n\treturn err\n}", "func (d *DmarcARRContext) Close() (err error) {\n\tif d.Stmt != nil {\n\t\td.Stmt.Close()\n\t}\n\n\tif d.Txn != nil {\n\t\terr = d.Txn.Commit()\n\t}\n\treturn\n}", "func (_m *Repository) CommitTransaction(ctx context.Context) {\n\t_m.Called(ctx)\n}", "func (_Logger *LoggerTransactor) LogRollbackFinalisation(opts *bind.TransactOpts, totalBatchesSlashed *big.Int) (*types.Transaction, error) {\n\treturn _Logger.contract.Transact(opts, \"logRollbackFinalisation\", totalBatchesSlashed)\n}", "func (s *session) Close() {\n\tctx := context.TODO()\n\ts.RollbackTxn(ctx)\n}", "func (tangle *Tangle) SetTransactionFinalized(transactionID transaction.ID) (modified bool, err error) {\n\treturn tangle.setTransactionFinalized(transactionID, EventSourceTangle)\n}", "func (s *Session) CommitTransaction(context.Context) error {\n\t// acquire lock\n\ts.mutex.Lock()\n\tdefer s.mutex.Unlock()\n\n\t// check if ended\n\tif s.ended {\n\t\treturn ErrSessionEnded\n\t}\n\n\t// check transaction\n\tif s.txn == nil {\n\t\treturn fmt.Errorf(\"missing transaction\")\n\t}\n\n\t// get and unset transaction\n\ttxn := s.txn\n\ts.txn = nil\n\n\t// commit transaction\n\terr := s.engine.Commit(txn)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (i *Item) Close(ctx context.Context) (err error) {\n\tif i.tx == nil {\n\t\treturn nil\n\t}\n\terr = i.tx.Commit(ctx)\n\ti.tx = nil\n\treturn err\n}", "func (p DbPlugin) AfterRequest(c *revel.Controller) {\n\tif err := c.Txn.Commit(); err != nil {\n\t\tif err != sql.ErrTxDone {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\tc.Txn = nil\n}", "func (cdt *SqlDBTx) Rollback() error {\n\treturn nil\n}", "func (tx *transaction) Commit() error {\n\tif tx.terminated {\n\t\treturn errors.Wrap(engine.ErrTransactionDiscarded)\n\t}\n\n\tif !tx.writable {\n\t\treturn errors.Wrap(engine.ErrTransactionReadOnly)\n\t}\n\n\tselect {\n\tcase <-tx.ctx.Done():\n\t\treturn tx.Rollback()\n\tdefault:\n\t}\n\n\ttx.terminated = true\n\n\tfor _, fn := range tx.onCommit {\n\t\tfn()\n\t}\n\n\treturn nil\n}", "func Cleanup(t *testing.T, db db.DB) {\n\tif db != nil {\n\t\tassert.NoError(t, db.Close())\n\t}\n}", "func (m *Manager) finalCleanup() {\n\tm.lock()\n\tfor m.state.cleaned != initiallyCleanedState {\n\t\tif m.state.cleaned == finallyCleanedState {\n\t\t\tm.unlock()\n\t\t\treturn // do not clean if already cleaned\n\t\t} else if m.state.cleaned == noCleanedState {\n\t\t\tm.state.cleanWait.Wait() // wait for wake up from `cleanup` or other `finalCleanup` method\n\t\t}\n\t}\n\n\tglog.Infof(\"%s %s has started a final cleanup\", cmn.DSortName, m.ManagerUUID)\n\tnow := time.Now()\n\n\tif err := m.cleanupStreams(); err != nil {\n\t\tglog.Error(err)\n\t}\n\n\tif err := m.dsorter.finalCleanup(); err != nil {\n\t\tglog.Error(err)\n\t}\n\n\t// The reason why this is not in regular cleanup is because we are only sure\n\t// that this can be freed once we cleanup streams - streams are asynchronous\n\t// and we may have race between in-flight request and cleanup.\n\tm.recManager.Cleanup()\n\n\tm.creationPhase.metadata.SendOrder = nil\n\tm.creationPhase.metadata.Shards = nil\n\n\tm.finishedAck.m = nil\n\n\t// Update clean state\n\tm.state.cleaned = finallyCleanedState\n\tm.state.cleanWait.Signal() // if there is another `finalCleanup` waiting it should be woken up to check the state and exit\n\tm.unlock()\n\n\tManagers.persist(m.ManagerUUID)\n\tglog.Infof(\"%s %s final cleanup has been finished in %v\", cmn.DSortName, m.ManagerUUID, time.Since(now))\n}", "func cleanup(c *context.Context, t *testing.T) {\n\tdbPath := c.DB.Path()\n\n\terr := c.DB.Close()\n\tif err != nil {\n\t\tt.Logf(\"Could not close db: %v\", err)\n\t}\n\n\terr = os.Remove(dbPath)\n\tif err != nil {\n\t\tt.Logf(\"Could not remove db (%v): %v\", dbPath, err)\n\t}\n}", "func RollbackTransaction(orm orm.Ormer) {\n\terr := orm.Rollback()\n\tif err != nil {\n\t\tklog.Errorf(\"failed to rollback transaction, err: %v\", err)\n\t}\n}", "func (mc *MysqlConn) cleanup() {\n\t// Makes cleanup idempotent\n\tif mc.netConn != nil {\n\t\tif err := mc.netConn.Close(); err != nil {\n\t\t\terrLog.Print(err)\n\t\t}\n\t\tmc.netConn = nil\n\t}\n\tmc.cfg = nil\n}", "func finalizeBlock(block *protocol.Block) error {\n\t//Check if we have a slashing proof that we can add to the block.\n\t//The slashingDict is updated when a new block is received and when a slashing proof is provided.\n\tlogger.Printf(\"-- Start Finalize\")\n\tif len(slashingDict) != 0 {\n\t\t//Get the first slashing proof.\n\t\tfor hash, slashingProof := range slashingDict {\n\t\t\tblock.SlashedAddress = hash\n\t\t\tblock.ConflictingBlockHash1 = slashingProof.ConflictingBlockHash1\n\t\t\tblock.ConflictingBlockHash2 = slashingProof.ConflictingBlockHash2\n\t\t\tblock.ConflictingBlockHashWithoutTx1 = slashingProof.ConflictingBlockHashWithoutTx1\n\t\t\tblock.ConflictingBlockHashWithoutTx2 = slashingProof.ConflictingBlockHashWithoutTx2\n\t\t\tbreak\n\t\t}\n\t}\n\n\t//Merkle tree includes the hashes of all txs in this block\n\tblock.MerkleRoot = protocol.BuildMerkleTree(block).MerkleRoot()\n\tvalidatorAcc, err := storage.GetAccount(protocol.SerializeHashContent(validatorAccAddress))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvalidatorAccHash := validatorAcc.Hash()\n\tcopy(block.Beneficiary[:], validatorAccHash[:])\n\n\t// Cryptographic Sortition for PoS in Bazo\n\t// The commitment proof stores a signed message of the Height that this block was created at.\n\tcommitmentProof, err := crypto.SignMessageWithRSAKey(commPrivKey, fmt.Sprint(block.Height))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\n\t//Block hash with MerkleTree and therefore, including all transactions\n\tpartialHash := block.HashBlock()\n\n\t//Block hash without MerkleTree and therefore, without any transactions\n\tpartialHashWithoutMerkleRoot := block.HashBlockWithoutMerkleRoot()\n\n\tprevProofs := GetLatestProofs(activeParameters.num_included_prev_proofs, block)\n\tnonce, err := proofOfStake(getDifficulty(), block.PrevHash, prevProofs, block.Height, validatorAcc.Balance, commitmentProof)\n\tif err != nil {\n\t\t//Delete all partially added transactions.\n\t\tif nonce == -2 {\n\t\t\tfor _, tx := range storage.FundsTxBeforeAggregation {\n\t\t\t\tstorage.WriteOpenTx(tx)\n\t\t\t}\n\t\t\tstorage.DeleteAllFundsTxBeforeAggregation()\n\t\t}\n\t\treturn err\n\t}\n\n\tvar nonceBuf [8]byte\n\tbinary.BigEndian.PutUint64(nonceBuf[:], uint64(nonce))\n\tblock.Nonce = nonceBuf\n\tblock.Timestamp = nonce\n\n\t//Put pieces together to get the final hash.\n\tblock.Hash = sha3.Sum256(append(nonceBuf[:], partialHash[:]...))\n\tblock.HashWithoutTx = sha3.Sum256(append(nonceBuf[:], partialHashWithoutMerkleRoot[:]...))\n\n\t//This doesn't need to be hashed, because we already have the merkle tree taking care of consistency.\n\tblock.NrAccTx = uint16(len(block.AccTxData))\n\tblock.NrFundsTx = uint16(len(block.FundsTxData))\n\tblock.NrConfigTx = uint8(len(block.ConfigTxData))\n\tblock.NrStakeTx = uint16(len(block.StakeTxData))\n\tblock.NrAggTx = uint16(len(block.AggTxData))\n\n\tcopy(block.CommitmentProof[0:crypto.COMM_PROOF_LENGTH], commitmentProof[:])\n\tlogger.Printf(\"-- End Finalization\")\n\treturn nil\n}", "func finalizer(a *Adapter) {\n\ta.db.Destroy()\n}", "func (tx *MockTx) Close() error {\n\treturn tx.Tx.Rollback()\n}", "func (t *Transaction) Close() error {\n\t// lock\n\tt.mu.Lock()\n\t// check if already closed\n\tif !t.alive {\n\t\tt.mu.Unlock()\n\t\treturn status.ErrClosed\n\t}\n\tt.alive = false\n\tt.shutdownOnce.Do(func() {\n\t\tclose(t.shutdown)\n\t})\n\tt.mu.Unlock()\n\t<-t.signalClosed\n\treturn nil\n}", "func (_LvStreamRightsHolder *LvStreamRightsHolderTransactor) RunFinalize(opts *bind.TransactOpts, arg0 *big.Int, arg1 *big.Int) (*types.Transaction, error) {\n\treturn _LvStreamRightsHolder.contract.Transact(opts, \"runFinalize\", arg0, arg1)\n}", "func FinalizeTransition(tr module.Transition, opt int, noFlush bool) error {\n\ttst := tr.(*transition)\n\tif opt&module.FinalizeNormalTransaction == module.FinalizeNormalTransaction && !noFlush {\n\t\tif err := tst.finalizeNormalTransaction(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif opt&module.FinalizePatchTransaction == module.FinalizePatchTransaction && !noFlush {\n\t\tif err := tst.finalizePatchTransaction(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif opt&module.FinalizeResult == module.FinalizeResult {\n\t\tif err := tst.finalizeResult(noFlush); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (ec *eConn) initiateFinalization() {\n\twithLock(&ec.dieCtl.Mutex, func() {\n\t\tec.conn.Close()\n\t\tec.eConnManager.removeEConn(ec)\n\n\t\tif ec.dieCtl.closed {\n\t\t\treturn\n\t\t}\n\n\t\tec.dieCtl.closed = true\n\t\tec.dieCtl.closeType = closeTypeLocal\n\n\t\tgo func() { // Execute these in a goroutine to release the lock on dieCtl\n\t\t\tec.writeBufMutex.Lock() // Ensure the following signal is received\n\t\t\tec.writeBufCond.Signal()\n\t\t\tec.writeBufMutex.Unlock()\n\t\t}()\n\t})\n\n}", "func (dbi *DB) Rollback() {\r\n\tif dbi.status {\r\n\t\tdbi.createOperation(\"DB_ROLLBACK\")\r\n\t\tdbi.data.commPrepare()\r\n\t\tdbi.data.comm()\r\n\t\tdbi.data.commParse()\r\n\t\tif Show {\r\n\t\t\tmylog.Println(\"[\" + dbi.Sid + \"]SQL=ROLLBACK;\")\r\n\t\t}\r\n\t}\r\n}", "func (c *conn) Commit() (err error) {\n\tif atomic.LoadInt32(&c.closed) != 0 {\n\t\treturn driver.ErrBadConn\n\t}\n\tif c.txCtx == nil {\n\t\treturn sql.ErrTxDone\n\t}\n\tctx := c.txCtx\n\tstmts := c.stmts\n\tc.txCtx = nil\n\tc.stmts = stmts[:0]\n\n\tif len(stmts) == 0 {\n\t\treturn nil\n\t}\n\tfor _, stmt := range stmts {\n\t\tc.log(\"commit statement: \", stmt.prefix, stmt.pattern)\n\t\tif err = stmt.commit(ctx); err != nil {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn\n}", "func (t *transaction) Rollback() {\n\tt.Request.URL = t.url\n\n\tif t.body != nil {\n\t\tt.body.Seek(0, 0) // rewind\n\t}\n}", "func (p *HelloWorld) AfterExecuteTransactions(b *types.Block, ctw *types.ContextWrapper) error {\n\treturn nil\n}", "func (t *Transaction) Rollback() {\n\tif err := t.tx.Rollback(); err != nil {\n\t\terr = checkErr(err)\n\t\tlogging.SystemErrorf(\"Unable to rollback %s transaction: %v\", t.dbType, err)\n\t}\n}", "func (runner *statementRunner) closeConnection() error {\n\tif runner.connection != nil {\n\t\trunner.lastBookmark = runner.connection.LastBookmark()\n\n\t\terr := runner.connection.Close()\n\t\trunner.connection = nil\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (txC *TxConnection) Rollback() (err error) {\n\treturn txC.Tx.Rollback()\n}", "func (ingest *Ingestion) Close() error {\n\treturn ingest.commit()\n}", "func (cdt *SqlDBTx) TxEnd(txFunc func() error) error {\n\treturn nil\n}", "func connectionFinalizer(c *Connection) {\n\tc.Close()\n}", "func connectionFinalizer(c *Connection) {\n\tc.Close()\n}", "func (s *SqlDb) Commit(ctx context.Context, txid TransactionID) {\n\tvar c *SqlContext\n\ti := ctx.Value(goradd.SqlContext)\n\tif i == nil {\n\t\tpanic(\"Can't use transactions without pre-loading a context\")\n\t} else {\n\t\tc = i.(*SqlContext)\n\t}\n\n\tif c.txCount != int(txid) {\n\t\tpanic(\"Missing Rollback after previous Begin\")\n\t}\n\n\tif c.txCount == 0 {\n\t\tpanic(\"Called Commit without a matching Begin\")\n\t}\n\tif c.txCount == 1 {\n\t\terr := c.tx.Commit()\n\t\tif err != nil {\n\t\t\tpanic(err.Error())\n\t\t}\n\t\tc.tx = nil\n\t}\n\tc.txCount--\n}", "func (tx *Tx) Rollback(ctx context.Context) error {\n\terr := tx.t.Rollback(ctx)\n\tif tx.c != nil {\n\t\ttx.c.Release()\n\t\ttx.c = nil\n\t}\n\treturn err\n}", "func (s *sqlite) Close() error {\n\tif s != nil {\n\t\tif err := s.insertStmt.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif err := s.queryStmt.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (s *FileSnapshotSink) finalize() error {\n\t// Flush any remaining data\n\tif err := s.buffered.Flush(); err != nil {\n\t\treturn err\n\t}\n\n\t// Get the file size\n\tstat, statErr := s.stateFile.Stat()\n\n\t// Close the file\n\tif err := s.stateFile.Close(); err != nil {\n\t\treturn err\n\t}\n\n\t// Set the file size, check after we close\n\tif statErr != nil {\n\t\treturn statErr\n\t}\n\ts.meta.Size = stat.Size()\n\n\t// Set the CRC\n\ts.meta.CRC = s.stateHash.Sum(nil)\n\treturn nil\n}", "func (c *ProjectFinalizerController) finalize(namespace *v1.Namespace) error {\n\t// if namespace is not terminating, ignore it\n\tif namespace.Status.Phase != v1.NamespaceTerminating {\n\t\treturn nil\n\t}\n\n\t// if we already processed this namespace, ignore it\n\tif projectutil.Finalized(namespace) {\n\t\treturn nil\n\t}\n\n\t// we have removed content, so mark it finalized by us\n\t_, err := projectutil.Finalize(c.client, namespace)\n\treturn err\n}", "func (t *Transaction) Commit() {\n\tif err := t.tx.Commit(); err != nil {\n\t\terr = checkErr(err)\n\t\tlogging.SystemErrorf(\"Unable to commit %s transaction: %v\", t.dbType, err)\n\t}\n}", "func (_Logger *LoggerTransactorSession) LogRollbackFinalisation(totalBatchesSlashed *big.Int) (*types.Transaction, error) {\n\treturn _Logger.Contract.LogRollbackFinalisation(&_Logger.TransactOpts, totalBatchesSlashed)\n}", "func (b *Block) Finalize(endorsements []*endorsement.Endorsement, ts time.Time) error {\n\tif len(b.endorsements) != 0 {\n\t\treturn errors.New(\"the block has been finalized\")\n\t}\n\tb.endorsements = endorsements\n\tb.commitTime = ts\n\n\treturn nil\n}", "func (cs *ConsensusState) finalizeCommit(height int64) {\n\tif cs.Height != height || cs.Step != ttypes.RoundStepCommit {\n\t\tqbftlog.Debug(fmt.Sprintf(\"finalizeCommit(%v): Invalid args. Current step: %v/%v/%v\", height, cs.Height, cs.Round, cs.Step))\n\t\treturn\n\t}\n\n\tblock := cs.ProposalBlock\n\tstateCopy := cs.state.Copy()\n\tqbftlog.Debug(\"finalizeCommit validators of statecopy\", \"validators\", stateCopy.Validators)\n\t// NOTE: the block.AppHash wont reflect these txs until the next block\n\tvar err error\n\tstateCopy, err = cs.blockExec.ApplyBlock(stateCopy, ttypes.BlockID{QbftBlockID: &tmtypes.QbftBlockID{Hash: block.Hash()}}, block)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"finalizeCommit ApplyBlock fail: %v\", err))\n\t}\n\n\t// commit block\n\tcommitBlock := cs.ProposalBlock.Data\n\tif !DetachExec() {\n\t\tcs.client.CommitBlock(commitBlock.Clone())\n\t}\n\tif bytes.Equal(cs.privValidator.GetAddress(), block.QbftBlock.Header.ProposerAddr) {\n\t\tqbftlog.Info(fmt.Sprintf(\"Proposer reach consensus. Current: %v/%v/%v\", cs.Height, cs.Round, cs.Step),\n\t\t\t\"CommitRound\", cs.CommitRound, \"tx-len\", len(commitBlock.Txs), \"cost\", types.Since(cs.begCons),\n\t\t\t\"proposer-addr\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(block.QbftBlock.Header.ProposerAddr)),\n\t\t\t\"seq\", block.Header.Sequence)\n\t} else {\n\t\tqbftlog.Info(fmt.Sprintf(\"Not-Proposer reach consensus. Current: %v/%v/%v\", cs.Height, cs.Round, cs.Step),\n\t\t\t\"CommitRound\", cs.CommitRound, \"tx-len\", len(commitBlock.Txs), \"cost\", types.Since(cs.begCons),\n\t\t\t\"proposer-addr\", fmt.Sprintf(\"%X\", ttypes.Fingerprint(block.QbftBlock.Header.ProposerAddr)),\n\t\t\t\"seq\", block.Header.Sequence)\n\t}\n\treqblock := cs.client.WaitBlock(height)\n\tstateCopy.LastResultsHash = reqblock.Hash(cs.client.GetAPI().GetConfig())\n\n\t//check whether need update validator nodes\n\tqbftNodes, err := cs.client.QueryValidatorsByHeight(block.Header.Height)\n\tif err == nil && qbftNodes != nil {\n\t\tif len(qbftNodes.Nodes) > 0 {\n\t\t\tqbftlog.Info(\"finalizeCommit validators of statecopy update\", \"update-qbftNodes\", qbftNodes)\n\t\t\tprevValSet := stateCopy.LastValidators.Copy()\n\t\t\tnextValSet := prevValSet.Copy()\n\t\t\terr := updateValidators(nextValSet, qbftNodes.Nodes)\n\t\t\tif err != nil {\n\t\t\t\tqbftlog.Error(\"Error changing validator set\", \"error\", err)\n\t\t\t}\n\t\t\t// change results from this height but only applies to the next height\n\t\t\tstateCopy.LastHeightValidatorsChanged = block.Header.Height + 1\n\t\t\tstateCopy.Validators = nextValSet\n\t\t}\n\t}\n\tqbftlog.Debug(\"finalizeCommit validators of statecopy\", \"validators\", stateCopy.Validators.String())\n\n\t// save local state and seen commit\n\tvar seenCommit *tmtypes.QbftCommit\n\tif cs.ProposalBlock.Header.Sequence == 0 {\n\t\tprecommits := cs.Votes.Precommits(cs.CommitRound)\n\t\tseenCommit = precommits.MakeCommit()\n\t} else {\n\t\tprevotes := cs.Votes.Prevotes(cs.CommitRound)\n\t\tseenCommit = prevotes.MakeCommonCommit()\n\t}\n\tnewState := SaveState(stateCopy)\n\terr = cs.client.csStore.SaveConsensusState(height, newState, seenCommit)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"finalizeCommit SaveSeenCommit fail: %v\", err))\n\t}\n\tqbftlog.Info(fmt.Sprintf(\"Save consensus state. Current: %v/%v/%v\", cs.Height, cs.CommitRound, cs.Step), \"cost\", types.Since(cs.begCons))\n\n\t// NewHeightStep!\n\tcs.updateToState(stateCopy)\n\n\t// cs.StartTime is already set.\n\t// Schedule Round0 to start soon.\n\tcs.scheduleRound0(&cs.RoundState)\n\n\t// By here,\n\t// * cs.Height has been increment to height+1\n\t// * cs.Step is now ttypes.RoundStepNewHeight\n\t// * cs.StartTime is set to when we will start round0.\n\t// Execute and commit the block, update and save the state, and update the mempool.\n\n}", "func (r *Request) finalize() error {\n\tfor len(r.funcs) > 0 {\n\t\tf := r.funcs[0]\n\t\tr.funcs = r.funcs[1:len(r.funcs)]\n\t\tf(r)\n\t}\n\tif len(r.errs) > 0 {\n\t\treturn r.errs[0]\n\t}\n\treturn nil\n}", "func (b *BluetoothAdapter) Finalize() (errs []error) {\n\treturn\n}", "func (c *GormController) Commit() r.Result {\n\tif c.Txn == nil {\n\t\treturn nil\n\t}\n\tc.Txn.Commit()\n\tif err := c.Txn.Error; err != nil && err != sql.ErrTxDone {\n\t\tpanic(err)\n\t}\n\tc.Txn = nil\n\treturn nil\n}", "func (c *GormController) Commit() revel.Result {\n if c.Txn == nil {\n return nil\n }\n c.Txn.Commit()\n if err := c.Txn.Error; err != nil && err != sql.ErrTxDone {\n panic(err)\n }\n c.Txn = nil\n return nil\n}", "func (_BaseAccessControlGroup *BaseAccessControlGroupTransactor) CleanUpAll(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _BaseAccessControlGroup.contract.Transact(opts, \"cleanUpAll\")\n}", "func (scs *StorageContractSet) Close() (err error) {\n\tscs.db.Close()\n\t_, err = scs.wal.CloseIncomplete()\n\treturn\n}", "func (session *Session) Close() {\n\tfor _, v := range session.stmtCache {\n\t\tv.Close()\n\t}\n\n\tif session.db != nil {\n\t\t// When Close be called, if session is a transaction and do not call\n\t\t// Commit or Rollback, then call Rollback.\n\t\tif session.Tx != nil && !session.IsCommitedOrRollbacked {\n\t\t\tsession.Rollback()\n\t\t}\n\t\tsession.Tx = nil\n\t\tsession.stmtCache = nil\n\t\tsession.db = nil\n\n\t\tsession.Statement.Init()\n\t\tsession.IsAutoCommit = true\n\t\tsession.IsCommitedOrRollbacked = false\n\t\tsession.IsAutoClose = false\n\t\tsession.AutoResetStatement = true\n\t\tsession.prepareStmt = false\n\n\t\t// processors\n\t\tsession.afterInsertBeans = nil\n\t\tsession.afterUpdateBeans = nil\n\t\tsession.afterDeleteBeans = nil\n\t\tsession.beforeClosures = nil\n\t\tsession.afterClosures = nil\n\t}\n}", "func (tx *transaction) Commit() error {\n\t// Prevent commits on managed transactions.\n\tif tx.managed {\n\t\ttx.close()\n\t\tpanic(\"managed transaction commit not allowed\")\n\t}\n\n\t// Ensure transaction state is valid.\n\tif err := tx.checkClosed(); err != nil {\n\t\treturn err\n\t}\n\n\t// Regardless of whether the commit succeeds, the transaction is closed\n\t// on return.\n\tdefer tx.close()\n\n\t// Ensure the transaction is writable.\n\tif !tx.writable {\n\t\tstr := \"Commit requires a writable database transaction\"\n\t\treturn database.MakeError(database.ErrTxNotWritable, str, nil)\n\t}\n\n\t// Write pending data. The function will rollback if any errors occur.\n\treturn tx.writePendingAndCommit()\n}" ]
[ "0.67475665", "0.67263764", "0.6428609", "0.64016694", "0.6389165", "0.6334275", "0.6285352", "0.6233909", "0.6224551", "0.6189027", "0.6181659", "0.61503744", "0.6138814", "0.60948277", "0.6061247", "0.6054606", "0.6040649", "0.60332334", "0.60171646", "0.60053915", "0.5936391", "0.58637124", "0.58557117", "0.58541596", "0.58209", "0.5760764", "0.5759793", "0.5668726", "0.5659762", "0.5645022", "0.5620354", "0.5548461", "0.55375355", "0.55180794", "0.5498715", "0.5491186", "0.54870474", "0.54870474", "0.5463653", "0.54561454", "0.5414731", "0.53799856", "0.5379003", "0.5377472", "0.5356428", "0.5327333", "0.53107935", "0.529866", "0.5279367", "0.52700055", "0.526913", "0.52545106", "0.5250018", "0.52466595", "0.52425057", "0.52337754", "0.520453", "0.52001107", "0.51993173", "0.5193506", "0.5189938", "0.5188129", "0.5184587", "0.5174695", "0.5174476", "0.5160198", "0.51573634", "0.5154523", "0.51520383", "0.51480675", "0.51454246", "0.5132982", "0.51272553", "0.51216376", "0.51199406", "0.51185286", "0.51169413", "0.5111133", "0.51029456", "0.5099059", "0.5097887", "0.50897086", "0.50897086", "0.50759494", "0.506972", "0.5050832", "0.50378627", "0.5030908", "0.5029061", "0.50282305", "0.5027011", "0.5011879", "0.50047296", "0.50021523", "0.50009507", "0.5000416", "0.49988857", "0.49936292", "0.49915725", "0.4982105" ]
0.8192969
0
NewStateManager constructs a new StateManager which saves data at the location specified in cfg and operates under the given options. The returned StateManager will not save more often than every 10 seconds and will not reliably return errors with Save, but will log them appropriately.
func NewStateManager(cfg *config.Config, options ...Option) (StateManager, error) { fi, err := os.Stat(cfg.DataDir) if err != nil { return nil, err } if !fi.IsDir() { return nil, errors.New("State manager DataDir must exist") } state := &state{ Data: make(saveableState), Version: ECSDataVersion, } manager := &basicStateManager{ statePath: cfg.DataDir, state: state, } for _, option := range options { option(manager) } manager.platformDependencies = newPlatformDependencies() return manager, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewStateManager(etcdEndpoints []string, dialTimeout, requestTimeout time.Duration) (StateManager, error) {\n\n\tsm := stateManager{\n\t\tstopChan: make([]chan interface{}, 0),\n\t\trequestTimeout: requestTimeout,\n\t}\n\n\tif cli, err := clientv3.New(clientv3.Config{\n\t\tDialTimeout: dialTimeout,\n\t\tEndpoints: etcdEndpoints,\n\t}); err == nil {\n\t\tsm.cli = cli\n\t} else {\n\t\treturn nil, err\n\t}\n\tsm.kv = clientv3.NewKV(sm.cli)\n\n\treturn &sm, nil\n}", "func NewStateManager() *StateManager {\n\treturn &StateManager{\n\t\tServersMap: make(map[string]Server),\n\t}\n}", "func NewStateMgr() *StateMgr {\n\treturn &StateMgr{current: Running, ime: true}\n}", "func NewStateManager(conf config.Config) (Manager, error) {\n\tname := \"postgres\"\n\tif conf.IsSet(\"state_manager\") {\n\t\tname = conf.GetString(\"state_manager\")\n\t}\n\n\tswitch name {\n\tcase \"postgres\":\n\t\tpgm := &SQLStateManager{}\n\t\terr := pgm.Initialize(conf)\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"problem initializing SQLStateManager\")\n\t\t}\n\t\treturn pgm, nil\n\tdefault:\n\t\treturn nil, errors.Errorf(\"state.Manager named [%s] not found\", name)\n\t}\n}", "func New(cfg *Config) (*SyncManager, error) {\n\tsm := SyncManager{\n\t\tcfg: *cfg,\n\t\ttxMemPool: make(map[common.Uint256]struct{}),\n\t\trequestedTxns: make(map[common.Uint256]struct{}),\n\t\trequestedBlocks: make(map[common.Uint256]struct{}),\n\t\tpeerStates: make(map[*peer.Peer]*peerSyncState),\n\t\tmsgChan: make(chan interface{}, cfg.MaxPeers*3),\n\t\tquit: make(chan struct{}),\n\t}\n\n\treturn &sm, nil\n}", "func NewStore(logger configmap.Logger, onAfterStore ...func(name string, value interface{})) *Store {\n\ts := &Store{}\n\n\t// Append an update function to run after a ConfigMap has updated to update the\n\t// current state of the Config.\n\tonAfterStore = append(onAfterStore, func(_ string, _ interface{}) {\n\t\ts.current.Store(&Config{\n\t\t\tTracing: s.UntypedLoad(tracingconfig.ConfigName).(*tracingconfig.Config).DeepCopy(),\n\t\t})\n\t})\n\ts.UntypedStore = configmap.NewUntypedStore(\n\t\t\"activator\",\n\t\tlogger,\n\t\tconfigmap.Constructors{\n\t\t\ttracingconfig.ConfigName: tracingconfig.NewTracingConfigFromConfigMap,\n\t\t},\n\t\tonAfterStore...,\n\t)\n\treturn s\n}", "func New(appConfig string) *StateMachine {\n\treturn &StateMachine{newCache(), services.StartService(appConfig)}\n}", "func NewState(opts ...Option) *LState {\r\n\tdo := &Options{\r\n\t\tHotfixTime: DefaultHotfix,\r\n\t\tCallStackSize: DefaultCallStackSize,\r\n\t\tRegistrySize: DefaultRegistrySize,\r\n\t}\r\n\tfor _, option := range opts {\r\n\t\toption.f(do)\r\n\t}\r\n\treturn NewStateWithOpts(do)\r\n}", "func New(storage types.StateStorage) types.State {\n\treturn &stateManager{\n\t\tstorage: storage,\n\t}\n}", "func newStore(ts service.Service, config *Config) (*Store, error) {\n\tif config.Datastore == nil {\n\t\tdatastore, err := newDefaultDatastore(config.RepoPath, config.LowMem)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tconfig.Datastore = datastore\n\t}\n\tif config.EventCodec == nil {\n\t\tconfig.EventCodec = newDefaultEventCodec(config.JsonMode)\n\t}\n\tif !managedDatastore(config.Datastore) {\n\t\tif config.Debug {\n\t\t\tif err := util.SetLogLevels(map[string]logging.LogLevel{\"store\": logging.LevelDebug}); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\n\tctx, cancel := context.WithCancel(context.Background())\n\ts := &Store{\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\tdatastore: config.Datastore,\n\t\tdispatcher: newDispatcher(config.Datastore),\n\t\teventcodec: config.EventCodec,\n\t\tmodelNames: make(map[string]*Model),\n\t\tjsonMode: config.JsonMode,\n\t\tlocalEventsBus: &localEventsBus{bus: broadcast.NewBroadcaster(0)},\n\t\tstateChangedNotifee: &stateChangedNotifee{},\n\t\tservice: ts,\n\t}\n\n\tif s.jsonMode {\n\t\tif err := s.reregisterSchemas(); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\ts.dispatcher.Register(s)\n\treturn s, nil\n}", "func NewStateWithOpts(do *Options) *LState {\r\n\tgluaOpts := glua.Options{\r\n\t\tCallStackSize: do.CallStackSize,\r\n\t\tRegistrySize: do.RegistrySize,\r\n\t}\r\n\tctx, cancel := context.WithCancel(context.Background())\r\n\tL := &LState{\r\n\t\tctx: ctx,\r\n\t\tcancelFunc: cancel,\r\n\t\tgl: glua.NewState(gluaOpts),\r\n\t\topts: do,\r\n\t}\r\n\tL.openlibs()\r\n\t// hotfix\r\n\tif L.opts.NeedHotfix {\r\n\t\tL.hfMgr = newHotfixMgr(ctx, L.opts.NeedHotfixCoro, L.opts.HotfixTime)\r\n\t}\r\n\treturn L\r\n}", "func New(addr string) *State {\n\treturn &State{\n\t\tAddr: addr,\n\t\tMetrics: make(map[string]int, 10),\n\t}\n}", "func NewHelperStateManager(dashConfig config.Dash, options ...HelperStateManagerOption) *HelperStateManager {\n\thm := &HelperStateManager{\n\t\tdashConfig: dashConfig,\n\t\tpoller: NewInterruptiblePoller(\"buildInfo\"),\n\t}\n\n\thm.helperGenerateFunc = hm.generateContexts\n\n\tfor _, option := range options {\n\t\toption(hm)\n\t}\n\n\treturn hm\n}", "func New(raw []byte, id string, store storage.Store, opts ...func(*Manager)) (*Manager, error) {\n\n\tparsedConfig, err := config.ParseConfig(raw, id)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tservices, err := parseServicesConfig(parsedConfig.Services)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tm := &Manager{\n\t\tStore: store,\n\t\tConfig: parsedConfig,\n\t\tID: id,\n\t\tservices: services,\n\t}\n\n\tfor _, f := range opts {\n\t\tf(m)\n\t}\n\n\treturn m, nil\n}", "func NewManager(configStore *store.ConfigurationStore, changeStore *store.ChangeStore, deviceStore *topocache.DeviceStore,\n\tnetworkStore *store.NetworkStore, topoCh chan events.TopoEvent) (*Manager, error) {\n\tlog.Info(\"Creating Manager\")\n\tmodelReg := &modelregistry.ModelRegistry{\n\t\tModelPlugins: make(map[string]modelregistry.ModelPlugin),\n\t\tModelReadOnlyPaths: make(map[string]modelregistry.ReadOnlyPathMap),\n\t\tModelReadWritePaths: make(map[string]modelregistry.ReadWritePathMap),\n\t\tLocationStore: make(map[string]string),\n\t}\n\n\tmgr = Manager{\n\t\tConfigStore: configStore,\n\t\tChangeStore: changeStore,\n\t\tDeviceStore: deviceStore,\n\t\tNetworkStore: networkStore,\n\t\tTopoChannel: topoCh,\n\t\tModelRegistry: modelReg,\n\t\tChangesChannel: make(chan events.ConfigEvent, 10),\n\t\tOperationalStateChannel: make(chan events.OperationalStateEvent, 10),\n\t\tSouthboundErrorChan: make(chan events.DeviceResponse, 10),\n\t\tDispatcher: dispatcher.NewDispatcher(),\n\t\tOperationalStateCache: make(map[device.ID]change.TypedValueMap),\n\t}\n\n\tchangeIds := make([]string, 0)\n\t// Perform a sanity check on the change store\n\tfor changeID, changeObj := range changeStore.Store {\n\t\terr := changeObj.IsValid()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif changeID != store.B64(changeObj.ID) {\n\t\t\treturn nil, fmt.Errorf(\"ChangeID: %s must match %s\",\n\t\t\t\tchangeID, store.B64(changeObj.ID))\n\t\t}\n\t\tchangeIds = append(changeIds, changeID)\n\t}\n\n\tchangeIdsStr := strings.Join(changeIds, \",\")\n\n\tfor configID, configObj := range configStore.Store {\n\t\tfor _, chID := range configObj.Changes {\n\t\t\tif !strings.Contains(changeIdsStr, store.B64(chID)) {\n\t\t\t\treturn nil, fmt.Errorf(\n\t\t\t\t\t\"ChangeID %s from Config %s not found in change store\",\n\t\t\t\t\tstore.B64(chID), configID)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn &mgr, nil\n}", "func New(m map[string]interface{}, publisher events.Publisher) (datatx.DataTX, error) {\n\tc, err := parseConfig(m)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &manager{\n\t\tconf: c,\n\t\tpublisher: publisher,\n\t\tstatCache: cache.GetStatCache(c.Store, c.Nodes, c.Database, c.Table, time.Duration(c.TTL)*time.Second, c.Size),\n\t}, nil\n}", "func New(\n\tconfig Config,\n\tclk clock.Clock,\n\tlocalPeerID core.PeerID,\n\tnetevents networkevent.Producer,\n\tlogger *zap.SugaredLogger) *State {\n\n\tconfig = config.applyDefaults()\n\n\treturn &State{\n\t\tconfig: config,\n\t\tclk: clk,\n\t\tnetevents: netevents,\n\t\tlocalPeerID: localPeerID,\n\t\tlogger: logger,\n\t\tconns: make(map[core.InfoHash]map[core.PeerID]entry),\n\t\tblacklist: make(map[connKey]*blacklistEntry),\n\t}\n}", "func New(task metafora.Task, h StatefulHandler, ss StateStore, cl CommandListener, e ErrHandler) metafora.Handler {\n\tif e == nil {\n\t\te = DefaultErrHandler\n\t}\n\treturn &stateMachine{\n\t\ttask: task,\n\t\th: h,\n\t\tss: ss,\n\t\tcl: cl,\n\t\terrHandler: e,\n\t\tmu: &sync.RWMutex{},\n\t\tts: time.Now(),\n\t\tstopL: &sync.Mutex{},\n\t\tstopped: make(chan bool),\n\t}\n}", "func New(cfg *config.Config, ssmSvc ssmiface.SSMAPI) *Store {\n\treturn &Store{\n\t\tssmSvc: ssmSvc,\n\t\tcfg: cfg,\n\t}\n}", "func newManager(id int, wg *sync.WaitGroup, pr int64, cr float64, ps float64) *Manager {\n\tvar weather Weather\n\tweather.initializeWeather()\n\tweather.generateWeather()\n\tforecast, multiplier := weather.getWeather()\n\tfmt.Printf(\"\\nCURRENT FORECAST: %s\\n\", forecast)\n\n\tproductsRate = pr\n\tcustomerRate = cr * multiplier\n\tprocessSpeed = ps\n\n\tcustomerStatusChan = make(chan int, 256)\n\tcheckoutChangeStatusChan = make(chan int, 256)\n\n\t// Default to 1 Checkout when the store opens\n\tnumberOfCheckoutsOpen = 1\n\n\treturn &Manager{id: id, wg: wg}\n}", "func New(c *Config) *manager {\n\tctx, cancel := context.WithCancel(context.Background())\n\treturn &manager{\n\t\tlogger: c.Logger.Named(\"driver_mgr\"),\n\t\tstate: c.State,\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\tloader: c.Loader,\n\t\tpluginConfig: c.PluginConfig,\n\t\tupdater: c.Updater,\n\t\teventHandlerFactory: c.EventHandlerFactory,\n\t\tinstances: make(map[string]*instanceManager),\n\t\treattachConfigs: make(map[loader.PluginID]*pstructs.ReattachConfig),\n\t\tallowedDrivers: c.AllowedDrivers,\n\t\tblockedDrivers: c.BlockedDrivers,\n\t\treadyCh: make(chan struct{}),\n\t}\n}", "func NewState(storageInstance *storage.Storage, opts ...options.Option[State]) (state *State) {\n\treturn options.Apply(&State{\n\t\tEvents: NewEvents(),\n\t\trootBlocks: memstorage.NewSlotStorage[models.BlockID, commitment.ID](),\n\t\tlatestRootBlocks: ringbuffer.NewRingBuffer[models.BlockID](8),\n\t\tstorage: storageInstance,\n\t\tlastEvictedSlot: storageInstance.Settings.LatestCommitment().Index(),\n\t\toptsRootBlocksEvictionDelay: 3,\n\t}, opts)\n}", "func New(opt StoreOptions) *Store {\n\tstore := &Store{}\n\n\tfmt.Println(opt.toString())\n\tsession, err := mgo.Dial(opt.toString())\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tstore.cli = session\n\tstore.database = opt.Database\n\treturn store\n}", "func New(config *Config) *SyncManager {\n\tsm := SyncManager{\n\t\tpeerNotifier: config.PeerNotifier,\n\t\tchain: config.Chain,\n\t\tchainParams: config.ChainParams,\n\t\ttxMemPool: config.TxMemPool,\n\t\tblockMemPool: config.BlockMemPool,\n\t\trejectedTxns: make(map[common.Uint256]struct{}),\n\t\trequestedTxns: make(map[common.Uint256]struct{}),\n\t\trequestedBlocks: make(map[common.Uint256]struct{}),\n\t\trequestedConfirmedBlocks: make(map[common.Uint256]struct{}),\n\t\tpeerStates: make(map[*peer.Peer]*peerSyncState),\n\t\tmsgChan: make(chan interface{}, config.MaxPeers*3),\n\t\tquit: make(chan struct{}),\n\t}\n\n\tevents.Subscribe(sm.handleBlockchainEvents)\n\n\treturn &sm\n}", "func New(ctx context.Context, alias, path string, cfgdir string) (*Store, error) {\n\tpath = fsutil.CleanPath(path)\n\ts := &Store{\n\t\talias: alias,\n\t\tpath: path,\n\t\tsync: gitmock.New(),\n\t}\n\n\t// init store backend\n\tswitch backend.GetStoreBackend(ctx) {\n\tcase backend.FS:\n\t\ts.store = fs.New(path)\n\t\tout.Debug(ctx, \"Using Store Backend: fs\")\n\tcase backend.KVMock:\n\t\ts.store = kvmock.New()\n\t\tout.Debug(ctx, \"Using Store Backend: kvmock\")\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Unknown store backend\")\n\t}\n\n\t// init sync backend\n\tswitch backend.GetSyncBackend(ctx) {\n\tcase backend.GoGit:\n\t\tout.Cyan(ctx, \"WARNING: Using experimental sync backend 'go-git'\")\n\t\tgit, err := gogit.Open(path)\n\t\tif err != nil {\n\t\t\tout.Debug(ctx, \"Failed to initialize sync backend 'gogit': %s\", err)\n\t\t} else {\n\t\t\ts.sync = git\n\t\t\tout.Debug(ctx, \"Using Sync Backend: go-git\")\n\t\t}\n\tcase backend.GitCLI:\n\t\tgpgBin, _ := gpgcli.Binary(ctx, \"\")\n\t\tgit, err := gitcli.Open(path, gpgBin)\n\t\tif err != nil {\n\t\t\tout.Debug(ctx, \"Failed to initialize sync backend 'git': %s\", err)\n\t\t} else {\n\t\t\ts.sync = git\n\t\t\tout.Debug(ctx, \"Using Sync Backend: git-cli\")\n\t\t}\n\tcase backend.GitMock:\n\t\t// no-op\n\t\tout.Debug(ctx, \"Using Sync Backend: git-mock\")\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Unknown Sync Backend\")\n\t}\n\n\t// init crypto backend\n\tswitch backend.GetCryptoBackend(ctx) {\n\tcase backend.GPGCLI:\n\t\tgpg, err := gpgcli.New(ctx, gpgcli.Config{\n\t\t\tUmask: fsutil.Umask(),\n\t\t\tArgs: gpgcli.GPGOpts(),\n\t\t})\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ts.crypto = gpg\n\t\tout.Debug(ctx, \"Using Crypto Backend: gpg-cli\")\n\tcase backend.XC:\n\t\t//out.Red(ctx, \"WARNING: Using highly experimental crypto backend!\")\n\t\tcrypto, err := xc.New(cfgdir, client.New(cfgdir))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ts.crypto = crypto\n\t\tout.Debug(ctx, \"Using Crypto Backend: xc\")\n\tcase backend.GPGMock:\n\t\t//out.Red(ctx, \"WARNING: Using no-op crypto backend (NO ENCRYPTION)!\")\n\t\ts.crypto = gpgmock.New()\n\t\tout.Debug(ctx, \"Using Crypto Backend: gpg-mock\")\n\tcase backend.OpenPGP:\n\t\tcrypto, err := openpgp.New(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\ts.crypto = crypto\n\t\tout.Debug(ctx, \"Using Crypto Backend: openpgp\")\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"no valid crypto backend selected\")\n\t}\n\n\treturn s, nil\n}", "func New(opts ...Opt) (*Store, error) {\n\tbadgerOptions := getBadgerOptions(opts...)\n\n\tdb, err := badger.Open(badgerOptions)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not open database: %w\", err)\n\t}\n\n\tstore := &Store{db, newChangelog()}\n\tif err = store.setup(); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn store, nil\n}", "func New() *StateMachine {\n\treturn &StateMachine{database: make(map[string]string)}\n}", "func New(opts ...Option) (*Store, error) {\n\tvar s Store\n\n\tfor _, opt := range opts {\n\t\tif err := opt.Apply(&s); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn &s, nil\n}", "func NewStore(c Config) *Store {\n\treturn &Store{\n\t\tpath: c.Dir,\n\t\thost: c.Hostname,\n\t\taddr: c.BindAddress,\n\t\tdata: &Data{},\n\t\tHeartbeatTimeout: time.Duration(c.HeartbeatTimeout),\n\t\tElectionTimeout: time.Duration(c.ElectionTimeout),\n\t\tLeaderLeaseTimeout: time.Duration(c.LeaderLeaseTimeout),\n\t\tCommitTimeout: time.Duration(c.CommitTimeout),\n\t\tLogger: log.New(os.Stderr, \"\", log.LstdFlags),\n\t}\n}", "func NewHandler(k keeper.Keeper) sdk.Handler {\n\tdefer telemetry.MeasureSince(time.Now(), \"evm\", \"state_transition\")\n\n\treturn func(ctx sdk.Context, msg sdk.Msg) (*sdk.Result, error) {\n\t\tsnapshotStateDB := k.CommitStateDB.Copy()\n\n\t\t// The \"recover\" code here is used to solve the problem of dirty data\n\t\t// in CommitStateDB due to insufficient gas.\n\n\t\t// The following is a detailed description:\n\t\t// If the gas is insufficient during the execution of the \"handler\",\n\t\t// panic will be thrown from the function \"ConsumeGas\" and finally\n\t\t// caught by the function \"runTx\" from Cosmos. The function \"runTx\"\n\t\t// will think that the execution of Msg has failed and the modified\n\t\t// data in the Store will not take effect.\n\n\t\t// Stacktrace:runTx->runMsgs->handler->...->gaskv.Store.Set->ConsumeGas\n\n\t\t// The problem is that when the modified data in the Store does not take\n\t\t// effect, the data in the modified CommitStateDB is not rolled back,\n\t\t// they take effect, and dirty data is generated.\n\t\t// Therefore, the code here specifically deals with this situation.\n\t\t// See https://github.com/cosmos/ethermint/issues/668 for more information.\n\t\tdefer func() {\n\t\t\tif r := recover(); r != nil {\n\t\t\t\t// We first used \"k.CommitStateDB = snapshotStateDB\" to roll back\n\t\t\t\t// CommitStateDB, but this can only change the CommitStateDB in the\n\t\t\t\t// current Keeper object, but the Keeper object will be destroyed\n\t\t\t\t// soon, it is not a global variable, so the content pointed to by\n\t\t\t\t// the CommitStateDB pointer can be modified to take effect.\n\t\t\t\ttypes.CopyCommitStateDB(snapshotStateDB, k.CommitStateDB)\n\t\t\t\tpanic(r)\n\t\t\t}\n\t\t}()\n\t\tctx = ctx.WithEventManager(sdk.NewEventManager())\n\n\t\tswitch msg := msg.(type) {\n\t\tcase *types.MsgEthereumTx:\n\t\t\t// execute state transition\n\t\t\tres, err := k.EthereumTx(sdk.WrapSDKContext(ctx), msg)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tresult, err := sdk.WrapServiceResult(ctx, res, err)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\t// log state transition result\n\t\t\tvar recipientLog string\n\t\t\tif res.ContractAddress != \"\" {\n\t\t\t\trecipientLog = fmt.Sprintf(\"contract address %s\", res.ContractAddress)\n\t\t\t} else {\n\t\t\t\trecipientLog = fmt.Sprintf(\"recipient address %s\", msg.Data.Recipient)\n\t\t\t}\n\n\t\t\tsender := ethcmn.BytesToAddress(msg.GetFrom().Bytes())\n\n\t\t\tlog := fmt.Sprintf(\n\t\t\t\t\"executed EVM state transition; sender address %s; %s\", sender, recipientLog,\n\t\t\t)\n\n\t\t\tk.Logger(ctx).Info(log)\n\t\t\tresult.Log = log\n\n\t\t\treturn result, nil\n\n\t\tdefault:\n\t\t\treturn nil, sdkerrors.Wrapf(sdkerrors.ErrUnknownRequest, \"unrecognized %s message type: %T\", types.ModuleName, msg)\n\t\t}\n\t}\n}", "func NewManager(opts options.Options, cfg config.Config) (*Manager, error) {\n\tcloneOpts, err := opts.CloneOptions()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tm := &Manager{\n\t\tOpts: opts,\n\t\tConfig: cfg,\n\t\tCloneOptions: cloneOpts,\n\n\t\tstopChan: make(chan os.Signal, 1),\n\t\tleakChan: make(chan Leak),\n\t\tleakWG: &sync.WaitGroup{},\n\t\tleakCache: make(map[string]bool),\n\t\tmetaWG: &sync.WaitGroup{},\n\t\tmetadata: Metadata{\n\t\t\tRegexTime: make(map[string]int64),\n\t\t\ttimings: make(chan interface{}),\n\t\t\tdata: make(map[string]interface{}),\n\t\t\tmux: new(sync.Mutex),\n\t\t},\n\t}\n\n\tsignal.Notify(m.stopChan, os.Interrupt)\n\n\t// start receiving leaks and metadata\n\tgo m.receiveLeaks()\n\tgo m.receiveMetadata()\n\tgo m.receiveInterrupt()\n\n\treturn m, nil\n}", "func New(options Options) *Store {\n\tif options.Codec == nil {\n\t\toptions.Codec = DefaultOptions.Codec\n\t}\n\n\tif options.Interval == 0 {\n\t\toptions.Interval = DefaultOptions.Interval\n\t}\n\n\tif options.TableName == \"\" {\n\t\toptions.TableName = DefaultOptions.TableName\n\t}\n\n\tsql := newSqlSvr(options.User, options.Pwd, options.Host, options.Db, options.TableName, options.Split)\n\tif sql == nil {\n\t\treturn nil\n\t}\n\n\ts := &Store{\n\t\tSql: sql,\n\t\tCodec: options.Codec,\n\t}\n\n\t//go s.autoGC(options.Interval)\n\n\treturn s\n}", "func New(options Options) *Store {\n\t// Set default options\n\tif options.Codec == nil {\n\t\toptions.Codec = DefaultOptions.Codec\n\t}\n\n\ts := Store{\n\t\tm: make(map[string]*Item),\n\t\tlock: new(sync.RWMutex),\n\t\tcodec: options.Codec,\n\t}\n\n\tgo s.autoGC(options.Interval)\n\n\treturn &s\n}", "func NewStore(\n\tctx context.Context, cfg StoreConfig, eng storage.Engine, nodeDesc *roachpb.NodeDescriptor,\n) *Store {\n\t// TODO(tschottdorf): find better place to set these defaults.\n\tcfg.SetDefaults()\n\n\tif !cfg.Valid() {\n\t\tlog.Fatalf(ctx, \"invalid store configuration: %+v\", &cfg)\n\t}\n\ts := &Store{\n\t\tcfg: cfg,\n\t\tdb: cfg.DB, // TODO(tschottdorf): remove redundancy.\n\t\tengine: eng,\n\t\tnodeDesc: nodeDesc,\n\t\tmetrics: newStoreMetrics(cfg.HistogramWindowInterval),\n\t\tctSender: cfg.ClosedTimestampSender,\n\t}\n\tif cfg.RPCContext != nil {\n\t\ts.allocator = MakeAllocator(cfg.StorePool, cfg.RPCContext.RemoteClocks.Latency)\n\t} else {\n\t\ts.allocator = MakeAllocator(cfg.StorePool, func(string) (time.Duration, bool) {\n\t\t\treturn 0, false\n\t\t})\n\t}\n\ts.replRankings = newReplicaRankings()\n\n\ts.draining.Store(false)\n\ts.scheduler = newRaftScheduler(s.metrics, s, storeSchedulerConcurrency)\n\n\ts.raftEntryCache = raftentry.NewCache(cfg.RaftEntryCacheSize)\n\ts.metrics.registry.AddMetricStruct(s.raftEntryCache.Metrics())\n\n\ts.coalescedMu.Lock()\n\ts.coalescedMu.heartbeats = map[roachpb.StoreIdent][]RaftHeartbeat{}\n\ts.coalescedMu.heartbeatResponses = map[roachpb.StoreIdent][]RaftHeartbeat{}\n\ts.coalescedMu.Unlock()\n\n\ts.mu.Lock()\n\ts.mu.replicaPlaceholders = map[roachpb.RangeID]*ReplicaPlaceholder{}\n\ts.mu.replicasByKey = newStoreReplicaBTree()\n\ts.mu.uninitReplicas = map[roachpb.RangeID]*Replica{}\n\ts.mu.Unlock()\n\n\ts.unquiescedReplicas.Lock()\n\ts.unquiescedReplicas.m = map[roachpb.RangeID]struct{}{}\n\ts.unquiescedReplicas.Unlock()\n\n\ts.rangefeedReplicas.Lock()\n\ts.rangefeedReplicas.m = map[roachpb.RangeID]struct{}{}\n\ts.rangefeedReplicas.Unlock()\n\n\ts.tsCache = tscache.New(cfg.Clock)\n\ts.metrics.registry.AddMetricStruct(s.tsCache.Metrics())\n\n\ts.txnWaitMetrics = txnwait.NewMetrics(cfg.HistogramWindowInterval)\n\ts.metrics.registry.AddMetricStruct(s.txnWaitMetrics)\n\ts.snapshotApplySem = make(chan struct{}, cfg.concurrentSnapshotApplyLimit)\n\n\ts.renewableLeasesSignal = make(chan struct{})\n\n\ts.limiters.BulkIOWriteRate = rate.NewLimiter(rate.Limit(bulkIOWriteLimit.Get(&cfg.Settings.SV)), bulkIOWriteBurst)\n\tbulkIOWriteLimit.SetOnChange(&cfg.Settings.SV, func() {\n\t\ts.limiters.BulkIOWriteRate.SetLimit(rate.Limit(bulkIOWriteLimit.Get(&cfg.Settings.SV)))\n\t})\n\ts.limiters.ConcurrentImportRequests = limit.MakeConcurrentRequestLimiter(\n\t\t\"importRequestLimiter\", int(importRequestsLimit.Get(&cfg.Settings.SV)),\n\t)\n\timportRequestsLimit.SetOnChange(&cfg.Settings.SV, func() {\n\t\ts.limiters.ConcurrentImportRequests.SetLimit(int(importRequestsLimit.Get(&cfg.Settings.SV)))\n\t})\n\ts.limiters.ConcurrentExportRequests = limit.MakeConcurrentRequestLimiter(\n\t\t\"exportRequestLimiter\", int(ExportRequestsLimit.Get(&cfg.Settings.SV)),\n\t)\n\n\t// The snapshot storage is usually empty at this point since it is cleared\n\t// after each snapshot application, except when the node crashed right before\n\t// it can clean it up. If this fails it's not a correctness issue since the\n\t// storage is also cleared before receiving a snapshot.\n\ts.sstSnapshotStorage = NewSSTSnapshotStorage(s.engine, s.limiters.BulkIOWriteRate)\n\tif err := s.sstSnapshotStorage.Clear(); err != nil {\n\t\tlog.Warningf(ctx, \"failed to clear snapshot storage: %v\", err)\n\t}\n\ts.protectedtsCache = cfg.ProtectedTimestampCache\n\n\t// On low-CPU instances, a default limit value may still allow ExportRequests\n\t// to tie up all cores so cap limiter at cores-1 when setting value is higher.\n\texportCores := runtime.GOMAXPROCS(0) - 1\n\tif exportCores < 1 {\n\t\texportCores = 1\n\t}\n\tExportRequestsLimit.SetOnChange(&cfg.Settings.SV, func() {\n\t\tlimit := int(ExportRequestsLimit.Get(&cfg.Settings.SV))\n\t\tif limit > exportCores {\n\t\t\tlimit = exportCores\n\t\t}\n\t\ts.limiters.ConcurrentExportRequests.SetLimit(limit)\n\t})\n\ts.limiters.ConcurrentAddSSTableRequests = limit.MakeConcurrentRequestLimiter(\n\t\t\"addSSTableRequestLimiter\", int(addSSTableRequestLimit.Get(&cfg.Settings.SV)),\n\t)\n\taddSSTableRequestLimit.SetOnChange(&cfg.Settings.SV, func() {\n\t\ts.limiters.ConcurrentAddSSTableRequests.SetLimit(int(addSSTableRequestLimit.Get(&cfg.Settings.SV)))\n\t})\n\ts.limiters.ConcurrentRangefeedIters = limit.MakeConcurrentRequestLimiter(\n\t\t\"rangefeedIterLimiter\", int(concurrentRangefeedItersLimit.Get(&cfg.Settings.SV)),\n\t)\n\tconcurrentRangefeedItersLimit.SetOnChange(&cfg.Settings.SV, func() {\n\t\ts.limiters.ConcurrentRangefeedIters.SetLimit(\n\t\t\tint(concurrentRangefeedItersLimit.Get(&cfg.Settings.SV)))\n\t})\n\n\ts.tenantRateLimiters = tenantrate.NewLimiterFactory(cfg.Settings, &cfg.TestingKnobs.TenantRateKnobs)\n\ts.metrics.registry.AddMetricStruct(s.tenantRateLimiters.Metrics())\n\n\ts.systemConfigUpdateQueueRateLimiter = quotapool.NewRateLimiter(\n\t\t\"SystemConfigUpdateQueue\",\n\t\tquotapool.Limit(queueAdditionOnSystemConfigUpdateRate.Get(&cfg.Settings.SV)),\n\t\tqueueAdditionOnSystemConfigUpdateBurst.Get(&cfg.Settings.SV))\n\tupdateSystemConfigUpdateQueueLimits := func() {\n\t\ts.systemConfigUpdateQueueRateLimiter.UpdateLimit(\n\t\t\tquotapool.Limit(queueAdditionOnSystemConfigUpdateRate.Get(&cfg.Settings.SV)),\n\t\t\tqueueAdditionOnSystemConfigUpdateBurst.Get(&cfg.Settings.SV))\n\t}\n\tqueueAdditionOnSystemConfigUpdateRate.SetOnChange(&cfg.Settings.SV,\n\t\tupdateSystemConfigUpdateQueueLimits)\n\tqueueAdditionOnSystemConfigUpdateBurst.SetOnChange(&cfg.Settings.SV,\n\t\tupdateSystemConfigUpdateQueueLimits)\n\n\tif s.cfg.Gossip != nil {\n\t\t// Add range scanner and configure with queues.\n\t\ts.scanner = newReplicaScanner(\n\t\t\ts.cfg.AmbientCtx, s.cfg.Clock, cfg.ScanInterval,\n\t\t\tcfg.ScanMinIdleTime, cfg.ScanMaxIdleTime, newStoreReplicaVisitor(s),\n\t\t)\n\t\ts.gcQueue = newGCQueue(s, s.cfg.Gossip)\n\t\ts.mergeQueue = newMergeQueue(s, s.db, s.cfg.Gossip)\n\t\ts.splitQueue = newSplitQueue(s, s.db, s.cfg.Gossip)\n\t\ts.replicateQueue = newReplicateQueue(s, s.cfg.Gossip, s.allocator)\n\t\ts.replicaGCQueue = newReplicaGCQueue(s, s.db, s.cfg.Gossip)\n\t\ts.raftLogQueue = newRaftLogQueue(s, s.db, s.cfg.Gossip)\n\t\ts.raftSnapshotQueue = newRaftSnapshotQueue(s, s.cfg.Gossip)\n\t\ts.consistencyQueue = newConsistencyQueue(s, s.cfg.Gossip)\n\t\t// NOTE: If more queue types are added, please also add them to the list of\n\t\t// queues on the EnqueueRange debug page as defined in\n\t\t// pkg/ui/src/views/reports/containers/enqueueRange/index.tsx\n\t\ts.scanner.AddQueues(\n\t\t\ts.gcQueue, s.mergeQueue, s.splitQueue, s.replicateQueue, s.replicaGCQueue,\n\t\t\ts.raftLogQueue, s.raftSnapshotQueue, s.consistencyQueue)\n\t\ttsDS := s.cfg.TimeSeriesDataStore\n\t\tif s.cfg.TestingKnobs.TimeSeriesDataStore != nil {\n\t\t\ttsDS = s.cfg.TestingKnobs.TimeSeriesDataStore\n\t\t}\n\t\tif tsDS != nil {\n\t\t\ts.tsMaintenanceQueue = newTimeSeriesMaintenanceQueue(\n\t\t\t\ts, s.db, s.cfg.Gossip, tsDS,\n\t\t\t)\n\t\t\ts.scanner.AddQueues(s.tsMaintenanceQueue)\n\t\t}\n\t}\n\n\tif cfg.TestingKnobs.DisableGCQueue {\n\t\ts.setGCQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableMergeQueue {\n\t\ts.setMergeQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableRaftLogQueue {\n\t\ts.setRaftLogQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableReplicaGCQueue {\n\t\ts.setReplicaGCQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableReplicateQueue {\n\t\ts.SetReplicateQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableSplitQueue {\n\t\ts.setSplitQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableTimeSeriesMaintenanceQueue {\n\t\ts.setTimeSeriesMaintenanceQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableRaftSnapshotQueue {\n\t\ts.setRaftSnapshotQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableConsistencyQueue {\n\t\ts.setConsistencyQueueActive(false)\n\t}\n\tif cfg.TestingKnobs.DisableScanner {\n\t\ts.setScannerActive(false)\n\t}\n\n\treturn s\n}", "func New(ctx context.Context, cfg *config.Config, logger *logrus.Logger) (*Store, error) {\n\tvar store Store\n\n\tstore.config = cfg\n\tstore.logger = logger\n\n\t// connect to postgres\n\tpgConn, err := postgres.NewConnect(cfg.Postgres)\n\tif err != nil {\n\t\treturn &store, errpath.Err(err)\n\t}\n\tstore.Pg = pgConn\n\n\tif pgConn != nil {\n\t\tgo store.keepAlivePg()\n\t}\n\n\treturn &store, nil\n}", "func newStateSetter() *stateSetter {\n\tstateSetter := new(stateSetter)\n\tstateSetter.predefinedStrategy = v1alpha1.MemcachedSpec{Interval: 30, Threshold: 20, Size: 2}\n\treturn stateSetter\n}", "func New(nodeID, clusterID, advertiseAddr, storeType, runnerAddr, configFile string, adminMan *admin.Manager) (*Manager, error) {\n\n\t// Create a new manager instance\n\tm := &Manager{nodeID: nodeID, clusterID: clusterID, advertiseAddr: advertiseAddr, storeType: storeType, runnerAddr: runnerAddr, configFile: configFile, adminMan: adminMan}\n\n\t// Initialise the consul client if enabled\n\tswitch storeType {\n\tcase \"none\":\n\t\tm.services = []*service{{id: nodeID, addr: advertiseAddr}}\n\t\treturn m, nil\n\tcase \"kube\":\n\t\ts, err := NewKubeStore(clusterID)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tm.store = s\n\t\tm.store.Register()\n\tcase \"consul\":\n\t\ts, err := NewConsulStore(nodeID, clusterID, advertiseAddr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tm.store = s\n\t\tm.store.Register()\n\tcase \"etcd\":\n\t\ts, err := NewETCDStore(nodeID, clusterID, advertiseAddr)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tm.store = s\n\t\tm.store.Register()\n\t}\n\n\treturn m, nil\n}", "func New(path string, config *Config) (*Store, error) {\n\t// open the store\n\tf, err := os.OpenFile(path, os.O_RDWR|os.O_CREATE, 0755)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar dS = &Store{\n\t\tConfig: config,\n\t\tf: f,\n\t\trecords: map[string]interface{}{},\n\t}\n\n\tswitch config.StoreType {\n\tcase JSONStore:\n\t\tdS.Storer = JSONDataStore{}\n\tcase YAMLStore:\n\t\tdS.Storer = YAMLDataStore{}\n\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"unsupported store type: %v\", config.StoreType)\n\t}\n\n\t// read records into store\n\tif err := dS.Open(f, dS.records); err != nil {\n\t\treturn nil, err\n\t}\n\n\t// configure name\n\tif config.StoreName == \"\" {\n\t\tconfig.StoreName = dS.f.Name()\n\t}\n\n\treturn dS, nil\n}", "func NewState(t *testing.T) (*Mocker, *state.State) {\n\tm, se := NewSession(t)\n\n\tst, err := state.NewFromSession(se, new(state.NoopStore))\n\tif err != nil {\n\t\tpanic(err) // this should never happen\n\t}\n\n\treturn m, st\n}", "func (s *MemStateStore) NewState(url string) (string, error) {\n\tstate := s.valueGenerator.String()\n\terr := s.Add(state, url)\n\treturn state, err\n}", "func New(tracker *bestblock.Tracker, lag int64, config Config) (*Manager, error) {\n\tm := &Manager{\n\t\tconfig: config,\n\t\tlag: lag,\n\t\ttracker: tracker,\n\t\tclosed: false,\n\t\tstopChan: make(chan bool),\n\t}\n\n\tlog.Info(\"setting up redis connection\")\n\tm.redis = redis.NewClient(&redis.Options{\n\t\tAddr: config.RedisServer,\n\t\tPassword: config.RedisPassword,\n\t\tDB: 0,\n\t\tReadTimeout: time.Second * 1,\n\t})\n\n\terr := m.redis.Ping().Err()\n\tif err != nil {\n\t\tlog.Error(err)\n\t\treturn nil, err\n\t}\n\n\tlog.Info(\"connected to redis successfully\")\n\n\tgo m.watchNewBlocks()\n\n\treturn m, nil\n}", "func newStore(c *Config, httpAddr, raftAddr string) *store {\n\tinternalData := meta.Data{\n\t\tIndex: 1,\n\t}\n\ts := store{\n\t\tdata: &Data{\n\t\t\tData: internalData,\n\t\t},\n\t\tclosing: make(chan struct{}),\n\t\tdataChanged: make(chan struct{}),\n\t\tpath: c.Dir,\n\t\tconfig: c,\n\t\thttpAddr: httpAddr,\n\t\traftAddr: raftAddr,\n\t\tlogger: zap.New(zap.NullEncoder()),\n\t}\n\n\treturn &s\n}", "func newStore(c *Config) (*Store, error) {\n\tif c == nil {\n\t\tc = defaultConfig()\n\t}\n\tmutex := &sync.RWMutex{}\n\tstore := new(Store)\n\tstartTime := time.Now().UTC()\n\tfileWatcher, err := newWatcher(\".\")\n\tif err != nil {\n\t\tlog.Info(fmt.Sprintf(\"unable to init file watcher: %v\", err))\n\t}\n\tif c.Monitoring {\n\t\tmonitoring.Init()\n\t}\n\tstore.fileWatcher = fileWatcher\n\tstore.store = makeStorage(\"\")\n\tstore.keys = []string{}\n\tstore.compression = c.Compression\n\tstore.dbs = make(map[string]*DB)\n\tstore.lock = mutex\n\tstore.stat = new(stats.Statistics)\n\tstore.stat.Start = startTime\n\tstore.indexes = make(map[string]*index)\n\tc.setMissedValues()\n\tstore.config = c\n\tif c.LoadPath != \"\" {\n\t\terrLoad := loadData(store, c.LoadPath)\n\t\tif errLoad != nil {\n\t\t\treturn nil, fmt.Errorf(\"unable to load data: %v\", errLoad)\n\t\t}\n\t}\n\tstore.writer, err = newWriter(c.LoadPath)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to create writer: %v\", err)\n\t}\n\treturn store, nil\n}", "func New(cfg *config.Config) (*Deduplicator, error) {\n\tif err := os.MkdirAll(cfg.DedupDir, 0700); err != nil {\n\t\treturn nil, fmt.Errorf(\"create state dir: %w\", err)\n\t}\n\tdedup := &Deduplicator{config: *cfg}\n\tdata, err := ioutil.ReadFile(dedup.stateFileName())\n\tif err != nil {\n\t\tif !os.IsNotExist(err) {\n\t\t\treturn nil, fmt.Errorf(\"read state file: %w\", err)\n\t\t}\n\t} else if err := json.Unmarshal(data, &dedup.state); err != nil {\n\t\treturn nil, fmt.Errorf(\"parse state: %w\", err)\n\t}\n\treturn dedup, nil\n}", "func New(cfg *config.Config, rootPath string) (*Manager, error) {\n\tself := peer.NewSelf(cfg, rootPath)\n\tm := &Manager{\n\t\tself: self,\n\t}\n\tp := self.ToPeer()\n\trecent, err := p.ReadRecent()\n\tif log.If(err) {\n\t\tlog.Println(err)\n\t}\n\tm.recent, err = thread.NewList(self, recent)\n\tm.peers, err = p.ReadPeers()\n\tlog.If(err)\n\tif len(m.peers) == 0 {\n\t\tm.peers = self.LoadInitPeers()\n\t}\n\tm.tags, err = p.ReadTags()\n\tlog.If(err)\n\tm.spams, err = p.ReadSpams()\n\tlog.If(err)\n\tm.readThreads = self.ReadThreads()\n\tif err := self.StartListen(m.accept); log.If(err) {\n\t\treturn nil, err\n\t}\n\tgo func() {\n\t\tfor {\n\t\t\tm.cron()\n\t\t\ttime.Sleep(time.Hour)\n\t\t}\n\t}()\n\treturn m, nil\n}", "func newResultState(sharedConfig jsonio.GoldResults, config *GoldClientConfig) *resultState {\n\tgoldURL := config.OverrideGoldURL\n\tif goldURL == \"\" {\n\t\tgoldURL = getGoldInstanceURL(config.InstanceID)\n\t}\n\tbucket := config.OverrideBucket\n\tif bucket == \"\" {\n\t\tbucket = getBucket(config.InstanceID)\n\t}\n\n\tret := &resultState{\n\t\tSharedConfig: sharedConfig,\n\t\tPerTestPassFail: config.PassFailStep,\n\t\tFailureFile: config.FailureFile,\n\t\tInstanceID: config.InstanceID,\n\t\tUploadOnly: config.UploadOnly,\n\t\tGoldURL: goldURL,\n\t\tBucket: bucket,\n\t}\n\n\treturn ret\n}", "func NewStateMachineWrapper(clusterID uint64, nodeID uint64,\n\tfactory unsafe.Pointer, style uint64, smType pb.StateMachineType,\n\tdone <-chan struct{}) rsm.IManagedStateMachine {\n\tcClusterID := C.uint64_t(clusterID)\n\tcNodeID := C.uint64_t(nodeID)\n\tcFactory := factory\n\tcStyle := C.uint64_t(style)\n\tswitch smType {\n\tcase pb.RegularStateMachine:\n\t\treturn &RegularStateMachineWrapper{\n\t\t\tdataStore: C.CreateDBRegularStateMachine(cClusterID, cNodeID, cFactory, cStyle),\n\t\t\tdone: done,\n\t\t}\n\tcase pb.ConcurrentStateMachine:\n\t\treturn &ConcurrentStateMachineWrapper{\n\t\t\tdataStore: C.CreateDBConcurrentStateMachine(cClusterID, cNodeID, cFactory, cStyle),\n\t\t\tdone: done,\n\t\t}\n\tcase pb.OnDiskStateMachine:\n\t\treturn &OnDiskStateMachineWrapper{\n\t\t\tdataStore: C.CreateDBOnDiskStateMachine(cClusterID, cNodeID, cFactory, cStyle),\n\t\t\tdone: done,\n\t\t}\n\tdefault:\n\t\tpanic(\"unknown statemachine type\")\n\t}\n}", "func NewManager(ctx context.Context, config *types.Config) (*Manager, error) {\n\tm := &Manager{config: config}\n\n\tswitch config.Store {\n\tcase common.GRPCStore:\n\t\tcorestore.Init(ctx, config)\n\t\tstore := corestore.Get()\n\t\tif store == nil {\n\t\t\treturn nil, common.ErrGetStoreFailed\n\t\t}\n\t\tm.store = store\n\tcase common.MocksStore:\n\t\tm.store = storemocks.NewFakeStore()\n\tdefault:\n\t\treturn nil, common.ErrInvalidStoreType\n\t}\n\n\tnode, err := m.store.GetNode(ctx, config.HostName)\n\tif err != nil {\n\t\tlog.WithFunc(\"NewManager\").Errorf(ctx, err, \"failed to get node %s\", config.HostName)\n\t\treturn nil, err\n\t}\n\n\tnodeIP := utils.GetIP(node.Endpoint)\n\tif nodeIP == \"\" {\n\t\tnodeIP = common.LocalIP\n\t}\n\n\tswitch config.Runtime {\n\tcase common.DockerRuntime:\n\t\tdocker.InitClient(config, nodeIP)\n\t\tm.runtimeClient = docker.GetClient()\n\t\tif m.runtimeClient == nil {\n\t\t\treturn nil, common.ErrGetRuntimeFailed\n\t\t}\n\tcase common.YavirtRuntime:\n\t\tyavirt.InitClient(config)\n\t\tm.runtimeClient = yavirt.GetClient()\n\t\tif m.runtimeClient == nil {\n\t\t\treturn nil, common.ErrGetRuntimeFailed\n\t\t}\n\tcase common.MocksRuntime:\n\t\tm.runtimeClient = runtimemocks.FromTemplate()\n\tdefault:\n\t\treturn nil, common.ErrInvalidRuntimeType\n\t}\n\n\tm.logBroadcaster = newLogBroadcaster()\n\tm.forwards = utils.NewHashBackends(config.Log.Forwards)\n\tm.storeIdentifier = m.store.GetIdentifier(ctx)\n\tm.nodeIP = nodeIP\n\tm.checkWorkloadMutex = &sync.Mutex{}\n\tm.startingWorkloads = haxmap.New[string, *utils.RetryTask]()\n\n\treturn m, nil\n}", "func NewState(e *etcd.Client, ttl time.Duration, path ...string) State {\n\trealttl := 1 * time.Second\n\tif ttl.Seconds() > realttl.Seconds() {\n\t\trealttl = ttl\n\t}\n\treturn &state{\n\t\te: e,\n\t\tkey: strings.Join(path, \"/\"),\n\t\tttl: uint64(realttl.Seconds()),\n\t}\n}", "func New() backend.Backend {\n\treturn &remotestate.Backend{\n\t\tConfigureFunc: configure,\n\n\t\t// Set the schema\n\t\tBackend: &schema.Backend{\n\t\t\tSchema: map[string]*schema.Schema{\n\t\t\t\t\"lock_id\": &schema.Schema{\n\t\t\t\t\tType: schema.TypeString,\n\t\t\t\t\tOptional: true,\n\t\t\t\t\tDescription: \"initializes the state in a locked configuration\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func New(fns ...TxnFn) *Manager {\n\tvar m Manager\n\tm.fns = fns\n\treturn &m\n}", "func New(c *Config) *manager {\n\tctx, cancel := context.WithCancel(context.Background())\n\treturn &manager{\n\t\tlogger: c.Logger.Named(\"device_mgr\"),\n\t\tstate: c.State,\n\t\tctx: ctx,\n\t\tcancel: cancel,\n\t\tloader: c.Loader,\n\t\tpluginConfig: c.PluginConfig,\n\t\tupdater: c.Updater,\n\t\tstatsInterval: c.StatsInterval,\n\t\tinstances: make(map[loader.PluginID]*instanceManager),\n\t\treattachConfigs: make(map[loader.PluginID]*pstructs.ReattachConfig),\n\t\tfingerprintResCh: make(chan struct{}, 1),\n\t}\n}", "func New() (*State, error) {\n\tplayerDB, err := db.Connect()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to connect to player db\")\n\t}\n\n\n\tec, err := engine_client.New()\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"failed to create engine client\")\n\t}\n\n\tpeerAddresses := strings.Split(*peers, \",\")\n\tif len(peerAddresses) == 0 {\n\t\treturn nil, errors.New(\"at least one peer is required\")\n\t}\n\n\tvar peerClients []player.Client\n\tfor _, p := range peerAddresses {\n\t\tc, err := player_client.New(player_client.WithAddress(p))\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"failed to create player client\")\n\t\t}\n\t\tpeerClients = append(peerClients, c)\n\t}\n\n\treturn &State{\n\t\tplayerDB: playerDB,\n\t\tengineClient: ec,\n\t\tpeers: peerClients,\n\t}, nil\n}", "func NewStore(lg *zap.Logger, b backend.Backend, le lease.Lessor, cfg StoreConfig) *store {\n\tif lg == nil {\n\t\tlg = zap.NewNop()\n\t}\n\tif cfg.CompactionBatchLimit == 0 {\n\t\tcfg.CompactionBatchLimit = defaultCompactBatchLimit\n\t}\n\ts := &store{\n\t\tcfg: cfg,\n\t\tb: b,\n\t\tkvindex: newTreeIndex(lg),\n\n\t\tle: le,\n\n\t\tcurrentRev: 1,\n\t\tcompactMainRev: -1,\n\n\t\tfifoSched: schedule.NewFIFOScheduler(),\n\n\t\tstopc: make(chan struct{}),\n\n\t\tlg: lg,\n\t}\n\ts.hashes = newHashStorage(lg, s)\n\ts.ReadView = &readView{s}\n\ts.WriteView = &writeView{s}\n\tif s.le != nil {\n\t\ts.le.SetRangeDeleter(func() lease.TxnDelete { return s.Write(traceutil.TODO()) })\n\t}\n\n\ttx := s.b.BatchTx()\n\ttx.LockOutsideApply()\n\ttx.UnsafeCreateBucket(buckets.Key)\n\ttx.UnsafeCreateBucket(buckets.Meta)\n\ttx.Unlock()\n\ts.b.ForceCommit()\n\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tif err := s.restore(); err != nil {\n\t\t// TODO: return the error instead of panic here?\n\t\tpanic(\"failed to recover store from backend\")\n\t}\n\n\treturn s\n}", "func NewContext(opts *ContextOpts) (*Context, error) {\n\t// Validate the version requirement if it is given\n\tif opts.Module != nil {\n\t\tif err := CheckRequiredVersion(opts.Module); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Copy all the hooks and add our stop hook. We don't append directly\n\t// to the Config so that we're not modifying that in-place.\n\tsh := new(stopHook)\n\thooks := make([]Hook, len(opts.Hooks)+1)\n\tcopy(hooks, opts.Hooks)\n\thooks[len(opts.Hooks)] = sh\n\n\tstate := opts.State\n\tif state == nil {\n\t\tstate = new(State)\n\t\tstate.init()\n\t}\n\n\t// If our state is from the future, then error. Callers can avoid\n\t// this error by explicitly setting `StateFutureAllowed`.\n\tif !opts.StateFutureAllowed && state.FromFutureTerraform() {\n\t\treturn nil, fmt.Errorf(\n\t\t\t\"Terraform doesn't allow running any operations against a state\\n\"+\n\t\t\t\t\"that was written by a future Terraform version. The state is\\n\"+\n\t\t\t\t\"reporting it is written by Terraform '%s'.\\n\\n\"+\n\t\t\t\t\"Please run at least that version of Terraform to continue.\",\n\t\t\tstate.TFVersion)\n\t}\n\n\t// Explicitly reset our state version to our current version so that\n\t// any operations we do will write out that our latest version\n\t// has run.\n\tstate.TFVersion = version.Version\n\n\t// Determine parallelism, default to 10. We do this both to limit\n\t// CPU pressure but also to have an extra guard against rate throttling\n\t// from providers.\n\tpar := opts.Parallelism\n\tif par == 0 {\n\t\tpar = 10\n\t}\n\n\t// Set up the variables in the following sequence:\n\t// 0 - Take default values from the configuration\n\t// 1 - Take values from TF_VAR_x environment variables\n\t// 2 - Take values specified in -var flags, overriding values\n\t// set by environment variables if necessary. This includes\n\t// values taken from -var-file in addition.\n\tvariables := make(map[string]interface{})\n\tif opts.Module != nil {\n\t\tvar err error\n\t\tvariables, err = Variables(opts.Module, opts.Variables)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Bind available provider plugins to the constraints in config\n\tvar providers map[string]ResourceProviderFactory\n\tif opts.ProviderResolver != nil {\n\t\tvar err error\n\t\tdeps := ModuleTreeDependencies(opts.Module, state)\n\t\treqd := deps.AllPluginRequirements()\n\t\tif opts.ProviderSHA256s != nil && !opts.SkipProviderVerify {\n\t\t\treqd.LockExecutables(opts.ProviderSHA256s)\n\t\t}\n\t\tproviders, err = resourceProviderFactories(opts.ProviderResolver, reqd)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t} else {\n\t\tproviders = make(map[string]ResourceProviderFactory)\n\t}\n\n\tdiff := opts.Diff\n\tif diff == nil {\n\t\tdiff = &Diff{}\n\t}\n\n\treturn &Context{\n\t\tcomponents: &basicComponentFactory{\n\t\t\tproviders: providers,\n\t\t\tprovisioners: opts.Provisioners,\n\t\t},\n\t\tdestroy: opts.Destroy,\n\t\tdiff: diff,\n\t\thooks: hooks,\n\t\tmeta: opts.Meta,\n\t\tmodule: opts.Module,\n\t\tshadow: opts.Shadow,\n\t\tstate: state,\n\t\ttargets: opts.Targets,\n\t\tuiInput: opts.UIInput,\n\t\tvariables: variables,\n\n\t\tparallelSem: NewSemaphore(par),\n\t\tproviderInputConfig: make(map[string]map[string]interface{}),\n\t\tproviderSHA256s: opts.ProviderSHA256s,\n\t\tsh: sh,\n\t}, nil\n}", "func New(cfg Config) (*Bolt, error) {\n\tif err := cfg.CheckAndSetDefaults(); err != nil {\n\t\treturn nil, trace.Wrap(err)\n\t}\n\tb := &Bolt{\n\t\tConfig: cfg,\n\t}\n\n\toptions := &bolt.Options{Timeout: cfg.OpenTimeout, ReadOnly: false}\n\tdb, err := bolt.Open(cfg.Path, cfg.FileMask, options)\n\tif err != nil {\n\t\tif err == bolt.ErrTimeout {\n\t\t\treturn nil, trace.ConnectionProblem(\n\t\t\t\tnil, \"local storage at %v is locked. Another instance is running?\", cfg.Path)\n\t\t}\n\t\treturn nil, trace.Wrap(err)\n\t}\n\tb.db = db\n\tif !cfg.Readonly {\n\t\tlog.Infof(\"BOLT: locked %v\", b.Path)\n\t}\n\treturn b, nil\n}", "func NewStore(c *Config) *Store {\n\t// create a new store\n\tstore := Store{}\n\tstore.config = c\n\tstore.DBname = c.DBname\n\t// check if the file exists\n\tvar build bool\n\t_, err := os.Stat(c.DBname)\n\tif err != nil {\n\t\tlogger.Critical(\"error on stat , %s\", err)\n\t\tbuild = true\n\t}\n\t// if it is a new file build some tables\n\tif build {\n\t\tstore.Build(c)\n\t}\n\tstore.leases = Load(c.DBname)\n\treturn &store\n}", "func NewStore(config Config) *Store {\n\terr := os.MkdirAll(config.RootDir, os.FileMode(0700))\n\tif err != nil {\n\t\tlogger.Fatal(err)\n\t}\n\torm := models.NewORM(config.RootDir)\n\tethrpc, err := rpc.Dial(config.EthereumURL)\n\tif err != nil {\n\t\tlogger.Fatal(err)\n\t}\n\tkeyStore := NewKeyStore(config.KeysDir())\n\n\tht, err := NewHeadTracker(orm)\n\tif err != nil {\n\t\tlogger.Fatal(err)\n\t}\n\n\tstore := &Store{\n\t\tORM: orm,\n\t\tConfig: config,\n\t\tKeyStore: keyStore,\n\t\tExiter: os.Exit,\n\t\tClock: Clock{},\n\t\tHeadTracker: ht,\n\t\tTxManager: &TxManager{\n\t\t\tConfig: config,\n\t\t\tEthClient: &EthClient{ethrpc},\n\t\t\tKeyStore: keyStore,\n\t\t\tORM: orm,\n\t\t},\n\t}\n\treturn store\n}", "func (c *Consensus) CreateStateMachine(db database.DB, consensusTimeOut time.Duration, verifyFn consensus.CandidateVerificationFunc, executeFn consensus.ExecuteTxsFunc) (consensus.Phase, consensus.Controller, error) {\n\treturn CreateStateMachine(c.Emitter, db, consensusTimeOut, verifyFn, executeFn, c.Requestor)\n}", "func NewStateMachine(options ...SetupFunc) *State52 {\n\tsm := &State52{}\n\n\t// Apply passed options.\n\tfor _, option := range options {\n\t\tif err := option(sm); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\t// Always build states\n\tsm.states = mapStates(sm.events)\n\n\tsm.validate()\n\treturn sm\n}", "func New(s interface{}, options ...Option) *Agent {\n\tvar opts agentOptions\n\tfor _, option := range options {\n\t\toption(&opts)\n\t}\n\n\tvar atomOptions []atom.Option\n\tif opts.equalityFn != nil {\n\t\tatomOptions = append(atomOptions,\n\t\t\tatom.EqualityFunc(opts.equalityFn))\n\t}\n\n\treturn &Agent{\n\t\tstate: atom.New(s, atomOptions...),\n\t\tqueue: jobq.New(runAction),\n\t\topts: opts,\n\t}\n}", "func NewStore(ts service.Service, opts ...Option) (*Store, error) {\n\tconfig := &Config{}\n\tfor _, opt := range opts {\n\t\tif err := opt(config); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn newStore(ts, config)\n}", "func New(ctx context.Context, next http.Handler, config *Config, name string) (http.Handler, error) {\n\n\tif len(config.Name) == 0 {\n\t\treturn nil, fmt.Errorf(\"name cannot be null\")\n\t}\n\n\trequest, err := buildRequest(config.ServiceUrl, config.Name, config.Timeout, \"\", \"\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error while building request\")\n\t}\n\n\treturn &Manager{\n\t\tnext: next,\n\t\tname: config.Name,\n\t\trequest: request,\n\t\tserviceUrl: config.ServiceUrl,\n\t\ttimeout: config.Timeout,\n\t}, nil\n}", "func New(ctx context.Context, logger logger.Logger) (storage.Storage, error) {\n\tm := &memstore{\n\t\tteams: make(map[string]storage.Team),\n\t\tusersByTeam: make(map[string][]string),\n\t\tlogger: logger,\n\t}\n\n\tif err := m.EnsureAdminsTeam(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn m, nil\n}", "func CreateStateStore(\n\tstateStoreKind statestore.Kind,\n\tconfiguration *statestore.Config) (statestore.StateStore, error) {\n\n\tswitch stateStoreKind {\n\tcase statestore.KindFile:\n\t\treturn file.NewFileStore(configuration.Logger, configuration.BaseDir, configuration.StateFileUpdateInterval), nil\n\tcase statestore.KindInMemory:\n\t\treturn inmemory.NewInMemoryStore(configuration.Logger), nil\n\tdefault:\n\t\treturn nil, errors.New(\"Unknown state store type\")\n\t}\n\n}", "func (m *LocalManager) New(ctx context.Context, id string) (linker.Storage, error) {\n\tdb, err := NewLocalStorage(ctx, fmt.Sprintf(\"%s/db-%s\", m.path, id))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func stateAndStore(\n\tconfig *cfg.Config,\n\tpubKey crypto.PubKey,\n\tappVersion uint64,\n) (dbm.DB, sm.State, *mockBlockStore) {\n\tstateDB := dbm.NewMemDB()\n\tstateStore := sm.NewStore(stateDB, sm.StoreOptions{\n\t\tDiscardABCIResponses: false,\n\t})\n\tstate, _ := sm.MakeGenesisStateFromFile(config.GenesisFile())\n\tstate.Version.Consensus.App = appVersion\n\tstore := newMockBlockStore(config, state.ConsensusParams)\n\tif err := stateStore.Save(state); err != nil {\n\t\tpanic(err)\n\t}\n\treturn stateDB, state, store\n}", "func New(cfg *types.Store, sub []byte) queue.Module {\r\n\tbs := drivers.NewBaseStore(cfg)\r\n\tvar subcfg subConfig\r\n\tif sub != nil {\r\n\t\ttypes.MustDecode(sub, &subcfg)\r\n\t}\r\n\tmavls := &Store{bs, &sync.Map{}, subcfg.EnableMavlPrefix, subcfg.EnableMVCC,\r\n\t\tsubcfg.EnableMavlPrune, subcfg.PruneHeight, subcfg.EnableMemTree, subcfg.EnableMemVal}\r\n\tmavls.enableMavlPrefix = subcfg.EnableMavlPrefix\r\n\tmavls.enableMVCC = subcfg.EnableMVCC\r\n\tmavls.enableMavlPrune = subcfg.EnableMavlPrune\r\n\tmavls.pruneHeight = subcfg.PruneHeight\r\n\tmavls.enableMemTree = subcfg.EnableMemTree\r\n\tmavls.enableMemVal = subcfg.EnableMemVal\r\n\tmavl.EnableMavlPrefix(mavls.enableMavlPrefix)\r\n\tmavl.EnableMVCC(mavls.enableMVCC)\r\n\tmavl.EnablePrune(mavls.enableMavlPrune)\r\n\tmavl.SetPruneHeight(int(mavls.pruneHeight))\r\n\tmavl.EnableMemTree(mavls.enableMemTree)\r\n\tmavl.EnableMemVal(mavls.enableMemVal)\r\n\tbs.SetChild(mavls)\r\n\treturn mavls\r\n}", "func stateAndStore(config *cfg.Config, pubKey crypto.PubKey, appVersion version.Protocol) (dbm.DB, sm.State, *mockBlockStore) {\n\tstateDB := dbm.NewMemDB()\n\tstate, _ := sm.MakeGenesisStateFromFile(config.GenesisFile())\n\tstate.Version.Consensus.App = appVersion\n\tstore := NewMockBlockStore(config, state.ConsensusParams)\n\treturn stateDB, state, store\n}", "func NewManager(\n\tmetricsConfig config.MetricsConfig,\n\tflushInterval time.Duration,\n\tloggers ldlog.Loggers,\n) (*Manager, error) {\n\tmetricsRelayID := uuid.New()\n\n\texporters, err := registerExporters(allExporterTypes(), metricsConfig, loggers)\n\tif err != nil { // COVERAGE: can't make this happen in unit tests\n\t\treturn nil, err\n\t}\n\n\tregisterPublicViewsOnce.Do(func() {\n\t\terr = view.Register(getPublicViews()...)\n\t})\n\tif err != nil { // COVERAGE: can't make this happen in unit tests\n\t\treturn nil, errInitMetricsViews(err)\n\t}\n\tregisterPrivateViewsOnce.Do(func() {\n\t\terr = view.Register(getPrivateViews()...)\n\t})\n\tif err != nil { // COVERAGE: can't make this happen in unit tests\n\t\treturn nil, errInitMetricsViews(err)\n\t}\n\n\tctx, _ := tag.New(context.Background(), tag.Insert(relayIDTagKey, metricsRelayID))\n\n\tm := &Manager{\n\t\topenCensusCtx: ctx,\n\t\tmetricsRelayID: metricsRelayID,\n\t\texporters: exporters,\n\t\tflushInterval: flushInterval,\n\t\tloggers: loggers,\n\t}\n\tif m.flushInterval <= 0 {\n\t\tm.flushInterval = defaultFlushInterval\n\t}\n\n\treturn m, nil\n}", "func New(localAddress raft.ServerAddress, dialOptions []grpc.DialOption) *Manager {\n\treturn &Manager{\n\t\tlocalAddress: localAddress,\n\t\tdialOptions: dialOptions,\n\n\t\trpcChan: make(chan raft.RPC),\n\t\tconnections: map[raft.ServerID]*conn{},\n\t}\n}", "func New(h Handler, tick time.Duration) *FSM {\n\thm, s := h.Init()\n\tfsm := &FSM{\n\t\thandler: h,\n\t\thandlerMap: hm,\n\t\tstate: s,\n\t\ttransitionChan: make(chan *Transition),\n\t\ttickChan: time.Tick(tick),\n\t\tstateChan: make(chan chan string),\n\t}\n\t// Start working.\n\tgo fsm.backend()\n\treturn fsm\n}", "func NewConfigManager(namespace string, configMap string, client client.Client, settlePeriod time.Duration) *ConfigManager {\n\treturn &ConfigManager{\n\t\tnamespace: namespace,\n\t\tconfigMap: configMap,\n\t\tclient: client,\n\t\tpolicyList: make(map[ResourceIdentifier][]pomeriumconfig.Policy),\n\t\tsettleTicker: time.NewTicker(settlePeriod),\n\t}\n}", "func New(lsnCfg engine.Listener, router http.Handler, stapler stapler.Stapler,\n\tconnTck conntracker.ConnectionTracker, autoCertCache autocert.Cache, wg *sync.WaitGroup,\n\toptions proxy.Options,\n) (*T, error) {\n\n\tscopedRouter, err := newScopeRouter(lsnCfg.Scope, router)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &T{\n\t\tlsnCfg: lsnCfg,\n\t\trouter: router,\n\t\toptions: options,\n\n\t\tstapler: stapler,\n\t\tconnTck: connTck,\n\t\tautoCertCache: autoCertCache,\n\t\tserveWg: wg,\n\t\tscopedRouter: scopedRouter,\n\t\tstate: srvStateInit,\n\t}, nil\n}", "func NewConfigManager(storeCallback func([]*PeerConfig) error) *ConfigManager {\n\treturn &ConfigManager{\n\t\tstoreCallback: storeCallback,\n\t\tstoreOnChange: false,\n\t\tpeers: []*PeerConfig{},\n\t}\n}", "func New(config *Config) Store {\n\treturn newStore(config)\n}", "func New(opt Options, fsm FSM, storageDir string) (*Raft, error) {\n\tif err := opt.validate(); err != nil {\n\t\treturn nil, err\n\t}\n\tif opt.Logger == nil {\n\t\topt.Logger = nopLogger{}\n\t}\n\tif opt.Alerts == nil {\n\t\topt.Alerts = nopAlerts{}\n\t}\n\tstore, err := openStorage(storageDir, opt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif store.cid == 0 || store.nid == 0 {\n\t\treturn nil, ErrIdentityNotSet\n\t}\n\tsm := &stateMachine{\n\t\tFSM: fsm,\n\t\tid: store.nid,\n\t\tch: make(chan interface{}, 1024), // todo configurable capacity\n\t\tsnaps: store.snaps,\n\t}\n\tr := &Raft{\n\t\trtime: newRandTime(),\n\t\ttimer: newSafeTimer(),\n\t\trpcCh: make(chan *rpc),\n\t\tdisconnected: make(chan uint64, 20),\n\t\tfsm: sm,\n\t\tfsmRestoredCh: make(chan error, 5),\n\t\tsnapTimer: newSafeTimer(),\n\t\tsnapInterval: opt.SnapshotInterval,\n\t\tsnapThreshold: opt.SnapshotThreshold,\n\t\tstorage: store,\n\t\tstate: Follower,\n\t\thbTimeout: opt.HeartbeatTimeout,\n\t\tpromoteThreshold: opt.PromoteThreshold,\n\t\tshutdownOnRemove: opt.ShutdownOnRemove,\n\t\tlogger: opt.Logger,\n\t\talerts: opt.Alerts,\n\t\tbandwidth: opt.Bandwidth,\n\t\tdialFn: net.DialTimeout,\n\t\tconnPools: make(map[uint64]*connPool),\n\t\ttaskCh: make(chan Task),\n\t\tfsmTaskCh: make(chan FSMTask),\n\t\tnewEntryCh: make(chan *newEntry),\n\t\tclose: make(chan struct{}),\n\t\tclosed: make(chan struct{}),\n\t}\n\n\tr.resolver = &resolver{\n\t\tdelegate: opt.Resolver,\n\t\taddrs: make(map[uint64]string),\n\t\tlogger: r.logger,\n\t\talerts: r.alerts,\n\t}\n\tr.resolver.update(store.configs.Latest)\n\n\treturn r, nil\n}", "func NewStore(ctx context.Context, onAfterStore ...func(name string, value interface{})) *Store {\n\tlogger := logging.FromContext(ctx)\n\n\tstore := &Store{\n\t\tUntypedStore: configmap.NewUntypedStore(\n\t\t\t\"route\",\n\t\t\tlogger,\n\t\t\tconfigmap.Constructors{\n\t\t\t\tDomainConfigName: NewDomainFromConfigMap,\n\t\t\t\tgc.ConfigName: gc.NewConfigFromConfigMapFunc(ctx),\n\t\t\t\tnetcfg.ConfigMapName: network.NewConfigFromConfigMap,\n\t\t\t\tcfgmap.FeaturesConfigName: cfgmap.NewFeaturesConfigFromConfigMap,\n\t\t\t},\n\t\t\tonAfterStore...,\n\t\t),\n\t}\n\n\treturn store\n}", "func New(storeConfig config.Store) (*Store, error) {\n\tdb, err := bolt.Open(storeConfig.DBPath, 0644, &bolt.Options{Timeout: 1 * time.Second})\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not open bolt DB database\")\n\t}\n\tbucketName := []byte(\"shorted\")\n\terr = db.Update(func(tx *bolt.Tx) error {\n\t\t_, err := tx.CreateBucketIfNotExists(bucketName)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Store{\n\t\tdb: db,\n\t\tidLength: storeConfig.ShortedIDLength,\n\t\tbucketName: bucketName,\n\t}, nil\n}", "func NewManager(run, shutdown func(context.Context) error) *Manager {\n\tmgr := &Manager{\n\t\trunFunc: run,\n\t\tshutdownFunc: shutdown,\n\n\t\trunDone: make(chan struct{}),\n\t\tstartupDone: make(chan struct{}),\n\t\tshutdownDone: make(chan struct{}),\n\t\tpauseStart: make(chan struct{}),\n\t\tstatus: make(chan Status, 1),\n\t}\n\tmgr.status <- StatusUnknown\n\treturn mgr\n}", "func NewManager(ms map[string]func() bool, opts Options) (Manager, error) {\n\tt := &Target{\n\t\tmetrics: ms,\n\t\topts: &opts,\n\t}\n\t// Set default namespace\n\tif t.opts.Namespace == \"\" {\n\t\tt.opts.Namespace = defaultNamespace\n\t}\n\t// Set default format\n\tif t.opts.ExpositionFormat == \"\" {\n\t\tt.opts.ExpositionFormat = defaultFormat\n\t}\n\t// Validate format value\n\terr := validateFormat(t.opts.ExpositionFormat)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn t, err\n}", "func New(ctx context.Context, dbURL string) (models.LogStore, error) {\n\tlog := common.Logger(ctx)\n\tu, err := url.Parse(dbURL)\n\tif err != nil {\n\t\tlog.WithError(err).WithFields(logrus.Fields{\"url\": dbURL}).Fatal(\"bad DB URL\")\n\t}\n\tlog.WithFields(logrus.Fields{\"db\": u.Scheme}).Debug(\"creating log store\")\n\n\tfor _, p := range providers {\n\t\tif p.Supports(u) {\n\t\t\treturn p.New(ctx, u)\n\t\t}\n\t}\n\treturn nil, fmt.Errorf(\"no log store provider available for url %s\", dbURL)\n}", "func New(ds datastore.Datastore, parameters Parameters) (Group, error) {\n\thandler, err := NewFSMHandler(parameters)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\td := handler.(fsmHandler)\n\treturn &stateGroup{StateGroup: statemachine.New(ds, handler, parameters.StateType), d: d}, nil\n}", "func New(m map[string]interface{}) (share.Manager, error) {\n\tc, err := parseConfig(m)\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"error creating a new manager\")\n\t\treturn nil, err\n\t}\n\n\tif c.GatewayAddr == \"\" {\n\t\treturn nil, errors.New(\"share manager config is missing gateway address\")\n\t}\n\n\tc.init()\n\n\t// load or create file\n\tmodel, err := loadOrCreate(c.File)\n\tif err != nil {\n\t\terr = errors.Wrap(err, \"error loading the file containing the shares\")\n\t\treturn nil, err\n\t}\n\n\treturn &mgr{\n\t\tc: c,\n\t\tmodel: model,\n\t}, nil\n}", "func NewStore(schema Schema, options ...CeousOption) *BaseStore {\n\tstore := &BaseStore{\n\t\tschema: schema,\n\t}\n\tfor _, option := range options {\n\t\toption(store)\n\t}\n\tif !store.disableCache {\n\t\tstore.runner = store._runner\n\t} else {\n\t\tstore.runner = sq.NewStmtCacher(store._runner)\n\t}\n\treturn store\n}", "func New(rootHash, utxoRootHash *corecrypto.HashType, db storage.Table) (*StateDB, error) {\n\n\ttr, err := trie.New(rootHash, db)\n\tif err != nil {\n\t\tlogger.Warn(err)\n\t\treturn nil, err\n\t}\n\tutr, err := trie.New(utxoRootHash, db)\n\tif err != nil {\n\t\tlogger.Warn(err)\n\t\treturn nil, err\n\t}\n\treturn &StateDB{\n\t\tdb: db,\n\t\ttrie: tr,\n\t\tutxoTrie: utr,\n\t\tstateObjects: make(map[types.AddressHash]*stateObject),\n\t\tstateObjectsDirty: make(map[types.AddressHash]struct{}),\n\t\tlogs: make(map[corecrypto.HashType][]*types.Log),\n\t\tpreimages: make(map[corecrypto.HashType][]byte),\n\t\tjournal: newJournal(),\n\t}, nil\n}", "func New(db *db.DB) core.StepStore {\n\treturn &stepStore{db}\n}", "func New(folder string, cfg *Config) (*DB, error) {\n\tenv, err := lmdb.NewEnv()\n\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"env create failed\")\n\t}\n\n\terr = env.SetMaxDBs(cfg.MaxDBs)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"env config failed\")\n\t}\n\terr = env.SetMapSize(cfg.SizeMbs * 1024 * 1024)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"map size failed\")\n\t}\n\n\tif err = env.SetFlags(cfg.EnvFlags); err != nil {\n\t\treturn nil, errors.Wrap(err, \"set flag\")\n\t}\n\n\tos.MkdirAll(folder, os.ModePerm)\n\terr = env.Open(folder, 0, cfg.Mode)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"open env\")\n\t}\n\n\tvar staleReaders int\n\tif staleReaders, err = env.ReaderCheck(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"reader check\")\n\t}\n\tif staleReaders > 0 {\n\t\tlog.Printf(\"cleared %d reader slots from dead processes\", staleReaders)\n\t}\n\n\tvar dbi lmdb.DBI\n\terr = env.Update(func(txn *lmdb.Txn) (err error) {\n\t\tdbi, err = txn.CreateDBI(\"agg\")\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"create DB\")\n\t}\n\n\treturn &DB{env, dbi}, nil\n\n}", "func New(log zerolog.Logger, chain Chain, feed Feeder, index index.Writer, options ...func(*MapperConfig)) (*Mapper, error) {\n\n\t// We don't use a checkpoint by default. The options can set one, in which\n\t// case we will add the checkpoint as a finalized state commitment in our\n\t// trie registry.\n\tcfg := MapperConfig{\n\t\tCheckpointFile: \"\",\n\t\tPostProcessing: PostNoop,\n\t}\n\tfor _, option := range options {\n\t\toption(&cfg)\n\t}\n\n\t// Check if the checkpoint file exists.\n\tif cfg.CheckpointFile != \"\" {\n\t\tstat, err := os.Stat(cfg.CheckpointFile)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"invalid checkpoint file: %w\", err)\n\t\t}\n\t\tif stat.IsDir() {\n\t\t\treturn nil, fmt.Errorf(\"invalid checkpoint file: directory\")\n\t\t}\n\t}\n\n\ti := Mapper{\n\t\tlog: log,\n\t\tchain: chain,\n\t\tfeed: feed,\n\t\tindex: index,\n\t\tcheckpoint: cfg.CheckpointFile,\n\t\tpost: cfg.PostProcessing,\n\t\twg: &sync.WaitGroup{},\n\t\tstop: make(chan struct{}),\n\t}\n\n\treturn &i, nil\n}", "func NewState() State {\n\treturn State{\n\t\tTime: NewLocalTime(time.Now()),\n\t}\n}", "func New() Manager {\n\treturn Manager{\n\t\tState: make(map[string]string),\n\t\tClientHolder: make(map[string]utils.Set),\n\t\tClientQueue: make(map[string]utils.Queue),\n\t}\n}", "func NewStore() *Store {\n\treturn &Store{commands: make(map[string]*Config, 0)}\n}", "func New(provided Config) (*PeerStore, error) {\n\tcfg := provided.Validate()\n\n\tps := &PeerStore{\n\t\tshards: newShardContainer(cfg.ShardCountBits),\n\t\tclosed: make(chan struct{}),\n\t\tcfg: cfg,\n\t}\n\n\t// Start a goroutine for garbage collection.\n\tps.wg.Add(1)\n\tgo func() {\n\t\tdefer ps.wg.Done()\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ps.closed:\n\t\t\t\treturn\n\t\t\tcase <-time.After(cfg.GarbageCollectionInterval):\n\t\t\t\tcutoffTime := time.Now().Add(cfg.PeerLifetime * -1)\n\t\t\t\tlog.Debug(\"optmem: collecting garbage\", log.Fields{\"cutoffTime\": cutoffTime})\n\t\t\t\tps.collectGarbage(cutoffTime)\n\t\t\t\tlog.Debug(\"optmem: finished collecting garbage\")\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Start a goroutine for reporting statistics to Prometheus.\n\tps.wg.Add(1)\n\tgo func() {\n\t\tdefer ps.wg.Done()\n\t\tt := time.NewTicker(cfg.PrometheusReportingInterval)\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-ps.closed:\n\t\t\t\tt.Stop()\n\t\t\t\treturn\n\t\t\tcase <-t.C:\n\t\t\t\tbefore := time.Now()\n\t\t\t\tlog.Debug(\"optmem: populating prometheus...\")\n\t\t\t\tps.populateProm()\n\t\t\t\tlog.Debug(\"storage: populateProm() finished\", log.Fields{\"timeTaken\": time.Since(before)})\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn ps, nil\n}", "func New(path string, opts *opt.Options) (*Store, error) {\n\tvar (\n\t\terr error\n\t\tdb *leveldb.DB\n\t)\n\n\tif path == \"\" {\n\t\tdb, err = leveldb.Open(ldbStorage.NewMemStorage(), opts)\n\t} else {\n\t\tdb, err = leveldb.OpenFile(path, opts)\n\t\tif ldbErrors.IsCorrupted(err) && !opts.GetReadOnly() {\n\t\t\tdb, err = leveldb.RecoverFile(path, opts)\n\t\t}\n\t}\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Store{\n\t\tdb: db,\n\t\tpath: path,\n\t}, nil\n}", "func NewManager(t cluster.Target, config *cmn.Config, st stats.Tracker) *Manager {\n\tecClient := cmn.NewClient(cmn.TransportArgs{\n\t\tTimeout: config.Client.Timeout.D(),\n\t\tUseHTTPS: config.Net.HTTP.UseHTTPS,\n\t\tSkipVerify: config.Net.HTTP.SkipVerify,\n\t})\n\treb := &Manager{\n\t\tt: t,\n\t\tfilterGFN: filter.NewDefaultFilter(),\n\t\tstatTracker: st,\n\t\tstages: newNodeStages(),\n\t\tecClient: ecClient,\n\t}\n\trebcfg := &config.Rebalance\n\tdmExtra := bundle.Extra{\n\t\tRecvAck: reb.recvAck,\n\t\tCompression: rebcfg.Compression,\n\t\tMultiplier: int(rebcfg.Multiplier),\n\t}\n\tdm, err := bundle.NewDataMover(t, rebTrname, reb.recvObj, cluster.Migrated, dmExtra)\n\tif err != nil {\n\t\tcos.ExitLogf(\"%v\", err)\n\t}\n\treb.dm = dm\n\treb.registerRecv()\n\treturn reb\n}", "func New(ctx context.Context, log logger.Logger, db *db.Store, cache *cache.Cache) (*Store, error) { // nolint:gocognit\n\ts := &Store{\n\t\tlog: log,\n\t\tcache: cache,\n\t}\n\n\t// Set configuration\n\ts.setConfig()\n\n\tvar err error\n\n\tswitch s.typeStore {\n\tcase \"postgres\":\n\t\ts.store, err = postgres.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"mongo\":\n\t\ts.store, err = mongo.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"redis\":\n\t\ts.store, err = redis.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"dgraph\":\n\t\ts.store, err = dgraph.New(ctx, db, log)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"leveldb\":\n\t\ts.store, err = leveldb.New(ctx, db)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"badger\":\n\t\ts.store, err = badger.New(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\tcase \"ram\":\n\t\tfallthrough\n\tdefault:\n\t\ts.store, err = ram.New(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\tlog.Info(\"init linkStore\", field.Fields{\n\t\t\"db\": s.typeStore,\n\t})\n\n\treturn s, nil\n}", "func New(dir string) (*Store, error) {\n\tpath := filepath.Clean(dir)\n\tif err := os.MkdirAll(path, 0700); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Store{dir: path}, nil\n}", "func NewStore(cfgFilepath string) (*Store, error) {\n\tcfg, err := getConfig(cfgFilepath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb, err := CreateDB(*cfg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Store{db}, nil\n}", "func New(ctx context.Context, db *sql.DB, m map[string]string) (*Store, error) {\n\tstore := &Store{db: db}\n\terr := store.InitTable(ctx, m)\n\treturn store, err\n}", "func New(s *consensus.State, tpool modules.TransactionPool, w modules.Wallet) (m *Miner, err error) {\n\tif s == nil {\n\t\terr = errors.New(\"miner cannot use a nil state\")\n\t\treturn\n\t}\n\tif tpool == nil {\n\t\terr = errors.New(\"miner cannot use a nil transaction pool\")\n\t\treturn\n\t}\n\tif w == nil {\n\t\terr = errors.New(\"miner cannot use a nil wallet\")\n\t\treturn\n\t}\n\n\tm = &Miner{\n\t\tstate: s,\n\t\ttpool: tpool,\n\t\twallet: w,\n\n\t\tparent: s.CurrentBlock().ID(),\n\t\ttarget: s.CurrentTarget(),\n\t\tearliestTimestamp: s.EarliestTimestamp(),\n\n\t\tthreads: 1,\n\t\titerationsPerAttempt: 16 * 1024,\n\t}\n\n\taddr, _, err := m.wallet.CoinAddress()\n\tif err != nil {\n\t\treturn\n\t}\n\tm.address = addr\n\n\tm.tpool.TransactionPoolSubscribe(m)\n\treturn\n}" ]
[ "0.58048505", "0.5739119", "0.5596712", "0.54425263", "0.5434711", "0.53853023", "0.53579974", "0.53098", "0.53037655", "0.5286296", "0.52441293", "0.52342004", "0.5219029", "0.5151147", "0.5101644", "0.50777805", "0.5030564", "0.5018116", "0.50093615", "0.49994498", "0.49910754", "0.498474", "0.49529254", "0.4945203", "0.49236366", "0.4912949", "0.49046293", "0.48965716", "0.4881642", "0.4869332", "0.48455903", "0.48407128", "0.4829387", "0.4828855", "0.4826705", "0.48245144", "0.48164147", "0.4807218", "0.47937068", "0.47822574", "0.47710818", "0.4762538", "0.4762102", "0.4743732", "0.47381276", "0.47359788", "0.47327513", "0.47289285", "0.47259802", "0.47254264", "0.47136784", "0.4711325", "0.4704249", "0.4694144", "0.46898985", "0.46713805", "0.46687496", "0.4664575", "0.4656525", "0.46453166", "0.46381345", "0.46363464", "0.4628286", "0.46102223", "0.46056747", "0.46018806", "0.458702", "0.45812565", "0.4571634", "0.45712593", "0.45677888", "0.45654133", "0.45650575", "0.45639098", "0.4560518", "0.45508474", "0.45493224", "0.45447573", "0.45417538", "0.45386398", "0.45378992", "0.45314243", "0.4527724", "0.45241028", "0.45225647", "0.45188686", "0.4505416", "0.44991297", "0.44919083", "0.449128", "0.448832", "0.44861567", "0.4469771", "0.44688463", "0.44597906", "0.44579995", "0.44575998", "0.44538113", "0.4451306", "0.4449035" ]
0.694828
0
AddSaveable is an option that adds a given saveable as one that should be saved under the given name. The name must be the same across uses of the statemanager (e.g. program invocations) for it to be serialized and deserialized correctly.
func AddSaveable(name string, saveable Saveable) Option { return (Option)(func(m StateManager) { manager, ok := m.(*basicStateManager) if !ok { seelog.Critical("Unable to add to state manager; unknown instantiation") return } manager.state.Data[name] = &saveable }) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Save(name string, data interface{}) error {\n\tp := jsonPath(name)\n\tb, err := json.MarshalIndent(data, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := ioutil.WriteFile(p, b, os.ModePerm); err != nil {\n\t\treturn err\n\t}\n\n\tlog.Printf(\"Saved %s\\n\", p)\n\treturn nil\n}", "func Save() {\n\tdata := Savedata{\n\t\tName: GS.current.name,\n\t\tGamestate: GS.current,\n\t}\n\n\tf, err := json.MarshalIndent(data, \"\", \" \")\n\tcheck(err)\n\tioutil.WriteFile(\"data/savegame.json\", f, 0644)\n}", "func (p FilePersister) Save(v *viper.Viper, basename string) error {\n\tv.SetConfigType(\"json\")\n\tv.AddConfigPath(p.Dir)\n\tv.SetConfigName(basename)\n\n\tif _, err := os.Stat(p.Dir); os.IsNotExist(err) {\n\t\tif err := os.MkdirAll(p.Dir, os.FileMode(0755)); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// WriteConfig is broken.\n\t// Someone proposed a fix in https://github.com/spf13/viper/pull/503,\n\t// but the fix doesn't work yet.\n\t// When it's fixed and merged we can get rid of `path`\n\t// and use viperConfig.WriteConfig() directly.\n\tpath := filepath.Join(p.Dir, fmt.Sprintf(\"%s.json\", basename))\n\treturn v.WriteConfigAs(path)\n}", "func (s Store) Save(param ParameterSet) error {\n\tbytes, err := json.MarshalIndent(param, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn s.backingStore.Save(ItemType, \"\", param.Name, bytes)\n}", "func (self *Repository) Save(message string) error {\n\t// TODO: Validate message, maybe even use a ruleset file so a project can\n\t// define rules for commits\n\t// Adds occur automatically since we are watching all files for changes always\n\t// to rebuild and such sop we just need to commit\n\treturn nil\n}", "func (self *Repository) Save(message string) error {\n\t// TODO: Validate message, maybe even use a ruleset file so a project can\n\t// define rules for commits\n\t// Adds occur automatically since we are watching all files for changes always\n\t// to rebuild and such sop we just need to commit\n\treturn nil\n}", "func (m *stateManager) Save() (err error) {\n\terr = m.storage.Write(m.state)\n\n\tif err == nil {\n\t\tm.stateChanged = false\n\t\tm.stateLoaded = true\n\t}\n\n\treturn\n}", "func Save(r io.ReadWriter, a audio.Audio) error {\n\treturn fmt.Errorf(\"unsupported Functionality\")\n}", "func (m SaveHelper) SaveGame(autosave bool) (ret string, err error) {\n\tif b, e := json.Marshal(m.values); e != nil {\n\t\terr = e\n\t} else {\n\t\tvar saveId string\n\t\tif autosave {\n\t\t\tsaveId = \"autosave\"\n\t\t}\n\t\tif r, e := m.saver.SaveBlob(saveId, b); e != nil {\n\t\t\terr = e\n\t\t} else {\n\t\t\tret = r\n\t\t}\n\t}\n\treturn\n}", "func (m *Manager) Add(name string, bot *Bot) error {\n\t// Check if key already exists\n\tif _, ok := m.bots[name]; ok {\n\t\treturn ErrBotExists\n\t}\n\n\t// Check if we should start the bot\n\tif m.running {\n\t\tif err := bot.Connect(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Add the bot to the manager\n\tm.bots[name] = bot\n\n\treturn nil\n}", "func Save(path string, x interface{}) error {\n\tmtx.Lock()\n\tdefer mtx.Unlock()\n\n\tfile, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\tdata, err := toJSON(x)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = io.Copy(file, data)\n\treturn err\n}", "func (f *File) Save(name string) error {\n\treturn ioutil.WriteFile(name, []byte(f.String()), 0666)\n}", "func (e *ObservableEditableBuffer) SaveableAndDirty() bool {\n\tsad := e.Dirty() && !e.IsDirOrScratch()\n\treturn e.details.Name != \"\" && sad\n}", "func (os OptionSet) Add(o *Option) {\n\tos[o.Name] = o\n}", "func (a *ArtifactImpl) Save(ctx context.Context, path string) (bool, error) {\n\treturn a.artifact.Save(ctx, path)\n}", "func (tc *ThingConfig) Save(name string) error {\n\n\td, err := yaml.Marshal(&tc)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn errors.Wrap(ioutil.WriteFile(fmt.Sprintf(\"%s.yaml\", name), d, 0600), \"Failed to save thing config\")\n}", "func Save(path string, v interface{}) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\tr, err := Marshal(v)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = io.Copy(f, r)\n\treturn err\n}", "func Save(path string, v interface{}) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\tr, err := Marshal(v)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = io.Copy(f, r)\n\treturn err\n}", "func Save(path string, v interface{}) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\tr, err := Marshal(v)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = io.Copy(f, r)\n\treturn err\n}", "func (f *File) Save(fileName string, offset int64, xestatus string) error {\n\tif f.file == nil {\n\t\treturn errors.New(\"state file not open\")\n\t}\n\n\terr := writeState(f.file, fileName, offset, xestatus)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"writeStatus\")\n\t}\n\n\treturn nil\n}", "func (r *Registry) Add(obj client.Object) error {\n\tif obj == nil || reflect.ValueOf(obj) == reflect.Zero(reflect.TypeOf(obj)) {\n\t\treturn nil\n\t}\n\n\tobjectName, err := r.objectName(obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfilename := objectName + \".yaml\"\n\n\tif _, ok := r.nameToObject[filename]; ok {\n\t\treturn fmt.Errorf(\"duplicate filename in registry: %q\", filename)\n\t}\n\n\tserializationYAML, err := runtime.Encode(r.codec, obj)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tr.nameToObject[filename] = &object{\n\t\tobj: obj,\n\t\tserialization: serializationYAML,\n\t}\n\n\treturn nil\n}", "func (l *localLinker) Save(ctx context.Context, req *pbd.SaveRequest) (*pbd.Empty, error) {\n\tl.store[req.Key] = req.Value\n\treturn &pbd.Empty{}, nil\n}", "func (s *saver) Save(journey models.Journey) error {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.state == closed {\n\t\treturn ErrSaverIsClosed\n\t}\n\n\tif len(s.buffer) == cap(s.buffer) {\n\t\treturn ErrInternalBufferIsFull\n\t}\n\n\ts.buffer = append(s.buffer, journey)\n\treturn nil\n}", "func (s *State) Save() error {\n\tf, err := os.Create(getPathForUsername(s.Username))\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\treturn json.NewEncoder(f).Encode(s)\n}", "func AddSonobuoyConfigFlag(cfg *SonobuoyConfig, cmd *cobra.Command) {\n\tcmd.PersistentFlags().Var(\n\t\tcfg, \"config\",\n\t\t\"path to a sonobuoy configuration JSON file. Overrides --mode\",\n\t)\n\tcmd.MarkFlagFilename(\"config\", \"json\")\n}", "func (s *Stash) Save(key string, value interface{}) error {\n\tswitch s.version {\n\tcase version1:\n\t\tmarshalledData, err := json.Marshal(value)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"error marshalling value\")\n\t\t}\n\t\ts.mutex.Lock()\n\t\ts.data.(v1Data)[key] = marshalledData\n\t\ts.mutex.Unlock()\n\n\t\tif s.autoFlush {\n\t\t\treturn s.Flush()\n\t\t} else {\n\t\t\treturn nil\n\t\t}\n\tdefault:\n\t\treturn UnknownVersionError{s.version}\n\t}\n}", "func (s *Storage) SaveGame(g *game.Game) error {\n\ts.mux.Lock()\n\tdefer s.mux.Unlock()\n\ts.games[g.ID] = g\n\treturn nil\n}", "func (m *Machine) Save(n string) error {\n\tf, err := os.Create(n)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to create vm config file\")\n\t}\n\tdefer f.Close()\n\n\tif err := json.NewEncoder(f).Encode(m); err != nil {\n\t\treturn errors.Wrap(err, \"failed to serialize machine object\")\n\t}\n\n\treturn nil\n}", "func save(novel *Novel) {\n\t//if novel exist history\n\ttag := false\n\tfor index, historyNovel := range historyNovels {\n\t\tif historyNovel.Name == novel.Name {\n\t\t\thistoryNovels[index] = novel\n\t\t\ttag = true\n\t\t}\n\t}\n\tif !tag {\n\t\thistoryNovels = append(historyNovels, novel)\n\t}\n\tSaveHistory(historyNovels)\n\tfmt.Println(\"Save complete...\")\n}", "func (l Level) save(path string) error {\n\tf, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY|os.O_TRUNC, 0777)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"open file to save level: %w\", err)\n\t}\n\tdefer f.Close()\n\tencoder := json.NewEncoder(f)\n\tencoder.SetIndent(\"\", \" \")\n\terr = encoder.Encode(l)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"save level: %w\", err)\n\t}\n\treturn nil\n}", "func addMaybeNamedOrString(s *scope, name string, obj pyObject, anon func(core.BuildInput), named func(string, core.BuildInput), systemAllowed, tool bool) {\n\tif obj == nil {\n\t\treturn\n\t}\n\tif str, ok := asString(obj); ok {\n\t\tif bi := parseBuildInput(s, str, name, systemAllowed, tool); bi != nil {\n\t\t\tanon(bi)\n\t\t}\n\t\treturn\n\t}\n\taddMaybeNamed(s, name, obj, anon, named, systemAllowed, tool)\n}", "func Save(w io.Writer, v interface{}) error {\n\tdata, err := json.Marshal(v)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = io.WriteString(w, string(data))\n\treturn err\n}", "func (n *nameHistory) Save() error {\n\tif !n.isChanged {\n\t\treturn nil\n\t}\n\n\tfp, err := os.OpenFile(n.filepath, os.O_CREATE|os.O_WRONLY|os.O_TRUNC, os.ModePerm)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not open %q file: %v\", n.filepath, err)\n\t}\n\tdefer fp.Close()\n\n\tif err := yaml.NewEncoder(fp).Encode(n.entries); err != nil {\n\t\treturn fmt.Errorf(\"could not save table name history: %v\", err)\n\t}\n\n\tif err := utils.SyncFileAndDirectory(fp); err != nil {\n\t\treturn fmt.Errorf(\"could not sync oid to name map file: %v\", err)\n\t}\n\n\tn.isChanged = false\n\n\treturn nil\n}", "func Save(fileName string, dst interface{}) error {\n\t// Create all directories\n\tif err := os.MkdirAll(filepath.Dir(fileName), os.ModePerm); err != nil {\n\t\treturn err\n\t}\n\tfile, err := os.Create(fileName)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\tif err == nil {\n\t\tencoder := gob.NewEncoder(file)\n\t\tif err = encoder.Encode(dst); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func AddResource(name string, observable, editable bool) (*Resource, bool) {\n\tif !utils.IsEmpty(name) && !HasResource(name) {\n\t\tr := Resource{Name: name, Observable: observable, Editable: editable}\n\t\tResources[name] = r\n\t\treturn &r, true\n\t}\n\treturn nil, false\n}", "func (m Db) Save(n int, p architecture.Version) {\n\tm.Dbm[n] = p\n}", "func (se *StateEngine) AddManager(m StateManager) {\n\tse.mgrLock.Lock()\n\tdefer se.mgrLock.Unlock()\n\tse.managers = append(se.managers, m)\n}", "func (defaultStorage) Save() error {\n\tpanic(noConfigStorage)\n}", "func (ws *WalletStore) Save() {\n\tvar buffer bytes.Buffer\n\tgob.Register(elliptic.P256())\n\tencoder := gob.NewEncoder(&buffer)\n\terr := encoder.Encode(ws.Wallets)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfile := ws.Config.GetWalletStoreFile(ws.NodeID)\n\terr = ioutil.WriteFile(file, buffer.Bytes(), 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (s *Store) Save() error {\n\tbk, err := os.OpenFile(filepath.Join(s.rwDirPath, storeBkName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer bk.Close()\n\n\tdst, err := os.OpenFile(filepath.Join(s.rwDirPath, storeName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer dst.Close()\n\n\t// backing up current store\n\t_, err = io.Copy(bk, dst)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\n\tenc := gob.NewEncoder(dst)\n\tbook := s.Clone()\n\terr = enc.Encode(book)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\treturn nil\n}", "func (fb *FileBackend) save(state *storage.State) error {\n\tout, err := proto.Marshal(state)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to encode state: %w\", err)\n\t}\n\ttmp := fmt.Sprintf(fb.path+\".%v\", time.Now())\n\tif err := ioutil.WriteFile(tmp, out, 0600); err != nil {\n\t\treturn fmt.Errorf(\"failed to write state: %w\", err)\n\t}\n\terr = os.Rename(tmp, fb.path)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to move state: %w\", err)\n\t}\n\treturn nil\n}", "func (b *Base) Add(key string, value interface{}) Serializer {\n\treturn b.AddIf(alwaysTrue, key, value)\n}", "func (r *Reserve) Save(s *Store) error {\n\tdata, _ := json.Marshal(s)\n\tif err := ioutil.WriteFile(r.path, data, 0644); err != nil {\n\t\treturn fmt.Errorf(\"Failed to set %s: %s\", r.name, err)\n\t}\n\treturn nil\n}", "func (c *Component) Save(n string, i interface{}) {\n\tc.GM.DB.Save(n, i)\n}", "func (c *Config) Save(name, key, value string) (err error) {\n\treturn c.SaveGlobal(name, key, value)\n}", "func (b Banai) Save(fileName string) (string, error) {\n\tabs, e := filepath.Abs(fileName)\n\tif e != nil {\n\t\treturn \"\", e\n\t}\n\tstashID := uuid.NewString()\n\n\te = fsutils.CopyfsItem(abs, stashID)\n\tif e != nil {\n\t\treturn \"\", e\n\t}\n\treturn stashID, nil\n}", "func (p *DefaultParser) Save(buf *bytes.Buffer, filename string) error {\n\terr := ioutil.WriteFile(filename, buf.Bytes(), 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func Save(path string, object interface{}) error {\n\tfile, err := os.Create(path)\n\tif err == nil {\n\t\tencoder := gob.NewEncoder(file)\n\t\tencoder.Encode(object)\n\t}\n\tfile.Close()\n\treturn err\n}", "func Save(path string, object interface{}) error {\n\tfile, err := os.Create(path)\n\tdefer file.Close()\n\tif err == nil {\n\t\tencoder := gob.NewEncoder(file)\n\t\tencoder.Encode(object)\n\t}\n\treturn err\n}", "func Save(path string, v interface{}) error {\n\tformatter, err := parsePath(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfile, err := os.OpenFile(path, os.O_TRUNC|os.O_CREATE|os.O_WRONLY, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\treturn formatter.Encode(file, v)\n}", "func Save(v interface{}, saveTo string) error {\n\tkv, err := reflectutil.ToMapWithDefault(v)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// iterate over map setting key => value in ini file\n\tcfg := ini.Empty()\n\tfor k, v := range kv {\n\t\tvar data string\n\t\tswitch v.(type) {\n\t\tcase float64, float32:\n\t\t\tdata = strconv.FormatFloat(v.(float64), 'f', -1, 64)\n\t\tdefault:\n\t\t\tdata = fmt.Sprint(v)\n\t\t}\n\t\tcfg.Section(\"\").NewKey(k, data)\n\t}\n\treturn cfg.SaveTo(saveTo)\n}", "func (w *Watcher) Add(name string) error { return w.AddWith(name) }", "func Add(name, url string) func(*types.Cmd) {\n\treturn func(g *types.Cmd) {\n\t\tg.AddOptions(\"add\")\n\t\tg.AddOptions(name)\n\t\tg.AddOptions(url)\n\t}\n}", "func (g *GamePlayerHand) Save() error {\n\tif g.ID == 0 {\n\t\treturn g.insert()\n\t}\n\n\treturn g.update()\n}", "func Save(fname string, scheme base16.Scheme, perm os.FileMode, writerArg ...Writer) error {\n\tvar fileWriter Writer = &FileWriter{}\n\tbase16Yaml := toBase16Yaml(scheme)\n\n\tif len(writerArg) == 1 {\n\t\tfileWriter = writerArg[0]\n\t}\n\n\tdata, err := MarshalBase16Yaml(base16Yaml)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = fileWriter.WriteFile(fname, data, perm)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Cluster) AddSyncable(toml []byte, snap *bridge.Snapshot) error {\n\tname, syncable, err := syncable.ParseSyncable(\"toml\", strings.NewReader(string(toml)), c.Data.Databases)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Router could not create syncable\")\n\t}\n\n\tlog.Printf(\"[node %d] About to initialize syncable: %s...\\n\", c.id, name)\n\tif err := syncable.Init(context.Background()); err != nil {\n\t\treturn errors.Wrapf(err, \"could not initialize syncable %s\", name)\n\t}\n\tlog.Printf(\"[node %d] ...Initialized syncable: %s\\n\", c.id, name)\n\n\tbridge, err := bridgeFactory.New(name, syncable, c.Data.Topics, c.leader, c, snap)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"Router could not create bridge\")\n\t}\n\tgo func() {\n\t\t// TODO Make the 5 seconds tunable in the config file\n\t\terr := bridge.Init(context.Background(), c.errorC, 5*time.Second)\n\t\tif err != nil {\n\t\t\tc.errorC <- err\n\t\t}\n\t}()\n\n\tc.mu.Lock()\n\tc.TOML.Syncables[name] = string(toml)\n\tc.Data.Syncables[name] = syncable\n\tc.Data.Bridges[name] = bridge\n\tc.mu.Unlock()\n\n\treturn nil\n}", "func (c *Controller) Save(name string, f file.Any) error {\n\tb, err := json.Marshal(f)\n\tif err != nil {\n\t\treturn err\n\t}\n\tutil.Log(string(b))\n\n\tr := bytes.NewReader(b)\n\n\tinput := &s3.PutObjectInput{\n\t\tBody: aws.ReadSeekCloser(r),\n\t\tBucket: aws.String(c.bucket),\n\t\tKey: aws.String(name),\n\t\tServerSideEncryption: aws.String(\"AES256\"),\n\t}\n\n\tswitch f.(type) {\n\tcase file.BananasMon:\n\t\tinput.Tagging = aws.String(\"App Name=hit-the-bananas\")\n\tcase file.D2sVendorDays:\n\t\tinput.Tagging = aws.String(\"App Name=drive-2-sku\")\n\t}\n\n\tresult, err := c.c3svc.PutObject(input)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc.verIDs[name] = result.VersionId\n\n\treturn nil\n}", "func (r *Room) Save() error {\n\tdata, err := json.Marshal(r.Data)\n\tif err != nil {\n\t\treturn err\n\t}\n\tos.Mkdir(fmt.Sprintf(\"%s/rooms\", config.GetString(\"save_path\")), 0755)\n\terr = ioutil.WriteFile(fmt.Sprintf(\"%s/rooms/%s\", config.GetString(\"save_path\"), r.Data.UUID), data, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (e *Store) Add(id string, Config *Config) {\n\te.Lock()\n\te.commands[id] = Config\n\te.Unlock()\n}", "func (r *Recipe) Save(fname string) (err error) {\n\tb, err := json.MarshalIndent(r, \"\", \" \")\n\tif err != nil {\n\t\treturn\n\t}\n\terr = ioutil.WriteFile(fname, b, 0644)\n\treturn\n}", "func Save(conn redis.Conn, key, setKey string, value []byte) error {\n\t_, err := conn.Do(\"SET\", key, value)\n\tif len(setKey) > 0 && !strings.HasPrefix(key, \"player\") {\n\t\tSadd(conn, setKey, key)\n\t}\n\treturn err\n}", "func Add(name string, b []byte) {\n\tnogos[name] = b\n}", "func (app *service) Save(state State) error {\n\tjs, err := app.adapter.ToJSON(state)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tchainHash := state.Chain()\n\tindex := state.Height()\n\tpath := filePath(chainHash, index)\n\treturn app.fileService.Save(path, js)\n}", "func (g GameOptions) Save() error {\n\t// Validate all option types\n\terr := g.validate()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trecords, err := GetGameOptionRecordsForGameType(g.GameTypeID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// TODO: remove any that no longer apply\n\n\t// save all options\n\tfor i := range records {\n\t\tvalue, ok := g.Options[records[i].Name]\n\t\tif !ok {\n\t\t\tvalue = records[i].DefaultValue\n\t\t}\n\n\t\trecordValue := getStringForType(records[i].Type, value)\n\t\tg.save(records[i].GameTypeGameOptionID, recordValue)\n\t}\n\n\treturn nil\n}", "func SaveJSON(f string, iface interface{}) error {\n\tif !PathExists(filepath.Dir(f)) {\n\t\treturn fmt.Errorf(\"Missing dir; %s\", f)\n\t}\n\n\tcontent, err := json.MarshalIndent(iface, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(f, content, 0644)\n}", "func (p *XrayCommand) Save() bool {\n\treturn p.saveInfo(p)\n}", "func (manager *basicStateManager) Save() error {\n\tdefer metrics.MetricsEngineGlobal.RecordStateManagerMetric(\"SAVE\")()\n\tmanager.saveTimesLock.Lock()\n\tdefer manager.saveTimesLock.Unlock()\n\tif time.Since(manager.lastSave) >= minSaveInterval {\n\t\t// we can just save\n\t\terr := manager.ForceSave()\n\t\tmanager.lastSave = time.Now()\n\t\tmanager.nextPlannedSave = time.Time{} // re-zero it; assume all pending desires to save are fulfilled\n\t\treturn err\n\t} else if manager.nextPlannedSave.IsZero() {\n\t\t// No save planned yet, we should plan one.\n\t\tnext := manager.lastSave.Add(minSaveInterval)\n\t\tmanager.nextPlannedSave = next\n\t\tgo func() {\n\t\t\ttime.Sleep(time.Until(next))\n\t\t\tmanager.Save()\n\t\t}()\n\t}\n\t// else nextPlannedSave wasn't Zero so there's a save planned elsewhere that'll\n\t// fulfill this\n\treturn nil\n}", "func (i *SinkFlags) AddWithFlagName(cmd *cobra.Command, fname, short string) {\n\tflag := \"--\" + fname\n\tif short == \"\" {\n\t\tcmd.Flags().StringVar(&i.Sink, fname, \"\", \"\")\n\t} else {\n\t\tcmd.Flags().StringVarP(&i.Sink, fname, short, \"\", \"\")\n\t}\n\tcmd.Flag(fname).Usage = \"Addressable sink for events. \" +\n\t\t\"You can specify a broker, channel, Knative service or URI. \" +\n\t\t\"Examples: '\" + flag + \" broker:nest' for a broker 'nest', \" +\n\t\t\"'\" + flag + \" channel:pipe' for a channel 'pipe', \" +\n\t\t\"'\" + flag + \" ksvc:mysvc:mynamespace' for a Knative service 'mysvc' in another namespace 'mynamespace', \" +\n\t\t\"'\" + flag + \" https://event.receiver.uri' for an HTTP URI, \" +\n\t\t\"'\" + flag + \" ksvc:receiver' or simply '\" + flag + \" receiver' for a Knative service 'receiver' in the current namespace. \" +\n\t\t\"'\" + flag + \" special.eventing.dev/v1alpha1/channels:pipe' for GroupVersionResource of v1alpha1 'pipe'. \" +\n\t\t\"If a prefix is not provided, it is considered as a Knative service in the current namespace.\"\n\t// Use default mapping if empty\n\tif i.SinkMappings == nil {\n\t\ti.SinkMappings = defaultSinkMappings\n\t}\n\tfor _, p := range config.GlobalConfig.SinkMappings() {\n\t\t//user configuration might override the default configuration\n\t\ti.SinkMappings[p.Prefix] = schema.GroupVersionResource{\n\t\t\tResource: p.Resource,\n\t\t\tGroup: p.Group,\n\t\t\tVersion: p.Version,\n\t\t}\n\t}\n}", "func (s *Saver) Save(ctx context.Context, shoutout *Shoutout) error {\n\titem, err := dynamodbattribute.MarshalMap(shoutout)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to marshal shoutout for storage: %s\", err)\n\t}\n\n\tinput := &dynamodb.PutItemInput{\n\t\tItem: item,\n\t\tTableName: aws.String(os.Getenv(\"TABLE_NAME\")),\n\t}\n\n\tif _, err = s.ddb.PutItemWithContext(ctx, input); err != nil {\n\t\treturn fmt.Errorf(\"failed to save shoutout: %s\", err)\n\t}\n\n\treturn nil\n}", "func (self *JsonConfig) Save() (err error) {\n\tb, err := json.Marshal(self.Configurable.All())\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := ioutil.WriteFile(self.Path, b, 0600); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *reflectorStore) Add(obj interface{}) error {\n\tmetaObj := obj.(metav1.Object)\n\tentity := r.parser.Parse(obj)\n\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\n\tr.hasSynced = true\n\tr.seen[string(metaObj.GetUID())] = entity.GetID()\n\tr.wlmetaStore.Notify([]workloadmeta.CollectorEvent{\n\t\t{\n\t\t\tType: workloadmeta.EventTypeSet,\n\t\t\tSource: collectorID,\n\t\t\tEntity: entity,\n\t\t},\n\t})\n\n\treturn nil\n}", "func SaveBinary(filename string, data []byte, mode os.FileMode) error {\n\t// Write the new file to a temporary\n\ttmpname := filename + \".tmp\"\n\tif err := ioutil.WriteFile(tmpname, data, mode); err != nil {\n\t\treturn err\n\t}\n\n\t// Write the new file to the target wallet file\n\tif err := ioutil.WriteFile(filename, data, mode); err != nil {\n\t\treturn err\n\t}\n\n\t// Remove the tmp file\n\treturn os.Remove(tmpname)\n}", "func (d *dataUsageCache) save(ctx context.Context, store objectIO, name string) error {\n\tpr, pw := io.Pipe()\n\tgo func() {\n\t\tpw.CloseWithError(d.serializeTo(pw))\n\t}()\n\tdefer pr.Close()\n\n\tr, err := hash.NewReader(pr, -1, \"\", \"\", -1)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Abandon if more than 5 minutes, so we don't hold up scanner.\n\tctx, cancel := context.WithTimeout(ctx, 5*time.Minute)\n\tdefer cancel()\n\t_, err = store.PutObject(ctx,\n\t\tdataUsageBucket,\n\t\tname,\n\t\tNewPutObjReader(r),\n\t\tObjectOptions{})\n\tif isErrBucketNotFound(err) {\n\t\treturn nil\n\t}\n\treturn err\n}", "func Save(fromVar interface{}, toPath string) {\n\tdata, e := json.Marshal(fromVar)\n\terr.Panic(e)\n\te = ioutil.WriteFile(toPath, data, 0664)\n\terr.Panic(e)\n}", "func (s *storage) Save(d *entity.Deck) error {\n\td.Id = uuid.NewString()\n\ts.decks[d.Id] = d\n\n\treturn nil\n}", "func (s *VarlinkInterface) ImageSave(ctx context.Context, c VarlinkCall, options_ ImageSaveOptions) error {\n\treturn c.ReplyMethodNotImplemented(ctx, \"io.podman.ImageSave\")\n}", "func (s *defaultModelSetter) Add(m Table, n ...string) {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\tif len(n) == 0 {\n\t\tn = []string{viper.GetString(\"app.name\")}\n\t}\n\ts.m[n[0]] = append(s.m[n[0]], m)\n}", "func addMaybeNamed(s *scope, name string, obj pyObject, anon func(core.BuildInput), named func(string, core.BuildInput), systemAllowed, tool bool) {\n\tif obj == nil {\n\t\treturn\n\t}\n\tif l, ok := asList(obj); ok {\n\t\tfor _, li := range l {\n\t\t\tif bi := parseBuildInput(s, li, name, systemAllowed, tool); bi != nil {\n\t\t\t\tanon(bi)\n\t\t\t}\n\t\t}\n\t} else if d, ok := asDict(obj); ok {\n\t\ts.Assert(named != nil, \"%s cannot be given as a dict\", name)\n\t\tfor k, v := range d {\n\t\t\tif v != None {\n\t\t\t\tif l, ok := asList(v); ok {\n\t\t\t\t\tfor _, li := range l {\n\t\t\t\t\t\tif bi := parseBuildInput(s, li, name, systemAllowed, tool); bi != nil {\n\t\t\t\t\t\t\tnamed(k, bi)\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif str, ok := asString(v); ok {\n\t\t\t\t\tif bi := parseBuildInput(s, str, name, systemAllowed, tool); bi != nil {\n\t\t\t\t\t\tnamed(k, bi)\n\t\t\t\t\t}\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\ts.Assert(ok, \"Values of %s must be a string or lists of strings\", name)\n\t\t\t}\n\t\t}\n\t} else if obj != None {\n\t\ts.Assert(false, \"Argument %s must be a list or dict, not %s\", name, obj.Type())\n\t}\n}", "func (h *FakeConfigMap) Save(configMap *v1.ConfigMap) (*v1.ConfigMap, error) {\n\th.configMaps[configMap.Name] = *configMap\n\treturn configMap, nil\n}", "func (s *fsStore) Add(ns *namespace.Namespace, name string) error {\n\tif _, err := os.Stat(ns.FileName()); err != nil {\n\t\treturn err\n\t}\n\tsrc := ns.FileName()\n\n\ttrgt := s.targetPath(name, ns.Type())\n\n\tif _, err := os.Stat(trgt); err == nil {\n\t\treturn store.ErrExists\n\t}\n\n\tf, err := os.Create(trgt)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\treturn unix.Mount(src, trgt, \"\", unix.MS_BIND, \"\")\n}", "func (s *ServiceState) save() {\n\tlog.Lvl3(\"Saving service\")\n\tb, err := network.Marshal(s.Storage)\n\tif err != nil {\n\t\tlog.Error(\"Couldn't marshal service:\", err)\n\t} else {\n\t\terr = ioutil.WriteFile(s.path+\"/prifi.bin\", b, 0660)\n\t\tif err != nil {\n\t\t\tlog.Error(\"Couldn't save file:\", err)\n\t\t}\n\t}\n}", "func (o *Operator) Add(f manager.Runnable) error {\n\treturn o.mgr.Add(f)\n}", "func (w *rNibWriterInstance) SaveRanLoadInformation(inventoryName string, ranLoadInformation *entities.RanLoadInformation) error {\n\n\tkey, rnibErr := common.ValidateAndBuildRanLoadInformationKey(inventoryName)\n\n\tif rnibErr != nil {\n\t\treturn rnibErr\n\t}\n\n\tdata, err := proto.Marshal(ranLoadInformation)\n\n\tif err != nil {\n\t\treturn common.NewInternalError(err)\n\t}\n\n\tvar pairs []interface{}\n\tpairs = append(pairs, key, data)\n\n\terr = w.sdl.Set(pairs)\n\n\tif err != nil {\n\t\treturn common.NewInternalError(err)\n\t}\n\n\treturn nil\n}", "func (d *Database) Save() error {\n\tb, err := json.Marshal(d.State)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Save\n\tif err := ioutil.WriteFile(d.FilePath, b, os.ModePerm); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (b *Backend) addEntry(s string) error {\n\tfp, err := os.OpenFile(b.config.omwFile, os.O_APPEND|os.O_RDWR|os.O_CREATE, 0644)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"can't open or create %s: %q\", b.config.omwFile, err)\n\t}\n\tdefer fp.Close()\n\tdata := SavedItems{}\n\tentry := SavedEntry{}\n\tentry.ID = uuid.New().String()\n\tentry.End = time.Now()\n\tentry.Task = s\n\tdata.Entries = append(data.Entries, entry)\n\tentriesBytes, err := toml.Marshal(data)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"can't marshal data\")\n\t}\n\ttoSave := string(entriesBytes)\n\tfileLock := flock.New(b.config.omwFile)\n\tlocked, err := fileLock.TryLock()\n\tdefer fileLock.Unlock()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"unable to get file lock\")\n\t}\n\tif !locked {\n\t\treturn errors.New(\"unable to get file lock\")\n\t}\n\t_, err = fp.WriteString(toSave)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"error saving new data\")\n\t}\n\treturn nil\n}", "func (p *Manager) Add(proxy Proxy) {\n\tp.Lock()\n\tdefer p.Unlock()\n\tif proxy.Type == \"\" {\n\t\tproxy.Type = DefaultType\n\t}\n\tp.List[p.WriteIndex] = proxy\n\tp.WriteIndex = len(p.List)\n}", "func (m *Manager) AddInterface(iface Interface) {\n\tm.interfaces = append(m.interfaces, iface)\n}", "func (r Runner) Add(job Job) {\n\tr.goJobs[job.Name] = job\n\n\tif job.Name == \"\" {\n\t\tpanic(fmt.Errorf(\"invalid job name %q\", job.Name))\n\t}\n\n\tr.jobRunProviders[job.Name] = func(st State) JobRun {\n\t\treturn func(ctx context.Context) error {\n\t\t\treturn job.Run(ctx, st)\n\t\t}\n\t}\n}", "func (self *Repository) Add(path string) error { return nil }", "func (self *Repository) Add(path string) error { return nil }", "func AddBuiltin(id string, fn func(args ...Object) Object) error {\n\tif builtins == nil {\n\t\tbuiltins = make(map[string]*Builtin)\n\t}\n\tif i := strings.Index(id, \"(\"); i >= 0 {\n\t\tname, params, err := parseFunction(id)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%s: %w\", id, err)\n\t\t}\n\t\torigFn := fn\n\t\tfn = func(args ...Object) Object {\n\t\t\tif err := checkArgs(params, args...); err != nil {\n\t\t\t\treturn Errorf(\"%w\", err)\n\t\t\t}\n\t\t\treturn origFn(args...)\n\t\t}\n\t\tid = name\n\t}\n\tif _, ok := builtins[id]; ok {\n\t\treturn fmt.Errorf(\"%s: %w\", id, ErrExists)\n\t}\n\tbuiltins[id] = &Builtin{Fn: fn}\n\treturn nil\n}", "func (p *ProfileCommand) Save() bool {\n\treturn p.saveInfo(p)\n}", "func AddSystem(name string, f func() System) {\n\tsystems[name] = f\n}", "func (b *Bucket) Save(key string, data []byte) (*brazier.Item, error) {\n\terr := b.node.Set(\"items\", key, data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &brazier.Item{\n\t\tKey: key,\n\t\tData: data,\n\t}, nil\n}", "func (c *Container) Save(i interface{}) cipher.SHA256 {\n\treturn c.db.AddAutoKey(encoder.Serialize(i))\n}", "func (p *Command) Save() bool {\n\treturn p.saveInfo(p)\n}", "func (s store) Save() {\n\ts.writeToDisk()\n}", "func saveGame(board *chess.Board, name string) {\n\tname = strings.Trim(name, \"\\n\")\n\tfile, _ := os.Create(name)\n\twriter := bufio.NewWriter(file)\n\tfmt.Println(\"Name: \", name)\n\tfor i, v := range board.History {\n\t\twriter.WriteString(strconv.Itoa(i+1) + \": \")\n\t\twriter.WriteString(v)\n\t\twriter.WriteString(\"\\n\")\n\t}\n\twriter.Flush()\n\tfile.Close()\n\n}", "func (c *Keystoreclient) Save(ctx context.Context, key string, message proto.Message) error {\n\tbytes, err := proto.Marshal(message)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = c.linker.Save(ctx, &pbd.SaveRequest{Key: key, Value: &google_protobuf.Any{Value: bytes}})\n\treturn err\n}", "func Save(willSave bool, popSaved uint, delSaved uint) {\n\tos.Exit(0)\n}" ]
[ "0.48275378", "0.481457", "0.4562074", "0.4471351", "0.44657955", "0.44657955", "0.44586483", "0.44475654", "0.44427806", "0.4442019", "0.44117406", "0.43987346", "0.439824", "0.43905815", "0.4367456", "0.4355819", "0.4354556", "0.4354556", "0.4354556", "0.43503666", "0.43464643", "0.43289828", "0.43183202", "0.4316226", "0.4291765", "0.42906785", "0.42899927", "0.4279614", "0.42766723", "0.4259018", "0.42525214", "0.42507285", "0.4250387", "0.42221433", "0.4220503", "0.42140955", "0.42135715", "0.42113262", "0.42066428", "0.42025173", "0.41965583", "0.41945305", "0.41938388", "0.41892597", "0.416967", "0.41677275", "0.41539818", "0.41500202", "0.41489804", "0.41484982", "0.41425815", "0.41390297", "0.4138875", "0.41323158", "0.41320565", "0.41270807", "0.41227812", "0.41197318", "0.41194904", "0.41080156", "0.41068625", "0.41049516", "0.4100449", "0.4087411", "0.408409", "0.40818417", "0.4077906", "0.40686622", "0.40643516", "0.4064233", "0.4057307", "0.4056068", "0.40498817", "0.40467015", "0.4044857", "0.40326807", "0.40303278", "0.40285665", "0.40152526", "0.4009556", "0.40092215", "0.40033823", "0.40001193", "0.39996827", "0.3991817", "0.39915168", "0.39872414", "0.39863193", "0.3982521", "0.3982521", "0.3977927", "0.39758524", "0.39736283", "0.3972706", "0.39708346", "0.39704323", "0.39658472", "0.3964171", "0.39625028", "0.39580086" ]
0.85015166
0
Save triggers a save to file, though respects a minimum save interval to wait between saves.
func (manager *basicStateManager) Save() error { defer metrics.MetricsEngineGlobal.RecordStateManagerMetric("SAVE")() manager.saveTimesLock.Lock() defer manager.saveTimesLock.Unlock() if time.Since(manager.lastSave) >= minSaveInterval { // we can just save err := manager.ForceSave() manager.lastSave = time.Now() manager.nextPlannedSave = time.Time{} // re-zero it; assume all pending desires to save are fulfilled return err } else if manager.nextPlannedSave.IsZero() { // No save planned yet, we should plan one. next := manager.lastSave.Add(minSaveInterval) manager.nextPlannedSave = next go func() { time.Sleep(time.Until(next)) manager.Save() }() } // else nextPlannedSave wasn't Zero so there's a save planned elsewhere that'll // fulfill this return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (p *plugin) periodicSave() {\n\tfor {\n\t\tselect {\n\t\tcase <-p.quit:\n\t\t\treturn\n\n\t\tcase <-time.After(SaveInterval):\n\t\t\tp.saveFile()\n\t\t}\n\t}\n}", "func (s *stepSaver) Save() error {\n\tvar wg sync.WaitGroup\n\tfor _, f := range s.LogFiles() {\n\t\twg.Add(1)\n\t\tgo func(f *logFile) {\n\t\t\tdefer wg.Done()\n\n\t\t\terr := f.Save()\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"ERROR saving logfile:\", err)\n\t\t\t}\n\t\t}(f)\n\t}\n\n\twg.Wait()\n\treturn nil\n}", "func (m *SynapsesPersist) Save() {\n\tif m.changed {\n\t\tfmt.Println(\"Saving synaptic data...\")\n\t\tindentedJSON, _ := json.MarshalIndent(m.Synapses, \"\", \" \")\n\n\t\tdataPath, err := filepath.Abs(m.relativePath)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\terr = ioutil.WriteFile(dataPath+m.file, indentedJSON, 0644)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(\"ERROR:\", err)\n\t\t}\n\n\t\tm.Clean()\n\t\tfmt.Println(\"Synaptic data saved\")\n\t}\n}", "func (s store) Save() {\n\ts.writeToDisk()\n}", "func (file *File) Save() {\n\tif file.autoFmt {\n\t\terr := file.Fmt()\n\t\tif err != nil {\n\t\t\tfile.NotifyUser(err.Error())\n\t\t}\n\t}\n\tfile.SnapshotSaved()\n\tcontents := []byte(file.ToString())\n\terr := ioutil.WriteFile(file.Name, contents, file.fileMode)\n\tif err != nil {\n\t\tfile.NotifyUser(\"Save Failed: \" + err.Error())\n\t} else {\n\t\tfile.savedBuffer.ReplaceBuffer(file.buffer.DeepDup())\n\t\tfile.NotifyUser(\"Saved.\")\n\t\tfile.modTime = time.Now()\n\t\tfile.md5sum = md5.Sum(contents)\n\t}\n}", "func (p *DefaultParser) Save(buf *bytes.Buffer, filename string) error {\n\terr := ioutil.WriteFile(filename, buf.Bytes(), 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (s *saver) Save(journey models.Journey) error {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.state == closed {\n\t\treturn ErrSaverIsClosed\n\t}\n\n\tif len(s.buffer) == cap(s.buffer) {\n\t\treturn ErrInternalBufferIsFull\n\t}\n\n\ts.buffer = append(s.buffer, journey)\n\treturn nil\n}", "func (file *File) RequestSave() {\n\tselect {\n\tcase file.saveChan <- struct{}{}:\n\tdefault:\n\t}\n}", "func (f *File) Save(fileName string, offset int64, xestatus string) error {\n\tif f.file == nil {\n\t\treturn errors.New(\"state file not open\")\n\t}\n\n\terr := writeState(f.file, fileName, offset, xestatus)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"writeStatus\")\n\t}\n\n\treturn nil\n}", "func (s *saver) Save(survey models.Survey) {\n\ts.save <- survey\n}", "func Save(willSave bool, popSaved uint, delSaved uint) {\n\tos.Exit(0)\n}", "func (app *service) Save(state State) error {\n\tjs, err := app.adapter.ToJSON(state)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tchainHash := state.Chain()\n\tindex := state.Height()\n\tpath := filePath(chainHash, index)\n\treturn app.fileService.Save(path, js)\n}", "func (ossDriver *OSSArtifactDriver) Save(path string, outputArtifact *wfv1.Artifact) error {\n\terr := wait.ExponentialBackoff(wait.Backoff{Duration: time.Second * 2, Factor: 2.0, Steps: 5, Jitter: 0.1},\n\t\tfunc() (bool, error) {\n\t\t\tlog.Infof(\"OSS Save path: %s, key: %s\", path, outputArtifact.OSS.Key)\n\t\t\tosscli, err := ossDriver.newOSSClient()\n\t\t\tif err != nil {\n\t\t\t\tlog.Warnf(\"Failed to create new OSS client: %v\", err)\n\t\t\t\treturn false, nil\n\t\t\t}\n\t\t\tbucketName := outputArtifact.OSS.Bucket\n\t\t\tbucket, err := osscli.Bucket(bucketName)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\tobjectName := outputArtifact.OSS.Key\n\t\t\terr = bucket.PutObjectFromFile(objectName, path)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\treturn true, nil\n\t\t})\n\treturn err\n}", "func (w *Worker) save() (n int, err error) {\n\tif w.position == 0 {\n\t\treturn 0, nil\n\t}\n\tif w.fileExists() {\n\t\tn, err = w.fileRoot.Write(w.buffer[0:w.position])\n\t\tif err == nil {\n\t\t\tw.position = 0\n\t\t}\n\t} else {\n\t\tw.errorCallback()\n\t}\n\treturn n, err\n}", "func (k *FileKeystore) Save() error {\n\tk.Lock()\n\terr := k.doSave(true)\n\tk.Unlock()\n\treturn err\n}", "func (l *LiveLog) Save(ctx context.Context) error {\n\tl.Lock()\n\tdefer l.Unlock()\n\treturn l.flush()\n}", "func (f *File) Save(path string) (err error) {\n\tdefer func() {\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"File.Save(%s): %w\", path, err)\n\t\t}\n\t}()\n\ttarget, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer func() {\n\t\tif ie := target.Close(); ie != nil {\n\t\t\terr = fmt.Errorf(\"write:%+v close:%w\", err, ie)\n\t\t}\n\t}()\n\terr = f.Write(target)\n\treturn\n}", "func (settings *Settings) Save(path string) error {\n\tcontent, err := json.Marshal(settings)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(path, content, 0644)\n}", "func (g *Generator) Save(w io.Writer) error {\n\tdata, err := json.MarshalIndent(g.spec, \"\", \"\\t\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t_, err = w.Write(data)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (buf *Buf) Save(file string) error {\n\tif file == \"\" {\n\t\treturn errors.New(\"No filename given\")\n\t}\n\n\tbs := []byte(buf.Text())\n\terr := ioutil.WriteFile(file, bs, 0644)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%s\", err)\n\t}\n\n\tbuf.Name = file\n\tbuf.ClearDirty()\n\treturn nil\n}", "func (spy *StoreSpy) Save(r *http.Request, w http.ResponseWriter, s *sessions.Session) error {\n\tspy.SaveCalls++\n\treturn nil\n}", "func (f *File) Save(name string) error {\n\treturn ioutil.WriteFile(name, []byte(f.String()), 0666)\n}", "func (r *CheckpointStore) Save(checkpoint core.Checkpoint) error {\n\tcheckpointSaveCallsMeter.Mark(1)\n\n\tif checkpoint.Region == r.config.LocalRegion {\n\t\treturn errors.New(\"cannot save checkpoint for local region\")\n\t}\n\n\tif !r.updateState(checkpoint) {\n\t\treturn nil\n\t}\n\n\tkey := r.formatMessageKey(checkpoint)\n\tvalue, err := r.marshalCheckpoint(checkpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn r.pubsub.Publish(r.config.Topic, key, value)\n}", "func (d *DiskStorage) Save() error {\n\n\tvar file, err = os.OpenFile(d.path, os.O_RDWR, 0644)\n\tif d.isError(err) {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\t_, errWrite := file.WriteString(d.String())\n\treturn errWrite\n}", "func (block *SimpleQueueBlock) Save(ctx context.Context, q *SimpleQueue) (err *mft.Error) {\n\tif q.MetaStorage == nil {\n\t\treturn nil\n\t}\n\tif !block.mxFileSave.TryLock(ctx) {\n\t\treturn GenerateError(10013000)\n\t}\n\tdefer block.mxFileSave.Unlock()\n\n\tif !block.mx.RTryLock(ctx) {\n\t\treturn GenerateError(10013001)\n\t}\n\n\tif block.SaveRv == block.ChangesRv {\n\t\tblock.mx.RUnlock()\n\t\treturn nil\n\t}\n\n\tchangesRv := block.ChangesRv\n\tdata, errMarshal := json.MarshalIndent(block.Data, \"\", \"\\t\")\n\tchLen := len(block.SaveWait)\n\n\tblock.mx.RUnlock()\n\n\tif errMarshal != nil {\n\t\treturn GenerateErrorE(10013002, errMarshal)\n\t}\n\n\tfileName := block.blockFileName()\n\n\tst, err := q.getStorageLock(ctx, block.Mark)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = st.Save(ctx, fileName, data)\n\tif err != nil {\n\t\treturn GenerateErrorE(10013003, err, fileName)\n\t}\n\n\tif !block.mx.TryLock(ctx) {\n\t\treturn GenerateError(10013004)\n\t}\n\tblock.SaveRv = changesRv\n\tif len(block.SaveWait) > 0 && chLen > 0 {\n\t\tsaveWait := make([]chan bool, 0)\n\t\tfor i := chLen; i < len(block.SaveWait); i++ {\n\t\t\tsaveWait = append(saveWait, block.SaveWait[i])\n\t\t}\n\t\tapprove := block.SaveWait[0:chLen]\n\n\t\tblock.SaveWait = saveWait\n\t\tgo func() {\n\t\t\tfor _, v := range approve {\n\t\t\t\tv <- true\n\t\t\t}\n\t\t}()\n\t}\n\n\tblock.mx.Unlock()\n\n\treturn nil\n}", "func (r *Reserve) Save(s *Store) error {\n\tdata, _ := json.Marshal(s)\n\tif err := ioutil.WriteFile(r.path, data, 0644); err != nil {\n\t\treturn fmt.Errorf(\"Failed to set %s: %s\", r.name, err)\n\t}\n\treturn nil\n}", "func (b *Bookmarks) Save() error {\n\treturn b.WriteToFile(b.Filename)\n}", "func Save() error {\n\treturn nil\n}", "func (scheduleAPI *scheduleAPIServer) saveScheduleWorker() {\n\tfor {\n\t\tselect {\n\t\tcase <-scheduleAPI.ctx.Done():\n\t\t\treturn\n\t\tcase <-time.After(time.Duration(5 * time.Minute)):\n\t\t\tfunc() {\n\t\t\t\tf, err := os.OpenFile(\"snapshot\", os.O_CREATE|os.O_WRONLY|os.O_TRUNC, 066)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Log.Warn(\"error while saving file\", zap.Error(err))\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t// Lock the mutex\n\t\t\t\tscheduleAPI.muSchedule.Lock()\n\n\t\t\t\tbs, err := proto.Marshal(&scheduleAPI.weeklySchedule)\n\t\t\t\tif err != nil {\n\t\t\t\t\tlogger.Log.Error(\"error while marshaling file\", zap.Error(err))\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t// Unlock the mutex\n\t\t\t\tscheduleAPI.muSchedule.Unlock()\n\n\t\t\t\t_, err = f.Write(bs)\n\t\t\t\tif err != nil {\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tlogger.Log.Error(\"error while writing to file\", zap.Error(err))\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t}()\n\t\t}\n\t}\n}", "func (app *service) Save(genesis Genesis) error {\n\tjs, err := app.adapter.ToJSON(genesis)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn app.fileService.Save(app.fileNameWithExt, js)\n}", "func savingFile(states AllStates, ID string) {\n\tfile, err := os.Create(\"elevator_states.txt\") //Creates file that will only contain latest data\n\t//checks for errors and saves to file as JSON\n\tcheck(err)\n\te := json.NewEncoder(file).Encode(states) //saves the AllStates struct to file\n\tcheck(e)\n}", "func (q *SimpleQueue) Save(ctx context.Context, user cn.CapUser) (err *mft.Error) {\n\tif q.MetaStorage == nil {\n\t\treturn nil\n\t}\n\tif !q.mxFileSave.TryLock(ctx) {\n\t\treturn GenerateError(10012000)\n\t}\n\tdefer q.mxFileSave.Unlock()\n\n\tif !q.mx.RTryLock(ctx) {\n\t\treturn GenerateError(10012001)\n\t}\n\n\tif q.SaveRv == q.ChangesRv {\n\t\tq.mx.RUnlock()\n\t\treturn nil\n\t}\n\n\tchangesRv := q.ChangesRv\n\tdata, errMarshal := json.MarshalIndent(q, \"\", \"\\t\")\n\tchLen := len(q.SaveWait)\n\n\tq.mx.RUnlock()\n\n\tif errMarshal != nil {\n\t\treturn GenerateErrorE(10012002, errMarshal)\n\t}\n\n\terr = q.MetaStorage.Save(ctx, MetaDataFileName, data)\n\tif err != nil {\n\t\treturn GenerateErrorE(10012003, err, MetaDataFileName)\n\t}\n\n\tif !q.mx.TryLock(ctx) {\n\t\treturn GenerateError(10012004)\n\t}\n\tq.SaveRv = changesRv\n\tif len(q.SaveWait) > 0 && chLen > 0 {\n\t\tsaveWait := make([]chan bool, 0)\n\t\tfor i := chLen; i < len(q.SaveWait); i++ {\n\t\t\tsaveWait = append(saveWait, q.SaveWait[i])\n\t\t}\n\t\tapprove := q.SaveWait[0:chLen]\n\n\t\tq.SaveWait = saveWait\n\t\tgo func() {\n\t\t\tfor _, v := range approve {\n\t\t\t\tv <- true\n\t\t\t}\n\t\t}()\n\t}\n\n\tq.mx.Unlock()\n\n\treturn nil\n}", "func (s *TreeSaver) Save(ctx context.Context, snPath string, node *restic.Node, nodes []FutureNode) FutureTree {\n\tch := make(chan saveTreeResponse, 1)\n\tjob := saveTreeJob{\n\t\tsnPath: snPath,\n\t\tnode: node,\n\t\tnodes: nodes,\n\t\tch: ch,\n\t}\n\tselect {\n\tcase s.ch <- job:\n\tcase <-ctx.Done():\n\t\tdebug.Log(\"not saving tree, context is cancelled\")\n\t\tclose(ch)\n\t\treturn FutureTree{ch: ch}\n\t}\n\n\treturn FutureTree{ch: ch}\n}", "func Save() {\n\tgo db.save()\n}", "func (fb *FileBackend) save(state *storage.State) error {\n\tout, err := proto.Marshal(state)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to encode state: %w\", err)\n\t}\n\ttmp := fmt.Sprintf(fb.path+\".%v\", time.Now())\n\tif err := ioutil.WriteFile(tmp, out, 0600); err != nil {\n\t\treturn fmt.Errorf(\"failed to write state: %w\", err)\n\t}\n\terr = os.Rename(tmp, fb.path)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to move state: %w\", err)\n\t}\n\treturn nil\n}", "func (r *RepoGob) Save(s *models.Counter) error {\n\n\tif s == nil {\n\t\treturn errors.New(\"counter struct not specified\")\n\t}\n\n\tbuf := new(bytes.Buffer)\n\tenc := gob.NewEncoder(buf)\n\terr := enc.Encode(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = ioutil.WriteFile(r.path, buf.Bytes(), 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Save(obj any, file string) error {\n\tfp, err := os.Create(file)\n\tdefer fp.Close()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\tbw := bufio.NewWriter(fp)\n\terr = Write(obj, bw)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\terr = bw.Flush()\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\treturn err\n}", "func (s *Script) Save(fn string) {\n\n\t// fmt.Println(s.String())\n\n\tBOM := \"\\uFEFF\"\n\tf, err := os.Create(fn)\n\tif err != nil {\n\t\tpanic(fmt.Errorf(\"writer: failed saving subtitle file: %s\", err))\n\t}\n\tdefer f.Close()\n\n\ts.MetaFilename = fn\n\n\tn, err := f.WriteString(BOM + s.String())\n\tif err != nil {\n\t\tfmt.Println(n, err)\n\t}\n\n\t// save changes\n\t// err =\n\tf.Sync()\n}", "func (tb *TextBuf) AutoSave() error {\n\tif tb.HasFlag(int(TextBufAutoSaving)) {\n\t\treturn nil\n\t}\n\ttb.SetFlag(int(TextBufAutoSaving))\n\tasfn := tb.AutoSaveFilename()\n\tb := tb.LinesToBytesCopy()\n\terr := ioutil.WriteFile(asfn, b, 0644)\n\tif err != nil {\n\t\tlog.Printf(\"giv.TextBuf: Could not AutoSave file: %v, error: %v\\n\", asfn, err)\n\t}\n\ttb.ClearFlag(int(TextBufAutoSaving))\n\treturn err\n}", "func (cfg *Configuration) Save() error {\n\tcfg.locker.Lock()\n\tdefer cfg.locker.Unlock()\n\tif cfg.FilePath == \"\" {\n\t\treturn errors.New(\"Configuration.FilePath was not set\")\n\t}\n\treturn gonfig.Write(cfg, true)\n}", "func (self *Repository) Save(message string) error {\n\t// TODO: Validate message, maybe even use a ruleset file so a project can\n\t// define rules for commits\n\t// Adds occur automatically since we are watching all files for changes always\n\t// to rebuild and such sop we just need to commit\n\treturn nil\n}", "func (self *Repository) Save(message string) error {\n\t// TODO: Validate message, maybe even use a ruleset file so a project can\n\t// define rules for commits\n\t// Adds occur automatically since we are watching all files for changes always\n\t// to rebuild and such sop we just need to commit\n\treturn nil\n}", "func Save() {\n\tdata := Savedata{\n\t\tName: GS.current.name,\n\t\tGamestate: GS.current,\n\t}\n\n\tf, err := json.MarshalIndent(data, \"\", \" \")\n\tcheck(err)\n\tioutil.WriteFile(\"data/savegame.json\", f, 0644)\n}", "func (c *ConfigurationFile) Save() error {\n\tcontent, err := json.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(c.location.get(), content, 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func Save(filename string, getBytes func() (string, int)) (msg string, stillDirty bool) {\n\tfp, e := os.OpenFile(filename, os.O_RDWR|os.O_CREATE|os.O_TRUNC, 0644)\n\tif e != nil {\n\t\treturn fmt.Sprintf(\"Can't save! file open error %s\", e), true\n\t}\n\tdefer fp.Close()\n\tbuf, length := getBytes()\n\tstillDirty = true\n\tn, err := io.WriteString(fp, buf)\n\tif err == nil {\n\t\tif n == length {\n\t\t\tstillDirty = false\n\t\t\tmsg = fmt.Sprintf(\"%d bytes written to disk\", length)\n\t\t} else {\n\t\t\tmsg = fmt.Sprintf(\"wanted to write %d bytes to file, wrote %d\", length, n)\n\t\t}\n\t\treturn msg, stillDirty\n\t}\n\treturn fmt.Sprintf(\"Can't save! I/O error %s\", err), stillDirty\n}", "func (f *File) Save() error {\n\tif f.Id == 0 {\n\t\treturn dbAccess.Insert(f)\n\t} else {\n\t\t_, err := dbAccess.Update(f)\n\t\treturn err\n\t}\n}", "func (connection *Connection) Save(path string) error {\n\tjson, err := json.Marshal(connection)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(path, json, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (a *App) save() {\n\ta.TodoStore.Save(a.TodoList.Data)\n\tif a.TodoList.IsSynced {\n\t\ta.EventLogger.ProcessEvents()\n\n\t\tsynchronizer := NewQuietSynchronizer()\n\t\tsynchronizer.ExecSyncInBackground()\n\t}\n}", "func (wj *WsubJob) Save(db XODB) error {\n\tif wj.Exists() {\n\t\treturn wj.Update(db)\n\t}\n\n\treturn wj.Insert(db)\n}", "func (bck *FileBackup) Save() error {\n\tconfig.Logger.Info(\"Writing database to file\")\n\tstart := time.Now()\n\n\tsize, err := bck.writeFile()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to save database\")\n\t}\n\n\tconfig.Logger.Info(\"Wrote database\", zap.Int(\"size (bytes)\", size), zap.Int(\"hashes\", bck.db.Hashes()), zap.Duration(\"duration\", time.Since(start)))\n\n\treturn nil\n}", "func (p *process) startSaveFile(path string, body string, opt *Options) error {\n\n\terr := ioutil.WriteFile(path, []byte(body), 0666)\n\tif err != nil {\n\t\treturn err\n\t}\n\tp.run = nil\n\treturn nil\n}", "func (d *dataUpdateTracker) startSaver(ctx context.Context, interval time.Duration, drives []string) {\n\tsaveNow := d.save\n\texited := make(chan struct{})\n\td.saveExited = exited\n\td.mu.Unlock()\n\tt := time.NewTicker(interval)\n\tdefer t.Stop()\n\tdefer close(exited)\n\tvar buf bytes.Buffer\n\tfor {\n\t\tvar exit bool\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\texit = true\n\t\tcase <-t.C:\n\t\tcase <-saveNow:\n\t\t}\n\t\tbuf.Reset()\n\t\td.mu.Lock()\n\t\tif !d.dirty {\n\t\t\td.mu.Unlock()\n\t\t\tif exit {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\td.Saved = UTCNow()\n\t\terr := d.serialize(&buf)\n\t\tif d.debug {\n\t\t\tconsole.Debugf(color.Green(\"dataUpdateTracker:\")+\" Saving: %v bytes, Current idx: %v\\n\", buf.Len(), d.Current.idx)\n\t\t}\n\t\td.dirty = false\n\t\td.mu.Unlock()\n\t\tif err != nil {\n\t\t\tlogger.LogIf(ctx, err, \"Error serializing usage tracker data\")\n\t\t\tif exit {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tif buf.Len() == 0 {\n\t\t\tlogger.LogIf(ctx, errors.New(\"zero sized output, skipping save\"))\n\t\t\tcontinue\n\t\t}\n\t\tfor _, drive := range drives {\n\t\t\tcacheFormatPath := pathJoin(drive, dataUpdateTrackerFilename)\n\t\t\terr := ioutil.WriteFile(cacheFormatPath, buf.Bytes(), os.ModePerm)\n\t\t\tif err != nil {\n\t\t\t\tif osIsNotExist(err) {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tlogger.LogIf(ctx, err)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t}\n\t\tif exit {\n\t\t\treturn\n\t\t}\n\t}\n}", "func (n *mockAgent) save() (s persistapi.AgentState) {\n\treturn\n}", "func (s *ServiceState) save() {\n\tlog.Lvl3(\"Saving service\")\n\tb, err := network.Marshal(s.Storage)\n\tif err != nil {\n\t\tlog.Error(\"Couldn't marshal service:\", err)\n\t} else {\n\t\terr = ioutil.WriteFile(s.path+\"/prifi.bin\", b, 0660)\n\t\tif err != nil {\n\t\t\tlog.Error(\"Couldn't save file:\", err)\n\t\t}\n\t}\n}", "func (tb *TextBuf) Save() error {\n\tif tb.Filename == \"\" {\n\t\treturn fmt.Errorf(\"giv.TextBuf: filename is empty for Save\")\n\t}\n\ttb.EditDone()\n\tinfo, err := os.Stat(string(tb.Filename))\n\tif err == nil && info.ModTime() != time.Time(tb.Info.ModTime) {\n\t\tvp := tb.ViewportFromView()\n\t\tgi.ChoiceDialog(vp, gi.DlgOpts{Title: \"File Changed on Disk\",\n\t\t\tPrompt: fmt.Sprintf(\"File has changed on disk since being opened or saved by you -- what do you want to do? File: %v\", tb.Filename)},\n\t\t\t[]string{\"Save To Different File\", \"Open From Disk, Losing Changes\", \"Save File, Overwriting\"},\n\t\t\ttb.This(), func(recv, send ki.Ki, sig int64, data interface{}) {\n\t\t\t\tswitch sig {\n\t\t\t\tcase 0:\n\t\t\t\t\tCallMethod(tb, \"SaveAs\", vp)\n\t\t\t\tcase 1:\n\t\t\t\t\ttb.Revert()\n\t\t\t\tcase 2:\n\t\t\t\t\ttb.SaveFile(tb.Filename)\n\t\t\t\t}\n\t\t\t})\n\t}\n\treturn tb.SaveFile(tb.Filename)\n}", "func (rcv *Storage) Save(data []byte, filename string, strategy string, offset int) {\n\tlog.Printf(\"Save file %s with strategy %s and offset %d\", filename, strategy, offset)\n\tchunks := rcv.splitter.Split(data, filename, strategy, offset)\n\n\tfor _, c := range chunks {\n\t\trcv.saveChunk(c)\n\t}\n\t//TODO add logic to handle table\n}", "func (p *SimplePlugin) Save() ([]byte, error) {\n\tif p.SaveFunc != nil {\n\t\treturn p.SaveFunc()\n\t}\n\treturn nil, nil\n}", "func (j Journal) Save() error {\n\t// marshal journal to JSON\n\tdata, err := json.Marshal(j)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// truncate and rewrite to the journal file\n\tjournalFilePath, err := getJournalPath(j.Name)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tjournalFile, err := os.OpenFile(journalFilePath, os.O_TRUNC|os.O_CREATE|os.O_RDWR, 0660)\n\tdefer journalFile.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif _, err := journalFile.Write(data); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *settings) saveToDisk() error {\n\tb, err := json.Marshal(m.redirects)\n\tif err != nil {\n\t\tlog.Printf(\"Error marshalling %s\", err)\n\t\treturn fmt.Errorf(\"error marshalling %s\", err)\n\t}\n\n\tif err := ioutil.WriteFile(m.filename, b, 0644); err != nil {\n\t\treturn fmt.Errorf(\"unable to open file %s\", err)\n\t}\n\tlog.Printf(\"saving to disk.\")\n\treturn nil\n}", "func Save(path string, v interface{}) error {\n\tformatter, err := parsePath(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfile, err := os.OpenFile(path, os.O_TRUNC|os.O_CREATE|os.O_WRONLY, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\treturn formatter.Encode(file, v)\n}", "func Save() {\n\tc := Config{viper.GetString(\"email\"), viper.GetString(\"platform\"), viper.GetDuration(\"timeout\")}\n\tdata, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn\n\t}\n\n\t_ = ioutil.WriteFile(viper.GetString(\"file_config\"), data, 0600)\n}", "func (c *Config) Save(filename string) (err error) {\n\tlog.Println(\"[DEBUG] Save\", filename)\n\n\tbody, err := json.MarshalIndent(c, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = c.writeFile(filename, body)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *TreeSaver) Save(ctx context.Context, snPath string, target string, node *restic.Node, nodes []FutureNode, complete CompleteFunc) FutureNode {\n\tfn, ch := newFutureNode()\n\tjob := saveTreeJob{\n\t\tsnPath: snPath,\n\t\ttarget: target,\n\t\tnode: node,\n\t\tnodes: nodes,\n\t\tch: ch,\n\t\tcomplete: complete,\n\t}\n\tselect {\n\tcase s.ch <- job:\n\tcase <-ctx.Done():\n\t\tdebug.Log(\"not saving tree, context is cancelled\")\n\t\tclose(ch)\n\t}\n\n\treturn fn\n}", "func (ifile *Indexfile) Save(lfindex *Index) error {\n\tifile.Open(CREATE | WRITE_ONLY | APPEND)\n\tdefer ifile.Close()\n\t_, err := ifile.LockedWriteAt(lfindex.ToBytes(), 0)\n\treturn err\n}", "func (c *Config) Save(file *os.File) error {\n\tif file == nil && c.file != nil {\n\t\tfile = c.file\n\t}\n\n\tif err := file.Truncate(0); err != nil {\n\t\treturn err\n\t}\n\tif _, err := file.Seek(0, 0); err != nil {\n\t\treturn err\n\t}\n\tif err := yaml.NewEncoder(file).Encode(c); err != nil {\n\t\treturn err\n\t}\n\n\treturn file.Sync()\n}", "func (p *plugin) saveFile() {\n\tp.m.RLock()\n\terr := util.WriteFile(p.file, p.users, true)\n\tp.m.RUnlock()\n\n\tif err != nil {\n\t\tlog.Println(\"[stats] save:\", err)\n\t}\n}", "func Save() {\n\tenforcer.SavePolicy()\n}", "func (config *Config) Save(file string) error {\n\tbts, err := json.Marshal(*config)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar out bytes.Buffer\n\tjson.Indent(&out, bts, \"\", \"\\t\")\n\n\treturn ioutil.WriteFile(file, out.Bytes(), 0600)\n}", "func (f *FilePersist) Save(ctx context.Context, data map[string]string) error {\n\tfr, err := os.OpenFile(f.filename, os.O_WRONLY, 0644)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to open file for persist: %w\", err)\n\t}\n\tdefer fr.Close()\n\tif err := json.NewEncoder(fr).Encode(data); err != nil {\n\t\treturn fmt.Errorf(\"unable to encode: %w\", err)\n\t}\n\treturn nil\n}", "func saveWorker(output Output, nodes *runtime.Nodes, saveInterval time.Duration) {\n\tticker := time.NewTicker(saveInterval)\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\toutput.Save(nodes)\n\t\tcase <-quit:\n\t\t\twg.Done()\n\t\t\tticker.Stop()\n\t\t\treturn\n\t\t}\n\t}\n}", "func (b *BlockCreator) save() error {\n\treturn persist.SaveJSON(settingsMetadata, b.persist, filepath.Join(b.persistDir, settingsFile))\n}", "func (c *Configfile) Save(filename string) error {\n\tif filename == \"\" {\n\t\tfilename = c.filename\n\t}\n\tb, err := yaml.Marshal(c)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn ioutil.WriteFile(filename, b, 0755)\n}", "func (c Configuration) Save() {\n\tbuf := new(bytes.Buffer)\n\tif err := toml.NewEncoder(buf).Encode(c); err != nil {\n\t\tlog.Fatalln(\"Failed to encode config\", err)\n\t}\n\tf, err := os.Create(configFile)\n\tif err != nil {\n\t\tlog.Fatalln(\"Failed to create file\", err)\n\t\treturn\n\t}\n\n\tw := bufio.NewWriter(f)\n\tbuf.WriteTo(w)\n\tw.Flush()\n}", "func (c *Keystoreclient) HardSave(ctx context.Context, key string, message proto.Message) error {\n\tfor i := 0; i < c.retries; i++ {\n\t\terr := c.Save(ctx, key, message)\n\t\tif err == nil {\n\t\t\treturn err\n\t\t}\n\n\t\ttime.Sleep(c.backoffTime / time.Duration(c.retries))\n\t}\n\n\treturn errors.New(\"Unable to perform hard save\")\n}", "func (xsml *XfileServiceMetricLog) Save(db XODB) error {\n\tif xsml.Exists() {\n\t\treturn xsml.Update(db)\n\t}\n\n\treturn xsml.Replace(db)\n}", "func Save(fromVar interface{}, toPath string) {\n\tdata, e := json.Marshal(fromVar)\n\terr.Panic(e)\n\te = ioutil.WriteFile(toPath, data, 0664)\n\terr.Panic(e)\n}", "func (c *Chain) Save() bool {\n\tc.mux.Lock()\n\tdefer c.mux.Unlock()\n\tbytes := parser.ParseToJSON(c.Chain)\n\te := saveToHDD(c.Path, bytes)\n\tif e != nil {\n\t\tpanic(e)\n\t}\n\treturn true\n}", "func (s *Store) Save() error {\n\tbk, err := os.OpenFile(filepath.Join(s.rwDirPath, storeBkName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer bk.Close()\n\n\tdst, err := os.OpenFile(filepath.Join(s.rwDirPath, storeName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer dst.Close()\n\n\t// backing up current store\n\t_, err = io.Copy(bk, dst)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\n\tenc := gob.NewEncoder(dst)\n\tbook := s.Clone()\n\terr = enc.Encode(book)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\treturn nil\n}", "func (s *StressReport) Save(filename string, name string) error {\n\tvar reports StressReports\n\tif err := reports.Load(filename); err != nil {\n\t\treports = map[string]*StressReport{\n\t\t\tname: s,\n\t\t}\n\t} else {\n\t\treports[name] = s\n\t}\n\n\tfmt.Printf(\"Writing reports in %s\\n\", filename)\n\n\tj, _ := json.Marshal(reports)\n\treturn os.WriteFile(filename, j, 0644)\n}", "func saveFile(conn *contact.Connection, file string, wg *sync.WaitGroup) {\n\tdefer wg.Done()\n\tfullPath := syncDir + file\n\n\t// Create the file and any parent directories\n\tfilePtr, err := synth.CreateFile(fullPath)\n\tif err != nil {\n\t\tlog.Printf(\"error creating file %s\\n\", fullPath)\n\t\treturn\n\t}\n\tdefer filePtr.Close()\n\n\tbuffer := bufio.NewWriter(filePtr)\n\tfor {\n\t\t// Get the next file block from the server\n\t\t_, data, err := conn.Read()\n\t\tif err != nil {\n\t\t\tlog.Printf(\"error reading file %s contents from connection\\n\", fullPath)\n\t\t\treturn\n\t\t}\n\n\t\t// An empty message indicates the end of the file\n\t\tif len(data) == 0 {\n\t\t\tbreak\n\t\t}\n\n\t\t// Remove any trailing NUL bytes\n\t\tdata = bytes.TrimRight(data, \"\\x00\")\n\n\t\t// Write the block to disk\n\t\t_, err = buffer.Write(data)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"error writing file %s contents to disk\\n\", fullPath)\n\t\t\treturn\n\t\t}\n\t}\n\n\tbuffer.Flush()\n\tfmt.Printf(\"File %s saved to disk.\\n\", file)\n}", "func Save(path string, v interface{}, opts Options) (err error) {\n\tvar (\n\t\tfile *os.File\n\t\ttmp = path + \".tmp.\" + cmn.GenTie()\n\t)\n\tif file, err = cmn.CreateFile(tmp); err != nil {\n\t\treturn\n\t}\n\tdefer func() {\n\t\t_ = file.Close()\n\t\tif err != nil {\n\t\t\tos.Remove(tmp)\n\t\t}\n\t}()\n\tif err = Encode(file, v, opts); err != nil {\n\t\treturn\n\t}\n\tif err = file.Close(); err != nil {\n\t\treturn\n\t}\n\terr = os.Rename(tmp, path)\n\treturn\n}", "func (ws *WalletStore) Save() {\n\tvar buffer bytes.Buffer\n\tgob.Register(elliptic.P256())\n\tencoder := gob.NewEncoder(&buffer)\n\terr := encoder.Encode(ws.Wallets)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfile := ws.Config.GetWalletStoreFile(ws.NodeID)\n\terr = ioutil.WriteFile(file, buffer.Bytes(), 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (cfg *Config) Save(fpath string) error {\n\t// create a copy\n\tconfig2 := cfg\n\t// clear, setup setting\n\tconfig2.Password = \"\"\n\tconfig2.Iterations = ConfigHashIterations\n\n\t// save to file\n\tbyteDat2, err := json.MarshalIndent(config2, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(fpath, byteDat2, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *MockConfig) SaveTimeout(timeout time.Duration) error {\n\targs := m.Called(timeout)\n\treturn args.Error(0)\n}", "func (c *Command) save() {\n\tif len(c.Args) == 2 {\n\t\tioutil.WriteFile(c.Args[0], []byte(c.Args[1]), os.ModePerm)\n\t}\n\tc.done()\n}", "func save(bytes []byte, fh *os.File) error {\n\t_, err := fh.Write(bytes)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = fh.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (d *Dump) Save() error {\n\td.mutex.RLock()\n\tdefer d.mutex.RUnlock()\n\n\treturn d.save()\n}", "func (q *Queue) SaveToDisk() {\n\tvar jobfile JobFile\n\tif q.HasOnDisk() {\n\t\tq.PopOnDiskJobs(&jobfile)\n\t}\n\tfor {\n\t\tselect {\n\t\tcase job := <-q.entries:\n\t\t\tjobfile.Jobs = append(jobfile.Jobs, job)\n\t\tdefault:\n\t\t\tif len(jobfile.Jobs) < 1 {\n\t\t\t\treturn\n\t\t\t}\n\t\t\tw, err := os.OpenFile(q.ondiskfile, os.O_CREATE|os.O_RDWR, 0666)\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\tw.Write(jobfile.Jobs.Serialize())\n\t\t\tq.rw.Lock()\n\t\t\tdefer q.rw.Unlock()\n\t\t\tq.ondisk = true\n\t\t\treturn\n\t\t}\n\t}\n}", "func (b *bookMark) save(event win_eventlog.EvtHandle) error {\n\tnewBookmark, err := win_eventlog.UpdateBookmark(b.handle, event, b.buf)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := b.file.Truncate(0); err != nil {\n\t\treturn err\n\t}\n\tif _, err := b.file.Seek(0, 0); err != nil {\n\t\treturn err\n\t}\n\t_, err = b.file.WriteString(newBookmark)\n\treturn err\n}", "func (ss *SuffixSnapshot) Save(key string, value string, ts typeutil.Timestamp) error {\n\t// if ts == 0, act like MetaKv\n\t// will not update lastestTs since ts not not valid\n\tif ts == 0 {\n\t\treturn ss.MetaKv.Save(key, value)\n\t}\n\n\tss.Lock()\n\tdefer ss.Unlock()\n\n\ttsKey := ss.composeTSKey(key, ts)\n\n\t// provided key value is latest\n\t// stores both tsKey and original key\n\tafter, err := ss.checkKeyTS(key, ts)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif after {\n\t\terr := ss.MetaKv.MultiSave(map[string]string{\n\t\t\tkey: value,\n\t\t\ttsKey: value,\n\t\t})\n\t\t// update latestTS only when MultiSave succeeds\n\t\tif err == nil {\n\t\t\tss.lastestTS[key] = ts\n\t\t}\n\t\treturn err\n\t}\n\n\t// modifying history key, just save tskey-value\n\treturn ss.MetaKv.Save(tsKey, value)\n}", "func (ts *FileTokenStorage) Save(token Token) error {\n\tencodedToken := base64.StdEncoding.EncodeToString([]byte(token.tokenString))\n\tts.token = encodedToken\n\treturn ioutil.WriteFile(ts.tokenFileName, []byte(encodedToken), 0644)\n}", "func (m *Machine) Save() error {\n\tm.State = driver.Saved\n\tfmt.Printf(\"Save %s: %s\\n\", m.Name, m.State)\n\treturn nil\n}", "func (g *Generator) SaveToFile(path string) error {\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\treturn g.Save(f)\n}", "func (c *Config) Save(cfgPath string) error {\n\tcfgFile, err := yaml.Marshal(c)\n\tif err == nil {\n\t\terr = ioutil.WriteFile(cfgPath, cfgFile, 0600)\n\t}\n\treturn err\n}", "func (s *LocalSnapStore) Save(snap brtypes.Snapshot, rc io.ReadCloser) error {\n\tdefer rc.Close()\n\terr := os.MkdirAll(path.Join(s.prefix, snap.SnapDir), 0700)\n\tif err != nil && !os.IsExist(err) {\n\t\treturn err\n\t}\n\tf, err := os.Create(path.Join(s.prefix, snap.SnapDir, snap.SnapName))\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\t_, err = io.Copy(f, rc)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn f.Sync()\n}", "func (j *Job) Save(ctx context.Context) (err error) {\n\tt := utils.FromTaskContext(ctx)\n\n\tcontent, err := msgpack.Marshal(j)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = contexts.DB.Put(constants.FormatJobKey(t, j.Path), content, nil)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (fcb *FileCacheBackend) Save(ob types.OutgoingBatch) error {\n\tfilename := fcb.getFilename(fcb.getCacheFilename(ob))\n\tlog.Printf(\"Saving to %s\", filename)\n\tfile, err := os.OpenFile(filename, os.O_CREATE|os.O_WRONLY|os.O_EXCL, 0600)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to save to %s - %s\", filename, err)\n\t}\n\tdefer file.Close()\n\tfor _, item := range ob.Values {\n\t\tfile.WriteString(item + \"\\n\")\n\t}\n\treturn nil\n}", "func (ctl *taskController) Save() error {\n\treturn ctl.saveImpl(true)\n}", "func (us *UpdateService) save() error {\n\tus.Updated = time.Now()\n\tcontent, _ := json.Marshal(us)\n\tkey := fmt.Sprintf(\"%s/%s/%s/%s/%s\", us.Proto, us.Version, us.Namespace, us.Repository, defaultMetaFileName)\n\terr := us.store.Put(key, content)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif us.km != nil {\n\t\t// write sign file, don't popup error even fail to saveSign\n\t\tus.saveSign(content)\n\t}\n\n\treturn nil\n}", "func (a *Network) Save(path string) {\n\tioutil.WriteFile(path, []byte(a.outputFormat()), 0666)\n}" ]
[ "0.70738846", "0.6883251", "0.65653616", "0.650868", "0.64468044", "0.6215792", "0.62057865", "0.6168609", "0.61583966", "0.61407447", "0.6125752", "0.6065374", "0.605968", "0.6021444", "0.60167485", "0.5978794", "0.59745365", "0.5957318", "0.5943211", "0.59240586", "0.5922223", "0.59110206", "0.59072614", "0.5901446", "0.5898372", "0.58689773", "0.58634335", "0.5861888", "0.58495677", "0.58404386", "0.583813", "0.5830092", "0.58268225", "0.5826267", "0.58236", "0.5810628", "0.58092856", "0.5807059", "0.5800384", "0.57809776", "0.5773616", "0.5773616", "0.5753383", "0.5752569", "0.57476795", "0.57419735", "0.5739191", "0.5738162", "0.5735322", "0.5732438", "0.5720978", "0.5718196", "0.5703609", "0.57017195", "0.5700982", "0.56947553", "0.56806576", "0.5679499", "0.56755555", "0.56674993", "0.5650703", "0.56489456", "0.5644274", "0.5625602", "0.56203294", "0.56188744", "0.5618611", "0.56169033", "0.5608304", "0.56005305", "0.55825764", "0.55781215", "0.5575675", "0.5568057", "0.5563", "0.5552835", "0.55524635", "0.5545495", "0.5542721", "0.5539532", "0.5538093", "0.5532025", "0.55316156", "0.5527616", "0.5523692", "0.55079114", "0.5503022", "0.55011904", "0.5496777", "0.54948694", "0.54913455", "0.54892516", "0.5480088", "0.5478308", "0.54768884", "0.5474091", "0.54740846", "0.547058", "0.54695344", "0.54691106" ]
0.6801875
2
ForceSave saves the given State to a file. It is an atomic operation on POSIX systems (by Renaming over the target file). This function logs errors at will and does not necessarily expect the caller to handle the error because there's little a caller can do in general other than just keep going. In addition, the StateManager internally buffers save requests in order to only save at most every STATE_SAVE_INTERVAL.
func (manager *basicStateManager) ForceSave() error { manager.savingLock.Lock() defer manager.savingLock.Unlock() seelog.Info("Saving state!") s := manager.state s.Version = ECSDataVersion data, err := json.Marshal(s) if err != nil { seelog.Error("Error saving state; could not marshal data; this is odd", "err", err) return err } return manager.writeFile(data) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (f *File) Save(fileName string, offset int64, xestatus string) error {\n\tif f.file == nil {\n\t\treturn errors.New(\"state file not open\")\n\t}\n\n\terr := writeState(f.file, fileName, offset, xestatus)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"writeStatus\")\n\t}\n\n\treturn nil\n}", "func (manager *basicStateManager) Save() error {\n\tdefer metrics.MetricsEngineGlobal.RecordStateManagerMetric(\"SAVE\")()\n\tmanager.saveTimesLock.Lock()\n\tdefer manager.saveTimesLock.Unlock()\n\tif time.Since(manager.lastSave) >= minSaveInterval {\n\t\t// we can just save\n\t\terr := manager.ForceSave()\n\t\tmanager.lastSave = time.Now()\n\t\tmanager.nextPlannedSave = time.Time{} // re-zero it; assume all pending desires to save are fulfilled\n\t\treturn err\n\t} else if manager.nextPlannedSave.IsZero() {\n\t\t// No save planned yet, we should plan one.\n\t\tnext := manager.lastSave.Add(minSaveInterval)\n\t\tmanager.nextPlannedSave = next\n\t\tgo func() {\n\t\t\ttime.Sleep(time.Until(next))\n\t\t\tmanager.Save()\n\t\t}()\n\t}\n\t// else nextPlannedSave wasn't Zero so there's a save planned elsewhere that'll\n\t// fulfill this\n\treturn nil\n}", "func (fb *FileBackend) save(state *storage.State) error {\n\tout, err := proto.Marshal(state)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to encode state: %w\", err)\n\t}\n\ttmp := fmt.Sprintf(fb.path+\".%v\", time.Now())\n\tif err := ioutil.WriteFile(tmp, out, 0600); err != nil {\n\t\treturn fmt.Errorf(\"failed to write state: %w\", err)\n\t}\n\terr = os.Rename(tmp, fb.path)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to move state: %w\", err)\n\t}\n\treturn nil\n}", "func (s *stepSaver) Save() error {\n\tvar wg sync.WaitGroup\n\tfor _, f := range s.LogFiles() {\n\t\twg.Add(1)\n\t\tgo func(f *logFile) {\n\t\t\tdefer wg.Done()\n\n\t\t\terr := f.Save()\n\t\t\tif err != nil {\n\t\t\t\tlog.Println(\"ERROR saving logfile:\", err)\n\t\t\t}\n\t\t}(f)\n\t}\n\n\twg.Wait()\n\treturn nil\n}", "func (m *stateManager) Save() (err error) {\n\terr = m.storage.Write(m.state)\n\n\tif err == nil {\n\t\tm.stateChanged = false\n\t\tm.stateLoaded = true\n\t}\n\n\treturn\n}", "func (s *State) Save() error {\n\tf, err := os.Create(getPathForUsername(s.Username))\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\treturn json.NewEncoder(f).Encode(s)\n}", "func (g *Gossiper) SaveState() {\n\tobj, e := json.MarshalIndent(g, \"\", \"\\t\")\n\tutils.HandleError(e)\n\t_ = os.Mkdir(utils.STATE_FOLDER, os.ModePerm)\n\tcwd, _ := os.Getwd()\n\te = ioutil.WriteFile(filepath.Join(cwd, utils.STATE_FOLDER, fmt.Sprint(g.Name, \".json\")), obj, 0644)\n\tutils.HandleError(e)\n}", "func (app *service) Save(state State) error {\n\tjs, err := app.adapter.ToJSON(state)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tchainHash := state.Chain()\n\tindex := state.Height()\n\tpath := filePath(chainHash, index)\n\treturn app.fileService.Save(path, js)\n}", "func (k *FileKeystore) doSave(override bool) error {\n\tif k.dirty == false {\n\t\treturn nil\n\t}\n\n\ttemporaryPath := fmt.Sprintf(\"%s.tmp\", k.Path)\n\n\tw := new(bytes.Buffer)\n\tjsonEncoder := json.NewEncoder(w)\n\tif err := jsonEncoder.Encode(k.secrets); err != nil {\n\t\treturn fmt.Errorf(\"cannot serialize the keystore before saving it to disk: %v\", err)\n\t}\n\n\tencrypted, err := k.encrypt(w)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot encrypt the keystore: %v\", err)\n\t}\n\n\tflags := os.O_RDWR | os.O_CREATE\n\tif override {\n\t\tflags |= os.O_TRUNC\n\t} else {\n\t\tflags |= os.O_EXCL\n\t}\n\n\tf, err := os.OpenFile(temporaryPath, flags, filePermission)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"cannot open file to save the keystore to '%s', error: %s\", k.Path, err)\n\t}\n\n\tf.Write(version)\n\tbase64Encoder := base64.NewEncoder(base64.StdEncoding, f)\n\tio.Copy(base64Encoder, encrypted)\n\tbase64Encoder.Close()\n\tf.Sync()\n\tf.Close()\n\n\terr = file.SafeFileRotate(k.Path, temporaryPath)\n\tif err != nil {\n\t\tos.Remove(temporaryPath)\n\t\treturn fmt.Errorf(\"cannot replace the existing keystore, with the new keystore file at '%s', error: %s\", k.Path, err)\n\t}\n\tos.Remove(temporaryPath)\n\n\tk.dirty = false\n\treturn nil\n}", "func (d *Database) Save() error {\n\tb, err := json.Marshal(d.State)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Save\n\tif err := ioutil.WriteFile(d.FilePath, b, os.ModePerm); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (s *saver) Save(journey models.Journey) error {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.state == closed {\n\t\treturn ErrSaverIsClosed\n\t}\n\n\tif len(s.buffer) == cap(s.buffer) {\n\t\treturn ErrInternalBufferIsFull\n\t}\n\n\ts.buffer = append(s.buffer, journey)\n\treturn nil\n}", "func (k *FileKeystore) Save() error {\n\tk.Lock()\n\terr := k.doSave(true)\n\tk.Unlock()\n\treturn err\n}", "func (c *Checkpoint) flush() error {\n\tc.fileLock.Lock()\n\tdefer c.fileLock.Unlock()\n\n\ttempFile := c.file + \".new\"\n\tfile, err := create(tempFile)\n\tif os.IsNotExist(err) {\n\t\t// Try to create directory if it does not exist.\n\t\tif createDirErr := c.createDir(); createDirErr == nil {\n\t\t\tfile, err = create(tempFile)\n\t\t}\n\t}\n\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to flush state to disk. %v\", err)\n\t}\n\n\t// Sort persisted eventLogs by name.\n\tc.sort = c.sort[:0]\n\tfor k := range c.states {\n\t\tc.sort = append(c.sort, k)\n\t}\n\tsort.Strings(c.sort)\n\n\tps := PersistedState{\n\t\tUpdateTime: time.Now().UTC(),\n\t\tStates: make([]JournalState, len(c.sort)),\n\t}\n\tfor i, name := range c.sort {\n\t\tps.States[i] = c.states[name]\n\t}\n\n\tdata, err := yaml.Marshal(ps)\n\tif err != nil {\n\t\tfile.Close()\n\t\treturn fmt.Errorf(\"Failed to flush state to disk. Could not marshal \"+\n\t\t\t\"data to YAML. %v\", err)\n\t}\n\n\t_, err = file.Write(data)\n\tif err != nil {\n\t\tfile.Close()\n\t\treturn fmt.Errorf(\"Failed to flush state to disk. Could not write to \"+\n\t\t\t\"%s. %v\", tempFile, err)\n\t}\n\n\tfile.Close()\n\terr = os.Rename(tempFile, c.file)\n\treturn err\n}", "func (c *Action) SaveState(k, v string) {\n\tfmt.Fprintf(c.w, saveStateFmt, k, escapeData(v))\n}", "func MustSave(filePath string, data any, optFns ...OpenOptionFunc) {\n\tbasefn.MustOK(SaveFile(filePath, data, optFns...))\n}", "func (c *Keystoreclient) HardSave(ctx context.Context, key string, message proto.Message) error {\n\tfor i := 0; i < c.retries; i++ {\n\t\terr := c.Save(ctx, key, message)\n\t\tif err == nil {\n\t\t\treturn err\n\t\t}\n\n\t\ttime.Sleep(c.backoffTime / time.Duration(c.retries))\n\t}\n\n\treturn errors.New(\"Unable to perform hard save\")\n}", "func (file *File) Save() {\n\tif file.autoFmt {\n\t\terr := file.Fmt()\n\t\tif err != nil {\n\t\t\tfile.NotifyUser(err.Error())\n\t\t}\n\t}\n\tfile.SnapshotSaved()\n\tcontents := []byte(file.ToString())\n\terr := ioutil.WriteFile(file.Name, contents, file.fileMode)\n\tif err != nil {\n\t\tfile.NotifyUser(\"Save Failed: \" + err.Error())\n\t} else {\n\t\tfile.savedBuffer.ReplaceBuffer(file.buffer.DeepDup())\n\t\tfile.NotifyUser(\"Saved.\")\n\t\tfile.modTime = time.Now()\n\t\tfile.md5sum = md5.Sum(contents)\n\t}\n}", "func (s *Store) Save() error {\n\tbk, err := os.OpenFile(filepath.Join(s.rwDirPath, storeBkName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer bk.Close()\n\n\tdst, err := os.OpenFile(filepath.Join(s.rwDirPath, storeName), os.O_RDWR|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer dst.Close()\n\n\t// backing up current store\n\t_, err = io.Copy(bk, dst)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\n\tenc := gob.NewEncoder(dst)\n\tbook := s.Clone()\n\terr = enc.Encode(book)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\treturn nil\n}", "func Save(filename string, getBytes func() (string, int)) (msg string, stillDirty bool) {\n\tfp, e := os.OpenFile(filename, os.O_RDWR|os.O_CREATE|os.O_TRUNC, 0644)\n\tif e != nil {\n\t\treturn fmt.Sprintf(\"Can't save! file open error %s\", e), true\n\t}\n\tdefer fp.Close()\n\tbuf, length := getBytes()\n\tstillDirty = true\n\tn, err := io.WriteString(fp, buf)\n\tif err == nil {\n\t\tif n == length {\n\t\t\tstillDirty = false\n\t\t\tmsg = fmt.Sprintf(\"%d bytes written to disk\", length)\n\t\t} else {\n\t\t\tmsg = fmt.Sprintf(\"wanted to write %d bytes to file, wrote %d\", length, n)\n\t\t}\n\t\treturn msg, stillDirty\n\t}\n\treturn fmt.Sprintf(\"Can't save! I/O error %s\", err), stillDirty\n}", "func MustSave(path string, f *os.File) {\n\tc, err := New()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\terr = c.SaveFile(path, f)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (m *MockManager) Save(arg0 state.State) error {\n\tret := m.ctrl.Call(m, \"Save\", arg0)\n\tret0, _ := ret[0].(error)\n\treturn ret0\n}", "func (s *ServiceState) save() {\n\tlog.Lvl3(\"Saving service\")\n\tb, err := network.Marshal(s.Storage)\n\tif err != nil {\n\t\tlog.Error(\"Couldn't marshal service:\", err)\n\t} else {\n\t\terr = ioutil.WriteFile(s.path+\"/prifi.bin\", b, 0660)\n\t\tif err != nil {\n\t\t\tlog.Error(\"Couldn't save file:\", err)\n\t\t}\n\t}\n}", "func (ifile *Indexfile) Save(lfindex *Index) error {\n\tifile.Open(CREATE | WRITE_ONLY | APPEND)\n\tdefer ifile.Close()\n\t_, err := ifile.LockedWriteAt(lfindex.ToBytes(), 0)\n\treturn err\n}", "func (service *HTTPRestService) saveState() error {\n\tlog.Printf(\"[Azure CNS] saveState\")\n\n\t// Skip if a store is not provided.\n\tif service.store == nil {\n\t\tlog.Printf(\"[Azure CNS] store not initialized.\")\n\t\treturn nil\n\t}\n\n\t// Update time stamp.\n\tservice.state.TimeStamp = time.Now()\n\terr := service.store.Write(storeKey, &service.state)\n\tif err == nil {\n\t\tlog.Printf(\"[Azure CNS] State saved successfully.\\n\")\n\t} else {\n\t\tlog.Errorf(\"[Azure CNS] Failed to save state., err:%v\\n\", err)\n\t}\n\n\treturn err\n}", "func (r *CheckpointStore) Save(checkpoint core.Checkpoint) error {\n\tcheckpointSaveCallsMeter.Mark(1)\n\n\tif checkpoint.Region == r.config.LocalRegion {\n\t\treturn errors.New(\"cannot save checkpoint for local region\")\n\t}\n\n\tif !r.updateState(checkpoint) {\n\t\treturn nil\n\t}\n\n\tkey := r.formatMessageKey(checkpoint)\n\tvalue, err := r.marshalCheckpoint(checkpoint)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn r.pubsub.Publish(r.config.Topic, key, value)\n}", "func (m SaveHelper) SaveGame(autosave bool) (ret string, err error) {\n\tif b, e := json.Marshal(m.values); e != nil {\n\t\terr = e\n\t} else {\n\t\tvar saveId string\n\t\tif autosave {\n\t\t\tsaveId = \"autosave\"\n\t\t}\n\t\tif r, e := m.saver.SaveBlob(saveId, b); e != nil {\n\t\t\terr = e\n\t\t} else {\n\t\t\tret = r\n\t\t}\n\t}\n\treturn\n}", "func (f *File) Save(path string) (err error) {\n\tdefer func() {\n\t\tif err != nil {\n\t\t\terr = fmt.Errorf(\"File.Save(%s): %w\", path, err)\n\t\t}\n\t}()\n\ttarget, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer func() {\n\t\tif ie := target.Close(); ie != nil {\n\t\t\terr = fmt.Errorf(\"write:%+v close:%w\", err, ie)\n\t\t}\n\t}()\n\terr = f.Write(target)\n\treturn\n}", "func (a *Agent) persistCheckState(check *checks.CheckTTL, status, output string) error {\n\t// Create the persisted state\n\tstate := persistedCheckState{\n\t\tCheckID: check.CheckID,\n\t\tStatus: status,\n\t\tOutput: output,\n\t\tExpires: time.Now().Add(check.TTL).Unix(),\n\t}\n\n\t// Encode the state\n\tbuf, err := json.Marshal(state)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Create the state dir if it doesn't exist\n\tdir := filepath.Join(a.config.DataDir, checkStateDir)\n\tif err := os.MkdirAll(dir, 0700); err != nil {\n\t\treturn fmt.Errorf(\"failed creating check state dir %q: %s\", dir, err)\n\t}\n\n\t// Write the state to the file\n\tfile := filepath.Join(dir, checkIDHash(check.CheckID))\n\n\t// Create temp file in same dir, to make more likely atomic\n\ttempFile := file + \".tmp\"\n\n\t// persistCheckState is called frequently, so don't use writeFileAtomic to avoid calling fsync here\n\tif err := ioutil.WriteFile(tempFile, buf, 0600); err != nil {\n\t\treturn fmt.Errorf(\"failed writing temp file %q: %s\", tempFile, err)\n\t}\n\tif err := os.Rename(tempFile, file); err != nil {\n\t\treturn fmt.Errorf(\"failed to rename temp file from %q to %q: %s\", tempFile, file, err)\n\t}\n\n\treturn nil\n}", "func Save(willSave bool, popSaved uint, delSaved uint) {\n\tos.Exit(0)\n}", "func (s store) Save() {\n\ts.writeToDisk()\n}", "func (l *LiveLog) Save(ctx context.Context) error {\n\tl.Lock()\n\tdefer l.Unlock()\n\treturn l.flush()\n}", "func SaveState(backupFile string, noteAll *[]*notes.Note) {\n\tdat, err := json.Marshal(noteAll)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\terr = ioutil.WriteFile(backupFile, dat, 0644)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n}", "func (t *WindowedThroughput) SaveState() ([]byte, error) {\n\treturn nil, nil\n}", "func savingFile(states AllStates, ID string) {\n\tfile, err := os.Create(\"elevator_states.txt\") //Creates file that will only contain latest data\n\t//checks for errors and saves to file as JSON\n\tcheck(err)\n\te := json.NewEncoder(file).Encode(states) //saves the AllStates struct to file\n\tcheck(e)\n}", "func (f *File) Save(name string) error {\n\treturn ioutil.WriteFile(name, []byte(f.String()), 0666)\n}", "func (l *LogDB) SaveRaftState(updates []pb.Update, shardID uint64) error {\n\tif l.collection.multiplexedLog() {\n\t\treturn l.concurrentSaveState(updates, shardID)\n\t}\n\treturn l.sequentialSaveState(updates, shardID)\n}", "func saveStore(s dhtStore) {\n\tif s.path == \"\" {\n\t\treturn\n\t}\n\ttmp, err := ioutil.TempFile(s.path, \"marconi\")\n\tif err != nil {\n\t\tlog.Println(\"saveStore tempfile:\", err)\n\t\treturn\n\t}\n\terr = json.NewEncoder(tmp).Encode(s)\n\t// The file has to be closed already otherwise it can't be renamed on\n\t// Windows.\n\ttmp.Close()\n\tif err != nil {\n\t\tlog.Println(\"saveStore json encoding:\", err)\n\t\treturn\n\t}\n\n\t// Write worked, so replace the existing file. That's atomic in Linux, but\n\t// not on Windows.\n\tp := fmt.Sprintf(\"%v-%v\", s.path+\"/dht\", s.Port)\n\tif err := os.Rename(tmp.Name(), p); err != nil {\n\t\t// if os.IsExist(err) {\n\t\t// Not working for Windows:\n\t\t// http://code.google.com/p/go/issues/detail?id=3828\n\n\t\t// It's not possible to atomically rename files on Windows, so I\n\t\t// have to delete it and try again. If the program crashes between\n\t\t// the unlink and the rename operation, it loses the configuration,\n\t\t// unfortunately.\n\t\tif err := os.Remove(p); err != nil {\n\t\t\tlog.Println(\"saveStore failed to remove the existing config:\", err)\n\t\t\treturn\n\t\t}\n\t\tif err := os.Rename(tmp.Name(), p); err != nil {\n\t\t\tlog.Println(\"saveStore failed to rename file after deleting the original config:\", err)\n\t\t\treturn\n\t\t}\n\t\t// } else {\n\t\t// \tlog.Println(\"saveStore failed when replacing existing config:\", err)\n\t\t// }\n\t} else {\n\t\t// log.Println(\"Saved DHT routing table to the filesystem.\")\n\t}\n}", "func Save(ctx context.Context) {\n\t// If test setup failed, then the output dir may not exist.\n\tdir, ok := testing.ContextOutDir(ctx)\n\tif !ok || dir == \"\" {\n\t\ttesting.ContextLog(ctx, \"Failed to get name of directory\")\n\t\treturn\n\t}\n\tif _, err := os.Stat(dir); err != nil {\n\t\treturn\n\t}\n\n\tdir = filepath.Join(dir, \"faillog\")\n\tif err := os.MkdirAll(dir, 0755); err != nil {\n\t\treturn\n\t}\n\n\tSaveToDir(ctx, dir)\n}", "func (d *DiskStorage) Save() error {\n\n\tvar file, err = os.OpenFile(d.path, os.O_RDWR, 0644)\n\tif d.isError(err) {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\t_, errWrite := file.WriteString(d.String())\n\treturn errWrite\n}", "func (ws *WalletStore) Save() {\n\tvar buffer bytes.Buffer\n\tgob.Register(elliptic.P256())\n\tencoder := gob.NewEncoder(&buffer)\n\terr := encoder.Encode(ws.Wallets)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tfile := ws.Config.GetWalletStoreFile(ws.NodeID)\n\terr = ioutil.WriteFile(file, buffer.Bytes(), 0644)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func SaveJSONSafe(filename string, thing interface{}, mode os.FileMode) error {\n\tb, err := json.MarshalIndent(thing, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\tflags := os.O_WRONLY | os.O_CREATE | os.O_EXCL\n\tf, err := os.OpenFile(filename, flags, mode)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\tn, err := f.Write(b)\n\tif n != len(b) && err != nil {\n\t\terr = errors.New(\"Failed to save complete file\")\n\t}\n\tif err != nil {\n\t\tos.Remove(filename)\n\t}\n\treturn err\n}", "func Save() {\n\tdata := Savedata{\n\t\tName: GS.current.name,\n\t\tGamestate: GS.current,\n\t}\n\n\tf, err := json.MarshalIndent(data, \"\", \" \")\n\tcheck(err)\n\tioutil.WriteFile(\"data/savegame.json\", f, 0644)\n}", "func (m *Machine) Save() error {\n\tm.State = driver.Saved\n\tfmt.Printf(\"Save %s: %s\\n\", m.Name, m.State)\n\treturn nil\n}", "func (k *KeyStore) SaveKeyStore() error {\n\n\t// Check if key is empty.\n\tif len(k.key) == 0 {\n\t\treturn fmt.Errorf(\"keystore.SaveKeyStore - keystore not initialized\")\n\t}\n\n\tf, err := os.Create(k.path)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"keystore.SaveKeyStore os.Create: %s\", err.Error())\n\t}\n\tdefer f.Close()\n\n\tsw, err := k.Encrypter(f)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"keystore.SaveKeyStore k.Encrypter: %s\", err.Error())\n\t}\n\tdefer sw.Close()\n\n\t// Write values to file.\n\tif err := k.serialize(sw); err != nil {\n\t\treturn fmt.Errorf(\"keystore.SaveKeyStore k.serialize: %s\", err.Error())\n\t}\n\n\treturn nil\n}", "func (w *Worker) save() (n int, err error) {\n\tif w.position == 0 {\n\t\treturn 0, nil\n\t}\n\tif w.fileExists() {\n\t\tn, err = w.fileRoot.Write(w.buffer[0:w.position])\n\t\tif err == nil {\n\t\t\tw.position = 0\n\t\t}\n\t} else {\n\t\tw.errorCallback()\n\t}\n\treturn n, err\n}", "func (aws *ActionWithState) Save() error {\n\tdb, err := sqlconn.GetConnection()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer db.Close()\n\tupdate, err := db.Query(\"\" +\n\t\t\"UPDATE `action` \" +\n\t\t\"\tSET `state` = ?, \" +\n\t\t\"\t\t`processed` = ? \" +\n\t\t\"\tWHERE `id` = ?\", aws.FSM.Current(), aws.Processed, aws.id)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer update.Close()\n\treturn nil\n}", "func (p *DefaultParser) Save(buf *bytes.Buffer, filename string) error {\n\terr := ioutil.WriteFile(filename, buf.Bytes(), 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (env *Solo) SaveSnapshot(fname string) {\n\tenv.glbMutex.Lock()\n\tdefer env.glbMutex.Unlock()\n\n\tsnapshot := soloSnapshot{\n\t\tUtxoDB: env.utxoDB.State(),\n\t}\n\n\tfor _, ch := range env.chains {\n\t\tchainSnapshot := soloChainSnapshot{\n\t\t\tName: ch.Name,\n\t\t\tStateControllerKeyPair: rwutil.WriteToBytes(ch.StateControllerKeyPair),\n\t\t\tChainID: ch.ChainID.Bytes(),\n\t\t\tOriginatorPrivateKey: rwutil.WriteToBytes(ch.OriginatorPrivateKey),\n\t\t\tValidatorFeeTarget: ch.ValidatorFeeTarget.Bytes(),\n\t\t}\n\n\t\terr := ch.db.Iterate(kvstore.EmptyPrefix, func(k, v []byte) bool {\n\t\t\tchainSnapshot.DB = append(chainSnapshot.DB, k, v)\n\t\t\treturn true\n\t\t})\n\t\trequire.NoError(env.T, err)\n\n\t\tsnapshot.Chains = append(snapshot.Chains, chainSnapshot)\n\t}\n\n\tb, err := json.Marshal(snapshot)\n\trequire.NoError(env.T, err)\n\terr = os.WriteFile(fname, b, 0o600)\n\trequire.NoError(env.T, err)\n}", "func Save(obj any, file string) error {\n\tfp, err := os.Create(file)\n\tdefer fp.Close()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\tbw := bufio.NewWriter(fp)\n\terr = Write(obj, bw)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn err\n\t}\n\terr = bw.Flush()\n\tif err != nil {\n\t\tlog.Println(err)\n\t}\n\treturn err\n}", "func (xsml *XfileServiceMetricLog) Save(db XODB) error {\n\tif xsml.Exists() {\n\t\treturn xsml.Update(db)\n\t}\n\n\treturn xsml.Replace(db)\n}", "func (n *NoOP) SaveSnapshot(w io.Writer,\n\tfileCollection statemachine.ISnapshotFileCollection,\n\tdone <-chan struct{}) (uint64, error) {\n\tdata, err := json.Marshal(n)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t_, err = w.Write(data)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn uint64(len(data)), nil\n}", "func (cm *fileConfigManager) Save() error {\n\tkuttilog.Printf(\n\t\t4,\n\t\t\"[DEBUG]Saving config file '%s'...\",\n\t\tcm.configfilename,\n\t)\n\tdata, err := cm.configdata.Serialize()\n\tif err != nil {\n\t\tkuttilog.Printf(\n\t\t\t4,\n\t\t\t\"[DEBUG]Error Saving config file '%s': %v.\",\n\t\t\tcm.configfilename,\n\t\t\terr,\n\t\t)\n\t\treturn err\n\t}\n\n\treturn SaveConfigfile(cm.configfilename, data)\n}", "func (n *nameHistory) Save() error {\n\tif !n.isChanged {\n\t\treturn nil\n\t}\n\n\tfp, err := os.OpenFile(n.filepath, os.O_CREATE|os.O_WRONLY|os.O_TRUNC, os.ModePerm)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not open %q file: %v\", n.filepath, err)\n\t}\n\tdefer fp.Close()\n\n\tif err := yaml.NewEncoder(fp).Encode(n.entries); err != nil {\n\t\treturn fmt.Errorf(\"could not save table name history: %v\", err)\n\t}\n\n\tif err := utils.SyncFileAndDirectory(fp); err != nil {\n\t\treturn fmt.Errorf(\"could not sync oid to name map file: %v\", err)\n\t}\n\n\tn.isChanged = false\n\n\treturn nil\n}", "func (block *SimpleQueueBlock) Save(ctx context.Context, q *SimpleQueue) (err *mft.Error) {\n\tif q.MetaStorage == nil {\n\t\treturn nil\n\t}\n\tif !block.mxFileSave.TryLock(ctx) {\n\t\treturn GenerateError(10013000)\n\t}\n\tdefer block.mxFileSave.Unlock()\n\n\tif !block.mx.RTryLock(ctx) {\n\t\treturn GenerateError(10013001)\n\t}\n\n\tif block.SaveRv == block.ChangesRv {\n\t\tblock.mx.RUnlock()\n\t\treturn nil\n\t}\n\n\tchangesRv := block.ChangesRv\n\tdata, errMarshal := json.MarshalIndent(block.Data, \"\", \"\\t\")\n\tchLen := len(block.SaveWait)\n\n\tblock.mx.RUnlock()\n\n\tif errMarshal != nil {\n\t\treturn GenerateErrorE(10013002, errMarshal)\n\t}\n\n\tfileName := block.blockFileName()\n\n\tst, err := q.getStorageLock(ctx, block.Mark)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = st.Save(ctx, fileName, data)\n\tif err != nil {\n\t\treturn GenerateErrorE(10013003, err, fileName)\n\t}\n\n\tif !block.mx.TryLock(ctx) {\n\t\treturn GenerateError(10013004)\n\t}\n\tblock.SaveRv = changesRv\n\tif len(block.SaveWait) > 0 && chLen > 0 {\n\t\tsaveWait := make([]chan bool, 0)\n\t\tfor i := chLen; i < len(block.SaveWait); i++ {\n\t\t\tsaveWait = append(saveWait, block.SaveWait[i])\n\t\t}\n\t\tapprove := block.SaveWait[0:chLen]\n\n\t\tblock.SaveWait = saveWait\n\t\tgo func() {\n\t\t\tfor _, v := range approve {\n\t\t\t\tv <- true\n\t\t\t}\n\t\t}()\n\t}\n\n\tblock.mx.Unlock()\n\n\treturn nil\n}", "func (s *State) Save(db XODB) error {\n\tif s.Exists() {\n\t\treturn s.Update(db)\n\t}\n\n\treturn s.Insert(db)\n}", "func (m *MockConfig) SaveTimeout(timeout time.Duration) error {\n\targs := m.Called(timeout)\n\treturn args.Error(0)\n}", "func (cfg *Configuration) Save() error {\n\tcfg.locker.Lock()\n\tdefer cfg.locker.Unlock()\n\tif cfg.FilePath == \"\" {\n\t\treturn errors.New(\"Configuration.FilePath was not set\")\n\t}\n\treturn gonfig.Write(cfg, true)\n}", "func (s *Siegfried) Save(path string) error {\n\tls := persist.NewLoadSaver(nil)\n\tls.SaveTime(s.C)\n\ts.em.Save(ls)\n\ts.mm.Save(ls)\n\ts.cm.Save(ls)\n\ts.bm.Save(ls)\n\ts.tm.Save(ls)\n\tls.SaveTinyUInt(len(s.ids))\n\tfor _, i := range s.ids {\n\t\ti.Save(ls)\n\t}\n\tif ls.Err != nil {\n\t\treturn ls.Err\n\t}\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\t_, err = f.Write(append(config.Magic(), byte(config.Version()[0]), byte(config.Version()[1])))\n\tif err != nil {\n\t\treturn err\n\t}\n\tz, err := flate.NewWriter(f, 1)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = z.Write(ls.Bytes())\n\tz.Close()\n\treturn err\n}", "func (mr *MockTimeLogStoreMockRecorder) Save(ctx, timelog interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Save\", reflect.TypeOf((*MockTimeLogStore)(nil).Save), ctx, timelog)\n}", "func (tb *TextBuf) Save() error {\n\tif tb.Filename == \"\" {\n\t\treturn fmt.Errorf(\"giv.TextBuf: filename is empty for Save\")\n\t}\n\ttb.EditDone()\n\tinfo, err := os.Stat(string(tb.Filename))\n\tif err == nil && info.ModTime() != time.Time(tb.Info.ModTime) {\n\t\tvp := tb.ViewportFromView()\n\t\tgi.ChoiceDialog(vp, gi.DlgOpts{Title: \"File Changed on Disk\",\n\t\t\tPrompt: fmt.Sprintf(\"File has changed on disk since being opened or saved by you -- what do you want to do? File: %v\", tb.Filename)},\n\t\t\t[]string{\"Save To Different File\", \"Open From Disk, Losing Changes\", \"Save File, Overwriting\"},\n\t\t\ttb.This(), func(recv, send ki.Ki, sig int64, data interface{}) {\n\t\t\t\tswitch sig {\n\t\t\t\tcase 0:\n\t\t\t\t\tCallMethod(tb, \"SaveAs\", vp)\n\t\t\t\tcase 1:\n\t\t\t\t\ttb.Revert()\n\t\t\t\tcase 2:\n\t\t\t\t\ttb.SaveFile(tb.Filename)\n\t\t\t\t}\n\t\t\t})\n\t}\n\treturn tb.SaveFile(tb.Filename)\n}", "func (this *DatastoreOperations) ScheduleFlushIfNeeded(state *DatastoreState, maxDelay int64) {\n\t// If a flush is already scheduled\n\tif state.FlushScheduler.FlushScheduled() {\n\t\t// Return immediately\n\t\treturn\n\t}\n\n\tflushIfNeeded := func() {\n\t\tstartTime := MonoUnixTimeMilli()\n\n\t\tdidFlush, err := state.FlushScheduler.EnsureFlush(state.File, time.Duration(maxDelay)*time.Millisecond)\n\t\tif err == nil {\n\t\t\tif !didFlush {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Log a success message\n\t\t\tthis.ParentServer.Logf(1, \"Flushed datastore '%s' %dms after written\", this.Name, MonoUnixTimeMilli()-startTime)\n\t\t} else { // Otherwise,\n\t\t\t// Log a failure message\n\t\t\tthis.ParentServer.Logf(1, \"Error flushing datastore '%s'. %s\", this.Name, err.Error())\n\t\t}\n\t}\n\n\t// If maximum delay is 0\n\tif maxDelay == 0 {\n\t\t// Flush synchronously\n\t\tflushIfNeeded()\n\t} else { // Otherwise\n\t\t// Flush asynchonously\n\t\t//\n\t\t// Increment the reference count to ensure the file isn't closed before or while the goroutine\n\t\t// executes\n\t\tstate.Increment()\n\n\t\tgo func() {\n\t\t\tflushIfNeeded()\n\n\t\t\t// Decrement the reference count\n\t\t\tstate.Decrement()\n\t\t}()\n\t}\n}", "func (s *TreeSaver) Save(ctx context.Context, snPath string, node *restic.Node, nodes []FutureNode) FutureTree {\n\tch := make(chan saveTreeResponse, 1)\n\tjob := saveTreeJob{\n\t\tsnPath: snPath,\n\t\tnode: node,\n\t\tnodes: nodes,\n\t\tch: ch,\n\t}\n\tselect {\n\tcase s.ch <- job:\n\tcase <-ctx.Done():\n\t\tdebug.Log(\"not saving tree, context is cancelled\")\n\t\tclose(ch)\n\t\treturn FutureTree{ch: ch}\n\t}\n\n\treturn FutureTree{ch: ch}\n}", "func (r *Reserve) Save(s *Store) error {\n\tdata, _ := json.Marshal(s)\n\tif err := ioutil.WriteFile(r.path, data, 0644); err != nil {\n\t\treturn fmt.Errorf(\"Failed to set %s: %s\", r.name, err)\n\t}\n\treturn nil\n}", "func (s *Store) SaveState() (err error) {\n\tif err = s.writePluginIDMap(); err != nil {\n\t\tslog.WithError(err).Error(\"can't write plugin id maps\")\n\t}\n\treturn\n}", "func (m *SynapsesPersist) Save() {\n\tif m.changed {\n\t\tfmt.Println(\"Saving synaptic data...\")\n\t\tindentedJSON, _ := json.MarshalIndent(m.Synapses, \"\", \" \")\n\n\t\tdataPath, err := filepath.Abs(m.relativePath)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\terr = ioutil.WriteFile(dataPath+m.file, indentedJSON, 0644)\n\t\tif err != nil {\n\t\t\tlog.Fatalln(\"ERROR:\", err)\n\t\t}\n\n\t\tm.Clean()\n\t\tfmt.Println(\"Synaptic data saved\")\n\t}\n}", "func (ns *NodeStore) save() error {\r\n\tb, err := json.Marshal(ns)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\terr = ioutil.WriteFile(\"nodestore.json\", b, 0660)\r\n\tif err != nil {\r\n\t\treturn err\r\n\t}\r\n\treturn nil\r\n}", "func (mgr *Manager) SaveNodeDef(kind string, force bool) error {\n\tatomic.AddUint64(&mgr.stats.TotSaveNodeDef, 1)\n\n\tif mgr.cfg == nil {\n\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefNil, 1)\n\t\treturn nil // Occurs during testing.\n\t}\n\n\tnodeDef := &NodeDef{\n\t\tHostPort: mgr.bindHttp,\n\t\tUUID: mgr.uuid,\n\t\tImplVersion: mgr.version,\n\t\tTags: mgr.tags,\n\t\tContainer: mgr.container,\n\t\tWeight: mgr.weight,\n\t\tExtras: mgr.extras,\n\t}\n\n\tfor {\n\t\tnodeDefs, cas, err := CfgGetNodeDefs(mgr.cfg, kind)\n\t\tif err != nil {\n\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefGetErr, 1)\n\t\t\treturn err\n\t\t}\n\t\tif nodeDefs == nil {\n\t\t\tnodeDefs = NewNodeDefs(mgr.version)\n\t\t}\n\t\tnodeDefPrev, exists := nodeDefs.NodeDefs[mgr.uuid]\n\t\tif exists && !force {\n\t\t\tif reflect.DeepEqual(nodeDefPrev, nodeDef) {\n\t\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefSame, 1)\n\t\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefOk, 1)\n\t\t\t\treturn nil // No changes, so leave the existing nodeDef.\n\t\t\t}\n\t\t}\n\n\t\tnodeDefs.UUID = NewUUID()\n\t\tnodeDefs.NodeDefs[mgr.uuid] = nodeDef\n\t\tnodeDefs.ImplVersion = CfgGetVersion(mgr.cfg)\n\t\tlog.Printf(\"manager: setting the nodeDefs implVersion \"+\n\t\t\t\"to %s\", nodeDefs.ImplVersion)\n\n\t\t_, err = CfgSetNodeDefs(mgr.cfg, kind, nodeDefs, cas)\n\t\tif err != nil {\n\t\t\tif _, ok := err.(*CfgCASError); ok {\n\t\t\t\t// Retry if it was a CAS mismatch, as perhaps\n\t\t\t\t// multiple nodes are all racing to register themselves,\n\t\t\t\t// such as in a full datacenter power restart.\n\t\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefRetry, 1)\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefSetErr, 1)\n\t\t\treturn err\n\t\t}\n\t\tbreak\n\t}\n\tatomic.AddUint64(&mgr.stats.TotSaveNodeDefOk, 1)\n\treturn nil\n}", "func (service ProjectService) SaveStatus() SaveStatus {\n\tvar status SaveStatus\n\tstatus.mod = service.mod\n\tstatus.FilesModified = len(service.mod.ModifiedFilenames())\n\tif status.FilesModified > 0 {\n\t\tlastChangeTime := service.mod.LastChangeTime()\n\n\t\tif service.ModHasStorageLocation() && !lastChangeTime.IsZero() {\n\t\t\tstatus.SavePending = true\n\t\t\tsaveAt := lastChangeTime.Add(time.Duration(autosaveTimeoutSec) * time.Second)\n\t\t\tstatus.SaveIn = time.Until(saveAt)\n\t\t\tif status.SaveIn <= 0 {\n\t\t\t\tstatus.SaveIn = 0\n\t\t\t}\n\t\t}\n\t}\n\treturn status\n}", "func (c *Client) SafeSave(events []eventhus.Event, version int) error {\n\treturn c.save(events, version, true)\n}", "func (m *FileBackend) Save(b BackendData) (err error) {\n\tvar f *os.File\n\tif f, err = os.OpenFile(m.path, os.O_RDWR, 0744); err != nil {\n\t\terr = fmt.Errorf(\"error opening json for saving FileBackend: %v\", err)\n\t\treturn\n\t}\n\tdefer f.Close()\n\n\t// Set FileBackend index to 0\n\tif err = f.Truncate(0); err != nil {\n\t\terr = fmt.Errorf(\"error truncating FileBackend: %v\", err)\n\t\treturn\n\t}\n\n\t// Create flat store from store data\n\tfs := b.NewFlatRecords()\n\n\t// Encode flat store as JSON to disk\n\treturn json.NewEncoder(f).Encode(fs)\n}", "func (s *State) Persist() error {\n\tmempool := make([]Transaction, len(s.txMempool))\n\tcopy(mempool, s.txMempool)\n\n\tfor i := 0; i < len(mempool); i++ {\n\t\ttxJson, err := json.Marshal(mempool[i])\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif _, err = s.dbFile.Write(append(txJson, '\\n')); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Remove the Transaction written to a file from the mempool\n\t\ts.txMempool = s.txMempool[1:]\n\t}\n\n\treturn nil\n}", "func (s yamlStore) Save() error {\n\treturn s.SaveTo(DefaultPath)\n}", "func Save() error {\n\treturn nil\n}", "func (buf *Buf) Save(file string) error {\n\tif file == \"\" {\n\t\treturn errors.New(\"No filename given\")\n\t}\n\n\tbs := []byte(buf.Text())\n\terr := ioutil.WriteFile(file, bs, 0644)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"%s\", err)\n\t}\n\n\tbuf.Name = file\n\tbuf.ClearDirty()\n\treturn nil\n}", "func (cfg *Config) Save(fpath string) error {\n\t// create a copy\n\tconfig2 := cfg\n\t// clear, setup setting\n\tconfig2.Password = \"\"\n\tconfig2.Iterations = ConfigHashIterations\n\n\t// save to file\n\tbyteDat2, err := json.MarshalIndent(config2, \"\", \" \")\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(fpath, byteDat2, 0644)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Save(cfg *Config) error {\n\tfile, err := Location()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdir := filepath.Dir(file)\n\terr = os.MkdirAll(dir, os.FileMode(0755))\n\tif err != nil {\n\t\treturn fmt.Errorf(\"can't create directory %s: %v\", dir, err)\n\t}\n\tdata, err := json.MarshalIndent(cfg, \"\", \" \")\n\tif err != nil {\n\t\treturn fmt.Errorf(\"can't marshal config: %v\", err)\n\t}\n\terr = os.WriteFile(file, data, 0600)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"can't write file '%s': %v\", file, err)\n\t}\n\treturn nil\n}", "func (mr *MockManagerMockRecorder) Save(arg0 interface{}) *gomock.Call {\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Save\", reflect.TypeOf((*MockManager)(nil).Save), arg0)\n}", "func (s LocalBackend) WriteState(st *State) error {\n\tlog.Debugf(\"Writing state to %s\\n\", s.Path)\n\tdata, err := json.MarshalIndent(st, \"\", \" \")\n\tif err != nil {\n\t\tlog.Fatal(\"Failed to Load State for Writing\")\n\t}\n\terr = ioutil.WriteFile(s.Path, data, 0644)\n\tif err != nil {\n\t\tlog.Fatal(\"Failed to write state to file\")\n\t}\n\treturn nil\n}", "func (c *ConfigFile) Save() error {\n\trw, err := NewConfigReadWriter(c.version)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfilename := rw.Filename(c)\n\tif filename == \"\" {\n\t\treturn fmt.Errorf(\"Can't save config with empty filename\")\n\t}\n\tif err := os.MkdirAll(filepath.Dir(filename), 0700); err != nil {\n\t\treturn err\n\t}\n\tfile, err := os.OpenFile(filename, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0600)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\treturn rw.SaveToWriter(file, c)\n}", "func (ts *FileTokenStorage) Save(token Token) error {\n\tencodedToken := base64.StdEncoding.EncodeToString([]byte(token.tokenString))\n\tts.token = encodedToken\n\treturn ioutil.WriteFile(ts.tokenFileName, []byte(encodedToken), 0644)\n}", "func (n *mockAgent) save() (s persistapi.AgentState) {\n\treturn\n}", "func (c *Passward) Save() error {\n\n\tif !util.DirectoryExists(c.Path) {\n\t\tif err := os.MkdirAll(c.Path, 0700); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tfile, err := os.Create(c.configPath())\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\tif err := toml.NewEncoder(file).Encode(c); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c Configuration) Save() {\n\tbuf := new(bytes.Buffer)\n\tif err := toml.NewEncoder(buf).Encode(c); err != nil {\n\t\tlog.Fatalln(\"Failed to encode config\", err)\n\t}\n\tf, err := os.Create(configFile)\n\tif err != nil {\n\t\tlog.Fatalln(\"Failed to create file\", err)\n\t\treturn\n\t}\n\n\tw := bufio.NewWriter(f)\n\tbuf.WriteTo(w)\n\tw.Flush()\n}", "func (m *settings) saveToDisk() error {\n\tb, err := json.Marshal(m.redirects)\n\tif err != nil {\n\t\tlog.Printf(\"Error marshalling %s\", err)\n\t\treturn fmt.Errorf(\"error marshalling %s\", err)\n\t}\n\n\tif err := ioutil.WriteFile(m.filename, b, 0644); err != nil {\n\t\treturn fmt.Errorf(\"unable to open file %s\", err)\n\t}\n\tlog.Printf(\"saving to disk.\")\n\treturn nil\n}", "func (fcb *FileCacheBackend) Save(ob types.OutgoingBatch) error {\n\tfilename := fcb.getFilename(fcb.getCacheFilename(ob))\n\tlog.Printf(\"Saving to %s\", filename)\n\tfile, err := os.OpenFile(filename, os.O_CREATE|os.O_WRONLY|os.O_EXCL, 0600)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Unable to save to %s - %s\", filename, err)\n\t}\n\tdefer file.Close()\n\tfor _, item := range ob.Values {\n\t\tfile.WriteString(item + \"\\n\")\n\t}\n\treturn nil\n}", "func (mr *MockMutantStorageMockRecorder) Save(val interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Save\", reflect.TypeOf((*MockMutantStorage)(nil).Save), val)\n}", "func SaveSyncStatus() error {\n\tif Status == nil {\n\t\treturn nil\n\t}\n\n\tdata, err := json.Marshal(Status)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn ioutil.WriteFile(StatusFilePath, data, 0)\n}", "func (r *Registry) SaveUnitState(jobName string, unitState *unit.UnitState) {\n\tkey := path.Join(r.keyPrefix, statePrefix, jobName)\n\t//TODO: Handle the error generated by marshal\n\tjson, _ := marshal(unitState)\n\tr.etcd.Set(key, json, 0)\n}", "func Save(path string, v interface{}, opts Options) (err error) {\n\tvar (\n\t\tfile *os.File\n\t\ttmp = path + \".tmp.\" + cmn.GenTie()\n\t)\n\tif file, err = cmn.CreateFile(tmp); err != nil {\n\t\treturn\n\t}\n\tdefer func() {\n\t\t_ = file.Close()\n\t\tif err != nil {\n\t\t\tos.Remove(tmp)\n\t\t}\n\t}()\n\tif err = Encode(file, v, opts); err != nil {\n\t\treturn\n\t}\n\tif err = file.Close(); err != nil {\n\t\treturn\n\t}\n\terr = os.Rename(tmp, path)\n\treturn\n}", "func (f *File) Save() error {\n\tif f.Id == 0 {\n\t\treturn dbAccess.Insert(f)\n\t} else {\n\t\t_, err := dbAccess.Update(f)\n\t\treturn err\n\t}\n}", "func (s yamlStore) SaveTo(path string) error {\n\t// If yamlStore is unset, none of New, Load, or LoadFrom were called successfully\n\tif s.cfg == nil {\n\t\treturn store.SaveError{Err: fmt.Errorf(\"undefined config, use one of the initializers: New, Load, LoadFrom\")}\n\t}\n\n\t// If it is a new configuration, the path should not exist yet\n\tif s.mustNotExist {\n\t\t// Lets check that the file doesn't exist\n\t\t_, err := s.fs.Stat(path)\n\t\tif os.IsNotExist(err) {\n\t\t\t// This is exactly what we want\n\t\t} else if err == nil || os.IsExist(err) {\n\t\t\treturn store.SaveError{Err: fmt.Errorf(\"configuration already exists in %q\", path)}\n\t\t} else {\n\t\t\treturn store.SaveError{Err: fmt.Errorf(\"unable to check for file prior existence: %w\", err)}\n\t\t}\n\t}\n\n\t// Marshall into YAML\n\tcontent, err := s.cfg.MarshalYAML()\n\tif err != nil {\n\t\treturn store.SaveError{Err: fmt.Errorf(\"unable to marshal to YAML: %w\", err)}\n\t}\n\n\t// Prepend warning comment for the 'PROJECT' file\n\tcontent = append([]byte(commentStr), content...)\n\n\t// Write the marshalled configuration\n\terr = afero.WriteFile(s.fs, path, content, 0600)\n\tif err != nil {\n\t\treturn store.SaveError{Err: fmt.Errorf(\"failed to save configuration to %q: %w\", path, err)}\n\t}\n\n\treturn nil\n}", "func executeSave(\n\ttx *bolt.Tx,\n\tns string,\n\tdoc *document.Document,\n\tforce bool,\n) error {\n\ts, err := database.CreateStore(tx, ns)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\trec, exists, err := s.TryGetRecord(doc.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !force && doc.Revision != rec.GetRevision() {\n\t\treturn &protavo.OptimisticLockError{\n\t\t\tDocumentID: doc.ID,\n\t\t\tGivenRev: doc.Revision,\n\t\t\tActualRev: rec.GetRevision(),\n\t\t\tOperation: \"save\",\n\t\t}\n\t}\n\n\tvar new *database.Record\n\tif exists {\n\t\tnew, err = updateRecord(s, doc, rec)\n\t} else {\n\t\tnew, err = createRecord(s, doc)\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc, err := marshalContent(doc)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := s.PutContent(doc.ID, c); err != nil {\n\t\treturn err\n\t}\n\n\treturn unmarshalRecordManagedFields(new, doc)\n}", "func (status SaveStatus) ConfirmPendingSave() {\n\tif status.mod == nil {\n\t\treturn\n\t}\n\tstatus.mod.ResetLastChangeTime()\n}", "func SkipSave(b bool) {\n\tdefaultConfig.SkipSave(b)\n}", "func saver(ticker *time.Ticker) {\n\tfor {\n\t\tselect {\n\t\tcase <-ticker.C:\n\t\t\tprintln(\"Writing state\")\n\t\t\tlock.Lock()\n\t\t\tif dirty {\n\t\t\t\tdirty = false\n\t\t\t\tstate := State{folders}\n\n\t\t\t\tbytes, err := json.Marshal(state)\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\n\t\t\t\t// write to S3\n\t\t\t\tauth, err := aws.EnvAuth()\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t\ts := s3.New(auth, aws.USEast)\n\t\t\t\tbucket := s.Bucket(s3_bucket)\n\t\t\t\terr = bucket.Put(\"rss.json\", bytes, \"application/json\", s3.ACL(\"private\"))\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t}\n\t\t\tlock.Unlock()\n\t\t}\n\t}\n}", "func (ach *AppConfigHelper) Save() error {\n\n\tif ach.LoadedConfig == nil {\n\t\treturn fmt.Errorf(\"Loaded Config was nil\")\n\t}\n\n\tif len(ach.FilePath) <= 0 {\n\t\treturn fmt.Errorf(\"Config File Path was not set could not save\")\n\t}\n\n\tif err := ach.LoadedConfig.Save(ach.FilePath); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (defaultStorage) Save() error {\n\tpanic(noConfigStorage)\n}", "func (mr *MockICommunicationMethodFieldServiceMockRecorder) DoSave(arg0, arg1 interface{}) *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"DoSave\", reflect.TypeOf((*MockICommunicationMethodFieldService)(nil).DoSave), arg0, arg1)\n}", "func (c *Config) Save() error {\r\n\tlog.Debug().Msg(\"[Config] Saving configuration...\")\r\n\tc.Validate()\r\n\r\n\treturn c.SaveFile(EnvManagerConfigFile)\r\n}", "func (bot *LeagueAnnouncerBot) persist() error {\n\tlog.Debug(\"Writing bot state to disk\")\n\t// Write the current state of the bot to disk, to load for next time\n\tstorageBytes, err := bot.storage.store()\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = ioutil.WriteFile(BOT_STATE, storageBytes, 0600)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}" ]
[ "0.617797", "0.5997442", "0.5781807", "0.5622699", "0.5534226", "0.52477074", "0.5205175", "0.5161928", "0.5143698", "0.51407456", "0.5077695", "0.50723076", "0.50583434", "0.5027826", "0.49810305", "0.4951463", "0.48769078", "0.4840408", "0.48384455", "0.48160774", "0.48149258", "0.4809999", "0.48070952", "0.48061866", "0.47798806", "0.47585565", "0.47568762", "0.47510087", "0.47433984", "0.47323632", "0.4725662", "0.47245556", "0.47142953", "0.47125885", "0.4711447", "0.46935552", "0.46929893", "0.46892774", "0.4648199", "0.46439198", "0.46374926", "0.46366808", "0.46188068", "0.46053532", "0.45957884", "0.45917013", "0.45908824", "0.45900816", "0.45844555", "0.4582278", "0.45763788", "0.45731914", "0.45645657", "0.45632282", "0.45385537", "0.45312577", "0.45282567", "0.4526536", "0.4506546", "0.44947344", "0.4481415", "0.44773415", "0.4477029", "0.44618088", "0.44570875", "0.4455463", "0.445147", "0.44471866", "0.44471118", "0.44469994", "0.4439637", "0.44355834", "0.44317535", "0.44270113", "0.44204816", "0.44173625", "0.44172266", "0.44136304", "0.44125834", "0.4411306", "0.44084024", "0.44016123", "0.43998542", "0.43995383", "0.4398216", "0.43896213", "0.43854487", "0.43853617", "0.43848696", "0.43831182", "0.43817806", "0.4380272", "0.4375305", "0.436929", "0.43671304", "0.43661475", "0.43651432", "0.43638614", "0.4363718", "0.4349553" ]
0.74332714
0
Load reads state off the disk from the wellknown filepath and loads it into the passed State object.
func (manager *basicStateManager) Load() error { s := manager.state seelog.Info("Loading state!") data, err := manager.readFile() if err != nil { seelog.Error("Error reading existing state file", "err", err) return err } if data == nil { return nil } // Dry-run to make sure this is a version we can understand err = manager.dryRun(data) if err != nil { return err } // Now load it into the actual state. The reason we do this with the // intermediate state is that we *must* unmarshal directly into the // "saveable" pointers we were given in AddSaveable; if we unmarshal // directly into a map with values of pointers, those pointers are lost. // We *must* unmarshal this way because the existing pointers could have // semi-initialized data (and are actually expected to) var intermediate intermediateState err = json.Unmarshal(data, &intermediate) if err != nil { seelog.Debug("Could not unmarshal into intermediate") return err } for key, rawJSON := range intermediate.Data { actualPointer, ok := manager.state.Data[key] if !ok { seelog.Error("Loading state: potentially malformed json key of " + key) continue } err = json.Unmarshal(rawJSON, actualPointer) if err != nil { seelog.Debug("Could not unmarshal into actual") return err } } seelog.Debug("Loaded state!", "state", s) return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (d *Database) Load() error {\n\tb, err := ioutil.ReadFile(d.FilePath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err := json.Unmarshal(b, &d.State); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func LoadState(username string) (*State, error) {\n\tf, err := os.Open(getPathForUsername(username))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer f.Close()\n\n\tvar s State\n\tif err := json.NewDecoder(f).Decode(&s); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &s, nil\n}", "func (m *stateManager) Load() (err error) {\n\tm.state, err = m.storage.Read()\n\n\tif err == nil {\n\t\tm.stateChanged = false\n\t\tm.stateLoaded = true\n\t}\n\n\treturn\n}", "func (self *LSHforest) Load(path string) error {\n\tb, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn msgpack.Unmarshal(b, self)\n}", "func (r *Registrar) LoadState() {\n\tif existing, e := os.Open(r.registryFile); e == nil {\n\t\tdefer existing.Close()\n\t\tfmt.Printf(\"Loading registrar data from %s\\n\", r.registryFile)\n\t\tdecoder := json.NewDecoder(existing)\n\t\tdecoder.Decode(&r.State)\n\t}\n}", "func (s *Store) Load() error {\n\tfReader, err := os.Open(filepath.Join(s.rwDirPath, storeName))\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\tdefer fReader.Close()\n\n\tdec := gob.NewDecoder(fReader)\n\n\ts.Lock()\n\tdefer s.Unlock()\n\n\terr = dec.Decode(&s.data)\n\tif err != nil {\n\t\treturn errors.Wrap(err, err.Error())\n\t}\n\treturn nil\n}", "func LoadState(id string) *Gossiper {\n\tg := Gossiper{}\n\tcwd, _ := os.Getwd()\n\tf, e := ioutil.ReadFile(filepath.Join(cwd, utils.STATE_FOLDER, fmt.Sprint(id, \".json\")))\n\tutils.HandleError(e)\n\tjson.Unmarshal(f, &g)\n\tfmt.Printf(\"Got this: %+v\\n\", &g)\n\treturn &g\n}", "func LoadFromState(board Board, tiles Tiles, players []Player, turn Turn) Game {\n\n\tdict, err := LoadDictionary(dictPath)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn Game{\n\t\tboard: board,\n\t\tplayers: players,\n\t\tTiles: tiles,\n\t\tTurn: turn,\n\t\tDictionary: dict,\n\t}\n}", "func Load(path string, v interface{}) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tf, err := os.Open(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\treturn Unmarshal(f, v)\n}", "func Load(path string, v interface{}) error {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tf, err := os.Open(path)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn err\n\t\t}\n\t}\n\tdefer f.Close()\n\treturn Unmarshal(f, v)\n}", "func (s *DataStore) Load() error {\n\tfile, err := os.Open(s.path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\ts.Lock()\n\tdefer s.Unlock()\n\n\terr = json.NewDecoder(file).Decode(&s.model)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func Load(path string, object interface{}) error {\n\tfile, err := os.Open(path)\n\tdefer file.Close()\n\tif err != nil {\n\t\tlog.Error(\"Was not able to open file\", \"path\", path, \"error\", err)\n\t\treturn err\n\t}\n\tdecoder := gob.NewDecoder(file)\n\terr = decoder.Decode(object)\n\tif err != nil {\n\t\tlog.Error(\"Was not able to decode file.\", \"path\", path, \"error\", err)\n\t}\n\treturn err\n}", "func (b Banai) Load(stashID string) ([]byte, error) {\n\tpath := filepath.Join(b.stashFolder, stashID)\n\t_, e := os.Stat(path)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\tf, e := ioutil.ReadFile(path)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\treturn f, nil\n\n}", "func (g *Gonf) Load() error {\n\tb, err := ioutil.ReadFile(g.path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcfg := map[string]interface{}{}\n\tif err := json.Unmarshal(b, &cfg); err != nil {\n\t\treturn err\n\t}\n\n\t// forbids access to Get/Set local and HTTP while reloading\n\tg.confLock.Lock()\n\tdefer g.confLock.Unlock()\n\tg.conf = cfg\n\n\treturn g.load(cfg, []string{}, []otoFlag{})\n}", "func (wlt *Wallet) Load(dir string) error {\n\tr := &ReadableWallet{}\n\tif err := r.Load(filepath.Join(dir, wlt.GetFilename())); err != nil {\n\t\treturn err\n\t}\n\tr.Meta[\"filename\"] = wlt.GetFilename()\n\t*wlt = NewWalletFromReadable(r)\n\treturn nil\n}", "func (bot *LeagueAnnouncerBot) loadState() error {\n\t// Look for a state file and read it in\n\tbytes, err := ioutil.ReadFile(BOT_STATE)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Load the state into memory, loading summoner details and discord channels\n\tbotState := botStorage{MonitoredSummoners: make(map[string]*SummonerDetails)}\n\terr = botState.load(bytes, bot.riot, bot.discord)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Overwrite the bots current state with what we loaded from disk\n\tbot.storage = botState\n\tlog.Debugf(\"Loaded bot state from disk, Loaded %d users\", len(botState.MonitoredSummoners))\n\treturn nil\n}", "func LoadContainerState(filename string) (*State, error) {\n\tif filename == \"\" || filename == \"-\" {\n\t\treturn ReadContainerState(os.Stdin)\n\t}\n\n\tinputFile, err := os.Open(filename)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to open file: %v\", err)\n\t}\n\tdefer inputFile.Close()\n\n\treturn ReadContainerState(inputFile)\n}", "func (s *JsonSource) Load() error {\n\n\tfile, err := ioutil.ReadFile(s.Path)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = json.Unmarshal([]byte(file), s.TargetStruct)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (bs *BoxStorage) Load() error {\n\n\tlog.Println(\"Loading file from\", bs.fileLocation+\"/\"+BoxStorageFileName)\n\n\tfh, err := os.Open(bs.fileLocation + \"/\" + BoxStorageFileName)\n\tdefer fh.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\tdec := gob.NewDecoder(fh)\n\terr = dec.Decode(&bs)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func Load(path string) (db DB, err error) {\n\tdb.filePath = path\n\n\tb, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn\n\t}\n\n\terr = json.Unmarshal(b, &db)\n\n\treturn\n}", "func (ColourSketchStore *ColourSketchStore) Load(path string) error {\n\tb, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = msgpack.Unmarshal(b, ColourSketchStore)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (t *TOMLLoader) Load(s interface{}) error {\n\tvar r io.Reader\n\n\tif t.Reader != nil {\n\t\tr = t.Reader\n\t} else if t.Path != \"\" {\n\t\tfile, err := getConfig(t.Path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\t\tr = file\n\t} else {\n\t\treturn ErrSourceNotSet\n\t}\n\n\tif _, err := toml.DecodeReader(r, s); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func LoadState(backupFile string, noteAll *[]*notes.Note) {\n\tif _, err := os.Stat(backupFile); os.IsNotExist(err) {\n\t\tfmt.Println(\"No backup file\")\n\t\treturn\n\t}\n\tcontent, err := ioutil.ReadFile(backupFile)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\tvar initNote []*notes.Note\n\terr = json.Unmarshal(content, &initNote)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\t*noteAll = initNote\n\tfmt.Println(initNote)\n}", "func Load(p string) (cfg *Config, err error) {\n\tvar d []byte\n\td, err = ioutil.ReadFile(p)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tcfg = new(Config)\n\tif err = yaml.Unmarshal(d, cfg); err != nil {\n\t\treturn nil, err\n\t}\n\n\tif cfg.ClusterState == nil {\n\t\tcfg.ClusterState = &ClusterState{}\n\t}\n\tif cfg.ALBIngressController == nil {\n\t\tcfg.ALBIngressController = &ALBIngressController{}\n\t}\n\n\tcfg.ConfigPath, err = filepath.Abs(p)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif cfg.ClusterState.UpTook != \"\" {\n\t\tcfg.ClusterState.upTook, err = time.ParseDuration(cfg.ClusterState.UpTook)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\tif cfg.ALBIngressController.IngressUpTook != \"\" {\n\t\tcfg.ALBIngressController.ingressUpTook, err = time.ParseDuration(cfg.ALBIngressController.IngressUpTook)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn cfg, nil\n}", "func (lm *LocalMeta) Load() error {\n\t// initialize gset\n\tvar err error\n\tlm.gset, err = gtid.ParserGTID(lm.flavor, \"\")\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\n\tfile, err := os.Open(lm.filename)\n\tif err != nil && !os.IsNotExist(errors.Cause(err)) {\n\t\treturn errors.Trace(err)\n\t}\n\n\tif os.IsNotExist(errors.Cause(err)) {\n\t\treturn nil\n\t}\n\tdefer file.Close()\n\n\t_, err = toml.DecodeReader(file, lm)\n\tif err != nil {\n\t\treturn errors.Trace(err)\n\t}\n\n\tlm.gset, err = gtid.ParserGTID(lm.flavor, lm.BinlogGTID)\n\treturn errors.Trace(err)\n}", "func Load(path string) (*OBJFile, error) {\n\tin, err := os.Open(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer in.Close()\n\treturn Decode(in)\n}", "func Load(path string) ([]byte, error) {\n\tmtx.Lock()\n\tdefer mtx.Unlock()\n\n\tdata, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn data, nil\n}", "func Load(path string, object interface{}) error {\n\tfile, err := os.Open(path)\n\tif err == nil {\n\t\tdecoder := json.NewDecoder(file)\n\t\terr = decoder.Decode(object)\n\t}\n\tfile.Close()\n\treturn err\n}", "func (d *Dump) Load() error {\n\td.mutex.Lock()\n\tdefer d.mutex.Unlock()\n\n\tvar (\n\t\tdata []byte\n\t\terr error\n\t)\n\n\tif data, err = ioutil.ReadFile(d.filename); err != nil {\n\t\treturn err\n\t}\n\n\treturn d.decodeGob(data)\n}", "func Load(filepath string) error {\n\tonce.Do(func() {\n\t\tloadError = load(filepath)\n\t})\n\n\treturn loadError\n}", "func Load(path string, target interface{}) error {\n\tformatter, err := parsePath(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdata, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn load(formatter, data, target)\n}", "func (n *nameHistory) Load() error {\n\tfp, err := os.OpenFile(n.filepath, os.O_RDONLY, os.ModePerm)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not open %q file: %v\", n.filepath, err)\n\t}\n\tdefer fp.Close()\n\n\tif err := yaml.NewDecoder(fp).Decode(&n.entries); err != nil {\n\t\treturn fmt.Errorf(\"could not decode file: %v\", err)\n\t}\n\n\tn.isChanged = false\n\n\treturn nil\n}", "func (s *YAMLFileSource) Load(_ *schema.StructValidator) (err error) {\n\tif s.path == \"\" {\n\t\treturn errors.New(\"invalid yaml path source configuration\")\n\t}\n\n\treturn s.koanf.Load(file.Provider(s.path), yaml.Parser())\n}", "func (s *yamlStore) LoadFrom(path string) error {\n\ts.mustNotExist = false\n\n\t// Read the file\n\tin, err := afero.ReadFile(s.fs, path)\n\tif err != nil {\n\t\treturn store.LoadError{Err: fmt.Errorf(\"unable to read %q file: %w\", path, err)}\n\t}\n\n\t// Check the file version\n\tvar versioned versionedConfig\n\tif err := yaml.Unmarshal(in, &versioned); err != nil {\n\t\treturn store.LoadError{Err: fmt.Errorf(\"unable to determine config version: %w\", err)}\n\t}\n\n\t// Create the config object\n\tvar cfg config.Config\n\tcfg, err = config.New(versioned.Version)\n\tif err != nil {\n\t\treturn store.LoadError{Err: fmt.Errorf(\"unable to create config for version %q: %w\", versioned.Version, err)}\n\t}\n\n\t// Unmarshal the file content\n\tif err := cfg.UnmarshalYAML(in); err != nil {\n\t\treturn store.LoadError{Err: fmt.Errorf(\"unable to unmarshal config at %q: %w\", path, err)}\n\t}\n\n\ts.cfg = cfg\n\treturn nil\n}", "func Load() {\n\tvar err error\n\n\tconfLen := len(FilePath)\n\tif confLen != 0 {\n\t\terr = readFromJSON(FilePath)\n\t}\n\tif err == nil {\n\t\terr = readFromENV()\n\t}\n\tif err != nil {\n\t\tpanic(`Configuration not found. Please specify configuration`)\n\t}\n}", "func loadStateFromJSON(fileName string) (*resultState, error) {\n\tret := &resultState{}\n\texists, err := loadJSONFile(fileName, ret)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif !exists {\n\t\treturn nil, skerr.Fmt(\"The state file %q doesn't exist.\", fileName)\n\t}\n\treturn ret, nil\n}", "func (ws *WalletStore) Load(filename string) error {\n\tcontent, err := ioutil.ReadFile(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\treader := bytes.NewReader(content)\n\tgob.Register(elliptic.P256())\n\tdecoder := gob.NewDecoder(reader)\n\terr = decoder.Decode(&ws.Wallets)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *Info) Load() error {\n\tb, err := ioutil.ReadFile(c.path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tc.Lock()\n\terr = json.Unmarshal(b, c)\n\tc.Unlock()\n\n\treturn err\n}", "func (r *Reserve) Load() *Store {\n\tret := NewStore()\n\tdata, err := ioutil.ReadFile(r.path)\n\tif err != nil {\n\t\tlog.Printf(\"Failed to load store %s: %s\\n\", r.name, err)\n\t\treturn &Store{}\n\t}\n\tjson.Unmarshal(data, &ret)\n\treturn ret\n}", "func (t *WindowedThroughput) LoadState(state []byte) error {\n\treturn nil\n}", "func Load() {\n\tjsonFile, err := os.Open(\"data/savegame.json\")\n\tcheck(err)\n\tfmt.Println(\"Loaded the file.\")\n\tdefer jsonFile.Close()\n\n\tbyteValue, err := ioutil.ReadAll(jsonFile)\n\tcheck(err)\n\tvar data Savedata\n\tjson.Unmarshal(byteValue, &data)\n\n\tGS.paused = false\n\tswitch loadedGamestate := data.Name; loadedGamestate {\n\tcase \"Flying\":\n\t\tGS.current = flying\n\t\tUI76.currentPanel = MenuFlying\n\tdefault:\n\t\tGS.current = landed\n\t\tUI76.currentPanel = MenuLanded\n\t}\n}", "func (wd WorkDir) Load(file string) ([]byte, error) {\n\tfh, err := os.Open(wd.Join(file))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer fh.Close()\n\treturn ioutil.ReadAll(fh)\n}", "func Load(path string) (*Directory, error) {\n\t// Check if valid path\n\tfi, err := os.Stat(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tcontext, dir := createStartingState(path, fi)\n\tif err := filepath.Walk(path, context.directoryWalker); err != nil {\n\t\treturn dir, err\n\t}\n\treturn dir, nil\n}", "func Load() error {\n\treturn def.Load()\n}", "func Load(path string) (c Cache, err error) {\n\tvar unlock filemu.UnlockFunc\n\tif unlock, err = filemu.RLock(context.Background(), lockPath()); err != nil {\n\t\treturn\n\t}\n\tdefer func() {\n\t\tif e := unlock(); err == nil {\n\t\t\terr = e\n\t\t}\n\t}()\n\n\tvar f *os.File\n\tif f, err = os.Open(path); err != nil {\n\t\treturn\n\t}\n\tdefer func() {\n\t\tif e := f.Close(); err == nil {\n\t\t\terr = e\n\t\t}\n\t}()\n\n\tvar w wrapper\n\tif err = yaml.NewDecoder(f).Decode(&w); err == nil {\n\t\tc = &cache{\n\t\t\tchannel: w.Channel,\n\t\t\tlastCheckedAt: w.LastCheckedAt,\n\t\t\tlatestRelease: w.LatestRelease,\n\t\t\tinvalidVer: w.InvalidVer,\n\t\t}\n\t}\n\n\treturn\n}", "func (function *Function) Load() (err error) {\n\tdefinition, err := ioutil.ReadFile(function.Path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfunction.Definition = string(definition)\n\n\treturn\n}", "func (s *yamlStore) Load() error {\n\treturn s.LoadFrom(DefaultPath)\n}", "func Load(filePath string, t Tomler) error {\n\ttomlValue := t.TOMLValue()\n\tvar err error\n\tif _, err = toml.DecodeFile(filePath, tomlValue); err != nil {\n\t\treturn err\n\t}\n\treturn t.FromTOML(tomlValue)\n}", "func (n *mockAgent) load(s persistapi.AgentState) {}", "func (z ZSTDStore) Load(path string) ([]byte, error) {\n\td, err := ioutil.ReadFile(gzPath(path))\n\tif err != nil {\n\t\treturn ioutil.ReadFile(strings.TrimSuffix(path, \".gz\"))\n\t}\n\treturn zstd.Decompress(nil, d)\n}", "func Load(fileName string, src interface{}) error {\n\tfile, err := os.Open(fileName)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\tif err == nil {\n\t\tdecoder := gob.NewDecoder(file)\n\t\tif err = decoder.Decode(src); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\t// Restore parameters\n\tswitch src.(type) {\n\tcase Model:\n\t\tmodel := src.(Model)\n\t\tmodel.SetParams(model.GetParams())\n\tdefault:\n\t\treturn errors.New(\"the file is not a model dump\")\n\t}\n\treturn nil\n}", "func (h *handler) Load(state []byte) (core.Template, error) {\n\tdec := scale.NewDecoder(bytes.NewBuffer(state))\n\tvault := &Vault{}\n\tif _, err := vault.DecodeScale(dec); err != nil {\n\t\treturn nil, fmt.Errorf(\"%w: %w\", core.ErrInternal, err)\n\t}\n\treturn vault, nil\n}", "func (p *Platform) ReadStateFromFile(filename string) (*Platform, error) {\n\tfile, err := os.Open(filename)\n\tdefer file.Close()\n\tif err != nil {\n\t\treturn p, err\n\t}\n\treturn p.ReadState(file)\n}", "func (m *Model) Load(path string) error {\n\tdata, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn errors.Wrap(err, \"load model\")\n\t}\n\tif err := json.Unmarshal(data, m); err != nil {\n\t\treturn errors.Wrap(err, \"load model\")\n\t}\n\treturn nil\n}", "func Load(p string) (Spec, error) {\n\tvar spec Spec\n\n\tbuf, err := os.ReadFile(p)\n\tif err != nil {\n\t\treturn spec, fmt.Errorf(\"failed to read file: %s - %w\", p, err)\n\t}\n\n\terr = yaml.Unmarshal(buf, &spec)\n\tif err != nil {\n\t\treturn spec, fmt.Errorf(\"failed to parse spec: %s - %w\", p, err)\n\t}\n\n\tspec.Path = p\n\n\treturn spec, nil\n}", "func (l *OptionalTOMLLoader) Load(s interface{}) error {\n\tif _, err := os.Stat(l.Path); err == nil {\n\t\treturn l.TOMLLoader.Load(s)\n\t}\n\treturn nil\n}", "func (r *PebbleFileRegistry) Load() error {\n\tr.mu.Lock()\n\tdefer r.mu.Unlock()\n\tr.mu.currProto = &enginepb.FileRegistry{}\n\tr.registryFilename = r.FS.PathJoin(r.DBDir, fileRegistryFilename)\n\tf, err := r.FS.Open(r.registryFilename)\n\tif oserror.IsNotExist(err) {\n\t\treturn nil\n\t}\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\tvar b []byte\n\tif b, err = ioutil.ReadAll(f); err != nil {\n\t\treturn err\n\t}\n\tif err = protoutil.Unmarshal(b, r.mu.currProto); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (ds *RegularStateMachineWrapper) Loaded(from rsm.From) {\n\tds.mu.Lock()\n\tdefer ds.mu.Unlock()\n\tds.SetLoaded(from)\n}", "func (e *Definition) Load(path, entity string) error {\n\n\tfullPath := filepath.Join(path, entity)\n\n\tb, err := ioutil.ReadFile(fullPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tcontent := string(b)\n\te.json = jsonutil.NewFromString(content)\n\te.byteRead = 0\n\n\ts, err := schema.Read(e)\n\tif err != nil {\n\t\tlog.Printf(\"failed to read schema: %s\", err)\n\t\treturn err\n\t}\n\n\te.schema = s\n\tv := validator.New(s)\n\te.validator = v\n\treturn nil\n}", "func (b *BrBuilder) Load() error {\n\tfpath := filepath.Join(b.StorePath, adminDir, branchBin)\n\tfd, err := os.OpenFile(fpath, os.O_RDONLY, 666)\n\tif err != nil {\n\t\t//log.Error(2, \"[Branch] Load branch %s failed: %v.\", b.Name(), err)\n\t\treturn err\n\t}\n\n\tdefer fd.Close()\n\treturn gob.NewDecoder(fd).Decode(&b.Branch)\n}", "func (this *DatastoreOperations) Load() (*DatastoreState, error) {\n\tvar err error\n\n\t// Initialize a blank state object\n\tstate := &DatastoreState{\n\t\tFlushScheduler: NewDatastoreFlushScheduler(),\n\t}\n\n\t// Open the datastore file\n\tstate.File, err = FileDescriptors.OpenAndIncrement(this.FilePath, os.O_RDWR, 0666)\n\n\t// If an error occurred while opening the file\n\tif err != nil {\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// Get file size\n\tfileSize, err := state.GetFileSize()\n\n\t// If an error occurred while getting the file size\n\tif err != nil {\n\t\t// Close the file\n\t\tstate.Decrement()\n\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// Create a new index object\n\tstate.Index = NewDatastoreIndex()\n\n\t// Add new entries to the index by scanning the datastore file\n\terr = state.Index.AppendFromEntryStream(NewPrefetchingReaderAt(state.File), 0, fileSize, func(iteratorResult *EntryStreamIteratorResult) error {\n\t\tchecksumErr := iteratorResult.VerifyAllChecksums()\n\t\tif checksumErr != nil {\n\t\t\treturn checksumErr\n\t\t}\n\n\t\t// If the entry is the first one\n\t\tif iteratorResult.Offset == 0 {\n\t\t\t// Verify it is a valid head entry\n\t\t\theadEntryErr := iteratorResult.VerifyValidHeadEntry()\n\t\t\tif headEntryErr != nil {\n\t\t\t\treturn headEntryErr\n\t\t\t}\n\t\t}\n\n\t\t// If the current entry is the last one but it doesn't have a transaction end flag\n\t\tif iteratorResult.EndOffset() == fileSize && !iteratorResult.HasTransactionEndFlag() {\n\t\t\t// Return an unexpected end of stream error\n\t\t\treturn io.ErrUnexpectedEOF\n\t\t}\n\n\t\treturn nil\n\t})\n\n\t// If an error occurred while appending to the index, or the file was empty\n\tif err != nil || fileSize == 0 {\n\t\t// If file ended unexpectedly, was corrupted or last entry didn't include a transaction end marker\n\t\tif fileSize == 0 || err == io.ErrUnexpectedEOF || err == ErrCorruptedEntry || err == ErrInvalidHeadEntry {\n\t\t\t// Log message\n\t\t\tthis.ParentServer.Logf(1, \"An incomplete or corrupted transcacion found in datastore '%s'. Attempting repair..\", this.Name)\n\n\t\t\t// Attempt to roll back to last succesful transaction\n\t\t\trepairedState, err := this.Repair(state)\n\n\t\t\t// Release original file (the repaired file would have a different descriptor)\n\t\t\tstate.Decrement()\n\n\t\t\t// If an error occurred while repairing the datastore file\n\t\t\tif err != nil {\n\t\t\t\t// Return the error\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\t// Set the repaired datastore as the loaded datastore\n\t\t\tstate = repairedState\n\t\t} else { // Otherwise, index creation failed for some other reason\n\t\t\t// Release file\n\t\t\tstate.Decrement()\n\n\t\t\t// Return the error\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Load head entry\n\terr = state.LoadHeadEntry()\n\n\t// If some error occurred while loading the head entry\n\tif err != nil {\n\t\t// An error here would be highly unexpected since the datastore has just been checked for corruption\n\t\t// which included validation of its head entry.\n\t\tthis.ParentServer.Logf(1, \"Datastore '%s' cannot be opened due to an unexpected error while trying to load its head entry: %s\", this.Name, err.Error())\n\n\t\t// Release file\n\t\tstate.Decrement()\n\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// If this is a cached datastore, load its content to memory\n\tif this.IsCached {\n\t\t// Load and deserialize the file's content\n\t\terr = state.UpdateDataCache(state.File, 0, state.Size())\n\n\t\t// If some error occured while trying load the file's content\n\t\tif err != nil {\n\t\t\t// Release file\n\t\t\tstate.Decrement()\n\n\t\t\t// Return the error\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treturn state, nil\n}", "func (env *Environment) Load(source, code string) error {\n\treturn nil\n}", "func (env *Environment) Load(source, code string) error {\n\treturn nil\n}", "func (l *loaderImpl) Load(location string) ([]byte, error) {\n\tscheme, err := l.getSchemeLoader(location)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfullLocation, err := scheme.FullLocation(l.root, location)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn scheme.Load(fullLocation)\n}", "func (s *Bench) load(path string) error {\n\tf, err := os.ReadFile(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn yaml.Unmarshal(f, &s)\n}", "func LoadFromFile(path string, v interface{}) (err error) {\n\tr, err := os.Open(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer utils.CheckClose(r, fmt.Sprintf(\"error closing %q\", path), &err)\n\treturn Load(r, v)\n}", "func (ts *FileTokenStorage) Load() (*Token, error) {\n\tif _,err := os.Stat(ts.tokenFileName); err == nil {\n\t\tout, err := ioutil.ReadFile(ts.tokenFileName)\n\t\tts.token = string(out)\n\t\treturn NewToken(ts.token), err\n\t}\n\ttoken := NewToken(\"\")\n\ttoken.Invalidate()\n\treturn token, nil\n}", "func Load(path string) (content []byte, err error) {\n\tif content, err = ioutil.ReadFile(path); err == nil {\n\t\tif err != nil {\n\t\t\tcontent = nil\n\t\t}\n\t}\n\treturn\n}", "func (d *Drawer) load() error {\n\tdata, err := ioutil.ReadFile(d.path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif len(data) > 0 {\n\t\tvar payload interface{}\n\t\tpayload, err = d.serializer.Deserialize(data)\n\t\tif !isPointer(payload) {\n\t\t\tpanic(NonPointerErr)\n\t\t}\n\t\td.payload = payload\n\t} else {\n\t\td.payload = nil\n\t}\n\treturn err\n}", "func (component *Component) Load(filename string) error {\n\treturn util.LoadYAML(filename, component)\n}", "func Load(path string) (*LevelDB, error) {\n\treturn LoadWithOptions(path, DefaultOptions)\n}", "func Load(state *state.State, driverName string, name string, config map[string]string, logger logger.Logger, volIDFunc func(volType VolumeType, volName string) (int64, error), commonRulesFunc func() map[string]func(string) error) (Driver, error) {\n\t// Locate the driver loader.\n\tdriverFunc, ok := drivers[driverName]\n\tif !ok {\n\t\treturn nil, ErrUnknownDriver\n\t}\n\n\td := driverFunc()\n\terr := d.init(state, name, config, logger, volIDFunc, commonRulesFunc)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn d, nil\n}", "func loadFromPath(path string, out interface{}) error {\n\tvar err error\n\n\tf, err := os.Open(path)\n\tif f != nil {\n\t\tdefer func() {\n\t\t\tferr := f.Close()\n\t\t\tif ferr != nil {\n\t\t\t\tlog.Println(ferr)\n\t\t\t}\n\t\t}()\n\t}\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata, err := ioutil.ReadAll(f)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn toml.Unmarshal(data, out)\n}", "func (r *CheckpointStore) Load(region, topic string, partition uint32) *core.Checkpoint {\n\tif region == r.config.LocalRegion {\n\t\treturn nil\n\t}\n\n\tr.mutex.RLock()\n\n\tkey := checkpointKey{\n\t\tRegion: region,\n\t\tTopic: topic,\n\t\tPartition: partition,\n\t}\n\n\tresult := r.state[key]\n\n\tr.mutex.RUnlock()\n\treturn result\n}", "func Load(filePath string) error {\n\tConfig = new(config)\n\n\tif err := Config.load(filePath); err != nil {\n\t\treturn err\n\t}\n\n\treturn Config.verifySettings()\n}", "func (s *Storage) Load(file string) error {\n\tnewDB, err := openDB(file)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.mux.Lock()\n\tdefer s.mux.Unlock()\n\n\toldDB := s.getDB()\n\ts.db.Store(newDB)\n\tif oldDB != nil {\n\t\toldDB.Close()\n\t}\n\treturn nil\n}", "func (s *Storage) Load(file string) error {\n\tnewDB, err := openDB(file)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts.mux.Lock()\n\tdefer s.mux.Unlock()\n\n\toldDB := s.getDB()\n\ts.db.Store(newDB)\n\tif oldDB != nil {\n\t\toldDB.Close()\n\t}\n\treturn nil\n}", "func (tfl tiltfileLoader) Load(ctx context.Context, tf *corev1alpha1.Tiltfile, prevResult *TiltfileLoadResult) TiltfileLoadResult {\n\tstart := time.Now()\n\tfilename := tf.Spec.Path\n\tabsFilename, err := ospath.RealAbs(tf.Spec.Path)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn TiltfileLoadResult{\n\t\t\t\tConfigFiles: []string{filename},\n\t\t\t\tError: fmt.Errorf(\"No Tiltfile found at paths '%s'. Check out https://docs.tilt.dev/tutorial.html\", filename),\n\t\t\t}\n\t\t}\n\t\tabsFilename, _ = filepath.Abs(filename)\n\t\treturn TiltfileLoadResult{\n\t\t\tConfigFiles: []string{absFilename},\n\t\t\tError: err,\n\t\t}\n\t}\n\n\ttiltignorePath := watch.TiltignorePath(absFilename)\n\ttlr := TiltfileLoadResult{\n\t\tConfigFiles: []string{absFilename, tiltignorePath},\n\t}\n\n\ttiltignore, err := watch.ReadTiltignore(tiltignorePath)\n\n\t// missing tiltignore is fine, but a filesystem error is not\n\tif err != nil {\n\t\ttlr.Error = err\n\t\treturn tlr\n\t}\n\n\ttlr.Tiltignore = tiltignore\n\n\ts := newTiltfileState(ctx, tfl.dcCli, tfl.webHost, tfl.execer, tfl.k8sContextPlugin, tfl.versionPlugin,\n\t\ttfl.configPlugin, tfl.extensionPlugin, tfl.ciSettingsPlugin, feature.FromDefaults(tfl.fDefaults))\n\n\tmanifests, result, err := s.loadManifests(tf)\n\n\ttlr.BuiltinCalls = result.BuiltinCalls\n\ttlr.DefaultRegistry = s.defaultReg\n\n\t// All data models are loaded with GetState. We ignore the error if the state\n\t// isn't properly loaded. This is necessary for handling partial Tiltfile\n\t// execution correctly, where some state is correctly assembled but other\n\t// state is not (and should be assumed empty).\n\tws, _ := watch.GetState(result)\n\ttlr.WatchSettings = ws\n\n\t// NOTE(maia): if/when add secret settings that affect the engine, add them to tlr here\n\tss, _ := secretsettings.GetState(result)\n\ts.secretSettings = ss\n\n\tioState, _ := io.GetState(result)\n\n\ttlr.ConfigFiles = append(tlr.ConfigFiles, ioState.Paths...)\n\ttlr.ConfigFiles = append(tlr.ConfigFiles, s.postExecReadFiles...)\n\ttlr.ConfigFiles = sliceutils.DedupedAndSorted(tlr.ConfigFiles)\n\n\tdps, _ := dockerprune.GetState(result)\n\ttlr.DockerPruneSettings = dps\n\n\taSettings, _ := tiltfileanalytics.GetState(result)\n\ttlr.AnalyticsOpt = aSettings.Opt\n\n\ttlr.Secrets = s.extractSecrets()\n\ttlr.FeatureFlags = s.features.ToEnabled()\n\ttlr.Error = err\n\ttlr.Manifests = manifests\n\ttlr.TeamID = s.teamID\n\n\tobjectSet, _ := v1alpha1.GetState(result)\n\ttlr.ObjectSet = objectSet\n\n\tvs, _ := version.GetState(result)\n\ttlr.VersionSettings = vs\n\n\ttelemetrySettings, _ := telemetry.GetState(result)\n\ttlr.TelemetrySettings = telemetrySettings\n\n\tus, _ := updatesettings.GetState(result)\n\ttlr.UpdateSettings = us\n\n\tci, _ := cisettings.GetState(result)\n\ttlr.CISettings = ci\n\n\tconfigSettings, _ := config.GetState(result)\n\tif tlr.Error == nil {\n\t\ttlr.EnabledManifests, tlr.Error = configSettings.EnabledResources(tf, manifests)\n\t}\n\n\tduration := time.Since(start)\n\tif tlr.Error == nil {\n\t\ts.logger.Infof(\"Successfully loaded Tiltfile (%s)\", duration)\n\t}\n\textState, _ := tiltextension.GetState(result)\n\thashState, _ := hasher.GetState(result)\n\n\tvar prevHashes hasher.Hashes\n\tif prevResult != nil {\n\t\tprevHashes = prevResult.Hashes\n\t}\n\ttlr.Hashes = hashState.GetHashes()\n\n\ttfl.reportTiltfileLoaded(s.builtinCallCounts, s.builtinArgCounts, duration,\n\t\textState.ExtsLoaded, prevHashes, tlr.Hashes)\n\n\tif len(aSettings.CustomTagsToReport) > 0 {\n\t\treportCustomTags(tfl.analytics, aSettings.CustomTagsToReport)\n\t}\n\n\treturn tlr\n}", "func (r *LocalRegistry) Load() {\n\tvar (\n\t\tregBytes []byte\n\t\terr error\n\t)\n\t// check if localRepo file exist\n\t_, err = os.Stat(r.file())\n\tif err != nil {\n\t\t// then assume localRepo.json is not there: try and create it\n\t\tr.save()\n\t} else {\n\t\tregBytes, err = ioutil.ReadFile(r.file())\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\terr = json.Unmarshal(regBytes, r)\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t}\n}", "func (l *Level) load(path string) error {\n\t*l = NewLevel()\n\tf, err := os.Open(path)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"open file to load level: %w\", err)\n\t}\n\tdefer f.Close()\n\tdecoder := json.NewDecoder(f)\n\terr = decoder.Decode(l)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"decoding level from file: %w\", err)\n\t}\n\tfor _, a := range l.Art {\n\t\terr := a.Load()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"load %v: %w\", a.Path, err)\n\t\t}\n\t}\n\tif l.PlayerArt != nil {\n\t\terr := l.PlayerArt.Load()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"load player art %v: %w\", l.PlayerArt.Path, err)\n\t\t}\n\t}\n\tif l.BGArt != nil {\n\t\terr := l.BGArt.Load()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"load BG art %v: %w\", l.BGArt.Path, err)\n\t\t}\n\t}\n\tif l.BGAudio != nil {\n\t\terr = l.BGAudio.Load()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"load bg audio: %w\", err)\n\t\t}\n\t}\n\tfor n, t := range l.Triggers {\n\t\terr := t.Load()\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"load trigger '%v': %w\", n, err)\n\t\t}\n\t\tl.Triggers[n] = t\n\t}\n\treturn nil\n}", "func (k *KMP) Load(filepath, format string) error {\n\tvar err error\n\tswitch format {\n\tcase \"json\":\n\t\tf, err := os.Open(filepath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer f.Close()\n\t\tb, err := ioutil.ReadAll(f)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = json.Unmarshal(b, k)\n\tdefault:\n\t\treturn fmt.Errorf(\"invalid load format, %s\", format)\n\t}\n\treturn err\n}", "func (i *Interpreter) Load(v string) (*big.Int, bool) {\n\tx, ok := i.state[v]\n\treturn x, ok\n}", "func Load(fromPath string, toVar interface{}) {\n\tfile, e := ioutil.ReadFile(fromPath)\n\terr.Panic(e)\n\tfile = RemoveComment(file)\n\terr.Panic(json.Unmarshal(file, toVar))\n}", "func (a *Agent) loadCheckState(check *structs.HealthCheck) error {\n\t// Try to read the persisted state for this check\n\tfile := filepath.Join(a.config.DataDir, checkStateDir, checkIDHash(check.CheckID))\n\tbuf, err := ioutil.ReadFile(file)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\treturn fmt.Errorf(\"failed reading file %q: %s\", file, err)\n\t}\n\n\t// Decode the state data\n\tvar p persistedCheckState\n\tif err := json.Unmarshal(buf, &p); err != nil {\n\t\ta.logger.Printf(\"[ERR] agent: failed decoding check state: %s\", err)\n\t\treturn a.purgeCheckState(check.CheckID)\n\t}\n\n\t// Check if the state has expired\n\tif time.Now().Unix() >= p.Expires {\n\t\ta.logger.Printf(\"[DEBUG] agent: check state expired for %q, not restoring\", check.CheckID)\n\t\treturn a.purgeCheckState(check.CheckID)\n\t}\n\n\t// Restore the fields from the state\n\tcheck.Output = p.Output\n\tcheck.Status = p.Status\n\treturn nil\n}", "func (l *Loader) Load() ([]byte, error) {\n\treturn ioutil.ReadFile(l.filename)\n}", "func Load() (cfg *Config, err error) {\n\tfile, err := Location()\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = os.Stat(file)\n\tif os.IsNotExist(err) {\n\t\tcfg = &Config{}\n\t\terr = nil\n\t\treturn\n\t}\n\tif err != nil {\n\t\terr = fmt.Errorf(\"can't check if config file '%s' exists: %v\", file, err)\n\t\treturn\n\t}\n\t// #nosec G304\n\tdata, err := os.ReadFile(file)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"can't read config file '%s': %v\", file, err)\n\t\treturn\n\t}\n\tcfg = &Config{}\n\tif len(data) == 0 {\n\t\treturn\n\t}\n\terr = json.Unmarshal(data, cfg)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"can't parse config file '%s': %v\", file, err)\n\t\treturn\n\t}\n\treturn\n}", "func (this *DatastoreOperations) LoadIfNeeded(increment bool) (*DatastoreState, error) {\n\t// Lock the state mutex\n\tthis.stateLock.Lock()\n\n\t// Unlock the state mutex whenever the function exits\n\tdefer this.stateLock.Unlock()\n\n\t// If the file was previously known not to be found and that result was cached, immediately\n\t// return with that error.\n\tif this.lastPathError != nil {\n\t\treturn nil, this.lastPathError\n\t}\n\n\t// Get existing state object\n\texistingState := this.State\n\n\t// If it exists (I.e. the datastore is already open)\n\tif existingState != nil {\n\t\t// If the returned state's reference count should be incremented\n\t\tif increment == true {\n\t\t\t// Increment the current state's file reference count\n\t\t\texistingState.Increment()\n\t\t}\n\n\t\t// Return the existing state object\n\t\treturn existingState, nil\n\t}\n\t// Otherwise\n\n\t// Start measuring the operation time\n\tstartTime := MonoUnixTimeMilliFloat()\n\n\t// Load the datastore\n\tstate, err := this.Load()\n\n\t// If an error occurred while loading the datastore\n\tif err != nil {\n\t\t// If the file wasn't found, cache that outcome to prevent unnecessary future calls to the file system.\n\t\t// This gives a significant performance boost, especially when trying to load dedicated configuration\n\t\t// datastores, which are quite common to be missing.\n\t\t//\n\t\t// However a disadvantage is that files that are copied to the storage directory after the server has\n\t\t// been started are not guaranteed to be seen by the server. One way to counteract this would be to reset\n\t\t// the cached value to nil every time interval (say every a second or two)\n\t\tif pathError, ok := err.(*os.PathError); ok {\n\t\t\tthis.lastPathError = pathError\n\t\t}\n\n\t\t// Return the error\n\t\treturn nil, err\n\t}\n\n\t// If the returned state's reference count should be incremented\n\tif increment == true {\n\t\t// Increment the current file's reference count\n\t\tstate.Increment()\n\t}\n\n\t// Set the resulting state object to the current state object\n\tthis.State = state\n\n\t// Log a completion message\n\tthis.ParentServer.Logf(1, \"Loaded datastore '%s' in %fms\", this.Name, MonoUnixTimeMilliFloat()-startTime)\n\n\treturn state, nil\n}", "func (sm *StateMachine) LoadState(in *Msg) {\n\ttmp, err := json.Marshal(sm.state)\n\tif err != nil {\n\t\tsm.plugin.Log.Info(\"failed to marshal state for db.\", err)\n\t\treturn\n\t}\n\n\t// Using upsert to either insert and return a value or on conflict to\n\t// update and return a value doesn't work, leading to this longer form.\n\t// Could it be a Postgres bug? This can and should be optimized.\n\tif in.User.ID > 0 {\n\t\tq := `INSERT INTO states\n\t\t (key, userid, value, pluginname) VALUES ($1, $2, $3, $4)`\n\t\t_, err = sm.plugin.DB.Exec(q, StateKey, in.User.ID, tmp,\n\t\t\tsm.plugin.Config.Name)\n\t} else {\n\t\tq := `INSERT INTO states\n\t\t (key, flexid, flexidtype, value, pluginname) VALUES ($1, $2, $3, $4, $5)`\n\t\t_, err = sm.plugin.DB.Exec(q, StateKey, in.User.FlexID,\n\t\t\tin.User.FlexIDType, tmp, sm.plugin.Config.Name)\n\t}\n\tif err != nil {\n\t\tif err.Error() != `pq: duplicate key value violates unique constraint \"states_userid_pkgname_key_key\"` &&\n\t\t\terr.Error() != `pq: duplicate key value violates unique constraint \"states_flexid_flexidtype_pluginname_key_key\"` {\n\t\t\tsm.plugin.Log.Info(\"could not insert value into states.\", err)\n\t\t\tsm.state = 0\n\t\t\treturn\n\t\t}\n\t\tif in.User.ID > 0 {\n\t\t\tq := `SELECT value FROM states\n\t\t\t WHERE userid=$1 AND key=$2 AND pluginname=$3`\n\t\t\terr = sm.plugin.DB.Get(&tmp, q, in.User.ID, StateKey,\n\t\t\t\tsm.plugin.Config.Name)\n\t\t} else {\n\t\t\tq := `SELECT value FROM states\n\t\t\t WHERE flexid=$1 AND flexidtype=$2 AND key=$3 AND pluginname=$4`\n\t\t\terr = sm.plugin.DB.Get(&tmp, q, in.User.FlexID,\n\t\t\t\tin.User.FlexIDType, StateKey, sm.plugin.Config.Name)\n\t\t}\n\t\tif err != nil {\n\t\t\tsm.plugin.Log.Info(\"failed to get value from state.\", err)\n\t\t\treturn\n\t\t}\n\t}\n\tvar val int\n\tif err = json.Unmarshal(tmp, &val); err != nil {\n\t\tsm.plugin.Log.Info(\"failed unmarshaling state from db.\", err)\n\t\treturn\n\t}\n\tsm.state = val\n\n\t// Have we already entered a state?\n\tsm.stateEntered = sm.plugin.GetMemory(in, stateEnteredKey).Bool()\n\treturn\n}", "func Load(path string) (*Config, error) {\n\tconfigFile, err := os.Open(path)\n\tdefer configFile.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar config Config\n\tjsonParser := json.NewDecoder(configFile)\n\tparseErr := jsonParser.Decode(&config)\n\n\tif parseErr != nil {\n\t\treturn nil, parseErr\n\t}\n\tconfig.fillFlags()\n\treturn &config, nil\n}", "func (sd *StateDB) loadStateObject(addr types.Address) (*StateObject, error) {\n\tdata, err := sd.trie.TryGet(addr.ToBytes())\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"get state from trie err: %v\", err)\n\t}\n\tif data == nil {\n\t\treturn nil, nil\n\t}\n\tvar state StateObject\n\terr = state.Decode(data, sd)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"decode err: %v\", err)\n\t}\n\treturn &state, nil\n}", "func MustLoadFromFile(path string) *Game {\n\tdata, err := ioutil.ReadFile(path)\n\tif err != nil {\n\t\tlog.Fatalf(\"Could not read %q: %v\", path, err)\n\t}\n\n\tg, err := New(data)\n\tif err != nil {\n\t\tlog.Fatalf(\"Could not parse %q: %v\", path, err)\n\t}\n\n\treturn g\n}", "func Load(filePath string, struc interface{}) error{\n\n\t//Build OS reader\n\treader := new(ReaderOS)\n\t\n\t//Reading File\n\tcontentBytes, err := reader.Read(filePath)\n\tif err != nil{\n\t\tlog.Println(TAG, err)\n\t\treturn err\n\t}\n\n\t//Get Parser\n\tparser := GetParser()\n\n\t//Parsing\n\tif err := parser.Parse(contentBytes, struc); err != nil {\n\t\tlog.Println(TAG, err)\n\t\treturn ErrParserFile\n\t}\n\n\treturn nil\n}", "func (s *store) Load() error {\n\t// if the config doesn't exist, create it.\n\tconfigExists, err := s.configExists()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif !configExists {\n\t\tc := &Config{\n\t\t\tGames: make([]Game, 0),\n\t\t}\n\n\t\t// Write a new config with default settings.\n\t\treturn s.Write(c)\n\t}\n\n\treturn nil\n}", "func Load (path string, unsafe...bool) (*Parser, error) {\n if j, e := gjson.Load(path, unsafe...); e == nil {\n return &Parser{j}, nil\n } else {\n return nil, e\n }\n}", "func Load(r io.Reader) error {\n\treturn DefaultInstance.Load(r)\n}", "func (m *SynapsesPersist) Load() {\n\tdataPath, err := filepath.Abs(m.relativePath)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\teFile, err := os.Open(dataPath + m.file)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tdefer eFile.Close()\n\n\tbytes, err := ioutil.ReadAll(eFile)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\terr = json.Unmarshal(bytes, &m.Synapses)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (m *MMU) Load(rom string) error {\n\tf, err := os.Open(rom)\n\tdefer f.Close()\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif e := m.LoadFrom(f); e != nil {\n\t\treturn e\n\t}\n\n\treturn nil\n}", "func Load(cfg interface{}, configPath string) error {\n\tif err := readConfigFile(configPath, cfg); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (y *YAMLLoader) Load(s interface{}) error {\n\tvar r io.Reader\n\n\tif y.Reader != nil {\n\t\tr = y.Reader\n\t} else if y.Path != \"\" {\n\t\tfile, err := getConfig(y.Path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\t\tr = file\n\t} else {\n\t\treturn ErrSourceNotSet\n\t}\n\n\tdata, err := ioutil.ReadAll(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn yaml.Unmarshal(data, s)\n}", "func (s *Store) Load(path string) (*Example, error) {\n\tparts := strings.Split(path, \"/\")\n\tif len(parts) == 2 && parts[1] == \"\" {\n\t\treturn emptyExample, nil\n\t} else if len(parts) != 3 {\n\t\treturn nil, errcode.New(http.StatusNotFound, \"Invalid URL: %s\", path)\n\t}\n\n\tcategory := s.DB.FindCategory(parts[1])\n\tif category == nil {\n\t\treturn nil, errcode.New(http.StatusNotFound, \"Could not find category: %s\", parts[1])\n\t}\n\texample := category.FindExample(parts[2])\n\tif example == nil {\n\t\treturn nil, errcode.New(http.StatusNotFound, \"Could not find example: %s\", parts[2])\n\t}\n\treturn example, nil\n}" ]
[ "0.7483943", "0.7456567", "0.73585737", "0.7178908", "0.7054577", "0.7017278", "0.69987273", "0.67917466", "0.6788214", "0.67881906", "0.6699514", "0.6665143", "0.66472304", "0.6567202", "0.6515057", "0.6506626", "0.6451997", "0.6351073", "0.6348336", "0.6344872", "0.6328139", "0.6311906", "0.6307294", "0.6287706", "0.62839067", "0.6283271", "0.6282947", "0.6279704", "0.6277421", "0.6273944", "0.6255252", "0.6245625", "0.62451065", "0.6234865", "0.62058824", "0.6201328", "0.61952", "0.6191466", "0.6185143", "0.61770713", "0.6175452", "0.6171341", "0.6114403", "0.6094779", "0.6089022", "0.60810286", "0.60747266", "0.6070526", "0.6068484", "0.6047389", "0.60403395", "0.6013719", "0.6012844", "0.60115695", "0.6010269", "0.59976625", "0.59881574", "0.59711677", "0.5967562", "0.5957839", "0.59506", "0.5950393", "0.5950393", "0.5930574", "0.5921984", "0.5915478", "0.5906937", "0.5896123", "0.58890325", "0.5883974", "0.5883638", "0.58724815", "0.58713603", "0.58660036", "0.5860444", "0.5860183", "0.5860183", "0.58544374", "0.58502656", "0.5847791", "0.5838755", "0.58338606", "0.5831226", "0.58294255", "0.5828173", "0.5818994", "0.58155", "0.58143294", "0.5804827", "0.5804702", "0.57950985", "0.5788705", "0.57884735", "0.5786755", "0.5776027", "0.5774293", "0.57626224", "0.5756577", "0.57557565", "0.5750986" ]
0.77769524
0
ZipDirectory zips a whole directory and its contents (NOT recursive!) See:
func ZipDirectory(dirPath string) ([]byte, error) { buf := new(bytes.Buffer) w := zip.NewWriter(buf) // Recurse for all files in the dir files, err := ioutil.ReadDir(dirPath) if err != nil { return nil, err } for _, file := range files { filePath := filepath.Join(dirPath, file.Name()) // Create file in zip header := &zip.FileHeader{ Name: file.Name(), Method: zip.Deflate, Modified: time.Now(), } f, err := w.CreateHeader(header) if err != nil { return nil, err } // Read file contents contents, err := ioutil.ReadFile(filePath) if err != nil { return nil, err } // Write file contents to zip _, err = f.Write(contents) if err != nil { return nil, err } } // Make sure to check the error on Close. err = w.Close() if err != nil { return nil, err } // Return zip data return buf.Bytes(), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func zipDirectory(sourceDir string) (*bytes.Buffer, error) {\n\t// Make sure the source directory exists and is a directory\n\tinfo, err := os.Stat(sourceDir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif !info.IsDir() {\n\t\t// XXX\n\t\treturn nil, nil\n\t}\n\n\t// Create a memory backed zip file for the contents\n\tbuf := new(bytes.Buffer)\n\tw := zip.NewWriter(buf)\n\n\tfilepath.Walk(sourceDir, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\tlog.Print(err)\n\t\t\treturn nil\n\t\t}\n\n\t\t// The zip archive nees to contain relative filenames\n\t\trel, _ := filepath.Rel(sourceDir, path)\n\t\tif rel == \".\" {\n\t\t\treturn nil\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(info)\n\n\t\t// Directories get a header entry with a trailing \"/\"\n\t\tif info.IsDir() {\n\t\t\t// fmt.Printf(\"Adding directory %s as %s\\n\", path, rel)\n\t\t\theader.Name = rel + \"/\"\n\t\t\t_, err := w.CreateHeader(header)\n\t\t\treturn err\n\t\t}\n\n\t\t// Add the file to the zip archive\n\t\t// fmt.Printf(\"Adding file %s as %s\\n\", path, rel)\n\t\theader.Name = rel\n\t\theader.Method = zip.Deflate\n\t\twriter, err := w.CreateHeader(header)\n\t\tf, err := os.Open(path)\n\t\tdefer f.Close()\n\t\t_, err = io.Copy(writer, f)\n\t\treturn err\n\t})\n\n\t// Done creating the zip file\n\tw.Close()\n\n\t// Write it out as a test\n\tf, err := os.Create(\"foo.zip\")\n\tf.Write(buf.Bytes())\n\tf.Close()\n\treturn buf, nil\n}", "func ZipDir(source, dest string) error {\n\tf, err := os.OpenFile(dest, os.O_WRONLY|os.O_CREATE, 0666)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\treturn MZipDir(source, f)\n}", "func ZipDir(dirPath string, w io.Writer) (err error) {\n\tzipWriter := zip.NewWriter(w)\n\terr = walkAndZip(dirPath, \"\", zipWriter)\n\tif err != nil {\n\t\treturn\n\t}\n\terr = zipWriter.Close()\n\treturn\n}", "func (a *Archiver) ZipDirectory(src, dest string) error {\n\tf, err := os.Create(dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\tw := zip.NewWriter(f)\n\tdefer w.Close()\n\n\twalker := func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\t\tfl, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer fl.Close()\n\n\t\tf, err := w.Create(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\ta.BeforeCopyCallback(path, info)\n\t\t_, err = io.Copy(f, fl)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\ta.AfterCopyCallback(path, info)\n\n\t\treturn nil\n\t}\n\treturn filepath.Walk(src, walker)\n}", "func (c *connection) Zipdir(path string) error {\n\tprogress := func(archivePath string) {\n\t}\n\treturn zip.Archive(path, c.Conn, progress)\n}", "func zipd(dir string, w io.Writer) error {\n\tz := zip.NewWriter(w)\n\tdefer func() {\n\t\tif err := z.Flush(); err != nil {\n\t\t\tfmt.Printf(\"Error flushing ZIP writer: %v\\n\", err)\n\t\t}\n\t\tif err := z.Close(); err != nil {\n\t\t\tfmt.Printf(\"Error closing ZIP writer: %v\\n\", err)\n\t\t}\n\t}()\n\tfilepath.Walk(dir, func(path string, info os.FileInfo, err error) error {\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\t\trel, err := filepath.Rel(dir, path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tr, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer r.Close()\n\t\tzw, err := z.Create(rel)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = io.Copy(zw, r)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n\n\treturn nil\n}", "func ZipDir(dirToZip, zipFilePath string) error {\n\tif !file.IsSafeFilePath(dirToZip) || !file.IsSafeFilePath(zipFilePath) {\n\t\treturn errors.New(\"file or dir path params are illegal\")\n\t}\n\tallFiles, err := file.ListFiles(dirToZip, true)\n\tif err != nil {\n\t\treturn err\n\t}\n\tzipFile, err := os.Create(zipFilePath)\n\tdefer func() { _ = zipFile.Close() }()\n\tif err != nil {\n\t\treturn nil\n\t}\n\tzipWriter := zip.NewWriter(zipFile)\n\tdefer func() { _ = zipWriter.Close() }()\n\tfor _, fileItem := range allFiles {\n\t\tfileInfo, e := os.Stat(fileItem)\n\t\tif e != nil {\n\t\t\treturn e\n\t\t}\n\t\tif fileInfo.IsDir() {\n\t\t\tcontinue\n\t\t}\n\t\tvar fileHeader *zip.FileHeader\n\t\tif fileHeader, e = zip.FileInfoHeader(fileInfo); e != nil {\n\t\t\treturn e\n\t\t}\n\t\tfileHeader.Name = fileItem[len(dirToZip)+1:]\n\t\tvar entryWriter io.Writer\n\t\tif entryWriter, e = zipWriter.CreateHeader(fileHeader); e != nil {\n\t\t\treturn e\n\t\t}\n\t\tvar entryFile *os.File\n\t\tif entryFile, e = os.Open(fileItem); e != nil {\n\t\t\treturn e\n\t\t}\n\t\tif e = transformFileTo(entryWriter, entryFile); e != nil {\n\t\t\treturn e\n\t\t}\n\t}\n\treturn nil\n}", "func zipDir(w http.ResponseWriter, r *http.Request) {\n\tzipPath := r.URL.Query().Get(\"zipPath\")\n\tzipName := r.URL.Query().Get(\"zipName\")\n\twz := zip.NewWriter(w)\n\tw.Header().Add(\"Content-Disposition\", \"attachment; filename=\\\"\"+zipName+\".zip\\\"\")\n\n\tvalidZipPath, err := validPath(zipPath)\n\tif err != nil {\n\t\tsendHTTPResp(w, 500, err)\n\t\treturn\n\t}\n\terr = walkZip(wz, validZipPath+\"/\", \"\")\n\tif err != nil {\n\t\tsendHTTPResp(w, 500, err)\n\t\treturn\n\t}\n\twz.Close()\n}", "func MZipDir(source string, dest io.Writer) error {\n\n\t// Add a slash to the end of the path\n\t// So the prefix is trimmed properly later on\n\tsource = filepath.Clean(source) + \"/\"\n\n\tzwr := zip.NewWriter(dest)\n\terr := Walk(source, func(p string, info os.FileInfo) error {\n\n\t\t// remove the root folder name from the archive\n\t\tnpath := strings.TrimPrefix(p, source)\n\n\t\theader, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Remove the root directory name from the archive\n\t\theader.Name = filepath.Join(npath, info.Name())\n\n\t\t// List the file as a directory in the archive\n\t\tif info.IsDir() {\n\t\t\theader.Name += \"/\"\n\t\t} else {\n\t\t\theader.Method = zip.Deflate\n\t\t}\n\n\t\thdr, err := zwr.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// The file is a directory, we do not need to copy anything into it\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tf, err := os.OpenFile(filepath.Join(p, info.Name()), os.O_RDONLY, 0666)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer f.Close()\n\n\t\t_, err = io.Copy(hdr, f)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = zwr.Close(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func processDirectory(dirpath string) {\r\n\r\n\tvar wg sync.WaitGroup\r\n\r\n\terr := filepath.Walk(dirpath, func(path string, info os.FileInfo, err error) error {\r\n\t\tif info.IsDir() {\r\n\t\t\treturn nil\r\n\t\t}\r\n\t\twg.Add(1)\r\n\t\tprocessFile(path, dirpath, &wg)\r\n\t\treturn nil\r\n\t})\r\n\r\n\twg.Wait()\r\n\r\n\tif err != nil {\r\n\t\tlog.Fatal(err)\r\n\t\ttime.Sleep(10 * time.Second)\r\n\t\treturn\r\n\t}\r\n\r\n\t// compress\r\n\tcandiname := dirpath\r\n\text := \".zip\"\r\n\t// add _re until not exist file\r\n\tfor {\r\n\t\tif _, err := os.Stat(candiname + ext); os.IsNotExist(err) {\r\n\t\t\tbreak\r\n\t\t}\r\n\t\tcandiname += \"_re\"\r\n\t}\r\n\r\n\tfmt.Println(dirpath + \"_re\")\r\n\t// bandizip.exe doesn't save top folder name, but bc.exe does, so use *.* with -r\r\n\tcmd := exec.Command(zipApp, \"c\", \"-r\", candiname+ext, dirpath+\"_re\\\\*.*\")\r\n\texecErr := cmd.Run()\r\n\tif execErr != nil {\r\n\t\tpanic(execErr)\r\n\t}\r\n\r\n\t// delete folder A_re\r\n\tos.RemoveAll(dirpath + \"_re\")\r\n}", "func Zip(ctx context.Context, opts *Options) error {\n\tlogger := opts.Logger\n\tif logger == nil {\n\t\tlogger = hclog.NewNullLogger()\n\t}\n\n\t// Setup our root directory with the given files.\n\troot, err := createRoot(ctx, logger, opts)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer os.RemoveAll(root)\n\n\t// Make our command for creating the archive\n\tcmd, err := dittoCmd(ctx, opts.BaseCmd)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tcmd.Args = []string{\n\t\tfilepath.Base(cmd.Path),\n\t\t\"-c\", // create an archive\n\t\t\"-k\", // create a PKZip archive, not CPIO\n\t}\n\tcmd.Args = append(cmd.Args, root)\n\tcmd.Args = append(cmd.Args, opts.OutputPath)\n\n\t// We store all output in out for logging and in case there is an error\n\tvar out bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = cmd.Stdout\n\n\t// Log what we're going to execute\n\tlogger.Info(\"executing ditto for zip archive creation\",\n\t\t\"output_path\", opts.OutputPath,\n\t\t\"command_path\", cmd.Path,\n\t\t\"command_args\", cmd.Args,\n\t)\n\n\t// Execute\n\tif err = cmd.Run(); err != nil {\n\t\tlogger.Error(\"error creating zip archive\", \"err\", err, \"output\", out.String())\n\t\treturn err\n\t}\n\n\tlogger.Info(\"zip archive creation complete\", \"output\", out.String())\n\treturn nil\n}", "func ToZip(dir, dest string) error {\n\tinfo, err := os.Stat(dir)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !(info.IsDir()) {\n\t\treturn fmt.Errorf(\"%s is not a directory\", dir)\n\t}\n\tarchive, err := os.Create(dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer archive.Close()\n\n\tz := zip.NewWriter(archive)\n\tdefer z.Close()\n\n\terr = filepath.Walk(dir, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// do not include srcDir into archive\n\t\tif info.Name() == filepath.Base(dir) {\n\t\t\treturn nil\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// adjust header.Name to preserve folder strulture\n\t\theader.Name = strings.TrimPrefix(path, dir)\n\n\t\tif info.IsDir() {\n\t\t\theader.Name += \"/\"\n\t\t} else {\n\t\t\theader.Method = zip.Deflate\n\t\t}\n\n\t\twriter, err := z.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tfile, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\t\t_, err = io.Copy(writer, file)\n\t\treturn err\n\t})\n\n\treturn err\n}", "func compressDirectory(source, target string) error {\n\t// Create file with target name and defer its closing\n\tzipfile, err := os.Create(target)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer zipfile.Close()\n\n\t// Create a new zip writer and pass our zipfile in\n\tarchive := zip.NewWriter(zipfile)\n\tdefer archive.Close()\n\n\t// Get os.FileInfo about our source\n\tinfo, err := os.Stat(source)\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\t// Check to see if the provided source file is a directory and set baseDir if so\n\tvar baseDir string\n\tif info.IsDir() {\n\t\tbaseDir = filepath.Base(source)\n\t}\n\n\t// Walk the whole filepath\n\tfilepath.Walk(source, func(path string, info os.FileInfo, err error) error {\n\t\text := filepath.Ext(path)\n\t\t_, ok := fileExtWhitelist[ext]\n\n\t\tif ok || (info.IsDir() && (ext != \".git\" && path != \"node_modules\")) {\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Creates a partially-populated FileHeader from an os.FileInfo\n\t\t\theader, err := zip.FileInfoHeader(info)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Check if baseDir has been set (from the IsDir check) and if it has not been\n\t\t\t// set, update the header.Name to reflect the correct path\n\t\t\tif baseDir != \"\" {\n\t\t\t\theader.Name = filepath.Join(baseDir, strings.TrimPrefix(path, source))\n\t\t\t}\n\n\t\t\t// Check if the file we are iterating is a directory and update the header.Name\n\t\t\t// or the header.Method appropriately\n\t\t\tif info.IsDir() {\n\t\t\t\theader.Name += \"/\"\n\t\t\t} else {\n\t\t\t\theader.Method = zip.Deflate\n\t\t\t}\n\n\t\t\t// Add a file to the zip archive using the provided FileHeader for the file metadata\n\t\t\twriter, err := archive.CreateHeader(header)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\t// Return nil if at this point if info is a directory\n\t\t\tif info.IsDir() {\n\t\t\t\treturn nil\n\t\t\t}\n\n\t\t\t// If it was not a directory, we open the file and copy it into the archive writer\n\t\t\t// ingore zip files\n\t\t\tfile, err := os.Open(path)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer file.Close()\n\t\t\t_, err = io.Copy(writer, file)\n\t\t}\n\n\t\treturn err\n\t})\n\n\treturn err\n}", "func createZIP(path, dir string) error {\n\tf, err := os.Create(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer f.Close()\n\n\tw := zip.NewWriter(f)\n\tdefer w.Close()\n\n\tprefix := filepath.Dir(dir)\n\twalker := func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tr, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer r.Close()\n\n\t\tname := strings.TrimLeft(strings.TrimPrefix(path, prefix), \"/\")\n\t\tf, err := w.Create(name)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = io.Copy(f, r)\n\t\treturn err\n\t}\n\treturn filepath.Walk(dir, walker)\n}", "func ExtractZip(data *[]byte, targetDir string) error {\n\tzipbuf := bytes.NewReader(*data)\n\tz, err := zip.NewReader(zipbuf, zipbuf.Size())\n\tif err != nil {\n\t\treturn err\n\t}\n\tdirPerm := os.FileMode(0750)\n\t//\tif err = os.MkdirAll(targetDir, dirPerm); err != nil {\n\t//\t\treturn err\n\t//\t}\n\n\t// Check if all files in zip are in the same directory.\n\t// this might be the case if the plugin zip contains the whole plugin dir\n\t// instead of its content.\n\tvar prefix string\n\tallPrefixed := false\n\tfor i, f := range z.File {\n\t\tparts := strings.Split(f.Name, \"/\")\n\t\tif i == 0 {\n\t\t\tprefix = parts[0]\n\t\t} else if parts[0] != prefix {\n\t\t\tallPrefixed = false\n\t\t\tbreak\n\t\t} else {\n\t\t\t// switch to true since we have at least a second file\n\t\t\tallPrefixed = true\n\t\t}\n\t}\n\t// Install files and directory's\n\tfor _, f := range z.File {\n\t\tparts := strings.Split(f.Name, \"/\")\n\t\tif allPrefixed {\n\t\t\tparts = parts[1:]\n\t\t}\n\n\t\ttargetName := filepath.Join(targetDir, filepath.Join(parts...))\n\t\tif f.FileInfo().IsDir() {\n\t\t\tif err := os.MkdirAll(targetName, dirPerm); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\tbasepath := filepath.Dir(targetName)\n\n\t\t\tif err := os.MkdirAll(basepath, dirPerm); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tcontent, err := f.Open()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer content.Close()\n\t\t\ttarget, err := os.Create(targetName)\n\t\t\tif err != nil {\n\t\t\t\tif strings.Contains(err.Error(), \"permission denied\") {\n\t\t\t\t\t// Skip files with permission denied\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer target.Close()\n\t\t\tif _, err = io.Copy(target, content); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func ZipFiles(filename string, folder string) error {\n\n\tnewZipFile, err := os.Create(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer newZipFile.Close()\n\n\tzipWriter := zip.NewWriter(newZipFile)\n\tdefer zipWriter.Close()\n\n\tcurrDir, err := os.Getwd()\n\n\tlog.Println(currDir, filepath.Base(folder))\n\n\t// Add files to zip\n\tif info, _ := os.Stat(folder); info.IsDir() {\n\t\tlocalFiles := []string{}\n\t\terr = filepath.Walk(folder, func(path string, info os.FileInfo, err error) error {\n\t\t\t// if a file\n\t\t\tif fileinfo, err := os.Stat(path); fileinfo.Mode().IsRegular() && err == nil {\n\t\t\t\tlocalFiles = append(localFiles, path)\n\t\t\t\t// log.Println(\"Added\", path)\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\n\t\tcount := int64(len(localFiles))\n\t\tbar := progressbar.Default(count)\n\n\t\tlog.Println(\"Number of files\", count)\n\n\t\t// Important\n\t\t// or it will become highly nested\n\t\tupOne, err := filepath.Abs(filepath.Join(folder, \"..\"))\n\t\tos.Chdir(upOne)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tfor _, loc := range localFiles {\n\t\t\tbar.Add(1)\n\t\t\t// log.Println(\"DAMNIT\")\n\t\t\trelpath, err := filepath.Rel(upOne, loc)\n\t\t\t// log.Println(relpath)\n\t\t\tif err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t\tif err = addFileToZip(zipWriter, filepath.Join(relpath)); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\tos.Chdir(currDir)\n\t\treturn nil\n\t}\n\treturn errors.New(\"not a directory\")\n}", "func walkArchiveZip(r io.ReaderAt, size int64, f WalkArchiveFunc) error {\n\tzipReader, err := zip.NewReader(r, size)\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar skippedDirPrefixes []string\n\tseenDirs := newSet[string]()\n\tprocessHeader := func(fileInfo fs.FileInfo, dir string) error {\n\t\tfor _, skippedDirPrefix := range skippedDirPrefixes {\n\t\t\tif strings.HasPrefix(dir, skippedDirPrefix) {\n\t\t\t\treturn fs.SkipDir\n\t\t\t}\n\t\t}\n\t\tif seenDirs.contains(dir) {\n\t\t\treturn nil\n\t\t}\n\t\tseenDirs.add(dir)\n\t\tname := strings.TrimSuffix(dir, \"/\")\n\t\tdirFileInfo := implicitDirHeader(dir, fileInfo.ModTime()).FileInfo()\n\t\tswitch err := f(name, dirFileInfo, nil, \"\"); {\n\t\tcase errors.Is(err, fs.SkipDir):\n\t\t\tskippedDirPrefixes = append(skippedDirPrefixes, dir)\n\t\t\treturn err\n\t\tcase err != nil:\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\nFILE:\n\tfor _, zipFile := range zipReader.File {\n\t\tzipFileReader, err := zipFile.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tname := path.Clean(zipFile.Name)\n\t\tif strings.HasPrefix(name, \"../\") || strings.Contains(name, \"/../\") {\n\t\t\treturn fmt.Errorf(\"%s: invalid filename\", zipFile.Name)\n\t\t}\n\n\t\tfor _, skippedDirPrefix := range skippedDirPrefixes {\n\t\t\tif strings.HasPrefix(zipFile.Name, skippedDirPrefix) {\n\t\t\t\tcontinue FILE\n\t\t\t}\n\t\t}\n\n\t\tswitch fileInfo := zipFile.FileInfo(); fileInfo.Mode() & fs.ModeType {\n\t\tcase 0:\n\t\t\tdirs, _ := path.Split(name)\n\t\t\tdirComponents := strings.Split(strings.TrimSuffix(dirs, \"/\"), \"/\")\n\t\t\tfor i := range dirComponents {\n\t\t\t\tdir := strings.Join(dirComponents[0:i+1], \"/\")\n\t\t\t\tif len(dir) > 0 {\n\t\t\t\t\tswitch err := processHeader(fileInfo, dir+\"/\"); {\n\t\t\t\t\tcase errors.Is(err, fs.SkipDir):\n\t\t\t\t\t\tcontinue FILE\n\t\t\t\t\tcase errors.Is(err, fs.SkipAll):\n\t\t\t\t\t\treturn nil\n\t\t\t\t\tcase err != nil:\n\t\t\t\t\t\treturn err\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\t\terr = f(name, fileInfo, zipFileReader, \"\")\n\t\tcase fs.ModeDir:\n\t\t\terr = processHeader(fileInfo, name+\"/\")\n\t\tcase fs.ModeSymlink:\n\t\t\tvar linknameBytes []byte\n\t\t\tlinknameBytes, err = io.ReadAll(zipFileReader)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\terr = f(name, fileInfo, nil, string(linknameBytes))\n\t\t}\n\n\t\terr2 := zipFileReader.Close()\n\n\t\tswitch {\n\t\tcase errors.Is(err, fs.SkipDir):\n\t\t\tskippedDirPrefixes = append(skippedDirPrefixes, zipFile.Name+\"/\")\n\t\tcase errors.Is(err, fs.SkipAll):\n\t\t\treturn nil\n\t\tcase err != nil:\n\t\t\treturn err\n\t\t}\n\n\t\tif err2 != nil {\n\t\t\treturn err2\n\t\t}\n\t}\n\treturn nil\n}", "func zipit(source, target string) error {\n\tzipfile, err := os.Create(target)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer zipfile.Close()\n\n\tarchive := zip.NewWriter(zipfile)\n\tdefer archive.Close()\n\n\tinfo, err := os.Stat(source)\n\tif err != nil {\n\t\treturn nil\n\t}\n\n\tvar baseDir string\n\tif info.IsDir() {\n\t\tbaseDir = filepath.Base(source)\n\t}\n\n\tfilepath.Walk(source, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif baseDir != \"\" {\n\t\t\theader.Name = filepath.Join(baseDir, strings.TrimPrefix(path, source))\n\t\t}\n\n\t\tif info.IsDir() {\n\t\t\theader.Name += \"/\"\n\t\t} else {\n\t\t\theader.Method = zip.Deflate\n\t\t}\n\n\t\twriter, err := archive.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tfile, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\t\t_, err = io.Copy(writer, file)\n\t\treturn err\n\t})\n\n\treturn err\n}", "func Zipper(folderPath string, outputFile string) error {\n\n\toutputPathArray := strings.Split(outputFile, \"/\")\n\toutputFileName := outputPathArray[len(outputPathArray)-1]\n\tif len(outputPathArray) > 1 {\n\t\toutputPathArray = outputPathArray[:len(outputPathArray)-1]\n\t\tcreateOutputPathIfNotExist(outputPathArray)\n\t}\n\n\tif err := checkExtOfOutputFile(outputFileName); err != nil {\n\t\treturn err\n\t}\n\tif err := checkIfFileNotExiste(outputFile); err != nil {\n\t\treturn err\n\t}\n\n\tnewZipFile, err := os.Create(outputFile)\n\tif err != nil {\n\t\tfmt.Println(\"hear\")\n\t\treturn err\n\t}\n\tdefer newZipFile.Close()\n\n\treturn nil\n}", "func downloadZip(rsp http.ResponseWriter, req *http.Request, u *url.URL, dir *os.FileInfo, localPath string) {\n\t// Generate a decent filename based on the folder URL:\n\tfullName := removeIfStartsWith(localPath, jailRoot)\n\tfullName = removeIfStartsWith(fullName, \"/\")\n\t// Translate '/' separators into '-':\n\tfullName = strings.Map(func(i rune) rune {\n\t\tif i == '/' {\n\t\t\treturn '-'\n\t\t} else {\n\t\t\treturn i\n\t\t}\n\t}, fullName)\n\n\tvar fis []os.FileInfo\n\t{\n\t\t// Open the directory to read its contents:\n\t\tdf, err := os.Open(localPath)\n\t\tif err != nil {\n\t\t\tdoError(req, rsp, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t\tdefer df.Close()\n\n\t\t// Read the directory entries:\n\t\tfis, err = df.Readdir(0)\n\t\tif err != nil {\n\t\t\tdoError(req, rsp, err.Error(), http.StatusInternalServerError)\n\t\t\treturn\n\t\t}\n\t}\n\n\t// Make sure filenames are in ascending order:\n\tsort.Sort(ByName{fis, sortAscending})\n\n\t// Start with a 200 status and set up the download:\n\th := rsp.Header()\n\th.Set(\"Content-Type\", \"application/zip\")\n\th.Set(\"Content-Description\", \"File Transfer\")\n\t// NOTE(jsd): Need proper HTTP value encoding here!\n\th.Set(\"Content-Disposition\", \"attachment; filename=\\\"\"+fullName+\".zip\\\"\")\n\th.Set(\"Content-Transfer-Encoding\", \"binary\")\n\n\t// Here we estimate the final length of the ZIP file being streamed:\n\tconst (\n\t\tfileHeaderLen = 30 // + filename + extra\n\t\tdataDescriptorLen = 16 // four uint32: descriptor signature, crc32, compressed size, size\n\t\tdirectoryHeaderLen = 46 // + filename + extra + comment\n\t\tdirectoryEndLen = 22 // + comment\n\t)\n\n\tzipLength := 0\n\tfor _, fi := range fis {\n\t\tif fi.IsDir() {\n\t\t\tcontinue\n\t\t}\n\t\tzipLength += fileHeaderLen\n\t\tzipLength += len(fi.Name())\n\t\t// + extra\n\n\t\t// TODO(jsd): ZIP64 support\n\t\tsize := fi.Size()\n\t\tzipLength += int(size)\n\t\tzipLength += dataDescriptorLen\n\n\t\t// Directory entries:\n\t\tzipLength += directoryHeaderLen\n\t\tzipLength += len(fi.Name())\n\t\t// + extra\n\t\t// + comment\n\t}\n\tzipLength += directoryEndLen\n\n\th.Set(\"Content-Length\", fmt.Sprintf(\"%d\", zipLength))\n\n\trsp.WriteHeader(http.StatusOK)\n\n\t// Create a zip stream writing to the HTTP response:\n\tzw := zip.NewWriter(rsp)\n\tfor _, fi := range fis {\n\t\tif fi.IsDir() {\n\t\t\tcontinue\n\t\t}\n\t\tname := fi.Name()\n\t\tif name[0] == '.' {\n\t\t\tcontinue\n\t\t}\n\n\t\t// Dereference symlinks, if applicable:\n\t\tfi = followSymlink(localPath, fi)\n\t\tfiPath := path.Join(localPath, name)\n\n\t\t// Open the source file for reading:\n\t\tlf, err := os.Open(fiPath)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\tdefer lf.Close()\n\n\t\t// Create the ZIP entry to write to:\n\t\tzfh, err := zip.FileInfoHeader(fi)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\t// Don't bother compressing the file:\n\t\tzfh.Method = zip.Store\n\n\t\tzf, err := zw.CreateHeader(zfh)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\t// Copy the file contents into the ZIP:\n\t\t_, err = io.Copy(zf, lf)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\t// Mark the end of the ZIP stream:\n\tzw.Close()\n\treturn\n}", "func RecursiveZip(pathToZip, destinationPath string) error {\n\tdestinationFile, err := os.Create(destinationPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tmyZip := zip.NewWriter(destinationFile)\n\terr = filepath.Walk(pathToZip, func(filePath string, info os.FileInfo, err error) error {\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\trelPath := strings.TrimPrefix(filePath, filepath.Dir(pathToZip))\n\t\tzipFile, err := myZip.Create(relPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfsFile, err := os.Open(filePath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = io.Copy(zipFile, fsFile)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\terr = myZip.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func zipit(source, target, prefix string) error {\n\tzipfile, err := os.Create(target)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer zipfile.Close()\n\n\tif prefix != \"\" {\n\t\t_, err = io.WriteString(zipfile, prefix)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tarchive := zip.NewWriter(zipfile)\n\tdefer archive.Close()\n\n\terr = filepath.Walk(source, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif path == source {\n\t\t\treturn nil\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader.Name = strings.TrimPrefix(path, source+string(filepath.Separator))\n\n\t\tif info.IsDir() {\n\t\t\theader.Name += string(os.PathSeparator)\n\t\t\theader.SetMode(0755)\n\t\t} else {\n\t\t\theader.Method = zip.Deflate\n\t\t\theader.SetMode(0744)\n\t\t}\n\n\t\twriter, err := archive.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tfile, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\n\t\t_, err = io.Copy(writer, file)\n\t\treturn err\n\t})\n\n\treturn err\n}", "func Unzip(src []byte, dest string) error {\n\tr := openZip(src)\n\tos.MkdirAll(dest, 0755)\n\t// Closure to address file descriptors issue with all the deferred .Close() methods\n\textractAndWriteFile := func(f *zip.File) error {\n\n\t\tpath := filepath.Join(dest, f.Name)\n\t\tisLink := f.FileInfo().Mode()&os.ModeSymlink == os.ModeSymlink\n\n\t\t// dir\n\t\tif f.FileInfo().IsDir() && !isLink {\n\t\t\treturn os.MkdirAll(path, f.Mode())\n\t\t}\n\n\t\t// open file\n\t\trc, err := f.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer rc.Close()\n\n\t\t// link\n\t\tif isLink {\n\t\t\tbuf, err := ioutil.ReadAll(rc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\treturn os.Symlink(string(buf), path)\n\t\t}\n\n\t\t// file\n\t\t// eventually create a missing ddir\n\t\terr = os.MkdirAll(filepath.Dir(path), 0755)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfile, err := os.OpenFile(path, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, f.Mode())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\t\t_, err = io.Copy(file, rc)\n\t\treturn err\n\t}\n\tfor _, f := range r.File {\n\t\terr := extractAndWriteFile(f)\n\t\tif err != nil {\n\t\t\tlog.Println(err)\n\t\t}\n\t}\n\treturn nil\n}", "func Zip(srcFile string, destZip string) error {\n\tzipfile, err := os.Create(destZip)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer zipfile.Close()\n\n\tarchive := zip.NewWriter(zipfile)\n\tdefer archive.Close()\n\n\terr = filepath.Walk(srcFile, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader.Name = path\n\t\tif info.IsDir() {\n\t\t\theader.Name += \"/\"\n\t\t} else {\n\t\t\theader.Method = zip.Deflate\n\t\t}\n\n\t\twriter, err := archive.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif !info.IsDir() {\n\t\t\tfile, err := os.Open(path)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer file.Close()\n\t\t\t_, err = io.Copy(writer, file)\n\t\t}\n\t\treturn err\n\t})\n\n\treturn err\n}", "func zipit(source, target, prefix string) error {\n\tzipfile, err := os.Create(target)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer zipfile.Close()\n\n\tif prefix != \"\" {\n\t\t_, err = io.WriteString(zipfile, prefix)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tarchive := zip.NewWriter(zipfile)\n\tdefer archive.Close()\n\n\terr = filepath.Walk(source, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader.Name = strings.TrimPrefix(path, source)\n\n\t\tif info.IsDir() {\n\t\t\theader.Name += string(os.PathSeparator)\n\t\t} else {\n\t\t\theader.Method = zip.Deflate\n\t\t}\n\n\t\twriter, err := archive.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tfile, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\n\t\t_, err = io.Copy(writer, file)\n\t\treturn err\n\t})\n\n\treturn err\n}", "func ZipFile(filename string, options options.Options) string {\n\tzfilename := filename + \".zip\"\n\tlog.Debugf(\"The file name is \" + zfilename)\n\tnewZipFile, err := os.Create(zfilename)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tdefer newZipFile.Close()\n\n\tzipWriter := zip.NewWriter(newZipFile)\n\tdefer zipWriter.Close()\n\n\tzipfile, err := os.Open(filename)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tdefer zipfile.Close()\n\n\t// Get the file information\n\tinfo, err := zipfile.Stat()\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\n\theader, err := zip.FileInfoHeader(info)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\n\t// Using FileInfoHeader() above only uses the basename of the file. If we want\n\t// to preserve the folder structure we can overwrite this with the full path.\n\theader.Name = strings.Replace(filename, options.Directory, \"\", -1)\n\n\t// Change to deflate to gain better compression\n\t// see http://golang.org/pkg/archive/zip/#pkg-constants\n\theader.Method = zip.Deflate\n\n\twriter, err := zipWriter.CreateHeader(header)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tif _, err = io.Copy(writer, zipfile); err != nil {\n\t\tlog.Error(err)\n\t}\n\treturn zfilename\n}", "func ZipFolder(folder string, zipFile string) error {\n\n\toutFile, err := os.Create(zipFile)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer outFile.Close()\n\n\tw := zip.NewWriter(outFile)\n\n\tzipFilePath := strings.Split(zipFile, \"/\")\n\terr = addFiles(w, folder, \"\", zipFilePath[len(zipFilePath)-1])\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = w.Close()\n\treturn err\n}", "func GitZip(dir, ref string) (io.ReadCloser, error) {\n\treturn GitArchive(dir, ref, \"zip\")\n}", "func makeZip(targ, workdir string) error {\n\tf, err := os.Create(targ)\n\tif err != nil {\n\t\treturn err\n\t}\n\tzw := zip.NewWriter(f)\n\n\terr = filepath.Walk(workdir, func(path string, fi os.FileInfo, err error) error {\n\t\tif !strings.HasPrefix(path, workdir) {\n\t\t\tlog.Panicf(\"walked filename %q doesn't begin with workdir %q\", path, workdir)\n\t\t}\n\t\tname := path[len(workdir):]\n\n\t\t// Convert to Unix-style named paths, as that's the\n\t\t// type of zip file that archive/zip creates.\n\t\tname = strings.Replace(name, \"\\\\\", \"/\", -1)\n\t\t// Chop of any leading / from filename, leftover from removing workdir.\n\t\tif strings.HasPrefix(name, \"/\") {\n\t\t\tname = name[1:]\n\t\t}\n\n\t\tif name == \"\" {\n\t\t\treturn nil\n\t\t}\n\n\t\tfh, err := zip.FileInfoHeader(fi)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfh.Name = name\n\t\tfh.Method = zip.Deflate\n\t\tif fi.IsDir() {\n\t\t\tfh.Name += \"/\" // append trailing slash\n\t\t\tfh.Method = zip.Store // no need to deflate 0 byte files\n\t\t}\n\t\tw, err := zw.CreateHeader(fh)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif fi.IsDir() {\n\t\t\treturn nil\n\t\t}\n\t\tr, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer r.Close()\n\t\t_, err = io.Copy(w, r)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\tif err := zw.Close(); err != nil {\n\t\treturn err\n\t}\n\treturn f.Close()\n}", "func zipr(pathToZip, destinationPath string) error {\n\tdestinationFile, err := os.Create(destinationPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tzipWriter := zip.NewWriter(destinationFile)\n\n\terr = filepath.Walk(pathToZip, func(filePath string, info os.FileInfo, err error) error {\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\trelPath := strings.TrimPrefix(filePath, pathToZip)\n\t\trelPath = strings.TrimPrefix(relPath, \"/\") // remove the slash if it exists\n\n\t\tfmt.Print(\" adding \" + relPath)\n\n\t\tfileToZip, err := os.Open(filePath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader.Name = relPath // preserve the relative path\n\t\theader.Method = zip.Deflate // compress the file\n\n\t\twriter, err := zipWriter.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\twritten, err := io.Copy(writer, fileToZip)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// TODO print the compression ratio (original size / compressed size)\n\t\tcompressedSize := 1.0 // TODO calculate\n\t\tcompressionRatio := compressedSize / float64(written) // TODO handle file size 0\n\t\tfmt.Printf(\" (compression %0.f%%)\\n\", compressionRatio)\n\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = zipWriter.Close()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func processArchive(path string) {\r\n\text := filepath.Ext(path)\r\n\r\n\t// -o:outputpath\r\n\textractpath := strings.TrimSpace(path[:len(path)-len(ext)])\r\n\r\n\tcmd := exec.Command(zipApp, \"x\", \"-o:\"+extractpath, path)\r\n\r\n\texecErr := cmd.Run()\r\n\tif execErr != nil {\r\n\t\tpanic(execErr)\r\n\t}\r\n\r\n\tprocessDirectory(extractpath)\r\n\r\n\tos.RemoveAll(extractpath)\r\n}", "func (o LayerVersionOutput) ZipFile() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *LayerVersion) pulumi.StringPtrOutput { return v.ZipFile }).(pulumi.StringPtrOutput)\n}", "func (t *Tar) Compress(directory files.Directory, writer *bytes.Buffer) error {\n\tgzipWriter, err := gzip.NewWriterLevel(writer, flate.BestCompression)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttarWriter := tar.NewWriter(gzipWriter)\n\n\tfiles.WalkFileTree(directory, func(file files.File) {\n\t\tbuffer := &bytes.Buffer{}\n\t\tif err := file.CopyContent(buffer); err != nil {\n\t\t\treturn\n\t\t}\n\n\t\ttarHeader := &tar.Header{\n\t\t\tName: filepath.ToSlash(file.AbsolutePath().String()),\n\t\t\tMode: int64(file.PermissionSet()),\n\t\t\tSize: int64(buffer.Len()),\n\t\t\tTypeflag: tar.TypeReg,\n\t\t}\n\n\t\tif err := tarWriter.WriteHeader(tarHeader); err != nil {\n\t\t\treturn\n\t\t}\n\t\tif _, err := tarWriter.Write(buffer.Bytes()); err != nil {\n\t\t\treturn\n\t\t}\n\t})\n\n\tfiles.WalkDirectoryTree(directory, func(d files.Directory) {\n\t\ttarHeader := &tar.Header{\n\t\t\tName: filepath.ToSlash(d.AbsolutePath().String()),\n\t\t\tMode: int64(d.PermissionSet()),\n\t\t\tTypeflag: tar.TypeDir,\n\t\t}\n\n\t\tif err := tarWriter.WriteHeader(tarHeader); err != nil {\n\t\t\treturn\n\t\t}\n\t})\n\n\tif err := gzipWriter.Close(); err != nil {\n\t\treturn err\n\t}\n\n\treturn tarWriter.Close()\n}", "func UnZipFile(filename string, destination string) string {\n\tlog.Debugf(\"Unzipping file %s\", filename)\n\treturnFn := filename\n\tif !strings.HasSuffix(filename, \".zip\") {\n\t\tlog.Warnf(\"Skipping file because it is not a zip file, %s\", filename)\n\t\treturn returnFn\n\t}\n\n\tzReader, err := zip.OpenReader(filename)\n\tif err != nil {\n\t\tlog.Error(err)\n\t}\n\tdefer zReader.Close()\n\tfor _, file := range zReader.Reader.File {\n\n\t\tzippedFile, err := file.Open()\n\t\tif err != nil {\n\t\t\tlog.Fatal(err)\n\t\t}\n\t\tdefer zippedFile.Close()\n\t\tlog.Debugf(\"this is the files name from zreader \" + file.Name)\n\t\textractedFilePath := filepath.Join(\n\t\t\tdestination,\n\t\t\tfile.Name,\n\t\t)\n\t\tlog.Debugf(\"\\tExtracted path: %s\", extractedFilePath)\n\t\tif file.FileInfo().IsDir() {\n\t\t\tlog.Println(\"Directory Created:\", extractedFilePath)\n\t\t\tos.MkdirAll(extractedFilePath, file.Mode())\n\t\t} else {\n\t\t\tlog.Println(\"\\tFile extracted:\", file.Name)\n\n\t\t\textractDir := filepath.Dir(extractedFilePath)\n\t\t\tos.MkdirAll(extractDir, os.ModePerm)\n\t\t\toutputFile, err := os.OpenFile(\n\t\t\t\textractedFilePath,\n\t\t\t\tos.O_WRONLY|os.O_CREATE|os.O_TRUNC,\n\t\t\t\tfile.Mode(),\n\t\t\t)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\n\t\t\t_, err = io.Copy(outputFile, zippedFile)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t}\n\t\t\treturnFn = outputFile.Name()\n\t\t\toutputFile.Close()\n\t\t}\n\t}\n\tlog.Debugf(\"\\tUnzip returning file name %s\", returnFn)\n\treturn returnFn\n}", "func (l *Local) Unzip(archive, target, name string) error {\n\ttd := filepath.Join(target, name)\n\tif err := os.MkdirAll(td, 0755); err != nil {\n\t\treturn xerrors.Errorf(\"Unable to create target directory: %w\", err)\n\t}\n\n\tcmd := exec.Command(\"unzip\", \"-qq\", archive, \"-d\", td)\n\n\tif err := cmd.Run(); err != nil {\n\t\treturn xerrors.Errorf(\"Unable to unzip archive: %w\", err)\n\t}\n\n\tcmd.ProcessState.Sys()\n\treturn nil\n}", "func Unzip(src, dest string) (err error) {\n\tdest = filepath.Clean(dest) + string(os.PathSeparator)\n\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer fs.CheckClose(r, &err)\n\n\tif err := file.MkdirAll(dest, 0755); err != nil {\n\t\treturn err\n\t}\n\n\t// Closure to address file descriptors issue with all the deferred .Close() methods\n\textractAndWriteFile := func(f *zip.File) error {\n\t\tpath := filepath.Join(dest, f.Name)\n\t\t// Check for Zip Slip: https://github.com/rclone/rclone/issues/3529\n\t\tif !strings.HasPrefix(path, dest) {\n\t\t\treturn fmt.Errorf(\"%s: illegal file path\", path)\n\t\t}\n\n\t\trc, err := f.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer fs.CheckClose(rc, &err)\n\n\t\tif f.FileInfo().IsDir() {\n\t\t\tif err := file.MkdirAll(path, 0755); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\tif err := file.MkdirAll(filepath.Dir(path), 0755); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tf, err := file.OpenFile(path, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0644)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer fs.CheckClose(f, &err)\n\n\t\t\t_, err = io.Copy(f, rc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\tfor _, f := range r.File {\n\t\terr := extractAndWriteFile(f)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func Zip(src string, dest string) error {\n\tdestinationFile, err := os.Create(dest)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tzipDestFile := zip.NewWriter(destinationFile)\n\tif err := filepath.Walk(src, func(filePath string, info os.FileInfo, err error) error {\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tif filePath == dest {\n\t\t\treturn nil\n\t\t}\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\trelPath := strings.TrimPrefix(filePath, src)\n\t\tzipFile, err := zipDestFile.Create(relPath)\n\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tfsFile, err := os.Open(filePath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = io.Copy(zipFile, fsFile)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := fsFile.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t}); err != nil {\n\t\treturn err\n\t}\n\n\tif err := zipDestFile.Close(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func ExtractZip(source string, destination string, stripComponents int) error {\n\tz, err := zip.OpenReader(source)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer z.Close()\n\n\tfor _, f := range z.File {\n\t\ttarget := strippedPath(f.Name, destination, stripComponents)\n\t\tif target == \"\" {\n\t\t\tcontinue\n\t\t}\n\n\t\tif f.FileInfo().IsDir() {\n\t\t\tif err := os.MkdirAll(target, 0755); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\tif err := writeFile(f, target); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func createZipFile(t *testing.T) []byte {\n\t// Create a buffer to write our archive to.\n\tbuf := new(bytes.Buffer)\n\n\t// Create a new zip archive.\n\tw := zip.NewWriter(buf)\n\n\t// Add some files to the archive.\n\tfor _, file := range files {\n\t\tf, err := w.Create(file.Name)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Unexpected error %v\", err)\n\t\t}\n\t\t_, err = f.Write([]byte(file.Body))\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Unexpected error %v\", err)\n\t\t}\n\t}\n\n\t// Make sure to check the error on Close.\n\terr := w.Close()\n\tif err != nil {\n\t\tt.Errorf(\"Unexpected error %v\", err)\n\t}\n\treturn buf.Bytes()\n}", "func TarDirectory(tw *tar.Writer, dir string) error {\n\treturn filepath.Walk(dir, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\thdr, err := tar.FileInfoHeader(info, path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\thdr.Name = path[len(dir)+1:]\n\t\tif err := tw.WriteHeader(hdr); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfi, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = io.Copy(tw, fi)\n\t\treturn err\n\t})\n}", "func Compress(path, target string) error {\n\t//TODO this is...dense. cyclomatic complexity >10\n\t//TODOAY we refactor this to use go channels\n\n\t//Verify directory exists\n\ts, err := os.Stat(path)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"fs: compress failed to detect archive from path\")\n\t}\n\n\t//TODO i'm not convinced this will work on individual files sooooo...\n\tif !s.IsDir() {\n\t\treturn errors.New(\"fs: compress operation requires path to a directory\")\n\t}\n\n\t//Get the archives parent for a default storage location\n\tparentPath, err := filepath.Abs(target)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"fs: compress failed to extract absolute path of archive\")\n\t}\n\n\t//Open the zip archive for writing\n\tarchiveBuffer, err := os.OpenFile(parentPath+string(os.PathSeparator)+s.Name()+\".zip\", os.O_RDWR|os.O_CREATE, 0755)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"fs: compress failed to open \"+archiveBuffer.Name()+\".zip\")\n\t}\n\n\t//Initialize compression writer\n\tz := zip.NewWriter(archiveBuffer)\n\n\t//walk the provided directory\n\tw := walker.New(path)\n\tif err := w.Walk(); err != nil {\n\t\treturn errors.Wrap(err, \"fs: compress failed to walk archive\")\n\t}\n\n\tfor _, file := range w.Archive() {\n\t\t//Get the files relative path in the archive\n\t\trelPath, err := filepath.Rel(path, file)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"fs: compress failed to extract relative file path in archive\")\n\t\t}\n\t\t//Create zip file buffer for compression storage\n\t\tzipFile, err := z.Create(s.Name() + string(os.PathSeparator) + relPath)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"fs: compress failed to create zip file \"+relPath)\n\t\t}\n\t\t//Open file for copying\n\t\tf, err := os.Open(file)\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"fs: compress failed to open \"+file)\n\t\t}\n\n\t\t//copy file into zip archive\n\t\tif _, err := io.Copy(zipFile, f); err != nil {\n\t\t\treturn errors.Wrap(err, \"fs: failed to write file to zip folder\")\n\t\t}\n\t\t//close file opened in iteration\n\t\tif err := f.Close(); err != nil {\n\t\t\treturn errors.Wrap(err, \"fs: compress failed to close \"+file)\n\t\t}\n\t}\n\n\t//close zip archive\n\tif err := z.Close(); err != nil {\n\t\treturn errors.Wrap(err, \"fs: compress failed to close zip writer\")\n\t}\n\t//close archive buffer\n\tif err := archiveBuffer.Close(); err != nil {\n\t\treturn errors.Wrap(err, \"fs: compress failed to close archive buffer\")\n\t}\n\n\treturn nil\n\n}", "func dumpZip(tree Stream, path string, t *testing.T) {\n\tfp, err := os.Create(path)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif err := tree.ToZip(zip.NewWriter(fp)); err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func Archive(inFilePath string, writer io.Writer, progress ProgressFunc) (err error) {\n\tvar zipWriter *zip.Writer\n\tvar zipFileWriter io.Writer\n\tvar zipHeader *zip.FileHeader\n\tvar archivePath, relativeFilePath, basePath string\n\tvar file *os.File\n\n\tzipWriter = zip.NewWriter(writer)\n\n\tbasePath = filepath.Dir(inFilePath)\n\n\terr = filepath.Walk(inFilePath, func(filePath string, fileInfo os.FileInfo, walkErr error) (err error) {\n\t\tif walkErr == nil {\n\t\t\tif !fileInfo.IsDir() {\n\n\t\t\t\trelativeFilePath, err = filepath.Rel(basePath, filePath)\n\t\t\t\tif err == nil {\n\n\t\t\t\t\tarchivePath = filepath.ToSlash(relativeFilePath)\n\n\t\t\t\t\tif progress != nil {\n\t\t\t\t\t\tprogress(archivePath)\n\t\t\t\t\t}\n\n\t\t\t\t\tzipHeader, err = zip.FileInfoHeader(fileInfo)\n\t\t\t\t\tif err == nil {\n\t\t\t\t\t\tzipHeader.Name = archivePath\n\t\t\t\t\t\tzipHeader.Method = zip.Deflate\n\t\t\t\t\t\t// fmt.Printf(\"archive path [%s]\\n\", archivePath)\n\t\t\t\t\t\tzipFileWriter, err = zipWriter.CreateHeader(zipHeader)\n\t\t\t\t\t\tif err == nil {\n\t\t\t\t\t\t\tfile, err = os.Open(filePath)\n\t\t\t\t\t\t\tif err == nil {\n\t\t\t\t\t\t\t\t_, err = io.Copy(zipFileWriter, file)\n\t\t\t\t\t\t\t\tfile.Close()\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\terr = walkErr\n\t\t}\n\t\treturn\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn zipWriter.Close()\n}", "func Unzip(src, dest string) error {\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer func() {\n\t\tif err := r.Close(); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}()\n\n\tos.MkdirAll(dest, 0755)\n\n\t// Closure to address file descriptors issue with all the deferred .Close() methods\n\textractAndWriteFile := func(f *zip.File) error {\n\t\trc, err := f.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer func() {\n\t\t\tif err := rc.Close(); err != nil {\n\t\t\t\tpanic(err)\n\t\t\t}\n\t\t}()\n\n\t\tthePath := filepath.Join(dest, f.Name)\n\n\t\tif f.FileInfo().IsDir() {\n\t\t\tos.MkdirAll(thePath, f.Mode())\n\t\t} else {\n\t\t\tos.MkdirAll(filepath.Dir(thePath), f.Mode())\n\t\t\tf, err := os.OpenFile(thePath, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, f.Mode())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tdefer func() {\n\t\t\t\tif err := f.Close(); err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t}()\n\n\t\t\t_, err = io.Copy(f, rc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tfor _, f := range r.File {\n\t\terr := extractAndWriteFile(f)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func ZipFiles(path string, output string, files []string) (string, error) {\n\n\tzipStorage := fmt.Sprintf(\"%s/.cfbuilds\", path)\n\n\tif _, err := os.Stat(zipStorage); os.IsNotExist(err) {\n\t\tos.Mkdir(zipStorage, 0755)\n\t}\n\n\toutPutZipDirectory := fmt.Sprintf(\"%s/%s.zip\", zipStorage, output)\n\tnewZip, err := os.Create(outPutZipDirectory)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tdefer newZip.Close()\n\n\tzipWriter := zip.NewWriter(newZip)\n\tdefer zipWriter.Close()\n\n\tbaseDir := filepath.Base(path)\n\n\tfor _, file := range files {\n\n\t\tif err = addFileToZip(zipWriter, file, baseDir); err != nil {\n\t\t\tfmt.Println(err)\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\n\treturn outPutZipDirectory, nil\n}", "func (p *CLIPacker) packFiles(w *zip.Writer, filePath string, zipPath string) error {\n\tvar bytes int\n\tfiles, err := ioutil.ReadDir(filePath)\n\n\tif err == nil {\n\t\tfor _, file := range files {\n\t\t\tif !file.IsDir() {\n\t\t\t\tvar data []byte\n\t\t\t\tdiskfile := filepath.Join(filePath, file.Name())\n\t\t\t\tif data, err = ioutil.ReadFile(diskfile); err == nil {\n\t\t\t\t\tvar f io.Writer\n\t\t\t\t\tif f, err = w.Create(path.Join(zipPath, file.Name())); err == nil {\n\t\t\t\t\t\tif bytes, err = f.Write(data); err == nil {\n\t\t\t\t\t\t\tfmt.Fprintln(p.LogOut, fmt.Sprintf(\"Writing %v bytes for %v\",\n\t\t\t\t\t\t\t\tbytes, diskfile))\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else if file.IsDir() {\n\t\t\t\t// Path separator in zipfile is always '/'\n\t\t\t\tp.packFiles(w, filepath.Join(filePath, file.Name()),\n\t\t\t\t\tpath.Join(zipPath, file.Name()))\n\t\t\t}\n\t\t}\n\t}\n\n\treturn err\n}", "func UnzipImpl(reader *zip.Reader, dest string, verbose bool) error {\n for _, f := range reader.File {\n zipped, err := f.Open()\n if err != nil {\n return errors.New(\"unzip: Unable to open [\" + f.Name + \"]\")\n }\n\n defer zipped.Close()\n\n path := filepath.Join(dest, f.Name)\n if f.FileInfo().IsDir() {\n os.MkdirAll(path, f.Mode())\n if verbose {\n fmt.Println(\"Creating directory\", path)\n }\n } else {\n // Ensure we create the parent folder\n err := os.MkdirAll(filepath.Dir(path), os.ModePerm)\n if err != nil {\n return errors.New(\"unzip: Unable to create parent folder [\" + path + \"]\")\n }\n\n writer, err := os.OpenFile(path, os.O_WRONLY|os.O_CREATE, f.Mode())\n if err != nil {\n return errors.New(\"unzip: Unable to create [\" + path + \"]\")\n }\n\n defer writer.Close()\n\n if _, err = io.Copy(writer, zipped); err != nil {\n return errors.New(\"unzip: Unable to create content in [\" + path + \"]\")\n }\n\n if verbose {\n fmt.Println(\"Decompressing : \", path)\n }\n }\n }\n return nil\n}", "func unzip(src string) error {\n r, err := zip.OpenReader(src)\n if err != nil {\n return err\n }\n defer func() {\n if err := r.Close(); err != nil {\n panic(err)\n }\n }()\n dest := filepath.Dir(src)\n\n // Closure to address file descriptors issue with all the deferred .Close() methods\n extractAndWriteFile := func(f *zip.File) error {\n rc, err := f.Open()\n if err != nil {\n return err\n }\n defer func() {\n if err := rc.Close(); err != nil {\n panic(err)\n }\n }()\n path := filepath.Join(dest, f.Name)\n\n if f.FileInfo().IsDir() {\n os.MkdirAll(path, f.Mode())\n } else {\n f, err := os.OpenFile(path, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, f.Mode())\n if err != nil {\n return err\n }\n defer func() {\n if err := f.Close(); err != nil {\n panic(err)\n }\n }()\n _, err = io.Copy(f, rc)\n if err != nil {\n return err\n }\n }\n return nil\n }\n\n for _, f := range r.File {\n err := extractAndWriteFile(f)\n if err != nil {\n return err\n }\n }\n\n return nil\n}", "func Untargzip(path string, r io.Reader) error {\n\tungzipped, err := gzip.NewReader(r)\n\tif err != nil {\n\t\treturn err\n\t}\n\ttarball := tar.NewReader(ungzipped)\n\n\tdefer ungzipped.Close()\n\n\t// We have to treat things differently for git-archives\n\tisGitArchive := false\n\n\t// Alright, things seem in order, let's make the base directory\n\tos.MkdirAll(path, 0755)\n\tfor {\n\t\thdr, err := tarball.Next()\n\t\tif err == io.EOF {\n\t\t\t// finished the tar\n\t\t\tbreak\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Skip the base dir\n\t\tif hdr.Name == \"./\" {\n\t\t\tcontinue\n\t\t}\n\n\t\t// If this was made with git-archive it will be in kinda an ugly\n\t\t// format, but we can identify it by the pax_global_header \"file\"\n\t\tname := hdr.Name\n\t\tif name == \"pax_global_header\" {\n\t\t\tisGitArchive = true\n\t\t\tcontinue\n\t\t}\n\n\t\t// It will also contain an extra subdir that we will automatically strip\n\t\tif isGitArchive {\n\t\t\tparts := strings.Split(name, \"/\")\n\t\t\tname = strings.Join(parts[1:], \"/\")\n\t\t}\n\n\t\tfpath := filepath.Join(path, name)\n\t\tif hdr.FileInfo().IsDir() {\n\t\t\terr = os.MkdirAll(fpath, 0755)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tfile, err := os.OpenFile(fpath, os.O_WRONLY|os.O_CREATE, hdr.FileInfo().Mode())\n\t\tdefer file.Close()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = io.Copy(file, tarball)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfile.Close()\n\t}\n\treturn nil\n}", "func ZipFiles(filename string, files []string) error {\n\t// Make the parent directory\n\tfilenameParts := strings.Split(filename, \"/\")\n\tparentDir := filenameParts[len(filenameParts)-2]\n\n\tfmt.Println(\"creating directory if not exists \" + parentDir)\n\n\terr := os.MkdirAll(fmt.Sprintf(\"%s/%s\", strings.Join(filenameParts[:len(filenameParts)-2], \"/\"), parentDir), 0755)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnewZipFile, err := os.Create(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer newZipFile.Close()\n\n\tzipWriter := zip.NewWriter(newZipFile)\n\tdefer zipWriter.Close()\n\n\t// Add files to zip\n\tfor _, file := range files {\n\t\tif err = AddFileToZip(zipWriter, file); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (c *cloner) zipPath() (string, error) {\n\tzipFiles, err := c.listZipFiles()\n\tif err != nil {\n\t\treturn \"\", nil\n\t}\n\tcacheCandidates := zipFiles\n\tif c.noCache {\n\t\tcacheCandidates = nil\n\t}\n\tzipPath, cacheHit, err := c.downloadZip(cacheCandidates)\n\tif err != nil {\n\t\tif cacheHit {\n\t\t\tc.cli.printWarning(err)\n\t\t} else {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\tif cacheHit {\n\t\tlog.Print(color.YellowString(\"Using cached sample apps ...\"))\n\t}\n\t// Remove obsolete files\n\tfor _, zf := range zipFiles {\n\t\tif zf.path != zipPath {\n\t\t\tos.Remove(zf.path)\n\t\t}\n\t}\n\treturn zipPath, nil\n}", "func (z *zipMaker) Zip(w http.ResponseWriter, r *http.Request, nameZip string, files []S3NameAndObjectID) error {\n\tw.Header().Add(\"Content-Disposition\", \"attachment; filename=\\\"\"+nameZip+\"\\\"\")\n\tw.Header().Add(\"Content-Type\", \"application/zip\")\n\n\t// Loop over files, add them to the\n\tzipWriter := zip.NewWriter(w)\n\tfor _, file := range files {\n\n\t\tif file.ObjectID == \"\" {\n\t\t\tlog.Printf(\"Missing path for file: %v\", file)\n\t\t\tcontinue\n\t\t}\n\n\t\t// Build safe file file name\n\t\tsafeFileName := makeSafeFileName.ReplaceAllString(file.Name, \"\")\n\t\tif safeFileName == \"\" { // Unlikely but just in case\n\t\t\tsafeFileName = \"file\"\n\t\t}\n\n\t\t// Read file from S3, log any errors\n\t\trdr, err := z.aws_bucket.GetReader(file.ObjectID)\n\t\tif err != nil {\n\t\t\tswitch t := err.(type) {\n\t\t\tcase *s3.Error:\n\t\t\t\tif t.StatusCode == 404 {\n\t\t\t\t\tlog.Printf(\"File not found. %s\", file.ObjectID)\n\t\t\t\t}\n\t\t\tdefault:\n\t\t\t\tlog.Printf(\"Error downloading \\\"%s\\\" - %s\", file.ObjectID, err.Error())\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\t// Build a good path for the file within the zip\n\t\tzipPath := safeFileName\n\t\t// zipPath := \"\"\n\t\t// // Prefix project Id and name, if any (remove if you don't need)\n\t\t// if file.ProjectId > 0 {\n\t\t// \tzipPath += strconv.FormatInt(file.ProjectId, 10) + \".\"\n\t\t// \t// Build Safe Project Name\n\t\t// \tfile.ProjectName = makeSafeFileName.ReplaceAllString(file.ProjectName, \"\")\n\t\t// \tif file.ProjectName == \"\" { // Unlikely but just in case\n\t\t// \t\tfile.ProjectName = \"Project\"\n\t\t// \t}\n\t\t// \tzipPath += file.ProjectName + \"/\"\n\t\t// }\n\t\t// // Prefix folder name, if any\n\t\t// if file.Folder != \"\" {\n\t\t// \tzipPath += file.Folder\n\t\t// \tif !strings.HasSuffix(zipPath, \"/\") {\n\t\t// \t\tzipPath += \"/\"\n\t\t// \t}\n\t\t// }\n\t\t// zipPath += safeFileName\n\n\t\t// We have to set a special flag so zip files recognize utf file names\n\t\t// See http://stackoverflow.com/questions/30026083/creating-a-zip-archive-with-unicode-filenames-using-gos-archive-zip\n\t\th := &zip.FileHeader{\n\t\t\tName: zipPath,\n\t\t\tMethod: zip.Deflate,\n\t\t\tFlags: 0x800,\n\t\t}\n\n\t\t// if file.Modified != \"\" {\n\t\t// \th.SetModTime(file.ModifiedTime)\n\t\t// }\n\n\t\tf, _ := zipWriter.CreateHeader(h)\n\n\t\tio.Copy(f, rdr)\n\t\trdr.Close()\n\t}\n\n\tzipWriter.Close()\n\n\t// log.Printf(\"%s\\t%s\\t%s\", r.Method, r.RequestURI, time.Since(start))\n\treturn nil\n}", "func zip(source, zipfilename string) (int64, error) {\n\tfn := fmt.Sprintf(\"zip(%s, %s)\", source, zipfilename) // keep original args for errmsg\n\tsource, err := filepath.Abs(source)\n\tif err != nil {\n\t\tlog.Printf(\"%s: Failed to get abs path for source directory in function '%s': %v\", lpStorage, fn, err)\n\t\treturn -1, err\n\t}\n\n\tzipfilename, err = filepath.Abs(zipfilename)\n\tif err != nil {\n\t\tlog.Printf(\"%s: Failed to get abs path for target zip file in function '%s': %v\", lpStorage, fn, err)\n\t\treturn -1, err\n\t}\n\n\t// Create zip file IO writer for MakeZip function\n\tzipfp, err := os.Create(zipfilename)\n\tif err != nil {\n\t\tlog.Printf(\"%s: Failed to create zip file for writing in function '%s': %v\", lpStorage, fn, err)\n\t\treturn -1, err\n\t}\n\tdefer zipfp.Close()\n\t// Change into clone directory to make the paths in the zip archive repo\n\t// root-relative.\n\torigdir, err := os.Getwd()\n\tif err != nil {\n\t\tlog.Printf(\"%s: Failed to get working directory in function '%s': %v\", lpStorage, fn, err)\n\t\treturn -1, err\n\t}\n\tdefer os.Chdir(origdir)\n\tif err := os.Chdir(source); err != nil {\n\t\tlog.Printf(\"%s: Failed to change to source directory to make zip file in function '%s': %v\", lpStorage, fn, err)\n\t\treturn -1, err\n\t}\n\n\tif err := archive.MakeZip(zipfp, \".\"); err != nil {\n\t\tlog.Printf(\"%s: Failed to create zip file in function '%s': %v\", lpStorage, fn, err)\n\t\treturn -1, err\n\t}\n\n\tstat, _ := zipfp.Stat()\n\treturn stat.Size(), nil\n}", "func Extract(log esl.Logger, archivePath, destPath string) error {\n\tl := log.With(esl.String(\"archivePath\", archivePath), esl.String(\"destPath\", destPath))\n\tdestAbsPath, err := filepath.Abs(destPath)\n\tif err != nil {\n\t\tl.Debug(\"Unable to compute abs dest path\", esl.Error(err))\n\t\treturn err\n\t}\n\n\tzr, err := zip.OpenReader(archivePath)\n\tif err != nil {\n\t\tl.Debug(\"Unable to open the archive\")\n\t\treturn err\n\t}\n\tdefer func() {\n\t\t_ = zr.Close()\n\t}()\n\n\textractFile := func(zf *zip.File) error {\n\t\tif zf.FileInfo().IsDir() {\n\t\t\tl.Debug(\"Skip folder\", esl.String(\"name\", zf.Name))\n\t\t\treturn nil\n\t\t}\n\n\t\tfileFolder := filepath.Join(destAbsPath, filepath.Dir(zf.Name))\n\t\tll := l.With(esl.String(\"fileFolder\", fileFolder), esl.String(\"filePath\", zf.Name))\n\n\t\tzr, err := zf.Open()\n\t\tif err != nil {\n\t\t\tl.Debug(\"Unable to read the file\", esl.Error(err))\n\t\t\treturn err\n\t\t}\n\t\tdefer func() {\n\t\t\t_ = zr.Close()\n\t\t}()\n\n\t\tfileFolderInfo, err := os.Lstat(fileFolder)\n\t\tswitch {\n\t\tcase os.IsNotExist(err):\n\t\t\tll.Debug(\"Try create a folder\")\n\t\t\tif err = os.MkdirAll(fileFolder, 0755); err != nil {\n\t\t\t\tll.Debug(\"Unable to create the folder\", esl.Error(err))\n\t\t\t\treturn err\n\t\t\t}\n\t\tcase fileFolderInfo != nil && !fileFolderInfo.IsDir():\n\t\t\tll.Debug(\"Path conflict with a file\")\n\t\t\treturn ErrorConflict\n\t\tcase err == nil:\n\t\t\tll.Debug(\"The folder found\")\n\t\tdefault:\n\t\t\tll.Debug(\"Unable to determine the folder\", esl.Error(err))\n\t\t\treturn err\n\t\t}\n\n\t\tfilePath := filepath.Join(destAbsPath, zf.Name)\n\t\tf, err := os.Create(filePath)\n\t\tif err != nil {\n\t\t\tll.Debug(\"Unable to create the file\", esl.Error(err))\n\t\t\treturn err\n\t\t}\n\t\tdefer func() {\n\t\t\t_ = f.Close()\n\t\t}()\n\n\t\tsize, err := io.Copy(f, zr)\n\t\tif err != nil {\n\t\t\tll.Debug(\"Unable to copy\", esl.Error(err))\n\t\t\treturn err\n\t\t}\n\t\tll.Debug(\"Extract completed\", esl.Int64(\"size\", size))\n\t\treturn nil\n\t}\n\n\tfor _, zf := range zr.File {\n\t\t// CWE-22, https://cwe.mitre.org/data/definitions/22.html\n\t\tif strings.Contains(zf.Name, \"..\") {\n\t\t\tl.Debug(\"Skip the entry that contains '..' to avoid CWE-22\", esl.String(\"name\", zf.Name))\n\t\t\tcontinue\n\t\t}\n\t\tif err := extractFile(zf); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func Unzip(src, destdir string) (paths []string, err error) {\n\tif err = os.MkdirAll(destdir, 0774); err != nil {\n\t\treturn nil, err\n\t}\n\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer r.Close()\n\n\tpaths = make([]string, 0)\n\n\tfor _, f := range r.File {\n\t\tif mgnstr.ContainsAny(f.Name, Excludes) {\n\t\t\tcontinue\n\t\t}\n\n\t\trc, err := f.Open()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdefer rc.Close()\n\n\t\t// dir無しでくるときがあるのでつくってしまう\n\t\ttpath := filepath.Join(destdir, f.Name)\n\t\tos.MkdirAll(path.Dir(tpath), 0777)\n\n\t\tif f.FileInfo().IsDir() {\n\t\t\tos.MkdirAll(tpath, f.Mode())\n\t\t} else {\n\t\t\tpaths = append(paths, tpath)\n\n\t\t\tfo, err := os.OpenFile(\n\t\t\t\ttpath, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, f.Mode())\n\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tdefer fo.Close()\n\n\t\t\tif _, err = io.Copy(fo, rc); err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn paths, nil\n}", "func zipFiles(source, destination string) error {\n\tbase := \"\"\n\n\tzipFile, err := os.Create(destination)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer zipFile.Close()\n\n\tw := zip.NewWriter(zipFile)\n\tdefer w.Close()\n\n\tf, err := os.Stat(source)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif f.IsDir() {\n\t\tbase = filepath.Base(source)\n\t}\n\n\terr = filepath.Walk(source, func(path string, f os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\theader, err := zip.FileInfoHeader(f)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif base != \"\" {\n\t\t\theader.Name = filepath.Join(base, strings.TrimPrefix(path, source))\n\t\t}\n\n\t\tif f.IsDir() {\n\t\t\theader.Name += \"/\"\n\t\t} else {\n\t\t\theader.Method = zip.Deflate\n\t\t}\n\n\t\twriter, err := w.CreateHeader(header)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif f.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tfile, err := os.Open(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer file.Close()\n\t\t_, err = io.Copy(writer, file)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func unzip(src string, dest string) error {\n\tzipReader, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\ttargetDir := dest\n\tfor _, file := range zipReader.Reader.File {\n\t\textractedFilePath := filepath.Join(\n\t\t\ttargetDir,\n\t\t\tfile.Name,\n\t\t)\n\t\tif file.FileInfo().IsDir() {\n\t\t\tif err := os.MkdirAll(extractedFilePath, file.Mode()); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\tif err := unzipFile(extractedFilePath, file); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func FromZip(src, dir string) error {\n\tz, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif err = os.MkdirAll(dir, 0755); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, file := range z.File {\n\t\tpath := filepath.Join(dir, file.Name)\n\t\tif file.FileInfo().IsDir() {\n\t\t\tif err = os.MkdirAll(path, file.Mode()); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tfileReader, err := file.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\ttargetFile, err := os.OpenFile(path, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, file.Mode())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif _, err := io.Copy(targetFile, fileReader); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err = fileReader.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err = targetFile.Close(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func Compress(dirName string) (string, error) {\n\tdir, err := os.Open(dirName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer dir.Close()\n\n\tdirInfo, err := dir.Stat()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tif !dirInfo.IsDir() {\n\t\treturn \"\", errors.New(\"not a dir: \" + \"dirName\")\n\t}\n\n\tfiles, err := dir.Readdir(0)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif len(files) == 0 {\n\t\tlog.Warning.Println(\"dir=\", dirName, \"is empty\")\n\t\treturn \"\", errors.New(\"dir is empty\")\n\t}\n\n\tgzFileName := dirName + \".tgz\"\n\ttgzfile, err := os.Create(gzFileName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer tgzfile.Close()\n\n\tvar fileWriter io.WriteCloser = tgzfile\n\tarchiver := gzip.NewWriter(fileWriter)\n\tarchiver.Name = gzFileName\n\tdefer archiver.Close()\n\n\ttarfileWriter := tar.NewWriter(archiver)\n\tdefer tarfileWriter.Close()\n\n\tfor _, fileInfo := range files {\n\t\tif fileInfo.IsDir() {\n\t\t\treturn \"\", errors.New(\"unexpected dir in\" + dirName)\n\t\t}\n\t\tfile, err := os.Open(filepath.Join(dir.Name(), fileInfo.Name()))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tdefer file.Close()\n\n\t\theader, err := tar.FileInfoHeader(fileInfo, fileInfo.Name())\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\terr = tarfileWriter.WriteHeader(header)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\t_, err = io.Copy(tarfileWriter, file)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\treturn gzFileName, nil\n}", "func GetArchiveZip(w http.ResponseWriter, r *http.Request) {\n\tstrFiles, _ := requestedFilesToRelativePaths(r.URL.Query()[\"f\"]) // TODO handle error\n\n\tpr, pw := io.Pipe()\n\tgo func() {\n\t\tif err := files.ZipWithPipes(strFiles, pw); err != nil {\n\t\t\tlog.Println(\"failed to create archive\", err.Error())\n\t\t}\n\t\tdefer pw.Close()\n\t}()\n\n\tw.Header().Add(conf.HeaderContentType, \"application/zip\")\n\tw.Header().Set(conf.HeaderContentDisposition, \"attachment; filename=archive.zip\")\n\n\twritePipeContent(w, pr)\n}", "func AddDirectory(a *Archive, from, to string, recursive bool) (int, error) {\n\treturn AddGlob(a, filepath.Join(from, \"*\"), to, recursive)\n}", "func ExtractDir(tarFile io.Reader, dir string, opts *Opts) error {\n\tif opts == nil {\n\t\topts = &Opts{}\n\t}\n\n\t// Simulate a \"cd\" to another directory.\n\tif !filepath.IsAbs(dir) {\n\t\tdir = filepath.Join(opts.ChangeDirectory, dir)\n\t}\n\n\tfi, err := os.Stat(dir)\n\tif os.IsNotExist(err) {\n\t\tif err := os.Mkdir(dir, os.ModePerm); err != nil {\n\t\t\treturn fmt.Errorf(\"could not create directory %s: %v\", dir, err)\n\t\t}\n\t} else if err != nil || !fi.IsDir() {\n\t\treturn fmt.Errorf(\"could not stat directory %s: %v\", dir, err)\n\t}\n\n\treturn applyToArchive(tarFile, func(tr *tar.Reader, hdr *tar.Header) error {\n\t\tif !passesFilters(hdr, opts.Filters) {\n\t\t\treturn nil\n\t\t}\n\t\treturn createFileInRoot(hdr, tr, dir)\n\t})\n}", "func AddDirToArchive(tw TarWriter, dir string) error {\n\tdir = filepath.Clean(dir)\n\n\treturn filepath.Walk(dir, func(file string, fi os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn AddFileToArchive(tw, file, fi)\n\t})\n}", "func walkAndZip(src string, dst string, zipWriter *zip.Writer) (err error) {\n\tcontents, err := ioutil.ReadDir(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, content := range contents {\n\t\tcontentName := content.Name()\n\t\tcontentPath := path.Join(src, contentName)\n\t\tif content.IsDir() {\n\t\t\tif err = walkAndZip(contentPath, dst+contentName+\"/\", zipWriter); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\terrChan := make(chan error)\n\t\t\t// In order to close the file immediately it is read completely,\n\t\t\t// a goroutine is created. Otherwise the opened file will not be closed\n\t\t\t// until the outer function `walkAndZip` exits (by `defer originFile.Close()`).\n\t\t\t// This could cause a lot file descriptors remain in the kernel.\n\t\t\t// A new goroutine might not be necessary,\n\t\t\t// because We can also use a function which returns `error` here\n\t\t\tgo func() {\n\t\t\t\tbuf := make([]byte, 1024)\n\t\t\t\toriginFile, err := os.Open(contentPath)\n\t\t\t\tif err != nil {\n\t\t\t\t\terrChan <- err\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tdefer originFile.Close()\n\t\t\t\tfileWriter, err := zipWriter.Create(dst + contentName) // for a file\n\t\t\t\tfor { // read and write\n\t\t\t\t\tnr, err := originFile.Read(buf)\n\t\t\t\t\tif nr < 0 {\n\t\t\t\t\t\terrChan <- err\n\t\t\t\t\t\treturn\n\t\t\t\t\t} else if nr == 0 { // EOF\n\t\t\t\t\t\tbreak // Do not use switch-case or break won't go out of the loop\n\t\t\t\t\t} else {\n\t\t\t\t\t\tnw, err := fileWriter.Write(buf[0:nr])\n\t\t\t\t\t\tif nw != nr {\n\t\t\t\t\t\t\terrChan <- fmt.Errorf(\"write buff length is not equal to read buff\")\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif err != nil && err != io.EOF {\n\t\t\t\t\t\t\terrChan <- err\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\terrChan <- nil\n\t\t\t}()\n\t\t\terr = <-errChan\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\treturn\n}", "func Unzip(src, dest string) error {\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer r.Close()\n\n\tfor _, f := range r.File {\n\t\trc, err := f.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer rc.Close()\n\n\t\tfpath := filepath.Join(dest, f.Name)\n\t\tif f.FileInfo().IsDir() {\n\t\t\tos.MkdirAll(fpath, f.Mode())\n\t\t} else {\n\t\t\tvar fdir string\n\t\t\tif lastIndex := strings.LastIndex(fpath, string(os.PathSeparator)); lastIndex > -1 {\n\t\t\t\tfdir = fpath[:lastIndex]\n\t\t\t}\n\n\t\t\terr = os.MkdirAll(fdir, f.Mode())\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tf, err := os.OpenFile(\n\t\t\t\tfpath, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, f.Mode())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer f.Close()\n\n\t\t\t_, err = io.Copy(f, rc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func main() {\n\terr := zipr(\"/Users/matthew/go/src/exercises/zip\", \"/Users/matthew/Desktop/files.zip\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func unzip(src, dest string) error {\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer r.Close()\n\n\t// Make File\n\tif err := os.MkdirAll(dest, os.ModePerm); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, zipFile := range r.File {\n\t\t// Check for ZipSlip\n\t\tfileName := strings.ReplaceAll(zipFile.Name, \"..\", \"\")\n\n\t\tif fileName != zipFile.Name {\n\t\t\tgraceLog(\"ignore zip file: %s\", zipFile.Name)\n\t\t\tcontinue\n\t\t}\n\n\t\ttargetPath := filepath.Join(dest, fileName)\n\n\t\tif err := writeZipFile(targetPath, zipFile); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (e *Extractor) Unzip(source, destination, manifest string) error {\n\te.Log.Info(\"extracting application\")\n\te.Log.Debugf(`parameters for extractor:\n\tsource: %+v\n\tdestination: %+v`, source, destination)\n\n\terr := e.FileSystem.MkdirAll(destination, 0755)\n\tif err != nil {\n\t\treturn CreateDirectoryError{err}\n\t}\n\n\tfile, err := e.FileSystem.Open(source)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\tfileStat, err := file.Stat()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t//reader, err := zip.NewReader(file, fileStat.Size())\n\treader, err := zip.NewReader(file, fileStat.Size())\n\tif err != nil {\n\t\treturn OpenZipError{source, err}\n\t}\n\n\tfor _, file := range reader.File {\n\t\terr := e.unzipFile(destination, file)\n\t\tif err != nil {\n\t\t\treturn ExtractFileError{file.Name, err}\n\t\t}\n\t}\n\n\tif manifest != \"\" {\n\t\tmanifestFile, err := e.FileSystem.OpenFile(path.Join(destination, \"manifest.yml\"), os.O_WRONLY|os.O_CREATE|os.O_TRUNC, 0600)\n\t\tif err != nil {\n\t\t\treturn OpenManifestError{err}\n\t\t}\n\t\tdefer manifestFile.Close()\n\n\t\t_, err = fmt.Fprint(manifestFile, manifest)\n\t\tif err != nil {\n\t\t\treturn PrintToManifestError{err}\n\t\t}\n\t}\n\n\te.Log.Info(\"extract was successful\")\n\treturn nil\n}", "func (f *File) WriteDir(filename string) error {\n\tfilename = path.Join(f.Prefix, filename)\n\tfilename += \"/\" // Must have trailing slash to tell it it's a directory.\n\tfh := zip.FileHeader{\n\t\tName: filename,\n\t\tMethod: zip.Store,\n\t}\n\tfh.SetModTime(modTime)\n\tif _, err := f.w.CreateHeader(&fh); err != nil {\n\t\treturn err\n\t}\n\tf.addExistingFile(filename, filename, 0, 0, 0)\n\treturn nil\n}", "func (f *FileStore) extract(r io.Reader, dir string) (string, error) {\n\ttemplateDir := path.Join(f.baseDir, dir)\n\tgr, err := gzip.NewReader(r)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer gr.Close()\n\n\ttr := tar.NewReader(gr)\n\tfor {\n\t\thdr, err := tr.Next()\n\t\tif err == io.EOF {\n\t\t\tbreak\n\t\t}\n\n\t\t// ignore links\n\t\tif hdr.Mode == tar.TypeSymlink {\n\t\t\tcontinue\n\t\t}\n\n\t\tp := path.Clean(path.Join(templateDir, hdr.Name))\n\t\tif strings.HasPrefix(p, \"..\") {\n\t\t\t// no relative paths\n\t\t\tcontinue\n\t\t}\n\t\terr = os.MkdirAll(path.Dir(p), os.FileMode(0700))\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tf, err := os.Create(p)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\n\t\tdefer f.Close()\n\t\t_, err = io.Copy(f, tr)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t}\n\treturn templateDir, nil\n}", "func createZipFile(filename string, file string) error {\n\terr := ioutil.WriteFile(file, []byte(\"package test\"), 0755)\n\n\tnewZipFile, _ := os.Create(filename)\n\tdefer newZipFile.Close()\n\n\tzipWriter := zip.NewWriter(newZipFile)\n\tdefer zipWriter.Close()\n\n\terr = addFileToZip(zipWriter, file)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = removeFile(file)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Context) Archive() {\n\ttarget := path.Join(c.AppDir(), fmt.Sprintf(\"%v.%v\", c.Tournament.Name, fileExtension))\n\tf, err := os.Create(target)\n\tif err != nil {\n\t\tc.Log.Error(err.Error())\n\t\treturn\n\t}\n\tdefer f.Close()\n\t// Create a buffer to write our archive to.\n\tbuf := bufio.NewWriter(f)\n\n\t// Create a new zip archive.\n\tw := zip.NewWriter(buf)\n\tdefer w.Close()\n\n\t// Add some files to the archive.\n\tvar files = []string{\n\t\t\"config.yml\",\n\t}\n\tfor _, fileName := range c.Songs.FileNames() {\n\t\tfiles = append(files, fmt.Sprintf(\"media/%v\", fileName))\n\t}\n\tfor _, file := range files {\n\t\tzf, err := w.Create(file)\n\t\tif err != nil {\n\t\t\tc.Log.Error(err.Error())\n\t\t}\n\t\tp := path.Join(c.StorageDir(), file)\n\t\tdata, _ := ioutil.ReadFile(p)\n\t\t_, err = zf.Write([]byte(data))\n\t\tif err != nil {\n\t\t\tc.Log.Error(err.Error())\n\t\t}\n\t}\n}", "func Unzip(src, dest string) ([]string, error) {\n\n\tvar filenames []string\n\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn filenames, err\n\t}\n\tdefer r.Close()\n\n\tfor _, f := range r.File {\n\n\t\trc, err := f.Open()\n\t\tif err != nil {\n\t\t\treturn filenames, err\n\t\t}\n\t\tdefer rc.Close()\n\n\t\t// Store filename/path for returning and using later on\n\t\tfpath := filepath.Join(dest, f.Name)\n\t\tfilenames = append(filenames, fpath)\n\n\t\tif f.FileInfo().IsDir() {\n\n\t\t\t// Make Folder\n\t\t\tos.MkdirAll(fpath, os.ModePerm)\n\n\t\t} else {\n\n\t\t\t// Make File\n\t\t\tvar fdir string\n\t\t\tif lastIndex := strings.LastIndex(fpath, string(os.PathSeparator)); lastIndex > -1 {\n\t\t\t\tfdir = fpath[:lastIndex]\n\t\t\t}\n\n\t\t\terr = os.MkdirAll(fdir, os.ModePerm)\n\t\t\tif err != nil {\n\t\t\t\tlog.Fatal(err)\n\t\t\t\treturn filenames, err\n\t\t\t}\n\t\t\tf, err := os.OpenFile(\n\t\t\t\tfpath, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, f.Mode())\n\t\t\tif err != nil {\n\t\t\t\treturn filenames, err\n\t\t\t}\n\t\t\tdefer f.Close()\n\n\t\t\t_, err = io.Copy(f, rc)\n\t\t\tif err != nil {\n\t\t\t\treturn filenames, err\n\t\t\t}\n\n\t\t}\n\t}\n\treturn filenames, nil\n}", "func (c *cloner) listZipFiles() ([]zipFile, error) {\n\tdirEntries, err := os.ReadDir(c.cli.config.cacheDir)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar zipFiles []zipFile\n\tfor _, entry := range dirEntries {\n\t\text := filepath.Ext(entry.Name())\n\t\tif ext != \".zip\" {\n\t\t\tcontinue\n\t\t}\n\t\tif !strings.HasPrefix(entry.Name(), sampleAppsNamePrefix) {\n\t\t\tcontinue\n\t\t}\n\t\tfi, err := entry.Info()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tname := fi.Name()\n\t\tetag := \"\"\n\t\tparts := strings.Split(name, \"_\")\n\t\tif len(parts) == 2 {\n\t\t\tetag = strings.TrimSuffix(parts[1], ext)\n\t\t}\n\t\tzipFiles = append(zipFiles, zipFile{\n\t\t\tpath: filepath.Join(c.cli.config.cacheDir, name),\n\t\t\tetag: etag,\n\t\t\tmodTime: fi.ModTime(),\n\t\t})\n\t}\n\treturn zipFiles, nil\n}", "func TestZipSlip(t *testing.T) {\n\tt.Parallel()\n\n\ttmpDir, err := ioutil.TempDir(\"\", \"astro-test\")\n\trequire.NoError(t, err)\n\tdefer os.RemoveAll(tmpDir)\n\n\t// Create zip path\n\ttmpZipFileName := filepath.Join(tmpDir, \"1/bad.zip\")\n\terr = os.MkdirAll(filepath.Dir(tmpZipFileName), 0755)\n\trequire.NoError(t, err)\n\n\t// Create zip file\n\ttmpZipFile, err := os.Create(tmpZipFileName)\n\trequire.NoError(t, err)\n\tdefer tmpZipFile.Close()\n\n\tzipWriter := zip.NewWriter(tmpZipFile)\n\tdefer zipWriter.Close()\n\n\t// Add some files\n\treadmeFile, err := zipWriter.Create(\"README.txt\")\n\trequire.NoError(t, err)\n\t_, err = readmeFile.Write([]byte(\"This is a zip file for testing.\"))\n\trequire.NoError(t, err)\n\n\t// Add a naughty file\n\tbadFile, err := zipWriter.Create(\"../naughty.txt\")\n\trequire.NoError(t, err)\n\t_, err = badFile.Write([]byte(\"This file should never be extracted.\"))\n\trequire.NoError(t, err)\n\n\t// Write zip\n\trequire.NoError(t, zipWriter.Close())\n\n\t// Test that extracting this zip file causes an error\n\ttmpDir, err = ioutil.TempDir(\"\", \"astro-test\")\n\trequire.NoError(t, err)\n\tdefer os.RemoveAll(tmpDir)\n\n\terr = unzip(tmpZipFile.Name(), tmpDir)\n\tassert.Error(t, err)\n\tassert.Contains(t, err.Error(), \"illegal file path in zip\")\n}", "func (fc *FakeCompressor) CompressFilesInDir(dir string) (path string, err error) {\n\tfilesInDir, _ := fc.fs.Glob(filepath.Join(dir, \"*\"))\n\tarchive := map[string][]byte{}\n\n\tfor _, file := range filesInDir {\n\t\tfileContents, err := fc.fs.ReadFile(filepath.Join(dir, file))\n\t\tif err != nil {\n\t\t\treturn \"\", fmt.Errorf(\"reading file to compress: %s\", err.Error())\n\t\t}\n\t\tarchive[file] = fileContents\n\t}\n\tcompressedArchive, err := json.Marshal(archive)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"marshalling json: %s\", err.Error())\n\t}\n\tcompressedFile, _ := fc.fs.TempFile(\"stemcell-tgz\")\n\tpath = compressedFile.Name()\n\terr = fc.fs.WriteFile(compressedFile.Name(), compressedArchive)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"writing file: %s\", err.Error())\n\t}\n\treturn path, nil\n}", "func (c *Config) ArchiveDir() string {\n\treturn filepath.Join(c.Dir, \"archive\")\n}", "func zipped(c *gin.Context) {\n\tpath := c.Params.ByName(\"name\")\n\tdata, err := store.Multi(path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tbuf := new(bytes.Buffer)\n\tw := zip.NewWriter(buf)\n\tfor name, data := range data {\n\t\tfmt.Println(\"adding \" + name)\n\t\tf, err := w.Create(name[1:])\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t\t_, err = f.Write(data)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\terr = w.Close()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tsize := int64(buf.Len())\n\tc.Writer.Header().Set(\"Content-Length\", strconv.FormatInt(size, 10))\n\tc.Writer.Header().Set(\"Content-Type\", \"application/zip\")\n\tio.Copy(c.Writer, buf)\n}", "func extractToDir(dir, prefix string, img v1.Image, imageName string) error {\n\tlogrus.Infof(\"Extracting %q %q to %q\", imageName, prefix, dir)\n\tif err := os.MkdirAll(filepath.Dir(dir), 0755); err != nil {\n\t\treturn err\n\t}\n\n\ttempDir, err := ioutil.TempDir(filepath.Split(dir))\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer os.RemoveAll(tempDir)\n\n\timageReader := mutate.Extract(img)\n\tdefer imageReader.Close()\n\n\t// Extract content to temporary directory.\n\tif err := extract(imageName, tempDir, prefix, imageReader); err != nil {\n\t\treturn err\n\t}\n\n\t// Try to rename the temp dir into its target location.\n\tif err := os.Rename(tempDir, dir); err == nil {\n\t\t// Successfully renamed into place, nothing else to do.\n\t\treturn nil\n\t} else if !os.IsExist(err) {\n\t\t// Failed to rename, but not because the destination already exists.\n\t\treturn err\n\t}\n\n\t// Target directory already exists (got ErrExist above), fall back list/rename files into place.\n\tfiles, err := ioutil.ReadDir(tempDir)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tvar errs []error\n\tfor _, file := range files {\n\t\tsrc := filepath.Join(tempDir, file.Name())\n\t\tdst := filepath.Join(dir, file.Name())\n\t\tif err := os.Rename(src, dst); os.IsExist(err) {\n\t\t\t// Can't rename because dst already exists, remove it...\n\t\t\tif err = os.RemoveAll(dst); err != nil {\n\t\t\t\terrs = append(errs, errors2.Wrapf(err, \"failed to remove %q\", dst))\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// ...then try renaming again\n\t\t\tif err = os.Rename(src, dst); err != nil {\n\t\t\t\terrs = append(errs, errors2.Wrapf(err, \"failed to rename %q to %q\", src, dst))\n\t\t\t}\n\t\t} else if err != nil {\n\t\t\t// Other error while renaming src to dst.\n\t\t\terrs = append(errs, errors2.Wrapf(err, \"failed to rename %q to %q\", src, dst))\n\t\t}\n\t}\n\treturn merr.NewErrors(errs...)\n}", "func NewSubDirZipWriter(baseFolder string, zipWriter *zip.Writer) types.ArchiveWriter {\n\treturn &subDirZipWriter{\n\t\t// Zip entries must not start with a slash.\n\t\tbaseFolder: strings.Trim(baseFolder, \"/\"),\n\t\tzipWriter: zipWriter,\n\t}\n}", "func (z *Zip) UnZip(src, dest string) error {\n\tr, err := zip.OpenReader(file.ReplacePathSeparator(src))\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor _, fileOpenedOnZip := range r.File {\n\t\tcontentFileOpenedOnZip, err := fileOpenedOnZip.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = z.createFileAndFolderToUnZip(file.ReplacePathSeparator(dest), contentFileOpenedOnZip, fileOpenedOnZip)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn r.Close()\n}", "func Uncompress(src string, dest string) (filenames []string, err error) {\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn filenames, err\n\t}\n\tdefer r.Close()\n\n\tfor _, f := range r.File {\n\t\t//Create anonymous function to avoid leaving too many files open\n\t\tcopyZipFile := func(f *zip.File) ([]string, error) {\n\t\t\trc, err := f.Open()\n\t\t\tif err != nil {\n\t\t\t\treturn filenames, err\n\t\t\t}\n\n\t\t\tdefer rc.Close()\n\n\t\t\t// Store filename/path for returning and using later on\n\t\t\tfpath := filepath.Join(dest, f.Name)\n\n\t\t\t// Check for ZipSlip exploit\n\t\t\tif !strings.HasPrefix(fpath, filepath.Clean(dest+string(os.PathSeparator))) {\n\t\t\t\treturn filenames, fmt.Errorf(\"%s: illegal file path\", fpath)\n\t\t\t}\n\n\t\t\tfilenames = append(filenames, fpath)\n\n\t\t\tif f.FileInfo().IsDir() {\n\t\t\t\t// Make Folder\n\t\t\t\terr = os.MkdirAll(fpath, os.ModePerm)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn filenames, err\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tif err = os.MkdirAll(filepath.Dir(fpath), os.ModePerm); err != nil {\n\t\t\t\t\treturn filenames, err\n\t\t\t\t}\n\t\t\t\tbody, err := ioutil.ReadAll(rc)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn filenames, err\n\t\t\t\t}\n\t\t\t\terr = ioutil.WriteFile(fpath, body, 0644)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn filenames, err\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn filenames, nil\n\t\t}\n\n\t\tfilenames, err = copyZipFile(f)\n\t\tif err != nil {\n\t\t\treturn filenames, err\n\t\t}\n\t}\n\treturn filenames, nil\n}", "func genZIP(debug, logger *log.Logger, noMtime, noCompress bool, src string) (string, error) {\n\tbuf := &bytes.Buffer{}\n\n\tw := zip.NewWriter(buf)\n\terr := filepath.Walk(src, func(path string, fi os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Ignore directories and hidden files.\n\t\t// No entry is needed for directories in a zip file.\n\t\t// Each file is represented with a path, no directory\n\t\t// entities are required to build the hierarchy.\n\t\tif fi.IsDir() || strings.HasPrefix(fi.Name(), \".\") {\n\t\t\treturn nil\n\t\t}\n\t\trelPath, err := filepath.Rel(src, path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t/* #nosec */\n\t\tb, err := ioutil.ReadFile(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfHeader, err := zip.FileInfoHeader(fi)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif noMtime {\n\t\t\t// Always use the same modification time so that\n\t\t\t// the output is deterministic with respect to the file contents.\n\t\t\tfHeader.SetModTime(mtimeDate)\n\t\t}\n\t\tfHeader.Name = filepath.ToSlash(relPath)\n\t\tif !noCompress {\n\t\t\tfHeader.Method = zip.Deflate\n\t\t}\n\t\tf, err := w.CreateHeader(fHeader)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t_, err = f.Write(b)\n\t\treturn err\n\t})\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\terr = w.Close()\n\ts := &strings.Builder{}\n\tfor _, b := range buf.Bytes() {\n\t\tif b == '\\n' {\n\t\t\t_, err = s.WriteString(`\\n`)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Println(\"Error writing to output string\")\n\t\t\t\tos.Exit(exSoftware)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tif b == '\\\\' {\n\t\t\t_, err = s.WriteString(`\\\\`)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Println(\"Error writing to output string\")\n\t\t\t\tos.Exit(exSoftware)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tif b == '\"' {\n\t\t\t_, err = s.WriteString(`\\\"`)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Println(\"Error writing to output string\")\n\t\t\t\tos.Exit(exSoftware)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tif (b >= 32 && b <= 126) || b == '\\t' {\n\t\t\terr = s.WriteByte(b)\n\t\t\tif err != nil {\n\t\t\t\tlogger.Println(\"Error writing byte to output string\")\n\t\t\t\tos.Exit(exSoftware)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tfmt.Fprintf(s, \"\\\\x%02x\", b)\n\t}\n\treturn s.String(), err\n}", "func MUnzip(from io.ReaderAt, size int64, to string) error {\n\trd, err := zip.NewReader(from, size)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tos.MkdirAll(to, 0666)\n\n\tfor _, v := range rd.File {\n\t\t// Do not unzip directories\n\t\tif v.FileInfo().IsDir() {\n\t\t\tos.MkdirAll(filepath.Join(to, v.Name), 0666)\n\t\t\tcontinue\n\t\t}\n\n\t\tzf, err := v.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer zf.Close()\n\t\tfpath := filepath.Join(to, v.Name)\n\n\t\tdf, err := os.OpenFile(fpath, os.O_WRONLY|os.O_CREATE, 0666)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer df.Close()\n\n\t\t_, err = io.Copy(df, zf)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func ExtractZipFile(r io.Reader) (io.ReadCloser, error) {\n\n\t//create a buffer and copy r to it\n\tbuf := new(bytes.Buffer)\n\t_, err := io.Copy(buf, r)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t//read bytes\n\tbr := bytes.NewReader(buf.Bytes())\n\tzip, err := zip.NewReader(br, br.Size())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t//check that the file is not empty\n\tif len(zip.File) == 0 {\n\t\treturn nil, fmt.Errorf(\"zip: archive is empty: %s\", \"unreachable\")\n\t}\n\n\t//access the file's contnts\n\treturn zip.File[0].Open()\n}", "func NewZipTree(name string) (map[string]MemFile, error) {\n\tfr, err := eos.Open(name)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tstats, _ := fr.Stat()\n\n\tr, err := zip.NewReader(fr, stats.Size())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tout := map[string]MemFile{}\n\tfor _, f := range r.File {\n\t\tif strings.HasSuffix(f.Name, \"/\") {\n\t\t\tcontinue\n\t\t}\n\t\tn := filepath.Clean(f.Name)\n\n\t\tzf := &ZipFile{\n\t\t\tzipFile: f,\n\t\t}\n\t\tout[n] = zf\n\t}\n\treturn out, nil\n}", "func unzipArchive(zippedArchive, dstPath string) (*unzippedContents, error) {\n\n\tuz := unzip.New(zippedArchive, dstPath+\"/\"+\"new_build/\")\n\terr := uz.Extract()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_, fileExt := getUserOS()\n\n\tcontents := &unzippedContents{\n\t\tnewMollyBinaryPath: dstPath + \"/\" + \"new_build/mollywallet\" + fileExt,\n\t\tupdateBinaryPath: dstPath + \"/\" + \"new_build/update\" + fileExt,\n\t}\n\n\treturn contents, err\n}", "func CreateZipFile(files []ZipItem, username string) (string, error) {\n\t// create zip archive\n\tzipFile, err := os.Create(GetOrderFolder() + username + \".zip\")\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn \"\", err\n\t}\n\tdefer zipFile.Close()\n\n\t// add all items to zip to the zip\n\twriter := zip.NewWriter(zipFile)\n\tfor _, file := range files {\n\t\tfor i := 0; i < file.Amount; i++ {\n\t\t\tname := file.Name\n\t\t\tname = strings.Replace(name, \".\", \"-\"+file.Format+\".\", 1)\n\t\t\tif i != 0 {\n\t\t\t\tname = strings.Replace(name, \".\", \"-\"+strconv.Itoa(i)+\".\", 1)\n\t\t\t}\n\t\t\taddFile(writer, \"./static\"+file.Path, name)\n\t\t}\n\t}\n\n\t// close the writer\n\terr = writer.Close()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn \"\", err\n\t}\n\treturn zipFile.Name(), nil\n}", "func (a *App) ProcessZipFile(ctx context.Context, fsys fs.FS, path string, archive string) error {\n\tf, err := fsys.Open(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ts, err := f.Stat()\n\tif err != nil {\n\t\treturn err\n\t}\n\tzfs, err := myzipfs.Reader(f, s.Size())\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn a.ProcessArchive(ctx, zfs, path)\n}", "func Unzip(src, dest string) error {\n\tr, err := zip.OpenReader(src)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer r.Close()\n\n\tfor _, f := range r.File {\n\t\trc, err := f.Open()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer rc.Close()\n\n\t\tpath := filepath.Join(dest, f.Name)\n\t\tif f.FileInfo().IsDir() {\n\t\t\tos.MkdirAll(path, f.Mode())\n\t\t} else {\n\t\t\tf, err := os.OpenFile(\n\t\t\t\tpath, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, f.Mode())\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tdefer f.Close()\n\t\t\t// kopiert aus reader in writer\n\t\t\t_, err = io.Copy(f, rc)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func createRoot(ctx context.Context, logger hclog.Logger, opts *Options) (string, error) {\n\t// Build our copy command\n\tcmd, err := dittoCmd(ctx, opts.BaseCmd)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Create our root directory\n\troot, err := ioutil.TempDir(\"\", \"gon-createzip\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Setup our args to copy our files into the root\n\tcmd.Args = []string{\n\t\tfilepath.Base(cmd.Path),\n\t}\n\tcmd.Args = append(cmd.Args, opts.Files...)\n\tcmd.Args = append(cmd.Args, root)\n\n\t// We store all output in out for logging and in case there is an error\n\tvar out bytes.Buffer\n\tcmd.Stdout = &out\n\tcmd.Stderr = cmd.Stdout\n\n\t// Log what we're going to execute\n\tlogger.Info(\"executing ditto to copy files for archiving\",\n\t\t\"output_path\", opts.OutputPath,\n\t\t\"command_path\", cmd.Path,\n\t\t\"command_args\", cmd.Args,\n\t)\n\n\t// Execute copy\n\tif err = cmd.Run(); err != nil {\n\t\tos.RemoveAll(root)\n\n\t\tlogger.Error(\n\t\t\t\"error copying source files to create zip archive\",\n\t\t\t\"err\", err,\n\t\t\t\"output\", out.String(),\n\t\t)\n\t\treturn \"\", err\n\t}\n\n\treturn root, nil\n}", "func zipContent(args []string) (string, error) {\n\tfmt.Println(\"Adding the following items to a zip file:\",\n\t\tstrings.Join(args, \" \"))\n\tzip := new(archivex.ZipFile)\n\ttmpfile, err := ioutil.TempFile(\"\", \"qr-filetransfer\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\ttmpfile.Close()\n\tif err := os.Rename(tmpfile.Name(), tmpfile.Name()+\".zip\"); err != nil {\n\t\treturn \"\", err\n\t}\n\tzip.Create(tmpfile.Name() + \".zip\")\n\tfor _, item := range args {\n\t\tf, err := os.Stat(item)\n\t\tif err != nil {\n\t\t\treturn \"\", err\n\t\t}\n\t\tif f.IsDir() == true {\n\t\t\tzip.AddAll(item, true)\n\t\t} else {\n\t\t\tzip.AddFile(item)\n\t\t}\n\t}\n\tif err := zip.Close(); err != nil {\n\t\treturn \"\", nil\n\t}\n\treturn zip.Name, nil\n}", "func ExtractFile(zf *zip.File, d string) error {\n\tif zf == nil {\n\t\treturn fmt.Errorf(\"Could not extract nil *zip.File\")\n\t}\n\tr, e := zf.Open()\n\tif e != nil {\n\t\treturn errors.NewUtil(zf, \"opening zipfile\", e)\n\t}\n\tdefer r.Close()\n\tp := filepath.Join(d, zf.Name)\n\tif zf.FileInfo().IsDir() {\n\t\tif e = os.MkdirAll(p, DPERM); e != nil {\n\t\t\treturn errors.NewUtil(p, \"creating directory\", e)\n\t\t}\n\t\treturn nil\n\t}\n\tif e = os.MkdirAll(filepath.Dir(p), DPERM); e != nil {\n\t\treturn errors.NewUtil(p, \"creating directory\", e)\n\t}\n\tf, e := os.OpenFile(p, os.O_WRONLY|os.O_CREATE|os.O_TRUNC, zf.Mode())\n\tif e != nil {\n\t\treturn errors.NewUtil(p, \"opening\", e)\n\t}\n\tdefer f.Close()\n\tif _, e = io.Copy(f, r); e != nil {\n\t\treturn errors.NewUtil(f, \"copying to\", e)\n\t}\n\treturn nil\n}", "func (a *TarArchiver) Unarchive(ctx context.Context, path string, rep Reporter, fn func(k string, w io.WriterAt) error) error {\n\t// We need to check the target directory first to avoid downloading data if there is a problem\n\terr := a.checkTargetDir(path)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttmpf, clean, err := a.tempFile()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer clean()\n\n\terr = fn(slashpath.Join(a.keyPrefix, TarArchiverKey), tmpf)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to download to temporary file\")\n\t}\n\n\t_, err = tmpf.Seek(0, 0)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to seek to the beginning of file\")\n\t}\n\n\tfi, err := tmpf.Stat()\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to stat temporary file\")\n\t}\n\n\tpr := rep.StartUnarchivingProgress(tmpf.Name(), fi.Size(), tmpf)\n\tdefer rep.StopUnarchivingProgress()\n\n\ttr := tar.NewReader(pr)\n\tfor {\n\t\thdr, err := tr.Next()\n\t\tswitch {\n\t\tcase err == io.EOF:\n\t\t\treturn nil //EOF we're done here\n\t\tcase err != nil:\n\t\t\treturn errors.Wrap(err, \"failed to read next header\")\n\t\tcase hdr == nil:\n\t\t\tcontinue\n\t\t}\n\n\t\t// the target location where the dir/file should be created\n\t\tparts := []string{path}\n\t\tparts = append(parts, strings.Split(hdr.Name, TarArchiverPathSeparator)...)\n\t\ttarget := filepath.Join(parts...)\n\n\t\tswitch hdr.Typeflag {\n\t\tcase tar.TypeDir: //if its a dir and it doesn't exist create it, no-op if it exists already\n\t\t\terr = os.MkdirAll(target, hdr.FileInfo().Mode())\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, \"failed to create directory for entry found in tar file\")\n\t\t\t}\n\n\t\tcase tar.TypeReg: //regular file is written, must not exist yet\n\t\t\tif err = func() (err error) {\n\t\t\t\tf, err := os.OpenFile(target, os.O_WRONLY|os.O_CREATE|os.O_EXCL, hdr.FileInfo().Mode())\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn errors.Wrap(err, \"failed to open new file for tar entry \")\n\t\t\t\t}\n\n\t\t\t\tdefer f.Close()\n\t\t\t\tif _, err := io.Copy(f, tr); err != nil {\n\t\t\t\t\treturn errors.Wrap(err, \"failed to copy archived file content\")\n\t\t\t\t}\n\n\t\t\t\treturn nil\n\t\t\t}(); err != nil {\n\t\t\t\treturn errors.Wrap(err, \"failed to extract file\")\n\t\t\t}\n\t\t}\n\t}\n}", "func packRelease(folder string) (targetFileName string, err error) {\n\tlog.Println(\"Packing release folder:\", folder)\n\n\ttargetFileName = filepath.Join(targetFolder, filepath.Base(folder)+\".zip\")\n\tlog.Println(\"To:\", targetFileName)\n\n\troot := filepath.Dir(folder)\n\n\tf, err := os.Create(targetFileName)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tdefer f.Close()\n\n\tzw := zip.NewWriter(f)\n\tdefer zw.Close()\n\n\terr = filepath.Walk(folder, func(path string, info os.FileInfo, err error) error {\n\t\tif info.IsDir() {\n\t\t\treturn nil\n\t\t}\n\n\t\tcontent, err := ioutil.ReadFile(path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfh, err := zip.FileInfoHeader(info)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\trelPath, err := filepath.Rel(root, path)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Zip expects forward slashes ('/'), so replace os dependant separators\n\t\trelPath = filepath.ToSlash(relPath)\n\t\t// fh.Name is only the fienme, so:\n\t\tfh.Name = relPath\n\n\t\tfh.Method = zip.Deflate\n\t\tw, err := zw.CreateHeader(fh)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif _, err = w.Write(content); err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t})\n\n\treturn\n}", "func NeuraxZIPSelf() error {\n\tarchive_name := os.Args[0] + \".zip\"\n\tfiles_to_zip := []string{os.Args[0]}\n\treturn cf.MakeZip(archive_name, files_to_zip)\n}", "func extractArchive(f *os.File, name, dir, stripPrefix string) (err error) {\n\tif strings.HasSuffix(name, \".zip\") {\n\t\treturn extractZip(f, name, dir, stripPrefix)\n\t}\n\tif strings.HasSuffix(name, \".tar.gz\") {\n\t\tzr, err := gzip.NewReader(f)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"extracting %s: %w\", name, err)\n\t\t}\n\t\tdefer func() {\n\t\t\tif cerr := zr.Close(); err == nil && cerr != nil {\n\t\t\t\terr = cerr\n\t\t\t}\n\t\t}()\n\t\treturn extractTar(zr, name, dir, stripPrefix)\n\t}\n\treturn fmt.Errorf(\"could not determine archive format from extension: %s\", name)\n}", "func (z *CompressApp) DoZipFile(filename string) error {\n\t//check exist file\n\n\tif z.fileUtils.Exists(filename + \".zip\") {\n\t\terr := fileutils.ErrFileAlreadyExists\n\t\tz.log.Error(err.Error())\n\t\treturn err\n\t}\n\t//check exist file end\n\tnewFile, err := os.Create(filename + \".zip\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer z.checkErrorFunc(newFile.Close)\n\n\tzipIt := zip.NewWriter(newFile)\n\tdefer z.checkErrorFunc(zipIt.Close)\n\n\tzipFile, err := os.Open(filename)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer z.checkErrorFunc(zipFile.Close)\n\n\t// get the file information\n\tinfo, err := zipFile.Stat()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\theader, err := zip.FileInfoHeader(info)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\theader.Method = zip.Deflate\n\n\twriter, err := zipIt.CreateHeader(header)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = io.Copy(writer, zipFile)\n\treturn err\n\n}", "func DecompressFiles(b []byte, dir string) error {\n\tif env.Debug {\n\t\tlog.Printf(\"Decompressing %d bytes to %s\", len(b), dir)\n\t}\n\treturn archiver.Zip.Read(bytes.NewBuffer(b), dir)\n}", "func unzip(path string, file io.Reader) {\n\tzipfile, err := zip.OpenReader(path)\n\tif err != nil {\n\t\treturn\n\t}\n\tdefer zipfile.Close()\n\tfor _, f := range zipfile.File {\n\t\tworkResult <- WorkResult{path, f.Name}\n\t}\n}" ]
[ "0.77978015", "0.7659449", "0.7599635", "0.7598706", "0.7559721", "0.72150767", "0.71371746", "0.69366485", "0.6818529", "0.6695745", "0.66736376", "0.6618481", "0.657473", "0.6563791", "0.63425386", "0.6162173", "0.6131495", "0.61223537", "0.6112147", "0.60838836", "0.607063", "0.60370266", "0.600981", "0.5990727", "0.59709024", "0.59390616", "0.59311134", "0.59115875", "0.58893937", "0.58483535", "0.5846171", "0.57379985", "0.5734622", "0.5699833", "0.56854296", "0.56774724", "0.56735164", "0.56547284", "0.56345063", "0.5597779", "0.5587068", "0.5553652", "0.5553504", "0.5507315", "0.55053484", "0.5475638", "0.54746413", "0.54743975", "0.54725474", "0.5464759", "0.54619986", "0.54594564", "0.5450662", "0.5443256", "0.5434778", "0.5433146", "0.5419053", "0.5391661", "0.53852516", "0.5373284", "0.5364802", "0.5360317", "0.5350918", "0.53363836", "0.5328507", "0.5324698", "0.5313211", "0.5287906", "0.52802444", "0.5273904", "0.5271275", "0.5227939", "0.5220773", "0.5209862", "0.52062386", "0.51895666", "0.51749456", "0.5154684", "0.5152058", "0.5137765", "0.5131762", "0.5122373", "0.51094025", "0.5100883", "0.5100097", "0.5099191", "0.50973463", "0.5088498", "0.5069424", "0.5024771", "0.5018106", "0.5011031", "0.5010436", "0.50082874", "0.49956322", "0.4994488", "0.49867758", "0.49818805", "0.49798334", "0.49733773" ]
0.75142837
5
return a random electionTimeout between 150ms~450ms according to guidance: Because the tester limits you to 10 heartbeats per second, you will have to use an election timeout larger than the paper's 150 to 300 milliseconds, but not too large, because then you may fail to elect a leader within five seconds.
func getElectionTimeout() time.Duration { return time.Duration(rand.Intn(300) + 150) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func electionTimeout() int64 {\n\treturn int64(rand.Intn(MAXELECTIMEOUT- MINELECTIMEOUT) + MINELECTIMEOUT)\n}", "func (node *Node) randElectionTimeout() time.Duration {\n\treturn time.Duration(150+rand.Intn(150)) * time.Millisecond\n}", "func GetRandomElectionTimeout() time.Duration {\n\treturn time.Duration(minElectionTimeout+rand.Intn(maxElectionTimeout-minElectionTimeout)) * time.Millisecond\n}", "func generateElectionTime() int {\n rand.Seed(time.Now().UnixNano())\n return rand.Intn(150)*2 + 300\n}", "func (rf *Raft) resetElectionTimeout() time.Duration {\n\trand.Seed(time.Now().UTC().UnixNano())\n\trf.randomizedElectionTimeout = rf.electionTimeout + time.Duration(rand.Int63n(rf.electionTimeout.Nanoseconds()))\n\treturn rf.randomizedElectionTimeout\n}", "func setRandomizedElectionTimeout(r *raft, v int) {\n\tr.randomizedElectionTimeout = v\n}", "func randomTimeOut() time.Duration {\n\tt := time.Duration(rand.Intn(150)+150) * time.Millisecond // rand [150,300) ms to time out\n\treturn t\n}", "func (rf *Raft) resetElectionTimeout() {\n\trf.electionTimeoutStartTime = time.Now()\n\t// randomize election timeout, 300~400ms\n\trf.electionTimeoutInterval = time.Duration(time.Millisecond * time.Duration(500+rand.Intn(300)))\n}", "func testNonleaderElectionTimeoutRandomized(t *testing.T, state StateType) {\n\tet := 10\n\tr := newTestRaft(1, []uint64{1, 2, 3}, et, 1, NewMemoryStorage())\n\tdefer closeAndFreeRaft(r)\n\ttimeouts := make(map[int]bool)\n\tfor round := 0; round < 50*et; round++ {\n\t\tswitch state {\n\t\tcase StateFollower:\n\t\t\tr.becomeFollower(r.Term+1, 2)\n\t\tcase StateCandidate:\n\t\t\tr.becomeCandidate()\n\t\t}\n\n\t\ttime := 0\n\t\tfor len(r.readMessages()) == 0 {\n\t\t\tr.tick()\n\t\t\ttime++\n\t\t}\n\t\ttimeouts[time] = true\n\t}\n\n\tfor d := et + 1; d < 2*et; d++ {\n\t\tif !timeouts[d] {\n\t\t\tt.Errorf(\"timeout in %d ticks should happen\", d)\n\t\t}\n\t}\n}", "func TestLearnerElectionTimeout(t *testing.T) {\n\tn1 := newTestLearnerRaft(1, []uint64{1}, []uint64{2}, 10, 1, NewMemoryStorage())\n\tn2 := newTestLearnerRaft(2, []uint64{1}, []uint64{2}, 10, 1, NewMemoryStorage())\n\tdefer closeAndFreeRaft(n1)\n\tdefer closeAndFreeRaft(n2)\n\n\tn1.becomeFollower(1, None)\n\tn2.becomeFollower(1, None)\n\n\t// n2 is learner. Learner should not start election even when times out.\n\tsetRandomizedElectionTimeout(n2, n2.electionTimeout)\n\tfor i := 0; i < n2.electionTimeout; i++ {\n\t\tn2.tick()\n\t}\n\n\tif n2.state != StateFollower {\n\t\tt.Errorf(\"peer 2 state: %s, want %s\", n2.state, StateFollower)\n\t}\n}", "func properTimeDuration(state int) time.Duration {\n\tif state == LEADER {\n\t\treturn time.Millisecond * HEARTBEAT_INTERVAL\n\t}\n\treturn time.Millisecond * time.Duration(\n\t\tMIN_ELECTION_INTERVAL+rand.Intn(MAX_ELECTION_INTERVAL-MIN_ELECTION_INTERVAL))\n}", "func (c *Clock) AfterElectionTimeout() <-chan chan struct{} {\n\td := c.ElectionTimeout + time.Duration(rand.Intn(int(c.ElectionTimeout)))\n\treturn newClockChan(d)\n}", "func randomTimeout(minVal, maxVal time.Duration) <-chan time.Time {\n\textra := time.Duration(rand.Int()) % maxVal\n\treturn time.After((minVal + extra) % maxVal)\n}", "func TestClock_AfterElectionTimeout(t *testing.T) {\n\tc := raft.NewClock()\n\tc.ElectionTimeout = 10 * time.Millisecond\n\tt0 := time.Now()\n\t<-c.AfterElectionTimeout()\n\tif d := time.Since(t0); d < c.ElectionTimeout {\n\t\tt.Fatalf(\"channel fired too soon: %v\", d)\n\t}\n}", "func runElectionTimeoutThread(\n\ttimeSinceLastUpdate * time.Time,\n\tisElection * bool,\n\tstate * ServerState,\n\tvoteChannels *[8]chan Vote,\n\tonWinChannel * chan bool,\n\telectionThreadSleepTime time.Duration,\n) {\n\tfor {\n\t\ttimeElapsed := time.Now().Sub(*timeSinceLastUpdate)\n\t\tif timeElapsed.Milliseconds() > ElectionTimeOut { //implements C4.\n\t\t\t*isElection = true // restarts election\n\t\t}\n\n\t\tif *isElection {\n\t\t\t*timeSinceLastUpdate = time.Now()\n\t\t\tgo elect(state, voteChannels, *onWinChannel)\n\t\t}\n\n\t\ttime.Sleep(electionThreadSleepTime)\n\t}\n}", "func (c *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn 120 * time.Second, 2 * time.Second\n}", "func (r *Raft) SetElectionTimer() {\n\tif r.Id==0 {\n\t\tElectionTimer = time.NewTimer(time.Millisecond*8000)\n\t} else if r.Id==1 {\n\t\tElectionTimer = time.NewTimer(time.Millisecond*2500)\n\t} else if r.Id==2 {\n\t\tElectionTimer = time.NewTimer(time.Millisecond*8500)\n\t} else if r.Id==3 {\n\t\tElectionTimer = time.NewTimer(time.Millisecond*9000)\t\t\n\t} else if r.Id==4 {\n\t\tElectionTimer = time.NewTimer(time.Millisecond*9500)\t\t\n\t}\n}", "func (p *MockProvisionerClient) Timeout() time.Duration {\n\treturn 30 * time.Second\n}", "func RandomDelay(_ uint, _ error, config *Config) time.Duration {\n\treturn time.Duration(rand.Int63n(int64(config.maxJitter)))\n}", "func new_time() time.Duration {\n\treturn time.Duration((rand.Intn(300) + 150)) * time.Millisecond\n}", "func testNonleadersElectionTimeoutNonconflict(t *testing.T, state StateType) {\n\tet := 10\n\tsize := 5\n\trs := make([]*raft, size)\n\tids := idsBySize(size)\n\tfor k := range rs {\n\t\trs[k] = newTestRaft(ids[k], ids, et, 1, NewMemoryStorage())\n\t}\n\tdefer func() {\n\t\tfor k := range rs {\n\t\t\tcloseAndFreeRaft(rs[k])\n\t\t}\n\t}()\n\tconflicts := 0\n\tfor round := 0; round < 1000; round++ {\n\t\tfor _, r := range rs {\n\t\t\tswitch state {\n\t\t\tcase StateFollower:\n\t\t\t\tr.becomeFollower(r.Term+1, None)\n\t\t\tcase StateCandidate:\n\t\t\t\tr.becomeCandidate()\n\t\t\t}\n\t\t}\n\n\t\ttimeoutNum := 0\n\t\tfor timeoutNum == 0 {\n\t\t\tfor _, r := range rs {\n\t\t\t\tr.tick()\n\t\t\t\tif len(r.readMessages()) > 0 {\n\t\t\t\t\ttimeoutNum++\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\t// several rafts time out at the same tick\n\t\tif timeoutNum > 1 {\n\t\t\tconflicts++\n\t\t}\n\t}\n\n\tif g := float64(conflicts) / 1000; g > 0.3 {\n\t\tt.Errorf(\"probability of conflicts = %v, want <= 0.3\", g)\n\t}\n}", "func (ck *Clerk) randomChooseLeader() int {\n\tn := len(ck.servers)\n\treturn (ck.leaderId+1) % n\n}", "func Test_TaskOption_LeadershipTimeout(t *testing.T) {\n\t// given\n\toption := crontask.LeadershipTimeout(time.Second)\n\toptions := &crontask.TaskOptions{LeadershipTimeout: time.Hour}\n\n\t// when\n\toption(options)\n\n\t// then\n\tif options.LeadershipTimeout != time.Second {\n\t\tt.Errorf(\"leadership timeout not correctly applied, got %s\", options.LeadershipTimeout)\n\t}\n}", "func (d *DDL) Timeout() time.Duration {\n\tif d.Metadata.Timeout == 0 {\n\t\treturn 10 * time.Second\n\t}\n\n\treturn time.Second * time.Duration(d.Metadata.Timeout)\n}", "func randomDuration() time.Duration {\n\treturn time.Duration(rand.Int63n(1e9))\n}", "func (m *exampleTimeout) Timeout(tType fab.TimeoutType) time.Duration {\n\tt, ok := defaultTypes[tType]\n\tif !ok {\n\t\treturn time.Second * 30 // general default if type is not found\n\t}\n\treturn t\n}", "func (d *DDL) Timeout() time.Duration {\n\tif d.Metadata.Timeout == 0 {\n\t\treturn time.Duration(10 * time.Second)\n\t}\n\n\treturn time.Duration(time.Second * time.Duration(d.Metadata.Timeout))\n}", "func (g *groupFailover) Timeout() time.Duration {\n\treturn g.timeout\n}", "func calculateRandomExpiry(validFrom, validUntil time.Time) time.Duration {\n\tduration := validUntil.Sub(validFrom)\n\n\tvar randomExpiry int\n\t// Our lower bound of renewal will be half of the total expiration time\n\tminValidity := int(duration.Minutes() * CertLowerRotationRange)\n\t// Our upper bound of renewal will be 80% of the total expiration time\n\tmaxValidity := int(duration.Minutes() * CertUpperRotationRange)\n\t// Let's select a random number of minutes between min and max, and set our retry for that\n\t// Using randomly selected rotation allows us to avoid certificate thundering herds.\n\tif maxValidity-minValidity < 1 {\n\t\trandomExpiry = minValidity\n\t} else {\n\t\trandomExpiry = rand.Intn(maxValidity-minValidity) + minValidity\n\t}\n\n\texpiry := time.Until(validFrom.Add(time.Duration(randomExpiry) * time.Minute))\n\tif expiry < 0 {\n\t\treturn 0\n\t}\n\treturn expiry\n}", "func (s *Conn) Timeout() time.Duration {\n\tif s.state == stateClosed {\n\t\treturn -1\n\t}\n\tnow := s.timeFn()\n\ts.logLossTimer(now)\n\tvar deadline time.Time\n\tif !s.drainingTimer.IsZero() {\n\t\tdeadline = s.drainingTimer\n\t} else if !s.recovery.lossDetectionTimer.IsZero() {\n\t\t// Minimum of loss and idle timer\n\t\tdeadline = s.recovery.lossDetectionTimer\n\t\tif !s.idleTimer.IsZero() && deadline.After(s.idleTimer) {\n\t\t\tdeadline = s.idleTimer\n\t\t}\n\t} else if !s.idleTimer.IsZero() {\n\t\tdeadline = s.idleTimer\n\t} else {\n\t\treturn -1\n\t}\n\ttimeout := deadline.Sub(now)\n\tif timeout < 0 {\n\t\ttimeout = 0\n\t}\n\treturn timeout\n}", "func getSleepTime() time.Duration {\n\treturn time.Duration(150+rand.Int31n(200)) * time.Millisecond\n}", "func randomDuration() Duration {\n\tx := 1.0 / (rand.Float64() * 1000)\n\treturn Duration(x*1000000+1) * Millisecond\n}", "func Timeout(d time.Duration) func(*Attacker) {\n\treturn func(a *Attacker) {\n\t\ta.client.Timeout = d\n\t}\n}", "func (p *Peer) HeartbeatTimeout() time.Duration {\n\treturn p.heartbeatTimer.MinDuration()\n}", "func (p *Peer) HeartbeatTimeout() time.Duration {\n\treturn p.heartbeatTimer.MinDuration()\n}", "func (r *Resolver) Timeout() time.Duration {\n\treturn time.Duration(30) * time.Second\n}", "func (p *partitionFailover) Timeout() time.Duration {\n\treturn p.timeout\n}", "func (p *Pool) Timeout() time.Duration {\n\treturn p.timeout\n}", "func (s stdlib) Timeout(time.Duration) {}", "func (d *DNSProviderPublic) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func getRandomDuration(min, max time.Duration) time.Duration {\n\treturn min + time.Duration(rand.Int63n(int64(max-min)))\n}", "func computeMaximumTimeout(cfg *gatewayv1.Timeouts) time.Duration {\n\tif cfg == nil {\n\t\treturn timeouts.DefaultTimeout\n\t}\n\n\tret := cfg.Default.AsDuration()\n\tfor _, e := range cfg.Overrides {\n\t\toverride := e.Timeout.AsDuration()\n\t\tif ret == 0 || override == 0 {\n\t\t\treturn 0\n\t\t}\n\n\t\tif override > ret {\n\t\t\tret = override\n\t\t}\n\t}\n\n\treturn ret\n}", "func (s *Server) RandServerHearbeat() time.Duration {\n\treturn (minServerHeartbeat + time.Duration(rand.Int63n(int64(maxServerHeartbeat-minServerHeartbeat))))\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func (d *DNSProvider) Timeout() (timeout, interval time.Duration) {\n\treturn d.config.PropagationTimeout, d.config.PollingInterval\n}", "func randomSleep(maxSleep int) (napTime time.Duration, napFunc func()) {\n\tnapTime = time.Duration(rand.Intn(maxSleep)) * time.Millisecond\n\tnapFunc = func() {\n\t\ttime.Sleep(napTime)\n\t}\n\treturn\n}", "func (client *BaseClient) Timeout() time.Duration {\n\treturn client.timeout\n}", "func (rf *Raft) heartbeatTimeoutCheck() {\n\t// get heartbeat check start time\n\tlastHeartbeatCheck := time.Now()\n\ti := 0\n\tfor !rf.killed() {\n\t\trf.mu.Lock()\n\t\tif rf.electionTimeout > 0 && rf.state == Follower {\n\t\t\tcurrentTime := time.Now()\n\t\t\trf.electionTimeout -= (currentTime.Sub(lastHeartbeatCheck))\n\t\t\tlastHeartbeatCheck = currentTime\n\t\t\tif i%10 == 0 { // decrease log density\n\t\t\t\trf.Log(LogDebug, \"timeout remaining:\", rf.electionTimeout)\n\t\t\t}\n\t\t} else if rf.state == Follower {\n\t\t\t// election needs to occur\n\t\t\t// quit this function and run the election\n\t\t\trf.Log(LogInfo, \"timed out as follower, running election.\")\n\t\t\trf.mu.Unlock()\n\t\t\tgo rf.runElection()\n\t\t\treturn\n\t\t}\n\t\trf.mu.Unlock()\n\t\ti++\n\t\ttime.Sleep(defaultPollInterval)\n\t}\n}", "func randomCandidate(r *rand.Rand) Candidate {\n\tvar status CandidateStatus\n\tif r.Float64() < float64(0.5) {\n\t\tstatus = Bonded\n\t} else {\n\t\tstatus = Unbonded\n\t}\n\tassets := sdk.NewRat(int64(r.Int31n(10000)))\n\tliabilities := sdk.NewRat(int64(r.Int31n(10000)))\n\treturn Candidate{\n\t\tStatus: status,\n\t\tAddress: addrs[0],\n\t\tPubKey: pks[0],\n\t\tAssets: assets,\n\t\tLiabilities: liabilities,\n\t}\n}", "func Timeout() time.Duration { return note.Timeout }", "func (c *connAttrs) Timeout() time.Duration { c.mu.RLock(); defer c.mu.RUnlock(); return c._timeout }", "func deliveryTime(min int, max int) int {\n\trand.Seed(time.Now().UTC().UnixNano())\n\treturn min + rand.Intn(max-min)\n}", "func (l *Logger) SeedTimeout(d core.Digest, infoHash core.InfoHash) {\n\tl.zap.Debug(\n\t\t\"Seed timeout\",\n\t\tzap.String(\"name\", d.Hex()),\n\t\tzap.String(\"info_hash\", infoHash.String()))\n}", "func getJitteredNetworkRetryTime() time.Duration {\n\treturn time.Duration(900+rand.Intn(100)) * time.Millisecond\n}", "func TestSelectVoterMaxVarious(t *testing.T) {\n\thash := 0\n\tfor minMaxRate := 1; minMaxRate <= 100000000; minMaxRate *= 10000 {\n\t\tt.Logf(\"<<< min: 100, max: %d >>>\", 100*minMaxRate)\n\t\tfor validators := 16; validators <= 256; validators *= 4 {\n\t\t\tfor voters := 1; voters <= validators; voters += 10 {\n\t\t\t\tvalSet, _ := randValidatorSetWithMinMax(PrivKeyEd25519, validators, 100, 100*int64(minMaxRate))\n\t\t\t\tvoterSet := SelectVoter(valSet, []byte{byte(hash)}, &VoterParams{int32(voters), 20})\n\t\t\t\tif voterSet.Size() < voters {\n\t\t\t\t\tt.Logf(\"Cannot elect voters up to MaxVoters: validators=%d, MaxVoters=%d, actual voters=%d\",\n\t\t\t\t\t\tvalidators, voters, voterSet.Size())\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\thash++\n\t\t\t}\n\t\t}\n\t}\n}", "func Timeout(timeout time.Duration) OptionFunc {\n\treturn func(tc *TracedClient) error {\n\t\tif timeout <= 0 {\n\t\t\treturn errors.New(\"timeout must be positive\")\n\t\t}\n\t\ttc.cl.Timeout = timeout\n\t\treturn nil\n\t}\n}", "func Timeout(d time.Duration) func(*Server) {\n\treturn func(s *Server) {\n\t\ts.timeout = d\n\t}\n}", "func (state *ServerState) TimeoutStateTransition() {\n\n\tif state.IsFollower() {\n\t\tstate.curState = CANDIDATE\n\t\tstate.votedFor = -1\n\t\tstate.currentTerm++\n\t} else if state.IsCandidate() {\n\t\t//Candidates who timeout keep being candidates\n\t\tstate.votedFor = -1\n\t\tstate.currentTerm++\n\t} else if state.IsLeader() {\n\t\tfmt.Println(\"WARNING: timedout as a leader\")\n\t\t//Leaders should not timeout\n\t}\n}", "func (conf Config) TickTimeout() time.Duration {\n\t// If a tick blocks longer than the interval of ticking, we may need to break it and retry.\n\treturn conf.TickDuration\n}", "func LeaseTimeout(duration time.Duration) func(*Config) error {\n\treturn func(c *Config) error {\n\t\tc.LeaseTimeout = duration\n\t\treturn nil\n\t}\n}", "func ConfigurableRacer(a, b string, timeout time.Duration) (winner string, error error) {\n\tselect {\n\tcase <-ping(a):\n\t\treturn a, nil\n\tcase <-ping(b):\n\t\treturn b, nil\n\tcase <-time.After(timeout):\n\t\treturn \"\", fmt.Errorf(\"timed out waiting for %s and %s\", a, b)\n\t}\n}", "func (r Replicator) Timeout() int {\n\treturn r.server.Timeout()\n}", "func (o GetAppTemplateContainerLivenessProbeOutput) Timeout() pulumi.IntOutput {\n\treturn o.ApplyT(func(v GetAppTemplateContainerLivenessProbe) int { return v.Timeout }).(pulumi.IntOutput)\n}", "func randomize(d time.Duration) time.Duration {\n\tmaxAdded := float64(d) / 4\n\tamount := rand.Float64() * maxAdded\n\n\treturn time.Duration(float64(d) + amount)\n}", "func (rf *Raft) startElectionTimer() {\n\tfor {\n\t\tif rf.killed() {\n\t\t\treturn\n\t\t}\n\t\trf.mu.Lock()\n\t\telectionTimeout := rf.electionTimeout\n\t\tlastHeard := rf.lastHeard\n\t\trf.mu.Unlock()\n\t\tnow := time.Now()\n\t\tif now.After(lastHeard.Add(electionTimeout)) {\n\t\t\tgo rf.candidate()\n\t\t\treturn\n\t\t}\n\t\ttime.Sleep(10 * time.Millisecond)\n\t}\n}", "func (r *Raft) candidate(timeout int) int {\n\twaitTime := timeout //added for passing timeout from outside--In SingleServerBinary\n\tresendTime := 5 //should be much smaller than waitTime\n\tElectionTimer := r.StartTimer(ElectionTimeout, waitTime)\n\t//This loop is for election process which keeps on going until a leader is elected\n\tfor {\n\t\t//reset the Votes else it will reflect the Votes received in last Term\n\t\tr.resetVotes()\n\t\tr.myCV.CurrentTerm += 1 //increment current Term\n\t\tr.myCV.VotedFor = r.Myconfig.Id //Vote for self\n\t\tr.WriteCVToDisk() //write Current Term and VotedFor to disk\n\t\tr.f_specific[r.Myconfig.Id].Vote = true //vote true\n\t\treqVoteObj := r.prepRequestVote() //prepare request Vote obj\n\t\tr.sendToAll(reqVoteObj) //send requests for Vote to all servers\n\t\tResendVoteTimer := r.StartTimer(ResendVoteTimeOut, resendTime)\n\t\tfor { //this loop for reading responses from all servers\n\t\t\treq := r.receive()\n\t\t\tswitch req.(type) {\n\t\t\tcase ClientAppendReq: ///candidate must also respond as false just like follower\n\t\t\t\trequest := req.(ClientAppendReq) //explicit typecasting\n\t\t\t\tresponse := ClientAppendResponse{}\n\t\t\t\tlogItem := LogItem{r.CurrentLogEntryCnt, false, request.Data} //lsn is count started from 0\n\t\t\t\tr.CurrentLogEntryCnt += 1\n\t\t\t\tresponse.LogEntry = logItem\n\t\t\t\tr.CommitCh <- &response.LogEntry\n\t\t\tcase RequestVoteResponse: //got the Vote response\n\t\t\t\tresponse := req.(RequestVoteResponse) //explicit typecasting so that fields of struct can be used\n\t\t\t\tif response.VoteGranted {\n\t\t\t\t\tr.f_specific[response.Id].Vote = true\n\t\t\t\t}\n\t\t\t\tVoteCount := r.countVotes()\n\t\t\t\tif VoteCount >= majority {\n\t\t\t\t\tResendVoteTimer.Stop()\n\t\t\t\t\tElectionTimer.Stop()\n\t\t\t\t\tr.LeaderConfig.Id = r.Myconfig.Id //update leader details\n\t\t\t\t\treturn leader //become the leader\n\t\t\t\t}\n\n\t\t\tcase AppendEntriesReq: //received an AE request instead of Votes, i.e. some other leader has been elected\n\t\t\t\trequest := req.(AppendEntriesReq)\n\t\t\t\tretVal := r.serviceAppendEntriesReq(request, nil, 0, candidate)\n\t\t\t\tif retVal == follower {\n\t\t\t\t\tResendVoteTimer.Stop()\n\t\t\t\t\tElectionTimer.Stop()\n\t\t\t\t\treturn follower\n\t\t\t\t}\n\n\t\t\tcase RequestVote:\n\t\t\t\trequest := req.(RequestVote)\n\t\t\t\t//==Can be shared with service request vote with additinal param of caller(candidate or follower)\n\t\t\t\tresponse := RequestVoteResponse{} //prep response object,for responding back to requester\n\t\t\t\tcandidateId := request.CandidateId\n\t\t\t\tresponse.Id = r.Myconfig.Id\n\t\t\t\tif r.isDeservingCandidate(request) {\n\t\t\t\t\tresponse.VoteGranted = true\n\t\t\t\t\tr.myCV.VotedFor = candidateId\n\t\t\t\t\tr.myCV.CurrentTerm = request.Term\n\t\t\t\t\tif request.Term > r.myCV.CurrentTerm { //write to disk only when value has changed\n\t\t\t\t\t\tr.WriteCVToDisk()\n\t\t\t\t\t}\n\t\t\t\t\tResendVoteTimer.Stop()\n\t\t\t\t\tElectionTimer.Stop()\n\t\t\t\t\treturn follower\n\t\t\t\t} else {\n\t\t\t\t\tresponse.VoteGranted = false\n\t\t\t\t}\n\t\t\t\tresponse.Term = r.myCV.CurrentTerm\n\t\t\t\tr.send(candidateId, response)\n\n\t\t\tcase int:\n\t\t\t\ttimeout := req.(int)\n\t\t\t\tif timeout == ResendVoteTimeOut {\n\t\t\t\t\trT := msecs * time.Duration(resendTime)\n\t\t\t\t\tResendVoteTimer.Reset(rT)\n\t\t\t\t\treqVoteObj := r.prepRequestVote() //prepare request Vote agn and send to all, ones rcvg the vote agn will vote true agn so won't matter and countVotes func counts no.of true entries\n\t\t\t\t\tr.sendToAll(reqVoteObj)\n\t\t\t\t} else if timeout == ElectionTimeout {\n\t\t\t\t\twaitTime_msecs := msecs * time.Duration(waitTime)\n\t\t\t\t\tElectionTimer.Reset(waitTime_msecs)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func Timeout(timeout int64) Option {\n\treturn func(opts *options) {\n\t\topts.timeout = time.Duration(timeout) * time.Second\n\t}\n}", "func (node *Node) startElectionTimer() {\n\telectionTimeout := node.randElectionTimeout()\n\n\tnode.mu.Lock()\n\ttimerStartTerm := node.currentTerm\n\tnode.mu.Unlock()\n\n\tticker := time.NewTicker(10 * time.Millisecond)\n\tdefer ticker.Stop()\n\n\t// This loops wakes every 10ms and checks if the conditions are conducive\n\t// for starting an election. This is not the most efficient and\n\t// theoretically we could just wake up every electionTimeout, but this\n\t// reduces testability/log readability.\n\tfor {\n\t\t<-ticker.C\n\n\t\tnode.mu.Lock()\n\t\tif node.state != candidate && node.state != follower {\n\t\t\tlog.Printf(\"The node is in the %s state, no need to run election\", node.state)\n\t\t\tnode.mu.Unlock()\n\t\t\treturn\n\t\t}\n\n\t\t// If the timer was started in a previous term then we can back off\n\t\t// because a newer go routine would have been spawned cooresponding to\n\t\t// the new term.\n\t\tif node.currentTerm != timerStartTerm {\n\t\t\tlog.Printf(\"Election timer started in term %d but now node has latest term %d, so we can back off\", timerStartTerm, node.currentTerm)\n\t\t\treturn\n\t\t}\n\n\t\t// Run an election if we have reached the election timeout.\n\t\tif timePassed := time.Since(node.timeSinceTillLastReset); timePassed > electionTimeout {\n\t\t\tnode.runElection()\n\t\t\tnode.mu.Unlock()\n\t\t\treturn\n\t\t}\n\n\t\tnode.mu.Unlock()\n\t}\n}", "func testNonleaderStartElection(t *testing.T, state StateType) {\n\t// election timeout\n\tet := 10\n\tr := newTestRaft(1, []uint64{1, 2, 3}, et, 1, NewMemoryStorage())\n\tdefer closeAndFreeRaft(r)\n\tswitch state {\n\tcase StateFollower:\n\t\tr.becomeFollower(1, 2)\n\tcase StateCandidate:\n\t\tr.becomeCandidate()\n\t}\n\n\tfor i := 1; i < 2*et; i++ {\n\t\tr.tick()\n\t}\n\n\tif r.Term != 2 {\n\t\tt.Errorf(\"term = %d, want 2\", r.Term)\n\t}\n\tif r.state != StateCandidate {\n\t\tt.Errorf(\"state = %s, want %s\", r.state, StateCandidate)\n\t}\n\tif !r.votes[r.id] {\n\t\tt.Errorf(\"vote for self = false, want true\")\n\t}\n\tmsgs := r.readMessages()\n\tsort.Sort(messageSlice(msgs))\n\twmsgs := []pb.Message{\n\t\t{From: 1, FromGroup: pb.Group{NodeId: 1, GroupId: 1, RaftReplicaId: 1},\n\t\t\tTo: 2, ToGroup: pb.Group{NodeId: 2, GroupId: 1, RaftReplicaId: 2},\n\t\t\tTerm: 2, Type: pb.MsgVote},\n\t\t{From: 1, FromGroup: pb.Group{NodeId: 1, GroupId: 1, RaftReplicaId: 1},\n\t\t\tTo: 3, ToGroup: pb.Group{NodeId: 3, GroupId: 1, RaftReplicaId: 3},\n\t\t\tTerm: 2, Type: pb.MsgVote},\n\t}\n\tif !reflect.DeepEqual(msgs, wmsgs) {\n\t\tt.Errorf(\"msgs = %v, want %v\", msgs, wmsgs)\n\t}\n}", "func RandomTime(t time.Time, r time.Duration) time.Time {\n\treturn t.Add(-time.Duration(float64(r) * rand.Float64()))\n}", "func getDurationWithin(min, max int) time.Duration {\n\tjitter := max - min\n\tif jitter <= 0 {\n\t\tjitter = 1\n\t}\n\tduration := rand.Intn(jitter) + min\n\treturn time.Duration(duration) * time.Second\n}", "func Timeout(o int) interface {\n\ttimeoutOptionSetter\n} {\n\treturn &timeoutOption{o}\n}", "func remainingTime(t *testing.T) time.Duration {\n\tdeadline, hasDeadline := t.Deadline()\n\tif hasDeadline {\n\t\treturn time.Until(deadline) - time.Second // give us 1 second to clean up\n\t}\n\n\treturn DefaultResourceTimeout\n}", "func IdleTimeout(d time.Duration) Option {\n\treturn func(conf *config) {\n\t\tconf.idleTimeout = d\n\t}\n}", "func LoadTimeout() int {\n\treturn viper.GetInt(\"timeout\")\n}", "func getRandomPeer() *Peer {\n\t// iterate to count number of live peers\n\tlivepeers := 0\n\tpeerListLock.Lock()\n\tfor i := 0; i < MaxPeers; i++ {\n\t\t// if peer is not dead\n\t\tif peerList[i].expirationTimer != 0 {\n\t\t\t// increment livepeers\n\t\t\tlivepeers++\n\t\t}\n\t}\n\tpeerListLock.Unlock()\n\n\tif livepeers == 0 {\n\t\treturn &Peer{expirationTimer:0}\n\t}\n\n\t// generate random number\n\trand.Seed(time.Now().UnixNano())\n\t// rnum is the distance from the start of the list through all the live peers.\n\t// it will give us a \"kind of\" random peer selection\n\trnum := rand.Intn(MaxPeers)+10\n\tpeerListLock.Lock()\n\tfor {\n\t\tfor i := 0; i < MaxPeers; i++ {\n\t\t\t// if peer is not dead\n\t\t\tif peerList[i].expirationTimer != 0 {\n\t\t\t\t// decrement rnum\n\t\t\t\trnum--\n\n\t\t\t\t// if last peer\n\t\t\t\tif rnum == 0 {\n\t\t\t\t\t// return current peer\n\t\t\t\t\tpeerListLock.Unlock()\n\t\t\t\t\treturn &peerList[i]\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func (o MrScalarProvisioningTimeoutOutput) Timeout() pulumi.IntOutput {\n\treturn o.ApplyT(func(v MrScalarProvisioningTimeout) int { return v.Timeout }).(pulumi.IntOutput)\n}", "func (m *Machine) timedOut() error {\n\t// no max time specified, i.e. we can go on forever.\n\tif m.MaxRuntime == 0 {\n\t\treturn nil\n\t}\n\n\truntime := time.Now().Sub(m.StartedAt)\n\n\tif runtime > m.MaxRuntime {\n\t\treturn fmt.Errorf(\"Timed out after %f seconds\", runtime.Seconds())\n\t}\n\n\treturn nil\n}", "func (c *Communicator) Timeout() time.Duration {\n\treturn c.connInfo.TimeoutVal\n}", "func (cfg *Config) Timeout(msg hotstuff.TimeoutMsg) {\n\tif cfg.cfg == nil {\n\t\treturn\n\t}\n\tvar ctx context.Context\n\tcfg.timeoutCancel()\n\tctx, cfg.timeoutCancel = context.WithCancel(context.Background())\n\tcfg.cfg.Timeout(ctx, proto.TimeoutMsgToProto(msg), gorums.WithNoSendWaiting())\n}", "func randDuration(d time.Duration) time.Duration {\n\treturn time.Duration(rand.Int63n(int64(d)))\n}", "func (m *MockConfig) Timeout() time.Duration {\n\targs := m.Called()\n\treturn args.Get(0).(time.Duration)\n}", "func (cp *Pool) IdleTimeout() time.Duration {\n\tp := cp.pool()\n\tif p == nil {\n\t\treturn 0\n\t}\n\treturn p.IdleTimeout()\n}", "func Timeout(timeout time.Duration, timeoutFunction OnTimeout) crOption {\n\treturn func(cr *ConsumerRegistration) *ConsumerRegistration {\n\t\tcr.timeout = timeout\n\t\tcr.onTimeout = timeoutFunction\n\t\treturn cr\n\t}\n}", "func (rf *Raft) electionTicker() {\n\tfor rf.killed() == false {\n\t\ttime.Sleep(NewElectionTimeout())\n\n\t\trf.mu.Lock()\n\t\trole := rf.currentRole\n\t\tlastHeard := time.Since(rf.lastHeardFromLeader)\n\t\trf.mu.Unlock()\n\n\t\tswitch role {\n\t\tcase Candidate:\n\t\t\trf.startElection()\n\t\tcase Follower:\n\t\t\tif lastHeard > electionTimeoutMax {\n\t\t\t\trf.startElection()\n\t\t\t}\n\t\t}\n\t}\n}", "func (s *NamespaceWebhook) TimeoutSeconds() int32 { return 2 }", "func (o *BasicOptions) Timeout() int {\n\tif o.AdvancedOptions.Timeout == 0 {\n\t\treturn int(defaultTimeout)\n\t}\n\treturn o.AdvancedOptions.Timeout\n}", "func LeaseTimeout(duration time.Duration) func(*Locker) error {\n\treturn func(l *Locker) error {\n\t\tl.LeaseTimeout = duration\n\t\treturn nil\n\t}\n}", "func NewElection(url string, desc string, frozenAt string, name string,\n\topenreg bool, questions []*Question, shortName string,\n\tuseVoterAliases bool, votersHash string, votingEnd string,\n\tvotingStart string, k *Key) (*Election, *big.Int, error) {\n\tuuid, err := GenUUID()\n\tif err != nil {\n\t\t// glog.Error(\"Couldn't generate an election UUID\")\n\t\treturn nil, nil, err\n\t}\n\n\t// var pk *Key\n\t// var secret *big.Int\n\t// if k == nil {\n\t// \tif pk, secret, err = NewKey(); err != nil {\n\t// \t\t// glog.Error(\"Couldn't generate a new key for the election\")\n\t// \t\treturn nil, nil, err\n\t// \t}\n\t// } else {\n\t// \t// Take the public params from k to generate the key.\n\t// \tif pk, secret, err = NewKeyFromParams(k.Generator, k.Prime, k.ExponentPrime); err != nil {\n\t// \t\t// glog.Error(\"Couldn't generate a new key for the election\")\n\t// \t\treturn nil, nil, err\n\t// \t}\n\t// }\n\n\te := &Election{\n\t\tCastURL: url,\n\t\tDescription: desc,\n\t\tFrozenAt: frozenAt,\n\t\tName: name,\n\t\tOpenreg: openreg,\n\t\tPublicKey: nil,\n\t\tQuestions: questions,\n\t\tShortName: shortName,\n\t\tUseVoterAliases: useVoterAliases,\n\t\tUuid: uuid,\n\t\tVotersHash: votersHash,\n\t\tVotingEndsAt: votingEnd,\n\t\tVotingStartsAt: votingStart,\n\t}\n\n\t// Compute the JSON of the election and compute its hash\n\t//json, err := MarshalJSON(e)\n\t//if err != nil {\n\t//\tglog.Error(\"Couldn't marshal the election as JSON\")\n\t//\treturn nil, nil, err\n\t//}\n\t//\n\t//h := sha256.Sum256(json)\n\t//encodedHash := base64.StdEncoding.EncodeToString(h[:])\n\t//e.ElectionHash = encodedHash[:len(encodedHash)-1]\n\t//e.JSON = json\n\treturn e, nil, nil\n\t// return e, secret, nil\n}", "func Timeout(d time.Duration) ConfigOpt {\n\treturn func(c *Config) {\n\t\tc.transport.ResponseHeaderTimeout = d\n\t\tc.transport.TLSHandshakeTimeout = d\n\t\tc.dialer.Timeout = d\n\t}\n}", "func Timeout(t time.Duration) Option {\n\treturn func(c *Config) Option {\n\t\tprevious := c.Timeout\n\t\tc.Timeout = t\n\t\treturn Timeout(previous)\n\t}\n}" ]
[ "0.87743986", "0.8494453", "0.8336411", "0.7417603", "0.70064396", "0.6965721", "0.68317294", "0.67236114", "0.6643873", "0.6371365", "0.6225482", "0.6209789", "0.6203307", "0.60753953", "0.60102594", "0.60018104", "0.58360827", "0.57980955", "0.5686369", "0.568286", "0.5628036", "0.5601557", "0.5483584", "0.5468691", "0.5463967", "0.54460704", "0.54367745", "0.5427088", "0.5404799", "0.53473127", "0.5342691", "0.5295177", "0.5282705", "0.52048826", "0.52048826", "0.51980686", "0.51894116", "0.5187358", "0.5161108", "0.51457244", "0.51299185", "0.5122955", "0.5106949", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5096922", "0.5083765", "0.5069338", "0.50680065", "0.50606316", "0.50346076", "0.5029177", "0.501232", "0.50074464", "0.4998427", "0.497983", "0.49779364", "0.49767688", "0.49687016", "0.49597627", "0.49308723", "0.4926264", "0.49197268", "0.49187946", "0.49183822", "0.48969257", "0.4884286", "0.48747277", "0.48562014", "0.48512715", "0.48416445", "0.48397267", "0.483949", "0.48185635", "0.48159266", "0.48148572", "0.48074925", "0.48024315", "0.4790902", "0.4782937", "0.47777525", "0.47717297", "0.4769083", "0.47524095", "0.47478002", "0.4740447", "0.4736398", "0.47265917", "0.47191814", "0.47073805", "0.47015274", "0.46945593" ]
0.8699186
1
return true if log1 is strictly more uptodate than log2
func moreUpToDate(lastLogIndex1 int, lastLogTerm1 int, lastLogIndex2 int, lastLogTerm2 int) bool { ans := false if lastLogTerm1 != lastLogTerm2 { ans = lastLogTerm1 > lastLogTerm2 } else { ans = lastLogIndex1 > lastLogIndex2 } DPrintf("[moreuptodate] %v %v , %v %v, ans=%v", lastLogIndex1, lastLogTerm1, lastLogIndex2, lastLogTerm2, ans) return ans }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func compareLog(lastTermA int, lastIndexA int, lastTermB int, lastIndexB int) bool {\n if lastTermA != lastTermB {\n return lastTermA < lastTermB\n }\n return lastIndexA < lastIndexB\n}", "func (rf *Raft) LogUpToDate(lastIndex int, lastTerm int) bool {\n\t// if empty log, check snapshot\n\tif len(rf.log) == 0 {\n\t\treturn (lastTerm > rf.lastIncludedTerm) || (lastTerm == rf.lastIncludedTerm && lastIndex >= rf.lastIncludedIndex)\n\t}\n\tlastEntry := rf.log[len(rf.log)-1]\n\treturn (lastTerm > lastEntry.Term) || (lastTerm == lastEntry.Term && lastIndex >= lastEntry.Index)\n}", "func (rf *Raft) atLeastUptodate(candidateLastLogIdx int , candidateLastLogTerm int) bool{\n\trevLastLogIdx, revLastLogTerm := rf.lastLogIdxAndTerm()\n\n\tif candidateLastLogTerm > revLastLogTerm {\n\t\treturn true\n\t}else if candidateLastLogTerm < revLastLogTerm {\n\n\t\treturn false\n\t}else{\n\t\t//candidateLastLogTerm == revLastLogTerm\n\t\treturn candidateLastLogIdx >= revLastLogIdx\n\t}\n}", "func (l *Log) checkEquivalence(log2 *Log, wl *warningList) {\n\t// Description and STH comparison are omitted.\n\tif !bytes.Equal(l.Key, log2.Key) {\n\t\twl.addWarning(fmt.Sprintf(\n\t\t\t\"Log %q and log %q have different keys.\",\n\t\t\tl.Description, log2.Description))\n\t}\n\tif l.MaximumMergeDelay != log2.MaximumMergeDelay {\n\t\twl.addWarning(fmt.Sprintf(\n\t\t\t\"Maximum merge delay mismatch for logs %q and %q: %d != %d.\",\n\t\t\tl.Description, log2.Description, l.MaximumMergeDelay,\n\t\t\tlog2.MaximumMergeDelay))\n\t}\n\t// Strong assumption: operators IDs are semantically same across logs.\n\tlog1Ops := l.OperatedBy\n\tlog2Ops := log2.OperatedBy\n\tsort.IntSlice(log1Ops).Sort()\n\tsort.IntSlice(log2Ops).Sort()\n\tif !reflect.DeepEqual(log1Ops, log2Ops) {\n\t\twl.addWarning(fmt.Sprintf(\n\t\t\t\"Operators mismatch for logs %q and %q.\",\n\t\t\tl.Description, log2.Description))\n\t}\n\tif l.URL != log2.URL {\n\t\twl.addWarning(fmt.Sprintf(\n\t\t\t\"URL mismatch for logs %q and %q: %s != %s.\",\n\t\t\tl.Description, log2.Description, l.URL, log2.URL))\n\t}\n\tif l.DisqualifiedAt != log2.DisqualifiedAt {\n\t\twl.addWarning(fmt.Sprintf(\n\t\t\t\"Disqualified-at-timing mismatch for logs %q and %q: %v != %v.\",\n\t\t\tl.Description, log2.Description,\n\t\t\tct.TimestampToTime(uint64(l.DisqualifiedAt)),\n\t\t\tct.TimestampToTime(uint64(log2.DisqualifiedAt))))\n\t}\n\tif l.DNSAPIEndpoint != log2.DNSAPIEndpoint {\n\t\twl.addWarning(fmt.Sprintf(\n\t\t\t\"DNS API mismatch for logs %q and %q: %s != %s.\",\n\t\t\tl.Description, log2.Description, l.DNSAPIEndpoint,\n\t\t\tlog2.DNSAPIEndpoint))\n\t}\n}", "func isNewerLog(aTerm, aIdx int, bTerm, bIdx int) bool {\n\tif aTerm != bTerm {\n\t\treturn aTerm > bTerm\n\t}\n\treturn aIdx >= bIdx\n}", "func noUpdates(lastSrvBytes *uint, currSrvBytes *uint) bool {\n\tif *currSrvBytes > *lastSrvBytes {\n\t\t// \"new\" threshold\n\t\t*lastSrvBytes = *currSrvBytes\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (rf *Raft) isCandidateUpToDate(args *RequestVoteArgs) bool {\n\t/*\n\tRaft determines which of two logs is more up-to-date by\n\tcomparing the index and term of the last entries in the logs.\n\tIf the logs have last entries with different terms, then the log with the\n\tlater term is more up-to-date. If the logs end with the same term,\n\tthen whichever log is longer is more up-to-date.\n\t */\n\tif args.Term < rf.currentTerm {\n\t\treturn false\n\t}\n\tif args.LastLogTerm < rf.currentTerm {\n\t\treturn false\n\t}\n\tif args.LastLogIndex < len(rf.log) - 1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func moreOrLessEqual(a, b, err float64) bool {\n\treturn math.Abs(a-b) < err\n}", "func logCompare(mainLog string, currLog string) {\r\n\t//if current log is empty (no violators); then cat /dev/null > mainlog\r\n\tif _, err := os.Stat(currLog); err != nil {\r\n\t\tif err := os.Truncate(mainLog, 0); err != nil {\r\n\t\t\tlog.Printf(\"Failed to truncate main log: %v\", err)\r\n\t\t}\r\n\t} else {\r\n\t\tinput, err := ioutil.ReadFile(mainLog)\r\n\t\tcheck(err)\r\n\t\tlines := strings.Split(string(input), \"\\n\")\r\n\t\tvar newLines []string\r\n\t\tfor i, line := range lines {\r\n\t\t\tword := strings.Split(lines[i], \",\")\r\n\t\t\tUID := word[0]\r\n\t\t\tcheckId := IsExist(UID, currLog)\r\n\t\t\tif checkId {\r\n\t\t\t\tnewLines = append(newLines, line)\r\n\t\t\t}\r\n\t\t}\r\n\t\toutput := strings.Join(newLines, \"\\n\")\r\n\t\terr = ioutil.WriteFile(mainLog, []byte(output), 0644)\r\n\t\tcheck(err)\r\n\t}\r\n\r\n}", "func cmpTransferLog(a types.Log, b types.Log) int {\n\tif a.BlockNumber > b.BlockNumber {\n\t\treturn 1\n\t} else if a.BlockNumber < b.BlockNumber {\n\t\treturn -1\n\t}\n\tif a.TxIndex > b.TxIndex {\n\t\treturn 1\n\t} else if a.TxIndex < b.TxIndex {\n\t\treturn -1\n\t}\n\treturn 0\n}", "func loggersEqual(actual, expected *Logger) error {\n\tif actual.out != expected.out {\n\t\treturn fmt.Errorf(\"output does not match\")\n\t}\n\n\tif err := approxInt64(actual.timestamp, expected.timestamp, 100 /* ms */); err != nil {\n\t\treturn fmt.Errorf(\"timestamp %v\", err)\n\t}\n\n\treturn nil\n}", "func (r *RaftNode) mergeLogEntries(req *AppendEntriesRequest) (success bool) {\n\n\tr.leaderMutex.Lock()\n\tdefer r.leaderMutex.Unlock()\n\n\tentries := req.GetEntries()\n\n\t// if prevLogIndex is out of range, cannot merge\n\tif req.GetPrevLogIndex() < 0 || req.GetPrevLogIndex() > r.getLastLogIndex() {\n\t\tr.Out(\"MERGING: Couldn't find prev\")\n\t\treturn false\n\t}\n\n\t// if log doesn't contain entry at prevLogIndex with term PrevLogTerm, cannot merge\n\tfollowerPrevLog := r.getLogEntry(req.GetPrevLogIndex())\n\tif followerPrevLog.TermId != req.GetPrevLogTerm() {\n\t\tr.Out(\"MERGING: Couldn't find prevEntry with term = %d; index = %d\", req.GetPrevLogTerm(), req.GetPrevLogIndex())\n\t\treturn false\n\t}\n\n\t// if there are entries present, merge them\n\tif entries != nil && len(entries) != 0 {\n\t\tfor i := range entries {\n\t\t\t// index of where we would insert the new item\n\t\t\tinsertAt := uint64(i) + req.GetPrevLogIndex() + 1\n\t\t\tif entries[i].GetIndex() != insertAt {\n\t\t\t\tr.Error(\"Request doesn't have correct index!! State corrupted.\")\n\t\t\t}\n\t\t\tr.Out(\"Merging logs: adding %v\", entries[i])\n\t\t\tif insertAt <= r.getLastLogIndex() {\n\t\t\t\tr.truncateLog(insertAt)\n\t\t\t\tr.appendLogEntry(*entries[i])\n\t\t\t} else {\n\t\t\t\t// if we go past the end of the log (or remove entries above), keep appending\n\t\t\t\tr.appendLogEntry(*entries[i])\n\t\t\t}\n\t\t}\n\t}\n\n\t// apply all logEntries up until leader's commitIndex to statemachine\n\tif req.GetLeaderCommit() > r.commitIndex {\n\t\tnewCommitIndex := min(req.GetLeaderCommit(), r.getLastLogIndex())\n\t\t// start at +1 since commitIndex has already been committed\n\t\tfor i := r.commitIndex + 1; i <= newCommitIndex; i++ {\n\t\t\tentry := r.getLogEntry(i)\n\t\t\tr.Out(\"COMMITTING index=%v;term=%v\", entry.GetIndex(), entry.GetTermId())\n\t\t\tif entry.Type == CommandType_STATE_MACHINE_COMMAND {\n\t\t\t\tresponse, err := r.stateMachine.ApplyCommand(entry.Command, entry.Data)\n\t\t\t\tif err != nil {\n\t\t\t\t\tr.Error(\"State machine error: %v (response: %s)\", err, response)\n\t\t\t\t}\n\t\t\t}\n\t\t\tr.lastApplied = i\n\t\t}\n\t\tr.commitIndex = newCommitIndex\n\t}\n\n\tr.Verbose(\"Merge successful\")\n\treturn true\n}", "func (l Log) Less(i, j int) bool {\n\treturn time.Time(l[i].Committer.Time).Before(time.Time(l[j].Committer.Time))\n}", "func (rf *Raft) AtLeastAsUpToDate(candidate RequestVoteArgs) bool {\n\tlastLogEntry := rf.lastLogEntry() // NOTE: this could be \"zero\" struct\n\tswitch {\n\tcase candidate.LastLogTerm > lastLogEntry.Term:\n\t\treturn true\n\tcase candidate.LastLogTerm == lastLogEntry.Term:\n\t\treturn candidate.LastLogIndex >= rf.lastApplied // is lastApplied correct here?\n\tcase candidate.LastLogTerm < lastLogEntry.Term:\n\t\treturn false\n\tdefault: // TODO need this?\n\t\treturn false\n\t}\n}", "func bookChanged(timestamps1, timestamps2 []float64) bool {\n\tfor i := 0; i < 40; i++ {\n\t\tif math.Abs(timestamps1[i]-timestamps2[i]) > .5 {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func diffLoggingCmp(logger *zap.SugaredLogger) cmpFunc {\n\treturn func(x, y interface{}, opts ...cmp.Option) bool {\n\t\tif diff := cmp.Diff(x, y, opts...); diff != \"\" {\n\t\t\tlogger.Debug(\"Event subscriptions differ (-want, +got)\\n\" + diff)\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t}\n}", "func isUpToDate(ctx context.Context, kube client.Client, in *v1alpha1.ReleaseParameters, observed *release.Release, s v1alpha1.ReleaseStatus) (bool, error) {\n\toc := observed.Chart\n\tif oc == nil {\n\t\treturn false, errors.New(errChartNilInObservedRelease)\n\t}\n\n\tocm := oc.Metadata\n\tif ocm == nil {\n\t\treturn false, errors.New(errChartMetaNilInObservedRelease)\n\t}\n\tif in.Chart.Name != ocm.Name {\n\t\treturn false, nil\n\t}\n\tif in.Chart.Version != ocm.Version {\n\t\treturn false, nil\n\t}\n\tdesiredConfig, err := composeValuesFromSpec(ctx, kube, in.ValuesSpec)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, errFailedToComposeValues)\n\t}\n\n\tif !reflect.DeepEqual(desiredConfig, observed.Config) {\n\t\treturn false, nil\n\t}\n\n\tchanged, err := newPatcher().hasUpdates(ctx, kube, in.PatchesFrom, s)\n\tif err != nil {\n\t\treturn false, errors.Wrap(err, errFailedToLoadPatches)\n\t}\n\n\tif changed {\n\t\treturn false, nil\n\t}\n\n\treturn true, nil\n}", "func almostEquals(a, b float64) bool {\n\treturn math.Abs(a-b) < threshold\n}", "func (rf *Raft) correctPrevLogEntry(PrevLogIndex int, PrevLogTerm int) bool {\n\t// if no log, have to check lastIncludedIndex and lastIncludedTerm\n\tif len(rf.log) == 0 {\n\t\treturn PrevLogIndex == rf.lastIncludedIndex && PrevLogTerm == rf.lastIncludedTerm\n\t}\n\tprevRaftLogIndex := rf.getTrimmedLogIndex(PrevLogIndex)\n\t// the leader nextIndex is ahead of us\n\tif prevRaftLogIndex >= len(rf.log) {\n\t\treturn false\n\t}\n\n\t// NOTE:\n\t// if prevRaftLogIndex == -1 ... this should never happen?\n\t// We know length of rf.log > 0 (see where this function is called), so this\n\t// would only occur if leader nextIndex for this server preceded our snapshot;\n\t// but on leader election, nextIndex is set to the end of the leader log,\n\t// including all committed entries.\n\t// However, our snapshot includes AT MOST all committed entries,\n\t// so nextIndex should never precede it.\n\tif prevRaftLogIndex == -1 && len(rf.log) > 0 {\n\t\trf.Log(LogInfo, \"AppendEntries call has PrevLogIndex preceding our log!\")\n\t\treturn true\n\t}\n\n\t// we must have an entry at the given index (see above note for why\n\t// PrevLogIndex will never precede our snapshot), so just return a bool for whether\n\t// or not the term of this entry is correct\n\treturn rf.log[prevRaftLogIndex].Term == PrevLogTerm\n\n}", "func (t *Timestamp) IsUpToDate() (bool, error) {\n\tif len(t.Sources) == 0 || len(t.Generates) == 0 {\n\t\treturn false, nil\n\t}\n\n\tsources, err := glob(t.Dir, t.Sources)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\tgenerates, err := glob(t.Dir, t.Generates)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\n\tsourcesMaxTime, err := getMaxTime(sources...)\n\tif err != nil || sourcesMaxTime.IsZero() {\n\t\treturn false, nil\n\t}\n\n\tgeneratesMinTime, err := getMinTime(generates...)\n\tif err != nil || generatesMinTime.IsZero() {\n\t\treturn false, nil\n\t}\n\n\treturn !generatesMinTime.Before(sourcesMaxTime), nil\n}", "func (d dataUpdateTrackerHistory) sort() bool {\n\tif len(d) == 0 {\n\t\treturn true\n\t}\n\tsort.Slice(d, func(i, j int) bool {\n\t\treturn d[i].idx > d[j].idx\n\t})\n\treturn d[0].idx-d[len(d)-1].idx == uint64(len(d))\n}", "func logdist(a, b common.Hash) int {\n\tlz := 0\n\tfor i := range a {\n\t\tx := a[i] ^ b[i]\n\t\tif x == 0 {\n\t\t\tlz += 8\n\t\t} else {\n\t\t\tlz += lzcount[x]\n\t\t\tbreak\n\t\t}\n\t}\n\treturn len(a)*8 - lz\n}", "func (l *RaftLog) appendEntries(prevTerm uint64, prevIndex uint64, commitIndex uint64, ents []pb.Entry) bool {\n\tif prevIndex > l.LastIndex() {\n\t\treturn false\n\t}\n\tt, err := l.Term(prevIndex)\n\tif err != nil {\n\t\treturn false\n\t}\n\tif t != prevTerm {\n\t\treturn false\n\t}\n\tif len(ents) <= 0 {\n\t\tif commitIndex > l.committed {\n\t\t\tcanCommit := mathutil.MinUint64Val(commitIndex, prevIndex+uint64(len(ents)))\n\t\t\tl.committed = mathutil.MaxUint64(l.committed, canCommit)\n\t\t}\n\t\treturn true\n\t}\n\tchangeEnts := findMergeEntries(l.entries, ents)\n\tl.pendingEntries = mergeEntries(l.pendingEntries, changeEnts)\n\tl.entries = mergeEntries(l.entries, changeEnts)\n\tif commitIndex > l.committed {\n\t\tcanCommit := mathutil.MinUint64Val(commitIndex, prevIndex+uint64(len(ents)))\n\t\tl.committed = mathutil.MaxUint64(l.committed, canCommit)\n\t}\n\treturn true\n}", "func almostEqual64(a, b, tol float64) bool {\n\treturn math.Abs(a-b) <= tol\n}", "func (c *HBComp) logVersion() {\n\tnow := time.Now()\n\tif now.Sub(c.lastVersionLogTime) >= c.app.VersionLogPeriod {\n\t\tc.Log.Info(c.app.InvocationArgs)\n\t\tc.lastVersionLogTime = now\n\t}\n}", "func Lesser(x, y *big.Float) bool {\n\treturn x.Cmp(y) == -1\n}", "func (n *Node) checkIfOldestFuturesCanReplace(oldestFutureMsgs []*hbft.OldestFutureMsg, currentSeqID, currentViewID uint64) (bool, common.Hash, bool) {\r\n\tdifferentFutureCntMap := make(map[common.Hash]int)\r\n\tdifferentFutureMsgMap := make(map[common.Hash]*hbft.OldestFutureMsg)\r\n\tdifferentFuturePrimMap := make(map[string]int)\r\n\r\n\t// Check view and signature\r\n\tcnt := 0\r\n\tfor _, oldestFutureMsg := range oldestFutureMsgs {\r\n\t\tif oldestFutureMsg.SequenceID == currentSeqID &&\r\n\t\t\toldestFutureMsg.ViewID == currentViewID {\r\n\t\t\tif err := n.checkMsgSignature(oldestFutureMsg); err != nil {\r\n\t\t\t\tn.HBFTDebugInfo(fmt.Sprintf(\"checkIfOldestFuturesCanReplace failed, check signature failed, %s\", err.Error()))\r\n\t\t\t\treturn false, common.Hash{}, false\r\n\t\t\t}\r\n\t\t\tif _, ok := differentFuturePrimMap[oldestFutureMsg.ViewPrimary]; ok {\r\n\t\t\t\tn.HBFTDebugInfo(\"checkIfOldestFuturesCanReplace failed, duplicated prim\")\r\n\t\t\t\treturn false, common.Hash{}, false\r\n\t\t\t} else {\r\n\t\t\t\tdifferentFuturePrimMap[oldestFutureMsg.ViewPrimary] = 1\r\n\t\t\t}\r\n\r\n\t\t\tif oldestFutureMsg.OldestFuture != nil {\r\n\t\t\t\tdifferentFutureMsgMap[oldestFutureMsg.OldestFuture.Hash()] = oldestFutureMsg\r\n\t\t\t} else {\r\n\t\t\t\tdifferentFutureMsgMap[common.Hash{}] = oldestFutureMsg\r\n\t\t\t}\r\n\t\t\tcnt++\r\n\t\t}\r\n\t}\r\n\tif cnt <= n.EleBackend.Get2fRealSealersCnt() {\r\n\t\tn.HBFTDebugInfo(\"checkIfOldestFuturesCanReplace failed, check view failed\")\r\n\t\treturn false, common.Hash{}, false\r\n\t}\r\n\r\n\tfor _, oldestFutureMsg := range oldestFutureMsgs {\r\n\t\tif !oldestFutureMsg.NoAnyFutures {\r\n\t\t\tif _, ok := differentFutureCntMap[oldestFutureMsg.OldestFuture.Hash()]; !ok {\r\n\t\t\t\tdifferentFutureCntMap[oldestFutureMsg.OldestFuture.Hash()] = 1\r\n\t\t\t} else {\r\n\t\t\t\tdifferentFutureCntMap[oldestFutureMsg.OldestFuture.Hash()] += 1\r\n\t\t\t}\r\n\t\t} else {\r\n\t\t\tcontinue\r\n\t\t}\r\n\t}\r\n\r\n\tif len(differentFutureCntMap) == 1 {\r\n\t\tfor hash, cnt := range differentFutureCntMap {\r\n\t\t\tif cnt > n.EleBackend.Get2fRealSealersCnt()/2 {\r\n\t\t\t\treturn true, hash, true\r\n\t\t\t} else {\r\n\t\t\t\treturn true, hash, false\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n\r\n\toldestFutureBlockToReturn := common.Hash{}\r\n\tmaxValue := uint64(0)\r\n\tfor hash, cnt := range differentFutureCntMap {\r\n\t\tif cnt > n.EleBackend.Get2fRealSealersCnt()/2 {\r\n\t\t\treturn true, hash, true\r\n\t\t} else {\r\n\t\t\tif differentFutureMsgMap[hash].Completed.ReqTimeStamp > maxValue && !common.EmptyHash(hash) {\r\n\t\t\t\tmaxValue = differentFutureMsgMap[hash].Completed.ReqTimeStamp\r\n\t\t\t\toldestFutureBlockToReturn = hash\r\n\t\t\t}\r\n\t\t\t//if differentFutureMsgMap[hash].Completed.BlockNum == 1 {\r\n\t\t\t//\tif differentFutureMsgMap[hash].Completed.ReqTimeStamp > maxValue {\r\n\t\t\t//\t\tmaxValue = differentFutureMsgMap[hash].Completed.ReqTimeStamp\r\n\t\t\t//\t\toldestFutureBlockToReturn = hash\r\n\t\t\t//\t}\r\n\t\t\t//} else {\r\n\t\t\t//\tif differentFutureMsgMap[hash].Completed.ReqTimeStamp > maxValue {\r\n\t\t\t//\t\tmaxValue = differentFutureMsgMap[hash].Completed.ReqTimeStamp\r\n\t\t\t//\t\toldestFutureBlockToReturn = hash\r\n\t\t\t//\t}\r\n\t\t\t//}\r\n\t\t}\r\n\t}\r\n\tif !common.EmptyHash(oldestFutureBlockToReturn) {\r\n\t\treturn true, oldestFutureBlockToReturn, true\r\n\t}\r\n\r\n\tn.HBFTDebugInfo(\"checkIfOldestFuturesCanReplace failed\")\r\n\treturn false, common.Hash{}, false\r\n}", "func (r *Raft) updateCommitted() bool {\n\t// update committed\n\tvar matches []int\n\tfor _, v := range r.Prs {\n\t\tmatches = append(matches, int(v.Match))\n\t}\n\tsort.Ints(matches)\n\tm := uint64(matches[((len(matches) - 1) / 2)])\n\n\tt, err := r.RaftLog.Term(m)\n\tif err != nil {\n\t\tpanic(\"term not found\")\n\t}\n\n\tif t != r.Term {\n\t\treturn false\n\t}\n\n\tif m > r.RaftLog.committed {\n\t\tr.RaftLog.committed = m\n\t\treturn true\n\t}\n\treturn false\n}", "func semvVerAGreaterThanB(a, b string) bool {\n\tverA, aErr := semver.NewVersion(a)\n\tverB, bErr := semver.NewVersion(b)\n\n\treturn aErr == nil && bErr == nil && verA.GreaterThan(verB)\n}", "func VectorChangesBefore(c1, c2 *ChangeEntry) bool {\n\n\t// Empty value handling\n\tif c1 == nil || c2 == nil {\n\t\tif c1 != nil {\n\t\t\treturn true\n\t\t} else {\n\t\t\treturn false\n\t\t}\n\t}\n\n\t// We want the following prioritization for changes entries:\n\t// 1. Backfill in progress\n\t// a. If both entries represent entries from an active backfill, compare by vb.seq\n\t// 2. No backfill in progress\n\t// a. Non-backfill and backfill_pending are prioritized over backfill_complete notifications, when vbseq are equal\n\t// b. If both entries are non-backfill, compare by vb.seq\n\t// c. If both entries are backfill_complete notifications, compare by vb.seq\n\ts1 := c1.Seq\n\ts2 := c2.Seq\n\n\t// Case 1 checks\n\tif s1.TriggeredByClock != nil {\n\t\tif s2.TriggeredByClock == nil {\n\t\t\t// c1 is backfill, c2 isn't\n\t\t\treturn true\n\t\t} else {\n\t\t\t// Both backfill, Case 1a.\n\t\t\treturn s1.VbucketSequenceBefore(s2.vbNo, s2.Seq)\n\t\t}\n\t}\n\tif s2.TriggeredByClock != nil {\n\t\t// c2 is backfill, c1 isn't\n\t\treturn false\n\t}\n\n\t// Case 2 checks\n\tif c1.backfill == BackfillFlag_Complete {\n\t\tif c2.backfill == BackfillFlag_Complete {\n\t\t\t// Case 2c. Both backfill complete notifications\n\t\t\tif s1.vbNo == s2.vbNo {\n\t\t\t\treturn s1.Seq < s2.Seq\n\t\t\t} else {\n\t\t\t\treturn s1.vbNo < s2.vbNo\n\t\t\t}\n\t\t} else {\n\t\t\t// 2a. if equal, give priority to non-backfill only when vb.seq equal\n\t\t\tif s1.vbNo == s2.vbNo && s1.Seq == s2.Seq {\n\t\t\t\treturn false\n\t\t\t} else {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\tif c2.backfill == BackfillFlag_Complete {\n\t\t// 2a - if equal, give priority to non-backfill\n\t\tif s1.vbNo == s2.vbNo && s1.Seq == s2.Seq {\n\t\t\treturn true\n\t\t} else {\n\t\t\treturn false\n\t\t}\n\t}\n\n\t// Case 2b.\n\treturn s1.VbucketSequenceBefore(s2.vbNo, s2.Seq)\n\n}", "func equal(a, b float64) bool {\n\treturn math.Abs(a-b) <= equalityThreshold\n}", "func (checker *TimestampChecker) IsUpToDate(t *taskfile.Task) (bool, error) {\n\tif len(t.Sources) == 0 {\n\t\treturn false, nil\n\t}\n\n\tsources, err := Globs(t.Dir, t.Sources)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\tgenerates, err := Globs(t.Dir, t.Generates)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\n\ttimestampFile := checker.timestampFilePath(t)\n\n\t// If the file exists, add the file path to the generates.\n\t// If the generate file is old, the task will be executed.\n\t_, err = os.Stat(timestampFile)\n\tif err == nil {\n\t\tgenerates = append(generates, timestampFile)\n\t} else {\n\t\t// Create the timestamp file for the next execution when the file does not exist.\n\t\tif !checker.dry {\n\t\t\tif err := os.MkdirAll(filepath.Dir(timestampFile), 0o755); err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\tf, err := os.Create(timestampFile)\n\t\t\tif err != nil {\n\t\t\t\treturn false, err\n\t\t\t}\n\t\t\tf.Close()\n\t\t}\n\t}\n\n\ttaskTime := time.Now()\n\n\t// Compare the time of the generates and sources. If the generates are old, the task will be executed.\n\n\t// Get the max time of the generates.\n\tgenerateMaxTime, err := getMaxTime(generates...)\n\tif err != nil || generateMaxTime.IsZero() {\n\t\treturn false, nil\n\t}\n\n\t// Check if any of the source files is newer than the max time of the generates.\n\tshouldUpdate, err := anyFileNewerThan(sources, generateMaxTime)\n\tif err != nil {\n\t\treturn false, nil\n\t}\n\n\t// Modify the metadata of the file to the the current time.\n\tif !checker.dry {\n\t\tif err := os.Chtimes(timestampFile, taskTime, taskTime); err != nil {\n\t\t\treturn false, err\n\t\t}\n\t}\n\n\treturn !shouldUpdate, nil\n}", "func (e *Entry) timeToUpdate() bool {\n\tnow := e.clk.Now()\n\te.mu.RLock()\n\tdefer e.mu.RUnlock()\n\t// no response or nextUpdate is in the past\n\tif e.response == nil || e.nextUpdate.Before(now) {\n\t\te.info(\"Stale response, updating immediately\")\n\t\treturn true\n\t}\n\tif e.maxAge > 0 {\n\t\t// cache max age has expired\n\t\tif e.lastSync.Add(e.maxAge).Before(now) {\n\t\t\te.info(\"max-age has expired, updating immediately\")\n\t\t\treturn true\n\t\t}\n\t}\n\n\t// update window is last quarter of NextUpdate - ThisUpdate\n\t// TODO: support using NextPublish instead of ThisUpdate if provided\n\t// in responses\n\twindowSize := e.nextUpdate.Sub(e.thisUpdate) / 4\n\tupdateWindowStarts := e.nextUpdate.Add(-windowSize)\n\tif updateWindowStarts.After(now) {\n\t\treturn false\n\t}\n\n\t// randomly pick time in update window\n\tupdateTime := updateWindowStarts.Add(time.Second * time.Duration(mrand.Intn(int(windowSize.Seconds()))))\n\tif updateTime.Before(now) {\n\t\te.info(\"Time to update\")\n\t\treturn true\n\t}\n\treturn false\n}", "func IsUpToDate(p v1beta1.QueueParameters, attributes map[string]string, tags map[string]string) bool { // nolint:gocyclo\n\tif len(p.Tags) != len(tags) {\n\t\treturn false\n\t}\n\n\tfor k, v := range p.Tags {\n\t\tpVal, ok := tags[k]\n\t\tif !ok || !strings.EqualFold(pVal, v) {\n\t\t\treturn false\n\t\t}\n\t}\n\n\tif aws.ToInt64(p.DelaySeconds) != toInt64(attributes[v1beta1.AttributeDelaySeconds]) {\n\t\treturn false\n\t}\n\tif aws.ToInt64(p.KMSDataKeyReusePeriodSeconds) != toInt64(attributes[v1beta1.AttributeKmsDataKeyReusePeriodSeconds]) {\n\t\treturn false\n\t}\n\tif aws.ToInt64(p.MaximumMessageSize) != toInt64(attributes[v1beta1.AttributeMaximumMessageSize]) {\n\t\treturn false\n\t}\n\tif aws.ToInt64(p.MessageRetentionPeriod) != toInt64(attributes[v1beta1.AttributeMessageRetentionPeriod]) {\n\t\treturn false\n\t}\n\tif aws.ToInt64(p.ReceiveMessageWaitTimeSeconds) != toInt64(attributes[v1beta1.AttributeReceiveMessageWaitTimeSeconds]) {\n\t\treturn false\n\t}\n\tif aws.ToInt64(p.VisibilityTimeout) != toInt64(attributes[v1beta1.AttributeVisibilityTimeout]) {\n\t\treturn false\n\t}\n\tif !cmp.Equal(aws.ToString(p.KMSMasterKeyID), attributes[v1beta1.AttributeKmsMasterKeyID]) {\n\t\treturn false\n\t}\n\tif !cmp.Equal(aws.ToString(p.Policy), attributes[v1beta1.AttributePolicy]) {\n\t\treturn false\n\t}\n\tif attributes[v1beta1.AttributeContentBasedDeduplication] != \"\" && strconv.FormatBool(aws.ToBool(p.ContentBasedDeduplication)) != attributes[v1beta1.AttributeContentBasedDeduplication] {\n\t\treturn false\n\t}\n\tif p.RedrivePolicy != nil {\n\t\tr := map[string]interface{}{\n\t\t\t\"deadLetterTargetArn\": p.RedrivePolicy.DeadLetterTargetARN,\n\t\t\t\"maxReceiveCount\": p.RedrivePolicy.MaxReceiveCount,\n\t\t}\n\t\tval, err := json.Marshal(r)\n\t\tif err == nil {\n\t\t\tif string(val) != attributes[v1beta1.AttributeRedrivePolicy] {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\treturn true\n}", "func (cr *ConsulRegistry) checkUpdate(key string, index uint64) bool {\n\tvar lastIndex uint64 = 0\n\tloaded, ok := cr.lastIndexes.Load(key)\n\tif ok {\n\t\tlastIndex = loaded.(uint64)\n\t}\n\tif index <= lastIndex {\n\t\treturn false\n\t}\n\tcr.lastIndexes.Store(key, index)\n\treturn true\n}", "func compare_clocks(new_clock ClockVector, last_clock ClockVector) bool {\n\treturn new_clock.X >= last_clock.X && new_clock.Y >= last_clock.Y && new_clock.Z >= last_clock.Z\n}", "func equalBubbledMetadata(md1, md2 siadir.Metadata) error {\n\t// Check AggregateHealth\n\tif md1.AggregateHealth != md2.AggregateHealth {\n\t\treturn fmt.Errorf(\"AggregateHealth not equal, %v and %v\", md1.AggregateHealth, md2.AggregateHealth)\n\t}\n\t// Check AggregateNumFiles\n\tif md1.AggregateNumFiles != md2.AggregateNumFiles {\n\t\treturn fmt.Errorf(\"AggregateNumFiles not equal, %v and %v\", md1.AggregateNumFiles, md2.AggregateNumFiles)\n\t}\n\t// Check Size\n\tif md1.AggregateSize != md2.AggregateSize {\n\t\treturn fmt.Errorf(\"aggregate sizes not equal, %v and %v\", md1.AggregateSize, md2.AggregateSize)\n\t}\n\t// Check Health\n\tif md1.Health != md2.Health {\n\t\treturn fmt.Errorf(\"healths not equal, %v and %v\", md1.Health, md2.Health)\n\t}\n\t// Check LastHealthCheckTimes\n\tif md2.LastHealthCheckTime != md1.LastHealthCheckTime {\n\t\treturn fmt.Errorf(\"LastHealthCheckTimes not equal %v and %v\", md2.LastHealthCheckTime, md1.LastHealthCheckTime)\n\t}\n\t// Check MinRedundancy\n\tif md1.MinRedundancy != md2.MinRedundancy {\n\t\treturn fmt.Errorf(\"MinRedundancy not equal, %v and %v\", md1.MinRedundancy, md2.MinRedundancy)\n\t}\n\t// Check Mod Times\n\tif md2.ModTime != md1.ModTime {\n\t\treturn fmt.Errorf(\"ModTimes not equal %v and %v\", md2.ModTime, md1.ModTime)\n\t}\n\t// Check NumFiles\n\tif md1.NumFiles != md2.NumFiles {\n\t\treturn fmt.Errorf(\"NumFiles not equal, %v and %v\", md1.NumFiles, md2.NumFiles)\n\t}\n\t// Check NumStuckChunks\n\tif md1.NumStuckChunks != md2.NumStuckChunks {\n\t\treturn fmt.Errorf(\"NumStuckChunks not equal, %v and %v\", md1.NumStuckChunks, md2.NumStuckChunks)\n\t}\n\t// Check NumSubDirs\n\tif md1.NumSubDirs != md2.NumSubDirs {\n\t\treturn fmt.Errorf(\"NumSubDirs not equal, %v and %v\", md1.NumSubDirs, md2.NumSubDirs)\n\t}\n\t// Check StuckHealth\n\tif md1.StuckHealth != md2.StuckHealth {\n\t\treturn fmt.Errorf(\"stuck healths not equal, %v and %v\", md1.StuckHealth, md2.StuckHealth)\n\t}\n\treturn nil\n}", "func (a result) beats(b result) bool {\n\treturn a.n < b.n || (a.n == b.n && a.s < b.s)\n}", "func RevisionIsNewer(revision1, revision2 uint64) bool {\n\tif revision1 < v2MinRevision || revision2 < v2MinRevision {\n\t\treturn revision1 > revision2\n\t}\n\n\tresult, isSameCluster := revisionCompare(revision1, revision2)\n\tif isSameCluster {\n\t\treturn result > 0\n\t}\n\t// if from different cluster, allow sent result == 0\n\treturn result >= 0\n}", "func UpToDate(spec *v1alpha1.RecordParameters, o cloudflare.DNSRecord) bool { //nolint:gocyclo\n\t// NOTE(bagricola): The complexity here is simply repeated\n\t// if statements checking for updated fields. You should think\n\t// before adding further complexity to this method, but adding\n\t// more field checks should not be an issue.\n\tif spec == nil {\n\t\treturn true\n\t}\n\n\t// Check if mutable fields are up to date with resource\n\n\t// If the Spec Name doesn't have the zone name on the end of it\n\t// Add it on the end when checking the result from the API\n\t// As CF returns the name as the full DNS record (including zone name)\n\tfn := spec.Name\n\tif !strings.HasSuffix(fn, o.ZoneName) {\n\t\tfn = fn + \".\" + o.ZoneName\n\t}\n\n\tif fn != o.Name {\n\t\treturn false\n\t}\n\n\tif spec.Content != o.Content {\n\t\treturn false\n\t}\n\n\tif spec.TTL != nil && *spec.TTL != int64(o.TTL) {\n\t\treturn false\n\t}\n\n\tif spec.Proxied != nil && o.Proxied != nil && *spec.Proxied != *o.Proxied {\n\t\treturn false\n\t}\n\n\tif spec.Priority != nil && o.Priority != nil && *spec.Priority != int32(*o.Priority) {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (rf *Raft) CheckLogs() {\n\trf.mu.Lock()\n\tstate := rf.state\n\trf.mu.Unlock()\n\tfor state == LEADER {\n\t\t//DPrintf(\"CHECKLOGS ON NODE %d: logs %s\", rf.me, rf.logs)\n\t\t//appendChan := make(chan AppendResult, len(rf.peers))\n\t\tfor peerId := range rf.peers {\n\t\t\tif peerId == rf.me {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\trf.mu.Lock()\n\t\t\tlogLen := len(rf.logs)\n\t\t\tnextIndex := rf.nextIndex[peerId]\n\t\t\trf.mu.Unlock()\n\t\t\tif logLen > nextIndex {\n\t\t\t\tgo func(peerId int) {\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\tprevLogIndex := rf.matchIndex[peerId]\n\t\t\t\t\tprevLogTerm := rf.logs[prevLogIndex].Term\n\t\t\t\t\targs := AppendEntriesArgs{rf.currentTerm, rf.me,\n\t\t\t\t\t\tprevLogIndex, prevLogTerm,\n\t\t\t\t\t\trf.logs[prevLogIndex+1:], rf.commitIndex}\n\t\t\t\t\t\t//DPrintf(\"[BEFOREAPPEND] ENTRIES %s PREV %d LOGS %s\", args.Entries, args.PrevLogIndex, rf.logs)\n\t\t\t\t\trepl := AppendResult{}\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\tfor rf.state == LEADER {\n\t\t\t\t\t\trf.sendAppendEntries(peerId, &args, &repl)\n\t\t\t\t\t\t//DPrintf(\"[CHECKAPPENDENTRIES REPLY]me: %d Term %d send to %d args: %s repl %s\", rf.me, rf.currentTerm, peerId, args, repl)\n\t\t\t\t\t\tif repl.Success && rf.state == LEADER{\n\t\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\t\trf.nextIndex[peerId] = args.PrevLogIndex + len(args.Entries) + 1\n\t\t\t\t\t\t\trf.matchIndex[peerId] = args.PrevLogIndex + len(args.Entries)\n\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\tif repl.Term > rf.currentTerm {\n\t\t\t\t\t\t\trf.currentTerm = repl.Term\n\t\t\t\t\t\t\trf.state = FOLLOWER\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\tif args.PrevLogIndex > 0 {\n\t\t\t\t\t\t\targs.PrevLogIndex -= 1\n\t\t\t\t\t\t\targs.PrevLogTerm = rf.logs[args.PrevLogIndex].Term\n\t\t\t\t\t\t\targs.Entries = rf.logs[args.PrevLogIndex+1:]\n\t\t\t\t\t\t}\n\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t}\n\t\t\t\t}(peerId)\n\t\t\t}\n\t\t}\n\t\ttime.Sleep(time.Duration(50) * time.Millisecond)\n\t\t// sleep for a while\n\t}\n}", "func hasToBeUpdated(home, foreign []corev1.EndpointSubset) bool {\n\tif len(home) != len(foreign) {\n\t\tklog.V(6).Info(\"the ep has to be updated because home and foreign subsets lengths are different\")\n\t\treturn true\n\t}\n\tfor i := 0; i < len(home); i++ {\n\t\tif len(home[i].Addresses) != len(foreign[i].Addresses) {\n\t\t\tklog.V(6).Info(\"the ep has to be updated because home and foreign addresses lengths are different\")\n\t\t\treturn true\n\t\t}\n\t\tfor j := 0; j < len(home[i].Addresses); j++ {\n\t\t\tif home[i].Addresses[j].IP != foreign[i].Addresses[j].IP {\n\t\t\t\tklog.V(6).Info(\"the ep has to be updated because home and foreign IPs are different\")\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false\n}", "func generateVerifySessionHasUpToDateProfileTraceLogs(ctx *middlewares.AutheliaCtx, userSession *session.UserSession,\n\tdetails *authentication.UserDetails) {\n\tgroupsAdded, groupsRemoved := utils.StringSlicesDelta(userSession.Groups, details.Groups)\n\temailsAdded, emailsRemoved := utils.StringSlicesDelta(userSession.Emails, details.Emails)\n\tnameDelta := userSession.DisplayName != details.DisplayName\n\n\tfields := map[string]any{\"username\": userSession.Username}\n\tmsg := \"User session groups are current\"\n\n\tif len(groupsAdded) != 0 || len(groupsRemoved) != 0 {\n\t\tif len(groupsAdded) != 0 {\n\t\t\tfields[\"added\"] = groupsAdded\n\t\t}\n\n\t\tif len(groupsRemoved) != 0 {\n\t\t\tfields[\"removed\"] = groupsRemoved\n\t\t}\n\n\t\tmsg = \"User session groups were updated\"\n\t}\n\n\tctx.Logger.WithFields(fields).Trace(msg)\n\n\tif len(emailsAdded) != 0 || len(emailsRemoved) != 0 {\n\t\tif len(emailsAdded) != 0 {\n\t\t\tfields[\"added\"] = emailsAdded\n\t\t} else {\n\t\t\tdelete(fields, \"added\")\n\t\t}\n\n\t\tif len(emailsRemoved) != 0 {\n\t\t\tfields[\"removed\"] = emailsRemoved\n\t\t} else {\n\t\t\tdelete(fields, \"removed\")\n\t\t}\n\n\t\tmsg = \"User session emails were updated\"\n\t} else {\n\t\tmsg = \"User session emails are current\"\n\n\t\tdelete(fields, \"added\")\n\t\tdelete(fields, \"removed\")\n\t}\n\n\tctx.Logger.WithFields(fields).Trace(msg)\n\n\tif nameDelta {\n\t\tctx.Logger.\n\t\t\tWithFields(map[string]any{\n\t\t\t\t\"username\": userSession.Username,\n\t\t\t\t\"before\": userSession.DisplayName,\n\t\t\t\t\"after\": details.DisplayName,\n\t\t\t}).\n\t\t\tTrace(\"User session display name updated\")\n\t} else {\n\t\tctx.Logger.Trace(\"User session display name is current\")\n\t}\n}", "func TestAddInfoSameKeyGreaterTimestamp(t *testing.T) {\n\tdefer leaktest.AfterTest(t)()\n\tstopper := stop.NewStopper()\n\tdefer stopper.Stop()\n\tis := newInfoStore(1, emptyAddr, stopper)\n\tinfo1 := is.newInfo(nil, time.Second)\n\tinfo2 := is.newInfo(nil, time.Second)\n\tif err1, err2 := is.addInfo(\"a\", info1), is.addInfo(\"a\", info2); err1 != nil || err2 != nil {\n\t\tt.Error(err1, err2)\n\t}\n}", "func isNewer(file1, file2 string) bool {\n\tif file1 == file2 {\n\t\treturn false\n\t}\n\tparts1 := strings.Split(strings.TrimSuffix(file1, \".zip\"), \"-\")\n\tparts2 := strings.Split(strings.TrimSuffix(file2, \".zip\"), \"-\")\n\tif len(parts1) != 3 || len(parts2) != 3 {\n\t\treturn false // Uncomparable.\n\t}\n\tif parts1[1] == parts2[1] {\n\t\t// If date parts are identical, then compare time parts.\n\t\ttime1, _ := strconv.Atoi(parts1[2])\n\t\ttime2, _ := strconv.Atoi(parts2[2])\n\t\treturn time1 > time2\n\t}\n\t// At this point, the date parts are not identical, so we compare them directly.\n\tdate1, _ := strconv.Atoi(parts1[1])\n\tdate2, _ := strconv.Atoi(parts2[1])\n\treturn date1 > date2\n}", "func (at *AnnotatedTable) checkModifiedAfter(ctx context.Context, other *AnnotatedTable) error {\n\t// If the source table is older than the destination table, then\n\t// don't overwrite it.\n\tthisMeta, err := at.CachedMeta(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Note that if other doesn't actually exist, its LastModifiedTime will be the time zero value,\n\t// so this will generally work as intended.\n\tif thisMeta.LastModifiedTime.Before(other.LastModifiedTime(ctx)) {\n\t\t// TODO should perhaps delete the source table?\n\t\treturn ErrSrcOlderThanDest\n\t}\n\treturn nil\n}", "func TestLogReplication2(t *testing.T) {\n\tack := make(chan bool)\n\n\t//Kill one server\n\traft.KillServer(1)\n\n\t//Append log to server\n\ttime.AfterFunc(1*time.Second, func() {\n\t\t//Get leader\n\t\tleaderId := raft.GetLeaderId()\n\n\t\t//Append a log entry to leader as client\n\t\traft.InsertFakeLogEntry(leaderId)\n\t})\n\n\t//Resurrect old server after enough time for other to move on\n\ttime.AfterFunc(2*time.Second, func() {\n\t\t//Resurrect old server\n\t\traft.ResurrectServer(1)\n\t})\n\n\t//Check log after some time to see if it matches with current leader\n\ttime.AfterFunc(3*time.Second, func() {\n\t\tleaderId := raft.GetLeaderId()\n\t\tleaderLog := raft.GetLogAsString(leaderId)\n\t\tserverLog := raft.GetLogAsString(1)\n\n\t\tcheckIfExpected(t, serverLog, leaderLog)\n\n\t\tack <- true\n\t})\n\n\t<-ack\n\n}", "func EqualFloat64(a, b, precision float64) bool {\n\treturn math.Abs(a-b) < precision\n}", "func (this *DeployLock) validateLatest(value int) bool {\n\tthis.mutex.Lock()\n\tdefer this.mutex.Unlock()\n\treturn this.numStarted == this.numFinished && this.numStarted == value\n}", "func (f ByAtime) Less(i, j int) bool {\n\tst1 := memo[i]\n\tif st1.Atim.Sec == 0 {\n\t\tif err := syscall.Stat(f[i].AbsPath, &st1); err != nil {\n\t\t\tlog.Fatal(\"Error on stat(\" + f[i].AbsPath + \"): \" + err.Error())\n\t\t} else {\n\t\t\tmemo[i] = st1\n\t\t}\n\t}\n\tst2 := memo[j]\n\tif st2.Atim.Sec == 0 {\n\t\tif err := syscall.Stat(f[j].AbsPath, &st2); err != nil {\n\t\t\tlog.Fatal(\"Error on stat(\" + f[j].AbsPath + \"): \" + err.Error())\n\t\t} else {\n\t\t\tmemo[j] = st2\n\t\t}\n\t}\n\treturn st1.Atim.Sec < st2.Atim.Sec\n}", "func itsConsistent(its *assert.Assertions, old, new []string, maxDelta int) {\n\tvar delta int\n\tfor _, oldItem := range old {\n\t\tif !findItem(oldItem, new) {\n\t\t\tdelta++\n\t\t}\n\t}\n\tits.True(delta < maxDelta, fmt.Sprintf(\"The item arrays should differ only by %d items (differed by %d items)\\n\\told: %s\\n\\tnew: %s\\n\", maxDelta, delta, strings.Join(old, \", \"), strings.Join(new, \", \")))\n}", "func isMateriallyGreater(a, b float64) bool {\n\treturn (a - b) > 0.001\n}", "func almostEqual(a, b float64) bool {\n\t// NaN has no equality but for testing we still want to know whether both values\n\t// are NaN.\n\tif math.IsNaN(a) && math.IsNaN(b) {\n\t\treturn true\n\t}\n\n\t// Cf. http://floating-point-gui.de/errors/comparison/\n\tif a == b {\n\t\treturn true\n\t}\n\n\tdiff := math.Abs(a - b)\n\n\tif a == 0 || b == 0 || diff < minNormal {\n\t\treturn diff < epsilon*minNormal\n\t}\n\treturn diff/(math.Abs(a)+math.Abs(b)) < epsilon\n}", "func (e *Entry) timeToUpdate() bool {\n\tnow := e.clk.Now()\n\te.mu.RLock()\n\tdefer e.mu.RUnlock()\n\tif e.response == nil {\n\t\t// not fetched anything previously\n\t\treturn true\n\t}\n\tif e.nextUpdate.Before(now) {\n\t\te.info(\"Stale response, updating immediately\")\n\t\treturn true\n\t}\n\tif e.maxAge > 0 {\n\t\t// cache max age has expired\n\t\tif e.lastSync.Add(e.maxAge).Before(now) {\n\t\t\te.info(\"max-age has expired, updating immediately\")\n\t\t\treturn true\n\t\t}\n\t}\n\n\t// update window is last quarter of NextUpdate - ThisUpdate\n\t// TODO: support using NextPublish instead of ThisUpdate if provided\n\t// in responses\n\twindowSize := e.nextUpdate.Sub(e.thisUpdate) / 4\n\tupdateWindowStarts := e.nextUpdate.Add(-windowSize)\n\tif updateWindowStarts.After(now) {\n\t\treturn false\n\t}\n\n\t// randomly pick time in update window\n\tupdateTime := updateWindowStarts.Add(time.Second * time.Duration(mrand.Intn(int(windowSize.Seconds()))))\n\tif updateTime.Before(now) {\n\t\te.info(\"Time to update\")\n\t\treturn true\n\t}\n\treturn false\n}", "func Same(t1, t2 *tree.Tree) bool {\n c1, c2 := make(chan int), make(chan int)\n go Walk(t1, c1)\n go Walk(t2, c2)\n\n for v1 := range c1 {\n v2 := <- c2\n if v1 != v2 {\n return false\n }\n }\n return true\n}", "func InstanceConfigUpToDate(instanceConfig, poolConfig *InstanceConfig) bool {\n\tif instanceConfig.UserData != poolConfig.UserData {\n\t\treturn false\n\t}\n\n\tif !util.Contains(strings.Split(poolConfig.ImageID, \",\"), instanceConfig.ImageID) {\n\t\treturn false\n\t}\n\n\tfor k, v := range poolConfig.Tags {\n\t\tif instanceValue, ok := instanceConfig.Tags[k]; !ok || v != instanceValue {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (t *tableCommon) canSkipUpdateBinlog(col *table.Column, value types.Datum) bool {\n\ttrace_util_0.Count(_tables_00000, 349)\n\tif col.IsGenerated() && !col.GeneratedStored {\n\t\ttrace_util_0.Count(_tables_00000, 351)\n\t\treturn true\n\t}\n\ttrace_util_0.Count(_tables_00000, 350)\n\treturn false\n}", "func Same(t1, t2 *tree.Tree) bool {\n\tvalues1 := make(chan int)\n\tvalues2 := make(chan int)\n\n\tgo Walk(t1, values1)\n\tgo Walk(t2, values2)\n\n\tfor val := range values1 {\n\t\tif val != <-values2 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (ms *MetricSet) haveFilesChanged() (bool, error) {\n\tvar stats syscall.Stat_t\n\tfor _, path := range ms.userFiles {\n\t\tif err := syscall.Stat(path, &stats); err != nil {\n\t\t\treturn true, errors.Wrapf(err, \"failed to stat %v\", path)\n\t\t}\n\n\t\tctime := time.Unix(int64(stats.Ctim.Sec), int64(stats.Ctim.Nsec))\n\t\tif ms.lastRead.Before(ctime) {\n\t\t\tms.log.Debugf(\"File changed: %v (lastRead=%v, ctime=%v)\", path, ms.lastRead, ctime)\n\n\t\t\treturn true, nil\n\t\t}\n\t}\n\n\treturn false, nil\n}", "func (c config) IsLastUpdateCheckTimeRecent(d time.Duration) bool {\n\tpath, err := c.updateCheckTouchPath()\n\tif err != nil {\n\t\tc.log.Errorf(\"Error getting check path: %s\", err)\n\t\treturn true\n\t}\n\tt, err := util.FileModTime(path)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\tc.log.Infof(\"No last update time\")\n\t\t} else {\n\t\t\tc.log.Errorf(\"Error getting last update time: %s\", err)\n\t\t}\n\t\treturn true\n\t}\n\trecent := time.Since(t) < d\n\tc.log.Debugf(\"Last update time (is recent? %s): %s\", strconv.FormatBool(recent), t)\n\treturn recent\n}", "func wasRunAfter(p1 PlanStatus, p2 PlanStatus) bool {\n\tif p1.Status == ExecutionNeverRun || p2.Status == ExecutionNeverRun || p1.LastUpdatedTimestamp == nil || p2.LastUpdatedTimestamp == nil {\n\t\treturn false\n\t}\n\treturn p1.LastUpdatedTimestamp.Time.After(p2.LastUpdatedTimestamp.Time)\n}", "func Same(t1, t2 *tree.Tree) bool {\n\tch1 := make(chan int)\n\tch2 := make(chan int)\n\n\tgo Walk(t1, ch1, true)\n\tgo Walk(t2, ch2, true)\n\n\t// all nodes from tree1 the same as in the tree2\n\tvar difFound bool\n\tfor v1 := range ch1 {\n\t\tv2, closed := <-ch2\n\t\tif !closed || v1 != v2 {\n\t\t\tdifFound = true\n\t\t}\n\t}\n\n\tresult := !difFound\n\n\t// if there more values left in tree2 after tree1 has been walked\n\tfor range ch2 {\n\t\tresult = false\n\t}\n\n\treturn result\n}", "func equalFloat64(a, b float64) bool {\n\t// Compare up to 6 decimal digits\n\teps := 1e-6\n\n\t// Check if difference between numbers falls within band of width 2 x epsilon\n\treturn (a-b) < eps && (b-a) < eps\n}", "func (rh *RumourHolder) CheckIfUpToDate(sp *StatusPacket) bool {\n\twant := sp.Want\n\tcounterMap := make(map[string]uint32)\n\tfor _, ps := range want {\n\t\tcounterMap[ps.Identifier] = ps.NextID\n\t}\n\tn1 := len(counterMap)\n\tn2 := len(rh.Rumours)\n\tif n1 != n2 {\n\t\treturn false\n\t}\n\n\tfor name, rfp := range rh.Rumours {\n\t\tnxt, ok := counterMap[name]\n\t\tif !ok {\n\t\t\treturn false\n\t\t}\n\t\tif nxt != rfp.Count+1 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func hasChanges(old *networking.Ingress, current *networking.Ingress) bool {\n\told.Status.LoadBalancer.Ingress = current.Status.LoadBalancer.Ingress\n\told.ResourceVersion = current.ResourceVersion\n\treturn !reflect.DeepEqual(old, current)\n}", "func IsUpToDate(d *v1alpha1.Device, p *packngo.Device) bool {\n\tif d.Spec.ForProvider.Hostname != p.Hostname {\n\t\treturn false\n\t}\n\tif d.Spec.ForProvider.Locked != p.Locked {\n\t\treturn false\n\t}\n\tif d.Spec.ForProvider.UserData != p.UserData {\n\t\treturn false\n\t}\n\tif d.Spec.ForProvider.IPXEScriptURL != p.IPXEScriptURL {\n\t\treturn false\n\t}\n\tif d.Spec.ForProvider.AlwaysPXE != p.AlwaysPXE {\n\t\treturn false\n\t}\n\tif !reflect.DeepEqual(d.Spec.ForProvider.Tags, p.Tags) {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func isOldEnough(pr *github.PullRequest) bool {\n\treturn time.Now().Unix()-pr.CreatedAt.Unix() > config.TimeForCreatingJIRATicket\n}", "func diff(chain1, chain2 []*x509.Certificate) bool {\n\t// Check if bundled one is different from the input.\n\tdiff := false\n\tif len(chain1) != len(chain2) {\n\t\tdiff = true\n\t} else {\n\t\tfor i := 0; i < len(chain1); i++ {\n\t\t\tcert1 := chain1[i]\n\t\t\tcert2 := chain2[i]\n\t\t\t// Use signature to differentiate.\n\t\t\tif !bytes.Equal(cert1.Signature, cert2.Signature) {\n\t\t\t\tdiff = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\treturn diff\n}", "func TestRevisionIsNewerWithTimeStamp(t *testing.T) {\n\toriginalAllowedNTPDiff := allowedNTPDiffInMilliSecond\n\ttestRevisionIsNewer(t, 0)\n\ttestRevisionIsNewer(t, 500)\n\ttestRevisionIsNewer(t, 1000)\n\ttestRevisionIsNewer(t, 10000)\n\tallowedNTPDiffInMilliSecond = originalAllowedNTPDiff\n}", "func (r *Raft) shouldSnapshot() bool {\n\t// Check the last snapshot index\n\tlastSnap, _ := r.getLastSnapshot()\n\n\t// Check the last log index\n\tlastIdx, err := r.logs.LastIndex()\n\tif err != nil {\n\t\tr.logger.Error(\"failed to get last log index\", \"error\", err)\n\t\treturn false\n\t}\n\n\t// Compare the delta to the threshold\n\tdelta := lastIdx - lastSnap\n\treturn delta >= r.config().SnapshotThreshold\n}", "func (cfg *config) checkConsistency(server int, index int, logEntry logEntry) {\n\tfor j := 0; j < len(cfg.logs); j++ {\n\t\te, ok := cfg.logs[j][index]\n\t\tif ok && (e != logEntry) {\n\t\t\tlog.Fatalf(\"server %d is trying to commit logEntry: %+v at index: %d \"+\n\t\t\t\t\"but server j: %d has already committed e: %+v for the same index\",\n\t\t\t\tserver, logEntry, index, j, e)\n\t\t}\n\t}\n}", "func isSnapTsConsistentOrAhead(snapTs, reqTs *common.TsVbuuid, strict_chk_threshold int) (bool, bool) {\n\n\tisSnapAhead := false\n\n\tlag := uint64(0) // Number by which KV Timestamp is behind snapshot Timestamp\n\tif snapTs == nil || reqTs == nil {\n\t\treturn false, isSnapAhead\n\t}\n\tif snapTs.Bucket != reqTs.Bucket {\n\t\treturn false, isSnapAhead\n\t}\n\tif len(snapTs.Seqnos) > len(reqTs.Seqnos) {\n\t\treturn false, isSnapAhead\n\t}\n\tfor i, seqno := range snapTs.Seqnos {\n\t\tif seqno < reqTs.Seqnos[i] {\n\t\t\treturn false, isSnapAhead\n\t\t}\n\n\t\tif seqno > reqTs.Seqnos[i] {\n\t\t\tlag += (seqno - reqTs.Seqnos[i])\n\t\t}\n\t}\n\n\tif lag > uint64(strict_chk_threshold) {\n\t\t// snapTs is significantly ahead of reqTS, this indicates that\n\t\t// there is a possible KV data loss. Switch to slow path that checks\n\t\t// for vbuuids along with vbseqnos so that stale results will not be returned\n\t\tisSnapAhead = true\n\t}\n\n\treturn true, isSnapAhead\n}", "func CompareAndSwapUint64(addr *uint64, old, new uint64) (swapped bool)", "func Stale(lastVersionCheck string, dur string) bool {\n\tttl, err := time.ParseDuration(dur)\n\tif err != nil {\n\t\t// If there is no duration provided, then we should presume the loading of\n\t\t// remote configuration failed and that we should retry that operation.\n\t\treturn true\n\t}\n\n\tlastChecked, _ := time.Parse(time.RFC3339, lastVersionCheck)\n\treturn lastChecked.Add(ttl).Before(time.Now())\n}", "func checkMasterLogsMatchBranch(master *LogList, branch *LogList, wl *warningList) {\n\tfor _, log := range branch.Logs {\n\t\tif masterEntry := master.FindLogByKey(log.Key); masterEntry != nil {\n\t\t\tmasterEntry.checkEquivalence(&log, wl)\n\t\t}\n\t}\n}", "func equal(a, b float64) bool {\n\tif math.IsNaN(a) && math.IsNaN(b) {\n\t\treturn true\n\t}\n\tif !math.IsNaN(a) && !math.IsNaN(b) {\n\t\treturn math.Abs(a-b) < eps\n\t}\n\treturn false\n}", "func CheckEquivalent(t1, t2 *tree.Tree) bool {\n\tc1 := make(chan int)\n\tc2:= make(chan int)\n\tgo Walk(t1, c1)\n\tgo Walk(t2, c2)\n\n\tx1, x2, ok1, ok2 := 0, 0, true, true\n\n\tfor ok1 && ok2 {\n\t\tx1, ok1 = <- c1\n\t\tx2, ok2 = <- c2\n\n\t\tif ok1 != ok2 || x1 != x2 {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func Equal(t1, t2 Token) bool {\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\n\tif t1 == nil && t2 == nil {\n\t\treturn true\n\t}\n\n\t// we already checked for t1 == t2 == nil, so safe to do this\n\tif t1 == nil || t2 == nil {\n\t\treturn false\n\t}\n\n\tm1, err := t1.AsMap(ctx)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\tfor iter := t2.Iterate(ctx); iter.Next(ctx); {\n\t\tpair := iter.Pair()\n\n\t\tv1 := m1[pair.Key.(string)]\n\t\tv2 := pair.Value\n\t\tswitch tmp := v1.(type) {\n\t\tcase time.Time:\n\t\t\ttmp2, ok := v2.(time.Time)\n\t\t\tif !ok {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\ttmp = tmp.Round(0).Truncate(time.Second)\n\t\t\ttmp2 = tmp2.Round(0).Truncate(time.Second)\n\t\t\tif !tmp.Equal(tmp2) {\n\t\t\t\treturn false\n\t\t\t}\n\t\tdefault:\n\t\t\tif v1 != v2 {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\tdelete(m1, pair.Key.(string))\n\t}\n\n\treturn len(m1) == 0\n}", "func (l *FileLog) mightNeedRotation() bool {\n\tif l.file == nil {\n\t\treturn true\n\t}\n\n\t// determine the timestamp for the current log file rounded to the day.\n\tfileTime := l.Clock.Now().UTC().Truncate(24 * time.Hour)\n\n\treturn l.fileTime.Before(fileTime)\n}", "func (n *Uint256) Lt(n2 *Uint256) bool {\n\tvar borrow uint64\n\t_, borrow = bits.Sub64(n.n[0], n2.n[0], borrow)\n\t_, borrow = bits.Sub64(n.n[1], n2.n[1], borrow)\n\t_, borrow = bits.Sub64(n.n[2], n2.n[2], borrow)\n\t_, borrow = bits.Sub64(n.n[3], n2.n[3], borrow)\n\treturn borrow != 0\n}", "func isSnapshotConsistentOrAhead(ss IndexSnapshot, reqTs *common.TsVbuuid,\n\tcons common.Consistency, strict_chk_threshold int) (bool, bool) {\n\n\tsnapTsConsistent, snapTsAhead := false, false\n\n\tif snapTs := ss.Timestamp(); snapTs != nil {\n\t\tif ss.IsEpoch() && reqTs.IsEpoch() {\n\t\t\treturn true, false\n\t\t}\n\n\t\tsnapTsConsistent, snapTsAhead = isSnapTsConsistentOrAhead(snapTs, reqTs, strict_chk_threshold)\n\t\tif snapTs.CheckCrc64(reqTs) && snapTsConsistent {\n\t\t\treturn true, snapTsAhead\n\t\t}\n\n\t\t// don't return error because client might be ahead of\n\t\t// in receiving a rollback.\n\t\t// return nil, ErrVbuuidMismatch\n\t\treturn false, false\n\t}\n\treturn false, false\n}", "func (dr *DeviceRoutes) IsUpToDate(ld *voltha.LogicalDevice) bool {\n\tdr.routeBuildLock.Lock()\n\tdefer dr.routeBuildLock.Unlock()\n\tnumNNI, numUNI := 0, 0\n\tif ld != nil {\n\t\tif len(dr.logicalPorts) != len(ld.Ports) {\n\t\t\treturn false\n\t\t}\n\t\tnumNNI = len(dr.RootPorts)\n\t\tnumUNI = len(ld.Ports) - numNNI\n\t}\n\treturn len(dr.Routes) == numNNI*numUNI*2\n}", "func IsPowOfTwoUseLog(number float64) bool {\n\tif number == 0 || math.Round(number) == math.MaxInt64 {\n\t\treturn false\n\t}\n\tlog := math.Log2(number)\n\treturn log == math.Round(log)\n}", "func isMoreRecentOwner(old *PhoneNumberInfo, new *PhoneNumberInfo) bool {\n\tnewActivationDate := *(new.getActivationDate())\n\toldActivationDate := *(old.getActivationDate())\n\tif newActivationDate.After(oldActivationDate) && old.DeactivationDate != new.ActivationDate {\n\t\treturn true\n\t}\n\treturn false\n}", "func (rf *Raft) buildAppendEntriesReplyWhenNotSuccess(reply *AppendEntriesReply, PrevLogIndex int, PrevLogTerm int) {\n\tif PrevLogIndex > rf.getLastIndex() {\n\t\t// this raft do not know about the PrevLogIndex\n\t\treply.SuggestPrevLogIndex = rf.getLastIndex()\n\t\treply.SuggestPrevLogTerm = rf.getLastTerm()\n\t} else {\n\t\t// there is conflict!\n\t\tConflictTerm := rf.getTermForIndex(PrevLogIndex)\n\t\tAssertF(ConflictTerm != PrevLogTerm, \"\")\n\t\tAssertF(PrevLogIndex > rf.commitIndex, \"\")\n\n\t\t// TODO: change to (ConflictTerm, FirstIndex)\n\t\tif ConflictTerm > PrevLogTerm {\n\t\t\t// T1 -- PrevLogTerm, T2 -- ConflictTerm, T1<T2\n\t\t\t// any (i1,t1) in leaders log, if i1<=PrevLogIndex, then t1<=PrevLogTerm\n\t\t\t// Then we find SuggestPrevLogIndex, in tuple (SuggestPrevLogIndex, t2),\n\t\t\t// that satisfies t2<=T1, and SuggestPrevLogIndex is the large one\n\t\t\t// suggestTerm = the max index ( <= PrevLogTerm )\n\t\t\treply.SuggestPrevLogIndex = PrevLogIndex\n\t\t\tfor ; reply.SuggestPrevLogIndex > rf.commitIndex && rf.getTermForIndex(reply.SuggestPrevLogIndex) > PrevLogTerm; reply.SuggestPrevLogIndex-- {\n\t\t\t}\n\t\t\treply.SuggestPrevLogTerm = rf.getTermForIndex(reply.SuggestPrevLogIndex) // term 0 if index 0\n\t\t} else {\n\t\t\treply.SuggestPrevLogIndex = PrevLogIndex - 1\n\t\t\treply.SuggestPrevLogTerm = rf.getTermForIndex(reply.SuggestPrevLogIndex) // term 0 if index 0\n\t\t}\n\n\t\tAssertF(reply.SuggestPrevLogIndex >= rf.commitIndex,\n\t\t\t\"reply.SuggestPrevLogIndex {%d} >= rf.commitIndex {%d}\",\n\t\t\treply.SuggestPrevLogIndex, rf.commitIndex)\n\t}\n\tAssertF(reply.SuggestPrevLogIndex < PrevLogIndex,\n\t\t\"reply.SuggestPrevLogIndex {%d} < PrevLogIndex {%d}\",\n\t\treply.SuggestPrevLogIndex, PrevLogIndex)\n}", "func (n *kvNamespace) logChange(key string, val *kvNsValue,\n\toldver uint64, op uint) {\n\n\t// check if oldver or key is in the log, if so remove that change\n\tsz := len(n.ChangeLog)\n\toldi := sz\n\tif oldver != 0 {\n\t\toldi = sort.Search(sz,\n\t\t\tfunc(i int) bool {\n\t\t\t\treturn n.ChangeLog[i].Value.Version >= oldver\n\t\t\t})\n\t\tif oldi < sz && n.ChangeLog[oldi].Value.Version != oldver {\n\t\t\t// not a match\n\t\t\toldi = sz\n\t\t}\n\t} else {\n\t\t// check all keys in changelog since we don't know old version\n\t\tfor i := sz - 1; i >= 0; i-- {\n\t\t\tif n.ChangeLog[i].Key == key {\n\t\t\t\toldi = i\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\tif oldi < sz {\n\t\tif oldi < sz-1 { // move elements to left one position\n\t\t\tcopy(n.ChangeLog[oldi:], n.ChangeLog[oldi+1:])\n\t\t}\n\t\t// truncate last element\n\t\tn.ChangeLog = n.ChangeLog[0 : sz-1]\n\t}\n\n\t// purge older half of change log if it has become too long\n\tif len(n.ChangeLog) >= 2*n.maxRetained {\n\t\tn.lastExpiredChange = n.ChangeLog[n.maxRetained-1].Value.Version\n\t\tn.ChangeLog = n.ChangeLog[n.maxRetained:]\n\t}\n\tn.ChangeLog = append(n.ChangeLog,\n\t\t&kvChange{Key: key, Value: val, Action: op})\n\t/*\n\t\tfmt.Printf(\"LastVersion: %v, ChangeLog: \", n.LastVersion)\n\t\tfor i := 0; i < len(n.ChangeLog); i++ {\n\t\t\tfmt.Printf(\"%v:%v(%d) \", n.ChangeLog[i].Key,\n\t\t\t\tn.ChangeLog[i].Value.Version, n.ChangeLog[i].Action)\n\t\t}\n\t\tfmt.Printf(\"\\n\")\n\t*/\n}", "func sameConfigDB(loggerOne, loggerTwo backend.JSONConfigurationDB) bool {\n\treturn (loggerOne.Host == loggerTwo.Host) && (loggerOne.Port == loggerTwo.Port) && (loggerOne.Name == loggerTwo.Name)\n}", "func (s ClusterLogTarget) Equal(t ClusterLogTarget, opts ...Options) bool {\n\tif !equalPointers(s.Address, t.Address) {\n\t\treturn false\n\t}\n\n\tif s.LogFormat != t.LogFormat {\n\t\treturn false\n\t}\n\n\tif !equalPointers(s.Port, t.Port) {\n\t\treturn false\n\t}\n\n\tif !equalPointers(s.Protocol, t.Protocol) {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (l *PersistableLedger) expired(entryTs int64) bool {\n\texpirationTime := time.Now().Unix() - l.expiration\n\treturn entryTs < expirationTime\n}", "func (fm FinalModelTimestamp) Verify() (bool, int) {\n if (fm.buffer.Offset() + fm.FBEOffset() + fm.FBESize()) > fm.buffer.Size() {\n return false, 0\n }\n\n return true, fm.FBESize()\n}", "func le(a, b swarm.Address) bool {\n\treturn bytes.Compare(a.Bytes(), b.Bytes()) == -1\n}", "func certsChanged(origCertChecksum []byte, origKeyChecksum []byte, origAuthCAChecksum []byte) bool {\n\t// Check if all files exist.\n\tcertNotEmpty, err := fileExistsAndNotEmpty(tlsCert)\n\tif err != nil {\n\t\tklog.Warningf(\"error checking if changed TLS cert file empty/exists: %v\", err)\n\t\treturn false\n\t}\n\tkeyNotEmpty, err := fileExistsAndNotEmpty(tlsKey)\n\tif err != nil {\n\t\tklog.Warningf(\"error checking if changed TLS key file empty/exists: %v\", err)\n\t\treturn false\n\t}\n\tcaNotEmpty, err := fileExistsAndNotEmpty(authCAFile)\n\tif err != nil {\n\t\tklog.Warningf(\"error checking if changed auth CA file empty/exists: %v\", err)\n\t\treturn false\n\t}\n\n\tif !certNotEmpty || !keyNotEmpty || !caNotEmpty {\n\t\t// One of the files is missing despite some file event.\n\t\tklog.V(1).Infof(\"certificate, key or auth CA is missing or empty, certificates will not be rotated\")\n\t\treturn false\n\t}\n\tcurrentCertChecksum := checksumFile(tlsCert)\n\tcurrentKeyChecksum := checksumFile(tlsKey)\n\tcurrentAuthCAChecksum := checksumFile(authCAFile)\n\n\tklog.V(2).Infof(\"certificate checksums before: %x, %x, %x. checksums after: %x, %x, %x\",\n\t\torigCertChecksum, origKeyChecksum, origAuthCAChecksum, currentCertChecksum, currentKeyChecksum, currentAuthCAChecksum)\n\t// Check if the non-empty certificate/key or auth CA files have actually changed.\n\tif !bytes.Equal(origCertChecksum, currentCertChecksum) && !bytes.Equal(origKeyChecksum, currentKeyChecksum) ||\n\t\t!bytes.Equal(origAuthCAChecksum, currentAuthCAChecksum) {\n\t\tklog.Infof(\"cert and key or auth CA changed, need to restart the metrics server\")\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (tk *timekeeper) checkStreamReadyToMerge(cmd Message) bool {\n\n\tstreamId := cmd.(*MsgMutMgrFlushDone).GetStreamId()\n\tbucket := cmd.(*MsgMutMgrFlushDone).GetBucket()\n\tflushTs := cmd.(*MsgMutMgrFlushDone).GetTS()\n\n\tif streamId == common.INIT_STREAM {\n\n\t\tfor _, buildInfo := range tk.indexBuildInfo {\n\t\t\t//if index belongs to the flushed bucket and in CATCHUP state\n\t\t\tidx := buildInfo.indexInst\n\t\t\tif idx.Defn.Bucket == bucket &&\n\t\t\t\tidx.State == common.INDEX_STATE_CATCHUP {\n\n\t\t\t\t//if the flushTs is past the lastFlushTs of this bucket in MAINT_STREAM,\n\t\t\t\t//this index can be merged to MAINT_STREAM\n\t\t\t\tbucketLastTsFlushedMap := tk.streamBucketLastTsFlushedMap[common.MAINT_STREAM]\n\t\t\t\tlastFlushedTs := (*bucketLastTsFlushedMap)[idx.Defn.Bucket]\n\n\t\t\t\tif flushTs.GreaterThanEqual(lastFlushedTs) {\n\t\t\t\t\t//disable flush for MAINT_STREAM for this bucket, so it doesn't\n\t\t\t\t\t//move ahead till merge is complete\n\t\t\t\t\tbucketFlushEnabledMap := tk.streamBucketFlushEnabledMap[common.MAINT_STREAM]\n\t\t\t\t\t(*bucketFlushEnabledMap)[idx.Defn.Bucket] = false\n\n\t\t\t\t\t//change state of all indexes of this bucket to ACTIVE\n\t\t\t\t\t//these indexes get removed later as part of merge message\n\t\t\t\t\t//from indexer\n\t\t\t\t\ttk.changeIndexStateForBucket(bucket, common.INDEX_STATE_ACTIVE)\n\n\t\t\t\t\tcommon.Debugf(\"Timekeeper::checkStreamReadyToMerge \\n\\tIndex Ready To Merge. \"+\n\t\t\t\t\t\t\"Index: %v Stream: %v Bucket: %v LastFlushTS: %v\", idx.InstId, streamId,\n\t\t\t\t\t\tbucket, lastFlushedTs)\n\n\t\t\t\t\ttk.supvRespch <- &MsgTKMergeStream{\n\t\t\t\t\t\tstreamId: streamId,\n\t\t\t\t\t\tbucket: bucket,\n\t\t\t\t\t\tmergeTs: flushTs}\n\n\t\t\t\t\ttk.supvCmdch <- &MsgSuccess{}\n\t\t\t\t\treturn true\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func TestLeaderElectionOverwriteNewerLogs(t *testing.T) {\n\ttestLeaderElectionOverwriteNewerLogs(t, false)\n}", "func (s *Seed) IsUpToDate(masterVersions kubermatic.Versions) bool {\n\treturn s.Status.Versions.Kubermatic == masterVersions.KubermaticCommit\n}", "func Same(t1, t2 *tree.Tree) bool {\n\tch1 := make(chan int)\n\tch2 := make(chan int)\n\n\tgo Walk(t1, ch1)\n\tgo Walk(t2, ch2)\n\n\tfor i := 0; i < 10; i++ {\n\t\tif <-ch1 != <-ch2 {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func Same(t1, t2 *tree.Tree) bool {\n\tch1 := make(chan int, 10)\n\tch2 := make(chan int, 10)\n\tgo Walk(t1, ch1)\n\tgo Walk(t2, ch2)\n\n\tv1, v2 := 0, 0\n\tok1, ok2 := true, true\n\n\tfor ok1 && ok2 {\n\t\tv1, ok1 = <-ch1\n\t\tv2, ok2 = <-ch2\n\t\tif ok1 != ok2 || v1 != v2 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func Same(t1, t2 *tree.Tree) bool{\n\tchannel1 := make(chan int)\n\tchannel2 := make(chan int)\n\tgo func(){\n\t\tWalk(t1,channel1)\n\t\tchannel1<-0\n\t}()\n\tgo func(){\n\t\tWalk(t2,channel2)\n\t\tchannel2<-0\n\t}()\n\tfor {\n t1 := <-channel1\n t2 := <-channel2\n if t1 == 0 && t2 == 0 {\n return true;\n }\n \n if t1 == t2 {\n continue;\n } else {\n return false;\n }\n \t}\n\t return true\n\t\n}", "func (src Version) IsGTOrE2(des Version) bool {\n\trelation := src.compareWith(des)\n\treturn relation == Greater || relation == Equal\n}", "func Same(t1, t2 *tree.Tree) bool {\n\tch1 := make(chan int)\n\tgo DoWalk(t1, ch1)\n\tch2 := make(chan int)\n\tgo DoWalk(t2, ch2)\n\n\tfor {\n\t\tv1, ok1 := <-ch1\n\t\tv2, ok2 := <-ch2\n\n\t\tif v1 != v2 || ok1 && !ok2 || !ok1 && ok2 {\n\t\t\treturn false\n\t\t}\n\n\t\tif !ok1 && !ok2 {\n\t\t\tbreak\n\t\t}\n\t}\n\treturn true\n}" ]
[ "0.6547263", "0.64683026", "0.642574", "0.6350501", "0.6076717", "0.5779664", "0.553337", "0.5500449", "0.5485863", "0.54752135", "0.54620296", "0.5398634", "0.53852415", "0.5334269", "0.53047884", "0.529602", "0.52464086", "0.52167886", "0.51762223", "0.51707363", "0.5127219", "0.5067506", "0.49789232", "0.49359047", "0.49295974", "0.4924391", "0.49020535", "0.48985153", "0.48910105", "0.48887247", "0.48818752", "0.4868519", "0.4847741", "0.4846965", "0.48445782", "0.48348147", "0.48261902", "0.48240498", "0.4814769", "0.48146445", "0.48132962", "0.4811258", "0.47953564", "0.4794355", "0.47889453", "0.47849512", "0.4781886", "0.47663775", "0.47657824", "0.47645304", "0.47612938", "0.47557646", "0.47459757", "0.47446218", "0.47338286", "0.47318345", "0.4729378", "0.4722264", "0.47222093", "0.47149223", "0.4710188", "0.47089145", "0.4701672", "0.4698153", "0.46976608", "0.46888247", "0.4687673", "0.4686655", "0.4678968", "0.46771866", "0.46619526", "0.46592262", "0.4657384", "0.4656439", "0.46518654", "0.4644428", "0.46443325", "0.46443", "0.46269634", "0.46242157", "0.4622452", "0.4621372", "0.46202266", "0.46199337", "0.46141475", "0.46090123", "0.46057972", "0.4604681", "0.46036232", "0.45995703", "0.45942175", "0.45890436", "0.4588688", "0.45863432", "0.45798293", "0.45784038", "0.4575565", "0.45710185", "0.4569782", "0.45684728" ]
0.7410175
0
initialization some variables when rf become a leader
func (rf *Raft) leaderInitialization() { rf.nextIndex = make([]int, len(rf.peers)) rf.matchIndex = make([]int, len(rf.peers)) for server := range rf.nextIndex { rf.nextIndex[server] = rf.getLastLogIndex() + 1 } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (le *LeaderElector) initElection() {\n\thighestRank := false\n\t//Poll servers with higher rank\n\tfor SID, serv := range le.ThisServer.GroupInfoPtr.GroupMembers {\n\t\tif SID < le.ThisServer.SID {\n\t\t\t//Has Higher rank, SID 0 > SID 1 > SID 2 ....\n\t\t\tok := call(serv, \"LeaderElector.ChangeLeader\", new(interface{}), &highestRank)\n\t\t\tif ok && highestRank == true {\n\t\t\t\t//Theres a server with higher rank, let go\n\t\t\t\tdebug(\"[*] Info : LeaderElector : There is Another Server - %s- With Higher Rank.Backing off. \", serv)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\t//No server with higher rank, become leader\n\tle.becomeLeader()\n}", "func (r *Raft) becomeLeader() {\n\t// Your Code Here (2A).\n\t// NOTE: Leader should propose a noop entry on its term\n\tif _, ok := r.Prs[r.id]; !ok {\n\t\treturn\n\t}\n\tlog.DInfo(\"r %d becomes the leader in term %d\", r.id, r.Term)\n\tr.State = StateLeader\n\tr.Lead = r.id\n\tr.Vote = r.id\n\tr.heartbeatElapsed = 0\n\tr.electionElapsed = 0\n\tr.actualElectionTimeout = rand.Intn(r.electionTimeout) + r.electionTimeout\n\t// 成为 leader 以后要重新设置日志同步信息,注意自己的日志同步信息应该一直是最新的,否则会影响 commit 计算\n\tfor k := range r.Prs {\n\t\tif k == r.id { // 不可以超出 peers 的范围\n\t\t\tr.Prs[r.id] = &Progress{\n\t\t\t\tMatch: r.RaftLog.LastIndex(),\n\t\t\t\tNext: r.RaftLog.LastIndex() + 1,\n\t\t\t}\n\t\t} else {\n\t\t\tr.Prs[k] = &Progress{\n\t\t\t\tMatch: 0,\n\t\t\t\tNext: r.RaftLog.LastIndex() + 1,\n\t\t\t}\n\t\t}\n\t}\n\t// raft 要求 leader 不能提交之前任期的日志条目,或者说,提交的日志条目必须包含自己的任期\n\t// 为了在本任期没有收到同步请求的情况下也要能提交之前的日志,应当在成为 leader 的时候立刻 propose 一个空条目并 append 下去\n\t_ = r.Step(pb.Message{\n\t\tMsgType: pb.MessageType_MsgPropose,\n\t\tTo: r.id,\n\t\tFrom: r.id,\n\t\tTerm: r.Term,\n\t\tEntries: []*pb.Entry{{\n\t\t\tEntryType: pb.EntryType_EntryNormal,\n\t\t\tTerm: 0,\n\t\t\tIndex: 0,\n\t\t\tData: nil,\n\t\t}},\n\t})\n}", "func (s *raftServer) initLeader(followers []int) (*utils.SyncIntIntMap, *utils.SyncIntIntMap, *utils.SyncIntIntMap) {\n\tnextIndex := utils.CreateSyncIntMap()\n\tmatchIndex := utils.CreateSyncIntMap()\n\taeToken := utils.CreateSyncIntMap() // acts like mutex in producer-consumer\n\tnextLogEntry := s.localLog.TailIndex() + 1\n\tfor _, f := range followers {\n\t\tnextIndex.Set(f, nextLogEntry)\n\t\tmatchIndex.Set(f, 0)\n\t\taeToken.Set(f, 1)\n\t}\n\treturn nextIndex, matchIndex, aeToken\n}", "func (rf *Raft) initialServer() {\n\tfor {\n\t\tif !rf.killed() {\n\t\t\ttimeOut := time.Millisecond * time.Duration(ElectionTimeOut+rand.Intn(ElectionTimeOut)) //warn 必须sleep timeOut,避免server在一个ElectionTimeOut结束后连续发起选举\n\t\t\ttime.Sleep(timeOut)\n\t\t\t//DPrintf(\"rf%v state%v\",rf.me,rf.state)\n\t\t\trf.mu.Lock()\n\t\t\tswitch rf.state {\n\t\t\tcase Candidate:\n\t\t\t\tDPrintf(\"candidate [%v] startElection detail:%v\", rf.me, rf)\n\t\t\t\tgo rf.startElection()\n\t\t\tcase Follower:\n\t\t\t\t//if time.Now().Sub(rf.lastReceiveHeartBeat) > (timeOut - time.Millisecond*time.Duration(10)) {\n\t\t\t\tif time.Now().Sub(rf.lastReceiveHeartBeat) > timeOut {\n\t\t\t\t\trf.state = Candidate\n\t\t\t\t\t//DPrintf(\"follower %v startElection\",rf)\n\t\t\t\t\tgo rf.startElection() // warn follower转为candidate后应立即startElection,不再等待新的ElectionTimeOUt\n\t\t\t\t}\n\t\t\t}\n\t\t\trf.mu.Unlock()\n\t\t} else {\n\t\t\treturn\n\t\t}\n\t}\n}", "func (rf *Raft) initRaftNodeToFollower(logCapacity int) {\n rf.mu.Lock()\n defer rf.mu.Unlock()\n\n rf.state = \"Follower\"\n\n rf.currentTerm = 0\n rf.votedFor = -1\n rf.log = make([]Entry, 1, logCapacity)\n rf.log[0].Term = 0\n\n rf.commitIndex = 0\n rf.lastApplied = 0\n\n rf.electionTime = generateElectionTime()\n rf.electionTimer = time.NewTimer(time.Duration(rf.electionTime) * time.Millisecond)\n\n rf.nextIndex = make([]int, len(rf.peers))\n rf.matchIndex = make([]int, len(rf.peers))\n for i:=0; i<len(rf.peers); i++ {\n rf.nextIndex[i] = len(rf.log)\n rf.matchIndex[i] = 0\n }\n\n rf.snapshottedIndex = 0\n}", "func newRaft(c *Config) *Raft {\n\tif err := c.validate(); err != nil {\n\t\tpanic(err.Error())\n\t}\n\t// Your Code Here (2A).\n\tl := newLog(c.Storage)\n\t// 恢复 hardState\n\thardState, confState, _ := c.Storage.InitialState()\n\tif hardState.Commit > 0 {\n\t\tl.committed = hardState.Commit\n\t}\n\tvar peers []uint64\n\tif len(confState.Nodes) > 0 {\n\t\tpeers = confState.Nodes\n\t} else {\n\t\tpeers = c.peers\n\t}\n\tr := &Raft{\n\t\tid: c.ID,\n\t\tTerm: hardState.Term,\n\t\tVote: hardState.Vote,\n\t\tRaftLog: l,\n\t\tPrs: make(map[uint64]*Progress),\n\t\tState: StateFollower,\n\t\tvotes: make(map[uint64]bool),\n\t\tmsgs: nil,\n\t\tLead: None,\n\t\theartbeatTimeout: c.HeartbeatTick,\n\t\telectionTimeout: c.ElectionTick,\n\t\theartbeatElapsed: 0,\n\t\telectionElapsed: 0,\n\t\tleadTransferee: 0,\n\t\tPendingConfIndex: 0,\n\t}\n\tif c.Applied > 0 {\n\t\tr.RaftLog.applied = c.Applied\n\t}\n\tfor _, v := range peers {\n\t\tif v == r.id { // 不可以超出 peers 的范围\n\t\t\tr.Prs[r.id] = &Progress{\n\t\t\t\tMatch: r.RaftLog.LastIndex(),\n\t\t\t\tNext: r.RaftLog.LastIndex() + 1,\n\t\t\t}\n\t\t} else {\n\t\t\tr.Prs[v] = &Progress{\n\t\t\t\tMatch: 0,\n\t\t\t\tNext: r.RaftLog.LastIndex() + 1,\n\t\t\t}\n\t\t}\n\t\tr.votes[v] = false\n\t}\n\tr.votes[r.Vote] = true\n\t// 此处不可以使用 becomeFollower,否则会丢失 Vote 变量\n\tr.actualElectionTimeout = rand.Intn(r.electionTimeout) + r.electionTimeout\n\tlog.DInfo(\"'Raft Created' %d=%+v\", r.id, r)\n\treturn r\n}", "func (rf *Raft) convertToLeader() {\n rf.mu.Lock()\n DLCPrintf(\"Server (%d)[state=%s, term=%d, votedFor=%d] convert to Leader\", rf.me, rf.state, rf.currentTerm, rf.votedFor)\n rf.electionTimer.Stop() \n rf.state = \"Leader\"\n for i:=0; i<len(rf.peers); i++ {\n rf.nextIndex[i] = rf.convertToGlobalViewIndex(len(rf.log))\n rf.matchIndex[i] = rf.convertToGlobalViewIndex(0)\n }\n rf.mu.Unlock()\n // 启动一个线程,定时给各个Follower发送HeartBeat Request \n time.Sleep(50 * time.Millisecond)\n go rf.sendAppendEntriesToMultipleFollowers()\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n persister *Persister, applyCh chan ApplyMsg) *Raft {\n rf := &Raft{}\n rf.peers = peers\n rf.persister = persister\n rf.me = me\n rf.applyCh = applyCh\n\n // Your initialization code here (2A, 2B, 2C).\n rf.dead = 0\n\n rf.currentTerm = 0\n rf.votedFor = -1\n rf.commitIndex = -1\n rf.lastApplied = -1\n rf.state = Follower\n rf.gotHeartbeat = false\n\n // initialize from state persisted before a crash\n rf.readPersist(persister.ReadRaftState())\n\n // Start Peer State Machine\n go func() {\n // Run forver\n for {\n \n if rf.killed() {\n fmt.Printf(\"*** Peer %d term %d: I have been terminated. Bye.\",rf.me, rf.currentTerm)\n return \n }\n \n rf.mu.Lock()\n state := rf.state\n rf.mu.Unlock()\n \n switch state {\n case Follower:\n fmt.Printf(\"-- peer %d term %d, status update: I am follolwer.\\n\",rf.me, rf.currentTerm)\n snoozeTime := rand.Float64()*(RANDOM_TIMER_MAX-RANDOM_TIMER_MIN) + RANDOM_TIMER_MIN\n fmt.Printf(\" peer %d term %d -- follower -- : Set election timer to time %f\\n\", rf.me, rf.currentTerm, snoozeTime)\n time.Sleep(time.Duration(snoozeTime) * time.Millisecond) \n \n rf.mu.Lock() \n fmt.Printf(\" peer %d term %d -- follower -- : my election timer had elapsed.\\n\",rf.me, rf.currentTerm)\n if (!rf.gotHeartbeat) {\n fmt.Printf(\"-> Peer %d term %d -- follower --: did not get heartbeat during the election timer. Starting election!\\n\",rf.me, rf.currentTerm) \n rf.state = Candidate\n }\n rf.gotHeartbeat = false\n rf.mu.Unlock()\n \n\n case Candidate:\n rf.mu.Lock()\n rf.currentTerm++\n fmt.Printf(\"-- peer %d: I am candidate! Starting election term %d\\n\",rf.me, rf.currentTerm)\n numPeers := len(rf.peers) // TODO: figure out what to with mutex when reading. Atomic? Lock?\n rf.votedFor = rf.me\n rf.mu.Unlock()\n \n voteCount := 1\n var replies = make([]RequestVoteReply, numPeers)\n rf.sendVoteRequests(replies, numPeers)\n\n snoozeTime := rand.Float64()*(RANDOM_TIMER_MAX-RANDOM_TIMER_MIN) + RANDOM_TIMER_MIN\n fmt.Printf(\" peer %d term %d -- candidate -- :Set snooze timer to time %f\\n\", rf.me, rf.currentTerm, snoozeTime)\n time.Sleep(time.Duration(snoozeTime) * time.Millisecond) \n \n rf.mu.Lock()\n fmt.Printf(\" peer %d term %d -- candidate -- :Waking up from snooze to count votes. %f\\n\", rf.me, rf.currentTerm, snoozeTime)\n if (rf.state != Follower) {\n fmt.Printf(\"-> Peer %d term %d -- candidate --: Start Counting votes...\\n\\n\",rf.me, rf.currentTerm)\n \n for id:=0; id < numPeers; id++ {\n if id != rf.me && replies[id].VoteGranted {\n voteCount++\n } \n }\n\n if voteCount > numPeers/2 {\n // Initialize leader nextIndex and match index\n for id:=0; id< (len(rf.peers)-1); id++{\n rf.nextIndex[id] = len(rf.log)\n rf.matchIndex[id] = 0\n }\n\n fmt.Printf(\"-- peer %d candidate: I am elected leader for term %d. voteCount:%d majority_treshold %d\\n\\n\",rf.me,rf.currentTerm, voteCount, numPeers/2)\n rf.state = Leader\n fmt.Printf(\"-> Peer %d leader of term %d: I send first heartbeat round to assert my authority.\\n\\n\",rf.me, rf.currentTerm)\n go rf.sendHeartbeats()\n // sanity check: (if there is another leader in this term then it cannot be get the majority of votes)\n if rf.gotHeartbeat {\n log.Fatal(\"Two leaders won election in the same term!\")\n }\n } else if rf.gotHeartbeat {\n fmt.Printf(\"-- peer %d candidate of term %d: I got heartbeat from a leader. So I step down :) \\n\",rf.me, rf.currentTerm)\n rf.state = Follower\n } else {\n fmt.Printf(\"-- peer %d candidate term %d: Did not have enough votes. Moving to a new election term.\\n\\n\",rf.me,rf.currentTerm)\n } \n } \n rf.mu.Unlock()\n \n\n case Leader:\n fmt.Printf(\"-- Peer %d term %d: I am leader.\\n\\n\",rf.me, rf.currentTerm)\n snoozeTime := (1/HEARTBEAT_RATE)*1000 \n fmt.Printf(\" Leader %d term %d: snooze for %f\\n\", rf.me, rf.currentTerm, snoozeTime)\n \n time.Sleep(time.Duration(snoozeTime) * time.Millisecond)\n \n rf.mu.Lock()\n if rf.state != Follower {\n\n if rf.gotHeartbeat {\n log.Fatal(\"Fatal Error: Have two leaders in the same term!!!\")\n }\n fmt.Printf(\" peer %d term %d --leader-- : I send periodic heartbeat.\\n\",rf.me, rf.currentTerm)\n go rf.sendHeartbeats()\n } \n rf.mu.Unlock()\n\n }\n }\n } ()\n \n\n return rf\n}", "func (r *Raft) candidate() int {\n\t//myId := r.Myconfig.Id\n\t//fmt.Println(\"Election started!I am\", myId)\n\n\t//reset the votes else it will reflect the votes received in last term\n\tr.resetVotes()\n\n\t//--start election timer for election-time out time, so when responses stop coming it must restart the election\n\n\twaitTime := 10\n\t//fmt.Println(\"ELection timeout is\", waitTime)\n\tElectionTimer := r.StartTimer(ElectionTimeout, waitTime)\n\t//This loop is for election process which keeps on going until a leader is elected\n\tfor {\n\t\tr.currentTerm = r.currentTerm + 1 //increment current term\n\t\t//fmt.Println(\"I am candidate\", r.Myconfig.Id, \"and current term is now:\", r.currentTerm)\n\n\t\tr.votedFor = r.Myconfig.Id //vote for self\n\t\tr.WriteCVToDisk() //write Current term and votedFor to disk\n\t\tr.f_specific[r.Myconfig.Id].vote = true\n\n\t\t//fmt.Println(\"before calling prepRV\")\n\t\treqVoteObj := r.prepRequestVote() //prepare request vote obj\n\t\t//fmt.Println(\"after calling prepRV\")\n\t\tr.sendToAll(reqVoteObj) //send requests for vote to all servers\n\t\t//this loop for reading responses from all servers\n\t\tfor {\n\t\t\treq := r.receive()\n\t\t\tswitch req.(type) {\n\t\t\tcase RequestVoteResponse: //got the vote response\n\t\t\t\tresponse := req.(RequestVoteResponse) //explicit typecasting so that fields of struct can be used\n\t\t\t\t//fmt.Println(\"Got the vote\", response.voteGranted)\n\t\t\t\tif response.voteGranted {\n\t\t\t\t\t//\t\t\t\t\ttemp := r.f_specific[response.id] //NOT ABLE TO DO THIS--WHY??--WORK THIS WAY\n\t\t\t\t\t//\t\t\t\t\ttemp.vote = true\n\n\t\t\t\t\tr.f_specific[response.id].vote = true\n\t\t\t\t\t//r.voteCount = r.voteCount + 1\n\t\t\t\t}\n\t\t\t\tvoteCount := r.countVotes()\n\t\t\t\t//fmt.Println(\"I am:\", r.Myconfig.Id, \"Votecount is\", voteCount)\n\t\t\t\tif voteCount >= majority {\n\t\t\t\t\t//fmt.Println(\"Votecount is majority, I am new leader\", r.Myconfig.Id)\n\t\t\t\t\tElectionTimer.Stop()\n\t\t\t\t\tr.LeaderConfig.Id = r.Myconfig.Id //update leader details\n\t\t\t\t\treturn leader //become the leader\n\t\t\t\t}\n\n\t\t\tcase AppendEntriesReq: //received an AE request instead of votes, i.e. some other leader has been elected\n\t\t\t\trequest := req.(AppendEntriesReq)\n\t\t\t\t//Can be clubbed with serviceAppendEntriesReq with few additions!--SEE LATER\n\n\t\t\t\t//fmt.Println(\"I am \", r.Myconfig.Id, \"candidate,got AE_Req from\", request.leaderId, \"terms my,leader are\", r.currentTerm, request.term)\n\t\t\t\twaitTime_secs := secs * time.Duration(waitTime)\n\t\t\t\tappEntriesResponse := AppendEntriesResponse{}\n\t\t\t\tappEntriesResponse.followerId = r.Myconfig.Id\n\t\t\t\tappEntriesResponse.success = false //false by default, in case of heartbeat or invalid leader\n\t\t\t\tif request.term >= r.currentTerm { //valid leader\n\t\t\t\t\tr.LeaderConfig.Id = request.leaderId //update leader info\n\t\t\t\t\tElectionTimer.Reset(waitTime_secs) //reset the timer\n\t\t\t\t\tvar myLastIndexTerm int\n\t\t\t\t\tif len(r.myLog) == 0 {\n\t\t\t\t\t\tmyLastIndexTerm = -1\n\n\t\t\t\t\t} else {\n\t\t\t\t\t\tmyLastIndexTerm = r.myLog[r.myMetaData.lastLogIndex].Term\n\t\t\t\t\t}\n\t\t\t\t\tif request.leaderLastLogIndex == r.myMetaData.lastLogIndex && request.term == myLastIndexTerm { //this is heartbeat from a valid leader\n\t\t\t\t\t\tappEntriesResponse.success = true\n\t\t\t\t\t}\n\t\t\t\t\tsend(request.leaderId, appEntriesResponse)\n\t\t\t\t\treturn follower\n\t\t\t\t} else {\n\t\t\t\t\t//check if log is same\n\t\t\t\t\t//fmt.Println(\"In candidate, AE_Req-else\")\n\t\t\t\t\tsend(request.leaderId, appEntriesResponse)\n\t\t\t\t}\n\t\t\tcase int:\n\t\t\t\twaitTime_secs := secs * time.Duration(waitTime)\n\t\t\t\tElectionTimer.Reset(waitTime_secs)\n\t\t\t\tbreak //come out of inner loop i.e. restart the election process\n\t\t\t\t//default: if something else comes, then ideally it should ignore that and again wait for correct type of response on channel\n\t\t\t\t//it does this, in the present code structure\n\t\t\t}\n\t\t}\n\t}\n}", "func newRaft(c *Config) *Raft {\n\tif err := c.validate(); err != nil {\n\t\tpanic(err.Error())\n\t}\n\t// Your Code Here (2A).\n\tr := &Raft{\n\t\tid: c.ID,\n\t\tPrs: make(map[uint64]*Progress),\n\t\tvotes: make(map[uint64]bool),\n\t\theartbeatTimeout: c.HeartbeatTick,\n\t\telectionTimeout: c.ElectionTick,\n\t\tRaftLog: newLog(c.Storage),\n\t}\n\thardSt, confSt, _ := r.RaftLog.storage.InitialState()\n\tif c.peers == nil {\n\t\tc.peers = confSt.Nodes\n\t}\n\tlastIndex := r.RaftLog.LastIndex()\n\tfor _, peer := range c.peers {\n\t\tif peer == r.id {\n\t\t\tr.Prs[peer] = &Progress{Next: lastIndex + 1, Match: lastIndex}\n\t\t} else {\n\t\t\tr.Prs[peer] = &Progress{Next: lastIndex + 1}\n\t\t}\n\t}\n\tr.becomeFollower(0, None)\n\tr.randomElectionTimeout = r.electionTimeout + rand.Intn(r.electionTimeout)\n\tr.Term, r.Vote, r.RaftLog.committed = hardSt.GetTerm(), hardSt.GetVote(), hardSt.GetCommit()\n\tif c.Applied > 0 {\n\t\tr.RaftLog.applied = c.Applied\n\t}\n\treturn r\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\n\trf.currentTerm = 0\n\trf.voteFor = VOTENULL\n\trf.logOfRaft = make([]LogOfRaft, 0)\n\trf.logOfRaft = append(rf.logOfRaft, LogOfRaft{0, 0, nil})\n\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\n\trf.nextIndex = make([]int, len(peers))\n\trf.matchIndex = make([]int, len(peers))\n\n\trf.state = Follower\n\n\trf.heartbeatInterval = time.Duration(HeartbeatInteval) * time.Millisecond\n\n\trf.grantVoteCh = make(chan bool, 1)\n\trf.appendEntryCh = make(chan bool, 1)\n\trf.leaderCh = make(chan bool, 1)\n\trf.exitCh = make(chan bool, 1)\n\n\trf.applyCh = applyCh\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\t//log.Printf(\" %v is start, its currentTerm is %v, votefor is %v, log is %v\", rf.me, rf.currentTerm, rf.voteFor, rf.logOfRaft)\n\n\tgo func() {\n\t\tfor {\n\t\t\tDPrintf(\" %v is alive and its state is %v\", rf.me, rf.state)\n\t\t\tselect {\n\t\t\tcase <-rf.exitCh:\n\t\t\t\tDPrintf(\" Exit Server(%v)\", rf.me)\n\t\t\t\tlog.Printf(\" Exit Server(%v)\", rf.me)\n\t\t\t\treturn\n\n\t\t\t\t//break\n\t\t\tdefault:\n\t\t\t}\n\t\t\telectionTimeout := GetRandomElectionTimeout()\n\t\t\trf.mu.Lock()\n\t\t\tstate := rf.state\n\t\t\t//log.Printf(\" Server(%d) state:%v, electionTimeout:%v\", rf.me, state, electionTimeout)\n\t\t\trf.mu.Unlock()\n\n\t\t\tswitch state {\n\t\t\tcase Follower:\n\t\t\t\tselect {\n\t\t\t\tcase <-rf.appendEntryCh:\n\t\t\t\tcase <-rf.grantVoteCh:\n\t\t\t\tcase <-time.After(electionTimeout):\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\trf.convertToCandidate()\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t}\n\t\t\tcase Candidate:\n\t\t\t\tgo rf.leaderElection()\n\t\t\t\tselect {\n\t\t\t\tcase <-rf.appendEntryCh:\n\t\t\t\tcase <-rf.grantVoteCh:\n\t\t\t\tcase <-rf.leaderCh:\n\t\t\t\tcase <-time.After(electionTimeout):\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\trf.convertToCandidate()\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t}\n\t\t\tcase Leader:\n\t\t\t\trf.startAppendEntries()\n\t\t\t\ttime.Sleep(rf.heartbeatInterval)\n\t\t\t}\n\n\t\t}\n\t}()\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n persister *Persister, applyCh chan ApplyMsg) *Raft {\n rf := &Raft{}\n rf.peers = peers\n rf.persister = persister\n rf.me = me\n\n // Your initialization code here (2A, 2B, 2C).\n rf.state = StateFollower\n rf.commitIndex = 0\n rf.votedFor = nilIndex\n rf.lastApplied = 0\n rf.currentTerm = 0\n // rf.log contains a dummy head\n rf.log = []LogEntry{LogEntry{rf.currentTerm, nil}}\n\n // initialize from state persisted before a crash\n rf.readPersist(persister.ReadRaftState())\n\n rf.print(\"Initialize\")\n // All servers\n go func() {\n for {\n if rf.isKilled {\n return\n }\n rf.mu.Lock()\n for rf.commitIndex > rf.lastApplied {\n rf.lastApplied++\n applyMsg := ApplyMsg{rf.lastApplied, rf.log[rf.lastApplied].Command, false, nil}\n applyCh <- applyMsg\n rf.print(\"applied log entry %d:%v\", rf.lastApplied, rf.log[rf.lastApplied])\n // Apply rf.log[lastApplied] into its state machine\n }\n rf.mu.Unlock()\n time.Sleep(50 * time.Millisecond)\n }\n }()\n\n // candidate thread\n go func() {\n var counterLock sync.Mutex\n for {\n if rf.isKilled {\n return\n }\n rf.mu.Lock()\n\t\t\tif rf.state == StateFollower { // ONLY follower would have election timeout\n\t\t\t\trf.state = StateCandidate\n\t\t\t}\n rf.mu.Unlock()\n duration := time.Duration(electionTimeout +\n Random(-electionRandomFactor, electionRandomFactor))\n time.Sleep(duration * time.Millisecond)\n rf.mu.Lock()\n\n if rf.state == StateCandidate {\n rf.print(\"start to request votes for term %d\", rf.currentTerm+1)\n counter := 0\n logLen := len(rf.log)\n lastTerm := 0\n lastIndex := logLen-1\n requestTerm := rf.currentTerm+1\n if logLen > 0 {\n lastTerm = rf.log[logLen-1].Term\n }\n rvArgs := RequestVoteArgs{requestTerm, rf.me, lastIndex, lastTerm}\n rvReplies := make([]RequestVoteReply, len(rf.peers))\n\n for index := range rf.peers {\n go func(index int) {\n ok := rf.sendRequestVote(index, &rvArgs, &rvReplies[index])\n rf.mu.Lock()\n if rvReplies[index].Term > rf.currentTerm {\n rf.currentTerm = rvReplies[index].Term\n rf.state = StateFollower\n rf.persist()\n }else if ok && (rvArgs.Term == rf.currentTerm) && rvReplies[index].VoteGranted {\n counterLock.Lock()\n counter++\n if counter > len(rf.peers)/2 && rf.state != StateLeader {\n rf.state = StateLeader\n rf.currentTerm = requestTerm\n rf.nextIndex = make([]int, len(rf.peers))\n rf.matchIndex = make([]int, len(rf.peers))\n // immediately send heartbeats to others to stop election\n for i := range rf.peers {\n rf.nextIndex[i] = len(rf.log)\n }\n rf.persist()\n\n rf.print(\"become leader for term %d, nextIndex = %v, rvArgs = %v\", rf.currentTerm, rf.nextIndex, rvArgs)\n }\n counterLock.Unlock()\n }\n rf.mu.Unlock()\n }(index)\n }\n }\n rf.mu.Unlock()\n }\n }()\n\n // leader thread\n go func(){\n for {\n if rf.isKilled {\n return\n }\n time.Sleep(heartbeatTimeout * time.Millisecond)\n rf.mu.Lock()\n // send AppendEntries(as heartbeats) RPC\n if rf.state == StateLeader {\n currentTerm := rf.currentTerm\n for index := range rf.peers {\n go func(index int) {\n // decrease rf.nextIndex[index] in loop till append success\n for {\n if index == rf.me || rf.state != StateLeader {\n break\n }\n // if rf.nextIndex[index] <= 0 || rf.nextIndex[index] > len(rf.log){\n // rf.print(\"Error: rf.nextIndex[%d] = %d, logLen = %d\", index, rf.nextIndex[index], len(rf.log))\n // }\n rf.mu.Lock()\n logLen := len(rf.log)\n appendEntries := rf.log[rf.nextIndex[index]:]\n prevIndex := rf.nextIndex[index]-1\n aeArgs := AppendEntriesArgs{currentTerm, rf.me,\n prevIndex, rf.log[prevIndex].Term,\n appendEntries, rf.commitIndex}\n aeReply := AppendEntriesReply{}\n rf.mu.Unlock()\n\n ok := rf.sendAppendEntries(index, &aeArgs, &aeReply)\n rf.mu.Lock()\n if ok && rf.currentTerm == aeArgs.Term { // ensure the reply is not outdated\n if aeReply.Success {\n rf.matchIndex[index] = logLen-1\n rf.nextIndex[index] = logLen\n rf.mu.Unlock()\n break\n }else {\n if aeReply.Term > rf.currentTerm { // this leader node is outdated\n rf.currentTerm = aeReply.Term\n rf.state = StateFollower\n rf.persist()\n rf.mu.Unlock()\n break\n }else{ // prevIndex not match, decrease prevIndex\n // rf.nextIndex[index]--\n // if aeReply.ConflictFromIndex <= 0 || aeReply.ConflictFromIndex >= logLen{\n // rf.print(\"Error: aeReply.ConflictFromIndex from %d = %d, logLen = %d\", aeReply.ConflictFromIndex, index, logLen)\n // }\n rf.nextIndex[index] = aeReply.ConflictFromIndex\n }\n }\n }\n rf.mu.Unlock()\n }\n }(index)\n }\n\n // Find logs that has appended to majority and update commitIndex\n for N := rf.commitIndex+1; N<len(rf.log); N++ {\n // To eliminate problems like the one in Figure 8,\n // Raft never commits log entries from previous terms by count- ing replicas. \n if rf.log[N].Term < rf.currentTerm{\n continue\n }else if rf.log[N].Term > rf.currentTerm{\n break\n }\n followerHas := 0\n for index := range rf.peers {\n if rf.matchIndex[index] >= N{\n followerHas++\n }\n }\n // If majority has the log entry of index N\n if followerHas > len(rf.peers) / 2 {\n rf.print(\"set commitIndex to %d, matchIndex = %v\", N, rf.matchIndex)\n rf.commitIndex = N\n }\n }\n }\n rf.mu.Unlock()\n }\n }()\n\n return rf\n}", "func (r *Raft) becomeLeader() {\n\t// NOTE: Leader should propose a noop entry on its term\n\tr.State = StateLeader\n\tr.Vote = 0\n\tfor p, _ := range r.Prs {\n\t\tif p == r.id {\n\t\t\tcontinue\n\t\t}\n\t\tr.Prs[p].Match = 0\n\t\tr.Prs[p].Next = r.RaftLog.LastIndex() + 1\n\t}\n\t//r.initializeProgress()\n\n\t// send heartbeat\n\t//m := pb.Message{MsgType: pb.MessageType_MsgBeat, From: r.id, To: r.id}\n\t//r.sendMsgLocally(m)\n\t// send noop message\n\tr.sendInitialAppend()\n\tr.electionElapsed = 0\n}", "func (m *Manager) init(ledgerIndex iotago.MilestoneIndex) error {\n\tm.currentLock.Lock()\n\tdefer m.currentLock.Unlock()\n\n\tcurrentProtoParams, err := m.storage.ProtocolParameters(ledgerIndex)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tm.current = currentProtoParams\n\tm.loadPending(ledgerIndex)\n\n\treturn nil\n}", "func Make(peers []*labrpc.ClientEnd, me int, persister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{} //一个raft实例\n\trf.peers = peers //一个raft实例包含的所有servers\n\trf.persister = persister //存放这台机器的持久状态persistent state\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.currentTerm = 0 //initialized to 0 on first boot\n\trf.state = \"follower\"\n\trf.voteFor = -1 // null if none\n\trf.log = make([]Entry, 1)\n\trf.commitIndex = 0 //initialized to 0\n\trf.lastApplied = 0 //initialized to 0\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\t//DPrintf(\"rf.commitIndex:%d, rf.lastApplied:%d\", rf.commitIndex, rf.lastApplied)\n\tr := time.Duration(rand.Int63()) % ElectionTimeout\n\trf.electionTimer = time.NewTimer(ElectionTimeout + r)\n\trf.appendEntriesTimers = make([]*time.Timer, len(rf.peers))\n\tfor peer := range(rf.peers) {\n\t\trf.appendEntriesTimers[peer] = time.NewTimer(ElectionTimeout)\n\t}\n\trf.applySignal = make(chan bool, 100)\n\trf.applyCh = applyCh\n\trf.applyTimer = time.NewTimer(ApplyLogTimeout)\n\n\t// 选举定时器\n\tgo func() {\n\t\t//DPrintf(\"选举定时器\")\n\t\tfor {\n\t\t\t//if rf.state != \"leader\" {\n\t\t\t//\t<-rf.electionTimer.timer.C // 定时器\n\t\t\t//\t//DPrintf(\"%d is %s, and change to candidate\", rf.me, rf.state)\n\t\t\t//\t//if rf.state == \"follower\" {\n\t\t\t//\trf.changeState(\"candidate\")\n\t\t\t//\t//}\n\t\t\t//\t//rf.mu.Unlock()\n\t\t\t//} else {\n\t\t\t//\trf.electionTimer.timer.Stop()\n\t\t\t//}\n\t\t\t// 即使被选为leader,选举定时器也不能停止,因为如果一旦有peer down出现,并且达不到quorum 法定人数,则不允许有leader被选出\n\t\t\t<-rf.electionTimer.C // 定时器\n\t\t\trf.changeState(\"candidate\")\n\t\t}\n\t}()\n\n\t// 发送appendEntries定时器\n\tfor peer := range(rf.peers) {\n\t\tif peer == rf.me {\n\t\t\tcontinue\n\t\t}\n\t\tgo func(peer int) {\n\t\t\tfor {\n\t\t\t\t<-rf.appendEntriesTimers[peer].C\n\t\t\t\tif rf.state == \"leader\" {\n\t\t\t\t\trf.appendEntries2Peer(peer)\n\t\t\t\t}\n\t\t\t}\n\t\t}(peer)\n\t}\n\n\t// commit 定时器\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-rf.applyTimer.C:\n\t\t\t\trf.applySignal <- true\n\t\t\tcase <-rf.applySignal:\n\t\t\t\trf.apply()\n\t\t\t}\n\t\t}\n\n\t}()\n\n\n\tgo func() {\n\t\tfor !rf.killed() {\n\t\t\ttime.Sleep(2000 * time.Millisecond)\n\t\t\tif rf.lockName != \"\" {\n\t\t\t\tDPrintf(\"%d who has lock: %s; iskilled:%v; duration: %v; MaxLockTime is :%v; rf.loclkStart: %v; rf.lockEnd: %v\\n\", rf.me, rf.lockName, rf.killed(), rf.lockEnd.Sub(rf.lockStart).Nanoseconds()/1e6, MaxLockTime, rf.lockStart, rf.lockEnd)\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\n\trf := &Raft{}\n\trf.peers = peers\n\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.state = FOLLOWER\n\trf.votedFor = -1\n\trf.log = make([]LogEntry, 1) //[]LogEntry{{0, nil}} // first index is 1, first term is also 1 , {LogTerm, Command}\n\trf.currentTerm = 0\n\n\t//rf.leaderCh = make(chan interface{})\n\t//rf.commitCh = make(chan interface{})\n\n\t//rf.electionTimerResetChan = make(chan bool)\n\t//rf.heartbeatResetChan = make(chan bool)\n\t// nextIndex和matchIndex只有leader用\n\trf.nextIndex = make([]int, len(peers))\n\tfor i := range rf.nextIndex {\n\t\trf.nextIndex[i] = rf.getLastIndex() + 1\n\t}\n\trf.matchIndex = make([]int, len(peers))\n\n\trf.applyCh = applyCh\n\n\trf.lastApplied = 0\n\trf.commitIndex = 0\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\t// Modify Make() to create a background goroutine that will kick off leader election\n\t// periodically by sending out RequestVote RPCs when it hasn't heard from another peer for a while.\n\t// This way a peer will learn who is the leader, if there is already a leader, or become the leader itself.\n\n\trf.electionTimer = time.NewTimer(rf.getRandomElectionTimeOut())\n\trf.heartbeatTimer = time.NewTimer(HeartbeatInterval)\n\trf.stopCh = make(chan struct{})\n\trf.notifyApplyCh = make(chan struct{})\n\tgo func() {\n\n\t\tfor !rf.killed() {\n\n\t\t\tselect {\n\t\t\tcase <-rf.electionTimer.C:\n\t\t\t\trf.lock(\"electionTimer\")\n\n\t\t\t\tif rf.state == LEADER {\n\t\t\t\t\trf.unlock(\"electionTimer1\")\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tif rf.state == FOLLOWER || rf.state == CANDIDATE {\n\t\t\t\t\t//DPrintf(\"ElectionTimer time out\")\n\t\t\t\t\trf.convertTo(CANDIDATE)\n\t\t\t\t\t// when the raft server becomes candidate\n\t\t\t\t\t// we should put the currentTerm update\n\t\t\t\t\t// and votedFor update in broadcastAppendEntries\n\t\t\t\t\t// or we have not time to win in the split vote\n\t\t\t\t\t// situation\n\n\n\t\t\t\t}\n\t\t\t\trf.unlock(\"electionTimer\")\n\n\t\t\tcase <-rf.heartbeatTimer.C:\n\t\t\t\trf.lock(\"heartbeatTimer\")\n\n\t\t\t\tif rf.state == LEADER {\n\t\t\t\t\t// rf.heartbeatTimer.Stop()\n\t\t\t\t\trf.stop(rf.heartbeatTimer)\n\t\t\t\t\trf.heartbeatTimer.Reset(HeartbeatInterval)\n\t\t\t\t\tgo rf.broadcastAppendEntries(rf.currentTerm)\n\t\t\t\t}\n\t\t\t\trf.unlock(\"heartbeatTimer\")\n\n\t\t\tcase <-rf.stopCh:\n\t\t\t\treturn\n\t\t\t}\n\n\t\t}\n\t}()\n\n\tgo func() {\n\t\tfor !rf.killed() {\n\t\t\tselect {\n\t\t\tcase <-rf.notifyApplyCh:\n\t\t\t\tgo rf.updateLastApplied()\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t//为了高可用需要存储在稳定介质中的数据\n\trf.votedFor = -1\n\trf.status = FOLLOWER\n\trf.currentTerm = 0\n\trf.logEntries = make([]LogEntry, 0, 10)\n\trf.logEntries = append(rf.logEntries, LogEntry{Term: 0})\n\n\t// failure发生以后重新计算的数据\n\trf.commitIndex = 0\n\trf.lastApplied = 0 //状态机丢失,从头开始计算得到状态机\n\n\t//成为leader后重新计算的数据\n\trf.nextIndex = make([]int, len(peers))\n\tfor i := range rf.nextIndex {\n\t\trf.nextIndex[i] = len(rf.logEntries)\n\t}\n\trf.matchIndex = make([]int, len(peers))\n\n\t//TODO: 学习Go time包中定时器的使用\n\tduration := rf.getElectionTimeout()\n\trf.electionTimer = time.NewTimer(duration)\n\n\trf.giveVoteCh = make(chan struct{}, 1)\n\trf.heartBeatCh = make(chan struct{}, 1)\n\n\trf.revCommand = make(chan struct{}, 1) //TODO:此处是否需要使用有缓存的channel\n\trf.stop = make(chan struct{}, 1)\n\trf.stateChangeCh = make(chan struct{}, 1)\n\trf.applyCh = applyCh\n\t// Your initialization code here (2A, 2B, 2C).\n\n\tgo func() {\n\t\trf.loop()\n\t}()\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\treturn rf\n}", "func (rf *Raft) startElection() {\n\tDPrintf(\"%v become-candidate.start-election\", rf.raftInfo())\n\trf.state = RaftCandidate\n\trf.resetElectionTimeoutTicks()\n\n\trf.term++\n\trf.votedFor = rf.me\n\trf.persist()\n\n\trf.voteGranted = map[int]bool{}\n\trf.broadcastRequestVote()\n}", "func (r *Raft) becomeLeader() {\n\n\tr.State = StateLeader\n\tr.heartbeatElapsed = 0\n\tr.Vote = None\n\n\t// init prs\n\tprs := make(map[uint64]*Progress)\n\tfor _, nodeId := range r.nodes {\n\t\tprs[nodeId] = &Progress{\n\t\t\tMatch: r.RaftLog.LastIndex(),\n\t\t\tNext: r.RaftLog.LastIndex() + 1,\n\t\t}\n\t}\n\tr.Prs = prs\n\n\t// NOTE: Leader should propose a noop entry on its term\n\tr.addNoopEntryToLog()\n\tif len(r.nodes) > 1 {\n\t\tr.sendMsgToAll(r.sendAppendWrap)\n\t} else {\n\t\tr.updateCommitted()\n\t}\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\t/// Start as a follower.\n\trf.currentState = \"FOLLOWER\"\n\trf.commitIndex=1\n\trf.lastApplied=1 // Initializing this to 1 as we have added dummy 0th entry\n\trf.votedFor=-1\n\t//05/12\n\t//Let the leader start with 0. When a candidate transitions to the candidate state it increments this value.\n\trf.currentTerm=0\n\t//Initialize the log.\n\t//This is a dummy entry\n\trf.log = append(rf.log, LogEntry{LastLogTerm: 0})\n\trf.applyCh = applyCh\n\trf.debug(\"++++++++++++++++++++++++++Length of the log during initialization---> %d \\n\",len(rf.log))\n\trf.electionTimer = time.NewTimer((400 + time.Duration(rand.Intn(300))) * time.Millisecond)\n\t// Your initialization code here (2A, 2B, 2C).\n\tgo rf.conductElection()\n\t//Send heart beat to everybody else\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{\n\t\tmu: sync.Mutex{},\n\t\tpeers: peers,\n\t\tPersister: persister,\n\n\t\tme: me,\n\t\tdead: 0,\n\t\tVoteFor: -1,\n\t\tTerm: 0,\n\t\tCommitIndex: -1,\n\t\tappliedLast: -1,\n\t\tLog: nil,\n\t\tnextIndex: make([]int, len(peers)),\n\t\tmatchIndex: make([]int, len(peers)),\n\t\tlastReceiveHeartBeat: time.Now(),\n\t\tstate: Follower,\n\t\tapplyCh: applyCh,\n\t\tLastIncludedIndex: -1,\n\t\tLastIncludedTerm: 0,\n\t\tappendEntryCh: make(chan bool),\n\t\tfollowerTime: make(map[int]time.Time),\n\t}\n\tfor i, _ := range rf.peers {\n\t\tif i != rf.me {\n\t\t\trf.followerTime[i] = time.Now()\n\t\t}\n\t}\n\tserverNum := len(rf.peers)\n\tif serverNum%2 == 1 {\n\t\trf.majorityNum = serverNum/2 + 1\n\t} else {\n\t\trf.majorityNum = serverNum / 2\n\t}\n\trf.readPersist(persister.ReadRaftState()) // warn rf restart ,restore persistent\n\tDPrintf(\"rf [%v] restart detail : %v\", rf.me, rf)\n\tgo rf.initialServer()\n\tgo rf.doSendHeartBeat()\n\t// Your initialization code here (2A, 2B, 2C).\n\t//fmt.Printf(\"rf initialize begin\")\n\t// initialize from state persisted before a crash\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\n\trf.votedFor = -1\n\trf.state = FOLLOWER\n\n\trf.logs = append(rf.logs, LogEntry{Term:0}) // add dummy\n\trf.currentTerm = 0\n\trf.heartBeatChan = make(chan bool, 100)\n\trf.leaderChan = make(chan bool, 100)\n\trf.voteChan = make(chan bool, 100)\n\trf.commitChan = make(chan bool, 100)\n\trf.chanApply = applyCh\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\trf.readSnapshot(persister.ReadSnapshot())\n\n\tDPrintf(\"[init] server %v(%p) initialization. %+v\", rf.me, rf, rf)\n\n\t// background go routine to track raft state\n\tgo func() {\n\t\tfor {\n\t\t\tswitch rf.state {\n\t\t\tcase FOLLOWER:\n\t\t\t\tselect {\n\t\t\t\tcase <- rf.heartBeatChan:\n\t\t\t\t\tDPrintf(\"[heartbeat]: server %v(%p) receive heartbeat. term: %v\", rf.me, rf, rf.currentTerm)\n\t\t\t\tcase <- rf.voteChan:\n\t\t\t\tcase <- time.After(time.Duration(rand.Int63() % 333 + 550) * time.Millisecond):\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\trf.state = CANDIDATE\n\t\t\t\t\tDPrintf(\"[state change]: server %v(%p), from FOLLOWER become a CANDIDATE. term: %v\", rf.me, rf, rf.currentTerm)\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t}\n\t\t\tcase CANDIDATE:\n\t\t\t\trf.mu.Lock()\n\t\t\t\trf.currentTerm++\n\t\t\t\trf.votedFor = me\n\t\t\t\trf.voteCounter = 1\n\t\t\t\trf.persist()\n\t\t\t\trf.mu.Unlock()\n\n\t\t\t\tDPrintf(\"[report]: server %v(%p) is CANDIDATE. term: %v\", rf.me, rf, rf.currentTerm)\n\n\t\t\t\tgo rf.broadcastVoteReq()\n\n\t\t\t\tselect {\n\t\t\t\tcase <- rf.heartBeatChan:\n\t\t\t\t\t//rf.mu.Lock()\n\t\t\t\t\trf.state = FOLLOWER\n\t\t\t\t\tDPrintf(\"[state change]: server %v(%p), CANDIDATE receive heartbeat, become a FOLLOWER. term: %v\", rf.me, rf, rf.currentTerm)\n\t\t\t\t\t//rf.mu.Unlock()\n\t\t\t\tcase <- rf.leaderChan:\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\trf.state = LEADER\n\t\t\t\t\trf.nextIndex = make([]int, len(rf.peers))\n\t\t\t\t\trf.matchIndex = make([]int, len(rf.peers))\n\t\t\t\t\t// 初始化每个follower的log匹配信息\n\t\t\t\t\tfor i := range rf.peers {\n\t\t\t\t\t\trf.nextIndex[i] = rf.getLastLogIndex() + 1\n\t\t\t\t\t\trf.matchIndex[i] = 0\n\t\t\t\t\t}\n\n\t\t\t\t\tDPrintf(\"[state change]: server %v(%p), from CANDIDATE become a LEADER. term: %v\", rf.me, rf, rf.currentTerm)\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\tcase <- time.After(time.Duration(rand.Int63() % 333 + 500) * time.Millisecond):\n\t\t\t\t}\n\n\t\t\tcase LEADER:\n\t\t\t\trf.broadcastAppendEntriesReq()\n\t\t\t\ttime.Sleep(50 * time.Millisecond) // 每秒20次\n\t\t\t}\n\t\t}\n\n\n\t}()\n\n\t// commit log\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <- rf.commitChan:\n\t\t\t\trf.mu.Lock()\n\t\t\t\tcommitIndex := rf.commitIndex\n\t\t\t\tbaseIndex := rf.logs[0].Index\n\n\t\t\t\tfor i := rf.lastApplied + 1; i <= commitIndex; i++ {\n\t\t\t\t\tmsg := ApplyMsg{CommandIndex:i, Command: rf.logs[i - baseIndex].Command, CommandValid: true}\n\t\t\t\t\tDPrintf(\"[commit log]: server %v, command: %v\", rf.me, msg.Command)\n\t\t\t\t\tapplyCh <- msg\n\t\t\t\t\trf.lastApplied = i\n\t\t\t\t}\n\n\t\t\t\trf.mu.Unlock()\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func (r *Raft) becomeLeader() {\n\t// leader 先发送一个空数据\n\tr.State = StateLeader\n\tr.Lead = r.id\n\tlastIndex := r.RaftLog.LastIndex()\n\tr.heartbeatElapsed = 0\n\tfor peer := range r.Prs {\n\t\tif peer == r.id {\n\t\t\tr.Prs[peer].Next = lastIndex + 2\n\t\t\tr.Prs[peer].Match = lastIndex + 1\n\t\t} else {\n\t\t\tr.Prs[peer].Next = lastIndex + 1\n\t\t}\n\t}\n\tr.RaftLog.entries = append(r.RaftLog.entries, pb.Entry{Term: r.Term, Index: r.RaftLog.LastIndex() + 1})\n\tr.bcastAppend()\n\tif len(r.Prs) == 1 {\n\t\tr.RaftLog.committed = r.Prs[r.id].Match\n\t}\n}", "func (r *Raft) leader() int {\n\tr.setNextIndex_All() //so that new leader sets it map\n\tr.sendAppendEntriesRPC() //send Heartbeats\n\twaitTime := 1 //duration between two heartbeats\n\twaitTime_msecs := msecs * time.Duration(waitTime)\n\tHeartbeatTimer := r.StartTimer(HeartbeatTimeout, waitTime) //starts the timer and places timeout object on the channel\n\twaitStepDown := 7\n\tRetryTimer := r.StartTimer(RetryTimeOut, waitStepDown)\n\tresponseCount := 0\n\ttotalCount := 0\n\tfor {\n\t\treq := r.receive() //wait for client append req,extract the msg received on self EventCh\n\t\tswitch req.(type) {\n\t\tcase ClientAppendReq:\n\t\t\t//reset the heartbeat timer, now this sendRPC will maintain the authority of the leader\n\t\t\tHeartbeatTimer.Reset(waitTime_msecs)\n\t\t\trequest := req.(ClientAppendReq)\n\t\t\tData := request.Data\n\t\t\t//No check for semantics of cmd before appending to log?\n\t\t\tr.AppendToLog_Leader(Data) //append to self log as byte array\n\t\t\tr.sendAppendEntriesRPC()\n\t\t\tresponseCount = 0 //for RetryTimer\n\t\tcase AppendEntriesResponse:\n\t\t\tresponse := req.(AppendEntriesResponse)\n\t\t\tresponseCount += 1\n\t\t\tif responseCount >= majority-1 { //excluding self\n\t\t\t\twaitTime_retry := msecs * time.Duration(waitStepDown)\n\t\t\t\tRetryTimer.Reset(waitTime_retry)\n\t\t\t}\n\t\t\tif !response.IsHeartBeat {\n\t\t\t\tretVal := r.serviceAppendEntriesResp(response, HeartbeatTimer, waitTime)\n\t\t\t\tif retVal == follower {\n\t\t\t\t\treturn follower\n\t\t\t\t}\n\t\t\t}\n\t\tcase AppendEntriesReq: // in case some other leader is also in function, it must fall back or remain leader\n\t\t\trequest := req.(AppendEntriesReq)\n\t\t\tif request.Term > r.myCV.CurrentTerm {\n\t\t\t\tr.myCV.CurrentTerm = request.Term //update self Term and step down\n\t\t\t\tr.myCV.VotedFor = -1 //since Term has increased so VotedFor must be reset to reflect for this Term\n\t\t\t\tr.WriteCVToDisk()\n\t\t\t\treturn follower //sender server is the latest leader, become follower\n\t\t\t} else {\n\t\t\t\t//reject the request sending false\n\t\t\t\treply := AppendEntriesResponse{r.myCV.CurrentTerm, false, r.Myconfig.Id, false, r.MyMetaData.LastLogIndex}\n\t\t\t\tr.send(request.LeaderId, reply)\n\t\t\t}\n\n\t\tcase RequestVote:\n\t\t\trequest := req.(RequestVote)\n\t\t\ttotalCount = responseCount + totalCount + 1 //till responses are coming, network is good to go!\n\t\t\tif totalCount >= majority {\n\t\t\t\twaitTime_retry := msecs * time.Duration(waitStepDown)\n\t\t\t\tRetryTimer.Reset(waitTime_retry)\n\t\t\t}\n\t\t\tr.serviceRequestVote(request, leader)\n\n\t\tcase int: //Time out-time to send Heartbeats!\n\t\t\ttimeout := req.(int)\n\t\t\tif timeout == RetryTimeOut { //that means responses are not being received--means partitioned so become follower\n\t\t\t\tRetryTimer.Stop()\n\t\t\t\treturn follower\n\t\t\t}\n\t\t\tif timeout == HeartbeatTimeout {\n\t\t\t\tHeartbeatTimer.Reset(waitTime_msecs)\n\t\t\t\tresponseCount = 0 //since new heartbeat is now being sent\n\t\t\t\t//it depends on nextIndex which is correctly read in prepAE_Req method,since it was AE other than HB(last entry), it would have already modified the nextIndex map\n\t\t\t\tr.sendAppendEntriesRPC()\n\t\t\t}\n\t\t}\n\t}\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.role = Follower\n\trf.votedFor = -1\n\trf.nextIndex = make([]int, len(rf.peers))\n\trf.matchIndex = make([]int, len(rf.peers))\n\trf.validRpcTimestamp = time.Now()\n\trf.applyCh = applyCh\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\t// Your initialization code here (2A, 2B, 2C).\n\tgo func() {\n\t\tfor {\n\t\t\trf.mu.Lock()\n\t\t\trole := rf.role\n\n\t\t\tswitch role {\n\t\t\tcase Follower:\n\t\t\t\tts := rf.validRpcTimestamp\n\t\t\t\trf.mu.Unlock()\n\t\t\t\trf.runAsFollower(ts)\n\t\t\tcase Candidate:\n\t\t\t\tlastLogIndex := len(rf.log) + rf.compactIndex\n\t\t\t\targs := &RequestVoteArgs{\n\t\t\t\t\tTerm: rf.currentTerm,\n\t\t\t\t\tCandidateID: rf.me,\n\t\t\t\t\tLastLogIndex: lastLogIndex,\n\t\t\t\t}\n\t\t\t\tif lastLogIndex > rf.compactIndex {\n\t\t\t\t\targs.LastLogTerm = rf.log[lastLogIndex - rf.compactIndex -1].Term\n\t\t\t\t}\n\t\t\t\trf.mu.Unlock()\n\t\t\t\trf.runAsCandidate(args)\n\t\t\tcase Leader:\n\t\t\t\tterm := rf.currentTerm\n\t\t\t\trf.mu.Unlock()\n\t\t\t\trf.runAsLeader(term)\n\t\t\t}\n\n\t\t\tif rf.killed() {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Wait for commitIndex to be updated.\n\tgo func() {\n\t\trf.mu.Lock()\n\t\tif rf.compactIndex != 0 {\n\t\t\tapplyCh <- ApplyMsg{\n\t\t\t\tCommandValid: false,\n\t\t\t\tSnapshot: persister.ReadSnapshot(),\n\t\t\t}\n\t\t\trf.lastApplied = rf.compactIndex\n\t\t\tDPrintf(\"instance %d after restart rf.lastApplied is %d\", rf.me, rf.lastApplied)\n\t\t}\n\t\trf.mu.Unlock()\n\n\t\tfor {\n\t\t\ttime.Sleep(10 * time.Millisecond)\n\n\t\t\tvar (\n\t\t\t\tentries []LogEntry\n\t\t\t\trecord int\n\t\t\t)\n\t\t\trf.mu.Lock()\n\t\t\tDPrintf(\"instance %d rf.commitIndex is %d, rf.lastApplied is %d, rf.compactIndex is %d, len(rf.log) is %d\", rf.me, rf.commitIndex, rf.lastApplied, rf.compactIndex, len(rf.log))\n\t\t\tif rf.commitIndex > rf.lastApplied {\n\t\t\t\tentries = rf.log[rf.lastApplied-rf.compactIndex : rf.commitIndex-rf.compactIndex]\n\t\t\t\trecord = rf.lastApplied\n\t\t\t\trf.lastApplied = rf.commitIndex\n\t\t\t}\n\t\t\trf.mu.Unlock()\n\n\t\t\tfor i := 0; i < len(entries); i++ {\n\t\t\t\trecord = record + 1\n\t\t\t\tapplyCh <- ApplyMsg{\n\t\t\t\t\tCommandValid: true,\n\t\t\t\t\tCommand: entries[i].Command,\n\t\t\t\t\tCommandIndex: record,\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif rf.killed() {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func (room *GameRoom) init() {\n\tif room.whiteList == nil {\n\t\troom.whiteList = make(map[string]*User)\n\t\troom.blackList = make(map[string]*User)\n\t\troom.inEffectCard = []*Card{}\n\t\troom.turnPlayers = []*User{}\n\t\troom.Turn = 0\n\t\troom.Clock = 11\n\t\troom.IsRouletteTurn = true\n\t\troom.IsRouletteTurn = false\n\t\troom.Level = 1\n\t}\n}", "func initilize() bool {\n\t// Create self entry\n\tLocalIP = getLocalIP().String()\n\tLogger = NewSsmsLogger(LocalIP)\n\ttimestamp := time.Now().UnixNano()\n\tstate := StateAlive\n\tCurrentMember = &Member{uint64(timestamp), ip2int(getLocalIP()), uint8(state)}\n\n\t// Create member list\n\tCurrentList = NewMemberList(20)\n\n\t// Make necessary tables\n\tPingAckTimeout = make(map[uint16]*time.Timer)\n\tFailureTimeout = make(map[[2]uint64]*time.Timer)\n\tDuplicateUpdateCaches = make(map[uint64]uint8)\n\tTTLCaches = NewTtlCache()\n\n\treturn true\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.state = Follower\n\trf.currentTerm = 0\n\trf.votedFor = NULL\n\n\trf.log = make([]Log, 1)\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\trf.nextIndex = make([]int, len(peers))\n\trf.matchIndex = make([]int, len(peers))\n\n\trf.applyCh = applyCh\n\trf.voteCh = make(chan bool, 1)\n\trf.appendLogCh = make(chan bool, 1)\n\trf.killCh = make(chan bool, 1)\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\theartBeatTime := time.Duration(100) * time.Millisecond\n\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <-rf.killCh:\n\t\t\t\treturn\n\t\t\tdefault:\n\t\t\t}\n\t\t\telectionTime := time.Duration(rand.Intn(200)+300) * time.Millisecond\n\t\t\trf.mu.Lock()\n\t\t\tstate := rf.state\n\t\t\trf.mu.Unlock()\n\t\t\tswitch state {\n\t\t\tcase Follower, Candidate:\n\t\t\t\tselect {\n\t\t\t\tcase <-rf.voteCh:\n\t\t\t\tcase <-rf.appendLogCh:\n\t\t\t\tcase <-time.After(electionTime):\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\trf.beCandidate()\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t}\n\t\t\tcase Leader:\n\t\t\t\trf.startAppendLog()\n\t\t\t\ttime.Sleep(heartBeatTime)\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.currentTerm = 0\n\trf.votedFor = -1\n\trf.electionTimer = time.NewTimer(0)\n\t//first index is 1?\n\trf.applyCh = applyCh\n\trf.log = []LogEntries{LogEntries{0, 0}}\n\n\t//rf.applyCh <- ApplyMsg{true, 0, 0}\n\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\n\trf.nextIndex = make([]int, len(rf.peers))\n\trf.matchIndex = make([]int, len(rf.peers))\n\n\trf.resetLeaderNextIndex()\n\trf.resetLeaderMatchIndex()\n\n\trf.raftState = Follower\n\trf.killHeartBeatLoop = make(chan struct{}, 1)\n\trf.killElectionEventLoop = make(chan struct{}, 1)\n\n\t//may be need modification later\n\trf.lastIncludedIndex = 0\n\trf.lastIncludedTerm = 0\n\n\trf.notifyApplyCh = make(chan struct{}, 100)\n\trf.shutdownApply = make(chan struct{}, 1)\n\n\tgo rf.leaderElectionEventLoop()\n\tgo rf.heartbeatEventLoop()\n\tgo rf.apply()\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(rf.persister.ReadRaftState())\n\tDPrintf(\"after readPersist me is %v rf lastindex is %v commitindex is %v log is %v \",\n\t\trf.me, rf.lastIncludedIndex, rf.commitIndex, rf.log)\n\treturn rf\n}", "func leaderElection(nodeCtx *NodeCtx) {\n\t// The paper doesnt specifically mention any leader election protocols, so we assume that the leader election protocol\n\t// used in bootstrap is also used in the normal protocol, with the adition of iteration (unless the same leader would\n\t// be selected).\n\n\t// TODO actually add a setup phase where one must publish their hash. This way there will always\n\t// be a leader even if some nodes are offline. But with the assumption that every node is online\n\t// this works fine.\n\n\t// get current randomness\n\trecBlock := nodeCtx.blockchain.getLastReconfigurationBlock()\n\trnd := recBlock.Randomness\n\n\t// get current iteration\n\t_currIteration := nodeCtx.i.getI()\n\tcurrI := make([]byte, 8)\n\tbinary.LittleEndian.PutUint64(currI, uint64(_currIteration))\n\n\tlistOfHashes := make([]byte32sortHelper, len(nodeCtx.committee.Members))\n\t// calculate hash(id | rnd | currI) for every member\n\tii := 0\n\tfor _, m := range nodeCtx.committee.Members {\n\t\tconnoctated := byteSliceAppend(m.Pub.Bytes[:], rnd[:], currI)\n\t\thsh := hash(connoctated)\n\t\tlistOfHashes[ii] = byte32sortHelper{m.Pub.Bytes, hsh}\n\t\tii++\n\t}\n\n\t// sort list\n\tlistOfHashes = sortListOfByte32SortHelper(listOfHashes)\n\n\t// calculate hash of self\n\tselfHash := hash(byteSliceAppend(nodeCtx.self.Priv.Pub.Bytes[:], rnd[:], currI))\n\t// fmt.Println(\"self: \", bytes32ToString(selfHash), bytes32ToString(nodeCtx.self.Priv.Pub.Bytes))\n\t// for i, lof := range listOfHashes {\n\t// \tfmt.Println(i, bytes32ToString(lof.toSort), bytes32ToString(lof.original))\n\t// }\n\n\t// the leader is the lowest in list except if selfHash is lower than that.\n\t// fmt.Println(byte32Operations(selfHash, \"<\", listOfHashes[0].toSort))\n\tif byte32Operations(selfHash, \"<\", listOfHashes[0].toSort) {\n\t\tnodeCtx.committee.CurrentLeader = nodeCtx.self.Priv.Pub\n\t\tlog.Println(\"I am leader!\", nodeCtx.amILeader())\n\t} else {\n\t\tleader := listOfHashes[0].original\n\t\tnodeCtx.committee.CurrentLeader = nodeCtx.committee.Members[leader].Pub\n\t}\n}", "func (r *Raft) CallElection(){\n\t\n\tr.CurrentTerm+=1 // increase the current term by 1 to avoid conflict\n\tVoteAckcount:=1 // Number of vote received, initialised to 1 as own vote fo candiate is positive\n\tr.IsLeader = 0 // Set the state of server as candiate\n\tvar VoteCount =make (chan int,(len(r.ClusterConfigV.Servers)-1))\n\t//fmt.Println(\"Sending vote requests for:\",r.Id)\n\t\n\tfor _,server := range r.ClusterConfigV.Servers {\t\t\t\n\t\t\t\tif server.Id != r.Id{\n\t\t\t\t\tgo r.sendVoteRequestRpc(server,VoteCount) \t\t\t\t\t\n\t\t\t\t}}\n\n\tfor i:=0;i< len(r.ClusterConfigV.Servers)-1;i++ {\n\t\t\t\t\tVoteAckcount = VoteAckcount+ <- VoteCount \n\t\t\t\t\t// if Candiate gets majoirty, declare candiate as Leader and send immediae heartbeat to followers declaring\n\t\t\t\t\t// election of new leader\n\t\t\t\tif VoteAckcount > (len(r.ClusterConfigV.Servers)/2) && r.IsLeader == 0 { \n\t\t\t\t\tlog.Println(\"New leader is:\",r.Id)\n\t\t\t\t\tr.IsLeader=1\n\t\t\t\t\tr.LeaderId=r.Id\n\t\t\t\t\traft.SendImmediateHeartBit <- 1\n\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t}\t\t\n\t\tif r.IsLeader==1{\n\t\t\t// initlised next index to lastlog index, and match index to 0 fro all servers\n\t\tfor _,server := range r.ClusterConfigV.Servers {\n\t\t\t\tr.NextIndex[server.Id]=len(r.Log)\n\t\t\t\tr.MatchIndex[server.Id]=0\n\t\t\t\tr.ResetTimer()\n\t\t\t}\n\t\t}else{ \n\t\t\t// Is candidate fails to get elected, fall back to follower state and reset timer for reelection \n\t\t\tr.IsLeader=2\n\t\t\tr.ResetTimer()\n\t\t}\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.applyChan = applyCh\n\n\t// Your initialization code here (2A, 2B, 2C).\n\tDPrintf(\"Peer-%d begins to initialize.\\n\", rf.me)\n\trf.currentTerm = 0\n\trf.voteFor = -1\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\trf.state = Follower\n\ttimeout := RaftElectionTimeout\n\trf.heartbeatInterval = time.Duration(timeout / 2)\n\trf.electionTimeout = time.Duration(timeout)\n\trf.eventChan = make(chan Event, 1) // heartbeat should have 1 entry for message, this can void deadlock.\n\trf.log = make([]LogEntry, 1)\n\trf.nextIndex = make(map[int]int)\n\trf.matchIndex = make(map[int]int)\n\trf.maxAttempts = 3\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\t// init nextIndex and matchIndex\n\tlogSize := len(rf.log)\n\tfor key, _ := range rf.peers {\n\t\trf.nextIndex[key] = logSize\n\t\trf.matchIndex[key] = 0\n\t}\n\n\tDPrintf(\"Peer-%d is initialized. log=%v, term=%d, leader=%d\\n\", rf.me, rf.log, rf.currentTerm, rf.state)\n\n\t// start services.\n\tDPrintf(\"Peer-%d start services\\n\", rf.me)\n\tgo rf.electionService()\n\tgo rf.applyService()\n\tgo rf.logSyncService()\n\n\t// for rf.state != End {\n\t// \tsleep(1000)\n\t// }\n\n\treturn rf\n}", "func (a *acceptor) init() error {\n\tinstanceID, err := a.state.load()\n\tif err != nil {\n\t\tlNLErr(\"Load State fail, error: %v\", err)\n\t\treturn err\n\t}\n\n\tif instanceID == 0 {\n\t\tlPLGImp(a.conf.groupIdx, \"Empty database\")\n\t}\n\n\ta.setInstanceID(instanceID)\n\n\tlPLGImp(a.conf.groupIdx, \"OK\")\n\n\treturn nil\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\tnPeers := len(peers)\n\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.currentTerm = 0\n\trf.votedFor = -1\n\trf.logs = make([]*LogEntry, 0)\n\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\n\trf.nextIndex = make([]int, nPeers)\n\trf.matchIndex = make([]int, nPeers)\n\n\trf.state = Follower\n\trf.hasHeartbeat = false\n\n\trf.replicateCount = make([]int, 0)\n\trf.replicateStart = 0\n\n\trf.appliedCond = sync.NewCond(&rf.mu)\n\n\t// goroutine that track leader's heartbeat\n\tgo func(){\n\t\t// as stated in the paper, BroadcastTime << electionTimeOut\n\t\t// the example value in the paper is\n\t\t// BroadcastTime: 0.5~20 ms, electionTimeOut 150~300ms\n\t\t// since we have BroadcastTime of 0.1s, and we need to constrain\n\t\t// multiple rounds to be less than 5s\n\t\t// electionTimeOut should be 1 second-ish, i.e. 0.8 ~ 1.2 s\n\t\trand.Seed(time.Now().UTC().UnixNano())\n\t\tvar electionTimeOut time.Duration\n\t\tStateTransition:\n\t\tfor true {\n\t\t\telectionTimeOut = (time.Duration(rand.Intn(400) + 800)) * time.Millisecond\n\t\t\tif rf.state == Leader {\n\t\t\t\tfor i := 0; i < nPeers; i++{\n\t\t\t\t\tif i == me {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t// parameters & reply\n\t\t\t\t\tlastLogIndex := len(rf.logs)\n\t\t\t\t\tprevLogIndex := rf.nextIndex[i] - 1\n\t\t\t\t\tvar prevLogTerm int\n\t\t\t\t\tif prevLogIndex >= 1 {\n\t\t\t\t\t\tprevLogTerm = rf.logs[prevLogIndex - 1].Term\n\t\t\t\t\t} else {\n\t\t\t\t\t\tprevLogTerm = -1\n\t\t\t\t\t}\n\t\t\t\t\targs := &AppendEntriesArgs{\n\t\t\t\t\t\tTerm: rf.currentTerm,\n\t\t\t\t\t\tLeaderId: rf.me,\n\t\t\t\t\t\tPrevLogIndex: prevLogIndex,\n\t\t\t\t\t\tPrevLogTerm: prevLogTerm,\n\t\t\t\t\t\tLeaderCommit: rf.commitIndex,\n\t\t\t\t\t}\n\t\t\t\t\treply := &AppendEntriesReply{}\n\n\t\t\t\t\tif len(rf.logs) > 0 && len(rf.logs) >= rf.nextIndex[i] {\n\t\t\t\t\t\targs.Entries = rf.logs[prevLogIndex:]\n\t\t\t\t\t\t// send AppendEntries RPC with logs\n\t\t\t\t\t\tgo func(server int){\n\t\t\t\t\t\t\tDPrintf(\"Peer %d: Sending new entries(%d ~ %d) to Peer %d\\n\", rf.me, prevLogIndex + 1, len(rf.logs), server)\n\t\t\t\t\t\t\tif ok := rf.sendAppendEntriesRPC(server, args, reply); ok {\n\t\t\t\t\t\t\t\t// term, success\n\t\t\t\t\t\t\t\tif reply.Success {\n\t\t\t\t\t\t\t\t\trf.nextIndex[server] = lastLogIndex + 1\n\t\t\t\t\t\t\t\t\trf.matchIndex[server] = lastLogIndex\n\t\t\t\t\t\t\t\t\tfor index := prevLogIndex + 1; index <= lastLogIndex; index++ {\n\t\t\t\t\t\t\t\t\t\tDPrintf(\"Peer %d: access replicateCount[%d] (replicateStart = %d)\", rf.me, index - 1 -rf.replicateStart, rf.replicateStart)\n\t\t\t\t\t\t\t\t\t\taccessIndex := index - 1 -rf.replicateStart\n\t\t\t\t\t\t\t\t\t\tif accessIndex < 0 || rf.replicateCount[accessIndex] > len(rf.peers) / 2 {\n\t\t\t\t\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t\trf.replicateCount[accessIndex]++;\n\t\t\t\t\t\t\t\t\t\tif rf.replicateCount[accessIndex] > len(rf.peers) / 2 {\n\t\t\t\t\t\t\t\t\t\t\trf.commitIndex = index\n\t\t\t\t\t\t\t\t\t\t\tDPrintf(\"Peer %d: entry %d has been replicated to majority, sending apply signal\\n\", me, index)\n\t\t\t\t\t\t\t\t\t\t\trf.appliedCond.Signal()\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\tif updated := rf.updateTerm(reply.Term); !updated {\n\t\t\t\t\t\t\t\t\t\t// rejected because log consistency\n\t\t\t\t\t\t\t\t\t\trf.nextIndex[server]--;\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t// DPrintf(\"Peer %d: Can't reach Peer %d, AppendEntries RPC returned false!\\n\", me, server)\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}(i)\n\t\t\t\t\t} else {\n\t\t\t\t\t\t// Send Heartbeat\n\t\t\t\t\t\t// DPrintf(\"Peer %d: Sending heartbeat to Peer %d\", rf.me, i)\n\t\t\t\t\t\tgo rf.sendAppendEntriesRPC(i, args, reply)\n\t\t\t\t\t}\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t}\n\t\t\t\ttime.Sleep(BroadcastIntv)\n\t\t\t} else if rf.state == Candidate {\n\t\t\t\trf.currentTerm++\n\t\t\t\trf.votedFor = me\n\t\t\t\tvoteCount := 1\n\t\t\t\tvoteCh := make(chan int)\n\n\t\t\t\tfor i := 0; i < nPeers; i++ {\n\t\t\t\t\tif i == me {\n\t\t\t\t\t\tcontinue\n\t\t\t\t\t}\n\t\t\t\t\tgo func(server int){\n\t\t\t\t\t\tlastLogIndex := len(rf.logs)\n\t\t\t\t\t\tvar lastLogTerm int\n\t\t\t\t\t\tif lastLogIndex > 0 {\n\t\t\t\t\t\t\tlastLogTerm = rf.logs[lastLogIndex - 1].Term\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\tlastLogTerm = -1\n\t\t\t\t\t\t}\n\n\t\t\t\t\t\targs := &RequestVoteArgs{\n\t\t\t\t\t\t\tTerm: rf.currentTerm,\n\t\t\t\t\t\t\tCandidateId: me,\n\t\t\t\t\t\t\tLastLogIndex: lastLogIndex,\n\t\t\t\t\t\t\tLastLogTerm: lastLogTerm,\n\t\t\t\t\t\t}\n\t\t\t\t\t\treply := &RequestVoteReply{}\n\n\t\t\t\t\t\t// Send RequestVote RPC\n\t\t\t\t\t\t// DPrintf(\"Peer %d: Sending RequestVote RPC to peer %d.\\n\", me, server)\n\t\t\t\t\t\tok := rf.sendRequestVote(server, args, reply)\n\t\t\t\t\t\tif ok {\n\t\t\t\t\t\t\trf.updateTerm(reply.VotersTerm)\n\t\t\t\t\t\t\tif reply.VoteGranted {\n\t\t\t\t\t\t\t\t// DPrintf(\"Peer %d: Receive grant vote from Peer %d.\\n\", me, server)\n\t\t\t\t\t\t\t\tvoteCh <- 1\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t// DPrintf(\"Peer %d: Can't reach Peer %d, RPC returned false!\\n\", me, server)\n\t\t\t\t\t\t}\n\t\t\t\t\t}(i)\n\t\t\t\t}\n\n\t\t\t\ttimeOutCh := time.After(electionTimeOut)\n\t\t\t\t// DPrintf(\"Peer %d(@%s state) will wait %v for votes.\\n\", me, rf.state, electionTimeOut)\n\t\t\t\tvoteLoop:\n\t\t\t\tfor {\n\t\t\t\t\tselect {\n\t\t\t\t\tcase <- voteCh:\n\t\t\t\t\t\tvoteCount++\n\t\t\t\t\t\tif rf.state == Candidate && voteCount > nPeers / 2 {\n\t\t\t\t\t\t\tDPrintf(\"Peer %d: Received majority votes, become leader now!\\n\", me)\n\t\t\t\t\t\t\trf.state = Leader\n\t\t\t\t\t\t\t// before become a leader, init nextIndex[] and matchIndex[]\n\t\t\t\t\t\t\tfor i := 0; i < nPeers; i++ {\n\t\t\t\t\t\t\t\tcurrentIndex := len(rf.logs)\n\t\t\t\t\t\t\t\trf.nextIndex[i] = currentIndex + 1\n\t\t\t\t\t\t\t\trf.matchIndex[i] = currentIndex // optimistic guessing\n\n\t\t\t\t\t\t\t\t// replicateCount\n\t\t\t\t\t\t\t\tuncommitedLogs := len(rf.logs) - rf.commitIndex\n\t\t\t\t\t\t\t\trf.replicateCount = make([]int, uncommitedLogs, 1)\n\t\t\t\t\t\t\t\trf.replicateStart = rf.commitIndex\n\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tbreak voteLoop\n\t\t\t\t\t\t}\n\t\t\t\t\tcase <- timeOutCh:\n\t\t\t\t\t\tDPrintf(\"Peer %d(@%s): Election timed out.\\n\", me, rf.state)\n\t\t\t\t\t\tbreak voteLoop\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t} else if rf.state == Follower {\n\t\t\t\ttime.Sleep(BroadcastIntv)\n\t\t\t\tif !rf.hasHeartbeat {\n\t\t\t\t\t// DPrintf(\"Peer %d(@%s state) hasn't receive heartbeat, will wait heartbeat for %v!\\n\", me, rf.state, electionTimeOut)\n\t\t\t\t\ttimer := time.After(electionTimeOut)\n\t\t\t\t\tStartElectionTimer:\n\t\t\t\t\tfor {\n\t\t\t\t\t\tselect{\n\t\t\t\t\t\tcase <- timer:\n\t\t\t\t\t\t\tDPrintf(\"Peer %d: Leader timed Out!\\n\", me)\n\t\t\t\t\t\t\trf.state = Candidate\n\t\t\t\t\t\t\tcontinue StateTransition\n\t\t\t\t\t\tdefault:\n\t\t\t\t\t\t\tif rf.hasHeartbeat {\n\t\t\t\t\t\t\t\tbreak StartElectionTimer\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\trf.hasHeartbeat = false\n\t\t\t}\n\t\t}\n\t}()\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\tgo func(){\n\t\trf.mu.Lock() // 不能放在for循环里,否则循环只能执行一遍\n\t\tfor {\n\t\t\tfor rf.lastApplied < rf.commitIndex {\n\t\t\t\tindex := rf.lastApplied + 1\n\t\t\t\tapplyMsg := ApplyMsg{\n\t\t\t\t\tCommandValid: true,\n\t\t\t\t\tCommand: rf.logs[index - 1].Command,\n\t\t\t\t\tCommandIndex: index,\n\t\t\t\t}\n\t\t\t\tapplyCh <- applyMsg\n\t\t\t\trf.lastApplied = index\n\t\t\t\tvar printlog []interface{}\n\t\t\t\tfor _, v := range rf.logs {\n\t\t\t\t\tprintlog = append(printlog, v.Command)\n\t\t\t\t}\n\t\t\t\tDPrintf(\"Peer %d: applied entry %d to local state machine. logs: %v\", rf.me, index, printlog)\n\t\t\t}\n\t\t\trf.appliedCond.Wait()\n\t\t}\n\t}()\n\n\treturn rf\n}", "func (r *Raft) leader() int {\n\t//fmt.Println(\"In leader(), I am: \", r.Myconfig.Id)\n\n\tr.sendAppendEntriesRPC() //send Heartbeats\n\t//waitTime := 4 //duration between two heartbeats\n\twaitTime := 1\n\twaitTime_secs := secs * time.Duration(waitTime)\n\t//fmt.Println(\"Heartbeat time out is\", waitTime)\n\n\twaitTimeAE := 5 //max time to wait for AE_Response\n\tHeartbeatTimer := r.StartTimer(HeartbeatTimeout, waitTime) //starts the timer and places timeout object on the channel\n\t//var AppendEntriesTimer *time.Timer\n\twaitStepDown := 7\n\tRetryTimer := r.StartTimer(RetryTimeOut, waitStepDown)\n\t//fmt.Println(\"I am\", r.Myconfig.Id, \"timer created\", AppendEntriesTimer)\n\tresponseCount := 0\n\tfor {\n\n\t\treq := r.receive() //wait for client append req,extract the msg received on self eventCh\n\t\tswitch req.(type) {\n\t\tcase ClientAppendReq:\n\t\t\t//reset the heartbeat timer, now this sendRPC will maintain the authority of the leader\n\t\t\tHeartbeatTimer.Reset(waitTime_secs)\n\t\t\trequest := req.(ClientAppendReq)\n\t\t\tdata := request.data\n\t\t\t//fmt.Println(\"Received CA request,cmd is: \", string(data))\n\t\t\t//No check for semantics of cmd before appending to log?\n\t\t\tr.AppendToLog_Leader(data) //append to self log as byte array\n\t\t\tr.sendAppendEntriesRPC()\n\t\t\tresponseCount = 0 //for RetryTimer\n\t\t\t//AppendEntriesTimer = r.StartTimer(AppendEntriesTimeOut, waitTimeAE) //Can be written in HeartBeatTimer too\n\t\t\t//fmt.Println(\"I am\", r.Myconfig.Id, \"Timer assigned a value\", AppendEntriesTimer)\n\t\tcase AppendEntriesResponse:\n\t\t\tresponse := req.(AppendEntriesResponse)\n\t\t\t//fmt.Println(\"got AE_Response! from : \", response.followerId, response)\n\t\t\tresponseCount += 1\n\t\t\tif responseCount >= majority {\n\t\t\t\twaitTime_retry := secs * time.Duration(waitStepDown)\n\t\t\t\tRetryTimer.Reset(waitTime_retry)\n\t\t\t}\n\t\t\t//when isHeartBeat is true then success is also true according to the code in serviceAEReq so case wont be there when isHB is true and success is false\n\t\t\t// isHB true means it is a succeeded heartbeat hence no work to do if it is AE req then only proceed else do nothing and continue\n\t\t\t//So when follower's log is stale or he is more latest, it would set isHB false\n\t\t\tif !response.isHeartBeat {\n\t\t\t\tretVal := r.serviceAppendEntriesResp(response, HeartbeatTimer, waitTimeAE, waitTime)\n\t\t\t\tif retVal == follower {\n\t\t\t\t\treturn follower\n\t\t\t\t}\n\n\t\t\t}\n\n\t\tcase AppendEntriesReq: // in case some other leader is also in function, it must fall back or remain leader\n\t\t\trequest := req.(AppendEntriesReq)\n\t\t\tif request.term > r.currentTerm {\n\t\t\t\t//fmt.Println(\"In leader,AE_Req case, I am \", r.Myconfig.Id, \"becoming follower now, because request.term, r.currentTerm\", request.term, r.currentTerm)\n\t\t\t\tr.currentTerm = request.term //update self term and step down\n\t\t\t\tr.votedFor = -1 //since term has increased so votedFor must be reset to reflect for this term\n\t\t\t\tr.WriteCVToDisk()\n\t\t\t\treturn follower //sender server is the latest leader, become follower\n\t\t\t} else {\n\t\t\t\t//reject the request sending false\n\t\t\t\treply := AppendEntriesResponse{r.currentTerm, false, r.Myconfig.Id, false, r.myMetaData.lastLogIndex}\n\t\t\t\tsend(request.leaderId, reply)\n\t\t\t}\n\n\t\tcase int: //Time out-time to send Heartbeats!\n\t\t\ttimeout := req.(int)\n\t\t\tif timeout == RetryTimeOut {\n\t\t\t\tRetryTimer.Stop()\n\t\t\t\treturn follower\n\t\t\t}\n\t\t\t//fmt.Println(\"Timeout of\", r.Myconfig.Id, \"is of type:\", timeout)\n\n\t\t\t//waitTime_secs := secs * time.Duration(waitTime)\n\t\t\tif timeout == HeartbeatTimeout {\n\t\t\t\t//fmt.Println(\"Leader:Reseting HB timer\")\n\t\t\t\tHeartbeatTimer.Reset(waitTime_secs)\n\t\t\t\tresponseCount = 0 //since new heartbeat is now being sent\n\t\t\t\t//it depends on nextIndex which is correctly read in prepAE_Req method,\n\t\t\t\t//since it was AE other than HB(last entry), it would have already modified the nextIndex map\n\t\t\t\tr.sendAppendEntriesRPC() //This either sends Heartbeats or retries the failed AE due to which the timeout happened,\n\t\t\t\t//HeartbeatTimer.Reset(secs * time.Duration(8)) //for checking leader change, setting timer of f4 to 8s--DOESN'T work..-_CHECK\n\t\t\t}\n\n\t\t}\n\t}\n}", "func (r *Raft) Init(config *ClusterConfig, thisServerId int) {\n\tgo r.AcceptConnection(r.ClusterConfigV.Servers[thisServerId].ClientPort) // done\n\tgo r.AcceptRPC(r.ClusterConfigV.Servers[thisServerId].LogPort)\t//\n\tgo SendHeartbeat() // NOT Done TODO\n\tgo Evaluator()\t//\n\tgo AppendCaller()\n\tgo CommitCaller()\n\tgo DataWriter()\n\tr.SetElectionTimer()\n\tgo Loop() //if he is not leader TODO\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.numVote = 0\n\trf.log = make([]LogItem, 0)\n\trf.votedFor = -1\n\trf.currentTerm = 0\n\trf.state = Follower\n\trf.nextIndex = make([]int, len(peers))\n\trf.matchIndex = make([]int, len(peers))\n\trf.commitIndex = -1\n\trf.lastApplied = -1\n\n\t// random generate a duration\n\t//randomTimeOut := rand.Intn(MAXELECTIONTIMEOUT - MINELECTIONTIMEOUT) + MINELECTIONTIMEOUT\n\n\t//rf.timer = time.NewTimer(time.Duration(randomTimeOut))\n\n\t// Your initialization code here (2A, 2B, 2C).\n\t// 触发一定时间内没有收到heartbeat进行选举\n\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\trf.resetTimer()\n\n\n\n\treturn rf\n}", "func (r *Raft) candidate(timeout int) int {\n\twaitTime := timeout //added for passing timeout from outside--In SingleServerBinary\n\tresendTime := 5 //should be much smaller than waitTime\n\tElectionTimer := r.StartTimer(ElectionTimeout, waitTime)\n\t//This loop is for election process which keeps on going until a leader is elected\n\tfor {\n\t\t//reset the Votes else it will reflect the Votes received in last Term\n\t\tr.resetVotes()\n\t\tr.myCV.CurrentTerm += 1 //increment current Term\n\t\tr.myCV.VotedFor = r.Myconfig.Id //Vote for self\n\t\tr.WriteCVToDisk() //write Current Term and VotedFor to disk\n\t\tr.f_specific[r.Myconfig.Id].Vote = true //vote true\n\t\treqVoteObj := r.prepRequestVote() //prepare request Vote obj\n\t\tr.sendToAll(reqVoteObj) //send requests for Vote to all servers\n\t\tResendVoteTimer := r.StartTimer(ResendVoteTimeOut, resendTime)\n\t\tfor { //this loop for reading responses from all servers\n\t\t\treq := r.receive()\n\t\t\tswitch req.(type) {\n\t\t\tcase ClientAppendReq: ///candidate must also respond as false just like follower\n\t\t\t\trequest := req.(ClientAppendReq) //explicit typecasting\n\t\t\t\tresponse := ClientAppendResponse{}\n\t\t\t\tlogItem := LogItem{r.CurrentLogEntryCnt, false, request.Data} //lsn is count started from 0\n\t\t\t\tr.CurrentLogEntryCnt += 1\n\t\t\t\tresponse.LogEntry = logItem\n\t\t\t\tr.CommitCh <- &response.LogEntry\n\t\t\tcase RequestVoteResponse: //got the Vote response\n\t\t\t\tresponse := req.(RequestVoteResponse) //explicit typecasting so that fields of struct can be used\n\t\t\t\tif response.VoteGranted {\n\t\t\t\t\tr.f_specific[response.Id].Vote = true\n\t\t\t\t}\n\t\t\t\tVoteCount := r.countVotes()\n\t\t\t\tif VoteCount >= majority {\n\t\t\t\t\tResendVoteTimer.Stop()\n\t\t\t\t\tElectionTimer.Stop()\n\t\t\t\t\tr.LeaderConfig.Id = r.Myconfig.Id //update leader details\n\t\t\t\t\treturn leader //become the leader\n\t\t\t\t}\n\n\t\t\tcase AppendEntriesReq: //received an AE request instead of Votes, i.e. some other leader has been elected\n\t\t\t\trequest := req.(AppendEntriesReq)\n\t\t\t\tretVal := r.serviceAppendEntriesReq(request, nil, 0, candidate)\n\t\t\t\tif retVal == follower {\n\t\t\t\t\tResendVoteTimer.Stop()\n\t\t\t\t\tElectionTimer.Stop()\n\t\t\t\t\treturn follower\n\t\t\t\t}\n\n\t\t\tcase RequestVote:\n\t\t\t\trequest := req.(RequestVote)\n\t\t\t\t//==Can be shared with service request vote with additinal param of caller(candidate or follower)\n\t\t\t\tresponse := RequestVoteResponse{} //prep response object,for responding back to requester\n\t\t\t\tcandidateId := request.CandidateId\n\t\t\t\tresponse.Id = r.Myconfig.Id\n\t\t\t\tif r.isDeservingCandidate(request) {\n\t\t\t\t\tresponse.VoteGranted = true\n\t\t\t\t\tr.myCV.VotedFor = candidateId\n\t\t\t\t\tr.myCV.CurrentTerm = request.Term\n\t\t\t\t\tif request.Term > r.myCV.CurrentTerm { //write to disk only when value has changed\n\t\t\t\t\t\tr.WriteCVToDisk()\n\t\t\t\t\t}\n\t\t\t\t\tResendVoteTimer.Stop()\n\t\t\t\t\tElectionTimer.Stop()\n\t\t\t\t\treturn follower\n\t\t\t\t} else {\n\t\t\t\t\tresponse.VoteGranted = false\n\t\t\t\t}\n\t\t\t\tresponse.Term = r.myCV.CurrentTerm\n\t\t\t\tr.send(candidateId, response)\n\n\t\t\tcase int:\n\t\t\t\ttimeout := req.(int)\n\t\t\t\tif timeout == ResendVoteTimeOut {\n\t\t\t\t\trT := msecs * time.Duration(resendTime)\n\t\t\t\t\tResendVoteTimer.Reset(rT)\n\t\t\t\t\treqVoteObj := r.prepRequestVote() //prepare request Vote agn and send to all, ones rcvg the vote agn will vote true agn so won't matter and countVotes func counts no.of true entries\n\t\t\t\t\tr.sendToAll(reqVoteObj)\n\t\t\t\t} else if timeout == ElectionTimeout {\n\t\t\t\t\twaitTime_msecs := msecs * time.Duration(waitTime)\n\t\t\t\t\tElectionTimer.Reset(waitTime_msecs)\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n}", "func (rf *Raft) tryToBeLeader() {\n\t//Step 1\n\tvar maxVoteNum, currentSuccessNum int\n\trf.mu.Lock()\n\trf.currentTerm++\n\trf.votedFor = rf.me\n\trf.role = Candidate\n\tmaxVoteNum = len(rf.peers)\n\trf.mu.Unlock()\n\trf.persist()\n\n\tcurrentSuccessNum = 1\n\tvar mutex sync.Mutex\n\tfor i := 0; i < maxVoteNum; i++ {\n\t\tif i != rf.me {\n\t\t\tgo func(idx int) {\n\t\t\t\tvar templateArgs RequestVoteArgs\n\t\t\t\trf.mu.Lock()\n\t\t\t\taLeaderComeUp := rf.role == Follower || rf.role == Leader\n\n\t\t\t\tif aLeaderComeUp {\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\ttemplateArgs.Term = rf.currentTerm\n\t\t\t\ttemplateArgs.CandidateID = rf.me\n\t\t\t\ttemplateArgs.LastLogTerm = rf.logs[len(rf.logs)-1].Term\n\t\t\t\ttemplateArgs.LastLogIndex = len(rf.logs) - 1\n\t\t\t\trf.mu.Unlock()\n\n\t\t\t\targs := templateArgs\n\t\t\t\tvar reply RequestVoteReply\n\t\t\t\tok := rf.sendRequestVote(idx, &args, &reply)\n\n\t\t\t\trf.mu.Lock()\n\t\t\t\taLeaderComeUp = rf.role == Follower || rf.role == Leader || rf.role == None\n\t\t\t\trf.mu.Unlock()\n\t\t\t\tif aLeaderComeUp {\n\t\t\t\t\treturn\n\t\t\t\t} else {\n\t\t\t\t\tif ok {\n\t\t\t\t\t\tmutex.Lock()\n\t\t\t\t\t\tcurrentSuccessNum++\n\t\t\t\t\t\tmutex.Unlock()\n\t\t\t\t\t\tif currentSuccessNum >= maxVoteNum/2+1 {\n\t\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\t\trf.role = Leader\n\t\t\t\t\t\t\tfor i := 0; i < len(rf.peers); i++ {\n\t\t\t\t\t\t\t\trf.nextIndex[i] = len(rf.logs)\n\t\t\t\t\t\t\t\trf.matchIndex[i] = 0\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\tgo rf.logDuplicate()\n\t\t\t\t\t\t\trf.msgChan <- BecomeLeader\n\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}(i)\n\t\t}\n\t}\n\n}", "func init(){\n\n\tcustomers = loadAllCustomers()\n\tBuildRank()\n\tBuildTable()\n\n}", "func Make(peers []*labrpc.ClientEnd, me int, persister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.stopCh = make(chan struct{})\n\trf.applyCh = applyCh\n\trf.electionTimer = time.NewTimer(rf.randElectionTimeout())\n\trf.appendEntryTimer = make([]*time.Timer, len(peers))\n\tfor i,_ := range rf.peers {\n\t\trf.appendEntryTimer[i] = time.NewTimer(HeartBeatTimeout)\n\t}\n\n\trf.state = Follower\n\trf.currentTerm = 0\n\trf.votedFor = -1\n\t// 初始时日志长度为1\n\trf.logs = make([]LogEntry, 1)\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\trf.lastSnapshotIndex = 0\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\trf.nextIndex = make([]int, len(peers))\n\trf.matchIndex = make([]int, len(peers))\n\n\t// 发起投票\n\tgo func() {\n\t\tfor {\n\t\t\tselect {\n\t\t\tcase <- rf.stopCh:\n\t\t\t\treturn\n\t\t\tcase <- rf.electionTimer.C:\n\t\t\t\tgo rf.startElection()\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int, persister *Persister, applyCh chan ApplyMsg, gid ...int) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.applyCh = applyCh\n\n\tif len(gid) != 0 {\n\t\trf.gid = gid[0]\n\t} else {\n\t\trf.gid = -1\n\t}\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.currentTerm = 0\n\trf.voteFor = -1\n\trf.role = Follower\n\trf.logEntries = make([]LogEntry, 1)\n\trf.readPersist(persister.ReadRaftState())\n\n\trf.electionTimer = time.NewTimer(randElectionTimeout())\n\trf.appendEntriesTimers = make([]*time.Timer, len(rf.peers))\n\tfor i := range rf.peers {\n\t\trf.appendEntriesTimers[i] = time.NewTimer(HeartBeatTimeout)\n\t}\n\trf.applyTimer = time.NewTimer(ApplyInterval)\n\trf.notifyApplyCh = make(chan struct{}, 100)\n\n\t// apply log\n\tgo func() {\n\t\tfor !rf.killed() {\n\t\t\tselect {\n\t\t\tcase <-rf.applyTimer.C:\n\t\t\t\trf.notifyApplyCh <- struct{}{}\n\t\t\tcase <-rf.notifyApplyCh:\n\t\t\t\trf.startApplyLogs()\n\t\t\t}\n\t\t}\n\t}()\n\t// start election\n\tgo func() {\n\t\tfor !rf.killed() {\n\t\t\t<-rf.electionTimer.C\n\t\t\trf.startElection()\n\t\t}\n\t}()\n\t// leader send logs\n\tfor i := range peers {\n\t\tif i == rf.me {\n\t\t\tcontinue\n\t\t}\n\t\tgo func(index int) {\n\t\t\tfor !rf.killed() {\n\t\t\t\t<-rf.appendEntriesTimers[index].C\n\t\t\t\trf.sendAppendEntries(index)\n\t\t\t}\n\t\t}(i)\n\t}\n\n\treturn rf\n}", "func (rf *Raft) Start(command interface{}) (int, int, bool) {\n\tDPrintf(\"peer-%d ----------------------Start()-----------------------\", rf.me)\n\tindex := -1\n\tterm := -1\n\tisLeader := true\n\n\t// Your code here (2B).\n\t//term, isLeader = rf.GetState()\n\trf.mu.Lock()\n\tterm = rf.currentTerm\n\tif rf.state != Leader {\n\t\tisLeader = false\n\t}\n\tif isLeader {\n\t\t// Append the command into its own rf.log\n\t\tvar newlog LogEntry\n\t\tnewlog.Term = rf.currentTerm\n\t\tnewlog.Command = command\n\t\trf.log = append(rf.log, newlog)\n\t\trf.persist()\n\t\tindex = len(rf.log) // the 3rd return value.\n\t\trf.repCount[index] = 1\n\t\t// now the log entry is appended into leader's log.\n\t\trf.mu.Unlock()\n\n\t\t// start agreement and return immediately.\n\t\tfor peer_index, _ := range rf.peers {\n\t\t\tif peer_index == rf.me {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\t// send AppendEntries RPC to each peer. And decide when it is safe to apply a log entry to the state machine.\n\t\t\tgo func(i int) {\n\t\t\t\trf.mu.Lock()\n\t\t\t\tnextIndex_copy := make([]int, len(rf.peers))\n\t\t\t\tcopy(nextIndex_copy, rf.nextIndex)\n\t\t\t\trf.mu.Unlock()\n\t\t\t\tfor {\n\t\t\t\t\t// make a copy of current leader's state.\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t// we should not send RPC if rf.currentTerm != term, the log entry will be sent in later AE-RPCs in args.Entries.\n\t\t\t\t\tif rf.state != Leader || rf.currentTerm != term {\n\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\t\t\t\t\t// make a copy of leader's raft state.\n\t\t\t\t\tcommitIndex_copy := rf.commitIndex // during the agreement, commitIndex may increase.\n\t\t\t\t\tlog_copy := make([]LogEntry, len(rf.log)) // during the agreement, log could grow.\n\t\t\t\t\tcopy(log_copy, rf.log)\n\t\t\t\t\trf.mu.Unlock()\n\n\t\t\t\t\tvar args AppendEntriesArgs\n\t\t\t\t\tvar reply AppendEntriesReply\n\t\t\t\t\targs.Term = term\n\t\t\t\t\targs.LeaderId = rf.me\n\t\t\t\t\targs.LeaderCommit = commitIndex_copy\n\t\t\t\t\t// If last log index >= nextIndex for a follower: send AppendEntries RPC with log entries starting at nextIndex\n\t\t\t\t\t// NOTE: nextIndex is just a predication. not a precise value.\n\t\t\t\t\targs.PrevLogIndex = nextIndex_copy[i] - 1\n\t\t\t\t\tif args.PrevLogIndex > 0 {\n\t\t\t\t\t\t// FIXME: when will this case happen??\n\t\t\t\t\t\tif args.PrevLogIndex > len(log_copy) {\n\t\t\t\t\t\t\t// TDPrintf(\"adjust PrevLogIndex.\")\n\t\t\t\t\t\t\t//return\n\t\t\t\t\t\t\targs.PrevLogIndex = len(log_copy)\n\t\t\t\t\t\t}\n\t\t\t\t\t\targs.PrevLogTerm = log_copy[args.PrevLogIndex-1].Term\n\t\t\t\t\t}\n\t\t\t\t\targs.Entries = make([]LogEntry, len(log_copy)-args.PrevLogIndex)\n\t\t\t\t\tcopy(args.Entries, log_copy[args.PrevLogIndex:len(log_copy)])\n\t\t\t\t\tok := rf.sendAppendEntries(i, &args, &reply)\n\t\t\t\t\t// handle RPC reply in the same goroutine.\n\t\t\t\t\tif ok == true {\n\t\t\t\t\t\tif reply.Success == true {\n\t\t\t\t\t\t\t// this case means that the log entry is replicated successfully.\n\t\t\t\t\t\t\tDPrintf(\"peer-%d AppendEntries success!\", rf.me)\n\t\t\t\t\t\t\t// re-establish the assumption.\n\t\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\t\tif rf.state != Leader || rf.currentTerm != term {\n\t\t\t\t\t\t\t\t//Figure-8 and p-8~9: never commits log entries from previous terms by counting replicas!\n\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t// NOTE: TA's QA: nextIndex[i] should not decrease, so check and set.\n\t\t\t\t\t\t\tif index >= rf.nextIndex[i] {\n\t\t\t\t\t\t\t\trf.nextIndex[i] = index + 1\n\t\t\t\t\t\t\t\t// TA's QA\n\t\t\t\t\t\t\t\trf.matchIndex[i] = args.PrevLogIndex + len(args.Entries) // matchIndex is not used in my implementation.\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t// test whether we can update the leader's commitIndex.\n\t\t\t\t\t\t\trf.repCount[index]++\n\t\t\t\t\t\t\t// update leader's commitIndex! We can determine that Figure-8's case will not occur now,\n\t\t\t\t\t\t\t// because we have test rf.currentTerm == term_copy before, so we will never commit log entries from previous terms.\n\t\t\t\t\t\t\tif rf.commitIndex < index && rf.repCount[index] > len(rf.peers)/2 {\n\t\t\t\t\t\t\t\t// apply the command.\n\t\t\t\t\t\t\t\tDPrintf(\"peer-%d Leader moves its commitIndex from %d to %d.\", rf.me, rf.commitIndex, index)\n\t\t\t\t\t\t\t\t// NOTE: the Leader should commit one by one.\n\t\t\t\t\t\t\t\trf.commitIndex = index\n\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\t// now the command at commitIndex is committed.\n\t\t\t\t\t\t\t\tgo func() {\n\t\t\t\t\t\t\t\t\trf.canApplyCh <- true\n\t\t\t\t\t\t\t\t}()\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\treturn // jump out of the loop.\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t// AppendEntries RPC fails because of log inconsistency: Decrement nextIndex and retry\n\t\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\t\t// re-establish the assumption.\n\t\t\t\t\t\t\tif rf.state != Leader || rf.currentTerm != term {\n\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\tif reply.Term > rf.currentTerm {\n\t\t\t\t\t\t\t\trf.state = Follower\n\t\t\t\t\t\t\t\trf.currentTerm = reply.Term\n\t\t\t\t\t\t\t\trf.persist()\n\t\t\t\t\t\t\t\trf.resetElectionTimeout()\n\t\t\t\t\t\t\t\tDPrintf(\"peer-%d degenerate from Leader into Follower!!!\", rf.me)\n\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\trf.nonleaderCh <- true\n\t\t\t\t\t\t\t\t// don't try to send AppendEntries RPC to others then, rf is not the leader.\n\t\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t// NOTE: the nextIndex[i] should never < 1\n\t\t\t\t\t\t\t\tconflict_term := reply.ConflictTerm\n\t\t\t\t\t\t\t\tconflict_index := reply.ConflictIndex\n\t\t\t\t\t\t\t\t// refer to TA's guide blog.\n\t\t\t\t\t\t\t\t// first, try to find the first index of conflict_term in leader's log.\n\t\t\t\t\t\t\t\tfound := false\n\t\t\t\t\t\t\t\tnew_next_index := conflict_index // at least 1\n\t\t\t\t\t\t\t\tfor j := 0; j < len(rf.log); j++ {\n\t\t\t\t\t\t\t\t\tif rf.log[j].Term == conflict_term {\n\t\t\t\t\t\t\t\t\t\tfound = true\n\t\t\t\t\t\t\t\t\t} else if rf.log[j].Term > conflict_term {\n\t\t\t\t\t\t\t\t\t\tif found {\n\t\t\t\t\t\t\t\t\t\t\tnew_next_index = j + 1\n\t\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\tnextIndex_copy[i] = new_next_index\n\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\t// now retry to send AppendEntries RPC to peer-i.\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t} else {\n\t\t\t\t\t\t// RPC fails. Retry!\n\t\t\t\t\t\t// when network partition\n\t\t\t\t\t\ttime.Sleep(time.Millisecond * time.Duration(100))\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}(peer_index)\n\t\t}\n\t} else {\n\t\trf.mu.Unlock()\n\t}\n\n\treturn index, term, isLeader\n}", "func initServerFunctions() {\n NoErrors = serverError{nil, OK_CODE}\n Online = make(map[string]*chan message)\n // commands which server can execute\n ServerFunctions = make( map[string]feature )\n ServerFunctions[\"SIGN_UP\"] = feature{sign_up, 3, false}\n ServerFunctions[\"GET_TKN\"] = feature{sign_in, 1, false}\n ServerFunctions[\"QUIT\"] = feature{unlogin, 2, true}\n ServerFunctions[\"SEND_MSG\"] = feature{sendMsg, 4, true}\n ServerFunctions[\"GET_MSG\"] = feature{getNewMsg, 2, true}\n ServerFunctions[\"FIND_USR\"] = feature{findUsernames, 3, true}\n}", "func New(IP, firstContact, ID string, OutputTo io.Writer) *RaftMember {\n\t//try to get the existing data from file\n\tos.Mkdir(\"/tmp/raft/\", 0777)\n\tR, PStore, success := readPersist()\n\t//build the stateMachine\n\tpwd := \"/tmp\"\n\tos.Mkdir(pwd+\"/machines\", 0777) //make the machines directory\n\n\tR.hrt = new(time.Timer) //initialize our timer.\n\tR.giveStatus = make(chan StatusBall)\n\tR.outputTo = OutputTo\n\tR.needStatus = make(chan bool)\n\tR.giveLog = make(chan string)\n\tR.needLog = make(chan bool)\n\tR.nextIndex = make(map[string]int)\n\tR.matchIndex = make(map[string]int)\n\tR.appendRequests = make(chan string, 100)\n\tR.needScoreBoard = make(chan bool)\n\tR.suspend = make(chan int)\n\tR.GiveScoreboard = make(chan map[string]int)\n\tif !success {\n\t\t//load failed or files dont exist\n\t\tfmt.Println(\"Creating New Member\")\n\t\ttid, err := makeID()\n\t\tR.VoteLog = append(R.VoteLog, VoteRecord{Votes: make(map[string]bool), VotedFor: \"NIL\"}) //bootstrap the vote record for the 0 term\n\t\tterr.BreakError(\"ID Creation\", err)\n\t\tR.ID = tid\n\t\tR.Log = append(R.Log, LogEntry{Term: 0, Entry: \"init\"}) // make the first entry an initialize for counting purposes.\n\t\tR.Target = 1\n\t}\n\tR.Machine = stateMachine.CreateStateMachine(pwd+\"/machines/\"+R.ID[:6]+\".sm\", contentionLevel) //build or recreate the state machine.\n\tterr.VerbPrint(R.outputTo, 1, verb, \"Target Set to:\", R.Target)\n\tR.Machine.SetTarget(R.Target) //restores a target after a reboot\n\tif ID != \"\" {\n\t\tR.ID = ID\n\t}\n\n\tR.CM = CM.New(R.ID, IP, R.outputTo)\n\tif firstContact != \"\" {\n\t\tterr.VerbPrint(R.outputTo, 3, verb, \"connecting to\", firstContact)\n\t\tR.Connect(firstContact)\n\t}\n\tR.writePersist(PStore)\n\tR.et = timers.NewElectionTimer()\n\tR.run(PStore) //spawns a go routine to handle incomming messages\n\treturn R\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.n = len(peers)\n\trf.currentState = StateFollower\n\trf.currentTerm = 0\n\trf.votedFor = -1\n\trf.logs = make([]LogEntry, 1)\n\trf.logs[0] = LogEntry{\n\t\tTerm: 0,\n\t\tType: NoOp,\n\t\tCommand: nil,\n\t}\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\trf.resetElectionChan = make(chan int)\n\trf.numberOfGrantedVotes = 0\n\trf.nextIndex = make([]int, rf.n)\n\tfor i := range rf.nextIndex {\n\t\trf.nextIndex[i] = 1\n\t}\n\trf.matchIndex = make([]int, rf.n)\n\n\trf.appendEntriesReplyHandler = make(chan AppendEntriesReply)\n\trf.stopLeaderLogicHandler = make(chan int)\n\trf.requestVoteReplyHandler = make(chan RequestVoteReply)\n\n\t// Your initialization code here (2A, 2B, 2C).\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\t// start ticker goroutine to start elections\n\tgo rf.ticker()\n\n\tDPrintf(\"Raft node (%v) starts up...\\n\", me)\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {// {{{\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\t//rf.leaderId = -1\n\trf.cmtChan = make(chan bool, MAX_CHAN_DEPTH)\n\trf.applyChan = applyCh\n\t\n\t//initialization\n\trf.electTimer = time.NewTimer(TIMEOUT_MAX * time.Millisecond)\n\trf.state = FOLLOWER\n\trf.voteCnt = 0\n\trf.cmtIdx = 0\n\trf.applyIdx = 0\n\trf.nextIdx = make([]int, len(rf.peers)) //initialize when become leader\n\trf.matchIdx = make([]int, len(rf.peers))\n\tfor i := range rf.matchIdx {\n\t rf.matchIdx[i] = 0\n\t}\n\t// initialize from state persisted before a crash\n\trf.curTerm = 0\n\trf.votedFor = -1\n\trf.log = make([]LogEntry, 0)\n\trf.log = append(rf.log, LogEntry{Term: 0})\n\trf.readPersist(persister.ReadRaftState())\n\t\n\tgo rf.writeApplyCh()\n\tgo rf.run()\n\t\n\treturn rf\n}", "func (r *Raft) becomeCandidate() {\n\t// Your Code Here (2A).\n\tif _, ok := r.Prs[r.id]; !ok {\n\t\treturn\n\t}\n\tr.State = StateCandidate\n\tr.Term++\n\tr.Lead = None\n\tr.Vote = r.id\n\tr.votes = make(map[uint64]bool)\n\tr.votes[r.id] = true // 自己给自己投票\n\tr.actualElectionTimeout = rand.Intn(r.electionTimeout) + r.electionTimeout\n\tr.leadTransferee = None\n}", "func Make(peers []*labrpc.ClientEnd, me int,\r\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\r\n\trf := &Raft{}\r\n\trf.peers = peers\r\n\trf.persister = persister\r\n\trf.me = me\r\n\r\n\t// Your initialization code here.\r\n\trf.state = FOLLOWER\r\n\trf.votedFor = -1\r\n\trf.voteCh = make(chan struct{}) //define the v of the channal is the size of an empty struct{}\r\n\trf.appendCh = make(chan struct{})\r\n\t\r\n\t// initialize from state persisted before a crash\r\n\trf.readPersist(persister.ReadRaftState())\r\n\t\r\n\tgo rf.startLoop() // start election\r\n\r\n\r\n\treturn rf\r\n}", "func (rf *Raft) follower() {\n\tgo rf.startElectionTimer()\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.applyCh = applyCh\n\trf.candidateId = me\n\trf.votedFor = NILVOTE \n\trf.role = FOLLOWER\n\trf.currentTerm = 0\n\trf.resetTimeoutEvent = makeTimestamp()\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\tif(len(rf.log) == 0) {\n\t\tentry := Entry {0, nil, 0}\n\t\trf.log = append(rf.log, entry)\n\t\trf.persist()\n\t}\n\n\tAssert(len(rf.log) >= 1, \"raft log not bigger than 0\")\n\n\trf.nextIndex = make([]int, len(rf.peers))\n\trf.matchIndex = make([]int, len(rf.peers))\n\n\t// start ticker goroutine to start elections\n\tgo rf.ticker()\n\n\t// have the leader send heartbeats out periodically\n\tgo rf.sendHeartbeats()\n\treturn rf\n}", "func newRaft(c *Config) *Raft {\n\tif err := c.validate(); err != nil {\n\t\tpanic(err.Error())\n\t}\n\tpeer2Progress := make(map[uint64]*Progress, len(c.peers))\n\tpeer2Vote := make(map[uint64]bool, len(c.peers))\n\tfor _, s := range c.peers {\n\t\tpeer2Vote[s] = false\n\t\tpeer2Progress[s] = &Progress{0, 0}\n\t}\n\trand.Seed(time.Now().UnixNano())\n\thardState, _, _ := c.Storage.InitialState()\n\treturn &Raft{id: c.ID, Term: hardState.Term, Vote: hardState.Vote, RaftLog: newLog(c.Storage), State: StateFollower, Prs: peer2Progress, votes: peer2Vote, Lead: 0, heartbeatTimeout: c.HeartbeatTick, electionTimeout: c.ElectionTick, heartbeatElapsed: 0, electionElapsed: 0, actualElectionTimeout: 0}\n}", "func (rf *Raft) Start(command interface{}) (int, int, bool) {\n\t//Client 给复制状态机发送了一个command\n\trf.mu.Lock()\n\tindex := len(rf.log)\n\tterm := rf.currentTerm\n\tisLeader := true\n\n\t// Your code here (2B).\n\t//DPrintf(\"2B TEST: current server is %d, state is %s\", rf.me, rf.state)\n\tif rf.state != \"leader\" {\n\t\tisLeader = false\n\t}else{\n\t\t//if rf.log[0].Term == 0 && rf.log[0].Index == 0 && rf.log[0].Command == nil {\n\t\t//\trf.log[0].Term = rf.currentTerm\n\t\t//\trf.log[0].Command = command\n\t\t//}else{\n\t\t//\tindex = len(rf.log)\n\t\t//\tlogEntry := Entry{\n\t\t//\t\tTerm: rf.currentTerm,\n\t\t//\t\tIndex: index,\n\t\t//\t\tCommand: command,\n\t\t//\t}\n\t\t//\trf.log = append(rf.log, logEntry)\n\t\t//\trf.matchIndex[rf.me] = index\n\t\t//\tDPrintf(\"2B TEST: %d's log is %v; index is %d, term is %d, logEntry is %v\", rf.me, rf.log, index, term, logEntry)\n\t\t//}\n\t\tindex = len(rf.log)\n\t\tlogEntry := Entry{\n\t\t\tTerm: rf.currentTerm,\n\t\t\tIndex: index,\n\t\t\tCommand: command,\n\t\t}\n\t\trf.log = append(rf.log, logEntry)\n\t\trf.matchIndex[rf.me] = index\n\t\t//DPrintf(\"2B TEST: %d's log is %v; index is %d, term is %d, logEntry is %v\", rf.me, rf.log, index, term, logEntry)\n\t}\n\trf.resetHeartBeatTimers()\n\trf.mu.Unlock()\n\treturn index, term, isLeader\n}", "func (r *Raft) becomeFollower(term uint64, lead uint64) {\n\t// Your Code Here (2A).\n\tr.State = StateFollower\n\tr.Term = term\n\tr.Lead = lead\n\tr.Vote = r.Lead\n\tr.electionElapsed = 0\n\tr.actualElectionTimeout = rand.Intn(r.electionTimeout) + r.electionTimeout\n\tr.leadTransferee = None\n}", "func newRaft(c *Config) *Raft {\n\tif err := c.validate(); err != nil {\n\t\tpanic(err.Error())\n\t}\n\ts := c.Storage\n\thardStatus, confStatus, err := s.InitialState()\n\n\tif err != nil {\n\t\tpanic(err.Error())\n\t}\n\n\tif c.peers == nil {\n\t\tc.peers = confStatus.Nodes\n\t}\n\n\tnodes := c.peers\n\t// init vote to false\n\tvotes := make(map[uint64]bool)\n\tfor _, nodeId := range nodes {\n\t\tvotes[nodeId] = false\n\t}\n\n\treturn &Raft{\n\t\tid: c.ID,\n\t\tTerm: hardStatus.Commit,\n\t\tVote: hardStatus.Vote,\n\t\tRaftLog: newLog(c.Storage),\n\t\tPrs: nil,\n\t\t// init as a follower\n\t\tState: StateFollower,\n\t\tvotes: nil,\n\t\tmsgs: nil,\n\t\tLead: None,\n\t\theartbeatTimeout: c.HeartbeatTick,\n\t\telectionTimeout: c.ElectionTick,\n\t\trandomElectionTimeout: randomTimeout(c.ElectionTick),\n\t\theartbeatElapsed: 0,\n\t\telectionElapsed: 0,\n\t\tleadTransferee: 0,\n\t\tPendingConfIndex: 0,\n\t\tnodes: nodes,\n\t}\n}", "func (le *LeaderElector) adjustLeadership() {\n\t//Try to regain leadership if this is the server with highest rank\n\t//or discover server with highest rank\n\tle.Lock()\n\tle.adjustingLead = true\n\tle.CurrentLeader = \"\"\n\tle.LeaderSID = NO_LEADER\n\tle.Unlock()\n\t<-time.After(le.RegainLeadFreq * time.Second)\n\tdebug(\"[*] Info : LeaderElector : Adjusting LeaderShip Election Started.\")\n\tle.initElection()\n\tle.Lock()\n\tle.adjustingLead = false\n\tle.Unlock()\n}", "func initCallhome(ctx context.Context, objAPI ObjectLayer) {\n\tif !globalCallhomeConfig.Enabled() {\n\t\treturn\n\t}\n\n\tgo func() {\n\t\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\t\t// Leader node (that successfully acquires the lock inside runCallhome)\n\t\t// will keep performing the callhome. If the leader goes down for some reason,\n\t\t// the lock will be released and another node will acquire it and take over\n\t\t// because of this loop.\n\t\tfor {\n\t\t\tif !globalCallhomeConfig.Enabled() {\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif !runCallhome(ctx, objAPI) {\n\t\t\t\t// callhome was disabled or context was canceled\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// callhome running on a different node.\n\t\t\t// sleep for some time and try again.\n\t\t\tduration := time.Duration(r.Float64() * float64(globalCallhomeConfig.FrequencyDur()))\n\t\t\tif duration < time.Second {\n\t\t\t\t// Make sure to sleep atleast a second to avoid high CPU ticks.\n\t\t\t\tduration = time.Second\n\t\t\t}\n\t\t\ttime.Sleep(duration)\n\t\t}\n\t}()\n}", "func (rf *Raft) GetState() (int, bool) {\n\n var term int\n var isleader bool\n\n rf.mu.Lock()\n defer rf.mu.Unlock()\n term = rf.currentTerm\n isleader = false\n if rf.state == Leader {\n isleader = true\n } \n\n // Your code here (2A).\n return term, isleader\n}", "func (rf *Raft) reset(term int) {\n\trf.term = term\n\tfor idx := range rf.votes {\n\t\trf.votes[idx] = -1\n\t}\n\trf.lastHeartBeat = time.Now()\n\trf.lastElection = time.Now()\n\trf.vote = -1\n}", "func becomeLeader(target Node) error {\n\tlog.Info(\"Becoming the leader of \", target)\n\tmyFrame.CurrentLeader = me\n\tmyFrame.LastUpdated = time.Now()\n\tb, err := json.Marshal(&myFrame)\n\tif err != nil {\n\t\tlog.WithError(err).Error(\"Failed marshaling intro\")\n\t\treturn err\n\t}\n\n\tresp, err := http.Post(\"http://\"+target.Address+\"/lead\", \"application/json\", bytes.NewBuffer(b))\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn err\n\t}\n\n\tif resp.StatusCode != http.StatusOK {\n\t\treturn errors.New(\"Introduction gave non-200 status back\")\n\t}\n\treturn nil\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\tn := len(peers)\n\trf.App = make([] chan bool, n)\n\tfor i := range rf.App{\n\t\trf.App[i] = make(chan bool, 1)\n\t\trf.App[i] <- true\n\t}\n\trf.currentTerm = 0\n\trf.votedFor = -1\n\trf.leader = -1\n\trf.role = 0\n\trf.log = append(rf.log, LogEntry{Term:0, Index:0})\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\trf.nextIndex = make([]int, n)\n\trf.matchIndex = make([]int, n)\n\trf.Heartbeat = make(chan bool)\n\trf.GrantVote = make(chan bool)\n\trf.roleChan = make(chan int)\n\trf.applyCh = applyCh\n\t\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\trf.readSnapshot(persister.ReadSnapshot())\n\trf.randgene()\n\tgo rf.changestatus()\n\tgo rf.timeoutTimer()\n\n\treturn rf\n}", "func init() {\n\trkmDB[\"fweuler\"] = func() rkmethod { return new(FwEuler) }\n}", "func (r *Raft) becomeCandidate() {\n\t// Your Code Here (2A).\n\tr.State = StateCandidate\n\tr.Lead = None\n\tr.Term++\n\tr.Vote = r.id\n\tr.votes = make(map[uint64]bool)\n\tr.votes[r.id] = true\n}", "func (rf *Raft) convertToLeader() {\n\trf.state = Leader\n\tDPrintf(\"peer-%d becomes the new leader!!!\", rf.me)\n\t// when a leader first comes to power, it initializes all nextIndex values to the index just after the last one in its log. (Section 5.3)\n\trf.nextIndex = make([]int, len(rf.peers))\n\trf.matchIndex = make([]int, len(rf.peers))\n\tnext_index_initval := len(rf.log) + 1\n\tfor i := 0; i < len(rf.peers); i++ {\n\t\trf.nextIndex[i] = next_index_initval\n\t\trf.matchIndex[i] = 0 // increases monotonically.\n\t}\n\trf.repCount = make(map[int]int)\n\tDPrintf(\"peer-%d Leader's log array's length = %d.\", rf.me, len(rf.log))\n\trf.leaderCh <- true\n}", "func (rf *Raft) ticker() {\n\n\ttime.Sleep(time.Duration(1000000 * rand.Intn(150)))\n\n\tfor !rf.killed() {\n\n\t\t// Your code here to check if a leader election should\n\t\t// be started and to randomize sleeping time using\n\t\t// time.Sleep().\n\n\t\trf.mu.Lock()\n\t\tvar sleepTime time.Duration\n\t\tisLeader := rf.IsLeader()\n\t\tif isLeader {\n\t\t\t// rf.checkCommit()\n\t\t\t// rf.debugCommit()\n\t\t\trf.canSend = false\n\t\t\trf.broadcast() // heartbeat all followers\n\t\t\tsleepTime = time.Millisecond * 100 // fixed time, 10 times per second max\n\t\t} else {\n\t\t\tif !rf.gotContacted {\n\t\t\t\t// start election\n\t\t\t\trf.currentTerm += 1\n\t\t\t\trf.votedFor = rf.me\n\t\t\t\tatomic.StoreInt32(&rf.state, CANDIDATE)\n\t\t\t\t// rf.state = CANDIDATE\n\n\t\t\t\t// reset vote count\n\t\t\t\tfor i := range rf.voteCount {\n\t\t\t\t\trf.voteCount[i] = false\n\t\t\t\t}\n\t\t\t\trf.voteCount[rf.me] = true // vote for itself\n\n\t\t\t\trf.debug(\"Failed to be contacted, initiating election in term %v\\n\", rf.currentTerm)\n\n\t\t\t\t// request vote from all others\n\t\t\t\tfor i := range rf.peers {\n\t\t\t\t\tif i != rf.me {\n\t\t\t\t\t\tgo rf.candidateNotify(i, rf.currentTerm, rf.me, rf.lastEntryIndex(), rf.index(rf.lastEntryIndex()).Term)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tsleepTime = time.Duration(1000000 * (100 + rand.Intn(200))) // 100-300ms\n\t\t\t} else {\n\t\t\t\trf.debug(\"Got contacted and will sleep again...\\n\")\n\t\t\t\tsleepTime = time.Duration(1000000 * (600 + rand.Intn(350))) // 600-950ms\n\t\t\t}\n\t\t\trf.gotContacted = false\n\t\t}\n\n\t\trf.mu.Unlock()\n\n\t\ttime.Sleep(sleepTime)\n\t}\n}", "func InitReviewer() {\n\n}", "func leaderReset() time.Duration {\n\treturn time.Duration(time.Duration(50) * time.Millisecond)\n\n}", "func (n *Node) initFTable(alone bool) {\n\tfor i := 0; i < KeySize; i++ {\n\t\tn.fingers[i].start = n.ID.PowMod(int64(i), int64(KeySize))\n\t\tif alone {\n\t\t\tn.fingers[i].node = n.Rnode\n\t\t}\n\t}\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.mu = sync.Mutex{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\n\trf.currentTerm = 0\n\trf.votedFor = -1\n\trf.log = []LogEntry{}\n\trf.committedIndex = -1\n\trf.lastApplied = -1\n\n\trf.state = Follower\n\trf.applyCh = applyCh\n\n\trf.grantVoteCh = make(chan bool)\n\trf.heartBeatCh = make(chan bool)\n\trf.leaderCh = make(chan bool)\n\n\trf.timer = time.NewTimer(time.Duration(rand.Intn(300)+300) * time.Millisecond)\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\tgo rf.heartBeat()\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\trf.currentTerm = 0\n\trf.votedFor = -1\n\trf.commitIndex = -1\n\trf.lastApplied = -1\n\trf.leaderId = -1\n\n\tfor i := 0; i < len(peers); i++ {\n\t\trf.nextIndex = append(rf.nextIndex, -1)\n\t\trf.matchIndex = append(rf.matchIndex, -1)\n\t}\n\n\trf.role = FOLLOWER\n\tgo rf.electionTimer()\n\n\t// Your initialization code here (2A, 2B, 2C).\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\treturn rf\n}", "func (rf *Raft) GetState() (int, bool) {\n\n var term int\n var isleader bool\n // Your code here (2A).\n rf.mu.Lock()\n defer rf.mu.Unlock()\n term = rf.currentTerm\n isleader = false\n if rf.state == LEADER {\n isleader = true\n }\n return term, isleader\n}", "func init() {\n\tgo webhook.ProcessRouteStatus(controller)\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\trf.ConvertTo(Follower)\n\trf.CurrentTerm = 0\n\trf.Log = []LogEntry{}\n\n\trf.CommitIndex = 0\n\trf.LastApplied = 0\n\n\trf.NextIndex = []int{}\n\trf.MatchIndex = []int{}\n\t// Your initialization code here (2A, 2B, 2C).\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\tgo rf.ticker()\n\n\t// 每个结点应当检查自己的状态,\n\t// 如果是 leader 的话,就向其他节点发送心跳包\n\tgo rf.sendHeartBeats()\n\n\treturn rf\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\t// Your initialization code here (2A, 2B, 2C).\n\t// create a channel in Raft\n\trf.applyCh = applyCh\n\trf.state = Follower\n\trf.nonleaderCh = make(chan bool, 20)\n\trf.leaderCh = make(chan bool, 20)\n\trf.canApplyCh = make(chan bool, 20)\n\t// set election timeout\n\trf.voteCount = 0\n\trf.resetElectionTimeout()\n\n\t// Initialize volatile state on all servers.\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\trf.log = make([]LogEntry, 0)\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\t// seperate goroutine to apply command to statemachine.\n\tgo func() {\n\t\tfor {\n\t\t\t<-rf.canApplyCh\n\t\t\t// apply\n\t\t\trf.mu.Lock()\n\t\t\tcommitIndex_copy := rf.commitIndex\n\t\t\tlastApplied_copy := rf.lastApplied\n\t\t\tlog_copy := make([]LogEntry, len(rf.log))\n\t\t\tcopy(log_copy, rf.log)\n\t\t\trf.mu.Unlock()\n\t\t\tfor curr_index := lastApplied_copy + 1; curr_index <= commitIndex_copy; curr_index++ {\n\t\t\t\tDPrintf(\"peer-%d apply command-%d at index-%d.\", rf.me, log_copy[curr_index-1].Command.(int), curr_index)\n\t\t\t\tvar curr_command ApplyMsg\n\t\t\t\tcurr_command.CommandValid = true\n\t\t\t\tcurr_command.Command = log_copy[curr_index-1].Command\n\t\t\t\tcurr_command.CommandIndex = curr_index\n\t\t\t\trf.applyCh <- curr_command\n\t\t\t\trf.lastApplied = curr_index\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Leader's heartbeat long-running goroutine.\n\tgo func() {\n\t\tfor {\n\t\t\tif rf.state == Leader {\n\t\t\t\t// send heartbeats\n\t\t\t\trf.broadcastHeartbeats()\n\t\t\t\ttime.Sleep(time.Millisecond * time.Duration(100)) // 100ms per heartbeat. (heartbeat time interval << election timeout)\n\t\t\t} else {\n\t\t\t\t// block until be elected as the new leader.\n\t\t\t\tDPrintf(\"peer-%d leader's heartbeat long-running goroutine. block.\", rf.me)\n\t\t\t\t<-rf.leaderCh\n\t\t\t\tDPrintf(\"peer-%d leader's heartbeat long-running goroutine. get up.\", rf.me)\n\t\t\t}\n\t\t}\n\t}()\n\n\t// Nonleader's election timeout long-running goroutine.\n\tgo func() {\n\t\tfor {\n\t\t\t// check rf.state == Follower\n\t\t\tif rf.state != Leader {\n\t\t\t\t// begin tic-toc\n\t\t\t\ttime.Sleep(time.Millisecond * time.Duration(10))\n\t\t\t\tif rf.electionTimeout() {\n\t\t\t\t\tDPrintf(\"peer-%d kicks off an election!\\n\", rf.me)\n\t\t\t\t\t// election timeout! kick off an election.\n\t\t\t\t\t// convertion to a Candidate.\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\tDPrintf(\"peer-%d becomes a Candidate!!!\\n\", rf.me)\n\t\t\t\t\trf.state = Candidate\n\t\t\t\t\trf.currentTerm += 1\n\t\t\t\t\trf.persist()\n\t\t\t\t\tterm_copy := rf.currentTerm // create a copy of the term and it'll be used in RequestVote RPC.\n\t\t\t\t\t// vote for itself.\n\t\t\t\t\trf.voteCount = 1\n\t\t\t\t\trf.resetElectionTimeout()\n\t\t\t\t\t// send RequestVote RPCs to all other peers in seperate goroutines.\n\t\t\t\t\tlast_log_index_copy := len(rf.log)\n\t\t\t\t\tlast_log_term_copy := -1\n\t\t\t\t\tif last_log_index_copy > 0 {\n\t\t\t\t\t\tlast_log_term_copy = rf.log[last_log_index_copy-1].Term\n\t\t\t\t\t}\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\tfor peer_index, _ := range rf.peers {\n\t\t\t\t\t\tif peer_index == rf.me {\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t\t// create goroutine.\n\t\t\t\t\t\tgo func(i int) {\n\t\t\t\t\t\t\t// use a copy of the state of the rf peer\n\t\t\t\t\t\t\tvar args RequestVoteArgs\n\t\t\t\t\t\t\targs.Term = term_copy\n\t\t\t\t\t\t\targs.CandidateId = rf.me\n\t\t\t\t\t\t\targs.LastLogIndex = last_log_index_copy\n\t\t\t\t\t\t\targs.LastLogTerm = last_log_term_copy\n\t\t\t\t\t\t\tvar reply RequestVoteReply\n\t\t\t\t\t\t\tDPrintf(\"peer-%d send a sendRequestVote RPC to peer-%d\", rf.me, i)\n\t\t\t\t\t\t\t// reduce RPCs....\n\t\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\t\tif rf.state != Candidate || rf.currentTerm != term_copy {\n\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\treturn\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\tok := rf.sendRequestVote(i, &args, &reply)\n\t\t\t\t\t\t\t// handle the RPC reply in the same goroutine.\n\t\t\t\t\t\t\tif ok == true {\n\t\t\t\t\t\t\t\tif reply.VoteGranted == true {\n\t\t\t\t\t\t\t\t\t// whether the peer is still a Candidate and the previous term? if yes, increase rf.voteCount; if no, ignore.\n\t\t\t\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\t\t\t\t// re-establish the assumption.\n\t\t\t\t\t\t\t\t\tif rf.state == Candidate && term_copy == rf.currentTerm {\n\t\t\t\t\t\t\t\t\t\trf.voteCount += 1\n\t\t\t\t\t\t\t\t\t\tDPrintf(\"peer-%d gets a vote!\", rf.me)\n\t\t\t\t\t\t\t\t\t\tif rf.voteCount > len(rf.peers)/2 {\n\t\t\t\t\t\t\t\t\t\t\trf.convertToLeader()\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\t\t\t\t\t// re-establish the assumption.\n\t\t\t\t\t\t\t\t\tif rf.state == Candidate && term_copy == rf.currentTerm {\n\t\t\t\t\t\t\t\t\t\tif reply.Term > rf.currentTerm {\n\t\t\t\t\t\t\t\t\t\t\trf.state = Follower\n\t\t\t\t\t\t\t\t\t\t\trf.currentTerm = reply.Term\n\t\t\t\t\t\t\t\t\t\t\trf.persist()\n\t\t\t\t\t\t\t\t\t\t\trf.voteCount = 0\n\t\t\t\t\t\t\t\t\t\t\tDPrintf(\"peer-%d calm down from a Candidate to a Follower!!!\", rf.me)\n\t\t\t\t\t\t\t\t\t\t\trf.resetElectionTimeout()\n\t\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\t\t\t\t}\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}(peer_index)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\t// block until become a Follower or Candidate.\n\t\t\t\tDPrintf(\"peer-%d non-leader's election timeout long-running goroutine. block.\", rf.me)\n\t\t\t\t<-rf.nonleaderCh\n\t\t\t\tDPrintf(\"peer-%d non-leader's election timeout long-running goroutine. get up.\", rf.me)\n\t\t\t\trf.resetElectionTimeout()\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func (r *Raft) runCandidate() {\n\t// Start vote for us, and set a timeout\n\tvoteCh := r.electSelf()\n\telectionTimeout := randomTimeout(r.conf.ElectionTimeout, 2*r.conf.ElectionTimeout)\n\n\t// Tally the votes, need a simple majority\n\tgrantedVotes := 0\n\tquorum := r.quorumSize()\n\tr.logD.Printf(\"Cluster size: %d, votes needed: %d\", len(r.peers)+1, quorum)\n\n\ttransition := false\n\tfor !transition {\n\t\tselect {\n\t\tcase rpc := <-r.rpcCh:\n\t\t\tswitch cmd := rpc.Command.(type) {\n\t\t\tcase *AppendEntriesRequest:\n\t\t\t\ttransition = r.appendEntries(rpc, cmd)\n\t\t\tcase *RequestVoteRequest:\n\t\t\t\ttransition = r.requestVote(rpc, cmd)\n\t\t\tdefault:\n\t\t\t\tr.logE.Printf(\"Candidate state, got unexpected command: %#v\",\n\t\t\t\t\trpc.Command)\n\t\t\t\trpc.Respond(nil, fmt.Errorf(\"unexpected command\"))\n\t\t\t}\n\n\t\t// Got response from peers on voting request\n\t\tcase vote := <-voteCh:\n\t\t\t// Check if the term is greater than ours, bail\n\t\t\tif vote.Term > r.getCurrentTerm() {\n\t\t\t\tr.logD.Printf(\"Newer term discovered\")\n\t\t\t\tr.setState(Follower)\n\t\t\t\tif err := r.setCurrentTerm(vote.Term); err != nil {\n\t\t\t\t\tr.logE.Printf(\"Failed to update current term: %w\", err)\n\t\t\t\t}\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\t// Check if the vote is granted\n\t\t\tif vote.Granted {\n\t\t\t\tgrantedVotes++\n\t\t\t\tr.logD.Printf(\"Vote granted. Tally: %d\", grantedVotes)\n\t\t\t}\n\n\t\t\t// Check if we've become the leader\n\t\t\tif grantedVotes >= quorum {\n\t\t\t\tr.logD.Printf(\"Election won. Tally: %d\", grantedVotes)\n\t\t\t\tr.setState(Leader)\n\t\t\t\treturn\n\t\t\t}\n\t\tcase a := <-r.applyCh:\n\t\t\t// Reject any operations since we are not the leader\n\t\t\ta.response = ErrNotLeader\n\t\t\ta.Response()\n\n\t\tcase <-electionTimeout:\n\t\t\t// Election failed! Restart the election. We simply return,\n\t\t\t// which will kick us back into runCandidate\n\t\t\tr.logW.Printf(\"Election timeout reached, restarting election\")\n\t\t\treturn\n\n\t\tcase <-r.shutdownCh:\n\t\t\treturn\n\t\t}\n\t}\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.applyCh = applyCh\n\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.isLeader = false\n\trf.VotedFor = -1\n\n\trf.nextIndex = make([]int, len(peers))\n\trf.matchIndex = make([]int, len(peers))\n\n\trf.Logs = make([]LogEntry, 1)\n\trf.Logs[0] = LogEntry{\n\t\tTerm:0,\n\t\tCommand:nil,\n\t}\n\trf.shutdown = make(chan struct{})\n\tfor i := 0; i < len(peers); i++ {\n\t\trf.nextIndex[i] = len(rf.Logs)\n\t}\n\trf.resetTimer = make(chan struct{})\n\trf.commitCond = sync.NewCond(&rf.mu)\n\trf.newEntryCond = make([]*sync.Cond, len(peers))\n\tfor i := 0; i < len(peers); i++ {\n\t\trf.newEntryCond[i] = sync.NewCond(&rf.mu)\n\t}\n\trf.electionTimeout = time.Millisecond * (400 +\n\t\ttime.Duration(rand.Int63() % 400))\n\trf.electionTimer = time.NewTimer(rf.electionTimeout)\n\n\trf.hearBeatInterval = time.Millisecond * 200\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\tgo rf.electionDaemon()\n\n\tgo rf.applyEntryDaemon()\n\treturn rf\n}", "func main() {\n\tcountlog.Info(\"Start raft demo\")\n\ts1 := server.NewRaftServerWithEnv(core.NewServerConf(clusters[1], leader, clusters), env)\n\ts2 := server.NewRaftServerWithEnv(core.NewServerConf(clusters[2], leader, clusters), env)\n\ts3 := server.NewRaftServerWithEnv(core.NewServerConf(clusters[3], leader, clusters), env)\n\ts4 := server.NewRaftServerWithEnv(core.NewServerConf(clusters[4], leader, clusters), env)\n\tl1 := server.NewRaftServerWithEnv(core.NewServerConf(leader, leader, clusters), env)\n\tgo s1.Start()\n\tgo s2.Start()\n\tgo s3.Start()\n\tgo s4.Start()\n\tl1.Start()\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.leader = -1\n\trf.hasLeader = false\n\n\t// Your initialization code here (2A, 2B, 2C).\n\n\trf.log = []logItem{logItem{-1, 0}} // fill in an empty log\n\n\trf.votedTerms = map[int]int{}\n\n\t// init heartbeat channel\n\trf.stopHeartbeatCh = make(chan bool)\n\n\trf.stopBroadcastLogCh = make(chan bool)\n\n\t// start election clock\n\tgo rf.startElectionTimer()\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\trf.applyCh = applyCh\n\n\treturn rf\n}", "func (r *Raft) tick() {\n\t// Your Code Here (2A).\n\tr.electionElapsed++\n\tif r.electionElapsed >= r.actualElectionTimeout {\n\t\tr.electionElapsed = 0\n\t\t// 非 leader 一段时间没有收到 append/heartbeat 便发起选举\n\t\tif r.State != StateLeader {\n\t\t\tr.campaign()\n\t\t}\n\t}\n\tr.leaderTransferTimeout++\n\tif r.leaderTransferTimeout >= r.actualElectionTimeout {\n\t\tr.leaderTransferTimeout = 0\n\t\tr.leadTransferee = 0\n\t}\n\tif r.State == StateLeader {\n\t\tr.heartbeatElapsed++\n\t\tif r.heartbeatElapsed >= r.heartbeatTimeout {\n\t\t\tr.heartbeatElapsed = 0\n\t\t\tr.broadcastHeartbeat()\n\t\t}\n\t}\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.mu = sync.Mutex{}\n\n\t// Your initialization code here (2A, 2B, 2C).\n\tstate := Follower\n\trf.state = &state\n\trf.currentTerm = 0\n\trf.votedFor = \"\"\n\t// init the log entries\n\trf.logEntries = []LogEntry{\n\t\t{\n\t\t\tIndex: 0,\n\t\t\tTerm: 0,\n\t\t},\n\t}\n\trf.heartBeatCh = make(chan bool)\n\trf.grantCh = make(chan bool)\n\trf.heartBeatInterval = time.Millisecond * 150\n\trf.electionTimeout = 2 * rf.heartBeatInterval\n\trf.electAsLeaderCh = make(chan bool)\n\trf.peerId = strconv.Itoa(rf.me)\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\trf.applyCh = applyCh\n\n\tgo rf.applyEntries()\n\n\t// start ticker goroutine to start elections\n\tgo rf.runServer()\n\n\treturn rf\n}", "func (rf *Raft) convertToCandidate() {\n rf.mu.Lock()\n DLCPrintf(\"Server (%d)[state=%s, term=%d, votedFor=%d] convert to Candidate\", rf.me, rf.state, rf.currentTerm, rf.votedFor) \n rf.state = \"Candidate\"\n rf.currentTerm++\n rf.votedFor = rf.me\n rf.electionTime = generateElectionTime()\n rf.electionTimer.Reset(time.Duration(rf.electionTime) * time.Millisecond)\n rf.persist()\n DLCPrintf(\"Server (%d)[state=%s, term=%d, votedFor=%d, electionTime=%d] start request votes\", rf.me, rf.state, rf.currentTerm, rf.votedFor, rf.electionTime) \n rf.mu.Unlock()\n\n // 启动一个线程, requestVote\n go rf.requestForVotes()\n\n}", "func init() {\n\tparser.SharedParser().RegisterFabric(UpdateFabric{})\n}", "func (r *Raft) becomeCandidate() {\n\tr.State = StateCandidate\n\tr.Term += 1\n\tr.votes = make(map[uint64]bool)\n\t// vote for self\n\tr.votes[r.id] = true\n\tr.Vote = r.id\n\tr.electionElapsed = 0\n\tr.randomElectionTimeout = randomTimeout(r.electionTimeout)\n\tr.voteFailCount = 0\n\n\t// Your Code Here (2A).\n}", "func (rf *Raft) electionService() {\n\tfor {\n\t\trf.mu.Lock()\n\t\t// snapshot current state of raft, (state & term)\n\t\tcurrentState := rf.state\n\t\tcurrentTerm := rf.currentTerm\n\t\trf.mu.Unlock()\n\t\tswitch currentState {\n\t\tcase Follower:\n\t\t\t// clear raft state.\n\t\t\tselect {\n\t\t\tcase <-time.After(rf.electionTimeout + time.Duration(rand.Intn(500))):\n\t\t\t\tDPrintf(\"Peer-%d's election is timeout.\", rf.me)\n\t\t\t\trf.mu.Lock()\n\t\t\t\tDPrintf(\"Peer-%d's election hold the lock, now the currtentTerm=%d, rf.currentTerm=%d.\", rf.me, currentTerm, rf.currentTerm)\n\t\t\t\t// we should record the currentTerm for which the timer wait.\n\t\t\t\tif rf.state == Follower && rf.currentTerm == currentTerm {\n\t\t\t\t\trf.transitionState(Timeout)\n\t\t\t\t\tDPrintf(\"Peer-%d LSM has set state to candidate.\", rf.me)\n\t\t\t\t}\n\t\t\t\trf.mu.Unlock()\n\t\t\t\tDPrintf(\"Peer-%d turn state from %v to %v.\", rf.me, currentState, rf.state)\n\t\t\tcase event := <-rf.eventChan:\n\t\t\t\tswitch event {\n\t\t\t\tcase HeartBeat:\n\t\t\t\t\tDPrintf(\"Peer-%d Received heartbeat from leader, reset timer.\", rf.me)\n\t\t\t\tcase NewTerm:\n\t\t\t\t\tif rf.currentTerm == currentTerm {\n\t\t\t\t\t\tDPrintf(\"Peer-%d, waring: it received a NewTerm event, but term is not changed.\", rf.me)\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\tcase Candidate:\n\t\t\t// start a election.\n\t\t\t// using a thread to do election, and sending message to channel;\n\t\t\t// this can let the heartbeat to break the election progress.\n\t\t\tvoteDoneChan := make(chan bool)\n\t\t\tgo func() {\n\t\t\t\tTPrintf(\"Peer-%d becomes candidate, try to hold the lock.\", rf.me)\n\t\t\t\trf.mu.Lock()\n\t\t\t\tTPrintf(\"Peer-%d becomes candidate, has hold the lock, rf.voteFor=%d.\", rf.me, rf.voteFor)\n\t\t\t\t// check state first, if the state has changed, do not vote.\n\t\t\t\t// then check voteFor, if it has voteFor other peer in this term.\n\t\t\t\ttoVote := rf.state == Candidate && (rf.voteFor == -1 || rf.voteFor == rf.me)\n\t\t\t\tif toVote {\n\t\t\t\t\trf.voteFor = rf.me // should mark its voteFor when it begins to vote.\n\t\t\t\t\tDPrintf(\"Peer-%d set voteFor=%d.\", rf.me, rf.voteFor)\n\t\t\t\t}\n\t\t\t\trf.mu.Unlock()\n\t\t\t\tok := false\n\t\t\t\tif toVote {\n\t\t\t\t\tDPrintf(\"Peer-%d begin to vote.\", rf.me)\n\t\t\t\t\trequest := rf.createVoteRequest()\n\t\t\t\t\t// the process logic for each peer.\n\t\t\t\t\tprocess := func(server int) bool {\n\t\t\t\t\t\treply := new(RequestVoteReply)\n\t\t\t\t\t\trf.sendRequestVote(server, request, reply)\n\t\t\t\t\t\tok := rf.processVoteReply(reply)\n\t\t\t\t\t\treturn ok\n\t\t\t\t\t}\n\t\t\t\t\tok = rf.agreeWithServers(process)\n\t\t\t\t}\n\t\t\t\tvoteDoneChan <- ok\n\t\t\t}()\n\t\t\tselect {\n\t\t\tcase done := <-voteDoneChan:\n\t\t\t\tif done {\n\t\t\t\t\tDPrintf(\"Peer-%d win.\", rf.me)\n\t\t\t\t\t// if voting is success, we set state to leader.\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\tif rf.state == Candidate {\n\t\t\t\t\t\trf.transitionState(Win)\n\t\t\t\t\t}\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\tDPrintf(\"Peer-%d becomes the leader.\", rf.me)\n\t\t\t\t} else {\n\t\t\t\t\t// if voting is failed, we reset voteFor to -1, but do not reset state and term.\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\tif rf.state == Candidate {\n\t\t\t\t\t\trf.transitionState(Timeout)\n\t\t\t\t\t}\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\tsleep(rand.Intn(500))\n\t\t\t\t}\n\t\t\tcase event := <-rf.eventChan:\n\t\t\t\tswitch event {\n\t\t\t\tcase HeartBeat:\n\t\t\t\t\t// if another is win, we will receive heartbeat, so we shoul\n\t\t\t\t\tDPrintf(\"Peer-%d received heartbeat when voting, turn to follower, reset timer.\", rf.me)\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\trf.transitionState(NewLeader)\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\tcase NewTerm:\n\t\t\t\t\tDPrintf(\"Peer-%d received higher term when voting, stop waiting.\", rf.me)\n\t\t\t\t}\n\t\t\t}\n\t\tcase Leader:\n\t\t\t// start to send heartbeat.\n\t\t\tDPrintf(\"Peer-%d try to send heartbeat.\", rf.me)\n\t\t\trf.sendHeartbeat()\n\t\t\ttime.Sleep(rf.heartbeatInterval)\n\t\tcase End:\n\t\t\tDPrintf(\"Peer-%d is stopping.\\n\", rf.me)\n\t\t\treturn\n\t\tdefault:\n\t\t\tDPrintf(\"Do not support state: %v\\n\", currentState)\n\t\t}\n\t}\n}", "func (node *Node) init() {\n\n\t//step0. parse config\n\n\tif err := node.parseConfig(node.configFile); err != nil {\n\t\t//temp code\n\t\tlog.Panicf(\"Parse Config Error: %s\", err.Error())\n\t\treturn\n\t}\n\n\t//step1. init process runtime and update node state to NodeStateInit\n\n\t//step2. try to connecting other nodes and store the connection to internalConns\n\n\t//init finnal. update node state to NodeStateNormal\n\n}", "func init() {\n\tstart := time.Now()\n\tSetEnforcer(casbinPolicy, casbinModel)\n\tfmt.Printf(\"Took %v to load model and policy\", time.Since(start))\n}", "func (r *Raft) follower(timeout int) int {\n\t//myId := r.Myconfig.Id\n\t//fmt.Println(\"In follower()\", myId)\n\t//start heartbeat timer,timeout func wil place HeartbeatTimeout on channel\n\twaitTime := timeout //use random number after func is tested--PENDING\n\tHeartBeatTimer := r.StartTimer(HeartbeatTimeout, waitTime)\n\n\tfor {\n\t\treq := r.receive()\n\t\tswitch req.(type) {\n\t\tcase AppendEntriesReq:\n\t\t\trequest := req.(AppendEntriesReq) //explicit typecasting\n\t\t\tr.serviceAppendEntriesReq(request, HeartBeatTimer, waitTime)\n\t\tcase RequestVote:\n\t\t\twaitTime_secs := secs * time.Duration(waitTime)\n\t\t\trequest := req.(RequestVote)\n\t\t\t//fmt.Println(\"Requestvote came to\", myId, \"from\", request.candidateId)\n\t\t\tHeartBeatTimer.Reset(waitTime_secs)\n\t\t\t//fmt.Println(\"Timer reset to:\", waitTime_secs)\n\t\t\tr.serviceRequestVote(request)\n\t\tcase ClientAppendReq: //follower can't handle clients and redirects to leader, sends upto commitCh as well as clientCh\n\t\t\t//fmt.Println(\"in client append\")\n\t\t\trequest := req.(ClientAppendReq) //explicit typecasting\n\t\t\tresponse := ClientAppendResponse{}\n\t\t\tlogItem := LogItem{r.CurrentLogEntryCnt, false, request.data} //lsn is count started from 0\n\t\t\tr.CurrentLogEntryCnt += 1\n\t\t\tresponse.logEntry = logItem\n\t\t\tr.commitCh <- &response.logEntry\n\t\t\t//var e error = ErrRedirect(r.LeaderConfig.Id)\n\t\t\t//response.ret_error = e\n\t\t\t//r.clientCh <- response //respond to client giving the leader Id--Should only be given leader id right?--CHECK\n\t\tcase int:\n\t\t\t//fmt.Println(\"In follower timeout\", r.Myconfig.Id, time.Now().Format(layout))\n\t\t\tHeartBeatTimer.Stop() //turn off timer as now election timer will start in candidate() mode\n\t\t\treturn candidate\n\t\t}\n\t}\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\n\trf.currentTerm = 0\n\trf.votedFor = NULL\n\trf.logEntries = make([]LogEntry, 1)\n\n\trf.state = Follower\n\n\trf.commitIndex = 0\n\trf.lastApplied = 0\n\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\n\trf.nextIndex = make([]int, len(rf.peers))\n\trf.matchIndex = make([]int, len(rf.peers))\n\n\trf.applyCh = applyCh\n\trf.leaderElected = make(chan bool)\n\trf.appendEntry = make(chan bool)\n\theartBeat := time.Duration(100)*time.Millisecond\n\t// You'll need to write code that takes actions periodically or after delays in time.\n\t// The easiest way to do this is to create a goroutine with a loop that calls time.Sleep().\n\t// Don't use Go's time.Timer or time.Ticker, which are difficult to use correctly.\n\tgo func() {\n\t\tfor {\n\t\t\tdelay := time.Duration(rand.Intn(300) + 300)*time.Millisecond\n\t\t\ttimer := time.NewTimer(delay)\n\t\t\trf.mu.Lock()\n\t\t\tstate := rf.state\n\t\t\trf.mu.Unlock()\n\t\t\tswitch state {\n\t\t\tcase Follower, Candidate:\n\t\t\t\tselect {\n\t\t\t\tcase <- timer.C:\n\t\t\t\t\trf.mu.Lock()\n\t\t\t\t\trf.BeCandidate()\n\t\t\t\t\trf.mu.Unlock()\n\t\t\t\t\tgo func() {\n\t\t\t\t\t\trf.StartElection()\n\t\t\t\t\t}()\n\t\t\t\tcase <- rf.leaderElected:\n\t\t\t\tcase <- rf.appendEntry:\n\n\t\t\t\t}\n\t\t\tcase Leader:\n\t\t\t\trf.StartAppendLog()\n\t\t\t\ttime.Sleep(heartBeat)\n\t\t\t}\n\t\t}\n\t}()\n\n\treturn rf\n}", "func init() {\n\tuserFields := schema.User{}.Fields()\n\t_ = userFields\n\t// userDescUsername is the schema descriptor for username field.\n\tuserDescUsername := userFields[0].Descriptor()\n\t// user.DefaultUsername holds the default value on creation for the username field.\n\tuser.DefaultUsername = userDescUsername.Default.(string)\n\t// userDescName is the schema descriptor for name field.\n\tuserDescName := userFields[1].Descriptor()\n\t// user.DefaultName holds the default value on creation for the name field.\n\tuser.DefaultName = userDescName.Default.(string)\n\t// userDescSurname is the schema descriptor for surname field.\n\tuserDescSurname := userFields[2].Descriptor()\n\t// user.DefaultSurname holds the default value on creation for the surname field.\n\tuser.DefaultSurname = userDescSurname.Default.(string)\n}", "func initialise() {\n\taction[0] = \"stone\"\n\taction[1] = \"paper\"\n\taction[2] = \"scissors\"\n\tcanDefeat[0] = 2\n\tcanDefeat[1] = 0\n\tcanDefeat[2] = 1\n\tp1TotalScore = 0\n\tp2TotalScore = 0\n\tp3TotalScore = 0\n\tp4TotalScore = 0\n\n}", "func (s *raftServer) startElection() {\n\ts.setState(CANDIDATE)\n\tpeers := s.server.Peers()\n\ts.writeToLog(\"Number of peers: \" + strconv.Itoa(len(peers)))\n\tvotes := make(map[int]bool) // map to store received votes\n\tvotes[s.server.Pid()] = true\n\ts.voteFor(s.server.Pid(), s.Term())\n\tfor s.State() == CANDIDATE {\n\t\ts.incrTerm() // increment term for current\n\t\tcandidateTimeout := time.Duration(s.duration + s.rng.Int63n(RandomTimeoutRange)) // random timeout used by Raft authors\n\t\ts.sendRequestVote()\n\t\ts.writeToLog(\"Sent RequestVote message \" + strconv.Itoa(int(candidateTimeout)))\n\t\ts.eTimeout.Reset(candidateTimeout * time.Millisecond) // start re-election timer\n\t\tfor {\n\t\t\tacc := false\n\t\t\tselect {\n\t\t\tcase e, _ := <-s.server.Inbox():\n\t\t\t\t// received a message on server's inbox\n\t\t\t\tmsg := e.Msg\n\t\t\t\tif ae, ok := msg.(AppendEntry); ok { // AppendEntry\n\t\t\t\t\tacc = s.handleAppendEntry(e.Pid, &ae)\n\t\t\t\t} else if rv, ok := msg.(RequestVote); ok { // RequestVote\n\t\t\t\t\tacc = s.handleRequestVote(e.Pid, &rv)\n\n\t\t\t\t} else if grantV, ok := msg.(GrantVote); ok && grantV.VoteGranted {\n\t\t\t\t\tvotes[e.Pid] = true\n\t\t\t\t\ts.writeToLog(\"Received grantVote message from \" + strconv.Itoa(e.Pid) + \" with term #\" + strconv.Itoa(grantV.Term))\n\t\t\t\t\ts.writeToLog(\"Votes received so far \" + strconv.Itoa(len(votes)))\n\t\t\t\t\tif len(votes) == len(peers)/2+1 { // received majority votes\n\t\t\t\t\t\ts.setState(LEADER)\n\t\t\t\t\t\ts.sendHeartBeat()\n\t\t\t\t\t\tacc = true\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\tcase <-s.eTimeout.C:\n\t\t\t\t// received timeout on election timer\n\t\t\t\ts.writeToLog(\"Received re-election timeout\")\n\t\t\t\tacc = true\n\t\t\tdefault:\n\t\t\t\ttime.Sleep(1 * time.Millisecond) // sleep to avoid busy looping\n\t\t\t}\n\n\t\t\tif acc {\n\t\t\t\ts.eTimeout.Reset(candidateTimeout * time.Millisecond) // start re-election timer\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n}", "func def_lure(bt *yboter,b *game.Board, r *game.Robot) game.Action {\n\t//move back to lure enermy\n\tnearby_count := count_enermies_oct(b,r)\n\tnearby_friend_count := count_friend_oct(b,r)\n\tdirection_backward := direction_back(b , r )\n\tdirection_forward := direction_forward(b , r )\n\n\tenermyloc := game.Loc{}\n\tenermyloc = r.Loc\n\tenermyloc = enermyloc.Add(direction_forward)\n\tloc_type := b.LocType(enermyloc)\n\tif (loc_type ==game.Valid){\n\t\topp_bot := b.At(enermyloc)\n\n\t\tloc := game.Loc{}\n\t\tloc = r.Loc\n\t\tloc = loc.Add(direction_backward)\n\t\tloc_type := b.LocType(loc)\n\t\tif (opp_bot !=nil){\n\t\t\tif (nearby_count >2 &&nearby_friend_count<=2 && loc_type== game.Valid){\n\t\t\t\t\tfuturehealth := bt.robot_positions[opp_bot.Loc].future_health\n\t\t\t\t\tif (futurehealth>10 && !friendAt(b,loc)){\n\t\t\t\t\t\treturn game.Action{\n\t\t\t\t\t\t\t\tKind: game.Move,\n\t\t\t\t\t\t\t\tDirection: direction_backward,\n\t\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n return game.Action{Kind: game.Wait}\n}", "func init() {\n\ttoken = nep17.Token{\n\t\tName: \"Awesome NEO Token\",\n\t\tSymbol: \"ANT\",\n\t\tDecimals: decimals,\n\t\tOwner: owner,\n\t\tTotalSupply: 11000000 * multiplier,\n\t\tCirculationKey: \"TokenCirculation\",\n\t}\n\tctx = storage.GetContext()\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\trf := &Raft{}\n\trf.peers = peers\n\trf.persister = persister\n\trf.me = me\n\trf.applyCh = applyCh\n\t// Your initialization code here (2A, 2B, 2C).\n\trf.logs = []*Entry{{0, nil}}\n\n\trf.nextIndex = make([]int, len(rf.peers))\n\trf.matchIndex = make([]int, len(rf.peers))\n\t// initialize from state persisted before a crash\n\trf.readPersist(persister.ReadRaftState())\n\trf.timer = time.NewTimer(randomTimeout())\n\trf.heartBeatsTimer = time.NewTimer(0)\n\tgo rf.electionWatcher()\n\tgo func() {\n\t\tfor range time.Tick(time.Second / 2) {\n\t\t\tif rf.killed() {\n\t\t\t\treturn\n\t\t\t}\n\t\t\t//log.Println(rf, rf.lockLocation, \"\\t\", rf.persister.RaftStateSize())\n\t\t}\n\t}()\n\treturn rf\n}", "func (rf *Raft) BeLeader() {\n\tif rf.state != Candidate {\n\t\treturn\n\t}\n\trf.state = Leader\n\trf.nextIndex = make([]int, len(rf.peers))\n\trf.matchIndex = make([]int, len(rf.peers))\n\n\tfor i := range rf.nextIndex {\n\t\trf.nextIndex[i] = rf.GetLastLogIndex() + 1\n\t}\n}", "func init() {\n\tShCache = &ShareCache{\n\t\tLRPC: &LeaderRpcAddr{\n\t\t\tAddr: \"\",\n\t\t\tPort: \"\",\n\t\t},\n\t}\n}", "func (rf *Raft) GetState() (int, bool) {\n\n var term int\n var isleader bool\n // Your code here (2A).\n rf.mu.Lock()\n term = rf.currentTerm\n isleader = (rf.state == StateLeader)\n rf.mu.Unlock()\n return term, isleader\n}", "func init() {\n\tSBC = data.NewBlockChain()\n\tid, _ := strconv.ParseInt(os.Args[1], 10, 32)\n\tPeers = data.NewPeerList( /*Register()*/ int32(id), 32) // Uses port number as ID since TA server is down\n\tprivateKey, publicKey = client.GenerateKeyPair()\n\tifStarted = false\n\tmpt.Initial()\n\tclientBalanceMap = make(map[string]int32)\n\tpendingTransaction = make(map[string]string)\n\ttransactionMpt.Initial()\n\tclientBalanceMap[string(client.PublicKeyToBytes(publicKey))] = 1000\n\thighestblockTransaction = 0\n}", "func Make(peers []*labrpc.ClientEnd, me int,\n\tpersister *Persister, applyCh chan ApplyMsg) *Raft {\n\n\trf := &Raft{\n\t\tpeers: peers,\n\t\tpersister: persister,\n\t\tme: me,\n\t\tstate: follower,\n\t\tcurrentTerm: 0,\n\t\tvotedFor: -1,\n\t\tlog: []LogEntry{{\n\t\t\tEntry: 0,\n\t\t\tTerm: 0,\n\t\t}},\n\t\tcommitIndex: 0,\n\t\tlastApplied: 0,\n\t\theartBeatTimer: time.Millisecond * 100,\n\t\tappendLogCh: make(chan struct{}),\n\t\tvoteCh: make(chan struct{}),\n\t\tkillChan: make(chan struct{}),\n\t\tapplyChan: applyCh,\n\t\tlastIncludedIndex: 0,\n\t\tlastIncludedTerm: 0,\n\t}\n\n\t// Your initialization code here (2A, 2B, 2C).\n\n\trand.Seed(time.Now().UnixNano())\n\trf.resetElectionTimer()\n\n\trf.readPersist(persister.ReadRaftState())\n\n\t// initialize from state persisted before a crash\n\tDPrintln(\"test is starting server \", me, \"log length\", rf.getLogLen(), \"last log entry\", rf.getLogEntry(rf.getLastLogIndex()),\n\t\t\"term\", rf.currentTerm, \" commitIndex\", rf.commitIndex,\n\t\t\"last applied\", rf.lastApplied, \"lastIncludedIndex\", rf.lastIncludedIndex,\n\t\t\"lastIncluded Term\", rf.lastIncludedTerm)\n\tgo rf.Run(me, peers)\n\n\treturn rf\n}", "func testCandidateResetTerm(t *testing.T, mt pb.MessageType) {\n\ta := newTestRaft(1, []uint64{1, 2, 3}, 10, 1, NewMemoryStorage())\n\tb := newTestRaft(2, []uint64{1, 2, 3}, 10, 1, NewMemoryStorage())\n\tc := newTestRaft(3, []uint64{1, 2, 3}, 10, 1, NewMemoryStorage())\n\n\tnt := newNetwork(a, b, c)\n\tdefer nt.closeAll()\n\n\tnt.send(pb.Message{From: 1, To: 1, Type: pb.MsgHup})\n\tif a.state != StateLeader {\n\t\tt.Errorf(\"state = %s, want %s\", a.state, StateLeader)\n\t}\n\tif b.state != StateFollower {\n\t\tt.Errorf(\"state = %s, want %s\", b.state, StateFollower)\n\t}\n\tif c.state != StateFollower {\n\t\tt.Errorf(\"state = %s, want %s\", c.state, StateFollower)\n\t}\n\n\t// isolate 3 and increase term in rest\n\tnt.isolate(3)\n\n\tnt.send(pb.Message{From: 2, To: 2, Type: pb.MsgHup})\n\tnt.send(pb.Message{From: 1, To: 1, Type: pb.MsgHup})\n\n\tif a.state != StateLeader {\n\t\tt.Errorf(\"state = %s, want %s\", a.state, StateLeader)\n\t}\n\tif b.state != StateFollower {\n\t\tt.Errorf(\"state = %s, want %s\", b.state, StateFollower)\n\t}\n\n\t// trigger campaign in isolated c\n\tc.resetRandomizedElectionTimeout()\n\tfor i := 0; i < c.randomizedElectionTimeout; i++ {\n\t\tc.tick()\n\t}\n\n\tif c.state != StateCandidate {\n\t\tt.Errorf(\"state = %s, want %s\", c.state, StateCandidate)\n\t}\n\n\tnt.recover()\n\n\t// leader sends to isolated candidate\n\t// and expects candidate to revert to follower\n\tnt.send(pb.Message{From: 1, To: 3, Term: a.Term, Type: mt})\n\n\tif c.state != StateFollower {\n\t\tt.Errorf(\"state = %s, want %s\", c.state, StateFollower)\n\t}\n\n\t// follower c term is reset with leader's\n\tif a.Term != c.Term {\n\t\tt.Errorf(\"follower term expected same term as leader's %d, got %d\", a.Term, c.Term)\n\t}\n}" ]
[ "0.63949907", "0.6267157", "0.6195712", "0.61056024", "0.6065081", "0.60086596", "0.5934397", "0.5931934", "0.59088296", "0.5904501", "0.59030974", "0.5869632", "0.5802007", "0.57828546", "0.5722769", "0.5721325", "0.5693509", "0.56461924", "0.56128496", "0.5611066", "0.55978256", "0.55848706", "0.55843496", "0.55809206", "0.5572329", "0.5553682", "0.5501328", "0.54776204", "0.54759747", "0.54719", "0.54590094", "0.5456636", "0.5433187", "0.54321927", "0.54190546", "0.5418869", "0.53993076", "0.5393227", "0.5389637", "0.538306", "0.5350783", "0.53448683", "0.53413665", "0.53369975", "0.53209317", "0.53165704", "0.5315797", "0.53126", "0.53039396", "0.52994084", "0.5291682", "0.5290106", "0.5274309", "0.5269219", "0.52555865", "0.5252356", "0.5240753", "0.522723", "0.52216", "0.52147675", "0.5212584", "0.52121174", "0.52110106", "0.51965624", "0.5193653", "0.519116", "0.51874864", "0.51871324", "0.5186086", "0.5182791", "0.5170617", "0.5168057", "0.51672506", "0.51482767", "0.51477295", "0.51408756", "0.513706", "0.5136404", "0.5132691", "0.512941", "0.5127633", "0.5124532", "0.51242113", "0.51212853", "0.5118807", "0.5115255", "0.51025933", "0.5090873", "0.5082684", "0.50823367", "0.50812334", "0.50808007", "0.5074314", "0.50743085", "0.5067422", "0.5060909", "0.5056498", "0.5055488", "0.50548816", "0.5049352" ]
0.7046132
0
/ param regionId: Region ID (Required)
func NewAddTemplateRequest( regionId string, ) *AddTemplateRequest { return &AddTemplateRequest{ JDCloudRequest: core.JDCloudRequest{ URL: "/regions/{regionId}/addTemplate", Method: "POST", Header: nil, Version: "v1", }, RegionId: regionId, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r CreateImageFromSnapshotsRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r OnlineSqlTaskQueryRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r InvokeRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r StatusReportRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r SearchLogContextRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r InvokeCommandRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r SetAuthConfigRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeAuditLogRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r *InvokeRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r AddTemplateRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r ReportTaskRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r LastDownsampleRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r QueryForbiddenInfoListRequest) GetRegionId() string {\n return \"\"\n}", "func (r AuthPrivilegeRequest) GetRegionId() string {\n return \"\"\n}", "func (r CreateCollectInfoRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r CreateBackupPlanRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r *StatusReportRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r GetLargeScreenDataRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r VerifyContractRequest) GetRegionId() string {\n return \"\"\n}", "func (r ModifyTaskRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r CreateVideoUploadTaskRequest) GetRegionId() string {\n return \"\"\n}", "func (r UpdateAsCronRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r ListDNSRecordsRequest) GetRegionId() string {\n return \"\"\n}", "func (r ModifyAppGeneralSettingByIdRequest) GetRegionId() string {\n return \"\"\n}", "func (r ListSmsSendOverviewUsingGETRequest) GetRegionId() string {\n return \"\"\n}", "func (r QueryDeviceStatusForPCdnRequest) GetRegionId() string {\n return \"\"\n}", "func (r DeleteVeditProjectRequest) GetRegionId() string {\n return \"\"\n}", "func (r SendMessagesByPinUsingPOSTRequest) GetRegionId() string {\n return \"\"\n}", "func (r CreateClusterRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeBandwidthsRequest) GetRegionId() string {\n return \"\"\n}", "func (r SetDevicePropertyRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeBillSummarysRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r GeneralAlterEventRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DeployRequest) GetRegionId() string {\n return \"\"\n}", "func (r GetBucketFilesRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r *AddTemplateRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r DescribeSlowLogRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DelSubDeviceWithCoreRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r UpdateListenerRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r AddPushStreamRuleRequest) GetRegionId() string {\n return \"\"\n}", "func (r DeleteCensorLibItemsRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeCacheInstancesRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r *UpdateListenerRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r AttachGroupPolicyRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeAlarmHistoryRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DeleteExtraCacheTimeRequest) GetRegionId() string {\n return \"\"\n}", "func (t *Toon) RegionID() int64 {\n\treturn t.Int(\"region\")\n}", "func (r *DescribeBillSummarysRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r DescribeThingShadowRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeDomainOnlineStreamRequest) GetRegionId() string {\n return \"\"\n}", "func SetRegion(r string) {\n\tcurrentRegion = r\n}", "func (r DescribeSnapshotPolicyDiskRelationsRequest) GetRegionId() string {\n return r.RegionId\n}", "func regionByID(regionID int64) *Region {\n\tif id := int(regionID); id >= 0 && id < len(Regions) {\n\t\treturn Regions[id]\n\t}\n\treturn RegionUnknown\n}", "func (this *BuoyController) Region() {\n\tbuoyBusiness.Region(&this.BaseController, this.GetString(\":region\"))\n}", "func (o *VmRestorePoint) SetRegionId(v string) {\n\to.RegionId = &v\n}", "func (o *VmRestorePoint) HasRegionId() bool {\n\tif o != nil && o.RegionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (t *Territory) Region(ctx context.Context, db DB) (*Region, error) {\n\treturn RegionByRegionID(ctx, db, t.RegionID)\n}", "func updateRegions() {\n\tlogrus.Debug(\"Updating regions.\")\n\n\tregions, err := getMarketRegions()\n\tif err != nil {\n\t\tlogrus.WithError(err).Error(\"Could not get regionIDs from ESI!\")\n\t\treturn\n\t}\n\n\tregionIDs = regions\n\tlogrus.Debug(\"Region update done.\")\n}", "func (o *Operator) RegionID() uint64 {\n\treturn o.regionID\n}", "func (o *VmRestorePoint) GetRegionId() string {\n\tif o == nil || o.RegionId == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.RegionId\n}", "func (c *Client) Region() (string, error) {\n\treturn c.get(\"/region\")\n}", "func (s *AvailableZoneForDescribeAvailableResourceOutput) SetRegionId(v string) *AvailableZoneForDescribeAvailableResourceOutput {\n\ts.RegionId = &v\n\treturn s\n}", "func (o *VirtualMachineToAlternativeRestoreOptions) SetRegionId(v string) {\n\to.RegionId = v\n}", "func refreshRegion() {\n\tdefer time.AfterFunc(regionRefreshDur, refreshRegion)\n\tupdateRegionIDCs()\n}", "func Get_region(ipaddress string) IP2Locationrecord {\n\treturn handleError(defaultDB.query(ipaddress, region))\n}", "func (s *KinesisService) RegionName() string {\n\treturn s.region.Name()\n}", "func (a *AdditionalGUTI) GetAMFRegionID() (aMFRegionID uint8) {}", "func (s *testPluginCodeSuite) TestGetRegionIDs(c *C) {\n\tregionIDs := schedule.GetRegionIDs(s.tc,\"757365727461626C653A7573657231773937383833313437333137333731323135\",\n\t\t\"757365727461626C653A7573657234443637353232383738383832303830353737\")\n\tc.Assert(len(regionIDs), Equals, 3)\n\tc.Assert(regionIDs[0], Equals, uint64(2))\n\tc.Assert(regionIDs[1], Equals, uint64(3))\n\tc.Assert(regionIDs[2], Equals, uint64(4))\n}", "func (o *DcimRacksListParams) SetRegionID(regionID *string) {\n\to.RegionID = regionID\n}", "func Region(v string) predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldRegion), v))\n\t})\n}", "func Region(ctx context.Context) (string, bool) {\n\tsecretID, ok := ctx.Value(ContextRegion).(string)\n\n\treturn secretID, ok\n}", "func (o *DcimSitesReadParams) SetRegionID(regionID *string) {\n\to.RegionID = regionID\n}", "func awsFormatRegion(r *string) aws.Region {\n\tvar region aws.Region\n\tswitch *r {\n\tcase \"us-gov-west-1\":\n\t\tregion = aws.USGovWest\n\tcase \"us-east-1\":\n\t\tregion = aws.USEast\n\tcase \"us-west-1\":\n\t\tregion = aws.USWest\n\tcase \"us-west-2\":\n\t\tregion = aws.USWest2\n\tcase \"eu-west-1\":\n\t\tregion = aws.EUWest\n\tcase \"ap-southeast-1\":\n\t\tregion = aws.APSoutheast\n\tcase \"ap-southeast-2\":\n\t\tregion = aws.APSoutheast2\n\tcase \"ap-northeast-1\":\n\t\tregion = aws.APNortheast\n\tcase \"sa-east-1\":\n\t\tregion = aws.SAEast\n\tcase \"\":\n\t\tregion = aws.USEast\n\tdefault:\n\t\tlog.Fatalf(\"Invalid Region: %s\\n\", *r)\n\t}\n\treturn region\n}", "func (b *ServiceClusterBuilder) Region(value string) *ServiceClusterBuilder {\n\tb.region = value\n\tb.bitmap_ |= 128\n\treturn b\n}", "func (service *Service) GetAllInRegion(d9CloudAccountID, awsRegionName string) (*[]CloudSecurityGroupResponse, *http.Response, error) {\n\tif d9CloudAccountID == \"\" && awsRegionName == \"\" {\n\t\treturn nil, nil, fmt.Errorf(\"d9 cloud account id and aws region name must be passed\")\n\t}\n\n\toptions := GetSecurityGroupQueryParameters{\n\t\tCloudAccountID: d9CloudAccountID,\n\t\tRegionID: awsRegionName,\n\t}\n\n\tv := new([]CloudSecurityGroupResponse)\n\tresp, err := service.Client.NewRequestDo(\"GET\", awsSgResourcePath, options, nil, v)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn v, resp, nil\n}", "func (api *LaborStatsAPI) QueryRegion() ([]Region, error) {\n\ta := RegionAPI{\n\t\tDebug: api.Debug,\n\t\tSecretKey: api.SecretKey,\n\t}\n\n\terr := a.sendRequest()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := a.unmarshalData()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n\n}", "func (b *SubscriptionMetricsBuilder) Region(value string) *SubscriptionMetricsBuilder {\n\tb.region = value\n\tb.bitmap_ |= 8192\n\treturn b\n}", "func (o *VirtualMachineToAlternativeRestoreOptions) GetRegionId() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.RegionId\n}", "func StringToRegion(regionKeyOrID string) (r Region, err error) {\n\tregionStr := strings.ToLower(regionKeyOrID)\n\t// check if region identifier is provided\n\tr = Region(regionStr)\n\t_, ok := regionRealm[r]\n\tif ok {\n\t\treturn\n\t}\n\n\t// check if region key is provided\n\tr, ok = shortNameRegion[regionStr]\n\tif ok {\n\t\treturn\n\t}\n\n\treturn \"\", fmt.Errorf(\"region named %s is not recognized\", regionKeyOrID)\n}", "func (b *Backend) InjectRegion(region string) {\n\tb.gatewayStatus.Region = region\n}", "func (a *AdditionalGUTI) SetAMFRegionID(aMFRegionID uint8) {}", "func (bc *BasicLineGraph) SearchRegion(regionKey []byte) *RegionInfo {\n\tbc.RLock()\n\tdefer bc.RUnlock()\n\treturn bc.Regions.SearchRegion(regionKey)\n}", "func (o GetSmartagFlowLogsLogOutput) SlsRegionId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetSmartagFlowLogsLog) string { return v.SlsRegionId }).(pulumi.StringOutput)\n}", "func TestGetAllRegions(t *testing.T) {\n\tawsRegionSample := []string{\"ap-southeast-1\", \"us-west-2\", \"ap-northeast-1\", \"eu-west-2\", \"eu-central-1\"}\n\tawsChinaSample := []string{\"cn-north-1\", \"cn-northwest-1\"}\n\n\tawsRegions := GetAllRegions()\n\tfor _, region := range awsRegionSample {\n\t\tif !stringInSlice(region, awsRegions) {\n\t\t\tt.Errorf(\"Could not find region %s in retrieved list: %v\", region, awsRegions)\n\t\t}\n\t}\n\n\t// Test the same for China\n\tawsRegions = GetAllChinaRegions()\n\tfor _, region := range awsChinaSample {\n\t\tif !stringInSlice(region, awsRegions) {\n\t\t\tt.Errorf(\"Could not find region %s in retrieved list: %v\", region, awsRegions)\n\t\t}\n\t}\n}", "func (d *DB) Get_region(ipaddress string) (IP2Locationrecord, error) {\n\treturn d.query(ipaddress, region)\n}", "func TestIntegrationRegion(t *testing.T) {\n\ttests := []struct {\n\t\tname string\n\t\tregion string\n\t\tregions []string\n\t\thaserror bool\n\t}{\n\t\t{\"test empty region\", \"\", genericRegions, true},\n\t\t{\"test good region\", genericRegions[0], genericRegions, false},\n\t}\n\tfor _, tt := range tests {\n\t\ttc := tt\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\ts := New(tc.region, tc.regions)\n\t\t\tss, err := s.Region(tc.region)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatalf(\"Region() failed: %v\", err)\n\t\t\t}\n\t\t\tif ss.Config == nil || ss.Config.Region == nil {\n\t\t\t\tt.Fatal(\"ss.Config or ss.Config.Region is nil\")\n\t\t\t}\n\t\t\tresult := *ss.Config.Region\n\t\t\tif result != tc.region {\n\t\t\t\tt.Fatalf(\"Region() region invalid, expected: %s, got: %s\", tc.region, result)\n\t\t\t}\n\t\t})\n\t}\n}", "func (o *GetMarketsRegionIDHistoryParams) SetRegionID(regionID int32) {\n\to.RegionID = regionID\n}", "func (s *Session) RegionByID(ctx context.Context, req RegionRequest, id types.ID) (*RegionResults, error) {\n\t// Build the URL\n\turl := s.APIURL + \"/\" + regionEndpoint + \"/\" + string(id)\n\n\t// Call and return\n\treturn s.region(ctx, url, req)\n}", "func (v *VCard) Region() string {\n\treturn v.getFirstAddressField(4)\n}", "func RegisterRegion(updateFunc func()) *Region {\n\treturn (*Region)(newRegion(updateFunc))\n}", "func RegionContains(v string) predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\ts.Where(sql.Contains(s.C(FieldRegion), v))\n\t})\n}", "func (c *MockAzureCloud) Region() string {\n\treturn c.Location\n}", "func SelectRegion(optRegion string) (region string, err error) {\n\tif optRegion == \"OS Environment 'AWS_REGION'\" {\n\t\tif os.Getenv(\"AWS_REGION\") != \"\" {\n\t\t\treturn os.Getenv(\"AWS_REGION\"), err\n\t\t}\n\t\treturn usEast1, err\n\t}\n\treturn optRegion, err\n}", "func ExampleServiceClient_RegionConfiguration() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armdatabox.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewServiceClient().RegionConfiguration(ctx, \"westus\", armdatabox.RegionConfigurationRequest{\n\t\tScheduleAvailabilityRequest: &armdatabox.ScheduleAvailabilityRequest{\n\t\t\tSKUName: to.Ptr(armdatabox.SKUNameDataBox),\n\t\t\tStorageLocation: to.Ptr(\"westus\"),\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.RegionConfigurationResponse = armdatabox.RegionConfigurationResponse{\n\t// \tScheduleAvailabilityResponse: &armdatabox.ScheduleAvailabilityResponse{\n\t// \t\tAvailableDates: []*time.Time{\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-11T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-12T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-13T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-14T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-15T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-16T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-17T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-18T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-19T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-20T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-21T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-22T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-23T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-24T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-25T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-26T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-27T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-28T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-29T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-30T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-31T00:00:00+00:00\"); return t}())},\n\t// \t\t},\n\t// \t}\n}", "func (t *Toon) Region() *Region {\n\treturn regionByID(t.RegionID())\n}", "func GetRegion(kpp string) int {\n\ti, _ := strconv.Atoi(kpp[:2])\n\treturn i\n}", "func (client *IdentityClient) SetRegion(region string) {\n\tclient.Host = common.StringToRegion(region).EndpointForTemplate(\"identity\", \"https://identity.{region}.oci.{secondLevelDomain}\")\n}", "func GetRegion(licenseKey string) string {\n\tmatches := regionLicenseRegex.FindStringSubmatch(licenseKey)\n\tif len(matches) > 1 {\n\t\treturn matches[1]\n\t}\n\n\treturn \"\"\n}", "func (e Endpoints) AddRegion(ctx context.Context, token string, region registry.Region) (err error) {\n\trequest := AddRegionRequest{\n\t\tRegion: region,\n\t\tToken: token,\n\t}\n\tresponse, err := e.AddRegionEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(AddRegionResponse).Err\n}", "func GetRegion(ctx *pulumi.Context) string {\n\tv, err := config.Try(ctx, \"newrelic:region\")\n\tif err == nil {\n\t\treturn v\n\t}\n\tvar value string\n\tif d := internal.GetEnvOrDefault(\"US\", nil, \"NEW_RELIC_REGION\"); d != nil {\n\t\tvalue = d.(string)\n\t}\n\treturn value\n}", "func (c *Client) Region() string {\n\treturn c.region.String()\n}" ]
[ "0.72846764", "0.72511977", "0.71837354", "0.7152639", "0.7145258", "0.70786595", "0.7025792", "0.70122856", "0.7010603", "0.69977754", "0.69926155", "0.69665456", "0.695628", "0.69529605", "0.69462615", "0.69448465", "0.68738157", "0.68721294", "0.68643606", "0.6850821", "0.6850792", "0.68428713", "0.6837523", "0.6836159", "0.68260694", "0.6813795", "0.6795255", "0.6794021", "0.6793947", "0.67649794", "0.6764175", "0.6761878", "0.6756104", "0.67513984", "0.6705868", "0.6691487", "0.6688884", "0.6681892", "0.66764474", "0.66504204", "0.65957457", "0.6593493", "0.6567724", "0.6548042", "0.6540865", "0.6538746", "0.6518333", "0.64796865", "0.6478394", "0.63962024", "0.6208138", "0.61931396", "0.6074013", "0.60204965", "0.6006099", "0.6005655", "0.5944226", "0.59208286", "0.58950955", "0.58835334", "0.5871577", "0.58655316", "0.58599913", "0.5827411", "0.5809107", "0.5797924", "0.5776207", "0.5703442", "0.5690993", "0.56801593", "0.5658537", "0.5658007", "0.56210697", "0.5615739", "0.55991524", "0.559838", "0.55941683", "0.55881566", "0.5571985", "0.5570214", "0.5567725", "0.55509186", "0.55249906", "0.5523768", "0.5516205", "0.5495234", "0.54900324", "0.5483179", "0.5441644", "0.54367644", "0.5408674", "0.5392602", "0.5382393", "0.5366695", "0.5328126", "0.53121775", "0.53085333", "0.53060913", "0.5304922", "0.52890635", "0.52826744" ]
0.0
-1
/ This constructor has better compatible ability when API parameters changed
func NewAddTemplateRequestWithoutParam() *AddTemplateRequest { return &AddTemplateRequest{ JDCloudRequest: core.JDCloudRequest{ URL: "/regions/{regionId}/addTemplate", Method: "POST", Header: nil, Version: "v1", }, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func New() *Client {\n return &Client{&API{}}\n}", "func (u Updates) construct() UpdatesClass { return &u }", "func (u UpdatesTooLong) construct() UpdatesClass { return &u }", "func (u UpdateShortMessage) construct() UpdatesClass { return &u }", "func (u UpdateShort) construct() UpdatesClass { return &u }", "func New(config Config) API {\n\treturn API{\n\t\tClient: &http.Client{},\n\t\tConfig: config,\n\t}\n}", "func (u UpdatesCombined) construct() UpdatesClass { return &u }", "func (u UpdateShortSentMessage) construct() UpdatesClass { return &u }", "func (u UpdateShortChatMessage) construct() UpdatesClass { return &u }", "func (s ServerDHParamsOk) construct() ServerDHParamsClass { return &s }", "func newClient(apiKey string) *Client {\n\tvar url *url.URL\n\turl, _ = url.Parse(\"https://vulners.com/api/v3\")\n\treturn &Client{baseURL: url, apiKey: apiKey}\n}", "func newApiDefinitionFromAPI(a utils.API) *api {\n\treturn &api{a.ID, a.Name, a.Context, a.Version, a.Provider,\n\t\ta.LifeCycleStatus}\n}", "func CreateServiceStatusRequest() (request *ServiceStatusRequest) {\nrequest = &ServiceStatusRequest{\nRpcRequest: &requests.RpcRequest{},\n}\nrequest.InitWithApiInfo(\"Yundun\", \"2015-04-16\", \"ServiceStatus\", \"yundun\", \"openAPI\")\nreturn\n}", "func New(config Config) IAPI {\n\tif config.BaseURL == \"\" {\n\t\tconfig.BaseURL = defaultAPIURL\n\t}\n\tif config.Timeout == nil {\n\t\tconfig.Timeout = newDuration(defaultTimeout)\n\t}\n\n\tapi := &API{config: config}\n\tapi.client = &http.Client{\n\t\tTransport: api,\n\t\tTimeout: *config.Timeout,\n\t}\n\treturn api\n}", "func New(value interface{}, unsafe...bool) *Parser {\n return &Parser{gjson.New(value, unsafe...)}\n}", "func New(value interface{}, unsafe...bool) *Parser {\n return &Parser{gjson.New(value, unsafe...)}\n}", "func (c *Objs) Initialize(i util.XapiClient) {\n c.Address = &addr.Addr{}\n c.Address.Initialize(i)\n\n c.AddressGroup = &addrgrp.AddrGrp{}\n c.AddressGroup.Initialize(i)\n\n c.Services = &srvc.Srvc{}\n c.Services.Initialize(i)\n}", "func New(key string) *Api {\n\treturn &Api{&http.Client{}, key, \"https://newsapi.org/v2/\"}\n}", "func newOAPI(c container.Container) (x OAPI, err error) {\n\tx = OAPI{c: c}\n\terr = json.Unmarshal(c.Bytes(), &x.o.OpenAPI)\n\treturn x, err\n}", "func (g *GoFlickr) newRequest(apiMethod string) ApiRequest {\n\n\treq := ApiRequest{\n\t\tMethodName: apiMethod,\n\t}\n\treq.addParam(\"api_key\", g.ApiKey)\n\treq.addParam(\"format\", \"json\")\n\treq.addParam(\"nojsoncallback\", \"1\")\n\treturn req\n\n}", "func newAPI(cfg *ClientConfig, options ...ClientOption) *Client {\n\tclient := &Client{\n\t\tConfig: cfg,\n\t\thttpClient: &http.Client{},\n\t}\n\n\tfor _, option := range options {\n\t\toption(client)\n\t}\n\n\treturn client\n}", "func NewApiResponse() *ApiResponse {\n this := ApiResponse{}\n return &this\n}", "func New(sess *session.Session, config *client.Config) *API {\n\treturn &API{\n\t\tV2010: v2010.New(sess, config),\n\t}\n}", "func NewAPI(log *util.Logger, identity *Identity, cache time.Duration) *API {\n\tv := &API{\n\t\tlog: log,\n\t\tidentity: identity,\n\t\tHelper: request.NewHelper(log),\n\t}\n\n\t// api is unbelievably slow when retrieving status\n\tv.Helper.Client.Timeout = 120 * time.Second\n\n\treturn v\n}", "func newBase() *base {\n\treturn &base{shared.NewUUID(), time.Now().UTC(), time.Now().UTC(), false/*, shared.NewUUID()*/}\n}", "func (s ServerDHParamsFail) construct() ServerDHParamsClass { return &s }", "func NewUnsafe(value...interface{}) *Parser {\n if len(value) > 0 {\n return &Parser{gjson.New(value[0], false)}\n }\n return &Parser{gjson.New(nil, false)}\n}", "func (c ChannelLocation) construct() ChannelLocationClass { return &c }", "func InitAPI() {\n\n}", "func New(apiKey string, logger *logging.StandardLogger) *Client {\n\tc := &http.Client{Timeout: 15 * time.Second}\n\n\tq := baseURL.Query() // Get a copy of the query values.\n\tq.Add(\"token\", apiKey) // Add a new value to the set.\n\tbaseURL.RawQuery = q.Encode() // Encode and assign back to the original query.\n\n\treturn &Client{\n\t\tc: c,\n\t\tapiKey: apiKey,\n\t\tsyncToken: \"*\",\n\t\tlogger: logger,\n\t}\n}", "func New(logger model.Logger) API {\n\treturn API{logger: logger}\n}", "func (a StoriesAllStoriesNotModified) construct() StoriesAllStoriesClass { return &a }", "func newAPIRequest(host, path, apiKey string, body []byte) (*http.Request, error) {\n\treturn newAPIRequestWithHost(host, path, apiKey, body)\n}", "func NewRequest(id string, method string, params interface{}) *Request {\n\tbuffer, err := json.Marshal(params)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn &Request{JsonRPC: \"2.0\", Id: id, Method: method, Params: buffer}\n}", "func (b BotAppNotModified) construct() BotAppClass { return &b }", "func Constructor() Codec {\n\treturn Codec{}\n}", "func New(pingPeriod, pongTimeout time.Duration) *API {\n\treturn &API{\n\t\tclients: make(map[uint][]*client),\n\t\tpingPeriod: pingPeriod,\n\t\tpongTimeout: pingPeriod + pongTimeout,\n\t}\n}", "func (o *FakeObject) New(args ...interface{}) Object { return o.Invoke(args) }", "func NewAPI(log *util.Logger, identity oauth2.TokenSource, brand, country string) *API {\n\tv := &API{\n\t\tHelper: request.NewHelper(log),\n\t\tbrand: brand,\n\t\tcountry: country,\n\t\tbaseURI: DefaultBaseURI,\n\t}\n\n\tv.Client.Transport = &oauth2.Transport{\n\t\tSource: identity,\n\t\tBase: v.Client.Transport,\n\t}\n\n\treturn v\n}", "func InitOrder(userId int, buy bool, marketOrder bool, companyTicker string,\n numberOfShares int,\n limitPrice LimitPrice, eventTime time.Time) *Order {\n order := Order{\n IdNumber: currentId,\n UserId: userId,\n Buy: buy,\n MarketOrder: marketOrder,\n CompanyTicker: companyTicker,\n NumberOfShares: numberOfShares,\n LimitPrice: limitPrice,\n EventTime: eventTime}\n currentId += 1\n return &order\n}", "func (e EncryptedChatRequested) construct() EncryptedChatClass { return &e }", "func New() *Parameter {\n\tparam := &Parameter{\n\t\tparameters: &sync.Map{},\n\t}\n\n\treturn param\n\n}", "func (e EncryptedChatDiscarded) construct() EncryptedChatClass { return &e }", "func (c *baseClient) New() *baseClient {\n\t// Copy headers\n\theader := make(http.Header)\n\tfor k, v := range c.header {\n\t\theader[k] = v\n\t}\n\n\treturn &baseClient{\n\t\thttpClient: c.httpClient,\n\t\tmethod: c.method,\n\t\turl: c.url,\n\t\theader: header,\n\t}\n}", "func Constructor() Twitter {\n \n}", "func (p ProxyTypeHTTP) construct() ProxyTypeClass { return &p }", "func newCreateParam4Create(req *http.Request) (*CreateParam, error) {\n\tparam := &CreateParam{}\n\terr := xreq.BindJSON(req, param)\n\treturn param, err\n}", "func (s *Saiyan) constructor(name string, power int) *Saiyan {\n\treturn &Saiyan{\n\t\tName: name,\n\t\tPower: power,\n\t}\n}", "func (c *Dg) Initialize(con util.XapiClient) {\n c.con = con\n}", "func init() {\n\t//todo...\n}", "func NewAPI(repository Source, sender Sender) API {\n\tapi := new(apiImpl)\n\tapi.repository = repository\n\tapi.sender = sender\n\treturn *api\n}", "func (p PhoneCallRequested) construct() PhoneCallClass { return &p }", "func newBaseClient() *baseClient {\n\treturn &baseClient{\n\t\thttpClient: http.DefaultClient,\n\t\tmethod: \"GET\",\n\t\theader: make(http.Header),\n\t}\n}", "func (this *NowStr) Constructor() FunctionConstructor { return NewNowStr }", "func New(log *util.Logger, user, password string, cache time.Duration, config Config) (*API, error) {\n\tif err := mergo.Merge(&config, defaults); err != nil {\n\t\treturn nil, err\n\t}\n\n\tv := &API{\n\t\tlog: log,\n\t\tHelper: request.NewHelper(log),\n\t\tconfig: config,\n\t\tuser: user,\n\t\tpassword: password,\n\t}\n\n\t// api is unbelievably slow when retrieving status\n\tv.Helper.Client.Timeout = 120 * time.Second\n\n\tv.apiG = provider.NewCached(v.statusAPI, cache).InterfaceGetter()\n\n\treturn v, nil\n}", "func New(apiKey string) (*API, error) {\n\tif apiKey == \"\" {\n\t\treturn nil, fmt.Errorf(\"empty API key\")\n\t}\n\n\tu, err := url.Parse(defaultURL)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"parsing URL\")\n\t}\n\n\treturn &API{Key: apiKey, URL: u}, nil\n}", "func NewAPI(url string, t string) API {\n\ta := new(API)\n\ta.baseURL = url\n\ta.token = t\n\treturn *a\n}", "func (c CallbackQueryPayloadGame) construct() CallbackQueryPayloadClass { return &c }", "func NewClient() *Client {\n c := &Client{}\n c.userService.URL = &c.URL\n return c\n}", "func (x *fastReflection_Params) New() protoreflect.Message {\n\treturn new(fastReflection_Params)\n}", "func New(client *httpclient.Client, baseURL string) Data {\n\td := Data{\n\t\tclient: client,\n\t\tbaseURL: baseURL,\n\t}\n\n\treturn d\n}", "func New(c *core.Core, r *mux.Router, version string) *API {\n\ta := &API{}\n\ta.AppVersion = version\n\ta.c = c\n\ta.reqDecoder = schema.NewDecoder()\n\twireupRoutes(r, a)\n\treturn a\n}", "func newRequest(method, url string, body string) *http.Request {\n\treq, err := http.NewRequest(method, url, strings.NewReader(body))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treq.Header.Set(\"X-API-Token\", \"token1\")\n\treturn req\n}", "func NewFoo() *Foo {\n return &Foo{}\n}", "func New() *API {\n\tapi := new(API)\n\tapi.msgCtrl = newMsgCtrl()\n\tapi.pingCtrl = newPingCtrl()\n\treturn api\n}", "func (s SecureValueTypeInternalPassport) construct() SecureValueTypeClass { return &s }", "func (a *Service) init(encryptionAlgorithm string, encryptionKey string, issuer string,\n\ttimeoutInHours time.Duration, maxRefreshInHours time.Duration) *error_utils.ApiError {\n\n\t// validations\n\t// validate encryption algorithm\n\tif err := a.validateEncryptionAlgorithm(encryptionAlgorithm); err != nil {\n\t\treturn error_utils.NewBadRequestError(fmt.Sprintf(\"Auth: %v\", err.Error()))\n\t}\n\n\t// validate issuer\n\tif strings.TrimSpace(issuer) == \"\"{\n\t\treturn error_utils.NewBadRequestError(\"Auth: issuer cannot be empty\")\n\t}\n\n\t// validate max refresh\n\tif maxRefreshInHours <= 0{\n\t\treturn error_utils.NewBadRequestError(\"Auth: max refresh should be greater than 0\")\n\t}\n\n\t// validate jwt secret key\n\tif len(strings.TrimSpace(encryptionKey)) == 0 {\n\t\treturn error_utils.NewBadRequestError(\"Auth: encryption key cannot be empty\")\n\t}\n\n\t// validations\n\n\ta.encryptionAlgorithm = encryptionAlgorithm\n\ta.encryptionKey = []byte(encryptionKey)\n\ta.issuer = issuer\n\ta.maxRefresh = time.Hour * maxRefreshInHours\n\n\t// set defaults\n\ta.timeFunc = time.Now\n\tif a.timeout <= 0 {\n\t\ta.timeout = time.Hour * timeoutInHours\n\t}\n\t// set defaults\n\n\treturn nil\n}", "func New(credentials config.UserCredential) *RestAPI {\n\treturn &RestAPI{credentials: credentials}\n\t//var api = new(RestAPI)\n\t//api.credentials = credentials\n\t//return api\n}", "func (p PhoneCall) construct() PhoneCallClass { return &p }", "func Initialize(\n\tctx context.Context,\n\t// configuration\n\tcfg *config.TemporalConfig,\n\tversion string,\n\topts Options,\n\tclients Clients,\n\t// API dependencies\n\tl *zap.SugaredLogger,\n\n) (*API, error) {\n\tvar (\n\t\terr error\n\t\trouter = gin.Default()\n\t)\n\t// update dev mode\n\tdev = opts.DevMode\n\tl = l.Named(\"api\")\n\tim, err := rtfs.NewManager(\n\t\tcfg.IPFS.APIConnection.Host+\":\"+cfg.IPFS.APIConnection.Port,\n\t\t\"\", time.Minute*60,\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\timCluster, err := rtfscluster.Initialize(\n\t\tctx,\n\t\tcfg.IPFSCluster.APIConnection.Host,\n\t\tcfg.IPFSCluster.APIConnection.Port,\n\t)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// set up API struct\n\tapi, err := new(cfg, router, l, clients, im, imCluster, opts.DebugLogging)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tapi.version = version\n\n\t// init routes\n\tif err = api.setupRoutes(); err != nil {\n\t\treturn nil, err\n\t}\n\tapi.l.Info(\"api initialization successful\")\n\n\t// return our configured API service\n\treturn api, nil\n}", "func (s AuthSentCodeTypeCall) construct() AuthSentCodeTypeClass { return &s }", "func (e EncryptedChatWaiting) construct() EncryptedChatClass { return &e }", "func New(apiKey string) *Client {\n\tc := &http.Client{Timeout: 30 * time.Second}\n\n\treturn &Client{\n\t\tc: c,\n\t\tapiKey: apiKey,\n\t}\n}", "func (i InputCheckPasswordSRP) construct() InputCheckPasswordSRPClass { return &i }", "func NewVersionsRequestBuilderInternal(pathParameters map[string]string, requestAdapter i2ae4187f7daee263371cb1c977df639813ab50ffa529013b7437480d1ec0158f.RequestAdapter)(*VersionsRequestBuilder) {\n m := &VersionsRequestBuilder{\n }\n m.urlTemplate = \"{+baseurl}/users/{user%2Did}/drives/{drive%2Did}/list/items/{listItem%2Did}/versions{?%24top,%24skip,%24search,%24filter,%24count,%24orderby,%24select,%24expand}\";\n urlTplParams := make(map[string]string)\n for idx, item := range pathParameters {\n urlTplParams[idx] = item\n }\n m.pathParameters = urlTplParams;\n m.requestAdapter = requestAdapter;\n return m\n}", "func (e EncryptedChat) construct() EncryptedChatClass { return &e }", "func (c CallbackQueryPayloadDataWithPassword) construct() CallbackQueryPayloadClass { return &c }", "func newRequest(req *http.Request) *Request {\n\trequest := &Request{\n\t\tRequest: req,\n\t}\n\n\treturn request\n}", "func (s AuthSentCode) construct() AuthSentCodeClass { return &s }", "func (c *Variable) Initialize(con util.XapiClient) {\n c.con = con\n}", "func New(token, api string) (*API, error) {\n\tu, err := url.Parse(api)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tswitch u.Scheme {\n\tcase \"http\":\n\t\tfallthrough\n\tcase \"https\":\n\t\treturn NewHTTP(token, api)\n\tcase \"ws\":\n\t\tfallthrough\n\tcase \"wss\":\n\t\treturn NewWebsocket(token, api)\n\tdefault:\n\t\treturn nil, errors.New(\"invalid api's scheme\")\n\t}\n}", "func New(z *zoho.Zoho) *API {\n\tid := func() string {\n\t\tvar id []byte\n\t\tkeyspace := \"abcdefghijklmnopqrutuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789\"\n\t\tfor i := 0; i < 25; i++ {\n\t\t\tid = append(id, keyspace[rand.Intn(len(keyspace))])\n\t\t}\n\t\treturn string(id)\n\t}()\n\tAPI := &API{\n\t\tZoho: z,\n\t\tid: id,\n\t}\n\treturn API\n}", "func new(master_key string) Ca {\n\tcatls, err := tls.LoadX509KeyPair(\"../storage/root-certificate/ca_cert.pem\", \"../storage/root-certificate/ca_key.pem\")\n\tcheck(err)\n\tfirst_start_time := time.Date(2021, 1, 1, 0, 0, 0, 0, time.UTC).AddDate(0, 0, 0)\n\treturn Ca{catls, master_key, first_start_time}\n}", "func NewAPI() API {\n\treturn API{}\n}", "func (c *PanoDev) Initialize(i util.XapiClient) {\n c.EmailServer = &emailsrv.PanoServer{}\n c.EmailServer.Initialize(i)\n\n c.EmailServerProfile = &email.PanoEmail{}\n c.EmailServerProfile.Initialize(i)\n\n c.HttpHeader = &header.PanoHeader{}\n c.HttpHeader.Initialize(i)\n\n c.HttpParam = &param.PanoParam{}\n c.HttpParam.Initialize(i)\n\n c.HttpServer = &httpsrv.PanoServer{}\n c.HttpServer.Initialize(i)\n\n c.HttpServerProfile = &http.PanoHttp{}\n c.HttpServerProfile.Initialize(i)\n\n c.SnmpServerProfile = &snmp.PanoSnmp{}\n c.SnmpServerProfile.Initialize(i)\n\n c.SnmpV2cServer = &v2c.PanoV2c{}\n c.SnmpV2cServer.Initialize(i)\n\n c.SnmpV3Server = &v3.PanoV3{}\n c.SnmpV3Server.Initialize(i)\n\n c.SyslogServer = &syslogsrv.PanoServer{}\n c.SyslogServer.Initialize(i)\n\n c.SyslogServerProfile = &syslog.PanoSyslog{}\n c.SyslogServerProfile.Initialize(i)\n}", "func New(token string) *API {\n\treturn &API{\n\t\tToken: token,\n\t\thasHandlers: false,\n\t}\n}", "func (a StoriesAllStories) construct() StoriesAllStoriesClass { return &a }", "func newRequest(req *http.Request) *Request {\n\treturn &Request{\n\t\tRequest: req,\n\t}\n}", "func newBuffer(buf []byte) *Buffer {\n\treturn &Buffer{data: buf}\n}", "func Constructor() WordDictionary {\n \n}", "func (a UserAuth) construct() UserAuthClass { return &a }", "func NewTrainCar() TrainCar {\n c := TrainCar{name: \"TrainCar\", vehicle: \"TrainCar\", speed: 30, capacity: 30, railway: \"CNR\"}\n return c\n}", "func (qiwi *PersonalAPI) newRequest(apiKey, method, spath string, data map[string]interface{}) (req *http.Request, err error) {\n\n\tvar path = APIURL + spath\n\n\tvar body io.Reader\n\n\tif len(data) > 0 {\n\n\t\ts, err := json.Marshal(data)\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tbody = bytes.NewBuffer(s)\n\n\t}\n\n\treq, err = http.NewRequest(method, path, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Bearer \"+apiKey)\n\n\treturn req, err\n}", "func (c ChannelsChannelParticipants) construct() ChannelsChannelParticipantsClass { return &c }", "func (c ChannelLocationEmpty) construct() ChannelLocationClass { return &c }", "func init() {}", "func init() {}", "func init() {}", "func init() {}", "func newFileFromParams(upload *Upload, params *common.File) *File {\n\tfile := &File{}\n\tfile.upload = upload\n\tfile.metadata = params\n\tfile.Name = params.Name\n\tfile.Size = params.Size\n\treturn file\n}", "func Init(domain, apiKey string, options *ClientOptions) ApiClient {\n\tclient := ApiClient{\n\t\tdomain: domain,\n\t\tapiKey: apiKey,\n\t}\n\tif options != nil {\n\t\tclient.logger = options.Logger\n\t}\n\tif client.logger != nil {\n\t\tclient.logger.Println(\"Freshservice Client initializing... Domain =\", domain, \"authorization =\", apiKey)\n\t}\n\tclient.Departments = newDepartmentManager(&client)\n\tclient.Tickets = newTicketManager(&client)\n\tclient.Requesters = newrequesterManager(&client)\n\tclient.ServiceRequests = newServiceRequestManager(&client)\n\treturn client\n}" ]
[ "0.6067332", "0.5754528", "0.57516235", "0.54003227", "0.5373023", "0.53507143", "0.5346712", "0.5340663", "0.5327761", "0.52998316", "0.5280727", "0.5264717", "0.526081", "0.5239495", "0.5230421", "0.5230421", "0.5197221", "0.5185121", "0.5176182", "0.5171889", "0.5158723", "0.5145489", "0.514032", "0.5124777", "0.5120952", "0.51199514", "0.5083054", "0.5076184", "0.5071839", "0.50704986", "0.5047437", "0.5046053", "0.5043979", "0.5036311", "0.5034735", "0.5031239", "0.50279653", "0.5012242", "0.50109524", "0.50088096", "0.4989665", "0.4989426", "0.49882707", "0.49880123", "0.49857002", "0.4983146", "0.49765834", "0.49698725", "0.495726", "0.49513316", "0.49503437", "0.4947504", "0.49449307", "0.49436033", "0.49282384", "0.4918568", "0.49150726", "0.49108613", "0.49006838", "0.48994255", "0.48983938", "0.48976675", "0.4894737", "0.48937866", "0.48916543", "0.4889628", "0.48884213", "0.4885713", "0.48790026", "0.48754352", "0.48744726", "0.48732635", "0.4868927", "0.48623124", "0.48604286", "0.48595634", "0.48587027", "0.48564336", "0.48509264", "0.48459142", "0.48428553", "0.48405167", "0.48389584", "0.48368466", "0.48364234", "0.48293954", "0.48288757", "0.48277548", "0.48262113", "0.48256397", "0.48221427", "0.48203284", "0.48179853", "0.48124915", "0.4809337", "0.48077765", "0.48077765", "0.48077765", "0.48077765", "0.48015347", "0.47967973" ]
0.0
-1
/ param regionId: Region ID(Required)
func (r *AddTemplateRequest) SetRegionId(regionId string) { r.RegionId = regionId }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r CreateImageFromSnapshotsRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r OnlineSqlTaskQueryRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r StatusReportRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r InvokeRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r SearchLogContextRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r InvokeCommandRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeAuditLogRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r AddTemplateRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r ReportTaskRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r LastDownsampleRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r SetAuthConfigRequest) GetRegionId() string {\n return \"\"\n}", "func (r QueryForbiddenInfoListRequest) GetRegionId() string {\n return \"\"\n}", "func (r AuthPrivilegeRequest) GetRegionId() string {\n return \"\"\n}", "func (r CreateBackupPlanRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r CreateCollectInfoRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r VerifyContractRequest) GetRegionId() string {\n return \"\"\n}", "func (r GetLargeScreenDataRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r *InvokeRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r ModifyAppGeneralSettingByIdRequest) GetRegionId() string {\n return \"\"\n}", "func (r ModifyTaskRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r UpdateAsCronRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r ListSmsSendOverviewUsingGETRequest) GetRegionId() string {\n return \"\"\n}", "func (r CreateVideoUploadTaskRequest) GetRegionId() string {\n return \"\"\n}", "func (r ListDNSRecordsRequest) GetRegionId() string {\n return \"\"\n}", "func (r QueryDeviceStatusForPCdnRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeBillSummarysRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r SendMessagesByPinUsingPOSTRequest) GetRegionId() string {\n return \"\"\n}", "func (r DeleteVeditProjectRequest) GetRegionId() string {\n return \"\"\n}", "func (r CreateClusterRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DeployRequest) GetRegionId() string {\n return \"\"\n}", "func (r GeneralAlterEventRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r *StatusReportRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r SetDevicePropertyRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeBandwidthsRequest) GetRegionId() string {\n return \"\"\n}", "func (r DelSubDeviceWithCoreRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeSlowLogRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r UpdateListenerRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r GetBucketFilesRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r AddPushStreamRuleRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeAlarmHistoryRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DeleteCensorLibItemsRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeCacheInstancesRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r AttachGroupPolicyRequest) GetRegionId() string {\n return \"\"\n}", "func (r DeleteExtraCacheTimeRequest) GetRegionId() string {\n return \"\"\n}", "func (t *Toon) RegionID() int64 {\n\treturn t.Int(\"region\")\n}", "func (r DescribeThingShadowRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r *UpdateListenerRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r *DescribeBillSummarysRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r DescribeDomainOnlineStreamRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeSnapshotPolicyDiskRelationsRequest) GetRegionId() string {\n return r.RegionId\n}", "func SetRegion(r string) {\n\tcurrentRegion = r\n}", "func regionByID(regionID int64) *Region {\n\tif id := int(regionID); id >= 0 && id < len(Regions) {\n\t\treturn Regions[id]\n\t}\n\treturn RegionUnknown\n}", "func (this *BuoyController) Region() {\n\tbuoyBusiness.Region(&this.BaseController, this.GetString(\":region\"))\n}", "func (o *VmRestorePoint) HasRegionId() bool {\n\tif o != nil && o.RegionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (t *Territory) Region(ctx context.Context, db DB) (*Region, error) {\n\treturn RegionByRegionID(ctx, db, t.RegionID)\n}", "func updateRegions() {\n\tlogrus.Debug(\"Updating regions.\")\n\n\tregions, err := getMarketRegions()\n\tif err != nil {\n\t\tlogrus.WithError(err).Error(\"Could not get regionIDs from ESI!\")\n\t\treturn\n\t}\n\n\tregionIDs = regions\n\tlogrus.Debug(\"Region update done.\")\n}", "func (o *VmRestorePoint) SetRegionId(v string) {\n\to.RegionId = &v\n}", "func (o *VmRestorePoint) GetRegionId() string {\n\tif o == nil || o.RegionId == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.RegionId\n}", "func refreshRegion() {\n\tdefer time.AfterFunc(regionRefreshDur, refreshRegion)\n\tupdateRegionIDCs()\n}", "func (o *Operator) RegionID() uint64 {\n\treturn o.regionID\n}", "func (c *Client) Region() (string, error) {\n\treturn c.get(\"/region\")\n}", "func Get_region(ipaddress string) IP2Locationrecord {\n\treturn handleError(defaultDB.query(ipaddress, region))\n}", "func (a *AdditionalGUTI) GetAMFRegionID() (aMFRegionID uint8) {}", "func (s *KinesisService) RegionName() string {\n\treturn s.region.Name()\n}", "func (s *AvailableZoneForDescribeAvailableResourceOutput) SetRegionId(v string) *AvailableZoneForDescribeAvailableResourceOutput {\n\ts.RegionId = &v\n\treturn s\n}", "func (o *VirtualMachineToAlternativeRestoreOptions) SetRegionId(v string) {\n\to.RegionId = v\n}", "func (s *testPluginCodeSuite) TestGetRegionIDs(c *C) {\n\tregionIDs := schedule.GetRegionIDs(s.tc,\"757365727461626C653A7573657231773937383833313437333137333731323135\",\n\t\t\"757365727461626C653A7573657234443637353232383738383832303830353737\")\n\tc.Assert(len(regionIDs), Equals, 3)\n\tc.Assert(regionIDs[0], Equals, uint64(2))\n\tc.Assert(regionIDs[1], Equals, uint64(3))\n\tc.Assert(regionIDs[2], Equals, uint64(4))\n}", "func Region(v string) predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\ts.Where(sql.EQ(s.C(FieldRegion), v))\n\t})\n}", "func Region(ctx context.Context) (string, bool) {\n\tsecretID, ok := ctx.Value(ContextRegion).(string)\n\n\treturn secretID, ok\n}", "func awsFormatRegion(r *string) aws.Region {\n\tvar region aws.Region\n\tswitch *r {\n\tcase \"us-gov-west-1\":\n\t\tregion = aws.USGovWest\n\tcase \"us-east-1\":\n\t\tregion = aws.USEast\n\tcase \"us-west-1\":\n\t\tregion = aws.USWest\n\tcase \"us-west-2\":\n\t\tregion = aws.USWest2\n\tcase \"eu-west-1\":\n\t\tregion = aws.EUWest\n\tcase \"ap-southeast-1\":\n\t\tregion = aws.APSoutheast\n\tcase \"ap-southeast-2\":\n\t\tregion = aws.APSoutheast2\n\tcase \"ap-northeast-1\":\n\t\tregion = aws.APNortheast\n\tcase \"sa-east-1\":\n\t\tregion = aws.SAEast\n\tcase \"\":\n\t\tregion = aws.USEast\n\tdefault:\n\t\tlog.Fatalf(\"Invalid Region: %s\\n\", *r)\n\t}\n\treturn region\n}", "func (b *ServiceClusterBuilder) Region(value string) *ServiceClusterBuilder {\n\tb.region = value\n\tb.bitmap_ |= 128\n\treturn b\n}", "func (api *LaborStatsAPI) QueryRegion() ([]Region, error) {\n\ta := RegionAPI{\n\t\tDebug: api.Debug,\n\t\tSecretKey: api.SecretKey,\n\t}\n\n\terr := a.sendRequest()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tres, err := a.unmarshalData()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn res, nil\n\n}", "func (service *Service) GetAllInRegion(d9CloudAccountID, awsRegionName string) (*[]CloudSecurityGroupResponse, *http.Response, error) {\n\tif d9CloudAccountID == \"\" && awsRegionName == \"\" {\n\t\treturn nil, nil, fmt.Errorf(\"d9 cloud account id and aws region name must be passed\")\n\t}\n\n\toptions := GetSecurityGroupQueryParameters{\n\t\tCloudAccountID: d9CloudAccountID,\n\t\tRegionID: awsRegionName,\n\t}\n\n\tv := new([]CloudSecurityGroupResponse)\n\tresp, err := service.Client.NewRequestDo(\"GET\", awsSgResourcePath, options, nil, v)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\treturn v, resp, nil\n}", "func StringToRegion(regionKeyOrID string) (r Region, err error) {\n\tregionStr := strings.ToLower(regionKeyOrID)\n\t// check if region identifier is provided\n\tr = Region(regionStr)\n\t_, ok := regionRealm[r]\n\tif ok {\n\t\treturn\n\t}\n\n\t// check if region key is provided\n\tr, ok = shortNameRegion[regionStr]\n\tif ok {\n\t\treturn\n\t}\n\n\treturn \"\", fmt.Errorf(\"region named %s is not recognized\", regionKeyOrID)\n}", "func (b *Backend) InjectRegion(region string) {\n\tb.gatewayStatus.Region = region\n}", "func (b *SubscriptionMetricsBuilder) Region(value string) *SubscriptionMetricsBuilder {\n\tb.region = value\n\tb.bitmap_ |= 8192\n\treturn b\n}", "func TestGetAllRegions(t *testing.T) {\n\tawsRegionSample := []string{\"ap-southeast-1\", \"us-west-2\", \"ap-northeast-1\", \"eu-west-2\", \"eu-central-1\"}\n\tawsChinaSample := []string{\"cn-north-1\", \"cn-northwest-1\"}\n\n\tawsRegions := GetAllRegions()\n\tfor _, region := range awsRegionSample {\n\t\tif !stringInSlice(region, awsRegions) {\n\t\t\tt.Errorf(\"Could not find region %s in retrieved list: %v\", region, awsRegions)\n\t\t}\n\t}\n\n\t// Test the same for China\n\tawsRegions = GetAllChinaRegions()\n\tfor _, region := range awsChinaSample {\n\t\tif !stringInSlice(region, awsRegions) {\n\t\t\tt.Errorf(\"Could not find region %s in retrieved list: %v\", region, awsRegions)\n\t\t}\n\t}\n}", "func (o *DcimRacksListParams) SetRegionID(regionID *string) {\n\to.RegionID = regionID\n}", "func (o GetSmartagFlowLogsLogOutput) SlsRegionId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetSmartagFlowLogsLog) string { return v.SlsRegionId }).(pulumi.StringOutput)\n}", "func (bc *BasicLineGraph) SearchRegion(regionKey []byte) *RegionInfo {\n\tbc.RLock()\n\tdefer bc.RUnlock()\n\treturn bc.Regions.SearchRegion(regionKey)\n}", "func (o *VirtualMachineToAlternativeRestoreOptions) GetRegionId() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.RegionId\n}", "func TestIntegrationRegion(t *testing.T) {\n\ttests := []struct {\n\t\tname string\n\t\tregion string\n\t\tregions []string\n\t\thaserror bool\n\t}{\n\t\t{\"test empty region\", \"\", genericRegions, true},\n\t\t{\"test good region\", genericRegions[0], genericRegions, false},\n\t}\n\tfor _, tt := range tests {\n\t\ttc := tt\n\t\tt.Run(tc.name, func(t *testing.T) {\n\t\t\ts := New(tc.region, tc.regions)\n\t\t\tss, err := s.Region(tc.region)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatalf(\"Region() failed: %v\", err)\n\t\t\t}\n\t\t\tif ss.Config == nil || ss.Config.Region == nil {\n\t\t\t\tt.Fatal(\"ss.Config or ss.Config.Region is nil\")\n\t\t\t}\n\t\t\tresult := *ss.Config.Region\n\t\t\tif result != tc.region {\n\t\t\t\tt.Fatalf(\"Region() region invalid, expected: %s, got: %s\", tc.region, result)\n\t\t\t}\n\t\t})\n\t}\n}", "func (d *DB) Get_region(ipaddress string) (IP2Locationrecord, error) {\n\treturn d.query(ipaddress, region)\n}", "func (o *DcimSitesReadParams) SetRegionID(regionID *string) {\n\to.RegionID = regionID\n}", "func (v *VCard) Region() string {\n\treturn v.getFirstAddressField(4)\n}", "func (a *AdditionalGUTI) SetAMFRegionID(aMFRegionID uint8) {}", "func (s *Session) RegionByID(ctx context.Context, req RegionRequest, id types.ID) (*RegionResults, error) {\n\t// Build the URL\n\turl := s.APIURL + \"/\" + regionEndpoint + \"/\" + string(id)\n\n\t// Call and return\n\treturn s.region(ctx, url, req)\n}", "func RegionContains(v string) predicate.Location {\n\treturn predicate.Location(func(s *sql.Selector) {\n\t\ts.Where(sql.Contains(s.C(FieldRegion), v))\n\t})\n}", "func (c *MockAzureCloud) Region() string {\n\treturn c.Location\n}", "func RegisterRegion(updateFunc func()) *Region {\n\treturn (*Region)(newRegion(updateFunc))\n}", "func (t *Toon) Region() *Region {\n\treturn regionByID(t.RegionID())\n}", "func (s *Service) RegionChange(c context.Context, rid, rand int, plat int8, build int, mobiApp string) (sis []*show.Item) {\n\tif rand < 0 {\n\t\trand = 0\n\t}\n\tvar (\n\t\tcnt = 4\n\t\tpn = rand + 1\n\t\tisOsea = model.IsOverseas(plat)\n\t\tbangumiType = 0\n\t\ttmp []*show.Item\n\t)\n\tif (mobiApp == \"iphone\" && build > 5600) || (mobiApp == \"android\" && build > 507000) {\n\t\tif _, isBangumi := _bangumiReids[rid]; isBangumi {\n\t\t\tif (plat == model.PlatIPhone && build > 6050) || (plat == model.PlatAndroid && build > 512007) {\n\t\t\t\tbangumiType = _bangumiEpisodeID\n\t\t\t} else {\n\t\t\t\tbangumiType = _bangumiSeasonID\n\t\t\t}\n\t\t}\n\t}\n\tif model.IsIPad(plat) {\n\t\tcnt = 8\n\t}\n\tas, aids, err := s.dyn.RegionDynamic(c, rid, pn, cnt)\n\tif err != nil {\n\t\tlog.Error(\"s.rcmmnd.RegionDynamic(%d, %d, %d) error(%v)\", rid, pn, cnt, err)\n\t\tsis = []*show.Item{}\n\t\treturn\n\t}\n\tif bangumiType != 0 {\n\t\ttmp = s.fromArchivesBangumiOsea(c, as, aids, isOsea, bangumiType)\n\t} else {\n\t\ttmp = s.fromArchivesOsea(as, isOsea)\n\t}\n\tsis = append(sis, tmp...)\n\treturn\n}", "func (o *GetMarketsRegionIDHistoryParams) SetRegionID(regionID int32) {\n\to.RegionID = regionID\n}", "func ExampleServiceClient_RegionConfiguration() {\n\tcred, err := azidentity.NewDefaultAzureCredential(nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to obtain a credential: %v\", err)\n\t}\n\tctx := context.Background()\n\tclientFactory, err := armdatabox.NewClientFactory(\"<subscription-id>\", cred, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to create client: %v\", err)\n\t}\n\tres, err := clientFactory.NewServiceClient().RegionConfiguration(ctx, \"westus\", armdatabox.RegionConfigurationRequest{\n\t\tScheduleAvailabilityRequest: &armdatabox.ScheduleAvailabilityRequest{\n\t\t\tSKUName: to.Ptr(armdatabox.SKUNameDataBox),\n\t\t\tStorageLocation: to.Ptr(\"westus\"),\n\t\t},\n\t}, nil)\n\tif err != nil {\n\t\tlog.Fatalf(\"failed to finish the request: %v\", err)\n\t}\n\t// You could use response here. We use blank identifier for just demo purposes.\n\t_ = res\n\t// If the HTTP response code is 200 as defined in example definition, your response structure would look as follows. Please pay attention that all the values in the output are fake values for just demo purposes.\n\t// res.RegionConfigurationResponse = armdatabox.RegionConfigurationResponse{\n\t// \tScheduleAvailabilityResponse: &armdatabox.ScheduleAvailabilityResponse{\n\t// \t\tAvailableDates: []*time.Time{\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-11T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-12T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-13T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-14T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-15T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-16T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-17T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-18T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-19T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-20T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-21T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-22T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-23T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-24T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-25T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-26T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-27T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-28T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-29T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-30T00:00:00+00:00\"); return t}()),\n\t// \t\t\tto.Ptr(func() time.Time { t, _ := time.Parse(time.RFC3339Nano, \"2019-07-31T00:00:00+00:00\"); return t}())},\n\t// \t\t},\n\t// \t}\n}", "func (g *Game) RegionIDs() ([]string, *Error) {\n\tvar result []string\n\n\tswitch asserted := g.RegionsData.(type) {\n\t// list of IDs (strings)\n\tcase []interface{}:\n\t\tfor _, something := range asserted {\n\t\t\tid, okay := something.(string)\n\t\t\tif okay {\n\t\t\t\tresult = append(result, id)\n\t\t\t}\n\t\t}\n\n\t// sub-resource due to embeds, aka \"{data:....}\"\n\t// TODO: skip the conversion back and forth and just assert our way through the available data\n\tcase map[string]interface{}:\n\t\tregions, err := g.Regions()\n\t\tif err != nil {\n\t\t\treturn result, err\n\t\t}\n\n\t\tfor _, region := range regions.Regions() {\n\t\t\tresult = append(result, region.ID)\n\t\t}\n\t}\n\n\treturn result, nil\n}", "func (s *Service) NewRegionList(c context.Context, plat int8, build int, ver, mobiApp, device, language string) (res []*region.Region, version string, err error) {\n\tip := metadata.String(c, metadata.RemoteIP)\n\tvar (\n\t\thantlanguage = \"hant\"\n\t)\n\tif ok := model.IsOverseas(plat); ok && language != _initlanguage && language != hantlanguage {\n\t\tlanguage = hantlanguage\n\t} else if language == \"\" {\n\t\tlanguage = _initlanguage\n\t}\n\tvar (\n\t\trs = s.cachelist[fmt.Sprintf(_initRegionKey, plat, language)]\n\t\tchild = map[int][]*region.Region{}\n\t\tridtmp = map[string]struct{}{}\n\t\tpids []string\n\t\tauths map[string]*locmdl.Auth\n\t)\n\tfor _, rtmp := range rs {\n\t\tif rtmp.Area != \"\" {\n\t\t\tpids = append(pids, rtmp.Area)\n\t\t}\n\t}\n\tif len(pids) > 0 {\n\t\tauths, _ = s.loc.AuthPIDs(c, strings.Join(pids, \",\"), ip)\n\t}\nLOOP:\n\tfor _, rtmp := range rs {\n\t\tr := &region.Region{}\n\t\t*r = *rtmp\n\t\tif _, isgbm := _isBangumiIndex[r.Rid]; isgbm {\n\t\t\tr.IsBangumi = 1\n\t\t}\n\t\tvar tmpl, limitshow bool\n\t\tif limit, ok := s.limitCache[r.ID]; ok {\n\t\t\tfor i, l := range limit {\n\t\t\t\tif i+1 <= len(limit)-1 {\n\t\t\t\t\tif ((l.Condition == \"gt\" && limit[i+1].Condition == \"lt\") && (l.Build < limit[i+1].Build)) ||\n\t\t\t\t\t\t((l.Condition == \"lt\" && limit[i+1].Condition == \"gt\") && (l.Build > limit[i+1].Build)) {\n\t\t\t\t\t\tif (l.Condition == \"gt\" && limit[i+1].Condition == \"lt\") &&\n\t\t\t\t\t\t\t(build > l.Build && build < limit[i+1].Build) {\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t} else if (l.Condition == \"lt\" && limit[i+1].Condition == \"gt\") &&\n\t\t\t\t\t\t\t(build < l.Build && build > limit[i+1].Build) {\n\t\t\t\t\t\t\tbreak\n\t\t\t\t\t\t} else {\n\t\t\t\t\t\t\ttmpl = true\n\t\t\t\t\t\t\tcontinue\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t\tif tmpl {\n\t\t\t\t\tif i == len(limit)-1 {\n\t\t\t\t\t\tlimitshow = true\n\t\t\t\t\t\t// continue LOOP\n\t\t\t\t\t\tbreak\n\t\t\t\t\t}\n\t\t\t\t\ttmpl = false\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif model.InvalidBuild(build, l.Build, l.Condition) {\n\t\t\t\t\tlimitshow = true\n\t\t\t\t\tcontinue\n\t\t\t\t\t// continue LOOP\n\t\t\t\t} else {\n\t\t\t\t\tlimitshow = false\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\tif limitshow {\n\t\t\tcontinue LOOP\n\t\t}\n\t\tif auth, ok := auths[r.Area]; ok && auth.Play == locmdl.Forbidden {\n\t\t\tlog.Warn(\"s.invalid area(%v) ip(%v) error(%v)\", r.Area, ip, err)\n\t\t\tcontinue\n\t\t}\n\t\tif isAudit := s.auditRegion(mobiApp, plat, build, r.Rid); isAudit {\n\t\t\tcontinue\n\t\t}\n\t\tif config, ok := s.configCache[r.ID]; ok {\n\t\t\tr.Config = config\n\t\t}\n\t\tkey := fmt.Sprintf(_regionRepeat, r.Rid, r.Reid)\n\t\tif _, ok := ridtmp[key]; !ok {\n\t\t\tridtmp[key] = struct{}{}\n\t\t} else {\n\t\t\tcontinue\n\t\t}\n\t\tif r.Reid != 0 {\n\t\t\tcl, ok := child[r.Reid]\n\t\t\tif !ok {\n\t\t\t\tcl = []*region.Region{}\n\t\t\t}\n\t\t\tcl = append(cl, r)\n\t\t\tchild[r.Reid] = cl\n\t\t} else {\n\t\t\tres = append(res, r)\n\t\t}\n\t}\n\tif len(res) == 0 {\n\t\tres = _emptyRegions\n\t} else {\n\t\tfor _, r := range res {\n\t\t\tr.Children = child[r.Rid]\n\t\t}\n\t}\n\tif version = s.hash(res); version == ver {\n\t\terr = ecode.NotModified\n\t\tres = nil\n\t}\n\treturn\n}", "func GetRegion(licenseKey string) string {\n\tmatches := regionLicenseRegex.FindStringSubmatch(licenseKey)\n\tif len(matches) > 1 {\n\t\treturn matches[1]\n\t}\n\n\treturn \"\"\n}", "func GetRegion(kpp string) int {\n\ti, _ := strconv.Atoi(kpp[:2])\n\treturn i\n}", "func SelectRegion(optRegion string) (region string, err error) {\n\tif optRegion == \"OS Environment 'AWS_REGION'\" {\n\t\tif os.Getenv(\"AWS_REGION\") != \"\" {\n\t\t\treturn os.Getenv(\"AWS_REGION\"), err\n\t\t}\n\t\treturn usEast1, err\n\t}\n\treturn optRegion, err\n}", "func (e Endpoints) AddRegion(ctx context.Context, token string, region registry.Region) (err error) {\n\trequest := AddRegionRequest{\n\t\tRegion: region,\n\t\tToken: token,\n\t}\n\tresponse, err := e.AddRegionEndpoint(ctx, request)\n\tif err != nil {\n\t\treturn\n\t}\n\treturn response.(AddRegionResponse).Err\n}" ]
[ "0.72940403", "0.7261987", "0.719279", "0.71777534", "0.7161425", "0.7092324", "0.70373225", "0.7020295", "0.7007854", "0.69990385", "0.6996247", "0.69877774", "0.697372", "0.6970635", "0.6954599", "0.6908058", "0.6894179", "0.6872303", "0.6870722", "0.6858112", "0.6849531", "0.68383706", "0.68236595", "0.6805839", "0.6796558", "0.6786098", "0.6782944", "0.6779622", "0.67704177", "0.67690307", "0.6762565", "0.6746733", "0.6717867", "0.67130655", "0.6692396", "0.6685987", "0.6674401", "0.66740376", "0.6635239", "0.66056055", "0.66008943", "0.655774", "0.65568775", "0.6529033", "0.6506193", "0.64930546", "0.64188707", "0.63757944", "0.63631636", "0.6219298", "0.6172336", "0.6131299", "0.6125028", "0.603584", "0.60283333", "0.60101485", "0.5889533", "0.5878875", "0.5875929", "0.58602726", "0.5851607", "0.58506095", "0.58290964", "0.57503504", "0.5742587", "0.57368565", "0.5732126", "0.568922", "0.56543285", "0.564192", "0.56357944", "0.5618371", "0.5615157", "0.56141204", "0.5604579", "0.55949384", "0.55863696", "0.5572343", "0.5571334", "0.5554884", "0.5530434", "0.5526289", "0.55238664", "0.55156267", "0.5507112", "0.55037", "0.54631037", "0.5455686", "0.5429848", "0.54276496", "0.54130965", "0.5410514", "0.54019177", "0.53880996", "0.53456086", "0.53233916", "0.53194505", "0.531267", "0.5306067", "0.5303944" ]
0.6539114
43
GetRegionId returns path parameter 'regionId' if exist, otherwise return empty string
func (r AddTemplateRequest) GetRegionId() string { return r.RegionId }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r SearchLogContextRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r InvokeRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DeployRequest) GetRegionId() string {\n return \"\"\n}", "func (r StatusReportRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r InvokeCommandRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r VerifyContractRequest) GetRegionId() string {\n return \"\"\n}", "func (r DeleteVeditProjectRequest) GetRegionId() string {\n return \"\"\n}", "func (r ReportTaskRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeAuditLogRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r ModifyAppGeneralSettingByIdRequest) GetRegionId() string {\n return \"\"\n}", "func (r AuthPrivilegeRequest) GetRegionId() string {\n return \"\"\n}", "func (r LastDownsampleRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r SetAuthConfigRequest) GetRegionId() string {\n return \"\"\n}", "func (r CreateImageFromSnapshotsRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r GeneralAlterEventRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r OnlineSqlTaskQueryRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r ModifyTaskRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r UpdateListenerRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r QueryForbiddenInfoListRequest) GetRegionId() string {\n return \"\"\n}", "func (r GetBucketFilesRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r CreateBackupPlanRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r ListDNSRecordsRequest) GetRegionId() string {\n return \"\"\n}", "func (r QueryDeviceStatusForPCdnRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeBandwidthsRequest) GetRegionId() string {\n return \"\"\n}", "func (r SetDevicePropertyRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r CreateCollectInfoRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r GetLargeScreenDataRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r UpdateAsCronRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r AttachGroupPolicyRequest) GetRegionId() string {\n return \"\"\n}", "func (r DeleteExtraCacheTimeRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeBillSummarysRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r CreateClusterRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeSlowLogRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DelSubDeviceWithCoreRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeThingShadowRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeSnapshotPolicyDiskRelationsRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeAlarmHistoryRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r AddPushStreamRuleRequest) GetRegionId() string {\n return \"\"\n}", "func (r ListSmsSendOverviewUsingGETRequest) GetRegionId() string {\n return \"\"\n}", "func (r CreateVideoUploadTaskRequest) GetRegionId() string {\n return \"\"\n}", "func (r DescribeCacheInstancesRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r SendMessagesByPinUsingPOSTRequest) GetRegionId() string {\n return \"\"\n}", "func (r DeleteCensorLibItemsRequest) GetRegionId() string {\n return r.RegionId\n}", "func (r DescribeDomainOnlineStreamRequest) GetRegionId() string {\n return \"\"\n}", "func (o *VmRestorePoint) GetRegionId() string {\n\tif o == nil || o.RegionId == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.RegionId\n}", "func GetRegion(ctx *pulumi.Context) string {\n\tv, err := config.Try(ctx, \"newrelic:region\")\n\tif err == nil {\n\t\treturn v\n\t}\n\tvar value string\n\tif d := internal.GetEnvOrDefault(\"US\", nil, \"NEW_RELIC_REGION\"); d != nil {\n\t\tvalue = d.(string)\n\t}\n\treturn value\n}", "func (config *Config) GetRegion() string {\n\tregion := config.Region\n\n\tif region == \"\" {\n\t\tregion = Region[\"bj\"]\n\t}\n\n\treturn region\n}", "func (o *VirtualMachineToAlternativeRestoreOptions) GetRegionId() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.RegionId\n}", "func (o *VmRestorePoint) GetRegionIdOk() (*string, bool) {\n\tif o == nil || o.RegionId == nil {\n\t\treturn nil, false\n\t}\n\treturn o.RegionId, true\n}", "func GetRegion(svc *ec2metadata.EC2Metadata) (string, error) {\n\tdoc, err := svc.GetInstanceIdentityDocument()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn doc.Region, nil\n}", "func GetRegion(metaEndPoint string, preferEnv bool) string {\n\tvar region string\n\tif preferEnv {\n\t\tregion = getRegionFromEnv()\n\t\tif region == \"\" {\n\t\t\tregion = getRegionFromInstanceDocument(metaEndPoint)\n\t\t}\n\t} else {\n\t\tregion = getRegionFromInstanceDocument(metaEndPoint)\n\t\tif region == \"\" {\n\t\t\tregion = getRegionFromEnv()\n\t\t}\n\t}\n\tif region == \"\" {\n\t\tlog.Println(\"No region information available. Defaulting to us-west-2\")\n\t\tregion = \"us-west-2\"\n\t}\n\treturn region\n}", "func Region(ctx context.Context) (string, bool) {\n\tsecretID, ok := ctx.Value(ContextRegion).(string)\n\n\treturn secretID, ok\n}", "func GetRegion(ctx *pulumi.Context) string {\n\tv, err := config.Try(ctx, \"aws:region\")\n\tif err == nil {\n\t\treturn v\n\t}\n\tif dv, ok := getEnvOrDefault(\"\", nil, \"AWS_REGION\", \"AWS_DEFAULT_REGION\").(string); ok {\n\t\treturn dv\n\t}\n\treturn v\n}", "func (s serverConfigV14) GetRegion() string {\n\tserverConfigMu.RLock()\n\tdefer serverConfigMu.RUnlock()\n\n\treturn s.Region\n}", "func GetRegion(kpp string) int {\n\ti, _ := strconv.Atoi(kpp[:2])\n\treturn i\n}", "func (m *metadata) GetRegion() string {\n\treturn m.region\n}", "func GetRegion(licenseKey string) string {\n\tmatches := regionLicenseRegex.FindStringSubmatch(licenseKey)\n\tif len(matches) > 1 {\n\t\treturn matches[1]\n\t}\n\n\treturn \"\"\n}", "func (o *VmRestorePoint) HasRegionId() bool {\n\tif o != nil && o.RegionId != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (c *Client) GetRegion() string {\n\treturn c.config.Region\n}", "func (t *Toon) RegionID() int64 {\n\treturn t.Int(\"region\")\n}", "func (a *AZs) GetRegion(az string) string {\n\tfor _, vaz := range *a {\n\t\tif az == vaz.Name {\n\t\t\treturn vaz.Region\n\t\t}\n\t}\n\treturn \"\"\n}", "func GetEndpointForRegion(\n\tfqdn string, //DU fqdn\n\tauth KeystoneAuth, // Auth info\n\tregion string, //region name\n\tserviceID string, // ID for regionInfo service\n) (string, error) {\n\n\tzap.S().Debug(\"Fetching endpoint for region: \", region)\n\n\t// Form the URL\n\turl := fmt.Sprintf(\"%s/keystone/v3/endpoints\", fqdn)\n\n\t// Generate the http client object\n\tclient := &http.Client{}\n\n\t// Create the context to invoke the service manager API.\n\te_api := EndpointManagerAPI{client, url, auth.Token}\n\n\t// Invoke the actual \"get services\" API.\n\tendpoint, err := e_api.GetEndpointForRegion_API(region, serviceID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tzap.S().Debug(\"Endpoint found: \", endpoint)\n\treturn endpoint, nil\n}", "func (c *Client) Region() (string, error) {\n\treturn c.get(\"/region\")\n}", "func (o *Workloadv1Location) GetRegion() string {\n\tif o == nil || o.Region == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Region\n}", "func Get_region(ipaddress string) IP2Locationrecord {\n\treturn handleError(defaultDB.query(ipaddress, region))\n}", "func (d *Downloader) getRegion() string {\n\tif d.region != \"\" {\n\t\treturn d.region\n\t}\n\n\tregion, err := d.metadata.Region()\n\tif err != nil {\n\t\tlog.Warn(\"Could not retrieve the region from EC2 Instance Metadata. Error: %s\", err.Error())\n\t\tregion = config.DefaultRegionName\n\t}\n\td.region = region\n\n\treturn d.region\n}", "func (o *VirtualMachineToAlternativeRestoreOptions) GetRegionIdOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.RegionId, true\n}", "func (m *Info) GetRegion() string {\n\treturn m.ec2Metadata.getRegion()\n}", "func (o *ReplicaMapping) GetTargetRegionId() string {\n\tif o == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\n\treturn o.TargetRegionId\n}", "func (o *DisplayInfo) GetRegion() string {\n\tif o == nil || o.Region == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Region\n}", "func (c *namespaceCluster) GetRegion(id uint64) *core.RegionInfo {\n\tr := c.Cluster.GetRegion(id)\n\tif r == nil || !c.checkRegion(r) {\n\t\treturn nil\n\t}\n\treturn r\n}", "func GetRegion(configuredRegion string) (string, error) {\n\tif configuredRegion != \"\" {\n\t\treturn configuredRegion, nil\n\t}\n\n\tsess, err := session.NewSessionWithOptions(session.Options{\n\t\tSharedConfigState: session.SharedConfigEnable,\n\t})\n\tif err != nil {\n\t\treturn \"\", errwrap.Wrapf(\"got error when starting session: {{err}}\", err)\n\t}\n\n\tregion := aws.StringValue(sess.Config.Region)\n\tif region != \"\" {\n\t\treturn region, nil\n\t}\n\n\tmetadata := ec2metadata.New(sess, &aws.Config{\n\t\tEndpoint: ec2Endpoint,\n\t\tEC2MetadataDisableTimeoutOverride: aws.Bool(true),\n\t\tHTTPClient: &http.Client{\n\t\t\tTimeout: time.Second,\n\t\t},\n\t})\n\tif !metadata.Available() {\n\t\treturn DefaultRegion, nil\n\t}\n\n\tregion, err = metadata.Region()\n\tif err != nil {\n\t\treturn \"\", errwrap.Wrapf(\"unable to retrieve region from instance metadata: {{err}}\", err)\n\t}\n\n\treturn region, nil\n}", "func (o *Operator) RegionID() uint64 {\n\treturn o.regionID\n}", "func (e Endpoint) Region() string {\n\tif e.Properties.CredentialScope.Region != \"\" {\n\t\treturn fmt.Sprintf(\"%q\", e.Properties.CredentialScope.Region)\n\t}\n\treturn \"region\"\n}", "func (d *DB) Get_region(ipaddress string) (IP2Locationrecord, error) {\n\treturn d.query(ipaddress, region)\n}", "func getAwsRegion() (region string) {\n\tregion, _ = getAwsRegionE()\n\treturn\n}", "func regionByID(regionID int64) *Region {\n\tif id := int(regionID); id >= 0 && id < len(Regions) {\n\t\treturn Regions[id]\n\t}\n\treturn RegionUnknown\n}", "func (t *TikvHandlerTool) GetRegionIDByKey(encodedKey []byte) (uint64, error) {\n\tkeyLocation, err := t.RegionCache.LocateKey(tikv.NewBackofferWithVars(context.Background(), 500, nil), encodedKey)\n\tif err != nil {\n\t\treturn 0, derr.ToTiDBErr(err)\n\t}\n\treturn keyLocation.Region.GetID(), nil\n}", "func (o GetEndpointResultOutput) Region() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetEndpointResult) string { return v.Region }).(pulumi.StringOutput)\n}", "func (r *InvokeRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (d *DynamoDBMetastore) GetRegionSuffix() string {\n\treturn d.regionSuffix\n}", "func (r *CountryRegionRequest) Get(ctx context.Context) (resObj *CountryRegion, err error) {\n\tvar query string\n\tif r.query != nil {\n\t\tquery = \"?\" + r.query.Encode()\n\t}\n\terr = r.JSONRequest(ctx, \"GET\", query, nil, &resObj)\n\treturn\n}", "func (bc *BasicLineGraph) GetRegion(regionID uint3264) *RegionInfo {\n\tbc.RLock()\n\tdefer bc.RUnlock()\n\treturn bc.Regions.GetRegion(regionID)\n}", "func (o GetSmartagFlowLogsLogOutput) SlsRegionId() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetSmartagFlowLogsLog) string { return v.SlsRegionId }).(pulumi.StringOutput)\n}", "func (r *StatusReportRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (r *reader) GetVyprvpnRegion() (region string, err error) {\n\treturn r.envParams.GetValueIfInside(\"REGION\", constants.VyprvpnRegionChoices())\n}", "func GetShardIDFromRegion(region string) string {\n\treturn regionByShards[region]\n}", "func NewGetProviderRegionByIDBadRequest() *GetProviderRegionByIDBadRequest {\n\treturn &GetProviderRegionByIDBadRequest{}\n}", "func (o PublicDelegatedPrefixPublicDelegatedSubPrefixResponseOutput) Region() pulumi.StringOutput {\n\treturn o.ApplyT(func(v PublicDelegatedPrefixPublicDelegatedSubPrefixResponse) string { return v.Region }).(pulumi.StringOutput)\n}", "func (a *AdditionalGUTI) GetAMFRegionID() (aMFRegionID uint8) {}", "func (v *VCard) Region() string {\n\treturn v.getFirstAddressField(4)\n}", "func GetRegionofIP(ipaddr string) (string, error) {\n\tif strings.Contains(ipaddr, \"[::1]\") {\n\t\treturn \"Uttarakhand\", nil\n\t}\n\n\tresp, err := http.Get(fmt.Sprintf(\"https://ipapi.co/%s/region/\", ipaddr))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\tdefer resp.Body.Close()\n\n\tb, err := ioutil.ReadAll(resp.Body)\n\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn string(b), err\n}", "func (m *VppToken) GetCountryOrRegion()(*string) {\n return m.countryOrRegion\n}", "func (o ApplicationOutput) LogicalRegionId() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Application) pulumi.StringPtrOutput { return v.LogicalRegionId }).(pulumi.StringPtrOutput)\n}", "func (o *ReplicaMapping) GetTargetRegionIdOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.TargetRegionId, true\n}", "func (f *Frontend) detectRegion(lang language.Tag, r *http.Request) language.Region {\n\treg, err := language.ParseRegion(strings.TrimSpace(r.FormValue(\"r\")))\n\tif err != nil {\n\t\treg, _ = lang.Region()\n\t}\n\n\treturn reg.Canonicalize()\n}", "func getRegion(name string) (aws.Region, error) {\n\tvar regions = map[string]aws.Region{\n\t\taws.APNortheast.Name: aws.APNortheast,\n\t\taws.APSoutheast.Name: aws.APSoutheast,\n\t\taws.APSoutheast2.Name: aws.APSoutheast2,\n\t\taws.EUCentral.Name: aws.EUCentral,\n\t\taws.EUWest.Name: aws.EUWest,\n\t\taws.USEast.Name: aws.USEast,\n\t\taws.USWest.Name: aws.USWest,\n\t\taws.USWest2.Name: aws.USWest2,\n\t\taws.USGovWest.Name: aws.USGovWest,\n\t\taws.SAEast.Name: aws.SAEast,\n\t}\n\tregion, ok := regions[name]\n\tif !ok {\n\t\treturn aws.Region{}, fmt.Errorf(\"No region matches %s\", name)\n\t}\n\treturn region, nil\n}", "func (r *UpdateListenerRequest) SetRegionId(regionId string) {\n r.RegionId = regionId\n}", "func (o *ReplicaMapping) GetSourceRegionIdOk() (*string, bool) {\n\tif o == nil {\n\t\treturn nil, false\n\t}\n\treturn &o.SourceRegionId, true\n}", "func getAwsRegionE() (region string, err error) {\n\n\tif os.Getenv(\"AWS_DEFAULT_REGION\") != \"\" {\n\t\tregion = os.Getenv(\"AWS_DEFAULT_REGION\")\n\t} else {\n\t\t// Grab it from this EC2 instace\n\t\tregion, err = ec2metadata.New(session.New()).Region()\n\t}\n\treturn\n}" ]
[ "0.7995377", "0.7943644", "0.791642", "0.7911354", "0.7890485", "0.77948856", "0.7780772", "0.777277", "0.7767606", "0.7762716", "0.7751107", "0.7685351", "0.76844555", "0.76183575", "0.75917506", "0.75853777", "0.75520635", "0.75412256", "0.7488131", "0.74709654", "0.7463794", "0.7462256", "0.7415943", "0.73814064", "0.7371244", "0.7367841", "0.7345188", "0.73432964", "0.73403054", "0.73211527", "0.7320389", "0.72889525", "0.7256485", "0.72488874", "0.7217192", "0.7205616", "0.7203658", "0.71601295", "0.71368575", "0.7114998", "0.70987767", "0.70733684", "0.7035486", "0.6973361", "0.69045633", "0.64936405", "0.6455567", "0.6414799", "0.6404857", "0.638722", "0.63233876", "0.6248989", "0.6241314", "0.62181604", "0.61905754", "0.6180508", "0.61461896", "0.6109462", "0.6049675", "0.6014335", "0.59911746", "0.5981517", "0.5979074", "0.5979063", "0.59618574", "0.5941423", "0.59164864", "0.58951646", "0.586568", "0.585639", "0.5842091", "0.5814401", "0.58114094", "0.5808264", "0.57987136", "0.5782437", "0.5748324", "0.57178247", "0.5717383", "0.56583935", "0.56543803", "0.5632223", "0.5616508", "0.5593543", "0.55902743", "0.55784947", "0.5565945", "0.553557", "0.55334306", "0.5521601", "0.5515392", "0.5513095", "0.5484462", "0.54732245", "0.5471264", "0.5450142", "0.544453", "0.5440043", "0.54385626", "0.5423792" ]
0.777523
7
Proverb should have a comment documenting it.
func Proverb(rhyme []string) []string { results := make([]string, len(rhyme)) if len(rhyme) == 0 { return []string{} } else if len(rhyme) == 1 { results[0] = "And all for the want of a " + rhyme[0] + "." return results } for i := 0; i < len(rhyme)-1; i++ { results[i] = "For want of a " + rhyme[i] + " the " + rhyme[i+1] + " was lost." } results[len(rhyme)-1] = "And all for the want of a " + rhyme[0] + "." return results }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Proverb() string {\n\treturn quote.Concurrency()\n}", "func comment() {\r\n\t// ini komentar single line\r\n}", "func (un *Decoder) Comment(c xml.Comment) error { return nil }", "func _() {\n\t// Comment should end in a period\n}", "func pr(diff string) {\n\tif githubToken == \"\" {\n\t\tlog.Printf(\"In order to add request comment, set the GITHUB_TOKEN input.\")\n\t\treturn\n\t}\n\n\tbody := \"[Goaction](https://github.com/posener/goaction) will apply the following changes after PR is merged.\\n\\n\" + diff\n\tif diff == \"\" {\n\t\tbody = \"[Goaction](https://github.com/posener/goaction) detected no required changes to Github action files.\"\n\t}\n\n\tctx := context.Background()\n\terr := actionutil.PRComment(ctx, githubToken, body)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func (t *liverpc) sectionComment(sectionTitle string) {\n\tt.P()\n\tt.P(`// `, strings.Repeat(\"=\", len(sectionTitle)))\n\tt.P(`// `, sectionTitle)\n\tt.P(`// `, strings.Repeat(\"=\", len(sectionTitle)))\n\tt.P()\n}", "func (*AddCommentReq) Descriptor() ([]byte, []int) {\n\treturn file_proto_comment_comment_proto_rawDescGZIP(), []int{6}\n}", "func PulumiComment(kind string) string {\n\tconst prefix = \"\\n\\n\"\n\n\tk := kinds.Kind(kind)\n\tswitch k {\n\tcase kinds.Deployment, kinds.Ingress, kinds.Pod, kinds.Service, kinds.StatefulSet:\n\t\treturn prefix + awaitComments(k)\n\tcase kinds.Job:\n\t\treturn prefix + awaitComments(k) + prefix + replaceUnreadyComment()\n\tcase kinds.Secret:\n\t\treturn prefix + helpfulLinkComments(k)\n\tdefault:\n\t\treturn \"\"\n\t}\n}", "func TestGoDocCommentBreakPass(t *testing.T) {\n\taccept(t, \"godoc_test.go\", \"TestGoDocCommentBreak\")\n}", "func (s *BasePCREListener) EnterComment(ctx *CommentContext) {}", "func (r *Resolver) Comment() generated.CommentResolver { return &commentResolver{r} }", "func (r *Resolver) Comment() generated.CommentResolver { return &commentResolver{r} }", "func pr(diff string) {\n\tif githubToken == \"\" {\n\t\tlog.Printf(\"In order to add request comment, set the GITHUB_TOKEN input.\")\n\t\treturn\n\t}\n\n\tbody := \"[goreadme](https://github.com/posener/goreadme) will not make any changes in this PR\"\n\tif diff != \"\" {\n\t\tbody = fmt.Sprintf(\n\t\t\t\"[goreadme](https://github.com/posener/goreadme) diff for %s file for this PR:\\n\\n%s\",\n\t\t\tpath,\n\t\t\tdiff)\n\t}\n\n\tctx := context.Background()\n\terr := actionutil.PRComment(ctx, githubToken, body)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n}", "func (*Comment) Descriptor() ([]byte, []int) {\n\treturn file_proto_comment_comment_proto_rawDescGZIP(), []int{0}\n}", "func (*Comment) Descriptor() ([]byte, []int) {\n\treturn file_proto_comments_proto_rawDescGZIP(), []int{0}\n}", "func Proverb(rhyme []string) []string {\n\toutput := []string{}\n\tif len(rhyme) == 0 {\n\t\treturn output\n\t}\n\tif len(rhyme) >= 2 {\n\t\tfor i := 0; i < len(rhyme)-1; i++ {\n\t\t\toutput = append(output, fmt.Sprintf(\"For want of a %s the %s was lost.\", rhyme[i], rhyme[i+1]))\n\t\t}\n\t}\n\toutput = append(output, fmt.Sprintf(\"And all for the want of a %s.\", rhyme[0]))\n\treturn output\n}", "func (env *Env) Comment(val Value, docs ...string) {\n\tenv.Commentary = append(env.Commentary, env.doc(val, docs...))\n}", "func (tp *TestParams) Comment() string {\n\treturn tp.Cmt\n}", "func main() {\n\t// Single line comment, commonly for code\n}", "func (*CreateCommentRequest_Comment) Descriptor() ([]byte, []int) {\n\treturn file_article_proto_rawDescGZIP(), []int{10, 0}\n}", "func (f *Formatter) printDoc(v proto.Visitee) {\n\tif hasDoc, ok := v.(proto.Documented); ok {\n\t\tif doc := hasDoc.Doc(); doc != nil {\n\t\t\tf.printComment(doc)\n\t\t}\n\t}\n}", "func Proverb(rhyme []string) (proverb []string) {\n\tfor i := 1; i < len(rhyme); i++ {\n\t\tproverb = append(proverb, fmt.Sprintf(PROVERB, rhyme[i-1], rhyme[i]))\n\t}\n\tif len(rhyme) > 0 {\n\t\tproverb = append(proverb, fmt.Sprintf(CONCLUSION, rhyme[0]))\n\t}\n\treturn proverb\n}", "func (vd *Verb_Delegate) Description() string {\n\treturn \"the method will delegate to another object.\"\n}", "func (n *Mod) Doc(text string) *Mod {\n\tn.Obj.ObjComment = NewComment(text)\n\tn.Obj.ObjComment.ObjParent = n\n\treturn n\n}", "func (*ListingComment_Comment) Descriptor() ([]byte, []int) {\n\treturn file_listing_comments_comments_proto_rawDescGZIP(), []int{1, 0}\n}", "func Verb(msg string) {\n\tif lvl <= ver {\n\t\tl.Print(\"[VERB ]: \" + msg)\n\t}\n}", "func GenComment(decl *ast.GenDecl) string {\n\tif decl.Doc != nil {\n\t\treturn decl.Doc.Text()\n\t}\n\treturn \"\"\n}", "func (s *prometheusruleWebhook) Doc() string {\n\treturn (docString)\n}", "func (*Comment) Descriptor() ([]byte, []int) {\n\treturn file_article_proto_rawDescGZIP(), []int{1}\n}", "func (p *plugin) printComment(comments ...string) {\n\n\tcomment := strings.Join(comments, \" \")\n\n\tif comment == \"\" {\n\t\treturn\n\t}\n\n\ttext := strings.TrimSuffix(comment, \"\\n\")\n\tfor _, line := range strings.Split(text, \"\\n\") {\n\t\tp.P(\"// \", strings.TrimPrefix(line, \" \"))\n\t}\n\n}", "func (*CreateCommentRequest) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{21}\n}", "func Proverb(rhyme []string) []string {\n\tparts := make([]string, 0, len(rhyme))\n\n\tfor i := 0; i < len(rhyme)-1; i++ {\n\t\tparts = append(parts, fmt.Sprintf(\"For want of a %s the %s was lost.\", rhyme[i], rhyme[i+1]))\n\t}\n\n\tif len(rhyme) > 0 {\n\t\tparts = append(parts, fmt.Sprintf(\"And all for the want of a %s.\", rhyme[0]))\n\t}\n\n\treturn parts\n}", "func (*Approve) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{4}\n}", "func (*CreateCommentResponse) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{22}\n}", "func (*GetCommentRequest) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{23}\n}", "func (p *Parser) parseComment() {\n p.lex.SkipToNewLine()\n}", "func (p *Doc) Comment(key, comments string) bool {\n\te, ok := p.props[key]\n\tif !ok {\n\t\treturn false\n\t}\n\n\t// 如果所有注释为空\n\tif comments == \"\" {\n\t\tp.lines.InsertBefore(&line{typo: '#', value: \"#\"}, e)\n\t\treturn true\n\t}\n\n\t// 创建一个新的Scanner\n\tscanner := bufio.NewScanner(strings.NewReader(comments))\n\tfor scanner.Scan() {\n\t\tp.lines.InsertBefore(&line{typo: '#', value: \"#\" + scanner.Text()}, e)\n\t}\n\n\treturn true\n}", "func Proverb(rhyme []string) []string {\n\tproverbs := make([]string, len(rhyme))\n\n\tfor i, phrase := range rhyme {\n\t\tif i == 0 {\n\t\t\tproverbs[len(rhyme)-1] = fmt.Sprintf(\"And all for the want of a %s.\", phrase)\n\t\t}\n\t\tif i+1 < len(rhyme) {\n\t\t\tproverbs[i] = fmt.Sprintf(\"For want of a %s the %s was lost.\", phrase, rhyme[i+1])\n\t\t}\n\t}\n\n\treturn proverbs\n}", "func (t *tester) commentBeginning(ctx context.Context, change *gerrit.ChangeInfo) error {\n\t// It would be nice to do a similar thing to the coordinator, using comment\n\t// threads that can be resolved, but that is slightly more complex than what\n\t// we really need to start with.\n\t//\n\t// Similarly it would be nice to comment links to logs earlier.\n\treturn t.gerrit.SetReview(ctx, change.ID, change.CurrentRevision, gerrit.ReviewInput{\n\t\tMessage: \"TryBots beginning\",\n\t})\n}", "func (*PresentProofRequest) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{37}\n}", "func (sm *StateMachine) Comment(comment string) *StateMachine {\n\tsm.comment = comment\n\treturn sm\n}", "func (me TAttlistCommentsCorrectionsRefType) IsCommentOn() bool { return me.String() == \"CommentOn\" }", "func Proverb(rhyme []string) (proverb []string) {\n\tproverb = make([]string, 0, len(rhyme))\n\tif len(rhyme) > 0 {\n\t\tproverb = append(proverb, fmt.Sprintf(\"And all for the want of a %s.\", rhyme[0]))\n\t}\n\tif len(rhyme) > 1 {\n\t\tlines := make([]string, 0, len(rhyme)-1)\n\t\titer := zipSlices(rhyme[:len(rhyme)-1], rhyme[1:])\n\t\tfor tuple := iter(); tuple != nil; tuple = iter() {\n\t\t\tlines = append(lines, fmt.Sprintf(\"For want of a %s the %s was lost.\", tuple[0], tuple[1]))\n\t\t}\n\t\tproverb = append(lines, proverb...)\n\t}\n\n\treturn proverb\n}", "func (*SimpleRepeatedDocument) Descriptor() ([]byte, []int) {\n\treturn file_examples_documents_example_proto_rawDescGZIP(), []int{15}\n}", "func (o *Poke) Doc() string {\n\n\treturn `When available, poke can be used to update various information about the parent.\nFor instance, for enforcers, poke will be used as the heartbeat.`\n}", "func Proverb(nouns []string) (rhymes []string) {\n\n\tif len(nouns) == 0 {\n\t\treturn rhymes\n\t}\n\n\trhymes = make([]string, len(nouns))\n\tfor i := 0; i < len(nouns)-1; i++ {\n\t\trhymes[i] = fmt.Sprintf(\"For want of a %s the %s was lost.\", nouns[i], nouns[i+1])\n\t}\n\n\trhymes[len(nouns)-1] = fmt.Sprintf(\"And all for the want of a %s.\", nouns[0])\n\treturn\n}", "func (*CreateCommentRequest) Descriptor() ([]byte, []int) {\n\treturn file_article_proto_rawDescGZIP(), []int{10}\n}", "func (ge *GollumEvent) Comment() string {\n\treturn \"\"\n}", "func (c VerifyCmd) Synopsis() string {\n\treturn \"Check if gtm satisfies a Semantic Version 2.0 constraint\"\n}", "func (_ AdverbProcessor) Process(c *Chunk) *Chunk {\n\tmsg := \"This is an adverb.\"\n\treturn doTextProcessor(proc.AdverbProcessor(), \"adverb\", c, msg)\n}", "func (dw *DrawingWand) Comment(comment string) {\n\tcscomment := C.CString(comment)\n\tdefer C.free(unsafe.Pointer(cscomment))\n\tC.MagickDrawComment(dw.dw, cscomment)\n}", "func (ctx *ShowCommentContext) OK(r *GoaComment) error {\n\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.goa.comment+json\")\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func Proverb(rhyme []string) (result []string) {\n\tnumWords := len(rhyme)\n\tresult = make([]string, numWords)\n\n\tfor i := 0; i < numWords-1; i++ {\n\t\tresult[i] = fmt.Sprintf(\"For want of a %s the %s was lost.\", rhyme[i], rhyme[i+1])\n\t}\n\tif numWords > 0 {\n\t\tresult[numWords-1] = fmt.Sprintf(\"And all for the want of a %s.\", rhyme[0])\n\t}\n\treturn\n}", "func (*GetCommentsByProductIdReq) Descriptor() ([]byte, []int) {\n\treturn file_proto_comment_comment_proto_rawDescGZIP(), []int{2}\n}", "func Comment(ctx context.Context, cfg *v1.Config, pr int, contents []byte) error {\n\tc := newClient(ctx, cfg.Github)\n\treturn c.CommentOnPR(pr, string(contents))\n}", "func (*ListingComment) Descriptor() ([]byte, []int) {\n\treturn file_listing_comments_comments_proto_rawDescGZIP(), []int{1}\n}", "func (*ArticleComment) Descriptor() ([]byte, []int) {\n\treturn file_proto_articleService_proto_rawDescGZIP(), []int{2}\n}", "func Proverb(rhyme []string) []string {\n\tproverbLen := len(rhyme)\n\tproverb := make([]string, proverbLen)\n\n\tif proverbLen == 0 {\n\t\treturn proverb\n\t}\n\n\tif proverbLen > 1 {\n\t\tfor i := 0; i < proverbLen-1; i++ {\n\t\t\tproverb[i] = fmt.Sprintf(\"For want of a %v the %v was lost.\", rhyme[i], rhyme[i+1])\n\t\t}\n\t}\n\n\tproverb[proverbLen-1] = fmt.Sprintf(\"And all for the want of a %v.\", rhyme[0])\n\n\treturn proverb\n}", "func (c *Command) Synopsis() string {\n\treturn \"\"\n}", "func (*Proof) Descriptor() ([]byte, []int) {\n\treturn file_proof_proto_rawDescGZIP(), []int{1}\n}", "func (*CommentResponse) Descriptor() ([]byte, []int) {\n\treturn file_article_proto_rawDescGZIP(), []int{16}\n}", "func (*UpdateCommentRequest) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{25}\n}", "func Proverb(rhyme []string) []string {\n\n\tsize := len(rhyme)\n\tproverb := make([]string, size)\n\tif size == 0 {\n\t\treturn proverb\n\t}\n\n\tfor i := 0; i < size; i++ {\n\t\tif i == size-1 {\n\t\t\tproverb[i] = \"And all for the want of a \" + rhyme[0] + \".\"\n\t\t} else {\n\t\t\tproverb[i] = \"For want of a \" + rhyme[i] + \" the \" + rhyme[i+1] + \" was lost.\"\n\t\t}\n\t}\n\n\treturn proverb\n}", "func Proverb(rhyme []string) []string {\n\tlength := len(rhyme)\n\trhymes := []string{}\n\n\tswitch length {\n\tcase 0:\n\t\treturn []string{}\n\tcase 1:\n\t\t\"And all for the want of a \"\n\t}\n\n\tfor index, item := range rhyme {\n\t\t// multiple strings, all but last like\n\t\t// \"For want of a <item> the <next-item> was lost.\"\n\t\t// last string like \"And all for the want of a <first-item>.\"\n\t\t// no input defaults to empty string?\n\t}\n}", "func (*Question) Descriptor() ([]byte, []int) {\n\treturn file_survey_proto_rawDescGZIP(), []int{1}\n}", "func main() {\n\t// fmt.Println(\"Golang is bad\")\n\tfmt.Println(\"Hello World\") // comment after statement\n}", "func (*UpdateCommentResponse) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{26}\n}", "func (*ProofRequest) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{35}\n}", "func (*DeleteCommentRequest) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{27}\n}", "func (o *APICheck) Doc() string {\n\n\treturn `Allows you to verify if a client identified by his token is allowed to do\nsome operations on some APIs.`\n}", "func (*CMsgDOTATriviaQuestion) Descriptor() ([]byte, []int) {\n\treturn file_dota_gcmessages_common_proto_rawDescGZIP(), []int{66}\n}", "func (*Resp) Descriptor() ([]byte, []int) {\n\treturn file_proto_comment_comment_proto_rawDescGZIP(), []int{1}\n}", "func (*GetCommentByOrderIdReq) Descriptor() ([]byte, []int) {\n\treturn file_proto_comment_comment_proto_rawDescGZIP(), []int{4}\n}", "func (prrce *PullRequestReviewCommentEvent) Say() string {\n\tcomment := prrce.Raw().Payload[\"comment\"].(map[string]interface{})\n\tfile := comment[\"path\"].(string)\n\tpr := prrce.Raw().Payload[\"pull_request\"].(map[string]interface{})\n\tprTitle := pr[\"title\"]\n\treturn fmt.Sprint(\"#{actor} just commented on a pull request with the title: \", prTitle, \",\\nThe comment was in the file: \", file, \",\\nHere's the comment: #{comment}\")\n}", "func (p *Package) Synopsis(text string) string {\n\ttext = firstSentence(text)\n\tlower := strings.ToLower(text)\n\tfor _, prefix := range IllegalPrefixes {\n\t\tif strings.HasPrefix(lower, prefix) {\n\t\t\treturn \"\"\n\t\t}\n\t}\n\tpr := p.Printer()\n\tpr.TextWidth = -1\n\td := p.Parser().Parse(text)\n\tif len(d.Content) == 0 {\n\t\treturn \"\"\n\t}\n\tif _, ok := d.Content[0].(*comment.Paragraph); !ok {\n\t\treturn \"\"\n\t}\n\td.Content = d.Content[:1] // might be blank lines, code blocks, etc in “first sentence”\n\treturn strings.TrimSpace(string(pr.Text(d)))\n}", "func (*AnchorDocumentRequest) Descriptor() ([]byte, []int) {\n\treturn file_p2p_p2p_proto_rawDescGZIP(), []int{4}\n}", "func (o *EnforcerReport) Doc() string {\n\n\treturn `Post a new defender statistics report.`\n}", "func (*GetCommentResponse) Descriptor() ([]byte, []int) {\n\treturn file_view_grpc_blog_api_proto_rawDescGZIP(), []int{24}\n}", "func (this *Router) Comment(comment string) *Router {\n\tthis.comment = comment\n\treturn this\n}", "func (*MoveAcknowledgment) Descriptor() ([]byte, []int) {\n\treturn file_FillerGame_proto_rawDescGZIP(), []int{4}\n}", "func (*CMsgDraftTrivia) Descriptor() ([]byte, []int) {\n\treturn file_dota_gcmessages_client_proto_rawDescGZIP(), []int{385}\n}", "func (c *ServeCommand) Synopsis() string {\n\treturn \"Spored Server\"\n}", "func exported_to_c_with_comment() {}", "func (r *Distribution) Comment() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"comment\"])\n}", "func (*Snippet) Descriptor() ([]byte, []int) {\n\treturn file_api_ocp_snippet_api_ocp_snippet_api_proto_rawDescGZIP(), []int{14}\n}", "func (*CommentsResponse) Descriptor() ([]byte, []int) {\n\treturn file_article_proto_rawDescGZIP(), []int{17}\n}", "func (*DocumentHighlightRequest) Descriptor() ([]byte, []int) {\n\treturn file_protocol_rpc_rpc_proto_rawDescGZIP(), []int{146}\n}", "func (*PresentProofResponse) Descriptor() ([]byte, []int) {\n\treturn file_messages_proto_rawDescGZIP(), []int{38}\n}", "func (*TalkVoice) Descriptor() ([]byte, []int) {\n\treturn file_msgdata_proto_rawDescGZIP(), []int{26}\n}", "func (c *RunCommand) Synopsis() string {\n\treturn \"コンパイル後、テストを実行する\"\n}", "func (*Prescription) Descriptor() ([]byte, []int) {\n\treturn file_heyrenee_v1_messages_prescription_proto_rawDescGZIP(), []int{0}\n}", "func (*ManualCpm) Descriptor() ([]byte, []int) {\n\treturn file_google_ads_googleads_v14_common_bidding_proto_rawDescGZIP(), []int{4}\n}", "func (opt OptComment) Option(d *bson.Document) error {\n\td.Append(bson.EC.String(\"comment\", string(opt)))\n\treturn nil\n}", "func (*ManualCpa) Descriptor() ([]byte, []int) {\n\treturn file_google_ads_googleads_v14_common_bidding_proto_rawDescGZIP(), []int{2}\n}", "func (*CMsgDOTASubmitTriviaQuestionAnswer) Descriptor() ([]byte, []int) {\n\treturn file_dota_gcmessages_client_proto_rawDescGZIP(), []int{235}\n}", "func (*Note) Descriptor() ([]byte, []int) {\n\treturn file_determined_project_v1_project_proto_rawDescGZIP(), []int{1}\n}", "func (*SchnorrProofPb) Descriptor() ([]byte, []int) {\n\treturn file_atse_proto_rawDescGZIP(), []int{3}\n}", "func (*CMsgDOTABotDebugInfo_Bot_Action) Descriptor() ([]byte, []int) {\n\treturn file_dota_gcmessages_common_proto_rawDescGZIP(), []int{41, 0, 1}\n}", "func (o LookupResponseHeadersPolicyResultOutput) Comment() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupResponseHeadersPolicyResult) string { return v.Comment }).(pulumi.StringOutput)\n}", "func (*Review) Descriptor() ([]byte, []int) {\n\treturn file_reviews_proto_rawDescGZIP(), []int{1}\n}" ]
[ "0.57563734", "0.5630922", "0.5606449", "0.55619377", "0.54658043", "0.54651946", "0.5421217", "0.5368903", "0.5354967", "0.53499293", "0.5338456", "0.5338456", "0.5326887", "0.5319632", "0.5307838", "0.529722", "0.52923405", "0.5290288", "0.52884936", "0.5268125", "0.5220723", "0.5218646", "0.5208466", "0.51970804", "0.5181834", "0.5180591", "0.5164736", "0.5162604", "0.51619184", "0.51603484", "0.51435035", "0.51286864", "0.5109863", "0.51095545", "0.50969255", "0.5095463", "0.50953835", "0.50878036", "0.5083413", "0.50771326", "0.507497", "0.50726104", "0.5069607", "0.50603336", "0.50577575", "0.5053144", "0.50459397", "0.50441515", "0.50438714", "0.5032576", "0.5025138", "0.50218105", "0.50212896", "0.5016855", "0.5010254", "0.50059366", "0.50033695", "0.5002187", "0.49979833", "0.49884966", "0.49861047", "0.49860513", "0.4982695", "0.49686313", "0.49603078", "0.4955565", "0.49553615", "0.49508804", "0.49432474", "0.49307784", "0.49271637", "0.4921886", "0.49206072", "0.49165314", "0.4908579", "0.49010533", "0.48984298", "0.48953447", "0.48928255", "0.48837003", "0.48793253", "0.48689505", "0.48664874", "0.48553532", "0.48526677", "0.48487872", "0.4846771", "0.48426402", "0.4837636", "0.48367977", "0.48360726", "0.4835687", "0.48319155", "0.48233157", "0.48207602", "0.4820688", "0.4811415", "0.48097938", "0.48072702", "0.48009688" ]
0.50858814
38
IsIsogram takes a string and returns a boolen if the string is an isogram
func IsIsogram(word string) bool { letters := make(map[rune]bool) for _, val := range word { if !unicode.IsLetter(val) { continue } if _, ok := letters[unicode.ToUpper(val)]; ok { return false } letters[unicode.ToUpper(val)] = true } return true }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func IsIsogram(str string) bool {\n\tr := strings.NewReplacer(\"-\", \"\", \" \", \"\")\n\tstr = r.Replace(str)\n\tstr = strings.ToLower(str)\n\n\tchars := map[string]bool{}\n\tfor _, x := range str {\n\t\tchars[string(x)] = true\n\t}\n\n\treturn len(str) == len(chars)\n}", "func IsIsogram(s string) bool {\n\tletters := map[rune]bool{}\n\tfor _, r := range strings.ToLower(s) {\n\t\tif unicode.IsLetter(r) {\n\t\t\tif letters[r] {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tletters[r] = true\n\t\t}\n\t}\n\treturn true\n}", "func IsIsogram(s string) bool {\n\tif s == \"\" {\n\t\treturn true\n\t}\n\n\tm := make(map[rune]bool)\n\n\tfor _, c := range s {\n\t\tif c == '-' || c == ' ' {\n\t\t\tcontinue\n\t\t}\n\n\t\tif m[c] {\n\t\t\treturn false\n\t\t}\n\n\t\tm[unicode.ToLower(c)] = true\n\t}\n\n\treturn true\n}", "func IsIsogram(in string) bool {\n\toutputCount := 0\n\tfor i := 0; i < len(in); i++ {\n\t\tc := strings.ToLower(string(in[i]))\n\t\toutputCount = 0\n\t\tfor j := 0; j < len(in); j++ {\n\t\t\tif c == strings.ToLower(string(in[j])) {\n\t\t\t\tif c == \"-\" || c == \" \" {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\toutputCount++\n\t\t\t\tif outputCount > 1 {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\treturn true\n\n}", "func IsIsogram(input string) bool {\n\tif input == \"\" {\n\t\treturn true\n\t}\n\n\tinput = strings.ToLower(input)\n\tfor i, el := range input {\n\t\tif strings.LastIndexByte(strings.ToLower(input), input[i]) != i {\n\t\t\tif unicode.IsLetter(el) != false {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\treturn true\n}", "func IsIsogram(s string) bool {\n\ta := []rune(s)\n\tm := make(map[rune]bool, len(s))\n\n\tfor _, l := range a {\n\t\tr := unicode.ToLower(l)\n\t\tif unicode.IsLetter(r) && m[r] {\n\t\t\treturn false\n\t\t}\n\t\tm[r] = true\n\t}\n\n\treturn true\n}", "func IsIsogram(input string) bool {\n\trunes := make(map[rune]bool)\n\tfor _, c := range input {\n\t\tif c == '-' || c == ' ' {\n\t\t\tcontinue\n\t\t}\n\t\tif runes[c] {\n\t\t\treturn false\n\t\t}\n\t\tc = unicode.ToUpper(c)\n\t\trunes[c] = true\n\t}\n\treturn true\n}", "func IsIsogram (s string)bool{\n\ts = strings.ToLower(s)\n\n\tfor i, j := range s {\n\t\tif unicode.IsLetter(j) && strings.ContainsRune(s[i+1:], j){\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsIsogram(text string) bool {\n\tr := make(map[rune]bool)\n\tfor _, c := range strings.ToLower(text) {\n\t\tif !unicode.IsLetter(c) {\n\t\t\tcontinue\n\t\t}\n\t\tif r[c] {\n\t\t\treturn false\n\t\t}\n\t\tr[c] = true\n\t}\n\treturn true\n}", "func IsIsogram(s string) bool {\n\tvar chars = map[rune]int{}\n\tfor _, r := range strings.ToLower(s) {\n\t\tif strings.Contains(\" -\", string(r)) {\n\t\t\tcontinue\n\t\t}\n\t\tif chars[r] == 1 {\n\t\t\treturn false\n\t\t}\n\t\tchars[r] = 1\n\t}\n\treturn true\n}", "func IsIsogram(input string) bool {\n\tif input == \"\" {\n\t\treturn true\n\t}\n\tinput = strings.ToLower(input)\n\tfor i, char := range input {\n\t\tlastIndex := strings.LastIndex(input, string(char))\n\t\tif lastIndex > i == true && unicode.IsLetter(char) == true {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsIsogram(s string) bool {\n\tappeared := make(map[rune]bool)\n\tfor _, c := range s {\n\t\tif !unicode.IsLetter(c) {\n\t\t\tcontinue\n\t\t}\n\t\tnormalizedChar := unicode.ToLower(c)\n\t\tif appeared[normalizedChar] {\n\t\t\treturn false\n\t\t}\n\t\tappeared[normalizedChar] = true\n\t}\n\treturn true\n}", "func IsIsogram(input string) bool {\n\tseenChars := map[rune]Void{}\n\tfor _, char := range input {\n\t\tif char == '-' || char == ' ' {\n\t\t\tcontinue\n\t\t}\n\t\tlowerChar := unicode.ToLower(char)\n\t\tif _, present := seenChars[lowerChar]; present {\n\t\t\treturn false\n\t\t}\n\t\tseenChars[lowerChar] = Void{}\n\t}\n\treturn true\n}", "func IsIsogram(input string) bool {\n\tletters := make(map[rune]bool)\n\n\tfor _, c := range input {\n\t\tc = unicode.ToLower(c)\n\n\t\tif c == '-' || c == ' ' {\n\t\t\tcontinue\n\t\t}\n\n\t\tif letters[c] {\n\t\t\treturn false\n\t\t}\n\n\t\tletters[c] = true\n\t}\n\n\treturn true\n}", "func IsIsogram(text string) bool {\n\tletters := map[rune]bool{}\n\n\tfor _, letter := range strings.ToLower(text) {\n\t\tif !unicode.IsLetter(letter) {\n\t\t\tcontinue\n\t\t}\n\t\tif letters[letter] {\n\t\t\treturn false\n\t\t}\n\t\tletters[letter] = true\n\t}\n\n\treturn true\n}", "func IsIsogram(word string) bool {\n\tcharexists := map[rune]bool{}\n\tfor _, char := range strings.ToLower(word) {\n\t\tif unicode.IsLetter(char) && charexists[char] {\n\t\t\treturn false\n\t\t}\n\t\tcharexists[char] = true\n\t}\n\treturn true\n}", "func IsIsogram(word string) bool {\n\tletters := filter(word)\n\tsort.Sort(letters)\n\n\tfor index := 0; index < len(letters)-1; index++ {\n\t\tif letters[index] == letters[index+1] {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func IsIsogram(text string) bool {\n\tletters := make(map[rune]bool, len(text))\n\n\tfor _, r := range text {\n\t\tif ignoredRune(r) {\n\t\t\tcontinue\n\t\t}\n\n\t\tr = unicode.ToLower(r)\n\n\t\tif letters[r] {\n\t\t\treturn false\n\t\t}\n\n\t\tletters[r] = true\n\t}\n\n\treturn true\n}", "func IsIsogram(word string) bool {\n\tvar counter = map[rune]bool{}\n\tfor _, letter := range word {\n\t\tletter = unicode.ToLower(letter)\n\t\tif unicode.IsLetter(letter) && counter[letter] == true {\n\t\t\treturn false\n\t\t}\n\t\tcounter[letter] = true\n\t}\n\treturn true\n}", "func IsIsogram(word string) bool {\n\tcountChars := make(map[rune]int)\n\tfor _, value := range word {\n\t\tif !unicode.IsLetter(value) {\n\t\t\tcontinue\n\t\t}\n\t\tvalue = unicode.ToLower(value)\n\t\tif countChars[value] >= 1 {\n\t\t\treturn false\n\t\t}\n\t\tcountChars[value]++\n\t}\n\treturn true\n}", "func IsIsogram(word string) bool {\n\n\tif len(word) == 0 {\n\t\treturn true // empty string is an isogram\n\t}\n\n\t// Create a Dynamic memory for the Counter Map\n\tletter_count_map := make(map[string]int)\n\n\t// Scrub off all `-` from the words\n\tscrubbed_word := strings.ReplaceAll(strings.ToLower(word), \"-\", \"\")\n\n\t// Remove spaces between words\n\tscrubbed_word = strings.ReplaceAll(scrubbed_word, \" \", \"\")\n\n\t// Iterate over all letters in the scrubbed word and create a count map\n\tfor _, each_letter := range scrubbed_word {\n\t\t// create a map of each letter and the count of occurence\n\t\tletter_count_map[string(each_letter)] = strings.Count(scrubbed_word, string(each_letter))\n\t}\n\n\t// parse through all the counted values and check if all are unique\n\tfor _, count := range letter_count_map {\n\t\tif count > 1 {\n\t\t\treturn false // if any letter occurs more than once -> not an isogram\n\t\t}\n\t}\n\treturn true\n}", "func IsIsogram(word string) bool {\n\tvar characters = map[rune]bool{}\n\n\tfor _, letter := range word {\n\t\tif !unicode.IsLetter(letter) {\n\t\t\tcontinue\n\t\t}\n\n\t\tb := unicode.ToUpper(letter)\n\t\tif characters[b] != false {\n\t\t\t// We have seen this letter before, early return\n\t\t\treturn false\n\t\t}\n\t\tcharacters[b] = true\n\t}\n\n\t// If we reached here, we haven't seen any character twice\n\t// which means that the word is an isogram.\n\treturn true\n}", "func IsIsogram(sentence string) bool {\n\tvar uniqueChars = make(map[rune]bool)\n\tvar numberOfUniqueChars = 0\n\tvar sentenceChars = charsOnlyFrom(sentence)\n\tfor _, char := range sentenceChars {\n\t\tvar lowerChar = unicode.ToLower(char)\n\t\tif uniqueChars[lowerChar] == false {\n\t\t\tnumberOfUniqueChars++\n\t\t\tuniqueChars[lowerChar] = true\n\t\t}\n\t}\n\n\treturn len(sentenceChars) == numberOfUniqueChars\n}", "func IsIsogram(word string) bool {\n\tseen := make(map[rune]int)\n\tfor _, ch := range word {\n\t\tif ch == '-' || ch == ' ' {\n\t\t\tcontinue\n\t\t}\n\t\tnewCh := unicode.ToLower(ch)\n\t\tif _, ok := seen[newCh]; ok {\n\t\t\tseen[newCh] += 1\n\t\t} else {\n\t\t\tseen[newCh] = 1\n\t\t}\n\t}\n\tfor _, count := range seen {\n\t\tif count > 1 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsIsogram(word string) bool {\n\t// declare empty arr\n\tif word == \"\" {\n\t\treturn true\n\t}\n\tvar letters []string\n\n\t// initialize empty letters array\n\t// for every character in the word\n\t// \t\tgo through every element in the array of letters\n\t// \t\t\tif the character in the word already exists in letters array\n\t// \t\t\t\treturn false\n\t// \t\tappend the character in the word to the letters array\n\n\tfor i := 0; i < len(word); i++ {\n\t\tchar := strings.ToLower(string(word[i]))\n\t\t// check if the letter is in the letters array already\n\t\tfor j := 0; j < len(letters); j++ {\n\t\t\tletterChar := strings.ToLower(string(letters[j]))\n\t\t\tif letterChar == char {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t\tif char == \"-\" || char == \" \" {\n\t\t\tcontinue\n\t\t}\n\t\tletters = append(letters, char)\n\t}\n\t// check if that char already exists in arr\n\t// return False\n\t// store the char in an arr\n\n\treturn true\n}", "func IsIsogram(input string) bool {\n\tmaps = make(map[rune]int)\n\tfor _, r := range input {\n\t\tvar char = unicode.ToUpper(r)\n\n\t\tif maps[char] > 0 {\n\t\t\treturn false\n\t\t}\n\n\t\tif char >= 65 && char <= 90 {\n\t\t\tmaps[char] = maps[char] + 1\n\t\t}\n\t}\n\treturn true\n}", "func IsIsogram(input string) bool {\n\ninput = strings.ToLower(input)\n\nindex := make(map[string]bool)\n\nfor i:= 0 ; i < len(input) ; i++ {\n //skip any non alphabets\n if !strings.ContainsAny(string(input[i]),\"abcdefghijklmnopqrstuvwxyz\") {\n continue;\n }\n _,letterFound:=index[string(input[i])]\n if letterFound {\n return false\n } else {\n index[string(input[i])] = true\n }\n}\n return true\n}", "func IsIsogramNested(input string) bool {\n\n\tfor i := range input {\n\t\tfor j := range input {\n\t\t\tif i == j || !unicode.IsLetter(rune(input[i])) || !unicode.IsLetter(rune(input[j])) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif unicode.ToLower(rune(input[i])) == unicode.ToLower(rune(input[j])) {\n\t\t\t\treturn false\n\t\t\t}\n\t\t}\n\t}\n\n\treturn true\n}", "func isImagStr(s string) bool {\n\treturn strings.Contains(s, \"i\") && s[len(s)-1] == 'i'\n}", "func IsUnique(s string) bool {\n\tseen := make([]int, 8) // every bit represent a character\n\tfor _, c := range s {\n\t\tif isSet(byte(c), seen) {\n\t\t\treturn false\n\t\t}\n\t\tset(byte(c), seen)\n\t}\n\treturn true\n}", "func IsUnique(str string) bool {\n\tif len(str) > 256 {\n\t\treturn false\n\t}\n\tvar charMap [256]bool\n\tfor _, c := range str {\n\t\tif charMap[int(c)] == true {\n\t\t\treturn false\n\t\t}\n\t\tcharMap[int(c)] = true\n\t}\n\treturn true\n}", "func isExported(s string) bool {\n\tif len(s) == 0 {\n\t\treturn false\n\t}\n\tl := string(byte(s[0]))\n\treturn strings.ToUpper(l) == l\n}", "func isChannel(str string) bool {\n return len(str) != 0 && strings.Index(\"#\", str[0:1]) != -1;\n}", "func isHex(str string) bool {\n\tif hasHexPrefix(str) {\n\t\tstr = str[2:]\n\t}\n\tfor _, c := range []byte(str) {\n\t\tif !isHexCharacter(c) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsIMEI(str string) bool {\n\treturn rxIMEI.MatchString(str)\n}", "func isFasta(str string) (bool, error) {\n\tif strings.Contains(str, \".fa\") || strings.Contains(str, \".fasta\") {\n\t\t// file exists\n\t\t_, err := os.Stat(str);\n\t\tif err == nil {\n\t\t\treturn true, nil\n\t\t} \n\t\treturn false, err\n\t}\n\treturn false, nil\n}", "func IsShouting(remark string) bool {\n\ts, _ := regexp.MatchString(\"[A-Z]{2,}\", remark)\n\tr, _ := regexp.MatchString(\"[a-z]\", remark)\n\treturn s && (r == false)\n}", "func isIsomorphic(s string, t string) bool {\n\tif len(s) != len(t) {\n\t\treturn false\n\t}\n\thms2t := make(map[byte]byte)\n\thmt2s := make(map[byte]byte)\n\tfor i := 0; i < len(s); i++ {\n\t\tvs2t, oks2s := hms2t[s[i]]\n\t\tvt2s, okt2s := hmt2s[t[i]]\n\t\tif !oks2s && !okt2s {\n\t\t\thms2t[s[i]] = t[i]\n\t\t\thmt2s[t[i]] = s[i]\n\t\t} else if okt2s && oks2s {\n\t\t\tif vs2t != t[i] || vt2s != s[i] {\n\t\t\t\treturn false\n\t\t\t}\n\t\t} else {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsSHA1(str string) bool {\n\treturn IsHash(str, \"sha1\")\n}", "func isPng(s io.Reader) bool {\n\th := make([]byte, 8)\n\t_, err := io.ReadFull(s, h)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn string(h) == magic\n}", "func isExist(str, filepath string) bool {\n\taccused, _ := ioutil.ReadFile(filepath)\n\tisExist, _ := regexp.Match(str, accused)\n\treturn isExist\n}", "func IsGraphic(r rune) bool", "func main() {\n\ts := \"anagram\"\n\tt := \"nagaram\"\n\n\tfmt.Println(isAnagram(s, t))\n}", "func instr(ch byte, st string) bool {\n for i := 0; i < len(st); i++ {\n if st[i] == ch { return true }\n }\n return false\n}", "func IsIA5String(raw []byte) bool {\n\tfor _, b := range raw {\n\t\ti := int(b)\n\t\tif i > 127 || i < 0 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsAnagram(a, b string) bool {\n\tba := ByteSlice(a)\n\tbb := ByteSlice(b)\n\tsort.Sort(ba)\n\tsort.Sort(bb)\n\tif bytes.Equal(ba, bb) {\n\t\treturn true\n\t} else {\n\t\treturn false\n\t}\n}", "func IsHex(s string) bool {\n\tif s == \"\" {\n\t\treturn false\n\t}\n\n\tfor _, r := range s {\n\t\tif (r < 48 || r > 57) && (r < 97 || r > 102) && (r < 65 || r > 70) && r != 45 {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func checkStr(s string) bool {\n\ts = strings.ToLower(s)\n\n\t// Ensure any embedded suffix / delimiter is removed as I'm using ReadString() function to capture.\n\t// In *nix, expect suffix '\\n', in Windows expect suffix '\\r\\n'\n\ts = strings.TrimSuffix(s, \"\\n\")\n\ts = strings.TrimSuffix(s, \"\\r\")\n\n\tif strings.HasPrefix(s, \"i\") && strings.HasSuffix(s, \"n\") && strings.Contains(s, \"a\") {\n\t\treturn true\n\t}\n\treturn false\n}", "func IsIdent(ch byte) bool {\n\treturn IsDigit(ch) || IsIdentStart(ch)\n}", "func IsHex(s string) bool {\n\tif len(s) != 24 {\n\t\treturn false\n\t}\n\n\t_, err := hex.DecodeString(s)\n\treturn err == nil\n}", "func isInterogative(a string) bool {\n\treturn a[len(a)-1] == '?'\n}", "func isAnagram(s string, t string) bool {\n\tif len(s) != len(t) {\n\t\treturn false\n\t}\n\tstatistics := [26]uint{}\n\tfor i := 0; i < len(s); i++ {\n\t\tstatistics[s[i]-'a'] += 1\n\t\tstatistics[t[i]-'a'] -= 1\n\t}\n\tfor _, count := range statistics {\n\t\tif count != 0 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (d Definition) IsString() bool {\n\tif k, ok := d.Output.(reflect.Kind); ok {\n\t\tif k == reflect.String {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func IsASCIIText(s string) bool {\n\tif len(s) == 0 {\n\t\treturn false\n\t}\n\tfor _, b := range []byte(s) {\n\t\tif 32 <= b && b <= 126 {\n\t\t\t// good\n\t\t} else {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func UUID(str string) bool {\n\tif len(str) != 36 {\n\t\treturn false\n\t}\n\n\tfor i, c := range str {\n\t\tif i == 8 || i == 13 || i == 18 || i == 23 {\n\t\t\tif c != '-' {\n\t\t\t\treturn false\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tif ('f' < c || c < 'a') && ('9' < c || c < '0') {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func IsASCII(str string) bool {\n\tif IsNull(str) {\n\t\treturn true\n\t}\n\treturn rxASCII.MatchString(str)\n}", "func IsString(input []byte) bool {\n\treturn len(input) >= 2 && input[0] == '\"' && input[len(input)-1] == '\"'\n\n}", "func isIsomorphic(s string, t string) bool {\n hashs := make(map[byte]int)\n hasht := make(map[byte]int)\n count := len(s)\n if count != len(t) { return false}\n for i:= 0; i < count ; i++ {\n a , b := hashs[s[i]] \n c , d := hasht[t[i]]\n if a != c || b != d { return false} \n hashs[s[i]] = i\n hasht[t[i]] = i\n }\n return true\n}", "func isChromosome(replicon string) (is bool) {\n\tif strings.Contains(replicon, \"chromosome\") {\n\t\tis = true\n\t} else {\n\t\tis = false\n\t}\n\treturn\n}", "func IsUUID(str string) bool {\n\treturn rxUUID.MatchString(str)\n}", "func IsUUID(str string) bool {\n\treturn rxUUID.MatchString(str)\n}", "func main() {\n\tfmt.Println(isMatch(\"mississippi\", \"mis*is*p*.\"))\n\tfmt.Println(isMatch(\"aab\", \"c*a*b\"))\n}", "func isPublic(myString string) bool {\n\ta := []rune(myString)\n\ta[0] = unicode.ToUpper(a[0])\n\treturn myString == string(a)\n}", "func (me TxsdMimeTypeSequenceType) IsImage() bool { return me.String() == \"image\" }", "func isExported(id string) bool {\n\tr, _ := utf8.DecodeRuneInString(id)\n\treturn unicode.IsUpper(r)\n}", "func IsPrintableASCII(str string) bool {\n\tif IsNull(str) {\n\t\treturn true\n\t}\n\treturn rxPrintableASCII.MatchString(str)\n}", "func (pmp *PngMediaParser) LooksLikeFormat(data []byte) bool {\n return bytes.Compare(data[:len(PngSignature)], PngSignature[:]) == 0\n}", "func contains(a string, x byte) bool {\n\t// Detects if a string contains a character (or sub string)\n\tfor _, n := range a {\n\t\tif x == byte(n) { // if the byte is in there\n\t\t\treturn true // then return true\n\t\t}\n\t}\n\treturn false // otherwise false\n}", "func isASCII(s string) bool {\n\tfor i := 0; i < len(s); i++ {\n\t\tif s[i] >= utf8.RuneSelf {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func isPrintable(s string) bool {\n\tfor _, r := range s {\n\t\tif r > unicode.MaxASCII {\n\t\t\tfmt.Println(\"FOUND NON PRINTABLE CHAR\")\n\t\t\tfmt.Println(r)\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsPrintable(s string) bool {\n\tfor _, r := range s {\n\t\tif r > unicode.MaxASCII || !unicode.IsPrint(r) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsExist(str, filepath string) bool {\r\n\tb, err := ioutil.ReadFile(filepath)\r\n\tcheck(err)\r\n\tisExist, err := regexp.Match(str, b)\r\n\tcheck(err)\r\n\treturn isExist\r\n}", "func IsDataURI(str string) bool {\n\tdataURI := strings.Split(str, \",\")\n\tif !rxDataURI.MatchString(dataURI[0]) {\n\t\treturn false\n\t}\n\treturn IsBase64(dataURI[1])\n}", "func (e DirectoryEntry) IsString() bool { return ((e.Name & 0x80000000) >> 31) > 0 }", "func Detect_If_String_Has_Unique_Chars(str string) bool {\n //1. itterate through all the chars and check to see if the char exist in the map if it does return false if not return true.\n //var myMap map[byte]bool\n myMap := make(map[byte]bool)\n\n //loop through string\n /*fmt.Println(\"Ivan was here\", str)\n for i := 0 ; i < len(str) ; i++ {\n println(i)\n fmt.Printf(\"%c \\n\" ,str[i])\n }\n */\n\n\n for i := 0 ; i < len(str) ; i++ {\n //println(i)\n fmt.Printf(\"%c \\n\" ,str[i])\n if myMap[str[i]] == false {\n myMap[str[i]] = true\n } else {\n return false\n }\n }\n return true\n}", "func isText(s []byte) bool {\n\tconst max = 1024 // at least utf8.UTFMax\n\tif len(s) > max {\n\t\ts = s[0:max]\n\t}\n\tfor i, c := range string(s) {\n\t\tif i+utf8.UTFMax > len(s) {\n\t\t\t// last char may be incomplete - ignore\n\t\t\tbreak\n\t\t}\n\t\tif c == 0xFFFD || c < ' ' && c != '\\n' && c != '\\t' && c != '\\f' {\n\t\t\t// decoding error or control character - not a text file\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsUDID(s string) bool {\n\tif len(s) != 36 {\n\t\treturn false\n\t}\n\tparts := strings.Split(s, \"-\")\n\tif len(parts) != 5 {\n\t\treturn false\n\t}\n\tfor k, v := range map[int]int{0: 8, 1: 4, 2: 4, 3: 4, 4: 12} {\n\t\tif len(parts[k]) != v {\n\t\t\treturn false\n\t\t}\n\t}\n\tfor _, r := range s {\n\t\tswitch r {\n\t\tcase 'a', 'b', 'c', 'd', 'e', 'f', '-', 'A', 'B', 'C', 'D', 'E', 'F', '0', '1', '2', '3', '4', '5', '6', '7', '8', '9':\n\t\tdefault:\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsExported(name string) bool {\n\tch, _ := utf8.DecodeRuneInString(name)\n\treturn unicode.IsUpper(ch)\n}", "func isExported(name string) bool {\n\tr, _ := utf8.DecodeRuneInString(name)\n\treturn unicode.IsUpper(r)\n}", "func isExported(name string) bool {\n\tr, _ := utf8.DecodeRuneInString(name)\n\treturn unicode.IsUpper(r)\n}", "func isExported(name string) bool {\n\tr, _ := utf8.DecodeRuneInString(name)\n\treturn unicode.IsUpper(r)\n}", "func isExported(name string) bool {\n\tr, _ := utf8.DecodeRuneInString(name)\n\treturn unicode.IsUpper(r)\n}", "func IsHash(str string, algorithm string) bool {\n\tlen := \"0\"\n\talgo := strings.ToLower(algorithm)\n\n\tif algo == \"crc32\" || algo == \"crc32b\" {\n\t\tlen = \"8\"\n\t} else if algo == \"md5\" || algo == \"md4\" || algo == \"ripemd128\" || algo == \"tiger128\" {\n\t\tlen = \"32\"\n\t} else if algo == \"sha1\" || algo == \"ripemd160\" || algo == \"tiger160\" {\n\t\tlen = \"40\"\n\t} else if algo == \"tiger192\" {\n\t\tlen = \"48\"\n\t} else if algo == \"sha256\" {\n\t\tlen = \"64\"\n\t} else if algo == \"sha384\" {\n\t\tlen = \"96\"\n\t} else if algo == \"sha512\" {\n\t\tlen = \"128\"\n\t} else {\n\t\treturn false\n\t}\n\n\treturn Matches(str, \"^[a-f0-9]{\"+len+\"}$\")\n}", "func isUniqueChars1(str string) bool {\n\tif len(str) > 128 {\n\t\treturn false\n\t}\n\tvar char_set [128]bool // every bool represents if a character was seen\n\tfor _, val := range str {\n //fmt.Println(reflect.TypeOf(val))\n\t\tif char_set[val] {\n\t\t\treturn false\n\t\t}\n\t\tchar_set[val] = true\n\t}\n\treturn true\n}", "func isExported(name string) bool {\n\ts, _ := utf8.DecodeRuneInString(name)\n\treturn unicode.IsUpper(s)\n}", "func IsByte(m string) bool { return m == \"b\" }", "func IsUniqueChar(str string) bool {\n\tchecker := 0\n\tstart := int('a')\n\tfor _, c := range str {\n\t\tval := uint(int(c) - start)\n\t\tif checker & (1 << val) > 0 {\n\t\t\treturn false\n\t\t}\n\t\tchecker = checker | (1 << val)\n\t}\n\treturn true\n}", "func IsUUID(s string) bool {\n\treturn s != \"\" && rxUUID.MatchString(s)\n}", "func IsFunction(s string) bool {\n\ts = strings.Replace(s, \"(*)\", \"\", -1)\n\treturn strings.Contains(s, \"(\")\n}", "func isMatch(s string, p string) bool {\n\n}", "func Detect_if_anagram(str1 string , str2 string) bool {\n\n\tvar length1 = len(str1)\n\tvar length2 = len(str2)\n\n\tif length1 != length2 {\n\t\t\t\treturn false\n\t} else {\n\t\t\t\tstr1 := Sort_String(str1)\n\t\t\t\tstr2 := Sort_String(str2)\n\n\t\t\t\tfmt.Println(\"1. \" , str1)\n\t\t\t\tfmt.Println(\"2. \" , str2)\n\n\t\t\t\tif str1 == str2 {\n\t\t\t\t\treturn true\n\t\t\t\t} else {\n\t\t\t\t\treturn false\n\t\t\t\t}\n\t\t\t return true\n\t}\n}", "func isAnagram(s string, t string) bool {\n\tif len(s) != len(t) {\n\t\treturn false\n\t}\n\tvar char26 [26]int\n\tfor i := range s {\n\t\tchar26[s[i]-'a']++\n\t\tchar26[t[i]-'a']--\n\t}\n\tfor i := range char26 {\n\t\tif char26[i] != 0 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsTitle(r rune) bool", "func IsExportedName(name string) bool {\n\trune, _ := utf8.DecodeRuneInString(name)\n\treturn unicode.IsUpper(rune)\n}", "func isASCII(bytes []byte) bool {\n\tfor _, b := range bytes {\n\t\tif b^128 == 128 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func IsRotatedString(s1 string, s2 string) bool {\r\n\tif len(s1) != len(s2) || len(s1) == 0 || len(s2) == 0 {\r\n\t\treturn false\r\n\t}\r\n\treturn strings.Contains(s2+s2, s1)\r\n}", "func isExported(name string) bool {\n rune, _ := utf8.DecodeRuneInString(name)\n return unicode.IsUpper(rune)\n}", "func IsString(data interface{}) bool {\n\treturn typeIs(data, reflect.String)\n}", "func Valid(s string) bool { return Convert(s) == s && s != \"\" }", "func IsASCII(s string) bool {\n\tfor i := 0; i < len(s); i++ {\n\t\tif s[i] > unicode.MaxASCII {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}" ]
[ "0.86452794", "0.85342157", "0.8521849", "0.84966177", "0.8453791", "0.84477234", "0.84153825", "0.84105974", "0.8389144", "0.8372251", "0.8368941", "0.8365187", "0.83498234", "0.8287772", "0.824803", "0.8166897", "0.8162717", "0.81117034", "0.8082223", "0.8057842", "0.80450237", "0.798012", "0.7976361", "0.7840107", "0.77699655", "0.7765536", "0.76544154", "0.68745416", "0.61542857", "0.5754048", "0.57100517", "0.566725", "0.5625305", "0.55379176", "0.55131197", "0.5433913", "0.54072714", "0.5392207", "0.53731114", "0.5359788", "0.535461", "0.5335164", "0.5319564", "0.5310038", "0.52800494", "0.5264132", "0.52637017", "0.52549624", "0.5254289", "0.5245767", "0.52084327", "0.5204243", "0.5200462", "0.5186141", "0.518448", "0.5160463", "0.5159384", "0.51573884", "0.51469886", "0.5142691", "0.5142691", "0.51397485", "0.51192945", "0.5099699", "0.5093089", "0.50808483", "0.5079126", "0.50745815", "0.5063592", "0.5060051", "0.50559974", "0.50531894", "0.50499547", "0.50484884", "0.50473666", "0.50439113", "0.5043375", "0.5040532", "0.5037828", "0.5037828", "0.5037828", "0.5037828", "0.5033042", "0.5032377", "0.5027111", "0.5010388", "0.5009091", "0.5006797", "0.5002268", "0.49994773", "0.4993732", "0.499333", "0.4991183", "0.49817342", "0.49777406", "0.49759784", "0.4973534", "0.49721038", "0.49713913", "0.49575153" ]
0.8056851
20
Initialize your data structure here.
func Constructor() MyHashSet { return MyHashSet{make([]list.List, base)} }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func initData() {\n\tallMap = make(map[string]int64, 0)\n\terrMap = make(map[string]int64, 0)\n}", "func InitializeData() {\n\t// First we do the black cards\n\tBlackCards = make(map[int]BlackCard)\n\tImportBlackCards()\n\t// Print the amount of Black Cards loaded into memory.\n\tfmt.Printf(\"%d Black Cards loaded!\\n\", len(BlackCards))\n\n\t// Then we do the white cards.\n\tWhiteCards = make(map[int]WhiteCard)\n\tImportWhiteCards()\n\t// Print the amount of White Cards loaded into memory.\n\tfmt.Printf(\"%d White Cards loaded!\\n\", len(WhiteCards))\n\n\t// Moving on, we now want to make the players map.\n\tPlayers = make(map[string]Player)\n\n\t// We want to clear the following variables..\n\tPlayerCount = 0\n\tHighScore = 0\n\tZar = 0\n\tRound = 0\n\tRounds = 10\n\tRoundText = \"\"\n\tRoundResults = nil\n\tCreatorID = \"\"\n\tZars = nil\n\tJudging = false\n}", "func init(){\n\n\tcustomers = loadAllCustomers()\n\tBuildRank()\n\tBuildTable()\n\n}", "func (m *MixtapeIndex) init() {\n\tm.Users = make(map[string]*User)\n\tm.Playlists = make(map[string]*Playlist)\n\tm.Songs = make(map[string]*Song)\n\tm.PlaylistIdUpperBound = -1\n}", "func (l *list) init() {\n\tl.nodes = 0\n\tl.begin = &node{\"\", nil, nil}\n}", "func initialize() {\n\n\thead.data = \"San Francisco\"\n\thead.next = nil\n\n\tvar Oakland *Node = &Node{data: \"Oakland\", next: nil}\n\thead.next = Oakland\n\n\tvar Berkely *Node = &Node{data: \"Berkely\", next: nil}\n\tOakland.next = Berkely\n\n\tvar Fremont *Node = &Node{data: \"Fremont\", next: nil}\n\tBerkely.next = Fremont\n\n\t// remember this pattern for linked list\n\t// make the last insertion, tail = last value\n\ttail = Fremont\n\ttail.next = nil\n\tBerkely.next = tail\n}", "func init() {\n\tSliceOfString = make([]string, 0, 10)\n\tMapOfString = make(map[string]string)\n\tMapOfBool = make(map[string]bool)\n}", "func initializeGuildData(guildID string) {\n\t_, guildDataExists := guildData[guildID]\n\tif !guildDataExists {\n\t\tguildData[guildID] = &GuildData{}\n\t\tguildData[guildID].Queries = make(map[string]*Query)\n\t}\n}", "func (l *log) init() {\n\tl.entries = make([]*Entry, 0)\n}", "func (s *Struct) Init() {\n\ts.Comments = make([]string, 0)\n\ts.Attributes = make([]*Attribute, 0)\n\ts.Fields = make([]*Field, 0)\n}", "func initializeConnectionData() {\n\t// TODO: Access if we need the clients variable\n\t// clients = make(map[*websocket.Conn]*ClientData)\n\tconnections = make(map[*websocket.Conn]int)\n\tclientIdMap = make(map[int]*ClientData)\n}", "func (l *Labels) init() {\n\tif l.lmap == nil {\n\t\tl.lmap = make(map[string]int64)\n\t\tl.offLabels = make(map[int64][]string)\n\t}\n}", "func (m *metricMysqlBufferPoolDataPages) init() {\n\tm.data.SetName(\"mysql.buffer_pool.data_pages\")\n\tm.data.SetDescription(\"The number of data pages in the InnoDB buffer pool.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.MetricAggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (this *Vector)init(){\n\tif this.v==nil{\n\t\tthis.v=make([]interface{},0,1)\n\t}\n}", "func (d *DB) Init() {\n\td.NumCell = 1024 // set num cell first!\n\td.root = d.NewNode()\n\t//log.Printf(\"+d.root: %v\\n\", d.root)\n}", "func init() {\n\t//allocate memory to maps\n\tusermap = make(map[string]UserData)\n\tstopwords = make(map[string]struct{})\n\tSubCommentMap = make(map[string]SubsCommentData)\n\tusagecmmts = make(map[string](map[string]SubsWordData))\n\tusagesubmstitle = make(map[string](map[string]SubsWordData))\n\tusagesubmsbody = make(map[string](map[string]SubsWordData))\n\n\t//load secret file\n\tconfigdata, err := os.Open(\"C:\\\\Users\\\\Myke\\\\Dropbox\\\\School\\\\CSCI164\\\\Project\\\\config.secret\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t//close file when program is done\n\tdefer configdata.Close()\n\n\t//load config file from json to config struct\n\tscanner := bufio.NewScanner(configdata)\n\tfor scanner.Scan() {\n\t\terr = json.Unmarshal(scanner.Bytes(), &config)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t}\n\n\t//load stopwords\n\tstopdata, err := os.Open(\"C:\\\\Users\\\\Myke\\\\Dropbox\\\\School\\\\CSCI164\\\\Project\\\\stopwords\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t//place stop words into a map\n\tscanner = bufio.NewScanner(stopdata)\n\tfor scanner.Scan() {\n\t\tvar z struct{}\n\t\tstopwords[scanner.Text()] = z\n\t}\n\n\t//close file when program is done\n\tdefer stopdata.Close()\n\n}", "func (d *Data) Init() {\n\t// (X) d = NewData()\n\t// This only updates its pointer\n\t// , not the Data itself\n\t//\n\t*d = *NewData()\n}", "func (db *Database) init(g []Group, e []Entry, opts *Options) {\n\t// Clear inputted key material.\n\tif db.cparams.ComputedKey == nil {\n\t\tpanic(\"key should have been precomputed\")\n\t}\n\tdb.cparams.Key.Password, db.cparams.Key.KeyFileHash = nil, nil\n\n\tdb.staticIV = opts.staticIV()\n\tdb.groups = make(map[uint32]*Group, len(g))\n\tdb.entries = make([]*Entry, 0, len(e))\n\tdb.rand = opts.getRand()\n\tdb.root = &Group{Name: \"Root\", db: db}\n\tfor i := range g {\n\t\tgg := &g[i]\n\t\tdb.groups[gg.ID] = gg\n\t}\n\tfor i := range e {\n\t\tee := &e[i]\n\t\tif ee.isMetaStream() {\n\t\t\tdb.meta = append(db.meta, ee)\n\t\t} else {\n\t\t\tdb.entries = append(db.entries, ee)\n\t\t}\n\t}\n}", "func (c *Collector) Init() error { return nil }", "func InitializeData() {\n\n\tshops, err := getShops()\n\tif err != nil {\n\t\tlogger.Error.Println(err)\n\t\treturn\n\t}\n\n\tfor _, shop := range *shops {\n\t\tif err := getOrders(shop.Code); err != nil {\n\t\t\tlogger.Error.Println(err)\n\t\t\treturn\n\t\t}\n\t}\n}", "func (m *metricRedisMemoryLua) init() {\n\tm.data.SetName(\"redis.memory.lua\")\n\tm.data.SetDescription(\"Number of bytes used by the Lua engine\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptyGauge()\n}", "func (c *Context) init(ctx *fasthttp.RequestCtx) {\n\tc.RequestCtx = ctx\n\tc.data = newDataMap()\n\tc.index = -1\n\tc.Serialize = Serialize\n}", "func (m *Mixtape) init() {\n\tm.Users = []User{}\n\tm.Playlists = []Playlist{}\n\tm.Songs = []Song{}\n}", "func init() {\n\tcdb = &classDB{classMap: make(map[int]Class)}\n\tif err := loadData(cdb); err != nil {\n\t\tlog.Print(\"Panic in loading data\")\n\t}\n\tlog.Print(\"Class Data from file is loaded!!\")\n\n}", "func (l *lru_cache) Init(capacity int) {\n\tl.dataMap = make(map[Key]*list.Element)\n\n\tl.capacity = capacity\n\tl.linkList.Init()\n}", "func NewData() *Data {\n d := &Data{}\n d.hashMap = make(map[int]string)\n return d\n}", "func (m *metricRedisMaxmemory) init() {\n\tm.data.SetName(\"redis.maxmemory\")\n\tm.data.SetDescription(\"The value of the maxmemory configuration directive\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptyGauge()\n}", "func init() {\n\t//todo...\n}", "func initData(contractNumber string) {\n\tUBSCookies = nil\n\t//var err error;\n\tUBSCookieJar, _ = cookiejar.New(nil)\n\n\tUSBContractNumber = contractNumber\n}", "func init() {\n\t// Initialization goes here\n}", "func (honest *Honest) initializeData(datasetName string, numberOfNodes int, epsilon float64, isPoisoning bool) {\n\n\tif datasetName == \"creditcard\" {\n\t\t\n\t\tuseTorch = false\n\t\n\t\tif isPoisoning {\n\t\t\toutLog.Println(\"Get the bad credit data.\")\n\t\t\thonest.ncol = pyInit(\"creditbad\", \"creditbad\", epsilon)\t\n\t\t} else {\n\t\t\thonest.ncol = pyInit(datasetName, datasetName + strconv.Itoa(honest.id), epsilon)\n\t\t}\t\n\n\t} else {\n\t\n\t\tuseTorch = true\n\t\n\t\tif isPoisoning {\n\t\t\toutLog.Println(\"Get the bad data.\")\n\t\t\thonest.ncol = pyInit(\"mnist\", \"mnist_bad\", epsilon)\t\n\t\t} else {\n\t\t\thonest.ncol = pyInit(datasetName, datasetName + strconv.Itoa(honest.id), epsilon)\n\t\t}\n\n\t}\n\t\t\n\thonest.dataset = datasetName\n\thonest.bc = NewBlockchain(honest.ncol)\n\n}", "func (m *metricAerospikeNamespaceMemoryUsage) init() {\n\tm.data.SetName(\"aerospike.namespace.memory.usage\")\n\tm.data.SetDescription(\"Memory currently used by each component of the namespace\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func initMockData() {\n\n\t// Generate Mock Data\n\tbooks = append(books, models.GenerateSampleBookRecord())\n\tbooks = append(books, models.GenerateSampleBookRecord())\n\tbooks = append(books, models.GenerateSampleBookRecord())\n\tbooks = append(books, models.GenerateSampleBookRecord())\n}", "func init() {\n\tMemory = &memoryStorage{\n\t\ttraces: make(map[string]tracer.Trace),\n\t\tservices: make(map[string]string),\n\t\tserviceDeps: make(map[string]*tracer.Dependencies),\n\t}\n}", "func (l *Loader) init() {\n\tif l.loading == nil {\n\t\tl.loading = stringset.New(1)\n\t\tl.sources = make(map[string]string, 1)\n\t\tl.symbols = make(map[string]*Struct, 1)\n\t}\n}", "func (room *GameRoom) init() {\n\tif room.whiteList == nil {\n\t\troom.whiteList = make(map[string]*User)\n\t\troom.blackList = make(map[string]*User)\n\t\troom.inEffectCard = []*Card{}\n\t\troom.turnPlayers = []*User{}\n\t\troom.Turn = 0\n\t\troom.Clock = 11\n\t\troom.IsRouletteTurn = true\n\t\troom.IsRouletteTurn = false\n\t\troom.Level = 1\n\t}\n}", "func (svc Service) Initialize() {\n\tlog.Info(\"initialize the database...\")\n\tlines := strings.Split(geodata, \"\\n\")\n\n\tlog.Info(\"processing %d data rows...\\n\", len(lines))\n\tfor i := 0; i < len(lines); i++ {\n\t\tfields := strings.Split(lines[i], \",\")\n\t\tzipcode := Ziptype(fields[0])\n\n\t\tlat, _ := strconv.ParseFloat(fields[1], 64)\n\t\tlng, _ := strconv.ParseFloat(fields[2], 64)\n\n\t\tkey := svc.CreateKey(lat, lng)\n\n\t\tcoord := Coord{lat, lng}\n\t\tzipMap[zipcode] = &coord\n\n\t\tzcoord := ZipcodeCoord{zipcode, coord}\n\t\tkeyMap[key] = append(keyMap[key], &zcoord)\n\t}\n\n\tlog.Info(\"processed %d rows...\\n\", len(lines))\n}", "func (m *metricRedisMemoryUsed) init() {\n\tm.data.SetName(\"redis.memory.used\")\n\tm.data.SetDescription(\"Total number of bytes allocated by Redis using its allocator\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptyGauge()\n}", "func (m *metricRedisMemoryRss) init() {\n\tm.data.SetName(\"redis.memory.rss\")\n\tm.data.SetDescription(\"Number of bytes that Redis allocated as seen by the operating system\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptyGauge()\n}", "func (m *metricRedisDbKeys) init() {\n\tm.data.SetName(\"redis.db.keys\")\n\tm.data.SetDescription(\"Number of keyspace keys\")\n\tm.data.SetUnit(\"{key}\")\n\tm.data.SetEmptyGauge()\n\tm.data.Gauge().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricBigipPoolMemberCount) init() {\n\tm.data.SetName(\"bigip.pool.member.count\")\n\tm.data.SetDescription(\"Total number of pool members.\")\n\tm.data.SetUnit(\"{members}\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricMysqlBufferPoolPages) init() {\n\tm.data.SetName(\"mysql.buffer_pool.pages\")\n\tm.data.SetDescription(\"The number of pages in the InnoDB buffer pool.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.MetricAggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func init() {\n\tstationsCache = make(map[string]Station)\n}", "func (l *List) Init() {\r\n\tl.list = list.New()\r\n}", "func (pd *pymtData) Init(p payment.PymtUpsert) {\n\tpd.Type = p.Type\n\tpd.Attrs = p.Attributes\n}", "func init() {\n\tSample = map[string]Personas{\n\t\t\"1\": Personas{Id: \"1\", Nombre: \"Luis\", Apellido: \"Perez\"},\n\t\t\"2\": Personas{Id: \"2\", Nombre: \"Maria\", Apellido: \"Romano\"},\n\t\t\"3\": Personas{Id: \"3\", Nombre: \"Nestor\", Apellido: \"Sanchez\"},\n\t}\n}", "func init() {\n\te, err := data.FetchEmployees(EmployeeDataUrl)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"%s - %s\", ErrFetchEmployees, err))\n\t}\n\n\temployees = e\n}", "func init() {\n\n\t//Welcome Message\n\tfmt.Println(\"------------------------------------------------------------------\")\n\tfmt.Println(\"Starting Natyla...\")\n\tfmt.Println(\"Version: 1.02\")\n\n\t//Set the thread quantity based on the number of CPU's\n\tcoreNum := runtime.NumCPU()\n\n\tfmt.Println(\"Number of cores: \", coreNum)\n\n\t//read the config file\n\treadConfig()\n\n\t//create the data directory\n\tcreateDataDir()\n\n\t//set max memory form config\n\tmaxMemBytes, _ = config[\"memory\"].(json.Number).Int64()\n\tfmt.Println(\"Max memory defined as: \", maxMemBytes/1024/1024, \" Mbytes\")\n\n\truntime.GOMAXPROCS(coreNum)\n\n\t//Create a new doble-linked list to act as LRU\n\tlruList = list.New()\n\n\t//Create the channels\n\tlisChan = make(chan int, 1)\n\tLRUChan = make(chan int, 1)\n\tcollectionChan = make(chan int, 1)\n\n\tcollections = make(map[string]collectionChannel)\n\n\t//Read collections from disk\n\tnRead := readAllFromDisk()\n\tfmt.Println(\"Read\", nRead, \"entries from disk\")\n\n\tfmt.Println(\"Ready, API Listening on http://localhost:8080, Telnet on port 8081\")\n\tfmt.Println(\"------------------------------------------------------------------\")\n}", "func (m *metricActiveDirectoryDsReplicationNetworkIo) init() {\n\tm.data.SetName(\"active_directory.ds.replication.network.io\")\n\tm.data.SetDescription(\"The amount of network data transmitted by the Directory Replication Agent.\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricActiveDirectoryDsReplicationObjectRate) init() {\n\tm.data.SetName(\"active_directory.ds.replication.object.rate\")\n\tm.data.SetDescription(\"The number of objects transmitted by the Directory Replication Agent per second.\")\n\tm.data.SetUnit(\"{objects}/s\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (g *generator) InitData(kt *kit.Kit) error {\n\theader := http.Header{}\n\theader.Set(constant.UserKey, constant.BKUserForTestPrefix+\"gen-data\")\n\theader.Set(constant.RidKey, kt.Rid)\n\theader.Set(constant.AppCodeKey, \"test\")\n\theader.Add(\"Cookie\", \"bk_token=\"+constant.BKTokenForTest)\n\n\tg.data = make(AppReleaseMeta, 0)\n\n\t//if err := g.initApp1(kt.Ctx, header); err != nil {\n\t//\treturn err\n\t//}\n\n\tif err := g.initApp2(kt.Ctx, header); err != nil {\n\t\treturn err\n\t}\n\n\t//if err := g.initApp3(kt.Ctx, header); err != nil {\n\t//\treturn err\n\t//}\n\n\treturn nil\n}", "func InitGraph() {\n\tgraph.Rooms = make(map[string]Room)\n\t// graph.steps = make(map[int]string)\n\n}", "func (g *Graph) init(numVertex int) {\n if g.vertex == nil {\n g.vertex = make([]*NodeG, numVertex)\n }\n}", "func (f *FibonacciHeap) init() {\n\tf.min = &(Node{})\n\tf.nodes = make(map[Value]*Node)\n\tf.size = 0\n}", "func (s *Store) Init(sessionID []byte, defaultExpiration time.Duration) {\n\ts.sessionID = sessionID\n\ts.defaultExpiration = defaultExpiration\n\n\tif s.data == nil { // Ensure the store always has a valid pointer of Dict\n\t\ts.data = new(Dict)\n\t}\n}", "func (kb *Kbucket) Init(nodeID ID) {\n\tkb.ContactList = make([]Contact, 0, k)\n\tkb.NodeID = nodeID\n}", "func (self *MessageStore) Init() {\n\tself.in = make(map[string]*MsgEntry)\n\tself.out = make(map[string]*MsgEntry)\n}", "func init() {\n\tpetFields := schema.Pet{}.Fields()\n\t_ = petFields\n\t// petDescHeight is the schema descriptor for height field.\n\tpetDescHeight := petFields[0].Descriptor()\n\t// pet.HeightValidator is a validator for the \"height\" field. It is called by the builders before save.\n\tpet.HeightValidator = petDescHeight.Validators[0].(func(int) error)\n\t// petDescWeight is the schema descriptor for weight field.\n\tpetDescWeight := petFields[1].Descriptor()\n\t// pet.WeightValidator is a validator for the \"weight\" field. It is called by the builders before save.\n\tpet.WeightValidator = petDescWeight.Validators[0].(func(float64) error)\n\t// petDescChip is the schema descriptor for chip field.\n\tpetDescChip := petFields[7].Descriptor()\n\t// pet.DefaultChip holds the default value on creation for the chip field.\n\tpet.DefaultChip = petDescChip.Default.(func() uuid.UUID)\n}", "func (m *metricMysqlSorts) init() {\n\tm.data.SetName(\"mysql.sorts\")\n\tm.data.SetDescription(\"The number of MySQL sorts.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.MetricAggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (s *SinglyLinkedList) Init() *SinglyLinkedList {\n s.length = 0\n return s\n}", "func (m *metricMysqlHandlers) init() {\n\tm.data.SetName(\"mysql.handlers\")\n\tm.data.SetDescription(\"The number of requests to various MySQL handlers.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.MetricAggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (d *SliceDataStore) Init(parameters ...interface{}) error {\n\td.slice = make([]Record, parameters[0].(int))\n\treturn nil\n}", "func (mdb *MemoryDB) Init() (err error) {\n\tmdb.table = make(map[string]*TableRow)\n\treturn nil\n}", "func (db *Database) Init() {\n\tdata, dbErr := tiedot.OpenDB(db.Location)\n\tif dbErr != nil {\n\t\tlog.Error(dbConnectionError{\n\t\t\tmsg: \"Failed to connect to the tiedot database\",\n\t\t\terr: dbErr,\n\t\t})\n\t}\n\n\t// Set up the collections - throw away the error for now.\n\tfor _, c := range db.Collections {\n\t\tdata.Create(c.Name)\n\t\tdata.Use(c.Name).Index(c.Index)\n\t}\n\n\tdb.Data = data\n}", "func (c *Const) Init() {\n\tc.Values = make([]*Pair, 0)\n}", "func init() {\n\tif err := mb.Registry.AddMetricSet(\"consulkv\", \"kv\", New); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (g *Graph) init(numVertex int) {\n if g.vertex == nil {\n g.vertex = make([]*NodeG, numVertex+1)\n }\n}", "func (pq *PrefixQueue) init() error {\n\t// Get the main prefix queue data.\n\tval, err := pq.db.Get(pq.getDataKey(), nil)\n\tif err == errors.ErrNotFound {\n\t\treturn nil\n\t} else if err != nil {\n\t\treturn err\n\t}\n\n\tpq.size = binary.BigEndian.Uint64(val)\n\treturn nil\n}", "func init() {\n\tcallbacks = make(map[ModuleType]*ConfigCallback, 8)\n\tmodules = make(map[string]ModuleType, 32)\n}", "func (c *Pnrm) Initialize(i util.XapiClient) {\n c.DeviceGroup = &dg.Dg{}\n c.DeviceGroup.Initialize(i)\n\n c.GcpAccount = &account.Account{}\n c.GcpAccount.Initialize(i)\n\n c.GkeCluster = &cluster.Cluster{}\n c.GkeCluster.Initialize(i)\n\n c.GkeClusterGroup = &group.Group{}\n c.GkeClusterGroup.Initialize(i)\n\n c.Template = &template.Template{}\n c.Template.Initialize(i)\n\n c.TemplateStack = &stack.Stack{}\n c.TemplateStack.Initialize(i)\n\n c.TemplateVariable = &variable.Variable{}\n c.TemplateVariable.Initialize(i)\n}", "func (m *metricActiveDirectoryDsThreadCount) init() {\n\tm.data.SetName(\"active_directory.ds.thread.count\")\n\tm.data.SetDescription(\"The number of threads in use by the directory service.\")\n\tm.data.SetUnit(\"{threads}\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n}", "func (m *metricBigipPoolAvailability) init() {\n\tm.data.SetName(\"bigip.pool.availability\")\n\tm.data.SetDescription(\"Availability of the pool.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptyGauge()\n\tm.data.Gauge().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricBigipNodeAvailability) init() {\n\tm.data.SetName(\"bigip.node.availability\")\n\tm.data.SetDescription(\"Availability of the node.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptyGauge()\n\tm.data.Gauge().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricMysqlBufferPoolUsage) init() {\n\tm.data.SetName(\"mysql.buffer_pool.usage\")\n\tm.data.SetDescription(\"The number of bytes in the InnoDB buffer pool.\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.MetricAggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricActiveDirectoryDsLdapSearchRate) init() {\n\tm.data.SetName(\"active_directory.ds.ldap.search.rate\")\n\tm.data.SetDescription(\"The number of LDAP searches per second.\")\n\tm.data.SetUnit(\"{searches}/s\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n}", "func initDatabase() {\n\tif dbPath == \"\" {\n\t\t// No path provided, use the default path\n\t\tdbPath = getDefaultDBPath()\n\t}\n\t// Start the database\n\tdb, err := poddata.New(dbPath)\n\tif err != nil {\n\t\tlogrus.Fatal(err.Error())\n\t}\n\tdata = db\n}", "func (m *clistModel) Init(t *rapid.T) {\n\tm.clist = clist.New()\n\tm.model = []*clist.CElement{}\n}", "func (d *Dok) init() *Dok {\n\tif d == nil {\n\t\treturn nil\n\t}\n\n\tdok := &Dok{\n\t\tprev: d.prev,\n\t\tkey: d.key,\n\t}\n\treturn dok\n}", "func (m *metricMysqlPageOperations) init() {\n\tm.data.SetName(\"mysql.page_operations\")\n\tm.data.SetDescription(\"The number of InnoDB page operations.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.MetricAggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func init() {\n\tbucketsecretFields := schema.BucketSecret{}.Fields()\n\t_ = bucketsecretFields\n\t// bucketsecretDescType is the schema descriptor for type field.\n\tbucketsecretDescType := bucketsecretFields[3].Descriptor()\n\t// bucketsecret.DefaultType holds the default value on creation for the type field.\n\tbucketsecret.DefaultType = bucketsecretDescType.Default.(int)\n}", "func (cust *custInfo) initializeValues(){\n cust.customerId = \"123424\"\n cust.bankName = \"Bank1\"\n cust.accountNumber = \"123456789\"\n cust.seqNumber = 0\n}", "func (x *FileExtractor) init() {\n\thead := x.aws.GetHeadObject(x.ctx, x.bucket, x.key)\n\tx.size = *head.ContentLength\n\tx.fileMap = make(map[string][]*File)\n}", "func (c *Objs) Initialize(i util.XapiClient) {\n c.Address = &addr.Addr{}\n c.Address.Initialize(i)\n\n c.AddressGroup = &addrgrp.AddrGrp{}\n c.AddressGroup.Initialize(i)\n\n c.Services = &srvc.Srvc{}\n c.Services.Initialize(i)\n}", "func (b *Uniform) Init(nArms int) {\n\tb.counts = make([]uint64, nArms, nArms)\n\tb.values = make([]float64, nArms, nArms)\n}", "func InitBook(ticker string) *Book {\n return &Book{\n BuyTree: binarytree.NewTree(),\n SellTree: binarytree.NewTree(),\n LowestSell: nil,\n HighestBuy: nil,\n OrderMap: make(map[int]*Order),\n BuyLimitMap: make(map[LimitPrice]*InfoAtLimit),\n SellLimitMap: make(map[LimitPrice]*InfoAtLimit),\n TickerOFBook: ticker}\n}", "func (m *metricBigipPoolPacketCount) init() {\n\tm.data.SetName(\"bigip.pool.packet.count\")\n\tm.data.SetDescription(\"Number of packets transmitted to and from the pool.\")\n\tm.data.SetUnit(\"{packets}\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricBigipPoolMemberAvailability) init() {\n\tm.data.SetName(\"bigip.pool_member.availability\")\n\tm.data.SetDescription(\"Availability of the pool member.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptyGauge()\n\tm.data.Gauge().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricBigipNodeDataTransmitted) init() {\n\tm.data.SetName(\"bigip.node.data.transmitted\")\n\tm.data.SetDescription(\"Amount of data transmitted to and from the node.\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (c *Carta) Init(id, tipo, era, puntos, edificioGratis, cartaRequerida int, nombre string, produce, requiere [CANTIDAD_RECURSOS]int){\n c.Id = id\n c.Tipo = tipo\n c.era = era\n c.puntos = puntos\n c.edificioGratis = edificioGratis\n c.Nombre = nombre\n c.Produce = produce\n c.requiere = requiere\n c.cartaRequerida = cartaRequerida\n}", "func (m *Tracker) Init(tag string, deepcopy bool) {\n\tm.kmap = make(map[common.Key]interface{})\n\tm.tag = tag\n\tm.deepcopy = deepcopy\n}", "func (scope *Scope) initialize() {\n\tif scope.variables == nil {\n\t\tscope.variables = make(map[string]Variable, 0)\n\t}\n}", "func (p *MemProvider) Init(maxLifetime int64, _ string) error {\n\tp.lock.Lock()\n\tp.list = list.New()\n\tp.data = make(map[string]*list.Element)\n\tp.maxLifetime = maxLifetime\n\tp.lock.Unlock()\n\treturn nil\n}", "func (m *metricMysqlLogOperations) init() {\n\tm.data.SetName(\"mysql.log_operations\")\n\tm.data.SetDescription(\"The number of InnoDB log operations.\")\n\tm.data.SetUnit(\"1\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.MetricAggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (m *metricBigipNodePacketCount) init() {\n\tm.data.SetName(\"bigip.node.packet.count\")\n\tm.data.SetDescription(\"Number of packets transmitted to and from the node.\")\n\tm.data.SetUnit(\"{packets}\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(true)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n\tm.data.Sum().DataPoints().EnsureCapacity(m.capacity)\n}", "func (e *Enum) Init() {\n\te.Values = make([]*Pair, 0)\n}", "func (m *metricFlinkMemoryManagedUsed) init() {\n\tm.data.SetName(\"flink.memory.managed.used\")\n\tm.data.SetDescription(\"The amount of managed memory currently used.\")\n\tm.data.SetUnit(\"By\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n}", "func (m *metricActiveDirectoryDsNotificationQueued) init() {\n\tm.data.SetName(\"active_directory.ds.notification.queued\")\n\tm.data.SetDescription(\"The number of pending update notifications that have been queued to push to clients.\")\n\tm.data.SetUnit(\"{notifications}\")\n\tm.data.SetEmptySum()\n\tm.data.Sum().SetIsMonotonic(false)\n\tm.data.Sum().SetAggregationTemporality(pmetric.AggregationTemporalityCumulative)\n}", "func (meta *Meta) init() {\n\tmeta.client = utils.CreateMongoDB(dbConfig.Str(\"address\"), log)\n\tmeta.database = meta.client.Database(dbConfig.Str(\"db\"))\n\tmeta.collection = meta.database.Collection(metaCollection)\n}", "func initAll()", "func (l *List) Init() *List {\n\tl.root.next = &l.root\n\tl.root.prev = &l.root\n\tl.Len = 0\n\treturn l\n}", "func (g *metadataGraph) init(root *api.PrefixMetadata) {\n\tif root != nil && root.Prefix != \"\" {\n\t\tpanic(\"the root node metadata should have empty prefix\")\n\t}\n\tg.root = &metadataNode{\n\t\tacls: make([]*packageACL, len(legacyRoles)),\n\t\tmd: root,\n\t}\n}" ]
[ "0.69604844", "0.65919334", "0.6388845", "0.6343592", "0.6315344", "0.6158609", "0.6151085", "0.6112397", "0.6111184", "0.60915977", "0.608717", "0.60630053", "0.6002261", "0.5988767", "0.5965372", "0.5960862", "0.5948927", "0.5948688", "0.59285384", "0.5924077", "0.59088767", "0.58968955", "0.58803606", "0.58794534", "0.58656687", "0.5855126", "0.58371425", "0.58316094", "0.5831425", "0.581032", "0.5808829", "0.58040464", "0.57544035", "0.5752273", "0.57479066", "0.5726707", "0.57224464", "0.57089067", "0.5705016", "0.5704673", "0.56765133", "0.5670558", "0.56682605", "0.5655288", "0.5648208", "0.5646799", "0.56463176", "0.5645802", "0.5628311", "0.5626587", "0.56201476", "0.56170297", "0.56128794", "0.56114835", "0.56067777", "0.5602395", "0.5596113", "0.5592338", "0.55835235", "0.5581426", "0.55732876", "0.5571557", "0.55709136", "0.5565492", "0.55618584", "0.5556607", "0.55461365", "0.55347425", "0.55340403", "0.55271447", "0.55168265", "0.55146164", "0.5514585", "0.5511459", "0.55113786", "0.55094105", "0.5506953", "0.55032897", "0.55029315", "0.549896", "0.54894805", "0.5488802", "0.5488603", "0.54865706", "0.54858434", "0.54833215", "0.5482559", "0.54817957", "0.5475416", "0.54752374", "0.5472837", "0.5466833", "0.54645735", "0.5463237", "0.5462346", "0.5458468", "0.5454669", "0.54527766", "0.5447303", "0.54429483", "0.54355234" ]
0.0
-1
Returns true if this set contains the specified element
func (mhs *MyHashSet) Contains(key int) bool { idx := mhs.hash(key) for item := mhs.data[idx].Front(); item != nil; item = item.Next() { if item.Value.(int) == key { return true } } return false }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (set *Set) Contains(element interface{}) bool {\n\tval, ok := set.elements[element]\n\tif val && ok {\n\t\treturn true\n\t}\n\treturn false\n}", "func (s *Set) Contains(element interface{}) bool {\n\t_, exists := (*s)[element]\n\treturn exists\n}", "func (s *Set) Contains(elem A) bool {\n\t_, found := s.m[elem]\n\treturn found\n}", "func (this *HashSet) Contains(ele interface{}) bool {\n\treturn this.mmap.ContainsKey(ele)\n}", "func (s *Set[T]) Contains(elt T) bool {\n\t_, contains := (*s)[elt]\n\treturn contains\n}", "func (s *Set) Contains(val interface{}) bool {\n\t_, contains := s.set[val]\n\treturn contains\n}", "func (s Set) Contain(el string) bool {\n\t_, in := s[el]\n\treturn in\n}", "func (vector *Vector) Contains(element interface{}) bool {\n\tvar index int = vector.Peek(element)\n\tif index == -1 {\n\t\treturn false\n\t}\n\treturn true\n}", "func (s *Set) Has(element string) (bool, error) {\n\tv, err := s.store.Do(s.key, set.Has, element)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tval, ok := v.(bool)\n\tif !ok {\n\t\treturn false, newTypeErr(val, v)\n\t}\n\n\treturn val, nil\n}", "func (s *Set) Contains(val interface{}) bool {\n\tif _, ok := s.vals[val]; ok {\n\t\treturn true\n\t}\n\treturn false\n}", "func (s *IntSet) Contains(e int) bool {\n\t_, flag := s.elem[e]\n\treturn flag\n}", "func (s *StrSet) Contains(element string) bool {\n\tif _, ok := s.els[element]; !ok {\n\t\treturn false\n\t}\n\treturn true\n}", "func (s Stream) Contains(element interface{}) (bool, *errors.Error) {\n\tcurrent := s.run()\n\tif current.err != nil {\n\t\treturn false, current.err\n\t}\n\tif current.itemsLen == 0 {\n\t\treturn false, nil\n\t}\n\treturn (&contains.Contains{Items: current.itemsValue, ItemsType: current.itemsType, Element: element}).Run()\n}", "func (s *HashSet) Contain(data interface{}) (bool, error) {\n\terr := s.checkT(data)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\t_, ok := s.set[data]\n\tif ok {\n\t\treturn true, nil\n\t} else {\n\t\treturn false, nil\n\t}\n}", "func (s *Set) Contains(item Value) bool {\n\t_, ok := s.backingMap.Load(item)\n\treturn ok\n}", "func (list *ArrayList[T]) Contains(ele T) bool {\n\tfor _, o := range list.elems {\n\t\tif o == ele {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (t tagSet) Contains(key string) bool {\n\tif len(t) == 0 {\n\t\treturn false\n\t}\n\ti := t.Search(key)\n\treturn i < len(t) && t[i].key == key\n}", "func (s *set) Contains(t *Term) bool {\n\treturn s.get(t) != nil\n}", "func (s Set) Has(element string) bool {\n\t_, exists := s[strings.ToLower(element)]\n\treturn exists\n}", "func (t TagSet) Contains(tag string) bool {\n\t_, ok := t[tag]\n\treturn ok\n}", "func Contains(set Interface, x interface{}) bool {\n\treturn set.Contains(x)\n}", "func (s *Set) Has(elements ...interface{}) (has bool) {\n\tif len(elements) == 0 {\n\t\treturn false\n\t}\n\n\ts.mutex.RLock()\n\tfor _, element := range elements {\n\t\tif _, has = s.m[element]; !has {\n\t\t\tbreak\n\t\t}\n\t}\n\ts.mutex.RUnlock()\n\n\treturn\n}", "func (this *MyHashSet) Contains(key int) bool {\n\treturn this.s[key] == true\n}", "func (s *Set) Contains(key interface{}) bool {\n\t_, ok := s.index[key]\n\treturn ok\n}", "func (s Set) Contains(val int) bool {\n\tfor i := 0; i < len(s.data); i++ {\n\t\tif s.data[i] == val {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (this *ExDomain) Contains(ele int) bool {\n\t_, exists := this.Values[ele]\n\treturn exists\n}", "func (set *IntSet) Contains(i int) bool {\n\t_, found := set.members[i]\n\treturn found //true if it existed already\n}", "func (s Set) Contains(value interface{}) bool {\n\thash, err := hashstructure.Hash(value, nil)\n\tif err != nil {\n\t\tlog.Errorf(\"type could not be hashed: %+v\", value)\n\t}\n\t_, found := s[hash]\n\treturn found\n}", "func (s FeatureSet) Contains(tag FeatureTag) (c bool) {\n\tfor t := range s {\n\t\tif c = t.Equals(tag); c {\n\t\t\treturn\n\t\t}\n\t}\n\treturn\n}", "func (s *Set) Add(element interface{}) bool {\n\t_, exists := (*s)[element]\n\t(*s)[element] = struct{}{}\n\treturn !exists\n}", "func (s Set) Contains(key string) bool {\n\t_, ok := s[key]\n\treturn ok\n}", "func Contains[T comparable](collection []T, element T) bool {\n\tfor _, item := range collection {\n\t\tif item == element {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func Contains[T comparable](collection []T, element T) bool {\n\tfor _, item := range collection {\n\t\tif item == element {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (s *Set) Contains(item string) bool {\n\treturn s.m[item]\n}", "func (s *stackImpl) Contains(t T) bool {\n\tfor _, n := range s.items {\n\t\tif t == n {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (e EngineTypeSet) Contains(tp EngineType) bool {\n\treturn uint(e)&uint(tp) != 0\n}", "func (s Set) Contains(k string) bool {\n\t_, ok := s.m[k]\n\treturn ok\n}", "func (this *MyHashSet) Contains(key int) bool {\n\tif len(this.items) == 0 {\n\t\treturn false\n\t}\n\tfor _, v := range this.items {\n\t\tif v == key {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (this *MyHashSet) Contains(key int) bool {\n\t_, ok := this.s[key]\n\treturn ok\n}", "func (h *HashSet) Contains(key interface{}) bool {\n\tif _, ok := h.innerMap[key]; ok {\n\t\treturn ok\n\t}\n\treturn false\n}", "func (ll *LinkedList) Contains(element int) bool {\n\tcurrent := ll.start\n\tfor current != nil {\n\t\tif current.value == element {\n\t\t\treturn true\n\t\t}\n\t\tcurrent = current.next\n\t}\n\treturn false\n}", "func (list *ArrayList) Contains(elements ...interface{}) bool {\n\tvar flag bool = true\n\tfor _, e := range elements {\n\t\tif !list.contain(e) {\n\t\t\tflag = false\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn flag\n}", "func (set *lalrSet) contains(item lalrItem) bool {\n\thash := item.hash()\n\n\tif items, ok := set.items[hash]; ok {\n\t\tfor _, item2 := range(items) {\n\t\t\tif item.equals(item2) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false\n}", "func (s Set) Contains(t Set) bool {\n\tif len(t) > len(s) {\n\t\treturn false\n\t}\n\tfor n := range t {\n\t\tif _, ok := s[n]; !ok {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (ss Set) Contains(k string) bool {\n\treturn ss[k]\n}", "func (es EntityIDSet) Contains(id EntityID) bool {\n\t_, ok := es[id]\n\treturn ok\n}", "func (this *MyHashSet) Contains(key int) bool {\n\tflag := false\n\tk := this.HashFunc(key)\n\tfor _, v := range this.value[k] {\n\t\tif v == key {\n\t\t\tflag = true\n\t\t}\n\t}\n\treturn flag\n}", "func (set StringSet) Contains(e string) bool {\n\tval, ok := set[e]\n\treturn ok && val\n}", "func (set *Set) Contains(items ...interface{}) bool {\n\tfor _, item := range items {\n\t\tif _, contains := set.tree.Get(item); !contains {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (v *Set) Contains(b []byte) bool {\n\tiRepr := new(big.Int).SetBytes(b)\n\t_, found := v.indexOf(iRepr)\n\treturn found\n}", "func (s IntSet) Contains(key int) bool {\n\t_, exists := s[key]\n\treturn exists\n}", "func (set AnswerSetUnion) Has(x interface{}) bool {\n\tfor i := range set {\n\t\tif set[i].Has(x) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (set *AppleSet) Contains(i Apple) bool {\n\tif set == nil {\n\t\treturn false\n\t}\n\n\tset.s.RLock()\n\tdefer set.s.RUnlock()\n\n\t_, found := set.m[i]\n\treturn found\n}", "func (s StringSet) Has(element string) bool {\n\t_, ok := s.Set[element]\n\treturn ok\n}", "func (s *Set) Contains(key interface{}) bool {\n\t_, ok := s.skiplist.Get(key)\n\treturn ok\n}", "func Contains(container []int, element int) bool {\n\tfor _, a := range container {\n\t\tif a == element {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (set AnswerSetIntersection) Has(x interface{}) bool {\n\tfor i := range set {\n\t\tif !set[i].Has(x) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func Contains(slice []string, element string) bool {\n\tfor _, v := range slice {\n\t\tif v == element {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (set ModuleIdentifierSet) Contains(id ModuleIdentifier) bool {\n\tfor _, sid := range set.identifiers {\n\t\tif sid == id {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(list []interface{}, elem interface{}) bool {\n\tfor _, t := range list {\n\t\tif t == elem {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (this *MyHashSet) Contains(key int) bool {\n\toffset := key >> Offset\n\treturn (this.Set[offset] & uint8(1<<(uint(key)&(Base-1)))) != 0\n}", "func contains(array []string, element string) bool {\n\tfor _, curElement := range array {\n\t\tif curElement == element {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (b *Bitset) Contains(index uint) bool {\n\treturn b.BS.Test(index)\n}", "func (s *IntSet) Contains(i int) bool {\n\t_, ok := s.members[i]\n\n\treturn ok\n}", "func (s IntSet) Has(e int) bool {\n\t_, ok := s[e]\n\treturn ok\n}", "func (this *MyHashSet) Contains(key int) bool {\n\tindex := hash(key)\n\tfor _, v := range this.bucket[index] {\n\t\tif v == key {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (this *MyHashSet) Contains(key int) bool {\n\tindex := hash(key)\n\tfor _, v := range this.bucket[index] {\n\t\tif v == key {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (this *LinkedList) Contains(ele interface{}) bool {\n\tif this.head == nil {\n\t\treturn false\n\t}\n\tpe := this.head\n\tfor pe != this.tail {\n\t\tif pe.elem == ele {\n\t\t\treturn true\n\t\t}\n\t\tpe = pe.next\n\t}\n\treturn false\n}", "func (this *MultiMap) Contains(value interface{}) bool {\n\tif this.tree.Find(value) != nil {\n\t\treturn true\n\t}\n\treturn false\n}", "func (s *Set) Has(value string) (bool, error) {\n\tif !s.host.rawUTF8 {\n\t\tEncode(&value)\n\t}\n\trows, err := s.host.db.Query(fmt.Sprintf(\"SELECT %s FROM %s WHERE %s = $1\", setCol, s.table, setCol), value)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tif rows == nil {\n\t\treturn false, errors.New(\"Set Has returned no rows for value \" + value)\n\t}\n\tdefer rows.Close()\n\tvar scanValue sql.NullString\n\t// Get the value. Should not loop more than once.\n\tcounter := 0\n\tfor rows.Next() {\n\t\terr = rows.Scan(&scanValue)\n\t\tif err != nil {\n\t\t\t// No rows\n\t\t\treturn false, err\n\t\t}\n\t\tcounter++\n\t}\n\tif err := rows.Err(); err != nil {\n\t\treturn false, err\n\t}\n\t//if counter > 1 {\n\t// more than one element that has the same *value* is fine!\n\t//}\n\treturn counter > 0, nil\n}", "func (set *SetThreadSafe) Contains(Items ...interface{}) bool {\n\tfor _, item := range Items {\n\t\tif _, contains := set.Items.Load(item); !contains {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (set ResultSet) Contains(result Result) bool {\n\tfor _, res := range set {\n\t\tif res == result {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (ts *TagSet) has(t ident.Ident) bool {\n\t// TODO: this could be much, much faster! Maybe build a set on\n\t// each TagSet as required (with sync.Once)?\n\n\tfor _, t2 := range ts.tags {\n\t\tif t2.Equals(t) {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func Contains(array []string, element string) bool {\r\n\tfor _, e := range array {\r\n\t\tif e == element {\r\n\t\t\treturn true\r\n\t\t}\r\n\t}\r\n\treturn false\r\n}", "func (d *Doubly) Contains(comparison func(data interface{}) (exists bool)) bool {\n\td.rwLock.RLock()\n\tdefer d.rwLock.RUnlock()\n\tfor tmp := d.head; tmp != nil; tmp = tmp.Next {\n\t\tif comparison(tmp.Data) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (dict *Dictionary) Contains(key DictKey) bool {\n\tdict.lock.RLock()\n\tdefer dict.lock.RUnlock()\n\tvar exists bool\n\t_, exists = dict.elements[key]\n\treturn exists\n}", "func (s *Set) Has(val int) bool {\n\t_, ok := s.set[val]\n\treturn ok\n}", "func Contains(data, elem any) (valid, found bool) {\n\tif data == nil {\n\t\treturn false, false\n\t}\n\n\tdataRv := reflect.ValueOf(data)\n\tdataRt := reflect.TypeOf(data)\n\tdataKind := dataRt.Kind()\n\n\t// string\n\tif dataKind == reflect.String {\n\t\treturn true, strings.Contains(dataRv.String(), fmt.Sprint(elem))\n\t}\n\n\t// map\n\tif dataKind == reflect.Map {\n\t\tmapKeys := dataRv.MapKeys()\n\t\tfor i := 0; i < len(mapKeys); i++ {\n\t\t\tif reflects.IsEqual(mapKeys[i].Interface(), elem) {\n\t\t\t\treturn true, true\n\t\t\t}\n\t\t}\n\t\treturn true, false\n\t}\n\n\t// array, slice - other return false\n\tif dataKind != reflect.Slice && dataKind != reflect.Array {\n\t\treturn false, false\n\t}\n\n\tfor i := 0; i < dataRv.Len(); i++ {\n\t\tif reflects.IsEqual(dataRv.Index(i).Interface(), elem) {\n\t\t\treturn true, true\n\t\t}\n\t}\n\treturn true, false\n}", "func (al *ArrayList) Contains(value interface{}) bool {\n\tvar flag bool\n\tfor i := 0; i < al.Size(); i++ {\n\t\tif al.slice[i] == value {\n\t\t\tflag = true\n\t\t\tbreak\n\t\t}\n\t}\n\treturn flag\n}", "func (s *SeriesIDSet) Contains(id uint64) bool {\n\ts.RLock()\n\tx := s.ContainsNoLock(id)\n\ts.RUnlock()\n\treturn x\n}", "func (s *Set) Remove(element interface{}) bool {\n\tif _, exists := (*s)[element]; exists {\n\t\tdelete(*s, element)\n\t\treturn true\n\t}\n\treturn false\n}", "func (d *DSU) Contains(x interface{}) bool {\n\t_, ok := d.nodes[x]\n\n\treturn ok\n}", "func (s *Set) has(c cid.Cid) bool {\n\t_, ok := s.set[c]\n\treturn ok\n}", "func (ts TagSet) Has(tag string) bool {\n\treturn ts.Find(tag) != -1\n}", "func (set *Set) Contains(items ...interface{}) bool {\n\tfor _, item := range items {\n\t\tif _, contains := set.items[item]; !contains {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (t *Set) Has(key []byte) bool {\n\t// test for empty tree\n\tif t.Empty() {\n\t\treturn false\n\t}\n\t// walk for best member\n\tp := t.root\n\n\tfor p.node != nil {\n\t\t// try next node\n\t\tp = p.node.child[p.node.dir(key)]\n\t}\n\t// check for membership\n\tklen := len(key)\n\tif klen != len(p.Key) {\n\t\treturn false\n\t}\n\tfor i, b := range p.Key {\n\t\tif b != key[i] {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (cs Set) Contains(c Code) bool {\n\tfor _, cc := range cs {\n\t\tif cc == c {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (ss *StringSet) Contain(s string) bool {\n\t_, exist := ss.set[s]\n\treturn exist\n}", "func (this *MyHashSet) Contains(key int) bool {\n return this.arr[key]\n\n}", "func (r Expansions) Contains(id string) bool { _, ok := r.Get(id); return ok }", "func (mb *MutableBag) Contains(key string) bool {\n\tif _, found := mb.values[key]; found {\n\t\treturn true\n\t}\n\n\treturn mb.parent.Contains(key)\n}", "func (s *Set) Has(c cid.Cid) bool {\n\ts.lk.Lock()\n\tdefer s.lk.Unlock()\n\t_, ok := s.set[c]\n\treturn ok\n}", "func (s Uint64Set) Has(item uint64) bool {\n\t_, contained := s[item]\n\treturn contained\n}", "func (s *MyHashSet) Contains(key int) bool {\n\tbit := key % 64\n\tlength := key / 64\n\tif length >= len(s.bitset) {\n\t\treturn false\n\t}\n\treturn s.bitset[length]&(1<<uint(bit)) != 0\n}", "func (s *intSet) has(i int) bool {\n\tif s == nil {\n\t\treturn false\n\t}\n\treturn s.members[i] > 0\n}", "func (ls *ListStack) Contains(item adts.ContainerElement) bool {\n\tif ls.threadSafe {\n\t\tls.lock.Lock()\n\t\tdefer ls.lock.Unlock()\n\t\treturn ls.containsHelper(item)\n\t}\n\n\treturn ls.containsHelper(item)\n}", "func (s *Set) Remove(elem A) bool {\n\t_, found := s.m[elem]\n\tif found {\n\t\tdelete(s.m, elem)\n\t}\n\treturn found\n}", "func Contains(s, sub any) bool {\n\tok, found := includeElement(s, sub)\n\n\t// ok == false: 's' could not be applied builtin len()\n\t// found == false: 's' does not contain 'sub'\n\treturn ok && found\n}", "func (set *IntSet) Contains(i int) bool {\n\treturn set.set[i]\n}", "func (this *MyHashSet) Contains(key int) bool {\n\tbucket := h(uint32(key)) % this.bucket\n\titer := this.Table[bucket].Head\n\tfor iter != nil {\n\t\tif iter.Val == key {\n\t\t\treturn true\n\t\t}\n\t\titer = iter.Next\n\t}\n\treturn false\n}" ]
[ "0.8465254", "0.8461474", "0.8148759", "0.8126085", "0.7917791", "0.76294225", "0.75558245", "0.75072116", "0.74703974", "0.7402724", "0.73940265", "0.73736686", "0.7315672", "0.7295919", "0.7293987", "0.7293061", "0.7264501", "0.72467923", "0.7189788", "0.7184183", "0.7132645", "0.71011126", "0.7095512", "0.70843375", "0.70782447", "0.7047052", "0.7035574", "0.70328265", "0.7009895", "0.70092416", "0.6999734", "0.69713664", "0.69713664", "0.6946021", "0.6940851", "0.6887125", "0.6884955", "0.68810415", "0.6873405", "0.6871504", "0.68688864", "0.6855816", "0.68203884", "0.68173206", "0.6814486", "0.68061864", "0.6804528", "0.6801287", "0.6792697", "0.67918116", "0.67878395", "0.67863584", "0.6781621", "0.6779252", "0.67492545", "0.67261815", "0.6711594", "0.67017686", "0.66850173", "0.66829073", "0.6682006", "0.66817737", "0.66743886", "0.66722256", "0.66680163", "0.66545033", "0.66545033", "0.66359806", "0.6633918", "0.6629863", "0.66225326", "0.66099924", "0.66026366", "0.65975976", "0.65955764", "0.6585019", "0.65756494", "0.6565292", "0.65450984", "0.6541888", "0.65376794", "0.6533785", "0.65181977", "0.6516554", "0.65118355", "0.6511221", "0.64932066", "0.64910257", "0.6474129", "0.6464776", "0.64631194", "0.6456167", "0.64496756", "0.64478403", "0.64468616", "0.6441748", "0.6424289", "0.6417574", "0.64156574", "0.64123076" ]
0.6651313
67
UnmarshalJSON implements the json.Unmarshal interface and unmarshals a JSON object in the Name struct or a string as just the subject common name.
func (n *Name) UnmarshalJSON(data []byte) error { if cn, ok := maybeString(data); ok { n.CommonName = cn return nil } type nameAlias Name var nn nameAlias if err := json.Unmarshal(data, &nn); err != nil { return errors.Wrap(err, "error unmarshaling json") } *n = Name(nn) return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Subject) UnmarshalJSON(data []byte) error {\n\tvar name Name\n\tif err := name.UnmarshalJSON(data); err != nil {\n\t\treturn err\n\t}\n\t*s = Subject(name)\n\treturn nil\n}", "func (n *ACName) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn err\n\t}\n\tnn, err := NewACName(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*n = *nn\n\treturn nil\n}", "func (n *Name) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &n.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &n.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *Name) UnmarshalJSON(data []byte) error {\n\tstr := string(data)\n\tstart := strings.Index(str, \"\\\"\")\n\tlast := strings.LastIndex(str, \"\\\"\")\n\tif start == 0 && last == len(str)-1 {\n\t\tstr = str[start+1 : last]\n\t} else {\n\t\treturn fmt.Errorf(\"bad-format-for-name:%v\", string(data))\n\t}\n\t*n = Name(str)\n\treturn nil\n}", "func ParseNameFromJSON(ctx []byte) (string, error) {\n\tjs, err := simplejson.NewJson(ctx)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"decode json in file failed, err: %v\", err)\n\t}\n\n\tjsMetaData := js.Get(\"metadata\")\n\tname, _ := jsMetaData.Get(\"name\").String()\n\tif name == \"\" {\n\t\treturn \"\", fmt.Errorf(\"parse name failed or json structure error\")\n\t}\n\treturn name, nil\n}", "func (c *ConnectionItemName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *ChannelName) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"ChannelName should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = ChannelNameString(s)\n\treturn err\n}", "func (this *NamespacedName) UnmarshalJSON(b []byte) error {\n\treturn CommonUnmarshaler.Unmarshal(bytes.NewReader(b), this)\n}", "func (s *SingleServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &s.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachine\":\n\t\t\terr = unpopulate(val, \"VirtualMachine\", &s.VirtualMachine)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SharedStorageResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"sharedStorageAccountName\":\n\t\t\terr = unpopulate(val, \"SharedStorageAccountName\", &s.SharedStorageAccountName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sharedStorageAccountPrivateEndPointName\":\n\t\t\terr = unpopulate(val, \"SharedStorageAccountPrivateEndPointName\", &s.SharedStorageAccountPrivateEndPointName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NameProvider) GetJSONName(subject interface{}, name string) (string, bool) {\n\ttpe := reflect.Indirect(reflect.ValueOf(subject)).Type()\n\treturn n.GetJSONNameForType(tpe, name)\n}", "func (r *ResourceName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *ThreeTierFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"applicationServer\":\n\t\t\terr = unpopulate(val, \"ApplicationServer\", &t.ApplicationServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"centralServer\":\n\t\t\terr = unpopulate(val, \"CentralServer\", &t.CentralServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"databaseServer\":\n\t\t\terr = unpopulate(val, \"DatabaseServer\", &t.DatabaseServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &t.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sharedStorage\":\n\t\t\terr = unpopulate(val, \"SharedStorage\", &t.SharedStorage)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SingleServerCustomResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &s.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckNameAvailabilityRequestBody) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (ln *LabelName) UnmarshalJSON(b []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(b, &s); err != nil {\n\t\treturn err\n\t}\n\tif !LabelName(s).IsValid() {\n\t\treturn fmt.Errorf(\"%q is not a valid label name\", s)\n\t}\n\t*ln = LabelName(s)\n\treturn nil\n}", "func (v *NetworkName) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer15(&r, v)\n\treturn r.Error()\n}", "func (c *CheckNameAvailabilityResponse) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &c.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nameAvailable\":\n\t\t\terr = unpopulate(val, \"NameAvailable\", &c.NameAvailable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reason\":\n\t\t\terr = unpopulate(val, \"Reason\", &c.Reason)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CentralServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"availabilitySetName\":\n\t\t\terr = unpopulate(val, \"AvailabilitySetName\", &c.AvailabilitySetName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"loadBalancer\":\n\t\t\terr = unpopulate(val, \"LoadBalancer\", &c.LoadBalancer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachines\":\n\t\t\terr = unpopulate(val, \"VirtualMachines\", &c.VirtualMachines)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UsageName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &u.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &u.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UsageName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &u.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &u.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckNameAvailabilityInput) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *ThreeTierCustomResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &t.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *ApplicationServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"availabilitySetName\":\n\t\t\terr = unpopulate(val, \"AvailabilitySetName\", &a.AvailabilitySetName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachines\":\n\t\t\terr = unpopulate(val, \"VirtualMachines\", &a.VirtualMachines)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckNameAvailabilityResponseBody) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"absCode\":\n\t\t\terr = unpopulate(val, \"AbsCode\", &c.AbsCode)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &c.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"valid\":\n\t\t\terr = unpopulate(val, \"Valid\", &c.Valid)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *ConsortiumMember) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"dateModified\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.DateModified)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, &c.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"joinDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.JoinDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"role\":\n\t\t\terr = unpopulate(val, &c.Role)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, &c.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, &c.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NameAvailabilityInformation) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &n.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nameAvailable\":\n\t\t\terr = unpopulate(val, \"NameAvailable\", &n.NameAvailable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reason\":\n\t\t\terr = unpopulate(val, \"Reason\", &n.Reason)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *Issuer) UnmarshalJSON(data []byte) error {\n\tvar name Name\n\tif err := name.UnmarshalJSON(data); err != nil {\n\t\treturn err\n\t}\n\t*i = Issuer(name)\n\treturn nil\n}", "func (n *NameAvailabilityInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &n.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nameAvailable\":\n\t\t\terr = unpopulate(val, \"NameAvailable\", &n.NameAvailable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reason\":\n\t\t\terr = unpopulate(val, \"Reason\", &n.Reason)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckNameAvailabilityRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CheckResourceNameResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, \"Status\", &c.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func ParseNamespaceNameFromJSON(ctx []byte) (string, string, error) {\n\tjs, err := simplejson.NewJson(ctx)\n\tif err != nil {\n\t\treturn \"\", \"\", fmt.Errorf(\"decode json in file failed, err: %v\", err)\n\t}\n\n\tjsMetaData := js.Get(\"metadata\")\n\tnamespace, _ := jsMetaData.Get(\"namespace\").String()\n\tname, _ := jsMetaData.Get(\"name\").String()\n\tif namespace == \"\" {\n\t\treturn \"\", \"\", fmt.Errorf(\"parse namespace failed or json structure error\")\n\t}\n\tif name == \"\" {\n\t\treturn \"\", \"\", fmt.Errorf(\"parse name failed or json structure error\")\n\t}\n\treturn namespace, name, nil\n}", "func (h *Hostname) UnmarshalJSON(data []byte) error {\n\tvar hstr string\n\tif err := json.Unmarshal(data, &hstr); err != nil {\n\t\treturn err\n\t}\n\t*h = Hostname(hstr)\n\treturn nil\n}", "func (v *SyntheticsTestRequestBodyType) UnmarshalJSON(src []byte) error {\n\tvar value string\n\terr := json.Unmarshal(src, &value)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*v = SyntheticsTestRequestBodyType(value)\n\treturn nil\n}", "func (r *TestCasing) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"TestCasing should be a string, got %s\", data)\n\t}\n\tv, ok := _TestCasingNameToValue[s]\n\tif !ok {\n\t\treturn fmt.Errorf(\"invalid TestCasing %q\", s)\n\t}\n\t*r = v\n\treturn nil\n}", "func (r *RegisteredAsn) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &r.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *FormulaAndFunctionApmDependencyStatName) UnmarshalJSON(src []byte) error {\n\tvar value string\n\terr := json.Unmarshal(src, &value)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*v = FormulaAndFunctionApmDependencyStatName(value)\n\treturn nil\n}", "func (h *Human) UnmarshalJSON(data []byte) error {\n\ttype Alias Human\n\taux := &struct {\n\t\t*Alias\n\t}{\n\t\tAlias: (*Alias)(h),\n\t}\n\n\tif err := json.Unmarshal(data, &aux); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (m *MetricName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &m.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &m.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", m, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *DocumentSubjectUpdate) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonDe066f3DecodeGithubComStudtoolDocumentsServiceModels3(&r, v)\n\treturn r.Error()\n}", "func (c *CopyAudio) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"label\":\n\t\t\terr = unpopulate(val, \"Label\", &c.Label)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &c.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (d *DatabaseServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"availabilitySetName\":\n\t\t\terr = unpopulate(val, \"AvailabilitySetName\", &d.AvailabilitySetName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"loadBalancer\":\n\t\t\terr = unpopulate(val, \"LoadBalancer\", &d.LoadBalancer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachines\":\n\t\t\terr = unpopulate(val, \"VirtualMachines\", &d.VirtualMachines)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UID) UnmarshalJSON(data []byte) error {\n\tu.Str = string(data[1 : len(data)-1])\n\treturn nil\n}", "func UnmarshalJSON(ct string, data []byte) (proto.Message, error) {\n\n\tctn, p, err := mime.ParseMediaType(ct)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn UnmarshalJSONParams(ctn, p, data)\n\n}", "func (m *gohaiMarshaler) UnmarshalJSON(bytes []byte) error {\n\tfirstUnmarshall := \"\"\n\terr := json.Unmarshal(bytes, &firstUnmarshall)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = json.Unmarshal([]byte(firstUnmarshall), &(m.gohai))\n\treturn err\n}", "func (v *Header) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson171edd05DecodeGithubComEmirmuminogluJwt(&r, v)\n\treturn r.Error()\n}", "func (c *CheckNameAvailabilityParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *Thing) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Name = &name\n\t\t\t}\n\t\tcase \"url\":\n\t\t\tif v != nil {\n\t\t\t\tvar URL string\n\t\t\t\terr = json.Unmarshal(*v, &URL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.URL = &URL\n\t\t\t}\n\t\tcase \"image\":\n\t\t\tif v != nil {\n\t\t\t\tvar imageVar ImageObject\n\t\t\t\terr = json.Unmarshal(*v, &imageVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Image = &imageVar\n\t\t\t}\n\t\tcase \"description\":\n\t\t\tif v != nil {\n\t\t\t\tvar description string\n\t\t\t\terr = json.Unmarshal(*v, &description)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Description = &description\n\t\t\t}\n\t\tcase \"entityPresentationInfo\":\n\t\t\tif v != nil {\n\t\t\t\tvar entityPresentationInfo EntitiesEntityPresentationInfo\n\t\t\t\terr = json.Unmarshal(*v, &entityPresentationInfo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.EntityPresentationInfo = &entityPresentationInfo\n\t\t\t}\n\t\tcase \"bingId\":\n\t\t\tif v != nil {\n\t\t\t\tvar bingID string\n\t\t\t\terr = json.Unmarshal(*v, &bingID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.BingID = &bingID\n\t\t\t}\n\t\tcase \"contractualRules\":\n\t\t\tif v != nil {\n\t\t\t\tcontractualRules, err := unmarshalBasicContractualRulesContractualRuleArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.ContractualRules = &contractualRules\n\t\t\t}\n\t\tcase \"webSearchUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar webSearchURL string\n\t\t\t\terr = json.Unmarshal(*v, &webSearchURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.WebSearchURL = &webSearchURL\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.ID = &ID\n\t\t\t}\n\t\tcase \"_type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar TypeBasicResponseBase\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Type = typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (o *Organization) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Name = &name\n\t\t\t}\n\t\tcase \"url\":\n\t\t\tif v != nil {\n\t\t\t\tvar URL string\n\t\t\t\terr = json.Unmarshal(*v, &URL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.URL = &URL\n\t\t\t}\n\t\tcase \"image\":\n\t\t\tif v != nil {\n\t\t\t\tvar imageVar ImageObject\n\t\t\t\terr = json.Unmarshal(*v, &imageVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Image = &imageVar\n\t\t\t}\n\t\tcase \"description\":\n\t\t\tif v != nil {\n\t\t\t\tvar description string\n\t\t\t\terr = json.Unmarshal(*v, &description)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Description = &description\n\t\t\t}\n\t\tcase \"entityPresentationInfo\":\n\t\t\tif v != nil {\n\t\t\t\tvar entityPresentationInfo EntitiesEntityPresentationInfo\n\t\t\t\terr = json.Unmarshal(*v, &entityPresentationInfo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.EntityPresentationInfo = &entityPresentationInfo\n\t\t\t}\n\t\tcase \"bingId\":\n\t\t\tif v != nil {\n\t\t\t\tvar bingID string\n\t\t\t\terr = json.Unmarshal(*v, &bingID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.BingID = &bingID\n\t\t\t}\n\t\tcase \"contractualRules\":\n\t\t\tif v != nil {\n\t\t\t\tcontractualRules, err := unmarshalBasicContractualRulesContractualRuleArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.ContractualRules = &contractualRules\n\t\t\t}\n\t\tcase \"webSearchUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar webSearchURL string\n\t\t\t\terr = json.Unmarshal(*v, &webSearchURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.WebSearchURL = &webSearchURL\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.ID = &ID\n\t\t\t}\n\t\tcase \"_type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar TypeBasicResponseBase\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Type = typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (h *HTTPHeader) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &h.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &h.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", h, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (ra *RegisteredAsn) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"properties\":\n\t\t\tif v != nil {\n\t\t\t\tvar registeredAsnProperties RegisteredAsnProperties\n\t\t\t\terr = json.Unmarshal(*v, &registeredAsnProperties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.RegisteredAsnProperties = &registeredAsnProperties\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.Name = &name\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.ID = &ID\n\t\t\t}\n\t\tcase \"type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar string\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.Type = &typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (e *EntityNameAvailabilityCheckOutput) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"message\":\n\t\t\terr = unpopulate(val, \"Message\", &e.Message)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"nameAvailable\":\n\t\t\terr = unpopulate(val, \"NameAvailable\", &e.NameAvailable)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reason\":\n\t\t\terr = unpopulate(val, \"Reason\", &e.Reason)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *NetworkInterfaceResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"networkInterfaceName\":\n\t\t\terr = unpopulate(val, \"NetworkInterfaceName\", &n.NetworkInterfaceName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *NetworkName) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer15(l, v)\n}", "func (v *Student) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests1(&r, v)\n\treturn r.Error()\n}", "func (m *Message) UnmarshalJSONObject(dec *gojay.Decoder, k string) error {\n\n\tswitch k {\n\tcase \"Id\":\n\t\treturn dec.Int(&m.Id)\n\n\tcase \"Name\":\n\t\treturn dec.String(&m.Name)\n\n\tcase \"Price\":\n\t\treturn dec.Float64(&m.Price)\n\n\tcase \"Ints\":\n\t\tvar aSlice = Ints{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.Ints = []int(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"Floats\":\n\t\tvar aSlice = Float32s{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.Floats = []float32(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"SubMessageX\":\n\t\tvar value = &SubMessage{}\n\t\terr := dec.Object(value)\n\t\tif err == nil {\n\t\t\tm.SubMessageX = value\n\t\t}\n\n\t\treturn err\n\n\tcase \"MessagesX\":\n\t\tvar aSlice = SubMessagesPtr{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.MessagesX = []*SubMessage(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"SubMessageY\":\n\t\terr := dec.Object(&m.SubMessageY)\n\n\t\treturn err\n\n\tcase \"MessagesY\":\n\t\tvar aSlice = SubMessages{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.MessagesY = []SubMessage(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"IsTrue\":\n\t\tvar value bool\n\t\terr := dec.Bool(&value)\n\t\tif err == nil {\n\t\t\tm.IsTrue = &value\n\t\t}\n\t\treturn err\n\n\tcase \"Payload\":\n\t\tvar value = gojay.EmbeddedJSON{}\n\t\terr := dec.AddEmbeddedJSON(&value)\n\t\tif err == nil && len(value) > 0 {\n\t\t\tm.Payload = []byte(value)\n\t\t}\n\t\treturn err\n\n\tcase \"SQLNullString\":\n\t\tvar value = &sql.NullString{}\n\t\terr := dec.SQLNullString(value)\n\t\tif err == nil {\n\t\t\tm.SQLNullString = value\n\t\t}\n\t\treturn err\n\n\t}\n\treturn nil\n}", "func (j *Message) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (v *MusicianFullInformation) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson62dc445bDecode20211NoskoolTeamInternalAppMusiciansModels1(&r, v)\n\treturn r.Error()\n}", "func (receiver *Type) UnmarshalJSON(src []byte) error {\n\tif nil == receiver {\n\t\treturn errNilReceiver\n\t}\n\n\tvar s string\n\tif err := json.Unmarshal(src, &s); nil != err {\n\t\treturn err\n\t}\n\n\tif err := receiver.Scan(s); nil != err {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (r *Energierichtung) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"Energierichtung should be a string, got %s\", data)\n\t}\n\tv, ok := _EnergierichtungNameToValue[s]\n\tif !ok {\n\t\treturn fmt.Errorf(\"invalid Energierichtung %q\", s)\n\t}\n\t*r = v\n\treturn nil\n}", "func (c *CommonJobDetails) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"actions\":\n\t\t\terr = unpopulate(val, \"Actions\", &c.Actions)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"chainOfCustodySasKey\":\n\t\t\terr = unpopulate(val, \"ChainOfCustodySasKey\", &c.ChainOfCustodySasKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"contactDetails\":\n\t\t\terr = unpopulate(val, \"ContactDetails\", &c.ContactDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"copyLogDetails\":\n\t\t\tc.CopyLogDetails, err = unmarshalCopyLogDetailsClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataCenterCode\":\n\t\t\terr = unpopulate(val, \"DataCenterCode\", &c.DataCenterCode)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataExportDetails\":\n\t\t\terr = unpopulate(val, \"DataExportDetails\", &c.DataExportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataImportDetails\":\n\t\t\terr = unpopulate(val, \"DataImportDetails\", &c.DataImportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"datacenterAddress\":\n\t\t\tc.DatacenterAddress, err = unmarshalDatacenterAddressResponseClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deliveryPackage\":\n\t\t\terr = unpopulate(val, \"DeliveryPackage\", &c.DeliveryPackage)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deviceErasureDetails\":\n\t\t\terr = unpopulate(val, \"DeviceErasureDetails\", &c.DeviceErasureDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expectedDataSizeInTeraBytes\":\n\t\t\terr = unpopulate(val, \"ExpectedDataSizeInTeraBytes\", &c.ExpectedDataSizeInTeraBytes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"jobDetailsType\":\n\t\t\terr = unpopulate(val, \"JobDetailsType\", &c.JobDetailsType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"jobStages\":\n\t\t\terr = unpopulate(val, \"JobStages\", &c.JobStages)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"keyEncryptionKey\":\n\t\t\terr = unpopulate(val, \"KeyEncryptionKey\", &c.KeyEncryptionKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastMitigationActionOnJob\":\n\t\t\terr = unpopulate(val, \"LastMitigationActionOnJob\", &c.LastMitigationActionOnJob)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"preferences\":\n\t\t\terr = unpopulate(val, \"Preferences\", &c.Preferences)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"returnPackage\":\n\t\t\terr = unpopulate(val, \"ReturnPackage\", &c.ReturnPackage)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reverseShipmentLabelSasKey\":\n\t\t\terr = unpopulate(val, \"ReverseShipmentLabelSasKey\", &c.ReverseShipmentLabelSasKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"shippingAddress\":\n\t\t\terr = unpopulate(val, \"ShippingAddress\", &c.ShippingAddress)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *CardType) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"CardType should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = CardTypeString(s)\n\treturn err\n}", "func (r *Geschaeftspartnerrolle) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"Geschaeftspartnerrolle should be a string, got %s\", data)\n\t}\n\tv, ok := _GeschaeftspartnerrolleNameToValue[s]\n\tif !ok {\n\t\treturn fmt.Errorf(\"invalid Geschaeftspartnerrolle %q\", s)\n\t}\n\t*r = v\n\treturn nil\n}", "func (j *Publisher) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (d *Delegation) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &d.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &d.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &d.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &d.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &d.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *ActivityReportTitleUpdated) UnmarshalJSON(b []byte) error {\n\tvar helper activityReportTitleUpdatedUnmarshalHelper\n\tif err := json.Unmarshal(b, &helper); err != nil {\n\t\treturn err\n\t}\n\t*a = ActivityReportTitleUpdated(helper.Attributes)\n\treturn nil\n}", "func (u *SSN) UnmarshalJSON(data []byte) error {\n\tif string(data) == jsonNull {\n\t\treturn nil\n\t}\n\tvar ustr string\n\tif err := json.Unmarshal(data, &ustr); err != nil {\n\t\treturn err\n\t}\n\t*u = SSN(ustr)\n\treturn nil\n}", "func (dst *SearchSuggestItemResponse) UnmarshalJSON(data []byte) error {\n\tvar err error\n\t// use discriminator value to speed up the lookup\n\tvar jsonDict map[string]interface{}\n\terr = newStrictDecoder(data).Decode(&jsonDict)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to unmarshal JSON into map for the discriminator lookup.\")\n\t}\n\n\t// check if the discriminator value is 'mo.DocumentCount'\n\tif jsonDict[\"ObjectType\"] == \"mo.DocumentCount\" {\n\t\t// try to unmarshal JSON data into MoDocumentCount\n\t\terr = json.Unmarshal(data, &dst.MoDocumentCount)\n\t\tif err == nil {\n\t\t\treturn nil // data stored in dst.MoDocumentCount, return on the first match\n\t\t} else {\n\t\t\tdst.MoDocumentCount = nil\n\t\t\treturn fmt.Errorf(\"Failed to unmarshal SearchSuggestItemResponse as MoDocumentCount: %s\", err.Error())\n\t\t}\n\t}\n\n\t// check if the discriminator value is 'search.SuggestItem.List'\n\tif jsonDict[\"ObjectType\"] == \"search.SuggestItem.List\" {\n\t\t// try to unmarshal JSON data into SearchSuggestItemList\n\t\terr = json.Unmarshal(data, &dst.SearchSuggestItemList)\n\t\tif err == nil {\n\t\t\treturn nil // data stored in dst.SearchSuggestItemList, return on the first match\n\t\t} else {\n\t\t\tdst.SearchSuggestItemList = nil\n\t\t\treturn fmt.Errorf(\"Failed to unmarshal SearchSuggestItemResponse as SearchSuggestItemList: %s\", err.Error())\n\t\t}\n\t}\n\n\treturn nil\n}", "func (n *NameProvider) GetJSONNames(subject interface{}) []string {\n\tn.lock.Lock()\n\tdefer n.lock.Unlock()\n\ttpe := reflect.Indirect(reflect.ValueOf(subject)).Type()\n\tnames, ok := n.index[tpe]\n\tif !ok {\n\t\tnames = n.makeNameIndex(tpe)\n\t}\n\n\tres := make([]string, 0, len(names.jsonNames))\n\tfor k := range names.jsonNames {\n\t\tres = append(res, k)\n\t}\n\treturn res\n}", "func (v *VirtualNetworkUsageName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &v.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &v.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Teacher) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests(&r, v)\n\treturn r.Error()\n}", "func (f *FlexString) UnmarshalJSON(b []byte) error {\n\tif b[0] != '\"' {\n\t\tvar i int\n\t\terr := json.Unmarshal(b, &i)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t*f = FlexString(strconv.Itoa(i))\n\t\treturn nil\n\t}\n\treturn json.Unmarshal(b, (*string)(f))\n}", "func (h *Header) DecodeJSON(v interface{}) error {\n\tif len(*h) < 7 {\n\t\treturn ErrNonJSONHeader\n\t}\n\n\treturn json.Unmarshal(*h, &v)\n}", "func (d *DeviceServiceCheckNameAvailabilityParameters) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &d.Name)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (this *Simple) UnmarshalJSON(b []byte) error {\n\treturn TypesUnmarshaler.Unmarshal(bytes.NewReader(b), this)\n}", "func (j *jsonNative) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (j *LuaString) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (s *StaticMember) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &s.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &s.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *EventMsg) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (sr *SearchResponse) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"queryContext\":\n\t\t\tif v != nil {\n\t\t\t\tvar queryContext QueryContext\n\t\t\t\terr = json.Unmarshal(*v, &queryContext)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.QueryContext = &queryContext\n\t\t\t}\n\t\tcase \"entities\":\n\t\t\tif v != nil {\n\t\t\t\tvar entities Entities\n\t\t\t\terr = json.Unmarshal(*v, &entities)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.Entities = &entities\n\t\t\t}\n\t\tcase \"places\":\n\t\t\tif v != nil {\n\t\t\t\tvar places Places\n\t\t\t\terr = json.Unmarshal(*v, &places)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.Places = &places\n\t\t\t}\n\t\tcase \"contractualRules\":\n\t\t\tif v != nil {\n\t\t\t\tcontractualRules, err := unmarshalBasicContractualRulesContractualRuleArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.ContractualRules = &contractualRules\n\t\t\t}\n\t\tcase \"webSearchUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar webSearchURL string\n\t\t\t\terr = json.Unmarshal(*v, &webSearchURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.WebSearchURL = &webSearchURL\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.ID = &ID\n\t\t\t}\n\t\tcase \"_type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar TypeBasicResponseBase\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.Type = typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (v *VirtualMachineResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"dataDiskNames\":\n\t\t\terr = unpopulate(val, \"DataDiskNames\", &v.DataDiskNames)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"hostName\":\n\t\t\terr = unpopulate(val, \"HostName\", &v.HostName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"networkInterfaces\":\n\t\t\terr = unpopulate(val, \"NetworkInterfaces\", &v.NetworkInterfaces)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"osDiskName\":\n\t\t\terr = unpopulate(val, \"OSDiskName\", &v.OSDiskName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"vmName\":\n\t\t\terr = unpopulate(val, \"VMName\", &v.VMName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", v, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *Type) UnmarshalJSON(b []byte) error {\n\tvar text string\n\tif err := json.Unmarshal(b, &text); err != nil {\n\t\treturn err\n\t}\n\n\treturn t.UnmarshalText([]byte(text))\n}", "func (m *SubMessage) UnmarshalJSONObject(dec *gojay.Decoder, k string) error {\n\n\tswitch k {\n\tcase \"Id\":\n\t\treturn dec.Int(&m.Id)\n\n\tcase \"Description\":\n\t\treturn dec.String(&m.Description)\n\n\tcase \"StartTime\":\n\t\tvar format = time.RFC3339\n\t\tvar value = time.Time{}\n\t\terr := dec.Time(&value, format)\n\t\tif err == nil {\n\t\t\tm.StartTime = value\n\t\t}\n\t\treturn err\n\n\tcase \"EndTime\":\n\t\tvar format = time.RFC3339\n\t\tvar value = &time.Time{}\n\t\terr := dec.Time(value, format)\n\t\tif err == nil {\n\t\t\tm.EndTime = value\n\t\t}\n\t\treturn err\n\n\t}\n\treturn nil\n}", "func (o *OneOfStringDoaGddGAStringDoaGddGABERs71N5) UnmarshalJSON(bytes []byte) error {\n\tvar myStringDoaGddGA StringDoaGddGA\n\tif err := json.Unmarshal(bytes, &myStringDoaGddGA); err == nil {\n\t\to.StringDoaGddGA = &myStringDoaGddGA\n\t\treturn nil\n\t}\n\tvar myStringDoaGddGA StringDoaGddGA\n\tif err := json.Unmarshal(bytes, &myStringDoaGddGA); err == nil {\n\t\to.StringDoaGddGA = &myStringDoaGddGA\n\t\treturn nil\n\t}\n\treturn errors.New(\"failed to unmarshal one of the object properties\")\n}", "func (s *StringBeginsWithAdvancedFilter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"key\":\n\t\t\terr = unpopulate(val, \"Key\", &s.Key)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"operatorType\":\n\t\t\terr = unpopulate(val, \"OperatorType\", &s.OperatorType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"values\":\n\t\t\terr = unpopulate(val, \"Values\", &s.Values)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *MessageType) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"MessageType should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = MessageTypeString(s)\n\treturn err\n}", "func (v *NullString) UnmarshalJSON(data []byte) error {\r\n\tvar x *string\r\n\tif err := json.Unmarshal(data, &x); err != nil {\r\n\t\treturn err\r\n\t}\r\n\tif x != nil {\r\n\t\tv.String = *x\r\n\t\tv.Valid = true\r\n\t} else {\r\n\t\tv.Valid = false\r\n\t}\r\n\treturn nil\r\n}", "func (v *commonRule) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson795c59c6DecodeGrapeGuardRules4(&r, v)\n\treturn r.Error()\n}", "func (s *SubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *OneLike) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\tdecodeOneLike(&r, v)\n\treturn r.Error()\n}", "func (self *StringTypeDef) UnmarshalJSON(b []byte) error {\n\tvar m rawStringTypeDef\n\terr := json.Unmarshal(b, &m)\n\tif err == nil {\n\t\to := StringTypeDef(m)\n\t\t*self = o\n\t\terr = self.Validate()\n\t}\n\treturn err\n}", "func (s *SKU) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (b *BgpServiceCommunity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", b, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &b.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &b.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &b.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &b.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &b.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &b.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", b, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *CBPerson) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonE242b40eDecodeGithubComExampleSample2(&r, v)\n\treturn r.Error()\n}", "func (p *Pet) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &p.Name)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *EventType) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"EventType should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = EventTypeString(s)\n\treturn err\n}", "func (j *Job) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &j.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &j.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &j.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &j.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &j.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *RegisterRespPacket) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (j *GetMessagesResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (j *AckMessagesResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}" ]
[ "0.72347057", "0.6651387", "0.66214377", "0.63795656", "0.6209195", "0.6206142", "0.6127572", "0.6107126", "0.60906345", "0.6036192", "0.60216796", "0.6014077", "0.60137457", "0.59821725", "0.5960711", "0.5931556", "0.5922516", "0.5893776", "0.587173", "0.58690596", "0.58690596", "0.5828894", "0.5748344", "0.57009315", "0.5653378", "0.5646445", "0.5633441", "0.56133306", "0.5583513", "0.55784726", "0.55746245", "0.55449915", "0.54922783", "0.54809254", "0.54489434", "0.5397592", "0.53723043", "0.5359464", "0.5333111", "0.5321447", "0.5320933", "0.531954", "0.52797526", "0.52770305", "0.52637494", "0.5259714", "0.52574944", "0.5236661", "0.5235879", "0.52255344", "0.5215106", "0.52137464", "0.52135533", "0.5208173", "0.51923776", "0.5183894", "0.51831836", "0.5174228", "0.5143651", "0.5139057", "0.5137242", "0.5122863", "0.51196647", "0.5108563", "0.5105925", "0.5102648", "0.50949633", "0.50938046", "0.5082007", "0.50745755", "0.5064276", "0.50540435", "0.5053724", "0.5039768", "0.5031846", "0.50313", "0.50243294", "0.50136036", "0.50131434", "0.5012557", "0.5010253", "0.5008396", "0.5006248", "0.50021994", "0.50000674", "0.4994918", "0.49919745", "0.4991964", "0.49882823", "0.49843284", "0.49840826", "0.4981377", "0.49771747", "0.49745148", "0.49739066", "0.49731973", "0.4971481", "0.49692142", "0.49686635", "0.49632445" ]
0.7498778
0
UnmarshalJSON implements the json.Unmarshal interface and unmarshals a JSON object in the Subject struct or a string as just the subject common name.
func (s *Subject) UnmarshalJSON(data []byte) error { var name Name if err := name.UnmarshalJSON(data); err != nil { return err } *s = Subject(name) return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (v *DocumentSubjectUpdate) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonDe066f3DecodeGithubComStudtoolDocumentsServiceModels3(&r, v)\n\treturn r.Error()\n}", "func (n *Name) UnmarshalJSON(data []byte) error {\n\tif cn, ok := maybeString(data); ok {\n\t\tn.CommonName = cn\n\t\treturn nil\n\t}\n\n\ttype nameAlias Name\n\tvar nn nameAlias\n\tif err := json.Unmarshal(data, &nn); err != nil {\n\t\treturn errors.Wrap(err, \"error unmarshaling json\")\n\t}\n\t*n = Name(nn)\n\treturn nil\n}", "func (v *Teacher) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests(&r, v)\n\treturn r.Error()\n}", "func (v *SyntheticsTestRequestBodyType) UnmarshalJSON(src []byte) error {\n\tvar value string\n\terr := json.Unmarshal(src, &value)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*v = SyntheticsTestRequestBodyType(value)\n\treturn nil\n}", "func (c *ConsortiumMember) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"dateModified\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.DateModified)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, &c.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"joinDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.JoinDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"role\":\n\t\t\terr = unpopulate(val, &c.Role)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, &c.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, &c.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (v *DocumentSubjectUpdate) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjsonDe066f3DecodeGithubComStudtoolDocumentsServiceModels3(l, v)\n}", "func (i *Issuer) UnmarshalJSON(data []byte) error {\n\tvar name Name\n\tif err := name.UnmarshalJSON(data); err != nil {\n\t\treturn err\n\t}\n\t*i = Issuer(name)\n\treturn nil\n}", "func (m *SubMessage) UnmarshalJSONObject(dec *gojay.Decoder, k string) error {\n\n\tswitch k {\n\tcase \"Id\":\n\t\treturn dec.Int(&m.Id)\n\n\tcase \"Description\":\n\t\treturn dec.String(&m.Description)\n\n\tcase \"StartTime\":\n\t\tvar format = time.RFC3339\n\t\tvar value = time.Time{}\n\t\terr := dec.Time(&value, format)\n\t\tif err == nil {\n\t\t\tm.StartTime = value\n\t\t}\n\t\treturn err\n\n\tcase \"EndTime\":\n\t\tvar format = time.RFC3339\n\t\tvar value = &time.Time{}\n\t\terr := dec.Time(value, format)\n\t\tif err == nil {\n\t\t\tm.EndTime = value\n\t\t}\n\t\treturn err\n\n\t}\n\treturn nil\n}", "func UnmarshalJSON(ct string, data []byte) (proto.Message, error) {\n\n\tctn, p, err := mime.ParseMediaType(ct)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn UnmarshalJSONParams(ctn, p, data)\n\n}", "func ScheduleUnmarshalJSON(b []byte) (schedule Schedule, err error) {\n\tvar mixed interface{}\n\tjson.Unmarshal(b, &mixed)\n\n\tfor key, value := range mixed.(map[string]interface{}) {\n\t\trawValue, _ := json.Marshal(value)\n\t\tswitch key {\n\t\tcase \"date\":\n\t\t\tvar date Date\n\t\t\terr = json.Unmarshal(rawValue, &date)\n\t\t\tschedule = date\n\t\tcase \"day\":\n\t\t\tvar day Day\n\t\t\terr = json.Unmarshal(rawValue, &day)\n\t\t\tschedule = day\n\t\tcase \"intersection\":\n\t\t\tvar intersection Intersection\n\t\t\terr = json.Unmarshal(rawValue, &intersection)\n\t\t\tschedule = intersection\n\t\tcase \"month\":\n\t\t\tvar month Month\n\t\t\terr = json.Unmarshal(rawValue, &month)\n\t\t\tschedule = month\n\t\tcase \"union\":\n\t\t\tvar union Union\n\t\t\terr = json.Unmarshal(rawValue, &union)\n\t\t\tschedule = union\n\t\tcase \"week\":\n\t\t\tvar week Week\n\t\t\terr = json.Unmarshal(rawValue, &week)\n\t\t\tschedule = week\n\t\tcase \"weekday\":\n\t\t\tvar weekday Weekday\n\t\t\terr = json.Unmarshal(rawValue, &weekday)\n\t\t\tschedule = weekday\n\t\tcase \"year\":\n\t\t\tvar year Year\n\t\t\terr = json.Unmarshal(rawValue, &year)\n\t\t\tschedule = year\n\t\tdefault:\n\t\t\terr = fmt.Errorf(\"%s is not a recognized schedule\", key)\n\t\t}\n\t}\n\treturn\n}", "func (s *Subscription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"authorizationSource\":\n\t\t\terr = unpopulate(val, \"AuthorizationSource\", &s.AuthorizationSource)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, \"DisplayName\", &s.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"state\":\n\t\t\terr = unpopulate(val, \"State\", &s.State)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, \"SubscriptionID\", &s.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionPolicies\":\n\t\t\terr = unpopulate(val, \"SubscriptionPolicies\", &s.SubscriptionPolicies)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *ACName) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn err\n\t}\n\tnn, err := NewACName(s)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*n = *nn\n\treturn nil\n}", "func (i *ChannelName) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"ChannelName should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = ChannelNameString(s)\n\treturn err\n}", "func (t *Topic) UnmarshalJSON(b []byte) error {\n\ts, err := unmarshalJSON(b)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tval, ok := topics[s]\n\tif !ok {\n\t\treturn ErrUnknownTopic\n\t}\n\t*t = val\n\treturn nil\n}", "func (c *CopyAudio) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"label\":\n\t\t\terr = unpopulate(val, \"Label\", &c.Label)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &c.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SingleServerCustomResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &s.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *Message) UnmarshalJSONObject(dec *gojay.Decoder, k string) error {\n\n\tswitch k {\n\tcase \"Id\":\n\t\treturn dec.Int(&m.Id)\n\n\tcase \"Name\":\n\t\treturn dec.String(&m.Name)\n\n\tcase \"Price\":\n\t\treturn dec.Float64(&m.Price)\n\n\tcase \"Ints\":\n\t\tvar aSlice = Ints{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.Ints = []int(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"Floats\":\n\t\tvar aSlice = Float32s{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.Floats = []float32(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"SubMessageX\":\n\t\tvar value = &SubMessage{}\n\t\terr := dec.Object(value)\n\t\tif err == nil {\n\t\t\tm.SubMessageX = value\n\t\t}\n\n\t\treturn err\n\n\tcase \"MessagesX\":\n\t\tvar aSlice = SubMessagesPtr{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.MessagesX = []*SubMessage(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"SubMessageY\":\n\t\terr := dec.Object(&m.SubMessageY)\n\n\t\treturn err\n\n\tcase \"MessagesY\":\n\t\tvar aSlice = SubMessages{}\n\t\terr := dec.Array(&aSlice)\n\t\tif err == nil && len(aSlice) > 0 {\n\t\t\tm.MessagesY = []SubMessage(aSlice)\n\t\t}\n\t\treturn err\n\n\tcase \"IsTrue\":\n\t\tvar value bool\n\t\terr := dec.Bool(&value)\n\t\tif err == nil {\n\t\t\tm.IsTrue = &value\n\t\t}\n\t\treturn err\n\n\tcase \"Payload\":\n\t\tvar value = gojay.EmbeddedJSON{}\n\t\terr := dec.AddEmbeddedJSON(&value)\n\t\tif err == nil && len(value) > 0 {\n\t\t\tm.Payload = []byte(value)\n\t\t}\n\t\treturn err\n\n\tcase \"SQLNullString\":\n\t\tvar value = &sql.NullString{}\n\t\terr := dec.SQLNullString(value)\n\t\tif err == nil {\n\t\t\tm.SQLNullString = value\n\t\t}\n\t\treturn err\n\n\t}\n\treturn nil\n}", "func (m *Message) UnmarshalJSON(j []byte) error {\n\tvar rawStrings map[string]string\n\n\terr := json.Unmarshal(j, &rawStrings)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor k, v := range rawStrings {\n\t\tif strings.ToLower(k) == \"to\" {\n\t\t\tm.To = v\n\t\t}\n\t\tif strings.ToLower(k) == \"from\" {\n\t\t\tm.From = v\n\t\t}\n\t\tif strings.ToLower(k) == \"title\" {\n\t\t\tm.Title = v\n\t\t}\n\t\tif strings.ToLower(k) == \"content\" {\n\t\t\tm.Content = v\n\t\t}\n\t\t// properly parse Date field\n\t\tif strings.ToLower(k) == \"date\" {\n\t\t\tt, err := time.Parse(\"Jan 2, 2006 at 3:04pm (MST)\", v)\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tm.Date = t\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Student) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests1(&r, v)\n\treturn r.Error()\n}", "func (s *SingleServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &s.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachine\":\n\t\t\terr = unpopulate(val, \"VirtualMachine\", &s.VirtualMachine)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (this *NamespacedName) UnmarshalJSON(b []byte) error {\n\treturn CommonUnmarshaler.Unmarshal(bytes.NewReader(b), this)\n}", "func (a *ActivityReportTitleUpdated) UnmarshalJSON(b []byte) error {\n\tvar helper activityReportTitleUpdatedUnmarshalHelper\n\tif err := json.Unmarshal(b, &helper); err != nil {\n\t\treturn err\n\t}\n\t*a = ActivityReportTitleUpdated(helper.Attributes)\n\treturn nil\n}", "func (j *PublishMessagesResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (i *MonthlySchedule) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"MonthlySchedule should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = MonthlyScheduleString(s)\n\treturn err\n}", "func (s *Subscription) UnmarshalJSON(data []byte) error {\n\tif id, ok := ParseID(data); ok {\n\t\ts.ID = id\n\t\treturn nil\n\t}\n\n\ttype subscription Subscription\n\tvar v subscription\n\tif err := json.Unmarshal(data, &v); err != nil {\n\t\treturn err\n\t}\n\n\t*s = Subscription(v)\n\treturn nil\n}", "func (mcc *MsgCreateCertificate) UnmarshalJSON(data []byte) error {\n var jsonMsg struct {\n Certificate *utils.JSONWrapper `json:\"certificate\"`\n }\n if err := json.Unmarshal(data, &jsonMsg); err != nil {\n return err\n }\n if certificateFactory, ok := certificateFactories[jsonMsg.Certificate.Type]; ok {\n certificate := certificateFactory()\n if err := json.Unmarshal(jsonMsg.Certificate.Value, certificate); err != nil {\n return err\n }\n mcc.Certificate = certificate\n } else {\n return errors.New(fmt.Sprintf(\"unknown certificate type: %s\", jsonMsg.Certificate.Type))\n }\n return nil\n}", "func (j *Publisher) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func subjectFromJWT(c *gin.Context) string {\n\tauthHeader := c.Request.Header.Get(\"Authorization\")\n\tprefix := \"Bearer \"\n\tif !strings.HasPrefix(authHeader, prefix) {\n\t\t// Incorrect Authorization header format.\n\t\treturn \"\"\n\t}\n\ttoken := authHeader[strings.Index(authHeader, prefix)+len(prefix):]\n\tif token == \"\" {\n\t\t// JWT not found.\n\t\treturn \"\"\n\t}\n\n\tvar payload jwt.Payload\n\t_, err := jwt.Verify([]byte(token), jwtKey, &payload)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn payload.Subject\n}", "func (n *Name) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &n.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &n.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", n, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Topic) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer5(&r, v)\n\treturn r.Error()\n}", "func (r *ResourceName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Teacher) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests(l, v)\n}", "func (c *ContactDetail) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"email\":\n\t\t\terr = unpopulate(val, \"Email\", &c.Email)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"phone\":\n\t\t\terr = unpopulate(val, \"Phone\", &c.Phone)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"role\":\n\t\t\terr = unpopulate(val, \"Role\", &c.Role)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RegisteredAsn) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &r.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UID) UnmarshalJSON(data []byte) error {\n\tu.Str = string(data[1 : len(data)-1])\n\treturn nil\n}", "func (j *AckMessagesResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (c *CommonJobDetails) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"actions\":\n\t\t\terr = unpopulate(val, \"Actions\", &c.Actions)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"chainOfCustodySasKey\":\n\t\t\terr = unpopulate(val, \"ChainOfCustodySasKey\", &c.ChainOfCustodySasKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"contactDetails\":\n\t\t\terr = unpopulate(val, \"ContactDetails\", &c.ContactDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"copyLogDetails\":\n\t\t\tc.CopyLogDetails, err = unmarshalCopyLogDetailsClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataCenterCode\":\n\t\t\terr = unpopulate(val, \"DataCenterCode\", &c.DataCenterCode)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataExportDetails\":\n\t\t\terr = unpopulate(val, \"DataExportDetails\", &c.DataExportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataImportDetails\":\n\t\t\terr = unpopulate(val, \"DataImportDetails\", &c.DataImportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"datacenterAddress\":\n\t\t\tc.DatacenterAddress, err = unmarshalDatacenterAddressResponseClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deliveryPackage\":\n\t\t\terr = unpopulate(val, \"DeliveryPackage\", &c.DeliveryPackage)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deviceErasureDetails\":\n\t\t\terr = unpopulate(val, \"DeviceErasureDetails\", &c.DeviceErasureDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expectedDataSizeInTeraBytes\":\n\t\t\terr = unpopulate(val, \"ExpectedDataSizeInTeraBytes\", &c.ExpectedDataSizeInTeraBytes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"jobDetailsType\":\n\t\t\terr = unpopulate(val, \"JobDetailsType\", &c.JobDetailsType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"jobStages\":\n\t\t\terr = unpopulate(val, \"JobStages\", &c.JobStages)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"keyEncryptionKey\":\n\t\t\terr = unpopulate(val, \"KeyEncryptionKey\", &c.KeyEncryptionKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastMitigationActionOnJob\":\n\t\t\terr = unpopulate(val, \"LastMitigationActionOnJob\", &c.LastMitigationActionOnJob)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"preferences\":\n\t\t\terr = unpopulate(val, \"Preferences\", &c.Preferences)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"returnPackage\":\n\t\t\terr = unpopulate(val, \"ReturnPackage\", &c.ReturnPackage)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reverseShipmentLabelSasKey\":\n\t\t\terr = unpopulate(val, \"ReverseShipmentLabelSasKey\", &c.ReverseShipmentLabelSasKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"shippingAddress\":\n\t\t\terr = unpopulate(val, \"ShippingAddress\", &c.ShippingAddress)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *ThreeTierFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"applicationServer\":\n\t\t\terr = unpopulate(val, \"ApplicationServer\", &t.ApplicationServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"centralServer\":\n\t\t\terr = unpopulate(val, \"CentralServer\", &t.CentralServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"databaseServer\":\n\t\t\terr = unpopulate(val, \"DatabaseServer\", &t.DatabaseServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &t.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sharedStorage\":\n\t\t\terr = unpopulate(val, \"SharedStorage\", &t.SharedStorage)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SubResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *schedules) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeGithubComIskiyRabotauaTelegramBotPkgRabotaua1(&r, v)\n\treturn r.Error()\n}", "func (j *GetMessagesResponse) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (j *Message) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (v *Messages) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer17(&r, v)\n\treturn r.Error()\n}", "func (s *SharedStorageResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"sharedStorageAccountName\":\n\t\t\terr = unpopulate(val, \"SharedStorageAccountName\", &s.SharedStorageAccountName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sharedStorageAccountPrivateEndPointName\":\n\t\t\terr = unpopulate(val, \"SharedStorageAccountPrivateEndPointName\", &s.SharedStorageAccountPrivateEndPointName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *ConnectionItemName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Message) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer18(&r, v)\n\treturn r.Error()\n}", "func (e *Email) UnmarshalJSON(data []byte) error {\n\tvar estr string\n\tif err := json.Unmarshal(data, &estr); err != nil {\n\t\treturn err\n\t}\n\t*e = Email(estr)\n\treturn nil\n}", "func (t *ThreeTierCustomResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &t.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Msg) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeBackendInternalModels6(&r, v)\n\treturn r.Error()\n}", "func (v *hubMessage) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson2802b09fDecodeGithubComPhilippseithSignalr2(&r, v)\n\treturn r.Error()\n}", "func (v *SubscriptionTopic) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson25363b2dDecodeGithubComDarkfoxs96OpenApiV3SdkOkexGoSdkApi40(&r, v)\n\treturn r.Error()\n}", "func (s *StringBeginsWithAdvancedFilter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"key\":\n\t\t\terr = unpopulate(val, \"Key\", &s.Key)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"operatorType\":\n\t\t\terr = unpopulate(val, \"OperatorType\", &s.OperatorType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"values\":\n\t\t\terr = unpopulate(val, \"Values\", &s.Values)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *JobDeliveryInfo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"scheduledDateTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"ScheduledDateTime\", &j.ScheduledDateTime)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *JobDetails) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"actions\":\n\t\t\terr = unpopulate(val, \"Actions\", &j.Actions)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"chainOfCustodySasKey\":\n\t\t\terr = unpopulate(val, \"ChainOfCustodySasKey\", &j.ChainOfCustodySasKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"contactDetails\":\n\t\t\terr = unpopulate(val, \"ContactDetails\", &j.ContactDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"copyLogDetails\":\n\t\t\tj.CopyLogDetails, err = unmarshalCopyLogDetailsClassificationArray(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"copyProgress\":\n\t\t\terr = unpopulate(val, \"CopyProgress\", &j.CopyProgress)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataCenterCode\":\n\t\t\terr = unpopulate(val, \"DataCenterCode\", &j.DataCenterCode)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataExportDetails\":\n\t\t\terr = unpopulate(val, \"DataExportDetails\", &j.DataExportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"dataImportDetails\":\n\t\t\terr = unpopulate(val, \"DataImportDetails\", &j.DataImportDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"datacenterAddress\":\n\t\t\tj.DatacenterAddress, err = unmarshalDatacenterAddressResponseClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deliveryPackage\":\n\t\t\terr = unpopulate(val, \"DeliveryPackage\", &j.DeliveryPackage)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"deviceErasureDetails\":\n\t\t\terr = unpopulate(val, \"DeviceErasureDetails\", &j.DeviceErasureDetails)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"devicePassword\":\n\t\t\terr = unpopulate(val, \"DevicePassword\", &j.DevicePassword)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expectedDataSizeInTeraBytes\":\n\t\t\terr = unpopulate(val, \"ExpectedDataSizeInTeraBytes\", &j.ExpectedDataSizeInTeraBytes)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"jobDetailsType\":\n\t\t\terr = unpopulate(val, \"JobDetailsType\", &j.JobDetailsType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"jobStages\":\n\t\t\terr = unpopulate(val, \"JobStages\", &j.JobStages)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"keyEncryptionKey\":\n\t\t\terr = unpopulate(val, \"KeyEncryptionKey\", &j.KeyEncryptionKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"lastMitigationActionOnJob\":\n\t\t\terr = unpopulate(val, \"LastMitigationActionOnJob\", &j.LastMitigationActionOnJob)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"preferences\":\n\t\t\terr = unpopulate(val, \"Preferences\", &j.Preferences)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"returnPackage\":\n\t\t\terr = unpopulate(val, \"ReturnPackage\", &j.ReturnPackage)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"reverseShipmentLabelSasKey\":\n\t\t\terr = unpopulate(val, \"ReverseShipmentLabelSasKey\", &j.ReverseShipmentLabelSasKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"shippingAddress\":\n\t\t\terr = unpopulate(val, \"ShippingAddress\", &j.ShippingAddress)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", j, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (h *Hostname) UnmarshalJSON(data []byte) error {\n\tvar hstr string\n\tif err := json.Unmarshal(data, &hstr); err != nil {\n\t\treturn err\n\t}\n\t*h = Hostname(hstr)\n\treturn nil\n}", "func (v *Claims) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson171edd05DecodeGithubComEmirmuminogluJwt1(&r, v)\n\treturn r.Error()\n}", "func (c *CertificateVerificationDescription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"certificate\":\n\t\t\terr = unpopulate(val, \"Certificate\", &c.Certificate)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (ra *RegisteredAsn) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"properties\":\n\t\t\tif v != nil {\n\t\t\t\tvar registeredAsnProperties RegisteredAsnProperties\n\t\t\t\terr = json.Unmarshal(*v, &registeredAsnProperties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.RegisteredAsnProperties = &registeredAsnProperties\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.Name = &name\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.ID = &ID\n\t\t\t}\n\t\tcase \"type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar string\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.Type = &typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (v *Message) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonC5a4559bDecodeGithubComChromedpCdproto1(&r, v)\n\treturn r.Error()\n}", "func (v *schedulesStruct) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecodeGithubComIskiyRabotauaTelegramBotPkgRabotaua(&r, v)\n\treturn r.Error()\n}", "func JSONGetString(ctx context.Context, subject map[string]interface{}, keys ...string) string {\n\tdata := MapInterfaceToJSONBytes(subject)\n\n\tvalue, err := jsonparser.GetString(data, keys...)\n\tif err != nil {\n\t\tlogger.Errorf(ctx, \"JSON util\", \"failed to retrieve key, %#v error:%s\", keys, err.Error())\n\t\treturn \"\"\n\t}\n\n\treturn value\n}", "func (s *SystemAssignedServiceIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &s.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tenantId\":\n\t\t\terr = unpopulate(val, \"TenantID\", &s.TenantID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CampaignAudience) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"expression\":\n\t\t\tif v != nil {\n\t\t\t\tvar expression string\n\t\t\t\terr = json.Unmarshal(*v, &expression)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.Expression = &expression\n\t\t\t}\n\t\tcase \"criteria\":\n\t\t\tif v != nil {\n\t\t\t\tvar criteria map[string]*Criterion\n\t\t\t\terr = json.Unmarshal(*v, &criteria)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.Criteria = criteria\n\t\t\t}\n\t\tcase \"filters\":\n\t\t\tif v != nil {\n\t\t\t\tfilters, err := unmarshalBasicFilterArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.Filters = &filters\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *AssetContainerSas) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"assetContainerSasUrls\":\n\t\t\terr = unpopulate(val, \"AssetContainerSasUrls\", &a.AssetContainerSasUrls)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *DCCSend) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer27(&r, v)\n\treturn r.Error()\n}", "func (s *StringContainsAdvancedFilter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"key\":\n\t\t\terr = unpopulate(val, \"Key\", &s.Key)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"operatorType\":\n\t\t\terr = unpopulate(val, \"OperatorType\", &s.OperatorType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"values\":\n\t\t\terr = unpopulate(val, \"Values\", &s.Values)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (jc *JSONCertificate) UnmarshalJSON(b []byte) error {\n\treturn errors.New(\"Do not unmarshal cert JSON directly, use JSONCertificateWithRaw or x509.ParseCertificate function\")\n}", "func (v *completionMessage) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson2802b09fDecodeGithubComPhilippseithSignalr5(&r, v)\n\treturn r.Error()\n}", "func (m *X509Certificate) UnmarshalJSON(raw []byte) error {\n\tvar data struct {\n\t\tEncoded strfmt.Base64 `json:\"encoded,omitempty\"`\n\n\t\tIssuerDN json.RawMessage `json:\"issuerDN,omitempty\"`\n\n\t\tNotAfter strfmt.DateTime `json:\"notAfter,omitempty\"`\n\n\t\tNotBefore strfmt.DateTime `json:\"notBefore,omitempty\"`\n\n\t\tPublicKey *PublicKey `json:\"publicKey,omitempty\"`\n\n\t\tSerialNumber int64 `json:\"serialNumber,omitempty\"`\n\n\t\tSigAlgName string `json:\"sigAlgName,omitempty\"`\n\n\t\tSigAlgOID string `json:\"sigAlgOID,omitempty\"`\n\n\t\tSigAlgParams strfmt.Base64 `json:\"sigAlgParams,omitempty\"`\n\n\t\tSubjectDN json.RawMessage `json:\"subjectDN,omitempty\"`\n\n\t\tVersion int32 `json:\"version,omitempty\"`\n\t}\n\tbuf := bytes.NewBuffer(raw)\n\tdec := json.NewDecoder(buf)\n\tdec.UseNumber()\n\n\tif err := dec.Decode(&data); err != nil {\n\t\treturn err\n\t}\n\n\tvar propIssuerDN Principal\n\tif string(data.IssuerDN) != \"null\" {\n\t\tissuerDN, err := UnmarshalPrincipal(bytes.NewBuffer(data.IssuerDN), runtime.JSONConsumer())\n\t\tif err != nil && err != io.EOF {\n\t\t\treturn err\n\t\t}\n\t\tpropIssuerDN = issuerDN\n\t}\n\tvar propSubjectDN Principal\n\tif string(data.SubjectDN) != \"null\" {\n\t\tsubjectDN, err := UnmarshalPrincipal(bytes.NewBuffer(data.SubjectDN), runtime.JSONConsumer())\n\t\tif err != nil && err != io.EOF {\n\t\t\treturn err\n\t\t}\n\t\tpropSubjectDN = subjectDN\n\t}\n\n\tvar result X509Certificate\n\n\t// encoded\n\tresult.Encoded = data.Encoded\n\n\t// issuerDN\n\tresult.issuerDNField = propIssuerDN\n\n\t// notAfter\n\tresult.NotAfter = data.NotAfter\n\n\t// notBefore\n\tresult.NotBefore = data.NotBefore\n\n\t// publicKey\n\tresult.PublicKey = data.PublicKey\n\n\t// serialNumber\n\tresult.SerialNumber = data.SerialNumber\n\n\t// sigAlgName\n\tresult.SigAlgName = data.SigAlgName\n\n\t// sigAlgOID\n\tresult.SigAlgOID = data.SigAlgOID\n\n\t// sigAlgParams\n\tresult.SigAlgParams = data.SigAlgParams\n\n\t// subjectDN\n\tresult.subjectDNField = propSubjectDN\n\n\t// version\n\tresult.Version = data.Version\n\n\t*m = result\n\n\treturn nil\n}", "func (s *GitEvent) UnmarshalJSON(b []byte) error {\n\tvar j string\n\terr := json.Unmarshal(b, &j)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Note that if the string cannot be found then it will be set to the zero value, 'Push' in this case.\n\t*s = toID[j]\n\treturn nil\n}", "func (v *Header) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson171edd05DecodeGithubComEmirmuminogluJwt(&r, v)\n\treturn r.Error()\n}", "func (i *MessageType) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"MessageType should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = MessageTypeString(s)\n\treturn err\n}", "func (this *Simple) UnmarshalJSON(b []byte) error {\n\treturn TypesUnmarshaler.Unmarshal(bytes.NewReader(b), this)\n}", "func (c *CertificateBodyDescription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"certificate\":\n\t\t\terr = unpopulate(val, \"Certificate\", &c.Certificate)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *EventMsg) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (a *Assignment) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"properties\":\n\t\t\tif v != nil {\n\t\t\t\tvar assignmentProperties AssignmentProperties\n\t\t\t\terr = json.Unmarshal(*v, &assignmentProperties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\ta.AssignmentProperties = &assignmentProperties\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\ta.ID = &ID\n\t\t\t}\n\t\tcase \"type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar string\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\ta.Type = &typeVar\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\ta.Name = &name\n\t\t\t}\n\t\tcase \"location\":\n\t\t\tif v != nil {\n\t\t\t\tvar location string\n\t\t\t\terr = json.Unmarshal(*v, &location)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\ta.Location = &location\n\t\t\t}\n\t\tcase \"identity\":\n\t\t\tif v != nil {\n\t\t\t\tvar identity Identity\n\t\t\t\terr = json.Unmarshal(*v, &identity)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\ta.Identity = &identity\n\t\t\t}\n\t\tcase \"systemData\":\n\t\t\tif v != nil {\n\t\t\t\tvar systemData SystemData\n\t\t\t\terr = json.Unmarshal(*v, &systemData)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\ta.SystemData = &systemData\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *TimeUnit) UnmarshalJSON(b []byte) error {\n\tvar j string\n\terr := json.Unmarshal(b, &j)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Note that if the string cannot be found then it will be set to the zero value, 'Created' in this case.\n\t*m = toTimeUnitID[j]\n\treturn nil\n}", "func (receiver *Type) UnmarshalJSON(src []byte) error {\n\tif nil == receiver {\n\t\treturn errNilReceiver\n\t}\n\n\tvar s string\n\tif err := json.Unmarshal(src, &s); nil != err {\n\t\treturn err\n\t}\n\n\tif err := receiver.Scan(s); nil != err {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (a *Assignment) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &a.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, \"Identity\", &a.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &a.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &a.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &a.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &a.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *ChatMessage) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecode20191OPGPlus2InternalPkgModels20(&r, v)\n\treturn r.Error()\n}", "func (c *Container) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &c.ID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (j *PublishMessagesRequest) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (v *Student) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests1(l, v)\n}", "func (s *SkypeChannel) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"channelName\":\n\t\t\terr = unpopulate(val, \"ChannelName\", &s.ChannelName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &s.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &s.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"provisioningState\":\n\t\t\terr = unpopulate(val, \"ProvisioningState\", &s.ProvisioningState)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (sr *SearchResponse) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"queryContext\":\n\t\t\tif v != nil {\n\t\t\t\tvar queryContext QueryContext\n\t\t\t\terr = json.Unmarshal(*v, &queryContext)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.QueryContext = &queryContext\n\t\t\t}\n\t\tcase \"entities\":\n\t\t\tif v != nil {\n\t\t\t\tvar entities Entities\n\t\t\t\terr = json.Unmarshal(*v, &entities)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.Entities = &entities\n\t\t\t}\n\t\tcase \"places\":\n\t\t\tif v != nil {\n\t\t\t\tvar places Places\n\t\t\t\terr = json.Unmarshal(*v, &places)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.Places = &places\n\t\t\t}\n\t\tcase \"contractualRules\":\n\t\t\tif v != nil {\n\t\t\t\tcontractualRules, err := unmarshalBasicContractualRulesContractualRuleArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.ContractualRules = &contractualRules\n\t\t\t}\n\t\tcase \"webSearchUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar webSearchURL string\n\t\t\t\terr = json.Unmarshal(*v, &webSearchURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.WebSearchURL = &webSearchURL\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.ID = &ID\n\t\t\t}\n\t\tcase \"_type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar TypeBasicResponseBase\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tsr.Type = typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (s *StringNotBeginsWithAdvancedFilter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"key\":\n\t\t\terr = unpopulate(val, \"Key\", &s.Key)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"operatorType\":\n\t\t\terr = unpopulate(val, \"OperatorType\", &s.OperatorType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"values\":\n\t\t\terr = unpopulate(val, \"Values\", &s.Values)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (t *Thing) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Name = &name\n\t\t\t}\n\t\tcase \"url\":\n\t\t\tif v != nil {\n\t\t\t\tvar URL string\n\t\t\t\terr = json.Unmarshal(*v, &URL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.URL = &URL\n\t\t\t}\n\t\tcase \"image\":\n\t\t\tif v != nil {\n\t\t\t\tvar imageVar ImageObject\n\t\t\t\terr = json.Unmarshal(*v, &imageVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Image = &imageVar\n\t\t\t}\n\t\tcase \"description\":\n\t\t\tif v != nil {\n\t\t\t\tvar description string\n\t\t\t\terr = json.Unmarshal(*v, &description)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Description = &description\n\t\t\t}\n\t\tcase \"entityPresentationInfo\":\n\t\t\tif v != nil {\n\t\t\t\tvar entityPresentationInfo EntitiesEntityPresentationInfo\n\t\t\t\terr = json.Unmarshal(*v, &entityPresentationInfo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.EntityPresentationInfo = &entityPresentationInfo\n\t\t\t}\n\t\tcase \"bingId\":\n\t\t\tif v != nil {\n\t\t\t\tvar bingID string\n\t\t\t\terr = json.Unmarshal(*v, &bingID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.BingID = &bingID\n\t\t\t}\n\t\tcase \"contractualRules\":\n\t\t\tif v != nil {\n\t\t\t\tcontractualRules, err := unmarshalBasicContractualRulesContractualRuleArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.ContractualRules = &contractualRules\n\t\t\t}\n\t\tcase \"webSearchUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar webSearchURL string\n\t\t\t\terr = json.Unmarshal(*v, &webSearchURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.WebSearchURL = &webSearchURL\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.ID = &ID\n\t\t\t}\n\t\tcase \"_type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar TypeBasicResponseBase\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tt.Type = typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (s *StringInAdvancedFilter) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"key\":\n\t\t\terr = unpopulate(val, \"Key\", &s.Key)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"operatorType\":\n\t\t\terr = unpopulate(val, \"OperatorType\", &s.OperatorType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"values\":\n\t\t\terr = unpopulate(val, \"Values\", &s.Values)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CopyVideo) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"label\":\n\t\t\terr = unpopulate(val, \"Label\", &c.Label)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"@odata.type\":\n\t\t\terr = unpopulate(val, \"ODataType\", &c.ODataType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (l *LiveEventTranscription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"inputTrackSelection\":\n\t\t\terr = unpopulate(val, \"InputTrackSelection\", &l.InputTrackSelection)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"language\":\n\t\t\terr = unpopulate(val, \"Language\", &l.Language)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"outputTranscriptionTrack\":\n\t\t\terr = unpopulate(val, \"OutputTranscriptionTrack\", &l.OutputTranscriptionTrack)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CentralServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"availabilitySetName\":\n\t\t\terr = unpopulate(val, \"AvailabilitySetName\", &c.AvailabilitySetName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"loadBalancer\":\n\t\t\terr = unpopulate(val, \"LoadBalancer\", &c.LoadBalancer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachines\":\n\t\t\terr = unpopulate(val, \"VirtualMachines\", &c.VirtualMachines)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SubscriptionIsAllowedToCreateJobValidationRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &s.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *NullableSyntheticsAssertion) UnmarshalJSON(src []byte) error {\n\tv.isSet = true\n\n\t// this object is nullable so check if the payload is null or empty string\n\tif string(src) == \"\" || string(src) == \"{}\" {\n\t\treturn nil\n\t}\n\n\treturn json.Unmarshal(src, &v.value)\n}", "func (n *NameProvider) GetJSONName(subject interface{}, name string) (string, bool) {\n\ttpe := reflect.Indirect(reflect.ValueOf(subject)).Type()\n\treturn n.GetJSONNameForType(tpe, name)\n}", "func (d *Submit) UnmarshalJSON(b []byte) error {\n\tif string(b) == \"null\" {\n\t\treturn nil\n\t}\n\ttype alias Submit\n\tal := struct {\n\t\tDcs byte `json:\"dcs\"`\n\t\tVp *jvp `json:\"vp,omitempty\"`\n\t\tUd *UserData `json:\"ud,omitempty\"`\n\t\t*alias\n\t}{\n\t\talias: (*alias)(d)}\n\tif e := json.Unmarshal(b, &al); e != nil {\n\t\treturn e\n\t}\n\td.DCS = UnmarshalDataCoding(al.Dcs)\n\tif al.Vp != nil {\n\t\td.VP = ValidityPeriodOf(al.Vp.T, al.Vp.S)\n\t}\n\tif al.Ud != nil {\n\t\td.UD = *al.Ud\n\t}\n\treturn nil\n}", "func (t *Transform) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &t.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &t.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &t.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"systemData\":\n\t\t\terr = unpopulate(val, \"SystemData\", &t.SystemData)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &t.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (dst *WorkflowTaskMetadataResponse) UnmarshalJSON(data []byte) error {\n\tvar err error\n\t// use discriminator value to speed up the lookup\n\tvar jsonDict map[string]interface{}\n\terr = newStrictDecoder(data).Decode(&jsonDict)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Failed to unmarshal JSON into map for the discriminator lookup.\")\n\t}\n\n\t// check if the discriminator value is 'mo.AggregateTransform'\n\tif jsonDict[\"ObjectType\"] == \"mo.AggregateTransform\" {\n\t\t// try to unmarshal JSON data into MoAggregateTransform\n\t\terr = json.Unmarshal(data, &dst.MoAggregateTransform)\n\t\tif err == nil {\n\t\t\treturn nil // data stored in dst.MoAggregateTransform, return on the first match\n\t\t} else {\n\t\t\tdst.MoAggregateTransform = nil\n\t\t\treturn fmt.Errorf(\"Failed to unmarshal WorkflowTaskMetadataResponse as MoAggregateTransform: %s\", err.Error())\n\t\t}\n\t}\n\n\t// check if the discriminator value is 'mo.DocumentCount'\n\tif jsonDict[\"ObjectType\"] == \"mo.DocumentCount\" {\n\t\t// try to unmarshal JSON data into MoDocumentCount\n\t\terr = json.Unmarshal(data, &dst.MoDocumentCount)\n\t\tif err == nil {\n\t\t\treturn nil // data stored in dst.MoDocumentCount, return on the first match\n\t\t} else {\n\t\t\tdst.MoDocumentCount = nil\n\t\t\treturn fmt.Errorf(\"Failed to unmarshal WorkflowTaskMetadataResponse as MoDocumentCount: %s\", err.Error())\n\t\t}\n\t}\n\n\t// check if the discriminator value is 'mo.TagSummary'\n\tif jsonDict[\"ObjectType\"] == \"mo.TagSummary\" {\n\t\t// try to unmarshal JSON data into MoTagSummary\n\t\terr = json.Unmarshal(data, &dst.MoTagSummary)\n\t\tif err == nil {\n\t\t\treturn nil // data stored in dst.MoTagSummary, return on the first match\n\t\t} else {\n\t\t\tdst.MoTagSummary = nil\n\t\t\treturn fmt.Errorf(\"Failed to unmarshal WorkflowTaskMetadataResponse as MoTagSummary: %s\", err.Error())\n\t\t}\n\t}\n\n\t// check if the discriminator value is 'workflow.TaskMetadata.List'\n\tif jsonDict[\"ObjectType\"] == \"workflow.TaskMetadata.List\" {\n\t\t// try to unmarshal JSON data into WorkflowTaskMetadataList\n\t\terr = json.Unmarshal(data, &dst.WorkflowTaskMetadataList)\n\t\tif err == nil {\n\t\t\treturn nil // data stored in dst.WorkflowTaskMetadataList, return on the first match\n\t\t} else {\n\t\t\tdst.WorkflowTaskMetadataList = nil\n\t\t\treturn fmt.Errorf(\"Failed to unmarshal WorkflowTaskMetadataResponse as WorkflowTaskMetadataList: %s\", err.Error())\n\t\t}\n\t}\n\n\treturn nil\n}" ]
[ "0.64975536", "0.5930238", "0.5804621", "0.58040327", "0.5792889", "0.57868314", "0.5741335", "0.57283896", "0.5703842", "0.5663885", "0.5605825", "0.5570453", "0.55606997", "0.5558341", "0.5534667", "0.5529278", "0.5501311", "0.5495489", "0.54909563", "0.5471521", "0.5440525", "0.54307836", "0.5427272", "0.5424187", "0.54237", "0.54195243", "0.5404429", "0.53964514", "0.5392672", "0.53513676", "0.5348741", "0.53457874", "0.53354305", "0.5330924", "0.53237194", "0.53127337", "0.53037477", "0.5300732", "0.52978164", "0.5282823", "0.5282823", "0.5282823", "0.52697504", "0.52676886", "0.5266452", "0.52621436", "0.5255873", "0.52501625", "0.523898", "0.52329266", "0.52255905", "0.5219272", "0.5217505", "0.5213358", "0.521236", "0.52116203", "0.521092", "0.5208492", "0.5208243", "0.5205733", "0.52053124", "0.51984435", "0.5190783", "0.51894736", "0.518169", "0.51780164", "0.517707", "0.51761574", "0.5172073", "0.5171085", "0.5166979", "0.5161979", "0.51608855", "0.5160706", "0.5160473", "0.5148935", "0.51427203", "0.51382273", "0.5134657", "0.5129878", "0.51277035", "0.5122738", "0.5120679", "0.51202315", "0.5120139", "0.5120109", "0.51171464", "0.51155984", "0.5115465", "0.51116896", "0.51085246", "0.510811", "0.51052487", "0.5095981", "0.5095594", "0.50954694", "0.5091557", "0.50848734", "0.5081076", "0.5078517" ]
0.80555296
0
Set sets the subject in the given certificate.
func (s Subject) Set(c *x509.Certificate) { c.Subject = pkix.Name{ Country: s.Country, Organization: s.Organization, OrganizationalUnit: s.OrganizationalUnit, Locality: s.Locality, Province: s.Province, StreetAddress: s.StreetAddress, PostalCode: s.PostalCode, SerialNumber: s.SerialNumber, CommonName: s.CommonName, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (m *WindowsInformationProtectionDataRecoveryCertificate) SetSubjectName(value *string)() {\n err := m.GetBackingStore().Set(\"subjectName\", value)\n if err != nil {\n panic(err)\n }\n}", "func SetDetails(cert *x509.Certificate, country, organisation, organisationUnit string) {\n\tcert.Subject = pkix.Name{\n\t\tCountry: []string{country},\n\t\tOrganization: []string{organisation},\n\t\tOrganizationalUnit: []string{organisationUnit},\n\t}\n}", "func (h *Header) SetSubject(s string) {\n\th.SetText(\"Subject\", s)\n}", "func (i Issuer) Set(c *x509.Certificate) {\n\tc.Issuer = pkix.Name{\n\t\tCountry: i.Country,\n\t\tOrganization: i.Organization,\n\t\tOrganizationalUnit: i.OrganizationalUnit,\n\t\tLocality: i.Locality,\n\t\tProvince: i.Province,\n\t\tStreetAddress: i.StreetAddress,\n\t\tPostalCode: i.PostalCode,\n\t\tSerialNumber: i.SerialNumber,\n\t\tCommonName: i.CommonName,\n\t}\n}", "func (m *ScheduleItem) SetSubject(value *string)() {\n err := m.GetBackingStore().Set(\"subject\", value)\n if err != nil {\n panic(err)\n }\n}", "func (info *Info) SetX509Cert(cert *x509.Certificate) {\n\tinfo.Attributes[\"sig.x509.subject\"] = x509tools.FormatSubject(cert)\n\tinfo.Attributes[\"sig.x509.issuer\"] = x509tools.FormatIssuer(cert)\n\td := crypto.SHA1.New()\n\td.Write(cert.Raw)\n\tinfo.Attributes[\"sig.x509.fingerprint\"] = fmt.Sprintf(\"%x\", d.Sum(nil))\n}", "func (o *IdentityCredentialIdentifierOidcProperties) SetSubject(v string) {\n\to.Subject = &v\n}", "func (o *LoginRequest) SetSubject(v string) {\n\to.Subject = &v\n}", "func (m *PrinterCreateOperation) SetCertificate(value *string)() {\n err := m.GetBackingStore().Set(\"certificate\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *PrinterCreateOperation) SetCertificate(value *string)() {\n m.certificate = value\n}", "func (in *ActionMailTemplateTranslationUpdateInput) SetSubject(value string) *ActionMailTemplateTranslationUpdateInput {\n\tin.Subject = value\n\n\tif in._selectedParameters == nil {\n\t\tin._selectedParameters = make(map[string]interface{})\n\t}\n\n\tin._selectedParameters[\"Subject\"] = nil\n\treturn in\n}", "func (k Keeper) setCert(ctx sdk.Context, addr sdk.AccAddress, cert Cert) {\n\tstore := ctx.KVStore(k.storeKey)\n\tbz := k.cdc.MustMarshalBinary(cert)\n\tstore.Set(KeyCert(addr, cert.Property, cert.Certifier), bz)\n}", "func (o *OAuth2ConsentRequest) SetSubject(v string) {\n\to.Subject = &v\n}", "func (t *Tracker) SetSubject(subject *Subject) {\n\tt.Subject = subject\n}", "func (o CertificateOutput) Subject() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Certificate) pulumi.StringOutput { return v.Subject }).(pulumi.StringOutput)\n}", "func (m *Reminder) SetEventSubject(value *string)() {\n err := m.GetBackingStore().Set(\"eventSubject\", value)\n if err != nil {\n panic(err)\n }\n}", "func (db *SQLStore) SetCert(cert *ssh.Certificate) error {\n\treturn db.SetRecord(parseCertificate(cert))\n}", "func (sou *SubjectsOfferedUpdate) SetSubject(s *Subject) *SubjectsOfferedUpdate {\n\treturn sou.SetSubjectID(s.ID)\n}", "func (souo *SubjectsOfferedUpdateOne) SetSubject(s *Subject) *SubjectsOfferedUpdateOne {\n\treturn souo.SetSubjectID(s.ID)\n}", "func ExamplePdfMaroto_SetSubject() {\n\tm := pdf.NewMaroto(consts.Portrait, consts.A4)\n\n\tm.SetSubject(\"subject\", true)\n\n\t// Do more things and save...\n}", "func (o ServiceCertificateOutput) Subject() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceCertificate) *string { return v.Subject }).(pulumi.StringPtrOutput)\n}", "func (m *WindowsInformationProtectionDataRecoveryCertificate) SetCertificate(value []byte)() {\n err := m.GetBackingStore().Set(\"certificate\", value)\n if err != nil {\n panic(err)\n }\n}", "func (c *customerio) Subject(subject string) Mailer {\n\tc.subject = subject\n\treturn c\n}", "func (p *MockPeer) SetEnrollmentCertificate(pem *pem.Block) {\r\n\tp.MockCert = pem\r\n}", "func (m *MailYak) Subject(sub string) {\n\tm.subject = mime.QEncoding.Encode(\"UTF-8\", m.trimRegex.ReplaceAllString(sub, \"\"))\n}", "func (s *Client) SetSSLCertificate(appName string, chain, key []byte) error {\n\tvar body = struct {\n\t\tChain string `json:\"certificate_chain\"`\n\t\tPrivateKey string `json:\"private_key\"`\n\t}{string(chain), string(key)}\n\n\tvar res struct{}\n\treturn s.Post(&res, fmt.Sprintf(\"/apps/%s/sni-endpoints\", appName), body)\n}", "func (s *CertificateSummary) SetSubject(v string) *CertificateSummary {\n\ts.Subject = &v\n\treturn s\n}", "func (b *Builder) SetSubject(subject string) *Builder {\n\tb.Subject = subject\n\treturn b\n}", "func (m *ChatMessage) SetSubject(value *string)() {\n m.subject = value\n}", "func (s *Certificate) SetSubject(v string) *Certificate {\n\ts.Subject = &v\n\treturn s\n}", "func (o *NotificationConfig) SetSubject(v string) {\n\to.Subject = &v\n}", "func (p *MockPeer) SetEnrollmentCertificate(pem *pem.Block) {\n\tp.MockCert = pem\n}", "func (s *SendNotificationActionDefinition) SetSubject(v string) *SendNotificationActionDefinition {\n\ts.Subject = &v\n\treturn s\n}", "func (c *Cache) Set(host string, cert *tls.Certificate) {\n\tc.m.Store(host, cert)\n}", "func (s *AssumeRoleWithSAMLOutput) SetSubject(v string) *AssumeRoleWithSAMLOutput {\n\ts.Subject = &v\n\treturn s\n}", "func (o KeystoresAliasesSelfSignedCertCertsInfoCertInfoOutput) Subject() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v KeystoresAliasesSelfSignedCertCertsInfoCertInfo) *string { return v.Subject }).(pulumi.StringPtrOutput)\n}", "func (void *VoidResponse) SetCertificate(path string) *VoidResponse {\n\tvoid.Request = void.Request.Type(gorequest.TypeMultipart).SendFile(path, \"\", \"certificate\")\n\n\treturn void\n}", "func (h *HTTP) SetTLSCert(certPem, keyPem []byte) error {\n\th.TLS = true\n\th.CertFile, h.KeyFile = \"\", \"\"\n\n\tkey, err := tls.X509KeyPair([]byte(certPem), []byte(keyPem))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif h.Server.TLSConfig == nil {\n\t\th.Server.TLSConfig = &tls.Config{}\n\t}\n\th.Server.TLSConfig.Certificates = []tls.Certificate{key}\n\treturn nil\n}", "func (st *State) SetRsyslogCert(caCert string) error {\n\tvar result params.ErrorResult\n\targs := params.SetRsyslogCertParams{\n\t\tCACert: []byte(caCert),\n\t}\n\terr := st.caller.Call(rsyslogAPI, \"\", \"SetRsyslogCert\", args, &result)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif result.Error != nil {\n\t\treturn result.Error\n\t}\n\treturn nil\n}", "func (e *Domain) SetCert(cert *Cert) {\n\te.Certificate = cert\n}", "func (s *Server) SetCert(cert *tls.Certificate) {\n\ts.cert = cert\n}", "func (b *EmailBuilder) Subject(subject string) *EmailBuilder {\n\tb.email.subject = subject\n\treturn b\n}", "func (s *Server) SetCertFromACME(cert *tls.Certificate, err error) {\n\tif err != nil {\n\t\ts.log.Printf(\"ACME FAILED! %s\", err.Error())\n\t\treturn\n\t}\n\n\ts.cert = cert\n}", "func (m *X509Certificate) SetSubjectDN(val Principal) {\n\tm.subjectDNField = val\n}", "func ControlSubject(verb Verb, name, id string) (string, error) {\n\tverbStr := verb.String()\n\tif verbStr == \"\" {\n\t\treturn \"\", fmt.Errorf(\"%w: %q\", ErrVerbNotSupported, verbStr)\n\t}\n\tif name == \"\" && id != \"\" {\n\t\treturn \"\", ErrServiceNameRequired\n\t}\n\tif name == \"\" && id == \"\" {\n\t\treturn fmt.Sprintf(\"%s.%s\", APIPrefix, verbStr), nil\n\t}\n\tif id == \"\" {\n\t\treturn fmt.Sprintf(\"%s.%s.%s\", APIPrefix, verbStr, name), nil\n\t}\n\treturn fmt.Sprintf(\"%s.%s.%s.%s\", APIPrefix, verbStr, name, id), nil\n}", "func (o *FeatureRelationship) SetSubject(exec boil.Executor, insert bool, related *Feature) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(exec); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"feature_relationship\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"subject_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, featureRelationshipPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.FeatureID, o.FeatureRelationshipID}\n\n\tif boil.DebugMode {\n\t\tfmt.Fprintln(boil.DebugWriter, updateQuery)\n\t\tfmt.Fprintln(boil.DebugWriter, values)\n\t}\n\n\tif _, err = exec.Exec(updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.SubjectID = related.FeatureID\n\n\tif o.R == nil {\n\t\to.R = &featureRelationshipR{\n\t\t\tSubject: related,\n\t\t}\n\t} else {\n\t\to.R.Subject = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &featureR{\n\t\t\tSubjectFeatureRelationship: o,\n\t\t}\n\t} else {\n\t\trelated.R.SubjectFeatureRelationship = o\n\t}\n\n\treturn nil\n}", "func (o *ScheduleSubject) SetScheduleSubject(ctx context.Context, exec boil.ContextExecutor, insert bool, related *ScheduleContent) error {\n\tvar err error\n\tif insert {\n\t\tif err = related.Insert(ctx, exec, boil.Infer()); err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to insert into foreign table\")\n\t\t}\n\t}\n\n\tupdateQuery := fmt.Sprintf(\n\t\t\"UPDATE \\\"schedule_subject\\\" SET %s WHERE %s\",\n\t\tstrmangle.SetParamNames(\"\\\"\", \"\\\"\", 1, []string{\"schedule__subject_id\"}),\n\t\tstrmangle.WhereClause(\"\\\"\", \"\\\"\", 2, scheduleSubjectPrimaryKeyColumns),\n\t)\n\tvalues := []interface{}{related.ID, o.ID}\n\n\tif boil.IsDebug(ctx) {\n\t\twriter := boil.DebugWriterFrom(ctx)\n\t\tfmt.Fprintln(writer, updateQuery)\n\t\tfmt.Fprintln(writer, values)\n\t}\n\tif _, err = exec.ExecContext(ctx, updateQuery, values...); err != nil {\n\t\treturn errors.Wrap(err, \"failed to update local table\")\n\t}\n\n\to.ScheduleSubjectID = related.ID\n\tif o.R == nil {\n\t\to.R = &scheduleSubjectR{\n\t\t\tScheduleSubject: related,\n\t\t}\n\t} else {\n\t\to.R.ScheduleSubject = related\n\t}\n\n\tif related.R == nil {\n\t\trelated.R = &scheduleContentR{\n\t\t\tScheduleSubjectScheduleSubjects: ScheduleSubjectSlice{o},\n\t\t}\n\t} else {\n\t\trelated.R.ScheduleSubjectScheduleSubjects = append(related.R.ScheduleSubjectScheduleSubjects, o)\n\t}\n\n\treturn nil\n}", "func (k Keeper) SetCertificates(ctx sdk.Context, certificates types.Certificates) {\n\tstore := prefix.NewStore(ctx.KVStore(k.storeKey), types.KeyPrefix(types.CertificatesKey))\n\tb := k.cdc.MustMarshalBinaryBare(&certificates)\n\tstore.Set(types.KeyPrefix(certificates.Identifier.Index()), b)\n}", "func (r ApiRevokeOAuth2LoginSessionsRequest) Subject(subject string) ApiRevokeOAuth2LoginSessionsRequest {\n\tr.subject = &subject\n\treturn r\n}", "func (me *TAttlistKeywordMajorTopicYN) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (m *TokenManager) Set(ctx context.Context, subject, value string) error {\n\t_, err := sqlContext.GetQueryer(ctx, m.db).ExecContext(ctx, persistToken, subject, value)\n\treturn err\n}", "func (r ApiRevokeOAuth2ConsentSessionsRequest) Subject(subject string) ApiRevokeOAuth2ConsentSessionsRequest {\n\tr.subject = &subject\n\treturn r\n}", "func (r ApiListOAuth2ConsentSessionsRequest) Subject(subject string) ApiListOAuth2ConsentSessionsRequest {\n\tr.subject = &subject\n\treturn r\n}", "func (client *KeyVaultClient) setCertificateIssuerCreateRequest(ctx context.Context, vaultBaseURL string, issuerName string, parameter CertificateIssuerSetParameters, options *KeyVaultClientSetCertificateIssuerOptions) (*policy.Request, error) {\n\thost := \"{vaultBaseUrl}\"\n\thost = strings.ReplaceAll(host, \"{vaultBaseUrl}\", vaultBaseURL)\n\turlPath := \"/certificates/issuers/{issuer-name}\"\n\tif issuerName == \"\" {\n\t\treturn nil, errors.New(\"parameter issuerName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{issuer-name}\", url.PathEscape(issuerName))\n\treq, err := runtime.NewRequest(ctx, http.MethodPut, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"7.2\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, runtime.MarshalAsJSON(req, parameter)\n}", "func (hi *HandshakeInfo) SetIdentityCertProvider(identity certprovider.Provider) {\n\thi.mu.Lock()\n\thi.identityProvider = identity\n\thi.mu.Unlock()\n}", "func (m *InternalDomainFederation) SetNextSigningCertificate(value *string)() {\n err := m.GetBackingStore().Set(\"nextSigningCertificate\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o EnvironmentCertificateOutput) SubjectName() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *EnvironmentCertificate) pulumi.StringOutput { return v.SubjectName }).(pulumi.StringOutput)\n}", "func (sou *SubjectsOfferedUpdate) SetSubjectID(id int) *SubjectsOfferedUpdate {\n\tsou.mutation.SetSubjectID(id)\n\treturn sou\n}", "func (o ServiceHostnameConfigurationManagementOutput) Subject() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceHostnameConfigurationManagement) *string { return v.Subject }).(pulumi.StringPtrOutput)\n}", "func SetClientCert(newCert []byte, userName string) error {\n\tSQLQuery := `\n\t\tUPDATE users\n\t\tSET client_cert = $1\n\t\tWHERE lower(user_name) = lower($2)`\n\tcommandTag, err := pdb.Exec(SQLQuery, newCert, userName)\n\tif err != nil {\n\t\tlog.Printf(\"Updating client certificate for '%s' failed: %v\\n\", userName, err)\n\t\treturn err\n\t}\n\tif numRows := commandTag.RowsAffected(); numRows != 1 {\n\t\terrMsg := fmt.Sprintf(\"Wrong number of rows affected (%v) when storing client cert for '%s'\\n\",\n\t\t\tnumRows, userName)\n\t\tlog.Printf(errMsg)\n\t\treturn errors.New(errMsg)\n\t}\n\treturn nil\n}", "func (me *TAttlistDescriptorNameMajorTopicYN) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func WithSubject(sub string) ParserOption {\n\treturn func(p *Parser) {\n\t\tp.validator.expectedSub = sub\n\t}\n}", "func (m *MemoryTokenManager) Set(ctx context.Context, subject, token string) error {\n\tm.cache.Store(subject, token)\n\treturn nil\n}", "func (c *BaseMail) SetSubject(subject string) *BaseMail {\n\tc.Subject = subject\n\treturn c\n}", "func (souo *SubjectsOfferedUpdateOne) SetSubjectID(id int) *SubjectsOfferedUpdateOne {\n\tsouo.mutation.SetSubjectID(id)\n\treturn souo\n}", "func (me *TSAFPTCNCode) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (m *Application) SetCertification(value Certificationable)() {\n m.certification = value\n}", "func (o ServiceHostnameConfigurationProxyOutput) Subject() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceHostnameConfigurationProxy) *string { return v.Subject }).(pulumi.StringPtrOutput)\n}", "func setCaKeyCert(client *api.Client, caConfigPath string, keyCert string) (*api.Secret, error) {\n\tm := map[string]interface{}{\n\t\t\"pem_bundle\": keyCert,\n\t}\n\n\tres, err := client.Logical().Write(caConfigPath, m)\n\tif err != nil {\n\t\tlog.Errorf(\"Write() failed (error %v)\", err)\n\t\treturn nil, err\n\t}\n\treturn res, nil\n}", "func (o FederatedIdentityCredentialOutput) Subject() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *FederatedIdentityCredential) pulumi.StringOutput { return v.Subject }).(pulumi.StringOutput)\n}", "func (o ServiceHostnameConfigurationScmOutput) Subject() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceHostnameConfigurationScm) *string { return v.Subject }).(pulumi.StringPtrOutput)\n}", "func (a *acmStore) Store(cert *certificate.Resource, domains []string) error {\n\tif cert == nil || cert.Certificate == nil {\n\t\treturn ErrCertificateMissing\n\t}\n\n\tdomainsListString := strings.Join(domains, \", \")\n\n\ta.log.Infof(\"[%s] acm: Retrieving server certificate\", domainsListString)\n\n\tserverCert, err := retrieveServerCertificate(cert.Certificate)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"acm: unable to retrieve server certificate\")\n\t}\n\n\ta.log.Infof(\"[%s] acm: Finding existing server certificate in ACM\", domainsListString)\n\n\texistingCert, err := a.findExistingCertificate(domains)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"acm: unable to find existing certificate\")\n\t}\n\n\t// Retrieve exising certificate ID\n\tvar certArn *string\n\tif existingCert != nil {\n\t\tcertArn = existingCert.CertificateArn\n\t}\n\n\tif certArn != nil {\n\t\ta.log.Infof(\"[%s] acm: Found existing server certificate in ACM with Arn = '%s'\", domainsListString, aws.StringValue(certArn))\n\t}\n\n\t// Init request parameters\n\tinput := &acm.ImportCertificateInput{\n\t\tCertificate: serverCert,\n\t\tCertificateArn: certArn,\n\t\tCertificateChain: cert.IssuerCertificate,\n\t\tPrivateKey: cert.PrivateKey,\n\t}\n\n\tresp, err := a.acm.ImportCertificate(input)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"acm: unable to store certificate into ACM\")\n\t}\n\n\ta.log.Infof(\"[%s] acm: Imported certificate data in ACM with Arn = '%s'\", domainsListString, aws.StringValue(resp.CertificateArn))\n\n\treturn nil\n}", "func (me *TAttlistIssnIssnType) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (me *TxsdCompanyID) Set(s string) { (*xsdt.String)(me).Set(s) }", "func SetupCertificate(mgr ctrl.Manager, l logging.Logger, rl workqueue.RateLimiter, poll time.Duration) error {\n\tname := managed.ControllerName(v1alpha1.CertificateGroupKind)\n\n\treturn ctrl.NewControllerManagedBy(mgr).\n\t\tNamed(name).\n\t\tWithOptions(controller.Options{\n\t\t\tRateLimiter: ratelimiter.NewController(rl),\n\t\t}).\n\t\tFor(&v1alpha1.Certificate{}).\n\t\tComplete(managed.NewReconciler(mgr,\n\t\t\tresource.ManagedKind(v1alpha1.CertificateGroupVersionKind),\n\t\t\tmanaged.WithExternalConnecter(&connector{client: mgr.GetClient(), newClientFn: acm.NewClient}),\n\t\t\tmanaged.WithConnectionPublishers(),\n\t\t\tmanaged.WithPollInterval(poll),\n\t\t\tmanaged.WithReferenceResolver(managed.NewAPISimpleReferenceResolver(mgr.GetClient())),\n\t\t\tmanaged.WithInitializers(),\n\t\t\tmanaged.WithLogger(l.WithValues(\"controller\", name)),\n\t\t\tmanaged.WithRecorder(event.NewAPIRecorder(mgr.GetEventRecorderFor(name)))))\n}", "func (o ServiceHostnameConfigurationDeveloperPortalOutput) Subject() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceHostnameConfigurationDeveloperPortal) *string { return v.Subject }).(pulumi.StringPtrOutput)\n}", "func (suo *SubjectUpdateOne) SetSubjects(s string) *SubjectUpdateOne {\n\tsuo.mutation.SetSubjects(s)\n\treturn suo\n}", "func (r *Request) AssertSubject(t *testing.T, subject string) *Request {\n\tif r.Subject != subject {\n\t\tt.Fatalf(\"expected subject to be %#v, but got %#v\", subject, r.Subject)\n\t}\n\treturn r\n}", "func (c *certManager) EnsureCertificate(domain string, certificateSubject string) (*store.Certificate, error) {\n\tvar sans []string\n\tvar err error\n\n\tif certificateSubject == \"\" {\n\t\tcertificateSubject, err = publicsuffix.EffectiveTLDPlusOne(domain)\n\t\tif err != nil {\n\t\t\treturn nil, logger.Errorex(\"can't get public suffix for domain\", err, golog.String(\"domain\", domain))\n\t\t}\n\t}\n\tif certificateSubject != domain {\n\t\tsans = append(sans, domain)\n\t}\n\tcert, err := c.addCertificate(certificateSubject, sans, false)\n\tif err != nil {\n\t\treturn nil, logger.Errore(err)\n\t}\n\treturn cert, nil\n}", "func (c *Container) SetHostname() {\n\tif c.Hostname == \"\" {\n\t\tc.Hostname = c.Digest[:12]\n\t}\n\tsyscall.Sethostname([]byte(c.Hostname))\n}", "func (o ServiceHostnameConfigurationPortalOutput) Subject() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceHostnameConfigurationPortal) *string { return v.Subject }).(pulumi.StringPtrOutput)\n}", "func (su *SubjectUpdate) SetSubjects(s string) *SubjectUpdate {\n\tsu.mutation.SetSubjects(s)\n\treturn su\n}", "func (m *IosDeviceFeaturesConfiguration) SetSingleSignOnExtensionPkinitCertificate(value IosCertificateProfileBaseable)() {\n err := m.GetBackingStore().Set(\"singleSignOnExtensionPkinitCertificate\", value)\n if err != nil {\n panic(err)\n }\n}", "func (me *TAttlistQualifierNameMajorTopicYN) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (h *Header) Subject() (string, error) {\n\treturn h.Text(\"Subject\")\n}", "func (fi *FeatureInfo) WithSubject(sub string) *FeatureInfo {\n\tfi.Subject = sub\n\treturn fi\n}", "func Subject(subject string) (string, string) {\n\treturn \"Subject\", subject\n}", "func (cm *CertificateManager) SetCertificates(certificates []tls.Certificate) {\n\tc := &tls.Config{\n\t\tCertificates: certificates,\n\t}\n\tc.BuildNameToCertificate()\n\tcm.m.Lock()\n\tcm.certificates = certificates\n\tcm.nameToCertificate = c.NameToCertificate\n\tcm.m.Unlock()\n}", "func (me *TxsdAssessmentOccurrence) Set(s string) { (*xsdt.Nmtoken)(me).Set(s) }", "func (m *RuleBasedSubjectSet) SetRule(value *string)() {\n err := m.GetBackingStore().Set(\"rule\", value)\n if err != nil {\n panic(err)\n }\n}", "func (me *TxsdIncidentPurpose) Set(s string) { (*xsdt.Nmtoken)(me).Set(s) }", "func (me *TSAFPTJournalID) Set(s string) { (*xsdt.String)(me).Set(s) }", "func (me *TAttlistArticlePubModel) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (o ServiceHostnameConfigurationScmOutput) Certificate() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceHostnameConfigurationScm) *string { return v.Certificate }).(pulumi.StringPtrOutput)\n}", "func (me *TxsdAddressSimpleContentExtensionCategory) Set(s string) { (*xsdt.Nmtoken)(me).Set(s) }", "func (c *Certificate) Sign(pkey ed25519.PrivateKey) {\n\t// (<resolver-pk> <client-magic> <serial> <ts-start> <ts-end> <extensions>)\n\tvar b []byte\n\tscratch := make([]byte, 4)\n\n\tb = append(b, c.PublicKey[:]...)\n\tb = append(b, c.ClientMagic...)\n\n\tbinary.BigEndian.PutUint32(scratch, c.Serial)\n\tb = append(b, scratch...)\n\n\tbinary.BigEndian.PutUint32(scratch, uint32(c.Start.Unix()))\n\tb = append(b, scratch...)\n\n\tbinary.BigEndian.PutUint32(scratch, uint32(c.End.Unix()))\n\tb = append(b, scratch...)\n\n\tb = append(b, c.Extensions...)\n\tc.Signature = ed25519.Sign(pkey, b)\n}", "func (e *Executor) SetCertificates(certificateSigning, certificationTransport authentication.Certificate) error {\n\te.SigningCert = certificateSigning\n\te.TransportCert = certificationTransport\n\treturn e.setupTLSCertificate(e.TransportCert.TLSCert())\n}", "func (c *BaseMail) ReplaceSubject(replaceFrom, replaceTo string) *BaseMail {\n\tc.Subject = strings.ReplaceAll(c.Subject, replaceFrom, replaceTo)\n\treturn c\n}", "func (o ServiceHostnameConfigurationManagementOutput) Certificate() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceHostnameConfigurationManagement) *string { return v.Certificate }).(pulumi.StringPtrOutput)\n}", "func (me *TAttlistAuthorValidYN) Set(s string) { (*xsdt.Token)(me).Set(s) }" ]
[ "0.7174654", "0.6621678", "0.656546", "0.6436655", "0.63066554", "0.6229996", "0.6004813", "0.5974351", "0.5948033", "0.59166783", "0.5915121", "0.5868232", "0.58379894", "0.58334905", "0.57591057", "0.5694514", "0.5670662", "0.5630902", "0.56193864", "0.56166315", "0.5581966", "0.55690897", "0.5542101", "0.5538936", "0.55184263", "0.55102575", "0.5472026", "0.54713297", "0.5465775", "0.54535043", "0.54215544", "0.53727514", "0.5313206", "0.5288808", "0.5287283", "0.5285546", "0.524108", "0.52212656", "0.51825166", "0.5166685", "0.5163365", "0.512952", "0.5106149", "0.50780654", "0.50353646", "0.50162023", "0.4981602", "0.49778262", "0.49764925", "0.49716553", "0.49690565", "0.49680907", "0.49645403", "0.4951807", "0.49385458", "0.49231708", "0.49203697", "0.4919171", "0.4917149", "0.490936", "0.4897196", "0.48920852", "0.48751423", "0.48625895", "0.48546353", "0.48535287", "0.48532706", "0.48496366", "0.4830407", "0.48196274", "0.48000905", "0.47971454", "0.47858506", "0.47776842", "0.47410065", "0.47407874", "0.4738075", "0.4735639", "0.4719094", "0.47164235", "0.4716054", "0.4704167", "0.46994522", "0.46827522", "0.46693745", "0.46591058", "0.46470687", "0.46450922", "0.46341243", "0.46300846", "0.46265253", "0.46245837", "0.46059465", "0.46053097", "0.4598919", "0.45767188", "0.45713967", "0.45708382", "0.45693216", "0.45654768" ]
0.8281303
0
UnmarshalJSON implements the json.Unmarshal interface and unmarshals a JSON object in the Issuer struct or a string as just the subject common name.
func (i *Issuer) UnmarshalJSON(data []byte) error { var name Name if err := name.UnmarshalJSON(data); err != nil { return err } *i = Issuer(name) return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Subject) UnmarshalJSON(data []byte) error {\n\tvar name Name\n\tif err := name.UnmarshalJSON(data); err != nil {\n\t\treturn err\n\t}\n\t*s = Subject(name)\n\treturn nil\n}", "func (i *IssuingAuthorization) UnmarshalJSON(data []byte) error {\n\tif id, ok := ParseID(data); ok {\n\t\ti.ID = id\n\t\treturn nil\n\t}\n\n\ttype issuingAuthorization IssuingAuthorization\n\tvar v issuingAuthorization\n\tif err := json.Unmarshal(data, &v); err != nil {\n\t\treturn err\n\t}\n\n\t*i = IssuingAuthorization(v)\n\treturn nil\n}", "func (v *Claims) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson171edd05DecodeGithubComEmirmuminogluJwt1(&r, v)\n\treturn r.Error()\n}", "func (c *ConsortiumMember) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"dateModified\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.DateModified)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, &c.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"joinDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.JoinDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"role\":\n\t\t\terr = unpopulate(val, &c.Role)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"status\":\n\t\t\terr = unpopulate(val, &c.Status)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, &c.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SKUCredential) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"acrServerUrl\":\n\t\t\terr = unpopulate(val, \"AcrServerURL\", &s.AcrServerURL)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"acrToken\":\n\t\t\terr = unpopulate(val, \"AcrToken\", &s.AcrToken)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"expiry\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"Expiry\", &s.Expiry)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"repositories\":\n\t\t\terr = unpopulate(val, \"Repositories\", &s.Repositories)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"username\":\n\t\t\terr = unpopulate(val, \"Username\", &s.Username)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RegisteredAsn) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &r.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &r.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *DocumentSubjectUpdate) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonDe066f3DecodeGithubComStudtoolDocumentsServiceModels3(&r, v)\n\treturn r.Error()\n}", "func (s *Subscription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"authorizationSource\":\n\t\t\terr = unpopulate(val, \"AuthorizationSource\", &s.AuthorizationSource)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"displayName\":\n\t\t\terr = unpopulate(val, \"DisplayName\", &s.DisplayName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"state\":\n\t\t\terr = unpopulate(val, \"State\", &s.State)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionId\":\n\t\t\terr = unpopulate(val, \"SubscriptionID\", &s.SubscriptionID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"subscriptionPolicies\":\n\t\t\terr = unpopulate(val, \"SubscriptionPolicies\", &s.SubscriptionPolicies)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (ra *RegisteredAsn) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"properties\":\n\t\t\tif v != nil {\n\t\t\t\tvar registeredAsnProperties RegisteredAsnProperties\n\t\t\t\terr = json.Unmarshal(*v, &registeredAsnProperties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.RegisteredAsnProperties = &registeredAsnProperties\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.Name = &name\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.ID = &ID\n\t\t\t}\n\t\tcase \"type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar string\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tra.Type = &typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (r *ResourceIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"useSystemAssignedIdentity\":\n\t\t\terr = unpopulate(val, \"UseSystemAssignedIdentity\", &r.UseSystemAssignedIdentity)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"userAssignedIdentity\":\n\t\t\terr = unpopulate(val, \"UserAssignedIdentity\", &r.UserAssignedIdentity)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CertificateVerificationDescription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"certificate\":\n\t\t\terr = unpopulate(val, \"Certificate\", &c.Certificate)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Header) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson171edd05DecodeGithubComEmirmuminogluJwt(&r, v)\n\treturn r.Error()\n}", "func (uc *UnencryptedCredentials) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"jobName\":\n\t\t\tif v != nil {\n\t\t\t\tvar jobName string\n\t\t\t\terr = json.Unmarshal(*v, &jobName)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tuc.JobName = &jobName\n\t\t\t}\n\t\tcase \"jobSecrets\":\n\t\t\tif v != nil {\n\t\t\t\tjobSecrets, err := unmarshalBasicJobSecrets(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tuc.JobSecrets = jobSecrets\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (s *SingleServerCustomResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &s.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (jc *JSONCertificate) UnmarshalJSON(b []byte) error {\n\treturn errors.New(\"Do not unmarshal cert JSON directly, use JSONCertificateWithRaw or x509.ParseCertificate function\")\n}", "func (s *SystemAssignedServiceIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &s.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tenantId\":\n\t\t\terr = unpopulate(val, \"TenantID\", &s.TenantID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicyTransportSecurity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"certificateAuthority\":\n\t\t\terr = unpopulate(val, \"CertificateAuthority\", &f.CertificateAuthority)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *SSN) UnmarshalJSON(data []byte) error {\n\tif string(data) == jsonNull {\n\t\treturn nil\n\t}\n\tvar ustr string\n\tif err := json.Unmarshal(data, &ustr); err != nil {\n\t\treturn err\n\t}\n\t*u = SSN(ustr)\n\treturn nil\n}", "func (u *UserAssignedIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"clientId\":\n\t\t\terr = unpopulate(val, \"ClientID\", &u.ClientID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &u.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UserAssignedIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"clientId\":\n\t\t\terr = unpopulate(val, \"ClientID\", &u.ClientID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &u.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UserAssignedIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"clientId\":\n\t\t\terr = unpopulate(val, \"ClientID\", &u.ClientID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &u.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *ResourceName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &r.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &r.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (m *X509Certificate) UnmarshalJSON(raw []byte) error {\n\tvar data struct {\n\t\tEncoded strfmt.Base64 `json:\"encoded,omitempty\"`\n\n\t\tIssuerDN json.RawMessage `json:\"issuerDN,omitempty\"`\n\n\t\tNotAfter strfmt.DateTime `json:\"notAfter,omitempty\"`\n\n\t\tNotBefore strfmt.DateTime `json:\"notBefore,omitempty\"`\n\n\t\tPublicKey *PublicKey `json:\"publicKey,omitempty\"`\n\n\t\tSerialNumber int64 `json:\"serialNumber,omitempty\"`\n\n\t\tSigAlgName string `json:\"sigAlgName,omitempty\"`\n\n\t\tSigAlgOID string `json:\"sigAlgOID,omitempty\"`\n\n\t\tSigAlgParams strfmt.Base64 `json:\"sigAlgParams,omitempty\"`\n\n\t\tSubjectDN json.RawMessage `json:\"subjectDN,omitempty\"`\n\n\t\tVersion int32 `json:\"version,omitempty\"`\n\t}\n\tbuf := bytes.NewBuffer(raw)\n\tdec := json.NewDecoder(buf)\n\tdec.UseNumber()\n\n\tif err := dec.Decode(&data); err != nil {\n\t\treturn err\n\t}\n\n\tvar propIssuerDN Principal\n\tif string(data.IssuerDN) != \"null\" {\n\t\tissuerDN, err := UnmarshalPrincipal(bytes.NewBuffer(data.IssuerDN), runtime.JSONConsumer())\n\t\tif err != nil && err != io.EOF {\n\t\t\treturn err\n\t\t}\n\t\tpropIssuerDN = issuerDN\n\t}\n\tvar propSubjectDN Principal\n\tif string(data.SubjectDN) != \"null\" {\n\t\tsubjectDN, err := UnmarshalPrincipal(bytes.NewBuffer(data.SubjectDN), runtime.JSONConsumer())\n\t\tif err != nil && err != io.EOF {\n\t\t\treturn err\n\t\t}\n\t\tpropSubjectDN = subjectDN\n\t}\n\n\tvar result X509Certificate\n\n\t// encoded\n\tresult.Encoded = data.Encoded\n\n\t// issuerDN\n\tresult.issuerDNField = propIssuerDN\n\n\t// notAfter\n\tresult.NotAfter = data.NotAfter\n\n\t// notBefore\n\tresult.NotBefore = data.NotBefore\n\n\t// publicKey\n\tresult.PublicKey = data.PublicKey\n\n\t// serialNumber\n\tresult.SerialNumber = data.SerialNumber\n\n\t// sigAlgName\n\tresult.SigAlgName = data.SigAlgName\n\n\t// sigAlgOID\n\tresult.SigAlgOID = data.SigAlgOID\n\n\t// sigAlgParams\n\tresult.SigAlgParams = data.SigAlgParams\n\n\t// subjectDN\n\tresult.subjectDNField = propSubjectDN\n\n\t// version\n\tresult.Version = data.Version\n\n\t*m = result\n\n\treturn nil\n}", "func (spk *SiaPublicKey) UnmarshalJSON(b []byte) error {\n\tspk.LoadString(string(bytes.Trim(b, `\"`)))\n\tif spk.Key == nil {\n\t\t// fallback to old (base64) encoding\n\t\tvar oldSPK struct {\n\t\t\tAlgorithm Specifier\n\t\t\tKey []byte\n\t\t}\n\t\tif err := json.Unmarshal(b, &oldSPK); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tspk.Algorithm, spk.Key = oldSPK.Algorithm, oldSPK.Key\n\t}\n\treturn nil\n}", "func (s *SharedStorageResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"sharedStorageAccountName\":\n\t\t\terr = unpopulate(val, \"SharedStorageAccountName\", &s.SharedStorageAccountName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sharedStorageAccountPrivateEndPointName\":\n\t\t\terr = unpopulate(val, \"SharedStorageAccountPrivateEndPointName\", &s.SharedStorageAccountPrivateEndPointName)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (n *Name) UnmarshalJSON(data []byte) error {\n\tif cn, ok := maybeString(data); ok {\n\t\tn.CommonName = cn\n\t\treturn nil\n\t}\n\n\ttype nameAlias Name\n\tvar nn nameAlias\n\tif err := json.Unmarshal(data, &nn); err != nil {\n\t\treturn errors.Wrap(err, \"error unmarshaling json\")\n\t}\n\t*n = Name(nn)\n\treturn nil\n}", "func (s *SubscriptionIsAllowedToCreateJobValidationRequest) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"validationType\":\n\t\t\terr = unpopulate(val, \"ValidationType\", &s.ValidationType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Claims) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjson171edd05DecodeGithubComEmirmuminogluJwt1(l, v)\n}", "func (c *CentralServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"availabilitySetName\":\n\t\t\terr = unpopulate(val, \"AvailabilitySetName\", &c.AvailabilitySetName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"loadBalancer\":\n\t\t\terr = unpopulate(val, \"LoadBalancer\", &c.LoadBalancer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachines\":\n\t\t\terr = unpopulate(val, \"VirtualMachines\", &c.VirtualMachines)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (d *DeliveryWithResourceIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"destination\":\n\t\t\td.Destination, err = unmarshalEventSubscriptionDestinationClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"identity\":\n\t\t\terr = unpopulate(val, \"Identity\", &d.Identity)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", d, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CampaignAudience) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"expression\":\n\t\t\tif v != nil {\n\t\t\t\tvar expression string\n\t\t\t\terr = json.Unmarshal(*v, &expression)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.Expression = &expression\n\t\t\t}\n\t\tcase \"criteria\":\n\t\t\tif v != nil {\n\t\t\t\tvar criteria map[string]*Criterion\n\t\t\t\terr = json.Unmarshal(*v, &criteria)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.Criteria = criteria\n\t\t\t}\n\t\tcase \"filters\":\n\t\t\tif v != nil {\n\t\t\t\tfilters, err := unmarshalBasicFilterArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tc.Filters = &filters\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (v *Student) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests1(&r, v)\n\treturn r.Error()\n}", "func (v *SyntheticsTestRequestBodyType) UnmarshalJSON(src []byte) error {\n\tvar value string\n\terr := json.Unmarshal(src, &value)\n\tif err != nil {\n\t\treturn err\n\t}\n\t*v = SyntheticsTestRequestBodyType(value)\n\treturn nil\n}", "func (t *TrustedIDProvider) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &t.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &t.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &t.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &t.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SecurityPartnerProvider) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &s.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &s.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &s.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SingleServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &s.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachine\":\n\t\t\terr = unpopulate(val, \"VirtualMachine\", &s.VirtualMachine)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AssetContainerSas) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"assetContainerSasUrls\":\n\t\t\terr = unpopulate(val, \"AssetContainerSasUrls\", &a.AssetContainerSasUrls)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CertificateBodyDescription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"certificate\":\n\t\t\terr = unpopulate(val, \"Certificate\", &c.Certificate)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *InboundSecurityRule) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &i.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &i.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &i.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &i.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &i.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (l *License) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"thumbnailUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar thumbnailURL string\n\t\t\t\terr = json.Unmarshal(*v, &thumbnailURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.ThumbnailURL = &thumbnailURL\n\t\t\t}\n\t\tcase \"provider\":\n\t\t\tif v != nil {\n\t\t\t\tprovider, err := unmarshalBasicThingArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.Provider = &provider\n\t\t\t}\n\t\tcase \"text\":\n\t\t\tif v != nil {\n\t\t\t\tvar textVar string\n\t\t\t\terr = json.Unmarshal(*v, &textVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.Text = &textVar\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.Name = &name\n\t\t\t}\n\t\tcase \"url\":\n\t\t\tif v != nil {\n\t\t\t\tvar URL string\n\t\t\t\terr = json.Unmarshal(*v, &URL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.URL = &URL\n\t\t\t}\n\t\tcase \"image\":\n\t\t\tif v != nil {\n\t\t\t\tvar imageVar ImageObject\n\t\t\t\terr = json.Unmarshal(*v, &imageVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.Image = &imageVar\n\t\t\t}\n\t\tcase \"description\":\n\t\t\tif v != nil {\n\t\t\t\tvar description string\n\t\t\t\terr = json.Unmarshal(*v, &description)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.Description = &description\n\t\t\t}\n\t\tcase \"entityPresentationInfo\":\n\t\t\tif v != nil {\n\t\t\t\tvar entityPresentationInfo EntitiesEntityPresentationInfo\n\t\t\t\terr = json.Unmarshal(*v, &entityPresentationInfo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.EntityPresentationInfo = &entityPresentationInfo\n\t\t\t}\n\t\tcase \"bingId\":\n\t\t\tif v != nil {\n\t\t\t\tvar bingID string\n\t\t\t\terr = json.Unmarshal(*v, &bingID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.BingID = &bingID\n\t\t\t}\n\t\tcase \"contractualRules\":\n\t\t\tif v != nil {\n\t\t\t\tcontractualRules, err := unmarshalBasicContractualRulesContractualRuleArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.ContractualRules = &contractualRules\n\t\t\t}\n\t\tcase \"webSearchUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar webSearchURL string\n\t\t\t\terr = json.Unmarshal(*v, &webSearchURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.WebSearchURL = &webSearchURL\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.ID = &ID\n\t\t\t}\n\t\tcase \"_type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar TypeBasicResponseBase\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tl.Type = typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *ApnsCredential) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *Subscription) UnmarshalJSON(data []byte) error {\n\tif id, ok := ParseID(data); ok {\n\t\ts.ID = id\n\t\treturn nil\n\t}\n\n\ttype subscription Subscription\n\tvar v subscription\n\tif err := json.Unmarshal(data, &v); err != nil {\n\t\treturn err\n\t}\n\n\t*s = Subscription(v)\n\treturn nil\n}", "func (i *Identity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &i.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tenantId\":\n\t\t\terr = unpopulate(val, \"TenantID\", &i.TenantID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &i.Type)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"userAssignedIdentities\":\n\t\t\terr = unpopulate(val, \"UserAssignedIdentities\", &i.UserAssignedIdentities)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *Identity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &i.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tenantId\":\n\t\t\terr = unpopulate(val, \"TenantID\", &i.TenantID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &i.Type)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"userAssignedIdentities\":\n\t\t\terr = unpopulate(val, \"UserAssignedIdentities\", &i.UserAssignedIdentities)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", i, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *ChannelName) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"ChannelName should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = ChannelNameString(s)\n\treturn err\n}", "func subjectFromJWT(c *gin.Context) string {\n\tauthHeader := c.Request.Header.Get(\"Authorization\")\n\tprefix := \"Bearer \"\n\tif !strings.HasPrefix(authHeader, prefix) {\n\t\t// Incorrect Authorization header format.\n\t\treturn \"\"\n\t}\n\ttoken := authHeader[strings.Index(authHeader, prefix)+len(prefix):]\n\tif token == \"\" {\n\t\t// JWT not found.\n\t\treturn \"\"\n\t}\n\n\tvar payload jwt.Payload\n\t_, err := jwt.Verify([]byte(token), jwtKey, &payload)\n\tif err != nil {\n\t\treturn \"\"\n\t}\n\treturn payload.Subject\n}", "func (this *ResourceRequirements) UnmarshalJSON(b []byte) error {\n\treturn CommonUnmarshaler.Unmarshal(bytes.NewReader(b), this)\n}", "func (u *UnencryptedCredentials) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"jobName\":\n\t\t\terr = unpopulate(val, \"JobName\", &u.JobName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"jobSecrets\":\n\t\t\tu.JobSecrets, err = unmarshalCommonJobSecretsClassification(val)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *CreditCard) UnmarshalJSON(data []byte) error {\n\tif string(data) == jsonNull {\n\t\treturn nil\n\t}\n\tvar ustr string\n\tif err := json.Unmarshal(data, &ustr); err != nil {\n\t\treturn err\n\t}\n\t*u = CreditCard(ustr)\n\treturn nil\n}", "func (e *Email) UnmarshalJSON(data []byte) error {\n\tvar estr string\n\tif err := json.Unmarshal(data, &estr); err != nil {\n\t\treturn err\n\t}\n\t*e = Email(estr)\n\treturn nil\n}", "func (receiver *Type) UnmarshalJSON(src []byte) error {\n\tif nil == receiver {\n\t\treturn errNilReceiver\n\t}\n\n\tvar s string\n\tif err := json.Unmarshal(src, &s); nil != err {\n\t\treturn err\n\t}\n\n\tif err := receiver.Scan(s); nil != err {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *Certificate) UnmarshalJSON(b []byte) error {\n\treturn errors.New(\"Do not unmarshal cert JSON directly, use JSONCertificateWithRaw or x509.ParseCertificate function\")\n}", "func (s *SecurityRule) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &s.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SharedAccessAuthorizationRuleResource) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &s.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"location\":\n\t\t\terr = unpopulate(val, \"Location\", &s.Location)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &s.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &s.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &s.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tags\":\n\t\t\terr = unpopulate(val, \"Tags\", &s.Tags)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &s.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UserDetails) UnmarshalJSONObject(d *gojay.Decoder, s string) error {\n\tswitch s {\n\tcase JsKeyEmail:\n\t\treturn d.String(&u.Email)\n\tcase JsKeyCreated:\n\t\treturn d.Time(&u.Created, time.RFC3339Nano)\n\tcase JsKeyProviders:\n\t\treturn d.Array(&u.Providers)\n\tcase JsKeySupports:\n\t\treturn d.Array(&u.Supports)\n\t}\n\n\treturn u.UserMeta.UnmarshalJSONObject(d, s)\n}", "func (s *ServicePrincipalProfile) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"clientId\":\n\t\t\terr = unpopulate(val, \"ClientID\", &s.ClientID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"clientSecret\":\n\t\t\terr = unpopulate(val, \"ClientSecret\", &s.ClientSecret)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *ExpressRouteCircuitAuthorization) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &e.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &e.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &e.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &e.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &e.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (a *AkamaiAccessControl) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"akamaiSignatureHeaderAuthenticationKeyList\":\n\t\t\terr = unpopulate(val, \"AkamaiSignatureHeaderAuthenticationKeyList\", &a.AkamaiSignatureHeaderAuthenticationKeyList)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (rrci *ResourceRateCardInfo) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"Currency\":\n\t\t\tif v != nil {\n\t\t\t\tvar currency string\n\t\t\t\terr = json.Unmarshal(*v, &currency)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\trrci.Currency = &currency\n\t\t\t}\n\t\tcase \"Locale\":\n\t\t\tif v != nil {\n\t\t\t\tvar locale string\n\t\t\t\terr = json.Unmarshal(*v, &locale)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\trrci.Locale = &locale\n\t\t\t}\n\t\tcase \"IsTaxIncluded\":\n\t\t\tif v != nil {\n\t\t\t\tvar isTaxIncluded bool\n\t\t\t\terr = json.Unmarshal(*v, &isTaxIncluded)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\trrci.IsTaxIncluded = &isTaxIncluded\n\t\t\t}\n\t\tcase \"OfferTerms\":\n\t\t\tif v != nil {\n\t\t\t\tofferTerms, err := unmarshalBasicOfferTermInfoArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\trrci.OfferTerms = &offerTerms\n\t\t\t}\n\t\tcase \"Meters\":\n\t\t\tif v != nil {\n\t\t\t\tvar meters []MeterInfo\n\t\t\t\terr = json.Unmarshal(*v, &meters)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\trrci.Meters = &meters\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *ApplicationGatewayTrustedClientCertificate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &a.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &a.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &a.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &a.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (f *FirewallPolicyCertificateAuthority) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"keyVaultSecretId\":\n\t\t\terr = unpopulate(val, \"KeyVaultSecretID\", &f.KeyVaultSecretID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &f.Name)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", f, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *CardType) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"CardType should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = CardTypeString(s)\n\treturn err\n}", "func (mcc *MsgCreateCertificate) UnmarshalJSON(data []byte) error {\n var jsonMsg struct {\n Certificate *utils.JSONWrapper `json:\"certificate\"`\n }\n if err := json.Unmarshal(data, &jsonMsg); err != nil {\n return err\n }\n if certificateFactory, ok := certificateFactories[jsonMsg.Certificate.Type]; ok {\n certificate := certificateFactory()\n if err := json.Unmarshal(jsonMsg.Certificate.Value, certificate); err != nil {\n return err\n }\n mcc.Certificate = certificate\n } else {\n return errors.New(fmt.Sprintf(\"unknown certificate type: %s\", jsonMsg.Certificate.Type))\n }\n return nil\n}", "func (this *NamespacedName) UnmarshalJSON(b []byte) error {\n\treturn CommonUnmarshaler.Unmarshal(bytes.NewReader(b), this)\n}", "func (u *UserAssignedServiceIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &u.Type)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"userAssignedIdentities\":\n\t\t\terr = unpopulate(val, \"UserAssignedIdentities\", &u.UserAssignedIdentities)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (i *MonthlySchedule) UnmarshalJSON(data []byte) error {\n\tvar s string\n\tif err := json.Unmarshal(data, &s); err != nil {\n\t\treturn fmt.Errorf(\"MonthlySchedule should be a string, got %s\", data)\n\t}\n\n\tvar err error\n\t*i, err = MonthlyScheduleString(s)\n\treturn err\n}", "func (l *ListContainerSasInput) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"expiryTime\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"ExpiryTime\", &l.ExpiryTime)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"permissions\":\n\t\t\terr = unpopulate(val, \"Permissions\", &l.Permissions)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", l, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Claims) UnmarshalJSON(b []byte) error {\n\ttokenClaims := struct {\n\t\tTenantString string `json:\"tenant\"`\n\t\tScopes string `json:\"scopes\"`\n\t\tConsumerID string `json:\"consumerID\"`\n\t\tConsumerType consumer.ConsumerType `json:\"consumerType\"`\n\t\tOnBehalfOf string `json:\"onBehalfOf\"`\n\t\tRegion string `json:\"region\"`\n\t\tTokenClientID string `json:\"tokenClientID\"`\n\t\tFlow oathkeeper.AuthFlow `json:\"flow\"`\n\t\tZID string `json:\"zid\"`\n\t\tjwt.StandardClaims\n\t}{}\n\n\terr := json.Unmarshal(b, &tokenClaims)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"while unmarshaling token claims:\")\n\t}\n\n\tc.Scopes = tokenClaims.Scopes\n\tc.ConsumerID = tokenClaims.ConsumerID\n\tc.ConsumerType = tokenClaims.ConsumerType\n\tc.OnBehalfOf = tokenClaims.OnBehalfOf\n\tc.Region = tokenClaims.Region\n\tc.TokenClientID = tokenClaims.TokenClientID\n\tc.Flow = tokenClaims.Flow\n\tc.ZID = tokenClaims.ZID\n\tc.StandardClaims = tokenClaims.StandardClaims\n\n\tif err := json.Unmarshal([]byte(tokenClaims.TenantString), &c.Tenant); err != nil {\n\t\tlog.D().Warnf(\"While unmarshaling tenants: %+v\", err)\n\t\tc.Tenant = make(map[string]string)\n\t}\n\n\treturn nil\n}", "func (t *ThreeTierCustomResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &t.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Social) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonE242b40eDecodeGithubComExampleSample(&r, v)\n\treturn r.Error()\n}", "func (v *Teacher) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonF1627ba7DecodeGithubComDuchiporexiaGoutilsXmsgTests(&r, v)\n\treturn r.Error()\n}", "func (u *ISBN) UnmarshalJSON(data []byte) error {\n\tif string(data) == jsonNull {\n\t\treturn nil\n\t}\n\tvar ustr string\n\tif err := json.Unmarshal(data, &ustr); err != nil {\n\t\treturn err\n\t}\n\t*u = ISBN(ustr)\n\treturn nil\n}", "func (v *License) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson967143c7DecodeGoldenModels(&r, v)\n\treturn r.Error()\n}", "func (ssr *SQLServerRegistration) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"properties\":\n\t\t\tif v != nil {\n\t\t\t\tvar SQLServerRegistrationProperties SQLServerRegistrationProperties\n\t\t\t\terr = json.Unmarshal(*v, &SQLServerRegistrationProperties)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tssr.SQLServerRegistrationProperties = &SQLServerRegistrationProperties\n\t\t\t}\n\t\tcase \"location\":\n\t\t\tif v != nil {\n\t\t\t\tvar location string\n\t\t\t\terr = json.Unmarshal(*v, &location)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tssr.Location = &location\n\t\t\t}\n\t\tcase \"tags\":\n\t\t\tif v != nil {\n\t\t\t\tvar tags map[string]*string\n\t\t\t\terr = json.Unmarshal(*v, &tags)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tssr.Tags = tags\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tssr.ID = &ID\n\t\t\t}\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tssr.Name = &name\n\t\t\t}\n\t\tcase \"type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar string\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\tssr.Type = &typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c *Certificate) jsonifySubjectKey() JSONSubjectKeyInfo {\n\tj := JSONSubjectKeyInfo{\n\t\tKeyAlgorithm: c.PublicKeyAlgorithm,\n\t\tSPKIFingerprint: c.SPKIFingerprint,\n\t}\n\n\tswitch key := c.PublicKey.(type) {\n\tcase *rsa.PublicKey:\n\t\trsaKey := new(jsonKeys.RSAPublicKey)\n\t\trsaKey.PublicKey = key\n\t\tj.RSAPublicKey = rsaKey\n\tcase *dsa.PublicKey:\n\t\tj.DSAPublicKey = &DSAPublicKeyJSON{\n\t\t\tP: key.P.Bytes(),\n\t\t\tQ: key.Q.Bytes(),\n\t\t\tG: key.G.Bytes(),\n\t\t\tY: key.Y.Bytes(),\n\t\t}\n\tcase *ecdsa.PublicKey:\n\t\tparams := key.Params()\n\t\tj.ECDSAPublicKey = &ECDSAPublicKeyJSON{\n\t\t\tP: params.P.Bytes(),\n\t\t\tN: params.N.Bytes(),\n\t\t\tB: params.B.Bytes(),\n\t\t\tGx: params.Gx.Bytes(),\n\t\t\tGy: params.Gy.Bytes(),\n\t\t\tX: key.X.Bytes(),\n\t\t\tY: key.Y.Bytes(),\n\t\t\tCurve: key.Curve.Params().Name,\n\t\t\tLength: key.Curve.Params().BitSize,\n\t\t}\n\tcase *AugmentedECDSA:\n\t\tparams := key.Pub.Params()\n\t\tj.ECDSAPublicKey = &ECDSAPublicKeyJSON{\n\t\t\tP: params.P.Bytes(),\n\t\t\tN: params.N.Bytes(),\n\t\t\tB: params.B.Bytes(),\n\t\t\tGx: params.Gx.Bytes(),\n\t\t\tGy: params.Gy.Bytes(),\n\t\t\tX: key.Pub.X.Bytes(),\n\t\t\tY: key.Pub.Y.Bytes(),\n\t\t\tCurve: key.Pub.Curve.Params().Name,\n\t\t\tLength: key.Pub.Curve.Params().BitSize,\n\t\t\tPub: key.Raw.Bytes,\n\t\t}\n\t}\n\treturn j\n}", "func (t *ThreeTierFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"applicationServer\":\n\t\t\terr = unpopulate(val, \"ApplicationServer\", &t.ApplicationServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"centralServer\":\n\t\t\terr = unpopulate(val, \"CentralServer\", &t.CentralServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"databaseServer\":\n\t\t\terr = unpopulate(val, \"DatabaseServer\", &t.DatabaseServer)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"namingPatternType\":\n\t\t\terr = unpopulate(val, \"NamingPatternType\", &t.NamingPatternType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sharedStorage\":\n\t\t\terr = unpopulate(val, \"SharedStorage\", &t.SharedStorage)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", t, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *ClientCert) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjson42239ddeDecodeGithubComKhliengDispatchServer29(&r, v)\n\treturn r.Error()\n}", "func (p *Partner) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"authorizationExpirationTimeInUtc\":\n\t\t\terr = unpopulateTimeRFC3339(val, \"AuthorizationExpirationTimeInUTC\", &p.AuthorizationExpirationTimeInUTC)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partnerName\":\n\t\t\terr = unpopulate(val, \"PartnerName\", &p.PartnerName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"partnerRegistrationImmutableId\":\n\t\t\terr = unpopulate(val, \"PartnerRegistrationImmutableID\", &p.PartnerRegistrationImmutableID)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", p, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (s *SharedAccessSignatureAuthorizationRule) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"keyName\":\n\t\t\terr = unpopulate(val, \"KeyName\", &s.KeyName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"primaryKey\":\n\t\t\terr = unpopulate(val, \"PrimaryKey\", &s.PrimaryKey)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"rights\":\n\t\t\terr = unpopulate(val, \"Rights\", &s.Rights)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"secondaryKey\":\n\t\t\terr = unpopulate(val, \"SecondaryKey\", &s.SecondaryKey)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", s, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *DocumentSubjectUpdate) UnmarshalEasyJSON(l *jlexer.Lexer) {\n\teasyjsonDe066f3DecodeGithubComStudtoolDocumentsServiceModels3(l, v)\n}", "func (u *UID) UnmarshalJSON(data []byte) error {\n\tu.Str = string(data[1 : len(data)-1])\n\treturn nil\n}", "func (a *ApplicationServerFullResourceNames) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"availabilitySetName\":\n\t\t\terr = unpopulate(val, \"AvailabilitySetName\", &a.AvailabilitySetName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"virtualMachines\":\n\t\t\terr = unpopulate(val, \"VirtualMachines\", &a.VirtualMachines)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *JwtData) UnmarshalJSON(data []byte) error {\n\tr := jlexer.Lexer{Data: data}\n\teasyjsonD2b7633eDecode20191OPGPlus2InternalPkgModels16(&r, v)\n\treturn r.Error()\n}", "func (a *AccountSKU) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"resourceType\":\n\t\t\terr = unpopulate(val, \"ResourceType\", &a.ResourceType)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"sku\":\n\t\t\terr = unpopulate(val, \"SKU\", &a.SKU)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *CertificateWithNonceDescription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &c.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &c.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &c.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (v *NullableSyntheticsAssertion) UnmarshalJSON(src []byte) error {\n\tv.isSet = true\n\n\t// this object is nullable so check if the payload is null or empty string\n\tif string(src) == \"\" || string(src) == \"{}\" {\n\t\treturn nil\n\t}\n\n\treturn json.Unmarshal(src, &v.value)\n}", "func (s *StandardClaim) Unmarshal(data []byte) error {\n\tvar json = jsoniter.ConfigCompatibleWithStandardLibrary\n\terr := json.Unmarshal(data, s)\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (c *CertificateDescription) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &c.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &c.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &c.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &c.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", c, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (r *RegisteredAsnListResult) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"nextLink\":\n\t\t\terr = unpopulate(val, \"NextLink\", &r.NextLink)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &r.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", r, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (certainty *Certainty) UnmarshalJSON(b []byte) error {\n\tvar value int\n\n\tif err := json.Unmarshal(b, &value); err != nil {\n\t\treturn err\n\t}\n\n\t*certainty = Certainty(value)\n\treturn nil\n}", "func (j *Publisher) UnmarshalJSON(input []byte) error {\n\tfs := fflib.NewFFLexer(input)\n\treturn j.UnmarshalJSONFFLexer(fs, fflib.FFParse_map_start)\n}", "func (self *ResourceAuth) UnmarshalJSON(b []byte) error {\n\tvar m rawResourceAuth\n\terr := json.Unmarshal(b, &m)\n\tif err == nil {\n\t\to := ResourceAuth(m)\n\t\t*self = o\n\t\terr = self.Validate()\n\t}\n\treturn err\n}", "func (a *ApplicationGatewaySSLCertificate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &a.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &a.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &a.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &a.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (e *EncryptionIdentity) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"principalId\":\n\t\t\terr = unpopulate(val, \"PrincipalID\", &e.PrincipalID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"tenantId\":\n\t\t\terr = unpopulate(val, \"TenantID\", &e.TenantID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &e.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", e, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *Tx509DataType) UnmarshalJSON(b []byte) error {\n\t// Declare temporary struct without functions to avoid recursive function call\n\ttype Alias Tx509DataType\n\n\t// Copy values into temporary struct\n\ttemp := &struct {\n\t\t*Alias\n\t}{\n\t\tAlias: (*Alias)(u),\n\t}\n\n\t// Unmarshal JSON\n\tif err := json.Unmarshal(b, &temp); err != nil {\n\t\treturn err\n\t}\n\n\t// Decode base64 value\n\tbyteValue, err := base64.StdEncoding.DecodeString(u.XsdGoPkgCDATA)\n\tif err != nil {\n\t\treturn err\n\t}\n\tu.XsdGoPkgCDATA = string(byteValue)\n\n\treturn nil\n}", "func (o *Organization) UnmarshalJSON(body []byte) error {\n\tvar m map[string]*json.RawMessage\n\terr := json.Unmarshal(body, &m)\n\tif err != nil {\n\t\treturn err\n\t}\n\tfor k, v := range m {\n\t\tswitch k {\n\t\tcase \"name\":\n\t\t\tif v != nil {\n\t\t\t\tvar name string\n\t\t\t\terr = json.Unmarshal(*v, &name)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Name = &name\n\t\t\t}\n\t\tcase \"url\":\n\t\t\tif v != nil {\n\t\t\t\tvar URL string\n\t\t\t\terr = json.Unmarshal(*v, &URL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.URL = &URL\n\t\t\t}\n\t\tcase \"image\":\n\t\t\tif v != nil {\n\t\t\t\tvar imageVar ImageObject\n\t\t\t\terr = json.Unmarshal(*v, &imageVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Image = &imageVar\n\t\t\t}\n\t\tcase \"description\":\n\t\t\tif v != nil {\n\t\t\t\tvar description string\n\t\t\t\terr = json.Unmarshal(*v, &description)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Description = &description\n\t\t\t}\n\t\tcase \"entityPresentationInfo\":\n\t\t\tif v != nil {\n\t\t\t\tvar entityPresentationInfo EntitiesEntityPresentationInfo\n\t\t\t\terr = json.Unmarshal(*v, &entityPresentationInfo)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.EntityPresentationInfo = &entityPresentationInfo\n\t\t\t}\n\t\tcase \"bingId\":\n\t\t\tif v != nil {\n\t\t\t\tvar bingID string\n\t\t\t\terr = json.Unmarshal(*v, &bingID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.BingID = &bingID\n\t\t\t}\n\t\tcase \"contractualRules\":\n\t\t\tif v != nil {\n\t\t\t\tcontractualRules, err := unmarshalBasicContractualRulesContractualRuleArray(*v)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.ContractualRules = &contractualRules\n\t\t\t}\n\t\tcase \"webSearchUrl\":\n\t\t\tif v != nil {\n\t\t\t\tvar webSearchURL string\n\t\t\t\terr = json.Unmarshal(*v, &webSearchURL)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.WebSearchURL = &webSearchURL\n\t\t\t}\n\t\tcase \"id\":\n\t\t\tif v != nil {\n\t\t\t\tvar ID string\n\t\t\t\terr = json.Unmarshal(*v, &ID)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.ID = &ID\n\t\t\t}\n\t\tcase \"_type\":\n\t\t\tif v != nil {\n\t\t\t\tvar typeVar TypeBasicResponseBase\n\t\t\t\terr = json.Unmarshal(*v, &typeVar)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\to.Type = typeVar\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func (a *ApplicationGatewayTrustedRootCertificate) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"etag\":\n\t\t\terr = unpopulate(val, \"Etag\", &a.Etag)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"id\":\n\t\t\terr = unpopulate(val, \"ID\", &a.ID)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"name\":\n\t\t\terr = unpopulate(val, \"Name\", &a.Name)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"properties\":\n\t\t\terr = unpopulate(val, \"Properties\", &a.Properties)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, \"Type\", &a.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", a, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UsageName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &u.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &u.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (u *UsageName) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"localizedValue\":\n\t\t\terr = unpopulate(val, \"LocalizedValue\", &u.LocalizedValue)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"value\":\n\t\t\terr = unpopulate(val, \"Value\", &u.Value)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unmarshalling type %T: %v\", u, err)\n\t\t}\n\t}\n\treturn nil\n}", "func (c *ClientCertAuthentication) UnmarshalJSON(data []byte) error {\n\tvar rawMsg map[string]json.RawMessage\n\tif err := json.Unmarshal(data, &rawMsg); err != nil {\n\t\treturn err\n\t}\n\tfor key, val := range rawMsg {\n\t\tvar err error\n\t\tswitch key {\n\t\tcase \"certificateExpirationDate\":\n\t\t\terr = unpopulateTimeRFC3339(val, &c.CertificateExpirationDate)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"certificateSubjectName\":\n\t\t\terr = unpopulate(val, &c.CertificateSubjectName)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"certificateThumbprint\":\n\t\t\terr = unpopulate(val, &c.CertificateThumbprint)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"password\":\n\t\t\terr = unpopulate(val, &c.Password)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"pfx\":\n\t\t\terr = unpopulate(val, &c.Pfx)\n\t\t\tdelete(rawMsg, key)\n\t\tcase \"type\":\n\t\t\terr = unpopulate(val, &c.Type)\n\t\t\tdelete(rawMsg, key)\n\t\t}\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}" ]
[ "0.68584406", "0.6105779", "0.59595776", "0.5927483", "0.59188044", "0.5742977", "0.5716191", "0.57083124", "0.56825536", "0.5634693", "0.55850786", "0.55462366", "0.5545028", "0.5543809", "0.5516314", "0.54736924", "0.5469089", "0.5468079", "0.5457318", "0.5457318", "0.5457318", "0.54404444", "0.5438114", "0.5425012", "0.54221165", "0.54206383", "0.5417908", "0.54143167", "0.53968227", "0.53921574", "0.5382305", "0.53624654", "0.5362419", "0.5360641", "0.535506", "0.5350819", "0.5347322", "0.5346954", "0.53456545", "0.5335717", "0.5334947", "0.5333904", "0.5332847", "0.5332847", "0.53311366", "0.5321637", "0.5318983", "0.5312561", "0.52929926", "0.52847874", "0.52810276", "0.5278764", "0.52731335", "0.5267963", "0.52644944", "0.5261193", "0.52583456", "0.5251911", "0.525094", "0.5244811", "0.52438927", "0.5241255", "0.5241166", "0.5227062", "0.5220476", "0.52057123", "0.5203328", "0.52022046", "0.5179209", "0.5178677", "0.5175727", "0.5175647", "0.51649994", "0.5164348", "0.51629406", "0.51616573", "0.5160812", "0.51563174", "0.5153445", "0.5148834", "0.5140258", "0.51240027", "0.5120657", "0.5120421", "0.51185435", "0.51163447", "0.5114871", "0.51107544", "0.51079774", "0.51032096", "0.5099508", "0.5095317", "0.50932807", "0.5089534", "0.50873667", "0.50861555", "0.5079315", "0.5077522", "0.5077522", "0.5077008" ]
0.75038767
0
Set sets the issuer in the given certificate.
func (i Issuer) Set(c *x509.Certificate) { c.Issuer = pkix.Name{ Country: i.Country, Organization: i.Organization, OrganizationalUnit: i.OrganizationalUnit, Locality: i.Locality, Province: i.Province, StreetAddress: i.StreetAddress, PostalCode: i.PostalCode, SerialNumber: i.SerialNumber, CommonName: i.CommonName, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (m CrossOrderCancelReplaceRequest) SetIssuer(v string) {\n\tm.Set(field.NewIssuer(v))\n}", "func (o *AppSso) SetIssuer(v string) {\n\to.Issuer = &v\n}", "func (m MarketDataSnapshotFullRefresh) SetIssuer(v string) {\n\tm.Set(field.NewIssuer(v))\n}", "func (m SecurityListRequest) SetIssuer(v string) {\n\tm.Set(field.NewIssuer(v))\n}", "func SetIss(issuer string) {\n\tiss = issuer\n}", "func (o *ModelsUser) SetIssuer(v string) {\n\to.Issuer = &v\n}", "func (m OrderStatusRequest) SetIssuer(v string) {\n\tm.Set(field.NewIssuer(v))\n}", "func (m OrderStatusRequest) SetIssuer(v string) {\n\tm.Set(field.NewIssuer(v))\n}", "func (s *OidcConfigForResponse) SetIssuer(v string) *OidcConfigForResponse {\n\ts.Issuer = &v\n\treturn s\n}", "func (s *Store) UpdateIssuer(issuer string) {\n\ts.write(\"/issuer\", issuer)\n}", "func (s Subject) Set(c *x509.Certificate) {\n\tc.Subject = pkix.Name{\n\t\tCountry: s.Country,\n\t\tOrganization: s.Organization,\n\t\tOrganizationalUnit: s.OrganizationalUnit,\n\t\tLocality: s.Locality,\n\t\tProvince: s.Province,\n\t\tStreetAddress: s.StreetAddress,\n\t\tPostalCode: s.PostalCode,\n\t\tSerialNumber: s.SerialNumber,\n\t\tCommonName: s.CommonName,\n\t}\n}", "func (s *OidcConfig) SetIssuer(v string) *OidcConfig {\n\ts.Issuer = &v\n\treturn s\n}", "func (s *AssumeRoleWithSAMLOutput) SetIssuer(v string) *AssumeRoleWithSAMLOutput {\n\ts.Issuer = &v\n\treturn s\n}", "func (client *KeyVaultClient) setCertificateIssuerHandleResponse(resp *http.Response) (KeyVaultClientSetCertificateIssuerResponse, error) {\n\tresult := KeyVaultClientSetCertificateIssuerResponse{RawResponse: resp}\n\tif err := runtime.UnmarshalAsJSON(resp, &result.IssuerBundle); err != nil {\n\t\treturn KeyVaultClientSetCertificateIssuerResponse{}, err\n\t}\n\treturn result, nil\n}", "func (s *CertificateSummary) SetIssuer(v string) *CertificateSummary {\n\ts.Issuer = &v\n\treturn s\n}", "func (s *OIDC) SetIssuer(v string) *OIDC {\n\ts.Issuer = &v\n\treturn s\n}", "func WithIssuer(name string) Option {\n\treturn issuerOption(name)\n}", "func (s *Certificate) SetIssuer(v string) *Certificate {\n\ts.Issuer = &v\n\treturn s\n}", "func SetTLSIssuer(ctx context.Context, issuer string) context.Context {\n\treturn context.WithValue(ctx, PouchTLSIssuer, issuer)\n}", "func (in *Issuer) DeepCopy() *Issuer {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Issuer)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *Issuer) DeepCopy() *Issuer {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(Issuer)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (s *Client) SetSSLCertificate(appName string, chain, key []byte) error {\n\tvar body = struct {\n\t\tChain string `json:\"certificate_chain\"`\n\t\tPrivateKey string `json:\"private_key\"`\n\t}{string(chain), string(key)}\n\n\tvar res struct{}\n\treturn s.Post(&res, fmt.Sprintf(\"/apps/%s/sni-endpoints\", appName), body)\n}", "func (client *KeyVaultClient) setCertificateIssuerCreateRequest(ctx context.Context, vaultBaseURL string, issuerName string, parameter CertificateIssuerSetParameters, options *KeyVaultClientSetCertificateIssuerOptions) (*policy.Request, error) {\n\thost := \"{vaultBaseUrl}\"\n\thost = strings.ReplaceAll(host, \"{vaultBaseUrl}\", vaultBaseURL)\n\turlPath := \"/certificates/issuers/{issuer-name}\"\n\tif issuerName == \"\" {\n\t\treturn nil, errors.New(\"parameter issuerName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{issuer-name}\", url.PathEscape(issuerName))\n\treq, err := runtime.NewRequest(ctx, http.MethodPut, runtime.JoinPaths(host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"7.2\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, runtime.MarshalAsJSON(req, parameter)\n}", "func (me *TAttlistIssnIssnType) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (p *MockPeer) SetEnrollmentCertificate(pem *pem.Block) {\r\n\tp.MockCert = pem\r\n}", "func (info *Info) SetX509Cert(cert *x509.Certificate) {\n\tinfo.Attributes[\"sig.x509.subject\"] = x509tools.FormatSubject(cert)\n\tinfo.Attributes[\"sig.x509.issuer\"] = x509tools.FormatIssuer(cert)\n\td := crypto.SHA1.New()\n\td.Write(cert.Raw)\n\tinfo.Attributes[\"sig.x509.fingerprint\"] = fmt.Sprintf(\"%x\", d.Sum(nil))\n}", "func TestConfigureIssuer(t *testing.T) {\n\twebDir = \"../../../../dex-assets\"\n\tenv := getTestEnv(t)\n\tapi := NewIdentityServer(EnvFromServiceEnv(env), false)\n\n\tserver := newDexWeb(EnvFromServiceEnv(env), api)\n\tdefer server.stopWebServer()\n\n\terr := env.GetDexDB().CreateConnector(dex_storage.Connector{ID: \"conn\", Type: \"github\"})\n\trequire.NoError(t, err)\n\n\t// request the OIDC configuration endpoint - the issuer is an empty string\n\treq := httptest.NewRequest(\"GET\", \"/.well-known/openid-configuration\", nil)\n\trecorder := httptest.NewRecorder()\n\tserver.ServeHTTP(recorder, req)\n\trequire.Equal(t, http.StatusOK, recorder.Result().StatusCode)\n\n\tvar oidcConfig map[string]interface{}\n\trequire.NoError(t, json.NewDecoder(recorder.Result().Body).Decode(&oidcConfig))\n\trequire.Equal(t, \"\", oidcConfig[\"issuer\"].(string))\n\n\t//reconfigure the issuer, the server should reload and serve the new issuer value\n\t_, err = api.SetIdentityServerConfig(context.Background(), &identity.SetIdentityServerConfigRequest{\n\t\tConfig: &identity.IdentityServerConfig{Issuer: \"http://example.com:1234\"},\n\t})\n\trequire.NoError(t, err)\n\n\trecorder = httptest.NewRecorder()\n\tserver.ServeHTTP(recorder, req)\n\trequire.Equal(t, http.StatusOK, recorder.Result().StatusCode)\n\n\trequire.NoError(t, json.NewDecoder(recorder.Result().Body).Decode(&oidcConfig))\n\trequire.Equal(t, \"http://example.com:1234\", oidcConfig[\"issuer\"].(string))\n}", "func (p *MockPeer) SetEnrollmentCertificate(pem *pem.Block) {\n\tp.MockCert = pem\n}", "func (v verifier) VerifyIssuer(raw []byte) error {\n\tsigned, err := cert.ParseSignedIssuer(raw)\n\tif err != nil {\n\t\treturn serrors.WrapStr(\"unable to parse signed issuer certificate\", err)\n\t}\n\treturn v.verifyIssuer(signed)\n}", "func (in *IssuerSpec) DeepCopy() *IssuerSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(IssuerSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *IssuerSpec) DeepCopy() *IssuerSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(IssuerSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func WithIssuer(iss string) ParserOption {\n\treturn func(p *Parser) {\n\t\tp.validator.expectedIss = iss\n\t}\n}", "func (m *InternalDomainFederation) SetNextSigningCertificate(value *string)() {\n err := m.GetBackingStore().Set(\"nextSigningCertificate\", value)\n if err != nil {\n panic(err)\n }\n}", "func SetDetails(cert *x509.Certificate, country, organisation, organisationUnit string) {\n\tcert.Subject = pkix.Name{\n\t\tCountry: []string{country},\n\t\tOrganization: []string{organisation},\n\t\tOrganizationalUnit: []string{organisationUnit},\n\t}\n}", "func (m *X509Certificate) SetIssuerDN(val Principal) {\n\tm.issuerDNField = val\n}", "func (me *TAttlistELocationIDValidYN) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (void *VoidResponse) SetCertificate(path string) *VoidResponse {\n\tvoid.Request = void.Request.Type(gorequest.TypeMultipart).SendFile(path, \"\", \"certificate\")\n\n\treturn void\n}", "func (s *Server) SetCertFromACME(cert *tls.Certificate, err error) {\n\tif err != nil {\n\t\ts.log.Printf(\"ACME FAILED! %s\", err.Error())\n\t\treturn\n\t}\n\n\ts.cert = cert\n}", "func (r ApiListTrustedOAuth2JwtGrantIssuersRequest) Issuer(issuer string) ApiListTrustedOAuth2JwtGrantIssuersRequest {\n\tr.issuer = &issuer\n\treturn r\n}", "func (in *AWSPCAIssuer) DeepCopy() *AWSPCAIssuer {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(AWSPCAIssuer)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (client *KeyVaultClient) updateCertificateIssuerHandleResponse(resp *http.Response) (KeyVaultClientUpdateCertificateIssuerResponse, error) {\n\tresult := KeyVaultClientUpdateCertificateIssuerResponse{RawResponse: resp}\n\tif err := runtime.UnmarshalAsJSON(resp, &result.IssuerBundle); err != nil {\n\t\treturn KeyVaultClientUpdateCertificateIssuerResponse{}, err\n\t}\n\treturn result, nil\n}", "func (_IRMAScheme *IRMASchemeSession) AddIssuer(_id string, _logoUrl string, _metadata []byte) (*types.Transaction, error) {\n\treturn _IRMAScheme.Contract.AddIssuer(&_IRMAScheme.TransactOpts, _id, _logoUrl, _metadata)\n}", "func (c Claims) SetIssuedAt(tm time.Time) Claims {\n\tc[\"iat\"] = tm.UTC().Unix()\n\treturn c\n}", "func (o *OutputManager) SetOutputIDSolidForIssuer(outputID string, issuer string) {\n\to.Lock()\n\tdefer o.Unlock()\n\n\tif _, ok := o.issuerSolidOutIDMap[issuer]; !ok {\n\t\to.issuerSolidOutIDMap[issuer] = make(map[string]types.Empty)\n\t}\n\to.issuerSolidOutIDMap[issuer][outputID] = types.Void\n}", "func (m *WindowsInformationProtectionDataRecoveryCertificate) SetCertificate(value []byte)() {\n err := m.GetBackingStore().Set(\"certificate\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m SecurityListRequest) SetEncodedIssuer(v string) {\n\tm.Set(field.NewEncodedIssuer(v))\n}", "func (m CrossOrderCancelReplaceRequest) SetEncodedIssuer(v string) {\n\tm.Set(field.NewEncodedIssuer(v))\n}", "func (m *ServicePrincipalRiskDetection) SetTokenIssuerType(value *TokenIssuerType)() {\n err := m.GetBackingStore().Set(\"tokenIssuerType\", value)\n if err != nil {\n panic(err)\n }\n}", "func (c *Cache) Set(host string, cert *tls.Certificate) {\n\tc.m.Store(host, cert)\n}", "func (k Keeper) setCert(ctx sdk.Context, addr sdk.AccAddress, cert Cert) {\n\tstore := ctx.KVStore(k.storeKey)\n\tbz := k.cdc.MustMarshalBinary(cert)\n\tstore.Set(KeyCert(addr, cert.Property, cert.Certifier), bz)\n}", "func (in *AWSPCAIssuerSpec) DeepCopy() *AWSPCAIssuerSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(AWSPCAIssuerSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (m *PrinterCreateOperation) SetCertificate(value *string)() {\n m.certificate = value\n}", "func (o ServerCertificateCommonNameResponseOutput) CertificateIssuerThumbprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ServerCertificateCommonNameResponse) string { return v.CertificateIssuerThumbprint }).(pulumi.StringOutput)\n}", "func (n *resPool) SetEntitlement(res *scalar.Resources) {\n\tn.Lock()\n\tdefer n.Unlock()\n\tn.entitlement = res\n\tlog.WithFields(log.Fields{\n\t\t\"respool_id\": n.id,\n\t\t\"entitlement\": res,\n\t}).Debug(\"Setting Entitlement\")\n}", "func (_IRMAScheme *IRMASchemeTransactor) AddIssuer(opts *bind.TransactOpts, _id string, _logoUrl string, _metadata []byte) (*types.Transaction, error) {\n\treturn _IRMAScheme.contract.Transact(opts, \"addIssuer\", _id, _logoUrl, _metadata)\n}", "func (in *IssuerRef) DeepCopy() *IssuerRef {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(IssuerRef)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *IssuerRef) DeepCopy() *IssuerRef {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(IssuerRef)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (_IRMAScheme *IRMASchemeTransactorSession) AddIssuer(_id string, _logoUrl string, _metadata []byte) (*types.Transaction, error) {\n\treturn _IRMAScheme.Contract.AddIssuer(&_IRMAScheme.TransactOpts, _id, _logoUrl, _metadata)\n}", "func (me *TAttlistAuthorValidYN) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (me *TAttlistMedlineCitationOwner) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (o ServerCertificateCommonNameOutput) CertificateIssuerThumbprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ServerCertificateCommonName) string { return v.CertificateIssuerThumbprint }).(pulumi.StringOutput)\n}", "func (a *acmStore) Store(cert *certificate.Resource, domains []string) error {\n\tif cert == nil || cert.Certificate == nil {\n\t\treturn ErrCertificateMissing\n\t}\n\n\tdomainsListString := strings.Join(domains, \", \")\n\n\ta.log.Infof(\"[%s] acm: Retrieving server certificate\", domainsListString)\n\n\tserverCert, err := retrieveServerCertificate(cert.Certificate)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"acm: unable to retrieve server certificate\")\n\t}\n\n\ta.log.Infof(\"[%s] acm: Finding existing server certificate in ACM\", domainsListString)\n\n\texistingCert, err := a.findExistingCertificate(domains)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"acm: unable to find existing certificate\")\n\t}\n\n\t// Retrieve exising certificate ID\n\tvar certArn *string\n\tif existingCert != nil {\n\t\tcertArn = existingCert.CertificateArn\n\t}\n\n\tif certArn != nil {\n\t\ta.log.Infof(\"[%s] acm: Found existing server certificate in ACM with Arn = '%s'\", domainsListString, aws.StringValue(certArn))\n\t}\n\n\t// Init request parameters\n\tinput := &acm.ImportCertificateInput{\n\t\tCertificate: serverCert,\n\t\tCertificateArn: certArn,\n\t\tCertificateChain: cert.IssuerCertificate,\n\t\tPrivateKey: cert.PrivateKey,\n\t}\n\n\tresp, err := a.acm.ImportCertificate(input)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"acm: unable to store certificate into ACM\")\n\t}\n\n\ta.log.Infof(\"[%s] acm: Imported certificate data in ACM with Arn = '%s'\", domainsListString, aws.StringValue(resp.CertificateArn))\n\n\treturn nil\n}", "func (hi *HandshakeInfo) SetIdentityCertProvider(identity certprovider.Provider) {\n\thi.mu.Lock()\n\thi.identityProvider = identity\n\thi.mu.Unlock()\n}", "func (s *Server) SetCert(cert *tls.Certificate) {\n\ts.cert = cert\n}", "func (e *Domain) SetCert(cert *Cert) {\n\te.Certificate = cert\n}", "func (o OauthIdpConfigOutput) Issuer() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *OauthIdpConfig) pulumi.StringOutput { return v.Issuer }).(pulumi.StringOutput)\n}", "func (o GetOidcOpenidConfigResultOutput) Issuer() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetOidcOpenidConfigResult) string { return v.Issuer }).(pulumi.StringOutput)\n}", "func (db *SQLStore) SetCert(cert *ssh.Certificate) error {\n\treturn db.SetRecord(parseCertificate(cert))\n}", "func (me *TAttlistELocationIDEIdType) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (m *PrinterCreateOperation) SetCertificate(value *string)() {\n err := m.GetBackingStore().Set(\"certificate\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o CertificateOutput) IssuerCertificateAuthority() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Certificate) pulumi.StringOutput { return v.IssuerCertificateAuthority }).(pulumi.StringOutput)\n}", "func (o ClientCertificateCommonNameResponseOutput) CertificateIssuerThumbprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ClientCertificateCommonNameResponse) string { return v.CertificateIssuerThumbprint }).(pulumi.StringOutput)\n}", "func (am *ACMEIssuer) setEmail(ctx context.Context, allowPrompts bool) error {\n\tleEmail := am.Email\n\n\t// First try package default email, or a discovered email address\n\tif leEmail == \"\" {\n\t\tleEmail = DefaultACME.Email\n\t}\n\tif leEmail == \"\" {\n\t\tdiscoveredEmailMu.Lock()\n\t\tleEmail = discoveredEmail\n\t\tdiscoveredEmailMu.Unlock()\n\t}\n\n\t// Then try to get most recent user email from storage\n\tvar gotRecentEmail bool\n\tif leEmail == \"\" {\n\t\tleEmail, gotRecentEmail = am.mostRecentAccountEmail(ctx, am.CA)\n\t}\n\tif !gotRecentEmail && leEmail == \"\" && allowPrompts {\n\t\t// Looks like there is no email address readily available,\n\t\t// so we will have to ask the user if we can.\n\t\tvar err error\n\t\tleEmail, err = am.promptUserForEmail()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// User might have just signified their agreement\n\t\tam.mu.Lock()\n\t\tam.agreed = DefaultACME.Agreed\n\t\tam.mu.Unlock()\n\t}\n\n\t// Save the email for later and ensure it is consistent\n\t// for repeated use; then update cfg with the email\n\tleEmail = strings.TrimSpace(strings.ToLower(leEmail))\n\tdiscoveredEmailMu.Lock()\n\tif discoveredEmail == \"\" {\n\t\tdiscoveredEmail = leEmail\n\t}\n\tdiscoveredEmailMu.Unlock()\n\n\t// The unexported email field is the one we use\n\t// because we have thread-safe control over it\n\tam.mu.Lock()\n\tam.email = leEmail\n\tam.mu.Unlock()\n\n\treturn nil\n}", "func (ci *CertificateIssuer) Cancel() {\n\tclose(ci.cancel)\n}", "func (m MarketDataSnapshotFullRefresh) SetEncodedIssuer(v string) {\n\tm.Set(field.NewEncodedIssuer(v))\n}", "func (m *X509Certificate) IssuerDN() Principal {\n\treturn m.issuerDNField\n}", "func (ec *EllipticCurve) Set(value string) error {\n\tswitch strings.ToUpper(value) {\n\tcase strEccP521, \"P-521\":\n\t\t*ec = EllipticCurveP521\n\tcase strEccP384, \"P-384\":\n\t\t*ec = EllipticCurveP384\n\tcase strEccP256, \"P-256\":\n\t\t*ec = EllipticCurveP256\n\tcase strEccED25519:\n\t\t*ec = EllipticCurveED25519\n\tdefault:\n\t\t*ec = EllipticCurveDefault\n\t}\n\n\treturn nil\n}", "func (in *ClusterIssuer) DeepCopy() *ClusterIssuer {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(ClusterIssuer)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func setCaKeyCert(client *api.Client, caConfigPath string, keyCert string) (*api.Secret, error) {\n\tm := map[string]interface{}{\n\t\t\"pem_bundle\": keyCert,\n\t}\n\n\tres, err := client.Logical().Write(caConfigPath, m)\n\tif err != nil {\n\t\tlog.Errorf(\"Write() failed (error %v)\", err)\n\t\treturn nil, err\n\t}\n\treturn res, nil\n}", "func (o FederatedIdentityCredentialOutput) Issuer() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *FederatedIdentityCredential) pulumi.StringOutput { return v.Issuer }).(pulumi.StringOutput)\n}", "func (o LookupSamlResultOutput) Issuer() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LookupSamlResult) string { return v.Issuer }).(pulumi.StringOutput)\n}", "func (m *PrivilegedAccessGroupEligibilitySchedule) SetPrincipal(value DirectoryObjectable)() {\n err := m.GetBackingStore().Set(\"principal\", value)\n if err != nil {\n panic(err)\n }\n}", "func (m *TenantMutation) SetSSOIssuer(s string) {\n\tm._SSOIssuer = &s\n}", "func (m *WindowsInformationProtectionDataRecoveryCertificate) SetSubjectName(value *string)() {\n err := m.GetBackingStore().Set(\"subjectName\", value)\n if err != nil {\n panic(err)\n }\n}", "func (client *KeyVaultClient) getCertificateIssuerHandleResponse(resp *http.Response) (KeyVaultClientGetCertificateIssuerResponse, error) {\n\tresult := KeyVaultClientGetCertificateIssuerResponse{RawResponse: resp}\n\tif err := runtime.UnmarshalAsJSON(resp, &result.IssuerBundle); err != nil {\n\t\treturn KeyVaultClientGetCertificateIssuerResponse{}, err\n\t}\n\treturn result, nil\n}", "func (o ClientCertificateCommonNameOutput) CertificateIssuerThumbprint() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ClientCertificateCommonName) string { return v.CertificateIssuerThumbprint }).(pulumi.StringOutput)\n}", "func (m OrderStatusRequest) SetEncodedIssuer(v string) {\n\tm.Set(field.NewEncodedIssuer(v))\n}", "func parseIssuerResponse(pres *presentproof.Presentation,\n\tvdriReg vdrapi.Registry, docLoader ld.DocumentLoader) (*verifiable.Credential, error) {\n\tif len(pres.PresentationsAttach) == 0 {\n\t\treturn nil, fmt.Errorf(\"%w: expected at least 1 attachment but got 0\", errInvalidCredential)\n\t}\n\n\tattachment := pres.PresentationsAttach[0]\n\n\tvpBytes, err := attachment.Data.Fetch()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to fetch contents of attachment with id %s : %w\", attachment.ID, err)\n\t}\n\n\tvp, err := verifiable.ParsePresentation(\n\t\tvpBytes,\n\t\tverifiable.WithPresPublicKeyFetcher(verifiable.NewVDRKeyResolver(vdriReg).PublicKeyFetcher()),\n\t\tverifiable.WithPresJSONLDDocumentLoader(docLoader),\n\t)\n\tif err != nil {\n\t\treturn nil,\n\t\t\tfmt.Errorf(\"%w: failed to parse verifiable presentation %s: %s\", errInvalidCredential, vpBytes, err.Error())\n\t}\n\n\tif len(vp.Credentials()) != 1 {\n\t\treturn nil, fmt.Errorf(\n\t\t\t\"%w: expected one credential in the issuer's VP but got %d\", errInvalidCredential, len(vp.Credentials()))\n\t}\n\n\trawCred, err := vp.MarshalledCredentials()\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to marshal issuer's vp credentials: %w\", err)\n\t}\n\n\tdata, err := verifiable.ParseCredential(\n\t\trawCred[0],\n\t\tverifiable.WithPublicKeyFetcher(verifiable.NewVDRKeyResolver(vdriReg).PublicKeyFetcher()),\n\t\tverifiable.WithJSONLDDocumentLoader(docLoader),\n\t)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to parse issuer's credential: %w\", err)\n\t}\n\n\treturn data, nil\n}", "func (me *TAttlistInvestigatorValidYN) Set(s string) { (*xsdt.Token)(me).Set(s) }", "func (s *service) SetEnforcer(ef *casbin.SyncedEnforcer) (err error) {\n\tif !s.cfg.Casbin.Enable {\n\t\treturn\n\t}\n\tif s.tool.PtrIsNil(ef) {\n\t\treturn fmt.Errorf(\"enforcer is nil\")\n\t}\n\ts.ef = ef\n\treturn\n}", "func (o *IntrospectedOAuth2Token) SetIss(v string) {\n\to.Iss = &v\n}", "func (c Claims) SetIssuedNow() Claims {\n\tc[\"iat\"] = EpochNow()\n\treturn c\n}", "func (m *MacOSEnterpriseWiFiConfiguration) SetRootCertificateForServerValidation(value MacOSTrustedRootCertificateable)() {\n err := m.GetBackingStore().Set(\"rootCertificateForServerValidation\", value)\n if err != nil {\n panic(err)\n }\n}", "func (o *AppSso) GetIssuer() string {\n\tif o == nil || o.Issuer == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Issuer\n}", "func (h *ValidationHelper) ValidateIssuer(iss string) error {\n\t// Always passes validation if issuer is not provided\n\tif h.issuer == nil {\n\t\treturn nil\n\t}\n\n\treturn h.ValidateIssuerAgainst(iss, *h.issuer)\n}", "func WithIssuerPassword(password []byte) Option {\n\treturn func(o *options) {\n\t\to.issuerPassword = password\n\t}\n}", "func (m *InternalDomainFederation) SetSignOutUri(value *string)() {\n err := m.GetBackingStore().Set(\"signOutUri\", value)\n if err != nil {\n panic(err)\n }\n}", "func (st *State) SetRsyslogCert(caCert string) error {\n\tvar result params.ErrorResult\n\targs := params.SetRsyslogCertParams{\n\t\tCACert: []byte(caCert),\n\t}\n\terr := st.caller.Call(rsyslogAPI, \"\", \"SetRsyslogCert\", args, &result)\n\tif err != nil {\n\t\treturn err\n\t}\n\tif result.Error != nil {\n\t\treturn result.Error\n\t}\n\treturn nil\n}", "func (mr *MockConfigurationMockRecorder) Issuer() *gomock.Call {\n\tmr.mock.ctrl.T.Helper()\n\treturn mr.mock.ctrl.RecordCallWithMethodType(mr.mock, \"Issuer\", reflect.TypeOf((*MockConfiguration)(nil).Issuer))\n}", "func (k *Key) Issuer() string {\n\treturn k.base.Issuer()\n}" ]
[ "0.6783625", "0.64646125", "0.64233357", "0.6402847", "0.6400131", "0.63534003", "0.633354", "0.633354", "0.6285115", "0.61741287", "0.6124071", "0.60192466", "0.60087657", "0.59796745", "0.5918731", "0.5877506", "0.5831812", "0.5810167", "0.5741675", "0.5730789", "0.5730789", "0.5657177", "0.56363577", "0.5628575", "0.5588858", "0.5587578", "0.54463965", "0.5437847", "0.5406956", "0.53468996", "0.53468996", "0.5336208", "0.53329825", "0.5282405", "0.527133", "0.5199091", "0.5178362", "0.51324207", "0.5126386", "0.5125373", "0.51024526", "0.5068148", "0.50577164", "0.5042806", "0.50367886", "0.50212127", "0.50103045", "0.50032246", "0.49812257", "0.4970541", "0.4963076", "0.49540716", "0.4950048", "0.49482894", "0.49477696", "0.49440658", "0.49440658", "0.49439836", "0.49051186", "0.4899775", "0.48824316", "0.48808047", "0.48796046", "0.48593873", "0.4859195", "0.48578998", "0.48545185", "0.4852787", "0.48527277", "0.48267788", "0.48251095", "0.4822193", "0.4813795", "0.481221", "0.48050803", "0.4792261", "0.47707668", "0.4767778", "0.47665334", "0.47618794", "0.47604787", "0.47387615", "0.4733301", "0.47315374", "0.4724541", "0.471719", "0.4715072", "0.4709894", "0.4676924", "0.46701849", "0.46561763", "0.46552557", "0.46281475", "0.46200654", "0.46118397", "0.46062514", "0.4605189", "0.4597781", "0.45825619", "0.45752144" ]
0.7840807
0
TODO(ilackarms): make this a global function
func TransitionFunction(original, desired *v1.Proxy) (bool, error) { if len(original.Listeners) != len(desired.Listeners) { return true, nil } for i := range original.Listeners { if !original.Listeners[i].Equal(desired.Listeners[i]) { return true, nil } } return false, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Location() {}", "func sampleFunc() {\n\t// TODO[TK]\n\t// todo\n\t// FIXME\n}", "func getFunctionName(fn interface{}) string {\n\tname := runtime.FuncForPC(reflect.ValueOf(fn).Pointer()).Name()\n\n\t// Method above returns functions in the form : main.foo\n\tparts := strings.Split(name, \".\")\n\treturn parts[len(parts)-1]\n}", "func (c *Compiler) getGlobalInfo(g *ssa.Global) globalInfo {\n\tinfo := globalInfo{}\n\tif strings.HasPrefix(g.Name(), \"C.\") {\n\t\t// Created by CGo: such a name cannot be created by regular C code.\n\t\tinfo.linkName = g.Name()[2:]\n\t\tinfo.extern = true\n\t} else {\n\t\t// Pick the default linkName.\n\t\tinfo.linkName = g.RelString(nil)\n\t\t// Check for //go: pragmas, which may change the link name (among\n\t\t// others).\n\t\tdoc := c.astComments[info.linkName]\n\t\tif doc != nil {\n\t\t\tinfo.parsePragmas(doc)\n\t\t}\n\t}\n\treturn info\n}", "func (logProxy *loggerProxy)getCallerName() string{\n pc := make([]uintptr, 1)\n //Skipping the functions that are part of loggerProxy to get right caller.\t\n runtime.Callers(4, pc)\n f := runtime.FuncForPC(pc[0])\n return f.Name()\n}", "func prepGlobals() {\n\tcurrent = directory.listChildren() //global current text\n\tcurrentSplit = strings.Split(current, \"\\n\") //global slice of current text.\n}", "func (f *Ifacer) getUniqueImportName(pkgName string, level int) string {\n\tstrs := strings.Split(pkgName, string(filepath.Separator))\n\n\tlenStr := len(strs)\n\n\tif lenStr-1 < level {\n\t\tf.logger.Panicf(\"%d out of range\", level)\n\t}\n\n\tvar importName string\n\tfor _, str := range strs[lenStr-level-1:] {\n\t\tif strings.Contains(str, \".\") {\n\t\t\tstr = strings.Replace(str, \".\", \"\", -1)\n\t\t}\n\t\timportName += str\n\t}\n\n\treturn importName\n}", "func curate(file *pe.File) []byte {\n\n\tm := make(map[string]interface{})\n\tfields := make([]string, 0)\n\n\tif file.HasDOSHdr {\n\t\tm[\"dos_header\"] = file.DOSHeader\n\t\tfields = append(fields, \"dos_header\")\n\t}\n\n\tif file.HasRichHdr {\n\t\tm[\"rich_header\"] = file.RichHeader\n\t\tfields = append(fields, \"rich_header\")\n\t}\n\n\tif file.HasCOFF {\n\t\tm[\"coff\"] = file.COFF\n\t\tfields = append(fields, \"coff\")\n\t}\n\n\tif file.HasNTHdr {\n\t\tm[\"nt_header\"] = file.NtHeader\n\t\tfields = append(fields, \"nt_header\")\n\t}\n\n\tif file.HasSections {\n\t\tm[\"sections\"] = file.Sections\n\t\tfields = append(fields, \"sections\")\n\t}\n\n\tif file.HasExport {\n\t\tm[\"export\"] = file.Export\n\t\tfields = append(fields, \"export\")\n\t}\n\n\tif file.HasImport {\n\t\tm[\"import\"] = file.Imports\n\t\tfields = append(fields, \"import\")\n\t}\n\n\tif file.HasResource {\n\t\tm[\"resource\"] = file.Resources\n\t\tfields = append(fields, \"resource\")\n\t}\n\n\tif file.HasException {\n\t\tm[\"exception\"] = file.Exceptions\n\t\tfields = append(fields, \"exception\")\n\t}\n\n\tif file.HasReloc {\n\t\tm[\"reloc\"] = file.Relocations\n\t\tfields = append(fields, \"reloc\")\n\t}\n\n\tif file.HasDebug {\n\t\tm[\"debug\"] = file.Debugs\n\t\tfields = append(fields, \"debug\")\n\t}\n\n\tif file.HasGlobalPtr {\n\t\tm[\"global_ptr\"] = file.GlobalPtr\n\t\tfields = append(fields, \"global_ptr\")\n\t}\n\n\tif file.HasTLS {\n\t\tm[\"tls\"] = file.TLS\n\t\tfields = append(fields, \"tls\")\n\t}\n\n\tif file.HasLoadCFG {\n\t\tm[\"load_config\"] = file.LoadConfig\n\t\tfields = append(fields, \"load_config\")\n\t}\n\n\tif file.HasBoundImp {\n\t\tm[\"bound_import\"] = file.BoundImports\n\t\tfields = append(fields, \"bound_import\")\n\t}\n\n\tif file.HasIAT {\n\t\tm[\"iat\"] = file.IAT\n\t\tfields = append(fields, \"iat\")\n\t}\n\n\tif file.HasDelayImp {\n\t\tm[\"delay_import\"] = file.DelayImports\n\t\tfields = append(fields, \"delay_import\")\n\t}\n\n\tif file.HasCLR {\n\t\tm[\"clr\"] = file.CLR\n\t\tfields = append(fields, \"clr\")\n\t}\n\n\tif file.HasSecurity {\n\t\tm[\"security\"] = file.Certificates\n\t\tif file.IsSigned {\n\t\t\tif file.Certificates.Verified {\n\t\t\t\tm[\"signature\"] = \"Signed file, valid signature\"\n\t\t\t} else {\n\t\t\t\tm[\"signature\"] = \"Signed file, invalid signature\"\n\t\t\t}\n\t\t}\n\t\tfields = append(fields, \"security\")\n\t} else {\n\t\tm[\"signature\"] = \"File is not signed\"\n\t}\n\n\tm[\"meta\"] = fields\n\treturn toJSON(m)\n}", "func getstring(org string) string {\n\tresult := \"\"\n\t// it is a little status machine( IN ESPCAE: '\"...\\', NOT IN ESPCE: '\"...'(\n\t// mormal string that do not have char '\\')\n\tIN_ESPCAE_STATUS := false\n\tfor _, v := range org[1 : len(org)-1] {\n\t\tif IN_ESPCAE_STATUS {\n\t\t\t// if now in espcae status, it will espcae char '\\', '\\\"', and \"\\'\"\n\t\t\tif v == '\\\\' || v == '\"' || v == '\\'' {\n\t\t\t\tresult += string(v)\n\t\t\t\t// back to normal status\n\t\t\t\tIN_ESPCAE_STATUS = false\n\t\t\t} else {\n\t\t\t\t// it do not matter, we do not need to espcae it, so we back the\n\t\t\t\t// normal string( e.g. \"\\a\" -> \"\\a\" )\n\t\t\t\tresult += string('\\\\')\n\t\t\t\tresult += string(v)\n\t\t\t}\n\t\t} else if v == '\\\\' {\n\t\t\t// if match the char '\\', then change self's status\n\t\t\tIN_ESPCAE_STATUS = true\n\t\t} else {\n\t\t\t// in normal status\n\t\t\tresult += string(v)\n\t\t}\n\t}\n\treturn result\n}", "func setup(){}", "func init() {\n\t//todo...\n}", "func getHelperMethodIfNeeded(functionName string, indent string) (string, bool) {\n\tswitch functionName {\n\tcase \"filebase64sha256\":\n\t\treturn `function computeFilebase64sha256(path string) string {\n\tconst fileData = Buffer.from(fs.readFileSync(path), 'binary')\n\treturn crypto.createHash('sha256').update(fileData).digest('hex')\n}`, true\n\tcase \"notImplemented\":\n\t\treturn fmt.Sprintf(\n\t\t\t`%sfunction notImplemented(message: string) {\n%s throw new Error(message);\n%s}`, indent, indent, indent), true\n\tcase \"singleOrNone\":\n\t\treturn fmt.Sprintf(\n\t\t\t`%sfunction singleOrNone<T>(elements: pulumi.Input<T>[]): pulumi.Input<T> {\n%s if (elements.length != 1) {\n%s throw new Error(\"singleOrNone expected input list to have a single element\");\n%s }\n%s return elements[0];\n%s}`, indent, indent, indent, indent, indent, indent), true\n\tdefault:\n\t\treturn \"\", false\n\t}\n}", "func _getFuncName() string {\n var ptrVal = _getFuncPtrVal()\n\n // Look up the function object\n var fun = runtime.FuncForPC(ptrVal)\n\n // returns: _/home/user/src/gohavenet/src.TestProcMgr_False\n var funcNamePath = fun.Name()\n\n // Split on the slash and return just the func name\n var pathElems = strings.Split(funcNamePath, \"/\")\n var index = len(pathElems) - 1\n if index < 0 {\n index = 0\n }\n\n return pathElems[index]\n}", "func findInterfaceName(i interface{}) (in string) {\n\tin = DefaultFunctionName\n\n\t// TODO: Fix this crap.\n\tname := reflect.ValueOf(i).Type().String()\n\telems := strings.Split(name, \".\")\n\tiname := elems[len(elems)-1]\n\tif len(iname) > 0 {\n\t\tin = iname\n\t}\n\treturn\n}", "func uniquePackageOf(fd *descriptor.FileDescriptorProto) string {\n\ts, ok := uniquePackageName[fd]\n\tif !ok {\n\t\tlog.Fatal(\"internal error: no package name defined for \" + fd.GetName())\n\t}\n\treturn s\n}", "func TestGetNone4A(t *testing.T) {\n}", "func exactFunction() {\n\treturn nil\n}", "func TestSinglePrewrite4A(t *testing.T) {\n}", "func getCallerSourceLocation() string {\n\t_, file, line, ok := runtime.Caller(2)\n\tresult := \"unknown:unknown\"\n\tif ok {\n\t\tresult = fmt.Sprintf(\"%s:%d\", file, line)\n\t}\n\treturn result\n}", "func generateByThirdPackage() {\n\n}", "func (b BananaPhone) getSysID(funcname string, ord uint32, useOrd bool) (uint16, error) {\n\tex, e := b.banana.Exports()\n\tif e != nil {\n\t\treturn 0, e\n\t}\n\n\tfor _, exp := range ex {\n\t\tif (useOrd && exp.Ordinal == ord) || // many bothans died for this feature (thanks awgh). Turns out that a value can be exported by ordinal, but not by name! man I love PE files. ha ha jk.\n\t\t\texp.Name == funcname {\n\t\t\toffset := rvaToOffset(b.banana, exp.VirtualAddress)\n\t\t\tb, e := b.banana.Bytes()\n\t\t\tif e != nil {\n\t\t\t\treturn 0, e\n\t\t\t}\n\t\t\tbuff := b[offset : offset+10]\n\n\t\t\treturn sysIDFromRawBytes(buff)\n\t\t}\n\t}\n\treturn 0, errors.New(\"Could not find syscall ID\")\n}", "func main() {\n var name_from_main_function string = \"Ravikanth Garimella\"\n\n fmt.Printf(\"Name from main Function:%s\\n \",name_from_main_function)\n\n fmt.Printf(\"Name from stringutil name.go file is %s\\n\", stringutil.Name)\n\n fmt.Printf(\"Name from the Function GetName() is %s\\n\", stringutil.GetName())\n\n fmt.Printf(\"Name from the Function GetNamePassVariable after passing the variable is %s\\n\", stringutil.GetNamePassVariable(\"Ravikanth Garimella\"))\n\n fmt.Printf(\"Name trying to call variable with lower case: %s\\n\", stringutil.GetNamePassVariable2(\"Nasser Hussain\"))\n\n// fmt.Printf(\"Name trying to call variable with lower case directly: %s\\n\",stringutil.get_name_lower_case(\"Lower Case\"))\n/* Above statement fails with error : # command-line-arguments\n./main.go:23: cannot refer to unexported name stringutil.get_name_lower_case */\n\n}", "func getHelperMethodIfNeeded(functionName string, indent string) (string, bool) {\n\tswitch functionName {\n\tcase \"filebase64\":\n\t\treturn `private static string ReadFileBase64(string path) {\n\t\treturn Convert.ToBase64String(System.Text.Encoding.UTF8.GetBytes(File.ReadAllText(path)));\n\t}`, true\n\tcase \"filebase64sha256\":\n\t\treturn `private static string ComputeFileBase64Sha256(string path) {\n\t\tvar fileData = System.Text.Encoding.UTF8.GetBytes(File.ReadAllText(path));\n\t\tvar hashData = SHA256.Create().ComputeHash(fileData);\n\t\treturn Convert.ToBase64String(hashData);\n\t}`, true\n\tcase \"sha1\":\n\t\treturn `private static string ComputeSHA1(string input) {\n\t\treturn BitConverter.ToString(\n\t\t\tSHA1.Create().ComputeHash(Encoding.UTF8.GetBytes(input))\n\t\t).Replace(\"-\",\"\").ToLowerInvariant());\n\t}`, true\n\tcase \"notImplemented\":\n\t\treturn fmt.Sprintf(`\n%sstatic object NotImplemented(string errorMessage) \n%s{\n%s throw new System.NotImplementedException(errorMessage);\n%s}`, indent, indent, indent, indent), true\n\tdefault:\n\t\treturn \"\", false\n\t}\n}", "func (objectIDs *DbObjectIDs) getUniqueID() (string, error) {\n\tid := objectIDs.ownerControllerName + \":\" + string(objectIDs.idsType.ownerObjectType)\n\tfor _, key := range objectIDs.idsType.GetExternalIDKeys() {\n\t\tvalue, ok := objectIDs.objectIDs[key]\n\t\tif !ok {\n\t\t\treturn \"\", fmt.Errorf(\"key %v is required but not present\", key)\n\t\t}\n\t\tid += \":\" + value\n\t}\n\treturn id, nil\n}", "func getCaller(skip int, shortFileName bool) string {\n\tvar b strings.Builder\n\t_, file, no, ok := runtime.Caller(skip)\n\tif ok {\n\t\tif shortFileName {\n\t\t\tif lastSlashIndex := strings.LastIndex(file, \"/\"); lastSlashIndex != -1 {\n\t\t\t\tfile = file[lastSlashIndex+1:]\n\t\t\t}\n\t\t}\n\t\tfmt.Fprintf(&b, \"%s:%d\", file, no)\n\t}\n\treturn b.String()\n}", "func Utils() string {\n\treturn \"Hi there, this is an example code file!\"\n}", "func fullFilePath2(fSys *afero.Afero, fn string) string {\n\txType := reflect.TypeOf(fSys)\n\txValue := reflect.ValueOf(fSys)\n\tfmt.Println(xType, xValue) // Os: *afero.Afero &{0x85d228} // MemMap: *afero.Afero &{0xc00007bb60}\n\tif fn == \"none\" || fn == \"off\" {\n\t\treturn fn\n\t}\n\tvar e error\n\tvar s string\n\t//if xValue < &{0xc000000000} {\n\ts, e = filepath.Abs(fn)\n\tmsg.InfoOnErr(e, fmt.Sprintf(\"failed to parse %s\\n\", fn))\n\t//}\n\treturn s\n}", "func (m *generateCommon) getGenerateCommon() *generateCommon {\n\treturn m\n}", "func getGlobalAuthEndpoint(baseURL string) string {\n\tparts := strings.Split(baseURL, \"api.konghq\")\n\treturn baseEndpointUS + parts[len(parts)-1] + authEndpointV2\n}", "func getUID(lib utils.PathIdentifier) string {\n\treturn lib.Key()[:5]\n}", "func Init() error {\n\n}", "func getCurrentAbPathByCaller() string {\n\tvar abPath string\n\t_, filename, _, ok := runtime.Caller(0)\n\tif ok {\n\t\tabPath = path.Dir(filename)\n\t}\n\treturn abPath\n}", "func getCurrentAbPathByCaller() string {\n\tvar abPath string\n\t_, filename, _, ok := runtime.Caller(0)\n\tif ok {\n\t\tabPath = path.Dir(filename)\n\t}\n\treturn abPath\n}", "func (j *juicefs) getUniqueId(ctx context.Context, volumeId string) (string, error) {\n\tif os.Getenv(\"STORAGE_CLASS_SHARE_MOUNT\") == \"true\" && !config.ByProcess {\n\t\tpv, err := j.K8sClient.GetPersistentVolume(ctx, volumeId)\n\t\t// In static provision, volumeId may not be PV name, it is expected that PV cannot be found by volumeId\n\t\tif err != nil && !k8serrors.IsNotFound(err) {\n\t\t\treturn \"\", err\n\t\t}\n\t\t// In dynamic provision, PV.spec.StorageClassName is which SC(StorageClass) it belongs to.\n\t\tif err == nil && pv.Spec.StorageClassName != \"\" {\n\t\t\treturn pv.Spec.StorageClassName, nil\n\t\t}\n\t}\n\treturn volumeId, nil\n}", "func getOID(oid string) (string, error) {\n\tif strings.HasPrefix(oid, \".\") {\n\t\treturn oid, nil\n\t}\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tfixed, ok := lookupOID[oid]\n\tif !ok {\n\t\treturn oid, errors.Errorf(\"no OID found for %s\", oid)\n\t}\n\treturn fixed, nil\n}", "func initAll()", "func getWrapper(t *testing.T, filename string) *CFGWrapper {\n\tvar config loader.Config\n\tf, err := config.ParseFile(filename, nil)\n\tif err != nil {\n\t\tt.Error(err.Error())\n\t\tt.FailNow()\n\t\treturn nil\n\t}\n\n\tconfig.CreateFromFiles(\"testing\", f)\n\n\tprog, err := config.Load()\n\n\tif err != nil {\n\t\tt.Error(err.Error())\n\t\tt.FailNow()\n\t\treturn nil\n\t}\n\n\tfirstFunc, ok := f.Decls[0].(*ast.FuncDecl)\n\tif !ok { // skip import decl if exists\n\t\tfirstFunc = f.Decls[1].(*ast.FuncDecl) // panic here if no first func\n\t}\n\tcfg := cfg.FromFunc(firstFunc)\n\tv := make(map[int]ast.Stmt)\n\tstmts := make(map[ast.Stmt]int)\n\tobjs := make(map[string]*types.Var)\n\tobjNames := make(map[*types.Var]string)\n\ti := 1\n\tast.Inspect(firstFunc, func(n ast.Node) bool {\n\t\tswitch x := n.(type) {\n\t\tcase *ast.Ident:\n\t\t\tif obj, ok := prog.Created[0].ObjectOf(x).(*types.Var); ok {\n\t\t\t\tobjs[obj.Name()] = obj\n\t\t\t\tobjNames[obj] = obj.Name()\n\t\t\t}\n\t\tcase ast.Stmt:\n\t\t\tswitch x.(type) {\n\t\t\tcase *ast.BlockStmt:\n\t\t\t\treturn true\n\t\t\t}\n\t\t\tv[i] = x\n\t\t\tstmts[x] = i\n\t\t\ti++\n\t\tcase *ast.FuncLit:\n\t\t\t// skip statements in anonymous functions\n\t\t\treturn false\n\t\t}\n\t\treturn true\n\t})\n\tv[END] = cfg.Exit\n\tv[START] = cfg.Entry\n\tstmts[cfg.Entry] = START\n\tstmts[cfg.Exit] = END\n\tif len(v) != len(cfg.GetBlocks()) {\n\t\tt.Logf(\"expected %d vertices, got %d --construction error\", len(v), len(cfg.GetBlocks()))\n\t}\n\n\treturn &CFGWrapper{\n\t\tcfg: cfg,\n\t\tprog: prog,\n\t\texp: v,\n\t\tstmts: stmts,\n\t\tobjs: objs,\n\t\tobjNames: objNames,\n\t\tfset: prog.Fset,\n\t\tf: f,\n\t}\n}", "func (d *Abstraction) getGeneratedName(s string, m string) string {\n\tvar buffer bytes.Buffer\n\tbuffer.WriteString(s)\n\tbuffer.WriteString(\".\")\n\tbuffer.WriteString(m)\n\treturn buffer.String()\n}", "func getDefLink(resourceName string) *string {\n\tret := defPrefix + resourceName\n\treturn &ret\n}", "func getSinglePath(v interface{}) (string, error) {\n\tif name, err := resourceName(v); err != nil {\n\t\treturn \"\", err\n\t} else if id, err := resourceId(v); err != nil {\n\t\treturn \"\", err\n\t} else {\n\t\treturn fmt.Sprintf(\"%s(%d)\", name, id), nil\n\t}\n}", "func getHelper(nodeRPC string, input string, fname string, cmd string) {\n\n\tcolorprint.Info(\">>>> Please enter the name of the file that you would like to obtain\")\n\tfmt.Scan(&fname)\n\tcolorprint.Debug(\"<<<< \" + fname)\n\tcolorprint.Info(\">>>> Please enter the address of the node you want to connect to\")\n\tfmt.Scan(&input)\n\tcolorprint.Debug(\"<<<< \" + input)\n\tnodeAddr := input\n\t// Connect to utility.Service via RPC // returns *Client, err\n\tavail, _, _ := CheckFileAvailability(fname, nodeAddr)\n\tif avail && (cmd == \"get\") {\n\t\tcolorprint.Info(\">>>> Would you like to get the file from the node[\" + nodeRPC + \"]?(y/n)\")\n\t\tfmt.Scan(&input)\n\t\tcolorprint.Debug(\"<<<< \" + input)\n\t\tif input == \"y\" {\n\t\t\t// TODO\n\t\t}\n\t}\n}", "func getHostName() (string){\n host, err := os.Hostname()\n if err != nil {\n fmt.Println(\"%s\", err)\n log.Error(\"get os.HostName error \", err)\n } else {\n //fmt.Println(\"%s\", host)\n log.Debugf(\"get os.HostName success host:\", host)\n }\n return host\n}", "func TestPrewriteMultiple4A(t *testing.T) {\n}", "func init() {}", "func init() {}", "func init() {}", "func init() {}", "func normalizeLocation(in string) string {\n\tout := in\n\tout = stackLocation.ReplaceAllString(out, \"$1\")\n\tout = functionArgs.ReplaceAllString(out, \"$1()\")\n\tout = testFailureOutput.ReplaceAllString(out, \"\")\n\tout = klogPrefix.ReplaceAllString(out, \"<klog> \")\n\treturn out\n}", "func (h *CookiejarHandler) getAddress(name string) string {\n\thashedName := Hexdigest(name)\n\treturn h.namespace + hashedName[:64]\n}", "func globalIntern(s string) string {\n\tglobalInternMu.Lock()\n\tdefer globalInternMu.Unlock()\n\treturn globalInternMap.Intern(s)\n}", "func getGlobalInfo() (globalInfo map[string]interface{}) {\n\tglobalInfo = map[string]interface{}{/*\n\t\t\"isDistXL\": globalIsDistXL,\n\t\t\"isXL\": globalIsXL,\n\t\t\"isBrowserEnabled\": globalIsBrowserEnabled,\n\t\t\"isWorm\": globalWORMEnabled,\n\t\t\"isEnvBrowser\": globalIsEnvBrowser,\n\t\t\"isEnvCreds\": globalIsEnvCreds,\n\t\t\"isEnvRegion\": globalIsEnvRegion,\n\t\t\"isSSL\": globalIsSSL,\n\t\t\"serverRegion\": globalServerRegion,\n\t\t// Add more relevant global settings here.*/\n\t}\n\n\treturn globalInfo\n}", "func getFullyQualifiedFunctionName(fn interface{}) string {\n\treturn runtime.FuncForPC(reflect.ValueOf(fn).Pointer()).Name()\n}", "func generateTableId(db *sql.DB, tablename string) string {\n\n\trand128bits := make([]byte, 16)\n\t_, err := rand.Read(rand128bits)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tid := hex.EncodeToString(rand128bits)\n\n\t/* Check for collision */\n\tvar available bool\n\ttest := fmt.Sprintf(\"SELECT COUNT(*) = 0 FROM %s WHERE id = %q\", tablename, id)\n\terr = db.QueryRow(test).Scan(&available)\n\n\tif err != nil {\n\t\tfmt.Println(\"err.Error()\")\n\t\treturn id\n\t}\n\n\tif(available){\n\t\treturn id\n\t} else {\n\t\treturn generateTableId(db, tablename)\n\t}\n}", "func FindLarge() { // accessible to inside as well as outside package\r\n\tfmt.Println(\"LArge find large\")\r\n}", "func (r *Render) getCurrentPrefix() string {\n\tif prefix, ok := r.livePrefixCallback(); ok {\n\t\treturn prefix\n\t}\n\treturn r.prefix\n}", "func getCallPriorityID(strPriorityName string, espXmlmc *apiLib.XmlmcInstStruct, buffer *bytes.Buffer) (string, string) {\n\tpriorityID := \"\"\n\tif swImportConf.PriorityMapping[strPriorityName] != nil {\n\t\tstrPriorityName = fmt.Sprintf(\"%s\", swImportConf.PriorityMapping[strPriorityName])\n\t\tif strPriorityName != \"\" {\n\t\t\tpriorityID = getPriorityID(strPriorityName, espXmlmc, buffer)\n\t\t}\n\t}\n\treturn priorityID, strPriorityName\n}", "func passthruLookup(name string) (string, bool) {\n\treturn \"\", false\n}", "func (p *callgrindProfiler) getFunctionParameters(function *lisp.LVal) (string, int, string) {\n\tvar source string\n\tline := 0\n\tif function.Source == nil {\n\t\tif cell := function.Cells[0]; cell != nil && cell.Source != nil {\n\t\t\tsource = cell.Source.File\n\t\t\tline = cell.Source.Line\n\t\t} else {\n\t\t\tsource = \"no-source\"\n\t\t}\n\t} else {\n\t\tsource = function.Source.File\n\t\tline = function.Source.Line\n\t}\n\tfName := fmt.Sprintf(\"%s:%s\", function.FunData().Package, getFunNameFromFID(p.runtime, function.FunData().FID))\n\treturn source, line, fName\n}", "func getCallerPackage() (string, bool) {\n\tvar pkg string\n\tpc, _, _, ok := runtime.Caller(2)\n\tif !ok {\n\t\treturn pkg, false\n\t}\n\tparts := strings.Split(runtime.FuncForPC(pc).Name(), \".\")\n\tif parts[len(parts)-2][0] == '(' {\n\t\tpkg = strings.Join(parts[0:len(parts)-2], \".\")\n\t} else {\n\t\tpkg = strings.Join(parts[0:len(parts)-1], \".\")\n\t}\n\treturn pkg, true\n}", "func checkSiteConfig(config Config) {\n\n}", "func getSenderPrefix(conv chat1.ConversationLocal, currentUsername, senderUsername string) (senderPrefix string) {\n\tif showSenderPrefix(conv) {\n\t\tif senderUsername == currentUsername {\n\t\t\tsenderPrefix = \"You: \"\n\t\t} else {\n\t\t\tsenderPrefix = fmt.Sprintf(\"%s: \", senderUsername)\n\t\t}\n\t}\n\treturn senderPrefix\n}", "func getCallerFuncName() (callerFuncName string) {\n\tpc, _, _, _ := runtime.Caller(2)\n\tdetails := runtime.FuncForPC(pc)\n\treturn details.Name()\n}", "func (alias *Alias) loadGlobalAlias() {\n\t//Identify defaults location.\n\tif runtime.GOOS == util.WindowsOS {\n\t\treadAliasFromBytes([]byte(globalDefaultYamlWindows), alias)\n\t} else { // Looking at Linux\n\t\treadAliasFromBytes([]byte(globalDefaultYamlLinux), alias)\n\t}\n}", "func getCaller(callDepth int, suffixesToIgnore ...string) (file string, line int) {\n\t// bump by 1 to ignore the getCaller (this) stackframe\n\tcallDepth++\nouter:\n\tfor {\n\t\tvar ok bool\n\t\t_, file, line, ok = runtime.Caller(callDepth)\n\t\tif !ok {\n\t\t\tfile = \"???\"\n\t\t\tline = 0\n\t\t\tbreak\n\t\t}\n\n\t\tfor _, s := range suffixesToIgnore {\n\t\t\tif strings.HasSuffix(file, s) {\n\t\t\t\tcallDepth++\n\t\t\t\tcontinue outer\n\t\t\t}\n\t\t}\n\t\tbreak\n\t}\n\treturn\n}", "func almostOkayFunction() {\n\treturn nil\n}", "func (a *api) getBaseURL(targetAppID string) string {\n\tendpoint, ok := a.universal.CompStore.GetHTTPEndpoint(targetAppID)\n\tif ok && endpoint.Name == targetAppID {\n\t\treturn endpoint.Spec.BaseURL\n\t}\n\treturn \"\"\n}", "func goodFunction() {\n\treturn nil\n}", "func getErrOrigin() string {\n\tvar origin string\n\tif pc, file, line, ok := runtime.Caller(2); ok {\n\t\tif f := runtime.FuncForPC(pc); f != nil {\n\t\t\tif wd, err := os.Getwd(); err == nil {\n\t\t\t\torigin = fmt.Sprintf(\"%s %s:%d\", strings.TrimPrefix(f.Name(), pkgPath), strings.TrimPrefix(file, wd+string(os.PathSeparator)), line)\n\t\t\t}\n\t\t}\n\t}\n\treturn origin\n}", "func (api *nodeAPI) getFullKey(tenant, name string) string {\n\tif tenant != \"\" {\n\t\treturn fmt.Sprint(globals.ConfigRootPrefix, \"/\", \"cluster\", \"/\", \"nodes\", \"/\", tenant, \"/\", name)\n\t}\n\treturn fmt.Sprint(globals.ConfigRootPrefix, \"/\", \"cluster\", \"/\", \"nodes\", \"/\", name)\n}", "func getFunctionName(fn interface{}) string {\n\treturn runtime.FuncForPC(reflect.ValueOf((fn)).Pointer()).Name()\n}", "func get_Progname(name string) string {\n var pos = 0\n for i, ch := range name {\n if ch == '/' { pos = i }\n }\n return name[pos:]\n}", "func mustBeUnique(label string) {\n\n\tif _, ok := functions[label]; ok {\n\t\tpanic(fmt.Sprintf(\"%s already defined as function\", label))\n\t}\n\n\tif _, ok := variables[label]; ok {\n\t\tpanic(fmt.Sprintf(\"%s already defined as variable\", label))\n\t}\n\n\tif _, ok := aliases[label]; ok {\n\t\tpanic(fmt.Sprintf(\"%s already defined as alias\", label))\n\t}\n}", "func init() {\n\t// algorithm taken from\n\t// https://github.com/zenazn/goji/blob/master/web/middleware/request_id.go#L44-L50\n\tvar buf [12]byte\n\tvar b64 string\n\tfor len(b64) < 10 {\n\t\trand.Read(buf[:])\n\t\tb64 = base64.StdEncoding.EncodeToString(buf[:])\n\t\tb64 = strings.NewReplacer(\"+\", \"\", \"/\", \"\").Replace(b64)\n\t}\n\treqPrefix = string(b64[0:10])\n}", "func funcName() string {\n\tpc, _, _, _ := runtime.Caller(1)\n\tcompleteFuncname := runtime.FuncForPC(pc).Name()\n\treturn strings.Split(completeFuncname, \".\")[len(strings.Split(completeFuncname, \".\"))-1]\n}", "func getCallerPackage() string {\n\tconst replayModeRecordModeCaller = 3\n\tc := caller.Get(replayModeRecordModeCaller)\n\tpkg := strings.SplitN(c, \".\", 2)[0]\n\treturn path.Base(pkg)\n}", "func getTeamID(env config.Environment, id string) string {\n\tvar teamIDLookup string\n\tif env.DebugOn {\n\t\tteamIDLookup = \"TGB941BGQ\"\n\t} else {\n\t\tteamIDLookup = id\n\t}\n\treturn teamIDLookup\n}", "func printErr (err error){ // this function is used for printing error\r\n if err != nil{\r\n fmt.Println(err)\r\n }\r\n }", "func initConfig() {\n}", "func initConfig() {\n}", "func initConfig() {\n}", "func initConfig() {\n}", "func initConfig() {\n}", "func initConfig() {\n}", "func TestPrewriteWritten4A(t *testing.T) {\n}", "func (t *SimpleChaincode) getLoc(stub *shim.ChaincodeStub , args []string) ([]byte,error) {\n \n \t \n \ts := []string{args[0], \"requester\"};\n s1 := strings.Join(s, \"_\");\n \t \n \n requester_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n \t//------------------------------------------------------------\n \ts = []string{args[0], \"beneficiary\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n beneficiary_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n \ts = []string{args[0], \"amount\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n amount_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"expiry_date\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n expiry_date_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"status\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n status_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"advising_bank\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n advising_bank_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"document_hash\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n document_hash_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"loc_filename\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n loc_filename_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n \ts = []string{args[0], \"contract_hash\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n contract_hash_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n\ts = []string{args[0], \"bol_hash\"};\n s1 = strings.Join(s, \"_\");\n \t \n \n bol_hash_string, err := stub.GetState(s1);\n \t\n \tif err != nil {\n\t\treturn nil, err\n\t}\n\t//--------------------------------------------------------------\n \t\n \ts = []string{string(requester_string),string(beneficiary_string),string(amount_string),string(expiry_date_string),string(status_string),string(advising_bank_string),string(document_hash_string),string(loc_filename_string),string(contract_hash_string),string(bol_hash_string)};\n \n // s=[]string{string(contract_hash_string),string(bol_hash_string)};\n final_string := strings.Join(s, \"|\");\n \t\n \t\n \t\n //s := strconv.Itoa(counter) ;\n //ret_s := []byte(s);\n return []byte(final_string), nil;\n \n }", "func PicUsefulLog() {\n\n}", "func (fOpenMode FileOpenMode) checkInitializeMaps(reInitialize bool) {\n\n if !reInitialize &&\n mFileOpenModeIntToString != nil &&\n len(mFileOpenModeIntToString) > 5 &&\n mFileOpenModeStringToInt != nil &&\n len(mFileOpenModeStringToInt) > 5 &&\n mFileOpenModeLwrCaseStringToInt != nil &&\n len(mFileOpenModeLwrCaseStringToInt) > 5 {\n return\n }\n\n var t = FOpenMode.ModeAppend()\n\n mFileOpenModeIntToString = make(map[int]string, 0)\n mFileOpenModeStringToInt = make(map[string]int, 0)\n mFileOpenModeLwrCaseStringToInt = make(map[string]int, 0)\n\n s := reflect.TypeOf(t)\n\n r := reflect.TypeOf(0) // int\n args := [1]reflect.Value{reflect.Zero(s)}\n\n for i := 0; i < s.NumMethod(); i++ {\n\n f := s.Method(i).Name\n\n if f == \"String\" ||\n f == \"ParseString\" ||\n f == \"Value\" ||\n f == \"IsValid\" ||\n f == \"checkInitializeMaps\" {\n continue\n }\n\n value := s.Method(i).Func.Call(args[:])[0].Convert(r).Int()\n x := int(value)\n mFileOpenModeIntToString[x] = f\n mFileOpenModeStringToInt[f] = x\n mFileOpenModeLwrCaseStringToInt[strings.ToLower(f)] = x\n }\n\n}", "func (t UrnTargets) getMatcher(glob string) *regexp.Regexp {\n\tif r := t.globs[glob]; r != nil {\n\t\treturn r\n\t}\n\tsegmentGlob := strings.Split(glob, \"**\")\n\tfor i, v := range segmentGlob {\n\t\tpart := strings.Split(v, \"*\")\n\t\tfor i, v := range part {\n\t\t\tpart[i] = regexp.QuoteMeta(v)\n\t\t}\n\t\tsegmentGlob[i] = strings.Join(part, \"[^:]*\")\n\t}\n\n\t// Because we have quoted all input, this is safe to compile.\n\tr := regexp.MustCompile(\"^\" + strings.Join(segmentGlob, \".*\") + \"$\")\n\n\t// We cache and return the matcher\n\tt.globs[glob] = r\n\treturn r\n}", "func getUniquePath(url string) string {\n\t// Include current timestamp in the hashed string to ensure that we can have multiple paths for the same url\n\thashedURL := sha256.Sum256([]byte(url + time.Now().Format(time.RFC850)))\n\treturn base64.URLEncoding.EncodeToString(hashedURL[:])[:9]\n}", "func (lId *LangId) getVarNames() {\n\n\tvar getVarNames = func(toCamel bool) []string {\n\n\t\tvar (\n\t\t\ttmpMap = make(map[string]string)\n\t\t\tout []string\n\t\t\ttmpName string\n\t\t)\n\t\tfor _, val := range lId.lines {\n\t\t\ttmpSl := strings.Split(val, \":\")\n\t\t\tif len(tmpSl) > 1 {\n\t\t\t\ttmpMap[tmpSl[0]] = \"\"\n\t\t\t}\n\t\t}\n\t\tfor name, _ := range tmpMap {\n\t\t\tif toCamel {\n\t\t\t\ttmpName = ToCamel(strings.TrimSpace(name))\n\t\t\t} else {\n\t\t\t\ttmpName = strings.TrimSpace(name)\n\t\t\t}\n\t\t\tif len(tmpName) > 25 || len(tmpName) == 0 {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tout = append(out, tmpName)\n\t\t}\n\t\t// Sort string preserving order ascendant\n\t\tsort.SliceStable(out, func(i, j int) bool {\n\t\t\treturn out[i] < out[j]\n\t\t})\n\t\treturn out\n\t}\n\n\tfmt.Println(\"To replace 'langIdEntry' structure\")\n\tfmt.Println(\"type langIdEntry struct {\")\n\tvarNames := getVarNames(true)\n\tfor idx, name := range varNames {\n\t\tfmt.Printf(\"\\t%s\", name)\n\t\tif idx < len(varNames)-1 {\n\t\t\tfmt.Println(\",\")\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\tfmt.Println(` string\n}\n`)\n\n\tfmt.Println(\"\\nTo replace 'varNamesList'\")\n\tfmt.Println(\"var varNamesList = []string{\")\n\tfor _, name := range getVarNames(false) {\n\t\tfmt.Printf(\"\\t\\\"%s:\\\",\\n\", name)\n\t}\n\tfmt.Println(\"}\")\n\n\tfmt.Println(\"\\nTo replace 'storeEntry' method\")\n\tvar cases string\n\tfor _, name := range getVarNames(false) {\n\t\tcases += buildCaseStatement(name)\n\t}\n\tfmt.Println(buildFuncContent(cases))\n}", "func GetBaseURIParameters(config CONFIGURATION) map[string]interface{} {\r\n kvpMap := map[string]interface{}{\r\n }\r\n return kvpMap;\r\n}", "func globalBasePath(service *compute.Service, project string) string {\n\treturn fmt.Sprintf(\"%s%s/global/\", service.BasePath, project)\n}", "func main_init()", "func getReqNameId(reqIIdList []cres.IID, driverNameId string) string {\n\tfor _, iid := range reqIIdList {\n\t\tif iid.SystemId == driverNameId {\n\t\t\treturn iid.NameId\n\t\t}\n\t}\n\treturn \"\"\n}", "func (api *hostAPI) getFullKey(tenant, name string) string {\n\tif tenant != \"\" {\n\t\treturn fmt.Sprint(globals.ConfigRootPrefix, \"/\", \"cluster\", \"/\", \"hosts\", \"/\", tenant, \"/\", name)\n\t}\n\treturn fmt.Sprint(globals.ConfigRootPrefix, \"/\", \"cluster\", \"/\", \"hosts\", \"/\", name)\n}", "func (api *objectAPI) getFullKey(tenant, name string) string {\n\tif tenant != \"\" {\n\t\treturn fmt.Sprint(globals.ConfigRootPrefix, \"/\", \"objstore\", \"/\", \"objects\", \"/\", tenant, \"/\", name)\n\t}\n\treturn fmt.Sprint(globals.ConfigRootPrefix, \"/\", \"objstore\", \"/\", \"objects\", \"/\", name)\n}", "func (a *App) getLowLoadInstance(load map[string]int) string {\n\tvar ss string\n\tvar low = -1\n\tfor s, l := range load {\n\t\tif low < 0 || l < low {\n\t\t\tss = s\n\t\t\tlow = l\n\t\t}\n\t}\n\treturn ss\n}", "func getCallerInfo() string {\n\ts := os.Getenv(\"SC_CALLER_INFO\")\n\treturn s\n}", "func getCallerPosition() string {\n\t_, file, line, ok := runtime.Caller(2)\n\tif !ok {\n\t\treturn \"\"\n\t}\n\tfile = path.Base(file)\n\treturn fmt.Sprintf(\"%s%s:%d%s\", colorWhere, file, line, colorClear)\n}", "func getLocale(locale string) string {\n\tvalidLocales := [...]string{\"en\", \"ar\"}\n\n\tfor _, l := range validLocales {\n\t\tif l == locale {\n\t\t\treturn l\n\t\t}\n\t}\n\treturn \"en\"\n}", "func TestSpec_GetBaseType(t *testing.T) {\n\tcode := `\npackage test\nfunc main() {\n var a int\n\tb := &a\n\t_ = b\n}`\n\ts := NewSpec(code)\n\ts.SearchKind = SearchAll\n\t//fmt.Println(s.GetBaseType(\"b\"))\n\tif s.GetBaseType(\"a\") == nil && s.GetBaseType(\"b\").String() == \"int\" {\n\t} else {\n\t\tt.Error(`test failed`)\n\t}\n}" ]
[ "0.47383794", "0.4736673", "0.47247332", "0.47122508", "0.46919596", "0.4691863", "0.46611983", "0.46438965", "0.4642302", "0.46224102", "0.4615945", "0.458921", "0.45803776", "0.45387766", "0.45320305", "0.4528254", "0.45200878", "0.4505518", "0.45018515", "0.44895494", "0.44842383", "0.44753283", "0.44689342", "0.44688687", "0.4468828", "0.44667143", "0.44616205", "0.44589883", "0.4458764", "0.44525284", "0.44521746", "0.4449987", "0.4449987", "0.44483295", "0.44343552", "0.4431844", "0.44312358", "0.44308224", "0.4429933", "0.44292128", "0.44163102", "0.44127855", "0.44057068", "0.43947047", "0.43947047", "0.43947047", "0.43947047", "0.43796027", "0.4377032", "0.4373182", "0.43723178", "0.43703458", "0.43702427", "0.4368387", "0.43683007", "0.4366094", "0.4365061", "0.43636274", "0.43605208", "0.43586275", "0.43549818", "0.43546376", "0.43528432", "0.43516254", "0.43515244", "0.43509454", "0.43494377", "0.43494284", "0.4341585", "0.4339438", "0.43382257", "0.43366623", "0.43364972", "0.43351886", "0.43328962", "0.43326148", "0.43323776", "0.43255574", "0.43255574", "0.43255574", "0.43255574", "0.43255574", "0.43255574", "0.43141043", "0.43129396", "0.4311154", "0.43095943", "0.4308182", "0.4307295", "0.430458", "0.43028623", "0.4299216", "0.42990014", "0.42981675", "0.4296555", "0.42910454", "0.42894232", "0.42876133", "0.4285661", "0.4285334", "0.4282888" ]
0.0
-1
Returns if stopCh is closed or Send() returns an error.
func (a *datadogAdapter) Start(ctx context.Context) error { a.logger.Info("Starting Datadog adapter") return a.ceClient.StartReceiver(ctx, a.dispatch) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *Stopper) IsStopped() <-chan struct{} {\n\tif s == nil {\n\t\treturn nil\n\t}\n\treturn s.stopped\n}", "func (s *Stopper) IsStopped() <-chan struct{} {\n\tif s == nil {\n\t\treturn nil\n\t}\n\treturn s.stopped\n}", "func (s *Server) Stopped() <-chan struct{} {\n\treturn s.stopCh\n}", "func (s *OpenHackSimulator) Stop() chan bool {\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tglog.V(0).Infof(\"recovered from panic while stopping closed channel: %v\", r)\n\t\t}\n\t}()\n\tclose(s.stopCh)\n\treturn s.done\n}", "func (s *Stopper) ShouldStop() <-chan struct{} {\n\tif s == nil {\n\t\t// A nil stopper will never signal ShouldStop, but will also never panic.\n\t\treturn nil\n\t}\n\treturn s.stopper\n}", "func (s *Stopper) ShouldStop() <-chan struct{} {\n\tif s == nil {\n\t\t// A nil stopper will never signal ShouldStop, but will also never panic.\n\t\treturn nil\n\t}\n\treturn s.stopper\n}", "func (w *WSConn) IsClose(err error) bool {\n\treturn WSIsClose(err)\n}", "func (s *Sender) Stop() {\n\tclose(s.inputChan)\n\t<-s.done\n}", "func (s *SocketClient) Stop() {\n\ts.stopCh <- 0\n}", "func (c *Check) Stop() { close(c.stopCh) }", "func IsErrClosing(err error) bool {\n\tif opErr, ok := err.(*net.OpError); ok {\n\t\terr = opErr.Err\n\t}\n\treturn \"use of closed network connection\" == err.Error()\n}", "func (s *Socket) HasBeenClosed() bool {\n\treturn s.hasBeenClosed\n}", "func (self Channel) Stop() error {\n\treturn boolToError(C.BASS_ChannelStop(self.cint()))\n}", "func (socket *Socket) IsClosed() bool {\n\tsocket.Lock()\n\tdefer socket.Unlock()\n\treturn socket.closed\n}", "func (this *Reporter) IsStop() bool {\n\treturn this.Status == REPORT_STATUS_STOPED\n}", "func (w *Watcher) sendError(err error) bool {\n\tselect {\n\tcase w.Errors <- err:\n\t\treturn true\n\tcase <-w.quit:\n\t}\n\treturn false\n}", "func (t *ipTransport) Stop() <-chan struct{} {\n\tt.cancel()\n\n\treturn t.stopped\n}", "func (h *Handler) Stopped() chan struct{} {\n\treturn h.handler.stopped\n}", "func (r *RecordStream) Closed() bool { return r.state == closed || r.state == serverLost }", "func isErrClosing(err error) bool {\n\tif opErr, ok := err.(*net.OpError); ok {\n\t\terr = opErr.Err\n\t}\n\treturn \"use of closed network connection\" == err.Error()\n}", "func (s *Sender) Stop() {\n\tif !s.running {\n\t\treturn\n\t}\n\n\tclose(s.quit)\n}", "func (client *Client) Closed() bool {\n\tselect {\n\tcase <-client.done:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (ce ConnectionEvent) IsClose() bool {\n\treturn ce == LocalClose || ce == RemoteClose ||\n\t\tce == OnReadErrClose || ce == OnWriteErrClose || ce == OnWriteTimeout\n}", "func (w *worker) stop() bool {\n\tif w.closedReqChan {\n\t\treturn false\n\t}\n\tclose (w.reqs)\n\tw.closedReqChan = true\n\t_ = <-w.joinChan\n\tclose(w.joinChan)\n\treturn true\n}", "func (d *D) stop() {\n\tclose(d.stopCh)\n}", "func (str *Stream) Stop() error {\n\treturn str.close(true)\n}", "func (w *Watcher) sendError(err error) bool {\n\tselect {\n\tcase w.Errors <- err:\n\t\treturn true\n\tcase <-w.done:\n\t\treturn false\n\t}\n}", "func isChanClosed(ch chan struct{}) bool {\n\tselect {\n\tcase <-ch:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (cli *Client) Closed() bool {\n\treturn cli.State() == Closed\n}", "func (q *MyQueue) IsClose() bool {\n\treturn q.closed\n}", "func (o *ChannelOutbound) Stop() error {\n\treturn o.once.Stop(o.stop)\n}", "func (c *StreamerController) IsClosed() bool {\n\tc.RLock()\n\tdefer c.RUnlock()\n\n\treturn c.closed\n}", "func (rr *RtcpReceiver) Close() {\n\trr.events <- rtcpReceiverEventTerminate{}\n\t<-rr.done\n}", "func (rs *RequestSender) Close() bool {\n\treturn rs.close\n}", "func (s *Server) closing() <-chan bool {\n\ts.Lock()\n\tdefer s.Unlock()\n\treturn s.getClosing()\n}", "func IsClosed() bool {\n\treturn server.IsClosed()\n}", "func (s *Server) IsClosed() bool {\n\treturn s.listener.IsClosed()\n}", "func (c *conn) Closed() bool {\n\treturn c.closed\n}", "func (client *SyncClient) IsClosed() bool {\n\treturn client.cClient == nil\n}", "func (s socket) Close() error {\n\ts.done <- true\n\treturn nil\n}", "func (l *DockerListener) Stopped() chan struct{} {\n\treturn l.stopped\n}", "func (ch *Channel) IsClosed() bool {\n\treturn atomic.LoadInt32(&ch.closed) == 1\n}", "func (a API) StopChk() (isNew bool) {\n\tselect {\n\tcase o := <-a.Ch.(chan StopRes):\n\t\tif o.Err != nil {\n\t\t\ta.Result = o.Err\n\t\t} else {\n\t\t\ta.Result = o.Res\n\t\t}\n\t\tisNew = true\n\tdefault:\n\t}\n\treturn\n}", "func isClosedErr(err error) bool {\n\treturn strings.HasSuffix(err.Error(), \"use of closed network connection\") // YUCK!\n}", "func (ws WaitStatus) Stopped() bool {\n\treturn ws&0xff == 0x7f\n}", "func (h *Hosts) Stopped() bool {\n\tselect {\n\tcase <-h.stopped:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (w *Watcher) Close() error {\n\tclose(w.commCh)\n\t<-w.closed\n\treturn nil\n}", "func (s *Sync) IsInterrupted() bool {\n\tselect {\n\tcase <-s.grp.Ch():\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (p *TCPProxy) Stop() error {\n\tatomic.StoreInt32(&p.closed, 1)\n\treturn p.listener.Close()\n}", "func (lc *Closer) HasBeenClosed() <-chan struct{} {\n\treturn lc.ctx.Done()\n}", "func (t *channelTransport) stop() {\n\tt.stopChan <- struct{}{}\n}", "func (c Chan) TryReceive() bool {\n\tselect {\n\tcase <-c:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (c *WaitCounter) Stopped() bool {\n\treturn atomic.LoadUint64(&c.Value) == math.MaxUint64\n}", "func (c *minecraftConn) Closed() bool {\n\tselect {\n\tcase <-c.closed:\n\t\treturn true\n\tdefault:\n\t\treturn false\n\t}\n}", "func (client *NpmClient) isStopped() bool {\n\tclient.Lock()\n\tdefer client.Unlock()\n\treturn client.stopped\n}", "func (m *MajsoulChannel) Close(err error) {\n\tm.mutexChannel.Lock()\n\tif m.isopen {\n\t\tm.exitValue = err\n\t\tm.isopen = false\n\t\tclose(m.stop)\n\t\tlog.Println(\"Majsoul channel closed.\")\n\t}\n\tm.mutexChannel.Unlock()\n}", "func (ctl *Control) ClosedDoneCh() <-chan struct{} {\n\treturn ctl.closedDoneCh\n}", "func (c *Channel) Close() error {\n\treturn c.exit(false)\n}", "func (c *client) Closed() <-chan struct{} {\n\treturn c.closeFlag\n}", "func (p *Poller) IsStopped() bool {\n\treturn p.isStopped\n}", "func IsSocketClosedError(e error) bool {\n\treturn strings.HasSuffix(e.Error(), \"use of closed network connection\")\n}", "func (p *pipe) Close() error {\n\tmsg := msg{\n\t\tdata: nil,\n\t\terr: io.EOF, // Indicates to the other side's Read function that session has ended.\n\t}\n\n\tselect {\n\tcase p.otherEnd.ch <- msg: // Sent msg to the other side's Read function.\n\t\treturn nil\n\tcase <-p.ctx.Done():\n\t\treturn p.ctx.Err()\n\t}\n}", "func (cd *CloseDetector) IsClosed() bool {\n\treturn cd.closed\n}", "func (s *T) Stop() {\n\tclose(s.stopCh)\n\ts.wg.Wait()\n}", "func chanClose(ch *channel) {\n\tif ch == nil {\n\t\t// Not allowed by the language spec.\n\t\truntimePanic(\"close of nil channel\")\n\t}\n\tswitch ch.state {\n\tcase chanStateClosed:\n\t\t// Not allowed by the language spec.\n\t\truntimePanic(\"close of closed channel\")\n\tcase chanStateSend:\n\t\t// This panic should ideally on the sending side, not in this goroutine.\n\t\t// But when a goroutine tries to send while the channel is being closed,\n\t\t// that is clearly invalid: the send should have been completed already\n\t\t// before the close.\n\t\truntimePanic(\"close channel during send\")\n\tcase chanStateRecv:\n\t\t// unblock all receivers with the zero value\n\t\tch.state = chanStateClosed\n\t\tfor ch.blocked != nil {\n\t\t\tch.resumeRX(false)\n\t\t}\n\tcase chanStateEmpty, chanStateBuf:\n\t\t// Easy case. No available sender or receiver.\n\t}\n\tch.state = chanStateClosed\n\tchanDebug(ch)\n}", "func (s *Server) getClosing() chan bool {\n\tif s.Closing == nil {\n\t\ts.Closing = make(chan bool)\n\t}\n\treturn s.Closing\n}", "func (l *remoteLoggingClient) Close() error {\n\tl.stream.CloseSend()\n\t<-l.doneCh\n\tif l.runErr != nil {\n\t\treturn errors.Wrap(l.runErr, \"remote logging background routine failed\")\n\t}\n\treturn nil\n}", "func (p *Poller) Stop() {\n\tp.stopMutex.Lock()\n\tdefer p.stopMutex.Unlock()\n\n\tp.isStopped = true\n\tclose(p.Channel)\n}", "func (consumer *Consumer) Closed() bool {\n\treturn consumer.closed\n}", "func (p *Processor) Stop() {\n\tc := p.stopChan\n\tp.stopChan = nil\n\tc <- struct{}{}\n\tp.Conn.Close()\n}", "func (dev *Device) Close() bool {\n\tdev.Ok = false\n\tclose(dev.chQuit)\n\tif err := dev.port.Close(); err != nil {\n\t\tlog.Printf(\"close err: %s\", err)\n\t\treturn false\n\t}\n\treturn true\n}", "func IsNetworkClosed(err error) bool {\n\topErr, ok := err.(*net.OpError)\n\tif ok {\n\t\treturn opErr.Err == io.EOF\n\t}\n\n\treturn err == io.EOF\n}", "func (eis *eventSocket) stop() error {\n\teis.log.Info(\"closing Chain IPC\")\n\terrs := wrappers.Errs{}\n\terrs.Add(eis.unregisterFn(), eis.socket.Close())\n\treturn errs.Err\n}", "func (s *Stopper) Stop() {\n\tclose(s.shouldStopC)\n\ts.wg.Wait()\n}", "func (o *BinReaderOption) Stop(header *BinEventHeader) bool {\n\tif o == nil {\n\t\treturn false\n\t} else if o.EndPos != 0 && o.EndPos < header.LogPos {\n\t\treturn true\n\t} else if !o.EndTime.IsZero() && o.EndTime.Unix() <= time.Unix(header.Timestamp, 0).Unix() {\n\t\treturn true\n\t}\n\treturn false\n}", "func (s *EchoServer) RequestStop() bool {\n\ts.mut.Lock()\n\tdefer s.mut.Unlock()\n\n\tselect {\n\tcase <-s.reqStop:\n\t\treturn false\n\tdefault:\n\t\tclose(s.reqStop)\n\t\treturn true\n\t}\n}", "func (l *Listener) waitForRecvOrStop(recv chan Message) bool {\r\n\tfor {\r\n\t\tselect {\r\n\t\tcase <-l.stop:\r\n\t\t\treturn true\r\n\t\tcase msg := <-recv:\r\n\t\t\tgo func() { l.handler.HandleMessage(msg) }()\r\n\t\t\treturn false\r\n\t\t}\r\n\t}\r\n}", "func IsErrNetClosing(err error) bool {\n\tif err == nil {\n\t\treturn false\n\t}\n\treturn strings.Contains(err.Error(), useOfClosedErrMsg)\n}", "func (r *Receiver) Close() error { return nil }", "func (s *Streamer) Stop() {\n\tclose(s.stopc)\n}", "func (c *Client) Close(_ context.Context) error {\n\t// Cancel the background goroutines,\n\t// and wait for them to finish\n\t// before closing the socket.\n\tc.cancel()\n\tpollError := <-c.pollError\n\tpingError := <-c.pingError\n\tcloseError := c.webSock.Close()\n\n\tswitch {\n\tcase closeError != nil:\n\t\treturn closeError\n\tcase pollError != nil:\n\t\treturn pollError\n\tcase pingError != nil:\n\t\treturn pingError\n\tdefault:\n\t\treturn nil\n\t}\n}", "func (sw *streamWatcher) stopping() bool {\n\tsw.Lock()\n\tdefer sw.Unlock()\n\treturn sw.stopped\n}", "func (e *Exporter) Stop() error {\n\te.mu.RLock()\n\tcc := e.grpcClientConn\n\tstarted := e.started\n\te.mu.RUnlock()\n\n\tif !started {\n\t\treturn nil\n\t}\n\n\t// Now close the underlying gRPC connection.\n\tvar err error\n\tif cc != nil {\n\t\terr = cc.Close()\n\t}\n\n\t// At this point we can change the state variable started\n\te.mu.Lock()\n\te.started = false\n\te.mu.Unlock()\n\tclose(e.stopCh)\n\n\t// Ensure that the backgroundConnector returns\n\t<-e.backgroundConnectionDoneCh\n\n\treturn err\n}", "func (c *MsgConnection) Closed() bool {\n\tc.cmtx.Lock()\n\tdefer c.cmtx.Unlock()\n\treturn c.closed\n}", "func (folderWatcher *FolderWatcher) Stopped() chan bool {\n\treturn folderWatcher.stopped\n}", "func (a *connectionMock) Closed() bool {\n\treturn a.closed\n}", "func (s *Socket) Stop() {\n\tlog.Error(\"Closing socket\")\n\tif s == nil {\n\t\treturn\n\t}\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\tfor cmd := range s.cmds.forEach() {\n\t\tclose(cmd.done)\n\t}\n\ts.con.Close()\n\tif s.state == open {\n\t\ts.state = closed\n\t\tclose(s.done)\n\t}\n}", "func (obj *ChannelMessageTracer) Stop() {\n\t//logger.Log(fmt.Sprint(\"Entering ChannelMessageTracer:Stop ...\"))\n\tif obj.isRunning {\n\t\t// Finish / Flush any remaining processing\n\t\ttraceFileWithSuffix := fmt.Sprintf(\"%s.%d\", obj.traceFileName, obj.currentSuffix)\n\t\tflag, _, _ := exists(traceFileWithSuffix)\n\t\tif flag {\n\t\t\t_ = obj.traceFile.Sync() // Flush\n\t\t\t_ = obj.traceFile.Close() // Close FD\n\t\t\tobj.traceFile = nil\n\t\t}\n\t\tobj.isRunning = false\n\t}\n\t//logger.Log(fmt.Sprint(\"Returning ChannelMessageTracer:Stop ...\"))\n}", "func (s *SendError) IsTerminallyUnderpriced() bool {\n\treturn s.is(TerminallyUnderpriced)\n}", "func Stop() {\n\tExitChannel <- true\n}", "func (_BondedECDSAKeep *BondedECDSAKeepCallerSession) IsClosed() (bool, error) {\n\treturn _BondedECDSAKeep.Contract.IsClosed(&_BondedECDSAKeep.CallOpts)\n}", "func (a *Attacker) Stop() {\n\tselect {\n\tcase <-a.stopch:\n\t\treturn\n\tdefault:\n\t\tclose(a.stopch)\n\t}\n}", "func (device *Device) isClosed() bool {\n\treturn device.deviceState() == deviceStateClosed\n}", "func (cb *CanBusClient) Stop(_ error) {\n\tclose(cb.stop)\n}", "func (d *D) Stop() error {\n\tif d.cancel == nil {\n\t\treturn fmt.Errorf(\"cannot call Stop() before Start()\")\n\t}\n\n\t// Call cancel received from Start()\n\td.cancel()\n\n\t// Close channels, clients are expected to drain and check for zero value\n\tclose(d.psChan)\n\n\treturn nil\n}", "func (o *Outbound) Stop() error {\n\treturn o.once.Stop(o.chooser.Stop)\n}", "func (w *Whisper) Stop() error {\n\tclose(w.quit)\n\tlog.Info(\"whisper stopped\")\n\treturn nil\n}", "func (s *Stream) Close() error {\n\tcloseStream := false\n\ts.stateLock.Lock()\n\tswitch s.state {\n\t// Opened means we need to signal a close\n\tcase streamSYNSent:\n\t\tfallthrough\n\tcase streamSYNReceived:\n\t\tfallthrough\n\tcase streamEstablished:\n\t\ts.state = streamLocalClose\n\t\tgoto SEND_CLOSE\n\n\tcase streamLocalClose:\n\tcase streamRemoteClose:\n\t\ts.state = streamClosed\n\t\tcloseStream = true\n\t\tgoto SEND_CLOSE\n\n\tcase streamClosed:\n\tcase streamReset:\n\tdefault:\n\t\tpanic(\"unhandled state\")\n\t}\n\ts.stateLock.Unlock()\n\treturn nil\nSEND_CLOSE:\n\t// This shouldn't happen (the more realistic scenario to cancel the\n\t// timer is via processFlags) but just in case this ever happens, we\n\t// cancel the timer to prevent dangling timers.\n\tif s.closeTimer != nil {\n\t\ts.closeTimer.Stop()\n\t\ts.closeTimer = nil\n\t}\n\n\t// If we have a StreamCloseTimeout set we start the timeout timer.\n\t// We do this only if we're not already closing the stream since that\n\t// means this was a graceful close.\n\t//\n\t// This prevents memory leaks if one side (this side) closes and the\n\t// remote side poorly behaves and never responds with a FIN to complete\n\t// the close. After the specified timeout, we clean our resources up no\n\t// matter what.\n\tif !closeStream && s.session.config.StreamCloseTimeout > 0 {\n\t\ts.closeTimer = time.AfterFunc(\n\t\t\ts.session.config.StreamCloseTimeout, s.closeTimeout)\n\t}\n\n\ts.stateLock.Unlock()\n\ts.sendClose()\n\ts.notifyWaiting()\n\tif closeStream {\n\t\ts.session.closeStream(s.id)\n\t}\n\treturn nil\n}", "func (s *Socket) Close() {\n\ts.mux.Lock()\n\tif s.stop != nil {\n\t\ts.stop <- true\n\t\t<-s.stop\n\t\ts.stop = nil\n\t}\n\ts.mux.Unlock()\n}", "func SetStopCh(stopCh chan struct{}) {\n\tstopChannel = stopCh\n}", "func IsErrNetClosing(err error) bool {\n\t// No other way to check\n\treturn strings.Contains(err.Error(), \"use of closed network connection\")\n}" ]
[ "0.5975818", "0.5975818", "0.5821494", "0.5494627", "0.548023", "0.548023", "0.54433465", "0.5399817", "0.53671", "0.536415", "0.53559923", "0.53453416", "0.53445506", "0.5313858", "0.53045666", "0.53027296", "0.5285623", "0.5274369", "0.52731997", "0.5271154", "0.5244728", "0.52446926", "0.52439517", "0.52427006", "0.5222376", "0.5203817", "0.52036405", "0.5194937", "0.5185391", "0.5177517", "0.5175435", "0.5153948", "0.51527804", "0.5149036", "0.5138186", "0.51378036", "0.5134181", "0.5111791", "0.5108769", "0.5104346", "0.5096782", "0.50870395", "0.5083707", "0.5083213", "0.50782704", "0.50770074", "0.50700545", "0.5043912", "0.5034249", "0.5030917", "0.5030213", "0.5023161", "0.5022815", "0.5022775", "0.5018657", "0.5003591", "0.49993974", "0.49924773", "0.4980324", "0.49774936", "0.49668136", "0.4966264", "0.49557817", "0.495481", "0.4953654", "0.4951436", "0.49485683", "0.49452323", "0.49380428", "0.49344003", "0.492441", "0.49242017", "0.4914158", "0.49130136", "0.49113762", "0.49080452", "0.49042895", "0.48968953", "0.48885277", "0.48777783", "0.48758566", "0.4871603", "0.48696136", "0.48686528", "0.48685336", "0.48676842", "0.48667535", "0.48553336", "0.48416016", "0.48364258", "0.48344904", "0.48289275", "0.48262566", "0.48236012", "0.4821514", "0.48198897", "0.48180747", "0.48121697", "0.48096466", "0.4809609", "0.4806922" ]
0.0
-1
newAPIRequest returns a POST http.Request that is ready to send to the Datadog generalpurpose API.
func newAPIRequest(host, path, apiKey string, body []byte) (*http.Request, error) { return newAPIRequestWithHost(host, path, apiKey, body) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (a *APIv2) NewAPIRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.SetBasicAuth(\"x\", a.key)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"X-Gms-Client-Version\", config.VERSION)\n\treq.Header.Add(\"X-Gms-Revision\", utils.GetCurrentRevision())\n\treq.Header.Add(\"X-Gms-Branch\", utils.GetCurrentBranch())\n\treturn req, nil\n}", "func (qiwi *PersonalAPI) newRequest(apiKey, method, spath string, data map[string]interface{}) (req *http.Request, err error) {\n\n\tvar path = APIURL + spath\n\n\tvar body io.Reader\n\n\tif len(data) > 0 {\n\n\t\ts, err := json.Marshal(data)\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tbody = bytes.NewBuffer(s)\n\n\t}\n\n\treq, err = http.NewRequest(method, path, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Bearer \"+apiKey)\n\n\treturn req, err\n}", "func (g *GoFlickr) newRequest(apiMethod string) ApiRequest {\n\n\treq := ApiRequest{\n\t\tMethodName: apiMethod,\n\t}\n\treq.addParam(\"api_key\", g.ApiKey)\n\treq.addParam(\"format\", \"json\")\n\treq.addParam(\"nojsoncallback\", \"1\")\n\treturn req\n\n}", "func (c *Client) newRequest(ctx context.Context, method, url string, body io.Reader) (*http.Request, error) {\n\t// Build new request with base URL.\n\treq, err := http.NewRequest(method, c.URL+url, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Set API key in header.\n\tif user := wtf.UserFromContext(ctx); user != nil && user.APIKey != \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bearer \"+user.APIKey)\n\t}\n\n\t// Default to JSON format.\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-type\", \"application/json\")\n\n\treturn req, nil\n}", "func (c *Client) newRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"api-name\", c.apiName)\n\treq.Header.Set(\"api-key\", c.apiKey)\n\treq.Header.Set(c.userHeader, c.user)\n\treturn req, nil\n}", "func newLogsAPIRequest(host, path, apiKey string, body []byte) (*http.Request, error) {\n\treturn newAPIRequestWithHost(host, path, apiKey, body)\n}", "func (c client) newRequest(ctx context.Context, method string, url string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequestWithContext(ctx, method, url, body)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to build request %w\", err)\n\t}\n\n\treturn req, nil\n}", "func (g *Group) newRequest(operation *request.Operation, input interface{}, output interface{}, handlers *request.Handlers) *request.Request {\n return request.New(g.Client, operation, input, output, handlers)\n}", "func (c *APIGateway) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func newRequest(method, url string, body string) *http.Request {\n\treq, err := http.NewRequest(method, url, strings.NewReader(body))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treq.Header.Set(\"X-API-Token\", \"token1\")\n\treturn req\n}", "func (s *Stein) newRequest(method string, path string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, path, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\treturn req, nil\n}", "func (c *apiClient) newRequest(ctx context.Context, method, resourcePath string,\n\tquery url.Values, body io.Reader) (*http.Request, error) {\n\tURL, err := url.Parse(c.BaseURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// BaseURL and resource URL is joined if they have a path\n\tURL.Path = c.joinURLPath(URL.Path, resourcePath)\n\tif query != nil {\n\t\tURL.RawQuery = query.Encode()\n\t}\n\trequest, err := http.NewRequestWithContext(ctx, method, URL.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trequest.Host = c.Host // allow cloudfronting\n\tif c.Authorization != \"\" {\n\t\trequest.Header.Set(\"Authorization\", c.Authorization)\n\t}\n\tif c.Accept != \"\" {\n\t\trequest.Header.Set(\"Accept\", c.Accept)\n\t}\n\trequest.Header.Set(\"User-Agent\", c.UserAgent)\n\treturn request, nil\n}", "func (c *InputService17ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (s *OidcService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (conn Connection) newRequest(method, cmd string, body io.Reader) *http.Request {\n\t// req, err := conn.jhReq(method, cmd, body)\n\treq, err := http.NewRequest(method, conn.HubURL+cmd, body)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Coulnd't generate HTTP request - %s\\n\", err.Error()))\n\t}\n\n\treq.Header.Add(\"Authorization\", fmt.Sprintf(\"token %s\", conn.Token))\n\n\treturn req\n}", "func (c *Client) NewRequest(method, apiPath string, body interface{}) (*http.Request, error) {\n loc, err := joinPaths(c.BaseURL, apiPath)\n if err != nil {\n return nil, err\n }\n\n buf := new(bytes.Buffer)\n if body != nil {\n err = json.NewEncoder(buf).Encode(body)\n if err != nil {\n return nil, err\n }\n }\n\n req, err := http.NewRequest(method, loc.String(), buf)\n if err != nil {\n return nil, err\n }\n\n req.Header.Add(\"Content-Type\", \"application/json\")\n req.Header.Add(\"Accept\", \"application/json, text/plain, */*\")\n req.Header.Add(\"User-Agent\", c.UserAgent)\n return req, nil\n}", "func (c *InputService14ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *Client) newRequest(method, path string, v interface{}, ctype string) (req *http.Request, err error) {\n\t// Build request JSON.\n\tbody, err := writeJson(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq, err = http.NewRequest(method, c.pathToEndPoint(path), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Add(\"X-Kii-AppID\", c.AppId)\n\treq.Header.Add(\"X-Kii-AppKey\", c.AppKey)\n\tif ctype != \"\" {\n\t\treq.Header.Add(\"Content-Type\", ctype)\n\t}\n\tif c.Authorization != \"\" {\n\t\treq.Header.Add(\"Authorization\", c.Authorization)\n\t}\n\treturn\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v7.1.0\")\n\treturn req, nil\n}", "func (adm AdminClient) newRequest(method string, reqData requestData) (req *http.Request, err error) {\n\t// If no method is supplied default to 'POST'.\n\tif method == \"\" {\n\t\tmethod = \"POST\"\n\t}\n\n\t// Default all requests to \"\"\n\tlocation := \"\"\n\n\t// Construct a new target URL.\n\ttargetURL, err := adm.makeTargetURL(reqData)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Initialize a new HTTP request for the method.\n\treq, err = http.NewRequest(method, targetURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tadm.setUserAgent(req)\n\tfor k, v := range reqData.customHeaders {\n\t\treq.Header.Set(k, v[0])\n\t}\n\tif length := len(reqData.content); length > 0 {\n\t\treq.ContentLength = int64(length)\n\t}\n\treq.Header.Set(\"X-Amz-Content-Sha256\", hex.EncodeToString(sum256(reqData.content)))\n\treq.Body = ioutil.NopCloser(bytes.NewReader(reqData.content))\n\n\treq = s3signer.SignV4(*req, adm.accessKeyID, adm.secretAccessKey, \"\", location)\n\treturn req, nil\n}", "func (c *IoT) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func (r *Requester) newRequest(endpoint string) (*http.Request, error) {\n req, err := http.NewRequest(\"GET\", endpoint, nil)\n if err != nil {\n return nil, err\n }\n\tbearer := fmt.Sprintf(\"Bearer %s\", r.bearer)\n req.Header.Add(\"Authorization\", bearer)\n\treq.Header.Add(\"Ocp-Apim-Subscription-Key\", apimKey)\n req.Header.Set(\"User-Agent\", \"hackacraic\")\n\treturn req, nil\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v5.4.1\")\n\treturn req, nil\n}", "func (g *baseGithub) newRequest(method, url string, body interface{}) (*http.Request, error) {\n\tvar err error\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\tenc := json.NewEncoder(buf)\n\t\tenc.SetEscapeHTML(false)\n\t\terr = enc.Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, url, buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\treq.Header.Set(\"Accept\", \"application/json\")\n\n\treturn req, nil\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v8.1.0\")\n\treturn req, nil\n}", "func newRequest(req *http.Request) *Request {\n\trequest := &Request{\n\t\tRequest: req,\n\t}\n\n\treturn request\n}", "func (c *InputService2ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService2ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService4ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService4ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService18ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (k *KeKahu) newRequest(method, endpoint string, body io.Reader) (*http.Request, error) {\n\n\t// Parse the endpoint\n\tep, err := url.Parse(endpoint)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not parse endpoint: %s\", err)\n\t}\n\n\t// Resolve the URL reference\n\tbaseURL, err := k.config.GetURL()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\turl := baseURL.ResolveReference(ep)\n\n\t// Construct the request\n\treq, err := http.NewRequest(method, url.String(), body)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not create request: %s\", err)\n\t}\n\n\t// Add the headers\n\treq.Header.Set(\"Authorization\", fmt.Sprintf(\"Bearer %s\", k.config.APIKey))\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Accept\", \"application/json\")\n\n\ttrace(\"created %s request to %s\", method, url)\n\treturn req, nil\n}", "func (c *InputService9ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService9ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService20ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService16ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (s *ApplicationsService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService1ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService1ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func newRequest(r *http.Request, vars map[string]string) *Request {\n\trequest := &Request{r, vars, nil}\n\treturn request\n}", "func (c *InputService15ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (k *API) NewRequest(url, method string, body url.Values) (*http.Request, error) {\n\n\tb := bytes.NewBufferString(body.Encode())\n\treq, err := http.NewRequest(method, url, b)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"make korbit req\")\n\t}\n\n\ttoken := fmt.Sprintf(\"%s %s\", k.Token.TokenType, k.Token.AccessToken)\n\treq.Header.Set(\"Authorization\", token)\n\n\tif method == \"POST\" {\n\t\treq.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\t}\n\n\treturn req, nil\n}", "func (c *InputService8ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService8ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (s *SitesService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService5ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService5ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func NewRequest() *Request {\n\treturn defaul.NewRequest()\n}", "func (c *InputService7ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService7ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService19ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService6ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService6ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService10ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func newRequest(w http.ResponseWriter, r *http.Request, app *Webapp) *Request {\n\treq := &Request{\n\t\tw: w,\n\t\tr: r,\n\t\tapp: app,\n\t\tstatus: 200,\n\t\tcontentLength: 0,\n\t\tcontentType: \"text/html; charset=utf-8\",\n\t\tdate: time.Now(),\n\t\treplied: false,\n\t}\n\treturn req\n}", "func (b *AuroraBackend) NewRequest(params *CallParams) (*http.Request, error) {\n\treq, err := http.NewRequest(params.Method, params.Path, params.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\theaders := params.Headers\n\tif headers == nil {\n\t\theaders = make(http.Header)\n\t}\n\n\tif params.Credentials != nil {\n\t\theaders.Add(\"X-Application-ID\", params.Credentials.AppID)\n\t\theaders.Add(\"X-Application-Token\", params.Credentials.AppToken)\n\t\theaders.Add(\"X-Device-ID\", params.Credentials.DeviceID)\n\t}\n\n\treq.Header = headers\n\treturn req, nil\n}", "func newRequest(db *Database) *Request {\n\treq := &Request{\n\t\tdb: db,\n\t\tpath: \"/\",\n\t\tquery: url.Values{},\n\t\theader: http.Header{},\n\t}\n\treturn req\n}", "func (c *InputService3ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService3ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (s *HighAvailabilityService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService13ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func newRequest(req *http.Request) *Request {\n\treturn &Request{\n\t\tRequest: req,\n\t}\n}", "func (c *InputService21ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (s APIv1) NewRequest(ctx context.Context, method, path string, data interface{}) (req *http.Request, err error) {\n\t// Resolve the URL reference from the path\n\tendpoint := s.endpoint.ResolveReference(&url.URL{Path: path})\n\n\tvar body io.ReadWriter\n\tif data != nil {\n\t\tbody = &bytes.Buffer{}\n\t\tif err = json.NewEncoder(body).Encode(data); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not serialize request data: %s\", err)\n\t\t}\n\t} else {\n\t\tbody = nil\n\t}\n\n\t// Create the http request\n\tif req, err = http.NewRequestWithContext(ctx, method, endpoint.String(), body); err != nil {\n\t\treturn nil, fmt.Errorf(\"could not create request: %s\", err)\n\t}\n\n\t// Set the headers on the request\n\treq.Header.Add(\"User-Agent\", \"Whisper/1.0\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Add(\"Accept-Language\", \"en-US,en\")\n\treq.Header.Add(\"Accept-Encoding\", \"gzip, deflate, br\")\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\n\treturn req, nil\n}", "func (c *InputService12ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService14ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func NewRequest(requestName string, params rata.Params, header http.Header, query url.Values, body ...io.Reader) Request {\n\tif header == nil {\n\t\theader = http.Header{}\n\t}\n\theader.Set(\"Accept\", \"application/json\")\n\n\trequest := Request{\n\t\tRequestName: requestName,\n\t\tParams: params,\n\t\tHeader: header,\n\t\tQuery: query,\n\t}\n\n\tif len(body) == 1 {\n\t\trequest.Body = body[0]\n\t}\n\n\treturn request\n}", "func (rt *registryRoundTripper) NewRequest(method, path string,\n\tquery *url.Values, body interface{}) (*http.Request, error) {\n\n\treq, err := rt.DefaultRequestDoer.NewRequest(method, path, query, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif tok := rt.holder.Token(); tok != \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bearer \"+tok)\n\t}\n\n\treq.Header.Set(\"User-Agent\", \"Torus-Daemon/\"+rt.version)\n\treq.Header.Set(\"X-Registry-Version\", rt.apiVersion)\n\n\treturn req, nil\n}", "func (c *InputService2ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *Client) newRequest(url string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"user-key\", c.key)\n\treq.Header.Add(\"Accept\", \"application/json\")\n\n\treturn req, nil\n}", "func NewRequest(token string) *Request {\n\treqID := fmt.Sprintf(\"alaudacli-%d\", time.Now().Unix())\n\trestyReq := resty.R()\n\trestyReq.SetHeader(\"Content-Type\", \"application/json\")\n\trestyReq.SetHeader(\"Alauda-Request-ID\", reqID)\n\n\tif token != \"\" {\n\t\trestyReq.SetHeader(\"Authorization\", fmt.Sprintf(\"Token %s\", token))\n\t}\n\n\treturn &Request{\n\t\trestyReq,\n\t}\n}", "func CreateBaseRequest(method, url string, body []byte, user m.AuthUser, verbose bool) *http.Request {\n\treq, err := http.NewRequest(method, url, bytes.NewBuffer(body))\n\treq.SetBasicAuth(user.Username, user.Password)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Accept\", \"application/json\")\n\tError(err, \"Error creating the request\")\n\n\tif verbose {\n\t\tfmt.Println(\"Request Url:\", req.URL)\n\t\tfmt.Println(\"Request Headers:\", req.Header)\n\t\tfmt.Println(\"Request Body:\", req.Body)\n\t}\n\n\treturn req\n}", "func (c *InputService9ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService4ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *apiHttp) NewRequest(command string) RequestBuilder {\n\theaders := make(map[string]string)\n\tif c.Headers != nil {\n\t\tfor k := range c.Headers {\n\t\t\theaders[k] = c.Headers.Get(k)\n\t\t}\n\t}\n\n\treturn &requestBuilder{\n\t\tcommand: command,\n\t\tclient: c,\n\t\theaders: headers,\n\t}\n}", "func (c *OutputService4ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService4ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService1ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService2ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService2ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (client *Client) NewRequest(messageDataBuffer *strings.Reader) (*http.Request, error) {\n\trequest, err := http.NewRequest(\"POST\", client.BaseURL, messageDataBuffer)\n\n\tif err != nil {\n\t\terrStr := fmt.Sprintf(\"Error constructing the HTTP network request ... here is the error %v\", err)\n\t\treturn &http.Request{}, &errorString{errStr}\n\t}\n\n\t// fmt.Printf(\"ACCOUNT SID \", client.AccountSID , \"Auth Token \", client.AuthToken)\n\trequest.SetBasicAuth(client.AccountSID, client.AuthToken) // Authenticating user credentials\n\n\t// Additional header fields to accept json media types which can be used for the response\n\trequest.Header.Add(\"Accept\", \"application/json\")\n\n\t// To indicate the media type that is being sent through the request\n\trequest.Header.Add(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\treturn request, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\trel, err := url.Parse(path.Join(apiVersion, urlStr))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.baseURL.ResolveReference(rel)\n\treq, err := http.NewRequest(method, u.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Authorization\", \"Bearer \"+c.config.accessToken)\n\tif c.userAgent != \"\" {\n\t\treq.Header.Set(\"User-Agent\", c.userAgent)\n\t}\n\treturn req, nil\n}", "func (c *InputService11ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *Client) newRequest(subdomain, method string, endpoint string, params Params, body interface{}) (*http.Request, error) {\n\n\tif subdomain == \"\" {\n\t\tsubdomain = \"www\"\n\t}\n\n\trequestURL := c.baseURL + \"/spi/v3/{endpoint}\"\n\trequestURL = strings.Replace(requestURL, \"{subdomain}\", subdomain, -1)\n\trequestURL = strings.Replace(requestURL, \"{domain}\", c.domain, -1)\n\trequestURL = strings.Replace(requestURL, \"{endpoint}\", endpoint, -1)\n\n\t// Query String\n\tqs := url.Values{}\n\tfor k, v := range params {\n\t\tqs.Add(k, fmt.Sprintf(\"%v\", v))\n\t}\n\n\tif len(qs) > 0 {\n\t\trequestURL += \"?\" + qs.Encode()\n\t}\n\n\treturn c.newRequestFromURL(requestURL, method, body)\n}", "func (client *Client) newRequest(method string, id string) (*http.Request, os.Error) {\n\trequest := new(http.Request)\n\tvar err os.Error\n\n\trequest.ProtoMajor = 1\n\trequest.ProtoMinor = 1\n\trequest.TransferEncoding = []string{\"chunked\"}\n\n\trequest.Method = method\n\n\t// Generate Resource-URI and parse it\n\turl := client.resource.String() + id\n\tif request.URL, err = http.ParseURL(url); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn request, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body io.Reader) (r *http.Request, err error) {\n\tif r, err = http.NewRequest(method, urlStr, body); err != nil {\n\t\treturn\n\t}\n\n\tr.Header.Set(\"User-Agent\", c.UserAgent)\n\tr.Header.Set(\"X-Requested-With\", \"XMLHttpRequest\")\n\n\tif c.ApiKey != nil {\n\t\tr.Header.Set(\"X-Sbss-Auth\", c.ApiKey.Login)\n\t\tr.AddCookie(c.ApiKey.Cookie)\n\t}\n\n\tif method == \"POST\" {\n\t\tr.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\t}\n\n\treturn\n}", "func (c *InputService22ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService7ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func NewRequest(method, url string, payload interface{}) (*http.Request, error) {\n\tvar buf io.Reader\n\tif payload != nil {\n\t\tvar b []byte\n\t\tb, err := json.Marshal(&payload)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbuf = bytes.NewBuffer(b)\n\t}\n\treturn http.NewRequest(method, url, buf)\n}", "func (c *OutputService1ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService1ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService9ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService15ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *Executor) newHTTPRequest(request *Request) (*http.Request, error) {\n\tif request.context == nil {\n\t\treturn nil, errNilContext\n\t}\n\tif !isSupportedHTTPMethod(request.method) {\n\t\treturn nil, fmt.Errorf(\"error executing request, found unsupport HTTP method %s\", request.method)\n\t}\n\n\t// JSON encode the object and use that as the body if specified, otherwise use the body as-is\n\treqBody := request.body\n\tif request.object != nil {\n\t\tb, err := encodeBody(request.object)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error executing request, failed to encode the request object to JSON: %w\", err)\n\t\t}\n\t\treqBody = b\n\t}\n\tu := path.Join(c.apiAddress, request.pathAndQuery)\n\n\tr, err := http.NewRequestWithContext(request.context, request.method, u, reqBody)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error executing request, failed to create a new underlying HTTP request: %w\", err)\n\t}\n\tr.Header.Set(\"User-Agent\", c.userAgent)\n\tif request.contentType != \"\" {\n\t\tr.Header.Set(\"Content-type\", request.contentType)\n\t}\n\tif request.contentLength != nil {\n\t\tr.ContentLength = *request.contentLength\n\t}\n\tfor k, v := range request.headers {\n\t\tr.Header.Set(k, v)\n\t}\n\n\treturn r, nil\n}", "func (c *InputService5ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService8ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (s *VirtualhostsService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService7ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService7ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *PaperTrailClient) NewRequest(url string) (*http.Request, error) {\n\treq, err := http.NewRequest(http.MethodGet, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"X-Papertrail-Token\", c.token)\n\treturn req, nil\n}" ]
[ "0.7601494", "0.7356932", "0.73531973", "0.73461235", "0.7335527", "0.72980696", "0.71913904", "0.7105686", "0.71012974", "0.7100697", "0.7054455", "0.7052163", "0.6941903", "0.6936371", "0.6926456", "0.6922782", "0.69117075", "0.68937415", "0.6887651", "0.6886106", "0.6883375", "0.6876941", "0.6876883", "0.6873908", "0.6873328", "0.6871219", "0.68707216", "0.68707216", "0.6862721", "0.6862721", "0.685456", "0.68494207", "0.6836224", "0.6836224", "0.6832308", "0.68285406", "0.6828389", "0.6827139", "0.6827139", "0.682538", "0.6823518", "0.6816359", "0.6792386", "0.6792386", "0.678892", "0.6786124", "0.6786124", "0.67799515", "0.6773788", "0.6773788", "0.67726696", "0.6768101", "0.6768101", "0.6760485", "0.67546296", "0.67477024", "0.67393947", "0.67329097", "0.67329097", "0.6729483", "0.672659", "0.6710993", "0.6703099", "0.6702191", "0.66813034", "0.6675333", "0.66644883", "0.6654745", "0.66533864", "0.66417474", "0.66377634", "0.6635827", "0.66299057", "0.6628454", "0.6626775", "0.6616933", "0.6616933", "0.66084003", "0.6601619", "0.6601619", "0.65998065", "0.6596877", "0.6595483", "0.65939176", "0.65919495", "0.65893316", "0.6589023", "0.65873516", "0.65868115", "0.6582976", "0.6582976", "0.65740776", "0.6571493", "0.65714073", "0.6570226", "0.6567472", "0.65666384", "0.6566204", "0.6566204", "0.65641266" ]
0.7916281
0
newLogsAPIRequest returns a POST http.Request that is ready to send to the Datadog logs API.
func newLogsAPIRequest(host, path, apiKey string, body []byte) (*http.Request, error) { return newAPIRequestWithHost(host, path, apiKey, body) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewLog(r *http.Request, sw *statusWriter) (logMessage *LogRequest) {\n\tlogMessage = &LogRequest{\n\t\tMethod: r.Method,\n\t\tPath: r.URL.Path,\n\t\tQueryString: r.URL.RawQuery,\n\t\tRemoteIP: realip.FromRequest(r),\n\t\tUserAgent: r.UserAgent(),\n\t}\n\n\tif sw != nil {\n\t\tlogMessage.Status = sw.status\n\t\tlogMessage.ContentLength = sw.length\n\t}\n\n\tdump, err := httputil.DumpRequest(r, true)\n\tif err == nil {\n\t\tlogMessage.Dump = dump\n\t}\n\n\tlogMessage.Params = r.Context().Value(\"params\")\n\n\treturn\n}", "func (f *CloudWatchLogsFormatter) Request() (*http.Request, error) {\n\tbody, err := newCloudWatchLogsBody(f.batch, f.eData, f.logGroupName, f.logStreamName, <-f.tokens)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr, err := http.NewRequest(http.MethodPost, f.url, body)\n\tif err != nil {\n\t\treturn r, err\n\t}\n\tr.Header.Add(\"Content-Type\", xAmazonJSON11)\n\tr.Header.Add(\"X-Amz-Target\", xAmazonTarget)\n\n\t_, err = f.signer.Sign(r, body, cloudWatchLogsServiceName, f.region, time.Now())\n\tf.ReadSeeker = body\n\treturn r, err\n}", "func newAPIRequest(host, path, apiKey string, body []byte) (*http.Request, error) {\n\treturn newAPIRequestWithHost(host, path, apiKey, body)\n}", "func (c *Client) BuildCreateRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: CreateLogPath()}\n\treq, err := http.NewRequest(\"POST\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"log\", \"create\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func CreateListApplicationLogsRequest() (request *ListApplicationLogsRequest) {\n\trequest = &ListApplicationLogsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"companyreg\", \"2019-05-08\", \"ListApplicationLogs\", \"companyreg\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func CreateDescribeGatewayLogsRequest() (request *DescribeGatewayLogsRequest) {\n\trequest = &DescribeGatewayLogsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"sgw\", \"2018-05-11\", \"DescribeGatewayLogs\", \"hcs_sgw\", \"openAPI\")\n\trequest.Method = requests.POST\n\treturn\n}", "func NewRequest(str string) (*Request, error) {\n\tallmatches := logRegexp.FindAllStringSubmatch(str, -1)\n\tif len(allmatches) == 0 {\n\t\treturn &Request{}, fmt.Errorf(\"no matches\")\n\t}\n\tmatches := allmatches[0]\n\n\tbytes, err := strconv.Atoi(matches[6])\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq_time, err := strconv.ParseFloat(matches[7], 64)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq := &Request{\n\t\tIp: matches[1],\n\t\tStatus: matches[5],\n\t\tBytesSent: bytes,\n\t\tReqTime: req_time,\n\t\tReferer: matches[8],\n\t\tAgent: matches[9],\n\t}\n\n\tparseTimestamp(matches[3], req)\n\tparseRequest(matches[4], req)\n\tfmt.Println(str)\n\treturn req, nil\n}", "func NewLogRequestFactory(options ...ClientOption) (RequestFactory, error) {\n\tf := &requestFactory{\n\t\tapiKeyHeader: apiKeyHeader,\n\t\tendpoint: \"log-api.newrelic.com\",\n\t\tpath: \"/log/v1\",\n\t\tuserAgent: defaultUserAgent,\n\t\tscheme: defaultScheme,\n\t\tzippers: newGzipPool(gzip.DefaultCompression),\n\t\tuncompressedBuffers: newUncompressedBufferPool(),\n\t}\n\terr := configure(f, options)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &hashRequestFactory{requestFactory: f}, nil\n}", "func (c *Client) newRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"api-name\", c.apiName)\n\treq.Header.Set(\"api-key\", c.apiKey)\n\treq.Header.Set(c.userHeader, c.user)\n\treturn req, nil\n}", "func (adm AdminClient) newRequest(method string, reqData requestData) (req *http.Request, err error) {\n\t// If no method is supplied default to 'POST'.\n\tif method == \"\" {\n\t\tmethod = \"POST\"\n\t}\n\n\t// Default all requests to \"\"\n\tlocation := \"\"\n\n\t// Construct a new target URL.\n\ttargetURL, err := adm.makeTargetURL(reqData)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Initialize a new HTTP request for the method.\n\treq, err = http.NewRequest(method, targetURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tadm.setUserAgent(req)\n\tfor k, v := range reqData.customHeaders {\n\t\treq.Header.Set(k, v[0])\n\t}\n\tif length := len(reqData.content); length > 0 {\n\t\treq.ContentLength = int64(length)\n\t}\n\treq.Header.Set(\"X-Amz-Content-Sha256\", hex.EncodeToString(sum256(reqData.content)))\n\treq.Body = ioutil.NopCloser(bytes.NewReader(reqData.content))\n\n\treq = s3signer.SignV4(*req, adm.accessKeyID, adm.secretAccessKey, \"\", location)\n\treturn req, nil\n}", "func newRequest(method, url string, body string) *http.Request {\n\treq, err := http.NewRequest(method, url, strings.NewReader(body))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treq.Header.Set(\"X-API-Token\", \"token1\")\n\treturn req\n}", "func CreateListAuthenticationLogsRequest() (request *ListAuthenticationLogsRequest) {\n\trequest = &ListAuthenticationLogsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"idaas-doraemon\", \"2021-05-20\", \"ListAuthenticationLogs\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func (a *APIv2) NewAPIRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.SetBasicAuth(\"x\", a.key)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"X-Gms-Client-Version\", config.VERSION)\n\treq.Header.Add(\"X-Gms-Revision\", utils.GetCurrentRevision())\n\treq.Header.Add(\"X-Gms-Branch\", utils.GetCurrentBranch())\n\treturn req, nil\n}", "func (c *PaperTrailClient) NewRequest(url string) (*http.Request, error) {\n\treq, err := http.NewRequest(http.MethodGet, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"X-Papertrail-Token\", c.token)\n\treturn req, nil\n}", "func newAppendEntriesRequest(term uint64, prevLogIndex uint64, prevLogTerm uint64, commitIndex uint64, leaderName string, entries []*LogEntry) *AppendEntriesRequest {\n\treturn &AppendEntriesRequest{\n\t\tTerm: term,\n\t\tPrevLogIndex: prevLogIndex,\n\t\tPrevLogTerm: prevLogTerm,\n\t\tCommitIndex: commitIndex,\n\t\tLeaderName: leaderName,\n\t\tEntries: entries,\n\t}\n}", "func CreateGetAlarmLogRequest() (request *GetAlarmLogRequest) {\n\trequest = &GetAlarmLogRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"emas-appmonitor\", \"2019-06-11\", \"GetAlarmLog\", \"\", \"\")\n\trequest.Method = requests.POST\n\treturn\n}", "func NewRequestLogger() heimdall.Plugin {\n return &requestLogger{}\n}", "func (c *Client) newRequest(ctx context.Context, method, url string, body io.Reader) (*http.Request, error) {\n\t// Build new request with base URL.\n\treq, err := http.NewRequest(method, c.URL+url, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Set API key in header.\n\tif user := wtf.UserFromContext(ctx); user != nil && user.APIKey != \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bearer \"+user.APIKey)\n\t}\n\n\t// Default to JSON format.\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-type\", \"application/json\")\n\n\treturn req, nil\n}", "func CreateContextQueryLogRequest() (request *ContextQueryLogRequest) {\n\trequest = &ContextQueryLogRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"Emr\", \"2016-04-08\", \"ContextQueryLog\", \"emr\", \"openAPI\")\n\treturn\n}", "func newRequest(req *http.Request) *Request {\n\trequest := &Request{\n\t\tRequest: req,\n\t}\n\n\treturn request\n}", "func (client *WebAppsClient) getWebSiteContainerLogsCreateRequest(ctx context.Context, resourceGroupName string, name string, options *WebAppsGetWebSiteContainerLogsOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/containerlogs\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodPost, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.SkipBodyDownload()\n\treq.Raw().Header.Set(\"Accept\", \"application/octet-stream\")\n\treturn req, nil\n}", "func (b *AuroraBackend) NewRequest(params *CallParams) (*http.Request, error) {\n\treq, err := http.NewRequest(params.Method, params.Path, params.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\theaders := params.Headers\n\tif headers == nil {\n\t\theaders = make(http.Header)\n\t}\n\n\tif params.Credentials != nil {\n\t\theaders.Add(\"X-Application-ID\", params.Credentials.AppID)\n\t\theaders.Add(\"X-Application-Token\", params.Credentials.AppToken)\n\t\theaders.Add(\"X-Device-ID\", params.Credentials.DeviceID)\n\t}\n\n\treq.Header = headers\n\treturn req, nil\n}", "func (s *Stein) newRequest(method string, path string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, path, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\treturn req, nil\n}", "func newRequest(db *Database) *Request {\n\treq := &Request{\n\t\tdb: db,\n\t\tpath: \"/\",\n\t\tquery: url.Values{},\n\t\theader: http.Header{},\n\t}\n\treturn req\n}", "func NewRequest() *Request {\n\treturn defaul.NewRequest()\n}", "func newRequest(r *http.Request, vars map[string]string) *Request {\n\trequest := &Request{r, vars, nil}\n\treturn request\n}", "func (dl DefaultLogger) LogRequest(*http.Request) {\n}", "func CreateGetHealthMonitorLogsRequest() (request *GetHealthMonitorLogsRequest) {\n\trequest = &GetHealthMonitorLogsRequest{\n\t\tRpcRequest: &requests.RpcRequest{},\n\t}\n\trequest.InitWithApiInfo(\"EHPC\", \"2018-04-12\", \"GetHealthMonitorLogs\", \"\", \"\")\n\trequest.Method = requests.GET\n\treturn\n}", "func (qiwi *PersonalAPI) newRequest(apiKey, method, spath string, data map[string]interface{}) (req *http.Request, err error) {\n\n\tvar path = APIURL + spath\n\n\tvar body io.Reader\n\n\tif len(data) > 0 {\n\n\t\ts, err := json.Marshal(data)\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tbody = bytes.NewBuffer(s)\n\n\t}\n\n\treq, err = http.NewRequest(method, path, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Bearer \"+apiKey)\n\n\treturn req, err\n}", "func (conn Connection) newRequest(method, cmd string, body io.Reader) *http.Request {\n\t// req, err := conn.jhReq(method, cmd, body)\n\treq, err := http.NewRequest(method, conn.HubURL+cmd, body)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Coulnd't generate HTTP request - %s\\n\", err.Error()))\n\t}\n\n\treq.Header.Add(\"Authorization\", fmt.Sprintf(\"token %s\", conn.Token))\n\n\treturn req\n}", "func (s *SitesService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c client) newRequest(ctx context.Context, method string, url string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequestWithContext(ctx, method, url, body)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to build request %w\", err)\n\t}\n\n\treturn req, nil\n}", "func newRequestScope(now time.Time, logger *logrus.Logger, request *http.Request, db *mongo.Database) RequestScope {\n\tl := log.NewLogger(logger, logrus.Fields{})\n\trequestID := request.Header.Get(\"X-Request-Id\")\n\tif requestID != \"\" {\n\t\tl.SetField(\"RequestID\", requestID)\n\t}\n\n\treturn &requestScope{\n\t\tLogger: l,\n\t\tnow: now,\n\t\trequestID: requestID,\n\t\tdb: db,\n\t\trequest: request,\n\t}\n}", "func NewRequestRecorder(r *http.Request) RequestRecorder {\n\tb, err := io.ReadAll(r.Body)\n\tif err == nil {\n\t\t_ = r.Body.Close()\n\t\tr.Body = io.NopCloser(bytes.NewReader(b))\n\t}\n\treturn RequestRecorder{\n\t\tRequest: r,\n\t\tPayload: b,\n\t}\n}", "func NewRequest(method, url string, payload interface{}) (*http.Request, error) {\n\tvar buf io.Reader\n\tif payload != nil {\n\t\tvar b []byte\n\t\tb, err := json.Marshal(&payload)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbuf = bytes.NewBuffer(b)\n\t}\n\treturn http.NewRequest(method, url, buf)\n}", "func NewDescribeAuditLogRequestWithoutParam() *DescribeAuditLogRequest {\n\n return &DescribeAuditLogRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/instances/{insId}/logs/{logId}\",\n Method: \"GET\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func (g *baseGithub) newRequest(method, url string, body interface{}) (*http.Request, error) {\n\tvar err error\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\tenc := json.NewEncoder(buf)\n\t\tenc.SetEscapeHTML(false)\n\t\terr = enc.Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, url, buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\treq.Header.Set(\"Accept\", \"application/json\")\n\n\treturn req, nil\n}", "func (s *VirtualhostsService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (s *AuthnReqListsService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *APIGateway) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func (client *WebAppsClient) getMSDeployLogCreateRequest(ctx context.Context, resourceGroupName string, name string, options *WebAppsGetMSDeployLogOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/extensions/MSDeploy/log\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (client *Client) NewRequest(messageDataBuffer *strings.Reader) (*http.Request, error) {\n\trequest, err := http.NewRequest(\"POST\", client.BaseURL, messageDataBuffer)\n\n\tif err != nil {\n\t\terrStr := fmt.Sprintf(\"Error constructing the HTTP network request ... here is the error %v\", err)\n\t\treturn &http.Request{}, &errorString{errStr}\n\t}\n\n\t// fmt.Printf(\"ACCOUNT SID \", client.AccountSID , \"Auth Token \", client.AuthToken)\n\trequest.SetBasicAuth(client.AccountSID, client.AuthToken) // Authenticating user credentials\n\n\t// Additional header fields to accept json media types which can be used for the response\n\trequest.Header.Add(\"Accept\", \"application/json\")\n\n\t// To indicate the media type that is being sent through the request\n\trequest.Header.Add(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\treturn request, nil\n}", "func newRequest(req *http.Request) *Request {\n\treturn &Request{\n\t\tRequest: req,\n\t}\n}", "func (c *Client) NewDataRecordsRequest(ctx context.Context, path string) (*http.Request, error) {\n\tscheme := c.Scheme\n\tif scheme == \"\" {\n\t\tscheme = \"https\"\n\t}\n\tu := url.URL{Host: c.Host, Scheme: scheme, Path: path}\n\treq, err := http.NewRequest(\"GET\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif c.JWTSigner != nil {\n\t\tif err := c.JWTSigner.Sign(req); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\treturn req, nil\n}", "func newRequestScope(now time.Time, logger *logrus.Logger, request *http.Request) RequestScope {\n\tl := NewLogger(logger, logrus.Fields{})\n\trequestID := request.Header.Get(\"X-Request-Id\")\n\tif requestID != \"\" {\n\t\tl.SetField(\"RequestID\", requestID)\n\t}\n\treturn &requestScope{\n\t\tLogger: l,\n\t\tnow: now,\n\t\trequestID: requestID,\n\t}\n}", "func (g *Group) newRequest(operation *request.Operation, input interface{}, output interface{}, handlers *request.Handlers) *request.Request {\n return request.New(g.Client, operation, input, output, handlers)\n}", "func NewRequest(requestName string, params rata.Params, header http.Header, query url.Values, body ...io.Reader) Request {\n\tif header == nil {\n\t\theader = http.Header{}\n\t}\n\theader.Set(\"Accept\", \"application/json\")\n\n\trequest := Request{\n\t\tRequestName: requestName,\n\t\tParams: params,\n\t\tHeader: header,\n\t\tQuery: query,\n\t}\n\n\tif len(body) == 1 {\n\t\trequest.Body = body[0]\n\t}\n\n\treturn request\n}", "func (client *WebAppsClient) listDeploymentLogCreateRequest(ctx context.Context, resourceGroupName string, name string, id string, options *WebAppsListDeploymentLogOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/deployments/{id}/log\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif id == \"\" {\n\t\treturn nil, errors.New(\"parameter id cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{id}\", url.PathEscape(id))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (c *Client) BuildListRequest(ctx context.Context, v interface{}) (*http.Request, error) {\n\tu := &url.URL{Scheme: c.scheme, Host: c.host, Path: ListLogPath()}\n\treq, err := http.NewRequest(\"POST\", u.String(), nil)\n\tif err != nil {\n\t\treturn nil, goahttp.ErrInvalidURL(\"log\", \"list\", u.String(), err)\n\t}\n\tif ctx != nil {\n\t\treq = req.WithContext(ctx)\n\t}\n\n\treturn req, nil\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v7.1.0\")\n\treturn req, nil\n}", "func NewRequest(r *http.Request) *Request {\n\tvar request Request\n\trequest.ID = atomic.AddUint32(&requestID, 1)\n\trequest.Method = r.Method\n\trequest.Body = r.Body\n\trequest.BodyBuff = new(bytes.Buffer)\n\trequest.BodyBuff.ReadFrom(r.Body)\n\trequest.RemoteAddr = r.Header.Get(\"X-Forwarded-For\")\n\trequest.Header = r.Header\n\tif request.RemoteAddr == \"\" {\n\t\trequest.RemoteAddr = r.RemoteAddr\n\t}\n\trequest.UrlParams = mux.Vars(r)\n\trequest.QueryParams = r.URL.Query()\n\treturn &request\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v8.1.0\")\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\trel, err := url.Parse(path.Join(apiVersion, urlStr))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.baseURL.ResolveReference(rel)\n\treq, err := http.NewRequest(method, u.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Authorization\", \"Bearer \"+c.config.accessToken)\n\tif c.userAgent != \"\" {\n\t\treq.Header.Set(\"User-Agent\", c.userAgent)\n\t}\n\treturn req, nil\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v5.4.1\")\n\treturn req, nil\n}", "func (s *ApplicationsService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func NewAppStoreVersionReleaseRequestsCreateInstanceRequest(server string, body AppStoreVersionReleaseRequestsCreateInstanceJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewAppStoreVersionReleaseRequestsCreateInstanceRequestWithBody(server, \"application/json\", bodyReader)\n}", "func (g *GoFlickr) newRequest(apiMethod string) ApiRequest {\n\n\treq := ApiRequest{\n\t\tMethodName: apiMethod,\n\t}\n\treq.addParam(\"api_key\", g.ApiKey)\n\treq.addParam(\"format\", \"json\")\n\treq.addParam(\"nojsoncallback\", \"1\")\n\treturn req\n\n}", "func NewRequest(apiKey string, query *Tag) *Tag {\n\treturn Request().Add(\n\t\tLogin().Opts(Opts{\n\t\t\tOptAuthenticationKey: apiKey,\n\t\t}),\n\t\tquery,\n\t)\n}", "func (s *OidcService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *Client) newRequest(method, path string, v interface{}, ctype string) (req *http.Request, err error) {\n\t// Build request JSON.\n\tbody, err := writeJson(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq, err = http.NewRequest(method, c.pathToEndPoint(path), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Add(\"X-Kii-AppID\", c.AppId)\n\treq.Header.Add(\"X-Kii-AppKey\", c.AppKey)\n\tif ctype != \"\" {\n\t\treq.Header.Add(\"Content-Type\", ctype)\n\t}\n\tif c.Authorization != \"\" {\n\t\treq.Header.Add(\"Authorization\", c.Authorization)\n\t}\n\treturn\n}", "func (c *apiHttp) NewRequest(command string) RequestBuilder {\n\theaders := make(map[string]string)\n\tif c.Headers != nil {\n\t\tfor k := range c.Headers {\n\t\t\theaders[k] = c.Headers.Get(k)\n\t\t}\n\t}\n\n\treturn &requestBuilder{\n\t\tcommand: command,\n\t\tclient: c,\n\t\theaders: headers,\n\t}\n}", "func RequestLogging(n janice.HandlerFunc) janice.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) error {\n\t\tp := r.URL.String()\n\n\t\tvar err error\n\t\tm := httpsnoop.CaptureMetricsFn(w, func(ww http.ResponseWriter) {\n\t\t\terr = n(ww, r)\n\t\t})\n\n\t\tle := Logger.WithFields(logrus.Fields{\n\t\t\t\"type\": \"request\",\n\t\t\t\"host\": r.Host,\n\t\t\t\"method\": r.Method,\n\t\t\t\"path\": p,\n\t\t\t\"code\": m.Code,\n\t\t\t\"duration\": m.Duration.String(),\n\t\t\t\"written\": m.Written,\n\t\t})\n\n\t\tif rid, ok := GetRequestID(r); ok {\n\t\t\tle = le.WithField(\"request\", rid)\n\t\t}\n\n\t\tle.Info()\n\n\t\treturn err\n\t}\n}", "func (c *InputService17ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (client *SyncGroupsClient) listLogsCreateRequest(ctx context.Context, resourceGroupName string, serverName string, databaseName string, syncGroupName string, startTime string, endTime string, typeParam SyncGroupsType, options *SyncGroupsClientListLogsOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Sql/servers/{serverName}/databases/{databaseName}/syncGroups/{syncGroupName}/logs\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif serverName == \"\" {\n\t\treturn nil, errors.New(\"parameter serverName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{serverName}\", url.PathEscape(serverName))\n\tif databaseName == \"\" {\n\t\treturn nil, errors.New(\"parameter databaseName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{databaseName}\", url.PathEscape(databaseName))\n\tif syncGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter syncGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{syncGroupName}\", url.PathEscape(syncGroupName))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.host, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"startTime\", startTime)\n\treqQP.Set(\"endTime\", endTime)\n\treqQP.Set(\"type\", string(typeParam))\n\tif options != nil && options.ContinuationToken != nil {\n\t\treqQP.Set(\"continuationToken\", *options.ContinuationToken)\n\t}\n\treqQP.Set(\"api-version\", \"2020-11-01-preview\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (rt *registryRoundTripper) NewRequest(method, path string,\n\tquery *url.Values, body interface{}) (*http.Request, error) {\n\n\treq, err := rt.DefaultRequestDoer.NewRequest(method, path, query, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif tok := rt.holder.Token(); tok != \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bearer \"+tok)\n\t}\n\n\treq.Header.Set(\"User-Agent\", \"Torus-Daemon/\"+rt.version)\n\treq.Header.Set(\"X-Registry-Version\", rt.apiVersion)\n\n\treturn req, nil\n}", "func (s *HighAvailabilityService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func buildHTTPRequest(request *MacaroonRequest) (*http.Request, error) {\n\tpayload := &jsonMacaroonRequest{\n\t\tCaveats: []string{\n\t\t\tfmt.Sprint(\"activity:\", strings.Join(request.Activities, \",\")),\n\t\t},\n\t}\n\n\tif request.Lifetime > 0 {\n\t\tbefore := time.Now().Add(request.Lifetime).UTC()\n\t\tpayload.Caveats = append(payload.Caveats, fmt.Sprint(\"before:\", before.Format(time.RFC3339)))\n\t}\n\n\tpayloadData, e := json.Marshal(payload)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\n\treq := &http.Request{\n\t\tMethod: \"POST\",\n\t\tHeader: http.Header{},\n\t}\n\treq.Header.Add(\"Content-Type\", \"application/macaroon-request\")\n\treq.URL, e = url.Parse(request.Resource)\n\tif e != nil {\n\t\treturn nil, e\n\t}\n\treq.Body = ioutil.NopCloser(bytes.NewReader(payloadData))\n\treq.ContentLength = int64(len(payloadData))\n\treturn req, nil\n}", "func NewRequest(r *http.Request, ns string, body []byte) (*Request, error) {\n\th, err := json.Marshal(r.Header)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Request{\n\t\tID: uuid.NewV4(),\n\t\tMethod: r.Method,\n\t\tHeaders: h,\n\t\tPayload: body,\n\t\tHost: r.Host,\n\t\tScheme: r.URL.Scheme,\n\t\tPath: r.URL.Path,\n\t\tNamespace: ns,\n\t\tRetries: 0,\n\t}, nil\n}", "func (client *WebAppsClient) getInstanceMSDeployLogCreateRequest(ctx context.Context, resourceGroupName string, name string, instanceID string, options *WebAppsGetInstanceMSDeployLogOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.Web/sites/{name}/instances/{instanceId}/extensions/MSDeploy/log\"\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif name == \"\" {\n\t\treturn nil, errors.New(\"parameter name cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{name}\", url.PathEscape(name))\n\tif instanceID == \"\" {\n\t\treturn nil, errors.New(\"parameter instanceID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{instanceId}\", url.PathEscape(instanceID))\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\treq, err := runtime.NewRequest(ctx, http.MethodGet, runtime.JoinPaths(client.ep, urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2021-02-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header.Set(\"Accept\", \"application/json\")\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body io.Reader) (r *http.Request, err error) {\n\tif r, err = http.NewRequest(method, urlStr, body); err != nil {\n\t\treturn\n\t}\n\n\tr.Header.Set(\"User-Agent\", c.UserAgent)\n\tr.Header.Set(\"X-Requested-With\", \"XMLHttpRequest\")\n\n\tif c.ApiKey != nil {\n\t\tr.Header.Set(\"X-Sbss-Auth\", c.ApiKey.Login)\n\t\tr.AddCookie(c.ApiKey.Cookie)\n\t}\n\n\tif method == \"POST\" {\n\t\tr.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\t}\n\n\treturn\n}", "func newRequest(w http.ResponseWriter, r *http.Request, app *Webapp) *Request {\n\treq := &Request{\n\t\tw: w,\n\t\tr: r,\n\t\tapp: app,\n\t\tstatus: 200,\n\t\tcontentLength: 0,\n\t\tcontentType: \"text/html; charset=utf-8\",\n\t\tdate: time.Now(),\n\t\treplied: false,\n\t}\n\treturn req\n}", "func NewSearchLogContextRequestWithoutParam() *SearchLogContextRequest {\n\n return &SearchLogContextRequest{\n JDCloudRequest: core.JDCloudRequest{\n URL: \"/regions/{regionId}/logsets/{logsetUID}/logtopics/{logtopicUID}/logcontext\",\n Method: \"POST\",\n Header: nil,\n Version: \"v1\",\n },\n }\n}", "func newRequest(w http.ResponseWriter, rq *http.Request) request {\n\tr := request{\n\t\tpath: rq.URL.Path,\n\t\tctx: rq.Context(),\n\t\tr: rq,\n\t\tw: w,\n\t\tstart: time.Now(),\n\t\trid: rand.Uint64(),\n\t}\n\tr.rid |= 1 << 63 // sacrifice one bit of entropy so they always have the same # digits\n\tr.ip = r.r.Header.Get(\"X-Forwarded-For\")\n\tr.port = r.r.Header.Get(\"X-Forwarded-Port\")\n\tif r.ip == \"\" {\n\t\tr.ip, r.port, _ = net.SplitHostPort(r.r.RemoteAddr)\n\t}\n\tr.log(\n\t\t\"ip\", r.ip,\n\t\t\"port\", r.port,\n\t\t\"raddr\", r.r.RemoteAddr,\n\t\t\"method\", r.r.Method,\n\t\t\"path\", r.r.URL.Path,\n\t\t\"ref\", r.r.Referer(),\n\t\t\"ua\", r.r.UserAgent(),\n\t)\n\treturn r\n}", "func (mdbc *MongoDbController) AddRequestLog(log *authUtils.RequestLogData) error {\n\tcollection, backCtx, cancel := mdbc.getCollection(\"logging\")\n\tdefer cancel()\n\n\tinsert := bson.D{\n\t\t{Key: \"timestamp\", Value: primitive.Timestamp{T: uint32(log.Timestamp.Unix())}},\n\t\t{Key: \"type\", Value: log.Type},\n\t\t{Key: \"clientIP\", Value: log.ClientIP},\n\t\t{Key: \"method\", Value: log.Method},\n\t\t{Key: \"path\", Value: log.Path},\n\t\t{Key: \"protocol\", Value: log.Protocol},\n\t\t{Key: \"statusCode\", Value: log.StatusCode},\n\t\t{Key: \"latency\", Value: log.Latency},\n\t\t{Key: \"userAgent\", Value: log.UserAgent},\n\t\t{Key: \"errorMessage\", Value: log.ErrorMessage},\n\t}\n\n\t_, mdbErr := collection.InsertOne(backCtx, insert)\n\n\tif mdbErr != nil {\n\t\treturn dbController.NewDBError(mdbErr.Error())\n\t}\n\n\treturn nil\n}", "func (c *InputService14ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *Client) NewRequest(method, apiPath string, body interface{}) (*http.Request, error) {\n loc, err := joinPaths(c.BaseURL, apiPath)\n if err != nil {\n return nil, err\n }\n\n buf := new(bytes.Buffer)\n if body != nil {\n err = json.NewEncoder(buf).Encode(body)\n if err != nil {\n return nil, err\n }\n }\n\n req, err := http.NewRequest(method, loc.String(), buf)\n if err != nil {\n return nil, err\n }\n\n req.Header.Add(\"Content-Type\", \"application/json\")\n req.Header.Add(\"Accept\", \"application/json, text/plain, */*\")\n req.Header.Add(\"User-Agent\", c.UserAgent)\n return req, nil\n}", "func (r *APIClientRepository) LogRequest(key, method, resource, ip string) error {\n\ttx, err := r.DB.Begin()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttx.Exec(`\n insert into api_requests (key, method, resource, requested_at, requested_by)\n values ($1, $2, $3, NOW(), $4)\n `,\n\t\tkey,\n\t\tmethod,\n\t\tresource,\n\t\tip,\n\t)\n\n\tif err = tx.Commit(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *IoT) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func NewRequest(method, url string) (*http.Request, error) {\n\treturn http.NewRequest(method, url, nil)\n}", "func NewRequest(client Client) (*Request, error) {\n\n\tvar req = Request{\n\t\tToken: utils.Generate(8),\n\t\tClient: client.ID,\n\t\tStatus: 1,\n\t\tScore: 0,\n\t}\n\n\terr := controller.Database.Create(&req).Error\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &req, nil\n}", "func CreateBaseRequest(method, url string, body []byte, user m.AuthUser, verbose bool) *http.Request {\n\treq, err := http.NewRequest(method, url, bytes.NewBuffer(body))\n\treq.SetBasicAuth(user.Username, user.Password)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Accept\", \"application/json\")\n\tError(err, \"Error creating the request\")\n\n\tif verbose {\n\t\tfmt.Println(\"Request Url:\", req.URL)\n\t\tfmt.Println(\"Request Headers:\", req.Header)\n\t\tfmt.Println(\"Request Body:\", req.Body)\n\t}\n\n\treturn req\n}", "func NewRequest() (r *Request) {\n\treturn &Request{}\n}", "func NewRequest(url string) *Request {\n\treturn client.NewRequest(url)\n}", "func (c *OutputService15ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService14ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService15ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (a *API) FunctionLogsApiPost(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\t// get dbType from url\n\tdbType := vars[\"dbType\"]\n\n\t// find correspond db http url\n\tdbCnf := a.getLogDBConfig(dbType)\n\n\tsvcUrl, err := url.Parse(dbCnf.httpURL)\n\tif err != nil {\n\t\tlog.Printf(\"Failed to establish proxy server for function logs: %v\", err)\n\t}\n\t// set up proxy server director\n\tdirector := func(req *http.Request) {\n\t\t// only replace url Scheme and Host to remote influxDB\n\t\t// and leave query string intact\n\t\treq.URL.Scheme = svcUrl.Scheme\n\t\treq.URL.Host = svcUrl.Host\n\t\treq.URL.Path = svcUrl.Path\n\t\t// set up http basic auth for database authentication\n\t\treq.SetBasicAuth(dbCnf.username, dbCnf.password)\n\t}\n\tproxy := &httputil.ReverseProxy{\n\t\tDirector: director,\n\t}\n\tproxy.ServeHTTP(w, r)\n}", "func (o *GetLogsParams) WriteToRequest(r runtime.ClientRequest, reg strfmt.Registry) error {\n\n\tif err := r.SetTimeout(o.timeout); err != nil {\n\t\treturn err\n\t}\n\tvar res []error\n\n\tif o.XRequestID != nil {\n\n\t\t// header param X-Request-Id\n\t\tif err := r.SetHeaderParam(\"X-Request-Id\", *o.XRequestID); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t}\n\n\tif o.Page != nil {\n\n\t\t// query param page\n\t\tvar qrPage int64\n\t\tif o.Page != nil {\n\t\t\tqrPage = *o.Page\n\t\t}\n\t\tqPage := swag.FormatInt64(qrPage)\n\t\tif qPage != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page\", qPage); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif o.PageSize != nil {\n\n\t\t// query param page_size\n\t\tvar qrPageSize int64\n\t\tif o.PageSize != nil {\n\t\t\tqrPageSize = *o.PageSize\n\t\t}\n\t\tqPageSize := swag.FormatInt64(qrPageSize)\n\t\tif qPageSize != \"\" {\n\t\t\tif err := r.SetQueryParam(\"page_size\", qPageSize); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\t// path param project_name\n\tif err := r.SetPathParam(\"project_name\", o.ProjectName); err != nil {\n\t\treturn err\n\t}\n\n\tif o.Q != nil {\n\n\t\t// query param q\n\t\tvar qrQ string\n\t\tif o.Q != nil {\n\t\t\tqrQ = *o.Q\n\t\t}\n\t\tqQ := qrQ\n\t\tif qQ != \"\" {\n\t\t\tif err := r.SetQueryParam(\"q\", qQ); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (conn Connection) newRequest(method, cmd string, body io.Reader) *http.Request {\n\treq, err := http.NewRequest(method, conn.ServiceURL+cmd, body)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Coulnd't generate HTTP request - %s\\n\", err.Error()))\n\t}\n\n\tfor k, v := range conn.Headers {\n\t\treq.Header.Add(k, v)\n\t}\n\n\treturn req\n}", "func (c *SQS) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func createRequest(t *testing.T, method string, path string, body io.Reader) (*http.Request, *httptest.ResponseRecorder, *bytes.Buffer) {\n\trecorder := httptest.NewRecorder()\n\treq, err := http.NewRequest(method, path, body)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tlogger, output := NewFakeLogger()\n\treq = req.WithContext(context.WithValue(req.Context(), middleware.LoggerKey, &logger))\n\treq = req.WithContext(context.WithValue(req.Context(), middleware.AuthUserKey, \"test@draupnir\"))\n\treq = req.WithContext(context.WithValue(req.Context(), middleware.RefreshTokenKey, \"refresh-token\"))\n\treq = req.WithContext(context.WithValue(req.Context(), middleware.UserIPAddressKey, \"1.2.3.4\"))\n\n\treturn req, recorder, output\n}", "func (c *InputService13ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (client *ScriptExecutionsClient) getExecutionLogsCreateRequest(ctx context.Context, resourceGroupName string, privateCloudName string, scriptExecutionName string, options *ScriptExecutionsClientGetExecutionLogsOptions) (*policy.Request, error) {\n\turlPath := \"/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AVS/privateClouds/{privateCloudName}/scriptExecutions/{scriptExecutionName}/getExecutionLogs\"\n\tif client.subscriptionID == \"\" {\n\t\treturn nil, errors.New(\"parameter client.subscriptionID cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{subscriptionId}\", url.PathEscape(client.subscriptionID))\n\tif resourceGroupName == \"\" {\n\t\treturn nil, errors.New(\"parameter resourceGroupName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{resourceGroupName}\", url.PathEscape(resourceGroupName))\n\tif privateCloudName == \"\" {\n\t\treturn nil, errors.New(\"parameter privateCloudName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{privateCloudName}\", url.PathEscape(privateCloudName))\n\tif scriptExecutionName == \"\" {\n\t\treturn nil, errors.New(\"parameter scriptExecutionName cannot be empty\")\n\t}\n\turlPath = strings.ReplaceAll(urlPath, \"{scriptExecutionName}\", url.PathEscape(scriptExecutionName))\n\treq, err := runtime.NewRequest(ctx, http.MethodPost, runtime.JoinPaths(client.internal.Endpoint(), urlPath))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treqQP := req.Raw().URL.Query()\n\treqQP.Set(\"api-version\", \"2023-03-01\")\n\treq.Raw().URL.RawQuery = reqQP.Encode()\n\treq.Raw().Header[\"Accept\"] = []string{\"application/json\"}\n\tif options != nil && options.ScriptOutputStreamType != nil {\n\t\tif err := runtime.MarshalAsJSON(req, options.ScriptOutputStreamType); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\treturn req, nil\n\t}\n\treturn req, nil\n}", "func (c *OutputService4ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService4ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (i *instance) NewRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Make a context for this request.\n\tc := &context{\n\t\treq: req,\n\t\tsession: newSessionID(),\n\t\tinstance: i,\n\t}\n\n\t// Associate this request.\n\trelease := appengine_internal.RegisterTestContext(req, c)\n\ti.relFuncs = append(i.relFuncs, release)\n\n\treturn req, nil\n}", "func (c *apiClient) newRequest(ctx context.Context, method, resourcePath string,\n\tquery url.Values, body io.Reader) (*http.Request, error) {\n\tURL, err := url.Parse(c.BaseURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// BaseURL and resource URL is joined if they have a path\n\tURL.Path = c.joinURLPath(URL.Path, resourcePath)\n\tif query != nil {\n\t\tURL.RawQuery = query.Encode()\n\t}\n\trequest, err := http.NewRequestWithContext(ctx, method, URL.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trequest.Host = c.Host // allow cloudfronting\n\tif c.Authorization != \"\" {\n\t\trequest.Header.Set(\"Authorization\", c.Authorization)\n\t}\n\tif c.Accept != \"\" {\n\t\trequest.Header.Set(\"Accept\", c.Accept)\n\t}\n\trequest.Header.Set(\"User-Agent\", c.UserAgent)\n\treturn request, nil\n}", "func (c *InputService9ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService9ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService4ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}" ]
[ "0.6749788", "0.63361615", "0.63235855", "0.61429274", "0.611278", "0.60289276", "0.6024031", "0.6008218", "0.6004079", "0.59973615", "0.5975654", "0.588213", "0.5871575", "0.5856058", "0.5855431", "0.58046454", "0.57759327", "0.5774671", "0.577183", "0.57710236", "0.5766473", "0.57561964", "0.5747855", "0.57313854", "0.5730542", "0.5723059", "0.5722645", "0.57178444", "0.5714399", "0.56989074", "0.5674266", "0.56704575", "0.56670666", "0.5660409", "0.5637996", "0.56374586", "0.56332135", "0.5626307", "0.55781364", "0.55781126", "0.5577463", "0.55765134", "0.5574986", "0.55689293", "0.5567455", "0.5561118", "0.5556086", "0.5530332", "0.5524869", "0.5523114", "0.5511816", "0.55008316", "0.5497008", "0.54953986", "0.5488256", "0.5481228", "0.54737914", "0.547376", "0.5473615", "0.54717886", "0.54693896", "0.54646736", "0.5461584", "0.54615283", "0.5458587", "0.5457425", "0.5456467", "0.54551953", "0.5453529", "0.5450772", "0.54496175", "0.5449478", "0.5446689", "0.54457116", "0.54370385", "0.54347515", "0.5427176", "0.54187703", "0.54167306", "0.54166734", "0.5416266", "0.54145247", "0.54129684", "0.54119533", "0.5411778", "0.541075", "0.54048955", "0.53991413", "0.5395437", "0.5395281", "0.5392396", "0.5391416", "0.53861165", "0.5384741", "0.5384741", "0.53846544", "0.53845125", "0.53819966", "0.53819966", "0.5377351" ]
0.8531254
0
newAPIRequestWithHost returns a POST http.Request that is ready to send to the Datadog API.
func newAPIRequestWithHost(host, path, apiKey string, body []byte) (*http.Request, error) { req, err := http.NewRequest(http.MethodPost, host+path, bytes.NewReader(body)) if err != nil { return nil, err } req.Header.Set(contentTypeHeader, contentTypeJSON) req.Header.Set(apiKeyHeader, apiKey) return req, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func newAPIRequest(host, path, apiKey string, body []byte) (*http.Request, error) {\n\treturn newAPIRequestWithHost(host, path, apiKey, body)\n}", "func newLogsAPIRequest(host, path, apiKey string, body []byte) (*http.Request, error) {\n\treturn newAPIRequestWithHost(host, path, apiKey, body)\n}", "func NewRequest(host string) *Request {\n\trequest := &Request{host, url.Values{}, http.Header{}, BasicAuth{}}\n\treturn request\n}", "func (c *apiClient) newRequest(ctx context.Context, method, resourcePath string,\n\tquery url.Values, body io.Reader) (*http.Request, error) {\n\tURL, err := url.Parse(c.BaseURL)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t// BaseURL and resource URL is joined if they have a path\n\tURL.Path = c.joinURLPath(URL.Path, resourcePath)\n\tif query != nil {\n\t\tURL.RawQuery = query.Encode()\n\t}\n\trequest, err := http.NewRequestWithContext(ctx, method, URL.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trequest.Host = c.Host // allow cloudfronting\n\tif c.Authorization != \"\" {\n\t\trequest.Header.Set(\"Authorization\", c.Authorization)\n\t}\n\tif c.Accept != \"\" {\n\t\trequest.Header.Set(\"Accept\", c.Accept)\n\t}\n\trequest.Header.Set(\"User-Agent\", c.UserAgent)\n\treturn request, nil\n}", "func (a *APIv2) NewAPIRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.SetBasicAuth(\"x\", a.key)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"X-Gms-Client-Version\", config.VERSION)\n\treq.Header.Add(\"X-Gms-Revision\", utils.GetCurrentRevision())\n\treq.Header.Add(\"X-Gms-Branch\", utils.GetCurrentBranch())\n\treturn req, nil\n}", "func (c *Client) newRequest(ctx context.Context, method, url string, body io.Reader) (*http.Request, error) {\n\t// Build new request with base URL.\n\treq, err := http.NewRequest(method, c.URL+url, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Set API key in header.\n\tif user := wtf.UserFromContext(ctx); user != nil && user.APIKey != \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bearer \"+user.APIKey)\n\t}\n\n\t// Default to JSON format.\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Content-type\", \"application/json\")\n\n\treturn req, nil\n}", "func (c *Client) newRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"api-name\", c.apiName)\n\treq.Header.Set(\"api-key\", c.apiKey)\n\treq.Header.Set(c.userHeader, c.user)\n\treturn req, nil\n}", "func newRequest(method, url string, body string) *http.Request {\n\treq, err := http.NewRequest(method, url, strings.NewReader(body))\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treq.Header.Set(\"X-API-Token\", \"token1\")\n\treturn req\n}", "func (rt *registryRoundTripper) NewRequest(method, path string,\n\tquery *url.Values, body interface{}) (*http.Request, error) {\n\n\treq, err := rt.DefaultRequestDoer.NewRequest(method, path, query, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif tok := rt.holder.Token(); tok != \"\" {\n\t\treq.Header.Set(\"Authorization\", \"Bearer \"+tok)\n\t}\n\n\treq.Header.Set(\"User-Agent\", \"Torus-Daemon/\"+rt.version)\n\treq.Header.Set(\"X-Registry-Version\", rt.apiVersion)\n\n\treturn req, nil\n}", "func (conn Connection) newRequest(method, cmd string, body io.Reader) *http.Request {\n\t// req, err := conn.jhReq(method, cmd, body)\n\treq, err := http.NewRequest(method, conn.HubURL+cmd, body)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Coulnd't generate HTTP request - %s\\n\", err.Error()))\n\t}\n\n\treq.Header.Add(\"Authorization\", fmt.Sprintf(\"token %s\", conn.Token))\n\n\treturn req\n}", "func (k *KeKahu) newRequest(method, endpoint string, body io.Reader) (*http.Request, error) {\n\n\t// Parse the endpoint\n\tep, err := url.Parse(endpoint)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not parse endpoint: %s\", err)\n\t}\n\n\t// Resolve the URL reference\n\tbaseURL, err := k.config.GetURL()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\turl := baseURL.ResolveReference(ep)\n\n\t// Construct the request\n\treq, err := http.NewRequest(method, url.String(), body)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"could not create request: %s\", err)\n\t}\n\n\t// Add the headers\n\treq.Header.Set(\"Authorization\", fmt.Sprintf(\"Bearer %s\", k.config.APIKey))\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Accept\", \"application/json\")\n\n\ttrace(\"created %s request to %s\", method, url)\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, path string) (*http.Request, error) {\n\tvar herokuAuthToken = os.Getenv(\"HKPG_HEROKU_AUTH_TOKEN\")\n\tif herokuAuthToken == \"\" {\n\t\tlog.Fatalf(\"HKPG_HEROKU_AUTH_TOKEN must be set\")\n\t}\n\n\tconst apiURL = \"https://postgres-api.heroku.com\"\n\n\treq, err := http.NewRequest(method, apiURL+path, nil)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.SetBasicAuth(\"\", herokuAuthToken)\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, apiPath string, body interface{}) (*http.Request, error) {\n loc, err := joinPaths(c.BaseURL, apiPath)\n if err != nil {\n return nil, err\n }\n\n buf := new(bytes.Buffer)\n if body != nil {\n err = json.NewEncoder(buf).Encode(body)\n if err != nil {\n return nil, err\n }\n }\n\n req, err := http.NewRequest(method, loc.String(), buf)\n if err != nil {\n return nil, err\n }\n\n req.Header.Add(\"Content-Type\", \"application/json\")\n req.Header.Add(\"Accept\", \"application/json, text/plain, */*\")\n req.Header.Add(\"User-Agent\", c.UserAgent)\n return req, nil\n}", "func (r *Requester) newRequest(endpoint string) (*http.Request, error) {\n req, err := http.NewRequest(\"GET\", endpoint, nil)\n if err != nil {\n return nil, err\n }\n\tbearer := fmt.Sprintf(\"Bearer %s\", r.bearer)\n req.Header.Add(\"Authorization\", bearer)\n\treq.Header.Add(\"Ocp-Apim-Subscription-Key\", apimKey)\n req.Header.Set(\"User-Agent\", \"hackacraic\")\n\treturn req, nil\n}", "func (b *AuroraBackend) NewRequest(params *CallParams) (*http.Request, error) {\n\treq, err := http.NewRequest(params.Method, params.Path, params.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\theaders := params.Headers\n\tif headers == nil {\n\t\theaders = make(http.Header)\n\t}\n\n\tif params.Credentials != nil {\n\t\theaders.Add(\"X-Application-ID\", params.Credentials.AppID)\n\t\theaders.Add(\"X-Application-Token\", params.Credentials.AppToken)\n\t\theaders.Add(\"X-Device-ID\", params.Credentials.DeviceID)\n\t}\n\n\treq.Header = headers\n\treturn req, nil\n}", "func (c *Client) newRequest(subdomain, method string, endpoint string, params Params, body interface{}) (*http.Request, error) {\n\n\tif subdomain == \"\" {\n\t\tsubdomain = \"www\"\n\t}\n\n\trequestURL := c.baseURL + \"/spi/v3/{endpoint}\"\n\trequestURL = strings.Replace(requestURL, \"{subdomain}\", subdomain, -1)\n\trequestURL = strings.Replace(requestURL, \"{domain}\", c.domain, -1)\n\trequestURL = strings.Replace(requestURL, \"{endpoint}\", endpoint, -1)\n\n\t// Query String\n\tqs := url.Values{}\n\tfor k, v := range params {\n\t\tqs.Add(k, fmt.Sprintf(\"%v\", v))\n\t}\n\n\tif len(qs) > 0 {\n\t\trequestURL += \"?\" + qs.Encode()\n\t}\n\n\treturn c.newRequestFromURL(requestURL, method, body)\n}", "func (c *apiHttp) NewRequest(command string) RequestBuilder {\n\theaders := make(map[string]string)\n\tif c.Headers != nil {\n\t\tfor k := range c.Headers {\n\t\t\theaders[k] = c.Headers.Get(k)\n\t\t}\n\t}\n\n\treturn &requestBuilder{\n\t\tcommand: command,\n\t\tclient: c,\n\t\theaders: headers,\n\t}\n}", "func (c client) newRequest(ctx context.Context, method string, url string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequestWithContext(ctx, method, url, body)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"unable to build request %w\", err)\n\t}\n\n\treturn req, nil\n}", "func newRequest(method, url string, headers map[string]string, bd io.Reader) *http.Request {\n\treq, err := http.NewRequest(method, url, bd)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\t// extract the escaped original host+path from url\n\t// http://localhost/path/here?v=1#frag -> //localhost/path/here\n\topaque := \"\"\n\tif i := len(req.URL.Scheme); i > 0 {\n\t\topaque = url[i+1:]\n\t}\n\n\tif i := strings.LastIndex(opaque, \"?\"); i > -1 {\n\t\topaque = opaque[:i]\n\t}\n\tif i := strings.LastIndex(opaque, \"#\"); i > -1 {\n\t\topaque = opaque[:i]\n\t}\n\t// Escaped host+path workaround as detailed in https://golang.org/pkg/net/url/#URL\n\t// for < 1.5 client side workaround\n\treq.URL.Opaque = opaque\n\n\t// Simulate writing to wire\n\tvar buff bytes.Buffer\n\treq.Write(&buff)\n\tioreader := bufio.NewReader(&buff)\n\n\t// Parse request off of 'wire'\n\treq, err = http.ReadRequest(ioreader)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tfor key, value := range headers {\n\t\treq.Header.Add(key, value)\n\t}\n\n\treturn req\n}", "func (qiwi *PersonalAPI) newRequest(apiKey, method, spath string, data map[string]interface{}) (req *http.Request, err error) {\n\n\tvar path = APIURL + spath\n\n\tvar body io.Reader\n\n\tif len(data) > 0 {\n\n\t\ts, err := json.Marshal(data)\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tbody = bytes.NewBuffer(s)\n\n\t}\n\n\treq, err = http.NewRequest(method, path, body)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"Bearer \"+apiKey)\n\n\treturn req, err\n}", "func (s *VirtualhostsService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v7.1.0\")\n\treturn req, nil\n}", "func NewRequest(r *http.Request, ns string, body []byte) (*Request, error) {\n\th, err := json.Marshal(r.Header)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Request{\n\t\tID: uuid.NewV4(),\n\t\tMethod: r.Method,\n\t\tHeaders: h,\n\t\tPayload: body,\n\t\tHost: r.Host,\n\t\tScheme: r.URL.Scheme,\n\t\tPath: r.URL.Path,\n\t\tNamespace: ns,\n\t\tRetries: 0,\n\t}, nil\n}", "func NewPost(url string) *Request { return NewRequest(\"POST\", url) }", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v8.1.0\")\n\treturn req, nil\n}", "func NewRequest(apiKey string, query *Tag) *Tag {\n\treturn Request().Add(\n\t\tLogin().Opts(Opts{\n\t\t\tOptAuthenticationKey: apiKey,\n\t\t}),\n\t\tquery,\n\t)\n}", "func newRequest(ctx context.Context, url string, reqBody io.Reader, contentType string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"POST\", url, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq = req.WithContext(ctx)\n\tif customHeader := getCustomHTTPReqHeaders(ctx); customHeader != nil {\n\t\treq.Header = customHeader\n\t}\n\treq.Header.Set(\"Accept\", contentType)\n\treq.Header.Set(\"Content-Type\", contentType)\n\treq.Header.Set(\"Twirp-Version\", \"v5.4.1\")\n\treturn req, nil\n}", "func (g *Group) newRequest(operation *request.Operation, input interface{}, output interface{}, handlers *request.Handlers) *request.Request {\n return request.New(g.Client, operation, input, output, handlers)\n}", "func (c *Client) newRequest(method, path string, v interface{}, ctype string) (req *http.Request, err error) {\n\t// Build request JSON.\n\tbody, err := writeJson(v)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq, err = http.NewRequest(method, c.pathToEndPoint(path), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Add(\"X-Kii-AppID\", c.AppId)\n\treq.Header.Add(\"X-Kii-AppKey\", c.AppKey)\n\tif ctype != \"\" {\n\t\treq.Header.Add(\"Content-Type\", ctype)\n\t}\n\tif c.Authorization != \"\" {\n\t\treq.Header.Add(\"Authorization\", c.Authorization)\n\t}\n\treturn\n}", "func (c *Client) newPost(endpoint string, reqBody []byte) (*http.Request, error) {\n\tcurl, err := c.getURL(endpoint)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq, err := http.NewRequest(http.MethodPost, curl, bytes.NewReader(reqBody))\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"Failed posting to %s\", curl)\n\t}\n\treturn req, nil\n}", "func CreateBaseRequest(method, url string, body []byte, user m.AuthUser, verbose bool) *http.Request {\n\treq, err := http.NewRequest(method, url, bytes.NewBuffer(body))\n\treq.SetBasicAuth(user.Username, user.Password)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Accept\", \"application/json\")\n\tError(err, \"Error creating the request\")\n\n\tif verbose {\n\t\tfmt.Println(\"Request Url:\", req.URL)\n\t\tfmt.Println(\"Request Headers:\", req.Header)\n\t\tfmt.Println(\"Request Body:\", req.Body)\n\t}\n\n\treturn req\n}", "func (c *Client) NewRequest(method, urlStr string, body io.Reader) (*http.Request, error) {\n\trel, err := url.Parse(path.Join(apiVersion, urlStr))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.baseURL.ResolveReference(rel)\n\treq, err := http.NewRequest(method, u.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Authorization\", \"Bearer \"+c.config.accessToken)\n\tif c.userAgent != \"\" {\n\t\treq.Header.Set(\"User-Agent\", c.userAgent)\n\t}\n\treturn req, nil\n}", "func (s *DefaultClient) NewRequest(method, endpoint string, payload io.Reader) (request *http.Request, err error) {\n\turl := fmt.Sprintf(\"%s/%s\", s.Host, endpoint)\n\trequest, err = http.NewRequest(method, url, payload)\n\tif err != nil {\n\t\treturn\n\t}\n\n\trequest.SetBasicAuth(s.Username, s.Password)\n\tif payload != nil {\n\t\trequest.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\n\treturn\n}", "func (c *Executor) newHTTPRequest(request *Request) (*http.Request, error) {\n\tif request.context == nil {\n\t\treturn nil, errNilContext\n\t}\n\tif !isSupportedHTTPMethod(request.method) {\n\t\treturn nil, fmt.Errorf(\"error executing request, found unsupport HTTP method %s\", request.method)\n\t}\n\n\t// JSON encode the object and use that as the body if specified, otherwise use the body as-is\n\treqBody := request.body\n\tif request.object != nil {\n\t\tb, err := encodeBody(request.object)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"error executing request, failed to encode the request object to JSON: %w\", err)\n\t\t}\n\t\treqBody = b\n\t}\n\tu := path.Join(c.apiAddress, request.pathAndQuery)\n\n\tr, err := http.NewRequestWithContext(request.context, request.method, u, reqBody)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error executing request, failed to create a new underlying HTTP request: %w\", err)\n\t}\n\tr.Header.Set(\"User-Agent\", c.userAgent)\n\tif request.contentType != \"\" {\n\t\tr.Header.Set(\"Content-type\", request.contentType)\n\t}\n\tif request.contentLength != nil {\n\t\tr.ContentLength = *request.contentLength\n\t}\n\tfor k, v := range request.headers {\n\t\tr.Header.Set(k, v)\n\t}\n\n\treturn r, nil\n}", "func newRestRequest(restConfig RestConfig, restreq RestRequest) (*http.Request, error) {\n\treq, err := http.NewRequest(restreq.Method, restreq.Path, bytes.NewReader(restreq.Body))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif restConfig.AppID != \"\" {\n\t\treq.Header.Add(\"X-LC-Id\", restConfig.AppID)\n\t\treq.Header.Add(\"X-LC-Key\", restConfig.RestKey)\n\t}\n\t//Append Token if given\n\tif restreq.Token != \"\" {\n\t\treq.Header.Add(\"X-LC-Session\", restreq.Token)\n\t}\n\n\tif restreq.Type != \"\" {\n\t\treq.Header.Add(\"Content-Type\", restreq.Type)\n\t} else {\n\t\t//fallback to text/plain\n\t\treq.Header.Add(\"Content-Type\", \"text/plain\")\n\t}\n\t//log.Println(req)\n\treturn req, nil\n}", "func (r *RequestBuilder) Build(httpVerb string, uri string, body io.Reader) (*http.Request, error) {\n\turl, err := r.getAbsoluteURL(uri)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trequest, err := http.NewRequest(httpVerb, url, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\trequest.Header.Add(\"Content-Type\", \"application/json\")\n\trequest.Header.Add(\"User-Agent\", r.userAgent)\n\n\tif len(r.token) > 0 {\n\t\trequest.Header.Add(\"Authorization\", fmt.Sprintf(\"Bearer %s\", r.token))\n\t}\n\n\treturn request, nil\n}", "func newRequest(r *http.Request, vars map[string]string) *Request {\n\trequest := &Request{r, vars, nil}\n\treturn request\n}", "func (c *APIClient) buildRequest(\n\tpath, method string,\n\tpostBody interface{},\n\theaders map[string]string,\n\tqueryParams url.Values) (localVarRequest *http.Request, err error) {\n\n\tvar body *bytes.Buffer\n\n\t// Detect postBody type and post.\n\tif postBody != nil {\n\t\tcontentType := headers[\"Content-Type\"]\n\t\tif contentType == \"\" {\n\t\t\tcontentType = detectContentType(postBody)\n\t\t\theaders[\"Content-Type\"] = contentType\n\t\t}\n\n\t\tbody, err = setBody(postBody, contentType)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Setup path and query parameters\n\turl, err := url.Parse(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\turl.Scheme = c.cfg.Scheme\n\turl.Host = c.cfg.Host\n\n\t// Adding Query Param\n\tquery := url.Query()\n\tfor k, v := range queryParams {\n\t\tfor _, iv := range v {\n\t\t\tquery.Add(k, iv)\n\t\t}\n\t}\n\n\t// Encode the parameters.\n\turl.RawQuery = query.Encode()\n\n\t// Generate a new request\n\tif body != nil {\n\t\tlocalVarRequest, err = http.NewRequest(method, url.String(), body)\n\t} else {\n\t\tlocalVarRequest, err = http.NewRequest(method, url.String(), nil)\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// add header parameters, if any\n\tif len(headers) > 0 {\n\t\tfinalheaders := http.Header{}\n\t\tfor h, v := range headers {\n\t\t\tfinalheaders.Set(h, v)\n\t\t}\n\t\tlocalVarRequest.Header = finalheaders\n\t}\n\n\t// Override request host, if applicable\n\tif c.cfg.Host != \"\" {\n\t\tlocalVarRequest.Host = c.cfg.Host\n\t}\n\n\t// Add the user agent to the request.\n\tlocalVarRequest.Header.Add(\"User-Agent\", c.cfg.UserAgent)\n\n\tfor header, value := range c.cfg.DefaultHeader {\n\t\tlocalVarRequest.Header.Add(header, value)\n\t}\n\n\treturn localVarRequest, nil\n}", "func (conn Connection) newRequest(method, cmd string, body io.Reader) *http.Request {\n\treq, err := http.NewRequest(method, conn.ServiceURL+cmd, body)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Coulnd't generate HTTP request - %s\\n\", err.Error()))\n\t}\n\n\tfor k, v := range conn.Headers {\n\t\treq.Header.Add(k, v)\n\t}\n\n\treturn req\n}", "func WithHost(host string) Preparer {\n\treturn func(r *http.Request) *http.Request {\n\t\tr.Host = host\n\t\treturn r\n\t}\n}", "func (c *Client) newRequest(url string) (*http.Request, error) {\n\treq, err := http.NewRequest(\"GET\", url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"user-key\", c.key)\n\treq.Header.Add(\"Accept\", \"application/json\")\n\n\treturn req, nil\n}", "func (k *API) NewRequest(url, method string, body url.Values) (*http.Request, error) {\n\n\tb := bytes.NewBufferString(body.Encode())\n\treq, err := http.NewRequest(method, url, b)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"make korbit req\")\n\t}\n\n\ttoken := fmt.Sprintf(\"%s %s\", k.Token.TokenType, k.Token.AccessToken)\n\treq.Header.Set(\"Authorization\", token)\n\n\tif method == \"POST\" {\n\t\treq.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\t}\n\n\treturn req, nil\n}", "func NewRequest(url string) *Request {\n\treturn client.NewRequest(url)\n}", "func (c *Client) NewRequest(method, urlStr string, body interface{}) (*http.Request, error) {\n\trel, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.BaseURL.ResolveReference(rel)\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\terr := json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tauthHeader := \"Bearer \" + c.Option.APIKey\n\n\treq.Header.Set(\"Authorization\", authHeader)\n\tif c.UserAgent != \"\" {\n\t\treq.Header.Add(\"User-Agent\", c.UserAgent)\n\t}\n\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, path string, reqBody interface{}, needAuth bool) (*http.Request, error) {\n\tvar body []byte\n\tvar err error\n\n\tif reqBody != nil {\n\t\tbody, err = json.Marshal(reqBody)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\ttarget := fmt.Sprintf(\"%s%s\", c.APIEndPoint, path)\n\treq, err := http.NewRequest(method, target, bytes.NewReader(body))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Inject headers\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", \"application/json;charset=utf-8\")\n\t}\n\treq.Header.Set(\"Authorization\", fmt.Sprintf(\"sso-key %s:%s\", c.APIKey, c.APISecret))\n\treq.Header.Set(\"Accept\", \"application/json\")\n\treq.Header.Set(\"User-Agent\", \"ExternalDNS/\"+externaldns.Version)\n\n\t// Send the request with requested timeout\n\tc.Client.Timeout = c.Timeout\n\n\treturn req, nil\n}", "func (g *baseGithub) newRequest(method, url string, body interface{}) (*http.Request, error) {\n\tvar err error\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\tenc := json.NewEncoder(buf)\n\t\tenc.SetEscapeHTML(false)\n\t\terr = enc.Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, url, buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\treq.Header.Set(\"Accept\", \"application/json\")\n\n\treturn req, nil\n}", "func (c *Client) newRequest(method, path string) (*request, error) {\n\tbase, _ := url.Parse(c.config.Address)\n\tu, err := url.Parse(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tr := &request{\n\t\tconfig: &c.config,\n\t\tmethod: method,\n\t\turl: &url.URL{\n\t\t\tScheme: base.Scheme,\n\t\t\tUser: base.User,\n\t\t\tHost: base.Host,\n\t\t\tPath: u.Path,\n\t\t\tRawPath: u.RawPath,\n\t\t},\n\t\theader: make(http.Header),\n\t\tparams: make(map[string][]string),\n\t}\n\tif c.config.Region != \"\" {\n\t\tr.params.Set(\"region\", c.config.Region)\n\t}\n\tif c.config.Namespace != \"\" {\n\t\tr.params.Set(\"namespace\", c.config.Namespace)\n\t}\n\tif c.config.WaitTime != 0 {\n\t\tr.params.Set(\"wait\", durToMsec(r.config.WaitTime))\n\t}\n\tif c.config.SecretID != \"\" {\n\t\tr.token = r.config.SecretID\n\t}\n\n\t// Add in the query parameters, if any\n\tfor key, values := range u.Query() {\n\t\tfor _, value := range values {\n\t\t\tr.params.Add(key, value)\n\t\t}\n\t}\n\n\tfor key, values := range c.config.Headers {\n\t\tr.header[key] = values\n\t}\n\n\treturn r, nil\n}", "func (s *Stein) newRequest(method string, path string, body io.Reader) (*http.Request, error) {\n\treq, err := http.NewRequest(method, path, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\n\treturn req, nil\n}", "func (g *GoFlickr) newRequest(apiMethod string) ApiRequest {\n\n\treq := ApiRequest{\n\t\tMethodName: apiMethod,\n\t}\n\treq.addParam(\"api_key\", g.ApiKey)\n\treq.addParam(\"format\", \"json\")\n\treq.addParam(\"nojsoncallback\", \"1\")\n\treturn req\n\n}", "func (s APIv1) NewRequest(ctx context.Context, method, path string, data interface{}) (req *http.Request, err error) {\n\t// Resolve the URL reference from the path\n\tendpoint := s.endpoint.ResolveReference(&url.URL{Path: path})\n\n\tvar body io.ReadWriter\n\tif data != nil {\n\t\tbody = &bytes.Buffer{}\n\t\tif err = json.NewEncoder(body).Encode(data); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"could not serialize request data: %s\", err)\n\t\t}\n\t} else {\n\t\tbody = nil\n\t}\n\n\t// Create the http request\n\tif req, err = http.NewRequestWithContext(ctx, method, endpoint.String(), body); err != nil {\n\t\treturn nil, fmt.Errorf(\"could not create request: %s\", err)\n\t}\n\n\t// Set the headers on the request\n\treq.Header.Add(\"User-Agent\", \"Whisper/1.0\")\n\treq.Header.Add(\"Accept\", \"application/json\")\n\treq.Header.Add(\"Accept-Language\", \"en-US,en\")\n\treq.Header.Add(\"Accept-Encoding\", \"gzip, deflate, br\")\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\n\treturn req, nil\n}", "func newRequest(db *Database) *Request {\n\treq := &Request{\n\t\tdb: db,\n\t\tpath: \"/\",\n\t\tquery: url.Values{},\n\t\theader: http.Header{},\n\t}\n\treturn req\n}", "func NewCreateanewDomainRequestWithBody(server string, contentType string, body io.Reader) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/domains\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"POST\", queryUrl.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", contentType)\n\treturn req, nil\n}", "func (adm AdminClient) newRequest(method string, reqData requestData) (req *http.Request, err error) {\n\t// If no method is supplied default to 'POST'.\n\tif method == \"\" {\n\t\tmethod = \"POST\"\n\t}\n\n\t// Default all requests to \"\"\n\tlocation := \"\"\n\n\t// Construct a new target URL.\n\ttargetURL, err := adm.makeTargetURL(reqData)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Initialize a new HTTP request for the method.\n\treq, err = http.NewRequest(method, targetURL.String(), nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tadm.setUserAgent(req)\n\tfor k, v := range reqData.customHeaders {\n\t\treq.Header.Set(k, v[0])\n\t}\n\tif length := len(reqData.content); length > 0 {\n\t\treq.ContentLength = int64(length)\n\t}\n\treq.Header.Set(\"X-Amz-Content-Sha256\", hex.EncodeToString(sum256(reqData.content)))\n\treq.Body = ioutil.NopCloser(bytes.NewReader(reqData.content))\n\n\treq = s3signer.SignV4(*req, adm.accessKeyID, adm.secretAccessKey, \"\", location)\n\treturn req, nil\n}", "func newRequest(method, url string, body io.Reader, headers http.Header) (*http.Request, error) {\n\treq, err := http.NewRequest(method, url, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfor key, value := range headers {\n\t\treq.Header.Add(key, value[0])\n\t}\n\treturn req, nil\n}", "func (self *Client) newRequest(url *url.URL, method string, body io.Reader) (*http.Request, error) {\n\trc, ok := body.(io.ReadCloser)\n\tif !ok && body != nil {\n\t\trc = ioutil.NopCloser(body)\n\t}\n\n\treq := &http.Request{\n\t\tMethod: method,\n\t\tURL: url,\n\t\tProto: \"HTTP/1.1\",\n\t\tProtoMajor: 1,\n\t\tProtoMinor: 1,\n\t\tHeader: make(http.Header),\n\t\tBody: rc,\n\t\tHost: url.Host,\n\t}\n\n\tif body != nil {\n\t\tswitch v := body.(type) {\n\t\tcase *bytes.Buffer:\n\t\t\treq.ContentLength = int64(v.Len())\n\t\tcase *bytes.Reader:\n\t\t\treq.ContentLength = int64(v.Len())\n\t\tcase *strings.Reader:\n\t\t\treq.ContentLength = int64(v.Len())\n\t\t}\n\t}\n\treturn req, nil\n}", "func (r *request) toHTTP() (*http.Request, error) {\n\t// Encode the query parameters\n\tr.url.RawQuery = r.params.Encode()\n\n\t// Check if we should encode the body\n\tif r.body == nil && r.obj != nil {\n\t\tif b, err := encodeBody(r.obj); err != nil {\n\t\t\treturn nil, err\n\t\t} else {\n\t\t\tr.body = b\n\t\t}\n\t}\n\n\tctx := func() context.Context {\n\t\tif r.ctx != nil {\n\t\t\treturn r.ctx\n\t\t}\n\t\treturn context.Background()\n\t}()\n\n\t// Create the HTTP request\n\treq, err := http.NewRequestWithContext(ctx, r.method, r.url.RequestURI(), r.body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header = r.header\n\n\t// Optionally configure HTTP basic authentication\n\tif r.url.User != nil {\n\t\tusername := r.url.User.Username()\n\t\tpassword, _ := r.url.User.Password()\n\t\treq.SetBasicAuth(username, password)\n\t} else if r.config.HttpAuth != nil {\n\t\treq.SetBasicAuth(r.config.HttpAuth.Username, r.config.HttpAuth.Password)\n\t}\n\n\treq.Header.Add(\"Accept-Encoding\", \"gzip\")\n\tif r.token != \"\" {\n\t\treq.Header.Set(\"X-Nomad-Token\", r.token)\n\t}\n\n\treq.URL.Host = r.url.Host\n\treq.URL.Scheme = r.url.Scheme\n\treq.Host = r.url.Host\n\treturn req, nil\n}", "func NewRequest(token string) *Request {\n\treqID := fmt.Sprintf(\"alaudacli-%d\", time.Now().Unix())\n\trestyReq := resty.R()\n\trestyReq.SetHeader(\"Content-Type\", \"application/json\")\n\trestyReq.SetHeader(\"Alauda-Request-ID\", reqID)\n\n\tif token != \"\" {\n\t\trestyReq.SetHeader(\"Authorization\", fmt.Sprintf(\"Token %s\", token))\n\t}\n\n\treturn &Request{\n\t\trestyReq,\n\t}\n}", "func (c Client) createRequest(method string, endpoint string, params url.Values) (req *http.Request, err error) {\n\tendpoint = APIBaseURL + endpoint\n\t// get current unix time in MS\n\tts := time.Now().UnixNano() / int64(time.Millisecond)\n\t// add required timestamp value\n\tparams.Add(\"timestamp\", fmt.Sprint(ts))\n\tif method == \"GET\" || method == \"DELETE\" {\n\t\treq, err = http.NewRequest(method, fmt.Sprintf(\"%s?%s\", endpoint, params.Encode()), nil)\n\t} else {\n\t\treq, err = http.NewRequest(method, endpoint, strings.NewReader(params.Encode()))\n\t\treq.Header.Add(\"Content-Length\", strconv.Itoa(len(params.Encode())))\n\t}\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif strings.Contains(endpoint, \"/private\") {\n\t\tc.authenticate(req, params)\n\t}\n\treq.Header.Set(\"User-Agent\", APIUserAgent)\n\treq.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\n\treturn req, nil\n}", "func NewRequest(method, url string, payload interface{}) (*http.Request, error) {\n\tvar buf io.Reader\n\tif payload != nil {\n\t\tvar b []byte\n\t\tb, err := json.Marshal(&payload)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tbuf = bytes.NewBuffer(b)\n\t}\n\treturn http.NewRequest(method, url, buf)\n}", "func newRequest(req *http.Request) *Request {\n\trequest := &Request{\n\t\tRequest: req,\n\t}\n\n\treturn request\n}", "func (c *Client) NewRequest(method, urlStr string, body interface{}) (*http.Request, error) {\n\tif !strings.HasSuffix(c.baseURL.Path, \"/\") {\n\t\treturn nil, fmt.Errorf(\"BaseURL must have a trailing slash, but %q does not\", c.baseURL)\n\t}\n\tu, err := c.baseURL.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\terr := json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\treq.Header.Set(\"Accept\", \"application/json; charset=utf-8\")\n\treq.Header.Set( \"x-api-key\", c.opts.apiKey)\n\treq.Header.Set(\"Host\", \"openapi.etsy.com\")\n\treq.Header.Set(\"User-Agent\", c.UserAgent)\n\n\treturn req, nil\n}", "func NewRequest(requestName string, params rata.Params, header http.Header, query url.Values, body ...io.Reader) Request {\n\tif header == nil {\n\t\theader = http.Header{}\n\t}\n\theader.Set(\"Accept\", \"application/json\")\n\n\trequest := Request{\n\t\tRequestName: requestName,\n\t\tParams: params,\n\t\tHeader: header,\n\t\tQuery: query,\n\t}\n\n\tif len(body) == 1 {\n\t\trequest.Body = body[0]\n\t}\n\n\treturn request\n}", "func (c *Client) NewRequest(method, urlStr string, body interface{}) (*http.Request, error) {\n\trel, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.baseURL.ResolveReference(rel)\n\n\tbuf := new(bytes.Buffer)\n\tif body != nil {\n\t\terr := json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Authorization\", \"Bearer \"+c.authToken)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treturn req, nil\n}", "func NewRequest(addr string, body []byte) *Request {\n\treturn &Request{\n\t\tAddr: addr,\n\t\tHeader: defaultHeader(len(body)),\n\t\tBody: body,\n\t}\n}", "func (h *HTTPClient) newReq(method, token, projID, path string, body io.Reader) (*http.Request, error) {\n\turlStr := fmt.Sprintf(\"%s://%s:%d/3/projects/%s/%s\", h.scheme, h.host, h.port, projID, path)\n\treq, err := http.NewRequest(method, urlStr, body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Authorization\", \"OAuth \"+token)\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body interface{}) (*http.Request, error) {\n\tvar s string\n\tif strings.HasPrefix(urlStr, \"/\") {\n\t\ts = fmt.Sprintf(\"/%s%s\", apiVersion, urlStr)\n\t} else {\n\t\ts = fmt.Sprintf(\"/%s/%s\", apiVersion, urlStr)\n\t}\n\trel, err := url.Parse(s)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.BaseURL.ResolveReference(rel)\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\terr = json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", \"application/json\")\n\t}\n\treturn req, nil\n}", "func (c *APIGateway) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func newP(method, url, token string, body interface{}) (*http.Request, error) {\n\treader, err := encode(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\trequest, err := NewRequest(method, url, reader)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbearer := fmt.Sprintf(\"Bearer %s\", token)\n\trequest.Header.Add(\"Authorization\", bearer)\n\trequest.Header.Add(\"Accept\", \"application/json\")\n\trequest.Header.Add(\"Content-Type\", \"application/json\")\n\treturn request, nil\n}", "func (r *vtmClient) buildAPIRequest(method, uri string, reader io.Reader) (request *http.Request, err error) {\n\t// Create the endpoint URL\n\turl := fmt.Sprintf(\"%s/%s\", r.config.URL, uri)\n\n\t// Make the http request to VTM\n\trequest, err = http.NewRequest(method, url, reader)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Add any basic auth and the content headers\n\tif r.config.HTTPBasicAuthUser != \"\" && r.config.HTTPBasicPassword != \"\" {\n\t\trequest.SetBasicAuth(r.config.HTTPBasicAuthUser, r.config.HTTPBasicPassword)\n\t}\n\n\trequest.Header.Add(\"Content-Type\", \"application/json\")\n\trequest.Header.Add(\"Accept\", \"application/json\")\n\n\treturn request, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body io.Reader) (r *http.Request, err error) {\n\tif r, err = http.NewRequest(method, urlStr, body); err != nil {\n\t\treturn\n\t}\n\n\tr.Header.Set(\"User-Agent\", c.UserAgent)\n\tr.Header.Set(\"X-Requested-With\", \"XMLHttpRequest\")\n\n\tif c.ApiKey != nil {\n\t\tr.Header.Set(\"X-Sbss-Auth\", c.ApiKey.Login)\n\t\tr.AddCookie(c.ApiKey.Cookie)\n\t}\n\n\tif method == \"POST\" {\n\t\tr.Header.Set(\"Content-Type\", \"application/x-www-form-urlencoded\")\n\t}\n\n\treturn\n}", "func (c *PaperTrailClient) NewRequest(url string) (*http.Request, error) {\n\treq, err := http.NewRequest(http.MethodGet, url, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"X-Papertrail-Token\", c.token)\n\treturn req, nil\n}", "func (s *SitesService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *DockerHubClient) NewRequest(method, path string, body interface{}) (*http.Request, error) {\n\trelPath, err := url.Parse(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.HubURL.ResolveReference(relPath)\n\n\tbuf := new(bytes.Buffer)\n\tif body != nil {\n\t\terr := json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Accept\", mediaType)\n\treq.Header.Add(\"Content-Type\", mediaType)\n\treq.Header.Add(\"User-Agent\", c.UserAgent)\n\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body interface{}) (*http.Request, error) {\n\trel, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.baseURL.ResolveReference(rel)\n\n\tvar data io.Reader\n\tif body != nil {\n\t\tb, err := json.Marshal(body)\n\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tdata = bytes.NewReader(b)\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"Accept\", \"application/json\")\n\n\t// Set session cookie if there is one\n\tif c.session != nil {\n\t\treq.Header.Set(\"Authorization\", c.session.Authorization)\n\t\t// TODO: remove this\n\t\t//fmt.Printf(c.session.Authorization)\n\t}\n\n\treturn req, nil\n}", "func (c *InputService7ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *InputService7ProtocolTest) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func BuildRequest(method, url string, body []byte, headers http.Header) *http.Request {\n\tvar bodyBuffer io.Reader\n\tif body != nil {\n\t\tbodyBuffer = bytes.NewBuffer(body)\n\t}\n\tr, err := http.NewRequest(method, url, bodyBuffer)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tr.Header = headers\n\treturn r\n}", "func newFileUploadRequest(hostBase string, organization string, application string, path string) (*http.Response, *string, error) {\n\tfile, err := os.Open(path)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tdefer file.Close()\n\n\tbody := &bytes.Buffer{}\n\twriter := multipart.NewWriter(body)\n\tpart, err := writer.CreateFormFile(\"file\", filepath.Base(path))\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\t_, err = io.Copy(part, file)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\twriter.WriteField(\"name\", application)\n\n\t//set the content type\n\twriter.FormDataContentType()\n\n\terr = writer.Close()\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\turi := fmt.Sprintf(\"%s/%s/apps\", hostBase, organization)\n\n\trequest, err := http.NewRequest(\"POST\", uri, body)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\trequest.Header.Set(\"Content-Type\", writer.FormDataContentType())\n\trequest.Header.Set(\"Authorization\", fmt.Sprintf(\"Bearer %s\", \"e30K.e30K.e30K\"))\n\n\tclient := &http.Client{\n\t\tTimeout: 120 * time.Second,\n\t}\n\n\tresponse, err := client.Do(request)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tbodyBytes, err := ioutil.ReadAll(response.Body)\n\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\n\tbodyResponse := string(bodyBytes)\n\n\treturn response, &bodyResponse, nil\n}", "func (c *Client) NewRequest(method, urlPath string, body interface{}) (*http.Request, error) {\n\t// Parse our URL.\n\trel, err := url.Parse(urlPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Resolve to absolute URI.\n\tu := c.BackendURL.ResolveReference(rel)\n\n\tbuf := new(bytes.Buffer)\n\tif body != nil {\n\t\terr = json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\t// Create the request.\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Add our packages UA.\n\treq.Header.Add(\"User-Agent\", c.UserAgent)\n\n\treturn req, nil\n}", "func (s *HighAvailabilityService) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := s.NewRequest(op, params, data)\n\n\treturn req\n}", "func NewCreateElasticIpRequest(server string, body CreateElasticIpJSONRequestBody) (*http.Request, error) {\n\tvar bodyReader io.Reader\n\tbuf, err := json.Marshal(body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tbodyReader = bytes.NewReader(buf)\n\treturn NewCreateElasticIpRequestWithBody(server, \"application/json\", bodyReader)\n}", "func (c *Client) NewRequest(method, urlStr string, opt interface{}, body interface{}) (*http.Request, error) {\n\trel, err := addOptions(urlStr, opt)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif AuthTest {\n\t\t// Add the auth_test param\n\t\tvalues := rel.Query()\n\t\tvalues.Add(\"auth_test\", strconv.FormatBool(AuthTest))\n\t\trel.RawQuery = values.Encode()\n\t}\n\n\tu := c.BaseURL.ResolveReference(rel)\n\n\tbuf := new(bytes.Buffer)\n\tif body != nil {\n\t\terr := json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Authorization\", \"Bearer \"+c.authToken)\n\treq.Header.Add(\"Content-Type\", \"application/json\")\n\treturn req, nil\n}", "func newRequest(req *http.Request) *Request {\n\treturn &Request{\n\t\tRequest: req,\n\t}\n}", "func buildHTTPRequest(method, url string, payload *bytes.Buffer) (*http.Request, error) {\n\t// Build whole HTTP request with payload data\n\tnewReq, err := http.NewRequest(method, url, payload)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Set default headers for new request\n\tnewReq.Header.Set(\"Accept\", \"application/json\")\n\tnewReq.Header.Set(\"Content-type\", \"application/json\")\n\n\treturn newReq, nil\n}", "func (c *Client) NewRequest(method, path string, body []byte) (*http.Request, error) {\n\n\tfullURI := baseURI + path\n\n\tvar reqBody io.Reader\n\tif body != nil {\n\t\treqBody = bytes.NewReader(body)\n\t}\n\n\treq, err := http.NewRequest(method, fullURI, reqBody)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Authorization\", \"Bearer: \"+c.token)\n\treq.Header.Add(\"User-Agent\", userAgent)\n\n\treturn req, nil\n}", "func newfileUploadRequest(uri string, headers map[string]string, paramName, path string) (*http.Request, error) {\n\tfile, err := os.Open(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdefer file.Close()\n\n\tbody := &bytes.Buffer{}\n\twriter := multipart.NewWriter(body)\n\tpart, err := writer.CreateFormFile(paramName, filepath.Base(path))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_, err = io.Copy(part, file)\n\n\terr = writer.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Does the request with the headers\n\treq, err := http.NewRequest(\"POST\", uri+\"/api/v2/upload\", body)\n\tfor key, val := range headers {\n\t\treq.Header.Set(key, val)\n\t}\n\treq.Header.Set(\"Content-Type\", writer.FormDataContentType())\n\treq.Header.Set(\"User-Agent\", \"Share-CLI/1.0\")\n\treturn req, err\n}", "func newRequest(method, path string, body io.Reader) (*http.Request, error) {\n\tr := http.Request{\n\t\tMethod: method,\n\t\tURL: &url.URL{Path: path},\n\t\tProto: \"HTTP/1.1\",\n\t\tProtoMajor: 1,\n\t\tProtoMinor: 1,\n\t\tHeader: make(http.Header),\n\t}\n\n\tif body != nil {\n\t\tswitch b := body.(type) {\n\t\tcase *bytes.Buffer:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t\tr.ContentLength = int64(b.Len())\n\t\tcase *bytes.Reader:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t\tr.ContentLength = int64(b.Len())\n\t\tcase *strings.Reader:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t\tr.ContentLength = int64(b.Len())\n\t\tdefault:\n\t\t\tr.Body = ioutil.NopCloser(body)\n\t\t}\n\t}\n\n\treturn &r, nil\n}", "func (c *Client) NewRequest(ctx context.Context, method, endpointURL string, body interface{}) (*http.Request, error) {\n\tu := c.requestURL(method, endpointURL)\n\n\tbuf := new(bytes.Buffer)\n\tif body != nil {\n\t\terr := json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", mediaType)\n\treq.Header.Set(\"User-Agent\", c.UserAgent)\n\n\treturn req, nil\n}", "func (c *Client) NewRequest(ctx context.Context, method string, urlStr string, body interface{}) (*http.Request, error) {\n\trelURL, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tu := c.BaseURL.ResolveReference(relURL)\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\terr = json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq = req.WithContext(ctx)\n\treq.Header.Add(\"Authorization\", fmt.Sprintf(\"Bearer %s\", c.token))\n\n\tif c.UserAgent != \"\" {\n\t\treq.Header.Set(\"User-Agent\", c.UserAgent)\n\t}\n\n\tif body != nil {\n\t\treq.Header.Set(\"Content-Type\", mediaTypeJSON)\n\t}\n\n\treturn req, err\n}", "func newRequest(t *testing.T, path string) *http.Request {\n\tif !strings.HasPrefix(path, \"/\") {\n\t\tpath = \"/\" + path\n\t}\n\tu := fmt.Sprintf(\"http://localhost:%d%s\", expectedPort, path)\n\treq, err := http.NewRequestWithContext(\n\t\tcontext.Background(),\n\t\thttp.MethodGet, u,\n\t\tnil,\n\t)\n\trequire.NoError(t, err)\n\treturn req\n}", "func setupRequest(apiID string, apiKey string, argMethod string, argURI string, argPostBody []byte) (*http.Request, error) {\n\turi := atlasURI + argURI\n\turl := atlasHost + uri\n\temptyRequest := http.Request{}\n\treq, err := http.NewRequest(argMethod, url, nil)\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\tclient := &http.Client{}\n\tresp, err := client.Do(req)\n\tif err != nil {\n\t\tlog.Printf(\"Error - setupRequest - Failed http response. Resp: %+v, Err: %+v\", resp, err)\n\t\treturn &emptyRequest, err\n\t}\n\tdefer resp.Body.Close()\n\tdigestParts := digestParts(resp)\n\tdigestParts[\"uri\"] = uri\n\tdigestParts[\"method\"] = argMethod\n\n\tusername := apiID\n\tif len(username) == 0 {\n\t\terr := fmt.Errorf(\"apiID variable not set\")\n\t\tlog.Printf(\"Error - setupRequest. Err: %+v\", err)\n\t\treturn &emptyRequest, err\n\t}\n\n\tpassword := apiKey\n\tif len(password) == 0 {\n\t\terr := fmt.Errorf(\"apiKey variable not set\")\n\t\tlog.Printf(\"Error - setupRequest. Err: %+v\", err)\n\t\treturn &emptyRequest, err\n\t}\n\n\tdigestParts[\"username\"] = username\n\tdigestParts[\"password\"] = password\n\tif argPostBody == nil {\n\t\treq, err = http.NewRequest(argMethod, url, nil)\n\t} else {\n\t\treq, err = http.NewRequest(argMethod, url, bytes.NewBuffer(argPostBody))\n\t}\n\treq.Header.Set(\"Authorization\", getDigestAuthrization(digestParts))\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treturn req, nil\n\n}", "func NewRequest(r *http.Request) *Request {\n\tvar request Request\n\trequest.ID = atomic.AddUint32(&requestID, 1)\n\trequest.Method = r.Method\n\trequest.Body = r.Body\n\trequest.BodyBuff = new(bytes.Buffer)\n\trequest.BodyBuff.ReadFrom(r.Body)\n\trequest.RemoteAddr = r.Header.Get(\"X-Forwarded-For\")\n\trequest.Header = r.Header\n\tif request.RemoteAddr == \"\" {\n\t\trequest.RemoteAddr = r.RemoteAddr\n\t}\n\trequest.UrlParams = mux.Vars(r)\n\trequest.QueryParams = r.URL.Query()\n\treturn &request\n}", "func (c *OutputService7ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func (c *OutputService7ProtocolTest) newRequest(op *aws.Operation, params, data interface{}) *aws.Request {\n\treq := c.NewRequest(op, params, data)\n\n\treturn req\n}", "func newfileUploadRequest(uri string, resource string, params map[string]string, path string) (*http.Request, error) {\n\tfile, err := os.Open(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfileContents, err := ioutil.ReadAll(file)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfile.Close()\n\n\trequest, err := http.NewRequest(\"POST\", uri+resource, bytes.NewBuffer([]byte(fileContents)))\n\n\tif err != nil {\n\t\tlog.Println(\"Could not allocate new request object: \", err)\n\t\treturn nil, err\n\t}\n\n\tvalues := request.URL.Query()\n\tfor key, val := range params {\n\t\tvalues.Add(key, val)\n\t}\n\n\trequest.URL.RawQuery = values.Encode()\n\n\trequest.Header.Add(\"Accept\", \"application/json\")\n\trequest.Header.Add(\"Authorization\", authorizationKey)\n\trequest.Header.Add(\"Content-Type\", \"text/csv\")\n\n\treturn request, err\n}", "func NewRequest(method, url string) (*http.Request, error) {\n\treturn http.NewRequest(method, url, nil)\n}", "func NewCreateElasticIpRequestWithBody(server string, contentType string, body io.Reader) (*http.Request, error) {\n\tvar err error\n\n\tqueryUrl, err := url.Parse(server)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tbasePath := fmt.Sprintf(\"/elastic-ip\")\n\tif basePath[0] == '/' {\n\t\tbasePath = basePath[1:]\n\t}\n\n\tqueryUrl, err = queryUrl.Parse(basePath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(\"POST\", queryUrl.String(), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Add(\"Content-Type\", contentType)\n\treturn req, nil\n}", "func (c *IoT) newRequest(op *request.Operation, params, data interface{}) *request.Request {\n\treq := c.NewRequest(op, params, data)\n\n\t// Run custom request initialization if present\n\tif initRequest != nil {\n\t\tinitRequest(req)\n\t}\n\n\treturn req\n}", "func (c *Client) NewRequest(method, urlPath string, body interface{}) (*http.Request, error) {\n\tpath, err := url.Parse(urlPath)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.baseURL.ResolveReference(path)\n\n\tbuf := new(bytes.Buffer)\n\tif body != nil {\n\t\tif err := json.NewEncoder(buf).Encode(body); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq.Header.Set(\"Content-Type\", \"application/json\")\n\treq.Header.Set(\"User-Agent\", c.userAgent)\n\treq.Header.Set(\"X-TrackerToken\", c.token)\n\treturn req, nil\n}", "func (c *Client) NewRequest(method, urlStr string, body interface{}) (*http.Request, error) {\n\trel, err := url.Parse(urlStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tu := c.BaseURL.ResolveReference(rel)\n\n\tvar buf io.ReadWriter\n\tif body != nil {\n\t\tbuf = new(bytes.Buffer)\n\t\terr := json.NewEncoder(buf).Encode(body)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t}\n\n\treq, err := http.NewRequest(method, u.String(), buf)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif c.accessToken != \"\" {\n\t\treq.Header.Add(\"Authorization\", \"Bearer \"+c.accessToken)\n\t}\n\n\treq.Header.Add(\"Accept\", mediaTypeV1)\n\tif c.UserAgent != \"\" {\n\t\treq.Header.Add(\"User-Agent\", c.UserAgent)\n\t}\n\treturn req, nil\n}" ]
[ "0.7647956", "0.72992545", "0.64595675", "0.6244778", "0.61958534", "0.6163761", "0.61544037", "0.6129627", "0.61294055", "0.61095285", "0.6086343", "0.6038118", "0.5998417", "0.5987045", "0.59082973", "0.5906049", "0.5901249", "0.5884452", "0.5871139", "0.5857211", "0.58516026", "0.58399236", "0.58319", "0.57928747", "0.57854813", "0.5777074", "0.57723874", "0.574915", "0.574274", "0.57194805", "0.5717038", "0.5696445", "0.5688682", "0.5676367", "0.56721795", "0.5664635", "0.5643292", "0.56377184", "0.5634601", "0.5631492", "0.5629857", "0.56276566", "0.56237483", "0.56154186", "0.5603702", "0.5603313", "0.5599737", "0.5594035", "0.5580918", "0.5572309", "0.55716825", "0.5568175", "0.5566381", "0.55635226", "0.5560588", "0.5559625", "0.5544744", "0.5543157", "0.5537151", "0.55250734", "0.55122197", "0.55112785", "0.55081177", "0.5506452", "0.5499369", "0.549494", "0.5489732", "0.5485167", "0.5482804", "0.54796803", "0.5473768", "0.5473726", "0.5471476", "0.54553187", "0.5433994", "0.5433994", "0.54313546", "0.542904", "0.542511", "0.541496", "0.54126555", "0.5411389", "0.54039633", "0.53983593", "0.5396154", "0.53827685", "0.53801346", "0.5375428", "0.5365494", "0.5353819", "0.5353197", "0.53493774", "0.5344646", "0.5344646", "0.53436726", "0.5340677", "0.5340223", "0.53399706", "0.53388995", "0.5333895" ]
0.8111466
0
MarshalJSON will be called by json.Marshal it will return json representation of server's status.
func (status ServerStatus) MarshalJSON() ([]byte, error) { var ( buffer bytes.Buffer data []byte err error ) if status.Role == "master" { data, err = ffjson.Marshal(unencodable(status)) } else { status.Role = "slave" data, err = ffjson.Marshal(status.BasicServerStatus) } err = json.Indent(&buffer, data, "", ServerStatusResponseIndentation) if err != nil { logger.Errorf("can't indent json %s: %s", data, err) return data, nil } return buffer.Bytes(), nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func getStatusJSON() ([]byte, error) {\n\t// Retrieve status\n\tstatus, err := common.GetServerStatus()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Marshal into JSON from request\n\tres, err := json.Marshal(status)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Return status\n\treturn res, nil\n}", "func (s *Server) getServerStatus(c echo.Context) error {\n\n\treturn c.JSON(http.StatusOK,\n\t\tstruct {\n\t\t\tStatus string `json:\"status\"`\n\t\t\tLastUpTime time.Time `json:\"last_up_time\"`\n\t\t\tLastDownTime time.Time `json:\"last_down_time\"`\n\t\t}{\n\t\t\tStatus: s.Status.String(),\n\t\t\tLastUpTime: s.LastUpTime,\n\t\t\tLastDownTime: s.LastDownTime,\n\t\t},\n\t)\n}", "func Status(version string, platform string, goVersion string, goPlatform string) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tjs, _ := json.Marshal(models.Server{\n\t\t\tVersion: version,\n\t\t\tPlatform: platform,\n\t\t\tGoVersion: goVersion,\n\t\t\tGoPlatform: goPlatform,\n\t\t})\n\n\t\tw.Write(js)\n\t}\n}", "func (s *Server) Status(c *gin.Context) {\n\tc.JSON(200, system.Info())\n}", "func (s *HTTPStatus) Status(w http.ResponseWriter, r *http.Request) {\n\tif err := json.NewEncoder(w).Encode(s); err != nil {\n\t\tlog.Error(\"Failed to encode status\", zap.Error(err), zap.Any(\"status\", *s))\n\t}\n}", "func jsonStatusOk() ([]byte, error) {\n\treturn json.Marshal(status{Status: \"ok\"})\n}", "func APIStatus(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write([]byte(GetStatus()))\n}", "func CheckStatus(w http.ResponseWriter, r *http.Request) {\n\tjson.NewEncoder(w).Encode(Status{Message : \"Alive\"})\n}", "func StatusHandler(w http.ResponseWriter, r *http.Request) {\n\tif r.Method == http.MethodOptions {\n\t\treturn\n\t}\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\tio.WriteString(w, `{\"status\": \"Connected\"}`)\n}", "func statusHandler(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tenc := json.NewEncoder(w)\n\terr := enc.Encode(getCurrentStatus())\n\tif err != nil {\n\t\tsklog.Errorf(\"Failed to write or encode output: %s\", err)\n\t\treturn\n\t}\n}", "func (o ApplicationOutput) ServingStatus() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Application) pulumi.StringOutput { return v.ServingStatus }).(pulumi.StringOutput)\n}", "func Status(response []byte) {\n\t// unmarshal into the SnapStatus struct, to get \"Status\" and\n\t// flatten the property set of the response\n\tvar status map[string]string\n\tjson.Unmarshal(response, &status)\n\n\tutils.PrintStatus(status[\"status\"], status[\"message\"])\n}", "func Status(c *gin.Context) {\n\tresp := gin.H{}\n\tfor _, s := range models.All {\n\t\tresp[s.Name] = s.Status\n\t}\n\tc.JSON(200, resp)\n}", "func (r *MachinePoolsAddServerResponse) Status(value int) *MachinePoolsAddServerResponse {\n\tr.status = value\n\treturn r\n}", "func (r *MachinePoolsListServerResponse) Status(value int) *MachinePoolsListServerResponse {\n\tr.status = value\n\treturn r\n}", "func EndpointGETStatus(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.WriteHeader(http.StatusOK)\n\n\t// Create the actual data response of the API call\n\tdata := Status{Name: \"AKTVE API Server\", Status: \"online\", Version: gAPIVersion}\n\tdata.Update()\n\n\t// Create a success response\n\tsuccess := Success{Success: true, Error: \"\"}\n\n\t// Combine the success and data structs so that they can be returned\n\treturnJSON := struct {\n\t\tSuccess\n\t\tStatus\n\t}{success, data}\n\n\tif err := json.NewEncoder(w).Encode(returnJSON); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (ps PoolStatus) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(ps.String())\n}", "func (k *Keybase) status() status {\n\tcmdOut, err := k.Exec(\"status\", \"-j\")\n\tif err != nil {\n\t\treturn status{}\n\t}\n\n\tvar s status\n\tjson.Unmarshal(cmdOut, &s)\n\n\treturn s\n}", "func RespondWithStatus(status string, data interface{}) string {\n response := make(map[string]interface{})\n response[\"status\"] = status\n response[\"data\"] = data\n edata, _ := json.Marshal(response)\n return string(edata)\n}", "func (jj *Juju) Status(user, service string) (*simplejson.Json, error) {\n //TODO Without id, should return every services matching {user}-*, not everything\n id := jj.id(user, service)\n args := []string{\"status\", \"--format\", \"json\"}\n if service != \"\" {\n args = append(args, id)\n }\n log.Infof(\"fetch juju status (%s)\\n\", id)\n\n cmd := exec.Command(\"juju\", args...)\n output, err := cmd.CombinedOutput()\n if err != nil {\n return EmptyJSON(), err\n }\n log.Debugf(\"successful request: %v\\n\", string(output))\n\n jsonOutput, err := simplejson.NewJson(output)\n if err != nil {\n return EmptyJSON(), err\n }\n\n mapping, _ := vmSshForward(user, jj.Controller, jsonOutput)\n jsonOutput.Set(\"ssh-port\", mapping)\n\n return jsonOutput, err\n}", "func (client *ClientRPCMethods) Status(in *string, response *ServerStatus) (err error) {\n\t*response = *client.client.callback.Status()\n\treturn nil\n}", "func (h HealthController) Status(c *gin.Context) {\n\tc.JSON(200, gin.H{\n\t\t\"status\": \"ok\",\n\t})\n}", "func (s Server)MarshalJSON() ([]byte, error){\n objectMap := make(map[string]interface{})\n if(s.Identity != nil) {\n objectMap[\"identity\"] = s.Identity\n }\n if(s.Sku != nil) {\n objectMap[\"sku\"] = s.Sku\n }\n if(s.ServerProperties != nil) {\n objectMap[\"properties\"] = s.ServerProperties\n }\n if(s.Tags != nil) {\n objectMap[\"tags\"] = s.Tags\n }\n if(s.Location != nil) {\n objectMap[\"location\"] = s.Location\n }\n return json.Marshal(objectMap)\n }", "func (s *Server) handleStatus(w http.ResponseWriter, req *http.Request) {\n\tif req.Method != http.MethodGet {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\ts.mu.Lock()\n\tst := s.status\n\ts.mu.Unlock()\n\tb, err := json.MarshalIndent(st, \"\", \" \")\n\tif err != nil {\n\t\tw.WriteHeader(500)\n\t\tw.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\tw.Write(b)\n}", "func Status(res http.ResponseWriter, req *http.Request) {\n\n\ttype statusJSON struct {\n\t\tKind string\n\t\tEntry interface{}\n\t}\n\n\tvar b []statusJSON\n\n\trepo, err := git2go.OpenRepository(\"/Users/ggordan/bootstrap\")\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tvar so git2go.StatusOptions\n\n\tstatusList, _ := repo.StatusList(&so)\n\n\tentries, _ := statusList.EntryCount()\n\n\tfor i := 0; i < entries; i++ {\n\t\tvar t string\n\t\tentry, _ := statusList.ByIndex(i)\n\n\t\tswitch entry.Status {\n\t\tcase git2go.StatusIndexNew:\n\t\t\tt = \"index_new\"\n\t\tcase git2go.StatusWtModified:\n\t\t\tfallthrough\n\t\tcase git2go.StatusIndexModified:\n\t\t\tt = \"index_modified\"\n\t\tcase git2go.StatusWtDeleted:\n\t\t\tfallthrough\n\t\tcase git2go.StatusIndexDeleted:\n\t\t\tt = \"index_deleted\"\n\t\t}\n\n\t\tb = append(b, statusJSON{\n\t\t\tKind: t,\n\t\t\tEntry: entry,\n\t\t})\n\t}\n\n\tfmt.Println(\"DATA\", statusList, entries)\n\n\tbb, _ := json.Marshal(b)\n\n\tres.Write(bb)\n}", "func (h *handler) writeJSONStatus(status int, value interface{}) {\n\tif !h.requestAccepts(\"application/json\") {\n\t\tbase.Warn(\"Client won't accept JSON, only %s\", h.rq.Header.Get(\"Accept\"))\n\t\th.writeStatus(http.StatusNotAcceptable, \"only application/json available\")\n\t\treturn\n\t}\n\t \n\tjsonOut, err := json.Marshal(value)\n\t \n\tif err != nil {\n\t\tbase.Warn(\"Couldn't serialize JSON for %v : %s\", value, err)\n\t\th.writeStatus(http.StatusInternalServerError, \"JSON serialization failed\")\n\t\treturn\n\t}\n\n\tif PrettyPrint {\n\t\tvar buffer bytes.Buffer\n\t\tjson.Indent(&buffer, jsonOut, \"\", \" \")\n\t\tjsonOut = append(buffer.Bytes(), '\\n')\n\t}\n\t\t\n\th.setHeader(\"Content-Type\", \"application/json\")\n\tif h.rq.Method != \"HEAD\" {\n\t\t//if len(jsonOut) < 1000 {\n\t\t//\th.disableResponseCompression()\n\t\t//}\n\t\th.setHeader(\"Content-Length\", fmt.Sprintf(\"%d\", len(jsonOut)))\n\t\tif status > 0 {\n\t\t\th.response.WriteHeader(status)\n\t\t\th.setStatus(status, \"\")\n\t\t}\n\t\th.response.Write(jsonOut)\n\t} else if status > 0 {\n\t\th.response.WriteHeader(status)\n\t\th.setStatus(status, \"\")\n\t}\n}", "func (v Status) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonD2b7633eEncodeGithubComMailcoursesTechnoparkDbmsForumGeneratedModels5(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (wallet *Wallet) ServerStatus() (status ServerStatus, err error) {\n\treturn wallet.connector.(*WebConnector).ServerStatus()\n}", "func statusCheck(w http.ResponseWriter, r *http.Request) {\n\tLog.Printf(\"/status\")\n\tproxiStatus.IndexSize = repo.GetIndexSize()\n\tstatus,_ := json.Marshal(proxiStatus)\n\tio.WriteString(w, string(status))\n}", "func (config *Configuration) ServerStatusString() string {\n pinfo, _ := Util.GetProcessInfo(os.Getpid())\n result := fmt.Sprintf(\"%16s:%d PID %d Elapsed %s CPU %1.1f%% Thr %d/%d Mem %s Messages: %s\",\n config.ServiceName,\n config.ServicePort,\n pinfo.PID,\n Util.HumanDuration(time.Since(pinfo.StartTime)),\n pinfo.CPUPercent,\n runtime.NumGoroutine(),\n runtime.NumCPU(),\n Util.HumanBytes(int64(pinfo.VMemory)),\n Util.HumanInt(int64(config.MessageCount)),\n )\n return result\n}", "func handleStatusRequest(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tstatusJSON, _ := json.Marshal(&globalStatus)\n\tfmt.Fprintf(w, \"%s\\n\", statusJSON)\n}", "func (s *status) format(status string) string {\n\t// We expect all JSON encodings succeed, since the possible values come\n\t// from a very small set controlled by us. Panic otherwise.\n\ts.msg.Status = status\n\tb, err := json.Marshal(s.msg)\n\tif err != nil {\n\t\tpanic(fmt.Sprint(\"JSON encoding status failed: \", err))\n\t}\n\treturn string(b)\n}", "func (s *Server) JSON(w http.ResponseWriter, status int, data interface{}) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(status)\n\tjson.NewEncoder(w).Encode(data)\n}", "func (u configExportMessage) JSON() string {\n\tu.Status = \"success\"\n\tstatusJSONBytes, e := json.MarshalIndent(u, \"\", \" \")\n\tfatalIf(probe.NewError(e), \"Unable to marshal into JSON.\")\n\n\treturn string(statusJSONBytes)\n}", "func (as *FirelensStatus) MarshalJSON() ([]byte, error) {\n\tif as == nil {\n\t\treturn nil, errors.New(\"firelens resource status is nil\")\n\t}\n\treturn []byte(`\"` + as.String() + `\"`), nil\n}", "func (server *FlexibleServer) GetStatus() genruntime.ConvertibleStatus {\n\treturn &server.Status\n}", "func (svc Service) statusHandler(w http.ResponseWriter, r *http.Request, ps httprouter.Params) {\n\tjson := GetStatusAsJSON()\n\tfmt.Println(json)\n\tfmt.Fprintf(w, \"%s\\n\\r\", json)\n}", "func (c *Client) Status() string {\n\n\treturn c.status.String()\n}", "func (c *ConsulServiceRegistry) Status() *supervisor.Status {\n\ts := &Status{}\n\n\t_, err := c.getClient()\n\tif err != nil {\n\t\ts.Health = err.Error()\n\t} else {\n\t\ts.Health = \"ready\"\n\t}\n\n\tc.statusMutex.Lock()\n\tserversNum := c.serversNum\n\tc.statusMutex.Unlock()\n\n\ts.ServersNum = serversNum\n\n\treturn &supervisor.Status{\n\t\tObjectStatus: s,\n\t}\n}", "func (hs HostStatus) MarshalJSON() ([]byte, error) {\n\tbuffer := bytes.NewBufferString(`\"`)\n\tbuffer.WriteString(hs.String())\n\tbuffer.WriteString(`\"`)\n\treturn buffer.Bytes(), nil\n}", "func (h PingController) Status(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\n\t\t\"message\": \"pong\",\n\t})\n}", "func (sg StatusGroup) Status(w http.ResponseWriter, r *http.Request) error {\n\tctx := r.Context()\n\n\thost, err := os.Hostname()\n\tif err != nil {\n\t\thost = \"unavailable\"\n\t}\n\n\tinfo := struct {\n\t\tStatus string `json:\"status\"`\n\t\tHost string `json:\"host\"`\n\t}{\n\t\tStatus: \"up\",\n\t\tHost: host,\n\t}\n\n\treturn respond(ctx, w, info, http.StatusOK)\n}", "func (srs SecondaryReplicatorStatus) MarshalJSON() ([]byte, error) {\n\tsrs.Kind = KindSecondaryReplicatorStatus\n\tobjectMap := make(map[string]interface{})\n\tif srs.ReplicationQueueStatus != nil {\n\t\tobjectMap[\"ReplicationQueueStatus\"] = srs.ReplicationQueueStatus\n\t}\n\tif srs.LastReplicationOperationReceivedTimeUtc != nil {\n\t\tobjectMap[\"LastReplicationOperationReceivedTimeUtc\"] = srs.LastReplicationOperationReceivedTimeUtc\n\t}\n\tif srs.IsInBuild != nil {\n\t\tobjectMap[\"IsInBuild\"] = srs.IsInBuild\n\t}\n\tif srs.CopyQueueStatus != nil {\n\t\tobjectMap[\"CopyQueueStatus\"] = srs.CopyQueueStatus\n\t}\n\tif srs.LastCopyOperationReceivedTimeUtc != nil {\n\t\tobjectMap[\"LastCopyOperationReceivedTimeUtc\"] = srs.LastCopyOperationReceivedTimeUtc\n\t}\n\tif srs.LastAcknowledgementSentTimeUtc != nil {\n\t\tobjectMap[\"LastAcknowledgementSentTimeUtc\"] = srs.LastAcknowledgementSentTimeUtc\n\t}\n\tif srs.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = srs.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (manager *Manager) GetServerStatus() string {\n\tif manager.Startup != nil && manager.Startup.Error {\n\t\treturn \"startup_error\"\n\t}\n\n\tif manager.ActiveServer == nil {\n\t\treturn cloud.StatusOff\n\t}\n\n\tswitch manager.ActiveServer.Status {\n\tcase cloud.StatusActive:\n\t\treturn cloud.StatusActive\n\tcase cloud.StatusStartup:\n\t\treturn cloud.StatusStartup\n\tdefault:\n\t\treturn cloud.StatusOff\n\t}\n}", "func (c *Client) Status(ctx context.Context) error {\n\treturn c.PingContext(ctx)\n}", "func sendStatus(c *Client, level string, msg string, detail string) {\n\tlog.Debugf(\"sendStatus: %s\", msg)\n\tvar resp = &manager.PlayerStatus{\n\t\tType: \"status\",\n\t\tStatus: statusString[gp_daemon_status],\n\t\tLevel: level,\n\t\tMsg: msg,\n\t\tDetail: detail,\n\t}\n\tj, _ := json.Marshal(resp)\n\t//c.send <- j\n\tsendToManager(c, j)\n\tlog.Debug(\"exit sendStatus\")\n}", "func GetStatusAsJSON(cfg *Config) string {\n\tstatus := GetStatus(cfg)\n\tblob, _ := json.Marshal(status)\n\n\treturn string(blob)\n}", "func GetServerStatus() (ServerStatus, error) {\n\t// Get system hostname\n\tvar hostname string\n\thostname, err := os.Hostname()\n\tif err != nil {\n\t\treturn ServerStatus{}, err\n\t}\n\n\t// Get current memory profile\n\tmem := &runtime.MemStats{}\n\truntime.ReadMemStats(mem)\n\n\t// Report memory usage in MB\n\tmemMb := float64((float64(mem.Alloc) / 1000) / 1000)\n\n\t// Current uptime\n\tuptime := time.Now().Unix() - Static.StartTime\n\n\t// API status\n\tapiStatus := TimedStats{\n\t\tatomic.LoadInt64(&Static.API.Minute),\n\t\tatomic.LoadInt64(&Static.API.HalfHour),\n\t\tatomic.LoadInt64(&Static.API.Hour),\n\t\tatomic.LoadInt64(&Static.API.Total),\n\t}\n\n\t// HTTP status\n\thttpStatus := TimedStats{\n\t\tatomic.LoadInt64(&Static.HTTP.Minute),\n\t\tatomic.LoadInt64(&Static.HTTP.HalfHour),\n\t\tatomic.LoadInt64(&Static.HTTP.Hour),\n\t\tatomic.LoadInt64(&Static.HTTP.Total),\n\t}\n\n\t// UDP status\n\tudpStatus := TimedStats{\n\t\tatomic.LoadInt64(&Static.UDP.Minute),\n\t\tatomic.LoadInt64(&Static.UDP.HalfHour),\n\t\tatomic.LoadInt64(&Static.UDP.Hour),\n\t\tatomic.LoadInt64(&Static.UDP.Total),\n\t}\n\n\t// Build status struct\n\tstatus := ServerStatus{\n\t\tos.Getpid(),\n\t\thostname,\n\t\truntime.GOOS,\n\t\truntime.GOARCH,\n\t\truntime.NumCPU(),\n\t\truntime.NumGoroutine(),\n\t\tmemMb,\n\t\tStatic.Maintenance,\n\t\tStatic.StatusMessage,\n\t\tuptime,\n\t\tapiStatus,\n\t\thttpStatus,\n\t\tudpStatus,\n\t}\n\n\t// Return status struct\n\treturn status, nil\n}", "func (s *Shell) Status(_ *cli.Context) error {\n\tresp, err := s.HTTP.Get(\"/health?full=1\", nil)\n\tif err != nil {\n\t\treturn s.errorOut(err)\n\t}\n\tdefer func() {\n\t\tif cerr := resp.Body.Close(); cerr != nil {\n\t\t\terr = multierr.Append(err, cerr)\n\t\t}\n\t}()\n\n\treturn s.renderAPIResponse(resp, &HealthCheckPresenters{})\n}", "func getHealthStatus(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"status\": \"ready\"})\n}", "func (s Server) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(struct {\n\t\tDuration int64 `json:\"duration,omitempty\"` // Measures the duration of the inbound HTTP request in ms\n\t}{\n\t\ts.Duration.Milliseconds(),\n\t})\n}", "func GetStatus(c echo.Context) error {\n\tserver := c.(*Server)\n\n\tcount, err := server.Repository.Games()\n\tif err != nil {\n\t\treturn err\n\t}\n\treturn c.JSON(http.StatusOK, Status{\n\t\tName: server.Name,\n\t\tGames: count,\n\t})\n}", "func (c *Seaweed) Status() (result *SystemStatus, err error) {\n\tdata, _, err := c.client.get(encodeURI(*c.master, \"/dir/status\", nil), nil)\n\tif err == nil {\n\t\tresult = &SystemStatus{}\n\t\terr = json.Unmarshal(data, result)\n\t}\n\treturn\n}", "func (api *API) Status(request *restful.Request, response *restful.Response) {\n\tresponse.WriteHeader(http.StatusOK)\n}", "func (s *Session) getStatus() (map[string]interface{}, error) {\n\tvar dummyMap = make(map[string]interface{})\n\terr := s.checkTank()\n\tif err != nil {\n\t\treturn dummyMap, err\n\t}\n\terr = s.checkName()\n\tif err != nil {\n\t\treturn dummyMap, err\n\t}\n\n\tresp, err := netClient.Get(fmt.Sprintf(\"%v/status?session=%v\", s.Tank.Url, s.Name))\n\tif err != nil {\n\t\terr = fmt.Errorf(\"http.GET failed: %w\", err)\n\t\ts.Status = \"disconnect\"\n\t\treturn dummyMap, err\n\t}\n\tdefer resp.Body.Close()\n\n\trespBody, err := checkResponseCode(*resp)\n\tif err != nil {\n\t\ts.setFailed([]string{err.Error()})\n\t\treturn dummyMap, err\n\t}\n\n\tvar respJson map[string]interface{}\n\terr = json.Unmarshal(respBody, &respJson)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"fail to unmarshal get status response: %w\", err)\n\t\treturn dummyMap, err\n\t}\n\n\tswitch stage := respJson[\"current_stage\"].(type) {\n\tcase string:\n\t\ts.Stage = stage\n\t}\n\tswitch status := respJson[\"status\"].(type) {\n\tcase string:\n\t\ts.Status = status\n\t}\n\treturn respJson, nil\n}", "func (m *MysqlProxy) GetStatus() (map[string]interface{}, error) {\n result := map[string]interface{}{}\n\n result[\"main\"] = redis.EncodeData(m)\n tables := []string{}\n shardDB := []string{}\n\n for _, table := range m.Tables {\n tables = append(tables, redis.EncodeData(table))\n }\n\n for _, db := range m.ShardDBs {\n shardDB = append(shardDB, redis.EncodeData(db))\n }\n\n result[\"tables\"] = tables\n result[\"sharddbs\"] = shardDB\n\n return result, nil\n}", "func (v Status) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjsonD2b7633eEncodeGithubComDbProjectPkgModels6(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (e LintStatus) MarshalJSON() ([]byte, error) {\n\ts := e.String()\n\treturn json.Marshal(s)\n}", "func (j JobStatus) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(j.String())\n}", "func (sr *statusResponse) getJson(s string) []byte {\n\tsr.Status = s\n\tjsr, _ := json.Marshal(sr)\n\treturn jsr\n}", "func (rs ReplicatorStatus) MarshalJSON() ([]byte, error) {\n\trs.Kind = KindReplicatorStatus\n\tobjectMap := make(map[string]interface{})\n\tif rs.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = rs.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (sars SecondaryActiveReplicatorStatus) MarshalJSON() ([]byte, error) {\n\tsars.Kind = KindActiveSecondary\n\tobjectMap := make(map[string]interface{})\n\tif sars.ReplicationQueueStatus != nil {\n\t\tobjectMap[\"ReplicationQueueStatus\"] = sars.ReplicationQueueStatus\n\t}\n\tif sars.LastReplicationOperationReceivedTimeUtc != nil {\n\t\tobjectMap[\"LastReplicationOperationReceivedTimeUtc\"] = sars.LastReplicationOperationReceivedTimeUtc\n\t}\n\tif sars.IsInBuild != nil {\n\t\tobjectMap[\"IsInBuild\"] = sars.IsInBuild\n\t}\n\tif sars.CopyQueueStatus != nil {\n\t\tobjectMap[\"CopyQueueStatus\"] = sars.CopyQueueStatus\n\t}\n\tif sars.LastCopyOperationReceivedTimeUtc != nil {\n\t\tobjectMap[\"LastCopyOperationReceivedTimeUtc\"] = sars.LastCopyOperationReceivedTimeUtc\n\t}\n\tif sars.LastAcknowledgementSentTimeUtc != nil {\n\t\tobjectMap[\"LastAcknowledgementSentTimeUtc\"] = sars.LastAcknowledgementSentTimeUtc\n\t}\n\tif sars.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = sars.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (s Status) WriteJSON(w http.ResponseWriter) error {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(s.Code)\n\t_, err := fmt.Fprintf(w, `{\"error\":%q}`, s.String())\n\treturn err\n}", "func WriteJSONStatus(w http.ResponseWriter, statusCode int, v interface{}) error {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(statusCode)\n\treturn json.NewEncoder(w).Encode(v)\n}", "func (o ApplicationStatusSyncOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApplicationStatusSync) string { return v.Status }).(pulumi.StringOutput)\n}", "func HealthCheckStatusHandler(w http.ResponseWriter, r *http.Request) {\n\tresponse := HealthCheckStatusResponse{\"OK\"}\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\tw.WriteHeader(http.StatusOK)\n\tif err := json.NewEncoder(w).Encode(response); err != nil {\n\t\tpanic(err)\n\t}\n}", "func (self *client) GetStatus() {\n\n}", "func HandleStatus(www http.ResponseWriter, req *http.Request) {\n\tif isTLS {\n\t\twww.Header().Add(\"Strict-Transport-Security\", \"max-age=63072000; includeSubDomains\")\n\t}\n\twww.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\twww.WriteHeader(http.StatusOK) // 200\n\tfmt.Fprintf(www, `{\"status\":\"success\"}`)\n\treturn\n}", "func (t TryjobStatus) MarshalJSON() ([]byte, error) {\n\treturn []byte(\"\\\"\" + t.String() + \"\\\"\"), nil\n}", "func checkStatus() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tc.JSON(http.StatusOK, \"Server is running successfully !!!!!\")\n\t}\n}", "func (s *Server) ServerStatus(ctx context.Context, req *observerpb.ServerStatusRequest) (*observerpb.ServerStatusResponse, error) {\n\tctx, cancel := context.WithCancel(context.Background())\n\tdefer cancel()\n\tg, ctx := errgroup.WithContext(ctx)\n\tgo func() {\n\t\tselect {\n\t\tcase <-s.stop:\n\t\t\tcancel()\n\t\tcase <-ctx.Done():\n\t\t}\n\t}()\n\n\tpeers := s.peerList()\n\tstatuses := make(chan *observerpb.ServerStatusResponse, len(peers))\n\tfor _, p := range peers {\n\t\tp := p\n\t\tif p.conn == nil || p.connErr != nil {\n\t\t\ts.log.WithField(\"address\", p.Address.String()).Infof(\n\t\t\t\t\"No connection to peer %s, skipping\", p.Name,\n\t\t\t)\n\t\t\tgo s.connectPeer(p.Name, p.Address.String())\n\t\t\tcontinue\n\t\t}\n\t\tg.Go(func() error {\n\t\t\tclient := observerpb.NewObserverClient(p.conn)\n\t\t\tstatus, err := client.ServerStatus(ctx, req)\n\t\t\tif err != nil {\n\t\t\t\ts.log.WithFields(logrus.Fields{\n\t\t\t\t\t\"error\": err,\n\t\t\t\t\t\"peer\": p,\n\t\t\t\t}).Warning(\"Failed to retrieve server status\")\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tselect {\n\t\t\tcase statuses <- status:\n\t\t\tcase <-ctx.Done():\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t}\n\tgo func() {\n\t\tg.Wait()\n\t\tclose(statuses)\n\t}()\n\tresp := &observerpb.ServerStatusResponse{}\n\tfor status := range statuses {\n\t\tif status == nil {\n\t\t\tcontinue\n\t\t}\n\t\tresp.MaxFlows += status.MaxFlows\n\t\tresp.NumFlows += status.NumFlows\n\t\tresp.SeenFlows += status.SeenFlows\n\t\t// use the oldest uptime as a reference for the uptime as cumulating\n\t\t// values would make little sense\n\t\tif resp.UptimeNs == 0 || resp.UptimeNs > status.UptimeNs {\n\t\t\tresp.UptimeNs = status.UptimeNs\n\t\t}\n\t}\n\treturn resp, g.Wait()\n}", "func (w *MonResponseWriter) Status() int {\n\treturn w.status\n}", "func (c *Client) Status() (*Status, error) {\n\tstatus := &Status{}\n\terr := c.Get().UsePath(\"/status\").Do().Unmarshal(status)\n\treturn status, err\n}", "func (as *AppStatusHandler) JSONEncoded() ([]byte, error) {\n\tas.RLock()\n\tdefer as.RUnlock()\n\treturn json.MarshalIndent(as.state, \"\", \" \")\n}", "func (j JobStatus) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"executionCount\", j.ExecutionCount)\n\tpopulate(objectMap, \"failureCount\", j.FailureCount)\n\tpopulate(objectMap, \"faultedCount\", j.FaultedCount)\n\tpopulateTimeRFC3339(objectMap, \"lastExecutionTime\", j.LastExecutionTime)\n\tpopulateTimeRFC3339(objectMap, \"nextExecutionTime\", j.NextExecutionTime)\n\treturn json.Marshal(objectMap)\n}", "func (ts TransferStatus) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(ts.String())\n}", "func (se *statusError) MarshalJSON() ([]byte, error) {\n\tm := map[string]interface{}{}\n\n\tif se == nil || se.error == nil {\n\t\treturn json.Marshal(m)\n\t}\n\n\tm[\"error\"] = se.Error()\n\tm[\"status\"] = se.code\n\treturn json.Marshal(m)\n}", "func (healthStatus ContainerHealthStatus) String() string {\n\treturn healthStatus.BackendStatus()\n}", "func (v OperationStatus) EncodeJSON(b []byte) []byte {\n\tb = append(b, `{\"status\":`...)\n\tb = json.AppendString(b, v.Status)\n\tb = append(b, `,\"successful\":`...)\n\tb = json.AppendBool(b, v.Successful)\n\treturn append(b, \"}\"...)\n}", "func (v SyncStatus) EncodeJSON(b []byte) []byte {\n\tb = append(b, \"{\"...)\n\tif v.CurrentIndex.Set {\n\t\tb = append(b, `\"current_index\":`...)\n\t\tb = json.AppendInt(b, v.CurrentIndex.Value)\n\t\tb = append(b, \",\"...)\n\t}\n\tif v.Stage.Set {\n\t\tb = append(b, `\"stage\":`...)\n\t\tb = json.AppendString(b, v.Stage.Value)\n\t\tb = append(b, \",\"...)\n\t}\n\tif v.Synced.Set {\n\t\tb = append(b, `\"synced\":`...)\n\t\tb = json.AppendBool(b, v.Synced.Value)\n\t\tb = append(b, \",\"...)\n\t}\n\tif v.TargetIndex.Set {\n\t\tb = append(b, `\"target_index\":`...)\n\t\tb = json.AppendInt(b, v.TargetIndex.Value)\n\t\tb = append(b, \",\"...)\n\t}\n\tb[len(b)-1] = '}'\n\treturn b\n}", "func (sh *StatusHandler) Status(ctx context.Context, in *empty.Empty, out *proto.StatusResponse) error {\n\tout.OK = true\n\tout.Address = sh.address\n\n\treturn nil\n}", "func (m MediaServiceOperationStatus) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulateTimeRFC3339(objectMap, \"endTime\", m.EndTime)\n\tpopulate(objectMap, \"error\", m.Error)\n\tpopulate(objectMap, \"id\", m.ID)\n\tpopulate(objectMap, \"name\", m.Name)\n\tpopulateTimeRFC3339(objectMap, \"startTime\", m.StartTime)\n\tpopulate(objectMap, \"status\", m.Status)\n\treturn json.Marshal(objectMap)\n}", "func ReadyHandler(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tw.WriteHeader(http.StatusOK)\n\n\tstatus := statusResponse{\n\t\tStatus: \"Greeter service is healthy\",\n\t}\n\n\tjson.NewEncoder(w).Encode(status)\n}", "func (this *MeshStatus) MarshalJSON() ([]byte, error) {\n\tstr, err := MeshMarshaler.MarshalToString(this)\n\treturn []byte(str), err\n}", "func (m *VMStatus) Marshal() []byte {\n\twriter := jspb.NewWriter()\n\tm.MarshalToWriter(writer)\n\treturn writer.GetResult()\n}", "func (v NetworkStatusResponse) EncodeJSON(b []byte) []byte {\n\tb = append(b, '{', '\"', 'c', 'u', 'r', 'r', 'e', 'n', 't', '_', 'b', 'l', 'o', 'c', 'k', '_', 'i', 'd', 'e', 'n', 't', 'i', 'f', 'i', 'e', 'r', '\"', ':')\n\tb = v.CurrentBlockIdentifier.EncodeJSON(b)\n\tb = append(b, ',', '\"', 'c', 'u', 'r', 'r', 'e', 'n', 't', '_', 'b', 'l', 'o', 'c', 'k', '_', 't', 'i', 'm', 'e', 's', 't', 'a', 'm', 'p', '\"', ':')\n\tb = json.AppendInt(b, int64(v.CurrentBlockTimestamp))\n\tb = append(b, ',', '\"', 'g', 'e', 'n', 'e', 's', 'i', 's', '_', 'b', 'l', 'o', 'c', 'k', '_', 'i', 'd', 'e', 'n', 't', 'i', 'f', 'i', 'e', 'r', '\"', ':')\n\tb = v.GenesisBlockIdentifier.EncodeJSON(b)\n\tb = append(b, \",\"...)\n\tif v.OldestBlockIdentifier.Set {\n\t\tb = append(b, '\"', 'o', 'l', 'd', 'e', 's', 't', '_', 'b', 'l', 'o', 'c', 'k', '_', 'i', 'd', 'e', 'n', 't', 'i', 'f', 'i', 'e', 'r', '\"', ':')\n\t\tb = v.OldestBlockIdentifier.Value.EncodeJSON(b)\n\t\tb = append(b, \",\"...)\n\t}\n\tb = append(b, `\"peers\":[`...)\n\tfor i, elem := range v.Peers {\n\t\tif i != 0 {\n\t\t\tb = append(b, \",\"...)\n\t\t}\n\t\tb = elem.EncodeJSON(b)\n\t}\n\tb = append(b, \"],\"...)\n\tif v.SyncStatus.Set {\n\t\tb = append(b, `\"sync_status\":`...)\n\t\tb = v.SyncStatus.Value.EncodeJSON(b)\n\t\tb = append(b, \",\"...)\n\t}\n\tb[len(b)-1] = '}'\n\treturn b\n}", "func (healthStatus *ContainerHealthStatus) MarshalJSON() ([]byte, error) {\n\tif healthStatus == nil {\n\t\treturn nil, nil\n\t}\n\treturn []byte(`\"` + healthStatus.String() + `\"`), nil\n}", "func (m ManagerDeploymentStatus) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulateTimeRFC3339(objectMap, \"commitTime\", m.CommitTime)\n\tpopulate(objectMap, \"configurationIds\", m.ConfigurationIDs)\n\tpopulate(objectMap, \"deploymentStatus\", m.DeploymentStatus)\n\tpopulate(objectMap, \"deploymentType\", m.DeploymentType)\n\tpopulate(objectMap, \"errorMessage\", m.ErrorMessage)\n\tpopulate(objectMap, \"region\", m.Region)\n\treturn json.Marshal(objectMap)\n}", "func (kvsrs KeyValueStoreReplicaStatus) MarshalJSON() ([]byte, error) {\n\tkvsrs.Kind = KindKeyValueStore\n\tobjectMap := make(map[string]interface{})\n\tif kvsrs.DatabaseRowCountEstimate != nil {\n\t\tobjectMap[\"DatabaseRowCountEstimate\"] = kvsrs.DatabaseRowCountEstimate\n\t}\n\tif kvsrs.DatabaseLogicalSizeEstimate != nil {\n\t\tobjectMap[\"DatabaseLogicalSizeEstimate\"] = kvsrs.DatabaseLogicalSizeEstimate\n\t}\n\tif kvsrs.CopyNotificationCurrentKeyFilter != nil {\n\t\tobjectMap[\"CopyNotificationCurrentKeyFilter\"] = kvsrs.CopyNotificationCurrentKeyFilter\n\t}\n\tif kvsrs.CopyNotificationCurrentProgress != nil {\n\t\tobjectMap[\"CopyNotificationCurrentProgress\"] = kvsrs.CopyNotificationCurrentProgress\n\t}\n\tif kvsrs.StatusDetails != nil {\n\t\tobjectMap[\"StatusDetails\"] = kvsrs.StatusDetails\n\t}\n\tif kvsrs.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = kvsrs.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func StatusHandler(cfg *config.Config, nomadClient *nomad.Client, consulClient *consul.Client) func(w http.ResponseWriter, r *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tvar healthy *bool\n\n\t\tstatus := struct {\n\t\t\tHealty *bool\n\t\t\tConsul map[string]interface{}\n\t\t\tNomad map[string]interface{}\n\t\t}{\n\t\t\tConsul: make(map[string]interface{}),\n\t\t\tNomad: make(map[string]interface{}),\n\t\t}\n\n\t\thealthy = nil\n\t\tif nomadClient != nil {\n\t\t\tleader, err := nomadClient.Status().Leader()\n\t\t\tstatus.Nomad[\"enabled\"] = true\n\t\t\tstatus.Nomad[\"status\"] = struct {\n\t\t\t\tLeader string\n\t\t\t\tError error\n\t\t\t}{leader, err}\n\n\t\t\tif err != nil {\n\t\t\t\thealthy = newBool(false)\n\t\t\t} else {\n\t\t\t\thealthy = newBool(true)\n\t\t\t}\n\t\t} else {\n\t\t\tstatus.Nomad[\"enabled\"] = false\n\t\t}\n\n\t\tif consulClient != nil {\n\t\t\tleader, err := consulClient.Status().Leader()\n\t\t\tstatus.Consul[\"enabled\"] = true\n\t\t\tstatus.Consul[\"status\"] = struct {\n\t\t\t\tLeader string\n\t\t\t\tError error\n\t\t\t}{leader, err}\n\n\t\t\tif err != nil {\n\t\t\t\thealthy = newBool(false)\n\t\t\t} else {\n\t\t\t\tif healthy != nil {\n\t\t\t\t\thealthy = newBool(*healthy && true)\n\t\t\t\t} else {\n\t\t\t\t\thealthy = newBool(true)\n\t\t\t\t}\n\t\t\t}\n\t\t} else {\n\t\t\tstatus.Consul[\"enabled\"] = false\n\t\t}\n\t\tif healthy != nil {\n\t\t\tstatus.Healty = healthy\n\t\t} else {\n\t\t\tstatus.Healty = newBool(false)\n\t\t}\n\t\tw.Header().Set(\"Content-Type\", \"application/json; charset=UTF-8\")\n\n\t\tif *healthy {\n\t\t\tw.WriteHeader(http.StatusOK)\n\t\t} else {\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\t}\n\n\t\tjson.NewEncoder(w).Encode(status)\n\t}\n}", "func (wAPI WalletAPI) Status() (*Status, error) {\n\tvar stat Status\n\t_, raw, err := wAPI.sendRequest(\n\t\t\"GET\",\n\t\twAPI.Host+\":\"+wAPI.Port+\"/status\",\n\t\t\"\",\n\t)\n\n\tif err == nil {\n\t\terr = json.Unmarshal(*raw, &stat)\n\t}\n\n\treturn &stat, err\n}", "func (fkw *FakeClientWrapper) Status() k8sCl.StatusWriter {\n\treturn fkw.client.Status()\n}", "func (s *Status) writeStatus() error {\n\tb, err := json.Marshal(s)\n\tif err != nil {\n\t\tlogrus.Errorf(\"Failed to marshal readiness: %s\", err)\n\t\treturn err\n\t}\n\n\t// Make sure the directory exists.\n\tif err := os.MkdirAll(filepath.Dir(s.statusFile), os.ModePerm); err != nil {\n\t\tlogrus.Errorf(\"Failed to prepare directory: %s\", err)\n\t\treturn err\n\t}\n\n\t// Write the file.\n\terr = os.WriteFile(s.statusFile, b, 0644)\n\tif err != nil {\n\t\tlogrus.Errorf(\"Failed to write readiness file: %s\", err)\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (client *Client) Status(handle string) (status *Status, err error) {\n\tif client.conn == nil {\n\t\treturn nil, ErrLostConn\n\t}\n\tvar mutex sync.Mutex\n\tmutex.Lock()\n\tclient.innerHandler.put(\"s\"+handle, func(resp *Response) {\n\t\tdefer mutex.Unlock()\n\t\tvar err error\n\t\tstatus, err = resp._status()\n\t\tif err != nil {\n\t\t\tclient.err(err)\n\t\t}\n\t})\n\treq := getRequest()\n\treq.DataType = dtGetStatus\n\treq.Data = []byte(handle)\n\tclient.write(req)\n\tmutex.Lock()\n\treturn\n}", "func (client *Client) Status(handle string) {\n job := newJob(common.REQ, common.GET_STATUS, []byte(handle))\n client.writeJob(job)\n}", "func (srv *Server) minerStatusHandler(w http.ResponseWriter, req *http.Request) {\n\twriteJSON(w, srv.miner.MinerInfo())\n}", "func (o SslCertificateManagedSslCertificateResponseOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v SslCertificateManagedSslCertificateResponse) string { return v.Status }).(pulumi.StringOutput)\n}", "func (abcr AppendBlobsCreateResponse) Status() string {\n\treturn abcr.rawResponse.Status\n}", "func (ccr ContainersCreateResponse) Status() string {\n\treturn ccr.rawResponse.Status\n}", "func (rsb ReplicaStatusBase) MarshalJSON() ([]byte, error) {\n\trsb.Kind = KindReplicaStatusBase\n\tobjectMap := make(map[string]interface{})\n\tif rsb.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = rsb.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}" ]
[ "0.75226384", "0.7062121", "0.6975611", "0.69731665", "0.69530994", "0.6858634", "0.6767639", "0.6518932", "0.6513278", "0.6463999", "0.6414882", "0.6347935", "0.633668", "0.6336468", "0.632129", "0.6243403", "0.62287056", "0.62213385", "0.621537", "0.62147", "0.6196064", "0.61926216", "0.61876744", "0.61761487", "0.6168704", "0.6163883", "0.6140225", "0.61363363", "0.61190253", "0.60789376", "0.607567", "0.6073026", "0.6061407", "0.6059738", "0.6053687", "0.6051751", "0.6047875", "0.6041023", "0.6038503", "0.6033932", "0.6024955", "0.60157293", "0.60032207", "0.5994333", "0.5994209", "0.5990358", "0.59875315", "0.59865105", "0.5983496", "0.5978398", "0.59747225", "0.5970503", "0.59686697", "0.5965932", "0.5955608", "0.5952774", "0.5941988", "0.5941177", "0.59180135", "0.59128666", "0.59034973", "0.5900752", "0.5892207", "0.5877866", "0.5869406", "0.5862822", "0.58605266", "0.58559215", "0.5855549", "0.5841772", "0.5841672", "0.5839518", "0.58392894", "0.5835521", "0.58309805", "0.5829035", "0.5822671", "0.58191943", "0.58160025", "0.58141255", "0.5811455", "0.58078104", "0.5806515", "0.5797301", "0.57830215", "0.5780701", "0.57793635", "0.5775182", "0.57728076", "0.57696575", "0.5766997", "0.5754964", "0.5753537", "0.57526636", "0.5752042", "0.5750393", "0.5743186", "0.5739679", "0.57361907", "0.57281846" ]
0.71358997
1
MarshalTOML will be called by toml.Marshal it will return toml representation of server's status.
func (status ServerStatus) MarshalTOML() ([]byte, error) { var buffer bytes.Buffer var err error encoder := toml.NewEncoder(&buffer) encoder.Indent = ServerStatusResponseIndentation if status.Role == "master" { err = encoder.Encode(unencodable(status)) } else { status.Role = "slave" err = encoder.Encode(status.BasicServerStatus) } return buffer.Bytes(), err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (o TransitRouterMulticastDomainOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *TransitRouterMulticastDomain) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o ApplicationOutput) ServingStatus() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Application) pulumi.StringOutput { return v.ServingStatus }).(pulumi.StringOutput)\n}", "func (o ServerPolicyOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ServerPolicy) pulumi.StringPtrOutput { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (o ApplicationStatusWorkflowOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusWorkflow) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (o ApplicationStatusSyncOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ApplicationStatusSync) string { return v.Status }).(pulumi.StringOutput)\n}", "func (config *Configuration) ServerStatusString() string {\n pinfo, _ := Util.GetProcessInfo(os.Getpid())\n result := fmt.Sprintf(\"%16s:%d PID %d Elapsed %s CPU %1.1f%% Thr %d/%d Mem %s Messages: %s\",\n config.ServiceName,\n config.ServicePort,\n pinfo.PID,\n Util.HumanDuration(time.Since(pinfo.StartTime)),\n pinfo.CPUPercent,\n runtime.NumGoroutine(),\n runtime.NumCPU(),\n Util.HumanBytes(int64(pinfo.VMemory)),\n Util.HumanInt(int64(config.MessageCount)),\n )\n return result\n}", "func (o ListenerOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Listener) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o WorkloadStatusConfigStaticOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v WorkloadStatusConfigStatic) string { return v.Status }).(pulumi.StringOutput)\n}", "func (s *Server) getServerStatus(c echo.Context) error {\n\n\treturn c.JSON(http.StatusOK,\n\t\tstruct {\n\t\t\tStatus string `json:\"status\"`\n\t\t\tLastUpTime time.Time `json:\"last_up_time\"`\n\t\t\tLastDownTime time.Time `json:\"last_down_time\"`\n\t\t}{\n\t\t\tStatus: s.Status.String(),\n\t\t\tLastUpTime: s.LastUpTime,\n\t\t\tLastDownTime: s.LastDownTime,\n\t\t},\n\t)\n}", "func (o BucketReplicationConfigRuleOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRule) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o NetworkInsightsAnalysisOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *NetworkInsightsAnalysis) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o ReplicationSetOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ReplicationSet) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (me THITStatus) String() string { return xsdt.String(me).String() }", "func (s *TiFlashSpec) Status(ctx context.Context, timeout time.Duration, tlsCfg *tls.Config, pdList ...string) string {\n\tstoreAddr := utils.JoinHostPort(s.Host, s.FlashServicePort)\n\tstate := checkStoreStatus(ctx, storeAddr, tlsCfg, pdList...)\n\tif s.Offline && strings.ToLower(state) == \"offline\" {\n\t\tstate = \"Pending Offline\" // avoid misleading\n\t}\n\treturn state\n}", "func (status ServerStatus) MarshalJSON() ([]byte, error) {\n\tvar (\n\t\tbuffer bytes.Buffer\n\t\tdata []byte\n\t\terr error\n\t)\n\n\tif status.Role == \"master\" {\n\t\tdata, err = ffjson.Marshal(unencodable(status))\n\t} else {\n\t\tstatus.Role = \"slave\"\n\t\tdata, err = ffjson.Marshal(status.BasicServerStatus)\n\t}\n\n\terr = json.Indent(&buffer, data, \"\", ServerStatusResponseIndentation)\n\tif err != nil {\n\t\tlogger.Errorf(\"can't indent json %s: %s\", data, err)\n\t\treturn data, nil\n\t}\n\n\treturn buffer.Bytes(), nil\n}", "func (me TpubStatus) String() string { return xsdt.String(me).String() }", "func (o BucketV2ReplicationConfigurationRuleOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketV2ReplicationConfigurationRule) string { return v.Status }).(pulumi.StringOutput)\n}", "func (s TiFlashSpec) Status(tlsCfg *tls.Config, pdList ...string) string {\n\tstoreAddr := fmt.Sprintf(\"%s:%d\", s.Host, s.FlashServicePort)\n\tstate := checkStoreStatus(storeAddr, tlsCfg, pdList...)\n\tif s.Offline && strings.ToLower(state) == \"offline\" {\n\t\tstate = \"Pending Offline\" // avoid misleading\n\t}\n\treturn state\n}", "func (o BucketIntelligentTieringConfigurationOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketIntelligentTieringConfiguration) pulumi.StringPtrOutput { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (o PartnerOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Partner) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (vscw *VerticalSplitCloneWorker) StatusAsHTML() template.HTML {\n\tvscw.mu.Lock()\n\tdefer vscw.mu.Unlock()\n\tresult := \"<b>Working on:</b> \" + vscw.destinationKeyspace + \"/\" + vscw.destinationShard + \"</br>\\n\"\n\tresult += \"<b>State:</b> \" + vscw.state + \"</br>\\n\"\n\tswitch vscw.state {\n\tcase stateVSCError:\n\t\tresult += \"<b>Error</b>: \" + vscw.err.Error() + \"</br>\\n\"\n\tcase stateVSCCopy:\n\t\tresult += \"<b>Running</b>:</br>\\n\"\n\t\tresult += strings.Join(vscw.copyLogs, \"</br>\\n\")\n\tcase stateVSCDone:\n\t\tresult += \"<b>Success</b>:</br>\\n\"\n\t\tresult += strings.Join(vscw.copyLogs, \"</br>\\n\")\n\t}\n\n\treturn template.HTML(result)\n}", "func (o EmailSenderOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *EmailSender) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o GetServerGroupsGroupOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetServerGroupsGroup) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o ApplicationStatusOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatus) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (status ServerStatus) MarshalHierarchical() []byte {\n\tvar hierarchy error\n\tif status.Address != \"\" {\n\t\thierarchy = hierr.Push(status.Address)\n\t} else {\n\t\thierarchy = hierr.Push(\"status\")\n\t}\n\n\tif status.Role != \"master\" {\n\t\tstatus.Role = \"slave\"\n\t}\n\n\thierarchy = hierr.Push(\n\t\thierarchy,\n\t\tfmt.Sprintf(\"role: %s\", status.Role),\n\t)\n\n\thierarchy = hierr.Push(\n\t\thierarchy,\n\t\tfmt.Sprintf(\"total: %d\", len(status.Mirrors)),\n\t)\n\n\tif status.HierarchicalError != \"\" {\n\t\thierarchy = hierr.Push(\n\t\t\thierarchy,\n\t\t\thierr.Push(\"error\", status.HierarchicalError),\n\t\t)\n\t}\n\n\tif len(status.Mirrors) > 0 {\n\t\tmirrors := errors.New(\"mirrors\")\n\t\tfor _, mirror := range status.Mirrors {\n\t\t\tmirrors = hierr.Push(mirrors, mirror.MarshalHierarchical())\n\t\t}\n\t\thierarchy = hierr.Push(hierarchy, mirrors)\n\t}\n\n\tif status.Role == \"master\" {\n\t\thierarchy = hierr.Push(hierarchy, status.Upstream.MarshalHierarchical())\n\t}\n\n\treturn []byte(hierr.String(hierarchy))\n}", "func (w *StatusWorker) StatusAsHTML() template.HTML {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\treturn template.HTML(\"<b>State:</b> \" + w.state.String() + \"</br>\\n\")\n}", "func (o StreamProcessorOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *StreamProcessor) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (m *VMStatus) MarshalToWriter(writer jspb.Writer) {\n\tif m == nil {\n\t\treturn\n\t}\n\n\tif m.MajorStatus != 0 {\n\t\twriter.WriteUint64(1, m.MajorStatus)\n\t}\n\n\tif m.HasSubStatus {\n\t\twriter.WriteBool(2, m.HasSubStatus)\n\t}\n\n\tif m.SubStatus != 0 {\n\t\twriter.WriteUint64(3, m.SubStatus)\n\t}\n\n\tif m.HasMessage {\n\t\twriter.WriteBool(4, m.HasMessage)\n\t}\n\n\tif len(m.Message) > 0 {\n\t\twriter.WriteString(5, m.Message)\n\t}\n\n\treturn\n}", "func (ls ListSchema) Status() string {\n\treturn ls.rawResponse.Status\n}", "func (o ApplicationOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Application) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o GetRulesRuleOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetRulesRule) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o GetRulesRuleOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetRulesRule) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o ApplicationStatusOperationStateSyncResultResourcesOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultResources) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (o AnycastEipAddressOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *AnycastEipAddress) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o ChangeSetOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ChangeSet) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (scw *SplitCloneWorker) StatusAsHTML() template.HTML {\n\tstate := scw.State()\n\n\tresult := \"<b>Working on:</b> \" + scw.keyspace + \"/\" + scw.shard + \"</br>\\n\"\n\tresult += \"<b>State:</b> \" + state.String() + \"</br>\\n\"\n\tswitch state {\n\tcase WorkerStateCloneOnline:\n\t\tresult += \"<b>Running:</b></br>\\n\"\n\t\tresult += \"<b>Copying from:</b> \" + scw.formatSources() + \"</br>\\n\"\n\t\tstatuses, eta := scw.tableStatusListOnline.format()\n\t\tresult += \"<b>ETA:</b> \" + eta.String() + \"</br>\\n\"\n\t\tresult += strings.Join(statuses, \"</br>\\n\")\n\tcase WorkerStateCloneOffline:\n\t\tresult += \"<b>Running:</b></br>\\n\"\n\t\tresult += \"<b>Copying from:</b> \" + scw.formatSources() + \"</br>\\n\"\n\t\tstatuses, eta := scw.tableStatusListOffline.format()\n\t\tresult += \"<b>ETA</b>: \" + eta.String() + \"</br>\\n\"\n\t\tresult += strings.Join(statuses, \"</br>\\n\")\n\t\tif scw.online {\n\t\t\tresult += \"</br>\\n\"\n\t\t\tresult += \"<b>Result from preceding Online Clone:</b></br>\\n\"\n\t\t\tstatuses, _ := scw.tableStatusListOnline.format()\n\t\t\tresult += strings.Join(statuses, \"</br>\\n\")\n\t\t}\n\tcase WorkerStateDone:\n\t\tresult += \"<b>Success</b>:</br>\\n\"\n\t\tif scw.online {\n\t\t\tresult += \"</br>\\n\"\n\t\t\tresult += \"<b>Online Clone Result:</b></br>\\n\"\n\t\t\tstatuses, _ := scw.tableStatusListOnline.format()\n\t\t\tresult += strings.Join(statuses, \"</br>\\n\")\n\t\t}\n\t\tif scw.offline {\n\t\t\tresult += \"</br>\\n\"\n\t\t\tresult += \"<b>Offline Clone Result:</b></br>\\n\"\n\t\t\tstatuses, _ := scw.tableStatusListOffline.format()\n\t\t\tresult += strings.Join(statuses, \"</br>\\n\")\n\t\t}\n\t}\n\n\treturn template.HTML(result)\n}", "func (o GetConfigurationRecordersRecorderOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetConfigurationRecordersRecorder) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o SslCertificateManagedSslCertificateResponseOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v SslCertificateManagedSslCertificateResponse) string { return v.Status }).(pulumi.StringOutput)\n}", "func (s *Server) Status(c *gin.Context) {\n\tc.JSON(200, system.Info())\n}", "func (o DomainOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Domain) pulumi.StringPtrOutput { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (o BgpIpOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *BgpIp) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (s VirtualNodeStatus) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o TlsCipherPolicyOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *TlsCipherPolicy) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o LaunchOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Launch) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (me TxsdWorkStatus) String() string { return xsdt.String(me).String() }", "func Status(version string, platform string, goVersion string, goPlatform string) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tjs, _ := json.Marshal(models.Server{\n\t\t\tVersion: version,\n\t\t\tPlatform: platform,\n\t\t\tGoVersion: goVersion,\n\t\t\tGoPlatform: goPlatform,\n\t\t})\n\n\t\tw.Write(js)\n\t}\n}", "func GetStatus(cfg *config.Config, c client.Client) string {\n\tvar status string\n\n\tvalStatus := GetValStatusFromDB(cfg, c)\n\tif valStatus == \"1\" {\n\t\tvalStatus = \"voting\"\n\t} else {\n\t\tvalStatus = \"jailed\"\n\t}\n\tstatus = fmt.Sprintf(\"Heimdall Node Status:\\n- Your validator is currently %s \\n\", valStatus)\n\n\tvalHeight := GetValidatorBlock(cfg, c) // get heimdall validator block height\n\tstatus = status + fmt.Sprintf(\"- Validator current block height %s \\n\", valHeight)\n\n\tnetworkHeight := GetNetworkBlock(cfg, c) // get heimdall network block height\n\tstatus = status + fmt.Sprintf(\"- Network current block height %s \\n\", networkHeight)\n\n\tvotingPower := GetVotingPowerFromDb(cfg, c) // get heimdall validator voting power\n\tstatus = status + fmt.Sprintf(\"- Voting power of your validator is %s \\n\", votingPower)\n\n\tborHeight := GetBorCurrentBlokHeight(cfg, c) // get bor validator block height\n\tstatus = status + fmt.Sprintf(\"\\nBor Node :\\n- Validator current block height %s \\n\", borHeight)\n\n\tspanID := GetBorSpanIDFromDb(cfg, c) // get bor latest span ID\n\tstatus = status + fmt.Sprintf(\"- Current span id is %s \\n\", spanID)\n\n\treturn status\n}", "func (p *LegoPort) Status() (string, error) {\n\treturn stringFrom(attributeOf(p, status))\n}", "func (o SignonOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *Signon) pulumi.StringPtrOutput { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (r *MachinePoolsListServerResponse) Status(value int) *MachinePoolsListServerResponse {\n\tr.status = value\n\treturn r\n}", "func (o InstanceOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Instance) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o InstanceOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Instance) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (s *Server) Status(c context.Context, em *empty.Empty) (t *proto.Top, err error) {\n\tt = s.Top()\n\treturn t, nil\n}", "func (o GetServerGroupServerAttachmentsAttachmentOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetServerGroupServerAttachmentsAttachment) string { return v.Status }).(pulumi.StringOutput)\n}", "func ServerStatusList(list []*data.Deployment_ServerStatus, opts Options) string {\n\treturn formatList(opts, list, func(i int) []kv {\n\t\tx := list[i]\n\t\td := []kv{\n\t\t\t{\"id\", x.GetId()},\n\t\t\t{\"description\", x.GetDescription()},\n\t\t\t{\"version\", x.GetVersion()},\n\t\t\t{\"type\", x.GetType()},\n\t\t\t{\"created-at\", formatTime(opts, x.GetCreatedAt())},\n\t\t\t{\"last-started-at\", formatTime(opts, x.GetLastStartedAt())},\n\t\t\t{\"creating\", formatBool(opts, x.GetCreating())},\n\t\t\t{\"ready\", formatBool(opts, x.GetReady())},\n\t\t\t{\"failed\", formatBool(opts, x.GetFailed())},\n\t\t\t{\"upgrading\", formatBool(opts, x.GetUpgrading())},\n\t\t\t{\"ok\", formatBool(opts, x.GetOk())},\n\t\t\t{\"member-of-cluster\", formatBool(opts, x.GetMemberOfCluster())},\n\t\t}\n\t\treturn d\n\t}, false)\n}", "func (o BucketReplicationConfigRuleDestinationReplicationTimeOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketReplicationConfigRuleDestinationReplicationTime) string { return v.Status }).(pulumi.StringOutput)\n}", "func (me THITReviewStatus) String() string { return xsdt.String(me).String() }", "func (w *StatusWorker) StatusAsText() string {\n\tw.mu.Lock()\n\tdefer w.mu.Unlock()\n\n\treturn \"State: \" + w.state.String() + \"\\n\"\n}", "func (o ApplicationStatusResourcesOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusResources) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func Status(response []byte) {\n\t// unmarshal into the SnapStatus struct, to get \"Status\" and\n\t// flatten the property set of the response\n\tvar status map[string]string\n\tjson.Unmarshal(response, &status)\n\n\tutils.PrintStatus(status[\"status\"], status[\"message\"])\n}", "func (o ApplicationStatusHealthOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusHealth) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (wallet *Wallet) ServerStatus() (status ServerStatus, err error) {\n\treturn wallet.connector.(*WebConnector).ServerStatus()\n}", "func (o GetAggregateConfigRulesRuleOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetAggregateConfigRulesRule) string { return v.Status }).(pulumi.StringOutput)\n}", "func (vscw *VerticalSplitCloneWorker) StatusAsText() string {\n\tvscw.mu.Lock()\n\tdefer vscw.mu.Unlock()\n\tresult := \"Working on: \" + vscw.destinationKeyspace + \"/\" + vscw.destinationShard + \"\\n\"\n\tresult += \"State: \" + vscw.state + \"\\n\"\n\tswitch vscw.state {\n\tcase stateVSCError:\n\t\tresult += \"Error: \" + vscw.err.Error() + \"\\n\"\n\tcase stateVSCCopy:\n\t\tresult += \"Running:\\n\"\n\t\tresult += strings.Join(vscw.copyLogs, \"\\n\")\n\tcase stateVSCDone:\n\t\tresult += \"Success:\\n\"\n\t\tresult += strings.Join(vscw.copyLogs, \"\\n\")\n\t}\n\treturn result\n}", "func (o BucketLifecycleConfigurationV2RuleOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketLifecycleConfigurationV2Rule) string { return v.Status }).(pulumi.StringOutput)\n}", "func GetServerStatus() (ServerStatus, error) {\n\t// Get system hostname\n\tvar hostname string\n\thostname, err := os.Hostname()\n\tif err != nil {\n\t\treturn ServerStatus{}, err\n\t}\n\n\t// Get current memory profile\n\tmem := &runtime.MemStats{}\n\truntime.ReadMemStats(mem)\n\n\t// Report memory usage in MB\n\tmemMb := float64((float64(mem.Alloc) / 1000) / 1000)\n\n\t// Current uptime\n\tuptime := time.Now().Unix() - Static.StartTime\n\n\t// API status\n\tapiStatus := TimedStats{\n\t\tatomic.LoadInt64(&Static.API.Minute),\n\t\tatomic.LoadInt64(&Static.API.HalfHour),\n\t\tatomic.LoadInt64(&Static.API.Hour),\n\t\tatomic.LoadInt64(&Static.API.Total),\n\t}\n\n\t// HTTP status\n\thttpStatus := TimedStats{\n\t\tatomic.LoadInt64(&Static.HTTP.Minute),\n\t\tatomic.LoadInt64(&Static.HTTP.HalfHour),\n\t\tatomic.LoadInt64(&Static.HTTP.Hour),\n\t\tatomic.LoadInt64(&Static.HTTP.Total),\n\t}\n\n\t// UDP status\n\tudpStatus := TimedStats{\n\t\tatomic.LoadInt64(&Static.UDP.Minute),\n\t\tatomic.LoadInt64(&Static.UDP.HalfHour),\n\t\tatomic.LoadInt64(&Static.UDP.Hour),\n\t\tatomic.LoadInt64(&Static.UDP.Total),\n\t}\n\n\t// Build status struct\n\tstatus := ServerStatus{\n\t\tos.Getpid(),\n\t\thostname,\n\t\truntime.GOOS,\n\t\truntime.GOARCH,\n\t\truntime.NumCPU(),\n\t\truntime.NumGoroutine(),\n\t\tmemMb,\n\t\tStatic.Maintenance,\n\t\tStatic.StatusMessage,\n\t\tuptime,\n\t\tapiStatus,\n\t\thttpStatus,\n\t\tudpStatus,\n\t}\n\n\t// Return status struct\n\treturn status, nil\n}", "func (b *ServiceClusterBuilder) Status(value string) *ServiceClusterBuilder {\n\tb.status = value\n\tb.bitmap_ |= 256\n\treturn b\n}", "func (s VirtualRouterStatus) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o ApplicationStatusWorkflowPtrOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusWorkflow) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Status\n\t}).(pulumi.StringPtrOutput)\n}", "func (o BareMetalAdminClusterOutput) Status() ResourceStatusResponseOutput {\n\treturn o.ApplyT(func(v *BareMetalAdminCluster) ResourceStatusResponseOutput { return v.Status }).(ResourceStatusResponseOutput)\n}", "func (r *MachinePoolsAddServerResponse) Status(value int) *MachinePoolsAddServerResponse {\n\tr.status = value\n\treturn r\n}", "func (o ServiceOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *Service) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o MachineInstanceStatusConditionsOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v MachineInstanceStatusConditions) string { return v.Status }).(pulumi.StringOutput)\n}", "func (lfs ListFilesystemSchema) Status() string {\n\treturn lfs.rawResponse.Status\n}", "func (o PublicAdvertisedPrefixPublicDelegatedPrefixResponseOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v PublicAdvertisedPrefixPublicDelegatedPrefixResponse) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o GeoReplicationStatsResponseOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GeoReplicationStatsResponse) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o ReservedInstanceOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ReservedInstance) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o ServiceAttachmentConnectedEndpointResponseOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v ServiceAttachmentConnectedEndpointResponse) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o DataSourceOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *DataSource) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (server *FlexibleServer) GetStatus() genruntime.ConvertibleStatus {\n\treturn &server.Status\n}", "func (o GetSmartagFlowLogsLogOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetSmartagFlowLogsLog) string { return v.Status }).(pulumi.StringOutput)\n}", "func (me TReviewableHITStatus) String() string { return xsdt.String(me).String() }", "func (s MeshStatus) String() string {\n\treturn awsutil.Prettify(s)\n}", "func Status(ops types.HTTPOptions, cfg *config.Config, c client.Client) {\n\tbp, err := db.CreateBatchPoints(cfg.InfluxDB.Database)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tstatus, err := scraper.GetStatus(ops)\n\tif err != nil {\n\t\t_ = db.WriteToInfluxDb(c, bp, \"heimdall_node_status\", map[string]string{}, map[string]interface{}{\"status\": 0})\n\n\t\tlog.Printf(\"Validator Error: %v\", err)\n\t\treturn\n\t}\n\n\tif &status.Result == nil {\n\t\tif strings.ToUpper(cfg.AlerterPreferences.NodeStatusAlert) == \"YES\" {\n\t\t\t_ = alerter.SendTelegramAlert(\"⚠️ Your heimdall validator instance is not running\", cfg)\n\t\t\t_ = alerter.SendEmailAlert(\"⚠️ Your heimdall validator instance is not running\", cfg)\n\t\t}\n\t\t_ = db.WriteToInfluxDb(c, bp, \"heimdall_node_status\", map[string]string{}, map[string]interface{}{\"status\": 0})\n\t\treturn\n\t}\n\n\terr = db.WriteToInfluxDb(c, bp, \"heimdall_node_status\", map[string]string{}, map[string]interface{}{\"status\": 1})\n\tif err != nil {\n\t\tlog.Printf(\"Error while writing node status into db : %v \", err)\n\t}\n\n\tvar bh int\n\tcurrentBlockHeight := status.Result.SyncInfo.LatestBlockHeight\n\tif currentBlockHeight != \"\" {\n\t\tbh, _ = strconv.Atoi(currentBlockHeight)\n\t\terr = db.WriteToInfluxDb(c, bp, \"heimdall_current_block_height\", map[string]string{}, map[string]interface{}{\"height\": bh})\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Error while stroing current block height : %v\", err)\n\t\t}\n\t}\n\n\t// Store validator details such as moniker, signer address and hex address\n\tmoniker := status.Result.NodeInfo.Moniker\n\thexAddress := status.Result.ValidatorInfo.Address\n\tsignerAddress := cfg.ValDetails.SignerAddress\n\t_ = db.WriteToInfluxDb(c, bp, \"heimdall_val_desc\", map[string]string{}, map[string]interface{}{\"moniker\": moniker, \"hex_address\": hexAddress, \"signer_address\": signerAddress, \"address\": signerAddress[2:]})\n\n\tlog.Printf(\"Moniker:%s \", moniker)\n}", "func (o GetListenersListenerOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetListenersListener) string { return v.Status }).(pulumi.StringOutput)\n}", "func (o ContainerV1Output) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ContainerV1) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func (o GetTransitRouterRouteTablePropagationsResultOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v GetTransitRouterRouteTablePropagationsResult) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (sh *StatusHandler) Status(ctx context.Context, in *empty.Empty, out *proto.StatusResponse) error {\n\tout.OK = true\n\tout.Address = sh.address\n\n\treturn nil\n}", "func (o RuleMfaOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *RuleMfa) pulumi.StringPtrOutput { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (o ApplicationStatusResourcesHealthOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusResourcesHealth) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (o EmailIdentityDkimSigningAttributesOutput) Status() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v EmailIdentityDkimSigningAttributes) *string { return v.Status }).(pulumi.StringPtrOutput)\n}", "func (s *Server) ServerStatus(ctx context.Context, req *observerpb.ServerStatusRequest) (*observerpb.ServerStatusResponse, error) {\n\tvar (\n\t\tcancel context.CancelFunc\n\t\tg *errgroup.Group\n\t)\n\tmd, ok := metadata.FromIncomingContext(ctx)\n\tif ok {\n\t\tctx = metadata.NewOutgoingContext(ctx, md)\n\t}\n\tctx, cancel = context.WithCancel(ctx)\n\tdefer cancel()\n\tg, ctx = errgroup.WithContext(ctx)\n\n\tpeers := s.peers.List()\n\tvar numConnectedNodes, numUnavailableNodes uint32\n\tvar unavailableNodes []string\n\tstatuses := make(chan *observerpb.ServerStatusResponse, len(peers))\n\tfor _, p := range peers {\n\t\tif !isAvailable(p.Conn) {\n\t\t\tnumUnavailableNodes++\n\t\t\ts.opts.log.WithField(\"address\", p.Address).Infof(\n\t\t\t\t\"No connection to peer %s, skipping\", p.Name,\n\t\t\t)\n\t\t\ts.peers.ReportOffline(p.Name)\n\t\t\tif len(unavailableNodes) < numUnavailableNodesReportMax {\n\t\t\t\tunavailableNodes = append(unavailableNodes, p.Name)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tnumConnectedNodes++\n\t\tp := p\n\t\tg.Go(func() error {\n\t\t\tclient := s.opts.ocb.observerClient(&p)\n\t\t\tstatus, err := client.ServerStatus(ctx, req)\n\t\t\tif err != nil {\n\t\t\t\ts.opts.log.WithFields(logrus.Fields{\n\t\t\t\t\t\"error\": err,\n\t\t\t\t\t\"peer\": p,\n\t\t\t\t}).Warning(\"Failed to retrieve server status\")\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\tselect {\n\t\t\tcase statuses <- status:\n\t\t\tcase <-ctx.Done():\n\t\t\t}\n\t\t\treturn nil\n\t\t})\n\t}\n\tgo func() {\n\t\tg.Wait()\n\t\tclose(statuses)\n\t}()\n\tresp := &observerpb.ServerStatusResponse{\n\t\tVersion: build.RelayVersion.String(),\n\t\tNumConnectedNodes: &wrapperspb.UInt32Value{\n\t\t\tValue: numConnectedNodes,\n\t\t},\n\t\tNumUnavailableNodes: &wrapperspb.UInt32Value{\n\t\t\tValue: numUnavailableNodes,\n\t\t},\n\t\tUnavailableNodes: unavailableNodes,\n\t}\n\tfor status := range statuses {\n\t\tif status == nil {\n\t\t\tcontinue\n\t\t}\n\t\tresp.MaxFlows += status.MaxFlows\n\t\tresp.NumFlows += status.NumFlows\n\t\tresp.SeenFlows += status.SeenFlows\n\t\t// use the oldest uptime as a reference for the uptime as cumulating\n\t\t// values would make little sense\n\t\tif resp.UptimeNs < status.UptimeNs {\n\t\t\tresp.UptimeNs = status.UptimeNs\n\t\t}\n\t}\n\treturn resp, g.Wait()\n}", "func (o GetLoadBalancersBalancerOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v GetLoadBalancersBalancer) string { return v.Status }).(pulumi.StringOutput)\n}", "func (s WarmPoolStatus) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (ctx *ShowStatusContext) OK(r *Status) error {\n\tctx.ResponseData.Header().Set(\"Content-Type\", \"application/vnd.status+json\")\n\treturn ctx.ResponseData.Service.Send(ctx.Context, 200, r)\n}", "func (o ResourceGroupOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *ResourceGroup) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}", "func statusCheck(w http.ResponseWriter, r *http.Request) {\n\tLog.Printf(\"/status\")\n\tproxiStatus.IndexSize = repo.GetIndexSize()\n\tstatus,_ := json.Marshal(proxiStatus)\n\tio.WriteString(w, string(status))\n}", "func (o RecommendationConfigurationPropertiesResponseOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v RecommendationConfigurationPropertiesResponse) string { return v.Status }).(pulumi.StringOutput)\n}", "func (s VirtualGatewayStatus) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (o EventDataStoreOutput) Status() pulumi.StringOutput {\n\treturn o.ApplyT(func(v *EventDataStore) pulumi.StringOutput { return v.Status }).(pulumi.StringOutput)\n}" ]
[ "0.59508634", "0.5944554", "0.5864982", "0.57805276", "0.5772215", "0.5771864", "0.57045937", "0.56820184", "0.5675158", "0.5670201", "0.5664504", "0.56315213", "0.5588605", "0.55804783", "0.5577197", "0.555757", "0.55561626", "0.55493563", "0.55350584", "0.55324584", "0.55291665", "0.55190766", "0.5514948", "0.5506187", "0.5497242", "0.54934597", "0.54882944", "0.5480481", "0.5479387", "0.54746246", "0.5474481", "0.5474481", "0.5459087", "0.5457267", "0.54542387", "0.545153", "0.5444678", "0.5429048", "0.5425568", "0.5421136", "0.5420837", "0.54205114", "0.5416356", "0.54067606", "0.5400365", "0.53917783", "0.53880304", "0.53851926", "0.5384847", "0.5376893", "0.5366723", "0.5366723", "0.5359539", "0.5358414", "0.53542465", "0.53515595", "0.534758", "0.53414196", "0.5332961", "0.533082", "0.5327495", "0.53265095", "0.5322625", "0.5321021", "0.5319704", "0.53107315", "0.5310019", "0.53027076", "0.53001076", "0.529638", "0.52837527", "0.52833873", "0.52715045", "0.526886", "0.5266672", "0.52655095", "0.5263773", "0.52628016", "0.5255066", "0.52470237", "0.5246274", "0.5242633", "0.5238793", "0.52196586", "0.5219273", "0.5217946", "0.5214749", "0.521251", "0.5209359", "0.5206369", "0.52048975", "0.5202982", "0.5200223", "0.51995486", "0.51941943", "0.5192975", "0.5192526", "0.51906323", "0.51880664", "0.51867735" ]
0.788723
0
MarshalHierarchical returns hierarchical string representation. It's very useful for humans.
func (status ServerStatus) MarshalHierarchical() []byte { var hierarchy error if status.Address != "" { hierarchy = hierr.Push(status.Address) } else { hierarchy = hierr.Push("status") } if status.Role != "master" { status.Role = "slave" } hierarchy = hierr.Push( hierarchy, fmt.Sprintf("role: %s", status.Role), ) hierarchy = hierr.Push( hierarchy, fmt.Sprintf("total: %d", len(status.Mirrors)), ) if status.HierarchicalError != "" { hierarchy = hierr.Push( hierarchy, hierr.Push("error", status.HierarchicalError), ) } if len(status.Mirrors) > 0 { mirrors := errors.New("mirrors") for _, mirror := range status.Mirrors { mirrors = hierr.Push(mirrors, mirror.MarshalHierarchical()) } hierarchy = hierr.Push(hierarchy, mirrors) } if status.Role == "master" { hierarchy = hierr.Push(hierarchy, status.Upstream.MarshalHierarchical()) } return []byte(hierr.String(hierarchy)) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (status MirrorStatus) MarshalHierarchical() string {\n\thierarchy := hierr.Push(\n\t\tstatus.Name,\n\t\tfmt.Sprintf(\"state: %s\", status.State),\n\t)\n\n\tif status.ModifyDate > 0 {\n\t\thierarchy = hierr.Push(\n\t\t\thierarchy,\n\t\t\tfmt.Sprintf(\"modify date: %v\", status.ModifyDate),\n\t\t)\n\t}\n\n\treturn hierr.String(hierarchy)\n}", "func (status UpstreamStatus) MarshalHierarchical() string {\n\thierarchy := hierr.Push(\n\t\t\"upstream\",\n\t\t\"total: \"+strconv.Itoa(status.Total),\n\t\tfmt.Sprintf(\n\t\t\t\"success: %d (%.2f%%)\",\n\t\t\tstatus.Success, status.SuccessPercent,\n\t\t),\n\t\tfmt.Sprintf(\n\t\t\t\"error: %d (%.2f%%)\",\n\t\t\tstatus.Error, status.ErrorPercent,\n\t\t),\n\t)\n\n\tif len(status.Slaves) > 0 {\n\t\tslaves := errors.New(\"slaves\")\n\t\tfor _, slave := range status.Slaves {\n\t\t\tslaves = hierr.Push(slaves, slave.MarshalHierarchical())\n\t\t}\n\n\t\thierarchy = hierr.Push(hierarchy, slaves)\n\t}\n\n\treturn hierr.String(hierarchy)\n}", "func (rht *RHT) Marshal() string {\n\tmembers := rht.Members()\n\n\tsize := len(members)\n\tkeys := make([]string, 0, size)\n\tfor k := range members {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\n\tsb := strings.Builder{}\n\tsb.WriteString(\"{\")\n\n\tidx := 0\n\tfor _, k := range keys {\n\t\tvalue := members[k]\n\t\tsb.WriteString(fmt.Sprintf(\"\\\"%s\\\":%s\", k, value.Marshal()))\n\t\tif size-1 != idx {\n\t\t\tsb.WriteString(\",\")\n\t\t}\n\t\tidx++\n\t}\n\tsb.WriteString(\"}\")\n\n\treturn sb.String()\n}", "func (s HierarchyLevel) String() string {\n\treturn awsutil.Prettify(s)\n}", "func encodeTree(hmt *Tree, finalTree *string) {\n\tif hmt == nil {\n\t\treturn\n\t}\n\t\n\tif hmt.LeftNode == nil && hmt.RightNode == nil{\n\t\t*finalTree += \"1\" + string(hmt.Char)\n\t} else {\n\t\t*finalTree += \"0\"\n\t}\n\tencodeTree(hmt.LeftNode, finalTree)\n\tencodeTree(hmt.RightNode, finalTree) \n}", "func (s HierarchyStructure) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (t *Tree) formatStr() []byte {\n\tvar s []byte\n\ts = append(s, t.format...)\n\tif t.parent != nil {\n\t\tt := t.parent.formatStr()\n\t\tt = append(t, s...)\n\t\ts = t\n\t}\n\treturn s\n}", "func stringify(n *Node, level int) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level)\n\t\tfmt.Printf(format+\"%d\\n\", n.key)\n\t\tstringify(n.right, level)\n\t}\n}", "func stringify(treeNode *TreeNode, level int) {\n\tif treeNode != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"***> \"\n\t\tlevel++\n\t\tstringify(treeNode.leftNode, level)\n\t\tfmt.Printf(format+\"%d\\n\", treeNode.key)\n\t\tstringify(treeNode.rightNode, level)\n\t}\n}", "func (w *RootWalker) String() (out string) {\n\ttabs := func(n int) {\n\t\tfor i := 0; i < n; i++ {\n\t\t\tout += \"\\t\"\n\t\t}\n\t}\n\tout += fmt.Sprint(\"Root\")\n\tsize := w.Size()\n\tif size == 0 {\n\t\treturn\n\t}\n\tout += fmt.Sprintf(\".Refs[%d] ->\\n\", w.stack[0].prevInFieldIndex)\n\tfor i, obj := range w.stack {\n\t\tschName := \"\"\n\t\ts, _ := w.r.SchemaByReference(obj.s)\n\t\tif s != nil {\n\t\t\tschName = s.Name()\n\t\t}\n\n\t\ttabs(i)\n\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\tout += fmt.Sprintf(` = \"%v\"`+\"\\n\", obj.p)\n\n\t\ttabs(i)\n\t\tif obj.next != nil {\n\t\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\t\tout += fmt.Sprintf(\".%s\", obj.next.prevFieldName)\n\t\t\tif obj.next.prevInFieldIndex != -1 {\n\t\t\t\tout += fmt.Sprintf(\"[%d]\", obj.next.prevInFieldIndex)\n\t\t\t}\n\t\t\tout += fmt.Sprint(\" ->\\n\")\n\t\t}\n\t}\n\treturn\n}", "func (w *RootWalker) String() (out string) {\n\ttabs := func(n int) {\n\t\tfor i := 0; i < n; i++ {\n\t\t\tout += \"\\t\"\n\t\t}\n\t}\n\tout += fmt.Sprint(\"Root\")\n\tsize := w.Size()\n\tif size == 0 {\n\t\treturn\n\t}\n\tout += fmt.Sprintf(\".Refs[%d] ->\\n\", w.stack[0].prevInFieldIndex)\n\tfor i, obj := range w.stack {\n\t\tschName := \"\"\n\t\ts, _ := w.r.SchemaByReference(obj.s)\n\t\tif s != nil {\n\t\t\tschName = s.Name()\n\t\t}\n\n\t\ttabs(i)\n\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\tout += fmt.Sprintf(` = \"%v\"`+\"\\n\", obj.p)\n\n\t\ttabs(i)\n\t\tif obj.next != nil {\n\t\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\t\tout += fmt.Sprintf(\".%s\", obj.next.prevFieldName)\n\t\t\tif obj.next.prevInFieldIndex != -1 {\n\t\t\t\tout += fmt.Sprintf(\"[%d]\", obj.next.prevInFieldIndex)\n\t\t\t}\n\t\t\tout += fmt.Sprint(\" ->\\n\")\n\t\t}\n\t}\n\treturn\n}", "func (rh ResourceHierarchy) String() string {\n\treturn string(rh)\n}", "func WriteTree(writer io.Writer, hierarchy *Hierarchy, includeEmpty bool) {\n\ttree := assembleTree(hierarchy)\n\tkeys := make([]string, len(tree))\n\ti := 0\n\tfor k := range tree {\n\t\tkeys[i] = k\n\t\ti++\n\t}\n\tsort.Strings(keys)\n\tfor _, key := range keys {\n\t\tb := tree[key]\n\t\twriteBranch(writer, b, \"\", hierarchy, includeEmpty)\n\t}\n}", "func hierarchy(t Tags) string {\n\t// Left pad tracknumber to two digits\n\ttrack := fmt.Sprintf(\"%02s\", t[\"tracknumber\"])\n\tartist := t[\"artist\"]\n\ts := path.Join(artist[0:1],\n\t\tartist,\n\t\tt[\"date\"] + \"--\" + t[\"album\"],\n\t\ttrack + \"-\" + t[\"title\"])\n\tlog.Printf(\"Group path for %q is %s\\n\", t, s)\n\treturn s\n}", "func ToString(ctx context.Context) (ret string) {\n\tfather, ok := ctx.Value(key).(*constNode)\n\tif !ok {\n\t\treturn \"\"\n\t}\n\tDone(ctx)\n\tconst fmtStr = \"%s%s (%dms %d%%)\\n\"\n\tvar levelPrint func(level int, node *constNode, prefix string)\n\tlevelPrint = func(level int, node *constNode, prefix string) {\n\t\tvar (\n\t\t\tlastTabs string\n\t\t\tnoLastTabs string\n\t\t)\n\t\tnoLastTabs = prefix + \"├─\"\n\t\tlastTabs = prefix + \"└─\"\n\t\tfor i, child := range node.child {\n\t\t\ttabs := noLastTabs\n\t\t\tif i == len(node.child)-1 {\n\t\t\t\ttabs = lastTabs\n\t\t\t}\n\t\t\tchildCostMs := child.cost().Milliseconds()\n\t\t\tfatherCostMs := node.cost().Milliseconds()\n\t\t\tradio := int64(0)\n\t\t\tif fatherCostMs > 0 {\n\t\t\t\tradio = childCostMs * 100 / fatherCostMs\n\t\t\t}\n\t\t\tret += fmt.Sprintf(fmtStr, tabs, child.title, childCostMs, radio)\n\t\t\tif len(child.child) > 0 {\n\t\t\t\tif i == len(node.child)-1 {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\" \")\n\t\t\t\t} else {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\"│ \")\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tret += fmt.Sprintf(fmtStr, \"\", father.title, father.cost().Milliseconds(), 100)\n\tlevelPrint(0, father, \"\")\n\treturn\n}", "func (s HierarchyGroup) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (this *Codec) serialize(root *TreeNode) string {\n var s string \n res:=helpSerialize(root,s)\n fmt.Println(res)\n return res\n}", "func (s HierarchyPath) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (p *page) marshal() string {\n\tb := &strings.Builder{}\n\tfmt.Fprintf(b, \"\\n%d 0 obj\\n<<\\n\", p.id)\n\tfmt.Fprintf(b, \" /MediaBox [ %f %f %f %f ]\\n\", p.mediaBox.llx, p.mediaBox.lly, p.mediaBox.urx, p.mediaBox.ury)\n\tfmt.Fprintf(b, \" /CropBox [ %f %f %f %f ]\\n\", p.cropBox.llx, p.cropBox.lly, p.cropBox.urx, p.cropBox.ury)\n\tfmt.Fprintf(b, \" /BleedBox [ %f %f %f %f ]\\n\", p.bleedBox.llx, p.bleedBox.lly, p.bleedBox.urx, p.bleedBox.ury)\n\tfmt.Fprintf(b, \" /TrimBox [ %f %f %f %f ]\\n\", p.trimBox.llx, p.trimBox.lly, p.trimBox.urx, p.trimBox.ury)\n\tfmt.Fprintf(b, \" /Contents [ \")\n\tfor _, cid := range p.contentIds {\n\t\tfmt.Fprintf(b, \" %d 0 R \", cid)\n\t}\n\tfmt.Fprintf(b, \" ]\\n\")\n\tfmt.Fprintf(b, \" /Parent %d 0 R\\n\", p.parentID)\n\tb.WriteString(p.raw)\n\tfmt.Fprintf(b, \"\\n>>\\nendobj\\n\")\n\treturn b.String()\n}", "func (m QueryHierarchyItem) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 1)\n\n\taO0, err := swag.WriteJSON(m.WorkItemTrackingResource)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\t// now for regular properties\n\tvar propsQueryHierarchyItem struct {\n\t\tChildren []*QueryHierarchyItem `json:\"children\"`\n\n\t\tClauses *WorkItemQueryClause `json:\"clauses,omitempty\"`\n\n\t\tColumns []*WorkItemFieldReference `json:\"columns\"`\n\n\t\tCreatedBy *IdentityReference `json:\"createdBy,omitempty\"`\n\n\t\tCreatedDate strfmt.DateTime `json:\"createdDate,omitempty\"`\n\n\t\tFilterOptions interface{} `json:\"filterOptions,omitempty\"`\n\n\t\tHasChildren bool `json:\"hasChildren,omitempty\"`\n\n\t\tID strfmt.UUID `json:\"id,omitempty\"`\n\n\t\tIsDeleted bool `json:\"isDeleted,omitempty\"`\n\n\t\tIsFolder bool `json:\"isFolder,omitempty\"`\n\n\t\tIsInvalidSyntax bool `json:\"isInvalidSyntax,omitempty\"`\n\n\t\tIsPublic bool `json:\"isPublic,omitempty\"`\n\n\t\tLastExecutedBy *IdentityReference `json:\"lastExecutedBy,omitempty\"`\n\n\t\tLastExecutedDate strfmt.DateTime `json:\"lastExecutedDate,omitempty\"`\n\n\t\tLastModifiedBy *IdentityReference `json:\"lastModifiedBy,omitempty\"`\n\n\t\tLastModifiedDate strfmt.DateTime `json:\"lastModifiedDate,omitempty\"`\n\n\t\tLinkClauses *WorkItemQueryClause `json:\"linkClauses,omitempty\"`\n\n\t\tName string `json:\"name,omitempty\"`\n\n\t\tPath string `json:\"path,omitempty\"`\n\n\t\tQueryRecursionOption interface{} `json:\"queryRecursionOption,omitempty\"`\n\n\t\tQueryType interface{} `json:\"queryType,omitempty\"`\n\n\t\tSortColumns []*WorkItemQuerySortColumn `json:\"sortColumns\"`\n\n\t\tSourceClauses *WorkItemQueryClause `json:\"sourceClauses,omitempty\"`\n\n\t\tTargetClauses *WorkItemQueryClause `json:\"targetClauses,omitempty\"`\n\n\t\tWiql string `json:\"wiql,omitempty\"`\n\t}\n\tpropsQueryHierarchyItem.Children = m.Children\n\n\tpropsQueryHierarchyItem.Clauses = m.Clauses\n\n\tpropsQueryHierarchyItem.Columns = m.Columns\n\n\tpropsQueryHierarchyItem.CreatedBy = m.CreatedBy\n\n\tpropsQueryHierarchyItem.CreatedDate = m.CreatedDate\n\n\tpropsQueryHierarchyItem.FilterOptions = m.FilterOptions\n\n\tpropsQueryHierarchyItem.HasChildren = m.HasChildren\n\n\tpropsQueryHierarchyItem.ID = m.ID\n\n\tpropsQueryHierarchyItem.IsDeleted = m.IsDeleted\n\n\tpropsQueryHierarchyItem.IsFolder = m.IsFolder\n\n\tpropsQueryHierarchyItem.IsInvalidSyntax = m.IsInvalidSyntax\n\n\tpropsQueryHierarchyItem.IsPublic = m.IsPublic\n\n\tpropsQueryHierarchyItem.LastExecutedBy = m.LastExecutedBy\n\n\tpropsQueryHierarchyItem.LastExecutedDate = m.LastExecutedDate\n\n\tpropsQueryHierarchyItem.LastModifiedBy = m.LastModifiedBy\n\n\tpropsQueryHierarchyItem.LastModifiedDate = m.LastModifiedDate\n\n\tpropsQueryHierarchyItem.LinkClauses = m.LinkClauses\n\n\tpropsQueryHierarchyItem.Name = m.Name\n\n\tpropsQueryHierarchyItem.Path = m.Path\n\n\tpropsQueryHierarchyItem.QueryRecursionOption = m.QueryRecursionOption\n\n\tpropsQueryHierarchyItem.QueryType = m.QueryType\n\n\tpropsQueryHierarchyItem.SortColumns = m.SortColumns\n\n\tpropsQueryHierarchyItem.SourceClauses = m.SourceClauses\n\n\tpropsQueryHierarchyItem.TargetClauses = m.TargetClauses\n\n\tpropsQueryHierarchyItem.Wiql = m.Wiql\n\n\tjsonDataPropsQueryHierarchyItem, errQueryHierarchyItem := swag.WriteJSON(propsQueryHierarchyItem)\n\tif errQueryHierarchyItem != nil {\n\t\treturn nil, errQueryHierarchyItem\n\t}\n\t_parts = append(_parts, jsonDataPropsQueryHierarchyItem)\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func Marshal(g *ag.Graph) ([]byte, error) {\n\tgv, err := graphviz.BuildGraph(g, graphviz.Options{ColoredTimeSteps: true})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn []byte(gv.String()), nil\n}", "func (s DescribeUserHierarchyStructureOutput) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.HierarchyStructure != nil {\n\t\tv := s.HierarchyStructure\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"HierarchyStructure\", v, metadata)\n\t}\n\treturn nil\n}", "func (l *Level) MarshalText() ([]byte, error) {\n\tif l == nil {\n\t\treturn nil, errMarshalNilLevel\n\t}\n\treturn []byte(l.String()), nil\n}", "func (l *Level) MarshalText() ([]byte, error) {\n\tif l == nil {\n\t\treturn nil, errMarshalNilLevel\n\t}\n\treturn []byte(l.String()), nil\n}", "func PrintObjectTree(htype *HType) string {\n\tif htype.children == nil {\n\t\treturn fmt.Sprintf(\"%d: %s\", htype.checkValue, htype.name)\n\t}\n\treturn fmt.Sprintf(\"%d: %s(%s)\", htype.checkValue, htype.name, strings.Join(forall(htype.children, PrintObjectTree), \", \"))\n}", "func (st *SlimTrie) Marshal() ([]byte, error) {\n\tvar buf []byte\n\twriter := bytes.NewBuffer(buf)\n\n\t_, err := pbcmpl.Marshal(writer, &versionedArray{&st.Children.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal children\")\n\t}\n\n\t_, err = pbcmpl.Marshal(writer, &versionedArray{&st.Steps.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal steps\")\n\t}\n\n\t_, err = pbcmpl.Marshal(writer, &versionedArray{&st.Leaves.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal leaves\")\n\t}\n\n\treturn writer.Bytes(), nil\n}", "func (f *Forest) ToString() string {\n\n\tfh := f.height\n\t// tree height should be 6 or less\n\tif fh > 6 {\n\t\treturn \"forest too big to print \"\n\t}\n\n\toutput := make([]string, (fh*2)+1)\n\tvar pos uint8\n\tfor h := uint8(0); h <= fh; h++ {\n\t\trowlen := uint8(1 << (fh - h))\n\n\t\tfor j := uint8(0); j < rowlen; j++ {\n\t\t\tvar valstring string\n\t\t\tok := f.data.size() >= uint64(pos)\n\t\t\tif ok {\n\t\t\t\tval := f.data.read(uint64(pos))\n\t\t\t\tif val != empty {\n\t\t\t\t\tvalstring = fmt.Sprintf(\"%x\", val[:2])\n\t\t\t\t}\n\t\t\t}\n\t\t\tif valstring != \"\" {\n\t\t\t\toutput[h*2] += fmt.Sprintf(\"%02d:%s \", pos, valstring)\n\t\t\t} else {\n\t\t\t\toutput[h*2] += fmt.Sprintf(\" \")\n\t\t\t}\n\t\t\tif h > 0 {\n\t\t\t\t//\t\t\t\tif x%2 == 0 {\n\t\t\t\toutput[(h*2)-1] += \"|-------\"\n\t\t\t\tfor q := uint8(0); q < ((1<<h)-1)/2; q++ {\n\t\t\t\t\toutput[(h*2)-1] += \"--------\"\n\t\t\t\t}\n\t\t\t\toutput[(h*2)-1] += \"\\\\ \"\n\t\t\t\tfor q := uint8(0); q < ((1<<h)-1)/2; q++ {\n\t\t\t\t\toutput[(h*2)-1] += \" \"\n\t\t\t\t}\n\n\t\t\t\t//\t\t\t\t}\n\n\t\t\t\tfor q := uint8(0); q < (1<<h)-1; q++ {\n\t\t\t\t\toutput[h*2] += \" \"\n\t\t\t\t}\n\n\t\t\t}\n\t\t\tpos++\n\t\t}\n\n\t}\n\tvar s string\n\tfor z := len(output) - 1; z >= 0; z-- {\n\t\ts += output[z] + \"\\n\"\n\t}\n\treturn s\n}", "func (s CreateUserHierarchyGroupOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (h Hls) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"fragmentsPerTsSegment\", h.FragmentsPerTsSegment)\n\treturn json.Marshal(objectMap)\n}", "func (ths *TreeHashStack) Serialize() ([]byte, error) {\n\tbuf := new(bytes.Buffer)\n\n\tif err := gob.NewEncoder(buf).Encode(ths); nil != err {\n\t\treturn nil, err\n\t}\n\n\treturn buf.Bytes(), nil\n}", "func (h Hash) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(h.String())\n}", "func (g *Group) Marshal() []byte {\n\tout := make([]byte, 0, 4*2*32 + 2*2*2*32)\n\tout = append(out, g.g1.Marshal()...)\n\tout = append(out, g.h.Marshal()...)\n\tout = append(out, g.u.Marshal()...)\n\tout = append(out, g.v.Marshal()...)\n\tout = append(out, g.g2.Marshal()...)\n\tout = append(out, g.w.Marshal()...)\n\treturn out\n}", "func (a Level) MarshalText() ([]byte, error) {\n\tswitch a {\n\tcase LevelNone:\n\t\treturn []byte(\"\"), nil\n\tcase LevelAnon:\n\t\treturn []byte(\"anon\"), nil\n\tcase LevelAuth:\n\t\treturn []byte(\"auth\"), nil\n\tcase LevelRoot:\n\t\treturn []byte(\"root\"), nil\n\tdefault:\n\t\treturn nil, errors.New(\"auth.Level: invalid level value\")\n\t}\n}", "func main() {\n\troot := TreeNode{\n\t\tVal: 1,\n\t\tLeft: &TreeNode{\n\t\t\tVal: 2,\n\t\t\tLeft: nil,\n\t\t\tRight: nil,\n\t\t},\n\t\tRight: &TreeNode{\n\t\t\tVal: 3,\n\t\t\tLeft: &TreeNode{\n\t\t\t\tVal: 4,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t\tRight: &TreeNode{\n\t\t\t\tVal: 5,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t},\n\t}\n\tobj := Constructor()\n\tdata := obj.serialize(&root)\n\tfmt.Println(data)\n}", "func (s DescribeUserHierarchyStructureOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeUserHierarchyStructureOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (node *Node) Pack() []byte {\n\tbfr := new(bytes.Buffer)\n\t// Write CATID\n\tbinary.Write(bfr, BIGEND, LBTYPE_CATID)\n\tbinary.Write(bfr, BIGEND, node.Id())\n\t// Write Catalog string key\n\tbinary.Write(bfr, BIGEND, LBTYPE_CATKEY)\n\tksz := AsLBUINT(len(node.Name()))\n\tbinary.Write(bfr, BIGEND, ksz)\n\tbfr.Write([]byte(node.Name()))\n\t// Write field map\n\tif node.HasFields() > 0 {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_MAP)\n\t\tbyts := node.FieldMap.ToBytes(node.debug)\n\t\tbinary.Write(bfr, BIGEND, AsLBUINT(len(byts)))\n\t\tbfr.Write(byts)\n\t} else {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_NIL)\n\t}\n\t// Write parents CATID set\n\tif node.HasParents() > 0 {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_CATID_SET)\n\t\tbyts := node.parents.ToBytes(node.debug)\n\t\tbinary.Write(bfr, BIGEND, AsLBUINT(len(byts)))\n\t\tbfr.Write(byts)\n\t} else {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_NIL)\n\t}\n\treturn bfr.Bytes()\n}", "func (t *ASCIITree) PrintTree(w io.Writer) {\n\tancestorPrefix := \"\"\n\tfor _, parent := range t.Ancestors() {\n\t\tif parent.Level() <= 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif parent.Last() {\n\t\t\tancestorPrefix += \" \"\n\t\t} else {\n\t\t\tancestorPrefix += \" │\"\n\t\t}\n\t}\n\n\tmyPrefix := \"\"\n\tmultilinePrefix := \"\"\n\tif t.Level() > 0 {\n\t\tif t.Last() {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" └── \"\n\t\t\t\tmultilinePrefix += \" \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" └─┬ \"\n\t\t\t\tmultilinePrefix += \" └─┬ \"\n\t\t\t}\n\t\t} else {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" ├── \"\n\t\t\t\tmultilinePrefix += \" │ \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" ├─┬ \"\n\t\t\t\tmultilinePrefix += \" │ │ \"\n\t\t\t}\n\t\t}\n\t}\n\n\tif t.Text != \"\" {\n\t\tlines := strings.Split(t.Text, \"\\n\")\n\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, myPrefix, lines[0])\n\t\tfor _, line := range lines[1:] {\n\t\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, multilinePrefix, line)\n\t\t}\n\t}\n\n\tfor _, child := range t.children {\n\t\tchild.PrintTree(w)\n\t}\n}", "func (m *binaryMaxentClassifier) marshal(path string) error {\n\tfolder := filepath.Join(path, \"Maxent\")\n\terr := os.Mkdir(folder, os.ModePerm)\n\tfor i, entry := range []string{\"labels\", \"mapping\", \"weights\"} {\n\t\tcomponent, _ := os.Create(filepath.Join(folder, entry+\".gob\"))\n\t\tencoder := gob.NewEncoder(component)\n\t\tif i == 0 {\n\t\t\tcheckError(encoder.Encode(m.labels))\n\t\t} else if i == 1 {\n\t\t\tcheckError(encoder.Encode(m.mapping))\n\t\t} else {\n\t\t\tcheckError(encoder.Encode(m.weights))\n\t\t}\n\t}\n\treturn err\n}", "func (v dHGroup) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson94b2531bEncodeGitRonaksoftComRiverWebWasmConnection1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (l Level) MarshalJSON() ([]byte, error) {\n\treturn []byte(`\"` + levels[l] + `\"`), nil\n}", "func stringify(n *BinarySearchNode, level int, builder *strings.Builder) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level, builder)\n\t\tbuilder.WriteString(fmt.Sprintf(format+\"%d\\n\", n.value))\n\t\tstringify(n.right, level, builder)\n\t}\n}", "func stringTree(builder *strings.Builder, depth int, node *TreeNode) (*strings.Builder, error) {\n\tvar err error\n\tfor i := 0; i < depth; i++ {\n\t\tif _, err = builder.WriteString(\" \"); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\tfmt.Fprintf(builder, \"└ %d\\n\", node.Val)\n\n\tif node.Left != nil {\n\t\tif builder, err = stringTree(builder, depth+1, node.Left); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\tif node.Right != nil {\n\t\tif builder, err = stringTree(builder, depth+1, node.Right); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\treturn builder, nil\n}", "func (s CreateUserHierarchyGroupInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func displayAsDefault(node *Node, level int) string {\n\tvar copies string\n\tif node.copies > 1 {\n\t\tcopies = fmt.Sprintf(\"| %3d %-6s\", node.copies, \"copies\")\n\t} else if node.copies == 1 {\n\t\tcopies = fmt.Sprintf(\"| %3d %-6s\", node.copies, \"copy\")\n\t}\n\n\tvar path = strings.TrimPrefix(node.fullPath, \"obj/\")\n\tpath = strings.TrimPrefix(path, \"lib/\")\n\tif level > 1 {\n\t\tpath = filepath.Base(path)\n\t}\n\tvar maxLen = 80 - 2*level\n\tif maxLen < 0 {\n\t\tmaxLen = 0\n\t}\n\tvar pathLength = len(path)\n\tif pathLength > maxLen {\n\t\tvar startPos = pathLength - maxLen + 3\n\t\tif startPos > pathLength {\n\t\t\tstartPos = pathLength\n\t\t}\n\t\tpath = \"...\" + path[startPos:]\n\t}\n\tpath = fmt.Sprintf(\"%s%s\", strings.Repeat(\" \", level), path)\n\tret := fmt.Sprintf(\"%-80s | %10s %10s\\n\", path, formatSize(node.size), copies)\n\n\t// Iterate over the childen in a sorted order.\n\tkeys := make([]string, 0, len(node.children))\n\tfor k := range node.children {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\n\tfor _, k := range keys {\n\t\tn := node.children[k]\n\t\tret += n.storageBreakdown(level + 1)\n\t}\n\treturn ret\n}", "func (m *MerkleTree) String() string {\n\ts := \"\"\n\tfor _, l := range m.Leafs {\n\t\ts += fmt.Sprint(l)\n\t\ts += \"\\n\"\n\t}\n\treturn s\n}", "func (f *FileChunk) Marshal() string {\n\treturn strings.Join([]string{\n\t\tf.name,\n\t\tf.uuid,\n\t\tfmt.Sprintf(\"%08X\", f.part),\n\t\tfmt.Sprintf(\"%08X\", f.total),\n\t\tf.base64Chunk,\n\t}, \":\")\n}", "func (s DescribeUserHierarchyGroupOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (uid MyULID) Marshal() ([]byte, error) {\n\tif len(uid) < 16 {\n\t\treturn []byte(ZeroULID), nil\n\t}\n\treturn []byte(uid)[:16], nil\n}", "func (t *smallFlatTable) Marshal(ctx MarshalContext, enc *marshal.Encoder) {\n\tt.initSubTables(ctx.ctx)\n\tenc.PutRawBytes(smallFlatMagic[:])\n\tenc.PutHash(t.hash)\n\tenc.PutGOB(&t.ast)\n\tenc.PutVarint(int64(len(t.subTables)))\n\tsubBufs := make([]marshal.Encoder, len(t.subTables))\n\tfor i := range t.subTables {\n\t\tt.subTables[i].Marshal(ctx, &subBufs[i])\n\t}\n\tfor i := range t.subTables {\n\t\tenc.PutRawBytes(subBufs[i].Bytes())\n\t}\n}", "func (s *RecursiveSchema) MarshalJSON() ([]byte, error) {\n\treturn []byte(fmt.Sprintf(`\"%s\"`, s.Actual.GetName())), nil\n}", "func (path DerivationPath) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(path.String())\n}", "func (a Level) String() string {\n\ts, err := a.MarshalText()\n\tif err != nil {\n\t\treturn \"unkn\"\n\t}\n\treturn string(s)\n}", "func (p *HJSON) Marshal(o map[string]interface{}) ([]byte, error) {\n\treturn hjson.Marshal(o)\n}", "func (pm Msg) pack() []byte {\n\tout := new(bytes.Buffer)\n\n\tfmt.Fprintf(out, \"%s\\n%s\\n%s\\n%s\\n%s\\n\", pm.Src, pm.Dst, pm.Wdir, pm.Type, pm.Attr.pack())\n\tln := len(pm.Data)\n\tfmt.Fprintf(out, \"%d\\n%s\", ln, pm.Data[:ln])\n\n\treturn out.Bytes()\n}", "func MarshalChaos(s interface{}) (string, error) {\n\tb, err := json.MarshalIndent(s, \"\", \" \")\n\tif err != nil {\n\t\treturn \"\", errors.Wrapf(err, \"failed to marshal indent\")\n\t}\n\treturn string(b), nil\n}", "func (h Hostname) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(string(h))\n}", "func (l *List) MarshalText() (text []byte, err error) {\n\theading := strings.TrimSpace(l.rootItem.desc)\n\ts := heading + \"\\n\"\n\ts += strings.Repeat(\"=\", len(heading)) + \"\\n\"\n\tfor _, li := range l.rootItem.items {\n\t\ts = recursiveMarshal(s, li, 0)\n\t}\n\treturn []byte(s), nil\n}", "func (npi NamedPartitionInformation) MarshalJSON() ([]byte, error) {\n\tnpi.ServicePartitionKind = ServicePartitionKindNamed1\n\tobjectMap := make(map[string]interface{})\n\tif npi.Name != nil {\n\t\tobjectMap[\"Name\"] = npi.Name\n\t}\n\tif npi.ID != nil {\n\t\tobjectMap[\"Id\"] = npi.ID\n\t}\n\tif npi.ServicePartitionKind != \"\" {\n\t\tobjectMap[\"ServicePartitionKind\"] = npi.ServicePartitionKind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func Marshal(c *configs.Config, schemas *terraform.Schemas) ([]byte, error) {\n\tvar output config\n\n\tpcs := make(map[string]providerConfig)\n\tmarshalProviderConfigs(c, schemas, pcs)\n\n\trootModule, err := marshalModule(c, schemas, \"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\toutput.RootModule = rootModule\n\n\tnormalizeModuleProviderKeys(&rootModule, pcs)\n\n\tfor name, pc := range pcs {\n\t\tif pc.parentKey != \"\" {\n\t\t\tdelete(pcs, name)\n\t\t}\n\t}\n\toutput.ProviderConfigs = pcs\n\n\tret, err := json.Marshal(output)\n\treturn ret, err\n}", "func (h HubRouteTable) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"etag\", h.Etag)\n\tpopulate(objectMap, \"id\", h.ID)\n\tpopulate(objectMap, \"name\", h.Name)\n\tpopulate(objectMap, \"properties\", h.Properties)\n\tpopulate(objectMap, \"type\", h.Type)\n\treturn json.Marshal(objectMap)\n}", "func (ll *LogLevel) MarshalText() ([]byte, error) {\n\treturn []byte(ll.String()), nil\n}", "func (s DescribeUserHierarchyStructureInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeUserHierarchyStructureInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (a Level) MarshalJSON() ([]byte, error) {\n\tres, err := a.MarshalText()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn append(append([]byte{'\"'}, res...), '\"'), nil\n}", "func (o OpenShiftCluster) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"id\", o.ID)\n\tpopulate(objectMap, \"location\", o.Location)\n\tpopulate(objectMap, \"name\", o.Name)\n\tpopulate(objectMap, \"properties\", o.Properties)\n\tpopulate(objectMap, \"systemData\", o.SystemData)\n\tpopulate(objectMap, \"tags\", o.Tags)\n\tpopulate(objectMap, \"type\", o.Type)\n\treturn json.Marshal(objectMap)\n}", "func (cfg frozenConfig) MarshalToString(val interface{}) (string, error) {\n buf, err := encoder.Encode(val, cfg.encoderOpts)\n return rt.Mem2Str(buf), err\n}", "func (this *Codec) serialize(root *TreeNode) string {\n if root == nil {\n return \"x\"\n }\n return strconv.Itoa(root.Val) + \",\" + this.serialize(root.Left)+ \",\" + this.serialize(root.Right)\n}", "func (s DescribeUserHierarchyGroupInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func Tree(e interface{}) string {\n\tswitch e := e.(type) {\n\tcase *gst.File:\n\t\treturn fmt.Sprintf(\"(package %s %s)\", e.PkgName, Tree(e.Decls))\n\tcase []gst.FuncDecl:\n\t\ts := \"\"\n\t\tfor _, fn := range e {\n\t\t\ts += fmt.Sprintf(\"func %s() {\\n%s\\n}\", fn.Name, Tree(fn.Body))\n\t\t}\n\t\treturn s\n\tcase *gst.RetStmt:\n\t\treturn fmt.Sprintf(\"ret\")\n\tcase *gst.ExprStmt:\n\t\treturn fmt.Sprintf(\"%s\", Tree(e.Exprs))\n\tcase value.Int:\n\t\treturn fmt.Sprintf(\"<int %s>\", e)\n\tcase variableExpr:\n\t\treturn fmt.Sprintf(\"<var %s>\", e.name)\n\tcase *unary:\n\t\treturn fmt.Sprintf(\"(%s %s)\", e.op, Tree(e.right))\n\tcase *binary:\n\t\t// Special case for [].\n\t\tif e.op == \"[]\" {\n\t\t\treturn fmt.Sprintf(\"(%s[%s])\", Tree(e.left), Tree(e.right))\n\t\t}\n\t\treturn fmt.Sprintf(\"(%s %s %s)\", Tree(e.left), e.op, Tree(e.right))\n\tcase sliceExpr:\n\t\ts := \"<TODO>\"\n\t\treturn s\n\tcase []value.Expr:\n\t\tif len(e) == 1 {\n\t\t\treturn Tree(e[0])\n\t\t}\n\t\ts := \"<\"\n\t\tfor i, expr := range e {\n\t\t\tif i > 0 {\n\t\t\t\ts += \"; \"\n\t\t\t}\n\t\t\ts += Tree(expr)\n\t\t}\n\t\ts += \">\"\n\t\treturn s\n\tdefault:\n\t\treturn fmt.Sprintf(\"%T\", e)\n\t}\n}", "func (d *DOM) Marshal() (string, error) {\n\tb, err := json.Marshal(d)\n\treturn string(b), err\n}", "func (h Hostname) MarshalText() ([]byte, error) {\n\treturn []byte(string(h)), nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\treturn dfsSerial(root, \"\")\n}", "func marshalData(t testing.TB, tree map[string]string) (map[string]string, []byte) {\n\tsha1tree := map[string]string{}\n\tentries := &dumbcaslib.Entry{}\n\tfor k, v := range tree {\n\t\th := dumbcaslib.Sha1Bytes([]byte(v))\n\t\tsha1tree[k] = h\n\t\te := entries\n\t\tparts := strings.Split(k, \"/\")\n\t\tfor i := 0; i < len(parts)-1; i++ {\n\t\t\tif e.Files == nil {\n\t\t\t\te.Files = map[string]*dumbcaslib.Entry{}\n\t\t\t}\n\t\t\tif e.Files[parts[i]] == nil {\n\t\t\t\te.Files[parts[i]] = &dumbcaslib.Entry{}\n\t\t\t}\n\t\t\te = e.Files[parts[i]]\n\t\t}\n\t\tif e.Files == nil {\n\t\t\te.Files = map[string]*dumbcaslib.Entry{}\n\t\t}\n\t\te.Files[parts[len(parts)-1]] = &dumbcaslib.Entry{\n\t\t\tSha1: h,\n\t\t\tSize: int64(len(v)),\n\t\t}\n\t}\n\n\t// Then process entries itself.\n\tdata, err := json.Marshal(entries)\n\tut.AssertEqual(t, nil, err)\n\treturn sha1tree, data\n}", "func (h Hash) MarshalText() ([]byte, error) {\n\treturn []byte(h.String()), nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tstrs := make([]string, 0)\n\tvar preorder func(*TreeNode)\n\tpreorder = func(node *TreeNode) {\n\t\tif node == nil {\n\t\t\tstrs = append(strs, this.null)\n\t\t\treturn\n\t\t}\n\n\t\tstrs = append(strs, strconv.Itoa(node.Val))\n\t\tpreorder(node.Left)\n\t\tpreorder(node.Right)\n\t}\n\tpreorder(root)\n\treturn strings.Join(strs, this.sep)\n}", "func (r *Key) marshal(c *Client) ([]byte, error) {\n\tm, err := expandKey(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling Key: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func TraverseLevelOrder(n *Node) string {\n\tvar s string\n\tif n == nil {\n\t\treturn \"\"\n\t}\n\tq := list.New()\n\tq.PushBack(n)\n\tfor q.Len() > 0 {\n\t\ttemp := q.Front()\n\t\tq.Remove(temp)\n\t\tcurrent := temp.Value.(*Node)\n\t\ts = s + strconv.Itoa(current.Data) + \" \"\n\t\tif current.Left != nil {\n\t\t\tq.PushBack(current.Left)\n\t\t}\n\t\tif current.Right != nil {\n\t\t\tq.PushBack(current.Right)\n\t\t}\n\t}\n\treturn strings.TrimSpace(s)\n}", "func (t ThreeTierFullResourceNames) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"applicationServer\", t.ApplicationServer)\n\tpopulate(objectMap, \"centralServer\", t.CentralServer)\n\tpopulate(objectMap, \"databaseServer\", t.DatabaseServer)\n\tobjectMap[\"namingPatternType\"] = NamingPatternTypeFullResourceName\n\tpopulate(objectMap, \"sharedStorage\", t.SharedStorage)\n\treturn json.Marshal(objectMap)\n}", "func (sh SubdirectoryHeader) toBytes() []byte {\n\tbuf := make([]byte, 0x27)\n\tbuf[0] = sh.TypeAndNameLength\n\tcopyBytes(buf[1:0x10], sh.SubdirectoryName[:])\n\tbuf[0x10] = sh.SeventyFive\n\tcopyBytes(buf[0x11:0x18], sh.Unused1[:])\n\tcopyBytes(buf[0x18:0x1c], sh.Creation.toBytes())\n\tbuf[0x1c] = sh.Version\n\tbuf[0x1d] = sh.MinVersion\n\tbuf[0x1e] = byte(sh.Access)\n\tbuf[0x1f] = sh.EntryLength\n\tbuf[0x20] = sh.EntriesPerBlock\n\tbinary.LittleEndian.PutUint16(buf[0x21:0x23], sh.FileCount)\n\tbinary.LittleEndian.PutUint16(buf[0x23:0x25], sh.ParentPointer)\n\tbuf[0x25] = sh.ParentEntry\n\tbuf[0x26] = sh.ParentEntryLength\n\treturn buf\n}", "func MarshallHcl(conf Config, path string) string {\n\tf := hclwrite.NewEmptyFile()\n\tgohcl.EncodeIntoBody(&conf, f.Body())\n\tr := regexp.MustCompile(\"[$]+\")\n\tb := r.ReplaceAllLiteral(f.Bytes(), []byte(\"$\"))\n\tdestination, err := os.Create(path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\t_, writeErr := destination.Write(b)\n\t// _, writeErr := f.WriteTo(destination)\n\tif writeErr != nil {\n\t\tlog.Error(\"write error\")\n\t}\n\tdestination.Close()\n\treturn path\n}", "func (p PropagatedRouteTable) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"ids\", p.IDs)\n\tpopulate(objectMap, \"labels\", p.Labels)\n\treturn json.Marshal(objectMap)\n}", "func (d Dir) MarshalJSON() ([]byte, error) {\n\tbuffer := bytes.NewBufferString(`\"`)\n\tbuffer.WriteString(toString[d])\n\tbuffer.WriteString(`\"`)\n\treturn buffer.Bytes(), nil\n}", "func (this *family) MarshalBinary() (data []byte, err error) {\n\t// var addrs = map[uintptr]uint64{}\n\tbuf := bytes.NewBuffer(make([]byte, 0, ((8 + len(this.LastName)) + (8 + ((len(this.Pets) * 15) + (len(this.Pets) * 9))))))\n\tif err = this.marshalTo(buf /*, addrs*/); err == nil {\n\t\tdata = buf.Bytes()\n\t}\n\treturn\n}", "func (r *HealthCheck) marshal(c *Client) ([]byte, error) {\n\tm, err := expandHealthCheck(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling HealthCheck: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func (h HubRouteTableProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"associatedConnections\", h.AssociatedConnections)\n\tpopulate(objectMap, \"labels\", h.Labels)\n\tpopulate(objectMap, \"propagatingConnections\", h.PropagatingConnections)\n\tpopulate(objectMap, \"provisioningState\", h.ProvisioningState)\n\tpopulate(objectMap, \"routes\", h.Routes)\n\treturn json.Marshal(objectMap)\n}", "func (m HyperflexNode) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 2)\n\n\taO0, err := swag.WriteJSON(m.MoBaseMo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\tvar dataAO1 struct {\n\t\tBuildNumber string `json:\"BuildNumber,omitempty\"`\n\n\t\tCluster *HyperflexClusterRef `json:\"Cluster,omitempty\"`\n\n\t\tClusterMember *AssetClusterMemberRef `json:\"ClusterMember,omitempty\"`\n\n\t\tDisplayVersion string `json:\"DisplayVersion,omitempty\"`\n\n\t\tHostName string `json:\"HostName,omitempty\"`\n\n\t\tHypervisor string `json:\"Hypervisor,omitempty\"`\n\n\t\tIdentity *HyperflexHxUuIDDt `json:\"Identity,omitempty\"`\n\n\t\tIP *HyperflexHxNetworkAddressDt `json:\"Ip,omitempty\"`\n\n\t\tLockdown *bool `json:\"Lockdown,omitempty\"`\n\n\t\tModelNumber string `json:\"ModelNumber,omitempty\"`\n\n\t\tPhysicalServer *ComputePhysicalRef `json:\"PhysicalServer,omitempty\"`\n\n\t\tRole string `json:\"Role,omitempty\"`\n\n\t\tSerialNumber string `json:\"SerialNumber,omitempty\"`\n\n\t\tStatus string `json:\"Status,omitempty\"`\n\n\t\tVersion string `json:\"Version,omitempty\"`\n\t}\n\n\tdataAO1.BuildNumber = m.BuildNumber\n\n\tdataAO1.Cluster = m.Cluster\n\n\tdataAO1.ClusterMember = m.ClusterMember\n\n\tdataAO1.DisplayVersion = m.DisplayVersion\n\n\tdataAO1.HostName = m.HostName\n\n\tdataAO1.Hypervisor = m.Hypervisor\n\n\tdataAO1.Identity = m.Identity\n\n\tdataAO1.IP = m.IP\n\n\tdataAO1.Lockdown = m.Lockdown\n\n\tdataAO1.ModelNumber = m.ModelNumber\n\n\tdataAO1.PhysicalServer = m.PhysicalServer\n\n\tdataAO1.Role = m.Role\n\n\tdataAO1.SerialNumber = m.SerialNumber\n\n\tdataAO1.Status = m.Status\n\n\tdataAO1.Version = m.Version\n\n\tjsonDataAO1, errAO1 := swag.WriteJSON(dataAO1)\n\tif errAO1 != nil {\n\t\treturn nil, errAO1\n\t}\n\t_parts = append(_parts, jsonDataAO1)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (s DescribeUserHierarchyStructureInput) MarshalFields(e protocol.FieldEncoder) error {\n\te.SetValue(protocol.HeaderTarget, \"Content-Type\", protocol.StringValue(\"application/json\"), protocol.Metadata{})\n\n\tif s.InstanceId != nil {\n\t\tv := *s.InstanceId\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.PathTarget, \"InstanceId\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\treturn nil\n}", "func (n Node) String() string {\n\tc := n.GetConfig()\n\n\ts := n.Info.Name() + \"\\n\"\n\tif n.getDepth() == 0 || c.FullPaths {\n\t\ts = n.Path + \"\\n\"\n\t}\n\n\tfor _, v := range n.Children {\n\t\tif !c.DisableIndentation {\n\t\t\ts += v.generatePrefix()\n\t\t}\n\t\ts += v.String()\n\t}\n\n\treturn s\n}", "func (h H) MarshalXML(e *xml.Encoder, start xml.StartElement) error {\n\tstart.Name = xml.Name{\n\t\tSpace: \"\",\n\t\tLocal: \"map\",\n\t}\n\tif err := e.EncodeToken(start); err != nil {\n\t\treturn err\n\t}\n\tfor key, value := range h {\n\t\telem := xml.StartElement{\n\t\t\tName: xml.Name{Space: \"\", Local: key},\n\t\t\tAttr: []xml.Attr{},\n\t\t}\n\t\tif err := e.EncodeElement(value, elem); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif err := e.EncodeToken(xml.EndElement{Name: start.Name}); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (h Horse) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"isAShowHorse\", h.IsAShowHorse)\n\tpopulate(objectMap, \"name\", h.Name)\n\treturn json.Marshal(objectMap)\n}", "func (o *Organism) MarshalBinary() ([]byte, error) {\n\tvar buf bytes.Buffer\n\t_, err := fmt.Fprintln(&buf, o.Fitness, o.Generation, o.highestFitness, o.isPopulationChampionChild, o.Genotype.Id)\n\to.Genotype.Write(&buf)\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\treturn buf.Bytes(), nil\n\t}\n}", "func (r *AWSEMRCluster) MarshalJSON() ([]byte, error) {\n\ttype Properties AWSEMRCluster\n\treturn json.Marshal(&struct {\n\t\tType string\n\t\tProperties Properties\n\t}{\n\t\tType: r.AWSCloudFormationType(),\n\t\tProperties: (Properties)(*r),\n\t})\n}", "func (e Level) MarshalJSON() ([]byte, error) {\n\treturn []byte(fmt.Sprintf(`\"%v\"`, e)), nil\n}", "func (set *Set) String() string {\n\tstr := \"TreeSet\\n\"\n\titems := []string{}\n\tfor _, v := range set.tree.Keys() {\n\t\titems = append(items, fmt.Sprintf(\"%v\", v))\n\t}\n\tstr += strings.Join(items, \", \")\n\treturn str\n}", "func (node *URLNode) generateTree(tabSize int) string {\n\tsubTree := \"\"\n\tfor _, child := range node.children {\n\t\tline := strings.Repeat(\"\\t\", tabSize)\n\t\tline += \"└── \"\n\t\tsubTree += line + child.generateTree(tabSize+1)\n\t}\n\treturn node.url + \"\\n\" + subTree\n}", "func (npsd NamedPartitionSchemeDescription) MarshalJSON() ([]byte, error) {\n\tnpsd.PartitionScheme = PartitionSchemeNamed1\n\tobjectMap := make(map[string]interface{})\n\tif npsd.Count != nil {\n\t\tobjectMap[\"Count\"] = npsd.Count\n\t}\n\tif npsd.Names != nil {\n\t\tobjectMap[\"Names\"] = npsd.Names\n\t}\n\tif npsd.PartitionScheme != \"\" {\n\t\tobjectMap[\"PartitionScheme\"] = npsd.PartitionScheme\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tif root == nil {\n\t\treturn \"\"\n\t}\n\tans := make([]string, 0, 10)\n\tserialize(root, &ans)\n\n\treturn strings.Join(ans, \",\")\n}", "func (id NodeID) Marshal() ([]byte, error) {\n\treturn id.Bytes(), nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tvar res string\n\n\tvar dfs func(node *TreeNode)\n\tdfs = func(node *TreeNode) {\n\t\tif node == nil {\n\t\t\tres += \"null,\"\n\t\t\treturn\n\t\t}\n\n\t\tres += fmt.Sprintf(\"%+v\", node.Val) + \",\"\n\t\tdfs(node.Left)\n\t\tdfs(node.Right)\n\t}\n\n\tdfs(root)\n\n\treturn res\n}" ]
[ "0.73076534", "0.71470153", "0.564921", "0.5278168", "0.52564615", "0.51511955", "0.5093083", "0.50881916", "0.50562584", "0.50160295", "0.50160295", "0.49893352", "0.49387497", "0.49149144", "0.4902577", "0.48992035", "0.488287", "0.48746312", "0.48374742", "0.48336658", "0.4802439", "0.47979125", "0.47972766", "0.47972766", "0.47807863", "0.47284326", "0.47249743", "0.46696383", "0.4662889", "0.46598524", "0.46467835", "0.46445036", "0.4643782", "0.4641022", "0.4628852", "0.4628852", "0.46248138", "0.46051297", "0.4602882", "0.4579795", "0.45608208", "0.4557029", "0.4541013", "0.45363593", "0.45279288", "0.45266986", "0.45249856", "0.45127046", "0.45105448", "0.4484179", "0.44804266", "0.44722936", "0.44654074", "0.44589603", "0.44587654", "0.44503808", "0.44461688", "0.44457498", "0.4442993", "0.44167617", "0.44145352", "0.4411263", "0.44103906", "0.44103906", "0.44054565", "0.4398999", "0.43974108", "0.43933988", "0.43879083", "0.4383588", "0.43818152", "0.43781936", "0.43729973", "0.43716815", "0.43657368", "0.43650118", "0.4362418", "0.43557957", "0.43399256", "0.4335358", "0.43329313", "0.433027", "0.43280244", "0.43254805", "0.43252853", "0.4321901", "0.43205002", "0.43193206", "0.43129233", "0.43121624", "0.43111157", "0.4309705", "0.43017998", "0.42997465", "0.4294033", "0.4291887", "0.42901626", "0.42894512", "0.42886543", "0.42886353" ]
0.69061786
2
MarshalHierarchical returns hierarchical string representation.
func (status MirrorStatus) MarshalHierarchical() string { hierarchy := hierr.Push( status.Name, fmt.Sprintf("state: %s", status.State), ) if status.ModifyDate > 0 { hierarchy = hierr.Push( hierarchy, fmt.Sprintf("modify date: %v", status.ModifyDate), ) } return hierr.String(hierarchy) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (status UpstreamStatus) MarshalHierarchical() string {\n\thierarchy := hierr.Push(\n\t\t\"upstream\",\n\t\t\"total: \"+strconv.Itoa(status.Total),\n\t\tfmt.Sprintf(\n\t\t\t\"success: %d (%.2f%%)\",\n\t\t\tstatus.Success, status.SuccessPercent,\n\t\t),\n\t\tfmt.Sprintf(\n\t\t\t\"error: %d (%.2f%%)\",\n\t\t\tstatus.Error, status.ErrorPercent,\n\t\t),\n\t)\n\n\tif len(status.Slaves) > 0 {\n\t\tslaves := errors.New(\"slaves\")\n\t\tfor _, slave := range status.Slaves {\n\t\t\tslaves = hierr.Push(slaves, slave.MarshalHierarchical())\n\t\t}\n\n\t\thierarchy = hierr.Push(hierarchy, slaves)\n\t}\n\n\treturn hierr.String(hierarchy)\n}", "func (status ServerStatus) MarshalHierarchical() []byte {\n\tvar hierarchy error\n\tif status.Address != \"\" {\n\t\thierarchy = hierr.Push(status.Address)\n\t} else {\n\t\thierarchy = hierr.Push(\"status\")\n\t}\n\n\tif status.Role != \"master\" {\n\t\tstatus.Role = \"slave\"\n\t}\n\n\thierarchy = hierr.Push(\n\t\thierarchy,\n\t\tfmt.Sprintf(\"role: %s\", status.Role),\n\t)\n\n\thierarchy = hierr.Push(\n\t\thierarchy,\n\t\tfmt.Sprintf(\"total: %d\", len(status.Mirrors)),\n\t)\n\n\tif status.HierarchicalError != \"\" {\n\t\thierarchy = hierr.Push(\n\t\t\thierarchy,\n\t\t\thierr.Push(\"error\", status.HierarchicalError),\n\t\t)\n\t}\n\n\tif len(status.Mirrors) > 0 {\n\t\tmirrors := errors.New(\"mirrors\")\n\t\tfor _, mirror := range status.Mirrors {\n\t\t\tmirrors = hierr.Push(mirrors, mirror.MarshalHierarchical())\n\t\t}\n\t\thierarchy = hierr.Push(hierarchy, mirrors)\n\t}\n\n\tif status.Role == \"master\" {\n\t\thierarchy = hierr.Push(hierarchy, status.Upstream.MarshalHierarchical())\n\t}\n\n\treturn []byte(hierr.String(hierarchy))\n}", "func (rht *RHT) Marshal() string {\n\tmembers := rht.Members()\n\n\tsize := len(members)\n\tkeys := make([]string, 0, size)\n\tfor k := range members {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\n\tsb := strings.Builder{}\n\tsb.WriteString(\"{\")\n\n\tidx := 0\n\tfor _, k := range keys {\n\t\tvalue := members[k]\n\t\tsb.WriteString(fmt.Sprintf(\"\\\"%s\\\":%s\", k, value.Marshal()))\n\t\tif size-1 != idx {\n\t\t\tsb.WriteString(\",\")\n\t\t}\n\t\tidx++\n\t}\n\tsb.WriteString(\"}\")\n\n\treturn sb.String()\n}", "func (s HierarchyLevel) String() string {\n\treturn awsutil.Prettify(s)\n}", "func encodeTree(hmt *Tree, finalTree *string) {\n\tif hmt == nil {\n\t\treturn\n\t}\n\t\n\tif hmt.LeftNode == nil && hmt.RightNode == nil{\n\t\t*finalTree += \"1\" + string(hmt.Char)\n\t} else {\n\t\t*finalTree += \"0\"\n\t}\n\tencodeTree(hmt.LeftNode, finalTree)\n\tencodeTree(hmt.RightNode, finalTree) \n}", "func (s HierarchyStructure) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (t *Tree) formatStr() []byte {\n\tvar s []byte\n\ts = append(s, t.format...)\n\tif t.parent != nil {\n\t\tt := t.parent.formatStr()\n\t\tt = append(t, s...)\n\t\ts = t\n\t}\n\treturn s\n}", "func (m QueryHierarchyItem) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 1)\n\n\taO0, err := swag.WriteJSON(m.WorkItemTrackingResource)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\t// now for regular properties\n\tvar propsQueryHierarchyItem struct {\n\t\tChildren []*QueryHierarchyItem `json:\"children\"`\n\n\t\tClauses *WorkItemQueryClause `json:\"clauses,omitempty\"`\n\n\t\tColumns []*WorkItemFieldReference `json:\"columns\"`\n\n\t\tCreatedBy *IdentityReference `json:\"createdBy,omitempty\"`\n\n\t\tCreatedDate strfmt.DateTime `json:\"createdDate,omitempty\"`\n\n\t\tFilterOptions interface{} `json:\"filterOptions,omitempty\"`\n\n\t\tHasChildren bool `json:\"hasChildren,omitempty\"`\n\n\t\tID strfmt.UUID `json:\"id,omitempty\"`\n\n\t\tIsDeleted bool `json:\"isDeleted,omitempty\"`\n\n\t\tIsFolder bool `json:\"isFolder,omitempty\"`\n\n\t\tIsInvalidSyntax bool `json:\"isInvalidSyntax,omitempty\"`\n\n\t\tIsPublic bool `json:\"isPublic,omitempty\"`\n\n\t\tLastExecutedBy *IdentityReference `json:\"lastExecutedBy,omitempty\"`\n\n\t\tLastExecutedDate strfmt.DateTime `json:\"lastExecutedDate,omitempty\"`\n\n\t\tLastModifiedBy *IdentityReference `json:\"lastModifiedBy,omitempty\"`\n\n\t\tLastModifiedDate strfmt.DateTime `json:\"lastModifiedDate,omitempty\"`\n\n\t\tLinkClauses *WorkItemQueryClause `json:\"linkClauses,omitempty\"`\n\n\t\tName string `json:\"name,omitempty\"`\n\n\t\tPath string `json:\"path,omitempty\"`\n\n\t\tQueryRecursionOption interface{} `json:\"queryRecursionOption,omitempty\"`\n\n\t\tQueryType interface{} `json:\"queryType,omitempty\"`\n\n\t\tSortColumns []*WorkItemQuerySortColumn `json:\"sortColumns\"`\n\n\t\tSourceClauses *WorkItemQueryClause `json:\"sourceClauses,omitempty\"`\n\n\t\tTargetClauses *WorkItemQueryClause `json:\"targetClauses,omitempty\"`\n\n\t\tWiql string `json:\"wiql,omitempty\"`\n\t}\n\tpropsQueryHierarchyItem.Children = m.Children\n\n\tpropsQueryHierarchyItem.Clauses = m.Clauses\n\n\tpropsQueryHierarchyItem.Columns = m.Columns\n\n\tpropsQueryHierarchyItem.CreatedBy = m.CreatedBy\n\n\tpropsQueryHierarchyItem.CreatedDate = m.CreatedDate\n\n\tpropsQueryHierarchyItem.FilterOptions = m.FilterOptions\n\n\tpropsQueryHierarchyItem.HasChildren = m.HasChildren\n\n\tpropsQueryHierarchyItem.ID = m.ID\n\n\tpropsQueryHierarchyItem.IsDeleted = m.IsDeleted\n\n\tpropsQueryHierarchyItem.IsFolder = m.IsFolder\n\n\tpropsQueryHierarchyItem.IsInvalidSyntax = m.IsInvalidSyntax\n\n\tpropsQueryHierarchyItem.IsPublic = m.IsPublic\n\n\tpropsQueryHierarchyItem.LastExecutedBy = m.LastExecutedBy\n\n\tpropsQueryHierarchyItem.LastExecutedDate = m.LastExecutedDate\n\n\tpropsQueryHierarchyItem.LastModifiedBy = m.LastModifiedBy\n\n\tpropsQueryHierarchyItem.LastModifiedDate = m.LastModifiedDate\n\n\tpropsQueryHierarchyItem.LinkClauses = m.LinkClauses\n\n\tpropsQueryHierarchyItem.Name = m.Name\n\n\tpropsQueryHierarchyItem.Path = m.Path\n\n\tpropsQueryHierarchyItem.QueryRecursionOption = m.QueryRecursionOption\n\n\tpropsQueryHierarchyItem.QueryType = m.QueryType\n\n\tpropsQueryHierarchyItem.SortColumns = m.SortColumns\n\n\tpropsQueryHierarchyItem.SourceClauses = m.SourceClauses\n\n\tpropsQueryHierarchyItem.TargetClauses = m.TargetClauses\n\n\tpropsQueryHierarchyItem.Wiql = m.Wiql\n\n\tjsonDataPropsQueryHierarchyItem, errQueryHierarchyItem := swag.WriteJSON(propsQueryHierarchyItem)\n\tif errQueryHierarchyItem != nil {\n\t\treturn nil, errQueryHierarchyItem\n\t}\n\t_parts = append(_parts, jsonDataPropsQueryHierarchyItem)\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (s HierarchyGroup) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (rh ResourceHierarchy) String() string {\n\treturn string(rh)\n}", "func (s HierarchyPath) String() string {\n\treturn awsutil.Prettify(s)\n}", "func WriteTree(writer io.Writer, hierarchy *Hierarchy, includeEmpty bool) {\n\ttree := assembleTree(hierarchy)\n\tkeys := make([]string, len(tree))\n\ti := 0\n\tfor k := range tree {\n\t\tkeys[i] = k\n\t\ti++\n\t}\n\tsort.Strings(keys)\n\tfor _, key := range keys {\n\t\tb := tree[key]\n\t\twriteBranch(writer, b, \"\", hierarchy, includeEmpty)\n\t}\n}", "func stringify(n *Node, level int) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level)\n\t\tfmt.Printf(format+\"%d\\n\", n.key)\n\t\tstringify(n.right, level)\n\t}\n}", "func (this *Codec) serialize(root *TreeNode) string {\n var s string \n res:=helpSerialize(root,s)\n fmt.Println(res)\n return res\n}", "func stringify(treeNode *TreeNode, level int) {\n\tif treeNode != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"***> \"\n\t\tlevel++\n\t\tstringify(treeNode.leftNode, level)\n\t\tfmt.Printf(format+\"%d\\n\", treeNode.key)\n\t\tstringify(treeNode.rightNode, level)\n\t}\n}", "func (s CreateUserHierarchyGroupOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (w *RootWalker) String() (out string) {\n\ttabs := func(n int) {\n\t\tfor i := 0; i < n; i++ {\n\t\t\tout += \"\\t\"\n\t\t}\n\t}\n\tout += fmt.Sprint(\"Root\")\n\tsize := w.Size()\n\tif size == 0 {\n\t\treturn\n\t}\n\tout += fmt.Sprintf(\".Refs[%d] ->\\n\", w.stack[0].prevInFieldIndex)\n\tfor i, obj := range w.stack {\n\t\tschName := \"\"\n\t\ts, _ := w.r.SchemaByReference(obj.s)\n\t\tif s != nil {\n\t\t\tschName = s.Name()\n\t\t}\n\n\t\ttabs(i)\n\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\tout += fmt.Sprintf(` = \"%v\"`+\"\\n\", obj.p)\n\n\t\ttabs(i)\n\t\tif obj.next != nil {\n\t\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\t\tout += fmt.Sprintf(\".%s\", obj.next.prevFieldName)\n\t\t\tif obj.next.prevInFieldIndex != -1 {\n\t\t\t\tout += fmt.Sprintf(\"[%d]\", obj.next.prevInFieldIndex)\n\t\t\t}\n\t\t\tout += fmt.Sprint(\" ->\\n\")\n\t\t}\n\t}\n\treturn\n}", "func (w *RootWalker) String() (out string) {\n\ttabs := func(n int) {\n\t\tfor i := 0; i < n; i++ {\n\t\t\tout += \"\\t\"\n\t\t}\n\t}\n\tout += fmt.Sprint(\"Root\")\n\tsize := w.Size()\n\tif size == 0 {\n\t\treturn\n\t}\n\tout += fmt.Sprintf(\".Refs[%d] ->\\n\", w.stack[0].prevInFieldIndex)\n\tfor i, obj := range w.stack {\n\t\tschName := \"\"\n\t\ts, _ := w.r.SchemaByReference(obj.s)\n\t\tif s != nil {\n\t\t\tschName = s.Name()\n\t\t}\n\n\t\ttabs(i)\n\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\tout += fmt.Sprintf(` = \"%v\"`+\"\\n\", obj.p)\n\n\t\ttabs(i)\n\t\tif obj.next != nil {\n\t\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\t\tout += fmt.Sprintf(\".%s\", obj.next.prevFieldName)\n\t\t\tif obj.next.prevInFieldIndex != -1 {\n\t\t\t\tout += fmt.Sprintf(\"[%d]\", obj.next.prevInFieldIndex)\n\t\t\t}\n\t\t\tout += fmt.Sprint(\" ->\\n\")\n\t\t}\n\t}\n\treturn\n}", "func (h Hls) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"fragmentsPerTsSegment\", h.FragmentsPerTsSegment)\n\treturn json.Marshal(objectMap)\n}", "func (s DescribeUserHierarchyStructureOutput) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.HierarchyStructure != nil {\n\t\tv := s.HierarchyStructure\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"HierarchyStructure\", v, metadata)\n\t}\n\treturn nil\n}", "func (ths *TreeHashStack) Serialize() ([]byte, error) {\n\tbuf := new(bytes.Buffer)\n\n\tif err := gob.NewEncoder(buf).Encode(ths); nil != err {\n\t\treturn nil, err\n\t}\n\n\treturn buf.Bytes(), nil\n}", "func (v dHGroup) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson94b2531bEncodeGitRonaksoftComRiverWebWasmConnection1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (p *page) marshal() string {\n\tb := &strings.Builder{}\n\tfmt.Fprintf(b, \"\\n%d 0 obj\\n<<\\n\", p.id)\n\tfmt.Fprintf(b, \" /MediaBox [ %f %f %f %f ]\\n\", p.mediaBox.llx, p.mediaBox.lly, p.mediaBox.urx, p.mediaBox.ury)\n\tfmt.Fprintf(b, \" /CropBox [ %f %f %f %f ]\\n\", p.cropBox.llx, p.cropBox.lly, p.cropBox.urx, p.cropBox.ury)\n\tfmt.Fprintf(b, \" /BleedBox [ %f %f %f %f ]\\n\", p.bleedBox.llx, p.bleedBox.lly, p.bleedBox.urx, p.bleedBox.ury)\n\tfmt.Fprintf(b, \" /TrimBox [ %f %f %f %f ]\\n\", p.trimBox.llx, p.trimBox.lly, p.trimBox.urx, p.trimBox.ury)\n\tfmt.Fprintf(b, \" /Contents [ \")\n\tfor _, cid := range p.contentIds {\n\t\tfmt.Fprintf(b, \" %d 0 R \", cid)\n\t}\n\tfmt.Fprintf(b, \" ]\\n\")\n\tfmt.Fprintf(b, \" /Parent %d 0 R\\n\", p.parentID)\n\tb.WriteString(p.raw)\n\tfmt.Fprintf(b, \"\\n>>\\nendobj\\n\")\n\treturn b.String()\n}", "func (l *Level) MarshalText() ([]byte, error) {\n\tif l == nil {\n\t\treturn nil, errMarshalNilLevel\n\t}\n\treturn []byte(l.String()), nil\n}", "func (l *Level) MarshalText() ([]byte, error) {\n\tif l == nil {\n\t\treturn nil, errMarshalNilLevel\n\t}\n\treturn []byte(l.String()), nil\n}", "func (h Hash) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(h.String())\n}", "func (st *SlimTrie) Marshal() ([]byte, error) {\n\tvar buf []byte\n\twriter := bytes.NewBuffer(buf)\n\n\t_, err := pbcmpl.Marshal(writer, &versionedArray{&st.Children.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal children\")\n\t}\n\n\t_, err = pbcmpl.Marshal(writer, &versionedArray{&st.Steps.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal steps\")\n\t}\n\n\t_, err = pbcmpl.Marshal(writer, &versionedArray{&st.Leaves.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal leaves\")\n\t}\n\n\treturn writer.Bytes(), nil\n}", "func (g *Group) Marshal() []byte {\n\tout := make([]byte, 0, 4*2*32 + 2*2*2*32)\n\tout = append(out, g.g1.Marshal()...)\n\tout = append(out, g.h.Marshal()...)\n\tout = append(out, g.u.Marshal()...)\n\tout = append(out, g.v.Marshal()...)\n\tout = append(out, g.g2.Marshal()...)\n\tout = append(out, g.w.Marshal()...)\n\treturn out\n}", "func (s DescribeUserHierarchyStructureOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeUserHierarchyStructureOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s CreateUserHierarchyGroupInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func ToString(ctx context.Context) (ret string) {\n\tfather, ok := ctx.Value(key).(*constNode)\n\tif !ok {\n\t\treturn \"\"\n\t}\n\tDone(ctx)\n\tconst fmtStr = \"%s%s (%dms %d%%)\\n\"\n\tvar levelPrint func(level int, node *constNode, prefix string)\n\tlevelPrint = func(level int, node *constNode, prefix string) {\n\t\tvar (\n\t\t\tlastTabs string\n\t\t\tnoLastTabs string\n\t\t)\n\t\tnoLastTabs = prefix + \"├─\"\n\t\tlastTabs = prefix + \"└─\"\n\t\tfor i, child := range node.child {\n\t\t\ttabs := noLastTabs\n\t\t\tif i == len(node.child)-1 {\n\t\t\t\ttabs = lastTabs\n\t\t\t}\n\t\t\tchildCostMs := child.cost().Milliseconds()\n\t\t\tfatherCostMs := node.cost().Milliseconds()\n\t\t\tradio := int64(0)\n\t\t\tif fatherCostMs > 0 {\n\t\t\t\tradio = childCostMs * 100 / fatherCostMs\n\t\t\t}\n\t\t\tret += fmt.Sprintf(fmtStr, tabs, child.title, childCostMs, radio)\n\t\t\tif len(child.child) > 0 {\n\t\t\t\tif i == len(node.child)-1 {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\" \")\n\t\t\t\t} else {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\"│ \")\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tret += fmt.Sprintf(fmtStr, \"\", father.title, father.cost().Milliseconds(), 100)\n\tlevelPrint(0, father, \"\")\n\treturn\n}", "func (s DescribeUserHierarchyGroupOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func Marshal(g *ag.Graph) ([]byte, error) {\n\tgv, err := graphviz.BuildGraph(g, graphviz.Options{ColoredTimeSteps: true})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn []byte(gv.String()), nil\n}", "func hierarchy(t Tags) string {\n\t// Left pad tracknumber to two digits\n\ttrack := fmt.Sprintf(\"%02s\", t[\"tracknumber\"])\n\tartist := t[\"artist\"]\n\ts := path.Join(artist[0:1],\n\t\tartist,\n\t\tt[\"date\"] + \"--\" + t[\"album\"],\n\t\ttrack + \"-\" + t[\"title\"])\n\tlog.Printf(\"Group path for %q is %s\\n\", t, s)\n\treturn s\n}", "func PrintObjectTree(htype *HType) string {\n\tif htype.children == nil {\n\t\treturn fmt.Sprintf(\"%d: %s\", htype.checkValue, htype.name)\n\t}\n\treturn fmt.Sprintf(\"%d: %s(%s)\", htype.checkValue, htype.name, strings.Join(forall(htype.children, PrintObjectTree), \", \"))\n}", "func (a Level) MarshalText() ([]byte, error) {\n\tswitch a {\n\tcase LevelNone:\n\t\treturn []byte(\"\"), nil\n\tcase LevelAnon:\n\t\treturn []byte(\"anon\"), nil\n\tcase LevelAuth:\n\t\treturn []byte(\"auth\"), nil\n\tcase LevelRoot:\n\t\treturn []byte(\"root\"), nil\n\tdefault:\n\t\treturn nil, errors.New(\"auth.Level: invalid level value\")\n\t}\n}", "func (m *binaryMaxentClassifier) marshal(path string) error {\n\tfolder := filepath.Join(path, \"Maxent\")\n\terr := os.Mkdir(folder, os.ModePerm)\n\tfor i, entry := range []string{\"labels\", \"mapping\", \"weights\"} {\n\t\tcomponent, _ := os.Create(filepath.Join(folder, entry+\".gob\"))\n\t\tencoder := gob.NewEncoder(component)\n\t\tif i == 0 {\n\t\t\tcheckError(encoder.Encode(m.labels))\n\t\t} else if i == 1 {\n\t\t\tcheckError(encoder.Encode(m.mapping))\n\t\t} else {\n\t\t\tcheckError(encoder.Encode(m.weights))\n\t\t}\n\t}\n\treturn err\n}", "func (path DerivationPath) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(path.String())\n}", "func (p *HJSON) Marshal(o map[string]interface{}) ([]byte, error) {\n\treturn hjson.Marshal(o)\n}", "func (s DescribeUserHierarchyGroupInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s *RecursiveSchema) MarshalJSON() ([]byte, error) {\n\treturn []byte(fmt.Sprintf(`\"%s\"`, s.Actual.GetName())), nil\n}", "func (l Level) MarshalJSON() ([]byte, error) {\n\treturn []byte(`\"` + levels[l] + `\"`), nil\n}", "func (o OpenShiftCluster) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"id\", o.ID)\n\tpopulate(objectMap, \"location\", o.Location)\n\tpopulate(objectMap, \"name\", o.Name)\n\tpopulate(objectMap, \"properties\", o.Properties)\n\tpopulate(objectMap, \"systemData\", o.SystemData)\n\tpopulate(objectMap, \"tags\", o.Tags)\n\tpopulate(objectMap, \"type\", o.Type)\n\treturn json.Marshal(objectMap)\n}", "func (node *Node) Pack() []byte {\n\tbfr := new(bytes.Buffer)\n\t// Write CATID\n\tbinary.Write(bfr, BIGEND, LBTYPE_CATID)\n\tbinary.Write(bfr, BIGEND, node.Id())\n\t// Write Catalog string key\n\tbinary.Write(bfr, BIGEND, LBTYPE_CATKEY)\n\tksz := AsLBUINT(len(node.Name()))\n\tbinary.Write(bfr, BIGEND, ksz)\n\tbfr.Write([]byte(node.Name()))\n\t// Write field map\n\tif node.HasFields() > 0 {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_MAP)\n\t\tbyts := node.FieldMap.ToBytes(node.debug)\n\t\tbinary.Write(bfr, BIGEND, AsLBUINT(len(byts)))\n\t\tbfr.Write(byts)\n\t} else {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_NIL)\n\t}\n\t// Write parents CATID set\n\tif node.HasParents() > 0 {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_CATID_SET)\n\t\tbyts := node.parents.ToBytes(node.debug)\n\t\tbinary.Write(bfr, BIGEND, AsLBUINT(len(byts)))\n\t\tbfr.Write(byts)\n\t} else {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_NIL)\n\t}\n\treturn bfr.Bytes()\n}", "func (t *smallFlatTable) Marshal(ctx MarshalContext, enc *marshal.Encoder) {\n\tt.initSubTables(ctx.ctx)\n\tenc.PutRawBytes(smallFlatMagic[:])\n\tenc.PutHash(t.hash)\n\tenc.PutGOB(&t.ast)\n\tenc.PutVarint(int64(len(t.subTables)))\n\tsubBufs := make([]marshal.Encoder, len(t.subTables))\n\tfor i := range t.subTables {\n\t\tt.subTables[i].Marshal(ctx, &subBufs[i])\n\t}\n\tfor i := range t.subTables {\n\t\tenc.PutRawBytes(subBufs[i].Bytes())\n\t}\n}", "func main() {\n\troot := TreeNode{\n\t\tVal: 1,\n\t\tLeft: &TreeNode{\n\t\t\tVal: 2,\n\t\t\tLeft: nil,\n\t\t\tRight: nil,\n\t\t},\n\t\tRight: &TreeNode{\n\t\t\tVal: 3,\n\t\t\tLeft: &TreeNode{\n\t\t\t\tVal: 4,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t\tRight: &TreeNode{\n\t\t\t\tVal: 5,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t},\n\t}\n\tobj := Constructor()\n\tdata := obj.serialize(&root)\n\tfmt.Println(data)\n}", "func Marshal(c *configs.Config, schemas *terraform.Schemas) ([]byte, error) {\n\tvar output config\n\n\tpcs := make(map[string]providerConfig)\n\tmarshalProviderConfigs(c, schemas, pcs)\n\n\trootModule, err := marshalModule(c, schemas, \"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\toutput.RootModule = rootModule\n\n\tnormalizeModuleProviderKeys(&rootModule, pcs)\n\n\tfor name, pc := range pcs {\n\t\tif pc.parentKey != \"\" {\n\t\t\tdelete(pcs, name)\n\t\t}\n\t}\n\toutput.ProviderConfigs = pcs\n\n\tret, err := json.Marshal(output)\n\treturn ret, err\n}", "func (h Hostname) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(string(h))\n}", "func (s DescribeUserHierarchyStructureInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeUserHierarchyStructureInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (f *FileChunk) Marshal() string {\n\treturn strings.Join([]string{\n\t\tf.name,\n\t\tf.uuid,\n\t\tfmt.Sprintf(\"%08X\", f.part),\n\t\tfmt.Sprintf(\"%08X\", f.total),\n\t\tf.base64Chunk,\n\t}, \":\")\n}", "func (h HubRouteTable) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"etag\", h.Etag)\n\tpopulate(objectMap, \"id\", h.ID)\n\tpopulate(objectMap, \"name\", h.Name)\n\tpopulate(objectMap, \"properties\", h.Properties)\n\tpopulate(objectMap, \"type\", h.Type)\n\treturn json.Marshal(objectMap)\n}", "func (m *MerkleTree) String() string {\n\ts := \"\"\n\tfor _, l := range m.Leafs {\n\t\ts += fmt.Sprint(l)\n\t\ts += \"\\n\"\n\t}\n\treturn s\n}", "func (r *Key) marshal(c *Client) ([]byte, error) {\n\tm, err := expandKey(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling Key: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func (p PropagatedRouteTable) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"ids\", p.IDs)\n\tpopulate(objectMap, \"labels\", p.Labels)\n\treturn json.Marshal(objectMap)\n}", "func (npi NamedPartitionInformation) MarshalJSON() ([]byte, error) {\n\tnpi.ServicePartitionKind = ServicePartitionKindNamed1\n\tobjectMap := make(map[string]interface{})\n\tif npi.Name != nil {\n\t\tobjectMap[\"Name\"] = npi.Name\n\t}\n\tif npi.ID != nil {\n\t\tobjectMap[\"Id\"] = npi.ID\n\t}\n\tif npi.ServicePartitionKind != \"\" {\n\t\tobjectMap[\"ServicePartitionKind\"] = npi.ServicePartitionKind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (h HubRouteTableProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"associatedConnections\", h.AssociatedConnections)\n\tpopulate(objectMap, \"labels\", h.Labels)\n\tpopulate(objectMap, \"propagatingConnections\", h.PropagatingConnections)\n\tpopulate(objectMap, \"provisioningState\", h.ProvisioningState)\n\tpopulate(objectMap, \"routes\", h.Routes)\n\treturn json.Marshal(objectMap)\n}", "func (sg *StorageGroup) Marshal(b ...[]byte) ([]byte, error) {\n\tvar buf []byte\n\tif len(b) > 0 {\n\t\tbuf = b[0]\n\t}\n\n\treturn (*storagegroup.StorageGroup)(sg).\n\t\tStableMarshal(buf)\n}", "func (t *ASCIITree) PrintTree(w io.Writer) {\n\tancestorPrefix := \"\"\n\tfor _, parent := range t.Ancestors() {\n\t\tif parent.Level() <= 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif parent.Last() {\n\t\t\tancestorPrefix += \" \"\n\t\t} else {\n\t\t\tancestorPrefix += \" │\"\n\t\t}\n\t}\n\n\tmyPrefix := \"\"\n\tmultilinePrefix := \"\"\n\tif t.Level() > 0 {\n\t\tif t.Last() {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" └── \"\n\t\t\t\tmultilinePrefix += \" \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" └─┬ \"\n\t\t\t\tmultilinePrefix += \" └─┬ \"\n\t\t\t}\n\t\t} else {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" ├── \"\n\t\t\t\tmultilinePrefix += \" │ \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" ├─┬ \"\n\t\t\t\tmultilinePrefix += \" │ │ \"\n\t\t\t}\n\t\t}\n\t}\n\n\tif t.Text != \"\" {\n\t\tlines := strings.Split(t.Text, \"\\n\")\n\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, myPrefix, lines[0])\n\t\tfor _, line := range lines[1:] {\n\t\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, multilinePrefix, line)\n\t\t}\n\t}\n\n\tfor _, child := range t.children {\n\t\tchild.PrintTree(w)\n\t}\n}", "func (uid MyULID) Marshal() ([]byte, error) {\n\tif len(uid) < 16 {\n\t\treturn []byte(ZeroULID), nil\n\t}\n\treturn []byte(uid)[:16], nil\n}", "func (f *Forest) ToString() string {\n\n\tfh := f.height\n\t// tree height should be 6 or less\n\tif fh > 6 {\n\t\treturn \"forest too big to print \"\n\t}\n\n\toutput := make([]string, (fh*2)+1)\n\tvar pos uint8\n\tfor h := uint8(0); h <= fh; h++ {\n\t\trowlen := uint8(1 << (fh - h))\n\n\t\tfor j := uint8(0); j < rowlen; j++ {\n\t\t\tvar valstring string\n\t\t\tok := f.data.size() >= uint64(pos)\n\t\t\tif ok {\n\t\t\t\tval := f.data.read(uint64(pos))\n\t\t\t\tif val != empty {\n\t\t\t\t\tvalstring = fmt.Sprintf(\"%x\", val[:2])\n\t\t\t\t}\n\t\t\t}\n\t\t\tif valstring != \"\" {\n\t\t\t\toutput[h*2] += fmt.Sprintf(\"%02d:%s \", pos, valstring)\n\t\t\t} else {\n\t\t\t\toutput[h*2] += fmt.Sprintf(\" \")\n\t\t\t}\n\t\t\tif h > 0 {\n\t\t\t\t//\t\t\t\tif x%2 == 0 {\n\t\t\t\toutput[(h*2)-1] += \"|-------\"\n\t\t\t\tfor q := uint8(0); q < ((1<<h)-1)/2; q++ {\n\t\t\t\t\toutput[(h*2)-1] += \"--------\"\n\t\t\t\t}\n\t\t\t\toutput[(h*2)-1] += \"\\\\ \"\n\t\t\t\tfor q := uint8(0); q < ((1<<h)-1)/2; q++ {\n\t\t\t\t\toutput[(h*2)-1] += \" \"\n\t\t\t\t}\n\n\t\t\t\t//\t\t\t\t}\n\n\t\t\t\tfor q := uint8(0); q < (1<<h)-1; q++ {\n\t\t\t\t\toutput[h*2] += \" \"\n\t\t\t\t}\n\n\t\t\t}\n\t\t\tpos++\n\t\t}\n\n\t}\n\tvar s string\n\tfor z := len(output) - 1; z >= 0; z-- {\n\t\ts += output[z] + \"\\n\"\n\t}\n\treturn s\n}", "func MarshalChaos(s interface{}) (string, error) {\n\tb, err := json.MarshalIndent(s, \"\", \" \")\n\tif err != nil {\n\t\treturn \"\", errors.Wrapf(err, \"failed to marshal indent\")\n\t}\n\treturn string(b), nil\n}", "func (h HubRoute) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"destinationType\", h.DestinationType)\n\tpopulate(objectMap, \"destinations\", h.Destinations)\n\tpopulate(objectMap, \"name\", h.Name)\n\tpopulate(objectMap, \"nextHop\", h.NextHop)\n\tpopulate(objectMap, \"nextHopType\", h.NextHopType)\n\treturn json.Marshal(objectMap)\n}", "func (m HyperflexNode) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 2)\n\n\taO0, err := swag.WriteJSON(m.MoBaseMo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\tvar dataAO1 struct {\n\t\tBuildNumber string `json:\"BuildNumber,omitempty\"`\n\n\t\tCluster *HyperflexClusterRef `json:\"Cluster,omitempty\"`\n\n\t\tClusterMember *AssetClusterMemberRef `json:\"ClusterMember,omitempty\"`\n\n\t\tDisplayVersion string `json:\"DisplayVersion,omitempty\"`\n\n\t\tHostName string `json:\"HostName,omitempty\"`\n\n\t\tHypervisor string `json:\"Hypervisor,omitempty\"`\n\n\t\tIdentity *HyperflexHxUuIDDt `json:\"Identity,omitempty\"`\n\n\t\tIP *HyperflexHxNetworkAddressDt `json:\"Ip,omitempty\"`\n\n\t\tLockdown *bool `json:\"Lockdown,omitempty\"`\n\n\t\tModelNumber string `json:\"ModelNumber,omitempty\"`\n\n\t\tPhysicalServer *ComputePhysicalRef `json:\"PhysicalServer,omitempty\"`\n\n\t\tRole string `json:\"Role,omitempty\"`\n\n\t\tSerialNumber string `json:\"SerialNumber,omitempty\"`\n\n\t\tStatus string `json:\"Status,omitempty\"`\n\n\t\tVersion string `json:\"Version,omitempty\"`\n\t}\n\n\tdataAO1.BuildNumber = m.BuildNumber\n\n\tdataAO1.Cluster = m.Cluster\n\n\tdataAO1.ClusterMember = m.ClusterMember\n\n\tdataAO1.DisplayVersion = m.DisplayVersion\n\n\tdataAO1.HostName = m.HostName\n\n\tdataAO1.Hypervisor = m.Hypervisor\n\n\tdataAO1.Identity = m.Identity\n\n\tdataAO1.IP = m.IP\n\n\tdataAO1.Lockdown = m.Lockdown\n\n\tdataAO1.ModelNumber = m.ModelNumber\n\n\tdataAO1.PhysicalServer = m.PhysicalServer\n\n\tdataAO1.Role = m.Role\n\n\tdataAO1.SerialNumber = m.SerialNumber\n\n\tdataAO1.Status = m.Status\n\n\tdataAO1.Version = m.Version\n\n\tjsonDataAO1, errAO1 := swag.WriteJSON(dataAO1)\n\tif errAO1 != nil {\n\t\treturn nil, errAO1\n\t}\n\t_parts = append(_parts, jsonDataAO1)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (r *AWSEMRCluster) MarshalJSON() ([]byte, error) {\n\ttype Properties AWSEMRCluster\n\treturn json.Marshal(&struct {\n\t\tType string\n\t\tProperties Properties\n\t}{\n\t\tType: r.AWSCloudFormationType(),\n\t\tProperties: (Properties)(*r),\n\t})\n}", "func (r *HealthCheck) marshal(c *Client) ([]byte, error) {\n\tm, err := expandHealthCheck(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling HealthCheck: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func (t ThreeTierFullResourceNames) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"applicationServer\", t.ApplicationServer)\n\tpopulate(objectMap, \"centralServer\", t.CentralServer)\n\tpopulate(objectMap, \"databaseServer\", t.DatabaseServer)\n\tobjectMap[\"namingPatternType\"] = NamingPatternTypeFullResourceName\n\tpopulate(objectMap, \"sharedStorage\", t.SharedStorage)\n\treturn json.Marshal(objectMap)\n}", "func (pm Msg) pack() []byte {\n\tout := new(bytes.Buffer)\n\n\tfmt.Fprintf(out, \"%s\\n%s\\n%s\\n%s\\n%s\\n\", pm.Src, pm.Dst, pm.Wdir, pm.Type, pm.Attr.pack())\n\tln := len(pm.Data)\n\tfmt.Fprintf(out, \"%d\\n%s\", ln, pm.Data[:ln])\n\n\treturn out.Bytes()\n}", "func (d *DOM) Marshal() (string, error) {\n\tb, err := json.Marshal(d)\n\treturn string(b), err\n}", "func (s SapLandscapeMonitorPropertiesGrouping) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"landscape\", s.Landscape)\n\tpopulate(objectMap, \"sapApplication\", s.SapApplication)\n\treturn json.Marshal(objectMap)\n}", "func stringify(n *BinarySearchNode, level int, builder *strings.Builder) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level, builder)\n\t\tbuilder.WriteString(fmt.Sprintf(format+\"%d\\n\", n.value))\n\t\tstringify(n.right, level, builder)\n\t}\n}", "func (phe PartitionHealthEvaluation) MarshalJSON() ([]byte, error) {\n\tphe.Kind = KindPartition\n\tobjectMap := make(map[string]interface{})\n\tif phe.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = phe.PartitionID\n\t}\n\tif phe.UnhealthyEvaluations != nil {\n\t\tobjectMap[\"UnhealthyEvaluations\"] = phe.UnhealthyEvaluations\n\t}\n\tif phe.AggregatedHealthState != \"\" {\n\t\tobjectMap[\"AggregatedHealthState\"] = phe.AggregatedHealthState\n\t}\n\tif phe.Description != nil {\n\t\tobjectMap[\"Description\"] = phe.Description\n\t}\n\tif phe.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = phe.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (h Hash) MarshalText() ([]byte, error) {\n\treturn []byte(h.String()), nil\n}", "func (a Level) MarshalJSON() ([]byte, error) {\n\tres, err := a.MarshalText()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn append(append([]byte{'\"'}, res...), '\"'), nil\n}", "func (o *Organism) MarshalBinary() ([]byte, error) {\n\tvar buf bytes.Buffer\n\t_, err := fmt.Fprintln(&buf, o.Fitness, o.Generation, o.highestFitness, o.isPopulationChampionChild, o.Genotype.Id)\n\to.Genotype.Write(&buf)\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\treturn buf.Bytes(), nil\n\t}\n}", "func (h Horse) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"isAShowHorse\", h.IsAShowHorse)\n\tpopulate(objectMap, \"name\", h.Name)\n\treturn json.Marshal(objectMap)\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\treturn dfsSerial(root, \"\")\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tstrs := make([]string, 0)\n\tvar preorder func(*TreeNode)\n\tpreorder = func(node *TreeNode) {\n\t\tif node == nil {\n\t\t\tstrs = append(strs, this.null)\n\t\t\treturn\n\t\t}\n\n\t\tstrs = append(strs, strconv.Itoa(node.Val))\n\t\tpreorder(node.Left)\n\t\tpreorder(node.Right)\n\t}\n\tpreorder(root)\n\treturn strings.Join(strs, this.sep)\n}", "func (l *List) MarshalText() (text []byte, err error) {\n\theading := strings.TrimSpace(l.rootItem.desc)\n\ts := heading + \"\\n\"\n\ts += strings.Repeat(\"=\", len(heading)) + \"\\n\"\n\tfor _, li := range l.rootItem.items {\n\t\ts = recursiveMarshal(s, li, 0)\n\t}\n\treturn []byte(s), nil\n}", "func (cfg frozenConfig) MarshalToString(val interface{}) (string, error) {\n buf, err := encoder.Encode(val, cfg.encoderOpts)\n return rt.Mem2Str(buf), err\n}", "func (m SdcardPartition) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 1)\n\n\tvar dataAO0 struct {\n\t\tType *string `json:\"Type,omitempty\"`\n\n\t\tVirtualDrives []*SdcardVirtualDrive `json:\"VirtualDrives\"`\n\t}\n\n\tdataAO0.Type = m.Type\n\n\tdataAO0.VirtualDrives = m.VirtualDrives\n\n\tjsonDataAO0, errAO0 := swag.WriteJSON(dataAO0)\n\tif errAO0 != nil {\n\t\treturn nil, errAO0\n\t}\n\t_parts = append(_parts, jsonDataAO0)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n if root == nil {\n return \"x\"\n }\n return strconv.Itoa(root.Val) + \",\" + this.serialize(root.Left)+ \",\" + this.serialize(root.Right)\n}", "func (s DescribeUserHierarchyStructureInput) MarshalFields(e protocol.FieldEncoder) error {\n\te.SetValue(protocol.HeaderTarget, \"Content-Type\", protocol.StringValue(\"application/json\"), protocol.Metadata{})\n\n\tif s.InstanceId != nil {\n\t\tv := *s.InstanceId\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.PathTarget, \"InstanceId\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\treturn nil\n}", "func (ll *LogLevel) MarshalText() ([]byte, error) {\n\treturn []byte(ll.String()), nil\n}", "func (h H) MarshalXML(e *xml.Encoder, start xml.StartElement) error {\n\tstart.Name = xml.Name{\n\t\tSpace: \"\",\n\t\tLocal: \"map\",\n\t}\n\tif err := e.EncodeToken(start); err != nil {\n\t\treturn err\n\t}\n\tfor key, value := range h {\n\t\telem := xml.StartElement{\n\t\t\tName: xml.Name{Space: \"\", Local: key},\n\t\t\tAttr: []xml.Attr{},\n\t\t}\n\t\tif err := e.EncodeElement(value, elem); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif err := e.EncodeToken(xml.EndElement{Name: start.Name}); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (d Dir) MarshalJSON() ([]byte, error) {\n\tbuffer := bytes.NewBufferString(`\"`)\n\tbuffer.WriteString(toString[d])\n\tbuffer.WriteString(`\"`)\n\treturn buffer.Bytes(), nil\n}", "func (host Host) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(host.String())\n}", "func (uh UnlockHash) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(uh.String())\n}", "func (npsd NamedPartitionSchemeDescription) MarshalJSON() ([]byte, error) {\n\tnpsd.PartitionScheme = PartitionSchemeNamed1\n\tobjectMap := make(map[string]interface{})\n\tif npsd.Count != nil {\n\t\tobjectMap[\"Count\"] = npsd.Count\n\t}\n\tif npsd.Names != nil {\n\t\tobjectMap[\"Names\"] = npsd.Names\n\t}\n\tif npsd.PartitionScheme != \"\" {\n\t\tobjectMap[\"PartitionScheme\"] = npsd.PartitionScheme\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (h Hostname) MarshalText() ([]byte, error) {\n\treturn []byte(string(h)), nil\n}", "func (m LogicalSwitch) MarshalJSON() ([]byte, error) {\n\tvar _parts [][]byte\n\n\taO0, err := swag.WriteJSON(m.ManagedResource)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\tvar data struct {\n\t\tAdminState *string `json:\"admin_state\"`\n\n\t\tIPPoolID string `json:\"ip_pool_id,omitempty\"`\n\n\t\tReplicationMode string `json:\"replication_mode,omitempty\"`\n\n\t\tTransportZoneID *string `json:\"transport_zone_id\"`\n\n\t\tVlan int64 `json:\"vlan,omitempty\"`\n\n\t\tVni int32 `json:\"vni,omitempty\"`\n\t}\n\n\tdata.AdminState = m.AdminState\n\n\tdata.IPPoolID = m.IPPoolID\n\n\tdata.ReplicationMode = m.ReplicationMode\n\n\tdata.TransportZoneID = m.TransportZoneID\n\n\tdata.Vlan = m.Vlan\n\n\tdata.Vni = m.Vni\n\n\tjsonData, err := swag.WriteJSON(data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, jsonData)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (s SystemTopic) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"id\", s.ID)\n\tpopulate(objectMap, \"identity\", s.Identity)\n\tpopulate(objectMap, \"location\", s.Location)\n\tpopulate(objectMap, \"name\", s.Name)\n\tpopulate(objectMap, \"properties\", s.Properties)\n\tpopulate(objectMap, \"systemData\", s.SystemData)\n\tpopulate(objectMap, \"tags\", s.Tags)\n\tpopulate(objectMap, \"type\", s.Type)\n\treturn json.Marshal(objectMap)\n}", "func (e Level) MarshalJSON() ([]byte, error) {\n\treturn []byte(fmt.Sprintf(`\"%v\"`, e)), nil\n}", "func (this *school) MarshalBinary() (data []byte, err error) {\n\t// var addrs = map[uintptr]uint64{}\n\tbuf := bytes.NewBuffer(make([]byte, 0, ((8 + (len(this.Teachers) * 507)) + (8 + (len(this.Pupils) * 507)))))\n\tif err = this.marshalTo(buf /*, addrs*/); err == nil {\n\t\tdata = buf.Bytes()\n\t}\n\treturn\n}", "func (this *NamespacedName) MarshalJSON() ([]byte, error) {\n\tstr, err := CommonMarshaler.MarshalToString(this)\n\treturn []byte(str), err\n}", "func (l LTIContextType) MarshalJSON() ([]byte, error) {\n\tvar ctxs []string\n\tvar ctxsStr string\n\n\tfor _, ctx := range l {\n\t\tctxs = append(ctxs, string(ctx))\n\t}\n\n\tctxsStr = strings.Join(ctxs, \",\")\n\tctxsStr = fmt.Sprintf(\"\\\"%s\\\"\", ctxsStr)\n\n\treturn []byte(ctxsStr), nil\n}", "func (epqsc EnsurePartitionQurumSafetyCheck) MarshalJSON() ([]byte, error) {\n\tepqsc.Kind = KindEnsurePartitionQuorum\n\tobjectMap := make(map[string]interface{})\n\tif epqsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = epqsc.PartitionID\n\t}\n\tif epqsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = epqsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func stringTree(builder *strings.Builder, depth int, node *TreeNode) (*strings.Builder, error) {\n\tvar err error\n\tfor i := 0; i < depth; i++ {\n\t\tif _, err = builder.WriteString(\" \"); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\tfmt.Fprintf(builder, \"└ %d\\n\", node.Val)\n\n\tif node.Left != nil {\n\t\tif builder, err = stringTree(builder, depth+1, node.Left); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\tif node.Right != nil {\n\t\tif builder, err = stringTree(builder, depth+1, node.Right); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\treturn builder, nil\n}", "func (r *Chapter) Marshal() ([]byte, error) {\n\treturn json.Marshal(r)\n}" ]
[ "0.71129423", "0.6860378", "0.5548206", "0.52124643", "0.51653415", "0.50677866", "0.4993746", "0.4927463", "0.49178874", "0.4913059", "0.48793265", "0.4859133", "0.48533386", "0.48482224", "0.48187178", "0.47770202", "0.4766247", "0.4766247", "0.4762357", "0.4755859", "0.475494", "0.47514337", "0.47279862", "0.47221738", "0.47221738", "0.4713375", "0.46628076", "0.46541762", "0.46465", "0.46465", "0.4643268", "0.46401712", "0.46364084", "0.46184868", "0.45753488", "0.45746037", "0.45392266", "0.45363152", "0.45340925", "0.45237577", "0.45207295", "0.45108944", "0.45103425", "0.45061898", "0.4490498", "0.44899294", "0.44618824", "0.44539604", "0.44449186", "0.4444734", "0.4444734", "0.44378555", "0.44356683", "0.44247687", "0.44246858", "0.4416261", "0.4407098", "0.43912312", "0.43875077", "0.43852565", "0.43818682", "0.43768424", "0.43767315", "0.4368136", "0.4366923", "0.43609485", "0.436092", "0.433976", "0.43381047", "0.4337389", "0.4329583", "0.43249846", "0.43246448", "0.43243548", "0.43189755", "0.43172085", "0.43150303", "0.4315013", "0.43120757", "0.43068147", "0.4303862", "0.4303151", "0.43006378", "0.43006375", "0.4299089", "0.4292772", "0.4281087", "0.42804724", "0.4274975", "0.42728513", "0.42691237", "0.42661998", "0.42650378", "0.4259356", "0.42558745", "0.42549577", "0.4252978", "0.42521575", "0.42475724", "0.42468274" ]
0.72192514
0
MarshalHierarchical returns hierarchical string representation.
func (status UpstreamStatus) MarshalHierarchical() string { hierarchy := hierr.Push( "upstream", "total: "+strconv.Itoa(status.Total), fmt.Sprintf( "success: %d (%.2f%%)", status.Success, status.SuccessPercent, ), fmt.Sprintf( "error: %d (%.2f%%)", status.Error, status.ErrorPercent, ), ) if len(status.Slaves) > 0 { slaves := errors.New("slaves") for _, slave := range status.Slaves { slaves = hierr.Push(slaves, slave.MarshalHierarchical()) } hierarchy = hierr.Push(hierarchy, slaves) } return hierr.String(hierarchy) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (status MirrorStatus) MarshalHierarchical() string {\n\thierarchy := hierr.Push(\n\t\tstatus.Name,\n\t\tfmt.Sprintf(\"state: %s\", status.State),\n\t)\n\n\tif status.ModifyDate > 0 {\n\t\thierarchy = hierr.Push(\n\t\t\thierarchy,\n\t\t\tfmt.Sprintf(\"modify date: %v\", status.ModifyDate),\n\t\t)\n\t}\n\n\treturn hierr.String(hierarchy)\n}", "func (status ServerStatus) MarshalHierarchical() []byte {\n\tvar hierarchy error\n\tif status.Address != \"\" {\n\t\thierarchy = hierr.Push(status.Address)\n\t} else {\n\t\thierarchy = hierr.Push(\"status\")\n\t}\n\n\tif status.Role != \"master\" {\n\t\tstatus.Role = \"slave\"\n\t}\n\n\thierarchy = hierr.Push(\n\t\thierarchy,\n\t\tfmt.Sprintf(\"role: %s\", status.Role),\n\t)\n\n\thierarchy = hierr.Push(\n\t\thierarchy,\n\t\tfmt.Sprintf(\"total: %d\", len(status.Mirrors)),\n\t)\n\n\tif status.HierarchicalError != \"\" {\n\t\thierarchy = hierr.Push(\n\t\t\thierarchy,\n\t\t\thierr.Push(\"error\", status.HierarchicalError),\n\t\t)\n\t}\n\n\tif len(status.Mirrors) > 0 {\n\t\tmirrors := errors.New(\"mirrors\")\n\t\tfor _, mirror := range status.Mirrors {\n\t\t\tmirrors = hierr.Push(mirrors, mirror.MarshalHierarchical())\n\t\t}\n\t\thierarchy = hierr.Push(hierarchy, mirrors)\n\t}\n\n\tif status.Role == \"master\" {\n\t\thierarchy = hierr.Push(hierarchy, status.Upstream.MarshalHierarchical())\n\t}\n\n\treturn []byte(hierr.String(hierarchy))\n}", "func (rht *RHT) Marshal() string {\n\tmembers := rht.Members()\n\n\tsize := len(members)\n\tkeys := make([]string, 0, size)\n\tfor k := range members {\n\t\tkeys = append(keys, k)\n\t}\n\tsort.Strings(keys)\n\n\tsb := strings.Builder{}\n\tsb.WriteString(\"{\")\n\n\tidx := 0\n\tfor _, k := range keys {\n\t\tvalue := members[k]\n\t\tsb.WriteString(fmt.Sprintf(\"\\\"%s\\\":%s\", k, value.Marshal()))\n\t\tif size-1 != idx {\n\t\t\tsb.WriteString(\",\")\n\t\t}\n\t\tidx++\n\t}\n\tsb.WriteString(\"}\")\n\n\treturn sb.String()\n}", "func (s HierarchyLevel) String() string {\n\treturn awsutil.Prettify(s)\n}", "func encodeTree(hmt *Tree, finalTree *string) {\n\tif hmt == nil {\n\t\treturn\n\t}\n\t\n\tif hmt.LeftNode == nil && hmt.RightNode == nil{\n\t\t*finalTree += \"1\" + string(hmt.Char)\n\t} else {\n\t\t*finalTree += \"0\"\n\t}\n\tencodeTree(hmt.LeftNode, finalTree)\n\tencodeTree(hmt.RightNode, finalTree) \n}", "func (s HierarchyStructure) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (t *Tree) formatStr() []byte {\n\tvar s []byte\n\ts = append(s, t.format...)\n\tif t.parent != nil {\n\t\tt := t.parent.formatStr()\n\t\tt = append(t, s...)\n\t\ts = t\n\t}\n\treturn s\n}", "func (m QueryHierarchyItem) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 1)\n\n\taO0, err := swag.WriteJSON(m.WorkItemTrackingResource)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\t// now for regular properties\n\tvar propsQueryHierarchyItem struct {\n\t\tChildren []*QueryHierarchyItem `json:\"children\"`\n\n\t\tClauses *WorkItemQueryClause `json:\"clauses,omitempty\"`\n\n\t\tColumns []*WorkItemFieldReference `json:\"columns\"`\n\n\t\tCreatedBy *IdentityReference `json:\"createdBy,omitempty\"`\n\n\t\tCreatedDate strfmt.DateTime `json:\"createdDate,omitempty\"`\n\n\t\tFilterOptions interface{} `json:\"filterOptions,omitempty\"`\n\n\t\tHasChildren bool `json:\"hasChildren,omitempty\"`\n\n\t\tID strfmt.UUID `json:\"id,omitempty\"`\n\n\t\tIsDeleted bool `json:\"isDeleted,omitempty\"`\n\n\t\tIsFolder bool `json:\"isFolder,omitempty\"`\n\n\t\tIsInvalidSyntax bool `json:\"isInvalidSyntax,omitempty\"`\n\n\t\tIsPublic bool `json:\"isPublic,omitempty\"`\n\n\t\tLastExecutedBy *IdentityReference `json:\"lastExecutedBy,omitempty\"`\n\n\t\tLastExecutedDate strfmt.DateTime `json:\"lastExecutedDate,omitempty\"`\n\n\t\tLastModifiedBy *IdentityReference `json:\"lastModifiedBy,omitempty\"`\n\n\t\tLastModifiedDate strfmt.DateTime `json:\"lastModifiedDate,omitempty\"`\n\n\t\tLinkClauses *WorkItemQueryClause `json:\"linkClauses,omitempty\"`\n\n\t\tName string `json:\"name,omitempty\"`\n\n\t\tPath string `json:\"path,omitempty\"`\n\n\t\tQueryRecursionOption interface{} `json:\"queryRecursionOption,omitempty\"`\n\n\t\tQueryType interface{} `json:\"queryType,omitempty\"`\n\n\t\tSortColumns []*WorkItemQuerySortColumn `json:\"sortColumns\"`\n\n\t\tSourceClauses *WorkItemQueryClause `json:\"sourceClauses,omitempty\"`\n\n\t\tTargetClauses *WorkItemQueryClause `json:\"targetClauses,omitempty\"`\n\n\t\tWiql string `json:\"wiql,omitempty\"`\n\t}\n\tpropsQueryHierarchyItem.Children = m.Children\n\n\tpropsQueryHierarchyItem.Clauses = m.Clauses\n\n\tpropsQueryHierarchyItem.Columns = m.Columns\n\n\tpropsQueryHierarchyItem.CreatedBy = m.CreatedBy\n\n\tpropsQueryHierarchyItem.CreatedDate = m.CreatedDate\n\n\tpropsQueryHierarchyItem.FilterOptions = m.FilterOptions\n\n\tpropsQueryHierarchyItem.HasChildren = m.HasChildren\n\n\tpropsQueryHierarchyItem.ID = m.ID\n\n\tpropsQueryHierarchyItem.IsDeleted = m.IsDeleted\n\n\tpropsQueryHierarchyItem.IsFolder = m.IsFolder\n\n\tpropsQueryHierarchyItem.IsInvalidSyntax = m.IsInvalidSyntax\n\n\tpropsQueryHierarchyItem.IsPublic = m.IsPublic\n\n\tpropsQueryHierarchyItem.LastExecutedBy = m.LastExecutedBy\n\n\tpropsQueryHierarchyItem.LastExecutedDate = m.LastExecutedDate\n\n\tpropsQueryHierarchyItem.LastModifiedBy = m.LastModifiedBy\n\n\tpropsQueryHierarchyItem.LastModifiedDate = m.LastModifiedDate\n\n\tpropsQueryHierarchyItem.LinkClauses = m.LinkClauses\n\n\tpropsQueryHierarchyItem.Name = m.Name\n\n\tpropsQueryHierarchyItem.Path = m.Path\n\n\tpropsQueryHierarchyItem.QueryRecursionOption = m.QueryRecursionOption\n\n\tpropsQueryHierarchyItem.QueryType = m.QueryType\n\n\tpropsQueryHierarchyItem.SortColumns = m.SortColumns\n\n\tpropsQueryHierarchyItem.SourceClauses = m.SourceClauses\n\n\tpropsQueryHierarchyItem.TargetClauses = m.TargetClauses\n\n\tpropsQueryHierarchyItem.Wiql = m.Wiql\n\n\tjsonDataPropsQueryHierarchyItem, errQueryHierarchyItem := swag.WriteJSON(propsQueryHierarchyItem)\n\tif errQueryHierarchyItem != nil {\n\t\treturn nil, errQueryHierarchyItem\n\t}\n\t_parts = append(_parts, jsonDataPropsQueryHierarchyItem)\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (s HierarchyGroup) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (rh ResourceHierarchy) String() string {\n\treturn string(rh)\n}", "func (s HierarchyPath) String() string {\n\treturn awsutil.Prettify(s)\n}", "func WriteTree(writer io.Writer, hierarchy *Hierarchy, includeEmpty bool) {\n\ttree := assembleTree(hierarchy)\n\tkeys := make([]string, len(tree))\n\ti := 0\n\tfor k := range tree {\n\t\tkeys[i] = k\n\t\ti++\n\t}\n\tsort.Strings(keys)\n\tfor _, key := range keys {\n\t\tb := tree[key]\n\t\twriteBranch(writer, b, \"\", hierarchy, includeEmpty)\n\t}\n}", "func stringify(n *Node, level int) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level)\n\t\tfmt.Printf(format+\"%d\\n\", n.key)\n\t\tstringify(n.right, level)\n\t}\n}", "func (this *Codec) serialize(root *TreeNode) string {\n var s string \n res:=helpSerialize(root,s)\n fmt.Println(res)\n return res\n}", "func stringify(treeNode *TreeNode, level int) {\n\tif treeNode != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"***> \"\n\t\tlevel++\n\t\tstringify(treeNode.leftNode, level)\n\t\tfmt.Printf(format+\"%d\\n\", treeNode.key)\n\t\tstringify(treeNode.rightNode, level)\n\t}\n}", "func (s CreateUserHierarchyGroupOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (w *RootWalker) String() (out string) {\n\ttabs := func(n int) {\n\t\tfor i := 0; i < n; i++ {\n\t\t\tout += \"\\t\"\n\t\t}\n\t}\n\tout += fmt.Sprint(\"Root\")\n\tsize := w.Size()\n\tif size == 0 {\n\t\treturn\n\t}\n\tout += fmt.Sprintf(\".Refs[%d] ->\\n\", w.stack[0].prevInFieldIndex)\n\tfor i, obj := range w.stack {\n\t\tschName := \"\"\n\t\ts, _ := w.r.SchemaByReference(obj.s)\n\t\tif s != nil {\n\t\t\tschName = s.Name()\n\t\t}\n\n\t\ttabs(i)\n\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\tout += fmt.Sprintf(` = \"%v\"`+\"\\n\", obj.p)\n\n\t\ttabs(i)\n\t\tif obj.next != nil {\n\t\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\t\tout += fmt.Sprintf(\".%s\", obj.next.prevFieldName)\n\t\t\tif obj.next.prevInFieldIndex != -1 {\n\t\t\t\tout += fmt.Sprintf(\"[%d]\", obj.next.prevInFieldIndex)\n\t\t\t}\n\t\t\tout += fmt.Sprint(\" ->\\n\")\n\t\t}\n\t}\n\treturn\n}", "func (w *RootWalker) String() (out string) {\n\ttabs := func(n int) {\n\t\tfor i := 0; i < n; i++ {\n\t\t\tout += \"\\t\"\n\t\t}\n\t}\n\tout += fmt.Sprint(\"Root\")\n\tsize := w.Size()\n\tif size == 0 {\n\t\treturn\n\t}\n\tout += fmt.Sprintf(\".Refs[%d] ->\\n\", w.stack[0].prevInFieldIndex)\n\tfor i, obj := range w.stack {\n\t\tschName := \"\"\n\t\ts, _ := w.r.SchemaByReference(obj.s)\n\t\tif s != nil {\n\t\t\tschName = s.Name()\n\t\t}\n\n\t\ttabs(i)\n\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\tout += fmt.Sprintf(` = \"%v\"`+\"\\n\", obj.p)\n\n\t\ttabs(i)\n\t\tif obj.next != nil {\n\t\t\tout += fmt.Sprintf(\" %s\", schName)\n\t\t\tout += fmt.Sprintf(\".%s\", obj.next.prevFieldName)\n\t\t\tif obj.next.prevInFieldIndex != -1 {\n\t\t\t\tout += fmt.Sprintf(\"[%d]\", obj.next.prevInFieldIndex)\n\t\t\t}\n\t\t\tout += fmt.Sprint(\" ->\\n\")\n\t\t}\n\t}\n\treturn\n}", "func (h Hls) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"fragmentsPerTsSegment\", h.FragmentsPerTsSegment)\n\treturn json.Marshal(objectMap)\n}", "func (s DescribeUserHierarchyStructureOutput) MarshalFields(e protocol.FieldEncoder) error {\n\tif s.HierarchyStructure != nil {\n\t\tv := s.HierarchyStructure\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetFields(protocol.BodyTarget, \"HierarchyStructure\", v, metadata)\n\t}\n\treturn nil\n}", "func (ths *TreeHashStack) Serialize() ([]byte, error) {\n\tbuf := new(bytes.Buffer)\n\n\tif err := gob.NewEncoder(buf).Encode(ths); nil != err {\n\t\treturn nil, err\n\t}\n\n\treturn buf.Bytes(), nil\n}", "func (v dHGroup) MarshalJSON() ([]byte, error) {\n\tw := jwriter.Writer{}\n\teasyjson94b2531bEncodeGitRonaksoftComRiverWebWasmConnection1(&w, v)\n\treturn w.Buffer.BuildBytes(), w.Error\n}", "func (p *page) marshal() string {\n\tb := &strings.Builder{}\n\tfmt.Fprintf(b, \"\\n%d 0 obj\\n<<\\n\", p.id)\n\tfmt.Fprintf(b, \" /MediaBox [ %f %f %f %f ]\\n\", p.mediaBox.llx, p.mediaBox.lly, p.mediaBox.urx, p.mediaBox.ury)\n\tfmt.Fprintf(b, \" /CropBox [ %f %f %f %f ]\\n\", p.cropBox.llx, p.cropBox.lly, p.cropBox.urx, p.cropBox.ury)\n\tfmt.Fprintf(b, \" /BleedBox [ %f %f %f %f ]\\n\", p.bleedBox.llx, p.bleedBox.lly, p.bleedBox.urx, p.bleedBox.ury)\n\tfmt.Fprintf(b, \" /TrimBox [ %f %f %f %f ]\\n\", p.trimBox.llx, p.trimBox.lly, p.trimBox.urx, p.trimBox.ury)\n\tfmt.Fprintf(b, \" /Contents [ \")\n\tfor _, cid := range p.contentIds {\n\t\tfmt.Fprintf(b, \" %d 0 R \", cid)\n\t}\n\tfmt.Fprintf(b, \" ]\\n\")\n\tfmt.Fprintf(b, \" /Parent %d 0 R\\n\", p.parentID)\n\tb.WriteString(p.raw)\n\tfmt.Fprintf(b, \"\\n>>\\nendobj\\n\")\n\treturn b.String()\n}", "func (l *Level) MarshalText() ([]byte, error) {\n\tif l == nil {\n\t\treturn nil, errMarshalNilLevel\n\t}\n\treturn []byte(l.String()), nil\n}", "func (l *Level) MarshalText() ([]byte, error) {\n\tif l == nil {\n\t\treturn nil, errMarshalNilLevel\n\t}\n\treturn []byte(l.String()), nil\n}", "func (h Hash) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(h.String())\n}", "func (st *SlimTrie) Marshal() ([]byte, error) {\n\tvar buf []byte\n\twriter := bytes.NewBuffer(buf)\n\n\t_, err := pbcmpl.Marshal(writer, &versionedArray{&st.Children.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal children\")\n\t}\n\n\t_, err = pbcmpl.Marshal(writer, &versionedArray{&st.Steps.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal steps\")\n\t}\n\n\t_, err = pbcmpl.Marshal(writer, &versionedArray{&st.Leaves.Base})\n\tif err != nil {\n\t\treturn nil, errors.WithMessage(err, \"failed to marshal leaves\")\n\t}\n\n\treturn writer.Bytes(), nil\n}", "func (g *Group) Marshal() []byte {\n\tout := make([]byte, 0, 4*2*32 + 2*2*2*32)\n\tout = append(out, g.g1.Marshal()...)\n\tout = append(out, g.h.Marshal()...)\n\tout = append(out, g.u.Marshal()...)\n\tout = append(out, g.v.Marshal()...)\n\tout = append(out, g.g2.Marshal()...)\n\tout = append(out, g.w.Marshal()...)\n\treturn out\n}", "func (s DescribeUserHierarchyStructureOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeUserHierarchyStructureOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s CreateUserHierarchyGroupInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func ToString(ctx context.Context) (ret string) {\n\tfather, ok := ctx.Value(key).(*constNode)\n\tif !ok {\n\t\treturn \"\"\n\t}\n\tDone(ctx)\n\tconst fmtStr = \"%s%s (%dms %d%%)\\n\"\n\tvar levelPrint func(level int, node *constNode, prefix string)\n\tlevelPrint = func(level int, node *constNode, prefix string) {\n\t\tvar (\n\t\t\tlastTabs string\n\t\t\tnoLastTabs string\n\t\t)\n\t\tnoLastTabs = prefix + \"├─\"\n\t\tlastTabs = prefix + \"└─\"\n\t\tfor i, child := range node.child {\n\t\t\ttabs := noLastTabs\n\t\t\tif i == len(node.child)-1 {\n\t\t\t\ttabs = lastTabs\n\t\t\t}\n\t\t\tchildCostMs := child.cost().Milliseconds()\n\t\t\tfatherCostMs := node.cost().Milliseconds()\n\t\t\tradio := int64(0)\n\t\t\tif fatherCostMs > 0 {\n\t\t\t\tradio = childCostMs * 100 / fatherCostMs\n\t\t\t}\n\t\t\tret += fmt.Sprintf(fmtStr, tabs, child.title, childCostMs, radio)\n\t\t\tif len(child.child) > 0 {\n\t\t\t\tif i == len(node.child)-1 {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\" \")\n\t\t\t\t} else {\n\t\t\t\t\tlevelPrint(level+1, child, prefix+\"│ \")\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\tret += fmt.Sprintf(fmtStr, \"\", father.title, father.cost().Milliseconds(), 100)\n\tlevelPrint(0, father, \"\")\n\treturn\n}", "func (s DescribeUserHierarchyGroupOutput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func Marshal(g *ag.Graph) ([]byte, error) {\n\tgv, err := graphviz.BuildGraph(g, graphviz.Options{ColoredTimeSteps: true})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn []byte(gv.String()), nil\n}", "func hierarchy(t Tags) string {\n\t// Left pad tracknumber to two digits\n\ttrack := fmt.Sprintf(\"%02s\", t[\"tracknumber\"])\n\tartist := t[\"artist\"]\n\ts := path.Join(artist[0:1],\n\t\tartist,\n\t\tt[\"date\"] + \"--\" + t[\"album\"],\n\t\ttrack + \"-\" + t[\"title\"])\n\tlog.Printf(\"Group path for %q is %s\\n\", t, s)\n\treturn s\n}", "func PrintObjectTree(htype *HType) string {\n\tif htype.children == nil {\n\t\treturn fmt.Sprintf(\"%d: %s\", htype.checkValue, htype.name)\n\t}\n\treturn fmt.Sprintf(\"%d: %s(%s)\", htype.checkValue, htype.name, strings.Join(forall(htype.children, PrintObjectTree), \", \"))\n}", "func (a Level) MarshalText() ([]byte, error) {\n\tswitch a {\n\tcase LevelNone:\n\t\treturn []byte(\"\"), nil\n\tcase LevelAnon:\n\t\treturn []byte(\"anon\"), nil\n\tcase LevelAuth:\n\t\treturn []byte(\"auth\"), nil\n\tcase LevelRoot:\n\t\treturn []byte(\"root\"), nil\n\tdefault:\n\t\treturn nil, errors.New(\"auth.Level: invalid level value\")\n\t}\n}", "func (m *binaryMaxentClassifier) marshal(path string) error {\n\tfolder := filepath.Join(path, \"Maxent\")\n\terr := os.Mkdir(folder, os.ModePerm)\n\tfor i, entry := range []string{\"labels\", \"mapping\", \"weights\"} {\n\t\tcomponent, _ := os.Create(filepath.Join(folder, entry+\".gob\"))\n\t\tencoder := gob.NewEncoder(component)\n\t\tif i == 0 {\n\t\t\tcheckError(encoder.Encode(m.labels))\n\t\t} else if i == 1 {\n\t\t\tcheckError(encoder.Encode(m.mapping))\n\t\t} else {\n\t\t\tcheckError(encoder.Encode(m.weights))\n\t\t}\n\t}\n\treturn err\n}", "func (path DerivationPath) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(path.String())\n}", "func (p *HJSON) Marshal(o map[string]interface{}) ([]byte, error) {\n\treturn hjson.Marshal(o)\n}", "func (s DescribeUserHierarchyGroupInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s *RecursiveSchema) MarshalJSON() ([]byte, error) {\n\treturn []byte(fmt.Sprintf(`\"%s\"`, s.Actual.GetName())), nil\n}", "func (l Level) MarshalJSON() ([]byte, error) {\n\treturn []byte(`\"` + levels[l] + `\"`), nil\n}", "func (o OpenShiftCluster) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"id\", o.ID)\n\tpopulate(objectMap, \"location\", o.Location)\n\tpopulate(objectMap, \"name\", o.Name)\n\tpopulate(objectMap, \"properties\", o.Properties)\n\tpopulate(objectMap, \"systemData\", o.SystemData)\n\tpopulate(objectMap, \"tags\", o.Tags)\n\tpopulate(objectMap, \"type\", o.Type)\n\treturn json.Marshal(objectMap)\n}", "func (node *Node) Pack() []byte {\n\tbfr := new(bytes.Buffer)\n\t// Write CATID\n\tbinary.Write(bfr, BIGEND, LBTYPE_CATID)\n\tbinary.Write(bfr, BIGEND, node.Id())\n\t// Write Catalog string key\n\tbinary.Write(bfr, BIGEND, LBTYPE_CATKEY)\n\tksz := AsLBUINT(len(node.Name()))\n\tbinary.Write(bfr, BIGEND, ksz)\n\tbfr.Write([]byte(node.Name()))\n\t// Write field map\n\tif node.HasFields() > 0 {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_MAP)\n\t\tbyts := node.FieldMap.ToBytes(node.debug)\n\t\tbinary.Write(bfr, BIGEND, AsLBUINT(len(byts)))\n\t\tbfr.Write(byts)\n\t} else {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_NIL)\n\t}\n\t// Write parents CATID set\n\tif node.HasParents() > 0 {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_CATID_SET)\n\t\tbyts := node.parents.ToBytes(node.debug)\n\t\tbinary.Write(bfr, BIGEND, AsLBUINT(len(byts)))\n\t\tbfr.Write(byts)\n\t} else {\n\t\tbinary.Write(bfr, BIGEND, LBTYPE_NIL)\n\t}\n\treturn bfr.Bytes()\n}", "func (t *smallFlatTable) Marshal(ctx MarshalContext, enc *marshal.Encoder) {\n\tt.initSubTables(ctx.ctx)\n\tenc.PutRawBytes(smallFlatMagic[:])\n\tenc.PutHash(t.hash)\n\tenc.PutGOB(&t.ast)\n\tenc.PutVarint(int64(len(t.subTables)))\n\tsubBufs := make([]marshal.Encoder, len(t.subTables))\n\tfor i := range t.subTables {\n\t\tt.subTables[i].Marshal(ctx, &subBufs[i])\n\t}\n\tfor i := range t.subTables {\n\t\tenc.PutRawBytes(subBufs[i].Bytes())\n\t}\n}", "func main() {\n\troot := TreeNode{\n\t\tVal: 1,\n\t\tLeft: &TreeNode{\n\t\t\tVal: 2,\n\t\t\tLeft: nil,\n\t\t\tRight: nil,\n\t\t},\n\t\tRight: &TreeNode{\n\t\t\tVal: 3,\n\t\t\tLeft: &TreeNode{\n\t\t\t\tVal: 4,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t\tRight: &TreeNode{\n\t\t\t\tVal: 5,\n\t\t\t\tRight: nil,\n\t\t\t\tLeft: nil,\n\t\t\t},\n\t\t},\n\t}\n\tobj := Constructor()\n\tdata := obj.serialize(&root)\n\tfmt.Println(data)\n}", "func Marshal(c *configs.Config, schemas *terraform.Schemas) ([]byte, error) {\n\tvar output config\n\n\tpcs := make(map[string]providerConfig)\n\tmarshalProviderConfigs(c, schemas, pcs)\n\n\trootModule, err := marshalModule(c, schemas, \"\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\toutput.RootModule = rootModule\n\n\tnormalizeModuleProviderKeys(&rootModule, pcs)\n\n\tfor name, pc := range pcs {\n\t\tif pc.parentKey != \"\" {\n\t\t\tdelete(pcs, name)\n\t\t}\n\t}\n\toutput.ProviderConfigs = pcs\n\n\tret, err := json.Marshal(output)\n\treturn ret, err\n}", "func (h Hostname) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(string(h))\n}", "func (s DescribeUserHierarchyStructureInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (s DescribeUserHierarchyStructureInput) String() string {\n\treturn awsutil.Prettify(s)\n}", "func (f *FileChunk) Marshal() string {\n\treturn strings.Join([]string{\n\t\tf.name,\n\t\tf.uuid,\n\t\tfmt.Sprintf(\"%08X\", f.part),\n\t\tfmt.Sprintf(\"%08X\", f.total),\n\t\tf.base64Chunk,\n\t}, \":\")\n}", "func (h HubRouteTable) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"etag\", h.Etag)\n\tpopulate(objectMap, \"id\", h.ID)\n\tpopulate(objectMap, \"name\", h.Name)\n\tpopulate(objectMap, \"properties\", h.Properties)\n\tpopulate(objectMap, \"type\", h.Type)\n\treturn json.Marshal(objectMap)\n}", "func (m *MerkleTree) String() string {\n\ts := \"\"\n\tfor _, l := range m.Leafs {\n\t\ts += fmt.Sprint(l)\n\t\ts += \"\\n\"\n\t}\n\treturn s\n}", "func (r *Key) marshal(c *Client) ([]byte, error) {\n\tm, err := expandKey(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling Key: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func (p PropagatedRouteTable) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"ids\", p.IDs)\n\tpopulate(objectMap, \"labels\", p.Labels)\n\treturn json.Marshal(objectMap)\n}", "func (npi NamedPartitionInformation) MarshalJSON() ([]byte, error) {\n\tnpi.ServicePartitionKind = ServicePartitionKindNamed1\n\tobjectMap := make(map[string]interface{})\n\tif npi.Name != nil {\n\t\tobjectMap[\"Name\"] = npi.Name\n\t}\n\tif npi.ID != nil {\n\t\tobjectMap[\"Id\"] = npi.ID\n\t}\n\tif npi.ServicePartitionKind != \"\" {\n\t\tobjectMap[\"ServicePartitionKind\"] = npi.ServicePartitionKind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (h HubRouteTableProperties) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"associatedConnections\", h.AssociatedConnections)\n\tpopulate(objectMap, \"labels\", h.Labels)\n\tpopulate(objectMap, \"propagatingConnections\", h.PropagatingConnections)\n\tpopulate(objectMap, \"provisioningState\", h.ProvisioningState)\n\tpopulate(objectMap, \"routes\", h.Routes)\n\treturn json.Marshal(objectMap)\n}", "func (sg *StorageGroup) Marshal(b ...[]byte) ([]byte, error) {\n\tvar buf []byte\n\tif len(b) > 0 {\n\t\tbuf = b[0]\n\t}\n\n\treturn (*storagegroup.StorageGroup)(sg).\n\t\tStableMarshal(buf)\n}", "func (t *ASCIITree) PrintTree(w io.Writer) {\n\tancestorPrefix := \"\"\n\tfor _, parent := range t.Ancestors() {\n\t\tif parent.Level() <= 0 {\n\t\t\tcontinue\n\t\t}\n\t\tif parent.Last() {\n\t\t\tancestorPrefix += \" \"\n\t\t} else {\n\t\t\tancestorPrefix += \" │\"\n\t\t}\n\t}\n\n\tmyPrefix := \"\"\n\tmultilinePrefix := \"\"\n\tif t.Level() > 0 {\n\t\tif t.Last() {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" └── \"\n\t\t\t\tmultilinePrefix += \" \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" └─┬ \"\n\t\t\t\tmultilinePrefix += \" └─┬ \"\n\t\t\t}\n\t\t} else {\n\t\t\tif t.Empty() {\n\t\t\t\tmyPrefix += \" ├── \"\n\t\t\t\tmultilinePrefix += \" │ \"\n\t\t\t} else {\n\t\t\t\tmyPrefix += \" ├─┬ \"\n\t\t\t\tmultilinePrefix += \" │ │ \"\n\t\t\t}\n\t\t}\n\t}\n\n\tif t.Text != \"\" {\n\t\tlines := strings.Split(t.Text, \"\\n\")\n\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, myPrefix, lines[0])\n\t\tfor _, line := range lines[1:] {\n\t\t\tfmt.Fprintf(w, \"%s%s%s\\n\", ancestorPrefix, multilinePrefix, line)\n\t\t}\n\t}\n\n\tfor _, child := range t.children {\n\t\tchild.PrintTree(w)\n\t}\n}", "func (uid MyULID) Marshal() ([]byte, error) {\n\tif len(uid) < 16 {\n\t\treturn []byte(ZeroULID), nil\n\t}\n\treturn []byte(uid)[:16], nil\n}", "func (f *Forest) ToString() string {\n\n\tfh := f.height\n\t// tree height should be 6 or less\n\tif fh > 6 {\n\t\treturn \"forest too big to print \"\n\t}\n\n\toutput := make([]string, (fh*2)+1)\n\tvar pos uint8\n\tfor h := uint8(0); h <= fh; h++ {\n\t\trowlen := uint8(1 << (fh - h))\n\n\t\tfor j := uint8(0); j < rowlen; j++ {\n\t\t\tvar valstring string\n\t\t\tok := f.data.size() >= uint64(pos)\n\t\t\tif ok {\n\t\t\t\tval := f.data.read(uint64(pos))\n\t\t\t\tif val != empty {\n\t\t\t\t\tvalstring = fmt.Sprintf(\"%x\", val[:2])\n\t\t\t\t}\n\t\t\t}\n\t\t\tif valstring != \"\" {\n\t\t\t\toutput[h*2] += fmt.Sprintf(\"%02d:%s \", pos, valstring)\n\t\t\t} else {\n\t\t\t\toutput[h*2] += fmt.Sprintf(\" \")\n\t\t\t}\n\t\t\tif h > 0 {\n\t\t\t\t//\t\t\t\tif x%2 == 0 {\n\t\t\t\toutput[(h*2)-1] += \"|-------\"\n\t\t\t\tfor q := uint8(0); q < ((1<<h)-1)/2; q++ {\n\t\t\t\t\toutput[(h*2)-1] += \"--------\"\n\t\t\t\t}\n\t\t\t\toutput[(h*2)-1] += \"\\\\ \"\n\t\t\t\tfor q := uint8(0); q < ((1<<h)-1)/2; q++ {\n\t\t\t\t\toutput[(h*2)-1] += \" \"\n\t\t\t\t}\n\n\t\t\t\t//\t\t\t\t}\n\n\t\t\t\tfor q := uint8(0); q < (1<<h)-1; q++ {\n\t\t\t\t\toutput[h*2] += \" \"\n\t\t\t\t}\n\n\t\t\t}\n\t\t\tpos++\n\t\t}\n\n\t}\n\tvar s string\n\tfor z := len(output) - 1; z >= 0; z-- {\n\t\ts += output[z] + \"\\n\"\n\t}\n\treturn s\n}", "func MarshalChaos(s interface{}) (string, error) {\n\tb, err := json.MarshalIndent(s, \"\", \" \")\n\tif err != nil {\n\t\treturn \"\", errors.Wrapf(err, \"failed to marshal indent\")\n\t}\n\treturn string(b), nil\n}", "func (h HubRoute) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"destinationType\", h.DestinationType)\n\tpopulate(objectMap, \"destinations\", h.Destinations)\n\tpopulate(objectMap, \"name\", h.Name)\n\tpopulate(objectMap, \"nextHop\", h.NextHop)\n\tpopulate(objectMap, \"nextHopType\", h.NextHopType)\n\treturn json.Marshal(objectMap)\n}", "func (m HyperflexNode) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 2)\n\n\taO0, err := swag.WriteJSON(m.MoBaseMo)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\tvar dataAO1 struct {\n\t\tBuildNumber string `json:\"BuildNumber,omitempty\"`\n\n\t\tCluster *HyperflexClusterRef `json:\"Cluster,omitempty\"`\n\n\t\tClusterMember *AssetClusterMemberRef `json:\"ClusterMember,omitempty\"`\n\n\t\tDisplayVersion string `json:\"DisplayVersion,omitempty\"`\n\n\t\tHostName string `json:\"HostName,omitempty\"`\n\n\t\tHypervisor string `json:\"Hypervisor,omitempty\"`\n\n\t\tIdentity *HyperflexHxUuIDDt `json:\"Identity,omitempty\"`\n\n\t\tIP *HyperflexHxNetworkAddressDt `json:\"Ip,omitempty\"`\n\n\t\tLockdown *bool `json:\"Lockdown,omitempty\"`\n\n\t\tModelNumber string `json:\"ModelNumber,omitempty\"`\n\n\t\tPhysicalServer *ComputePhysicalRef `json:\"PhysicalServer,omitempty\"`\n\n\t\tRole string `json:\"Role,omitempty\"`\n\n\t\tSerialNumber string `json:\"SerialNumber,omitempty\"`\n\n\t\tStatus string `json:\"Status,omitempty\"`\n\n\t\tVersion string `json:\"Version,omitempty\"`\n\t}\n\n\tdataAO1.BuildNumber = m.BuildNumber\n\n\tdataAO1.Cluster = m.Cluster\n\n\tdataAO1.ClusterMember = m.ClusterMember\n\n\tdataAO1.DisplayVersion = m.DisplayVersion\n\n\tdataAO1.HostName = m.HostName\n\n\tdataAO1.Hypervisor = m.Hypervisor\n\n\tdataAO1.Identity = m.Identity\n\n\tdataAO1.IP = m.IP\n\n\tdataAO1.Lockdown = m.Lockdown\n\n\tdataAO1.ModelNumber = m.ModelNumber\n\n\tdataAO1.PhysicalServer = m.PhysicalServer\n\n\tdataAO1.Role = m.Role\n\n\tdataAO1.SerialNumber = m.SerialNumber\n\n\tdataAO1.Status = m.Status\n\n\tdataAO1.Version = m.Version\n\n\tjsonDataAO1, errAO1 := swag.WriteJSON(dataAO1)\n\tif errAO1 != nil {\n\t\treturn nil, errAO1\n\t}\n\t_parts = append(_parts, jsonDataAO1)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (r *AWSEMRCluster) MarshalJSON() ([]byte, error) {\n\ttype Properties AWSEMRCluster\n\treturn json.Marshal(&struct {\n\t\tType string\n\t\tProperties Properties\n\t}{\n\t\tType: r.AWSCloudFormationType(),\n\t\tProperties: (Properties)(*r),\n\t})\n}", "func (r *HealthCheck) marshal(c *Client) ([]byte, error) {\n\tm, err := expandHealthCheck(c, r)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error marshalling HealthCheck: %w\", err)\n\t}\n\n\treturn json.Marshal(m)\n}", "func (t ThreeTierFullResourceNames) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"applicationServer\", t.ApplicationServer)\n\tpopulate(objectMap, \"centralServer\", t.CentralServer)\n\tpopulate(objectMap, \"databaseServer\", t.DatabaseServer)\n\tobjectMap[\"namingPatternType\"] = NamingPatternTypeFullResourceName\n\tpopulate(objectMap, \"sharedStorage\", t.SharedStorage)\n\treturn json.Marshal(objectMap)\n}", "func (pm Msg) pack() []byte {\n\tout := new(bytes.Buffer)\n\n\tfmt.Fprintf(out, \"%s\\n%s\\n%s\\n%s\\n%s\\n\", pm.Src, pm.Dst, pm.Wdir, pm.Type, pm.Attr.pack())\n\tln := len(pm.Data)\n\tfmt.Fprintf(out, \"%d\\n%s\", ln, pm.Data[:ln])\n\n\treturn out.Bytes()\n}", "func (d *DOM) Marshal() (string, error) {\n\tb, err := json.Marshal(d)\n\treturn string(b), err\n}", "func (s SapLandscapeMonitorPropertiesGrouping) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"landscape\", s.Landscape)\n\tpopulate(objectMap, \"sapApplication\", s.SapApplication)\n\treturn json.Marshal(objectMap)\n}", "func stringify(n *BinarySearchNode, level int, builder *strings.Builder) {\n\tif n != nil {\n\t\tformat := \"\"\n\t\tfor i := 0; i < level; i++ {\n\t\t\tformat += \" \"\n\t\t}\n\t\tformat += \"---[ \"\n\t\tlevel++\n\t\tstringify(n.left, level, builder)\n\t\tbuilder.WriteString(fmt.Sprintf(format+\"%d\\n\", n.value))\n\t\tstringify(n.right, level, builder)\n\t}\n}", "func (phe PartitionHealthEvaluation) MarshalJSON() ([]byte, error) {\n\tphe.Kind = KindPartition\n\tobjectMap := make(map[string]interface{})\n\tif phe.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = phe.PartitionID\n\t}\n\tif phe.UnhealthyEvaluations != nil {\n\t\tobjectMap[\"UnhealthyEvaluations\"] = phe.UnhealthyEvaluations\n\t}\n\tif phe.AggregatedHealthState != \"\" {\n\t\tobjectMap[\"AggregatedHealthState\"] = phe.AggregatedHealthState\n\t}\n\tif phe.Description != nil {\n\t\tobjectMap[\"Description\"] = phe.Description\n\t}\n\tif phe.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = phe.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (h Hash) MarshalText() ([]byte, error) {\n\treturn []byte(h.String()), nil\n}", "func (a Level) MarshalJSON() ([]byte, error) {\n\tres, err := a.MarshalText()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn append(append([]byte{'\"'}, res...), '\"'), nil\n}", "func (o *Organism) MarshalBinary() ([]byte, error) {\n\tvar buf bytes.Buffer\n\t_, err := fmt.Fprintln(&buf, o.Fitness, o.Generation, o.highestFitness, o.isPopulationChampionChild, o.Genotype.Id)\n\to.Genotype.Write(&buf)\n\tif err != nil {\n\t\treturn nil, err\n\t} else {\n\t\treturn buf.Bytes(), nil\n\t}\n}", "func (h Horse) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]any)\n\tpopulate(objectMap, \"isAShowHorse\", h.IsAShowHorse)\n\tpopulate(objectMap, \"name\", h.Name)\n\treturn json.Marshal(objectMap)\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\treturn dfsSerial(root, \"\")\n}", "func (this *Codec) serialize(root *TreeNode) string {\n\tstrs := make([]string, 0)\n\tvar preorder func(*TreeNode)\n\tpreorder = func(node *TreeNode) {\n\t\tif node == nil {\n\t\t\tstrs = append(strs, this.null)\n\t\t\treturn\n\t\t}\n\n\t\tstrs = append(strs, strconv.Itoa(node.Val))\n\t\tpreorder(node.Left)\n\t\tpreorder(node.Right)\n\t}\n\tpreorder(root)\n\treturn strings.Join(strs, this.sep)\n}", "func (l *List) MarshalText() (text []byte, err error) {\n\theading := strings.TrimSpace(l.rootItem.desc)\n\ts := heading + \"\\n\"\n\ts += strings.Repeat(\"=\", len(heading)) + \"\\n\"\n\tfor _, li := range l.rootItem.items {\n\t\ts = recursiveMarshal(s, li, 0)\n\t}\n\treturn []byte(s), nil\n}", "func (cfg frozenConfig) MarshalToString(val interface{}) (string, error) {\n buf, err := encoder.Encode(val, cfg.encoderOpts)\n return rt.Mem2Str(buf), err\n}", "func (m SdcardPartition) MarshalJSON() ([]byte, error) {\n\t_parts := make([][]byte, 0, 1)\n\n\tvar dataAO0 struct {\n\t\tType *string `json:\"Type,omitempty\"`\n\n\t\tVirtualDrives []*SdcardVirtualDrive `json:\"VirtualDrives\"`\n\t}\n\n\tdataAO0.Type = m.Type\n\n\tdataAO0.VirtualDrives = m.VirtualDrives\n\n\tjsonDataAO0, errAO0 := swag.WriteJSON(dataAO0)\n\tif errAO0 != nil {\n\t\treturn nil, errAO0\n\t}\n\t_parts = append(_parts, jsonDataAO0)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (this *Codec) serialize(root *TreeNode) string {\n if root == nil {\n return \"x\"\n }\n return strconv.Itoa(root.Val) + \",\" + this.serialize(root.Left)+ \",\" + this.serialize(root.Right)\n}", "func (s DescribeUserHierarchyStructureInput) MarshalFields(e protocol.FieldEncoder) error {\n\te.SetValue(protocol.HeaderTarget, \"Content-Type\", protocol.StringValue(\"application/json\"), protocol.Metadata{})\n\n\tif s.InstanceId != nil {\n\t\tv := *s.InstanceId\n\n\t\tmetadata := protocol.Metadata{}\n\t\te.SetValue(protocol.PathTarget, \"InstanceId\", protocol.QuotedValue{ValueMarshaler: protocol.StringValue(v)}, metadata)\n\t}\n\treturn nil\n}", "func (ll *LogLevel) MarshalText() ([]byte, error) {\n\treturn []byte(ll.String()), nil\n}", "func (h H) MarshalXML(e *xml.Encoder, start xml.StartElement) error {\n\tstart.Name = xml.Name{\n\t\tSpace: \"\",\n\t\tLocal: \"map\",\n\t}\n\tif err := e.EncodeToken(start); err != nil {\n\t\treturn err\n\t}\n\tfor key, value := range h {\n\t\telem := xml.StartElement{\n\t\t\tName: xml.Name{Space: \"\", Local: key},\n\t\t\tAttr: []xml.Attr{},\n\t\t}\n\t\tif err := e.EncodeElement(value, elem); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\tif err := e.EncodeToken(xml.EndElement{Name: start.Name}); err != nil {\n\t\treturn err\n\t}\n\treturn nil\n}", "func (d Dir) MarshalJSON() ([]byte, error) {\n\tbuffer := bytes.NewBufferString(`\"`)\n\tbuffer.WriteString(toString[d])\n\tbuffer.WriteString(`\"`)\n\treturn buffer.Bytes(), nil\n}", "func (host Host) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(host.String())\n}", "func (uh UnlockHash) MarshalJSON() ([]byte, error) {\n\treturn json.Marshal(uh.String())\n}", "func (npsd NamedPartitionSchemeDescription) MarshalJSON() ([]byte, error) {\n\tnpsd.PartitionScheme = PartitionSchemeNamed1\n\tobjectMap := make(map[string]interface{})\n\tif npsd.Count != nil {\n\t\tobjectMap[\"Count\"] = npsd.Count\n\t}\n\tif npsd.Names != nil {\n\t\tobjectMap[\"Names\"] = npsd.Names\n\t}\n\tif npsd.PartitionScheme != \"\" {\n\t\tobjectMap[\"PartitionScheme\"] = npsd.PartitionScheme\n\t}\n\treturn json.Marshal(objectMap)\n}", "func (h Hostname) MarshalText() ([]byte, error) {\n\treturn []byte(string(h)), nil\n}", "func (m LogicalSwitch) MarshalJSON() ([]byte, error) {\n\tvar _parts [][]byte\n\n\taO0, err := swag.WriteJSON(m.ManagedResource)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, aO0)\n\n\tvar data struct {\n\t\tAdminState *string `json:\"admin_state\"`\n\n\t\tIPPoolID string `json:\"ip_pool_id,omitempty\"`\n\n\t\tReplicationMode string `json:\"replication_mode,omitempty\"`\n\n\t\tTransportZoneID *string `json:\"transport_zone_id\"`\n\n\t\tVlan int64 `json:\"vlan,omitempty\"`\n\n\t\tVni int32 `json:\"vni,omitempty\"`\n\t}\n\n\tdata.AdminState = m.AdminState\n\n\tdata.IPPoolID = m.IPPoolID\n\n\tdata.ReplicationMode = m.ReplicationMode\n\n\tdata.TransportZoneID = m.TransportZoneID\n\n\tdata.Vlan = m.Vlan\n\n\tdata.Vni = m.Vni\n\n\tjsonData, err := swag.WriteJSON(data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_parts = append(_parts, jsonData)\n\n\treturn swag.ConcatJSON(_parts...), nil\n}", "func (s SystemTopic) MarshalJSON() ([]byte, error) {\n\tobjectMap := make(map[string]interface{})\n\tpopulate(objectMap, \"id\", s.ID)\n\tpopulate(objectMap, \"identity\", s.Identity)\n\tpopulate(objectMap, \"location\", s.Location)\n\tpopulate(objectMap, \"name\", s.Name)\n\tpopulate(objectMap, \"properties\", s.Properties)\n\tpopulate(objectMap, \"systemData\", s.SystemData)\n\tpopulate(objectMap, \"tags\", s.Tags)\n\tpopulate(objectMap, \"type\", s.Type)\n\treturn json.Marshal(objectMap)\n}", "func (e Level) MarshalJSON() ([]byte, error) {\n\treturn []byte(fmt.Sprintf(`\"%v\"`, e)), nil\n}", "func (this *school) MarshalBinary() (data []byte, err error) {\n\t// var addrs = map[uintptr]uint64{}\n\tbuf := bytes.NewBuffer(make([]byte, 0, ((8 + (len(this.Teachers) * 507)) + (8 + (len(this.Pupils) * 507)))))\n\tif err = this.marshalTo(buf /*, addrs*/); err == nil {\n\t\tdata = buf.Bytes()\n\t}\n\treturn\n}", "func (this *NamespacedName) MarshalJSON() ([]byte, error) {\n\tstr, err := CommonMarshaler.MarshalToString(this)\n\treturn []byte(str), err\n}", "func (l LTIContextType) MarshalJSON() ([]byte, error) {\n\tvar ctxs []string\n\tvar ctxsStr string\n\n\tfor _, ctx := range l {\n\t\tctxs = append(ctxs, string(ctx))\n\t}\n\n\tctxsStr = strings.Join(ctxs, \",\")\n\tctxsStr = fmt.Sprintf(\"\\\"%s\\\"\", ctxsStr)\n\n\treturn []byte(ctxsStr), nil\n}", "func (epqsc EnsurePartitionQurumSafetyCheck) MarshalJSON() ([]byte, error) {\n\tepqsc.Kind = KindEnsurePartitionQuorum\n\tobjectMap := make(map[string]interface{})\n\tif epqsc.PartitionID != nil {\n\t\tobjectMap[\"PartitionId\"] = epqsc.PartitionID\n\t}\n\tif epqsc.Kind != \"\" {\n\t\tobjectMap[\"Kind\"] = epqsc.Kind\n\t}\n\treturn json.Marshal(objectMap)\n}", "func stringTree(builder *strings.Builder, depth int, node *TreeNode) (*strings.Builder, error) {\n\tvar err error\n\tfor i := 0; i < depth; i++ {\n\t\tif _, err = builder.WriteString(\" \"); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\tfmt.Fprintf(builder, \"└ %d\\n\", node.Val)\n\n\tif node.Left != nil {\n\t\tif builder, err = stringTree(builder, depth+1, node.Left); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\tif node.Right != nil {\n\t\tif builder, err = stringTree(builder, depth+1, node.Right); err != nil {\n\t\t\treturn builder, err\n\t\t}\n\t}\n\treturn builder, nil\n}", "func (r *Chapter) Marshal() ([]byte, error) {\n\treturn json.Marshal(r)\n}" ]
[ "0.72192514", "0.6860378", "0.5548206", "0.52124643", "0.51653415", "0.50677866", "0.4993746", "0.4927463", "0.49178874", "0.4913059", "0.48793265", "0.4859133", "0.48533386", "0.48482224", "0.48187178", "0.47770202", "0.4766247", "0.4766247", "0.4762357", "0.4755859", "0.475494", "0.47514337", "0.47279862", "0.47221738", "0.47221738", "0.4713375", "0.46628076", "0.46541762", "0.46465", "0.46465", "0.4643268", "0.46401712", "0.46364084", "0.46184868", "0.45753488", "0.45746037", "0.45392266", "0.45363152", "0.45340925", "0.45237577", "0.45207295", "0.45108944", "0.45103425", "0.45061898", "0.4490498", "0.44899294", "0.44618824", "0.44539604", "0.44449186", "0.4444734", "0.4444734", "0.44378555", "0.44356683", "0.44247687", "0.44246858", "0.4416261", "0.4407098", "0.43912312", "0.43875077", "0.43852565", "0.43818682", "0.43768424", "0.43767315", "0.4368136", "0.4366923", "0.43609485", "0.436092", "0.433976", "0.43381047", "0.4337389", "0.4329583", "0.43249846", "0.43246448", "0.43243548", "0.43189755", "0.43172085", "0.43150303", "0.4315013", "0.43120757", "0.43068147", "0.4303862", "0.4303151", "0.43006378", "0.43006375", "0.4299089", "0.4292772", "0.4281087", "0.42804724", "0.4274975", "0.42728513", "0.42691237", "0.42661998", "0.42650378", "0.4259356", "0.42558745", "0.42549577", "0.4252978", "0.42521575", "0.42475724", "0.42468274" ]
0.71129423
1
New is a variadic function meaning it can take any number of arguments as input and create a new ordered list with them. The values provided may be out of order (unordered). The list we produce will be ordered.
func New(values ...uint16) (l *List) { l = &List{} // init the ptr for _, value := range values { l.Insert(value) } return l }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func New(vals ...interface{}) *List {\n\thead := list.New()\n\tfor _, v := range vals {\n\t\thead.PushBack(v)\n\t}\n\treturn &List{head}\n}", "func NewArgs(first *int32, after interface{}, last *int32, before interface{}, more ...interface{}) Args {\n\treturn append(Args{first, after, last, before}, more...)\n}", "func NewList(vs ...Value) Value {\n\treturn StrictPrepend(vs, EmptyList)\n}", "func New(values ...interface{}) *List {\n\tlist := &List{}\n\tif len(values) > 0 {\n\t\tlist.Add(values)\n\t}\n\treturn list\n}", "func New(elems ...interface{}) List {\n\tl := Mzero()\n\tfor _, elem := range elems {\n\t\tl = Cons(elem, l)\n\t}\n\treturn Reverse(l)\n}", "func New(values ...interface{}) *List {\n\tlist := &List{}\n\tif len(values) > 0 {\n\t\tlist.Add(values...)\n\t}\n\treturn list\n}", "func NewList(args ...interface{}) *List {\n\tl := List{}\n\tfor _, data := range args {\n\t\tl.PushBack(data)\n\t}\n\treturn &l\n}", "func NewList(args ...interface{}) *List {\n\tl := &List{}\n\tfor _, v := range args {\n\t\tl.PushBack(v)\n\t}\n\treturn l\n}", "func newArguments(arguments []string) *Arguments {\n\treturn &Arguments{\n\t\targs: arguments,\n\t\tcount: len(arguments),\n\t\tindex: 0,\n\t\trawMode: false,\n\t}\n}", "func New(args ...float64) Tuple {\n\treturn args\n}", "func newListFormulaArg(l []formulaArg) formulaArg {\n\treturn formulaArg{Type: ArgList, List: l}\n}", "func NewValues(kvs ...string) Values {\n\tif len(kvs)%2 == 1 {\n\t\tpanic(\"twister: even number args required for NewParam\")\n\t}\n\tm := make(Values)\n\tfor i := 0; i < len(kvs); i += 2 {\n\t\tm.Add(kvs[i], kvs[i+1])\n\t}\n\treturn m\n}", "func New(value ...interface{}) Stack {\n\ts := Stack{}\n\tfor _, e := range value {\n\t\ts.Push(e)\n\t}\n\treturn s\n}", "func NewList(values []interface{}) *List {\n\tvar items []Item\n\tlist := List{Len: len(values)}\n\tfor i := 0; i < list.Len; i++ {\n\t\titem := Item{Value: values[i]}\n\t\tif i > 0 {\n\t\t\t// i - 1 should reference to previous el before append\n\t\t\titem.Prev = &items[i-1]\n\t\t\titems[i-1].Next = &item\n\t\t}\n\t\titems = append(items, item)\n\t}\n\tif len(items) == 0 {\n\t\treturn &list\n\t}\n\tlist.First = &items[0]\n\tlist.Last = &items[len(items)-1]\n\treturn &list\n}", "func NewList(vs ...Value) *List {\n\tacc := EmptyList\n\tfor _, v := range vs {\n\t\tacc = acc.Append(v)\n\t}\n\treturn acc\n}", "func createList(arg string) []string {\n\tvar retObject = []string{arg}\n\treturn retObject\n}", "func NewList(v ...Value) List {\n\tr := emptyList\n\tfor i := len(v) - 1; i >= 0; i-- {\n\t\tr = &list{\n\t\t\tfirst: v[i],\n\t\t\trest: r,\n\t\t\tcount: r.count + 1,\n\t\t}\n\t}\n\treturn r\n}", "func New(values ...interface{}) Stack {\n\tlist := linkedlist.New(values)\n\tstack := Stack{list}\n\treturn stack\n}", "func NewList(vs ...Value) List {\n\treturn List{&vs}\n}", "func NewArguments(ps []PositionalArgument, ks []KeywordArgument, dicts []interface{}) Arguments {\n\treturn Arguments{\n\t\tpositionals: ps,\n\t\tkeywords: ks,\n\t\texpandedDicts: dicts,\n\t}\n}", "func NewOrderedValues(vals [][]string) OrderedValues {\n\tif len(vals) == 0 {\n\t\treturn nil\n\t}\n\tvar nov OrderedValues\n\tfor i := range vals {\n\t\tvar a [][]byte\n\t\tfor j := range vals[i] {\n\t\t\ta = append(a, []byte(vals[i][j]))\n\t\t}\n\t\tnov = append(nov, a)\n\t}\n\treturn nov\n}", "func New(l interface{}) list.Interface {\n\tif reflect.TypeOf(l).Kind() != reflect.Slice {\n\t\tpanic(fmt.Errorf(\"Param must be a slice\"))\n\t}\n\n\ts := reflect.ValueOf(l)\n\titems := make([]interface{}, s.Len())\n\n\tfor i := 0; i < len(items); i++ {\n\t\titems[i] = s.Index(i).Interface()\n\t}\n\n\treturn &randList{\n\t\tlist: items,\n\t\tperm: rand.Perm(len(items)),\n\t\toffset: 0,\n\t}\n}", "func New(values ...interface{}) Array {\n\tarray := Array{}\n\tarray = array.Push(values...)\n\treturn array\n}", "func NewList(list uint32, mode uint32) {\n\tsyscall.Syscall(gpNewList, 2, uintptr(list), uintptr(mode), 0)\n}", "func NewList(elems ...interface{}) *List {\n\tl := List{}\n\tfor _, elem := range elems {\n\t\tl.PushBack(elem)\n\t}\n\n\treturn &l\n}", "func New(vals ...datatype) (l *LinkedList) {\n\tl = &LinkedList{}\n\tvar cur, n *node = nil, nil\n\tfor _, v := range vals {\n\t\tn = &node{\n\t\t\tdata: v,\n\t\t}\n\t\tif cur == nil {\n\t\t\tcur = n\n\t\t\tl.head = n\n\t\t} else {\n\t\t\tcur.next = n\n\t\t\tcur = cur.next\n\t\t}\n\t}\n\treturn l\n}", "func ListConstructor(context RunContext, arguments []Argument) Value {\n\tvalues := make([]Value, len(arguments))\n\tfor i, arg := range arguments {\n\t\tvar value = EvalArgument(context, arg)\n\n\t\t// accept blocks within a list\n\t\t// as dictionaries in order to support\n\t\t// [{...} {...} ...] constructions\n\t\t//\n\t\tif value.Type() == TypeBlock {\n\t\t\tvalue = NewDictionaryWithBlock(context, value.(Block))\n\t\t}\n\n\t\tvalues[i] = value\n\t}\n\treturn NewListValue(values)\n}", "func New(elements []int) LinkedList {\n\tll := LinkedList{}\n\n\tfor _, item := range elements {\n\t\tll.Append(item)\n\t}\n\n\treturn ll\n}", "func NewList(first Term, rest ...Term) NonEmptyList {\n\treturn append(NonEmptyList{first}, rest...)\n}", "func ArgsortNew(src []float64) []int {\n\tinds := make([]int, len(src))\n\tfor i := range src {\n\t\tinds[i] = i\n\t}\n\tArgsort(src, inds)\n\treturn inds\n}", "func NewList(list uint32, mode uint32) {\n C.glowNewList(gpNewList, (C.GLuint)(list), (C.GLenum)(mode))\n}", "func NewQueue(args []func(http.ResponseWriter, *http.Request) (http.ResponseWriter, *http.Request)) *Queue {\n\tq := &Queue{}\n\tfor _, f := range args {\n\t\tq.list = append(q.list, f)\n\t}\n\treturn q\n}", "func newList(data interface{}) *List {\n\tnewL := new(List)\n\tnewL.Insert(data)\n\treturn newL\n}", "func NewList(elements ...interface{}) *List {\n\tl := &List{}\n\n\tfor _, el := range elements {\n\t\tl.PushBack(el)\n\t}\n\n\treturn l\n}", "func New(maxlevel int, cmpFn CompareFn) *List {\n\treturn NewCustom(maxlevel, DefaultProbability, cmpFn, time.Now().Unix())\n}", "func List(values []int) *ListExpression {\n\tif len(values) < 1 {\n\t\tpanic(\"schedule: invalid ListExpression expression\")\n\t}\n\tsort.Ints(values)\n\treturn &ListExpression{\n\t\tvalues: values,\n\t}\n}", "func (o *DailyBurstUsage) NewList() *Args {\n\to.Args.Verb = \"GET\"\n\to.Args.RequiresAuth = true\n\treturn o.Args\n}", "func New(values ...interface{}) *Set {\n\tset := &Set{items: make(map[interface{}]struct{})}\n\tif len(values) > 0 {\n\t\tset.Add(values...)\n\t}\n\treturn set\n}", "func variadic(typ int, op string, ods exprlist) *expr {\n\treturn &expr{\n\t\tsexp: append(exprlist{atomic(typ, op)}, ods...),\n\t}\n}", "func (e *exprHelper) NewList(elems ...ast.Expr) ast.Expr {\n\treturn e.exprFactory.NewList(e.nextMacroID(), elems, []int32{})\n}", "func ArgsList(args ...interface{}) Node {\n\tvar result Node\n\tfor _, arg := range args {\n\t\tif result.Node == nil {\n\t\t\tresult = ToNode(arg)\n\t\t} else {\n\t\t\tresult = Expr(\",\", result, arg)\n\t\t}\n\t}\n\treturn result\n}", "func (o *Transactions) NewList() *Args {\n\to.Args.Verb = \"GET\"\n\to.Args.RequiresAuth = true\n\treturn o.Args\n}", "func newList(vert bool, width, height float32) *List {\n\n\tli := new(List)\n\tli.initialize(vert, width, height)\n\treturn li\n}", "func NewList(list uint32, mode uint32) {\n\tC.glowNewList(gpNewList, (C.GLuint)(list), (C.GLenum)(mode))\n}", "func New(fields string, number bool, reverse bool, duplicate bool, files []string) *Sort {\n\treturn &Sort{\n\t\targs: args{\n\t\t\tfields: fields,\n\t\t\tnumber: number,\n\t\t\treverse: reverse,\n\t\t\tduplicate: duplicate,\n\t\t},\n\n\t\tfiles: files,\n\t}\n}", "func New(inOrder bool) *SkipList {\n\treturn NewWithMaxLevel(DefaultMaxLevel, inOrder)\n}", "func New(args ...func(*Impl)) *Impl {\n\tl := &Impl{os.Stderr, DefaultTimeFormat}\n\tfor _, arg := range args {\n\t\targ(l)\n\t}\n\treturn l\n}", "func NewList(capacity int) *List {\n\tif capacity < 0 {\n\t\tpanic(\"negative capacity\")\n\t}\n\n\tpids := make([]ID, capacity)\n\tfor i := 0; i < capacity; i++ {\n\t\tpids[i] = ID(i)\n\t}\n\n\treturn &List{\n\t\tentities: make([]*Entity, capacity+1), // + 1 due to the 'highest' variable\n\n\t\tlowest: -1,\n\t\thighest: -1,\n\n\t\tavailableIDs: pids,\n\t}\n}", "func NewListOpts(validator ValidatorFctType) ListOpts {\n var values []string\n return *NewListOptsRef(&values, validator)\n}", "func (list *List) Add(values ...interface{}){\n\tlist.growBy(len(values))\n\tfor _ , value := range values {\n\t\tlist.elements[list.size] = value\n\t\tlist.size++\n\t}\n}", "func listAppend(self BlObject, args ...BlObject) BlObject {\n var obj BlObject\n if blParseArguments(\"o\", args, &obj) == -1 {\n return nil\n }\n lobj := self.(*BlListObject)\n lobj.list = append(lobj.list, obj)\n lobj.lsize++\n\n return BlNil\n}", "func newLedger(entries []Entry) (ledger, error) {\n\tvar entriesCopy ledger\n\tfor _, e := range entries {\n\t\tt, err := time.Parse(\"2006-01-02\", e.Date)\n\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"Invalid date format: %s\", e.Date)\n\t\t}\n\n\t\te.dateTime = t\n\n\t\tif len(e.Description) > 25 {\n\t\t\te.Description = fmt.Sprintf(\"%s...\", e.Description[:22])\n\t\t}\n\n\t\tentriesCopy = append(entriesCopy, e)\n\t}\n\n\tsort.Sort(entriesCopy)\n\n\treturn entriesCopy, nil\n}", "func NewWithValues(r, c int, vals []Frac) M {\n\treturn M{r: r, c: c, values: vals}\n}", "func NewList(store *Tree) (*History, chan string, <-chan *messages.ChatMessage) {\n\tqueryChan := make(chan string)\n\toutChan := make(chan *messages.ChatMessage)\n\treturn &History{\n\t\tThreadView: vs.New(store),\n\t\tViewIDs: make(map[string]struct{}),\n\t\tQuery: queryChan,\n\t\tOutbound: outChan,\n\t}, queryChan, outChan\n}", "func NewListNode(values ...int) *ListNode {\n\tif len(values) == 0 {\n\t\treturn nil\n\t}\n\troot := &ListNode{Val: values[0]}\n\tl1 := root\n\n\tfor _, i := range values[1:] {\n\t\tl2 := l1.Append(i)\n\t\tl1 = l2\n\t}\n\n\treturn root\n\n}", "func NewListValue(list []interface{}) ListValue {\n\treturn ListValue(list)\n}", "func NewArgs(schema string, args []string) (*Args, error) {\n\ta := Args{\n\t\tschema: schema,\n\t\targs: args,\n\t\tunexpectedArguments: make([]rune, 0),\n\t\tmarhalers: map[rune]argumentMarshaler{},\n\t\tcurrentArgument: nil,\n\t}\n\n\treturn &a, a.parse()\n}", "func NewList(initial []W) UpdatableList {\n\tul := &updatableList{}\n\tul.Update(initial)\n\treturn ul\n}", "func (q *Queue) New(js ...Job) {\n\tfor _, j := range js {\n\t\tq.j <- j\n\t}\n}", "func NewLIFO(ts ...T) *LIFO {\n\ts := &LIFO{}\n\tfor i := range ts {\n\t\ts.Push(ts[i])\n\t}\n\treturn s\n}", "func (s Stream) NewValues(n int32) (capnp.Float64List, error) {\n\tl, err := capnp.NewFloat64List(s.Struct.Segment(), n)\n\tif err != nil {\n\t\treturn capnp.Float64List{}, err\n\t}\n\terr = s.Struct.SetPtr(1, l.List.ToPtr())\n\treturn l, err\n}", "func (list *List) Prepend(values ...interface{}) {\n\t// in reverse to keep passed order i.e. [\"c\",\"d\"] -> Prepend([\"a\",\"b\"]) -> [\"a\",\"b\",\"c\",d\"]\n\tfor v := len(values) - 1; v >= 0; v-- {\n\t\tnewElement := &element{value: values[v], next: list.first}\n\t\tlist.first = newElement\n\t\tif list.size == 0 {\n\t\t\tlist.last = newElement\n\t\t}\n\t\tlist.size++\n\t}\n}", "func (o *Discounts) NewList() *Args {\n\to.Args.Verb = \"GET\"\n\to.Args.RequiresAuth = true\n\treturn o.Args\n}", "func CloneListArg(n ListArg) ListArg {\n\tres := make(ListArg, 0, len(n))\n\tcopy(res, n)\n\treturn res\n}", "func NewOrderedMap(kvList ...*KV) (om *OrderedMap) {\n\tom = &OrderedMap{\n\t\tidxLookup: make(map[keyType]int),\n\t}\n\n\tfor i := 0; i < len(kvList); i++ {\n\t\tom.Set(kvList[i].Key, kvList[i].Value)\n\t}\n\treturn\n}", "func newOptions(opts ...Option) Options {\n\topt := Options{}\n\n\tfor _, o := range opts {\n\t\to(&opt)\n\t}\n\treturn opt\n}", "func NewList(g ...Getter) *List {\n\tlist := &List{\n\t\tlist: g,\n\t}\n\tlist.GetProxy = NewGetProxy(list) // self\n\treturn list\n}", "func makeParams(args ...interface{}) []rpcValue {\n\tif len(args) == 0 {\n\t\treturn nil\n\t}\n\tarr := make([]rpcValue, 0, len(args))\n\tfor _, v := range args {\n\t\tarr = append(arr, makeValue(v))\n\t}\n\treturn arr\n}", "func CloneOrderBy(n OrderBy) OrderBy {\n\tres := make(OrderBy, 0, len(n))\n\tfor _, x := range n {\n\t\tres = append(res, CloneRefOfOrder(x))\n\t}\n\treturn res\n}", "func generateTestOrdersList(count int) OrderList {\n\tl := make(OrderList, count)\n\tidx := 0\n\n\tfor kv := range generateTestOrders(count) {\n\t\tl[idx] = kv.Value.(*Order)\n\n\t\tidx++\n\t}\n\n\treturn l\n}", "func BuildList(vals []int) *ListNode {\n\thead := &ListNode{}\n\n\ttmp := head\n\tfor _, val := range vals {\n\t\ttmp.Next = &ListNode{Val: val}\n\t\ttmp = tmp.Next\n\t}\n\n\treturn head.Next\n}", "func New(a []int) *PTree {\n\tn := len(a)\n\tb := make([]int, n)\n\tfor i := 0; i < n; i++ {\n\t\tb[i] = i\n\t}\n\tsort.Sort(&arrayPerm{a, b})\n\tt := &PTree{}\n\tt.a = a\n\tt.n = len(a)\n\tt.root = make([]*node, n+1)\n\tt.root[n] = t.build(b)\n\tfor i := n - 1; i >= 0; i-- {\n\t\tt.root[i] = t.del(t.root[i+1], i)\n\t}\n\treturn t\n}", "func newOptions(opts ...Option) Options {\n\topt := Options{}\n\n\tfor _, o := range opts {\n\t\to(&opt)\n\t}\n\n\treturn opt\n}", "func newOptions(opts ...Option) Options {\n\topt := Options{}\n\n\tfor _, o := range opts {\n\t\to(&opt)\n\t}\n\n\treturn opt\n}", "func build_quote_from_list(args *List) *quote {\n\tif args == nil {\n\t\treturn Noop.unprotect()\n\t}\n\tsrc := newBuf(0)\n\tvar chead, ctail *sNode\n\tfor ; args != nil; args = args.Next {\n\t\tsrc.Write(args.Value.Ser().Bytes())\n\t\tsrc.WriteString(\" \")\n\t\tif chead != nil {\n\t\t\t//this is safe because when the vm rewrites it will\n\t\t\t//just blindly fill literals, so it doesn't matter what\n\t\t\t//the actual type is\n\t\t\tctail.next = &sNode{synLiteral, args.Value, nil}\n\t\t\tctail = ctail.next\n\t\t} else {\n\t\t\tchead = &sNode{synLiteral, args.Value, nil}\n\t\t\tctail = chead\n\t\t}\n\t}\n\treturn &quote{false, &command{chead, nil}, src.Bytes()}\n}", "func MakeList(first LangType, rest ...LangType) List {\n\tlst := List{}\n\n\tlst = lst.Append(first).(List)\n\n\tfor _, item := range rest {\n\t\tlst = lst.Append(item).(List)\n\t}\n\n\treturn lst\n}", "func NewFloatList(lst []float64) FloatList {\n\tcpy := []float64{}\n\ttotal := 0.0\n\n\tfor _, val := range lst {\n\t\tcpy = append(cpy, val)\n\t\ttotal += val\n\t}\n\n\tsort.Float64s(cpy)\n\n\treturn FloatList{cpy, len(cpy), total}\n}", "func NewFIFO(ts ...T) *FIFO {\n\ts := &FIFO{}\n\tfor i := range ts {\n\t\ts.Push(ts[i])\n\t}\n\treturn s\n}", "func New(raws []string) *Args {\n\ta := Args{raws: make([]string, len(raws))}\n\n\t// Raws\n\tcopy(a.raws, raws)\n\n\t// Values\n\tfor i := 0; i < len(raws); i++ {\n\t\traw := raws[i]\n\n\t\tif strings.HasPrefix(raw, \"-\") {\n\n\t\t\t// Option\n\t\t\tif strings.Contains(raw, \"=\") {\n\n\t\t\t\t// Binary Option\n\t\t\t\tif tokens := strings.SplitN(raw, \"=\", 2); len(tokens) == 2 {\n\t\t\t\t\tkey := strings.TrimPrefix(tokens[0], \"-\")\n\t\t\t\t\tbinOpt := keyVal{key, tokens[1]}\n\t\t\t\t\ta.binOpts = append(a.binOpts, binOpt)\n\t\t\t\t}\n\t\t\t} else {\n\n\t\t\t\t// Unary Option\n\t\t\t\tuniOpt := strings.TrimPrefix(raw, \"-\")\n\t\t\t\ta.uniOpts = append(a.uniOpts, uniOpt)\n\t\t\t}\n\t\t} else {\n\n\t\t\t// Non-Option\n\t\t\ta.vals = append(a.vals, raw)\n\t\t}\n\t}\n\n\treturn &a\n}", "func newListFromUIDs(uids []string) *CSPList {\n\treturn NewListBuilder().WithUIDs(uids...).List()\n}", "func (l List) With(keyvals ...interface{}) List {\n\tkeyvals = flattenFix(keyvals)\n\tlist := l.clone(len(l) + len(keyvals))\n\tlist = append(list, keyvals...)\n\treturn list\n}", "func CreateNewItems(f *Feed, gfItems []*gofeed.Item) []*Item {\n\tvar items []*Item\n\n\tfor i := range gfItems {\n\t\t// Reverse order so that, when timestamps are the same, sorting by\n\t\t// IDs will result in the right order.\n\t\tgfi := gfItems[len(gfItems)-i-1]\n\t\titems = append(items, createNewItem(gfi, f))\n\t}\n\n\thandleItemQuirks(items, gfItems, f)\n\n\tlog.Debugf(\"Created %d items for [Feed: %d]\", len(gfItems), f.ID())\n\treturn items\n}", "func CreateNew(val interface{}) (*Queue, error) {\r\n\treturn &Queue{\r\n\t\tQueueList: []interface{}{val},\r\n\t}, nil\r\n}", "func New(vals ...interface{}) Set {\n\ts := &setImpl{\n\t\tset: make(map[interface{}]struct{}, 0),\n\t}\n\tfor _, i := range vals {\n\t\ts.Insert(i)\n\t}\n\treturn s\n}", "func NewRepeated(less LessFunc) *List {\n\tl := New(less)\n\tl.repeat = true\n\treturn l\n}", "func NewTree(from []int) *Tree {\n\ttreeSize := calcTreeSize(len(from))\n\tnodes := make([]int, treeSize)\n\n\tt := &Tree{nodes, len(from)}\n\tt.build(from, 0, 0, len(from)-1)\n\n\treturn t\n}", "func newList() *List {\n\tl := &List{\n\t\tch: make(chan sh.QData),\n\t}\n\treturn l\n}", "func createALinkedList(input []int) *Node {\n\thead := &Node{\n\t\tval: input[0],\n\t\tnext: nil,\n\t}\n\tcurrentNode := head\n\tfor i := 1; i < len(input); i++ {\n\t\tnextNode := &Node{\n\t\t\tval: input[i],\n\t\t\tnext: nil,\n\t\t}\n\t\tcurrentNode.next = nextNode\n\t\tcurrentNode = nextNode\n\t}\n\treturn head\n}", "func New(cnt int) []Aggregator {\n\treturn make([]Aggregator, cnt)\n}", "func New(cnt int) []Aggregator {\n\treturn make([]Aggregator, cnt)\n}", "func Append(m []interface{}, args ...interface{}) []interface{} {\n\tfor _, v := range args {\n\t\tm = append(m, v)\n\t}\n\treturn m\n}", "func NewListOptsRef(values *[]string, validator ValidatorFctType) *ListOpts {\n return &ListOpts{\n values: values,\n validator: validator,\n }\n}", "func New() *List { return new(List).Init() }", "func newToDoList() toDoList {\n\treturn toDoList{}\n}", "func newOrderMutation(c config, op Op, opts ...orderOption) *OrderMutation {\n\tm := &OrderMutation{\n\t\tconfig: c,\n\t\top: op,\n\t\ttyp: TypeOrder,\n\t\tclearedFields: make(map[string]struct{}),\n\t}\n\tfor _, opt := range opts {\n\t\topt(m)\n\t}\n\treturn m\n}", "func newOrderMutation(c config, op Op, opts ...orderOption) *OrderMutation {\n\tm := &OrderMutation{\n\t\tconfig: c,\n\t\top: op,\n\t\ttyp: TypeOrder,\n\t\tclearedFields: make(map[string]struct{}),\n\t}\n\tfor _, opt := range opts {\n\t\topt(m)\n\t}\n\treturn m\n}", "func newObjectList() *ObjectList {\n\treturn &ObjectList{\n\t\tObjectIDs: make([]int, 0, 200),\n\t}\n}", "func NewList() *List {\n newObj := &List {\n counters : make(map[string]Counter),\n }\n\n return newObj\n}", "func (a Args) from(i int) Args {\n\tif i >= len(a.All) {\n\t\ti = len(a.All) - 1\n\t}\n\ta.All = a.All[i+1:]\n\n\tif i >= len(a.Completed) {\n\t\ti = len(a.Completed) - 1\n\t}\n\ta.Completed = a.Completed[i+1:]\n\treturn a\n}", "func NewListerWithOrderBy(tableName string, selectedColumns []string, orderByParams OrderByParams) Lister {\n\treturn &universalLister{\n\t\ttableName: tableName,\n\t\tselectedColumns: strings.Join(selectedColumns, \", \"),\n\t\torderByParams: orderByParams,\n\t}\n}" ]
[ "0.6274745", "0.62049556", "0.61423004", "0.6103346", "0.6102245", "0.61014", "0.5916176", "0.5880672", "0.5682654", "0.56018317", "0.5592861", "0.5583494", "0.5467811", "0.5385734", "0.53754604", "0.53720546", "0.5264382", "0.5236045", "0.5232234", "0.5227033", "0.5194634", "0.51709807", "0.51409507", "0.5129478", "0.51154107", "0.50921446", "0.5080382", "0.50668895", "0.5063183", "0.5037642", "0.5035549", "0.50338984", "0.5013308", "0.5003777", "0.49976692", "0.49965423", "0.49789372", "0.49563137", "0.49488258", "0.49438217", "0.4925095", "0.4858595", "0.48576283", "0.4854639", "0.485227", "0.48309216", "0.48258796", "0.4810789", "0.48051134", "0.4793891", "0.4744446", "0.47286466", "0.472364", "0.4718492", "0.47116667", "0.47082013", "0.47044095", "0.46974814", "0.4694523", "0.469253", "0.46803412", "0.4676759", "0.46538192", "0.46447524", "0.4637687", "0.46295056", "0.46196243", "0.46171692", "0.4615248", "0.4607803", "0.4604671", "0.46000054", "0.45957914", "0.45957914", "0.45740193", "0.4568795", "0.45581588", "0.45559436", "0.45550257", "0.454595", "0.4544385", "0.45420796", "0.45397878", "0.45357388", "0.45290646", "0.45224878", "0.45221147", "0.450666", "0.4501428", "0.4501428", "0.44814524", "0.4479744", "0.44763604", "0.4468537", "0.4466705", "0.4466705", "0.44622663", "0.44621786", "0.4461314", "0.44599038" ]
0.55874276
11
Insert will insert a new uint value into the list at the appropriate location If the value was already in the list (duplicate) an error will be returned.
func (l *List) Insert(n uint16) error { newNode := Node{Value: n} if l.Root == nil { // Establish the first node in the list l.Root = &newNode return nil // Node inserted we can exit! } // If we reached here, this means we had at least one node (root node) in the list var current, next *Node current = l.Root if newNode.Value < current.Value { // newNode should be the first in the list newNode.Next = current l.Root = &newNode return nil } for current != nil { next = current.Next if newNode.Value == current.Value { return fmt.Errorf("duplicate value %v cannot be inserted", newNode.Value) } if next == nil { // our new node is the last one in the list current.Next = &newNode return nil // inserted! } if newNode.Value > current.Value && newNode.Value < next.Value { // our new node goes in between two of the old ones newNode.Next = next current.Next = &newNode return nil // inserted! } current = next // progress to the next node in the list } return nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (s *SliceOfUint) Push(item uint) *SliceOfUint {\n\ts.items = append(s.items, item)\n\treturn s\n}", "func (s *SliceOfUint32) Push(item uint32) *SliceOfUint32 {\n\ts.items = append(s.items, item)\n\treturn s\n}", "func (m *Uint64) Insert(key interface{}, val *atomics.Uint64) {\n\tm.m.Store(key, val)\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\t_, ok := this.indices[val]\n\tif ok {\n\t\treturn false\n\t}\n\tthis.indices[val] = len(this.list)\n\tthis.list = append(this.list, val)\n\treturn true\n}", "func (hm *HashMap) Insert(h common.Hash, u types.Unit) {\n\thm.lock.Lock()\n\tdefer hm.lock.Unlock()\n\thm.data[h] = u\n}", "func (s *S43_SmallSetOfIntegers) Insert(n int, done chan bool) {\n\tdefer close(done)\n\n\ti := s.SEARCH(n)\n\tif i < s.size {\n\t\tdone <- false\n\t\treturn // nothing to do\n\t}\n\t// not found, insert to the array\n\tif i == s.size && s.size < 100 {\n\t\ts.content[s.size] = n\n\t\ts.size++\n\t\tdone <- true\n\t\treturn\n\t}\n\n\tdone <- false\n\treturn\n}", "func (rs *RandomizedSet) Insert(val int) bool {\n\tif _, ok := rs.hash[val]; ok {\n\t\treturn false\n\t}\n\trs.hash[val] = len(rs.arr)\n\trs.arr = append(rs.arr, val)\n\treturn true\n}", "func InsertUint64(slice []uint64, element, position uint64) []uint64 {\n\treturn append(slice[:position], append([]uint64{element}, slice[position:]...)...)\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.M[val]; ok {\n\t\treturn false\n\t}\n\tthis.M[val] = len(this.list)\n\tthis.list = append(this.list, val)\n\treturn true\n}", "func (s *Slab) Insert(val interface{}) int {\n\tkey := s.next\n\ts.insertAt(key, val)\n\treturn key\n}", "func (s Uint64Set) Insert(items ...uint64) {\n\tfor _, item := range items {\n\t\ts[item] = empty{}\n\t}\n}", "func (il *IntList) AppendUnique(v int) {\n last := il.Last()\n for it := il.First(); it != last; it = it.Next() {\n if it.Value() == v {\n return\n }\n }\n il.Append(v)\n}", "func (u *Uint) Add(n uint) (new uint) {\n\treturn uint(atomic.AddUintptr(&u.v, uintptr(n)))\n}", "func (p *PriorityQueue) Insert(v interface{}, priority float64) {\n\t_, ok := p.lookup[v]\n\tif ok {\n\t\treturn\n\t}\n\n\tnewItem := &item{\n\t\tvalue: v,\n\t\tpriority: priority,\n\t}\n\theap.Push(p.itemHeap, newItem)\n\tp.lookup[v] = newItem\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.set[val]; !ok {\n\t\tthis.data = append(this.data, val) // 从尾部加入\n\t\tthis.set[val] = this.length\n\t\tthis.length++\n\t\treturn true\n\t}\n\treturn false\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.record[val]; ok {\n\t\treturn false\n\t}\n\tthis.nums = append(this.nums, val)\n\tthis.record[val] = len(this.nums) - 1\n\treturn true\n}", "func (h *Heap) Insert(v ...int) {\n for _, val := range(v){\n h.data = append(h.data, val)\n }\n\n n := len(h.data)\n\n i := (float64) ((n / 2) - 1)\n h.num_nodes = (int) (math.Floor(i))\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.v[val]; ok {\n\t\treturn false\n\t}\n\tthis.arr = append(this.arr, val)\n\tthis.v[val] = len(this.arr) - 1\n\tthis.rand = 0\n\treturn true\n}", "func (d *Data) Insert(items ...interface{}) {\n\tfor _, value := range items {\n\t\td.Lock()\n\t\tv, ok := d.m[value]\n\t\td.Unlock()\n\t\tif ok {\n\t\t\td.Lock()\n\t\t\td.m[value] = v + 1\n\t\t\td.Unlock()\n\t\t\tcontinue\n\t\t}\n\t\td.Lock()\n\t\td.m[value] = 1\n\t\td.Unlock()\n\t}\n}", "func AddItem(list *List, valuePtr *uint64) {\n var newAddress uint64;\n var valueAddr uint64;\n if (list.capacity == list.itemCount) { //Grow list if its capacity doesn't suffice to add another item\n newAddress = Alloc(list.capacity * 2 * list.itemSize); //Double the capacity\n CopyMem(list.baseAddress, newAddress, list.capacity * list.itemSize); //Copy old list items\n list.baseAddress = newAddress; //Set new address as base address\n list.capacity = list.capacity * 2; //Update (increase) capacity\n }\n valueAddr = ToUint64FromUint64Ptr(valuePtr);\n CopyMem(valueAddr, list.baseAddress + list.itemSize * list.itemCount, list.itemSize); //Append the new value by copying its value into the memory of the corresponding list item\n list.itemCount = list.itemCount + 1; //Update item count\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.hashMap[val]; ok {\n\t\treturn false\n\t}\n\tvar idx int\n\tif len(this.arr) == this.length {\n\t\tthis.arr = append(this.arr, val)\n\t\tidx = len(this.arr) - 1\n\t} else {\n\t\tthis.arr[this.length] = val\n\t\tidx = this.length\n\t}\n\tthis.hashMap[val] = idx\n\tthis.length++\n\treturn true\n}", "func (h *binaryHeap) Insert(val int) {\n\tidx := h.len + 1\n\tif h.idxOutOfRange(idx) { // bug\n\t\treturn\n\t}\n\th.len++\n\th.tree[idx] = h.doNegative(val)\n\th.bubbleUp(idx)\n}", "func (u *Uintptr) Add(delta uintptr) (new uintptr) {\n\treturn atomic.AddUintptr(&u.v, delta)\n}", "func (l *List) Insert(pos int, v interface{}) error {\n\tif pos == 0 {\n\t\tl.head = &Node{v, l.head}\n\t\treturn nil\n\t}\n\n\tp := l.head\n\ti := pos - 1\n\tfor i != 0 {\n\t\tif p.next == nil {\n\t\t\treturn fmt.Errorf(\"%v is not a valid position for a %v long list\", pos, pos-i)\n\t\t}\n\t\tp = p.next\n\t\ti--\n\t}\n\n\tp.next = &Node{v, p.next}\n\treturn nil\n}", "func (d *DirectAddress) Insert(key int, value interface{}) {\n\tif err := d.validateKey(key); err != nil {\n\t\treturn\n\t}\n\td.array[key-d.uMin] = value\n}", "func (recv *ValueArray) Insert(index uint32, value *Value) *ValueArray {\n\tc_index_ := (C.guint)(index)\n\n\tc_value := (*C.GValue)(C.NULL)\n\tif value != nil {\n\t\tc_value = (*C.GValue)(value.ToC())\n\t}\n\n\tretC := C.g_value_array_insert((*C.GValueArray)(recv.native), c_index_, c_value)\n\tretGo := ValueArrayNewFromC(unsafe.Pointer(retC))\n\n\treturn retGo\n}", "func AddUint32(addr *uint32, delta uint32) (new uint32)", "func (this *RandomizedSet) Insert(val int) bool {\r\n\tif _, ok := this.dict[val]; ok {return false}\r\n\tthis.dict[val] = len(this.location)\r\n\tthis.location = append(this.location, val)\r\n\treturn true\r\n}", "func (h *binaryHeap) Insert(items ...int) {\n\tif h.size+len(items) >= len(h.items) {\n\t\th.resize(len(items))\n\t}\n\tfor _, element := range items {\n\t\th.items[h.size] = element\n\t\th.size++\n\t\th.siftup(h.size - 1)\n\t}\n}", "func (t *VLIntInt) RedisInsertValue(v VTIntInt) string {\n\treturn fmt.Sprintf(\"%d:%d\", v.Key, v.Value)\n}", "func (m MultiSet) Insert(val string){\n\tvar err error\n\tindex := rand.Intn(99999)\n\tfor m[strconv.Itoa(index)] != 0 {\n\t\tindex = rand.Intn(99999)\n\t}\n\tif m[strconv.Itoa(index)], err = strconv.Atoi(val); err != nil {\n\t\tfmt.Println(err)\n\t}\n}", "func (v *Set) Insert(b []byte) bool {\n\tiRepr := new(big.Int).SetBytes(b)\n\n\tl := len(*v)\n\tif l == 0 {\n\t\t*v = append(*v, iRepr)\n\t\treturn true\n\t}\n\n\tidx, found := v.indexOf(iRepr)\n\tif found {\n\t\treturn false\n\t}\n\n\t*v = append(*v, new(big.Int))\n\tcopy((*v)[idx+1:], (*v)[idx:])\n\n\t(*v)[idx] = iRepr\n\treturn true\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\t_, ok := this.m[val]\n\tif !ok {\n\t\tthis.m[val] = len(this.data)\n\t\tthis.data = append(this.data, val)\n\t}\n\n\treturn !ok\n}", "func (t *openAddressing) Insert(key string, value interface{}) {\n\tif t.loadFactor() > 0.5 {\n\t\tt.tableDouble()\n\t}\n\tround := 0\n\tfor round != len(t.values) {\n\t\thash := t.hash(key, round)\n\t\tif t.values[hash] == nil || t.values[hash].deleted {\n\t\t\tt.values[hash] = &deletablePair{\n\t\t\t\tpair: pair{key, value},\n\t\t\t\tdeleted: false,\n\t\t\t}\n\t\t\tt.len++\n\t\t\treturn\n\t\t}\n\t\tround++\n\t}\n}", "func (h *Heap) Insert(value int) {\n\tif h.count >= h.capacity {\n\t\treturn\n\t}\n\th.array[h.count] = value\n\th.heapify(h.count, h.cmpFunc)\n\th.count++\n}", "func (s *SliceOfUint) Append(item uint) *SliceOfUint {\n\ts.items = append(s.items, item)\n\treturn s\n}", "func (this *RandomizedCollection) Insert(val int) bool {\n\tvar found bool\n\tif _, found = this.m[val]; !found {\n\t\tthis.m[val] = []int{}\n\t}\n\tthis.m[val] = append(this.m[val], len(this.a))\n\tthis.a = append(this.a, val)\n\treturn !found\n}", "func (list *List) Insert(idx int, element interface{}) error {\n\tif list.Length() < idx || idx < 0 {\n\t\treturn fmt.Errorf(\"index out of range\")\n\t}\n\tlist_ := []interface{}(*list)\n\t*list = append(list_[:idx], append([]interface{}{element}, list_[idx:]...)...)\n\treturn nil\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\t_, ok := this.set[val]\n\tif ok {\n\t\treturn false\n\t}\n\tthis.set[val] = len(this.a)\n\tthis.a = append(this.a, val)\n\treturn true\n}", "func (a *Array) Insert(index uint, v int) error {\n\t// data is full\n\tif a.Len() == uint(cap(a.data)) {\n\t\treturn errors.New(\"array is full\")\n\t}\n\n\t// index is out of range\n\tif a.IsIndexOutOfRange(index) {\n\t\treturn errors.New(\"out of index range\")\n\t}\n\n\tfor i := a.Len(); i > index; i++ {\n\t\ta.data[i] = a.data[i-1]\n\t}\n\ta.data[index] = v\n\ta.length++\n\treturn nil\n}", "func (m *MsgMemoryBuffer) Insert(value interface{}) {\n\tm.buff.Value = value\n\tm.buff = m.buff.Next()\n}", "func (u *Uint) Store(new uint) {\n\tatomic.StoreUintptr(&u.v, uintptr(new))\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.m[val]; ok {\n\t\treturn false\n\t}\n\tn := len(this.arr)\n\tthis.m[val] = n\n\tthis.arr = append(this.arr, val)\n\treturn true\n}", "func (w *WindowedMap) Put(uid UID, value interface{}) {\n\tif _, ok := w.uidMap[uid]; ok {\n\t\tw.Remove(uid)\n\t}\n\titem := &uidItem{uid, w.clock.Now().Add(w.lifetime), -1, value}\n\theap.Push(&w.uidList, item)\n\tw.uidMap[uid] = item\n}", "func (s *RandomizedSet) Insert(val int) bool {\n\tif _, ok := s.table[val]; ok {\n\t\treturn false\n\t}\n\tn := len(s.arr)\n\ts.table[val] = n\n\ts.arr = append(s.arr, val)\n\n\treturn true\n}", "func (h *MinHeap) Insert(value int) {\n\t// return if the heap is full\n\tif h.IsFull() {\n\t\treturn\n\t}\n\n\th.used++\n\th.data[h.used] = value\n\n\tfor i := h.used; i>>1 > 0 && h.data[i] < h.data[i>>1]; i >>= 1 {\n\t\th.data[i], h.data[i>>1] = h.data[i>>1], h.data[i]\n\t}\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.table[val]; !ok {\n\t\tthis.arr = append(this.arr, val)\n\t\tthis.table[val] = len(this.arr) - 1\n\n\t\treturn true\n\t} else {\n\n\t\treturn false\n\t}\n}", "func (set *RandomizedSet) Insert(val int) bool {\n\tif v, ok := set.Value[val]; ok {\n\t\tif v {\n\t\t\treturn false\n\t\t}\n\t} else {\n\t\tset.Keys = append(set.Keys, val)\n\t}\n\n\tset.Value[val] = true\n\tset.Count++\n\treturn true\n}", "func (s *SliceOfUint32) Append(item uint32) *SliceOfUint32 {\n\ts.items = append(s.items, item)\n\treturn s\n}", "func (t *VSIntInt) RedisInsertValue(v VTIntInt) string {\n\treturn fmt.Sprintf(\"%d:%d\", v.Key, v.Value)\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.valsSet[val]; !ok {\n\t\tthis.valsSet[val] = valEl{val: val, ind: len(this.vals)}\n\t\tthis.vals = append(this.vals, val)\n\t\treturn true\n\t}\n\treturn false\n}", "func (s Int64) Insert(items ...int64) Int64 {\n\treturn Int64(cast(s).Insert(items...))\n}", "func (h *MaxHeap) Insert(value int) {\n\t// return if the heap is full\n\tif h.capacity == h.used+1 {\n\t\treturn\n\t}\n\n\th.used++\n\th.data[h.used] = value\n\n\tfor i := h.used; i>>1 > 0 && h.data[i] > h.data[i>>1]; i >>= 1 {\n\t\th.data[i], h.data[i>>1] = h.data[i>>1], h.data[i]\n\t}\n\n}", "func (q *quartileIndex) Insert(n int, at int) error {\n\tif n%4 != 0 {\n\t\tpanic(\"can only extend by nibbles (multiples of 4)\")\n\t}\n\terr := q.bits.Insert(n, at)\n\tif err != nil {\n\t\treturn err\n\t}\n\tnewlen := q.bits.Len()\n\tfor i := 0; i < 3; i++ {\n\t\tq.adjust(i, n, at, (newlen * (i + 1) / 4))\n\t}\n\treturn nil\n}", "func (t *VLDblInt) RedisInsertValue(v VTDblInt) string {\n\treturn fmt.Sprintf(\"%v:%d\", v.Key, v.Value)\n}", "func (u *Uint32) Add(delta uint32) (new uint32) {\n\treturn atomic.AddUint32(&u.v, delta)\n}", "func (t *VLStrInt) RedisInsertValue(v VTStrInt) string {\n\treturn fmt.Sprintf(\"%s:%d\", v.Key, v.Value)\n}", "func (t *VLIntDbl) RedisInsertValue(v VTIntDbl) string {\n\treturn fmt.Sprintf(\"%d:%v\", v.Key, v.Value)\n}", "func (v *Int32Vec) Insert(idx int, val ...int32) {\n\tdv := *v\n\tdv = append(dv, val...)\n\tcopy(dv[idx+len(val):], dv[idx:])\n\tcopy(dv[idx:], val)\n\t*v = dv\n}", "func (this *RandomizedSet) Insert(val int) bool {\n\tif _, ok := this.m[val]; ok{\n\t\treturn false\n\t}\n\tthis.m[val] = struct{}{}\n\treturn true\n}", "func (bh* BinomialHeap) Insert(value int) {\n bh.size += 1\n\n newnode := newBinomialHeapNode(value)\n bh.insert(newnode)\n}", "func (mgr *ResTypeMgr) Insert(restype string, bytes int, time int) {\n\tv := mgr.Find(restype)\n\tif v == nil {\n\t\tmgr.List = append(mgr.List, &ResTypeInfo{\n\t\t\tResType: restype,\n\t\t\tTotalBytes: bytes,\n\t\t\tTotalTime: time,\n\t\t\tTotalNums: 1,\n\t\t})\n\n\t\treturn\n\t}\n\n\tv.TotalBytes += bytes\n\tv.TotalTime += time\n\tv.TotalNums++\n}", "func (s *Storage) ListInsert(key string, items []string) (int, error) {\n\tshard := s.getShard(key)\n\n\tshard.mutex.Lock()\n\tdefer shard.mutex.Unlock()\n\n\tif item, ok := shard.keyValues[key]; ok {\n\t\tif isExpired(item.expiration) {\n\t\t\treturn 0, nil\n\t\t}\n\n\t\tif list, ok := item.value.([]string); ok {\n\t\t\tlist = append(list, items...)\n\t\t\titem.value = list\n\t\t\tshard.keyValues[key] = item\n\t\t\treturn len(list), nil\n\t\t}\n\t\treturn 0, newErrCustom(errWrongType)\n\t}\n\treturn 0, nil\n}", "func (t *VSDblInt) RedisInsertValue(v VTDblInt) string {\n\treturn fmt.Sprintf(\"%v:%d\", v.Key, v.Value)\n}", "func (q *PriorityQueue) Insert(value interface{}) {\n\tq.nodes = append(q.nodes, value)\n\tq.upHeap(len(q.nodes) - 1)\n}", "func (s Uint64Set) Add(item uint64) {\n\ts[item] = empty{}\n}", "func (head *XorNode) Insert(val int) *XorNode {\n\tnewHeadNode := &XorNode{Val: val}\n\tptr := unsafe.Pointer(head)\n\n\tnewHeadNode.Npx = 0 ^ uintptr(ptr)\n\tnewNodePtr := unsafe.Pointer(newHeadNode)\n\n\thead.Npx = uintptr(newNodePtr) ^ uintptr(head.Npx)\n\n\treturn newHeadNode\n}", "func putUInteger(log log.T, byteArray []byte, offset int, value uint32) (err error) {\n\treturn putInteger(log, byteArray, offset, int32(value))\n}", "func (s Byte) Insert(items ...byte) Byte {\n\tfor _, item := range items {\n\t\ts[item] = Empty{}\n\t}\n\treturn s\n}", "func Insert(a []int, value int) int {\n\tvar index = 1\n\tvar b = make([]int, index)\n\tcopy(b, a[:index])\n\tb = append(b, value)\n\tb = append(b, a[index:]...)\n\n\treturn index\n}", "func (heap *MinHeap) Insert(val int) {\n\theap.elements = append(heap.elements, val)\n\theap.minHeapify(heap.getParent(heap.lastIndex()))\n}", "func (t *VSIntDbl) RedisInsertValue(v VTIntDbl) string {\n\treturn fmt.Sprintf(\"%d:%v\", v.Key, v.Value)\n}", "func (t *VSStrInt) RedisInsertValue(v VTStrInt) string {\n\treturn fmt.Sprintf(\"%s:%d\", v.Key, v.Value)\n}", "func ScanInsert(list *[]float64, done *bool, value float64, h *float64) {\n *done = false\n for !*done {\n if value <= *h {\n *list = append(*list, value)\n *done = true\n }\n }\n}", "func (h *MinHeap) Insert(value int) {\n\th.Heap = append(h.Heap, value)\n\th.siftUp(len(h.Heap) - 1)\n}", "func (rs *RandomizedSet) Insert(val int) bool {\n\tif _, ok := rs.set[val]; ok {\n\t\treturn false\n\t}\n\trs.set[val] = len(rs.keys)\n\trs.keys = append(rs.keys, val)\n\treturn true\n}", "func (h *Heap) Insert(data interface{}) {\n\tif h.size == 0 {\n\t\th.values = append([]interface{}{nil}, data)\n\t\th.size++\n\t} else {\n\t\th.values = append(h.values, data)\n\t\th.size++\n\t\th.bubbleUp()\n\t}\n}", "func (h *heap) insert(a int) error {\n\tif h.n == cap(h.q) {\n\t\treturn errors.New(\"Error: heap overflow\")\n\t}\n\n\th.q[h.n] = a\n\tbubbleUp(h, h.n)\n\th.n++\n\n\treturn nil\n}", "func (a *Array) Insert(index uint, v int) error {\n\tif a.Len() == uint(cap(a.data)) {\n\t\treturn errors.New(\"full array\")\n\t}\n\t// Call the Insert function directly, index must be an existing subscript\n \n\tif index != a.length && a.isIndexOutOfRange(index) {\n\t\treturn errors.New(\"out of index range\")\n\t}\n\tfor i := a.length; i > index; i-- {\n\t\ta.data[i] = a.data[i-1]\n\t}\n\ta.data[index] = v\n\ta.length++\n\treturn nil\n}", "func (list *PyList) Insert(index int, obj *PyObject) error {\n\tif C.PyList_Insert(list.ptr, C.long(index), obj.ptr) == -1 {\n\t\treturn ErrCouldNotInsert\n\t}\n\n\treturn nil\n}", "func (l *List) Put(v interface{} /* val */) (*El, bool) {\n\tcur := l.search(v, false, true)\n\n\tif !l.repeat && cur != &l.zero && !l.less(cur.val, v) {\n\t\tcur.val = v\n\t\treturn cur, false\n\t}\n\n\treturn l.rndEl(v), true\n}", "func (t *VLInt) RedisInsertValue(v int64) string {\n\treturn fmt.Sprintf(\"%d\", v)\n}", "func (t *VLIntStr) RedisInsertValue(v VTIntStr) string {\n\treturn fmt.Sprintf(\"%d:%s\", v.Key, v.Value)\n}", "func (bl *blackList) insert(ip string, port int) {\n\tif bl.list.Len() >= bl.maxSize {\n\t\treturn\n\t}\n\n\tbl.list.Set(bl.genKey(ip, port), &blockedItem{\n\t\tip: ip,\n\t\tport: port,\n\t\tcreateTime: time.Now(),\n\t})\n}", "func (s *SliceOfUint64) Push(item uint64) *SliceOfUint64 {\n\ts.items = append(s.items, item)\n\treturn s\n}", "func (s ids) insert(id int) ids {\n\tindex := sort.SearchInts([]int(s), id)\n\tif index == len(s) {\n\t\ts = append(s, id)\n\t} else if s[index] != id {\n\t\ts = append(s, 0)\n\t\tcopy(s[index+1:], s[index:])\n\t\ts[index] = id\n\t}\n\treturn s\n}", "func (h *hashTable) insert(val []byte) {\n\tif h.search(val) {\n\t\treturn\n\t}\n\tif len(h.bucketSlice) == 0 {\n\t\th.bucketSlice = append(h.bucketSlice, newBucket())\n\t}\n\tprobeIdx := hashFunc(val) % uint32(bucketCnt)\n\tisInserted := false\nLoop:\n\tfor _, bucket := range h.bucketSlice {\n\t\t// if the bucket is already full, skip it\n\t\tif bucket.wrapped {\n\t\t\tcontinue\n\t\t}\n\t\t// if the index is not taken yet, map it\n\t\tif bucket.data[probeIdx] == nil {\n\t\t\tbucket.data[probeIdx] = val\n\t\t\tisInserted = true\n\t\t\tbreak\n\t\t}\n\t\t// linear probe\n\t\tfor idx, elem := range bucket.data {\n\t\t\tif uint32(idx) == probeIdx {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tif elem == nil {\n\t\t\t\tbucket.data[idx] = val\n\t\t\t\tisInserted = true\n\t\t\t\tbreak Loop\n\t\t\t}\n\t\t}\n\t\tbucket.wrapped = true\n\t}\n\tif !isInserted {\n\t\tnb := newBucket()\n\t\tnb.data[probeIdx] = val\n\t\th.bucketSlice = append(h.bucketSlice, nb)\n\t}\n}", "func (r *Row) AppendUIntValue(val uint64) {\n\tif len(r.fieldValues) < len(r.fields) {\n\t\tstr := strconv.FormatUint(val, 10)\n\t\tencodedVal := StringToLenencStr([]byte(str))\n\t\tr.fieldValues = append(r.fieldValues, str)\n\t\tr.fieldValuesCache = append(r.fieldValuesCache, encodedVal)\n\t}\n}", "func (q *Q) Insert(v interface{}, prio int32) {\n\n\trng := rdrand()\n\n\t// Implementing our own lock and spinning here is probably bad,\n\t// but sync.Mutex has no TryLock()\n\t// https://github.com/golang/go/issues/6123\n\tvar c uint32\n\tvar iter int\n\tfor {\n\t\trng = xorshiftMult64(rng)\n\t\tc = reduce(uint32(rng), len(q.locks))\n\t\tgotlock := q.locks[c].trylock()\n\t\tif gotlock {\n\t\t\tbreak\n\t\t}\n\t\titer++\n\t\tif iter >= len(q.locks) {\n\t\t\truntime.Gosched()\n\t\t}\n\t}\n\n\t// insert the item into priority queue c\n\theap.Push(&q.qs[c], &pq.Item{Value: v, Priority: prio})\n\n\t// update the stored minimum\n\tatomic.StoreInt32(&q.mins[c], q.qs[c][0].Priority)\n\n\t// unlock\n\tq.locks[c].unlock()\n}", "func (r *Recovery) Add(loc Location) error {\n\tr.mtx.Lock()\n\tdefer r.mtx.Unlock()\n\n\tsh := r.shards[loc.Shard]\n\tl := len(sh.data)\n\tif diff := int(loc.Slot/8) - l; diff >= 0 {\n\t\tsh.extend(diff + 1)\n\t\tfor i := 0; i <= diff; i++ {\n\t\t\tsh.data[l+i] = 0x0\n\t\t}\n\t}\n\tsh.push(loc.Slot)\n\treturn nil\n}", "func storeUIntIndex (txn *badger.Txn, key uint64, value []byte, prefix byte) error {\r\n\r\n\tindex := make([]byte, 8)\r\n\tbinary.LittleEndian.PutUint64(index, key)\r\n\tindex = append ([]byte{prefix}, index...)\r\n\r\n\treturn txn.Set(index, value)\r\n}", "func AddUintptr(addr *uintptr, delta uintptr) (new uintptr)", "func (list *List) Insert(id ID, entity *Entity) error {\n\tif err := list.checkBoundaries(id); err != nil {\n\t\treturn err\n\t}\n\n\tif int(id) >= list.highest {\n\t\tlist.highest = int(id)\n\t}\n\n\tif list.size == 0 {\n\t\tlist.lowest = int(id)\n\t} else if int(id) < list.lowest {\n\t\tlist.lowest = int(id)\n\t}\n\n\tlist.entities[id] = entity\n\tlist.size++\n\n\treturn nil\n}", "func (l *idList) insertValue(v doc.Metadata, at *idElement) *idElement {\n\te := l.Pool.get()\n\te.Value = v\n\treturn l.insert(e, at)\n}", "func (h *_heap) Insert(values ...interface{}) {\n\tif len(values) == 1 {\n\t\theap.Push(h, values[0])\n\t} else {\n\t\tvar correctedValues []containers.Container\n\t\tvar total int\n\t\tfor _, value := range values {\n\t\t\tcorrectedValues = append(correctedValues, h.datatype.Validate(value))\n\t\t\ttotal++\n\t\t}\n\t\th.data = append(h.data, correctedValues...)\n\t\th.size = h.size + total\n\t\theap.Init(h)\n\t}\n}", "func (t *VSIntStr) RedisInsertValue(v VTIntStr) string {\n\treturn fmt.Sprintf(\"%d:%s\", v.Key, v.Value)\n}", "func (p *IntVector) Insert(i int, x int)\t{ p.Vector.Insert(i, x) }", "func (d *DeadNonceList) Insert(name *ndn.Name, nonce []byte) bool {\n\twire, err := name.Encode().Wire()\n\tif err != nil {\n\t\treturn false\n\t}\n\thash := xxhash.Sum64(wire) + uint64(binary.BigEndian.Uint32(nonce))\n\t_, exists := d.list[hash]\n\n\tif !exists {\n\t\td.expiringEntries.PushBack(hash)\n\t\tgo func() {\n\t\t\ttime.Sleep(deadNonceListLifetime)\n\t\t\td.ExpirationTimer <- true\n\t\t}()\n\t}\n\treturn exists\n}", "func (b *Builder) Insert(key []byte, val uint64) error {\n\t// ensure items are added in lexicographic order\n\tif bytes.Compare(key, b.last) < 0 {\n\t\treturn ErrOutOfOrder\n\t}\n\tif len(key) == 0 {\n\t\tb.len = 1\n\t\tb.unfinished.setRootOutput(val)\n\t\treturn nil\n\t}\n\n\tprefixLen, out := b.unfinished.findCommonPrefixAndSetOutput(key, val)\n\tb.len++\n\terr := b.compileFrom(prefixLen)\n\tif err != nil {\n\t\treturn err\n\t}\n\tb.copyLastKey(key)\n\tb.unfinished.addSuffix(key[prefixLen:], out)\n\n\treturn nil\n}", "func (set *unconfirmedBlocks) Insert(index Uint64, hash bgmcommon.Hash) {\n\t// If a new block was mined locally, shift out any old enough blocks\n\tset.Shift(index)\n\n\t// Create the new item as its own ring\n\titem := ring.New(1)\n\titemPtr.Value = &unconfirmedBlock{\n\t\tindex: index,\n\t\thash: hash,\n\t}\n\t// Set as the initial ring or append to the end\n\tset.lock.Lock()\n\tdefer set.lock.Unlock()\n\n\tif set.blocks == nil {\n\t\tset.blocks = item\n\t} else {\n\t\tset.blocks.Move(-1).Link(item)\n\t}\n\t// Display a bgmlogs for the user to notify of a new mined block unconfirmed\n\tbgmlogs.Info(\"🔨 mined potential block\", \"number\", index, \"hash\", hash)\n}" ]
[ "0.59017307", "0.5847328", "0.5820235", "0.57712674", "0.5768164", "0.5740295", "0.57116437", "0.5704485", "0.56946474", "0.56795025", "0.56700283", "0.5634675", "0.5622456", "0.5621876", "0.5619987", "0.55716723", "0.5568672", "0.55686086", "0.55290586", "0.5510469", "0.5505629", "0.5494868", "0.5483282", "0.546276", "0.54568917", "0.5447819", "0.5447577", "0.5436268", "0.5434054", "0.5430627", "0.54277503", "0.5425442", "0.5413372", "0.5400504", "0.53779703", "0.53757656", "0.53628296", "0.53622824", "0.5349081", "0.5348776", "0.5346725", "0.5345502", "0.5344918", "0.534447", "0.53439087", "0.5339019", "0.53276855", "0.53213716", "0.53174657", "0.5308731", "0.5307736", "0.5302194", "0.52991796", "0.52928954", "0.5289788", "0.5287676", "0.5287476", "0.5287173", "0.5285865", "0.528477", "0.5284547", "0.5273913", "0.52665323", "0.526517", "0.52506655", "0.52460575", "0.5244878", "0.52430594", "0.523608", "0.5230114", "0.5224223", "0.521335", "0.5210067", "0.5194146", "0.51867115", "0.5183619", "0.5178678", "0.51756597", "0.51664263", "0.5161717", "0.51614755", "0.51595765", "0.51592666", "0.51539683", "0.5147507", "0.5146618", "0.51369107", "0.51364064", "0.51287127", "0.51258767", "0.5122016", "0.5117652", "0.5114821", "0.51098067", "0.5107753", "0.5104881", "0.51046044", "0.5100547", "0.5098979", "0.50987595" ]
0.5427049
31
Length returns the number of Nodes in a given array
func (l List) Length() (length uint) { current := l.Root for current != nil { length++ current = current.Next } return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (n Nodes) Len() int", "func (n nodes) Len() int { return len(n) }", "func (id ArrayNodeID) Length() int {\n\treturn len(id.id)\n}", "func (ns Nodes) Length() int {\n\treturn len(ns)\n}", "func (n Nodes) Len() int {\n\treturn len(n)\n}", "func (s NodeSlice) Len() int {\n\treturn len(s)\n}", "func nodeLen(data interface{}) int {\n\tif data == nil {\n\t\treturn 0\n\t}\n\tswitch d := data.(type) {\n\tcase []interface{}:\n\t\treturn len(d)\n\tcase map[string]interface{}:\n\t\treturn len(d)\n\tcase string, int, float64, bool:\n\t\treturn 1\n\t}\n\treturn 0\n}", "func (n Nodes) Len() int {\n\treturn len(n.nodes)\n}", "func (a ASTNode) Len() int {\n\tif !a.IsArr() {\n\t\tpanic(ConfErr{a.pos, errors.New(\"Not an array\")})\n\t}\n\treturn len(a.val.(astArr))\n}", "func numFactoredBinaryTrees(A []int) int {\n \n}", "func (t *AreaTopology) Len() int {\n\tn := 0\n\tfor _, area := range t.areas {\n\t\tn += len(area.nodes)\n\t}\n\treturn n\n}", "func (v ResourceNodes) Len() int {\n\treturn len(v)\n}", "func (n *Node) Length() int {\n\tlen := 0\n\tfor n != nil {\n\t\tlen++\n\t\tn = n.Next\n\t}\n\treturn len\n}", "func (seq *Sequence) Len() int { return len(seq.Nodes) }", "func (n *Node) Length() (length int) {\n\t// Step 1: Go through all the nodes\n\tfor i := n; i != nil; i = i.next {\n\t\tlength++\n\t}\n\n\t// Step 2: Return the length\n\treturn length\n}", "func (g *Graph) Len() int {\n\treturn len(g.nodes)\n}", "func (n *NodeSorter) Len() int {\n\treturn len(n.nodes)\n}", "func (s *nodeSorter) Len() int {\n\treturn len(s.nodes)\n}", "func (jz *Jzon) Length() (l int, err error) {\n\tif jz.Type == JzTypeArr {\n\t\treturn len(jz.data.([]*Jzon)), nil\n\t}\n\n\tif jz.Type == JzTypeObj {\n\t\treturn len(jz.data.(map[string]*Jzon)), nil\n\t}\n\n\treturn -1, errors.New(\"expect node of type JzTypeArr or JzTypeObj\" +\n\t\t\", but the real type is \" + typeStrings[jz.Type])\n}", "func (edges Edges) Len() int { return len(edges) }", "func (s *NodeSorter) Len() int {\n\treturn len(s.nodes)\n}", "func (h nodeList) Len() int {\n\treturn len(h)\n}", "func (f *Flow) Len() int {\n\treturn len(f.nodes)\n}", "func (wl *W3CNodeList) Length() int {\n\tif wl == nil {\n\t\treturn 0\n\t}\n\treturn len(wl.nodes)\n}", "func (da *DoubleArray) NumNodes() int {\n\treturn da.numNodes\n}", "func (a *Array) Len() int64 { return a.len }", "func (n *Node) Length() (l int) {\n\tif n == nil {\n\t\treturn\n\t}\n\n\tcn := n\n\n\tfor {\n\t\tif cn.next == nil {\n\t\t\tbreak\n\t\t}\n\n\t\tl++\n\t\tcn = cn.next\n\t}\n\n\treturn\n}", "func GetNumberOfNodes() int64 {\r\n\treturn int64(len(nodesByPosition))\r\n}", "func (this *Manager) GetNodesLen() int {\n\treturn len(this.nodes)\n}", "func (m *SplitNode_Children) Len() int {\n\tif m.Dense != nil {\n\t\tn := 0\n\t\tm.ForEach(func(_ int, _ int64) bool { n++; return true })\n\t\treturn n\n\t}\n\treturn len(m.Sparse)\n}", "func sumOfDistancesInTree(N int, edges [][]int) []int {\n \n}", "func (j *JSONData) ArrayLength() *int {\n\tif arr := j.getArray(); arr != nil {\n\t\tn := len(arr)\n\t\treturn &n\n\t}\n\treturn nil\n}", "func (ms *multiSorter) Len() int {\n\treturn len(ms.Nodes)\n}", "func (node *GoValueNode) Length() (int, error) {\n\tif node.IsArray() || node.IsMap() || node.IsString() {\n\n\t\treturn node.thisValue.Len(), nil\n\t}\n\n\treturn 0, fmt.Errorf(\"this node identified as \\\"%s\\\" is not referencing an array, slice, map or string\", node.IdentifiedAs())\n}", "func (tree *RedBlack[K, V]) Len() int {\n\treturn tree.size\n}", "func (array Array) Length() int {\n\treturn len(array)\n}", "func (s *StringArray) Len() int {\n\tl := 4\n\tfor _, ss := range s.Strings {\n\t\tl += ss.Len()\n\t}\n\n\treturn l\n}", "func (n *NumFreqSlice) Len() int {\n\treturn len(*n)\n}", "func (ca ComparableArray) Len() int { return len(ca) }", "func (ne nodeEntries) Len() int { return len(ne) }", "func (t *Array) Size() int32 { return t.T.Size() * t.N }", "func (array *Array) Length() int {\n\treturn len(array.data)\n}", "func (x circle) Len() int { return len(x) }", "func (v *Array) Len() int {\n\treturn int(C.mrb_ary_len(v.state, v.value))\n}", "func (g *Group) Len() int {\n\treturn len(g.Nodes)\n}", "func lengthOfLIS(arr []int) int {\n\tvar memoo [10][10]int\n\treturn LISHelper(-1, 0, arr, memoo)\n}", "func (p NodePools) Len() int { return len(p) }", "func (v Var) Len() (uint64, error) {\n\tdims, err := v.Dims()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tn := uint64(1)\n\tfor _, d := range dims {\n\t\tlen, err := d.Len()\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t\tn *= len\n\t}\n\treturn n, nil\n}", "func (arr *Array) Len() int {\n\treturn len(arr.elems)\n}", "func (th transactionArray) Len() int { return len(th) }", "func (this *JSONArray) Length() int {\n\ttheInnerArray := this.innerArray;\n\treturn len(theInnerArray)\n}", "func (list LinkedListNode) Len() (l int) {\n\tl = 0\n\tfor curr := &list; curr != nil; curr = curr.Next {\n\t\tl++\n\t}\n\treturn\n}", "func (t *BinaryTree) Size() int { return t.count }", "func (a nodesInRequestOrder) Len() int { return len(a) }", "func (bits *BitArray) Len() int {\n\treturn bits.length\n}", "func (arr SRArr) Len() int {\n\treturn len(arr)\n}", "func (c chainSlice) Len() int { return len(c.chains) }", "func ChildrenCount(n Node) int {\n\tswitch n := n.(type) {\n\tcase nil:\n\t\treturn 0\n\tcase Value:\n\t\treturn 0\n\tcase Array:\n\t\treturn len(n)\n\tcase Object:\n\t\tc := 0\n\t\tfor _, v := range n {\n\t\t\tswitch v := v.(type) {\n\t\t\tcase Object:\n\t\t\t\tc++\n\t\t\tcase Array:\n\t\t\t\tc += len(v)\n\t\t\t}\n\t\t}\n\t\treturn c\n\t}\n\treturn 0\n}", "func (ns Slice) Len() int {\n\treturn len(ns)\n}", "func (g *Graph) Size() int { return len(g.nodes) }", "func (t *Tree) Len() int { return t.Count }", "func (arr SortedRoutes) Len() int {\n\treturn len(arr)\n}", "func (p *IntArray) Length() int {\n\treturn len(*p)\n}", "func (candidates *LookupCandidates) Len() int {\n\treturn len(candidates.Nodelist)\n}", "func (candidates *LookupCandidates) Len() int {\n\treturn len(candidates.Nodelist)\n}", "func (idx *Tree) Len() (count int) {\n\tidx.Stop()\n\tcount = int(idx.liveObjects)\n\tfor _, a := range idx.allocators {\n\t\tcount += int(a.itemCounter)\n\t}\n\tidx.Start()\n\treturn\n}", "func ListLength(head *Node) int {\n\trunner := head\n\tlength := 0\n\tfor runner != nil {\n\t\tlength++\n\t\trunner = runner.Next\n\t}\n\treturn length\n}", "func (op *OptDomainSearch) Length() int {\n\tvar length int\n\tfor _, label := range op.DomainSearch {\n\t\tlength += len(label) + 2 // add the first and the last length bytes\n\t}\n\treturn length\n}", "func (r *RingT[T]) Len() int {\n\treturn int((r.head - r.tail) & r.mask)\n}", "func (r *RawNode) Len() int {\n\treturn len(r.raw)\n}", "func main() {\n\tarray := [12][4][7][10]float64{}\n\n\tx := len(array)\n\ty := len(array[0])\n\tz := len(array[0][0])\n\tw := len(array[0][0][0])\n\tfmt.Println(\"x:\", x, \"y:\", y, \"z:\", z, \"w:\", w)\n}", "func nodeWidth(n uint) uint {\n\treturn 2*(n-1) + 1\n}", "func (t *TagNode) Len() int {\n\treturn 1 + len(t.children)\n}", "func (hat *HashedArrayTree) Size() int {\n\treturn hat.size\n}", "func (aa *Array) Length() int {\n\taa.mutex.RLock()\n\tres := aa.length\n\taa.mutex.RUnlock()\n\n\treturn res\n}", "func (e Edges) Len() int {\n\treturn len(e.edges)\n}", "func (g *ItemGraph) GetLength() int {\n\tg.Lock.RLock()\n\tglength := len(g.Nodes)\n\tg.Lock.RUnlock()\n\t// log.Printf(\"Lenght: %i\", glength)\n\treturn glength\n}", "func (nl *NodeList) Len() int {\n\treturn nl.length\n}", "func (a *Array) Len() uint {\n\treturn a.length\n}", "func (a *Array) Len() int {\n\treturn a.length\n}", "func (t *TrieNode) Len() int {\n\tt.mx.RLock()\n\tdefer t.mx.RUnlock()\n\treturn t.LenHelper()\n}", "func (t *Tree) Len() int {\n\treturn t.Count\n}", "func (a *ArrayObject) length() int {\n\treturn len(a.Elements)\n}", "func (b *BTree) Len() int {\n\tb.mtx.Lock()\n\tdefer b.mtx.Unlock()\n\n\treturn int(b.rootNod.GetLength())\n}", "func (vn *VecN) Len() float64 {\n\tif vn == nil {\n\t\treturn float64(math.NaN())\n\t}\n\tif len(vn.vec) == 0 {\n\t\treturn 0\n\t}\n\n\treturn float64(math.Sqrt(float64(vn.Dot(vn))))\n}", "func (n *NodeList) Len() int {\n\treturn len(n.items)\n}", "func Size(node *Node) int {\n\tids := make(map[NodeId]bool)\n\tqueue := []*Node{node}\n\tfor len(queue) > 0 {\n\t\tnode = queue[0]\n\t\tqueue = queue[1:]\n\t\tids[node.Id] = true\n\t\tfor _, t := range node.Transitions {\n\t\t\tqueue = append(queue, t.Child)\n\t\t}\n\t}\n\treturn len(ids)\n}", "func (pool GenePool) Len() int {\n return len(pool)\n}", "func (b *Array) Len() int {\n\treturn b.Int.BitLen()\n}", "func len(v Type) int32 {}", "func (t *Trie) Len() uint32 {\n\tif t.root == nil {\n\t\treturn 0\n\t}\n\treturn t.root.count\n}", "func TotalLength(a []string) int {\n\ts := 0\n\tfor _, n := range a {\n\t\ts += len(n)\n\t}\n\treturn s\n}", "func (g *Graph) Len() int {\n\tg.RLock()\n\tnum := len(g.transactions)\n\tg.RUnlock()\n\n\treturn num\n}", "func (ons *orderedNodeSet) size() int {\n\treturn len(ons.nodes)\n}", "func countNodes(root *TreeNode) int {\n if root == nil {\n return 0\n }\n\n queue := []*TreeNode{root}\n count := 1\n for len(queue) > 0 {\n next := []*TreeNode{}\n for _, node := range queue {\n if node.Left != nil {\n next = append(next, node.Left)\n count++\n }\n\n if node.Right != nil {\n next = append(next, node.Right)\n count++\n }\n }\n\n queue = next\n }\n\n return count\n}", "func (t *MCTS) Nodes() int { return len(t.nodes) }", "func GetCountOfActiveNodes() int64 {\r\n\treturn int64(len(nodesByPosition))\r\n}", "func (c Categorical) Len() int {\n\treturn len(c.weights)\n}", "func (c Chain) Length() int {\n\tcount := 1\n\tcurrent := c.Rest\n\tfor {\n\t\tif current == nil {\n\t\t\tbreak\n\t\t}\n\t\tcount++\n\t\tcurrent = current.Rest\n\t}\n\treturn count\n}", "func (dll *DoublyLinkedList) Length() int32 {\n\tvar count int32 = 0\n\ttemp := dll.head\n\tfor temp != nil {\n\t\tcount += 1\n\t\ttemp = temp.next\n\t}\n\treturn count\n}", "func (ls *LookupList) Len() int {\n\treturn len(ls.Nodelist)\n}" ]
[ "0.73089707", "0.7082533", "0.7048946", "0.67685896", "0.67458403", "0.66961867", "0.66868424", "0.6594114", "0.6575953", "0.65673363", "0.65142375", "0.64836943", "0.6344854", "0.63161296", "0.6291031", "0.6233559", "0.6216792", "0.62127507", "0.6209047", "0.6203045", "0.6175145", "0.60947233", "0.609045", "0.60691977", "0.605265", "0.6034903", "0.6031159", "0.6012889", "0.6004245", "0.59904546", "0.5952708", "0.5947805", "0.594741", "0.59428114", "0.5923323", "0.5922826", "0.59070086", "0.59002584", "0.58953476", "0.58771133", "0.5865469", "0.5862454", "0.5847326", "0.58392006", "0.58255136", "0.5822534", "0.58209425", "0.5812228", "0.5811616", "0.58035403", "0.5801447", "0.5791699", "0.57910824", "0.57876587", "0.57806927", "0.57657945", "0.57637584", "0.574111", "0.57237494", "0.5720998", "0.57160276", "0.57083213", "0.5698532", "0.5695091", "0.5695091", "0.5685149", "0.5674949", "0.56631476", "0.5633378", "0.5631945", "0.56316763", "0.5630081", "0.56286925", "0.5627415", "0.56186306", "0.5609142", "0.56067026", "0.5604316", "0.5585236", "0.55794865", "0.55731124", "0.5561857", "0.5558838", "0.55469257", "0.5528082", "0.5520242", "0.5513438", "0.55125177", "0.55123967", "0.5498507", "0.5496067", "0.5467764", "0.54593134", "0.54589033", "0.5448409", "0.54478425", "0.5447797", "0.5444361", "0.5440038", "0.5439757", "0.5435705" ]
0.0
-1
ValueAt returns the value at the specified position index. The list is zero indexed meaning the first value is at 0. If an out of range index is provided, an error is returned.
func (l List) ValueAt(index int) (value uint16, err error) { if l.Root == nil { return value, fmt.Errorf("Oops! Looks like the list is empty") } current := l.Root currentIndex := 0 for current != nil { if currentIndex == index { return current.Value, err } else if current.Next == nil && index > currentIndex { return value, fmt.Errorf("Provided index %v is out of bounds", index) } currentIndex++ current = current.Next } return value, err }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (obj VECTOR_TYPE) ValueAt(i int) float64 {\n if i < 0 || i >= obj.Dim() {\n panic(\"index out of bounds\")\n }\n if v, ok := obj.values[i]; ok {\n return v.GetValue()\n } else {\n return 0.0\n }\n}", "func (jz *Jzon) ValueAt(i int) (v *Jzon, err error) {\n\tif jz.Type != JzTypeArr {\n\t\treturn v, expectTypeOf(JzTypeArr, jz.Type)\n\t}\n\n\tif i < 0 || i >= len(jz.data.([]*Jzon)) {\n\t\terr = errors.New(\"index is out of bound\")\n\t\treturn\n\t}\n\n\treturn jz.data.([]*Jzon)[i], nil\n}", "func (l LinkedList) ValueAt(index int) (interface{}, error) {\n\tif index >= l.Size {\n\t\treturn \"\", fmt.Errorf(\"Index %d out of range\", index)\n\t}\n\n\tvar val interface{}\n\tnode := l.Head\n\n\tfor i := 0; i < l.Size; i++ {\n\t\tif i == index {\n\t\t\tval = node.Data\n\t\t\tbreak\n\t\t}\n\n\t\tnode = node.Next\n\t}\n\n\treturn val, nil\n}", "func (obj *SparseRealVector) ValueAt(i int) float64 {\n if i < 0 || i >= obj.Dim() {\n panic(\"index out of bounds\")\n }\n if v, ok := obj.values[i]; ok {\n return v.GetValue()\n } else {\n return 0.0\n }\n}", "func (card *Card) ValueAt(cellName string) (int, error) {\n\tcell, err := card.cellAt(cellName)\n\tif err != nil {\n\t\treturn nan, err\n\t}\n\treturn cell.value, nil\n}", "func (seq Sequence) ValueAt(period int, e expr.Expr) (val float64, found bool) {\n\tif e.IsConstant() {\n\t\tval, found, _ = e.Get(nil)\n\t\treturn\n\t}\n\tif len(seq) == 0 {\n\t\treturn 0, false\n\t}\n\tif period < 0 {\n\t\treturn 0, false\n\t}\n\treturn seq.ValueAtOffset(period*e.EncodedWidth(), e)\n}", "func (items IntSlice) Value(index int) interface{} { return items[index] }", "func (a ValueArray) Get(index int) Value {\n\tif index < 0 || index >= len(a.data) {\n\t\treturn Null()\n\t}\n\treturn a.data[index]\n}", "func (a ValueArray) TryGet(index int) (Value, bool) {\n\tif index < 0 || index >= len(a.data) {\n\t\treturn Null(), false\n\t}\n\treturn a.data[index], true\n}", "func (ba *FilterBitArray) ValueAt(i uint) byte {\n\tif i < ba.Capacity() {\n\t\treturn (*ba)[i/byteSize] & (1 << (i % byteSize))\n\t}\n\treturn 0\n}", "func (seq Sequence) ValueAtOffset(offset int, e expr.Expr) (val float64, found bool) {\n\tif e.IsConstant() {\n\t\tval, found, _ = e.Get(nil)\n\t\treturn\n\t}\n\tif len(seq) == 0 {\n\t\treturn 0, false\n\t}\n\toffset = offset + Width64bits\n\tif offset >= len(seq) {\n\t\treturn 0, false\n\t}\n\tval, wasSet, _ := e.Get(seq[offset:])\n\treturn val, wasSet\n}", "func (seq List) Value(i int) interface{} { return seq[i] }", "func (node *GoValueNode) GetArrayValueAt(index int) (val reflect.Value, err error) {\n\tif node.IsArray() {\n\t\tdefer func() {\n\t\t\tif r := recover(); r != nil {\n\t\t\t\terr = fmt.Errorf(\"recovered : %v\", r)\n\t\t\t}\n\t\t}()\n\n\t\treturn node.thisValue.Index(index), err\n\t}\n\n\treturn reflect.Value{}, fmt.Errorf(\"this node identified as \\\"%s\\\" is not referring to an array or slice\", node.IdentifiedAs())\n}", "func (g *Grid) GetValueAtIdx(idxX int, idxY int) PointValue{\n\th:= g.Header\n\n\tif(idxX<0 || idxX>=h.Nx) || (idxY<0 || idxY>=h.Ny){\n\t\t//fmt.Println(\"Out of bounds\")\n\t\treturn PointValue{Point{h.Lo1 + float64(idxX)*h.Dx, h.La1 - float64(idxY)*h.Dy }, float32(-9999)}\n\t}\n\n\tv:= g.Data[idxY * h.Nx + idxX]\n\n\tx:= h.Lo1 + float64(idxX)*h.Dx\n\ty:= h.La1 - float64(idxY)*h.Dy\n\n\treturn PointValue{Point{x, y },v}\n}", "func (v valuer) Value(i int) float64 {\n\treturn v.data[i]\n}", "func (items Float64Slice) Value(index int) interface{} { return items[index] }", "func (v *Value) Index(i int) *Value {\n\tarr := v.Array()\n\tif len(arr) < i+1 {\n\t\treturn nil\n\t}\n\treturn arr[i]\n}", "func (d Datapoints) ValueAt(n int) float64 { return d[n].Value }", "func (this *Value) Index(index int) (*Value, error) {\n\t// aliases always have priority\n\tif this.alias != nil {\n\t\tresult, ok := this.alias[strconv.Itoa(index)]\n\t\tif ok {\n\t\t\treturn result, nil\n\t\t}\n\t}\n\t// next we already parsed, used that\n\tswitch parsedValue := this.parsedValue.(type) {\n\tcase []*Value:\n\t\tif index >= 0 && index < len(parsedValue) {\n\t\t\tresult := parsedValue[index]\n\t\t\treturn result, nil\n\t\t} else {\n\t\t\t// this way it behaves consistent with jsonpointer below\n\t\t\treturn nil, &Undefined{}\n\t\t}\n\t}\n\t// finally, consult the raw bytes\n\tif this.raw != nil {\n\t\tres, err := jsonpointer.Find(this.raw, \"/\"+strconv.Itoa(index))\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tif res != nil {\n\t\t\treturn NewValueFromBytes(res), nil\n\t\t}\n\t}\n\treturn nil, &Undefined{}\n}", "func valueAt(row []string, position int) *float64 {\n\tif row == nil {\n\t\treturn nil\n\t}\n\n\treturn stringToFloat(row[position])\n}", "func (s *VectorImplSlice) Get(i int) Value {\n\tif i < 0 || s.start+i >= s.stop {\n\t\tpanic(\"Index out of bounds\")\n\t}\n\n\treturn s.vector.Get(s.start + i)\n}", "func (l *Int32) Get(index int) int32 {\n\treturn l.values[index]\n}", "func (v Value) Index(i int) Value {\n\tpanic(message)\n}", "func (m *MockValues) ValueAt(arg0 int) float64 {\n\tm.ctrl.T.Helper()\n\tret := m.ctrl.Call(m, \"ValueAt\", arg0)\n\tret0, _ := ret[0].(float64)\n\treturn ret0\n}", "func (v variable) At(index int) interface{} {\n\tm, ok := v.store.Get(v.Name)\n\tif !ok {\n\t\treturn nil\n\t}\n\tif intArray, ok := m.([]interface{}); ok {\n\t\tif index < 1 || index > len(intArray) {\n\t\t\treturn nil\n\t\t}\n\t\treturn intArray[index-1]\n\t}\n\tif indexable, ok := m.(core.Indexable); ok {\n\t\treturn indexable.At(index)\n\t}\n\tif sequenceable, ok := m.(core.Sequenceable); ok {\n\t\treturn core.BuildSequence(sequenceable.S().At(index))\n\t}\n\treturn nil\n}", "func (p Polynom) ValueAt(x0 *big.Int) *big.Int {\n\tval := big.NewInt(0)\n\tfor i := len(p.coeff) - 1; i >= 0; i-- {\n\t\tval.Mul(val, x0)\n\t\tval.Add(val, p.coeff[i])\n\t\tval.Mod(val, p.mod)\n\t}\n\treturn val\n}", "func (v *VectorImpl) Get(i int) Value {\n\tif i < 0 || uint(i) >= v.len {\n\t\tpanic(\"Index out of bounds\")\n\t}\n\n\treturn v.sliceFor(uint(i))[i&shiftBitMask]\n}", "func (q *Deque) At(idx int) interface{} {\n\tif idx >= len(q.values) {\n\t\treturn nil\n\t}\n\tactualIdx := idx\n\tif q.front != 0 {\n\t\tactualIdx = (idx + q.front) % cap(q.values)\n\t}\n\treturn q.values[actualIdx]\n}", "func (v Vector) At(idx int) float64 {\n\treturn v[idx]\n}", "func (node *GoValueNode) GetMapValueAt(index reflect.Value) (reflect.Value, error) {\n\tif node.IsMap() {\n\t\tretVal := node.thisValue.MapIndex(index)\n\t\tif retVal.IsValid() {\n\n\t\t\treturn retVal, nil\n\t\t}\n\n\t\treturn reflect.Value{}, fmt.Errorf(\"this node identified as \\\"%s\\\" have no selector with specified key\", node.IdentifiedAs())\n\t}\n\n\treturn reflect.Value{}, fmt.Errorf(\"this node identified as \\\"%s\\\" is not referencing a map\", node.IdentifiedAs())\n}", "func (v *V) At(i int) float64 {\n\tif i < 0 || i >= v.Dim() {\n\t\tpanic(ErrIndex)\n\t}\n\treturn v.Data[i]\n}", "func getValueAt(index int, colName string, header *[]string, content *[][]string) (string, error) {\n\n\tif index == -1 {\n\t\treturn \"0.0\", nil\n\t}\n\tindexCol := findIndex(colName, header)\n\tif indexCol < 0 {\n\t\treturn \"\", fmt.Errorf(\"column %s not found\", colName)\n\t}\n\tval := (*content)[index][indexCol]\n\tvalF, err := strconv.ParseFloat(val, 64)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn fmt.Sprintf(\"%.2f\", valF), nil\n}", "func (arr *ArrayList) Get(index uint32) ItemType {\n if index < arr.length {\n return arr.data[index]\n }\n panic(\"out of bounds\")\n}", "func (v *Array) Get(idx int) (*MrbValue, error) {\n\tresult := C.mrb_ary_entry(v.value, C.mrb_int(idx))\n\n\tval := newValue(v.state, result)\n\tif val.Type() == TypeNil {\n\t\tval = nil\n\t}\n\n\treturn val, nil\n}", "func (enum Enum) Value(findIndex int) string {\n\tfor _, item := range enum.items {\n\t\tif item.index == findIndex {\n\t\t\treturn item.value\n\t\t}\n\t}\n\treturn \"ID not found\"\n}", "func (r *SlidingWindow) Get(index int) (interface{}, bool) {\n\tindex -= r.base\n\tif index < 0 || index >= r.Capacity() {return nil, false}\n\tindex = r.normalize(index + r.start)\n\tvalue := r.values[index]\n\treturn value.value, value.present\n}", "func (as *Addresses) Get(index int) (*Address, error) {\n\tif index < 0 || index >= len(as.values) {\n\t\treturn nil, errors.New(\"get: index out of range\")\n\t}\n\treturn &Address{as.values[index]}, nil\n}", "func (d *Dictionary) GetValueIndex(i int) int {\n\tindiceData := d.data.buffers[1].Bytes()\n\t// we know the value is non-negative per the spec, so\n\t// we can use the unsigned value regardless.\n\tswitch d.indices.DataType().ID() {\n\tcase arrow.UINT8, arrow.INT8:\n\t\treturn int(uint8(indiceData[d.data.offset+i]))\n\tcase arrow.UINT16, arrow.INT16:\n\t\treturn int(arrow.Uint16Traits.CastFromBytes(indiceData)[d.data.offset+i])\n\tcase arrow.UINT32, arrow.INT32:\n\t\tidx := arrow.Uint32Traits.CastFromBytes(indiceData)[d.data.offset+i]\n\t\tdebug.Assert(bits.UintSize == 64 || idx <= math.MaxInt32, \"arrow/dictionary: truncation of index value\")\n\t\treturn int(idx)\n\tcase arrow.UINT64, arrow.INT64:\n\t\tidx := arrow.Uint64Traits.CastFromBytes(indiceData)[d.data.offset+i]\n\t\tdebug.Assert((bits.UintSize == 32 && idx <= math.MaxInt32) || (bits.UintSize == 64 && idx <= math.MaxInt64), \"arrow/dictionary: truncation of index value\")\n\t\treturn int(idx)\n\t}\n\tdebug.Assert(false, \"unreachable dictionary index\")\n\treturn -1\n}", "func valueAtbit(num int, bit int) int {\n\treturn -1\n}", "func GetVal(ints []int, opcodeandparammodes []int, instructionpointer int, position int) int {\n\tmode := opcodeandparammodes[position]\n\tif mode == 0 { //position mode\n\t\treturn ints[ints[instructionpointer+position]]\n\t} else if mode == 1 { //immediate mode\n\t\treturn ints[instructionpointer+position]\n\t}\n\t//todo decide how to handle unsupported modes\n\treturn -1\n}", "func (s *SliceInt) At(index int) (int, error) {\n\tif s.data == nil || len(s.data) == 0 {\n\t\treturn 0, errors.New(\"SliceInt does not contain any elements\")\n\t}\n\n\tif index >= len(s.data) || index <= 0 {\n\t\treturn 0, fmt.Errorf(\"index %d outside the range of SliceInt\", index)\n\t}\n\n\treturn s.data[index], nil\n}", "func (m *matrix) GetValue(i, j int) (f float64, err error) {\n\tif ok, err := m.checkBounds(i, j); !ok {\n\t\treturn f, err\n\t}\n\treturn m.matrix[i][j], err\n}", "func (array *Array) GetAtIndex(index int) interface{} {\n\treturn array.data[index]\n}", "func (iter *SliceIterator) Value() interface{} {\n\treturn iter.s.At(iter.position)\n}", "func (aa *Array) Get(idx int) interface{} {\n\t// do not lock if not needed\n\tif idx < 0 || idx >= aa.length {\n\t\treturn nil\n\t}\n\n\taa.mutex.RLock()\n\tres := aa.items[idx]\n\taa.mutex.RUnlock()\n\treturn res\n}", "func (m *Mat) Value(index int) (val float32) {\n\tcursor := 0\n\tfor i := 0; i < len(m.W); i++ {\n\t\tfor f := 0; f < 4; f++ {\n\t\t\tif cursor >= index {\n\t\t\t\tval = m.W[i][f]\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tcursor++\n\t\t}\n\t}\n\treturn val\n}", "func (r *sparseRow) Value(feature int) float64 {\n\ti := search(r.ind, feature)\n\tif i >= 0 {\n\t\treturn r.val[i]\n\t}\n\treturn 0\n}", "func (list *List) Get(index int) (interface{}, bool) {\n\n\tif !list.withinRange(index) {\n\t\treturn nil, false\n\t}\n\n\telement := list.first\n\tfor e := 0; e != index; e, element = e+1, element.next {\n\t}\n\n\treturn element.value, true\n}", "func (vector *Vector) ElementAt(index int) interface{} {\n\treturn (*vector)[index]\n}", "func (a *DynamicArray) IndexAt(index int) (int, error) {\n\tif index < 0 || index >= a.len {\n\t\treturn InvalidValue, ErrArrIndexOutOfBound\n\t}\n\treturn a.data[index], nil\n}", "func (cache *Cache) GetAt(seqno uint16, index uint16, result []byte) uint16 {\n\tcache.mu.Lock()\n\tdefer cache.mu.Unlock()\n\n\tif int(index) >= len(cache.entries) {\n\t\treturn 0\n\t}\n\tif cache.entries[index].seqno != seqno {\n\t\treturn 0\n\t}\n\treturn uint16(copy(\n\t\tresult[:cache.entries[index].length()],\n\t\tcache.entries[index].buf[:]),\n\t)\n}", "func (uni *Uniform1fv) Get(pos int, v float32) float32 {\n\n\treturn uni.v[pos]\n}", "func (a *BooleanArchive) Value(entryIndex int) bool {\n\tif entryIndex >= a.size {\n\t\toutOfBoundsError := errors.New(\"index out of range\")\n\t\tpanic(outOfBoundsError)\n\t}\n\n\treturn a.deriveDetail(entryIndex).value\n}", "func (list elemlist) At(index int) interface{} {\n\tvar foundItem interface{}\n\n\tif index < len(list.elements) {\n\t\tfoundItem = list.elements[index]\n\t}\n\n\treturn foundItem\n}", "func (l *list) ElementAt(index int) interface{} {\n\treturn l.elements[index]\n}", "func (s VectOp) IndexOfVal(value float64) int {\n\treturn s.IndexOf(fs.EQ(value))\n}", "func (list *ArrayList) Get(index int) (interface{}, bool) {\n\n\tif !list.withinRange(index) {\n\t\treturn nil, false\n\t}\n\n\treturn list.elements[index], true\n}", "func (l *List) Get(index int) interface{} {\n\tif index < 0 || index >= l.len {\n\t\treturn -1\n\t}\n\n\tvar cur *Node\n\tif index < l.len/2 { // 从head开始查询\n\t\tcur = l.head\n\t\t// 空节点head的index是0\n\t\tfor i := 0; i < index; i++ {\n\t\t\tcur = cur.next\n\t\t}\n\t} else { // 从tail开始查询\n\t\tcur = l.tail\n\t\tfor i := l.len + 1; i > index; i-- {\n\t\t\tcur = cur.prev\n\t\t}\n\t}\n\n\treturn cur.value\n}", "func (r *Resultset) GetValue(row, column int) (interface{}, error) {\n\tif row >= len(r.Values) || row < 0 {\n\t\treturn nil, fmt.Errorf(\"invalid row index %d\", row)\n\t}\n\n\tif column >= len(r.Fields) || column < 0 {\n\t\treturn nil, fmt.Errorf(\"invalid column index %d\", column)\n\t}\n\n\treturn r.Values[row][column], nil\n}", "func(list *List) Get(index int) (interface{}, bool) {\n\tif !list.withinRange(index){\n\t\treturn nil, false\n\t}\n\n\treturn list.elements[index], true\n}", "func (da *cedar) get(key []byte, from, pos int) *int {\n\tfor ; pos < len(key); pos++ {\n\t\tif value := da.Array[from].Value; value >= 0 && value != ValueLimit {\n\t\t\tto := da.follow(from, 0)\n\t\t\tda.Array[to].Value = value\n\t\t}\n\t\tfrom = da.follow(from, key[pos])\n\t}\n\tto := from\n\tif da.Array[from].Value < 0 {\n\t\tto = da.follow(from, 0)\n\t}\n\treturn &da.Array[to].Value\n}", "func (m NumSeriesDistribution) Get(index int) *NumSeries {\n\tif index > -1 {\n\t\tif s, ok := m[index]; ok {\n\t\t\treturn s\n\t\t}\n\t}\n\treturn nil\n}", "func (v *Value) NumberAt(i int) (float64, error) {\n\tif v.kind != kindArray {\n\t\treturn 0.0, errors.New(\"JSON value is not an array\")\n\t}\n\tif i >= len(v.arrayContent) {\n\t\treturn 0.0, errors.New(\"Index is out of bounds of array\")\n\t}\n\tvalue, err := v.arrayContent[i].Number()\n\tif err != nil {\n\t\treturn 0.0, err\n\t}\n\treturn value, nil\n}", "func (g Grid) GetValueAt(p *Point) float32 {\n\n\tif(p.X<g.Header.Lo1 || p.X>g.Header.Lo2){\n\t\treturn float32(-9999);\n\t}\n\n\tif(p.Y>g.Header.La1 || p.Y<g.Header.La2){\n\t\treturn float32(-9999);\n\t}\n\n\tidxX := int(((p.X - g.Header.Lo1) / g.Width()) * float64(g.Header.Nx-1))\n\tidxY := int(((g.Header.La1 - p.Y) / g.Height()) * float64(g.Header.Ny-1))\n\n\tul := g.GetValueAtIdx(idxX, idxY)\n\tur := g.GetValueAtIdx(idxX+1, idxY)\n\tll := g.GetValueAtIdx(idxX, idxY+1)\n\tlr := g.GetValueAtIdx(idxX+1, idxY+1)\n\n\tv:=BilinearInterpolation(&ll,&ul,&lr,&ur,p)\n\n\treturn float32(v)\n}", "func (x IntSlice) Get(i int) interface{} {return x[i]}", "func (cc *CounterControl) ReadValueAtIndex(index int64) (*CounterData, error) {\n\tentity := cc.counter.ReadValueAtIndexRequest(index)\n\tentityList := []*p4V1.Entity{entity}\n\n\tres, err := cc.control.Client.ReadEntitiesSync(entityList)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif len(res) == 0 {\n\t\treturn nil, errors.New(\"No counter entries found at given index\")\n\t}\n\n\tresult := getCounterData(res[0])\n\treturn &result, nil\n}", "func (entry persistedEntry) Value(index int64) (*value, error) {\n\tif entry.DataLen > modules.RegistryDataSize {\n\t\terr := errors.New(\"Value: entry has a too big data len\")\n\t\tbuild.Critical(err)\n\t\treturn nil, err\n\t}\n\tspk, err := newSiaPublicKey(entry.Key)\n\tif err != nil {\n\t\treturn nil, errors.AddContext(err, \"Value: failed to convert compressed key to SiaPublicKey\")\n\t}\n\tswitch entry.Type {\n\tcase modules.RegistryTypeInvalid:\n\t\treturn nil, modules.ErrInvalidRegistryEntryType\n\tcase modules.RegistryTypeWithPubkey:\n\tcase modules.RegistryTypeWithoutPubkey:\n\tdefault:\n\t\treturn nil, modules.ErrInvalidRegistryEntryType\n\t}\n\treturn &value{\n\t\tentryType: entry.Type,\n\t\tkey: spk,\n\t\ttweak: entry.Tweak,\n\t\texpiry: types.BlockHeight(entry.Expiry),\n\t\tdata: entry.Data[:entry.DataLen],\n\t\trevision: entry.Revision,\n\t\tsignature: entry.Signature,\n\t\tstaticIndex: index,\n\t}, nil\n}", "func (s Serializer) indexValue(buf cmpbin.WriteableBytesBuffer, v any) (err error) {\n\tswitch t := v.(type) {\n\tcase nil:\n\tcase bool:\n\t\tb := byte(0)\n\t\tif t {\n\t\t\tb = 1\n\t\t}\n\t\terr = buf.WriteByte(b)\n\tcase int64:\n\t\t_, err = cmpbin.WriteInt(buf, t)\n\tcase float64:\n\t\t_, err = cmpbin.WriteFloat64(buf, t)\n\tcase string:\n\t\t_, err = cmpbin.WriteString(buf, t)\n\tcase []byte:\n\t\t_, err = cmpbin.WriteBytes(buf, t)\n\tcase GeoPoint:\n\t\terr = s.GeoPoint(buf, t)\n\tcase PropertyMap:\n\t\terr = s.PropertyMap(buf, t)\n\tcase *Key:\n\t\terr = s.Key(buf, t)\n\n\tdefault:\n\t\terr = fmt.Errorf(\"unsupported type: %T\", t)\n\t}\n\treturn\n}", "func (t Tuple) At(idx int) float64 {\n\treturn t[idx]\n}", "func (it *ScriptingReturnValueIterator) Value() (interface{}, variables.VariableType) {\n\tif it.inx < len(it.values.values) {\n\t\ta := it.values.values[it.inx]\n\t\tif a == nil {\n\t\t\treturn nil, variables.Undefined\n\t\t} else if v, ok := isNumericConstant(a); ok {\n\t\t\treturn v, variables.NumericType\n\t\t} else {\n\t\t\tT().Errorf(\"not yet implemented: type for %v\", a)\n\t\t}\n\t}\n\treturn nil, variables.Undefined\n}", "func (v *Value) StringAt(i int) (string, error) {\n\tif v.kind != kindArray {\n\t\treturn \"\", errors.New(\"JSON value is not an array\")\n\t}\n\tif i >= len(v.arrayContent) {\n\t\treturn \"\", errors.New(\"Index is out of bounds of array\")\n\t}\n\tvalue, err := v.arrayContent[i].String()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn value, nil\n}", "func (s Series) Val(i int) interface{} {\n\treturn s.elements.Elem(i).Val()\n}", "func (s *IntSliceSetting) Value() interface{} {\n\treturn *s.IntSliceValue\n}", "func (ps *PrimeStore) GetByIndex(nth uint64) (n uint64) {\n\tdefer Tracer(NewTrace(\"GetByIndex\"))\n\n\tn = 0\n\tif nth < ps.base || nth >= (ps.base+ps.count) {\n\t\tlog.Print(\"out of range.\", nth, \" \", ps)\n\t\treturn\n\t}\n\n\tn = ps.index[nth-ps.base]\n\treturn\n}", "func (mat *Matrix) GetValue(m, n int) float64 {\n\treturn mat.values[m][n]\n}", "func (t *FenwickTreeSimple) Get(index int) int {\n\treturn t.QueryRange(index, index)\n}", "func (p Point) At(idx int) float64 {\n\treturn p[idx]\n}", "func (r *Result) Getx(index int) interface{} {\n\tif index < 0 || index >= len(r.val.columns) {\n\t\tpanic(ErrorColumnNotFound{At: \"Getx\", Index: index})\n\t}\n\tbv := r.val.buffer[index]\n\treturn bv.Value\n}", "func (o *KeyValueOrdered) Value(key Key) (val *Value, idx int, ok bool) {\n\tidx, k := o.m[key]\n\tif !k {\n\t\treturn\n\t}\n\treturn &o.s[idx].V, idx, true\n}", "func (a *Array) Get(index int) interface{} {\n\treturn a.Data[index]\n}", "func (t *Tensor) ItemAt(pos ...int) *Tensor {\n\tif !t.idx.Validate(pos) {\n\t\tpanic(errorc.New(\"invalid position %v for %v\", pos, t.idx))\n\t}\n\n\treturn &Tensor{\n\t\tidx: t.idx.Scalar(pos),\n\t\tbuf: t.buf,\n\t}\n}", "func (cm CMap) Index(keyValue interface{}) interface{} {\n\tentry := cm.FindBy(keyValue, Equals)\n\n\tif entry == nil {\n\t\treturn nil\n\t}\n\treturn entry.Value\n}", "func (bc *BinaryCell) GetValue() float64 {\n\tif bc.free {\n\t\treturn 1\n\t}\n\treturn -1\n}", "func (v *Vector) Get(i int) float64 {\n\tswitch i {\n\tcase 0:\n\t\treturn v.X\n\tcase 1:\n\t\treturn v.Y\n\tcase 2:\n\t\treturn v.Z\n\t}\n\treturn 0.0\n}", "func (sll *SingleLinkedList) Get(index int) interface{} {\n\treturn sll.getNode(index).value\n}", "func (cs ContinuousSeries) GetValue(index int) (float64, float64) {\n\treturn cs.XValues[index], cs.YValues[index]\n}", "func (a *Args) Get(i int) string {\n\tif i < 0 || i >= a.Size() {\n\t\tfmt.Fprintf(os.Stderr, \"Args: Index out of range: %d [0, %d)\\n\", i, a.Size())\n\t\tos.Exit(1)\n\t}\n\n\treturn a.vals[i]\n}", "func (d *DynamicArr) Get(index int) (interface{}, error) {\n\tif index < 0 || index > (d.capacity-1) {\n\t\treturn nil, errors.New(\"Index out of range\")\n\t}\n\treturn d.array[index], nil\n}", "func (list *ArrayList) Get(index int) (interface{}, bool) {\n\tif !list.boundCheck(index) {\n\t\treturn nil, false\n\t}\n\n\treturn list.elements[index], true\n}", "func (i *IndexIterator) Value() int64 { return i.current.lpos }", "func (k *KVItem) Value(attribute string) (interface{}, error) {\n\tcAttribute := C.CString(attribute)\n\tdefer C.free(unsafe.Pointer(cAttribute))\n\tvar cValue unsafe.Pointer\n\tvar valueSize C.uint64_t\n\tvar valueType C.tiledb_datatype_t\n\tret := C.tiledb_kv_item_get_value(k.context.tiledbContext, k.tiledbKVItem, cAttribute, &cValue, &valueType, &valueSize)\n\n\tif ret != C.TILEDB_OK {\n\t\treturn nil, fmt.Errorf(\"Error getting value for KVItem: %s\", k.context.LastError())\n\t}\n\n\tswitch Datatype(valueType) {\n\tcase TILEDB_INT8:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_int8_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.int8_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]int8, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = int8(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int8(*(*C.int8_t)(cValue)), nil\n\tcase TILEDB_INT16:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_int16_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.int16_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]int16, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = int16(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int16(*(*C.int16_t)(cValue)), nil\n\tcase TILEDB_INT32:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_int32_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.int32_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]int32, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = int32(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int32(*(*C.int32_t)(cValue)), nil\n\tcase TILEDB_INT64:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_int64_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.int64_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]int64, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = int64(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int64(*(*C.int64_t)(cValue)), nil\n\tcase TILEDB_UINT8:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_uint8_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.uint8_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]uint8, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = uint8(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int8(*(*C.uint8_t)(cValue)), nil\n\tcase TILEDB_UINT16:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_uint16_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.uint16_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]uint16, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = uint16(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int16(*(*C.uint16_t)(cValue)), nil\n\tcase TILEDB_UINT32:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_uint32_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.uint32_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]uint32, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = uint32(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int32(*(*C.uint32_t)(cValue)), nil\n\tcase TILEDB_UINT64:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_uint64_t\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.uint64_t)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]uint64, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = uint64(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn int64(*(*C.uint64_t)(cValue)), nil\n\tcase TILEDB_FLOAT32:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_float\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.float)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]float32, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = float32(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn float32(*(*C.float)(cValue)), nil\n\tcase TILEDB_FLOAT64:\n\t\t// If the value size is greater than the size of a single value in bytes it is an array\n\t\telements := int(valueSize) / C.sizeof_double\n\t\tif elements > 1 {\n\t\t\ttmpslice := (*[1 << 30]C.double)(unsafe.Pointer(cValue))[:elements:elements]\n\t\t\tretSlice := make([]float64, elements)\n\t\t\tfor i, s := range tmpslice {\n\t\t\t\tretSlice[i] = float64(s)\n\t\t\t}\n\t\t\treturn retSlice, nil\n\t\t}\n\t\treturn float64(*(*C.double)(cValue)), nil\n\tcase TILEDB_CHAR:\n\t\telements := int(valueSize) / C.sizeof_char\n\t\treturn C.GoStringN((*C.char)(cValue), C.int(elements)), nil\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Unsupported tiledb value type: %v\", valueType)\n\t}\n\n\treturn nil, fmt.Errorf(\"Error getting value for KVItem\")\n}", "func (seq Sequence) ValueAtTime(t time.Time, e expr.Expr, resolution time.Duration) (val float64, found bool) {\n\tif e.IsConstant() {\n\t\tval, found, _ = e.Get(nil)\n\t\treturn\n\t}\n\tif len(seq) == 0 {\n\t\treturn 0, false\n\t}\n\tuntil := seq.Until()\n\tt = RoundTimeUntilUp(t, resolution, until)\n\tif t.After(until) {\n\t\treturn 0, false\n\t}\n\tperiod := int(until.Sub(t) / resolution)\n\treturn seq.ValueAt(period, e)\n}", "func (this *MyLinkedList) Get(index int) int {\n\tif index < 0 || index > this.size-1 {\n\t\treturn -1\n\t}\n\treturn this.listMap[index].Val\n}", "func (list *ArrayList[T]) Get(index int) (T, bool) {\n\tif index < 0 || index >= list.Size() {\n\t\tvar zero T\n\t\treturn zero, false\n\t}\n\treturn list.elems[index], true\n}", "func GetItem(slice []int, index int) int {\n\tif checkOutOfBounds(index, len(slice)) {\n\t\treturn -1\n\t}\n\treturn slice[index]\n}", "func (args *Args) at(index int) *Arg {\n\tif len(args.items) > index && index >= 0 {\n\t\treturn args.items[index]\n\t}\n\treturn nil\n}", "func (row Row) GetValue(column Column) interface{} {\n\tindex, ok := row.columnOrder[column]\n\tif !ok {\n\t\tlog.Printf(\"WARNING: %v is not found in %v\", column, row.columnOrder)\n\t}\n\treturn row.values[index]\n}", "func (s *SequenceItemValue) GetValue() interface{} { return s.elements }", "func (uni *UniformMatrix3f) Get(pos int) float32 {\n\n\treturn uni.v[pos]\n}", "func (cache *cache) GetValue(x int) (int, error) {\n\tfetchItem, err := cache.client.Get(strconv.Itoa(x))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tval, err := strconv.Atoi(string(fetchItem.Value))\n\treturn val, err\n}" ]
[ "0.75467837", "0.75191677", "0.7387684", "0.7204418", "0.6995977", "0.69785494", "0.6809637", "0.67981523", "0.659905", "0.6556795", "0.6541746", "0.648381", "0.6471923", "0.6410484", "0.6374607", "0.6345296", "0.63356483", "0.6318736", "0.63073444", "0.6300226", "0.6079921", "0.6041721", "0.60295594", "0.6012001", "0.60000026", "0.59945834", "0.59623075", "0.59379673", "0.5931072", "0.5919619", "0.5917956", "0.59092396", "0.59033996", "0.59032196", "0.58539945", "0.5852561", "0.5828481", "0.5827487", "0.58237827", "0.57510227", "0.57506883", "0.572387", "0.57122296", "0.5710157", "0.57050055", "0.5697055", "0.56905735", "0.56793123", "0.5679054", "0.5675515", "0.5656942", "0.56516415", "0.56402904", "0.56374806", "0.5632598", "0.5628831", "0.5626855", "0.56248045", "0.56192803", "0.5617153", "0.560704", "0.56052375", "0.56011736", "0.55921996", "0.5591341", "0.5590827", "0.5562174", "0.55584687", "0.5551486", "0.55433565", "0.55363613", "0.55306", "0.5516189", "0.5515836", "0.5507413", "0.5493445", "0.54927725", "0.54795104", "0.5476082", "0.5467083", "0.5455303", "0.54486114", "0.5442231", "0.54385746", "0.543297", "0.5422893", "0.5415431", "0.54059416", "0.5400061", "0.53892356", "0.53835315", "0.53803813", "0.5357302", "0.53533053", "0.5339397", "0.5331831", "0.53271383", "0.5323733", "0.53128827", "0.53082645" ]
0.7748144
0
Contain checks whether or not the value is in a given list
func (l List) Contain(val uint16) (c bool) { current := l.Root for current != nil { if current.Value == val { return true } current = current.Next } return }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Contains(list []string, val string) bool {\r\n\r\n\tfor _, el := range list {\r\n\r\n\t\tif el == val {\r\n\t\t\treturn true\r\n\t\t}\r\n\r\n\t}\r\n\r\n\treturn false\r\n}", "func Contains(list interface{}, value interface{}) bool {\n\tarr := reflect.ValueOf(list)\n\tval := reflect.ValueOf(value)\n\tif arr.Kind() != reflect.Slice {\n\t\tpanic(\"invalid data-type, occurred in sxutil package\")\n\t}\n\tif val.Kind() == reflect.Slice {\n\t\tpanic(\"invalid data-type, occurred in sxutil package\")\n\t}\n\n\tfor i := 0; i < arr.Len(); i++ {\n\t\tif arr.Index(i).Interface() == value {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(list []int, val int) bool {\n\tfor _, ele := range list {\n\t\tif ele == val {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(list []string, s string) bool {\n\tfor _, v := range list {\n\t\tif v == s {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (list *List) Contains(values ...interface{}) bool {\n\n\tif len(values) == 0 {\n\t\treturn true\n\t}\n\tif list.size == 0 {\n\t\treturn false\n\t}\n\tfor _, value := range values {\n\t\tfound := false\n\t\tfor element := list.first; element != nil; element = element.next {\n\t\t\tif element.value == value {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func Contains(list []interface{}, elem interface{}) bool {\n\tfor _, t := range list {\n\t\tif t == elem {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(list []string, wanted string) bool {\n\tfor _, found := range list {\n\t\tif found == wanted {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(list []string, target string) bool {\n\tfor _, s := range list {\n\t\tif s == target {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(list []string, a string) bool {\n\tfor _, b := range list {\n\t\tif b == a {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func InList(a string, list []string) bool {\n\tcheck := 0\n\tfor _, b := range list {\n\t\tif b == a {\n\t\t\tcheck += 1\n\t\t}\n\t}\n\tif check != 0 {\n\t\treturn true\n\t}\n\treturn false\n}", "func Contain(list []string, str string) bool {\n\tfor _, s := range list {\n\t\tif s == str {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(item string, list []string) bool {\n\tfor _, i := range list {\n\t\tif i == item {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (list *DoublyLinkedList) Contains(values ...interface{}) bool {\n\n\tif len(values) == 0 {\n\t\treturn true\n\t}\n\tif list.size == 0 {\n\t\treturn false\n\t}\n\tfor _, value := range values {\n\t\tif !list.contain(value) {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func isValuePresentInTheList(strArr []string, str string) bool {\n\tfor _, s := range strArr {\n\t\tif strings.Contains(s, str) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (exp *ListExpression) Contains(val int) bool {\n\tfirstV := exp.values[0]\n\tlastV := exp.values[len(exp.values)-1]\n\tif val < firstV || val > lastV {\n\t\treturn false\n\t}\n\tfor _, v := range exp.values {\n\t\tif val == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func ListContains(haystack interface{}, needle interface{}) bool {\n\tValueIface := reflect.ValueOf(haystack)\n\t// see if it's a pointer\n\tif ValueIface.Type().Kind() != reflect.Ptr {\n\t\t// convert it to a pointer\n\t\tValueIface = reflect.New(reflect.TypeOf(haystack))\n\t}\n\n\t// Ensure that the passed interface has field Items\n\titems := ValueIface.Elem().FieldByName(\"Items\")\n\tif !items.IsValid() {\n\t\treturn false\n\t}\n\n\t// Ensure that the Items field is a slice and loop over it\n\tswitch reflect.TypeOf(items).Kind() {\n\tcase reflect.Slice:\n\t\tfor i := 0; i < items.Len(); i++ {\n\t\t\tif reflect.DeepEqual(needle, items.Index(i)) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\n\treturn false\n}", "func Contains(list interface{}, elem interface{}) bool {\n\tvalue := reflect.ValueOf(list)\n\tif value.Kind() != reflect.Slice {\n\t\tlogger.Errorf(\"Input type is not an array or slice type: %v, kind:%s\", value, value.Kind())\n\n\t\treturn false\n\t}\n\n\tfor i:=0; i<value.Len();i++ {\n\t\tif value.Index(i).Interface() == elem {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(list []string, strToSearch string) bool {\n\tfor _, item := range list {\n\t\tif item == strToSearch {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(list []string, strToSearch string) bool {\n\tfor _, item := range list {\n\t\tif item == strToSearch {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func inList(needle string, haystack []string) bool {\n\tfor _, element := range haystack {\n\t\tif needle == element {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(list []*Record, r *Record) bool {\n\tfor _, record := range list {\n\t\tif compare(r, record) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (w *Whitelist) Contains(mid int64) (ok bool) {\n\tif mid > 0 {\n\t\t_, ok = w.list[mid]\n\t}\n\treturn\n}", "func contains(key string, search []string) bool {\n\tfor _, val := range search {\n\t\tif val == key {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(l []string, t string) bool {\n\tfor _, s := range l {\n\t\tif s == t {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(list []int64, item int64) bool {\n\tfor _, b := range list {\n\t\tif b == item {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func ListContains(list []string, str string) bool {\n\tfor i := range list {\n\t\tif list[i] == str {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func SInList(val string, list []string) bool {\n\tfor _, sval := range list {\n\t\tif val == sval {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s *Set) Contains(val interface{}) bool {\n\tif _, ok := s.vals[val]; ok {\n\t\treturn true\n\t}\n\treturn false\n}", "func Contains(h []string, n string) bool {\n\tfor _, s := range h {\n\t\tif s == n {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func ValueIn(value string, in []string) bool {\n\tfor _, v := range in {\n\t\tif value == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func checkInSlice(list []string, value string) bool {\n\tset := make(map[string]bool)\n\tfor _, v := range list {\n\t\tset[v] = true\n\t}\n\treturn set[value]\n}", "func Contains(nums []int, val int) bool {\n\n\treturn true\n}", "func Contains(target interface{}, list interface{}) (bool, error) {\n\tswitch list.(type) {\n\tdefault:\n\t\treturn false, fmt.Errorf(\"%v is an unsupported type\", reflect.TypeOf(list))\n\tcase []int:\n\t\trevert := list.([]int)\n\t\tfor _, r := range revert {\n\t\t\tif target == r {\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t\treturn false, nil\n\n\tcase []uint64:\n\t\trevert := list.([]uint64)\n\t\tfor _, r := range revert {\n\t\t\tif target == r {\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t\treturn false, nil\n\n\tcase []string:\n\t\trevert := list.([]string)\n\t\tfor _, r := range revert {\n\t\t\tif target == r {\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t\treturn false, nil\n\t}\n}", "func Contains(slice []string, val string) bool {\n\tfor _, item := range slice {\n\t\tif item == val {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(xs []float64, val float64) bool {\n\tfor _, x := range xs {\n\t\tif x == val {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(lst []Node, node Node) bool {\n\tfor _, n := range lst {\n\t\tif node == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func inList(element int, list []int) bool {\n\n\tfor _, temp := range list {\n\t\tif temp == element {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s *Set) Contains(val interface{}) bool {\n\t_, contains := s.set[val]\n\treturn contains\n}", "func Contains(haystack []Fingerprint, needle Fingerprint) bool {\n\tfor _, value := range haystack {\n\t\tif value == needle {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(a []string, x string) bool {\n\tfor _, value := range a {\n\t\tif x == value {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s Set) Contains(value interface{}) bool {\n\thash, err := hashstructure.Hash(value, nil)\n\tif err != nil {\n\t\tlog.Errorf(\"type could not be hashed: %+v\", value)\n\t}\n\t_, found := s[hash]\n\treturn found\n}", "func Contains(arr []string, val string) bool {\n\tfor _, a := range arr {\n\t\tif a == val {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func isIn(key []string, value []string) bool {\n\tset := make(map[string]bool)\n\n\tfor _, val := range value {\n\t\tset[val] = true\n\t}\n\n\tfor _, val := range key {\n\t\t_, found := set[val]\n\t\tif !found {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func inList(s string, list []string) bool {\n\treturn indexList(s, list) != -1\n}", "func Contains(val interface{}, array interface{}) (exists bool, index int) {\n\texists = false\n\tindex = -1\n\n\tswitch reflect.TypeOf(array).Kind() {\n\tcase reflect.Slice:\n\t\ts := reflect.ValueOf(array)\n\n\t\tfor i := 0; i < s.Len(); i++ {\n\t\t\tif reflect.DeepEqual(val, s.Index(i).Interface()) == true {\n\t\t\t\tindex = i\n\t\t\t\texists = true\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\n\treturn\n}", "func (k List) Contains(c Cipher) bool {\n\tcount := k.Count()\n\tfor i := 0; i < count; i++ {\n\t\tif k.Member(i) == c {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func in(slice []string, value string) bool {\n\tfor _, item := range slice {\n\t\tif value == item {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Includes(arr []string, value string) bool {\n\tfor i := range arr {\n\t\tif arr[i] == value {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s Slice) Contains(v uuid.UUID) bool {\n\tfor _, i := range s {\n\t\tif i == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s *HashSet) Contain(data interface{}) (bool, error) {\n\terr := s.checkT(data)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\t_, ok := s.set[data]\n\tif ok {\n\t\treturn true, nil\n\t} else {\n\t\treturn false, nil\n\t}\n}", "func contains(haystack []string, needle string)(bool) {\n\tfor _, hay := range haystack {\n\t\tif hay == needle {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func ExistIntList(list []int, val int) bool {\n\tif len(list) == 0 {\n\t\treturn false\n\t}\n\tfor _, i := range list {\n\t\tif i == val {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(c []string, want string) bool {\n\tfor _, e := range c {\n\t\tif e == want {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(slice, val interface{}) bool {\n\tfn, ok := containsOf(slice, val)\n\tif fn == nil {\n\t\tpanic(\"slice is not supported slice type\")\n\t}\n\tif !ok {\n\t\tpanic(\"val is not the same type as slice\")\n\t}\n\n\tsptr := noescape(ptrOf(slice))\n\tvptr := noescape(ptrOf(val))\n\treturn fn(sptr, vptr)\n}", "func contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(items []string, target string) bool {\n\tfor _, item := range items {\n\t\tif item == target {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(l []string, s string) bool {\n\tfor _, str := range l {\n\t\tif str == s {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func tokenListContainsValue(header http.Header, name string, value string) bool {\n\tfor _, s := range header[name] {\n\t\tif strings.Contains(s, value) {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func contains(arr []string, v string) bool {\r\n\tfor _, s := range arr {\r\n\t\tif s == v {\r\n\t\t\treturn true\r\n\t\t}\r\n\t}\r\n\treturn false\r\n}", "func Contains[T comparable](vs []T, needle T) bool {\n\tfor _, v := range vs {\n\t\tif v == needle {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(haystack []string, needle string) bool {\n\tfor _, v := range haystack {\n\t\tif v == needle {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (sl *LockFreeSkipList) Contains(value interface{}) bool {\n\tvar prevs [maxLevel]*node\n\tvar nexts [maxLevel]*node\n\treturn sl.find(value, &prevs, &nexts)\n}", "func Contains[T comparable](a []T, x T) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contain(s []int, target int) bool {\n\tfor _, v := range s {\n\t\tif v == target {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (al *ArrayList) Contains(value interface{}) bool {\n\tvar flag bool\n\tfor i := 0; i < al.Size(); i++ {\n\t\tif al.slice[i] == value {\n\t\t\tflag = true\n\t\t\tbreak\n\t\t}\n\t}\n\treturn flag\n}", "func isPresent(value string, list LWWNodeSlice) bool {\n\tfor _, element := range list {\n\t\tif element.Value == value {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(a []string, x string) bool {\n\tfor _, n := range a {\n\t\tif x == n {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains[T comparable](tt []T, item T) bool {\n\tfor _, t := range tt {\n\t\tif t == item {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(top3 []*VotingResult, cand *VotingResult) bool {\n\tfor _, c := range top3 {\n\t\tif cand == c {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains[T comparable](collection []T, element T) bool {\n\tfor _, item := range collection {\n\t\tif item == element {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func Contains[T comparable](collection []T, element T) bool {\n\tfor _, item := range collection {\n\t\tif item == element {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (list *ArrayList[T]) Contains(ele T) bool {\n\tfor _, o := range list.elems {\n\t\tif o == ele {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func deriveContains(list []int, item int) bool {\n\tfor _, v := range list {\n\t\tif v == item {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func contains(needle int, haystack []int) bool {\n\tfor _, i := range haystack {\n\t\tif needle == i {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(s string, set []string) bool {\n\tfor _, t := range set {\n\t\tif t == s {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (l *Int32) Contains(value int32) bool {\n\tfor _, v := range l.values {\n\t\tif v == value {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (this *ExDomain) Contains(ele int) bool {\n\t_, exists := this.Values[ele]\n\treturn exists\n}", "func Contains(heystack []string, needle string) bool {\n\tfor _, c := range heystack {\n\t\tif c == needle {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contain(s interface{}, v interface{}) bool {\n\tswitch slice := s.(type) {\n\tcase []string:\n\t\tfor _, val := range slice {\n\t\t\tif val == v {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\tcase []int:\n\t\tfor _, val := range slice {\n\t\t\tif val == v {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (lst stringList) Contains(s string) bool {\n\tfor _, elt := range lst {\n\t\tif s == elt {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (s Set) Contain(el string) bool {\n\t_, in := s[el]\n\treturn in\n}", "func (list *ArrayList) Contains(elements ...interface{}) bool {\n\tvar flag bool = true\n\tfor _, e := range elements {\n\t\tif !list.contain(e) {\n\t\t\tflag = false\n\t\t\tbreak\n\t\t}\n\t}\n\n\treturn flag\n}", "func (rlist Stats) Contains(s Stat) bool {\n\tfor _, r := range rlist {\n\t\tif r.ID == s.ID {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func HasIntersection(referenceList []string, items []string) bool {\n\tfor _, i := range items {\n\t\tif ListContainsString(referenceList, i) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (l *LinkedList) Contains(value datatype) bool {\n\t// If list is empty no point in searching.\n\tif l.head == nil {\n\t\treturn false\n\t}\n\n\tcurrent := l.head\n\tfound := false\n\tfor {\n\t\tif current.data == value {\n\t\t\tfound = true\n\t\t\tbreak\n\t\t} else {\n\t\t\tif current.next != nil {\n\t\t\t\tcurrent = current.next\n\n\t\t\t} else {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\treturn found\n}", "func IsContains(nl []int, n int) bool {\n\tfor _, i := range nl {\n\t\tif i == n {\n\t\t\tfmt.Println(\"Is contains\")\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func checkExists(list []string, v string) bool {\n\tfor _, a := range list {\n\t\tif a == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (array Array) Includes(value interface{}) bool {\n\tfor _, v := range array {\n\t\tif v == value {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func IsIn(str string, lst ...string) bool {\n\tfor _, l := range lst {\n\t\tif str == l {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func Contains(vs []string, t string) bool {\n\tfor _, v := range vs {\n\t\tif v == t {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func ListContainsElement(list []string, element string) bool {\n\tfor _, item := range list {\n\t\tif item == element {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func ListContainsElement(list []string, element string) bool {\n\tfor _, item := range list {\n\t\tif item == element {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func (s *stackImpl) Contains(t T) bool {\n\tfor _, n := range s.items {\n\t\tif t == n {\n\t\t\treturn true\n\t\t}\n\t}\n\n\treturn false\n}", "func Contains(aa []string, s string) bool {\n\tfor _, v := range aa {\n\t\tif s == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}" ]
[ "0.77087265", "0.7701496", "0.73847824", "0.7364822", "0.72933656", "0.72532976", "0.7242146", "0.72027147", "0.7201034", "0.71450686", "0.71242476", "0.71093506", "0.70753866", "0.7039177", "0.7035124", "0.70329887", "0.70328224", "0.69798446", "0.69798446", "0.6979638", "0.69444686", "0.6934607", "0.69339144", "0.69199336", "0.69003487", "0.68999213", "0.68987894", "0.68907493", "0.6878311", "0.6836463", "0.68299675", "0.6798745", "0.6788878", "0.6786311", "0.67780524", "0.6775818", "0.6774193", "0.6755129", "0.675434", "0.6713673", "0.6710465", "0.67090493", "0.6704373", "0.669771", "0.669139", "0.6680976", "0.66519576", "0.66517913", "0.6646936", "0.66420287", "0.664097", "0.66328067", "0.66209775", "0.66200227", "0.66198736", "0.66198736", "0.6613017", "0.6604039", "0.660044", "0.65891033", "0.65864956", "0.6581715", "0.6578445", "0.6566023", "0.65651095", "0.6559455", "0.6557036", "0.65386873", "0.65386873", "0.65386873", "0.65386873", "0.65386873", "0.65386873", "0.6536844", "0.6500408", "0.6499113", "0.6499113", "0.6490072", "0.64845103", "0.64754516", "0.64733636", "0.64666915", "0.64558965", "0.6455842", "0.64482653", "0.64350283", "0.64303905", "0.64262754", "0.6425491", "0.6425027", "0.6423451", "0.64220095", "0.64212424", "0.64164895", "0.6415868", "0.6406727", "0.6406551", "0.6406551", "0.64061", "0.64036924" ]
0.68923366
27
Remove the Node from the list why is List(pointers?)
func (l *List) Remove(n uint16) error { if l.Root == nil { return fmt.Errorf("value does not exist in List or the list is empty") } prev, curr := l.Root, l.Root if curr.Value == n { l.Root = curr.Next curr = nil return nil } for { // 1. Move cur to Next curr = curr.Next // 2. if cur is nil we have reached the end of the list and the value has // not been found; thus we error. If cur is not nil continue to step 3. if curr == nil { return fmt.Errorf("value did not exist in List") } // 3. if cur equals our n value, we found the value we wanted to remove. We // need to set prev.Next to cur.Next and cur.Next to nil; we have now // removed the value and can return. If cur did not equal our n value we // have not foud the value we wish to remove; proceed to step 4. if curr.Value == n { prev.Next = curr.Next curr.Next = nil return nil } // 4. We need to get prev and current to the same node (this was our initial // conditions for looping) - thus we make prev prev.Next. Prev and Cur // are now pointing at the same thing and we can loop. prev = prev.Next } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (l *List) remove(n *Node) {\n\tn.prev.next = n.next\n\tn.next.prev = n.prev\n\tn.next = nil\n\tn.prev = nil\n\tn.list = nil\n\tl.Size--\n}", "func removeNodeFromList(id string) {\n\ti := 0\n\tfor i < len(nodes) {\n\t\tcurrentNode := nodes[i]\n\t\tif currentNode.Id == id {\n\t\t\tnodes = append(nodes[:i], nodes[i+1:]...)\n\t\t} else {\n\t\t\ti++\n\t\t}\n\t}\n}", "func (ns Nodes) Remove() {\n\n}", "func (l *idList) remove(e *idElement) *idElement {\n\te.prev.next = e.next\n\te.next.prev = e.prev\n\te.next = nil // avoid memory leaks\n\te.prev = nil // avoid memory leaks\n\te.list = nil\n\tl.len--\n\treturn e\n}", "func (l *List) Remove(n *Node) interface{} {\n\tif n.list == l {\n\t\tl.remove(n)\n\t}\n\treturn n.Value\n}", "func (cache *Cache) removeNodeInfoFromList(name string) {\n\tni, ok := cache.nodes[name]\n\tif !ok {\n\t\tklog.Errorf(\"No NodeInfo with name %v found in the cache\", name)\n\t\treturn\n\t}\n\n\tif ni.prev != nil {\n\t\tni.prev.next = ni.next\n\t}\n\tif ni.next != nil {\n\t\tni.next.prev = ni.prev\n\t}\n\t// if the removed item was at the head, we must update the head.\n\tif ni == cache.headNode {\n\t\tcache.headNode = ni.next\n\t}\n\n\tdelete(cache.nodes, name)\n}", "func deleteNode_237(node *ListNode) {\n\tnode.Val = node.Next.Val\n\tnode.Next = node.Next.Next\n}", "func deleteNode(node *linkedlist.Node) {\n\t*node = *node.Next\n}", "func (this *MyHashMap) Remove(key int) {\n\ti := this.hashKey(key)\n\tnode := &this.list[i]\n\n\tif node.key == key {\n\n\t\tif node.next == nil {\n\t\t\t*node = LinkNode{\n\t\t\t\tvalue: -1,\n\t\t\t}\n\t\t} else {\n\t\t\t*node = *node.next\n\t\t}\n\n\t} else {\n\n\t\tupNode := node\n\n\t\tfor node.next != nil {\n\t\t\tupNode = node\n\t\t\tnode = node.next\n\t\t\tif node.key == key {\n\t\t\t\tupNode.next = node.next\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t}\n\t}\n}", "func removeNode(node *doubleListNode) {\n\tnode.right.left, node.left.right = node.left, node.right\n}", "func (l *SinglyLinkedList) remove(node *Node) *Node {\n\tvar previous *Node\n\tif l.head.next != nil {\n\t\tfor previous = l.head.next; previous.next != nil && previous.next != node; previous = previous.next {\n\n\t\t}\n\t}\n\n\tprevious.next = node.next\n\tnode.next = nil // avoid memory leaks\n\tnode.list = nil\n\tl.size--\n\tif previous.next == nil {\n\t\tl.last = previous\n\t} else {\n\t\tl.last = previous.next\n\t}\n\treturn node\n}", "func (l *SinglyLinkedList) Remove(node *Node) interface{} {\n\tif l.size != 0 && node.list == l {\n\t\treturn l.remove(node).Value\n\t}\n\treturn nil\n}", "func remove(nodes []*DSSNode, node *DSSNode) ([]*DSSNode, bool) {\n\tfor i, n := range nodes {\n\t\tif n == node {\n\t\t\tnodes[i] = nodes[len(nodes)-1]\n\t\t\tnodes[len(nodes)-1] = nil\n\t\t\tnodes = nodes[:len(nodes)-1]\n\t\t\treturn nodes, true\n\t\t}\n\t}\n\treturn nodes, false\n}", "func (l *List) Remove(value interface{}) {\n\tl.Lock()\n\tfor n := l.head; n != nil; n = n.next {\n\t\tif n.value == value {\n\t\t\tn.prev.next = n.next\n\t\t\tn.next.prev = n.prev\n\t\t\tl.length--\n\t\t}\n\t}\n\tl.Unlock()\n}", "func (list *LinkedList) Remove() {\n\tlink := list.firstLink\n\tif list.firstLink != nil {\n\t\tlist.firstLink = link.next\n\t}\n}", "func (l *List) Delete(n *Node) interface{} {\n\tif n.list == l {\n\t\tl.remove(n)\n\t}\n\treturn n.Data\n}", "func (l *List) DeleteList(name string) {\n\n\tl.size--\n\t/* single item in list */\n \tif l.head == l.tail {\n \t\tl.head = nil\n \t \tl.tail = nil\n \t\treturn\n \t} \n\n \t/* Find the entry to delete */\n\tcurrentNode := l.head\n\tvar prev *Node = nil\n\tfor currentNode != nil && \n\t\tstrings.Compare(strings.ToUpper(name), strings.ToUpper(currentNode.emp.name)) != 0 {\n\t\tprev = currentNode\n\t\tcurrentNode = currentNode.next\n\t}\n\n\t/* If entry not found */\n\tif currentNode == nil {\n\t\tfmt.Println(\"Node not found for name: %s\", name)\n\t\tl.size++\n\t\treturn\n\t}\n\n\t/*If the last entry to be removed */\n\tif (currentNode == l.tail) {\n\t\tprev.next = nil;\n\t\tl.tail = prev\n\t}\n\n\t/* if the first entry to be removed */\n\tif (currentNode == l.head) {\n\t\tl.head = currentNode.next\n\t\tcurrentNode.next = nil\n\t} else { /* middle entry to be removed */\n\t\tprev.next = currentNode.next\n\t}\n}", "func DelSlistNode(i *slist.Item) bool {\n\tif i == nil {\n\t\treturn false\n\t}\n\n\tn := i.Next\n\tif n == nil { // 如果i是单链表最后一个节点就无法处理\n\t\treturn false\n\t}\n\n\ti.Data = n.Data\n\ti.Next = n.Next\n\treturn true\n}", "func ListRemoveIf(l *List, data_ref interface{}) {\n\n\tfor l.Head != nil && l.Head.Data == data_ref {\n\t\tl.Head = l.Head.Next\n\t}\n\tif l.Head == nil {\n\t\treturn\n\t}\n\n\tcurrent := l.Head.Next\n\tparent := l.Head\n\tfor current != nil {\n\t\tif current.Data == data_ref {\n\t\t\tparent.Next = current.Next\n\t\t\tcurrent = current.Next\n\t\t\tcontinue\n\t\t}\n\t\tparent = parent.Next\n\t\tcurrent = current.Next\n\t}\n}", "func RemoveNode(ll *List, node *Node) (*Node, error) {\n\tll.Mux.Lock()\n\tif ll.Size == 0 {\n\t\tll.Mux.Unlock()\n\t\treturn nil, errors.New(\"List is empty\")\n\t} else if ll.Size == 1 {\n\t\tll.Mux.Unlock()\n\t\treturnNode, _ := RemoveLast(ll)\n\t\treturn returnNode, nil\n\t}\n\tll.Mux.Unlock()\n\n\tll.Mux.Lock()\n\tdefer ll.Mux.Unlock()\n\n\tprevNode := node.Prev\n\tnextNode := node.Next\n\n\tprevNode.Next = node.Next\n\tnextNode.Prev = node.Prev\n\n\tatomic.AddInt32(&ll.Size, -1)\n\n\treturn node, nil\n}", "func (list *Turn) Unlink(t *Turn) /*newList*/ *Turn {\n\t// If the item is not in any list, then unlinking is a no-op.\n\tif !t.IsList() {\n\t\tlog.V(3).Infof(\"%v: Unlink no-op %v\", list, t)\n\t\treturn list\n\t}\n\n\tassert.True(!list.IsEmpty(), \"Cannot unlink from an empty list.\")\n\tassert.True(list.IsList(), \"Cannot unlink from single item.\")\n\n\tlog.V(3).Infof(\"%v: Unlink %v\", list, t)\n\n\t// If the list contains only one item then it better be t.\n\tif list == list.next {\n\t\tassert.True(list == t, \"Expected item %v to be in list %v.\", t, list)\n\n\t\tt.next = nil // unlink the item from the list.\n\t\treturn Empty\n\t}\n\n\t// If t is the head then just remove it.\n\tif list.next == t {\n\t\tlist.next = t.next\n\t\tt.next = nil\n\t\treturn list\n\t}\n\n\t// Find the item (assuming it is in the list).\n\tbefore := list.next\n\tfor before.next != t {\n\t\tbefore = before.next\n\t\tassert.True(before != list, \"Expected item %v to be in list %v.\", t, list)\n\t}\n\tassert.True(before.next == t, \"Expected head %v to be t %v\", before.next, t)\n\n\tbefore.next = t.next\n\tt.next = nil\n\n\t// If we removed the list item, the before is the new list, otherwise list hasn't changed.\n\tif t == list {\n\t\treturn before\n\t}\n\treturn list\n}", "func (l *LinkedList) Remove() *Element {\n\telement := l.Head\n\tl.Head = element.Next\n\telement.Next = nil\n\treturn element\n}", "func (this *Ring) Remove(node string) *Ring {\n\thash := this.Hash(node)\n\tfor k,v := range(this.Ring) {\n\t\tif hash == v.Real {\n\t\t\tthis.Ring = append(this.Ring[:k], this.Ring[k:]...)\n\t\t}\n\t}\n\tdelete(this.Nodes, hash)\n\treturn this\n}", "func (s *SinglyLinkedList) Remove(n *Node) {\n if s.front == n {\n s.front = n.next\n s.length--\n } else {\n currentNode := s.front\n\n // search for node n\n for currentNode != nil && currentNode.next != nil && currentNode.next != n {\n currentNode = currentNode.next\n }\n\n // see if current's next node is n\n // if it's not n, then node n wasn't found in list s\n if currentNode.next == n {\n currentNode.next = currentNode.next.next\n s.length--\n }\n }\n}", "func (e *entry) remove() {\n\te.prev.next = e.next\n\te.next.prev = e.prev\n\te.prev = nil\n\te.next = nil\n}", "func (l *List) Remove(i Item) error {\r\n\tif i.parent != l {\r\n\t\treturn errors.New(\"Item must belong to list, which method is called\")\r\n\t}\r\n\tif *i.invalidated {\r\n\t\treturn errors.New(\"Item was deleted, therefore invalidated\")\r\n\t}\r\n\tif i.prev != nil && i.next != nil {\r\n\t\ti.prev.next = i.next\r\n\t\ti.next.prev = i.prev\r\n\t} else if i.prev == nil && i.next != nil {\r\n\t\ti.next.prev = nil\r\n\t\tl.first = i.next\r\n\r\n\t} else if i.next == nil && i.prev != nil {\r\n\t\ti.prev.next = nil\r\n\t\tl.last = i.prev\r\n\t} else {\r\n\t\tl.first = nil\r\n\t\tl.last = nil\r\n\t}\r\n\r\n\t*i.invalidated = true\r\n\tl.length--\r\n\treturn nil\r\n}", "func (l *DcmList) Remove() *DcmObject {\n\tif l.Empty() {\n\t\treturn nil\n\t} else if l.Valid() != true {\n\t\treturn nil\n\t} else {\n\t\ttmpnode := l.currentNode\n\t\tif l.currentNode.prevNode == nil {\n\t\t\tl.firstNode = l.currentNode.nextNode // delete first element\n\t\t} else {\n\t\t\tl.currentNode.prevNode.nextNode = l.currentNode.nextNode\n\t\t}\n\t\tif l.currentNode.nextNode == nil {\n\t\t\tl.lastNode = l.currentNode.prevNode // delete last element\n\t\t} else {\n\t\t\tl.currentNode.nextNode.prevNode = l.currentNode.prevNode\n\t\t}\n\t\tl.currentNode = l.currentNode.nextNode\n\t\ttmpobj := tmpnode.Value()\n\t\tl.cardinality = l.cardinality - 1\n\t\treturn tmpobj\n\t}\n}", "func (nl *NodeList) Remove(n *Node) *Node {\n\tif n == nil || nl.length == 0 {\n\t\treturn nil\n\t}\n\n\tif n.prev == nil {\n\t\tif nl.front == n {\n\t\t\tnl.front = n.Next\n\t\t}\n\t} else {\n\t\tn.prev.Next = n.Next\n\t}\n\n\tif n.Next == nil {\n\t\tif nl.back == n {\n\t\t\tnl.back = n.prev\n\t\t}\n\t} else {\n\t\tn.Next.prev = n.prev\n\t}\n\n\tn.Next = nil\n\tn.prev = nil\n\n\tnl.length--\n\treturn n\n}", "func (l *list) delete(i int) {\n\n\tif l.begin == nil {\n\t\tpanic(\"list empty\")\n\t}\n\n\t// List over/underflow\n\tif i > l.nodes || i < 0 {\n\t\tpanic(\"not exists\")\n\t}\n\n\t// Removing the last node\n\tif l.nodes == 1 && i == 0 {\n\t\tl.begin = nil\n\t\tl.nodes = 0\n\t\treturn\n\t}\n\n\t// Removing at the end of the list\n\tif i == l.nodes-1 {\n\t\tn := l.begin\n\n\t\tfor j := 0; j < l.nodes-1; j++ {\n\t\t\tn = n.right\n\t\t}\n\n\t\tn.left.right = nil\n\t\tn = nil\n\t\tl.nodes--\n\t\treturn\n\t}\n\n\t// Removing the first node\n\tif i == 0 {\n\t\tn := l.begin.right\n\t\tl.begin = n\n\t\tl.begin.left = nil\n\t\tl.nodes--\n\t\treturn\n\t}\n\n\n\t// Removing in somewhere between\n\tc := l.begin\n\n\tfor j := 0; j < i; j++ {\n\t\tc = c.right\n\t}\n\n\tc.left.right, c.right.left = c.right, c.left\n\tl.nodes--\n}", "func Delete(x int, l linklist.List) {\n\tp := FindPrevious(x, l)\n\tif p.Next != nil {\n\t\ttemp := p.Next\n\t\tp.Next = temp.Next\n\t}\n}", "func (l *List) Delete(i interface{}) {\n\tl = GetHead(l)\n\tfor l != nil {\n\t\tif l.value == i {\n\t\t\tif l.prev != nil {\n\t\t\t\tl.prev.next = l.next\n\t\t\t}\n\t\t\tif l.next != nil {\n\t\t\t\tl.next.prev = l.prev\n\t\t\t}\n\t\t\tl = nil\n\t\t\tbreak\n\t\t}\n\t\tl = l.next\n\t}\n}", "func (l *List) Del(v interface{} /* val */) *El {\n\tcur := l.search(v, true, true)\n\n\tif cur == nil || l.less(v, cur.val) {\n\t\treturn nil\n\t}\n\n\tl.len--\n\n\th := cur.height()\n\tfor i := h - 1; i >= 0; i-- {\n\t\t*l.up[i] = cur.nexti(i)\n\t}\n\n\tif l.autoreuse {\n\t\tReuse(cur)\n\t}\n\n\treturn cur\n}", "func (ns NodeSet) Remove(node Node) {\n\tdelete(ns, node)\n}", "func (list *LinkedList) remove(index int) interface{}{\n\tfor i:=0; i<index-1;i++{\n\t\tif(list==nil){\n\t\t\treturn errors.New(\"Out of Bounds\")\n\t\t}\n\t\tlist=list.next\n\t}\n\tpivot:=list.next\n\tif pivot==nil{\n\t\treturn nil\n\t}\n\tlist.next = pivot.next\n\tpivot.next=nil\n\treturn pivot.data\n\n\n}", "func (l *DoublyLinkedList) Remove(i int) {\n\tcur := l.head\n\n\tif cur == nil {\n\t\tpanic(\"DoublyLinkedList is empty!\")\n\t}\n\n\tif cur.value == i {\n\t\tl.head = &Node{}\n\t\treturn\n\t}\n\n\tfor cur.next != nil {\n\t\tcur = cur.next\n\n\t\tif cur.value == i {\n\t\t\tcur.prev.next = cur.next\n\t\t\treturn\n\t\t}\n\t}\n}", "func (l SList) Remove(prev int64, free func(off int64) error) error {\n\tif free != nil {\n\t\tif err := free(l.DataOff()); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif prev != 0 {\n\t\tnext, err := l.Next()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tp, err := l.OpenSList(prev)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tif err := p.setNext(next); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn l.Free(l.Off)\n}", "func removeElements(head *ListNode, val int) *ListNode {\n\tvar vhead = new(ListNode)\n\tvhead.Next = head\n\tvar cur, del = vhead, head\n\tfor del != nil {\n\t\tif del.Val == val {\n\t\t\tcur.Next = del.Next\n\t\t\tdel.Next = nil\n\t\t\tdel = cur.Next\n\t\t} else {\n\t\t\tcur = del\n\t\t\tdel = del.Next\n\t\t}\n\t}\n\n\treturn vhead.Next\n}", "func (nl *nodeList) delete(i int) *Node {\n\tend := len(nl.elements) - 1\n\tn := nl.elements[i]\n\t// Copy values from the deletion point to the left by one\n\tcopy(nl.elements[i:], nl.elements[i+1:])\n\t// Dereference the last value\n\tnl.elements[end] = nil\n\t// Truncate the slice\n\tnl.elements = nl.elements[:end]\n\n\treturn n\n}", "func removeNode(n *doc.Node) {\n\t// Check if this node exists.\n\tfor i, v := range localNodes {\n\t\tif n.ImportPath == v.ImportPath {\n\t\t\tlocalNodes = append(localNodes[:i], localNodes[i+1:]...)\n\t\t\treturn\n\t\t}\n\t}\n}", "func (this *MyHashMap) Remove(key int) {\n\tkey = this.hash(key)\n\tif this.ht[key] == nil {\n\t\treturn\n\t}\n\tif this.ht[key].k == key {\n\t\tthis.ht[key] = this.ht[key].next\n\t\treturn\n\t}\n\tfor node := this.ht[key]; node.next != nil; node = node.next {\n\t\tif node.next.k == key {\n\t\t\tnode.next = node.next.next\n\t\t\treturn\n\t\t}\n\t}\n}", "func (b *channelBlockedList) remove(old *channelBlockedList) *channelBlockedList {\n\tif b == old {\n\t\treturn b.next\n\t}\n\tc := b\n\tfor ; c != nil && c.next != old; c = c.next {\n\t}\n\tif c != nil {\n\t\tc.next = old.next\n\t}\n\treturn b\n}", "func removeElFromlist(p PointStruct, listp *[]PointStruct) {\n\n\tcopyInput := *listp\n\tfor idx, val := range copyInput{\n\n\t\tif val == p{\n\t\t\ttemp1:= copyInput[:idx]\n\t\t\ttemp2:= copyInput[idx:]\n\t\t\tif len(temp2) == 1{\n\t\t\t\t*listp = temp1\n\t\t\t\treturn\n\t\t\t}else{\n\t\t\t\ttemp2 = temp2[1:]\n\t\t\t\t*listp = append(temp1, temp2...)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n}", "func (p *LinkedList) DelNode(node *Cell) {\n\n\tif node == nil {\n\t\tcolor.Yellow.Println(\"Possibility list is empty or has reached the end.\")\n\t\treturn\n\t}\n\n\tif node == p.Head {\n\t\tp.Head = node.Next\n\t} else {\n\t\tif node != nil {\n\t\t\tif node.Next != nil {\n\t\t\t\tnode.Prev.Next = node.Next\n\t\t\t\tnode.Next.Prev = node.Prev\n\t\t\t} else {\n\t\t\t\tnode.Prev.Next = nil\n\t\t\t}\n\t\t}\n\t}\n}", "func Delete(list LWWNodeSlice, value string) LWWNodeSlice {\n\tnewList := LWWNodeSlice{}\n\tfor _, node := range list {\n\t\tif node.Value != value {\n\t\t\tnewList = append(newList, node)\n\t\t}\n\t}\n\treturn newList\n}", "func (this *MyHashMap) Remove(key int) {\n\tindex := this.nodeIndex(key)\n\tv := this.table[index]\n\tif v == nil {\n\t\treturn\n\t}\n\tif v.key == key {\n\t\tthis.table[index] = v.next\n\t\tthis.size--\n\t\treturn\n\t}\n\tfor v != nil && v.next != nil {\n\t\tif v.next.key == key {\n\t\t\tv.next = v.next.next\n\t\t\tthis.size--\n\t\t\treturn\n\t\t}\n\t\tv = v.next\n\t}\n}", "func (l *LRUCache) delete(node *Node) {\n\tnode.prev.next = node.next\n\tnode.next.prev = node.prev\n}", "func (l *List) Remove(index int) {\n\tif l.len == 0 {\n\t\tfmt.Println(\"list is empty\")\n\t\treturn\n\t}\n\tif !(index >= 0 && index < l.len) {\n\t\treturn\n\t}\n\tvar item *Item\n\n\titem = l.head\n\tfor i := 0; i != index; i, item = i+1, item.Next() {\n\t}\n\t\n\tif l.head == item {\n\t\tl.head = item.Next()\n\t}\n\tif l.tail == item {\n\t\tl.tail = item.Prev()\n\t}\n\tif item.prev != nil {\n\t\titem.prev.next = item.Next()\n\t}\n\tif item.next != nil {\n\t\titem.next.prev = item.Prev()\n\t}\n\t\n\tl.len--\n}", "func (n *node) nodeRemove() {\n\tn.evalTotal = n.evalTotal - n.entry.eval\n\tn.occupied = false\n\tparent := n.parent\n\tfor parent != nil {\n\t\tparent.evalTotal = parent.evalTotal - n.entry.eval\n\t\tparent = parent.parent\n\t}\n}", "func (l *Slist) RemoveFirst(d interface{}) {\n\tif l.len == 0 {\n\t\treturn\n\t}\n\n\t// the case the head data is equal d\n\tif l.head.data == d {\n\t\tl.head = l.head.next\n\t\tl.len--\n\t\treturn\n\t}\n\n\tprev := l.head\n\tcurrent := prev.next\n\tfor current != nil {\n\t\tif current.data == d {\n\t\t\t// headを持っているのがprevなので、prev.nextをcurrent.nextにする\n\t\t\t// current = current.nextだと、currentはheadを持っていないので、リムーブできない。\n\t\t\tprev.next = current.next\n\t\t\tl.len--\n\t\t\treturn\n\t\t}\n\t\tprev = current\n\t\tcurrent = current.next\n\t}\n}", "func (list *List) Remove(index int) {\n\n\tif !list.withinRange(index) {\n\t\treturn\n\t}\n\n\tif list.size == 1 {\n\t\tlist.Clear()\n\t\treturn\n\t}\n\n\tvar beforeElement *element\n\telement := list.first\n\tfor e := 0; e != index; e, element = e+1, element.next {\n\t\tbeforeElement = element\n\t}\n\n\tif element == list.first {\n\t\tlist.first = element.next\n\t}\n\tif element == list.last {\n\t\tlist.last = beforeElement\n\t}\n\tif beforeElement != nil {\n\t\tbeforeElement.next = element.next\n\t}\n\n\telement = nil\n\n\tlist.size--\n}", "func (node *circularLinkedNode) delete() {\n\tnode.prev.next, node.next.prev = node.next, node.prev\n}", "func (l *LinkedList) remove(index int) {\n\tprev := l.head\n\tif index > 0 {\n\t\tprev = l.get(index - 1)\n\t}\n\n\tcurrent := prev.next\n\tnext := current.next\n\n\tprev.next = next\n\tcurrent.next = nil\n\tl.length -= 1\n}", "func (l *list) delete(i int) {\n n := l.first\n\n j := 0\n for n.next != nil && j < i {\n n = n.next\n j++\n }\n\n if j == 0 && n != nil {\n l.first = n.next\n }\n\n tmp := n.next\n\n if (n.prev != nil) {\n n.prev.next = tmp\n }\n\n if (n.next != nil) {\n n.next.prev = n.prev\n }\n\n}", "func (hm *MyHashMap) Remove(key int) {\n\tvar prev *Node\n\tentry := hm.Arrays[hm.HashFunc(key)]\n\tfor entry != nil && entry.Key != key {\n\t\tprev = entry\n\t\tentry = entry.Next\n\t}\n\n\tif entry == nil {\n\t\treturn\n\t}\n\n\tif prev == nil {\n\t\thm.Arrays[hm.HashFunc(key)] = entry.Next\n\t} else {\n\t\tprev.Next = entry.Next\n\t}\n\n\treturn\n}", "func RemoveLast(ll *List) (*Node, error) {\n\t// Lock access\n\tll.Mux.Lock()\n\tdefer ll.Mux.Unlock()\n\n\tif ll.Size == 0 {\n\t\treturn nil, errors.New(\"List is empty\")\n\t}\n\t// Update reference pointers\n\tnodeToRemove := ll.Tail.Prev\n\n\tnodeToRemove.Prev.Next = ll.Tail\n\tll.Tail.Prev = nodeToRemove.Prev\n\n\tatomic.AddInt32(&ll.Size, -1)\n\n\treturn nodeToRemove, nil\n}", "func (n *Nodes) Remove(ids ...string) {\n\tif n == nil {\n\t\treturn\n\t}\n\n\tvar r Nodes\n\tfor k, v := range *n {\n\t\tvar found bool\n\t\tfor _, needle := range ids {\n\t\t\tif (*n)[k].ID() == needle {\n\t\t\t\tfound = true\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tif !found {\n\t\t\tr = append(r, v)\n\t\t}\n\t}\n\t*n = r\n}", "func (list *List) DeleteElement(data int) {\n // 1. Provide message to user if the list is empty and return\n if list.Size() == 0 {\n fmt.Println(\"Nothing to delete, the list is empty\")\n return\n }\n\n // 2. Get the current head of the list\n current := list.Head()\n\n // 3. Update the head if current head is the requested element and return\n if current.data == data {\n list.head = current.next\n current.next = nil\n list.size--\n return\n }\n\n // 4. Traverse the list, remove the requested element and return\n for current.next != nil {\n if current.next.data == data {\n tmp := current.next.next\n current.next.next = nil\n current.next = tmp\n list.size--\n return\n }\n current = current.next\n }\n\n // 5. Provide a message to user if the requested element is not found in list\n fmt.Println(\"Could not delete since the element requested does not exist in list\")\n}", "func (l *list) delete(i int) {\n\tif i < 0 || i >= l.size {\n\t\tpanic(\"list index out of bounds\")\n\t}\n\n\tvar n *node\n\tmid := l.size/2\n\tif mid >= i {\n\t\tn = l.root\n\t\tfor ; i!=0; i-- {\n\t\t\tn = n.next\n\t\t}\n\t} else {\n\t\tn = l.tail\n\t\tfor i=l.size-i-1; i!=0; i-- {\n\t\t\tn = n.prev\n\t\t}\n\t}\n\tif n.prev != nil {\n\t\tn.prev.next = n.next\n\t} else {\n\t\tl.root = n.next\n\t}\n\tif n.next != nil {\n\t\tn.next.prev = n.prev\n\t} else {\n\t\tl.tail = n.prev\n\t}\n\tl.size--\t\n}", "func (ons *orderedNodeSet) remove(node *Node) {\n\tif node == nil {\n\t\treturn\n\t}\n\n\t// Remove this node\n\ti, ok := ons.nodes[node]\n\tif ok {\n\t\tdelete(ons.order, i)\n\t}\n\tdelete(ons.nodes, node)\n}", "func (hm *HashMap) Remove(key []byte) error {\n\t// validate hashKey\n\terr := hm.np.validateKey(key)\n\tif err != nil {\n\t\treturn err\n\t}\n\t//1. calculate the hash num\n\thashNum := hm.hashFunc(key) % uint64(hm.haSize)\n\n\t//2. remove key from hashNode\n\thead := hm.ha[hashNum]\n\tif head == -1 {\n\t\treturn nil\n\t}\n\tnewHead := hm.np.del(head, key)\n\n\t//3. point to the new list head node\n\thm.ha[hashNum] = newHead\n\n\treturn nil\n}", "func (sl *SkipList) Delete(d int) *SkipNode {\n\tf := sl.head.getTop().find(d)\n\n\tn := f\n\tfor n != nil {\n\t\tn.prev.next = n.next\n\t\tn.next.prev = n.prev\n\t\tn = n.higher\n\t}\n\n\treturn f\n}", "func (this *MyHashMap) Remove(key int) {\n\thashCode := key % 1111\n\ttemp := &Node{\n\t\t0,\n\t\t0,\n\t\tthis.buckets[hashCode],\n\t}\n\tpre := temp\n\tcur := temp.Next\n\tfor cur != nil {\n\t\tif cur.K == key {\n\t\t\tpre.Next = cur.Next\n\t\t\tbreak\n\t\t}\n\t\tpre, cur = cur, cur.Next\n\t}\n\tthis.buckets[hashCode] = temp.Next\n}", "func (p *List) Reset() { *p = (*p)[0:0] }", "func removeBlockNodeFromSlice(nodes []*BlockNode, node *BlockNode) []*BlockNode {\n\tfor i := range nodes {\n\t\tif nodes[i].Hash.IsEqual(node.Hash) {\n\t\t\tcopy(nodes[i:], nodes[i+1:])\n\t\t\tnodes[len(nodes)-1] = nil\n\t\t\treturn nodes[:len(nodes)-1]\n\t\t}\n\t}\n\treturn nodes\n}", "func (cell * SpaceMapCell) Remove(el SpaceMapElement) (SpaceMapElement) {\n var e *list.Element\n for e = cell.Shapes.Front() ; e != nil ; e = e.Next() {\n val := e.Value\n if val.(SpaceMapElement) == el {\n cell.Shapes.Remove(e)\n return el\n }\n }\n return nil\n}", "func DeleteNode(n *Node) error {\n\tm := n.next\n\tif m == nil {\n\t\treturn errors.New(\"node is not in a list\")\n\t}\n\tn.next = m.next\n\tn.data = m.data\n\treturn nil\n}", "func (ll *LinkedList) Delete(element int) {\n\tif ll.IsEmpty() {\n\t\treturn\n\t}\n\n\tif ll.start.value == element {\n\t\tll.PopLeft()\n\n\t\treturn\n\t}\n\n\tif ll.end.value == element {\n\t\tll.PopRight()\n\n\t\treturn\n\t}\n\n\tcurrent := ll.start.next\n\tfor current != nil {\n\t\tif current.value == element {\n\t\t\tll.deleteNode(current)\n\t\t\treturn\n\t\t}\n\n\t\tcurrent = current.next\n\t}\n}", "func (e *entry) remove() {\n\tif e.logical != noAnchor {\n\t\tlog.Fatalf(\"may not remove anchor %q\", e.str)\n\t}\n\t// TODO: need to set e.prev.level to e.level if e.level is smaller?\n\te.elems = nil\n\tif !e.skipRemove {\n\t\tif e.prev != nil {\n\t\t\te.prev.next = e.next\n\t\t}\n\t\tif e.next != nil {\n\t\t\te.next.prev = e.prev\n\t\t}\n\t}\n\te.skipRemove = false\n}", "func (l *List) Remove(index int) error {\n\t// Does it make sense to make index as uint? On the other hand it's inconvenient to use uint everywhere\n\tif index < 0 || index > l.Len-1 {\n\t\treturn errors.New(\"index is put of range\")\n\t}\n\titem := l.First\n\tfor i := 0; i <= index; i++ {\n\t\tif i != index {\n\t\t\titem = item.Next\n\t\t\tcontinue\n\t\t}\n\t\tif i == 0 {\n\t\t\t// Set element as first and update the reference to the first element\n\t\t\tl.First = l.First.Next\n\t\t\t// check if element exists in case, when on previous step it was the last one\n\t\t\tif l.First != nil {\n\t\t\t\tl.First.Prev = nil\n\t\t\t}\n\t\t}\n\t\tif i == l.Len-1 {\n\t\t\t// Set element as last and update the reference to the last element\n\t\t\tl.Last = l.Last.Prev\n\t\t\t// check if element exists in case, when on previous step it was the first one\n\t\t\tif l.Last != nil {\n\t\t\t\tl.Last.Next = nil\n\t\t\t}\n\t\t}\n\t\tif i > 0 && i < l.Len-1 {\n\t\t\t// As far as I understood, item should be removed by GC from the memory, right?\n\t\t\titem.Prev = item.Next\n\t\t}\n\t\tl.Len--\n\t\treturn nil\n\t}\n\t// actually, impossible case\n\treturn errors.New(\"element with provided index was not found\")\n}", "func DeleteElement(l *list.List, mark *list.Element) *list.List {\r\n\tmark.Value = mark.Next.Value\r\n\tmark.Next = mark.Next.Next\r\n\tl.N--\r\n\treturn l\r\n}", "func (c *HashRing) RemoveNode(elt string) {\n\tc.Lock()\n\tdefer c.Unlock()\n\n\tweight := c.weights[elt]\n\tfor i := 0; i < c.numberOfCubes*weight; i++ {\n\t\tdelete(c.ring, c.generateHash(c.generateKey(elt, i)))\n\t}\n\tdelete(c.members, elt)\n\tdelete(c.weights, elt)\n\tc.updateSortedRing()\n}", "func removeN(head *node, n int) *node {\n\tret := head.next\n\n\tcur := head\n\tfor i := 0; i < n; i++ {\n\t\tcur = cur.next\n\t}\n\t// cur is now the last node we need to extract\n\thead.next = cur.next\n\tcur.next = ret\n\n\treturn ret\n}", "func (this *LinkedList) Remove(ele interface{}) bool {\n\tif this.head == nil && this.tail == nil {\n\t\treturn false\n\t}\n\tpe := this.head\n\tvar ps *entry\n\tps = nil\n\tfor pe != nil {\n\t\tif pe.elem == ele {\n\t\t\tif ps != nil {\n\t\t\t\tps.next = pe.next\n\t\t\t\tif pe == this.tail {\n\t\t\t\t\tthis.tail = ps\n\t\t\t\t}\n\t\t\t\tthis.size--\n\t\t\t} else {\n\t\t\t\t// remove first element\n\t\t\t\tthis.head = pe.next\n\t\t\t\t// if remove the only one element, tail must be change to nil\n\t\t\t\tif this.tail == pe {\n\t\t\t\t\tthis.tail = nil\n\t\t\t\t}\n\t\t\t\tthis.size--\n\t\t\t}\n\t\t\treturn true\n\t\t}\n\t\tps = pe\n\t\tpe = pe.next\n\t}\n\treturn false\n}", "func (l *LinkedList) Remove(index int) (*LLNode, error) {\n\tswitch {\n\tcase index > l.Size-1:\n\t\treturn nil, fmt.Errorf(\"Index %d out of range\", index)\n\n\tcase index == 0:\n\t\treturn l.Shift(), nil\n\n\tcase index == l.Size-1:\n\t\treturn l.Pop(), nil\n\n\tdefault:\n\t\tvar removedNode *LLNode\n\t\tcurrentNode, NextNode := l.Head, l.Head.Next\n\t\tfor i := 0; i < l.Size-1; i++ {\n\t\t\tif i+1 == index {\n\t\t\t\tremovedNode = NextNode\n\t\t\t\tcurrentNode.Next = NextNode.Next\n\t\t\t\tbreak\n\t\t\t}\n\n\t\t\tcurrentNode = NextNode\n\t\t\tNextNode = NextNode.Next\n\t\t}\n\n\t\tl.Size--\n\n\t\treturn removedNode, nil\n\t}\n}", "func (wr *WatchList) Remove() (w *Watch) {\n\tw = wr.current\n\tif w != nil {\n\t\tif wr.current != wr.first {\n\t\t\twr.current.Prev.Next = wr.current.Next\n\t\t} else {\n\t\t\twr.first = wr.current.Next\n\t\t}\n\t\tif wr.current != wr.last {\n\t\t\twr.current.Next.Prev = wr.current.Prev\n\t\t\twr.current = wr.current.Next\n\t\t} else {\n\t\t\twr.last = wr.current.Prev\n\t\t\twr.current = nil\n\t\t}\n\t\tw.Next = nil\n\t\tw.Prev = nil\n\t}\n\treturn\n}", "func (l *sampleList) Remove(sample *Sample) {\n\tprev := sample.prev\n\tnext := sample.next\n\tl.release(sample)\n\tif prev == nil {\n\t\tl.head = next\n\t} else {\n\t\tprev.next = next\n\t}\n\tif next == nil {\n\t\tl.tail = prev\n\t} else {\n\t\tnext.prev = prev\n\t}\n}", "func (l *LinkedList) Remove(v interface{}) bool {\n\thead := l.head\n\n\tif head.value == v {\n\t\tl.head = head.next\n\t\treturn true\n\t}\n\n\tcurrentHead := head\n\tfor {\n\t\t// 下一个节点是nil了\n\t\tif currentHead.next.next == nil {\n\t\t\tif currentHead.next.value == v {\n\t\t\t\t// 下一个节点是v ;找到了\n\t\t\t\tcurrentHead.next = nil\n\t\t\t\treturn true\n\t\t\t}\n\t\t\t// 没有找到\n\t\t\treturn false\n\t\t}\n\n\t\t// 正常的节点\n\t\tif currentHead.next.value == v {\n\t\t\tcurrentHead.next = currentHead.next.next\n\t\t\treturn true\n\t\t}\n\n\t\tcurrentHead = currentHead.next\n\t}\n}", "func (s Set) Remove(e graph.Node) {\n\tdelete(s, e.ID())\n}", "func (list *DoublyLinkedList) DeleteNode(index int) {\n\n\tif list.isEmpty() {\n\t\treturn\n\t}\n\tnode, _ := list.GetNodeAt(index)\n\tfmt.Println(node)\n\tfmt.Println(node.data.Name)\n\n\tif list.lenght == 1 {\n\t\tlist.head = nil\n\t} else if node == list.GetLastNode() { // Si es el ultimo al anterior se le apunta a nil\n\t\tnode.previous.next = nil\n\t} else if node == list.head { // Si es el primero al siguiente se le apunta a nil\n\t\ttemp := node.next\n\t\tlist.head = temp\n\t\tnode.next.previous = nil\n\n\t} else {\n\t\tnode.previous.next = node.next\n\t\tnode.next.previous = node.previous\n\t}\n\n\tlist.lenght--\n\n}", "func (l *list) pop() *listNode {\n\tif l.first == nil {\n\t\treturn nil\n\t}\n\n\tl.len--\n\tif l.first == l.tail {\n\t\taux := l.first\n\t\tl.first = nil\n\t\tl.tail = nil\n\t\treturn aux\n\t}\n\taux := l.first\n\tl.first = aux.next\n\treturn aux\n}", "func (list *List) DeleteFromBeginning() {\n // 1. Provide message to user if the list is empty and return\n if list.Size() == 0 {\n fmt.Println(\"Nothing to delete, the list is empty\")\n return\n }\n\n // 2. Get the current head and save it in temp\n oldHead := list.Head()\n\n // 3. Update the list's head to next element in list\n list.head = oldHead.next\n\n // 4. Remove the link from the old head\n oldHead.next = nil\n\n // 5. Decrement the list size\n list.size--\n}", "func (l *List) Clear() {\n\tl.head = nil\n}", "func (this *LinkedList) RemoveAt(index int) interface{} {\n\tif index < 0 || index >= this.Size() {\n\t\tpanic(\"index out of bound\")\n\t}\n\tpe := this.head\n\tvar ps *entry\n\tvar ele interface{}\n\tps = nil\n\tfor i := 0; i < index; i++ {\n\t\tps = pe\n\t\tpe = pe.next\n\t}\n\tif ps != nil {\n\t\tele = pe.elem\n\t\tps.next = pe.next\n\t\tif pe == this.tail {\n\t\t\t// remove the last element\n\t\t\tthis.tail = ps\n\t\t}\n\t} else {\n\t\t// remove the first element\n\t\tele = pe.elem\n\t\tthis.head = this.head.next\n\t\t// if remove the only one element, tail must be change to nil\n\t\tif this.tail == pe {\n\t\t\tthis.tail = nil\n\t\t}\n\t}\n\tthis.size--\n\treturn ele\n}", "func (r *Ring) Remove(nodeName string) {\n\tr.bst.remove(hash(nodeName))\n}", "func (s *nodeSet) remove(tn *treeNode) {\n\tif nPtr, has := s.itemMap[tn]; has {\n\t\tremoveNode(nPtr)\n\t\tdelete(s.itemMap, tn)\n\t}\n}", "func (l *LinkedList) Delete(data interface{}) *LLNode {\n\tvar deletedNode *LLNode\n\n\tswitch l.Size {\n\tcase 0:\n\t\treturn deletedNode\n\tcase 1:\n\t\tif l.Head.Data == data {\n\t\t\tdeletedNode = l.Head\n\t\t\tl.Head, l.Tail = nil, nil\n\t\t}\n\tdefault:\n\t\tif l.Head.Data == data {\n\t\t\tdeletedNode = l.Head\n\t\t\tl.Head = l.Head.Next\n\t\t} else {\n\t\t\tcurrentNode, NextNode := l.Head, l.Head.Next\n\t\t\tfor NextNode != nil {\n\t\t\t\tif NextNode.Data == data {\n\t\t\t\t\tdeletedNode = NextNode\n\t\t\t\t\tcurrentNode.Next = NextNode.Next\n\t\t\t\t\tbreak\n\t\t\t\t}\n\n\t\t\t\tcurrentNode = NextNode\n\t\t\t\tNextNode = NextNode.Next\n\t\t\t}\n\n\t\t\tif deletedNode == l.Tail {\n\t\t\t\tl.Tail = currentNode\n\t\t\t}\n\t\t}\n\t}\n\n\tif deletedNode != nil {\n\t\tl.Size--\n\t}\n\n\treturn deletedNode\n}", "func (l *LinkedList) Delete(val fmt.Stringer) (*Node, error) {\n\tif l.head == nil {\n\t\treturn nil, fmt.Errorf(\"Can't find %d in list\", val)\n\t}\n\tif l.head.data == val {\n\t\treturn l.DeleteFromFront()\n\t}\n\tnode := l.head\n\tfor node.next != nil {\n\t\tif node.next.data == val {\n\t\t\tdeleted := node.next\n\t\t\tnode.next = node.next.next\n\t\t\tl.size--\n\t\t\treturn deleted, nil\n\t\t}\n\t\tnode = node.next\n\t}\n\treturn nil, fmt.Errorf(\"Can't find %d in list\", val)\n}", "func (l *idList) Remove(e *idElement) doc.Metadata {\n\t// read the value before returning to the pool to avoid a data race with another goroutine getting access to the\n\t// list after it has been put back into the pool.\n\tv := e.Value\n\tif e.list == l {\n\t\t// if e.list == l, l must have been initialized when e was inserted\n\t\t// in l or l == nil (e is a zero Element) and l.remove will crash.\n\t\tl.remove(e)\n\t\tl.Pool.put(e)\n\t}\n\treturn v\n}", "func (list *LinkedList) DeleteNode(element interface{}) interface{} {\n\tif list.size == 0 {\n\t\treturn nil\n\t}\n\tif list.head.value == element {\n\t\tdeleted := list.head.value\n\t\tlist.head = list.head.next\n\t\tlist.size--\n\t\tif list.size == 0 {\n\t\t\tlist.tail = nil\n\t\t}\n\t\treturn deleted\n\t}\n\tresult := list.head.SearchPrevElement(element)\n\tif result.value == nil {\n\t\treturn nil\n\t}\n\tfmt.Printf(\"result = %+v\\n\", result)\n\tdeleted := result.next.value\n\t// if the deleted node is tail\n\tif result.next.next == nil {\n\t\tlist.tail = result\n\t}\n\tresult.next = result.next.next\n\tlist.size--\n\treturn deleted\n}", "func (ll *DoubleLinkedList) removeNode(node *node) {\n\tif node == ll.head {\n\t\tll.head = node.next\n\t\tll.head.prev = nil\n\t} else if node.next == nil {\n\t\tll.tail = node.prev\n\t\tll.tail.next = nil\n\t} else {\n\t\ttemp := ll.head\n\t\tfor temp.next != node {\n\t\t\ttemp = temp.next\n\t\t}\n\n\t\ttemp.next = node.next\n\t\tnode.next.prev = temp\n\t}\n\n\tnode = nil\n\tll.count--\n}", "func (list *LinkedList) removeDupes() {\n\tset := make(map[int]bool)\n\tset[list.head.value] = true\n\tfor node := list.head; node.next != nil; node = node.next {\n\t\tif set[node.next.value] {\n\t\t\t// delete the next node\n\t\t\tnode.next = node.next.next\n\t\t} else {\n\t\t\t// hello\n\t\t\tset[node.next.value] = true\n\t\t}\n\t}\n}", "func (s *slot) remove(c interface{}) {\n\ts.mux.Lock()\n\tdefer s.mux.Unlock()\n\tdelete(s.elements, c)\n}", "func (this *MyLinkedList) DeleteAtIndex(index int) {\n\tif this.head == nil || index < 0 {\n\t\treturn\n\t}\n\ttable := make(map[int]*ListNode)\n\ttail, i := this.head, 0\n\tfor tail != nil && i <= index {\n\t\ttable[i] = tail\n\t\ttail = tail.next\n\t\ti++\n\t}\n\tremoveElement := table[index]\n\tif removeElement == nil {\n\t\treturn\n\t}\n\tprev := table[index-1]\n\tvar next *ListNode\n\tif removeElement != nil {\n\t\tnext = removeElement.next\n\t}\n\tif prev == nil && next == nil {\n\t\tthis.head = nil\n\t\tthis.tail = nil\n\t} else {\n\t\tif prev == nil {\n\t\t\tthis.head = next\n\t\t} else {\n\t\t\tprev.next = next\n\t\t}\n\t}\n}", "func (l *LinkedList) Delete(elem string) {\n\tprevious := l.Head\n\tcurrent := l.Head\n\tfor current != nil {\n\t\tif current.Value == elem {\n\t\t\tprevious.Next = current.Next\n\t\t}\n\t\tprevious = current\n\t\tcurrent = current.Next\n\t}\n}", "func (l *LinkedList) Delete(val interface{}) {\n\tif l.Head == nil {\n\t\treturn\n\t}\n\n\tfor {\n\t\tif l.Head.Val == val {\n\t\t\tl.Head = l.Head.Next\n\t\t\tcontinue\n\t\t}\n\t\tbreak\n\t}\n\n\tp := l.Head\n\tfor p.Next != nil {\n\t\tif p.Next.Val == val {\n\t\t\tp.Next = p.Next.Next\n\t\t\tcontinue\n\t\t}\n\t\tp = p.Next\n\t}\n}", "func (rp *resourcePool) remove(e *resourcePoolElement) {\n\tif e == nil {\n\t\treturn\n\t}\n\n\tif e.prev != nil {\n\t\te.prev.next = e.next\n\t}\n\tif e.next != nil {\n\t\te.next.prev = e.prev\n\t}\n\tif e == rp.start {\n\t\trp.start = e.next\n\t}\n\tif e == rp.end {\n\t\trp.end = e.prev\n\t}\n\tatomicSubtract1Uint64(&rp.size)\n}", "func (l *List) Remove(element interface{}) bool {\n\titem, found := l.find(element)\n\tif !found {\n\t\treturn false\n\t}\n\tif l.Len() == 1 {\n\t\tl.firstElement = nil\n\t\tl.lastElement = nil\n\t}\n\tif item.Prev != nil {\n\t\titem.Prev.Next = item.Next\n\t} else {\n\t\tl.firstElement = item.Next\n\t\tl.firstElement.Prev = nil\n\t}\n\tif item.Next != nil {\n\t\titem.Next.Prev = item.Prev\n\t} else {\n\t\tl.lastElement = item.Prev\n\t\tl.lastElement.Next = nil\n\t}\n\tl.len = l.len - 1\n\treturn true\n}", "func (list *DoublyLinkedList) Remove(index int) {\n\n\tif !list.inRange(index) {\n\t\treturn\n\t}\n\n\tif list.size == 1 {\n\t\tlist.Clear()\n\t\treturn\n\t}\n\n\tvar element *doublyLinkedElement\n\t// determine traversal direction, last to first or first to last\n\tif list.size-index < index {\n\t\telement = list.last\n\t\tfor e := list.size - 1; e != index; e, element = e-1, element.prev {\n\t\t}\n\t} else {\n\t\telement = list.first\n\t\tfor e := 0; e != index; e, element = e+1, element.next {\n\t\t}\n\t}\n\n\tif element == list.first {\n\t\tlist.first = element.next\n\t}\n\tif element == list.last {\n\t\tlist.last = element.prev\n\t}\n\tif element.prev != nil {\n\t\telement.prev.next = element.next\n\t}\n\tif element.next != nil {\n\t\telement.next.prev = element.prev\n\t}\n\n\telement = nil\n\n\tlist.size--\n}", "func (q *Queue) Remove() {\n\tq.start = q.start.next\n}", "func (s *children) removeAt(index int) *node {\n\tn := (*s)[index]\n\tcopy((**s)[index:], (*s)[index+1:])\n\t(*s)[len(*s)-1] = nil\n\t*s = (*s)[:len(*s)-1]\n\treturn n\n}" ]
[ "0.7637457", "0.70456165", "0.69729555", "0.6835006", "0.6820128", "0.6781267", "0.6771515", "0.6749114", "0.6719199", "0.6703481", "0.66963655", "0.6665038", "0.65785605", "0.6566689", "0.6556581", "0.64979684", "0.64480686", "0.6443555", "0.6442345", "0.64389855", "0.64379364", "0.6423541", "0.6385108", "0.6380937", "0.6375891", "0.637321", "0.63580585", "0.6355754", "0.6343012", "0.63078415", "0.63048226", "0.6222956", "0.6197603", "0.61967146", "0.6196492", "0.61906993", "0.6172014", "0.6168949", "0.6161663", "0.6158594", "0.6152216", "0.6136937", "0.6117765", "0.6109168", "0.6092413", "0.60897124", "0.6089525", "0.60874635", "0.60831356", "0.6081773", "0.6072704", "0.6068282", "0.6050858", "0.6049262", "0.603657", "0.6028544", "0.6023778", "0.60218096", "0.6021274", "0.60186243", "0.6018261", "0.6015796", "0.6007875", "0.60043174", "0.60000074", "0.5997927", "0.5993846", "0.59670764", "0.59639937", "0.59546864", "0.5951202", "0.59441364", "0.59409237", "0.59335107", "0.5909085", "0.58898824", "0.58865047", "0.5879129", "0.5877282", "0.5874979", "0.5865101", "0.5863085", "0.58546484", "0.5848379", "0.584646", "0.5845768", "0.5845218", "0.5836388", "0.5824076", "0.5822938", "0.5821997", "0.5814146", "0.5795014", "0.5793879", "0.578434", "0.578383", "0.5772132", "0.5771725", "0.57705706", "0.5763983" ]
0.5889579
76
String will return a comma seperated representation of the list
func (l List) String() string { s := "<start>" e := "<end>" if l.Root == nil { s = s + e } if l.Root != nil { n := l.Root for n != nil { if n.Next == nil { // last one s = s + fmt.Sprintf("%v<end>", n.Value) } else { // not last s = s + fmt.Sprintf("%v, ", n.Value) } n = n.Next } } return s }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (f *StringList) String() string {\n\tif f == nil {\n\t\treturn \"\"\n\t}\n\treturn strings.Join(*f, \",\")\n}", "func (list *List) String() string {\n\tvar headerStrings []string\n\tfor _, element := range *list {\n\t\theader := element.ToString()\n\t\theaderStrings = append(headerStrings, header)\n\t}\n\treturn strings.Join(headerStrings, \", \")\n}", "func (pl List) String() string {\n\treturn strings.Join(pl.names(), \", \")\n}", "func (d deck) toString() string {\n\t//d deck to slice of string []string(d)\n\t//now the slice of string into a single string comma separated [\"a\",\"b\"]->(\"a,b\")\n\treturn strings.Join([]string(d), \",\")\n}", "func (el EnvNameList) String() string {\n\tl := make([]string, 0, len(el))\n\n\tfor i, e := range el {\n\t\tl[i] = string(e)\n\t}\n\n\treturn strings.Join(l, \", \")\n}", "func (l *List) String() string {\n\tif l == nil {\n\t\treturn \"nil\"\n\t}\n\tb := bytes.NewBuffer(nil)\n\tb.WriteString(\"[\")\n\ty := l\n\tfor {\n\t\tb.WriteString(fmt.Sprintf(\"%v\", y.val))\n\t\tif !y.End() {\n\t\t\tb.WriteString(\", \")\n\t\t} else {\n\t\t\tbreak\n\t\t}\n\t\ty = y.next\n\t}\n\tb.WriteString(\"]\")\n\n\treturn b.String()\n}", "func (l *List) String() string {\n\tvar b strings.Builder\n\tb.WriteByte('[')\n\tb.WriteString(strconv.Itoa(l.Len))\n\tb.Write([]byte{']', ' ', '{'})\n\tfor curr := l.head; curr != nil; curr = curr.next {\n\t\tfmt.Fprint(&b, curr.Val)\n\t\tb.WriteByte(',')\n\t}\n\tb.WriteByte('}')\n\treturn b.String()\n}", "func (lval *ListValue) String() string {\n\tbuf := new(strings.Builder)\n\tbuf.WriteByte('[')\n\tfor i, v := range lval.Values {\n\t\tif i > 0 {\n\t\t\tbuf.WriteString(\", \")\n\t\t}\n\t\tbuf.WriteString(v.String())\n\t}\n\tbuf.WriteByte(']')\n\treturn buf.String()\n}", "func (l *List) String() string {\n\twords := make([]string, 0, 3*l.Len())\n\tfor e := l.Front(); e != nil; e = e.Next() {\n\t\twords = append(words, fmt.Sprint(e.Value))\n\t}\n\treturn \"(\" + strings.Join(words, \" \") + \")\"\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\")\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\")\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\")\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\")\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\")\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\")\n}", "func (d deck) toString () string {\n\treturn strings.Join([]string(d), \",\")\n}", "func (list *ArrayList) String() string {\n\tstr := \"ArrayList\\n\"\n\tvalues := []string{}\n\tfor _, value := range list.elements[:list.size] {\n\t\tvalues = append(values, fmt.Sprintf(\"%v\", value))\n\t}\n\tstr += strings.Join(values, \", \")\n\treturn str\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\") //[]string(d) deck to slice of string conversion and then join all strings into a single string\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), \",\")\n\n}", "func (d deck) toString() string {\n\treturn strings.Join(d, \",\") // this will return a string seprated by comma\n}", "func (d deck) toString () string{\n\n\t// convert deck -> string[] -> string (comma separated)\n\treturn strings.Join([]string (d), \",\")\n\n}", "func CucumberTagsListToString(tags []string) string {\n\tvar tagList []string\n\tfor _, tag := range tags {\n\t\ttagList = append(tagList, \"@\"+tag)\n\t}\n\treturn strings.Join(tagList, \",\")\n}", "func StringList(strs []string) string {\n\tstrWithQuote := []string{}\n\tfor _, s := range strs {\n\t\tstrWithQuote = append(strWithQuote, fmt.Sprintf(`\"%s\"`, s))\n\t}\n\treturn strings.Join(strWithQuote, \", \")\n}", "func ListToString(list []string) string {\n\ts := \"\\n---------------------------------------\\n\"\n\n\tfor _, b := range list {\n\t\ts += b + \"\\n\"\n\t}\n\n\treturn s + \"---------------------------------------\"\n}", "func (this *List) String() string {\n\tes := make([]string, len(this.GetElems()))\n\tfor i, v := range this.GetElems() {\n\t\tes[i] = v.String()\n\t}\n\treturn this.Before.String() + listTypeToString(this.Type) + this.OpenCurly.String() + strings.Join(es, \"\") + this.CloseCurly.String()\n}", "func (rl ResolutionList) String() string {\n\tvar sb strings.Builder\n\tfmt.Fprintf(&sb, \"[\")\n\tsep := \"\"\n\tfor _, r := range rl {\n\t\tfmt.Fprintf(&sb, \"%s%s\", sep, r.asString())\n\t\tsep = \", \"\n\t}\n\tfmt.Fprintf(&sb, \"]\")\n\treturn sb.String()\n}", "func (list List) String() string {\n\tbuilder := strings.Builder{}\n\n\tfor _, t := range list {\n\t\tbuilder.WriteString(t.String())\n\t}\n\n\treturn builder.String()\n}", "func (l *FileList) String() string {\n\treturn strings.Join(*l, \", \")\n}", "func (l Slist) ToString() string {\n\ts := l.listToSlice()\n\treturn fmt.Sprint(s)\n}", "func (i Items) String() string {\n\tvar s []string\n\tfor _, item := range i.Items {\n\t\ts = append(s, fmt.Sprintf(\"[%v]: %v\", item.Code, item.Description))\n\t}\n\treturn strings.Join(s, \", \")\n}", "func (al *AddrList) String() string {\n\treturn al.csv\n}", "func (l *List) String() string {\n\tvar s = \"\"\n\tfor curr := l.lastElement; ; curr = curr.Prev {\n\t\ts = fmt.Sprintf(\"%v\", curr.Value) + s\n\t\tif curr.Prev == nil {\n\t\t\treturn s\n\t\t}\n\t}\n}", "func (list *ArrayList[T]) String() string {\n\treturn fmt.Sprint(list.elems)\n}", "func (a *StringSlice) String() string {\n\treturn strings.Join(*a, \",\")\n}", "func (rs Replacements) String() string {\n\tvar ss []string\n\tfor _, r := range rs {\n\t\ts := r.String()\n\t\tss = append(ss, s)\n\t}\n\treturn strings.Join(ss, \",\")\n}", "func (f *StringSlice) String() string {\n\treturn strings.Join(*f, \",\")\n}", "func (c Collection) String() string {\n\tif c == nil {\n\t\treturn \"\"\n\t}\n\n\tvalues := make([]string, 0, len(c))\n\tfor elem := range c {\n\t\tvalues = append(values, elem)\n\t}\n\n\tsort.Strings(values)\n\n\treturn strings.Join(values, \",\")\n}", "func (l *ExprList) String() string {\n\tvar buf bytes.Buffer\n\tbuf.WriteString(\"(\")\n\tfor i, expr := range l.Exprs {\n\t\tif i != 0 {\n\t\t\tbuf.WriteString(\", \")\n\t\t}\n\t\tbuf.WriteString(expr.String())\n\t}\n\tbuf.WriteString(\")\")\n\treturn buf.String()\n}", "func (rl RoarList) String() string {\n var result string\n for i, r := range roarList.Data() {\n result += fmt.Sprintf(\"[%d]: %v\\n\", i, r)\n }\n return result\n}", "func (s Set) FormatList() string {\n\tif len(s) == 0 {\n\t\treturn \"\"\n\t}\n\n\t// Collect members in a sorted list.\n\tm := s.SortedMembers()\n\n\t// Build list\n\tlist := \"\"\n\tfor len(m) > 0 {\n\t\t// Determine length of run.\n\t\tn := 0\n\t\tfor ; n < len(m) && m[n] == m[0]+uint(n); n++ {\n\t\t}\n\t\t// Append to list.\n\t\tif list != \"\" {\n\t\t\tlist += \",\"\n\t\t}\n\t\tlist += strconv.FormatUint(uint64(m[0]), 10)\n\t\tif n > 1 {\n\t\t\tlist += \"-\" + strconv.FormatUint(uint64(m[n-1]), 10)\n\t\t}\n\t\t// Advance.\n\t\tm = m[n:]\n\t}\n\n\treturn list\n}", "func (m *MultiString) String() string {\n\treturn strings.Join(*m, \",\")\n}", "func (l LiteralExprList) String() string {\n\tvar b strings.Builder\n\n\tb.WriteRune('[')\n\tfor i, e := range l {\n\t\tif i > 0 {\n\t\t\tb.WriteString(\", \")\n\t\t}\n\t\tb.WriteString(stringutil.Sprintf(\"%v\", e))\n\t}\n\tb.WriteRune(']')\n\n\treturn b.String()\n}", "func (nl *NodeList) String() string {\n\tif nl.front == nil {\n\t\treturn \"(Empty list)\"\n\t}\n\n\ts := \"\"\n\tfor n := nl.Front(); n != nil; n = n.Next {\n\t\ts += n.String()\n\t\tif n.Next != nil {\n\t\t\ts += \", \"\n\t\t}\n\t}\n\treturn s\n}", "func (s *Set) String() string {\n\treturn strings.Join(s.Slice(), \",\")\n}", "func getAddressListString(addresses []mail.Address) string {\n\tvar addressStrings []string\n\n\tfor _, address := range addresses {\n\t\taddressStrings = append(addressStrings, address.String())\n\t}\n\treturn strings.Join(addressStrings, \",\"+crlf+\" \")\n}", "func (list List) ToString() string {\n\tstr := \"List(\"\n\tfor _, val := range list {\n\t\tstr += fmt.Sprintf(\"%v\", val)\n\t}\n\treturn str + \")\"\n}", "func (a *ArrayString) String() string {\n\taEscaped := make([]string, len(*a))\n\tfor i, v := range *a {\n\t\tif strings.ContainsAny(v, `,'\"{[(`+\"\\n\") {\n\t\t\tv = fmt.Sprintf(\"%q\", v)\n\t\t}\n\t\taEscaped[i] = v\n\t}\n\treturn strings.Join(aEscaped, \",\")\n}", "func (d deck) toString() string {\n\t//strings is a package with function Join, which takes 2 arguments; we are joining the multiple strings w a comma to build one string\n\treturn strings.Join([]string(d), \",\")\n}", "func (list *LinkList) ToStr() string {\n\n\t// # c,cnt are limiters to make sure we don't go run away\n\t// # yes, we need cnt=self.count() and not -1, as we walk off list\n\tc, cnt := 0, list.Count()\n\tbuf := \"\"\n\tptr := list.head\n\tfor ptr != nil {\n\t\ts := fmt.Sprintf(\"%v\", ptr.value)\n\t\tbuf = buf + s + \",\"\n\t\tptr = ptr.next\n\t\tc++\n\t\tif c > cnt {\n\t\t\tfmt.Fprintf(os.Stderr, \"WAIT!!! Forever Loop!\\nRecursive LinkList/Node\\nbuf:[{buf}]\")\n\t\t\tos.Exit(1)\n\t\t}\n\t}\n\n\tif strings.HasSuffix(buf, \",\") {\n\t\tbuf = buf[:len(buf)-1]\n\t}\n\n\treturn buf\n}", "func intsToCommaString(ints []int) string {\n\ts := IntsToStrings(ints)\n\treturn strings.Join(s, \",\")\n}", "func (l HashChainLinkIdentityList) String() string {\n\tif len(l) == 0 {\n\t\treturn \"\"\n\t}\n\n\tvar b strings.Builder\n\tfor i, id := range l {\n\t\tb.WriteString(strconv.FormatInt(int64(i), 10))\n\t\tb.WriteString(\". Identity: \")\n\t\tb.WriteString(id.String())\n\t\tb.WriteString(\"\\n\")\n\t}\n\treturn b.String()\n}", "func (v *VList) String() string {\n\ts := \"(\"\n\tfor i, item := range v.value {\n\t\tif i > 0 {\n\t\t\ts += \" \"\n\t\t}\n\t\ts += item.String()\n\t}\n\treturn s + \")\"\n}", "func (me TransformListType) String() string { return xsdt.String(me).String() }", "func (v DataSlice) String() string { return StringSlice(\", \", \"[\", \"]\", v.Slice()...) }", "func (lst List) String() string {\n\titems := \"\"\n\tnode := lst.head\n\tcounter := Number(0)\n\tfor node != nil {\n\t\tif counter < lst.Len()-1 {\n\t\t\titems += fmt.Sprintf(\"%s \", node.value)\n\t\t} else {\n\t\t\titems += fmt.Sprint(node.value)\n\t\t}\n\t\tcounter++\n\t\tnode = node.next\n\t}\n\treturn fmt.Sprintf(\"(%s)\", items)\n}", "func (lis Config) StringList() string {\n\tvar buf strings.Builder\n\tfor _, o := range lis {\n\t\tbuf.WriteString(o.Space)\n\t\tbuf.WriteRune('\\n')\n\t}\n\treturn buf.String()\n}", "func (s Slice) String() string {\n\treturn s.ToString(\", \")\n}", "func (enums EnumSlice) String() string {\n\treturn strings.Join(enums.StringSlice(), \", \")\n}", "func (c ApiClientListResult) String() (out string) {\n\tfor k, v := range c {\n\t\tout += fmt.Sprintf(\"%s => %s\\n\", k, v)\n\t}\n\treturn out\n}", "func (list *List) String() string {\n\tstr := \"SinglyLinkedList\\n\"\n\tvalues := []string{}\n\tfor element := list.first; element != nil; element = element.next {\n\t\tvalues = append(values, fmt.Sprintf(\"%v\", element.value))\n\t}\n\tstr += strings.Join(values, \", \")\n\treturn str\n}", "func CucumberTagExclusionsListToString(tags []string) string {\n\tvar tagList []string\n\tfor _, tag := range tags {\n\t\ttagList = append(tagList, \"~@\"+tag)\n\t}\n\treturn strings.Join(tagList, \" && \")\n}", "func (vl *ListValue) String() string {\n\treturn fmt.Sprintf(\"%v\", vl.list)\n}", "func (m *MutipleValues) String() string {\n\tif m == nil {\n\t\treturn \"\"\n\t}\n\treturn strings.Join([]string(*m), \";\")\n}", "func (pl List) Format(buf *bytes.Buffer) {\n\tfor i, p := range pl {\n\t\tif i > 0 {\n\t\t\tbuf.WriteString(\", \")\n\t\t}\n\t\tbuf.WriteString(p.String())\n\t}\n}", "func SliceToString(imploded []string) string {\n\tret := \"\"\n\tfor i, item := range imploded {\n\t\tif i > 0 {\n\t\t\tret = ret + \",\"\n\t\t}\n\t\tret = ret + item\n\t}\n\treturn ret\n}", "func (el EpisodeList) String() string {\n\tscales := map[EpisodeType]int{}\n\n\tfor _, er := range el {\n\t\tif er == nil {\n\t\t\tcontinue\n\t\t}\n\n\t\ts := er.scale()\n\t\tif s > scales[er.Type] {\n\t\t\tscales[er.Type] = s\n\t\t}\n\t}\n\n\tparts := make([]string, len(el))\n\tfor i, er := range el {\n\t\tparts[i] = er.Format(scales[er.Type])\n\t}\n\n\treturn strings.Join(parts, \",\")\n}", "func (urls NodeURLs) String() string {\n\tvar xs []string\n\tfor _, u := range urls {\n\t\txs = append(xs, u.String())\n\t}\n\treturn strings.Join(xs, \",\")\n}", "func (o Strings) String() string {\n\treturn strings.Join(o, \", \")\n}", "func DeviceListString(l object.VirtualDeviceList) string {\n\tvar names []string\n\tfor _, d := range l {\n\t\tif d == nil {\n\t\t\tnames = append(names, \"<nil>\")\n\t\t} else {\n\t\t\tnames = append(names, l.Name(d))\n\t\t}\n\t}\n\treturn strings.Join(names, \",\")\n}", "func (l *List) String() string {\n\treturn fmt.Sprintf(\"%v\", l.entries)\n}", "func (c ContainerListResult) String() (out string) {\n\tfor k, v := range c {\n\t\tout += fmt.Sprintf(\"%s => %s\\n\", k, v)\n\t}\n\treturn out\n}", "func (g Group) String() (s string) {\n\tfor i := range g.str {\n\t\ts += g.str[i]\n\t\tif i < len(g.str)-1 {\n\t\t\ts += \", \"\n\t\t}\n\t}\n\treturn\n}", "func (vl ListValue) String() string {\n\treturn fmt.Sprintf(\"%v\", []interface{}(vl))\n}", "func (ls Set) String() string {\n\tselector := make([]string, 0, len(ls))\n\tfor key, value := range ls {\n\t\tselector = append(selector, key+\"=\"+value)\n\t}\n\t// Sort for determinism.\n\tsort.StringSlice(selector).Sort()\n\treturn strings.Join(selector, \",\")\n}", "func (ltype *ListType) String() string {\n\treturn \"[\" + ltype.Type.String() + \"]\"\n}", "func (r *DNSResolvers) String() string {\n\treturn strings.Join(*r, \",\")\n}", "func (c *Combination) ToString() string {\n\treturn strings.Join(c.Numbers, \", \")\n}", "func toAddrString(addresses []string) string {\n\n\tif len(addresses) < 1 {\n\t\treturn \"\"\n\t}\n\n\tvar ret = \"\"\n\t//loop through all but the last element in the address slice, appending them to the ret string with a comma afterwards\n\tfor _, v := range addresses[:len(addresses)-1] {\n\t\tret += v + \", \"\n\t}\n\t//append the last address to the ret string & return the list of addresses\n\tret += addresses[len(addresses)-1]\n\treturn ret\n}", "func (a *ArrayDuration) String() string {\n\tx := a.a\n\tformattedResults := make([]string, len(x))\n\tfor i, v := range x {\n\t\tformattedResults[i] = v.String()\n\t}\n\treturn strings.Join(formattedResults, \",\")\n}", "func (self dirset) String() string {\n\tkeys := make([]string, 0, len(self))\n\tfor key := range self {\n\t\tkeys = append(keys, key)\n\t}\n\treturn \"{\" + strings.Join(keys, \",\") + \"}\"\n}", "func (bil BootImageList) String() string {\n\ts := make([]string, 0, len(bil))\n\tfor _, image := range bil {\n\t\ts = append(s, image.String())\n\t}\n\treturn strings.Join(s, \", \")\n}", "func (me TFeatureListType) String() string { return xsdt.String(me).String() }", "func (l Links) String() string {\n\tvar strs []string\n\tfor _, link := range l {\n\t\tstrs = append(strs, link.String())\n\t}\n\treturn strings.Join(strs, \" , \")\n}", "func (list MyLinkedList) String() string {\n\tvals := []string{}\n\tp := list.head\n\tfor nil != p {\n\t\tvals = append(vals, fmt.Sprintf(\"%d\", p.val))\n\t\tp = p.next\n\t}\n\treturn strings.Join(vals, \", \")\n}", "func (a Amount) StringComma() string {\n\treturn strings.Replace(a.String(), \".\", \",\", -1)\n}", "func (f *Int64Slice) String() string {\n\tslice := make([]string, len(*f))\n\tfor i, v := range *f {\n\t\tslice[i] = strconv.FormatInt(v, 10)\n\t}\n\n\treturn strings.Join(slice, \",\")\n}", "func (l *List) JSON() string {\n\tstr := &strings.Builder{}\n\n\tstr.WriteByte('[')\n\n\tfor i, item := range l.value {\n\t\tif i > 0 {\n\t\t\tstr.WriteString(\", \")\n\t\t}\n\n\t\tstr.WriteString(item.JSON())\n\t}\n\n\tstr.WriteByte(']')\n\n\treturn str.String()\n}", "func (d deck) toString() string {\n\treturn strings.Join([]string(d), delimiterChar)\n}", "func (a Vector) String() string {\n\ts := make([]string, len(a))\n\tfor k := range a {\n\t\ts[k] = fmt.Sprintf(\"%g\", a[k])\n\t}\n\treturn fmt.Sprintf(\"[%s]\", strings.Join(s, \", \"))\n}", "func (tasklist TaskList) String() string {\n\tvar sb strings.Builder\n\tfor _, task := range tasklist {\n\t\tsb.WriteString(task.String())\n\t\tsb.WriteString(ys.NewLine)\n\t}\n\treturn sb.String()\n}", "func (as AddrRanges) String() string {\n\tvar ss []string\n\tfor _, a := range as {\n\t\ts := a.String()\n\t\tss = append(ss, s)\n\t}\n\treturn strings.Join(ss, \",\")\n}", "func (n NamespaceSearch) String() string {\n\tlis := make([]string, 0, len(n))\n\n\tfor _, v := range n {\n\t\tlis = append(lis, v.String())\n\t}\n\treturn strings.Join(lis, \",\")\n}", "func (l *LList) String() string {\n\tstr := \"\"\n\tcurr := l.head\n\n\tfor curr != nil {\n\t\tstr += strconv.Itoa(curr.val) + \" \"\n\t\tcurr = curr.next\n\t}\n\treturn strings.TrimRight(str, \" \")\n}", "func JoinNickComma(nl []IrcNick) string {\n\tnickformated := \"\"\n\tfor i, v := range nl {\n\t\tif i == 0 {\n\t\t\tnickformated = string(v)\n\t\t} else {\n\t\t\tnickformated += fmt.Sprintf(\",%s\", v)\n\t\t}\n\t}\n\n\treturn nickformated\n}", "func (f fieldSliceFlag) String() string {\n\tr := make([]string, len(f))\n\tfor i, s := range f {\n\t\tr[i] = string(s)\n\t}\n\treturn strings.Join(r, \", \")\n}", "func (s *StringList) String() string {\n\treturn fmt.Sprintf(\"%v\", *s)\n}", "func (l *lockConcurrentList) String() string {\n\tl.mutex.RLock()\n\tdefer l.mutex.RUnlock()\n\treturn fmt.Sprint(l.list)\n}", "func (s Permutation) String() string {\n\tres := make([]string, 2)\n\tres[0] = \"Uuid: \" + reform.Inspect(s.Uuid, true)\n\tres[1] = \"Data: \" + reform.Inspect(s.Data, true)\n\treturn strings.Join(res, \", \")\n}", "func CommaSeparatedStrings(list []string) string {\n\tvalues := make([]string, 0, len(list))\n\tfor _, value := range list {\n\t\tvalues = append(values, fmt.Sprintf(`\"%s\"`, value))\n\t}\n\treturn strings.Join(values, \", \")\n}", "func (i *IntList) String() string { return fmt.Sprintf(\"%v\", *i) }" ]
[ "0.73304474", "0.73099434", "0.7156719", "0.7094384", "0.7035619", "0.7017735", "0.6919301", "0.68560153", "0.6788458", "0.67519957", "0.67519957", "0.67519957", "0.67519957", "0.67519957", "0.67519957", "0.6743651", "0.6730801", "0.67237306", "0.67155063", "0.6709709", "0.6648324", "0.66353804", "0.6621236", "0.66211545", "0.6612107", "0.6598261", "0.6596988", "0.65760857", "0.6570902", "0.65692264", "0.65509367", "0.65034395", "0.6489023", "0.6476167", "0.6463698", "0.644733", "0.6439068", "0.6420889", "0.6395083", "0.6386434", "0.63746196", "0.63653946", "0.6350349", "0.63477254", "0.6346533", "0.63380283", "0.63350743", "0.6328872", "0.6324182", "0.63073945", "0.6285063", "0.62842983", "0.62820923", "0.6280925", "0.6279967", "0.627471", "0.6272113", "0.62513435", "0.62449646", "0.62374735", "0.62334377", "0.6229929", "0.62235194", "0.62233084", "0.6211701", "0.61988586", "0.61966795", "0.61852217", "0.6170548", "0.61596096", "0.6157324", "0.61554116", "0.6153645", "0.6128739", "0.6120058", "0.61025876", "0.6089625", "0.60677004", "0.60656863", "0.6042918", "0.6035736", "0.6033482", "0.60049134", "0.6004469", "0.6001196", "0.59909093", "0.5982335", "0.5978569", "0.59768087", "0.5945334", "0.5945266", "0.5934242", "0.59338146", "0.5928606", "0.59176713", "0.59164184", "0.59045094", "0.59009343", "0.5900845", "0.5891797" ]
0.5904232
97
cartFarmacy, cartSupermarket, cartStore, cartBanck
func initObjetos() { //objetos barbijo.FrameOX = 0 barbijo.FrameOY = 62 barbijo.FrameNum = 1 barbijo.FrameWidth = 38 barbijo.FrameHeight = 17 barbijo.X = float64(650) barbijo.Y = float64(150) barbijo.img, _, err = ebitenutil.NewImageFromFile(`data/objetos.png`, ebiten.FilterDefault) if err != nil { log.Fatal(err) } plasma.FrameOX = 0 plasma.FrameOY = 79 plasma.FrameNum = 1 plasma.FrameWidth = 30 plasma.FrameHeight = 48 plasma.X = float64(489) plasma.Y = float64(73) plasma.img = barbijo.img alchol.FrameOX = 0 alchol.FrameOY = 124 alchol.FrameNum = 1 alchol.FrameWidth = 30 alchol.FrameHeight = 39 alchol.X = float64(353) alchol.Y = float64(260) alchol.img = barbijo.img monedas.FrameOX = 0 monedas.FrameOY = 170 monedas.FrameNum = 1 monedas.FrameWidth = 24 monedas.FrameHeight = 30 monedas.X = float64(430) monedas.Y = float64(125) monedas.img = barbijo.img //carteles fondoNegro.FrameOX = 344 fondoNegro.FrameOY = 876 fondoNegro.FrameNum = 1 fondoNegro.FrameWidth = 905 fondoNegro.FrameHeight = 460 fondoNegro.X = float64(305) fondoNegro.Y = float64(230) fondoNegro.img, _, err = ebitenutil.NewImageFromFile(`data/mission-fondo.png`, ebiten.FilterDefault) if err != nil { log.Fatal(err) } fondoNegroPause.FrameOX = 344 fondoNegroPause.FrameOY = 145 fondoNegroPause.FrameNum = 1 fondoNegroPause.FrameWidth = 435 fondoNegroPause.FrameHeight = 155 fondoNegroPause.X = float64(300) fondoNegroPause.Y = float64(150) fondoNegroPause.img = fondoNegro.img fondoNegroVidas1.FrameOX = 0 fondoNegroVidas1.FrameOY = 485 fondoNegroVidas1.FrameNum = 1 fondoNegroVidas1.FrameWidth = 435 fondoNegroVidas1.FrameHeight = 155 fondoNegroVidas1.X = float64(0) fondoNegroVidas1.Y = float64(470) fondoNegroVidas1.img = fondoNegro.img fondoNegroVidas2.FrameOX = 0 fondoNegroVidas2.FrameOY = 485 fondoNegroVidas2.FrameNum = 1 fondoNegroVidas2.FrameWidth = 435 fondoNegroVidas2.FrameHeight = 155 fondoNegroVidas2.X = float64(750) fondoNegroVidas2.Y = float64(470) fondoNegroVidas2.img = fondoNegro.img fondoNegroInmune1.FrameOX = 0 fondoNegroInmune1.FrameOY = 500 fondoNegroInmune1.FrameNum = 1 fondoNegroInmune1.FrameWidth = 435 fondoNegroInmune1.FrameHeight = 50 fondoNegroInmune1.X = float64(0) fondoNegroInmune1.Y = float64(-10) fondoNegroInmune1.img = fondoNegro.img fondoNegroInmune2.FrameOX = 0 fondoNegroInmune2.FrameOY = 500 fondoNegroInmune2.FrameNum = 1 fondoNegroInmune2.FrameWidth = 435 fondoNegroInmune2.FrameHeight = 50 fondoNegroInmune2.X = float64(745) fondoNegroInmune2.Y = float64(-10) fondoNegroInmune2.img = fondoNegro.img fondoNegroFast1.FrameOX = 0 fondoNegroFast1.FrameOY = 500 fondoNegroFast1.FrameNum = 1 fondoNegroFast1.FrameWidth = 435 fondoNegroFast1.FrameHeight = 50 fondoNegroFast1.X = float64(200) fondoNegroFast1.Y = float64(-10) fondoNegroFast1.img = fondoNegro.img fondoNegroFast2.FrameOX = 0 fondoNegroFast2.FrameOY = 500 fondoNegroFast2.FrameNum = 1 fondoNegroFast2.FrameWidth = 435 fondoNegroFast2.FrameHeight = 50 fondoNegroFast2.X = float64(510) fondoNegroFast2.Y = float64(-10) fondoNegroFast2.img = fondoNegro.img fondoNegroCommans.FrameOX = 0 fondoNegroCommans.FrameOY = 0 fondoNegroCommans.FrameNum = 1 fondoNegroCommans.FrameWidth = 600 fondoNegroCommans.FrameHeight = 500 fondoNegroCommans.X = float64(180) fondoNegroCommans.Y = float64(0) fondoNegroCommans.img, _, err = ebitenutil.NewImageFromFile(`data/fondoNegro.png`, ebiten.FilterDefault) if err != nil { log.Fatal(err) } fondoNegroCompras.FrameOX = 0 fondoNegroCompras.FrameOX = 0 fondoNegroCompras.FrameOY = 0 fondoNegroCompras.FrameNum = 1 fondoNegroCompras.FrameWidth = 1100 fondoNegroCompras.FrameHeight = 300 fondoNegroCompras.X = float64(30) fondoNegroCompras.Y = float64(100) fondoNegroCompras.img, _, err = ebitenutil.NewImageFromFile(`data/fondoNegro1.png`, ebiten.FilterDefault) if err != nil { log.Fatal(err) } mhome.FrameOX = 0 mhome.FrameOY = 202 mhome.FrameNum = 1 mhome.FrameWidth = 144 mhome.FrameHeight = 287 mhome.X = float64(0) mhome.Y = float64(200) mhome.img = fondoNegro.img money.FrameOX = 0 money.FrameOY = 600 money.FrameNum = 1 money.FrameWidth = 290 money.FrameHeight = 232 money.X = float64(50) money.Y = float64(50) money.img = fondoNegro.img meds.FrameOX = 297 meds.FrameOY = 310 meds.FrameNum = 1 meds.FrameWidth = 300 meds.FrameHeight = 180 meds.X = float64(315) meds.Y = float64(310) meds.img = fondoNegro.img bread.FrameOX = 975 bread.FrameOY = 140 bread.FrameNum = 1 bread.FrameWidth = 140 bread.FrameHeight = 257 bread.X = float64(950) bread.Y = float64(145) bread.img = fondoNegro.img clothes.FrameOX = 278 clothes.FrameOY = 0 clothes.FrameNum = 1 clothes.FrameWidth = 387 clothes.FrameHeight = 142 clothes.X = float64(270) clothes.Y = float64(0) clothes.img = fondoNegro.img tpaper.FrameOX = 794 tpaper.FrameOY = 610 tpaper.FrameNum = 1 tpaper.FrameWidth = 210 tpaper.FrameHeight = 259 tpaper.X = float64(750) tpaper.Y = float64(35) tpaper.img = fondoNegro.img vaccine.FrameOX = 590 vaccine.FrameOY = 312 vaccine.FrameNum = 1 vaccine.FrameWidth = 360 vaccine.FrameHeight = 182 vaccine.X = float64(600) vaccine.Y = float64(335) vaccine.img = fondoNegro.img }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CartsList(c *cli.Context) {\n\tcolor.Allow(c)\n\n\tconf := config.GetConfig()\n\tdefer conf.Flush()\n\n\tindex := 1\n\tcache := make(map[string]string)\n\tfor _, cart := range conf.Carts {\n\t\tfmt.Printf(\"(%s) %s\\n\", color.ShortID(strconv.Itoa(index)), cart.Name)\n\t\tcache[strconv.Itoa(index)] = cart.Name\n\t\tindex++\n\t}\n\tconf.ResultCache[\"Carts\"] = cache\n}", "func GenerateUniqueCart(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func prepareCartWithDeliveries(t *testing.T, e *httpexpect.Expect) {\n\tt.Helper()\n\thelper.GraphQlRequest(t, e, loadGraphQL(t, \"cart_add_to_cart\", map[string]string{\"MARKETPLACE_CODE\": \"fake_simple\", \"DELIVERY_CODE\": \"delivery1\"})).Expect().Status(http.StatusOK)\n\thelper.GraphQlRequest(t, e, loadGraphQL(t, \"cart_add_to_cart\", map[string]string{\"MARKETPLACE_CODE\": \"fake_simple\", \"DELIVERY_CODE\": \"delivery2\"})).Expect().Status(http.StatusOK)\n}", "func CartList(c *gin.Context) {\n\tid, _ := strconv.Atoi(c.Query(\"cartId\"))\n\tvar CartitemViewModel []Cartitem\n\tif (id > 0) {\n\t\t//database.DBConn.Where(\"cart_id = ?\", id).Find(&Cartitems)\n\t\tdatabase.DBConn.Model(&Cartitem{}).Select(\" cart_id, item_id, count(cart_id) as count \").Group(` cart_id, item_id `).Having(\" cart_id = ?\", id).Find(&CartitemViewModel)\n\t} else {\n\t\tdatabase.DBConn.Model(&Cartitem{}).Select(\" cart_id, item_id, count(cart_id) as count \").Group(` cart_id, item_id `).Find(&CartitemViewModel)\n\t}\t\n\tc.JSON(200, CartitemViewModel)\n}", "func GetCart(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func ShowCart(w http.ResponseWriter, r *http.Request) {\n\tHomeVars := NewHomePageVars(r)\n\tutils.GenerateTemplate(w, HomeVars, \"cart\")\n}", "func (rest *RestApi) GetCart(w http.ResponseWriter, r *http.Request, id int64) error {\n gc := rest.GoCart\n cart := gc.GetCart(id)\n\n bytes, err := json.Marshal(cart)\n if err != nil {\n panic(err)\n }\n\n response := string(bytes)\n fmt.Fprintln(w, response)\n return nil\n}", "func (store *Store) Cart(ctx *gin.Context) (bool, error) {\n\tctx.String(200, \"You have requested the cart.\")\n\treturn true, nil\n}", "func (app *application) customerCart(w http.ResponseWriter, r *http.Request) {\n\tcustomerID := app.authenticatedCustomer(r)\n\tcart := app.carts[customerID]\n\n\tcartRowSlice := make([]cartRow, 0)\n\ttotal := 0\n\n\tfor listID, quantity := range cart {\n\t\tlisting, err := app.listings.Get(listID)\n\t\tif err != nil {\n\t\t\tapp.serverError(w, err)\n\t\t}\n\t\trow := cartRow{\n\t\t\tListingID: listing.ID,\n\t\t\tName: listing.Name,\n\t\t\tPrice: listing.Price,\n\t\t\tQuantity: quantity,\n\t\t\tAmount: quantity * listing.Price,\n\t\t}\n\t\tcartRowSlice = append(cartRowSlice, row)\n\t\ttotal += quantity * listing.Price\n\t}\n\tapp.render(w, r, \"customercart.page.tmpl\", &templateData{\n\t\tCart: cartRowSlice,\n\t\tCartTotal: total,\n\t})\n\treturn\n}", "func EmptyCart(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func (thisCheckout *inventory)CheckoutItems(){\n\tfor {\n\t\tif (thisCheckout.checkoutCokeCount >= 0 || thisCheckout.checkoutPepsiCount >= 0) {\n\t\t\ttime.Sleep(10 * time.Millisecond)\n\t\t\trand.Seed(time.Now().UnixNano())\n\t\t\tvar brandSelector int = (rand.Intn(2)) + 1\n\t\t\tif (brandSelector == 1) {\n\t\t\t\tthisCheckout.checkoutCokeCount = thisCheckout.checkoutCokeCount - 1\n\t\t\t\tthisCheckout.checkoutLedger += .55\n\t\t\t\tfmt.Println(\"1 Coke Can Sold at Checkout\")\n\t\t\t} else if (brandSelector == 2) {\n\t\t\t\tthisCheckout.checkoutPepsiCount = thisCheckout.checkoutPepsiCount - 1\n\t\t\t\tthisCheckout.checkoutLedger += .50\n\t\t\t\tfmt.Println(\"1 Pepsi Can Sold at Checkout\")\n\t\t\t}\n\t\t} else {\n\t\t\ttime.Sleep(10 * time.Millisecond)\n\t\t}\n\t}\n}", "func (rest *RestApi) AddToCart(w http.ResponseWriter, r *http.Request, cart_id int64, item_id int64, quantity int64) {\n\n //@ TODO: Need to check for quantity and increment if necessary\n\n cart := rest.GoCart.GetCart(cart_id)\n\n item := rest.GoCart.GetItem(item_id)\n item.SetItemQuantity(quantity)\n\n cart.Add(*item)\n rest.GoCart.SaveCart(*cart)\n}", "func AddItemToCart(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func handleFuncCart(w http.ResponseWriter, r *http.Request) {\n\tc := appengine.NewContext(r)\n\t\n\tc.Infof(\"handleFuncCarts\")\n\tval, err := handleCarts(c, r)\n\tif err == nil {\n\t\terr = json.NewEncoder(w).Encode(val)\n\t}\n\tif err != nil {\n\t\tw.WriteHeader(400)\n\t\tw.Write([]byte(fmt.Sprintf(\"api error: %#v\", err)))\n\t\treturn\t\n\t}\n}", "func (t *Track) CartAt(x, y int) (c *Cart, ok bool) {\n\tfor _, c := range t.Carts {\n\t\tif c.X == x && c.Y == y {\n\t\t\treturn c, true\n\t\t}\n\t}\n\treturn nil, false\n}", "func (store *Store) AddToCart(ctx *gin.Context) (bool, error) {\n\tctx.String(200, \"You are trying to add items to the cart.\")\n\treturn true, nil\n}", "func (c *Cart) Checkout() {\n\t// TODO\n\tfmt.Println(c.Items)\n fmt.Println(\"Total : \" + c.TotalPrice.getPriceInEuro())\n c = new(Cart)\n}", "func TestCartScenario(t *testing.T) {\n\t\n\te := httpexpect.New(t, API_URL)\n\t\n\tprintComment(\"SC20001\", \"Test Add 2 items of Unlimited 1 GB for $24.90\")\n\tcart := map[string]interface{}{\n\t\t\"code\": \"ult_small\",\n\t\t\"name\": \"Unlimited 1GB\",\n\t\t\"price\": 24.90,\n\t\t\"items\": 2,\n\t}\n\n\te.POST(\"/cart\").\n\t\tWithJSON(cart).\n\t\tExpect().\n\t\tStatus(http.StatusOK)\n\n}", "func (app *application) ShoppingCart(w http.ResponseWriter, r *http.Request) {\r\n\t// a seller does not have a shopping cart\r\n\tisSeller := app.isSeller(r)\r\n\tif isSeller {\r\n\t\tw.WriteHeader(http.StatusUnauthorized)\r\n\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\tError: http.StatusText(http.StatusUnauthorized),\r\n\t\t})\r\n\t\treturn\r\n\t}\r\n\t// retrieves userid from the session cookie\r\n\tuserID := app.session.GetString(r, \"userid\")\r\n\r\n\t// query shoppingcart table in the database\r\n\t// for the list of item in the user's cart\r\n\tcart, err := app.cart.Get(userID)\r\n\tif err != nil {\r\n\t\tw.WriteHeader(http.StatusInternalServerError)\r\n\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\tError: http.StatusText(http.StatusInternalServerError),\r\n\t\t})\r\n\t\treturn\r\n\t}\r\n\r\n\t// display the list of items\r\n\tapp.render(w, r, \"shoppingcart.page.tmpl\", &templateData{\r\n\t\tUser: &models.User{UserID: userID},\r\n\t\tShoppingCart: cart,\r\n\t\tDiscounts: models.Discount,\r\n\t})\r\n}", "func GetCart(c *gin.Context) {\r\n\tvar cart []Models.Cart\r\n\terr := Models.GetCart(&cart)\r\n\tif err != nil {\r\n\t\tc.AbortWithStatus(http.StatusNotFound)\r\n\t} else {\r\n\t\tc.JSON(http.StatusOK, cart)\r\n\t}\r\n}", "func TestCartScenario2(t *testing.T) {\n\t\n\te := httpexpect.New(t, API_URL)\n\t\n\tprintComment(\"SC20001\", \"Test Add 4 items of Unlimited 5 GB for $209.40\")\n\tcart := map[string]interface{}{\n\t\t\"code\": \"ult_large\",\n\t\t\"name\": \"Unlimited 5GB\",\n\t\t\"price\": 44.90,\n\t\t\"items\": 4,\n\t}\n\n\te.POST(\"/cart\").\n\t\tWithJSON(cart).\n\t\tExpect().\n\t\tStatus(http.StatusOK)\n\n}", "func NuevaCartaDeNinja() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Ninja\",\n\t\tDescripción: \"Ninja es considerado un mercenario tipo de guerrero japonés contratado para ejercer asesinatos caracterizado por su gran rapidez, su daño fisico y pobre defensa\",\n\t\tCosto: 1,\n\t\tDañoRojo: 2,\n\t\tDañoAzul: 1,\n\t\tCuración: 0,\n\t\tArmaduraRoja: 1,\n\t\tArmaduraAzul: 1,\n\t\tAntiVelocidad: 1,\n\t\tRango: 1,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 2,\n\t\t\tArmaduraAzulPorNivel: 1,\n\t\t\tCuraciónPorNivel: 0,\n\t\t\tDañoRojoPorNivel: 1,\n\t\t\tDañoAzulPorNivel: 0,\n\t\t},\n\t}\n}", "func CreateCart(name string) error {\n\tcart := domain.NewCart()\n\tcart.SetName(name)\n\tst := storage.NewMemoryStore()\n\treturn st.Save(cart.ID, cart.UncommitedChanges())\n}", "func NuevaCartaDeElfoArquero() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Elfo Arquero\",\n\t\tDescripción: \"Guerros del clan elfico que aprovecha las magia para sus ataques de larga distancia y portar un arco elfico con encantamientos de daño\",\n\t\tCosto: 1,\n\t\tDañoRojo: 6,\n\t\tDañoAzul: 1,\n\t\tCuración: 0,\n\t\tArmaduraRoja: 1,\n\t\tArmaduraAzul: 1,\n\t\tAntiVelocidad: 2,\n\t\tRango: 7,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 1,\n\t\t\tArmaduraAzulPorNivel: 1,\n\t\t\tCuraciónPorNivel: 0,\n\t\t\tDañoRojoPorNivel: 3,\n\t\t\tDañoAzulPorNivel: 0,\n\t\t},\n\t}\n}", "func NuevaCartaDeElfoMago() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Elfo Mago\",\n\t\tDescripción: \"Misteriosos hasta para los otros miembros de clan elfico, usando su magia para llegar hasta donde los otros elfos no han llegado\",\n\t\tCosto: 1,\n\t\tDañoRojo: 1,\n\t\tDañoAzul: 8,\n\t\tCuración: 0,\n\t\tArmaduraRoja: 2,\n\t\tArmaduraAzul: 3,\n\t\tAntiVelocidad: 3,\n\t\tRango: 8,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 1,\n\t\t\tArmaduraAzulPorNivel: 2,\n\t\t\tCuraciónPorNivel: 0,\n\t\t\tDañoRojoPorNivel: 0,\n\t\t\tDañoAzulPorNivel: 2,\n\t\t},\n\t}\n}", "func (p PricedItems) CartItems() map[string]domain.Price {\n\treturn p.cartItems\n}", "func Global(rw http.ResponseWriter, r *http.Request, enc encoding.Encoder, params martini.Params) string {\n\tvar err error\n\terr = setDB(r)\n\tif err != nil {\n\t\tapierror.GenerateError(\"Trouble getting brandID from query string\", err, rw, r)\n\t\treturn \"\"\n\t}\n\terr = setCustomerId(r)\n\tif err != nil {\n\t\tapierror.GenerateError(\"Trouble getting customer from api key\", err, rw, r)\n\t\treturn \"\"\n\t}\n\tpriceType := params[\"type\"]\n\tpercent, err := strconv.ParseFloat(params[\"percentage\"], 64)\n\tif err != nil {\n\t\tapierror.GenerateError(\"Trouble parsing percentage\", err, rw, r)\n\t\treturn \"\"\n\t}\n\tpercent = percent / 100\n\n\t//create partPriceMap\n\tprices, err := cartIntegration.GetPartPrices()\n\tif err != nil {\n\t\tapierror.GenerateError(\"Trouble getting part prices\", err, rw, r)\n\t\treturn \"\"\n\t}\n\tpriceMap := make(map[string]float64)\n\tfor _, p := range prices {\n\t\tkey := strconv.Itoa(p.PartID) + p.Type\n\t\tpriceMap[key] = p.Price\n\t}\n\n\t//get CustPrices\n\tcustPricesJson, err := cartIntegration.GetCustomerPrices(0, 0)\n\tif err != nil {\n\t\tapierror.GenerateError(\"Trouble getting prices by customer ID\", err, rw, r)\n\t\treturn \"\"\n\t}\n\tcustPrices := custPricesJson.Items\n\n\t//set to percentage\n\tfor i, _ := range custPrices {\n\t\tif custPrices[i].CustID == 0 {\n\t\t\tcustPrices[i].CustID = cartIntegration.Customer_ID\n\t\t}\n\t\tcustPrices[i].Price = priceMap[strconv.Itoa(custPrices[i].PartID)+priceType] * percent\n\t\tif custPrices[i].ID == 0 {\n\t\t\terr = custPrices[i].Create()\n\t\t} else {\n\t\t\terr = custPrices[i].Update()\n\n\t\t}\n\t\tif err != nil {\n\t\t\tapierror.GenerateError(\"Trouble updating price\", err, rw, r)\n\t\t\treturn \"\"\n\t\t}\n\t}\n\treturn encoding.Must(enc.Encode(custPrices))\n}", "func (app *application) checkout(w http.ResponseWriter, r *http.Request) {\n\tsession, err := app.sessionStore.Get(r, \"session-name\")\n\tif err != nil {\n\t\tapp.serverError(w, err)\n\t\treturn\n\t}\n\tcustomerID := session.Values[\"customerID\"].(int)\n\tcart := app.carts[customerID]\n\tvendorID := 0\n\n\t// check if all items in cart are from same vendor\n\tfor listingID := range cart {\n\t\tlisting, err := app.listings.Get(listingID)\n\t\tif err != nil {\n\t\t\tapp.serverError(w, err)\n\t\t\treturn\n\t\t}\n\t\tif vendorID == 0 {\n\t\t\tvendorID = listing.VendorID\n\t\t} else if vendorID != listing.VendorID {\n\t\t\tsession.AddFlash(\"Please select items from only one vendor.\")\n\t\t\terr = session.Save(r, w)\n\t\t\tif err != nil {\n\t\t\t\tapp.serverError(w, err)\n\t\t\t\treturn\n\t\t\t}\n\t\t\thttp.Redirect(w, r, \"/customer/checkout\", http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\t}\n\terr = r.ParseForm()\n\tif err != nil {\n\t\tapp.clientError(w, http.StatusBadRequest)\n\t\treturn\n\t}\n\tform := forms.New(r.PostForm)\n\tform.Required(\"drop_long\", \"drop_lat\")\n\tdropLat, err := strconv.ParseFloat(form.Get(\"drop_lat\"), 64)\n\tif err != nil {\n\t\tform.Errors.Add(\"drop_lat\", \"enter valid floating point number\")\n\t}\n\tdropLong, err := strconv.ParseFloat(form.Get(\"drop_long\"), 64)\n\tif err != nil {\n\t\tform.Errors.Add(\"drop_long\", \"enter valid floating point number\")\n\t}\n\tif !form.Valid() {\n\t\tapp.render(w, r, \"checkout.page.tmpl\", &templateData{Form: form})\n\t\treturn\n\t}\n\n\tdeliveryID, err := app.deliveries.Insert(customerID, vendorID, time.Now(), dropLat, dropLong)\n\tif err != nil {\n\t\tapp.serverError(w, err)\n\t\treturn\n\t}\n\tfor listingID, quantity := range cart {\n\t\tlisting, err := app.listings.Get(listingID)\n\t\tif err != nil {\n\t\t\tapp.serverError(w, err)\n\t\t}\n\t\terr = app.orders.Insert(deliveryID, listingID, quantity, listing.Price*quantity)\n\t\tif err != nil {\n\t\t\tapp.serverError(w, err)\n\t\t\treturn\n\t\t}\n\t}\n\n\tsession.AddFlash(\"Order placed! Track your order status here.\")\n\terr = session.Save(r, w)\n\tif err != nil {\n\t\tapp.serverError(w, err)\n\t\treturn\n\t}\n\thttp.Redirect(w, r, \"/customer/activeorders\", http.StatusSeeOther)\n}", "func (store *Store) EmptyCart(ctx *gin.Context) (bool, error) {\n\tctx.String(200, \"You are trying empty the cart.\")\n\treturn true, nil\n}", "func NuevaCartaDeMago() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Mago\",\n\t\tDescripción: \"Considerados por muchos como un hechiceros especializados en la magia y el conosimiento mistico Caracterizados por su daño magico y defensa magica\",\n\t\tCosto: 1,\n\t\tDañoRojo: 1,\n\t\tDañoAzul: 4,\n\t\tCuración: 0,\n\t\tArmaduraRoja: 1,\n\t\tArmaduraAzul: 4,\n\t\tAntiVelocidad: 4,\n\t\tRango: 3,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 2,\n\t\t\tArmaduraAzulPorNivel: 1,\n\t\t\tCuraciónPorNivel: 0,\n\t\t\tDañoRojoPorNivel: 0,\n\t\t\tDañoAzulPorNivel: 3,\n\t\t},\n\t}\n}", "func NuevaCartaDeOgro() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Ogro\",\n\t\tDescripción: \"Un ogro es el miembro de una raza de humanoides grandes, fieros y crueles que comen carne humana\",\n\t\tCosto: 1,\n\t\tDañoRojo: 2,\n\t\tDañoAzul: 0,\n\t\tCuración: 0,\n\t\tArmaduraRoja: 8,\n\t\tArmaduraAzul: 2,\n\t\tAntiVelocidad: 4,\n\t\tRango: 1,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 4,\n\t\t\tArmaduraAzulPorNivel: 2,\n\t\t\tCuraciónPorNivel: 0,\n\t\t\tDañoRojoPorNivel: 1,\n\t\t\tDañoAzulPorNivel: 0,\n\t\t},\n\t}\n}", "func (s *SmartContract) shipGoods(APIstub shim.ChaincodeStubInterface, args []string) sc.Response {\n \n\t lcID := struct {\n\t\t\t LcID string `json:\"lcID\"`\t\n\t\t\t Awb string `json:\"Awb\"`\n\t }{}\n \n\t err := json.Unmarshal([]byte(args[0]),&lcID)\n\t if err != nil {\n\t\t\t return shim.Error(\"Not able to parse args into LC\")\n\t }\n \n\t LCAsBytes, _ := APIstub.GetState(lcID.LcID)\n\t var lc LetterOfCredit\n\t err = json.Unmarshal(LCAsBytes, &lc)\n \n\t if err != nil {\n\t\t\t return shim.Error(\"Issue with LC json unmarshaling\")\n\t }\n \n \n\t // LC := LetterOfCredit{LCId: lc.LCId, ExpiryDate: lc.ExpiryDate, Buyer: lc.Buyer, SellerBank: lc.SellerBank, Seller: lc.Seller, Amount: lc.Amount, Status: \"Rejected\", BuyerBank: lc.BuyerBank}\n\t LC := LetterOfCredit{LCId: lc.LCId, ExpiryDate: lc.ExpiryDate, Buyer: lc.Buyer, SellerBank: lc.SellerBank, Seller: lc.Seller, Amount: lc.Amount,Awb : lcID.Awb , Status: \"Shipped\"}\n\t LCBytes, err := json.Marshal(LC)\n\t if err != nil {\n\t\t\t return shim.Error(\"Issue with LC json marshaling\")\n\t }\n \n\t APIstub.PutState(lc.LCId,LCBytes)\n\t fmt.Println(\"LC ShipStatus -> \", LC)\n \n\t return shim.Success(nil)\n }", "func TestCatalogueCheckout_NewCatalogueCheckout(t *testing.T) {\n\titem1 := mockPricing{price: 10}\n\n\tcheckout := checkout.NewCatalogueCheckout(map[string]pricing.PriceCalculator{\"A\": item1})\n\tassert.NotEmpty(t, checkout)\n}", "func computeBasket(w http.ResponseWriter, r *http.Request) {\n\t// List of id products separates by ,\n\tlistProducts := strings.Split(r.FormValue(\"products\"),\",\")\n\t// List of quantity of each product separates by ,\n\tproducts := make([]product,0,len(listProducts))\n\tfor i,q := range strings.Split(r.FormValue(\"quantities\"),\",\") {\n\t\tif quantity,err := strconv.ParseInt(q,10,32) ; err == nil {\n\t\t\tproducts = append(products,product{listProducts[i],int(quantity)})\n\t\t}\n\t}\n\n\tw.Write([]byte(\"0\"))\n\n}", "func RemoveItemFromCart(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func (m *PgModel) GetCartCouponsByCartUUID(ctx context.Context, cartUUID string) ([]*CartCouponJoinRow, error) {\n\t// 1. Check the cart exists\n\tq1 := \"SELECT id FROM cart WHERE uuid = $1\"\n\tvar cartID int\n\terr := m.db.QueryRowContext(ctx, q1, cartUUID).Scan(&cartID)\n\tif err == sql.ErrNoRows {\n\t\treturn nil, ErrCartNotFound\n\t}\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"postgres: query row context failed for q1=%q\", q1)\n\t}\n\n\t// 2. Get all the coupons applied to this cart.\n\tq2 := `\n\t\tSELECT\n\t\t p.id, p.uuid, p.cart_id, c.uuid as cart_uuid, p.coupon_id,\n\t\t u.uuid as coupon_uuid, u.coupon_code as coupon_code,\n\t\t r.uuid as promo_rule_uuid, p.created, p.modified\n\t\tFROM cart_coupon AS p\n\t\tINNER JOIN cart AS c\n\t\t ON c.id = p.cart_id\n\t\tINNER JOIN coupon AS u\n\t\t ON u.id = p.coupon_id\n\t\tINNER JOIN promo_rule AS r\n\t\t ON r.id = u.promo_rule_id\n\t\tWHERE p.cart_id = $1\n\t`\n\trows, err := m.db.QueryContext(ctx, q2, cartID)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"postgres: m.db.QueryContext(ctx) failed\")\n\t}\n\tdefer rows.Close()\n\n\tcartCoupons := make([]*CartCouponJoinRow, 0)\n\tfor rows.Next() {\n\t\tvar c CartCouponJoinRow\n\t\terr := rows.Scan(&c.id, &c.UUID, &c.cartID, &c.CartUUID, &c.couponID,\n\t\t\t&c.CouponUUID, &c.CouponCode, &c.PromoRuleUUID, &c.Created, &c.Modified)\n\t\tif err == sql.ErrNoRows {\n\t\t\treturn nil, ErrPriceListNotFound\n\t\t}\n\t\tif err != nil {\n\t\t\treturn nil, errors.Wrap(err, \"postgres: scan failed\")\n\t\t}\n\t\tcartCoupons = append(cartCoupons, &c)\n\t}\n\tif err := rows.Err(); err != nil {\n\t\treturn nil, errors.Wrap(err, \"postgres: rows.Err()\")\n\t}\n\treturn cartCoupons, nil\n\n}", "func (ch *CartHandler) Query(w http.ResponseWriter, r *http.Request) {\n\n\tp, err := ch.Cart.FetchAll()\n\tif err != nil {\n\t\tRespondWithError(w, http.StatusInternalServerError, err)\n\t}\n\n\tRespond(w, http.StatusOK, p)\n}", "func (storeShelf *inventory)CokeDelivery(){\n\tfor i := 0; i < 20; i++ {\n\t\tstoreShelf.inventoryCokeCount +=24\n\t\tstoreShelf.costCoke += 6.0\n\t\ttime.Sleep(5000 * time.Millisecond)\n\t\tfmt.Println(\"24 cans of Coke added to the shelf by the stocker\")\n\t}\n}", "func shop(s salesman) {\n\tfmt.Println(s)\n\ts.sell()\n\ts.show()\n}", "func (thisInv *inventory)StockShelf() { // Decrement inventory where getting it from it \n\tfor (thisInv.inventoryCokeCount >= 0 || thisInv.inventoryPepsiCount >= 0) {\n\t\ttime.Sleep(10 * time.Millisecond)\n\t\trand.Seed(time.Now().UnixNano())\n\t\tvar brandSelector int = (rand.Intn(2)) + 1\n\t\tif (brandSelector == 1) {\n\t\t\tthisInv.shelfCokeCount += 1\n\t\t\tfmt.Println(\"1 Coke Can Placed on Shelf\")\n\t\t} else if (brandSelector == 2) {\n\t\t\tthisInv.shelfPepsiCount += 1\n\t\t\tfmt.Println(\"1 Pepsi Can Placed on Shelf\")\n\t\t}\n\t}\n}", "func (app *application) CheckOut(w http.ResponseWriter, r *http.Request) {\r\n\t// retrieve userid from session cookie\r\n\tuserid := app.session.GetString(r, \"userid\")\r\n\r\n\t// check that every item in the cart has legal qty\r\n\t// retrieve information from database\r\n\tshoppingcart, err := app.cart.CheckOut(userid)\r\n\tif err != nil {\r\n\t\tw.WriteHeader(http.StatusInternalServerError)\r\n\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\tError: http.StatusText(http.StatusInternalServerError),\r\n\t\t})\r\n\t\treturn\r\n\t}\r\n\r\n\t// loop through the results to check if every item\r\n\t// passes the check\r\n\tvar pass bool = true\r\n\tfor _, item := range shoppingcart {\r\n\t\tif item.Invalid {\r\n\t\t\tpass = false\r\n\t\t}\r\n\t}\r\n\r\n\t// if any item fails the check, it is flagged\r\n\tif !pass {\r\n\t\tapp.render(w, r, \"shoppingcart.page.tmpl\", &templateData{\r\n\t\t\tShoppingCart: shoppingcart,\r\n\t\t})\r\n\t} else {\r\n\t\t// else proceed to create an order for every item\r\n\t\tfor _, v := range shoppingcart {\r\n\t\t\tv.UserID = userid\r\n\t\t\terr := app.orders.Create(v)\r\n\t\t\tif err != nil {\r\n\t\t\t\tw.WriteHeader(http.StatusInternalServerError)\r\n\t\t\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\t\t\tError: http.StatusText(http.StatusInternalServerError),\r\n\t\t\t\t})\r\n\t\t\t\treturn\r\n\t\t\t}\r\n\t\t}\r\n\r\n\t\t// then delete the user's shopping cart\r\n\t\terr = app.cart.DeleteAll(userid)\r\n\t\tif err != nil {\r\n\t\t\tw.WriteHeader(http.StatusInternalServerError)\r\n\t\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\t\tError: http.StatusText(http.StatusInternalServerError),\r\n\t\t\t})\r\n\t\t\treturn\r\n\t\t}\r\n\r\n\t\tapp.render(w, r, \"success.page.tmpl\", &templateData{\r\n\t\t\tUser: &models.User{UserID: userid},\r\n\t\t})\r\n\t}\r\n}", "func NuevaCartaDeArqueroHumano() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Arquero Humano\",\n\t\tDescripción: \"Humanos del antiguo clan de los Exiliados que disfrutanban llevar a la locura a sus victimas disparandoles flechzas hasta asesinarlos\",\n\t\tCosto: 1,\n\t\tDañoRojo: 8,\n\t\tDañoAzul: 1,\n\t\tCuración: 0,\n\t\tArmaduraRoja: 2,\n\t\tArmaduraAzul: 1,\n\t\tAntiVelocidad: 3,\n\t\tRango: 6,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 2,\n\t\t\tArmaduraAzulPorNivel: 1,\n\t\t\tCuraciónPorNivel: 0,\n\t\t\tDañoRojoPorNivel: 2,\n\t\t\tDañoAzulPorNivel: 0,\n\t\t},\n\t}\n}", "func client(propositions *Space,emballages *Space,transports *Space, sync *Space){\n\tvar productname,req string\n\tvar time,qty,cost int\n\tfmt.Printf(\"Que desirez vous produire ?\\n\")\n\tfmt.Scanln(&productname)\n\tfmt.Printf(\"Caractéristiques Techniques ?\\n\")\n\tfmt.Scanln(&req)\n\tfmt.Printf(\"en combien de temps ?\\n\")\n\tfmt.Scanln(&time)\n\tfmt.Printf(\"Combien d'exemplaires désirez vous produire ?\\n\")\n\tfmt.Scanln(&qty)\n\tfmt.Printf(\"Budget souhaité (en euros) ?\\n\")\n\tfmt.Scanln(&cost)\n\tpropositions.Put(productname,req,cost,time,qty)\n\tgo first_appel(propositions,sync)\n\tgo second_appel(emballages,sync)\n\tgo trois_appel(transports,sync)\n}", "func AddCart() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\trequestBody := model.ChatfuelCarts{}\n\t\tc.Bind(&requestBody)\n\n\t\tcart := model.Carts{\n\t\t\tMessengerUserID: requestBody.MessengerUserID,\n\t\t\tFirstName: requestBody.FirstName,\n\t\t\tProductID: requestBody.ProductID,\n\t\t\tProductName: requestBody.ProductName,\n\t\t\tQty: requestBody.Qty,\n\t\t\tPrice: requestBody.Price,\n\t\t}\n\n\t\tdb.Db.Create(&cart)\n\n\t\ttext := []model.Text{}\n\t\ttext = append(text, model.Text{\n\t\t\tText: \"加入購物車成功\",\n\t\t})\n\n\t\tmessage := model.Message{\n\t\t\tMessage: text,\n\t\t}\n\n\t\tc.JSON(http.StatusOK, message)\n\t}\n}", "func (c *Carta) Init(id, tipo, era, puntos, edificioGratis, cartaRequerida int, nombre string, produce, requiere [CANTIDAD_RECURSOS]int){\n c.Id = id\n c.Tipo = tipo\n c.era = era\n c.puntos = puntos\n c.edificioGratis = edificioGratis\n c.Nombre = nombre\n c.Produce = produce\n c.requiere = requiere\n c.cartaRequerida = cartaRequerida\n}", "func (handler *Handler) createShoppingCart(w http.ResponseWriter, r *http.Request, ps httprouter.Params) {\n\tuser, err := handler.authUser(r)\n\tif err != nil {\n\t\thandler.Error(w, r, err)\n\t\treturn\n\t}\n\n\tcart := shoppingcart.ShoppingCart{\n\t\tUserID: user.ID,\n\t}\n\n\tif err := handler.shoppingCartService.Create(r.Context(), &cart); err != nil {\n\t\thandler.Error(w, r, err)\n\t\tlogrus.Errorf(\"Unable to create shopping cart for user %d: %s\", user.ID, err)\n\t\treturn\n\t}\n\n\tw.WriteHeader(http.StatusCreated)\n\tif err := json.NewEncoder(w).Encode(cart); err != nil {\n\t\tlogrus.Errorf(\"Unable to respond with cart %s\", err)\n\t}\n}", "func AllskuObO(assets []structs.UniqueSpaces, skus structs.SKU) {\n\tfmt.Println(\"Here i Am\")\n\tleastSpillage := \"false\"\n\tqty := -1\n\tlastVol := -1\n\tlastSpillPercent := -1.00\n\tlastFreePercent := 100.00\n\titem_in_single_asset := 0\n\n\tfor i := 0; i <= len(assets); i++ {\n\n\t\tif i >= len(assets) {\n\t\t\tfmt.Println(\"Here will be the final answer\")\n\t\t\tfmt.Printf(`\n\t\t\t\t*****************************\n\t\t\t\tleastSpillage = %s\n\t\t\t\tlastVol = %d\n\t\t\t\tqty = %d\n\t\t\t\titem_in_single_asset = %d\n\t\t\t\tlastSpillPercent = %.2f\n\t\t\t\tlastFreePercent = %.2f\n\t\t\t\t*****************************\n\t\t\t`, leastSpillage, lastVol, qty, item_in_single_asset, lastSpillPercent, lastFreePercent)\n\t\t\t// break\n\t\t} else {\n\t\t\tcheckedFactors := checkFact(assets[i], skus, lastVol, lastFreePercent, lastSpillPercent)\n\t\t\tfmt.Println(checkedFactors)\n\t\t\tif checkedFactors.Success {\n\t\t\t\tfmt.Println(\"Ithhe bhi aaya \", assets[i].CType)\n\t\t\t\tleastSpillage = assets[i].CType\n\t\t\t\tlastVol = int(checkedFactors.Vol)\n\t\t\t\tqty = int(checkedFactors.Qty)\n\t\t\t\titem_in_single_asset = int(checkedFactors.ItemInSingleAsset)\n\t\t\t\tlastSpillPercent = checkedFactors.SpillPercent\n\t\t\t\tlastFreePercent = checkedFactors.FreePerc\n\t\t\t}\n\n\t\t}\n\n\t}\n\n}", "func NuevaCartaDeSacerdote() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Sacerdote\",\n\t\tDescripción: \"Los sacerdotes están entregados a lo espiritual sirviendo a la gente con su inquebrantable fe y sus dones místicos dedicados a sanar a sus compañeros en la guerra\",\n\t\tCosto: 1,\n\t\tDañoRojo: 0,\n\t\tDañoAzul: 1,\n\t\tCuración: 3,\n\t\tArmaduraRoja: 2,\n\t\tArmaduraAzul: 1,\n\t\tAntiVelocidad: 4,\n\t\tRango: 4,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 1,\n\t\t\tArmaduraAzulPorNivel: 1,\n\t\t\tCuraciónPorNivel: 2,\n\t\t\tDañoRojoPorNivel: 0,\n\t\t\tDañoAzulPorNivel: 0,\n\t\t},\n\t}\n}", "func NuevaCartaDeGuerrero() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Guerrero\",\n\t\tDescripción: \"Noble caballero de la Edad Media con sus armaduras y armas que se caracteriza por su resistencia y su ataque fisico\",\n\t\tCosto: 1,\n\t\tDañoRojo: 4,\n\t\tDañoAzul: 1,\n\t\tCuración: 0,\n\t\tArmaduraRoja: 4,\n\t\tArmaduraAzul: 2,\n\t\tAntiVelocidad: 4,\n\t\tRango: 1,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 2,\n\t\t\tArmaduraAzulPorNivel: 1,\n\t\t\tCuraciónPorNivel: 0,\n\t\t\tDañoRojoPorNivel: 2,\n\t\t\tDañoAzulPorNivel: 0,\n\t\t},\n\t}\n}", "func plants(propositions *Space, sync *Space) {\n\tvar item,req string\n\tvar time,qty,cost, rcost,rtime,rqty int\n\tt, _ := propositions.Query(&item,&req,&cost,&time,&qty)\n\tfor i := 1; i < NPLANT+1; i++ {\n\t\titem = (t.GetFieldAt(0)).(string)\n\t\treq = (t.GetFieldAt(1)).(string)\n\t\tcost = (t.GetFieldAt(2)).(int)\n\t\ttime = (t.GetFieldAt(3)).(int)\n\t\tqty = (t.GetFieldAt(4)).(int)\n\t\trtime = time-(rand.Intn(time)-1)\n\t\trqty = qty-(rand.Intn(qty)-1)\n\t\trcost = cost-(rand.Intn(cost)-1)\n\t\t//design\n\t\t//workshop\n\t\tpropositions.Put(i ,item, req, rcost, rtime, rqty)\n\t\tfmt.Printf(\"%d) Je propose %s, j'ai besoin de %s, je peux en fabriquer %d exemplaires en %d jours pour la modique somme de %d €\t\t||\t\",i, item , req , rqty , rtime , rcost)\n\t\tfmt.Printf(\"out(%d, %s, %s, %d, %d, %d )\\n\\n\",i, item , req , rcost , rtime , rqty)\n}\nsync.Put(\"appel1\")\n}", "func GetProductInCart() (interface{}, error) {\n\tvar shopping_carts []model.Shopping_cart\n\tif err := config.DB.Find(&shopping_carts).Error; err != nil {\n\t\treturn nil, err\n\t}\n\treturn shopping_carts, nil\n}", "func viewAnyOrderGet(c *gin.Context) { //admin also have the same view , later combine those two func TBD\n\tOrdID := c.Request.URL.Query()[\"ordid\"][0] // Getting Order ID passed with URL\n\t_, usrName := session.SessinStatus(c, \"user_session_cookie\")\n\tfmt.Println(\"Wnat to see the order details of order number \", OrdID)\n\toK, itemsList, date, status, PayMode, amt := db.GetSingleOredrDetails(OrdID)\n\tif !oK {\n\t\tfmt.Println(\"Something went wrong while picking Single Order Deatils ..Please have a look\")\n\t}\n\tfmt.Println(oK, itemsList, date, status, PayMode, amt)\n\t//\t\tsubTotalToFloat, _ := strconv.ParseFloat(singleCartItem.SubTotal, 64)\n\t//\t\tTotalAmt = TotalAmt + subTotalToFloat\n\t//\tTotalAmtInPaisa := TotalAmt * 100 // This is required while initate for payment in Razorpay\n\n\t//\tTotalAmtString := fmt.Sprintf(\"%.2f\", TotalAmt)\n\n\tc.HTML(\n\t\thttp.StatusOK,\n\t\t\"view_particular_order.html\",\n\t\tgin.H{\"title\": \"OrderDetail\",\n\t\t\t\"ItemsOrdered\": itemsList,\n\t\t\t\"OrdID\": OrdID,\n\t\t\t\"date\": date,\n\t\t\t\"PayMode\": PayMode,\n\t\t\t\"amt\": amt,\n\t\t\t\"OrdStatus\": status,\n\t\t\t\"usrName\": usrName,\n\n\t\t\t// \"TotalAmt\": TotalAmtString,\n\t\t\t// \"TotalAmtInPaisa\": TotalAmtInPaisa,\n\t\t},\n\t)\n\n}", "func NuevaCartaDeBrujo() *Carta {\n\treturn &Carta{\n\t\tNombre: \"Brujo\",\n\t\tDescripción: \"Los brujos son entrenados en las artes oscuras estos letales hechiceros usan su magia para ejercer dominacion sobre sus enemigos\",\n\t\tCosto: 1,\n\t\tDañoRojo: 1,\n\t\tDañoAzul: 4,\n\t\tCuración: 1,\n\t\tArmaduraRoja: 1,\n\t\tArmaduraAzul: 1,\n\t\tAntiVelocidad: 4,\n\t\tRango: 3,\n\t\tNivel: 1,\n\t\tExperiencia: 0,\n\t\tEstadísticasPorNivel: &EstadísticasPorNivel{\n\t\t\tCostoPorNivel: 1,\n\t\t\tVidaPorNivel: 1,\n\t\t\tArmaduraRojaPorNivel: 1,\n\t\t\tArmaduraAzulPorNivel: 1,\n\t\t\tCuraciónPorNivel: 1,\n\t\t\tDañoRojoPorNivel: 0,\n\t\t\tDañoAzulPorNivel: 2,\n\t\t},\n\t}\n}", "func GetCartByID(c *gin.Context) {\r\n\tid := c.Params.ByName(\"id\")\r\n\tvar cart []Models.Cart\r\n\terr := Models.GetCartByID(&cart, id)\r\n\tif err != nil {\r\n\t\tc.AbortWithStatus(http.StatusNotFound)\r\n\t} else {\r\n\t\tc.JSON(http.StatusOK, cart)\r\n\t}\r\n}", "func (r *CommerceCartQueryResolver) CommerceCart(ctx context.Context) (*dto.DecoratedCart, error) {\n\treq := web.RequestFromContext(ctx)\n\tdc, err := r.applicationCartReceiverService.ViewDecoratedCart(ctx, req.Session())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn dto.NewDecoratedCart(dc), nil\n}", "func (m *PgModel) AddCartCoupon(ctx context.Context, cartUUID, couponUUID string) (*CartCouponJoinRow, error) {\n\tcontextLogger := log.WithContext(ctx)\n\tcontextLogger.Debugf(\"postgres: AddCartCoupon(ctx context.Context, cartUUID=%q, couponUUID=%q string) started\", cartUUID, couponUUID)\n\n\t// 1. Check the cart exists\n\tq1 := \"SELECT id FROM cart WHERE uuid = $1\"\n\tvar cartID int\n\terr := m.db.QueryRowContext(ctx, q1, cartUUID).Scan(&cartID)\n\tif err == sql.ErrNoRows {\n\t\treturn nil, ErrCartNotFound\n\t}\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"postgres: query row context failed for q1=%q\", q1)\n\t}\n\n\t// 2. Check the coupon exists\n\tq2 := `\n\t\tSELECT\n\t\t c.id, c.coupon_code, void, reusable, spend_count, r.uuid as promo_rule_uuid,\n\t\t r.start_at, r.end_at\n\t\tFROM coupon AS c\n\t\tINNER JOIN promo_rule AS r\n\t\t ON r.id = c.promo_rule_id\n\t\tWHERE c.uuid = $1\n\t`\n\tvar void bool\n\tvar reusable bool\n\tvar spendCount int\n\tvar startAt *time.Time\n\tvar endAt *time.Time\n\n\tvar c CartCouponJoinRow\n\tc.cartID = cartID\n\tc.CartUUID = cartUUID\n\tc.CouponUUID = couponUUID\n\terr = m.db.QueryRowContext(ctx, q2, couponUUID).Scan(&c.couponID, &c.CouponCode, &void, &reusable, &spendCount, &c.PromoRuleUUID, &startAt, &endAt)\n\tif err == sql.ErrNoRows {\n\t\treturn nil, ErrCouponNotFound\n\t}\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"postgres: query row context failed for q2=%q\", q2)\n\t}\n\n\t// 3. Check the coupon has not already been applied to the cart.\n\tq3 := \"SELECT EXISTS(SELECT 1 FROM cart_coupon WHERE cart_id = $1 AND coupon_id = $2) AS exists\"\n\tvar exists bool\n\terr = m.db.QueryRowContext(ctx, q3, cartID, c.couponID).Scan(&exists)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"postgres: m.db.QueryRowContext(ctx, q3=%q, cartID=%d, c.couponID=%d).Scan(...) failed\", q3, cartID, c.couponID)\n\t}\n\n\tif exists {\n\t\treturn nil, ErrCartCouponExists\n\t}\n\n\t// Check if the coupon has been voided\n\tif void {\n\t\tcontextLogger.Debugf(\"postgres: coupon is void (couponUUID=%q)\", couponUUID)\n\t\treturn nil, ErrCouponVoid\n\t}\n\n\t// Check if the coupon has been used. (Only applies to non-reusable coupons)\n\tif !reusable && spendCount > 0 {\n\t\tcontextLogger.Debug(\"postgres: coupon is not reusable and spendCount > 0. The coupon has been already used.\")\n\t\treturn nil, ErrCouponUsed\n\t}\n\n\t// Check if the coupon has expired\n\tif startAt != nil && endAt != nil {\n\t\tnow := time.Now()\n\n\t\tdiffStart := now.Sub(*startAt)\n\n\t\t// if the difference is a negative value,\n\t\t// then the coupon hasn't yet started\n\t\thoursToStart := diffStart.Hours()\n\t\tif hoursToStart < 0 {\n\t\t\tcontextLogger.Infof(\"postgres: coupon is %.1f hours before start at date\", hoursToStart)\n\t\t\treturn nil, ErrCouponNotAtStartDate\n\t\t}\n\n\t\t// if the difference is a positive value,\n\t\t// then the coupon has expired.\n\t\tdiffEnd := now.Sub(*endAt)\n\t\thoursOverEnd := diffEnd.Hours()\n\t\tif hoursOverEnd > 0 {\n\t\t\tcontextLogger.Infof(\"postgres: coupon is %.1f hours over the end at date\", hoursOverEnd)\n\t\t\treturn nil, ErrCouponExpired\n\t\t}\n\n\t\tformat := \"2006-01-02 15:04:05 GMT\"\n\t\tcontextLogger.Infof(\"postgres: coupon is between %s and %s.\", startAt.In(loc).Format(format), endAt.In(loc).Format(format))\n\t} else {\n\t\tcontextLogger.Debugf(\"postgres: startAt for promo rule %q is nil\", c.PromoRuleUUID)\n\t}\n\n\t// 4. Insert the cart coupon.\n\tq4 := `\n\t\tINSERT INTO cart_coupon\n\t\t (cart_id, coupon_id)\n\t\tVALUES\n\t\t ($1, $2)\n\t\tRETURNING\n\t\t id, uuid, cart_id, coupon_id, created, modified\n\t`\n\trow := m.db.QueryRowContext(ctx, q4, cartID, c.couponID)\n\tif err := row.Scan(&c.id, &c.UUID, &c.cartID, &c.couponID, &c.Created, &c.Modified); err != nil {\n\t\treturn nil, errors.Wrapf(err, \"postgres: query scan failed q4=%q\", q4)\n\t}\n\n\treturn &c, nil\n}", "func NewCart() console.Cartridge {\n\treturn &cartridge{\n\t\tBaseCartridge: console.NewBaseCart(),\n\t}\n}", "func CreateCart(cr cart.Repository) http.Handler {\n\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tkey := r.URL.Query().Get(\"key\")\n\t\tif key == \"\" {\n\t\t\thttp.Error(w, \"missing key in query string\", http.StatusBadRequest)\n\t\t\treturn\n\t\t}\n\n\t\tnewCart := cart.New(key)\n\t\terr := cr.Store(newCart)\n\t\tif err != nil {\n\t\t\t//error handling\n\t\t}\n\t\tval := []byte{}\n\t\terr2 := json.Unmarshal(val, newCart)\n\t\tif err2 != nil {\n\t\t\t//\n\t\t}\n\t\tw.WriteHeader(http.StatusCreated)\n\t\tw.Write(val)\n\t})\n}", "func FindAllCartsByCardIDs(cartIDs []int) ([]*Cart, error) {\n\tvar cartItems []*Cart\n\terr := db.Model(&Cart{}).Where(\"ID IN ?\", cartIDs).Find(&cartItems).Error\n\tif err != nil && err != gorm.ErrRecordNotFound {\n\t\treturn nil, err\n\t}\n\n\treturn cartItems, nil\n}", "func (app *application) AddToCart(w http.ResponseWriter, r *http.Request) {\r\n\t// a seller does not have a shopping cart\r\n\tisSeller := app.isSeller(r)\r\n\tif isSeller {\r\n\t\tw.WriteHeader(http.StatusUnauthorized)\r\n\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\tError: http.StatusText(http.StatusUnauthorized),\r\n\t\t})\r\n\t\treturn\r\n\t}\r\n\r\n\t// retrieve userid from session cookie\r\n\tuserid := app.session.GetString(r, \"userid\")\r\n\r\n\t// retrieve ProductID from url\r\n\t// the ProducID should be valid\r\n\tproductID, err := strconv.Atoi(r.URL.Query().Get(\"productid\"))\r\n\tif err != nil {\r\n\t\tw.WriteHeader(http.StatusBadRequest)\r\n\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\tError: http.StatusText(http.StatusBadRequest),\r\n\t\t})\r\n\t\treturn\r\n\t}\r\n\r\n\t// perform the insert at the database\r\n\terr = app.cart.InsertItem(userid, productID)\r\n\tif err != nil {\r\n\t\tapp.errorLog.Println(err)\r\n\t\tw.WriteHeader(http.StatusInternalServerError)\r\n\t\tapp.render(w, r, \"error.page.tmpl\", &templateData{\r\n\t\t\tError: http.StatusText(http.StatusInternalServerError),\r\n\t\t})\r\n\t\treturn\r\n\t}\r\n\r\n\tapp.session.Put(r, \"flash\", \"Product successfully added to cart.\")\r\n\r\n\thttp.Redirect(w, r, r.Referer(), http.StatusSeeOther)\r\n}", "func RetrieveUserCart(service Service, userService users.Service) func(w http.ResponseWriter, r *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx := r.Context()\n\n\t\tusername, err := auth.GetLoggedInUsername(r)\n\t\tif err != nil {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusForbidden, errorcode.ErrorsInRequestData, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tuser, err := userService.RetrieveUserByUsername(ctx, username)\n\t\tif err != nil || user == nil {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusUnauthorized, errorcode.UserNotFound, \"User not found\")\n\t\t\treturn\n\t\t}\n\n\t\tcart, err := service.GetUserCart(ctx, user.ID)\n\t\tif err != nil {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusInternalServerError, \"internal_error\", err.Error())\n\t\t\treturn\n\t\t}\n\t\thttpresponse.RespondJSON(w, http.StatusOK, cart, nil)\n\t}\n}", "func (t *SimpleChaincode) querypkgbyshipper(stub shim.ChaincodeStubInterface, args []string) ([]byte, error){\n\n var jsonResp string\n var err error\n\n if len(args) != 1 {\n jsonResp = \"Error: Incorrect number of arguments. Need to pass Shipper \"\n return nil, errors.New(jsonResp)\n }\n\n valAsbytes, err := stub.GetState(\"PkgIdsKey\")\n if err != nil {\n jsonResp = \"Error: Failed to get state for PkgIdsKey \"\n return nil, errors.New(jsonResp)\n }\n\n var package_holder PKG_Holder\n err = json.Unmarshal(valAsbytes, &package_holder)\n if err != nil {\n fmt.Println(\"Could not marshal personal info object\", err)\n jsonResp = \" Error:Could not marshal personal info object\"\n return nil, errors.New(jsonResp)\n }\n\n var pkginfo PackageInfo\n\n result := \"[\"\n\n var temp []byte\n\n for _, PkgId := range package_holder.PkgIds {\n\n pkginfoasbytes, err := stub.GetState(PkgId)\n if err != nil {\n jsonResp = \"Error:Failed to get state for \" + PkgId\n return nil, errors.New(jsonResp)\n }\n\n err = json.Unmarshal(pkginfoasbytes, &pkginfo);\n if err != nil {\n fmt.Println(\"Could not marshal personal info object\", err)\n jsonResp = \" Error:Could not marshal personal info object\"\n return nil, errors.New(jsonResp)\n }\n\n// check for inout Shipper\n if pkginfo.Shipper == args[0] {\n temp = pkginfoasbytes\n result += string(temp) + \",\"\n }\n\n }\n\n if len(result) == 1 {\n result = \"[]\"\n } else {\n result = result[:len(result)-1] + \"]\"\n }\n\n return []byte(result), nil\n\n}", "func (cartData *CartData) CreateCartItem(CartID, UserID uint64) error {\n\tconn, err := db.MySQLConnect()\n\t// defer conn.Close()\n\n\tif err != nil {\n\t\tlog.Println(err)\n\t\tconn.Close()\n\t\treturn err\n\t}\n\n\tcartItem := CartItem{\n\t\tCartID: CartID,\n\t}\n\n\tlog.Printf(\"Duzina: %d\", len(cartData.Items))\n\tfor _, item := range cartData.Items {\n\t\tcartItem.ItemID = item.ItemID\n\t\tcartItem.Amount = item.Amount\n\t\tlog.Printf(\"ITEMID: %d\", item.ItemID)\n\t\tconn.Create(&cartItem)\n\t\t// if createErr != nil {\n\t\t// \tlog.Println(\"Drugi fail\")\n\t\t// \tlog.Println(createErr.Error)\n\t\t// \tconn.Close()\n\t\t// \treturn createErr.Error\n\t\t// }\n\t\t// portions := GetPortionByCategoryID(item.CategoryID)\n\t\t// image := GetImageByItemID(item.ItemID)\n\t\t// ingredients := GetIngredientsByItemID(item.ItemID)\n\t\t// homeItem := HomeItem{\n\t\t// \tItem: item,\n\t\t// \tPortion: portions,\n\t\t// \tIngredient: ingredients,\n\t\t// \tImage: image,\n\t\t// }\n\t\t// homeItems = append(homeItems, homeItem)\n\t}\n\n\tdeliveryAt := time.Now().Add(DefaultOrderWaitTime).Format(\"H:i:s\")\n\tparsedDeliveryAt, parseErr := time.Parse(\"H:i:s\", deliveryAt)\n\tif parseErr != nil {\n\t\tlog.Println(parseErr.Error())\n\t}\n\n\torder := Order{\n\t\tUserID: UserID,\n\t\tCartID: CartID,\n\t\tIsCanceled: 0,\n\t\tIsDelivered: 0,\n\t\tIsAccepted: \"pending\",\n\t\tDeliveryAt: parsedDeliveryAt,\n\t\tCreatedAt: time.Now(),\n\t}\n\n\tconn.Create(&order)\n\t// if createErr != nil {\n\t// \tlog.Println(\"CreateERR\")\n\t// \tlog.Println(createErr.Error)\n\t// }\n\tconn.Close()\n\n\treturn nil\n}", "func (s *SmartContract) updateProductCustodian(stub shim.ChaincodeStubInterface, args []string) peer.Response {\n\t//get user identity\n\tidentity, err := GetInvokerIdentity(stub)\n\tif err != nil {\n\t\tshim.Error(fmt.Sprintf(\"Error getting invoker identity: %s\\n\", err.Error()))\n\t}\n\n\tif len(args) != 2 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 2\")\n\t}\n\ttrackingID := args[0]\n\tnewLocation := args[1]\n\tnewCustodian := identity.Cert.Subject.String()\n\t//get state by id as key\n\texistingsBytes, _ := stub.GetState(trackingID)\n\n\t// return 404 is not found\n\tif len(existingsBytes) == 0 {\n\t\treturn peer.Response{\n\t\t\tStatus: 404,\n\t\t\tMessage: fmt.Sprintf(\"Item with trackingID %s not found\", trackingID),\n\t\t}\n\t}\n\n\t//try to unmarshal as container\n\tvar product Product\n\terr = json.Unmarshal(existingsBytes, &product)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\t//Ensure user is a participant\n\tif !(product.AccessibleBy(identity)) {\n\t\treturn peer.Response{\n\t\t\tStatus: 403,\n\t\t\tMessage: fmt.Sprintf(\"You are not authorized to perform this transaction, product not accesible by identity\"),\n\t\t}\n\t}\n\t//Ensure new custodian isnt the same as old one\n\tif newCustodian == product.Custodian {\n\t\treturn peer.Response{\n\t\t\tStatus: 403,\n\t\t\tMessage: fmt.Sprintf(\"You are already custodian\"),\n\t\t}\n\t}\n\n\t//make sure user cant claim a product separately from the container\n\tif product.ContainerID != \"\" {\n\t\tcontainerBytes, _ := stub.GetState(product.ContainerID)\n\t\tvar container Container\n\t\terr = json.Unmarshal(containerBytes, &container)\n\t\tif err != nil {\n\t\t\treturn shim.Error(err.Error())\n\t\t}\n\t\tif container.Custodian != newCustodian {\n\t\t\treturn peer.Response{\n\t\t\t\tStatus: 403,\n\t\t\t\tMessage: fmt.Sprintf(\"Product needs to be unpackaged before claiming a new owner\"),\n\t\t\t}\n\t\t}\n\t}\n\n\t//change custodian\n\tproduct.Custodian = newCustodian\n\tproduct.Location = newLocation\n\tproduct.Timestamp = int64(s.clock.Now().UTC().Unix())\n\n\tnewBytes, _ := json.Marshal(product)\n\n\tif err := stub.PutState(trackingID, newBytes); err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\ts.logger.Infof(\"Updated state: %s\\n\", trackingID)\n\treturn shim.Success([]byte(trackingID))\n\n}", "func Quantities(layers []string) (noodles int, sauce float64) {\n\tnoodlesPerLayer := 50\n\tsaucePerLayer := 0.2\n\tnoodles = 0\n\tsauce = 0.0\n\tfor _, layer := range layers {\n\t\tif layer == \"noodles\" {\n\t\t\tnoodles += noodlesPerLayer\n\t\t}\n\t\tif layer == \"sauce\" {\n\t\t\tsauce += saucePerLayer\n\t\t}\n\t}\n\treturn\n}", "func (s StorePaymentPurposeGiftedPremium) construct() StorePaymentPurposeClass { return &s }", "func (*GetCartReply_Cart) Descriptor() ([]byte, []int) {\n\treturn file_api_cart_sevrice_v1_cart_proto_rawDescGZIP(), []int{11, 0}\n}", "func (cq *CartQuery) All(ctx context.Context) ([]*Cart, error) {\n\tif err := cq.prepareQuery(ctx); err != nil {\n\t\treturn nil, err\n\t}\n\treturn cq.sqlAll(ctx)\n}", "func (thisShelf *inventory)MakeCustomers() {\n\tfor (thisShelf.shelfCokeCount >= 0 || thisShelf.shelfPepsiCount >= 0) {\n\t\ttime.Sleep(75 * time.Millisecond)\n\t\trand.Seed(time.Now().UnixNano())\n\t\tvar brandSelector int = (rand.Intn(2)) + 1\n\t\tif (brandSelector == 1) {\n\t\t\ttime.Sleep(75 * time.Millisecond)\n\t\t\trand.Seed(time.Now().UnixNano())\n\t\t\tvar canNumSelector int = (rand.Intn(4)) + 1\n\t\t\tvar actualNumCoke int\n\t\t\tif (canNumSelector == 1) {\n\t\t\t\tactualNumCoke = 6\n\t\t\t} else if (canNumSelector == 2) {\n\t\t\t\tactualNumCoke = 12\n\t\t\t} else if (canNumSelector == 3) {\n\t\t\t\tactualNumCoke = 18\n\t\t\t} else {\n\t\t\t\tactualNumCoke = 24\n\t\t\t}\n\t\t\tvar numRemovedCoke int = 0\n\t\t\tfor (numRemovedCoke == 0) {\n\t\t\t\tif (thisShelf.shelfCokeCount > actualNumCoke) {\n\t\t\t\t\tthisShelf.shelfCokeCount = thisShelf.shelfCokeCount - actualNumCoke\n\t\t\t\t\tnumRemovedCoke = actualNumCoke\n\t\t\t\t}\n\t\t\t\ttime.Sleep(10 * time.Millisecond)\n\t\t\t}\n\t\t\tthisShelf.checkoutCokeCount = thisShelf.checkoutCokeCount + numRemovedCoke\n\t\t\tfmt.Println(numRemovedCoke, \" Cans of Coke Removed from Shelf by Customer\")\n\t\t} else if (brandSelector == 2) {\n\t\t\ttime.Sleep(75 * time.Millisecond)\n\t\t\trand.Seed(time.Now().UnixNano())\n\t\t\tvar canNumSelector2 int = (rand.Intn(4)) + 1\n\t\t\tvar actualNumPepsi int\n\t\t\tif (canNumSelector2 == 1) {\n\t\t\t\tactualNumPepsi = 6\n\t\t\t} else if (canNumSelector2 == 2) {\n\t\t\t\tactualNumPepsi = 12\n\t\t\t} else if (canNumSelector2 == 3) {\n\t\t\t\tactualNumPepsi = 18\n\t\t\t} else {\n\t\t\t\tactualNumPepsi = 24\n\t\t\t}\n\t\t\tvar numRemovedPepsi int = 0\n\t\t\tfor (numRemovedPepsi == 0) {\n\t\t\t\tif (thisShelf.shelfPepsiCount > actualNumPepsi) {\n\t\t\t\t\tthisShelf.shelfPepsiCount = thisShelf.shelfPepsiCount - actualNumPepsi\n\t\t\t\t\tnumRemovedPepsi = actualNumPepsi\n\t\t\t\t}\n\t\t\t\ttime.Sleep(10 * time.Millisecond)\n\t\t\t}\n\t\t\tthisShelf.checkoutPepsiCount = thisShelf.checkoutPepsiCount + numRemovedPepsi\n\t\t\tfmt.Println(numRemovedPepsi, \" Cans of Pepsi Removed from Shelf by Customer\")\n\t\t}\n\t}\n}", "func Checkout(c *cli.Context) {\n\tapi := api.Create(c.GlobalString(\"locale\"))\n\n\tconf := config.GetConfig()\n\tdefer conf.Flush()\n\n\tcartName := conf.CartNameFromCache(c.Args().First())\n\n\tif cart, exists := conf.Carts[cartName]; exists {\n\t\tif getResponse, getErr := api.CartGet(cart.CartID, cart.HMAC); getErr == nil {\n\t\t\tdelete(conf.Carts, cartName)\n\t\t\tbrowser.OpenURL(getResponse.Cart.PurchaseURL)\n\t\t} else {\n\t\t\tpanic(getErr)\n\t\t}\n\t} else {\n\t\tfmt.Fprintf(os.Stderr, \"Cart %s is unknown\\n\", cartName)\n\t\tos.Exit(1)\n\t}\n}", "func mark_for_sale(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tvar err error\n\tfmt.Println(\"starting mark_for_sale\")\n\n\t// this is quirky\n\t// todo - get the \"company that authed the transfer\" from the certificate instead of an argument\n\t// should be possible since we can now add attributes to the enrollment cert\n\t// as is.. this is a bit broken (security wise), but it's much much easier to demo! holding off for demos sake\n\n\tif len(args) != 3 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 3\")\n\t}\n\n\t// input sanitation\n\terr = sanitize_arguments(args)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\tvar marble_id = args[0]\n\tvar authed_by_company = args[1]\n\tmin_price, err2 := strconv.Atoi(args[2])\n\n\tif err2 != nil {\n\t\treturn shim.Error(\"3rd argument must be a numeric string\")\n\t}\n\tfmt.Println(marble_id + \"->\" + strconv.Itoa(min_price) + \" - |\" + authed_by_company)\n\n\t// get marble's current state\n\tmarbleAsBytes, err := stub.GetState(marble_id)\n\tif err != nil {\n\t\treturn shim.Error(\"Failed to get marble\")\n\t}\n\tres := Marble{}\n\tjson.Unmarshal(marbleAsBytes, &res) //un stringify it aka JSON.parse()\n\n\t// check authorizing company\n\tif res.Owner.Company != authed_by_company {\n\t\treturn shim.Error(\"The company '\" + authed_by_company + \"' cannot authorize offer_for_sale for '\" + res.Owner.Company + \"'.\")\n\t}\n\n\t// mark the marble for sale\n\tres.IsForSale = true //set for Sale\n\tres.MinPrice = min_price // set minPrice\n\n\tjsonAsBytes, _ := json.Marshal(res) //convert to array of bytes\n\terr = stub.PutState(args[0], jsonAsBytes) //rewrite the marble with id as key\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n\tfmt.Println(\"- end mark_for_sale\")\n\treturn shim.Success(nil)\n\n}", "func (m *InMem) Add(c cart) error {\n\tm.AddQuantity(c.User, c.Product, c.Quantity)\n\treturn nil\n}", "func CreateNewOrder(res http.ResponseWriter, req *http.Request) {\n\tdefer func() { //to handle potential panic situation\n\t\tif err := recover(); err != nil {\n\t\t\tlog.Panic(\"Panic occured at create order:\", err)\n\t\t}\n\t}()\n\tmyUser := ses.GetUser(res, req)\n\tif !ses.AlreadyLoggedIn(req) {\n\t\thttp.Redirect(res, req, \"/\", http.StatusSeeOther)\n\t\treturn\n\t}\n\t//fmt.Println(Items)\n\tsortItems(ds.Items)\n\n\tvar newShoppingCart = []ds.Order{}\n\n\tif req.Method == http.MethodPost {\n\t\tnameRegExp := regexp.MustCompile(`^[\\w'\\-,.][^0-9_!¡?÷?¿/\\\\+=@#$%ˆ&*(){}|~<>;:[\\]]{2,30}$`) //name regexp to check for name pattern match\n\t\tname := strings.TrimSpace(req.FormValue(\"name\"))\n\t\tif !nameRegExp.MatchString(name) {\n\t\t\thttp.Error(res, \"You have entered an invalid name field.\", http.StatusBadRequest)\n\t\t\tlog.Warning(\"Invalid user input for name field\")\n\t\t\treturn\n\t\t}\n\t\tname = pol.Sanitize(name) //pol.Sanitize is used to sanitize inputs\n\n\t\taddRegExp := regexp.MustCompile(`^[\\w'\\-,.][^_!¡?÷?¿/\\\\+=$%ˆ&*(){}|~<>;:[\\]]{2,100}$`) ////address regexp to check for address pattern match\n\t\tadd := strings.TrimSpace(req.FormValue(\"address\"))\n\t\tif !addRegExp.MatchString(add) {\n\t\t\thttp.Error(res, \"You have entered an invalid address.\", http.StatusBadRequest)\n\t\t\tlog.Warning(\"Invalid user input for address field\")\n\t\t\treturn\n\t\t}\n\t\tadd = pol.Sanitize(add) //pol.Sanitize is used to sanitize inputs\n\n\t\tsday := req.FormValue(\"dday\") //sday is string day\n\t\tdayRegExp := regexp.MustCompile(`^[1-7]$`)\n\t\tif !dayRegExp.MatchString(sday) {\n\t\t\thttp.Error(res, \"You have entered an invalid delivery day.\", http.StatusBadRequest)\n\t\t\tlog.Warning(\"Invalid user input for delivery day\")\n\t\t\treturn\n\t\t}\n\n\t\tdday, _ := strconv.Atoi(sday)\n\n\t\tavailableDay := ds.IsDayAvailable(dday)\n\t\tif availableDay == false { //this checks if the order was placed on an unavailable day\n\t\t\terrorString := \"Sorry! There are no more available delivery slots for \" + ds.IntToDay(dday)\n\t\t\thttp.Error(res, errorString, http.StatusBadRequest)\n\t\t\tlog.Warning(\"There are no more available delivery slots for \" + ds.IntToDay(dday))\n\t\t\treturn\n\t\t}\n\n\t\torderQtyRegExp := regexp.MustCompile(`^[0-9]{1,2}$`) //order quantity reg exp to check for quantity pattern match\n\n\t\tfor i := 0; i < len(ds.Items); i++ {\n\t\t\tif !orderQtyRegExp.MatchString(req.FormValue(ds.Items[i].Name)) {\n\t\t\t\terrorString := \"You have entered an invalid order quantity for \" + ds.Items[i].Name + \".\"\n\t\t\t\thttp.Error(res, errorString, http.StatusBadRequest)\n\t\t\t\tlog.Warning(\"Invalid user input for order quantity\")\n\t\t\t\treturn\n\t\t\t}\n\t\t\tquantity, _ := strconv.Atoi(req.FormValue(ds.Items[i].Name)) //label for the form input is the item name, but returns a quantity of that item\n\t\t\tquantity64 := float64(quantity)\n\n\t\t\tif quantity64 > 0 {\n\t\t\t\titemAvailable := availableItem(ds.Items[i].Name)\n\t\t\t\tif itemAvailable == false { // this checks if the current item is in stock\n\t\t\t\t\terrorString := \"Oops, \" + ds.Items[i].Name + \" is no longer available for ordering.\"\n\t\t\t\t\thttp.Error(res, errorString, http.StatusBadRequest)\n\t\t\t\t\tlog.Warning(\"User overordered on item:\", ds.Items[i].Name)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tavailableBalance := isBalanceEnough(ds.Items[i].Name, quantity64)\n\t\t\t\tif availableBalance == false { //this checks if the user over ordered on the item\n\t\t\t\t\terrorString := \"Oops, there is no sufficient balance of\" + ds.Items[i].Name + \" for ordering..\"\n\t\t\t\t\thttp.Error(res, errorString, http.StatusBadRequest)\n\t\t\t\t\tlog.Warning(\"User overordered on item:\", ds.Items[i].Name)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tsingleCart := ds.Order{\n\t\t\t\t\tOrderItem: ds.Items[i].Name,\n\t\t\t\t\tQty: quantity64,\n\t\t\t\t}\n\t\t\t\tnewShoppingCart = append(newShoppingCart, singleCart)\n\t\t\t}\n\t\t}\n\n\t\tif len(newShoppingCart) == 0 {\n\t\t\thttp.Error(res, \"Error: You cannot submit an empty shopping cart.\", http.StatusBadRequest)\n\t\t\tlog.Warning(\"User entered empty shopping cart.\")\n\t\t\treturn\n\t\t}\n\n\t\ton := atomic.AddInt64(&OrderNumber, 1) // use of atomic function to prevent multiple clients from possibly creating identical order number\n\t\tamt := ds.CalculateAmount(newShoppingCart)\n\t\tnewOrder := ds.OrderInfo{\n\t\t\tUsername: myUser.Username,\n\t\t\tName: name,\n\t\t\tAddress: add,\n\t\t\tDeliveryDay: dday,\n\t\t\tOrderNum: on,\n\t\t\tShoppingCart: newShoppingCart,\n\t\t\tAmount: amt,\n\t\t}\n\n\t\tds.OrderList.AddOrder(newOrder)\n\t\terr := UpdateWeeklySchedule(ds.OrderList)\n\t\tif err != nil {\n\t\t\thttp.Error(res, err.Error(), http.StatusBadRequest)\n\t\t\tlog.Error(err)\n\t\t\treturn\n\t\t}\n\n\t\terr = UpdateWeeklyOrder(ds.OrderList)\n\t\tif err != nil {\n\t\t\thttp.Error(res, err.Error(), http.StatusBadRequest)\n\t\t\tlog.Error(err)\n\t\t\treturn\n\t\t}\n\n\t\t//fmt.Println(weeklySchedule)\n\t\t//fmt.Println(items)\n\t\t//orderList.printAllOrderNodes()\n\n\t\thttp.Redirect(res, req, \"/menu\", http.StatusSeeOther)\n\t\treturn\n\t}\n\n\ttype balanceStruct struct {\n\t\tItem string\n\t\tQuantity float64\n\t}\n\n\tvar itemData []balanceStruct\n\n\tfor i := 0; i < len(ds.Items); i++ {\n\t\tremainingQuantity := ds.Items[i].WeeklyCapacity - ds.Items[i].WeeklyOrder\n\t\td := balanceStruct{\n\t\t\tItem: ds.Items[i].Name,\n\t\t\tQuantity: remainingQuantity}\n\t\titemData = append(itemData, d)\n\t}\n\n\terr := tpl.ExecuteTemplate(res, \"createOrder.gohtml\", itemData)\n\tif err != nil {\n\t\thttp.Error(res, err.Error(), http.StatusBadRequest)\n\t\tlog.Fatalln(err)\n\t}\n\n\tViewAvailableDays(res, req)\n\tshowRemainingBalance(res, req)\n}", "func UpdateCartItem(c *gin.Context) {\n\tc.JSON(http.StatusOK, gin.H{\"message\": \"NOT IMPLEMENTED\"})\n}", "func (t *ManageMerchant) getMerchantsByIndustry(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\r\n\tvar jsonResp, industryName, errResp string\r\n\tvar err error\r\n\tvar merchantIndex []string\r\n\tvar valIndex Merchant\r\n\tfmt.Println(\"start getMerchantsByIndustry\")\r\n\tif len(args) != 1 {\r\n\t\terrMsg := \"{ \\\"message\\\" : \\\"Incorrect number of arguments. Expecting 'industryName' as an argument\\\", \\\"code\\\" : \\\"503\\\"}\"\r\n\t\terr = stub.SetEvent(\"errEvent\", []byte(errMsg))\r\n\t\tif err != nil {\r\n\t\t\treturn nil, err\r\n\t\t} \r\n\t\treturn nil, nil\r\n\t}\r\n\t// set merchantId\r\n\tindustryName = args[0]\r\n\r\n\tmerchantAsBytes, err := stub.GetState(MerchantIndexStr)\r\n\tif err != nil {\r\n\t\treturn nil, errors.New(\"Failed to get Merchant index string\")\r\n\t}\r\n\tjson.Unmarshal(merchantAsBytes, &merchantIndex)\t\t\t//un stringify it aka JSON.parse()\r\n\tfmt.Print(\"merchantIndex : \")\r\n\tfmt.Println(merchantIndex)\r\n\t\r\n\tjsonResp = \"{\"\r\n\tfor i,val := range merchantIndex{\r\n\t\tfmt.Println(strconv.Itoa(i) + \" - looking at \" + val + \" for getMerchantsByIndustry\")\r\n\t\tvalueAsBytes, err := stub.GetState(val)\r\n\t\tif err != nil {\r\n\t\t\terrResp = \"{\\\"Error\\\":\\\"Failed to get state for \" + val + \"\\\"}\"\r\n\t\t\treturn nil, errors.New(errResp)\r\n\t\t}\r\n\t\t//fmt.Print(\"valueAsBytes : \")\r\n\t\t//fmt.Println(valueAsBytes)\r\n\t\tjson.Unmarshal(valueAsBytes, &valIndex)\r\n\t\tfmt.Print(\"valIndex: \")\r\n\t\tfmt.Print(valIndex)\r\n\t\tif valIndex.MerchantIndustry == industryName{\r\n\t\t\tfmt.Println(\"Merchant found\")\r\n\t\t\tjsonResp = jsonResp + \"\\\"\"+ val + \"\\\":\" + string(valueAsBytes[:])\r\n\t\t\tfmt.Println(\"jsonResp inside if\")\r\n\t\t\tfmt.Println(jsonResp)\r\n\t\t\tif i < len(merchantIndex)-1 {\r\n\t\t\t\tjsonResp = jsonResp + \",\"\r\n\t\t\t}\r\n\t\t} \r\n\t}\r\n\tjsonResp = jsonResp + \"}\"\r\n\tif strings.Contains(jsonResp, \"},}\"){\r\n\t\tfmt.Println(\"in if for jsonResp contains wrong json\")\t\r\n\t\tjsonResp = strings.Replace(jsonResp, \"},}\", \"}}\", -1)\r\n\t}\r\n\tfmt.Println(\"jsonResp : \" + jsonResp)\r\n\tfmt.Print(\"jsonResp in bytes : \")\r\n\tfmt.Println([]byte(jsonResp))\r\n\tfmt.Println(\"end getMerchantsByIndustry\")\r\n\treturn []byte(jsonResp), nil\t\t\t\t\t\t\t\t\t\t\t//send it onward\r\n}", "func main() {\n\tserver, err := cloudstate.New(protocol.Config{\n\t\tServiceName: \"shopping-cart\",\n\t\tServiceVersion: \"0.1.0\",\n\t})\n\tif err != nil {\n\t\tlog.Fatalf(\"cloudstate.New failed: %v\", err)\n\t}\n\n\terr = server.RegisterEventSourced(&eventsourced.Entity{\n\t\tServiceName: \"com.example.shoppingcart.ShoppingCart\",\n\t\tPersistenceID: \"ShoppingCart\",\n\t\tEntityFunc: shoppingcart.NewShoppingCart,\n\t\tSnapshotEvery: 5,\n\t}, protocol.DescriptorConfig{\n\t\tService: \"shoppingcart.proto\",\n\t}.AddDomainDescriptor(\"domain.proto\"))\n\n\tif err != nil {\n\t\tlog.Fatalf(\"CloudState failed to register entity: %v\", err)\n\t}\n\terr = server.Run()\n\tif err != nil {\n\t\tlog.Fatalf(\"CloudState failed to run: %v\", err)\n\t}\n}", "func FindAllCartsByProductID(productID int) ([]*Cart, error) {\n\tcartItems := []*Cart{}\n\terr := db.Model(&Cart{}).Where(\"product_id = ?\", productID).Find(&cartItems).Error\n\tif err != nil && err != gorm.ErrRecordNotFound {\n\t\treturn nil, err\n\t}\n\n\treturn cartItems, nil\n}", "func (cs *checkoutService) emptyUserCart(ctx context.Context, userID string) error {\n\tif _, err := pb.NewCartServiceClient(cs.cartSvcConn).EmptyCart(ctx, &pb.EmptyCartRequest{UserId: userID}); err != nil {\n\t\treturn fmt.Errorf(\"failed to empty user cart during checkout: %+v\", err)\n\t}\n\treturn nil\n}", "func (rest *RestApi) Init() error {\n rest.GoCart.loadConfig()\n mysql := MysqlConnection{\n host: rest.GoCart.Config.Database.Host,\n port: rest.GoCart.Config.Database.Port,\n user: rest.GoCart.Config.Database.Username,\n password: rest.GoCart.Config.Database.Password,\n database: rest.GoCart.Config.Database.Database,\n table: rest.GoCart.Config.Database.Cart.Table,\n table_index: rest.GoCart.Config.Database.Cart.Mappings.Index,\n }\n mysql.EnsureCartTable()\n\n rest.GoCart = GoCart{\n Connection: mysql,\n }\n router := mux.NewRouter()\n\n /**\n * GET request\n */\n //http.HandleFunc(\"/gocart/getCart\", func(w http.ResponseWriter, r *http.Request) {\n router.HandleFunc(\"/gocart/getCart\", func(w http.ResponseWriter, r *http.Request) {\n cart_id, err := strconv.ParseInt(r.URL.Query().Get(\"cart_id\"), 10, 64)\n if err != nil {\n panic(err)\n }\n rest.GetCart(w, r, cart_id)\n }).Methods(\"GET\")\n\n /**\n * POST request\n */\n router.HandleFunc(\"/gocart/addToCart\", func(w http.ResponseWriter, r *http.Request) {\n cart_id, err := strconv.ParseInt(r.URL.Query().Get(\"cart_id\"), 10, 64)\n if err != nil {\n panic(err)\n }\n items_qsp := r.URL.Query().Get(\"items\")\n item_quantity := r.URL.Query().Get(\"quantity\")\n\n ids := strings.Split(items_qsp, \",\")\n for _, item_id := range ids {\n item_id, err := strconv.ParseInt(item_id, 10, 64)\n if err != nil {\n panic(err)\n }\n item_quantity, err := strconv.ParseInt(item_quantity, 10, 64)\n if err != nil {\n panic(err)\n }\n rest.AddToCart(w, r, cart_id, item_id, item_quantity)\n }\n // @TODO: Print some error/success message\n }).Methods(\"POST\")\n\n log.Fatal(http.ListenAndServe(\":9090\", router))\n return nil\n}", "func (cq *CartQuery) Only(ctx context.Context) (*Cart, error) {\n\tcs, err := cq.Limit(2).All(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tswitch len(cs) {\n\tcase 1:\n\t\treturn cs[0], nil\n\tcase 0:\n\t\treturn nil, &NotFoundError{cart.Label}\n\tdefault:\n\t\treturn nil, &NotSingularError{cart.Label}\n\t}\n}", "func (store *Store) RemoveFromCart(ctx *gin.Context) (bool, error) {\n\tctx.String(200, \"You are trying to remove items from the cart.\")\n\treturn true, nil\n}", "func CalculateInventory(){\n\treturn\n}", "func handleFuncProcess(w http.ResponseWriter, r *http.Request) {\n\tc := appengine.NewContext(r)\n\t\n\tpromo_code := r.FormValue(\"promo_code\")\n\t\n\t//Call initiated by browser when a JS or ajax attempts ot recompute the cart items\n\t//Get all items in the cart and process\n\t//For each item which has applicable pricing rules, include them in processing\n\t//Return total and list of items\n\t\n\tshop := newCartProc()\n\t//Rules for BuyThreePayTwoOnly\n\t//If you buy 3 items, you pay only two items\n\tc.Infof(\"Check_Rule_BuyThreePayTwoOnly\")\n\tshop.Check_Rule_BuyThreePayTwoOnly(w,r)\n\n\t//Rules for BulkDiscountMoreThanThree\n\t//The price will drop to $$ each for the first month, if the customer buys more than x items.\n\tc.Infof(\"Check_Rule_BulkDiscountMoreThanThree\")\n\tshop.Check_Rule_BulkDiscountMoreThanThree(w,r)\n\n\t//Rules for BundleFreeForEveryItemBought\n\t//We will bundle in a free item X free-of-charge with every Y sold\n\tc.Infof(\"Check_Rule_BundleFreeForEveryItemBought\")\n\tshop.Check_Rule_BundleFreeForEveryItemBought(w,r)\n\t\n\t//Process those items w/ no rules applied\n\tc.Infof(\"All_Others_No_Rule_Processing\")\n\tshop.All_Others_No_Rule_Processing(w,r)\n\t\n\t//Finally, see if any promos to be processed\n\t//Check rules if there are promos to be applied\n\tc.Infof(\"promo_code: %v\", promo_code)\n\tif promo_code != \"\" {\n\t\t//Rules for PromoCodeDiscount\n\t\t//Adding the promo code X will apply a $$ discount across the board.\n\t\tc.Infof(\"Check_Rule_PromoCodeDiscount\")\n\t\tshop.Check_Rule_PromoCodeDiscount(w,r,promo_code)\n\t}\n\t\n\t//return data as json\n\tcp := CartProc {\n\t\tTotal: shop.Total,\n\t\tRules: shop.Rules,\n\t\tCurrent: shop.Current,\n\t}\n\t\n\tdata,_ := json.MarshalIndent(cp, \"\", \" \")\n\tw.Write(data)\n\t\n}", "func (main *Main) GetCartByUserId(e echo.Context) (err error) {\n\tUserID := e.Param(\"code\")\n\n\tuser, cart, exc := CartModel.GetAll(\"user_id\", UserID)\n\tif exc != nil {\n\t\treturn rest.ConstructErrorResponse(e, exc)\n\t}\n\n\tdata := map[string]contract.Model{\n\t\t\"cart_detail\": cart,\n\t\t\"user_detail\": user,\n\t}\n\n\treturn rest.ConstructSuccessResponse(e, data)\n}", "func (liq *LineItemQuery) QueryCart() *CartQuery {\n\tquery := &CartQuery{config: liq.config}\n\tquery.path = func(ctx context.Context) (fromU *sql.Selector, err error) {\n\t\tif err := liq.prepareQuery(ctx); err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tstep := sqlgraph.NewStep(\n\t\t\tsqlgraph.From(lineitem.Table, lineitem.FieldID, liq.sqlQuery()),\n\t\t\tsqlgraph.To(cart.Table, cart.FieldID),\n\t\t\tsqlgraph.Edge(sqlgraph.M2O, true, lineitem.CartTable, lineitem.CartColumn),\n\t\t)\n\t\tfromU = sqlgraph.SetNeighbors(liq.driver.Dialect(), step)\n\t\treturn fromU, nil\n\t}\n\treturn query\n}", "func (cs *checkoutService) getUserCart(ctx context.Context, userID string) ([]*pb.CartItem, error) {\n\tcart, err := pb.NewCartServiceClient(cs.cartSvcConn).GetCart(ctx, &pb.GetCartRequest{UserId: userID})\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to get user cart during checkout: %+v\", err)\n\t}\n\treturn cart.GetItems(), nil\n}", "func (t *SimpleChaincode) addProduct(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tfmt.Println(\"Add product...\")\n\tvar product Product\n\tvar err error\n\t\n\tisExistAsBytes,_ := stub.GetState(args[0])\n\n\tif(isExistAsBytes != nil){\n\t\treturn shim.Error(err.Error())\n\t}\n\n\tproduct.SerialNo=args[0]\n\tproduct.BatchId=args[1]\n\tproduct.ModelNo=args[2]\n\tproduct.ModelName=args[3]\n\tproduct.DateOfMfg =args[4]\n\tproduct.CftId=args[5]\n\tproduct.Price=args[6]\n\tproduct.Specification=args[7]\n\tproduct.Status=\"MFG\"\n\tproduct.Ownership=\"Rolex Pvt Ltd\"\n\tproduct.DealerId=\"\"\n\tproduct.DealerName=\"\"\n\tproduct.ShippingDate =\"\"\n\tproduct.DateOfPurchase=\"\"\n\tproduct.Insurance=\"\"\n product.InsuranceExpiry=\"\"\n\tproduct.ServiceId=\"\"\n\tproduct.ServiceHistory=\"\"\n\t\n\tproductAsBytes, _ := json.Marshal(product)\n\terr = stub.PutState(product.SerialNo, productAsBytes)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\n//update all batches to the array whose key is \"AllWatches\"\n\tAllWatchesAsBytes, err := stub.GetState(\"AllWatches\")\n\n\tvar allWatches AllWatches\n\n\terr= json.Unmarshal(AllWatchesAsBytes, &allWatches)\n\tallWatches.AllWatches=append(allWatches.AllWatches,product.SerialNo)\n\n\tallwatchesAsBytes,_ :=json.Marshal(allWatches)\n\terr = stub.PutState(\"AllWatches\", allwatchesAsBytes)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n//-----------------------------------------------------------\n\treturn shim.Success(nil)\n}", "func init_product(stub shim.ChaincodeStubInterface, args []string) (pb.Response) {\n\tvar err error\n\tfmt.Println(\"starting init_product\")\n\n\n\tif len(args) != 11 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting 11\")\n\t}\n\n\n\t//input sanitation\n\terr = sanitize_arguments(args)\n\tif err != nil {\n\t\treturn shim.Error(err.Error())\n\t}\n\tvar product Product\n\n\tproduct.ObjectType = \"product\"\n\n\tproduct.GroupId = args[0]\n\tproduct.GroupName = strings.ToLower(args[1])\n\tproduct.RegisterUserId = args[2]\n\tproduct.RegisterUserName = strings.ToLower(args[3])\n\tproduct.ProductId = args[4]\n\tproduct.RegisterDate = strings.ToLower(args[5])\n\tproduct.ProductName =strings.ToLower(args[6])\n\tproduct.ProductWeight = strings.ToLower(args[7])\n\tproduct.ProductVolume = strings.ToLower(args[8])\n\tproduct.MainIngredientName = strings.ToLower(args[9])\n\tproduct.FORCoin,err = strconv.Atoi(args[10])\n\n\n\tproductAsBytes, _ :=json.Marshal(product)\n\n\tfmt.Println(\"writing product to state\")\n\n\tfmt.Println(string(productAsBytes))\n\tfmt.Println(\"productid\",product.ProductId)\n\terr = stub.PutState(product.ProductId, productAsBytes) //store owner by its Id\n\tif err != nil {\n\t\tfmt.Println(\"Could not store product\")\n\t\treturn shim.Error(err.Error())\n\t}\n\n//\tif err != nil {\n//\t\treturn shim.Error(\"3rd argument must be a numeric string\")\n//\t}\n\n\n\t//build the product json string manually\n//\tstr := `{\n//\t\t\"docType\":\"product\",\n//\t\t\"groupid\": \"` + groupid + `\",\n//\t\t\"groupname\": \"sarang\",\n//\t\t\"registeruserid\": \"sarng2\",\n//\t\t\"registerusername\": \"sarararng\",\n//\t\t\"productid\": \"` + productid + `\",\n//\t\t\"registerdate\": \"` + registerdate + `\",\n//\t\t\"productname\": \"` + productname + `\",\n//\t\t\"productweight\": \"` + productweight + `\",\n//\t\t\"productvolume\": \"` + productvolume + `\",\n//\t\t\"mainingredientname\": \"` + mainingredientname + `\", \n//\t\t\"forcoin\": ` + strconv.Itoa(forcoin) + `, \n//\t}`\n//\terr = stub.PutState(productid, []byte(str)) //store product with id as key\n//\n//\tAvals ,err :=stub.GetState(productid)\n//\tAvalsString :=string(Avals)\n//\n//\tfmt.Println(\"product?\",AvalsString)\n//\n//\n//\tif err != nil {\n//\t\treturn shim.Error(err.Error())\n//\t}\n//\n\tfmt.Println(\"- end init_product\")\n\n\treturn shim.Success(nil)\n}", "func (app *application) checkoutForm(w http.ResponseWriter, r *http.Request) {\n\tsession, err := app.sessionStore.Get(r, \"session-name\")\n\tif err != nil {\n\t\tapp.serverError(w, err)\n\t\treturn\n\t}\n\tcustomerID := session.Values[\"customerID\"].(int)\n\tcart := app.carts[customerID]\n\n\tcartRowSlice := make([]cartRow, 0)\n\ttotal := 0\n\n\tfor listID, quantity := range cart {\n\t\tlisting, err := app.listings.Get(listID)\n\t\tif err != nil {\n\t\t\tapp.serverError(w, err)\n\t\t\treturn\n\t\t}\n\t\trow := cartRow{\n\t\t\tListingID: listing.ID,\n\t\t\tName: listing.Name,\n\t\t\tPrice: listing.Price,\n\t\t\tQuantity: quantity,\n\t\t\tAmount: quantity * listing.Price,\n\t\t}\n\t\tcartRowSlice = append(cartRowSlice, row)\n\t\ttotal += quantity * listing.Price\n\t}\n\tapp.render(w, r, \"checkout.page.tmpl\", &templateData{\n\t\tCart: cartRowSlice,\n\t\tCartTotal: total,\n\t\tForm: forms.New(nil),\n\t})\n}", "func (_CraftingI *CraftingICallerSession) Items(arg0 *big.Int) (struct {\n\tBaseType uint8\n\tItemType uint8\n\tCrafted uint32\n\tCrafter *big.Int\n}, error) {\n\treturn _CraftingI.Contract.Items(&_CraftingI.CallOpts, arg0)\n}", "func (a *IntegrationApiService) ReturnCartItems(ctx _context.Context, customerSessionId string) apiReturnCartItemsRequest {\n\treturn apiReturnCartItemsRequest{\n\t\tapiService: a,\n\t\tctx: ctx,\n\t\tcustomerSessionId: customerSessionId,\n\t}\n}", "func (c *Customer) shop(readyForCheckoutChan chan int) {\n\n\tvar speedMultiplier float64\n\tspeedMultiplier = 1\n\n\t// Infinite loop of customer shopping\n\tfor {\n\t\tif c.getNumProducts() == int(productsRate) {\n\t\t\tbreak\n\t\t}\n\n\t\tif c.age > 65 {\n\t\t\tspeedMultiplier = 1.5\n\t\t}\n\n\t\tp := newProduct()\n\t\ttime.Sleep(time.Millisecond * time.Duration(p.getTime()*200*speedMultiplier))\n\t\tc.trolley.addProductToTrolley(p)\n\t\tc.shopTime += int64(p.getTime() * 200)\n\t\tif c.trolley.isFull() {\n\t\t\tbreak\n\t\t}\n\n\t\tif rand.Float64() < 0.05 {\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// Notify the channel in the supermarket FinishedShoppingListener() by sending the customer id to it\n\treadyForCheckoutChan <- c.id\n}", "func updateCart(locker *redislock.Client, c chan int, ctx context.Context, i int) {\n\n\tbackoff := redislock.LinearBackoff(10 * time.Millisecond)\n\tlock, err := locker.Obtain(ctx, \"my-key\", 2*time.Second, &redislock.Options{\n\t\tRetryStrategy: backoff,\n\t})\n\tif err == redislock.ErrNotObtained {\n\t\tfmt.Println(\"Could not obtain lock!\", lock)\n\t} else if err != nil {\n\t\tlog.Fatalln(err)\n\t}\n\n\tfmt.Println(\"lock saya\", i)\n\tlock.Release(ctx)\n\tfmt.Println(\"release\", i)\n\tc <- 1\n}", "func (t *ProductChaincode) initProduct(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n\t// For loop on quantity\n\tquantity, _ := strconv.Atoi(args[5])\n\n\tfor i := 0; i < quantity; i++ {\n\n\t\tid := rand.New(rand.NewSource(time.Now().UnixNano()))\n\n\t\ttoInt := id.Int()\n\t\ttagID := strconv.Itoa(toInt)\n\n\t\tfinalProduct := tagID\n\t\tvar finalArgs []string\n\t\tfinalArgs = append(finalArgs, finalProduct)\n\t\tfinalArgs = append(finalArgs, args[1])\n\t\tfinalArgs = append(finalArgs, args[2])\n\t\tfinalArgs = append(finalArgs, args[3])\n\t\tfinalArgs = append(finalArgs, args[4])\n\n\t\tvar product Product\n\t\tif err := product.FillFromArguments(finalArgs); err != nil {\n\t\t\treturn shim.Error(err.Error())\n\t\t}\n\n\t\tif product.ExistsIn(stub) {\n\t\t\tcompositeKey, _ := product.ToCompositeKey(stub)\n\t\t\treturn shim.Error(fmt.Sprintf(\"product with the key %s already exists\", compositeKey))\n\t\t}\n\n\t\t// TODO: set owner from GetCreatorOrg\n\t\tproduct.Value.State = stateRegistered\n\t\tproduct.Value.Name = args[0]\n\n\t\tif err := product.UpdateOrInsertIn(stub); err != nil {\n\t\t\treturn shim.Error(err.Error())\n\t\t}\n\n\t\t// TODO: think about index usability\n\t\t// ==== Index the product to enable state-based range queries, e.g. return all Active products ====\n\t\t// An 'index' is a normal key/value entry in state.\n\t\t// The key is a composite key, with the elements that you want to range query on listed first.\n\t\t// In our case, the composite key is based on stateIndexName~state~name.\n\t\t// This will enable very efficient state range queries based on composite keys matching stateIndexName~state~*\n\t\t//stateIndexKey, err := stub.CreateCompositeKey(stateIndexName, []string{strconv.Itoa(product.Value.State), product.Key.Name})\n\t\t//if err != nil {\n\t\t// return shim.Error(err.Error())\n\t\t//}\n\t\t//// Save index entry to state. Only the key name is needed, no need to store a duplicate copy of the product.\n\t\t//// Note - passing a 'nil' value will effectively delete the key from state, therefore we pass null character as value\n\t\t//value := []byte{0x00}\n\t\t//stub.PutState(stateIndexKey, value)\n\t}\n\treturn shim.Success(nil)\n\n}", "func (e *PlaceOrderServiceAdapter) PlaceGuestCart(ctx context.Context, cart *cartDomain.Cart, payment *placeorder.Payment) (placeorder.PlacedOrderInfos, error) {\n\treturn e.placeOrder(ctx, cart, payment)\n}", "func (i *Inventory) Process(cart map[string]uint32) (map[string]uint32, float32, notifier.Status, error) {\n\tvar cost float32\n\treceipt := make(map[string]uint32)\n\n\tfor item, count := range cart {\n\t\tprice, err := i.GetPrice(item)\n\t\tif err != nil {\n\t\t\treturn nil, 0, notifier.FailedInvalidCart, err\n\t\t}\n\t\tstock, _ := i.GetStock(item)\n\n\t\tswitch v := i.Deals[item].(type) {\n\t\tcase *Offer:\n\t\t\tgetCount := (count / v.Buy)\n\t\t\trem := count % v.Buy\n\t\t\tcost += float32(count) * price\n\t\t\treceipt[item] = getCount*(v.Buy+v.Get) + rem\n\t\tcase *PriceOffer:\n\t\t\tgetCount := (count / v.Buy)\n\t\t\trem := count % v.Buy\n\t\t\tcost += float32(getCount*v.Price)*price + float32(rem)*price\n\t\t\treceipt[item] = count\n\t\tdefault:\n\t\t\tcost += float32(count) * price\n\t\t\treceipt[item] = count\n\t\t}\n\n\t\tif stock != nil && *stock < receipt[item] {\n\t\t\treturn nil, 0, notifier.OutOfStock, fmt.Errorf(\"item, %v, out of stock\", item)\n\t\t}\n\t}\n\n\ti.updateStockPile(receipt)\n\treturn receipt, cost, notifier.Fulfilled, nil\n}", "func FindAllCartsByAccountID(accountID int) ([]*Cart, error) {\n\tcartItems := []*Cart{}\n\terr := db.Model(&Cart{}).Where(\"account_id = ?\", accountID).Find(&cartItems).Error\n\tif err != nil && err != gorm.ErrRecordNotFound {\n\t\treturn nil, err\n\t}\n\n\treturn cartItems, nil\n}", "func (t *SimpleChaincode) querybyrole(stub shim.ChaincodeStubInterface, args []string) ([]byte, error){\n\n var jsonResp string\n var err error\n\n if len(args) != 2 {\n jsonResp = \"Error:Incorrect number of arguments. Need to pass Role: Shipper, Provider, Insurer or Consignee & status value to be passed\"\n return nil, errors.New(jsonResp)\n }\n\n// validate role\n if args[0] == \"Shipper\"{\n fmt.Println(\"Shipper has been passed as Role\")\n } else if args[0] == \"Provider\" {\n fmt.Println(\"Provider has been passed as Role\")\n } else if args[0] == \"Insurer\" {\n fmt.Println(\"Insurer has been passed as Role\")\n } else if args[0] == \"Consignee\" {\n fmt.Println(\"Consignee has been passed as Role\")\n } else {\n jsonResp = \" Error:Incorrect Role has been passed, should be: Shipper, Provider, Insurer or Consignee\"\n return nil, errors.New(jsonResp)\n }\n\n valAsbytes, err := stub.GetState(\"PkgIdsKey\")\n if err != nil {\n jsonResp = \"Error:Failed to get state for PkgIdsKey \"\n return nil, errors.New(jsonResp)\n }\n\n var package_holder PKG_Holder\n err = json.Unmarshal(valAsbytes, &package_holder)\n if err != nil {\n fmt.Println(\"Could not marshal personal info object\", err)\n jsonResp = \"Error:Could not marshal personal info object\"\n return nil, errors.New(jsonResp)\n }\n\n var pkginfo PackageInfo\n\n result := \"[\"\n\n var temp []byte\n\n for _, PkgId := range package_holder.PkgIds {\n\n pkginfoasbytes, err := stub.GetState(PkgId)\n if err != nil {\n jsonResp = \"Error:Failed to get state for \" + PkgId\n return nil, errors.New(jsonResp)\n }\n\n err = json.Unmarshal(pkginfoasbytes, &pkginfo);\n if err != nil {\n fmt.Println(\"Could not marshal personal info object\", err)\n jsonResp = \"Error:Could not marshal personal info object\"\n return nil, errors.New(jsonResp)\n }\n\n // check for inout role\n if args[0] == \"Provider\"{\n if pkginfo.Provider == args[1] {\n temp = pkginfoasbytes\n result += string(temp) + \",\"\n }\n } else if args[0] == \"Shipper\" {\n if pkginfo.Shipper == args[1] {\n temp = pkginfoasbytes\n result += string(temp) + \",\"\n }\n } else if args[0] == \"Insurer\" {\n if pkginfo.Insurer == args[1] {\n temp = pkginfoasbytes\n result += string(temp) + \",\"\n }\n } else if args[0] == \"Consignee\" {\n if pkginfo.Consignee == args[1] {\n temp = pkginfoasbytes\n result += string(temp) + \",\"\n }\n }\n\n\n } // end of for loop\n\n if len(result) == 1 {\n result = \"[]\"\n } else {\n result = result[:len(result)-1] + \"]\"\n }\n\n return []byte(result), nil\n\n}", "func (storeShelf *inventory)PepsiDelivery(){\n\tfor i := 0; i < 20; i++ {\n\t\tstoreShelf.inventoryPepsiCount += 24\n\t\tstoreShelf.costPepsi += 4.8\n\t\ttime.Sleep(5000 * time.Millisecond)\n\t\tfmt.Println(\"24 cans of Pepsi added to the shelf by the stocker\")\n\t}\n}", "func AddCartItem(service Service, userService users.Service) func(w http.ResponseWriter, r *http.Request) {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tctx := r.Context()\n\t\tlogger := loglib.GetLogger(ctx)\n\t\tusername, err := auth.GetLoggedInUsername(r)\n\t\tif err != nil {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusForbidden, errorcode.ErrorsInRequestData, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tuser, err := userService.RetrieveUserByUsername(ctx, username)\n\t\tif err != nil || user == nil {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusUnauthorized, errorcode.UserNotFound, \"User not found\")\n\t\t\treturn\n\t\t}\n\n\t\tlogger.Infof(\"user is %v\", user.Username)\n\t\t// unmarshal request\n\t\treq := addCartItemRequest{}\n\t\tif err := json.NewDecoder(r.Body).Decode(&req); (err != nil || req == addCartItemRequest{}) {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusBadRequest, errorcode.ErrorsInRequestData, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\t// validate request\n\t\tif err := req.Validate(); err != nil {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusBadRequest, errorcode.ErrorsInRequestData, err.Error())\n\t\t\treturn\n\t\t}\n\n\t\tcart, err := service.AddItemCart(ctx, user.ID, req.ProductID, req.Quantity)\n\t\tif err != nil {\n\t\t\thttpresponse.ErrorResponseJSON(ctx, w, http.StatusInternalServerError, \"internal_error\", err.Error())\n\t\t\treturn\n\t\t}\n\n\t\thttpresponse.RespondJSON(w, http.StatusOK, cart, nil)\n\t}\n}", "func (cartItem CartItem) create() (Entity, error) {\n\n\t_item := cartItem\n\tif err := db.Create(&_item).Error; err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := _item.GetPreloadDb(false,false, nil).First(&_item,_item.Id).Error; err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := (Order{Id: _item.OrderId, AccountId: _item.AccountId}).UpdateDeliveryData(); err != nil {\n\t\tlog.Println(\"Error update cart item: \", err)\n\t}\n\tif err := (Order{Id: _item.OrderId, AccountId: _item.AccountId}).UpdateCost(); err != nil {\n\t\tlog.Println(\"Error update cart item: \", err)\n\t}\n\n\tvar entity Entity = &_item\n\n\treturn entity, nil\n}" ]
[ "0.617487", "0.6044997", "0.5967852", "0.57275903", "0.56581736", "0.5571226", "0.5502911", "0.5464738", "0.53850746", "0.5368966", "0.5366337", "0.53322273", "0.5254271", "0.519116", "0.513856", "0.5113555", "0.5099233", "0.5077585", "0.5071831", "0.504693", "0.49594086", "0.4950365", "0.4950159", "0.4903986", "0.49033773", "0.4900392", "0.48961383", "0.48860213", "0.48800382", "0.48763716", "0.48566106", "0.48526996", "0.48517132", "0.48483324", "0.48301443", "0.48288187", "0.4774005", "0.4773567", "0.47564483", "0.4754026", "0.47522718", "0.47502786", "0.47495607", "0.47394735", "0.47271588", "0.47267234", "0.472248", "0.47175768", "0.4685834", "0.4675684", "0.46579924", "0.46494767", "0.4648968", "0.46298894", "0.46231025", "0.46159446", "0.46065176", "0.46063027", "0.45901176", "0.45878407", "0.45862988", "0.4586189", "0.4583749", "0.4574451", "0.45741367", "0.4562881", "0.4560591", "0.45485675", "0.45448065", "0.45205322", "0.45171472", "0.45074934", "0.45035395", "0.4495614", "0.44855404", "0.44808447", "0.4478048", "0.44758305", "0.4465329", "0.44467184", "0.44419387", "0.44341773", "0.443379", "0.4424539", "0.4420671", "0.4419547", "0.44125032", "0.44077554", "0.4405624", "0.4403447", "0.43992168", "0.43985185", "0.43966544", "0.43886817", "0.43835738", "0.43834177", "0.43776712", "0.43769166", "0.43760592", "0.43696246", "0.43630368" ]
0.0
-1
NewDatabase create a new DatabaseObject
func NewDatabase(config *AppConfig) (*DB, error) { databaseConfigureation := &stdlib.DriverConfig{ ConnConfig: pgx.ConnConfig{ Host: config.DatabaseHost, User: config.DatabaseUser, Password: config.DatabasePass, Database: config.DatabaseName, Port: config.DatabasePort, }, } stdlib.RegisterDriverConfig(databaseConfigureation) database, err := sql.Open("pgx", databaseConfigureation.ConnectionString("")) if err != nil { return nil, err } database.SetMaxOpenConns(20) database.SetMaxIdleConns(10) // db.SetConnMaxLifetime(time.Second * 10) return &DB{database}, nil }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func newDatabase(info extraInfo, db *sql.DB) *database {\n\treturn &database{\n\t\tname: info.dbName,\n\t\tdriverName: info.driverName,\n\t\tdb: db,\n\t}\n}", "func TestNewDatabase(t *testing.T) {\n\tconf := config.GetConfig()\n\tdbClient, err := databases.NewClient(conf)\n\tassert.NoError(t, err)\n\tdb := databases.NewDatabase(conf, dbClient)\n\n\tassert.NotEmpty(t, db)\n}", "func newDatabase(s *Server) *Database {\n\treturn &Database{\n\t\tserver: s,\n\t\tusers: make(map[string]*DBUser),\n\t\tpolicies: make(map[string]*RetentionPolicy),\n\t\tshards: make(map[uint64]*Shard),\n\t\tseries: make(map[string]*Series),\n\t}\n}", "func NewDatabase(address string) *Database {\n\treturn &Database{address: address}\n}", "func NewDatabase(w Writer) *Database {\n\treturn &Database{Writer: w}\n}", "func NewDatabase(cfg *config.Config) (database *Database, err error) {\n\tdatabase = new(Database)\n\n\tif database.db, err = sql.Open(cfg.Database.Driver, cfg.Database.DataSource); err != nil {\n\t\treturn\n\t}\n\tif err = database.proposalState.prepare(database.db); err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func NewDatabase(dbType string, dbHost string, dbUser string, dbPassword string, dbName string, dbPort string) *DB {\n\tvar connectString string\n\tif dbType == \"postgres\" {\n\t\tconnectString = \"host=\" + dbHost + \" port=\" + dbPort + \" user=\" + dbUser + \" dbname=\" + dbName + \" password=\" + dbPassword\n\t}\n\tif dbType == \"mysql\" {\n\t\tconnectString = dbUser + \":\" + dbPassword + \"@\" + dbHost + \"/\" + dbName\n\t}\n\tif connectString != \"\" {\n\t\tdb, err := grom.Open(dbType, connectString)\n\t\tif err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\tif err = db.DB().Ping(); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\n\t\t// db.LogMode(true)\n\n\t\treturn &DB{db}\n\t} else {\n\t\tpanic(\"Must Define connection information\")\n\t}\n}", "func NewDatabase(name string, root *doltdb.RootValue, ddb *doltdb.DoltDB, rs *env.RepoState) *Database {\n\treturn &Database{\n\t\tname: name,\n\t\troot: root,\n\t\tddb: ddb,\n\t\trs: rs,\n\t\tbatchMode: single,\n\t\ttables: make(map[string]*DoltTable),\n\t}\n}", "func NewDatabase(ctx *pulumi.Context,\n\tname string, args *DatabaseArgs, opts ...pulumi.ResourceOption) (*Database, error) {\n\tif args == nil || args.Instance == nil {\n\t\treturn nil, errors.New(\"missing required argument 'Instance'\")\n\t}\n\tif args == nil {\n\t\targs = &DatabaseArgs{}\n\t}\n\tvar resource Database\n\terr := ctx.RegisterResource(\"gcp:sql/database:Database\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func New(dl logbook.Logbook, databaseName string) (*Database, error) {\n\tdb := &Database{\n\t\tDlog: dl,\n\t}\n\n\t// Check if data folder exists\n\t_, err := os.Stat(config.DBFolder)\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\t// If folder does not exist, create it\n\t\t\terr := os.Mkdir(config.DBFolder, 0777)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, errors.New(\"error\")\n\t\t\t}\n\t\t}\n\t}\n\n\terr = os.Chdir(config.DBFolder)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn nil, errors.New(\"error\")\n\t}\n\n\tf, err := os.Create(databaseName)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn nil, errors.New(\"error\")\n\t}\n\n\tdefer f.Close()\n\n\tdb.DatabaseName = databaseName\n\tdb.Version = 001\n\tdb.File = f\n\tdb.Data = map[string][]byte{}\n\n\twr := bufio.NewWriter(f)\n\n\t_, err = fmt.Fprintf(wr, \"DolceDB.%d\", config.DBVersion)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\twr.Flush()\n\n\terr = db.RebuildMap()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func dbNew() *DB {\n\treturn &DB{\n\t\tdata: make(map[string]string),\n\t}\n}", "func New(db *sql.DB) *Database {\n\treturn &Database{\n\t\tdb: db,\n\t}\n}", "func New(db *sql.DB) *Database {\n\treturn &Database{\n\t\tdb: db,\n\t}\n}", "func NewDatabase(vendor, model, path string) *Database {\n\t// Check files\n\tif _, err := os.Stat(path); os.IsNotExist(err) {\n\t\tlogrus.WithError(err).Fatal(\"[DB] Directory not found.\")\n\t\treturn nil\n\t}\n\n\td := &Database{\n\t\tVendor: vendor,\n\t\tModel: model,\n\t\tFilePath: fmt.Sprintf(\"%s/%s_%s.json\", path, vendor, model),\n\t}\n\n\t// Check file is not exists\n\tif _, err := os.Stat(d.FilePath); os.IsNotExist(err) {\n\t\tlogrus.Infof(\"[DB] File not found, Creating...\")\n\n\t\t// Create DB\n\t\td.State = generateState()\n\n\t\t// Save\n\t\tif err := d.Save(); err != nil {\n\t\t\tpanic(err)\n\t\t}\n\t}\n\n\t// Load\n\tif err := d.Load(); err != nil {\n\t\tpanic(err)\n\t}\n\n\t// Check State\n\t// if state has not exists, fill state from template.\n\treturn d\n}", "func NewDatabase(ctx context.Context, env env.Interface, uuid uuid.UUID, dbid string) (db *Database, err error) {\n\tdatabaseAccount, masterKey := env.CosmosDB(ctx)\n\n\th := &codec.JsonHandle{\n\t\tBasicHandle: codec.BasicHandle{\n\t\t\tDecodeOptions: codec.DecodeOptions{\n\t\t\t\tErrorIfNoField: true,\n\t\t\t},\n\t\t},\n\t}\n\n\terr = api.AddExtensions(&h.BasicHandle)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc := &http.Client{\n\t\tTransport: &http.Transport{\n\t\t\t// disable HTTP/2 for now: https://github.com/golang/go/issues/36026\n\t\t\tTLSNextProto: map[string]func(string, *tls.Conn) http.RoundTripper{},\n\t\t},\n\t\tTimeout: 30 * time.Second,\n\t}\n\n\tdbc, err := cosmosdb.NewDatabaseClient(c, h, databaseAccount, masterKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb = &Database{}\n\n\tdb.OpenShiftClusters, err = NewOpenShiftClusters(ctx, uuid, dbc, dbid, \"OpenShiftClusters\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Subscriptions, err = NewSubscriptions(ctx, uuid, dbc, dbid, \"Subscriptions\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func NewDatabase() *Database {\n\tdb := new(Database)\n\tdb.Taken = make(map[string]Appointment, 0)\n\treturn db\n}", "func NewDatabase() (*Database, error) {\n\tdatabase := &Database{}\n\n\t// Store a reference to the Logger\n\tdatabase.logger = logger.GetLogger()\n\n\t// Get the database path\n\tdbPath, err := getDatabasePath()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdatabase.Path = dbPath\n\n\treturn database, nil\n}", "func New(db *sql.DB) *Database {\n\treturn &Database{\n\t\tUsers: users.New(db),\n\t\tSessions: sessions.New(db),\n\t\tWorkouts: workouts.New(db),\n\t\tExercises: exercises.New(db),\n\t}\n}", "func NewDatabase(ctx context.Context, conf *Config) (cli *Database, err error) {\n\tclient, err := client(ctx, conf)\n\tif err != nil {\n\t\tfmt.Println(\"new client fail\", err)\n\t\treturn\n\t}\n\tdb := client.Database(conf.Database)\n\n\tcli = &Database{\n\t\tdatabase: db,\n\t}\n\treturn\n}", "func NewDatabase(ctx *pulumi.Context,\n\tname string, args *DatabaseArgs, opts ...pulumi.ResourceOption) (*Database, error) {\n\tif args == nil {\n\t\treturn nil, errors.New(\"missing one or more required arguments\")\n\t}\n\n\tif args.LocationId == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'LocationId'\")\n\t}\n\tif args.Type == nil {\n\t\treturn nil, errors.New(\"invalid value for required argument 'Type'\")\n\t}\n\topts = internal.PkgResourceDefaultOpts(opts)\n\tvar resource Database\n\terr := ctx.RegisterResource(\"gcp:firestore/database:Database\", name, args, &resource, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resource, nil\n}", "func NewDatabase(ctx context.Context, log *logrus.Entry, env env.Interface, m metrics.Interface, cipher encryption.Cipher, uuid string) (db *Database, err error) {\n\tdatabaseAccount, masterKey := env.CosmosDB()\n\n\th := newJSONHandle(cipher)\n\n\tc := &http.Client{\n\t\tTransport: dbmetrics.New(log, &http.Transport{\n\t\t\t// disable HTTP/2 for now: https://github.com/golang/go/issues/36026\n\t\t\tTLSNextProto: map[string]func(string, *tls.Conn) http.RoundTripper{},\n\t\t\tMaxIdleConnsPerHost: 20,\n\t\t}, m),\n\t\tTimeout: 30 * time.Second,\n\t}\n\n\tdbc, err := cosmosdb.NewDatabaseClient(log, c, h, databaseAccount, masterKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb = &Database{\n\t\tlog: log,\n\t\tm: m,\n\t}\n\n\tdb.AsyncOperations, err = NewAsyncOperations(uuid, dbc, env.DatabaseName(), \"AsyncOperations\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Billing, err = NewBilling(ctx, uuid, dbc, env.DatabaseName(), \"Billing\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Monitors, err = NewMonitors(ctx, uuid, dbc, env.DatabaseName(), \"Monitors\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.OpenShiftClusters, err = NewOpenShiftClusters(ctx, uuid, dbc, env.DatabaseName(), \"OpenShiftClusters\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Subscriptions, err = NewSubscriptions(ctx, uuid, dbc, env.DatabaseName(), \"Subscriptions\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func NewDatabase(ctx context.Context, log *logrus.Entry, env env.Interface, m metrics.Interface, uuid string) (db *Database, err error) {\n\tdatabaseAccount, masterKey := env.CosmosDB()\n\n\th := &codec.JsonHandle{\n\t\tBasicHandle: codec.BasicHandle{\n\t\t\tDecodeOptions: codec.DecodeOptions{\n\t\t\t\tErrorIfNoField: true,\n\t\t\t},\n\t\t},\n\t}\n\n\terr = api.AddExtensions(&h.BasicHandle)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tc := &http.Client{\n\t\tTransport: dbmetrics.New(log, &http.Transport{\n\t\t\t// disable HTTP/2 for now: https://github.com/golang/go/issues/36026\n\t\t\tTLSNextProto: map[string]func(string, *tls.Conn) http.RoundTripper{},\n\t\t\tMaxIdleConnsPerHost: 20,\n\t\t}, m),\n\t\tTimeout: 30 * time.Second,\n\t}\n\n\tdbc, err := cosmosdb.NewDatabaseClient(log, c, h, databaseAccount, masterKey)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb = &Database{\n\t\tlog: log,\n\t\tm: m,\n\t}\n\n\tdb.AsyncOperations, err = NewAsyncOperations(uuid, dbc, env.DatabaseName(), \"AsyncOperations\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Monitors, err = NewMonitors(ctx, uuid, dbc, env.DatabaseName(), \"Monitors\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.OpenShiftClusters, err = NewOpenShiftClusters(ctx, uuid, dbc, env.DatabaseName(), \"OpenShiftClusters\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdb.Subscriptions, err = NewSubscriptions(ctx, uuid, dbc, env.DatabaseName(), \"Subscriptions\")\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tgo db.emitMetrics(ctx)\n\n\treturn db, nil\n}", "func NewDatabase(host, port, name string) (Database, error) {\n\tdb := Database{host, port, name, nil,\n\t\tmap[string][]string{}, net.Dial, defaultChangeDelay}\n\tif !db.Running() {\n\t\treturn db, errNotRunning\n\t}\n\tif !db.Exists() {\n\t\tif err := db.createDatabase(); err != nil {\n\t\t\treturn db, err\n\t\t}\n\t}\n\treturn db, nil\n}", "func (s *Schema) NewDatabase(log MyLogger, name string, db *sql.DB, driver Driver) (*Database, error) {\n\tif _, ok := s.databases[name]; ok {\n\t\treturn nil, errors.Errorf(\"Duplicate database '%s'\", name)\n\t}\n\tdatabase := &Database{\n\t\tname: name,\n\t\tdb: db,\n\t\tbyLibrary: make(map[string][]Migration),\n\t\tmigrationIndex: make(map[MigrationName]Migration),\n\t\tparent: s,\n\t\tOptions: s.options,\n\t\tContext: s.context,\n\t\tdriver: driver,\n\t\tlog: log,\n\t}\n\ts.databases[name] = database\n\ts.databaseOrder = append(s.databaseOrder, database)\n\treturn database, nil\n}", "func NewDatabase(databaseName string, reset bool) *sql.DB {\n\t// Use default name if not specified\n\tif databaseName == \"\" {\n\t\t// check if the storage directory exist, if not create it\n\t\tstorage, err := os.Stat(\"./storage\")\n\t\tif err != nil {\n\t\t\tos.Mkdir(\"./storage\", os.ModeDir)\n\t\t} else if !storage.IsDir() {\n\t\t\t// check if the storage is indeed a directory or not\n\t\t\tlog.Fatal(\"\\\"storage\\\" exist but is not a directory\")\n\t\t}\n\t\tdatabaseName = \"./storage/db.sqlite\"\n\t}\n\tif reset {\n\t\tos.Remove(databaseName)\n\t}\n\tdb, err := sql.Open(\"sqlite3\", databaseName)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tsqlStmt := `CREATE TABLE IF NOT EXISTS User (\n nick TEXT NOT NULL PRIMARY KEY,\n email TEXT);`\n\t_, err = db.Exec(sqlStmt)\n\tif err != nil {\n\t\tlog.Fatalf(\"%q: %s\\n\", err, sqlStmt)\n\t}\n\n\tdbPtr = db\n\treturn db\n}", "func NewDatabase() *Database {\n\tdb := &Database{ctx: context.Background()}\n\tdb.Init()\n\n\treturn db\n}", "func NewDatabase(path, name string) *Database {\n\tdb, err := sql.Open(\"sqlite3\", path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tschema, ok := schemas[name]\n\tif !ok {\n\t\tpanic(fmt.Errorf(\"Unknown schema: %q\", name))\n\t}\n\tthis := &Database{db: db}\n\tthis.apply(schema)\n\treturn this\n}", "func NewDatabase(settings *settings.Settings) Database {\n\n\tgeneralDatabase := &generalDatabase{\n\t\tdriver: dbTypeToDriverMap[settings.DbType],\n\t\tsettings: settings,\n\t}\n\n\tswitch settings.DbType {\n\tcase \"mysql\":\n\t\treturn &mysql{generalDatabase}\n\tdefault: // pg\n\t\treturn &postgresql{generalDatabase}\n\t}\n}", "func newDatabase(count int) (*database, error) {\n\tdb, err := sql.Open(\"postgres\", fmt.Sprintf(`host=%s port=%s user=%s\n\t\tpassword=%s dbname=%s sslmode=disable`,\n\t\thost, port, user, password, dbname))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := db.Ping(); err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Printf(\"connected to psql client, host: %s\\n\", host)\n\n\treturn &database{\n\t\tdb: db,\n\t\terrChan: make(chan error, count),\n\t}, nil\n}", "func NewDatabase(store LeaseStorage) *Database {\n\treturn &Database{\n\t\tstore: store,\n\t\tl: log.Log,\n\t}\n}", "func New(ctx context.Context, ng engine.Engine) (*DB, error) {\n\treturn newDatabase(ctx, ng, database.Options{Codec: msgpack.NewCodec()})\n}", "func New(host string) *Database {\n\treturn &Database{\n\t\tHost: host,\n\t}\n}", "func createNewDB(log *logrus.Entry, cnf *Config) error {\n\tvar err error\n\n\tpsqlInfo := fmt.Sprintf(\"host=%s port=%d user=%s password=%s dbname=%s sslmode=disable\",\n\t\tcnf.DBHost, cnf.DBPort, cnf.DBUser, cnf.DBPassword, cnf.DBName)\n\n\tdb, err = sql.Open(\"postgres\", psqlInfo)\n\n\tif err != nil {\n\t\tlog.WithError(err).Fatalf(\"Failed to connect to db\")\n\t}\n\n\t//try to ping the db\n\terr = db.Ping()\n\tif err != nil {\n\t\tlog.WithError(err).Fatalf(\"Failed to ping db\")\n\t}\n\n\tif err = helpers.MigrateDB(db, cnf.SQLMigrationDir); err != nil {\n\t\treturn err\n\t}\n\n\tboil.SetDB(db)\n\n\treturn nil\n}", "func NewDatabase(dataDir string, opts ...Option) (*Database, error) {\n\toptions := Options{\n\t\tHost: \"localhost\",\n\t\tDialect: \"sqlite3\",\n\t}\n\tif err := options.Apply(opts...); err != nil {\n\t\treturn nil, err\n\t}\n\n\tdbPath := path.Join(dataDir, dbName)\n\tvar dialector gorm.Dialector\n\n\tswitch strings.ToLower(options.Dialect) {\n\tcase \"memory\":\n\t\tdbPath = \":memory:\"\n\t\tdialector = sqlite.Open(dbPath)\n\tcase \"mysql\":\n\t\tdbPath = fmt.Sprintf(\"%s:%s@(%s:%d)/%s?charset=utf8&parseTime=True\", options.User, options.Password, options.Host, options.Port, dbName)\n\t\tdialector = mysql.Open(dbPath)\n\tcase \"postgress\":\n\t\tdbPath = fmt.Sprintf(\"host=%s port=%d user=%s dbname=%s password=%s\", options.Host, options.Port, options.User, dbName, options.Password)\n\t\tdialector = postgres.Open(dbPath)\n\tcase \"sqlite3\":\n\t\tdbPath = dbPath + \".db\"\n\t\tdialector = sqlite.Open(dbPath)\n\t\tbreak\n\tdefault:\n\t\treturn nil, errors.New(\"unknown database dialect\")\n\t}\n\n\tdb, err := gorm.Open(dialector, &gorm.Config{\n\t\tAllowGlobalUpdate: true,\n\t\tLogger: silentLogger,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := db.AutoMigrate(&models.User{}, &models.Dataset{}, &models.Purchase{}, &models.Click{}); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &Database{db: db}, nil\n}", "func NewDatabase(dbname string) (*Database, error) {\n\tret := &Database{}\n\n\tdb, err := sql.Open(\"sqlite3\", dbname)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = db.Ping()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tret.db = db\n\n\treturn ret, nil\n}", "func newClientDatabase(database string) (*clientDatabase, error) {\n\tif database == \"\" {\n\t\tusr, err := user.Current()\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tdatabase = path.Join(usr.HomeDir, _DEFAULT_DATABASE_PATH)\n\t}\n\tos.MkdirAll(path.Dir(database), os.ModePerm)\n\tdb, err := sql.Open(\"sqlite\", database)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\t_, err = db.Exec(`\n\tCREATE TABLE IF NOT EXISTS identity (\n\t\tid BOOLEAN PRIMARY KEY CONSTRAINT one_row CHECK (id) NOT NULL,\n\t\tpublic BLOB NOT NULL,\n\t\tprivate BLOB NOT NULL\n\t);\n\n\tCREATE TABLE IF NOT EXISTS friend (\n \t\tpublic BLOB PRIMARY KEY NOT NULL,\n \tname TEXT NOT NULL\n\t);\n\n\tCREATE TABLE IF NOT EXISTS prekey (\n\t\tpublic BLOB PRIMARY KEY NOT NULL,\n\t\tprivate BLOB NOT NULL\n\t);\n\n\tCREATE TABLE IF NOT EXISTS onetime (\n\t\tpublic BLOB PRIMARY KEY NOT NUll,\n\t\tprivate BLOB NOT NULL\n\t);\n\t`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &clientDatabase{db}, nil\n}", "func New() (*Database, error) {\n\tvar err error\n\tvar db *memdb.MemDB\n\tonce.Do(func() {\n\t\tinstance = &Database{}\n\t\tdb, err = instance.createSchema()\n\t\tinstance.db = db\n\t\tif err == nil {\n\t\t\tinstance.loadDefaults()\n\t\t}\n\t})\n\treturn instance, err\n}", "func (userAuthorizationServerObject *userAuthorizationEngineServerObjectStruct) createNewDatabase() {\n\tvar err error\n\n\t// Create the database-file\n\tuserAuthorizationServerObject.logger.WithFields(logrus.Fields{\n\t\t\"Id\": \"4252df43-4f97-4bd7-bf38-7704482606e7\",\n\t\t\"userAuthorizationServerObject.databaseName\": userAuthorizationServerObject.databaseName,\n\t}).Info(\"Creating a new database\")\n\n\tdatabaseFile, err := os.Create(userAuthorizationServerObject.databaseName)\n\n\t// If not succeeded then exit program because something is not as intended.\n\tif err != nil {\n\t\tuserAuthorizationServerObject.logger.WithFields(logrus.Fields{\n\t\t\t\"Id\": \"860bdc01-4e47-481f-b8e8-bcfce399538c\",\n\t\t\t\"err.Error()\": err.Error(),\n\t\t}).Panic(\"Exiting because couldn't create a new database file\")\n\t} else {\n\n\t\t//Success in creating database\n\t\tuserAuthorizationServerObject.logger.WithFields(logrus.Fields{\n\t\t\t\"Id\": \"31a363a8-8c79-4775-a2c2-6cc60f2e872e\",\n\t\t\t\"userAuthorizationServerObject.databaseName\": userAuthorizationServerObject.databaseName,\n\t\t}).Debug(\"Success in creating a new database\")\n\n\t\t// Close file\n\t\terr = databaseFile.Close()\n\t\t// If not succeeded then exit program because something is not as intended.\n\t\tif err != nil {\n\t\t\tuserAuthorizationServerObject.logger.WithFields(logrus.Fields{\n\t\t\t\t\"Id\": \"7b9455ef-f06b-4af2-a875-201c83bb7f30\",\n\t\t\t\t\"err.Error()\": err.Error(),\n\t\t\t}).Panic(\"Exiting because couldn't close the new database file\")\n\t\t} else {\n\n\t\t\t// Fill newly create database with standardized data\n\t\t\tuserAuthorizationServerObject.FillDatabaseWithStandardizedData()\n\t\t}\n\t}\n\n}", "func NewDatabase(config Config) (db *gorm.DB, err error) {\n\tgorm.NowFunc = func() time.Time {\n\t\treturn time.Now().UTC()\n\t}\n\tdb, err = gorm.Open(\"mysql\", connectionString(config))\n\tif nil == err && config.ShowSQL {\n\t\tdb.LogMode(true)\n\t}\n\treturn\n}", "func NewDatabase() (ret OpenDB, err error) {\n\tdb := C.DatabaseNew()\n\tif db == nil {\n\t\tcErr := C.GoString(C.LastError)\n\t\tif len(cErr) > 0 {\n\t\t\terr = fmt.Errorf(\"%v\", C.GoString(C.LastError))\n\t\t} else {\n\t\t\terr = fmt.Errorf(\"Unknown error has occured\")\n\t\t}\n\t}\n\tret.db = db\n\treturn ret, err\n}", "func New(tb testing.TB, dir string, ver string, logger func(string)) (*DB, error) {\n\tdb, err := doNew(tb, dir, ver, logger)\n\tif err != nil && tb != nil {\n\t\ttb.Fatal(\"failed initializing database: \", err)\n\t}\n\treturn db, err\n}", "func New() (database, error) {\n\tdb, err := sql.Open(driverName, databaseName)\n\tif err != nil {\n\t\treturn database{}, err\n\t}\n\n\treturn database{db}, nil\n}", "func NewDatabase() Database {\n\tdb, err := sql.Open(\"postgres\", \"postgresql://root@localhost:26257?sslcert=%2Fhome%2Fubuntu%2Fnode1.cert&sslkey=%2Fhome%2Fubuntu%2Fnode1.key&sslmode=verify-full&sslrootcert=%2Fhome%2Fubuntu%2Fca.cert\")\n\tif err != nil {\n\t\tlog.Fatalln(\"database connection:\", err)\n\t}\n\tif _, err := db.Exec(\"CREATE DATABASE IF NOT EXISTS nfinite\"); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif _, err := db.Exec(\"SET DATABASE = nfinite\"); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif _, err = db.Exec(\"CREATE TABLE IF NOT EXISTS PartLookup (id SERIAL PRIMARY KEY, partId INT, ownerId INT);\"); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif _, err = db.Exec(\"CREATE TABLE IF NOT EXISTS FilePart (parentId INT, name string, id SERIAL PRIMARY KEY, fileIndex INT);\"); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif _, err = db.Exec(\"CREATE TABLE IF NOT EXISTS Client (id SERIAL, username string PRIMARY KEY, password string);\"); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\tif _, err = db.Exec(\"CREATE TABLE IF NOT EXISTS File (id SERIAL PRIMARY KEY, modified INT, name string, ownerId INT);\"); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn Database{db}\n}", "func NewDatabase(c config.Config, s *storage.Manager, k *k8s.Database) (Database, error) {\n\tif c.SideCar == nil || *c.SideCar {\n\t\tc.Database.Host = \"127.0.0.1\"\n\t}\n\tif c.Database.Type == constants.MARIADB {\n\t\treturn NewMariaDB(c, s, k)\n\t} else if c.Database.Type == constants.POSTGRES {\n\t\treturn NewPostgres(c, s)\n\t}\n\n\treturn nil, fmt.Errorf(\"unsupported database\")\n}", "func NewDatabase(env Env, logger Logger) Database {\n\n\tusername := env.DBUsername\n\tpassword := env.DBPassword\n\thost := env.DBHost\n\tport := env.DBPort\n\tdbname := env.DBName\n\n\turl := fmt.Sprintf(\"%s:%s@tcp(%s:%s)/%s?charset=utf8mb4&parseTime=True&loc=Local\", username, password, host, port, dbname)\n\n\tdb, err := gorm.Open(\"mysql\", url)\n\n\tif err != nil {\n\t\tlogger.Zap.Info(\"Url: \", url)\n\t\tlogger.Zap.Panic(err)\n\t}\n\n\tlogger.Zap.Info(\"Database connection established\")\n\n\treturn Database{\n\t\tDB: db,\n\t}\n}", "func NewDb(db *sql.DB, driverName string) *DB {\n return &DB{DB: db, driverName: driverName, Mapper: mapper()}\n}", "func NewDatabase(dataSourceName string) (*Database, error) {\n\tvar d Database\n\tvar err error\n\tif d.db, err = sql.Open(ScrapperDBDriver, dataSourceName); err != nil {\n\t\treturn nil, err\n\t}\n\tif err = d.scrapper.prepare(d.db); err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"complete storage setting\")\n\treturn &d, nil\n}", "func NewDatabase(location string) (*Database, error) {\n\tdb, err := sql.Open(\"sqlite3\", location)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Database{\n\t\tdb: db,\n\t}, nil\n}", "func (e *engine) CreateDatabase(databaseName string) (Database, error) {\n\tdbPath := filepath.Join(e.cfg.Dir, databaseName)\n\tif err := mkDirIfNotExist(dbPath); err != nil {\n\t\treturn nil, fmt.Errorf(\"create database[%s]'s path with error: %s\", databaseName, err)\n\t}\n\tcfgPath := optionsPath(dbPath)\n\tcfg := &databaseConfig{}\n\tif fileutil.Exist(cfgPath) {\n\t\tif err := decodeToml(cfgPath, cfg); err != nil {\n\t\t\treturn nil, fmt.Errorf(\"load database[%s] config from file[%s] with error: %s\",\n\t\t\t\tdatabaseName, cfgPath, err)\n\t\t}\n\t}\n\tdb, err := newDatabaseFunc(databaseName, dbPath, cfg, e.dataFlushChecker)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\te.databases.Store(databaseName, db)\n\treturn db, nil\n}", "func NewDatabase(host string, port int, username string, password string, dbname string) (*Database, error) {\n\tdsn := fmt.Sprintf(\"%s:%s@tcp(%s:%d)/%s?charset=utf8mb4&parseTime=True&loc=Local\",\n\t\tusername, password, host, port, dbname)\n\n\t// TODO: Setup logger for gorm here\n\t// I should implement GORM's logger interface on core.AppLogger and pass it to gorm config.\n\t// I should also pass log level to GORM.\n\n\t// dbconn, err := gorm.Open(mysql.Open(dsn), &gorm.Config{})\n\t// dbconn = dbconn.Debug()\n\tdbconn, err := gorm.Open(mysql.Open(dsn), &gorm.Config{Logger: logger.Default.LogMode(logger.Silent)})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tdbconn = dbconn.Session(&gorm.Session{})\n\tdb := Database{conn: dbconn}\n\treturn &db, nil\n}", "func MakeDatabase(driver, user, pass, uri string) (*Database, error) {\n\t// https://github.com/pxlsspace/Pxls/blob/master/src/main/java/space/pxls/data/Database.java#L49\n\tpURI, err := url.Parse(uri)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tconnConf := mysql.Config{\n\t\tNet: \"tcp\",\n\t\tAddr: pURI.Host,\n\t\tDBName: pURI.Path[1:],\n\t\tUser: user,\n\t\tPasswd: pass,\n\t\tMultiStatements: true,\n\t\tParseTime: true,\n\t\tAllowNativePasswords: true,\n\t}\n\tconn, err := gorm.Open(driver, connConf.FormatDSN())\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Generate tables and migrate them when a difference with the models is detected.\n\tconn.AutoMigrate(&DBPixel{}, &DBUser{}, &DBSession{})\n\n\treturn &Database{\n\t\tsql: conn,\n\t\tdriver: driver,\n\t}, nil\n}", "func New(ctx context.Context, ng engine.Engine, opts Options) (*Database, error) {\n\tif opts.Codec == nil {\n\t\treturn nil, errors.New(\"missing codec\")\n\t}\n\n\tdb := Database{\n\t\tng: ng,\n\t\tCodec: opts.Codec,\n\t}\n\n\tntx, err := db.ng.Begin(ctx, engine.TxOptions{\n\t\tWritable: true,\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer ntx.Rollback()\n\n\terr = db.initInternalStores(ntx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = ntx.Commit()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &db, nil\n}", "func NewDatabase(conf *util.ConfigReader) Database {\n\tdbConf := conf.MustGet(\"database\")\n\n\t// CONNECTION\n\n\tusername := dbConf.MustString(\"username\")\n\tpassword := dbConf.MustString(\"password\")\n\thostname := dbConf.MustString(\"hostname\")\n\tport := dbConf.MustInt(\"port\")\n\tdatabase := dbConf.MustString(\"database\")\n\tschema := dbConf.MustString(\"schema\")\n\toptions := dbConf.MustMap(\"options\")\n\toptions[\"search_path\"] = schema\n\toptionsString := toURLQuery(options)\n\tconnStr := fmt.Sprintf(\n\t\t\"postgresql://%s:%s@%s:%d/%s%s\",\n\t\tusername, password, hostname, port, database, optionsString)\n\tlog.Infof(\"Connecting to database using connection string: %s\", connStr)\n\tdb, err := sql.Open(\"postgres\", connStr)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\tpanic(err)\n\t}\n\n\t// EVOLUTION\n\n\terr = migrateDatabase(db, database, dbConf.MustGet(\"evolution\"), schema)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t\tpanic(err)\n\t}\n\n\t// SERVICE\n\n\tcheckQueriesOnInit := dbConf.UBool(\"check_queries_on_init\", false)\n\treturn &databaseService{db, checkQueriesOnInit}\n}", "func New(config *Config) (Database, error) {\n\tdb, err := connectToDB(config)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfmt.Println(\"Successfully connected to db\")\n\n\terr = migrateDB(config, db)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tfmt.Println(\"Successfully ran migrations\")\n\n\tsqlxDB := sqlx.NewDb(db, config.Driver)\n\n\tbaseDB := database{\n\t\tdb: sqlxDB,\n\t}\n\n\treturn &queries{\n\t\tauthorsQueries{baseDB},\n\t\tpostsQueries{baseDB},\n\t}, nil\n}", "func (s *API) CreateDatabase(req *CreateDatabaseRequest, opts ...scw.RequestOption) (*Database, error) {\n\tvar err error\n\n\tif req.Region == \"\" {\n\t\tdefaultRegion, _ := s.client.GetDefaultRegion()\n\t\treq.Region = defaultRegion\n\t}\n\n\tif fmt.Sprint(req.Region) == \"\" {\n\t\treturn nil, errors.New(\"field Region cannot be empty in request\")\n\t}\n\n\tif fmt.Sprint(req.InstanceID) == \"\" {\n\t\treturn nil, errors.New(\"field InstanceID cannot be empty in request\")\n\t}\n\n\tscwReq := &scw.ScalewayRequest{\n\t\tMethod: \"POST\",\n\t\tPath: \"/rdb/v1/regions/\" + fmt.Sprint(req.Region) + \"/instances/\" + fmt.Sprint(req.InstanceID) + \"/databases\",\n\t\tHeaders: http.Header{},\n\t}\n\n\terr = scwReq.SetBody(req)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar resp Database\n\n\terr = s.client.Do(scwReq, &resp, opts...)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &resp, nil\n}", "func CreateNew(db *mongo.Database) *DB {\n\treturn &DB{\n\t\tconn: db,\n\t}\n}", "func (rt *RestTester) CreateDatabase(dbName string, config DbConfig) *TestResponse {\n\tdbcJSON, err := base.JSONMarshal(config)\n\trequire.NoError(rt.TB, err)\n\tresp := rt.SendAdminRequest(http.MethodPut, fmt.Sprintf(\"/%s/\", dbName), string(dbcJSON))\n\treturn resp\n}", "func NewDatabase(mel *Melodious, addr string) (*Database, error) {\n\tdb, err := sql.Open(\"postgres\", addr)\n\n\t_, err = db.Exec(`CREATE SCHEMA IF NOT EXISTS melodious;`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create melodious schema\")\n\n\t_, err = db.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS melodious.accounts (\n\t\t\tid serial NOT NULL PRIMARY KEY,\n\t\t\tusername varchar(32) NOT NULL UNIQUE,\n\t\t\tpasshash varchar(64) NOT NULL,\n\t\t\towner BOOLEAN NOT NULL,\n\t\t\tbanned BOOLEAN NOT NULL DEFAULT false,\n\t\t\tip inet NOT NULL DEFAULT '0.0.0.0'\n\t\t);`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create accounts table\")\n\n\t_, err = db.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS melodious.channels (\n\t\t\tid serial NOT NULL PRIMARY KEY,\n\t\t\tname varchar(32) NOT NULL UNIQUE,\n\t\t\ttopic varchar(128) NOT NULL\n\t\t);`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create channels table\")\n\n\t_, err = db.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS melodious.messages (\n\t\t\tid serial NOT NULL PRIMARY KEY,\n\t\t\tchan_id int4 NOT NULL REFERENCES melodious.channels(id) ON DELETE CASCADE,\n\t\t\tauthor_id int4 NOT NULL REFERENCES melodious.accounts(id) ON DELETE CASCADE,\n\t\t\tmessage varchar(2048) NOT NULL,\n\t\t\tdt timestamp with time zone NOT NULL,\n\t\t\tpings varchar(32) []\n\t\t);`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create messages table\")\n\n\t_, err = db.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS melodious.groups (\n\t\t\tid serial NOT NULL PRIMARY KEY,\n\t\t\tname varchar(32) NOT NULL\n\t\t);`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create groups table\")\n\n\t_, err = db.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS melodious.group_holders (\n\t\t\tid serial NOT NULL PRIMARY KEY,\n\t\t\tgroup_id int4 NOT NULL REFERENCES melodious.groups(id) ON DELETE CASCADE,\n\t\t\tuser_id int4 REFERENCES melodious.accounts(id) ON DELETE CASCADE,\n\t\t\tchannel_id int4 REFERENCES melodious.channels(id) ON DELETE CASCADE,\n\t\t\tUNIQUE(group_id, user_id, channel_id)\n\t\t);`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create group_holders table\")\n\n\t_, err = db.Exec(`\n\t\tCREATE TABLE IF NOT EXISTS melodious.group_flags (\n\t\t\tid serial NOT NULL PRIMARY KEY,\n\t\t\tgroup_id int4 NOT NULL REFERENCES melodious.groups(id) ON DELETE CASCADE,\n\t\t\tname varchar(32) NOT NULL,\n\t\t\tflag jsonb NOT NULL,\n\t\t\tUNIQUE(group_id, name)\n\t\t);`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create group_flags table\")\n\n\t_, err = db.Exec(`\n\t\tCREATE OR REPLACE FUNCTION melodious.set_flag(group_name varchar(32), flag_name varchar(32), flag_data jsonb)\n\t\tRETURNS int4\n\t\tLANGUAGE plpgsql\n\t\tAS $$\n\t\tDECLARE\n\t\t\tfid int4 := NULL;\n\t\t\tgid int4 := NULL;\n\t\tBEGIN\n\t\t\tSELECT id INTO gid FROM melodious.groups WHERE name=group_name;\n\t\t\tSELECT id INTO fid FROM melodious.group_flags WHERE group_id=gid AND name=flag_name;\n\t\t\tIF fid IS NULL THEN\n\t\t\t\tINSERT INTO melodious.group_flags (group_id, name, flag) VALUES (gid, flag_name, flag_data) RETURNING id INTO fid;\n\t\t\t\tRETURN fid;\n\t\t\tELSE\n\t\t\t\tUPDATE melodious.group_flags SET flag=flag_data WHERE id=fid;\n\t\t\t\tRETURN fid;\n\t\t\tEND IF;\n\t\tEND;\n\t\t$$;\n\t`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create set_flag function\")\n\n\t_, err = db.Exec(`\n\t\tCREATE OR REPLACE PROCEDURE melodious.delete_flag(group_name varchar(32), flag_name varchar(32))\n\t\tLANGUAGE plpgsql\n\t\tAS $$\n\t\tDECLARE\n\t\t\tgid int4 := NULL;\n\t\tBEGIN\n\t\t\tSELECT id INTO gid FROM melodious.groups WHERE name=group_name;\n\t\t\tIF gid IS NOT NULL THEN\n\t\t\t\tDELETE FROM melodious.group_flags WHERE group_id=gid AND name=flag_name;\n\t\t\tEND IF;\n\t\tEND;\n\t\t$$;\n\t`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create delete_flag procedure\")\n\n\t_, err = db.Exec(`\n\t\tCREATE OR REPLACE FUNCTION melodious.insert_group_holder(group_name varchar(32), user_name varchar(32), chan_name varchar(32))\n\t\tRETURNS int4\n\t\tLANGUAGE plpgsql\n\t\tAS $$\n\t\tDECLARE\n\t\t\tgid int4 := NULL;\n\t\t\tuid int4 := NULL;\n\t\t\tcid int4 := NULL;\n\t\t\tghid int4 := NULL;\n\t\tBEGIN\n\t\t\tSELECT id INTO gid FROM melodious.groups WHERE name=group_name;\n\t\t\tIF gid IS NULL THEN\n\t\t\t\tRAISE EXCEPTION 'no such group';\n\t\t\tEND IF;\n\t\t\tIF user_name <> '' THEN\n\t\t\t\tSELECT id INTO uid FROM melodious.accounts WHERE username=user_name;\n\t\t\t\tIF uid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'no such user';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\t\tIF chan_name <> '' THEN\n\t\t\t\tSELECT id INTO cid FROM melodious.channels WHERE name=chan_name;\n\t\t\t\tIF cid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'no such channel';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\t\tINSERT INTO melodious.group_holders (group_id, user_id, channel_id) VALUES (gid, uid, cid) RETURNING id INTO ghid;\n\t\t\tRETURN ghid;\n\t\tEND;\n\t\t$$;\n\t`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create insert_group_holder function\")\n\n\t_, err = db.Exec(`\n\t\tCREATE OR REPLACE PROCEDURE melodious.delete_group_holders(group_name varchar(32), user_name varchar(32), chan_name varchar(32))\n\t\tLANGUAGE plpgsql\n\t\tAS $$\n\t\tDECLARE\n\t\t\tgid int4 := NULL;\n\t\t\tuid int4 := NULL;\n\t\t\tcid int4 := NULL;\n\t\tBEGIN\n\t\t\tIF group_name <> '' THEN\n\t\t\t\tSELECT id INTO gid FROM melodious.groups WHERE name=group_name;\n\t\t\t\tIF gid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'no such group';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\t\tIF user_name <> '' THEN\n\t\t\t\tSELECT id INTO uid FROM melodious.accounts WHERE username=user_name;\n\t\t\t\tIF uid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'no such user';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\t\tIF chan_name <> '' THEN\n\t\t\t\tSELECT id INTO cid FROM melodious.channels WHERE name=chan_name;\n\t\t\t\tIF cid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'no such channel';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\t\t\n\t\t\tIF group_name = '' AND user_name = '' AND chan_name = '' THEN\n\t\t\t\tDELETE FROM melodious.group_holders WHERE group_id=gid;\n\t\t\tELSIF group_name <> '' AND user_name <> '' AND chan_name = '' THEN\n\t\t\t\tDELETE FROM melodious.group_holders WHERE group_id=gid AND user_id=uid;\n\t\t\tELSIF group_name <> '' AND user_name = '' AND chan_name <> '' THEN\n\t\t\t\tDELETE FROM melodious.group_holders WHERE group_id=gid AND channel_id=cid;\n\t\t\tELSIF group_name <> '' AND user_name <> '' AND chan_name <> '' THEN\n\t\t\t\tDELETE FROM melodious.group_holders WHERE group_id=gid AND user_id=uid AND channel_id=cid;\n\t\t\tELSIF group_name = '' AND user_name = '' AND chan_name = '' THEN\n\t\t\t\tRAISE EXCEPTION 'cannot delete all group holders in a single request';\n\t\t\tELSIF group_name = '' AND user_name <> '' AND chan_name = '' THEN\n\t\t\t\tDELETE FROM melodious.group_holders WHERE user_id=uid;\n\t\t\tELSIF group_name = '' AND user_name = '' AND chan_name <> '' THEN\n\t\t\t\tDELETE FROM melodious.group_holders WHERE channel_id=cid;\n\t\t\tELSIF group_name = '' AND user_name <> '' AND chan_name <> '' THEN\n\t\t\t\tDELETE FROM melodious.group_holders WHERE user_id=uid AND channel_id=cid;\n\t\t\tEND IF;\n\t\tEND;\n\t\t$$;\n\t`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create delete_group_holders function\")\n\n\t_, err = db.Exec(`\n\t\tCREATE OR REPLACE FUNCTION melodious.query_flags(user_name varchar(32), chan_name varchar(32), igroup_name varchar(32), iflag_name varchar(32), flagcheck bool)\n\t\tRETURNS TABLE (\n\t\t\tgroup_holders jsonb [],\n\t\t\tflag_id int4,\n\t\t\tflag_name varchar(32),\n\t\t\tflag jsonb\n\t\t)\n\t\tLANGUAGE plpgsql\n\t\tAS $$\n\t\tDECLARE\n\t\t\tuid int4 := NULL;\n\t\t\tcid int4 := NULL;\n\t\t\tgid int4 := NULL;\n\t\tBEGIN\n\t\t\tIF user_name <> '' THEN\n\t\t\t\tSELECT id INTO uid FROM melodious.accounts WHERE username=user_name;\n\t\t\t\tIF uid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'no such user';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\t\tIF chan_name <> '' THEN\n\t\t\t\tSELECT id INTO cid FROM melodious.channels WHERE name=chan_name;\n\t\t\t\tIF cid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'no such channel';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\t\tIF igroup_name <> '' THEN\n\t\t\t\tSELECT id INTO gid FROM melodious.groups WHERE name=igroup_name;\n\t\t\t\tIF gid IS NULL THEN\n\t\t\t\t\tRAISE EXCEPTION 'go such group';\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\n\t\t\tIF igroup_name = '' THEN\n\t\t\t\tIF iflag_name = '' THEN\n\t\t\t\t\tIF user_name = '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (NOT flagcheck) OR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (gh.user_id = uid AND gh.channel_id = cid)\n\t\t\t\t\t\t\t OR (flagcheck AND gh.user_id = uid AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t OR (flagcheck AND gh.user_id IS NULL AND gh.channel_id = cid)\n\t\t\t\t\t\t\t OR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (NOT flagcheck AND gh.user_id = uid)\n\t\t\t\t\t\t\t OR (flagcheck AND gh.user_id = uid AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t OR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name = '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (NOT flagcheck AND gh.channel_id = cid)\n\t\t\t\t\t\t\t OR (flagcheck AND gh.user_id IS NULL AND gh.channel_id = cid)\n\t\t\t\t\t\t\t OR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tEND IF;\n\t\t\t\tELSE\n\t\t\t\t\tIF user_name = '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (gf.name = iflag_name)\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE gf.name = iflag_name\n\t\t\t\t\t\t AND (\n\t\t\t\t\t\t\t\t(gh.user_id = uid AND gh.channel_id = cid)\n\t\t\t\t\t\t\t \tOR (flagcheck AND gh.user_id = uid AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id = cid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE gf.name = iflag_name\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck AND gh.user_id = uid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id = uid AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name = '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB((SELECT name FROM melodious.groups WHERE id=gh.group_id LIMIT 1)))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE gf.name = iflag_name\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck AND gh.channel_id = cid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id = cid AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id IS NULL AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tEND IF;\n\t\t\t\tEND IF;\n\t\t\tELSE\n\t\t\t\tIF iflag_name = '' THEN\n\t\t\t\t\tIF user_name = '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE gh.group_id = gid\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE gh.group_id = gid\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(gh.channel_id = cid AND gh.user_id = uid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id = cid AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id IS NULL AND gh.user_id = uid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id IS NULL AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE gh.group_id = gid\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck AND gh.user_id = uid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id = uid AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name = '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE gh.group_id = gid\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck AND gh.channel_id = cid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id = cid AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id IS NULL AND gh.user_id IS NULL) \n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tEND IF;\n\t\t\t\tELSE\n\t\t\t\t\tIF user_name = '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tINNER JOIN melodious.groups g\n\t\t\t\t\t\tON g.id = gh.group_id\n\t\t\t\t\t\tWHERE (gf.name = iflag_name AND gh.group_id = gid)\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id IS NULL AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (gf.name = iflag_name AND gh.group_id = gid)\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(gh.user_id = uid AND gh.channel_id = cid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id = uid AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id = cid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name <> '' AND chan_name = '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (gh.group_id = gid AND gf.name = iflag_name)\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck AND gh.user_id = uid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id = uid AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.user_id IS NULL AND gh.channel_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tELSIF user_name = '' AND chan_name <> '' THEN\n\t\t\t\t\t\tRETURN QUERY SELECT\n\t\t\t\t\t\t\tARRAY_AGG(JSONB_SET(ROW_TO_JSON(gh)::JSONB, '{group_name}'::TEXT[], TO_JSONB(igroup_name))) AS group_holders,\n\t\t\t\t\t\t\tgf.id flag_id,\n\t\t\t\t\t\t\tgf.name flag_name,\n\t\t\t\t\t\t\tgf.flag flag\n\t\t\t\t\t\tFROM melodious.group_holders gh\n\t\t\t\t\t\tINNER JOIN melodious.group_flags gf\n\t\t\t\t\t\tON gh.group_id = gf.group_id\n\t\t\t\t\t\tWHERE (gf.name = iflag_name AND gh.group_id = gid)\n\t\t\t\t\t\t\tAND (\n\t\t\t\t\t\t\t\t(NOT flagcheck AND gh.channel_id = cid)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id = cid AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t\tOR (flagcheck AND gh.channel_id IS NULL AND gh.user_id IS NULL)\n\t\t\t\t\t\t\t)\n\t\t\t\t\t\tGROUP BY gf.id, gf.name;\n\t\t\t\t\tEND IF;\n\t\t\t\tEND IF;\n\t\t\tEND IF;\n\t\tEND;\n\t\t$$;\n\t`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create query_flag(5) function\")\n\n\t_, err = db.Exec(`\n\t\tCREATE OR REPLACE FUNCTION melodious.query_flags(user_name varchar(32), chan_name varchar(32), igroup_name varchar(32), iflag_name varchar(32))\n\t\tRETURNS TABLE (\n\t\t\tgroup_holders jsonb [],\n\t\t\tflag_id int4,\n\t\t\tflag_name varchar(32),\n\t\t\tflag jsonb\n\t\t)\n\t\tLANGUAGE plpgsql\n\t\tAS $$\n\t\tBEGIN\n\t\t\tRETURN QUERY SELECT melodious.query_flags(user_name, chan_name, igroup_name, iflag_name, false);\n\t\tEND;\n\t\t$$;\n\t`)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlog.Info(\"DB: check/create query_flag(4) function\")\n\n\tdbi := &Database{\n\t\tmel: mel,\n\t\tdb: db,\n\t}\n\n\tgo func() {\n\t\tdhe, err := time.ParseDuration(mel.Config.DeleteHistoryEvery)\n\t\tif err != nil {\n\t\t\tlog.WithField(\"err\", err).Fatal(\"cannot parse duration\")\n\t\t}\n\t\tshf := mel.Config.StoreHistoryFor\n\t\tfor {\n\t\t\tfunc() {\n\t\t\t\tvar err error\n\t\t\t\tdefer log.WithFields(log.Fields{\n\t\t\t\t\t\"storing-for\": shf,\n\t\t\t\t\t\"deleting-every\": dhe,\n\t\t\t\t}).Trace(\"deleting old messages\").Stop(&err)\n\t\t\t\terr = dbi.DeleteOldMessages(shf)\n\t\t\t}()\n\t\t\ttime.Sleep(dhe)\n\t\t}\n\t}()\n\n\treturn dbi, nil\n}", "func NewDB(e *entities.DB) *DB {\n\treturn &DB{e: e}\n}", "func NewDatabase() interfaces.IDatabase {\n\tdbConfig := config.Config.Database\n\tconnectionPath := fmt.Sprintf(\"host=%s port=%d user=%s dbname=%s password=%s sslmode=%s\",\n\t\tdbConfig.Host, dbConfig.Port, dbConfig.User, dbConfig.Name, dbConfig.Password, dbConfig.SSLMode)\n\n\tlogger.Info(connectionPath)\n\n\tdb, err := gorm.Open(\"postgres\", connectionPath)\n\tif err != nil {\n\t\tlogger.Fatal(\"Cannot connect to database: \", err)\n\t}\n\n\t// Set up connection pool\n\tdb.DB().SetMaxIdleConns(20)\n\tdb.DB().SetMaxOpenConns(200)\n\n\treturn &database{\n\t\tdb: db,\n\t}\n}", "func initDatabase() (*database.Database, error) {\n\tvar db *database.Database\n\n\tif *storage == MemoryStorage {\n\t\tdb = database.NewDatabase(database.NewMemoryWriter())\n\t} else if *storage == FileStorage {\n\t\tw, err := database.NewFileWriter(*storageFile)\n\t\tif err != nil {\n\t\t\treturn nil, fmt.Errorf(\"creating file storage error: %s\", err)\n\t\t}\n\t\tdb = database.NewDatabase(w)\n\t}\n\treturn db, nil\n}", "func NewDB() *DB {\n\treturn &DB{}\n}", "func New(config *connect.Config) Db {\n\tvar db Db\n\t// first find db in dbMap by DatabaseName\n\tdb = dbMap[config.DatabaseName]\n\t// find\n\tif db != nil {\n\t\treturn db\n\t}\n\t// not find in dbMap - New\n\tswitch config.DbType {\n\tcase connect.MONGODB:\n\t\t// init mongodb\n\t\tdb = mongo.New(config)\n\t}\n\tdbMap[config.DatabaseName] = db\n\treturn db\n}", "func NewDB(user, password, name string) *DB {\n\treturn &DB{\n\t\tdb: nil,\n\t\tuser: user,\n\t\tpassword: password,\n\t\tname: name}\n}", "func New(opts *Options) (*Database, error) {\n\tdb := new(Database)\n\tif err := opts.initCryptParams(&db.cparams); err != nil {\n\t\treturn nil, err\n\t}\n\tdb.init(nil, nil, opts)\n\treturn db, nil\n}", "func NewDB(dataSourceName string) (*DB, error) {\n\tdb, err := sql.Open(\"sqlite3\", dataSourceName)\n\terr = db.Ping()\n\tif err != nil {\n\t\treturn nil, errors.New(`Unable to open database: destination directory \"data\" does not exist\nPlease check your \"$HOME/.go22/\" directory or run \"go22 init\" to reinitalize the application.`)\n\t}\n\n\t_, err = db.Exec(connTable)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn &DB{db}, nil \n}", "func New(config *Config) (*Database, error) {\n\tdb, err := gorm.Open(\"postgres\", config.DatabaseURI)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"unable to connect to database\")\n\t}\n\treturn &Database{db}, nil\n}", "func (DBM *DBManager) createDatabase(sessionID int64, name string, path string, blocksize int64) (err error) {\n\tdb, err := data.NewDatabase(filepath.Join(path, name), blocksize)\n\tif err != nil {\n\t\treturn err\n\t}\n\tDBM.databases.Store(name, db)\n\treturn DBM.pair(sessionID, name)\n}", "func NewDatabase(path string) *Database {\n\tdb, err := sql.Open(\"sqlite3\", path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn &Database{db}\n}", "func NewDatabase(diskdb ethdb.Database) *Database {\n\treturn NewDatabaseWithCache(diskdb, 0)\n}", "func newDB(client *gorm.DB) (*DB, error) {\n\tif client == nil {\n\t\treturn nil, fmt.Errorf(\"Mysql: could not connect\")\n\t}\n\tdb := &DB{\n\t\tclient: client,\n\t}\n\treturn db, nil\n}", "func TestCreateDatabase(t *testing.T) {\n\n\tobj := testSQL{}\n\tslave := copain{}\n\n\tvar err error\n\n\tdb, err = saver.NewDbManager(\"root:mypass@tcp(127.0.0.1:3306)/greader\")\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\terr = db.CreateTable(&obj)\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n\n\terr = db.CreateTable(&slave)\n\tif err != nil {\n\t\tt.Error(err)\n\t}\n}", "func NewDB(dbfile string) (*DB, error) {\n\tdb, err := gorm.Open(\"sqlite3\", dbfile)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not connect to db\")\n\t}\n\n\tif err := db.AutoMigrate(&Lease{}).Error; err != nil {\n\t\treturn nil, errors.Wrap(err, \"while migrating database\")\n\t}\n\n\treturn &DB{db: db}, nil\n}", "func NewDB(dataSourceName string) (*DB, error) {\n\tdb, err := sql.Open(\"sqlite3\", dataSourceName)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &DB{db}, nil\n}", "func (b Factory) New(ctx context.Context, name string, l log.Interface, cfg *config.Config) (Database, error) {\n\tbackend, ok := b[name]\n\tif !ok {\n\t\treturn nil, errwrap.Wrap(ErrDatabaseNotFound, name)\n\t}\n\n\tdb, err := backend.New(ctx, l, cfg)\n\n\treturn db, errwrap.Wrap(err, \"failed to create backend\")\n}", "func New(ctx context.Context, config *Configuration) (*Database, error) {\n\tdbConfig := fmt.Sprintf(\n\t\t\"host=%s port=%d user=%s password='%s' dbname=%s search_path=%s sslmode=require\",\n\t\tconfig.Host,\n\t\tconfig.Port,\n\t\tconfig.User,\n\t\tconfig.Password,\n\t\tconfig.Name,\n\t\tconfig.Schema,\n\t)\n\n\tdb, err := sql.Open(config.Driver, dbConfig)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn nil, err\n\t}\n\n\treturn &Database{\n\t\tContext: ctx,\n\t\tDatabase: db,\n\t}, nil\n}", "func (s *RaftDatabase) createDatabase() error {\n\tvar volumeClaim *corev1.PersistentVolumeClaim\n\tstorageClass := GetArg(s.name, \"storage\", \"class\").String(\"\")\n\tif storageClass != \"\" {\n\t\tstorageSize := GetArg(s.name, \"storage\", \"size\").String(\"1G\")\n\t\tvolumeClaim = &corev1.PersistentVolumeClaim{\n\t\t\tSpec: corev1.PersistentVolumeClaimSpec{\n\t\t\t\tStorageClassName: &storageClass,\n\t\t\t\tAccessModes: []corev1.PersistentVolumeAccessMode{\n\t\t\t\t\t\"ReadWriteOnce\",\n\t\t\t\t},\n\t\t\t\tResources: corev1.ResourceRequirements{\n\t\t\t\t\tRequests: corev1.ResourceList{\n\t\t\t\t\t\t\"storage\": resource.MustParse(storageSize),\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t}\n\t}\n\n\tdatabase := &v1beta1.Database{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tName: s.Name(),\n\t\t\tNamespace: s.namespace,\n\t\t},\n\t\tSpec: v1beta1.DatabaseSpec{\n\t\t\tClusters: int32(s.Clusters()),\n\t\t\tTemplate: v1beta1.ClusterTemplateSpec{\n\t\t\t\tSpec: v1beta1.ClusterSpec{\n\t\t\t\t\tPartitions: int32(s.NumPartitions()),\n\t\t\t\t\tBackend: v1beta1.Backend{\n\t\t\t\t\t\tReplicas: int32(s.Replicas()),\n\t\t\t\t\t\tImage: s.Image(),\n\t\t\t\t\t\tImagePullPolicy: s.PullPolicy(),\n\t\t\t\t\t\tVolumeClaim: volumeClaim,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\t_, err := s.atomixClient.CloudV1beta1().Databases(s.namespace).Create(database)\n\treturn err\n}", "func NewDatabase(db Execer, dialect schema.Dialect, logger *log.Logger, wrapper interface{}) *Database {\n\tvar enabled int32 = 0\n\tif logger != nil {\n\t\tenabled = 1\n\t}\n\treturn &Database{db, dialect, logger, enabled, wrapper}\n}", "func newEmptyDB(ctx context.Context) (*sql.DB, error) {\n\tdb, err := sql.Open(\"mysql\", dataSource)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Create a randomly-named database and then connect using the new name.\n\tname := fmt.Sprintf(\"mono_%v\", time.Now().UnixNano())\n\n\tstmt := fmt.Sprintf(\"CREATE DATABASE %v\", name)\n\tif _, err := db.ExecContext(ctx, stmt); err != nil {\n\t\treturn nil, fmt.Errorf(\"error running statement %q: %v\", stmt, err)\n\t}\n\tdb.Close()\n\n\tdb, err = sql.Open(\"mysql\", dataSource+name+\"?parseTime=true\")\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to open new database %q: %v\", name, err)\n\t}\n\treturn db, db.Ping()\n}", "func New(host string, port string, user string, pass string, dbname string) (*DataBase, error) {\n\tdsn := \"host=\" + host + \" user=\" + user + \" password=\" + pass + \" dbname=\" + dbname + \" port=\" + port + \" sslmode=disable\" + \" TimeZone=America/Sao_Paulo\"\n\tdb, err := gorm.Open(postgres.Open(dsn), &gorm.Config{})\n\tif err != nil{\n\t\treturn nil, err\n\t}\t\n\tdb.AutoMigrate(&entity.SalesData{})\n\treturn &DataBase{\n\t\thost: host,\n\t\tport: port,\n\t\tuser: user,\n\t\tpass: pass,\n\t\tdbname: dbname,\n\t\tconnection: db,\n\t}, err\n}", "func NewDatabase(path string) *Database {\n\treturn &Database{\n\t\tdbFile: fmt.Sprintf(\"%s%sterraform-provider-ansible.json\", path, string(os.PathSeparator)),\n\t\tgroups: make(map[string]Group),\n\t}\n}", "func (c *Client) CreateDatabase(name string) (*meta.DatabaseInfo, error) {\n\tif db, _ := c.Database(name); db != nil {\n\t\treturn db, nil\n\t}\n\n\tcmd := &internal.CreateDatabaseCommand{\n\t\tName: proto.String(name),\n\t}\n\n\terr := c.retryUntilExec(internal.Command_CreateDatabaseCommand, internal.E_CreateDatabaseCommand_Command, cmd)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn c.Database(name)\n}", "func TestStore_CreateDatabase(t *testing.T) {\n\tt.Parallel()\n\ts := MustOpenStore()\n\tdefer s.Close()\n\n\t// Create database.\n\tif di, err := s.CreateDatabase(\"db0\"); err != nil {\n\t\tt.Fatal(err)\n\t} else if !reflect.DeepEqual(di, &meta.DatabaseInfo{Name: \"db0\"}) {\n\t\tt.Fatalf(\"unexpected database: %#v\", di)\n\t}\n\n\t// Create another database.\n\tif di, err := s.CreateDatabase(\"db1\"); err != nil {\n\t\tt.Fatal(err)\n\t} else if !reflect.DeepEqual(di, &meta.DatabaseInfo{Name: \"db1\"}) {\n\t\tt.Fatalf(\"unexpected database: %#v\", di)\n\t}\n}", "func newRecipeDB() (RecipeDB, error) {\n\tdb := make(RecipeDB)\n\tf, err := os.Open(filepath.Join(getDbDir(), dbFileName))\n\tif err != nil {\n\t\tif os.IsNotExist(err) {\n\t\t\treturn db, nil\n\t\t}\n\t\treturn nil, err\n\t}\n\tdefer f.Close()\n\tbytes, err := ioutil.ReadAll(f)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar recipelist []Recipe\n\tif err := json.Unmarshal(bytes, &recipelist); err != nil {\n\t\treturn nil, err\n\t}\n\tfor _, recipe := range recipelist {\n\t\tdb[recipe.Name] = recipe\n\t}\n\treturn db, nil\n}", "func (p *planner) CreateDatabase(n *parser.CreateDatabase) (planNode, error) {\n\tif n.Name == \"\" {\n\t\treturn nil, errEmptyDatabaseName\n\t}\n\n\tif p.user != security.RootUser {\n\t\treturn nil, fmt.Errorf(\"only %s is allowed to create databases\", security.RootUser)\n\t}\n\n\tdesc := makeDatabaseDesc(n)\n\n\tif err := p.createDescriptor(databaseKey{string(n.Name)}, &desc, n.IfNotExists); err != nil {\n\t\treturn nil, err\n\t}\n\treturn &valuesNode{}, nil\n}", "func NewDB(conf configuration.Ledger, opts *badger.Options) (*DB, error) {\n\topts = setOptions(opts)\n\tdir, err := filepath.Abs(conf.Storage.DataDirectory)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\topts.Dir = dir\n\topts.ValueDir = dir\n\n\tbdb, err := badger.Open(*opts)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"local database open failed\")\n\t}\n\n\tdb := &DB{\n\t\tdb: bdb,\n\t\ttxretiries: conf.Storage.TxRetriesOnConflict,\n\t\tidlocker: NewIDLocker(),\n\t\tnodeHistory: map[core.PulseNumber][]core.Node{},\n\t}\n\treturn db, nil\n}", "func newTestDB(t *testing.T) (*gorm.DB, func(...string) error) {\n\treturn dbutil.NewTestDB(t, AllTables...)\n}", "func NewDB(name string, backend BackendType, dir string) (DB, error) {\n\tdbCreator, ok := backends[backend]\n\tif !ok {\n\t\tkeys := make([]string, 0, len(backends))\n\t\tfor k := range backends {\n\t\t\tkeys = append(keys, string(k))\n\t\t}\n\t\treturn nil, fmt.Errorf(\"unknown db_backend %s, expected one of %v\",\n\t\t\tbackend, strings.Join(keys, \",\"))\n\t}\n\n\tdb, err := dbCreator(name, dir)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"failed to initialize database: %w\", err)\n\t}\n\treturn db, nil\n}", "func NewDataBase() *DataBase {\n\tres := DataBase{}\n\tres.db = make(map[string]*DataTable, initialNumTables)\n\tres.l = sync.RWMutex{}\n\treturn &res\n}", "func New(cfg config.DB) (*gorm.DB, error) {\n\tdb, err := database.New(cfg)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif err := db.AutoMigrate(\n\t\trepository.Client{},\n\t\trepository.Deal{},\n\t\trepository.Position{},\n\t\trepository.OHLCV{},\n\t); err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn db, nil\n}", "func NewDB() *DB {\n\td := new(DB)\n\td.Init()\n\treturn d\n}", "func NewDB(conf configuration.Ledger, opts *badger.Options) (*DB, error) {\n\topts = setOptions(opts)\n\tdir, err := filepath.Abs(conf.Storage.DataDirectory)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\topts.Dir = dir\n\topts.ValueDir = dir\n\n\tbdb, err := badger.Open(*opts)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"local database open failed\")\n\t}\n\n\tdb := &DB{\n\t\tdb: bdb,\n\t\ttxretiries: conf.Storage.TxRetriesOnConflict,\n\t\tidlocker: NewIDLocker(),\n\t}\n\treturn db, nil\n}", "func New(dsn string, maxConn int) dao.DB {\n\treturn &db{\n\t\tDB: open(dsn, maxConn),\n\t}\n}", "func NewDb() *Db {\n\treturn &Db{\n\t\tmake(map[string]*TitleInfo),\n\t\tmake(map[string]*EpisodeInfo),\n\t\tmake(map[string][]string),\n\t}\n}", "func (ctx *DefaultContext) NewDatabase() evm.DB {\n\treturn NewCache(ctx)\n}", "func NewDatabase(\n\to *DbOptions,\n\toa *AppOptions,\n\tl log.Logger,\n) (db.Session, error) {\n\tvar db *sqlx.DB\n\tvar err error\n\ttimeout := 15 * time.Second\n\tdeadline := time.Now().Add(timeout)\n\n\tmode := oa.Mode\n\t// register database driver proxy to log sql expresion\n\tproxyName := fmt.Sprintf(\"%s-proxy\", o.Code)\n\tonce.Do(func() {\n\t\tswitch o.Code {\n\t\tcase \"postgres\":\n\t\t\tsql.Register(proxyName, sqlhooks.Wrap(&pq.Driver{}, &hook{\n\t\t\t\tMode: mode,\n\t\t\t\tLogger: l,\n\t\t\t}))\n\t\tdefault:\n\t\t\tsql.Register(proxyName, sqlhooks.Wrap(&pq.Driver{}, &hook{\n\t\t\t\tMode: mode,\n\t\t\t\tLogger: l,\n\t\t\t}))\n\t\t}\n\t\tl.Info(context.Background(), \"SQL MODE: %s\", mode)\n\t})\n\n\t// connect to database server in 5 seconds\n\tfor {\n\t\tif time.Now().After(deadline) {\n\t\t\treturn nil, fmt.Errorf(\"database did not start up in %s (%v)\", timeout, err)\n\t\t}\n\t\t//db, err := sql.Open(\"postgres-proxy\", o.DSN)\n\t\tdb, err = sqlx.Open(proxyName, o.DSN)\n\t\tl.Debug(context.Background(), \"sqlx is openning...\")\n\t\tif err != nil {\n\t\t\ttime.Sleep(timeout / 5)\n\t\t\tcontinue\n\t\t}\n\t\terr = db.Ping()\n\t\tif err == nil {\n\t\t\tbreak\n\t\t}\n\t\ttime.Sleep(timeout / 5)\n\t}\n\n\tswitch o.Code {\n\tcase \"postgres\":\n\t\treturn bpostgres.New(db, l), nil\n\tdefault:\n\t\treturn bpostgres.New(db, l), nil\n\t}\n}", "func NewDatabase(config settings.PostgresDatabase, logger Logger) (*Database, error) {\n\tconnStr := \"postgres://\" + config.User + \":\" + config.Password +\n\t\t\"@\" + config.Address + \"/\" + config.Address + \"?sslmode=disable&connect_timeout=1\"\n\tdb, err := sql.Open(\"postgres\", connStr)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn &Database{\n\t\tsql: db,\n\t\tlogger: logger,\n\t}, nil\n}", "func NewDB(path string) (*gorm.DB, error) {\n\t// connect to the example sqlite, create if it doesn't exist\n\tdb, err := gorm.Open(\"sqlite3\", path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn db, nil\n}", "func CreateNew(dbFile, feeXPub string) error {\n\tlog.Infof(\"Initializing new database at %s\", dbFile)\n\n\tdb, err := bolt.Open(dbFile, 0600, &bolt.Options{Timeout: 1 * time.Second})\n\tif err != nil {\n\t\treturn fmt.Errorf(\"unable to open db file: %w\", err)\n\t}\n\n\tdefer db.Close()\n\n\t// Create all storage buckets of the VSP if they don't already exist.\n\terr = db.Update(func(tx *bolt.Tx) error {\n\t\t// Create parent bucket.\n\t\tvspBkt, err := tx.CreateBucket(vspBktK)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to create %s bucket: %w\", string(vspBktK), err)\n\t\t}\n\n\t\t// Initialize with initial database version (1).\n\t\terr = vspBkt.Put(versionK, uint32ToBytes(initialVersion))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tlog.Info(\"Generating ed25519 signing key\")\n\n\t\t// Generate ed25519 key\n\t\t_, signKey, err := ed25519.GenerateKey(rand.Reader)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to generate signing key: %w\", err)\n\t\t}\n\t\terr = vspBkt.Put(privateKeyK, signKey.Seed())\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Generate a secret key for initializing the cookie store.\n\t\tlog.Info(\"Generating cookie secret\")\n\t\tsecret := make([]byte, 32)\n\t\t_, err = rand.Read(secret)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\terr = vspBkt.Put(cookieSecretK, secret)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tlog.Info(\"Storing extended public key\")\n\t\t// Store fee xpub\n\t\terr = vspBkt.Put(feeXPubK, []byte(feeXPub))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t// Create ticket bucket.\n\t\t_, err = vspBkt.CreateBucket(ticketBktK)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to create %s bucket: %w\", string(ticketBktK), err)\n\t\t}\n\n\t\t// Create vote change bucket.\n\t\t_, err = vspBkt.CreateBucket(voteChangeBktK)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"failed to create %s bucket: %w\", string(voteChangeBktK), err)\n\t\t}\n\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog.Info(\"Database initialized\")\n\n\treturn nil\n}", "func NewDB(dpType, dbConnectionPath string) *DB {\n\treturn &DB{\n\t\tdbType: dpType,\n\t\tdbConnectionPath: dbConnectionPath,\n\t}\n}" ]
[ "0.7525549", "0.73108315", "0.72980404", "0.7078767", "0.7054522", "0.7045155", "0.7038624", "0.7013832", "0.7004014", "0.69702625", "0.695818", "0.69435066", "0.69435066", "0.6939298", "0.69298834", "0.6927611", "0.68909794", "0.6866389", "0.68505955", "0.6843715", "0.68333143", "0.68198353", "0.6815056", "0.68058205", "0.67987204", "0.678078", "0.6773446", "0.67395633", "0.6731411", "0.67170846", "0.6704581", "0.669639", "0.66920745", "0.6667495", "0.6656151", "0.664917", "0.6634278", "0.66235477", "0.661466", "0.6595808", "0.6595598", "0.65921813", "0.6590296", "0.65783125", "0.6570132", "0.65452355", "0.6538937", "0.6527284", "0.65025455", "0.6500726", "0.64852893", "0.64797765", "0.64791983", "0.6476864", "0.6468973", "0.6461576", "0.64587444", "0.6450832", "0.64463747", "0.643348", "0.64328134", "0.6432396", "0.6403574", "0.6399793", "0.63792837", "0.63756824", "0.6371533", "0.63714445", "0.6370109", "0.6368029", "0.636048", "0.6326417", "0.6310721", "0.62951845", "0.6290407", "0.62733155", "0.62730795", "0.62685657", "0.6266939", "0.62518346", "0.62503767", "0.6247029", "0.6232668", "0.62226236", "0.62223566", "0.6220395", "0.62142557", "0.6208846", "0.62085724", "0.62081826", "0.6205855", "0.61980003", "0.6196621", "0.61808157", "0.61764956", "0.61717033", "0.6168354", "0.6165991", "0.61639965", "0.6161209" ]
0.6359392
71
Evaluate will evaluate the entire policy chain using the provided input. As soon as it is discovered that the input violates one policy, Evaluate will return and not evaluate the rest of the chain.
func (c Chain) Evaluate(input Input) (string, string, bool) { for _, policyFunc := range c { reason, message, violationFound := policyFunc(input) if violationFound { return reason, message, violationFound } } return "", "", false }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (opa *client) EvaluatePolicy(policy string, input []byte) (*EvaluatePolicyResponse, error) {\n\tlog := opa.logger.Named(\"Evalute Policy\")\n\n\trequest, err := json.Marshal(&EvalutePolicyRequest{Input: input})\n\tif err != nil {\n\t\tlog.Error(\"failed to encode OPA input\", zap.Error(err), zap.String(\"input\", string(input)))\n\t\treturn nil, fmt.Errorf(\"failed to encode OPA input: %s\", err)\n\t}\n\n\thttpResponse, err := opa.httpClient.Post(opa.getDataQueryURL(getOpaPackagePath(policy)), \"application/json\", bytes.NewReader(request))\n\tif err != nil {\n\t\tlog.Error(\"http request to OPA failed\", zap.Error(err))\n\t\treturn nil, fmt.Errorf(\"http request to OPA failed: %s\", err)\n\t}\n\tif httpResponse.StatusCode != http.StatusOK {\n\t\tlog.Error(\"http response status from OPA not OK\", zap.Any(\"status\", httpResponse.Status))\n\t\treturn nil, fmt.Errorf(\"http response status not OK\")\n\t}\n\n\tresponse := &EvaluatePolicyResponse{}\n\terr = json.NewDecoder(httpResponse.Body).Decode(&response)\n\tif err != nil {\n\t\tlog.Error(\"failed to decode OPA result\", zap.Error(err))\n\t\treturn nil, fmt.Errorf(\"failed to decode OPA result: %s\", err)\n\t}\n\n\treturn response, nil\n}", "func (p *Policy) Evaluate(c *Context) (allow bool, match bool, err error) {\n\tfor _, statement := range p.Statements {\n\t\tstatementMatch, err := statement.match(c)\n\t\tif err != nil {\n\t\t\tbreak\n\t\t}\n\t\tif statementMatch {\n\t\t\tmatch = true\n\t\t\tif statement.Effect == Deny {\n\t\t\t\tallow = false\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tif statement.Effect == Allow {\n\t\t\t\tallow = true\n\t\t\t}\n\t\t}\n\t}\n\treturn allow, match, err\n}", "func (c *chainACL) Evaluate(signatureSet []*protoutil.SignedData) error {\n\tpolicy, ok := c.policyManager.GetPolicy(policies.ChannelWriters)\n\tif !ok {\n\t\treturn fmt.Errorf(\"could not find policy %s\", policies.ChannelWriters)\n\t}\n\n\terr := policy.EvaluateSignedData(signatureSet)\n\tif err != nil {\n\t\tc.Logger.Debugf(\"SigFilter evaluation failed: %s, policyName: %s\", err.Error(), policies.ChannelWriters)\n\t\treturn errors.Wrap(errors.WithStack(msgprocessor.ErrPermissionDenied), err.Error())\n\t}\n\treturn nil\n}", "func (e PolicyEngine) Process(r v1.Request) (res v1.Result) {\n\terrors := []v1.Error{}\n\trelevantEffects := []v1.Effect{}\n\n\tfor _, policy := range e.Policies {\n\t\tfor _, rule := range policy.Rules {\n\t\t\tif !matchResources(rule.Resources, r.Resources) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif !matchActions(rule.Actions, r.Actions) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif !matchActor(rule.Actors, r.Actor) {\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tconditionsMatched := true\n\t\t\tconditionErrors := []v1.Error{}\n\n\t\t\tfor _, cnd := range rule.Conditions {\n\t\t\t\tok, err := EvalCondition(cnd, r)\n\n\t\t\t\tif !ok {\n\t\t\t\t\tconditionsMatched = false\n\t\t\t\t}\n\n\t\t\t\tif err != nil {\n\t\t\t\t\tconditionErrors = append(conditionErrors, v1.Error{\n\t\t\t\t\t\tError: err,\n\t\t\t\t\t\tCode: \"idt_rule_fail\",\n\t\t\t\t\t\tMessage: \"indent: access(core): engine(EvalCondition): \" + err.Error(),\n\t\t\t\t\t})\n\t\t\t\t}\n\t\t\t}\n\n\t\t\teff := rule.Effect\n\n\t\t\tif conditionsMatched {\n\t\t\t\terrors = append(errors, conditionErrors...)\n\t\t\t} else {\n\t\t\t\teff = getOppositeEffect(eff)\n\t\t\t}\n\n\t\t\trelevantEffects = append(relevantEffects, eff)\n\t\t}\n\t}\n\n\tres.Effect = getMostLimitedEffect(relevantEffects)\n\tres.Errors = errors\n\n\tif len(res.Errors) != 0 {\n\t\tres.Effect = \"block\"\n\t}\n\n\treturn res\n}", "func (this *Self) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn item, nil\n}", "func (net *Network) Evaluate(inputValues []float64) float64 {\n\tinputLength := len(inputValues)\n\tfor i, n := range net.InputNodes {\n\t\tif i < inputLength {\n\t\t\tn.SetValue(inputValues[i])\n\t\t}\n\t}\n\tmaxIterationCounter := net.maxIterations\n\tif maxIterationCounter == 0 {\n\t\t// If max iterations has not been configured, use 100\n\t\tmaxIterationCounter = 100\n\t}\n\tresult, _ := net.OutputNode.evaluate(net.Weight, &maxIterationCounter)\n\treturn result\n}", "func (s *Strategy) Evaluate(pods []v1.Pod) []v1.Pod {\n\tlog.Printf(\"Evaluate called with %d pods\", len(pods))\n\tpods = s.eval(pods)\n\tlog.Printf(\"Evaluate exiting with %d pods\", len(pods))\n\treturn pods\n}", "func (s *candidate) Evaluate() (float64, error) {\n\tr, err := s.Schedule()\n\treturn r.Evaluate(), err\n}", "func (f *functionQuery) Evaluate(t iterator) interface{} {\n\treturn f.Func(f.Input, t)\n}", "func (p *prog) Eval(input any) (v ref.Val, det *EvalDetails, err error) {\n\t// Configure error recovery for unexpected panics during evaluation. Note, the use of named\n\t// return values makes it possible to modify the error response during the recovery\n\t// function.\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tswitch t := r.(type) {\n\t\t\tcase interpreter.EvalCancelledError:\n\t\t\t\terr = t\n\t\t\tdefault:\n\t\t\t\terr = fmt.Errorf(\"internal error: %v\", r)\n\t\t\t}\n\t\t}\n\t}()\n\t// Build a hierarchical activation if there are default vars set.\n\tvar vars interpreter.Activation\n\tswitch v := input.(type) {\n\tcase interpreter.Activation:\n\t\tvars = v\n\tcase map[string]any:\n\t\tvars = activationPool.Setup(v)\n\t\tdefer activationPool.Put(vars)\n\tdefault:\n\t\treturn nil, nil, fmt.Errorf(\"invalid input, wanted Activation or map[string]any, got: (%T)%v\", input, input)\n\t}\n\tif p.defaultVars != nil {\n\t\tvars = interpreter.NewHierarchicalActivation(p.defaultVars, vars)\n\t}\n\tv = p.interpretable.Eval(vars)\n\t// The output of an internal Eval may have a value (`v`) that is a types.Err. This step\n\t// translates the CEL value to a Go error response. This interface does not quite match the\n\t// RPC signature which allows for multiple errors to be returned, but should be sufficient.\n\tif types.IsError(v) {\n\t\terr = v.(*types.Err)\n\t}\n\treturn\n}", "func (this *Element) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.BinaryEval(this, item, context)\n}", "func (p *ApprovalPoll) Evaluate() ([]string, []CScore, error) {\n\tif p.candidates == nil || p.ballots == nil {\n\t\treturn []string{}, []CScore{}, errors.New(\"no candidates or no ballots\")\n\t}\n\twinners, ranks := p.getWinners()\n\treturn winners, ranks, nil\n}", "func (tsf *traceStateFilter) Evaluate(_ context.Context, _ pcommon.TraceID, trace *TraceData) (Decision, error) {\n\ttrace.Lock()\n\tdefer trace.Unlock()\n\tbatches := trace.ReceivedBatches\n\n\treturn hasSpanWithCondition(batches, func(span ptrace.Span) bool {\n\t\ttraceState, err := tracesdk.ParseTraceState(span.TraceState().AsRaw())\n\t\tif err != nil {\n\t\t\treturn false\n\t\t}\n\t\tif ok := tsf.matcher(traceState.Get(tsf.key)); ok {\n\t\t\treturn true\n\t\t}\n\t\treturn false\n\t}), nil\n}", "func (c *Contract) Evaluate(ctx TransactionContextInterface, jeweler string, paperNumber string, evaluator string, evalDateTime string) (*InventoryFinancingPaper, error) {\r\n\tpaper, err := ctx.GetPaperList().GetPaper(jeweler, paperNumber)\r\n\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\r\n\tif paper.GetEvaluator() == \"\" {\r\n\t\tpaper.SetEvaluator(evaluator)\r\n\t}\r\n\r\n\tif paper.IsReceived() {\r\n\t\tpaper.SetEvaluated()\r\n\t}\r\n\r\n\tif !paper.IsEvaluated() {\r\n\t\treturn nil, fmt.Errorf(\"inventory paper %s:%s is not yet evaluated, Current state = %s\", jeweler, paperNumber, paper.GetState())\r\n\t}\r\n\r\n\terr = ctx.GetPaperList().UpdatePaper(paper)\r\n\r\n\tif err != nil {\r\n\t\treturn nil, err\r\n\t}\r\n\tfmt.Printf(\"The evluator %q has evaluated the inventory financing paper %q:%q,the evaluate date is %q..\\n Current State is %q\", paper.GetEvaluator(), jeweler, paperNumber, evalDateTime, paper.GetState())\r\n\treturn paper, nil\r\n}", "func Evaluate(decision *Decision, maximiser bool) {\n\ts := whichStrategy(maximiser)\n\tdecision.Score = s.initial\n\tfor _, r := range decision.Responses {\n\t\tif r.Responses != nil {\n\t\t\t//\n\t\t\t// Evaluate the response from the opposite perspective.\n\t\t\t//\n\t\t\tEvaluate(r, !maximiser)\n\t\t}\n\t\tif s.compare(r.Score, decision.Score) {\n\t\t\tdecision.Score = r.Score\n\t\t}\n\t}\n}", "func (s *System) Evaluate(state []float32) []float32 {\n\tif len(state) > 0 {\n\t\treturn s.function(state, s.parametersVector)\n\t} else {\n\t\treturn s.function(s.stateVector, s.parametersVector)\n\t}\n}", "func (b *BaseIntent) Evaluate(ctx context.Context, evaluateCtx *EvaluateContext) error {\n\treturn nil\n}", "func (e *Evaluator) Evaluate(node ast.Node, env *object.Environment) object.Object {\n\te.Ctxt = node.Context()\n\tswitch node.(type) {\n\tcase *ast.Program:\n\t\tres := &object.StmtResults{}\n\t\tres.Results = []object.Object{}\n\n\t\t// adding statements\n\t\tfor _, stmt := range node.(*ast.Program).Statements {\n\t\t\tif ret, ok := stmt.(*ast.ReturnStatement); ok {\n\t\t\t\treturn e.Evaluate(ret, env)\n\t\t\t}\n\t\t\tresult := e.Evaluate(stmt, env)\n\t\t\tres.Results = append(res.Results, result)\n\t\t}\n\n\t\t// adding functions\n\t\t//todo: this should function differently than closures\n\t\tfor _, fn := range node.(*ast.Program).Functions {\n\t\t\tbody := fn.Body\n\t\t\tparams := fn.Params\n\t\t\tenv.Data[fn.Name.Value] = &object.Function{\n\t\t\t\tParams: params,\n\t\t\t\tBody: body,\n\t\t\t\tEnv: env,\n\t\t\t}\n\t\t}\n\n\t\t//todo: adding classes\n\n\t\treturn res\n\n\tcase ast.Statement:\n\t\tstmt := node.(ast.Statement)\n\n\t\tswitch node.(ast.Statement).(type) {\n\t\tcase *ast.LetStatement:\n\t\t\tletstmt := stmt.(*ast.LetStatement)\n\t\t\tval := e.Evaluate(letstmt.Value, env)\n\t\t\tenv.Set(letstmt.Name.Value, val)\n\t\t\treturn NULL\n\n\t\tcase *ast.ExprStatement:\n\t\t\texpr := stmt.(*ast.ExprStatement)\n\t\t\treturn e.Evaluate(expr.Expression, env)\n\n\t\tcase *ast.ReturnStatement:\n\t\t\tretstmt := stmt.(*ast.ReturnStatement)\n\t\t\tres := e.Evaluate(retstmt.Value, env)\n\t\t\treturn &object.Return{Inner: res}\n\n\t\tcase *ast.WhileStatement:\n\t\t\te.loopcount++\n\t\t\twhilestmt := stmt.(*ast.WhileStatement)\n\n\t\t\tvar result object.Object\n\n\t\t\tfor {\n\t\t\t\tval := e.Evaluate(whilestmt.Condition, env)\n\t\t\t\tif !evaluateTruthiness(val) {\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t\tresult = e.evalBlockStmt(whilestmt.Body, env)\n\t\t\t\tif object.IsErr(result) || object.IsBreak(result) {\n\t\t\t\t\tif object.IsBreak(result) {\n\t\t\t\t\t\te.loopcount--\n\t\t\t\t\t\treturn NULL\n\t\t\t\t\t}\n\t\t\t\t\treturn result\n\t\t\t\t}\n\t\t\t}\n\n\t\t\te.loopcount--\n\t\t\treturn result\n\n\t\tcase *ast.BreakStatement:\n\t\t\tif e.loopcount == 0 {\n\t\t\t\treturn &object.Exception{\n\t\t\t\t\tMsg: \"Cannot use break outside of loop\",\n\t\t\t\t\tCon: node.(ast.Statement).Context(),\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn &object.Break{}\n\n\t\tcase *ast.BlockStatement:\n\t\t\tblkstmt := stmt.(*ast.BlockStatement)\n\t\t\treturn e.evalBlockStmt(blkstmt, env)\n\n\t\tdefault:\n\t\t\treturn NULL\n\t\t}\n\n\tcase ast.Expression:\n\t\texpr := node.(ast.Expression)\n\n\t\tswitch node.(ast.Expression).(type) {\n\t\tcase *ast.Identifier:\n\t\t\tident := expr.(*ast.Identifier)\n\t\t\tif data, ok := env.Get(ident.Value); ok {\n\t\t\t\treturn data\n\t\t\t}\n\t\t\tif bltn, ok := builtins[ident.Value]; ok {\n\t\t\t\treturn bltn\n\t\t\t}\n\t\t\treturn &object.Exception{\n\t\t\t\tMsg: fmt.Sprintf(\"Could not find symbol %s\", ident.Value),\n\t\t\t\tCon: ident.Context(),\n\t\t\t}\n\n\t\tcase *ast.PrefixExpr:\n\t\t\tpexpr := expr.(*ast.PrefixExpr)\n\t\t\treturn e.evalPrefixExpr(pexpr, env)\n\n\t\tcase *ast.InfixExpr:\n\t\t\tiexpr := expr.(*ast.InfixExpr)\n\t\t\treturn e.evalInfixExpr(iexpr, env)\n\n\t\tcase *ast.IfExpression:\n\t\t\tifexpr := expr.(*ast.IfExpression)\n\t\t\tcondition := e.Evaluate(ifexpr.Condition, env)\n\t\t\tif condition == nil {\n\t\t\t\treturn &object.Exception{\n\t\t\t\t\tMsg: \"If condition returned nil\",\n\t\t\t\t\tCon: ifexpr.Context(),\n\t\t\t\t}\n\t\t\t}\n\t\t\tif evaluateTruthiness(condition) {\n\t\t\t\treturn e.Evaluate(ifexpr.Result, env)\n\t\t\t}\n\t\t\tif ifexpr.Alternative != nil {\n\t\t\t\tswitch ifexpr.Alternative.(type) {\n\t\t\t\tcase *ast.BlockStatement:\n\t\t\t\t\treturn e.Evaluate(ifexpr.Alternative.(*ast.BlockStatement), env)\n\t\t\t\tcase *ast.IfExpression:\n\t\t\t\t\treturn e.Evaluate(ifexpr.Alternative.(*ast.IfExpression), env)\n\t\t\t\tdefault:\n\t\t\t\t\treturn &object.Exception{\n\t\t\t\t\t\tMsg: \"Invalid else branch\",\n\t\t\t\t\t\tCon: ifexpr.Alternative.Context(),\n\t\t\t\t\t}\n\t\t\t\t}\n\t\t\t}\n\n\t\tcase *ast.FnLiteral:\n\t\t\tfnlit := expr.(*ast.FnLiteral)\n\t\t\tparams := fnlit.Params\n\t\t\tbody := fnlit.Body\n\t\t\treturn &object.Function{Params: params, Env: env, Body: body}\n\n\t\tcase *ast.FunctionCall:\n\t\t\t// asserting type\n\t\t\tfncall := expr.(*ast.FunctionCall)\n\n\t\t\t// resolving to object\n\t\t\tfunction := e.Evaluate(fncall.Ident, env)\n\t\t\tif object.IsErr(function) {\n\t\t\t\treturn function\n\t\t\t}\n\n\t\t\targs := e.evalExpressions(fncall.Params, env)\n\t\t\tif len(args) == 1 && object.IsErr(args[0]) {\n\t\t\t\treturn args[0]\n\t\t\t}\n\n\t\t\treturn e.applyFunction(function, args)\n\n\t\tcase *ast.DotExpression:\n\t\t\t//todo\n\t\t\treturn &object.Exception{\n\t\t\t\tMsg: \"DotExpr: unimplemented\",\n\t\t\t\tCon: node.Context(),\n\t\t\t}\n\n\t\tcase *ast.Int:\n\t\t\tintexpr := node.(ast.Expression).(*ast.Int)\n\t\t\treturn &object.Integer{Value: intexpr.Inner}\n\t\tcase *ast.Flt:\n\t\t\tfltexpr := node.(ast.Expression).(*ast.Flt)\n\t\t\treturn &object.Float{Value: fltexpr.Inner}\n\t\tcase *ast.Str:\n\t\t\tstrexpr := node.(ast.Expression).(*ast.Str)\n\t\t\treturn &object.String{Value: strexpr.Inner}\n\t\tcase *ast.Bool:\n\t\t\tboolexpr := node.(ast.Expression).(*ast.Bool)\n\t\t\treturn nativeBooltoObj(boolexpr.Inner)\n\t\tcase *ast.Array:\n\t\t\tarray := node.(ast.Expression).(*ast.Array)\n\t\t\tarr := &object.Array{}\n\n\t\t\t// preallocating so we don't have to waste cycles\n\t\t\t// reallocating every time we append\n\t\t\telements := make([]object.Object, 0, len(array.Elements))\n\n\t\t\tfor _, elem := range array.Elements {\n\t\t\t\telements = append(elements, e.Evaluate(elem, env))\n\t\t\t}\n\t\t\tarr.Elements = elements\n\n\t\t\treturn arr\n\n\t\tcase *ast.Map:\n\t\t\thash := node.(ast.Expression).(*ast.Map)\n\t\t\tnewmap := &object.Map{}\n\t\t\tnewmap.Elements = make(map[object.HashKey]object.Object)\n\n\t\t\tfor key, val := range hash.Elements {\n\t\t\t\tnkey, nval := e.Evaluate(key, env), e.Evaluate(val, env)\n\n\t\t\t\tif object.IsErr(nkey) {\n\t\t\t\t\treturn nkey\n\t\t\t\t}\n\t\t\t\tif object.IsErr(nval) {\n\t\t\t\t\treturn nval\n\t\t\t\t}\n\n\t\t\t\thashable, ok := nkey.(object.Hashable)\n\n\t\t\t\tif !ok {\n\t\t\t\t\treturn &object.Exception{\n\t\t\t\t\t\tMsg: fmt.Sprintf(\"Cannot use type %T as key for Map\", nkey),\n\t\t\t\t\t\tCon: hash.Context(),\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tnewmap.Elements[hashable.HashKey()] = nval\n\t\t\t}\n\n\t\t\treturn newmap\n\n\t\tcase *ast.IndexExpr:\n\t\t\tidx := node.(ast.Expression).(*ast.IndexExpr)\n\t\t\treturn e.evalIndexExpr(idx, env)\n\n\t\tdefault:\n\t\t\treturn NULL\n\t\t}\n\tdefault:\n\t\treturn &object.Exception{\n\t\t\tMsg: \"Unimplemented type\",\n\t\t\tCon: node.Context(),\n\t\t}\n\t}\n\treturn &object.Exception{\n\t\tMsg: fmt.Sprintf(\"Evaluate: unreachable code, got %T\", node),\n\t\tCon: node.Context(),\n\t}\n}", "func Evaluate(thing interface{}, env Environment) (error, Value, Environment) {\n\tswitch thing.(type) {\n\tcase Value:\n\t\treturn EvaluateValue(thing.(Value), env)\n\tcase SExpression:\n\t\tsexp := thing.(SExpression)\n\t\tif isSpecialForm(sexp.FormName.Contained) {\n\t\t\treturn EvaluateSpecialForm(sexp, env)\n\t\t} else {\n\t\t\treturn EvaluateSexp(thing.(SExpression), env)\n\t\t}\n\tdefault:\n\t\treturn errors.New(fmt.Sprintf(\"No way to evaluate %v\\n\", thing)), Value{}, env\n\t}\n}", "func (NullEvaluator) Evaluate(ctx context.Context, input *EvaluationInput) (*EvaluationOutput, error) {\n\treturn &EvaluationOutput{\n\t\tTraits: input.Traits,\n\t}, nil\n}", "func (a Application) Evaluate() Expression {\n\tvar f = a.Function.Evaluate()\n\tif l, ok := f.(Abstraction); ok {\n\t\treturn l.Body.Substitute(l.Argument, a.Argument).Evaluate()\n\t}\n\treturn Application{f, a.Argument}\n}", "func (m *Model) Evaluate(seq Sequence) {\n\tm.EvaluateAt(seq, 0)\n}", "func (c *Context) Evaluate(language, code string, stdins []string) ([]string, Message) {\n\tstdinGlob := glob(stdins)\n\tresults, msg := c.run(language, code, stdinGlob)\n\n\treturn unglob(results), msg\n}", "func (this *ObjectValues) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.UnaryEval(this, item, context)\n}", "func (this *ObjectValues) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.UnaryEval(this, item, context)\n}", "func Evaluate(item interface{}, passedContext interface{}) map[string]float64 {\n\t//fmt.Fprintf(os.Stderr, \"eval:: %v %T\\n\", item, item)\n\n\tif item != nil {\n\t\tswitch passedContext.(type) {\n\t\tcase *DimContext:\n\t\t\t{\n\t\t\t\t//fmt.Fprintf(os.Stderr, \"here before processMember %v\\n\", item)\n\t\t\t\tprocessMember(item, passedContext)\n\t\t\t\t//fmt.Fprintf(os.Stderr, \"here after processMember %v %T\\n\", item, item)\n\t\t\t}\n\t\t}\n\t\tswitch v := item.(type) {\n\t\tcase hasResults:\n\t\t\t{\n\t\t\t\treturn v.Results()\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (r RuleSet) Evaluate(packet Packet) (*Rule, []Rule, error) {\n\tmatches := []Rule{}\n\n\tfor _, rule := range r {\n\t\tif rule.Matches(packet) {\n\t\t\tmatches = append(matches, rule)\n\n\t\t\tif rule.Quick {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t}\n\n\tif len(matches) < 1 {\n\t\treturn nil, []Rule{}, nil\n\t}\n\n\treturn &matches[len(matches)-1], matches, nil\n}", "func (a Abstraction) Evaluate() Expression {\n\treturn Abstraction{a.Argument, a.Body.Evaluate()}\n}", "func (a *AlwaysReturn) Evaluate(r table.Row) (bool, error) {\n\treturn a.V, nil\n}", "func (ap *AttestationPolicy) EvaluateResults(raw *checker.RawResults) (PolicyResult, error) {\n\tlogger := sclog.NewLogger(sclog.DefaultLevel)\n\tif ap.PreventBinaryArtifacts {\n\t\tcheckResult, err := CheckPreventBinaryArtifacts(ap.AllowedBinaryArtifacts, raw, logger)\n\t\tif !checkResult || err != nil {\n\t\t\treturn checkResult, err\n\t\t}\n\t}\n\n\tif ap.PreventUnpinnedDependencies {\n\t\tcheckResult, err := CheckNoUnpinnedDependencies(ap.AllowedUnpinnedDependencies, raw, logger)\n\t\tif !checkResult || err != nil {\n\t\t\treturn checkResult, err\n\t\t}\n\t}\n\n\tif ap.PreventKnownVulnerabilities {\n\t\tcheckResult, err := CheckNoVulnerabilities(raw, logger)\n\t\tif !checkResult || err != nil {\n\t\t\treturn checkResult, err\n\t\t}\n\t}\n\n\tif ap.EnsureCodeReviewed {\n\t\t// By default, if code review reqs. aren't specified, we assume\n\t\t// the user wants there to be atleast one reviewer\n\t\tif len(ap.CodeReviewRequirements.RequiredApprovers) == 0 &&\n\t\t\tap.CodeReviewRequirements.MinReviewers == 0 {\n\t\t\tap.CodeReviewRequirements.MinReviewers = 1\n\t\t}\n\n\t\tcheckResult, err := CheckCodeReviewed(ap.CodeReviewRequirements, raw, logger)\n\t\tif !checkResult || err != nil {\n\t\t\treturn checkResult, err\n\t\t}\n\t}\n\n\treturn Pass, nil\n}", "func (this *ObjectInnerValues) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.UnaryEval(this, item, context)\n}", "func (state *KuhnGameState) Evaluate() float32 {\n\tcurrentActor := state.playerActor(state.CurrentActor().GetID())\n\tcurrentActorOpponent := state.playerActor(-state.CurrentActor().GetID())\n\tif state.causingAction.Name() == acting.Fold {\n\t\tcurrentActor.UpdateStack(currentActor.Stack + state.table.Pot)\n\t\treturn float32(currentActor.GetID()) * (state.table.Pot / 2)\n\t}\n\tcurrentActorHandValue := currentActor.EvaluateHand(state.table)\n\tcurrentActorOpponentHandValue := currentActorOpponent.EvaluateHand(state.table)\n\tif currentActorHandValue > currentActorOpponentHandValue {\n\t\tcurrentActor.UpdateStack(currentActor.Stack + state.table.Pot)\n\t\treturn float32(currentActor.GetID()) * (state.table.Pot / 2)\n\t} else {\n\t\tcurrentActorOpponent.UpdateStack(currentActorOpponent.Stack + state.table.Pot)\n\t\treturn float32(currentActorOpponent.GetID()) * (state.table.Pot / 2)\n\t}\n}", "func (this *Mod) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.BinaryEval(this, item, context)\n}", "func (v Variable) Evaluate() Expression {\n\treturn v\n}", "func (i *InOp) Evaluate(left, right EvalResult) (EvalResult, error) {\n\tpanic(\"implement me\")\n}", "func Evaluate(cards []Card) Hand {\n\tvar hands []Hand\n\tfor _, f := range evaluators {\n\t\thands = append(hands, f(cards))\n\t}\n\n\tvar best Hand\n\tfor _, h := range hands {\n\t\tif h > best {\n\t\t\tbest = h\n\t\t}\n\t}\n\n\treturn best\n}", "func (r *Resolver) Evaluate(args struct{ Expr string }) (Result, error) {\n\tvar result Result\n\tamount, err := calc.CalculateAmount(args.Expr)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\n\tunitName := amount.Units\n\tunit, err := NewUnit(unitName)\n\tif err != nil {\n\t\treturn result, err\n\t}\n\n\tresult = Result{amount.Value, unit, args.Expr}\n\tlog.Info(fmt.Sprintf(\"evaluate(%s) = %.2f %s\", args.Expr, result.value, result.units.pluralName))\n\treturn result, nil\n}", "func (a *AST) Evaluate(table ...map[string]ContextVar) *Expression {\n\tt := make(map[string]ContextVar)\n\tif len(table) > 0 && table[0] != nil {\n\t\tt = table[0]\n\t}\n\treturn a.Root.Evaluate(t)\n}", "func (this *ObjectPairs) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.UnaryEval(this, item, context)\n}", "func (this *ObjectPairs) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.UnaryEval(this, item, context)\n}", "func (gen *progGen) Eval(input any) (ref.Val, *EvalDetails, error) {\n\t// The factory based Eval() differs from the standard evaluation model in that it generates a\n\t// new EvalState instance for each call to ensure that unique evaluations yield unique stateful\n\t// results.\n\tstate := interpreter.NewEvalState()\n\tcostTracker := &interpreter.CostTracker{}\n\tdet := &EvalDetails{state: state, costTracker: costTracker}\n\n\t// Generate a new instance of the interpretable using the factory configured during the call to\n\t// newProgram(). It is incredibly unlikely that the factory call will generate an error given\n\t// the factory test performed within the Program() call.\n\tp, err := gen.factory(state, costTracker)\n\tif err != nil {\n\t\treturn nil, det, err\n\t}\n\n\t// Evaluate the input, returning the result and the 'state' within EvalDetails.\n\tv, _, err := p.Eval(input)\n\tif err != nil {\n\t\treturn v, det, err\n\t}\n\treturn v, det, nil\n}", "func (this *ObjectUnwrap) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.UnaryEval(this, item, context)\n}", "func (dn DoNothing) Evaluate(environment map[string]expr.Expression) map[string]expr.Expression {\n\treturn environment\n}", "func (u *comboUtility) Evaluate() float64 {\n\t// If the utility was reset, reevaluate it\n\tif u.reset {\n\t\tu.output = math.Min(1, math.Max(0, u.combinator(u.srcA.Evaluate(), u.srcB.Evaluate())))\n\t\tu.reset = false\n\t}\n\t// Return the currently set value\n\treturn u.output\n}", "func (l *LikeOp) Evaluate(left, right EvalResult) (EvalResult, error) {\n\tpanic(\"implement me\")\n}", "func (d *Division) Evaluate(left, right EvalResult) (EvalResult, error) {\n\treturn divideNumericWithError(left, right)\n}", "func (m *Manager) Evaluate(t *register.Transaction) (bool, error) {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\tvar matched []*Rule\n\tfor _, r := range m.rules {\n\t\tif r.Evaluate(t) {\n\t\t\tmatched = append(matched, r)\n\t\t}\n\t}\n\tif len(matched) == 0 {\n\t\treturn false, nil\n\t}\n\tif len(matched) > 1 {\n\t\tvar rules []string\n\t\tfor _, r := range matched {\n\t\t\trules = append(rules, r.Name)\n\t\t}\n\t\treturn false, fmt.Errorf(\"transaction %s matched multiple rules: %s\", t.ID, strings.Join(rules, \", \"))\n\t}\n\tif len(t.Category) > 0 {\n\t\treturn false, nil\n\t}\n\tt.Category = append(t.Category, &register.Category{Name: matched[0].Category, Amount: t.Amount})\n\treturn true, nil\n}", "func (policy QueryPolicy) EvalQuery(ctx context.Context, q ent.Query) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalQuery(ctx, q); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (policy QueryPolicy) EvalQuery(ctx context.Context, q ent.Query) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalQuery(ctx, q); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (policy QueryPolicy) EvalQuery(ctx context.Context, q ent.Query) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalQuery(ctx, q); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (policy QueryPolicy) EvalQuery(ctx context.Context, q ent.Query) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalQuery(ctx, q); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (this *ObjectInnerPairs) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.UnaryEval(this, item, context)\n}", "func (source DepthInfo) Evaluate(amount float64) Evaluation {\n\treturn Evaluation{Buy: sumDepth(amount, source.Asks),\n\t\tSell: sumDepth(amount, source.Bids)}\n}", "func Evaluate(ctx context.Context, proposals []*pb.MatchObject) ([]string, error) {\n\t// Map of approved and overloaded proposals. Using maps for easier lookup.\n\tapprovedProposals := map[string]bool{}\n\toverloadedProposals := map[string]bool{}\n\n\t// Map of all the players encountered in the proposals. Each entry maps a player id to\n\t// the first match in which the player was encountered.\n\tallPlayers := map[string]string{}\n\n\t// Iterate over each proposal to either add to approved map or overloaded map.\n\tfor _, proposal := range proposals {\n\t\tproposalID := proposal.Id\n\t\tapproved := true\n\t\tplayers := getPlayersInProposal(proposal)\n\t\t// Iterate over each player in the proposal to check if the player was encountered before.\n\t\tfor _, playerID := range players {\n\t\t\tif propID, found := allPlayers[playerID]; found {\n\t\t\t\t// Player was encountered in an earlier proposal. Mark the current proposal as overloaded (not approved).\n\t\t\t\t// Also, the first proposal where the player was encountered may have been marked approved. Remove that proposal\n\t\t\t\t// approved proposals and add to overloaded proposals since we encountered its player in current proposal too.\n\t\t\t\tapproved = false\n\t\t\t\tdelete(approvedProposals, propID)\n\t\t\t\toverloadedProposals[propID] = true\n\t\t\t} else {\n\t\t\t\t// Player encountered for the first time, add to all players map with the current proposal.\n\t\t\t\tallPlayers[playerID] = proposalID\n\t\t\t}\n\n\t\t\tif approved {\n\t\t\t\tapprovedProposals[proposalID] = true\n\t\t\t} else {\n\t\t\t\toverloadedProposals[proposalID] = true\n\t\t\t}\n\t\t}\n\t}\n\n\t// Convert the maps to lists of overloaded, approved proposals.\n\toverloadedList := []string{}\n\tapprovedList := []string{}\n\tfor k := range overloadedProposals {\n\t\toverloadedList = append(overloadedList, k)\n\t}\n\n\tfor k := range approvedProposals {\n\t\tapprovedList = append(approvedList, k)\n\t}\n\n\t// Select proposals to approve from the overloaded proposals list.\n\tchosen, err := chooseProposals(overloadedList)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"Failed to select approved list from overloaded proposals, %v\", err)\n\t}\n\n\t// Add the chosen proposals to the approved list.\n\tapprovedList = append(approvedList, chosen...)\n\treturn approvedList, nil\n}", "func (_if If) Evaluate(environment map[string]expr.Expression) map[string]expr.Expression {\n\tswitch c := _if.Condition.Evaluate(environment); c {\n\tcase expr.Boolean{Value: true}:\n\t\treturn _if.Consequence.Evaluate(environment)\n\tcase expr.Boolean{Value: false}:\n\t\treturn _if.Alternative.Evaluate(environment)\n\tdefault:\n\t\treturn environment\n\t}\n}", "func Evaluate(expression string) (float64, error) {\n\ttree, err := parse(expression)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn tree.evaluate()\n}", "func Eval(input string, context map[string]interface{}) float64 {\n\tnode, err := Parse(input)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\texpr := &expression{node, context}\n\treturn expr.eval(expr.ast)\n}", "func (r RuleDefinition) Evaluate(e UEvent) bool {\n\t// Compile if needed\n\tif r.rule == nil {\n\t\tif err := r.Compile(); err != nil {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn r.EvaluateAction(e.Action) && r.EvaluateEnv(e.Env)\n}", "func (a *Addition) Evaluate(left, right EvalResult) (EvalResult, error) {\n\treturn addNumericWithError(left, right)\n}", "func (p *Pool) Evaluate() {\n\tvar wg sync.WaitGroup\n\tfor i := 0; i < POOL_SIZE; i++ {\n\t\twg.Add(1)\n\t\tgo func(i int, track []tracks.Element) {\n\t\t\tp.Members[i].Score, p.Members[i].ScoreData = GetScore(track)\n\t\t\twg.Done()\n\t\t}(i, p.Members[i].Track)\n\t}\n\twg.Wait()\n\n\t// Assign fitness for every member. In the future, consider a smarter\n\t// algorithm higher score members a better chance of reproducing.\n\tfor i := 0; i < POOL_SIZE; i++ {\n\t\tp.Members[i].Fitness = float64(p.Members[i].Score)\n\t}\n}", "func (p Print) Evaluate(vars map[string]interface{}, ctx interface{}, funcs FunctionMap, quotes []string) (map[string]interface{}, interface{}, error) {\n\tvars, v, err := p.Node.Evaluate(vars, ctx, funcs, quotes)\n\tif err != nil {\n\t\treturn vars, ctx, err\n\t}\n\tfmt.Println(TryFormatLiteral(v, quotes, false, 0))\n\treturn vars, ctx, nil\n}", "func Evaluate(e ast.Node, genCtx *GenCtx) parser_driver.ValueExpr {\n\tswitch t := e.(type) {\n\tcase *ast.ParenthesesExpr:\n\t\treturn Evaluate(t.Expr, genCtx)\n\tcase *ast.BinaryOperationExpr:\n\t\tres, err := operator.BinaryOps.Eval(t.Op.String(), Evaluate(t.L, genCtx), Evaluate(t.R, genCtx))\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"error occurred on eval: %+v\", err))\n\t\t}\n\t\treturn res\n\tcase *ast.UnaryOperationExpr:\n\t\tres, err := operator.UnaryOps.Eval(t.Op.String(), Evaluate(t.V, genCtx))\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"error occurred on eval: %+v\", err))\n\t\t}\n\t\treturn res\n\tcase *ast.IsNullExpr:\n\t\tsubResult := Evaluate(t.Expr, genCtx)\n\t\tc := ConvertToBoolOrNull(subResult)\n\t\tr := parser_driver.ValueExpr{}\n\t\tr.SetInt64(0)\n\t\tif c == -1 {\n\t\t\tr.SetInt64(1)\n\t\t}\n\t\treturn r\n\tcase *ast.ColumnNameExpr:\n\t\tfor key, value := range genCtx.unwrapPivotRows {\n\t\t\toriginTableName := t.Name.Table.L\n\t\t\tfor k, v := range genCtx.TableAlias {\n\t\t\t\tif v == originTableName {\n\t\t\t\t\toriginTableName = k\n\t\t\t\t\tbreak\n\t\t\t\t}\n\t\t\t}\n\t\t\toriginColumnName := t.Name.Name.L\n\t\t\tif key == fmt.Sprintf(\"%s.%s\", originTableName, originColumnName) {\n\t\t\t\tv := parser_driver.ValueExpr{}\n\t\t\t\tv.SetValue(value)\n\t\t\t\tif tmpTable, ok := genCtx.TableAlias[t.Name.Table.L]; ok {\n\t\t\t\t\tt.Name.Table = model.NewCIStr(tmpTable)\n\t\t\t\t}\n\t\t\t\treturn v\n\t\t\t}\n\t\t}\n\t\tpanic(fmt.Sprintf(\"no such col %s in table %s\", t.Name, t.Name.Table))\n\tcase ast.ValueExpr:\n\t\tv := parser_driver.ValueExpr{}\n\t\tv.SetValue(t.GetValue())\n\t\tv.SetType(t.GetType())\n\t\treturn v\n\t}\n\n\t// is useless?\n\t// if e == nil {\n\t// \treturn trueValueExpr()\n\t// }\n\n\tpanic(\"not reachable\")\n\tv := parser_driver.ValueExpr{}\n\tv.SetNull()\n\treturn v\n}", "func (this *ObjectPut) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn this.TernaryEval(this, item, context)\n}", "func Evaluate(tpl string, data interface{}) (string, error) {\n\tt, err := Parse(tpl)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn Execute(t, data)\n}", "func Evaluate(c a.Context, s a.Sequence) a.Sequence {\n\treturn a.Map(c, compiler.Compile(c, s), evaluator)\n}", "func (policies Policies) EvalQuery(ctx context.Context, q ent.Query) error {\n\treturn policies.eval(ctx, func(policy ent.Policy) error {\n\t\treturn policy.EvalQuery(ctx, q)\n\t})\n}", "func (policies QueryPolicy) EvalQuery(ctx context.Context, q ent.Query) error {\n\tfor _, policy := range policies {\n\t\tswitch decision := policy.EvalQuery(ctx, q); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (session Runtime) Evaluate(code string, async bool, returnByValue bool) (interface{}, error) {\n\tresult, err := session.evaluate(code, session.currentContext(), async, returnByValue)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn result.Value, nil\n}", "func (f *CallExpression) Evaluate(ctx *Context) Value {\n\tcallable := f.Callable.Evaluate(ctx)\n\n\tif callable.Type == vtVariable {\n\t\tcallable = callable.Evaluate(ctx)\n\t}\n\n\tif callable.isCallable() {\n\t\tnewCtx := NewContext(\"\", nil)\n\t\targs := f.Args.EvaluateAll(ctx)\n\t\treturn callable.callable().Execute(newCtx, &args)\n\t}\n\n\tpanic(NewNotCallableError(callable))\n}", "func (expr *Expr) Evaluate(root NodeNavigator) interface{} {\n\tval := expr.q.Evaluate(iteratorFunc(func() NodeNavigator { return root }))\n\tswitch val.(type) {\n\tcase query:\n\t\treturn &NodeIterator{query: expr.q.Clone(), node: root}\n\t}\n\treturn val\n}", "func (s *Script) evaluate(tx *types.Transaction, txInIdx int) error {\n\tscript := *s\n\tscriptLen := len(script)\n\t// logger.Debugf(\"script len %d: %s\", scriptLen, s.Disasm())\n\n\tstack := newStack()\n\tfor pc, scriptPubKeyStart := 0, 0; pc < scriptLen; {\n\t\topCode, operand, newPc, err := s.parseNextOp(pc)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tpc = newPc\n\n\t\tif err := s.execOp(opCode, operand, tx, txInIdx, pc, &scriptPubKeyStart, stack); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Succeed if top stack item is true\n\treturn stack.validateTop()\n}", "func (policy Policy) Validate(bucketName string) error {\n\tif err := policy.isValid(); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, statement := range policy.Statements {\n\t\tif err := statement.Validate(bucketName); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func Evaluate(input string) (decimal.Decimal, error) {\n\tvar stack []decimal.Decimal\n\tinputs := strings.Split(input, \" \")\n\n\tfor _, command := range inputs {\n\t\tswitch command {\n\t\tcase \"+\", \"-\", \"*\", \"/\", \"%\", \"^\":\n\t\t\tif len(stack) < 2 {\n\t\t\t\treturn decimal.Zero, errors.New(\"stack overflow\")\n\t\t\t}\n\t\t\tlhs := stack[len(stack)-2]\n\t\t\trhs := stack[len(stack)-1]\n\t\t\tstack = stack[:len(stack)-1]\n\t\t\tswitch command {\n\t\t\tcase \"+\":\n\t\t\t\trhs = lhs.Add(rhs)\n\t\t\tcase \"-\":\n\t\t\t\trhs = lhs.Sub(rhs)\n\t\t\tcase \"*\":\n\t\t\t\trhs = lhs.Mul(rhs)\n\t\t\tcase \"/\":\n\t\t\t\trhs = lhs.Div(rhs)\n\t\t\tcase \"%\":\n\t\t\t\trhs = lhs.Mod(rhs)\n\t\t\tcase \"^\":\n\t\t\t\trhs = lhs.Pow(rhs)\n\t\t\t}\n\t\t\tstack[len(stack)-1] = rhs\n\t\tcase \"abs\", \"atan\", \"ceil\", \"cos\", \"floor\", \"neg\", \"sin\", \"tan\":\n\t\t\tif len(stack) < 1 {\n\t\t\t\treturn decimal.Zero, errors.New(\"stack overflow\")\n\t\t\t}\n\t\t\tval := stack[len(stack)-1]\n\t\t\tswitch command {\n\t\t\tcase \"abs\":\n\t\t\t\tval = val.Abs()\n\t\t\tcase \"atan\":\n\t\t\t\tval = val.Atan()\n\t\t\tcase \"ceil\":\n\t\t\t\tval = val.Ceil()\n\t\t\tcase \"cos\":\n\t\t\t\tval = val.Cos()\n\t\t\tcase \"floor\":\n\t\t\t\tval = val.Floor()\n\t\t\tcase \"neg\":\n\t\t\t\tval = val.Neg()\n\t\t\tcase \"sin\":\n\t\t\t\tval = val.Sin()\n\t\t\tcase \"tan\":\n\t\t\t\tval = val.Tan()\n\t\t\t}\n\t\t\tstack[len(stack)-1] = val\n\t\tdefault:\n\t\t\tval, err := decimal.NewFromString(command)\n\t\t\tif err != nil {\n\t\t\t\treturn val, err\n\t\t\t}\n\t\t\tstack = append(stack, val)\n\t\t}\n\t}\n\n\tif len(stack) != 1 {\n\t\treturn decimal.Zero, errors.New(\"unclean stack\")\n\t}\n\treturn stack[0], nil\n}", "func Eval(input string, env interface{}) (interface{}, error) {\n\tif _, ok := env.(Option); ok {\n\t\treturn nil, fmt.Errorf(\"misused expr.Eval: second argument (env) should be passed without expr.Env\")\n\t}\n\n\ttree, err := parser.Parse(input)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tprogram, err := compiler.Compile(tree, nil)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\toutput, err := vm.Run(program, env)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn output, nil\n}", "func (e *AccessorExpr) Evaluate(engine *Engine, input interface{}, args []*Statement) (interface{}, error) {\n\tin := reflect.ValueOf(input)\n\taccessor := e.Query[1:]\n\n\tif input == nil {\n\t\treturn nil, nil\n\t}\n\n\t// If it is a slice we need to Evaluate each one.\n\tif in.Kind() == reflect.Slice {\n\t\tt := TypeOfSliceElement(input)\n\t\tif t.Kind() == reflect.Ptr {\n\t\t\tt = t.Elem()\n\t\t}\n\t\treturnType := e.getReturnType(accessor, reflect.New(t).Interface())\n\n\t\tresults := reflect.MakeSlice(reflect.SliceOf(returnType), 0, 0)\n\n\t\tfor i := 0; i < in.Len(); i++ {\n\t\t\tresult, err := e.Evaluate(engine, in.Index(i).Interface(), nil)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\n\t\t\tresults = reflect.Append(results, reflect.ValueOf(result))\n\t\t}\n\n\t\treturn results.Interface(), nil\n\t}\n\n\tvar err error\n\tinput, err = e.evaluateAccessor(accessor, input)\n\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn input, nil\n}", "func Evaluate(expression *[]string, dispatchTable DispatchTable, stack *Stack) interface{} {\n\n\tfor idx, token := range *expression {\n\t\tvar dispatchFunction DispatchFunc\n\n\t\tif _, err := strconv.ParseFloat(token, 64); err == nil {\n\t\t\tdispatchFunction = dispatchTable[\"FLOAT\"]\n\t\t} else {\n\t\t\tvar evalsOk bool\n\t\t\tif dispatchFunction, evalsOk = dispatchTable[token]; !evalsOk {\n\t\t\t\tdispatchFunction = dispatchTable[\"__DEFAULT__\"]\n\t\t\t\t// delete token from expression\n\t\t\t\tcopy((*expression)[idx:], (*expression)[idx+1:])\n\t\t\t\t(*expression)[len(*expression)-1] = \"\"\n\t\t\t\t(*expression) = (*expression)[:len(*expression)-1]\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tdispatchFunction(token, stack)\n\t}\n\treturn stack.Pop()\n}", "func (r *RegexpOp) Evaluate(left, right EvalResult) (EvalResult, error) {\n\tpanic(\"implement me\")\n}", "func Evaluate(ctx context.Context, c client.Client, q qv1.QuayRegistry) ([]qv1.Condition, error) {\n\tvar conds []qv1.Condition\n\n\t// start by analysing the components that have no dependencies. we append their conditions\n\t// to the conditions slice we are going to return at the end and move on. the health of\n\t// any of these components don't affect other components health.\n\tfor _, component := range []Checker{\n\t\t&HPA{Client: c},\n\t\t&Route{Client: c},\n\t\t&Monitoring{Client: c},\n\t} {\n\t\tcond, err := component.Check(ctx, q)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\t\tconds = append(conds, cond)\n\t}\n\n\t// now analyse the components that the quay component depends on. if any of these is in\n\t// a faulty state then Quay won't be able to come up properly. we gather the name of\n\t// any faulty component in a slice of strings, all conditions are append to the slice\n\t// we return at the end of the process.\n\tvar failed []string\n\tfor _, component := range []Checker{\n\t\t&Postgres{Client: c},\n\t\t&ObjectStorage{Client: c},\n\t\t&Clair{Client: c},\n\t\t&ClairPostgres{Client: c},\n\t\t&TLS{Client: c},\n\t\t&Redis{Client: c},\n\t} {\n\t\tcond, err := component.Check(ctx, q)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t\tconds = append(conds, cond)\n\t\tif cond.Status != metav1.ConditionTrue {\n\t\t\tfailed = append(failed, component.Name())\n\t\t}\n\t}\n\n\t// if we found out any component in a faulty state we have to abort now. Quay component\n\t// must indicate which component is in a faulty state. as mirror component depends on\n\t// Quay component its status is also defined as faulty.\n\tif len(failed) > 0 {\n\t\tconds = append(\n\t\t\tconds,\n\t\t\tqv1.Condition{\n\t\t\t\tType: qv1.ComponentQuayReady,\n\t\t\t\tStatus: metav1.ConditionFalse,\n\t\t\t\tReason: qv1.ConditionReasonComponentNotReady,\n\t\t\t\tLastUpdateTime: metav1.NewTime(time.Now()),\n\t\t\t\tMessage: fmt.Sprintf(\n\t\t\t\t\t\"Awaiting for component %s to become available\",\n\t\t\t\t\tstrings.Join(failed, \",\"),\n\t\t\t\t),\n\t\t\t},\n\t\t\tqv1.Condition{\n\t\t\t\tType: qv1.ComponentMirrorReady,\n\t\t\t\tStatus: metav1.ConditionFalse,\n\t\t\t\tReason: qv1.ConditionReasonComponentNotReady,\n\t\t\t\tMessage: \"Awaiting for component quay to become available\",\n\t\t\t\tLastUpdateTime: metav1.NewTime(time.Now()),\n\t\t\t},\n\t\t)\n\t\treturn conds, nil\n\t}\n\n\t// checks now if the quay component is in a faulty state. if it is then sets mirror\n\t// component as faulty as well (awaiting for quay) and returns. quay condition is\n\t// append to the returned slice.\n\tquay := &Quay{Client: c}\n\tcond, err := quay.Check(ctx, q)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tconds = append(conds, cond)\n\n\t// if quay is in a faulty state then sets mirror as faulty and awaiting for quay. we\n\t// can return here as there is no need to check mirror status.\n\tif cond.Status != metav1.ConditionTrue {\n\t\tconds = append(\n\t\t\tconds,\n\t\t\tqv1.Condition{\n\t\t\t\tType: qv1.ComponentMirrorReady,\n\t\t\t\tStatus: metav1.ConditionFalse,\n\t\t\t\tReason: qv1.ConditionReasonComponentNotReady,\n\t\t\t\tMessage: \"Awaiting for component quay to become available\",\n\t\t\t},\n\t\t)\n\t\treturn conds, nil\n\t}\n\n\t// this is the last component we check the health for. it depends on quay component that\n\t// in turn depends on almost all other components.\n\tmirror := &Mirror{Client: c}\n\tcond, err = mirror.Check(ctx, q)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tconds = append(conds, cond)\n\treturn conds, nil\n}", "func (me *lmsEvaluator) Eval(game c4.State, p c4.Piece) float64 {\n\tvar bestScore, knownScore float64\n\n\t// Copy out the coefficients to reduce lock contention\n\tme.coeffsMutex.RLock()\n\tmyCoeffs := me.Coeffs\n\tme.coeffsMutex.RUnlock()\n\n\t// Estimate the game state's utility\n\tapproxScore, currentFeatures := BetterEval(myCoeffs, game, p)\n\n\t// Try to get a better estimate of the utility by looking one move ahead\n\t// with proven weights\n\tif game.GetTurn() != p {\n\t\tbestScore = math.Inf(-1)\n\t} else {\n\t\tbestScore = math.Inf(+1)\n\t}\n\n\tfor col := 0; col < c4.MaxColumns; col++ {\n\t\tif nextBoard, err := game.AfterMove(game.GetTurn(),\n\t\t\tcol); err == nil {\n\t\t\tnextScore, _ := BetterEval(\n\t\t\t\tmyCoeffs,\n\t\t\t\tnextBoard,\n\t\t\t\tnextBoard.GetTurn())\n\n\t\t\tif game.GetTurn() != p {\n\t\t\t\tif nextScore > bestScore {\n\t\t\t\t\tbestScore = nextScore\n\t\t\t\t}\n\t\t\t} else {\n\t\t\t\tif nextScore < bestScore {\n\t\t\t\t\tbestScore = nextScore\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\t// Use the evolved weights as a reference to prevent divergence\n\tknownScore, _ = BetterEval([6]float64{\n\t\t0.2502943943301069,\n\t\t-0.4952316649483701,\n\t\t0.3932539700819625,\n\t\t-0.2742452616759889,\n\t\t0.4746881137884282,\n\t\t0.2091091127191147}, game, p)\n\n\t// Change the coefficients according to the error\n\tme.count++\n\tif me.count%100000 == 0 {\n\t\tfmt.Println(me.count)\n\t\tfmt.Println(me.Coeffs)\n\t}\n\t// if !math.IsInf(bestScore, 0) {\n\t// \tfor j := 0; j < 6; j++ {\n\t// \t\tme.Coeffs[j] +=\n\t// \t\t\tmu * (bestScore - approxScore) * currentFeatures[j]\n\t// \t}\n\t// }\n\tgo func() {\n\t\tif !math.IsInf(bestScore, 0) {\n\t\t\tme.coeffsMutex.Lock()\n\t\t\tfor j := 0; j < 6; j++ {\n\t\t\t\tme.Coeffs[j] +=\n\t\t\t\t\tmu * (knownScore - approxScore) * currentFeatures[j]\n\t\t\t}\n\t\t\tme.coeffsMutex.Unlock()\n\t\t}\n\t}()\n\n\treturn approxScore\n}", "func (policy MutationPolicy) EvalMutation(ctx context.Context, m ent.Mutation) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalMutation(ctx, m); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (policy MutationPolicy) EvalMutation(ctx context.Context, m ent.Mutation) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalMutation(ctx, m); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (policy MutationPolicy) EvalMutation(ctx context.Context, m ent.Mutation) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalMutation(ctx, m); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (policy MutationPolicy) EvalMutation(ctx context.Context, m ent.Mutation) error {\n\tif decision, ok := decisionFromContext(ctx); ok {\n\t\treturn decision\n\t}\n\tfor _, rule := range policy {\n\t\tswitch decision := rule.EvalMutation(ctx, m); {\n\t\tcase decision == nil || errors.Is(decision, Skip):\n\t\tcase errors.Is(decision, Allow):\n\t\t\treturn nil\n\t\tdefault:\n\t\t\treturn decision\n\t\t}\n\t}\n\treturn nil\n}", "func (s *Subtraction) Evaluate(left, right EvalResult) (EvalResult, error) {\n\treturn subtractNumericWithError(left, right)\n}", "func (indis Individuals) Evaluate(parallel bool) error {\n\tif !parallel {\n\t\tvar err error\n\t\tfor i := range indis {\n\t\t\terr = indis[i].Evaluate()\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t}\n\n\tvar (\n\t\tnWorkers = runtime.GOMAXPROCS(-1)\n\t\tn = len(indis)\n\t\tchunkSize = (n + nWorkers - 1) / nWorkers\n\t\tg errgroup.Group\n\t)\n\n\tfor a := 0; a < n; a += chunkSize {\n\t\ta := a // https://golang.org/doc/faq#closures_and_goroutines\n\t\tvar b = minInt(a+chunkSize, n)\n\t\tg.Go(func() error {\n\t\t\treturn indis[a:b].Evaluate(false)\n\t\t})\n\t}\n\n\treturn g.Wait()\n}", "func (this *AggregateBase) evaluate(agg Aggregate, item value.Value,\n\tcontext expression.Context) (result value.Value, err error) {\n\tdefer func() {\n\t\tr := recover()\n\t\tif r != nil {\n\t\t\terr = fmt.Errorf(\"Error evaluating aggregate: %v.\", r)\n\t\t}\n\t}()\n\n\tav := item.(value.AnnotatedValue)\n\taggregates := av.GetAttachment(\"aggregates\")\n\tif aggregates != nil {\n\t\taggs := aggregates.(map[string]value.Value)\n\t\tresult = aggs[agg.String()]\n\t}\n\n\tif result == nil {\n\t\terr = fmt.Errorf(\"Aggregate %s not found.\", agg.String())\n\t}\n\n\treturn\n}", "func (this *MinVersion) Evaluate(item value.Value, context Context) (value.Value, error) {\n\treturn _MIN_VERSION_VALUE, nil\n}", "func (in InHandler) Evaluate(key, value interface{}) bool {\n\tswitch typedKey := key.(type) {\n\tcase string:\n\t\treturn in.validateValueWithStringPattern(typedKey, value)\n\tcase int, int32, int64, float32, float64, bool:\n\t\treturn in.validateValueWithStringPattern(fmt.Sprint(typedKey), value)\n\tcase []interface{}:\n\t\tvar stringSlice []string\n\t\tfor _, v := range typedKey {\n\t\t\tstringSlice = append(stringSlice, v.(string))\n\t\t}\n\t\treturn in.validateValueWithStringSetPattern(stringSlice, value)\n\tdefault:\n\t\tin.log.V(2).Info(\"Unsupported type\", \"value\", typedKey, \"type\", fmt.Sprintf(\"%T\", typedKey))\n\t\treturn false\n\t}\n}", "func (m *Multiplication) Evaluate(left, right EvalResult) (EvalResult, error) {\n\treturn multiplyNumericWithError(left, right)\n}", "func Evaluate(query string, values map[string]interface{}) interface{} {\n\ttokens := Parser(query)\n\trpn := ToPostfix(tokens)\n\touts := SolvePostfix(rpn, values)\n\treturn outs\n}", "func (s *service) BillingEvaluate(client req.Client, tenant string) (bool, error) {\n\tevaluation, _, err := client.BillingEvaluate(tenant)\n\tif err != nil {\n\t\treturn false, ErrEvaluate\n\t}\n\n\treturn evaluation.CanAccept, nil\n}", "func (p *policyd) Update(ctx context.Context) error {\n\tglg.Info(\"Updating policy\")\n\tdefer glg.Info(\"Updated policy\")\n\teg := errgroup.Group{}\n\trp := gache.New()\n\n\tfor _, domain := range p.athenzDomains {\n\t\tselect {\n\t\tcase <-ctx.Done():\n\t\t\tglg.Info(\"Update policy interrupted\")\n\t\t\treturn ctx.Err()\n\t\tdefault:\n\t\t\tdom := domain\n\t\t\teg.Go(func() error {\n\t\t\t\tselect {\n\t\t\t\tcase <-ctx.Done():\n\t\t\t\t\tglg.Info(\"Update policy interrupted\")\n\t\t\t\t\treturn ctx.Err()\n\t\t\t\tdefault:\n\t\t\t\t\treturn p.fetchAndCachePolicy(ctx, rp, dom)\n\t\t\t\t}\n\t\t\t})\n\t\t}\n\t}\n\n\tif err := eg.Wait(); err != nil {\n\t\treturn err\n\t}\n\n\trp.StartExpired(ctx, p.policyExpiredDuration).\n\t\tEnableExpiredHook().\n\t\tSetExpiredHook(func(ctx context.Context, key string) {\n\t\t\t//key = <domain>:role.<role>\n\t\t\tp.fetchAndCachePolicy(ctx, p.rolePolicies, strings.Split(key, \":role.\")[0])\n\t\t})\n\n\tp.rolePolicies, rp = rp, p.rolePolicies\n\trp.Stop()\n\trp.Clear()\n\n\treturn nil\n}", "func (net *Network) Evaluate2(inputValues []float64) (float64, error) {\n\tinputLength := len(inputValues)\n\tif inputLength > len(net.InputNodes) {\n\t\treturn 0.0, errors.New(\"Too many input values compared to input nodes\")\n\t}\n\tfor i, n := range net.InputNodes {\n\t\tif i < inputLength {\n\t\t\tn.SetValue(inputValues[i])\n\t\t}\n\t}\n\tmaxIterationCounter := net.maxIterations\n\tif maxIterationCounter == 0 {\n\t\t// If max iterations has not been configured, use 100\n\t\tmaxIterationCounter = 100\n\t}\n\tresult, _ := net.OutputNode.evaluate(net.Weight, &maxIterationCounter)\n\t//fmt.Println(\"Evaluation was stopped?\", stopped)\n\treturn result, nil //fmt.Errorf(\"evaluation was stopped after %d iterations\", maxIterationCounter)\n}", "func (iter ComputePolicyListResultIterator) Value() ComputePolicy {\n\tif !iter.page.NotDone() {\n\t\treturn ComputePolicy{}\n\t}\n\treturn iter.page.Values()[iter.i]\n}", "func Eval(node ast.Node, env *object.Environment, stop <-chan struct{}) object.Object {\n\tselect {\n\tcase <-stop:\n\t\treturn ConstNil\n\tdefault:\n\t}\n\n\tswitch node := node.(type) {\n\t// statements\n\tcase *ast.Program:\n\t\treturn evalProgram(node, env, stop)\n\tcase *ast.LetStatement:\n\t\tval := Eval(node.Value, env, stop)\n\t\tif isError(val) {\n\t\t\treturn val\n\t\t}\n\t\tenv.Set(node.Name.Value, val)\n\tcase *ast.ReturnStatement:\n\t\tval := Eval(node.Value, env, stop)\n\t\tif isError(val) {\n\t\t\treturn val\n\t\t}\n\t\treturn &object.ReturnValue{Value: val}\n\tcase *ast.BlockStatement:\n\t\treturn evalBlockStatement(node, env, stop)\n\tcase *ast.ExpressionStatement:\n\t\treturn Eval(node.Expression, env, stop)\n\n\t\t// expressions\n\tcase *ast.PrefixExpression:\n\t\tright := Eval(node.Right, env, stop)\n\t\tif isError(right) {\n\t\t\treturn right\n\t\t}\n\t\treturn evalPrefixExpr(node.Token, right)\n\tcase *ast.InfixExpression:\n\t\tif node.Operator == token.Assign {\n\t\t\treturn evalAssign(node, env, stop)\n\t\t}\n\n\t\tleft := Eval(node.Left, env, stop)\n\t\tif isError(left) {\n\t\t\treturn left\n\t\t}\n\t\tright := Eval(node.Right, env, stop)\n\t\tif isError(right) {\n\t\t\treturn right\n\t\t}\n\t\treturn evalInfixExpr(node.Token, left, right)\n\tcase *ast.IndexExpression:\n\t\tleft := Eval(node.Left, env, stop)\n\t\tif isError(left) {\n\t\t\treturn left\n\t\t}\n\n\t\tindex := Eval(node.Index, env, stop)\n\t\tif isError(index) {\n\t\t\treturn index\n\t\t}\n\t\treturn evalIndexExpr(node.Token, left, index)\n\tcase *ast.IfExpression:\n\t\treturn evalIfExpr(node, env, stop)\n\tcase *ast.WhileExpression:\n\t\treturn evalWhileExpr(node, env, stop)\n\tcase *ast.CallExpression:\n\t\tfunction := Eval(node.Func, env, stop)\n\t\tif isError(function) {\n\t\t\treturn function\n\t\t}\n\n\t\targs, err := evalExpressions(node.Args, env, stop)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn doFunction(node.Token, function, args, stop)\n\n\t\t// literals\n\tcase *ast.IntegerLiteral:\n\t\treturn &object.Integer{Value: node.Value}\n\tcase *ast.FloatLiteral:\n\t\treturn &object.Float{Value: node.Value}\n\tcase *ast.BooleanLiteral:\n\t\treturn boolToBoolean(node.Value)\n\tcase *ast.NilLiteral:\n\t\treturn ConstNil\n\tcase *ast.FunctionLiteral:\n\t\treturn &object.Function{Params: node.Params, Body: node.Body, Env: env}\n\tcase *ast.StringLiteral:\n\t\treturn &object.String{Value: node.Value}\n\tcase *ast.ArrayLiteral:\n\t\telems, err := evalExpressions(node.Elements, env, stop)\n\t\tif len(elems) == 1 && err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn &object.Array{Elements: elems}\n\tcase *ast.Identifier:\n\t\treturn evalIdentifier(node, env)\n\tcase *ast.AccessIdentifier:\n\t\treturn evalAccessIdentifier(node, env)\n\t}\n\treturn nil\n}", "func (pi *PolicyItem) computeEvalScore(service ServiceDefinition) {\n\tscore := ITEM_DEFAULT_SCORE\n\n\tif contains(pi.Groups, GroupPublic) {\n\t\tscore -= DISCOUNT_USERSGROUPS\n\t} else {\n\t\tcount := len(pi.Users) + len(pi.Groups)\n\t\tscore -= int(math.Min(float64(DISCOUNT_USERSGROUPS), float64(count)))\n\t}\n\n\tscore -= int(math.Round(float64((DISCOUNT_ACCESS_TYPES * len(pi.Accesses)) / len(service.AccessTypes))))\n\n\tcustomConditionsPenalty := CUSTOM_CONDITION_PENALTY * len(pi.Conditions)\n\tcustomConditionsDiscount := DISCOUNT_CUSTOM_CONDITIONS - customConditionsPenalty\n\n\tif customConditionsDiscount > 0 {\n\t\tscore -= customConditionsDiscount\n\t}\n\n\tpi.score = score\n}", "func (n *NotInOp) Evaluate(left, right EvalResult) (EvalResult, error) {\n\tpanic(\"implement me\")\n}", "func (e *Engine) Eval(ctx map[string]interface{}) ([]*model.DecisionValue, error) {\n\tinput := e.actPool.Get().(*activation)\n\tdefer e.actPool.Put(input)\n\tinput.vars = ctx\n\tvar decisions []*model.DecisionValue\n\tfor tmplName, insts := range e.instances {\n\t\trt, found := e.runtimes[tmplName]\n\t\tif !found {\n\t\t\t// Report an error\n\t\t\tcontinue\n\t\t}\n\t\tfor _, inst := range insts {\n\t\t\tif !e.selectInstance(inst, input) {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\tdecs, err := rt.Eval(inst, input)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tdecisions = append(decisions, decs...)\n\t\t}\n\t}\n\treturn decisions, nil\n}", "func ExampleEval() {\n\tfmt.Println(Eval(\"5\"))\n\tfmt.Println(Eval(\"1 + 2\"))\n\tfmt.Println(Eval(\"1 - 2 + 3\"))\n\tfmt.Println(Eval(\"3 * ( 3 + 1 * 3 ) / 2\"))\n\tfmt.Println(Eval(\"3 * ( ( 3 + 1 ) * 3 ) / 2\"))\n\t//OutPut:\n\t//5\n\t//3\n\t//2\n\t//9\n\t//18\n}" ]
[ "0.66251093", "0.6226339", "0.6088515", "0.5930903", "0.55782604", "0.55403847", "0.5462466", "0.54277873", "0.54017633", "0.5398092", "0.5365986", "0.53302735", "0.5328774", "0.52901286", "0.5252834", "0.5250405", "0.5220616", "0.52032036", "0.5192602", "0.5163183", "0.5145372", "0.5123805", "0.51151294", "0.5105302", "0.5105302", "0.5087724", "0.5069343", "0.50290555", "0.50149745", "0.4999898", "0.4987701", "0.49710584", "0.49573573", "0.49272192", "0.49066386", "0.48739234", "0.48570538", "0.48564392", "0.48556834", "0.48556834", "0.48370966", "0.48267275", "0.48139012", "0.4813686", "0.4810233", "0.48100615", "0.48042673", "0.47754592", "0.47754592", "0.47754592", "0.47754592", "0.47751048", "0.47597858", "0.47308198", "0.47240865", "0.47058597", "0.46938178", "0.46894225", "0.46868664", "0.46826744", "0.4677669", "0.4677171", "0.4673917", "0.46738628", "0.46724033", "0.46517172", "0.4646246", "0.46351755", "0.4620737", "0.46084487", "0.46059823", "0.46047655", "0.46016708", "0.4578055", "0.4560054", "0.45423466", "0.45310375", "0.45298657", "0.45245132", "0.4514632", "0.4514632", "0.4514632", "0.4514632", "0.4510055", "0.4495191", "0.44910693", "0.44695455", "0.446687", "0.44664115", "0.44659105", "0.44547674", "0.4453023", "0.44434032", "0.44252384", "0.4406182", "0.43883777", "0.43868378", "0.43746233", "0.43737423" ]
0.7003997
1
currentSecretValidForSpec is not actually registered as part of the policy chain and is instead called by currentCertificateRequestValidForSpec if no there is no existing CertificateRequest resource.
func currentSecretValidForSpec(input Input) (string, string, bool) { violations, err := certificates.SecretDataAltNamesMatchSpec(input.Secret, input.Certificate.Spec) if err != nil { // This case should never be reached as we already check the certificate data can // be parsed in an earlier policy check, but handle it anyway. // TODO: log a message return "", "", false } if len(violations) > 0 { return SecretMismatch, fmt.Sprintf("Existing issued Secret is not up to date for spec: %v", violations), true } return "", "", false }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (r *ResourceSpec) Valid() bool {\n\tif r.Name == \"\" {\n\t\tfmt.Println(\"no resource spec label\")\n\t\treturn false\n\t}\n\n\tfor _, c := range r.Credentials {\n\t\tif !c.Valid() {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func (ctl *Ctl) SpecIsValid() (bool, error) {\n\treturn true, nil\n}", "func (ctl *Ctl) SpecIsValid() (bool, error) {\n\treturn true, nil\n}", "func (r *Resource) Valid() bool {\n\tif r.Spec == nil {\n\t\tfmt.Println(\"no resource spec\")\n\t\treturn false\n\t}\n\n\treturn r.Spec.Valid()\n}", "func isSecret(resource v1alpha1.BackingServiceResource) bool {\n\treturn strings.ToLower(resource.Group+\".\"+resource.Version+\".\"+resource.Kind) == \".v1.secret\"\n}", "func (r *Resource) Valid() bool {\n\tif r.Name == \"\" {\n\t\tfmt.Println(\"no resource spec label\")\n\t\treturn false\n\t}\n\n\tfor _, c := range r.Credentials {\n\t\tif !c.Valid() {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func (self *Platform) ValidateSpec(spec *pb.ChaincodeSpec) error {\n\treturn nil\n}", "func (o *V1VolumeClaim) HasSpec() bool {\n\tif o != nil && o.Spec != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (r *NuxeoReconciler) secondarySecretIsCurrent(secondarySecret string, namespace string,\n\tresource v1alpha1.BackingServiceResource, resourceVersion string) bool {\n\tobj := corev1.Secret{}\n\tif err := r.Get(context.TODO(), types.NamespacedName{Name: secondarySecret, Namespace: namespace}, &obj); err != nil {\n\t\treturn false\n\t} else {\n\t\texpectedAnnotation := genAnnotationKey(resource)\n\t\tif existingResVer, ok := obj.Annotations[expectedAnnotation]; ok {\n\t\t\treturn existingResVer == resourceVersion\n\t\t}\n\t}\n\treturn false\n}", "func (m *SecretSpec) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.validateDriver(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateTemplating(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (s *RateLimitSuite) TestSpecIsOK(c *C) {\n\tc.Assert(plugin.NewRegistry().AddSpec(GetSpec()), IsNil)\n}", "func (s *RateLimitSuite) TestSpecIsOK(c *C) {\n\tc.Assert(plugin.NewRegistry().AddSpec(GetSpec()), IsNil)\n}", "func (in *SecretSpec) DeepCopy() *SecretSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(SecretSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (in *SecretSpec) DeepCopy() *SecretSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(SecretSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func validTlsSecret(t *testing.T) corev1.Secret {\n\tkey, err := rsa.GenerateKey(rand.Reader, 2048)\n\tif err != nil {\n\t\tt.Fatal(\"Private key cannot be created.\", err.Error())\n\t}\n\n\tcertTemplate := x509.Certificate{\n\t\tSerialNumber: big.NewInt(1337),\n\t}\n\tcert, err := x509.CreateCertificate(rand.Reader, &certTemplate, &certTemplate, &key.PublicKey, key)\n\tif err != nil {\n\t\tt.Fatal(\"Certificate cannot be created.\", err.Error())\n\t}\n\n\tca := &x509.Certificate{\n\t\tSerialNumber: big.NewInt(7331),\n\t\tIsCA: true,\n\t\tExtKeyUsage: []x509.ExtKeyUsage{x509.ExtKeyUsageClientAuth, x509.ExtKeyUsageServerAuth},\n\t\tKeyUsage: x509.KeyUsageDigitalSignature | x509.KeyUsageCertSign,\n\t}\n\n\tcaPrivKey, err := rsa.GenerateKey(rand.Reader, 4096)\n\tif err != nil {\n\t\tt.Fatal(\"CA private key cannot be created.\", err.Error())\n\t}\n\n\tcaBytes, err := x509.CreateCertificate(rand.Reader, ca, ca, &caPrivKey.PublicKey, caPrivKey)\n\tif err != nil {\n\t\tt.Fatal(\"CA certificate cannot be created.\", err.Error())\n\t}\n\n\tkeyPem := pem.EncodeToMemory(&pem.Block{\n\t\tType: \"RSA PRIVATE KEY\",\n\t\tBytes: x509.MarshalPKCS1PrivateKey(key),\n\t})\n\n\tcertPem := pem.EncodeToMemory(&pem.Block{\n\t\tType: \"CERTIFICATE\",\n\t\tBytes: cert,\n\t})\n\n\tcaPem := pem.EncodeToMemory(&pem.Block{\n\t\tType: \"CERTIFICATE\",\n\t\tBytes: caBytes,\n\t})\n\n\treturn corev1.Secret{\n\t\tData: map[string][]byte{\n\t\t\t\"certFile\": []byte(certPem),\n\t\t\t\"keyFile\": []byte(keyPem),\n\t\t\t\"caFile\": []byte(caPem),\n\t\t},\n\t}\n}", "func (s StatusCode) IsPrivateSpec() bool {\n\treturn s.In(StatusRangePrivate)\n}", "func ExpectValidCertificate(csr *certificatesv1.CertificateSigningRequest, _ crypto.Signer) error {\n\t_, err := pki.DecodeX509CertificateBytes(csr.Status.Certificate)\n\treturn err\n}", "func (r *reconciler) hasSecret(meta metav1.Object, o runtime.Object) bool {\n\tic := o.(*operatorv1.IngressController)\n\tsecretName := controller.RouterEffectiveDefaultCertificateSecretName(ic, r.operandNamespace)\n\tsecret := &corev1.Secret{}\n\tif err := r.client.Get(context.Background(), secretName, secret); err != nil {\n\t\tif errors.IsNotFound(err) {\n\t\t\treturn false\n\t\t}\n\t\tlog.Error(err, \"failed to look up secret for ingresscontroller\", \"name\", secretName, \"related\", meta.GetSelfLink())\n\t}\n\treturn true\n}", "func usesSecret(secret *corev1.Secret, podspec corev1.PodSpec) bool {\n\tvols := podspec.Volumes\n\n\tfor _, vol := range vols {\n\t\t// VolumeSource.Secret is a pointer, if it's uninitialized it should be nil\n\t\tif secretRef := vol.VolumeSource.Secret; secretRef != nil {\n\t\t\t// return true if we get a match\n\t\t\tif secretRef.SecretName == secret.GetName() {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\t// We didn't find a volume from a secret that matched our input secret.\n\treturn false\n}", "func validateEndpointRequestResource(req *request.Request, resource arn.Resource) error {\n\tresReq := s3shared.ResourceRequest{Request: req, Resource: resource}\n\n\tif len(resReq.Request.ClientInfo.PartitionID) != 0 && resReq.IsCrossPartition() {\n\t\treturn s3shared.NewClientPartitionMismatchError(resource,\n\t\t\treq.ClientInfo.PartitionID, aws.StringValue(req.Config.Region), nil)\n\t}\n\n\tif !resReq.AllowCrossRegion() && resReq.IsCrossRegion() {\n\t\treturn s3shared.NewClientRegionMismatchError(resource,\n\t\t\treq.ClientInfo.PartitionID, aws.StringValue(req.Config.Region), nil)\n\t}\n\n\t// Accelerate not supported\n\tif aws.BoolValue(req.Config.S3UseAccelerate) {\n\t\treturn s3shared.NewClientConfiguredForAccelerateError(resource,\n\t\t\treq.ClientInfo.PartitionID, aws.StringValue(req.Config.Region), nil)\n\t}\n\treturn nil\n}", "func vpaSpecNeedChange(hvpa *autoscalingv1alpha1.Hvpa, vpa *vpa_api.VerticalPodAutoscaler) bool {\n\treturn !reflect.DeepEqual(hvpa.Spec.Vpa.Template.Spec.ResourcePolicy, vpa.Spec.ResourcePolicy) ||\n\t\t*vpa.Spec.UpdatePolicy.UpdateMode != vpa_api.UpdateModeOff\n}", "func CertificateRequestInfoSupportsCertificate(cri *tls.CertificateRequestInfo, c *tls.Certificate,) error", "func ValidateControllerRegistrationSpec(spec *core.ControllerRegistrationSpec, fldPath *field.Path) field.ErrorList {\n\tallErrs := field.ErrorList{}\n\n\tvar (\n\t\tresourcesPath = fldPath.Child(\"resources\")\n\t\tdeploymentPath = fldPath.Child(\"deployment\")\n\n\t\tresources = make(map[string]string, len(spec.Resources))\n\t\tcontrolsResourcesPrimarily = false\n\t)\n\n\tfor i, resource := range spec.Resources {\n\t\tidxPath := resourcesPath.Index(i)\n\n\t\tif len(resource.Kind) == 0 {\n\t\t\tallErrs = append(allErrs, field.Required(idxPath.Child(\"kind\"), \"field is required\"))\n\t\t}\n\n\t\tif !extensionsv1alpha1.ExtensionKinds.Has(resource.Kind) {\n\t\t\tallErrs = append(allErrs, field.NotSupported(idxPath.Child(\"kind\"), resource.Kind, extensionsv1alpha1.ExtensionKinds.UnsortedList()))\n\t\t}\n\n\t\tif len(resource.Type) == 0 {\n\t\t\tallErrs = append(allErrs, field.Required(idxPath.Child(\"type\"), \"field is required\"))\n\t\t}\n\t\tif t, ok := resources[resource.Kind]; ok && t == resource.Type {\n\t\t\tallErrs = append(allErrs, field.Duplicate(idxPath, common.ExtensionID(resource.Kind, resource.Type)))\n\t\t}\n\t\tif resource.Kind != extensionsv1alpha1.ExtensionResource {\n\t\t\tif resource.GloballyEnabled != nil {\n\t\t\t\tallErrs = append(allErrs, field.Forbidden(idxPath.Child(\"globallyEnabled\"), fmt.Sprintf(\"field must not be set when kind != %s\", extensionsv1alpha1.ExtensionResource)))\n\t\t\t}\n\t\t\tif resource.ReconcileTimeout != nil {\n\t\t\t\tallErrs = append(allErrs, field.Forbidden(idxPath.Child(\"reconcileTimeout\"), fmt.Sprintf(\"field must not be set when kind != %s\", extensionsv1alpha1.ExtensionResource)))\n\t\t\t}\n\t\t}\n\n\t\tresources[resource.Kind] = resource.Type\n\t\tif resource.Primary == nil || *resource.Primary {\n\t\t\tcontrolsResourcesPrimarily = true\n\t\t}\n\t}\n\n\tif spec.Deployment != nil {\n\t\tif policy := spec.Deployment.Policy; policy != nil && !availablePolicies.Has(string(*policy)) {\n\t\t\tallErrs = append(allErrs, field.NotSupported(deploymentPath.Child(\"policy\"), *policy, availablePolicies.List()))\n\t\t}\n\n\t\tif spec.Deployment.SeedSelector != nil {\n\t\t\tif controlsResourcesPrimarily {\n\t\t\t\tallErrs = append(allErrs, field.Forbidden(deploymentPath.Child(\"seedSelector\"), \"specifying a seed selector is not allowed when controlling resources primarily\"))\n\t\t\t}\n\n\t\t\tallErrs = append(allErrs, metav1validation.ValidateLabelSelector(spec.Deployment.SeedSelector, deploymentPath.Child(\"seedSelector\"))...)\n\t\t}\n\t}\n\n\treturn allErrs\n}", "func (o *ProcessorSignalDecisionReportRequest) HasSecret() bool {\n\tif o != nil && o.Secret != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func checkSpec(spec *pb.ChaincodeSpec) error {\n\t// Don't allow nil value\n\tif spec == nil {\n\t\tfmt.Printf(\"Expected chaincode specification, nil received\")\n\t}\n\n\tplatform, err := platforms.Find(spec.Type)\n\tif err != nil {\n\t\tfmt.Printf(\"Failed to determine platform type: %s\", err)\n\t}\n\n\treturn platform.ValidateSpec(spec)\n}", "func (cs *CredentialSpecResource) DesiredTerminal() bool {\n\tcs.lock.RLock()\n\tdefer cs.lock.RUnlock()\n\n\treturn cs.desiredStatusUnsafe == resourcestatus.ResourceStatus(CredentialSpecRemoved)\n}", "func (ncc *NewCertConfig) Valid() (err error) {\n\tif ncc.CertTemplate == nil {\n\t\treturn fmt.Errorf(\"the template can't be empty\")\n\t}\n\n\tif ncc.Parent == nil {\n\t\tif ncc.KeyType == \"\" || ncc.KeyLength == \"\" {\n\t\t\tncc.KeyType = DefaultKeyType\n\t\t\tncc.KeyLength = DefaultKeyLength\n\t\t}\n\t\terr = ncc.genParent()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tif ncc.CertPool == nil {\n\t\tcertPool := x509.NewCertPool()\n\t\tcertPool.AddCert(ncc.Parent.Cert)\n\t\tncc.CertPool = certPool\n\t}\n\n\tif ncc.PublicKey == nil {\n\t\tncc.PublicKey, err = NewKeyPair(ncc.KeyType, ncc.KeyLength)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func isValidSecret(secret *v1.Secret) (bool, error) {\n\tswitch secret.Type {\n\t// We will accept TLS secrets that also have the 'ca.crt' payload.\n\tcase v1.SecretTypeTLS:\n\t\tdata, ok := secret.Data[v1.TLSCertKey]\n\t\tif !ok {\n\t\t\treturn false, errors.New(\"missing TLS certificate\")\n\t\t}\n\n\t\tif err := validateCertificate(data); err != nil {\n\t\t\treturn false, fmt.Errorf(\"invalid TLS certificate: %v\", err)\n\t\t}\n\n\t\tdata, ok = secret.Data[v1.TLSPrivateKeyKey]\n\t\tif !ok {\n\t\t\treturn false, errors.New(\"missing TLS private key\")\n\t\t}\n\n\t\tif err := validatePrivateKey(data); err != nil {\n\t\t\treturn false, fmt.Errorf(\"invalid TLS private key: %v\", err)\n\t\t}\n\n\t// Generic secrets may have a 'ca.crt' only.\n\tcase v1.SecretTypeOpaque, \"\":\n\t\tif _, ok := secret.Data[v1.TLSCertKey]; ok {\n\t\t\treturn false, nil\n\t\t}\n\n\t\tif _, ok := secret.Data[v1.TLSPrivateKeyKey]; ok {\n\t\t\treturn false, nil\n\t\t}\n\n\t\tif data := secret.Data[\"ca.crt\"]; len(data) == 0 {\n\t\t\treturn false, nil\n\t\t}\n\n\tdefault:\n\t\treturn false, nil\n\n\t}\n\n\t// If the secret we propose to accept has a CA bundle key,\n\t// validate that it is PEM certificate(s). Note that the\n\t// CA bundle on TLS secrets is allowed to be an empty string\n\t// (see https://github.com/projectcontour/contour/issues/1644).\n\tif data := secret.Data[\"ca.crt\"]; len(data) > 0 {\n\t\tif err := validateCertificate(data); err != nil {\n\t\t\treturn false, fmt.Errorf(\"invalid CA certificate bundle: %v\", err)\n\t\t}\n\t}\n\n\treturn true, nil\n}", "func (o *PartnerCustomerCreateRequest) HasSecret() bool {\n\tif o != nil && o.Secret != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (proxy *StandAloneProxyConfig) IsSecretConfiguredForCerts(secretName string) bool {\n\tfound := false\n\tfor _, server := range proxy.managementServers {\n\t\tif server.CertSecret == secretName {\n\t\t\tfound = true\n\t\t\tbreak\n\t\t}\n\t}\n\treturn found\n}", "func (c *CanaryDeployer) IsNewSpec(cd *flaggerv1.Canary) (bool, error) {\n\ttargetName := cd.Spec.TargetRef.Name\n\tcanary, err := c.kubeClient.AppsV1().Deployments(cd.Namespace).Get(targetName, metav1.GetOptions{})\n\tif err != nil {\n\t\tif errors.IsNotFound(err) {\n\t\t\treturn false, fmt.Errorf(\"deployment %s.%s not found\", targetName, cd.Namespace)\n\t\t}\n\t\treturn false, fmt.Errorf(\"deployment %s.%s query error %v\", targetName, cd.Namespace, err)\n\t}\n\n\tif cd.Status.LastAppliedSpec == \"\" {\n\t\treturn true, nil\n\t}\n\n\tnewSpec := &canary.Spec.Template.Spec\n\toldSpecJson, err := base64.StdEncoding.DecodeString(cd.Status.LastAppliedSpec)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"%s.%s decode error %v\", cd.Name, cd.Namespace, err)\n\t}\n\toldSpec := &corev1.PodSpec{}\n\terr = json.Unmarshal(oldSpecJson, oldSpec)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"%s.%s unmarshal error %v\", cd.Name, cd.Namespace, err)\n\t}\n\n\tif diff := cmp.Diff(*newSpec, *oldSpec, cmpopts.IgnoreUnexported(resource.Quantity{})); diff != \"\" {\n\t\t//fmt.Println(diff)\n\t\treturn true, nil\n\t}\n\n\treturn false, nil\n}", "func (c *controller) ensureSecretData(ctx context.Context, log logr.Logger, crt *cmapi.Certificate) error {\n\t// Retrieve the Secret which is associated with this Certificate.\n\tsecret, err := c.secretLister.Secrets(crt.Namespace).Get(crt.Spec.SecretName)\n\n\t// Secret doesn't exist so we can't do anything. The Certificate will be\n\t// marked for a re-issuance and the resulting Secret will be evaluated again.\n\tif apierrors.IsNotFound(err) {\n\t\tlog.V(logf.DebugLevel).Info(\"secret not found\", \"error\", err.Error())\n\t\treturn nil\n\t}\n\n\t// This error is transient, return error to be retried on the rate limiting\n\t// queue.\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tlog = log.WithValues(\"secret\", secret.Name)\n\n\t// If there is no certificate or private key data available at the target\n\t// Secret then exit early. The absense of these keys should cause an issuance\n\t// of the Certificate, so there is no need to run post issuance checks.\n\tif secret.Data == nil ||\n\t\tlen(secret.Data[corev1.TLSCertKey]) == 0 ||\n\t\tlen(secret.Data[corev1.TLSPrivateKeyKey]) == 0 {\n\t\tlog.V(logf.DebugLevel).Info(\"secret doesn't contain both certificate and private key data\",\n\t\t\t\"cert_data_len\", len(secret.Data[corev1.TLSCertKey]), \"key_data_len\", len(secret.Data[corev1.TLSPrivateKeyKey]))\n\t\treturn nil\n\t}\n\n\tdata := internal.SecretData{\n\t\tPrivateKey: secret.Data[corev1.TLSPrivateKeyKey],\n\t\tCertificate: secret.Data[corev1.TLSCertKey],\n\t\tCA: secret.Data[cmmeta.TLSCAKey],\n\t\tCertificateName: secret.Annotations[cmapi.CertificateNameKey],\n\t\tIssuerName: secret.Annotations[cmapi.IssuerNameAnnotationKey],\n\t\tIssuerKind: secret.Annotations[cmapi.IssuerKindAnnotationKey],\n\t\tIssuerGroup: secret.Annotations[cmapi.IssuerGroupAnnotationKey],\n\t}\n\n\t// Check whether the Certificate's Secret has correct output format and\n\t// metadata.\n\treason, message, isViolation := c.postIssuancePolicyChain.Evaluate(policies.Input{\n\t\tCertificate: crt,\n\t\tSecret: secret,\n\t})\n\n\tif isViolation {\n\t\tswitch reason {\n\t\tcase policies.InvalidCertificate, policies.ManagedFieldsParseError:\n\t\t\t//An error here indicates that the managed fields are malformed and the\n\t\t\t//decoder doesn't understand the managed fields on the Secret, or the\n\t\t\t//signed certificate data could not be decoded. There is nothing more the\n\t\t\t//controller can do here, so we exit nil so this controller doesn't end in\n\t\t\t//an infinite loop.\n\t\t\tlog.Error(errors.New(message), \"failed to determine whether the SecretTemplate matches Secret\")\n\t\t\treturn nil\n\t\tdefault:\n\n\t\t\t// Here the Certificate need to be re-reconciled.\n\t\t\tlog.Info(\"applying Secret data\", \"message\", message)\n\t\t\treturn c.secretsUpdateData(ctx, crt, data)\n\t\t}\n\t}\n\n\t// No Secret violations, nothing to do.\n\n\treturn nil\n}", "func TestHandleNewValidator(t *testing.T) {\n\tapp := simapp.Setup(false)\n\tctx := app.BaseApp.NewContext(false, tmproto.Header{})\n\n\taddrDels := simapp.AddTestAddrsIncremental(app, ctx, 1, sdk.TokensFromConsensusPower(200, sdk.DefaultPowerReduction))\n\tvalAddrs := simapp.ConvertAddrsToValAddrs(addrDels)\n\tpks := simapp.CreateTestPubKeys(1)\n\taddr, val := valAddrs[0], pks[0]\n\ttstaking := teststaking.NewHelper(t, ctx, app.CustomStakingKeeper, app.CustomGovKeeper)\n\tctx = ctx.WithBlockHeight(1)\n\n\t// Validator created\n\ttstaking.CreateValidator(addr, val, true)\n\n\tstaking.EndBlocker(ctx, app.CustomStakingKeeper)\n\n\t// Now a validator, for two blocks\n\tapp.CustomSlashingKeeper.HandleValidatorSignature(ctx, val.Address(), 100, true)\n\tctx = ctx.WithBlockHeight(2)\n\tapp.CustomSlashingKeeper.HandleValidatorSignature(ctx, val.Address(), 100, false)\n\n\tinfo, found := app.CustomSlashingKeeper.GetValidatorSigningInfo(ctx, sdk.ConsAddress(val.Address()))\n\trequire.True(t, found)\n\trequire.Equal(t, int64(1), info.StartHeight)\n\trequire.Equal(t, int64(1), info.MischanceConfidence)\n\trequire.Equal(t, int64(0), info.Mischance)\n\trequire.Equal(t, int64(1), info.MissedBlocksCounter)\n\trequire.Equal(t, int64(1), info.ProducedBlocksCounter)\n\trequire.Equal(t, time.Unix(0, 0).UTC(), info.InactiveUntil)\n\n\t// validator should be active still, should not have been inactivated\n\tvalidator, _ := app.CustomStakingKeeper.GetValidatorByConsAddr(ctx, sdk.GetConsAddress(val))\n\trequire.Equal(t, stakingtypes.Active, validator.GetStatus())\n}", "func (c *Cluster) RequiresSecret() bool {\n\treturn c.Spec.RootPasswordSecret == nil\n}", "func (m *Manager) IsTokenValid(ctx context.Context, token, resource, op string, attr map[string]string) (model.RequestParams, error) {\n\tm.lock.RLock()\n\tdefer m.lock.RUnlock()\n\n\tif !m.isProd {\n\t\treturn model.RequestParams{}, nil\n\t}\n\n\tclaims, err := m.parseToken(ctx, token)\n\tif err != nil {\n\t\treturn model.RequestParams{}, err\n\t}\n\n\t// Check if its an integration request and return the integration response if its an integration request\n\tres := m.integrationMan.HandleConfigAuth(ctx, resource, op, claims, attr)\n\tif res.CheckResponse() && res.Error() != nil {\n\t\treturn model.RequestParams{}, res.Error()\n\t}\n\n\t// Otherwise just return nil for backward compatibility\n\treturn model.RequestParams{Resource: resource, Op: op, Attributes: attr, Claims: claims}, nil\n}", "func _rtDealProposalIsInternallyValid(rt Runtime, dealP deal.StorageDealProposal) bool {\n\tif dealP.EndEpoch() <= dealP.StartEpoch() {\n\t\treturn false\n\t}\n\n\tif dealP.Duration() != dealP.EndEpoch()-dealP.StartEpoch() {\n\t\treturn false\n\t}\n\n\tIMPL_FINISH()\n\t// Get signature public key of client account actor.\n\tvar pk filcrypto.PublicKey\n\n\tIMPL_FINISH()\n\t// Determine which subset of DealProposal to use as the message to be signed by the client.\n\tvar m filcrypto.Message\n\n\t// Note: we do not verify the provider signature here, since this is implicit in the\n\t// authenticity of the on-chain message publishing the deal.\n\tsig := dealP.ClientSignature()\n\tsigVerified := vmr.RT_VerifySignature(rt, pk, sig, m)\n\tif !sigVerified {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func ExpectValidCertificate(expiresAfterDays int) HTTPCertificateExpectation {\n\treturn HTTPCertificateExpectation{\n\t\tExpiresAfterDays: expiresAfterDays,\n\t}\n}", "func NeedsResourceVersionUpdate(kind string) bool {\n\tif kind == \"SecurityContextConstraints\" ||\n\t\tkind == \"Service\" ||\n\t\tkind == \"ServiceMonitor\" ||\n\t\tkind == \"Route\" ||\n\t\tkind == \"Build\" ||\n\t\tkind == \"BuildRun\" ||\n\t\tkind == \"BuildConfig\" ||\n\t\tkind == \"ImageStream\" ||\n\t\tkind == \"PrometheusRule\" ||\n\t\tkind == \"CSIDriver\" ||\n\t\tkind == \"Issuer\" ||\n\t\tkind == \"CustomResourceDefinition\" ||\n\t\tkind == \"Certificate\" ||\n\t\tkind == \"SpecialResource\" ||\n\t\tkind == \"OperatorGroup\" ||\n\t\tkind == \"CertManager\" ||\n\t\tkind == \"MutatingWebhookConfiguration\" ||\n\t\tkind == \"ValidatingWebhookConfiguration\" ||\n\t\tkind == \"Deployment\" ||\n\t\tkind == \"ImagePolicy\" {\n\t\treturn true\n\t}\n\treturn false\n\n}", "func hasSecret(f Factory, spec *v1.PodSpec, ns, name string, wait bool) (bool, error) {\n\tfor _, c := range spec.InitContainers {\n\t\tif containerHasSecret(c, name) {\n\t\t\treturn true, nil\n\t\t}\n\t}\n\tfor _, c := range spec.Containers {\n\t\tif containerHasSecret(c, name) {\n\t\t\treturn true, nil\n\t\t}\n\t}\n\n\tsaName := spec.ServiceAccountName\n\tif saName != \"\" {\n\t\to, err := f.Get(\"v1/serviceaccounts\", client.FQN(ns, saName), wait, labels.Everything())\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\n\t\tvar sa v1.ServiceAccount\n\t\terr = runtime.DefaultUnstructuredConverter.FromUnstructured(o.(*unstructured.Unstructured).Object, &sa)\n\t\tif err != nil {\n\t\t\treturn false, errors.New(\"expecting ServiceAccount resource\")\n\t\t}\n\n\t\tfor _, ref := range sa.Secrets {\n\t\t\tif ref.Namespace == ns && ref.Name == name {\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t}\n\n\tfor _, v := range spec.Volumes {\n\t\tif sec := v.VolumeSource.Secret; sec != nil {\n\t\t\tif sec.SecretName == name {\n\t\t\t\treturn true, nil\n\t\t\t}\n\t\t}\n\t}\n\treturn false, nil\n}", "func (s SecretForTLSGeneratorV1) validate() error {\n\t// TODO: This is not strictly necessary. We can generate a self signed cert\n\t// if no key/cert is given. The only requiredment is that we either get both\n\t// or none. See test/e2e/ingress_utils for self signed cert generation.\n\tif len(s.Key) == 0 {\n\t\treturn fmt.Errorf(\"key must be specified.\")\n\t}\n\tif len(s.Cert) == 0 {\n\t\treturn fmt.Errorf(\"certificate must be specified.\")\n\t}\n\tif _, err := tls.LoadX509KeyPair(s.Cert, s.Key); err != nil {\n\t\treturn fmt.Errorf(\"failed to load key pair %v\", err)\n\t}\n\t// TODO: Add more validation.\n\t// 1. If the certificate contains intermediates, it is a valid chain.\n\t// 2. Format etc.\n\treturn nil\n}", "func TestAnySecurityRequirementMet(t *testing.T) {\n\t// Create of a map of scheme names and whether they are valid\n\tschemes := map[string]bool{\n\t\t\"a\": true,\n\t\t\"b\": true,\n\t\t\"c\": false,\n\t\t\"d\": false,\n\t}\n\n\t// Create the test cases\n\ttc := []struct {\n\t\tname string\n\t\tschemes []string\n\t\terror bool\n\t}{\n\t\t{\n\t\t\tname: \"/ok1\",\n\t\t\tschemes: []string{\"a\", \"b\"},\n\t\t\terror: false,\n\t\t},\n\t\t{\n\t\t\tname: \"/ok2\",\n\t\t\tschemes: []string{\"a\", \"c\"},\n\t\t\terror: false,\n\t\t},\n\t\t{\n\t\t\tname: \"/error\",\n\t\t\tschemes: []string{\"c\", \"d\"},\n\t\t\terror: true,\n\t\t},\n\t}\n\n\tdoc := openapi3.T{\n\t\tOpenAPI: \"3.0.0\",\n\t\tInfo: &openapi3.Info{\n\t\t\tTitle: \"MyAPI\",\n\t\t\tVersion: \"0.1\",\n\t\t},\n\t\tPaths: map[string]*openapi3.PathItem{},\n\t\tComponents: &openapi3.Components{\n\t\t\tSecuritySchemes: map[string]*openapi3.SecuritySchemeRef{},\n\t\t},\n\t}\n\n\t// Add the security schemes to the spec's components\n\tfor schemeName := range schemes {\n\t\tdoc.Components.SecuritySchemes[schemeName] = &openapi3.SecuritySchemeRef{\n\t\t\tValue: &openapi3.SecurityScheme{\n\t\t\t\tType: \"http\",\n\t\t\t\tScheme: \"basic\",\n\t\t\t},\n\t\t}\n\t}\n\n\t// Add the paths to the spec\n\tfor _, tc := range tc {\n\t\t// Create the security requirements from the test cases's schemes\n\t\tsecurityRequirements := openapi3.NewSecurityRequirements()\n\t\tfor _, scheme := range tc.schemes {\n\t\t\tsecurityRequirements.With(openapi3.SecurityRequirement{scheme: {}})\n\t\t}\n\n\t\t// Create the path with the security requirements\n\t\tdoc.Paths[tc.name] = &openapi3.PathItem{\n\t\t\tGet: &openapi3.Operation{\n\t\t\t\tSecurity: securityRequirements,\n\t\t\t\tResponses: openapi3.NewResponses(),\n\t\t\t},\n\t\t}\n\t}\n\n\terr := doc.Validate(context.Background())\n\trequire.NoError(t, err)\n\trouter, err := legacyrouter.NewRouter(&doc)\n\trequire.NoError(t, err)\n\n\t// Create the authentication function\n\tauthFunc := makeAuthFunc(schemes)\n\n\tfor _, tc := range tc {\n\t\t// Create the request input for the path\n\t\ttcURL, err := url.Parse(tc.name)\n\t\trequire.NoError(t, err)\n\t\thttpReq := httptest.NewRequest(http.MethodGet, tcURL.String(), nil)\n\t\troute, _, err := router.FindRoute(httpReq)\n\t\trequire.NoError(t, err)\n\t\treq := RequestValidationInput{\n\t\t\tRoute: route,\n\t\t\tOptions: &Options{\n\t\t\t\tAuthenticationFunc: authFunc,\n\t\t\t},\n\t\t}\n\n\t\t// Validate the security requirements\n\t\terr = ValidateSecurityRequirements(context.Background(), &req, *route.Operation.Security)\n\n\t\t// If there should have been an error\n\t\tif tc.error {\n\t\t\trequire.Errorf(t, err, \"an error is expected for path %q\", tc.name)\n\t\t} else {\n\t\t\trequire.NoErrorf(t, err, \"an error wasn't expected for path %q\", tc.name)\n\t\t}\n\t}\n}", "func ValidResource(api *kit.API, lookupOrgByResourceID func(context.Context, influxdb.ID) (influxdb.ID, error)) kit.Middleware {\n\treturn func(next http.Handler) http.Handler {\n\t\tfn := func(w http.ResponseWriter, r *http.Request) {\n\t\t\tstatusW := kit.NewStatusResponseWriter(w)\n\t\t\tid, err := influxdb.IDFromString(chi.URLParam(r, \"id\"))\n\t\t\tif err != nil {\n\t\t\t\tapi.Err(w, ErrCorruptID(err))\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tctx := r.Context()\n\n\t\t\torgID, err := lookupOrgByResourceID(ctx, *id)\n\t\t\tif err != nil {\n\t\t\t\tapi.Err(w, err)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tnext.ServeHTTP(statusW, r.WithContext(context.WithValue(ctx, ctxOrgKey, orgID)))\n\t\t}\n\t\treturn http.HandlerFunc(fn)\n\t}\n}", "func (o *AssetReportRefreshRequest) HasSecret() bool {\n\tif o != nil && o.Secret != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o PodSecurityPolicyPatchOutput) Spec() PodSecurityPolicySpecPatchPtrOutput {\n\treturn o.ApplyT(func(v *PodSecurityPolicyPatch) PodSecurityPolicySpecPatchPtrOutput { return v.Spec }).(PodSecurityPolicySpecPatchPtrOutput)\n}", "func (cs *CredentialSpecResource) updateAppliedStatusUnsafe(knownStatus resourcestatus.ResourceStatus) {\n\tif cs.appliedStatus == resourcestatus.ResourceStatus(CredentialSpecStatusNone) {\n\t\treturn\n\t}\n\n\t// Check if the resource transition has already finished\n\tif cs.appliedStatus <= knownStatus {\n\t\tcs.appliedStatus = resourcestatus.ResourceStatus(CredentialSpecStatusNone)\n\t}\n}", "func (e *Secret) CheckExisting(c *fi.Context) bool {\n\treturn true\n}", "func (c *AWSPCAIssuer) GetSpec() *AWSPCAIssuerSpec {\n\treturn &c.Spec\n}", "func TestIntegratedServiceManager_ValidateSpec(t *testing.T) {\n\ttests := []struct {\n\t\tname string\n\t\tspec integratedservices.IntegratedServiceSpec\n\t\tchecker func(err error) bool\n\t}{\n\t\t{\n\t\t\tname: \"initial test case\",\n\t\t\tspec: integratedservices.IntegratedServiceSpec{\n\t\t\t\t\"customAnchore\": obj{\n\t\t\t\t\t\"enabled\": true,\n\t\t\t\t\t\"url\": \"anchore.example.com\", // mandatory\n\t\t\t\t\t\"secretId\": \"mysecretid\", // mandatory\n\t\t\t\t},\n\t\t\t\t\"policy\": obj{\n\t\t\t\t\t\"policyId\": \"myPolicyID, select, from backend\",\n\t\t\t\t},\n\t\t\t\t\"releaseWhiteList\": []obj{ // optional\n\t\t\t\t\t{\n\t\t\t\t\t\t\"name\": \"name of release 1\", // mandatory\n\t\t\t\t\t\t\"reason\": \"reason of whitelisting\", // mandatory\n\t\t\t\t\t\t\"regexp\": \"whitelisted-[0-1]{2}.[a-z]{2,3}-releases\", // optional\n\t\t\t\t\t},\n\t\t\t\t\t{\n\t\t\t\t\t\t\"name\": \"name of release 2\",\n\t\t\t\t\t\t\"reason\": \"reason of whitelisting\",\n\t\t\t\t\t\t\"regexp\": \"whitelisted-[0-1]{2}.[a-z]{2,3}-releases\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\t\"webhookConfig\": obj{\n\t\t\t\t\t\"enabled\": true, //\n\t\t\t\t\t\"selector\": \"include or exclude\", // mandatory\n\t\t\t\t\t\"namespaces\": []string{\"default\", \"test\"},\n\t\t\t\t},\n\t\t\t},\n\t\t\tchecker: func(err error) bool {\n\t\t\t\treturn false\n\t\t\t},\n\t\t},\n\t\t// todo add more test fixtures here\n\t}\n\n\tctx := context.Background()\n\tintegratedServiceManager := MakeIntegratedServiceManager(nil, Config{})\n\tfor _, test := range tests {\n\t\tt.Run(test.name, func(t *testing.T) {\n\t\t\terr := integratedServiceManager.ValidateSpec(ctx, test.spec)\n\t\t\tif err != nil {\n\t\t\t\tt.Errorf(\"test failed with errors: %v\", err)\n\t\t\t}\n\t\t})\n\t}\n}", "func (mi *MinIOInstance) RequiresExternalCertSetup() bool {\n\treturn mi.Spec.ExternalCertSecret != nil\n}", "func (d Driver) Get(request secrets.Request) secrets.Response {\n\t// Handle them panics.\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tzap.S().Errorf(\"pki: error issuing certificate: %s\", r)\n\t\t}\n\t}()\n\n\tzap.S().Debugf(\"pki: got request for certificate: %#v\", request)\n\n\t// For now the secrets.Request.SecretLabels value does not get populated.\n\t// To work around this, the secret's labels are inspected on the daemon.\n\tmeta, _, err := d.client.SecretInspectWithRaw(context.Background(), request.SecretName)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"pki: error inspecting secret: %s\", err.Error())\n\n\t\t// The error is explicitly logged because Docker doesn't log the error returned.\n\t\tzap.S().Error(msg)\n\n\t\treturn secrets.Response{\n\t\t\tErr: msg,\n\t\t}\n\t}\n\n\tcertRequest := CertRequest{}\n\tif err := certRequest.FromSecretLabels(meta.Spec.Labels); err != nil {\n\t\tmsg := fmt.Sprintf(\"pki: error parsing secret's labels: %s\", err.Error())\n\n\t\t// The error is explicitly logged because Docker doesn't log the error returned.\n\t\tzap.S().Error(msg)\n\n\t\treturn secrets.Response{\n\t\t\tErr: msg,\n\t\t}\n\t}\n\n\tbundle, err := d.IssueCertificate(certRequest)\n\tif err != nil {\n\t\tmsg := fmt.Sprintf(\"pki: error issuing certificate: %s\", err.Error())\n\n\t\t// The error is explicitly logged because Docker doesn't log the error returned.\n\t\tzap.S().Error(msg)\n\n\t\treturn secrets.Response{\n\t\t\tErr: msg,\n\t\t}\n\t}\n\n\treturn secrets.Response{\n\t\tValue: bundle,\n\t\tDoNotReuse: true,\n\t}\n}", "func ValidFakeTemplateSpec() *batchv2alpha1.JobTemplateSpec {\n\treturn &batchv2alpha1.JobTemplateSpec{\n\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\tLabels: map[string]string{\n\t\t\t\t\"foo\": \"bar\",\n\t\t\t},\n\t\t},\n\t\tSpec: batchv1.JobSpec{\n\t\t\tTemplate: api.PodTemplateSpec{\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tLabels: map[string]string{\n\t\t\t\t\t\t\"foo\": \"bar\",\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t\tSpec: api.PodSpec{\n\t\t\t\t\tContainers: []api.Container{\n\t\t\t\t\t\t{\n\t\t\t\t\t\t\tName: \"baz\",\n\t\t\t\t\t\t\tImage: \"foo/bar\",\n\t\t\t\t\t\t\tImagePullPolicy: \"IfNotPresent\",\n\t\t\t\t\t\t},\n\t\t\t\t\t},\n\t\t\t\t\tRestartPolicy: \"OnFailure\",\n\t\t\t\t\tDNSPolicy: \"Default\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n}", "func (s *SpecGenerator) Validate() error {\n\t// Containers being added to a pod cannot have certain network attributes\n\t// associated with them because those should be on the infra container.\n\tif len(s.Pod) > 0 && s.NetNS.NSMode == FromPod {\n\t\tif len(s.Networks) > 0 {\n\t\t\treturn fmt.Errorf(\"networks must be defined when the pod is created: %w\", define.ErrNetworkOnPodContainer)\n\t\t}\n\t\tif len(s.PortMappings) > 0 || s.PublishExposedPorts {\n\t\t\treturn fmt.Errorf(\"published or exposed ports must be defined when the pod is created: %w\", define.ErrNetworkOnPodContainer)\n\t\t}\n\t\tif len(s.HostAdd) > 0 {\n\t\t\treturn fmt.Errorf(\"extra host entries must be specified on the pod: %w\", define.ErrNetworkOnPodContainer)\n\t\t}\n\t}\n\n\tif s.NetNS.IsContainer() && len(s.HostAdd) > 0 {\n\t\treturn fmt.Errorf(\"cannot set extra host entries when the container is joined to another containers network namespace: %w\", ErrInvalidSpecConfig)\n\t}\n\n\t//\n\t// ContainerBasicConfig\n\t//\n\t// Rootfs and Image cannot both populated\n\tif len(s.ContainerStorageConfig.Image) > 0 && len(s.ContainerStorageConfig.Rootfs) > 0 {\n\t\treturn fmt.Errorf(\"both image and rootfs cannot be simultaneously: %w\", ErrInvalidSpecConfig)\n\t}\n\t// Cannot set hostname and utsns\n\tif len(s.ContainerBasicConfig.Hostname) > 0 && !s.ContainerBasicConfig.UtsNS.IsPrivate() {\n\t\tif s.ContainerBasicConfig.UtsNS.IsPod() {\n\t\t\treturn fmt.Errorf(\"cannot set hostname when joining the pod UTS namespace: %w\", ErrInvalidSpecConfig)\n\t\t}\n\n\t\treturn fmt.Errorf(\"cannot set hostname when running in the host UTS namespace: %w\", ErrInvalidSpecConfig)\n\t}\n\t// systemd values must be true, false, or always\n\tif len(s.ContainerBasicConfig.Systemd) > 0 && !util.StringInSlice(strings.ToLower(s.ContainerBasicConfig.Systemd), SystemDValues) {\n\t\treturn fmt.Errorf(\"--systemd values must be one of %q: %w\", strings.Join(SystemDValues, \", \"), ErrInvalidSpecConfig)\n\t}\n\n\tif err := define.ValidateSdNotifyMode(s.ContainerBasicConfig.SdNotifyMode); err != nil {\n\t\treturn err\n\t}\n\n\t//\n\t// ContainerStorageConfig\n\t//\n\t// rootfs and image cannot both be set\n\tif len(s.ContainerStorageConfig.Image) > 0 && len(s.ContainerStorageConfig.Rootfs) > 0 {\n\t\treturn exclusiveOptions(\"rootfs\", \"image\")\n\t}\n\t// imagevolumemode must be one of ignore, tmpfs, or anonymous if given\n\tif len(s.ContainerStorageConfig.ImageVolumeMode) > 0 && !util.StringInSlice(strings.ToLower(s.ContainerStorageConfig.ImageVolumeMode), ImageVolumeModeValues) {\n\t\treturn fmt.Errorf(\"invalid ImageVolumeMode %q, value must be one of %s\",\n\t\t\ts.ContainerStorageConfig.ImageVolumeMode, strings.Join(ImageVolumeModeValues, \",\"))\n\t}\n\t// shmsize conflicts with IPC namespace\n\tif s.ContainerStorageConfig.ShmSize != nil && (s.ContainerStorageConfig.IpcNS.IsHost() || s.ContainerStorageConfig.IpcNS.IsNone()) {\n\t\treturn fmt.Errorf(\"cannot set shmsize when running in the %s IPC Namespace\", s.ContainerStorageConfig.IpcNS)\n\t}\n\n\t//\n\t// ContainerSecurityConfig\n\t//\n\t// userns and idmappings conflict\n\tif s.UserNS.IsPrivate() && s.IDMappings == nil {\n\t\treturn fmt.Errorf(\"IDMappings are required when not creating a User namespace: %w\", ErrInvalidSpecConfig)\n\t}\n\n\t//\n\t// ContainerCgroupConfig\n\t//\n\t//\n\t// None for now\n\n\t//\n\t// ContainerNetworkConfig\n\t//\n\t// useimageresolveconf conflicts with dnsserver, dnssearch, dnsoption\n\tif s.UseImageResolvConf {\n\t\tif len(s.DNSServers) > 0 {\n\t\t\treturn exclusiveOptions(\"UseImageResolvConf\", \"DNSServer\")\n\t\t}\n\t\tif len(s.DNSSearch) > 0 {\n\t\t\treturn exclusiveOptions(\"UseImageResolvConf\", \"DNSSearch\")\n\t\t}\n\t\tif len(s.DNSOptions) > 0 {\n\t\t\treturn exclusiveOptions(\"UseImageResolvConf\", \"DNSOption\")\n\t\t}\n\t}\n\t// UseImageHosts and HostAdd are exclusive\n\tif s.UseImageHosts && len(s.HostAdd) > 0 {\n\t\treturn exclusiveOptions(\"UseImageHosts\", \"HostAdd\")\n\t}\n\n\t// TODO the specgen does not appear to handle this? Should it\n\t// switch config.Cgroup.Cgroups {\n\t// case \"disabled\":\n\t//\tif addedResources {\n\t//\t\treturn errors.New(\"cannot specify resource limits when cgroups are disabled is specified\")\n\t//\t}\n\t//\tconfigSpec.Linux.Resources = &spec.LinuxResources{}\n\t// case \"enabled\", \"no-conmon\", \"\":\n\t//\t// Do nothing\n\t// default:\n\t//\treturn errors.New(\"unrecognized option for cgroups; supported are 'default', 'disabled', 'no-conmon'\")\n\t// }\n\t// Namespaces\n\tif err := s.UtsNS.validate(); err != nil {\n\t\treturn err\n\t}\n\tif err := validateIPCNS(&s.IpcNS); err != nil {\n\t\treturn err\n\t}\n\tif err := s.PidNS.validate(); err != nil {\n\t\treturn err\n\t}\n\tif err := s.CgroupNS.validate(); err != nil {\n\t\treturn err\n\t}\n\tif err := validateUserNS(&s.UserNS); err != nil {\n\t\treturn err\n\t}\n\n\tif err := validateNetNS(&s.NetNS); err != nil {\n\t\treturn err\n\t}\n\tif s.NetNS.NSMode != Bridge && len(s.Networks) > 0 {\n\t\t// Note that we also get the ip and mac in the networks map\n\t\treturn errors.New(\"networks and static ip/mac address can only be used with Bridge mode networking\")\n\t}\n\n\treturn nil\n}", "func (c *AWSPCAClusterIssuer) GetSpec() *AWSPCAIssuerSpec {\n\treturn &c.Spec\n}", "func (s *SsmSecret) isSecretValid() bool {\n\treturn int(time.Now().Unix()) <= s.expiry\n}", "func (service *PrivateLinkService) GetSpec() genruntime.ConvertibleSpec {\n\treturn &service.Spec\n}", "func (s BlockchainStatus) Valid() error {\n\tswitch s {\n\tcase Created, Preferred, Validating, Syncing:\n\t\treturn nil\n\tdefault:\n\t\treturn errUnknownBlockchainStatus\n\t}\n}", "func IsRequestValid(ctx context.Context, promises []*cAPI.Promise) (valid bool, signatureUUID bson.ObjectId, signers []Signer, senderIndex uint32) {\n\t// Due to specifications, there should be at least one promise (from the sender to himself)\n\tif len(promises) == 0 {\n\t\tvalid = false\n\t\treturn\n\t}\n\n\tok, expectedUUID, signers := IsPromiseSignedByPlatform(promises[0])\n\tif !ok {\n\t\tvalid = false\n\t\treturn\n\t}\n\n\tsender := net.GetClientHash(&ctx)\n\tif sender == nil {\n\t\tvalid = false\n\t\treturn\n\t}\n\n\tsenderIndex, err := GetIndexOfSigner(promises[0], sender)\n\tif err != nil {\n\t\tvalid = false\n\t\treturn\n\t}\n\n\t// To check that all the promises contain the same signed information, we only need to check that:\n\t// - it is correctly signed\n\t// - promises are consistent wrt at least one signed field\n\tfor _, promise := range promises {\n\t\tok, receivedUUID, _ := IsPromiseSignedByPlatform(promise)\n\t\tif !ok || (expectedUUID != receivedUUID) {\n\t\t\tvalid = false\n\t\t\treturn\n\t\t}\n\t}\n\n\treturn true, expectedUUID, signers, senderIndex\n}", "func (m SecurityListRequest) HasEncodedSecurityDesc() bool {\n\treturn m.Has(tag.EncodedSecurityDesc)\n}", "func (in *SecretTemplateSpec) DeepCopy() *SecretTemplateSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(SecretTemplateSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func (credential *FederatedIdentityCredential) GetSpec() genruntime.ConvertibleSpec {\n\treturn &credential.Spec\n}", "func ForContext(ctx context.Context, ds *did.Service, pks []did.DocPublicKey) (\n\t*ForContextData, error) {\n\t// NOTE(PN): Supporting only Secp251k1 keys for authentication for now\n\tkeyType := linkeddata.SuiteTypeSecp256k1Verification\n\n\treqTs, _ := ctx.Value(ReqTsCtxKey).(string)\n\tif reqTs == \"\" {\n\t\treturn nil, errors.New(\"no request ts passed in context\")\n\t}\n\tsignature, _ := ctx.Value(SignatureCtxKey).(string)\n\tif signature == \"\" {\n\t\treturn nil, errors.New(\"no signature passed in context\")\n\t}\n\tdidStr, _ := ctx.Value(DidCtxKey).(string)\n\n\tts, err := strconv.Atoi(reqTs)\n\tif err != nil {\n\t\treturn nil, errors.Wrap(err, \"could not convert ts to int\")\n\t}\n\n\t// altered grace period value can be passed in the context\n\tgracePeriod := DefaultRequestGracePeriodSecs\n\tgp, ok := ctx.Value(GracePeriodCtxKey).(int)\n\tif ok {\n\t\tgracePeriod = gp\n\t}\n\t// If did and key found, then pull doc for DID to check the signature\n\t// If no did and key passed, then check incoming list of pks to check signature\n\tif didStr != \"\" {\n\t\terr = VerifyEcdsaRequestSignatureWithDid(ds, keyType, signature, ts, didStr, gracePeriod)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t} else if pks != nil {\n\t\terr = VerifyEcdsaRequestSignatureWithPks(pks, keyType, signature, ts, \"\", gracePeriod)\n\t\tif err != nil {\n\t\t\treturn nil, err\n\t\t}\n\n\t} else {\n\t\treturn nil, errors.New(\"could not verify signature, no did or public keys\")\n\t}\n\n\treturn &ForContextData{Did: didStr}, nil\n}", "func (m JwtVapiClaims) Valid() error {\n\treturn nil\n}", "func outdatedSecretInUse(secretName, resourceVersion string, a map[string]string) bool {\n\tval, ok := a[secretResourceVersionAnnotation]\n\tif !ok {\n\t\t// the resourceVersion annotation wasn't on the object at all so\n\t\t// we have to assume the object needs a bounce.\n\t\treturn true\n\t}\n\n\tvar rvmap map[string]string\n\terr := json.Unmarshal([]byte(val), &rvmap)\n\tif err != nil {\n\t\t// it is expected that the value of this is a map. If it's not a map, we'll need\n\t\t// to clear it out and start again. We don't clear it out here. We let the update\n\t\t// logic handle that.\n\t\treturn true\n\t}\n\n\tsecretResourceVersion, ok := rvmap[secretName]\n\tif !ok {\n\t\t// we didn't find an entry for this object for the secret that's being reconciled\n\t\t// so we have to assume it needs updating and bounce the object.\n\t\treturn true\n\t}\n\n\tif secretResourceVersion != resourceVersion {\n\t\t// the resourceVersion on the secret does not match the resourceVersion for that secret that last\n\t\t// triggered a bounce, so we need t bounce it now.\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (l *lockRESTServer) IsValid(w http.ResponseWriter, r *http.Request) bool {\n\tif l.ll == nil {\n\t\tl.writeErrorResponse(w, errLockNotInitialized)\n\t\treturn false\n\t}\n\n\tif err := storageServerRequestValidate(r); err != nil {\n\t\tl.writeErrorResponse(w, err)\n\t\treturn false\n\t}\n\treturn true\n}", "func (e *DeployEngine) ActiveMatchesSpec() bool {\n\treturn podEquals(&e.Active.Spec.Template, &e.Deploy.Spec.Template)\n}", "func TestDesiredRouterCertsGlobalSecret(t *testing.T) {\n\ttype testInputs struct {\n\t\tingresses []operatorv1.IngressController\n\t\tsecrets []corev1.Secret\n\t}\n\ttype testOutputs struct {\n\t\tsecret *corev1.Secret\n\t}\n\tvar (\n\t\tdefaultCert = newSecret(\"router-certs-default\")\n\t\tdefaultCI = newIngressController(\"default\", \"\", \"apps.my.devcluster.openshift.com\")\n\n\t\tci1 = newIngressController(\"ci1\", \"s1\", \"dom1\")\n\t\tci2 = newIngressController(\"ci2\", \"s2\", \"dom2\")\n\t\ts1 = newSecret(\"s1\")\n\t\ts2 = newSecret(\"s2\")\n\t\t// data has the PEM for defaultCert, s1, and s2 (which all have\n\t\t// the same certificate and key).\n\t\tdata = bytes.Join([][]byte{\n\t\t\ts1.Data[\"tls.crt\"],\n\t\t\ts1.Data[\"tls.key\"],\n\t\t}, nil)\n\t)\n\ttestCases := []struct {\n\t\tdescription string\n\t\tinputs testInputs\n\t\toutput testOutputs\n\t}{\n\t\t{\n\t\t\tdescription: \"default configuration\",\n\t\t\tinputs: testInputs{\n\t\t\t\t[]operatorv1.IngressController{defaultCI},\n\t\t\t\t[]corev1.Secret{defaultCert},\n\t\t\t},\n\t\t\toutput: testOutputs{\n\t\t\t\t&corev1.Secret{\n\t\t\t\t\tData: map[string][]byte{\"apps.my.devcluster.openshift.com\": data},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tdescription: \"no ingresses\",\n\t\t\tinputs: testInputs{\n\t\t\t\t[]operatorv1.IngressController{},\n\t\t\t\t[]corev1.Secret{},\n\t\t\t},\n\t\t\toutput: testOutputs{nil},\n\t\t},\n\t\t{\n\t\t\tdescription: \"no secrets\",\n\t\t\tinputs: testInputs{\n\t\t\t\t[]operatorv1.IngressController{ci1},\n\t\t\t\t[]corev1.Secret{},\n\t\t\t},\n\t\t\toutput: testOutputs{nil},\n\t\t},\n\t\t{\n\t\t\tdescription: \"missing secret\",\n\t\t\tinputs: testInputs{\n\t\t\t\t[]operatorv1.IngressController{ci1, ci2},\n\t\t\t\t[]corev1.Secret{s1},\n\t\t\t},\n\t\t\toutput: testOutputs{\n\t\t\t\t&corev1.Secret{\n\t\t\t\t\tData: map[string][]byte{\"dom1\": data},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tdescription: \"extra secret\",\n\t\t\tinputs: testInputs{\n\t\t\t\t[]operatorv1.IngressController{ci2},\n\t\t\t\t[]corev1.Secret{s1, s2},\n\t\t\t},\n\t\t\toutput: testOutputs{\n\t\t\t\t&corev1.Secret{\n\t\t\t\t\tData: map[string][]byte{\"dom2\": data},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t{\n\t\t\tdescription: \"perfect match\",\n\t\t\tinputs: testInputs{\n\t\t\t\t[]operatorv1.IngressController{ci1, ci2},\n\t\t\t\t[]corev1.Secret{s1, s2},\n\t\t\t},\n\t\t\toutput: testOutputs{\n\t\t\t\t&corev1.Secret{\n\t\t\t\t\tData: map[string][]byte{\n\t\t\t\t\t\t\"dom1\": data,\n\t\t\t\t\t\t\"dom2\": data,\n\t\t\t\t\t},\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t}\n\n\tfor _, tc := range testCases {\n\t\texpected := tc.output.secret\n\t\tactual, err := desiredRouterCertsGlobalSecret(tc.inputs.secrets, tc.inputs.ingresses, \"openshift-ingress\")\n\t\tif err != nil {\n\t\t\tt.Errorf(\"failed to get desired router-ca global secret: %v\", err)\n\t\t\tcontinue\n\t\t}\n\t\tif expected == nil || actual == nil {\n\t\t\tif expected != nil {\n\t\t\t\tt.Errorf(\"%q: expected %v, got nil\", tc.description, expected)\n\t\t\t}\n\t\t\tif actual != nil {\n\t\t\t\tt.Errorf(\"%q: expected nil, got %v\", tc.description, actual)\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\t\tif !routerCertsSecretsEqual(expected, actual) {\n\t\t\tt.Errorf(\"%q: expected %v, got %v\", tc.description, expected, actual)\n\t\t}\n\t}\n}", "func (ctl *Ctl) CheckSpecFlags() error {\n\tfor _, registryJSON := range ctl.ScannerPodImageFacadeInternalRegistriesJSONSlice {\n\t\tregistry := &opssightv1.RegistryAuth{}\n\t\terr := json.Unmarshal([]byte(registryJSON), registry)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Invalid Registry Format\")\n\t\t}\n\t}\n\treturn nil\n}", "func (k Keeper) IsPocketSupportedBlockchain(ctx sdk.Ctx, chain string) bool {\n\t// loop through supported blockchains (network identifiers)\n\tfor _, c := range k.SupportedBlockchains(ctx) {\n\t\t// if contains chain return true\n\t\tif c == chain {\n\t\t\treturn true\n\t\t}\n\t}\n\t// else return false\n\treturn false\n}", "func TestDiscoveryResourceGate(t *testing.T) {\n\tresources := map[string][]metav1.APIResource{\n\t\t\"allLegacy\": {\n\t\t\t{Name: \"clusterpolicies\", Kind: \"ClusterPolicies\"},\n\t\t\t{Name: \"clusterpolicybindings\", Kind: \"ClusterPolicyBindings\"},\n\t\t\t{Name: \"policies\", Kind: \"Policies\"},\n\t\t\t{Name: \"policybindings\", Kind: \"PolicyBindings\"},\n\t\t\t{Name: \"foo\", Kind: \"Foo\"},\n\t\t},\n\t\t\"partialLegacy\": {\n\t\t\t{Name: \"clusterpolicies\", Kind: \"ClusterPolicies\"},\n\t\t\t{Name: \"clusterpolicybindings\", Kind: \"ClusterPolicyBindings\"},\n\t\t\t{Name: \"foo\", Kind: \"Foo\"},\n\t\t},\n\t\t\"noLegacy\": {\n\t\t\t{Name: \"foo\", Kind: \"Foo\"},\n\t\t\t{Name: \"bar\", Kind: \"Bar\"},\n\t\t},\n\t}\n\n\tlegacyTests := map[string]struct {\n\t\texistingResources *metav1.APIResourceList\n\t\texpectErrStr string\n\t}{\n\t\t\"scheme-legacy-all-supported\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: authorization.LegacySchemeGroupVersion.String(),\n\t\t\t\tAPIResources: resources[\"allLegacy\"],\n\t\t\t},\n\t\t\texpectErrStr: \"\",\n\t\t},\n\t\t\"scheme-legacy-some-supported\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: authorization.LegacySchemeGroupVersion.String(),\n\t\t\t\tAPIResources: resources[\"partialLegacy\"],\n\t\t\t},\n\t\t\texpectErrStr: \"the server does not support legacy policy resources\",\n\t\t},\n\t\t\"scheme-legacy-none-supported\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: authorization.LegacySchemeGroupVersion.String(),\n\t\t\t\tAPIResources: resources[\"noLegacy\"],\n\t\t\t},\n\t\t\texpectErrStr: \"the server does not support legacy policy resources\",\n\t\t},\n\t\t\"scheme-all-supported\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: authorization.SchemeGroupVersion.String(),\n\t\t\t\tAPIResources: resources[\"allLegacy\"],\n\t\t\t},\n\t\t\texpectErrStr: \"\",\n\t\t},\n\t\t\"scheme-some-supported\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: authorization.SchemeGroupVersion.String(),\n\t\t\t\tAPIResources: resources[\"partialLegacy\"],\n\t\t\t},\n\t\t\texpectErrStr: \"the server does not support legacy policy resources\",\n\t\t},\n\t\t\"scheme-none-supported\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: authorization.SchemeGroupVersion.String(),\n\t\t\t\tAPIResources: resources[\"noLegacy\"],\n\t\t\t},\n\t\t\texpectErrStr: \"the server does not support legacy policy resources\",\n\t\t},\n\t}\n\n\tdiscoveryTests := map[string]struct {\n\t\texistingResources *metav1.APIResourceList\n\t\tinputGVR []schema.GroupVersionResource\n\t\texpectedGVR []schema.GroupVersionResource\n\t\texpectedAll bool\n\t}{\n\t\t\"discovery-subset\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: \"v1\",\n\t\t\t\tAPIResources: resources[\"noLegacy\"],\n\t\t\t},\n\t\t\tinputGVR: []schema.GroupVersionResource{\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"foo\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"bar\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"noexist\",\n\t\t\t\t},\n\t\t\t},\n\t\t\texpectedGVR: []schema.GroupVersionResource{\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"foo\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"bar\",\n\t\t\t\t},\n\t\t\t},\n\t\t},\n\t\t\"discovery-none\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: \"v1\",\n\t\t\t\tAPIResources: resources[\"noLegacy\"],\n\t\t\t},\n\t\t\tinputGVR: []schema.GroupVersionResource{\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"noexist\",\n\t\t\t\t},\n\t\t\t},\n\t\t\texpectedGVR: []schema.GroupVersionResource{},\n\t\t},\n\t\t\"discovery-all\": {\n\t\t\texistingResources: &metav1.APIResourceList{\n\t\t\t\tGroupVersion: \"v1\",\n\t\t\t\tAPIResources: resources[\"noLegacy\"],\n\t\t\t},\n\t\t\tinputGVR: []schema.GroupVersionResource{\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"foo\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"bar\",\n\t\t\t\t},\n\t\t\t},\n\t\t\texpectedGVR: []schema.GroupVersionResource{\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"foo\",\n\t\t\t\t},\n\t\t\t\t{\n\t\t\t\t\tGroup: \"\",\n\t\t\t\t\tVersion: \"v1\",\n\t\t\t\t\tResource: \"bar\",\n\t\t\t\t},\n\t\t\t},\n\t\t\texpectedAll: true,\n\t\t},\n\t}\n\n\tfor tcName, tc := range discoveryTests {\n\t\tfunc() {\n\t\t\tserver := testServer(t, tc.existingResources)\n\t\t\tdefer server.Close()\n\t\t\tclient := discovery.NewDiscoveryClientForConfigOrDie(&restclient.Config{Host: server.URL})\n\n\t\t\tgot, all, err := DiscoverGroupVersionResources(client, tc.inputGVR...)\n\t\t\tif err != nil {\n\t\t\t\tt.Fatalf(\"myerr %s\", err.Error())\n\t\t\t}\n\t\t\tif !reflect.DeepEqual(got, tc.expectedGVR) {\n\t\t\t\tt.Fatalf(\"%s got %v, expected %v\", tcName, got, tc.expectedGVR)\n\t\t\t}\n\t\t\tif tc.expectedAll && !all {\n\t\t\t\tt.Fatalf(\"%s expected all\", tcName)\n\t\t\t}\n\t\t}()\n\t}\n\n\tfor tcName, tc := range legacyTests {\n\t\tfunc() {\n\t\t\tserver := testServer(t, tc.existingResources)\n\t\t\tdefer server.Close()\n\t\t\tclient := discovery.NewDiscoveryClientForConfigOrDie(&restclient.Config{Host: server.URL})\n\n\t\t\terr := LegacyPolicyResourceGate(client)\n\t\t\tif err != nil {\n\t\t\t\tif len(tc.expectErrStr) == 0 {\n\t\t\t\t\tt.Fatalf(\"%s unexpected err %s\\n\", tcName, err.Error())\n\t\t\t\t}\n\t\t\t\tif tc.expectErrStr != err.Error() {\n\t\t\t\t\tt.Fatalf(\"%s expected err %s, got %s\", tcName, tc.expectErrStr, err.Error())\n\t\t\t\t}\n\t\t\t}\n\t\t\tif err == nil && len(tc.expectErrStr) != 0 {\n\t\t\t\tt.Fatalf(\"%s expected err %s, got none\\n\", tcName, tc.expectErrStr)\n\t\t\t}\n\t\t}()\n\t}\n}", "func (in *SecretRefSpec) DeepCopy() *SecretRefSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(SecretRefSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func Validate(given secrets.Set, spec map[string]bundle.Credential, action string) error {\n\tfor name, cred := range spec {\n\t\tif !cred.AppliesTo(action) {\n\t\t\tcontinue\n\t\t}\n\n\t\tif !given.IsValid(name) && cred.Required {\n\t\t\treturn fmt.Errorf(\"bundle requires credential for %s\", name)\n\t\t}\n\t}\n\treturn nil\n}", "func (m SecurityListRequest) HasNoSecurityAltID() bool {\n\treturn m.Has(tag.NoSecurityAltID)\n}", "func ExpectValidPrivateKeyData(csr *certificatesv1.CertificateSigningRequest, key crypto.Signer) error {\n\tcert, err := pki.DecodeX509CertificateBytes(csr.Status.Certificate)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tequal := func() (bool, error) {\n\t\tswitch pub := key.Public().(type) {\n\t\tcase *rsa.PublicKey:\n\t\t\treturn pub.Equal(cert.PublicKey), nil\n\t\tcase *ecdsa.PublicKey:\n\t\t\treturn pub.Equal(cert.PublicKey), nil\n\t\tcase ed25519.PublicKey:\n\t\t\treturn pub.Equal(cert.PublicKey), nil\n\t\tdefault:\n\t\t\treturn false, fmt.Errorf(\"Unrecognised public key type: %T\", key)\n\t\t}\n\t}\n\n\tok, err := equal()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif !ok {\n\t\treturn errors.New(\"Expected signed certificate's public key to match requester's private key\")\n\t}\n\n\treturn nil\n}", "func IsSecretConfigured(ctx context.Context, conf *config.Configuration) bool {\n\tsecretId := getLicenseKeySecretId(conf)\n\tsecretValueInput := secretsmanager.GetSecretValueInput{SecretId: &secretId}\n\n\t_, err := secrets.GetSecretValueWithContext(ctx, &secretValueInput)\n\tif err != nil {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func createSupportBundleSpecSecret(app apptypes.AppType, sequence int64, kotsKinds *kotsutil.KotsKinds, secretName string, builtBundle *troubleshootv1beta2.SupportBundle, opts types.TroubleshootOptions, clientset kubernetes.Interface) error {\n\ts := serializer.NewYAMLSerializer(serializer.DefaultMetaFactory, scheme.Scheme, scheme.Scheme)\n\tvar b bytes.Buffer\n\tif err := s.Encode(builtBundle, &b); err != nil {\n\t\treturn errors.Wrap(err, \"failed to encode support bundle\")\n\t}\n\n\ttemplatedSpec := b.Bytes()\n\n\trenderedSpec, err := helper.RenderAppFile(app, &sequence, templatedSpec, kotsKinds, util.PodNamespace)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed render support bundle spec\")\n\t}\n\n\t// unmarshal the spec, look for image replacements in collectors and then remarshal\n\t// we do this after template rendering to support templating and then replacement\n\tsupportBundle, err := kotsutil.LoadSupportBundleFromContents(renderedSpec)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to unmarshal rendered support bundle spec\")\n\t}\n\n\tvar registrySettings registrytypes.RegistrySettings\n\tif !util.IsHelmManaged() {\n\t\ts, err := store.GetStore().GetRegistryDetailsForApp(app.GetID())\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to get registry settings for app\")\n\t\t}\n\t\tregistrySettings = s\n\t}\n\n\tcollectors, err := registry.UpdateCollectorSpecsWithRegistryData(supportBundle.Spec.Collectors, registrySettings, kotsKinds.Installation, kotsKinds.License, &kotsKinds.KotsApplication)\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to update collectors\")\n\t}\n\tsupportBundle.Spec.Collectors = collectors\n\tb.Reset()\n\tif err := s.Encode(supportBundle, &b); err != nil {\n\t\treturn errors.Wrap(err, \"failed to encode support bundle\")\n\t}\n\trenderedSpec = b.Bytes()\n\n\texistingSecret, err := clientset.CoreV1().Secrets(util.PodNamespace).Get(context.TODO(), secretName, metav1.GetOptions{})\n\tlabels := kotstypes.MergeLabels(kotstypes.GetKotsadmLabels(), kotstypes.GetTroubleshootLabels())\n\tif err != nil {\n\t\tif kuberneteserrors.IsNotFound(err) {\n\t\t\tsecret := &corev1.Secret{\n\t\t\t\tTypeMeta: metav1.TypeMeta{\n\t\t\t\t\tAPIVersion: \"v1\",\n\t\t\t\t\tKind: \"Secret\",\n\t\t\t\t},\n\t\t\t\tObjectMeta: metav1.ObjectMeta{\n\t\t\t\t\tName: secretName,\n\t\t\t\t\tNamespace: util.PodNamespace,\n\t\t\t\t\tLabels: labels,\n\t\t\t\t},\n\t\t\t\tData: map[string][]byte{\n\t\t\t\t\tSpecDataKey: renderedSpec,\n\t\t\t\t},\n\t\t\t}\n\n\t\t\t_, err = clientset.CoreV1().Secrets(util.PodNamespace).Create(context.TODO(), secret, metav1.CreateOptions{})\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, \"failed to create support bundle secret\")\n\t\t\t}\n\n\t\t\tlogger.Debugf(\"created %q support bundle spec secret\", secretName)\n\t\t} else {\n\t\t\treturn errors.Wrap(err, \"failed to read support bundle secret\")\n\t\t}\n\t} else {\n\t\tif existingSecret.Data == nil {\n\t\t\texistingSecret.Data = map[string][]byte{}\n\t\t}\n\t\texistingSecret.Data[SpecDataKey] = renderedSpec\n\t\texistingSecret.ObjectMeta.Labels = labels\n\n\t\t_, err = clientset.CoreV1().Secrets(util.PodNamespace).Update(context.TODO(), existingSecret, metav1.UpdateOptions{})\n\t\tif err != nil {\n\t\t\treturn errors.Wrap(err, \"failed to update support bundle secret\")\n\t\t}\n\t}\n\treturn nil\n}", "func (m *WireguardSpec) ContextValidate(ctx context.Context, formats strfmt.Registry) error {\n\treturn nil\n}", "func (o *ProcessorSignalDecisionReportRequest) GetSecretOk() (*string, bool) {\n\tif o == nil || o.Secret == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Secret, true\n}", "func (s *BareMetalMachineProviderSpec) IsValid() error {\n\tmissing := []string{}\n\tif s.CustomDeploy.Method == \"\" && s.Image.URL == \"\" {\n\t\tmissing = append(missing, \"Image.URL\")\n\t}\n\tif s.CustomDeploy.Method == \"\" && s.Image.Checksum == \"\" {\n\t\tmissing = append(missing, \"Image.Checksum\")\n\t}\n\tif len(missing) > 0 {\n\t\treturn fmt.Errorf(\"Missing fields from ProviderSpec: %v\", missing)\n\t}\n\treturn nil\n}", "func (ts Timespec) Valid() bool {\n\treturn !(ts.Sec < 0 || ts.Nsec < 0 || ts.Nsec >= int64(time.Second))\n}", "func (cs *CredentialSpecResource) KnownCreated() bool {\n\tcs.lock.RLock()\n\tdefer cs.lock.RUnlock()\n\n\treturn cs.knownStatusUnsafe == resourcestatus.ResourceStatus(CredentialSpecCreated)\n}", "func (req *WidgetPatch) Valid() bool {\n\treq.issues = []string{}\n\tif stringutil.IsWhiteSpace(req.Description) {\n\t\treq.issues = append(req.issues, \"Description cannot be blank\")\n\t}\n\treturn 0 == len(req.issues)\n}", "func (in *SecretAccessRequestSpec) DeepCopy() *SecretAccessRequestSpec {\n\tif in == nil {\n\t\treturn nil\n\t}\n\tout := new(SecretAccessRequestSpec)\n\tin.DeepCopyInto(out)\n\treturn out\n}", "func isSpecTrustedCASet(proxyConfig *configv1.ProxySpec) bool {\n\treturn len(proxyConfig.TrustedCA.Name) > 0\n}", "func (a *AmazonEC2) ValidateCloudSpecUpdate(oldSpec kubermaticv1.CloudSpec, newSpec kubermaticv1.CloudSpec) error {\n\treturn nil\n}", "func (o ValidatingAdmissionPolicyPatchTypeOutput) Spec() ValidatingAdmissionPolicySpecPatchPtrOutput {\n\treturn o.ApplyT(func(v ValidatingAdmissionPolicyPatchType) *ValidatingAdmissionPolicySpecPatch { return v.Spec }).(ValidatingAdmissionPolicySpecPatchPtrOutput)\n}", "func TestHandleAlreadyInactive(t *testing.T) {\n\t// initial setup\n\tapp := simapp.Setup(false)\n\tctx := app.BaseApp.NewContext(false, tmproto.Header{})\n\n\taddrDels := simapp.AddTestAddrsIncremental(app, ctx, 1, sdk.TokensFromConsensusPower(200, sdk.DefaultPowerReduction))\n\tvalAddrs := simapp.ConvertAddrsToValAddrs(addrDels)\n\tpks := simapp.CreateTestPubKeys(1)\n\taddr, val := valAddrs[0], pks[0]\n\tpower := int64(100)\n\ttstaking := teststaking.NewHelper(t, ctx, app.CustomStakingKeeper, app.CustomGovKeeper)\n\n\ttstaking.CreateValidator(addr, val, true)\n\n\tstaking.EndBlocker(ctx, app.CustomStakingKeeper)\n\n\t// 1000 first blocks OK\n\theight := int64(0)\n\tfor ; height < 1000; height++ {\n\t\tctx = ctx.WithBlockHeight(height)\n\t\tapp.CustomSlashingKeeper.HandleValidatorSignature(ctx, val.Address(), power, true)\n\t}\n\n\tproperties := app.CustomGovKeeper.GetNetworkProperties(ctx)\n\t// miss 11 blocks for mischance confidence\n\tfor ; height < 1000+int64(properties.MischanceConfidence)+1; height++ {\n\t\tctx = ctx.WithBlockHeight(height)\n\t\tapp.CustomSlashingKeeper.HandleValidatorSignature(ctx, val.Address(), power, false)\n\t}\n\n\t// info correctness after the overflow of mischance confidence\n\tinfo, found := app.CustomSlashingKeeper.GetValidatorSigningInfo(ctx, sdk.ConsAddress(val.Address()))\n\trequire.True(t, found)\n\trequire.Equal(t, int64(10), info.MischanceConfidence)\n\trequire.Equal(t, int64(1), info.Mischance)\n\trequire.Equal(t, int64(999), info.LastPresentBlock)\n\n\t// miss 110 blocks after mischance confidence happen\n\tfor ; height < 1000+int64(properties.MaxMischance+properties.MischanceConfidence)+1; height++ {\n\t\tctx = ctx.WithBlockHeight(height)\n\t\tapp.CustomSlashingKeeper.HandleValidatorSignature(ctx, val.Address(), power, false)\n\t}\n\n\t// end block\n\tstaking.EndBlocker(ctx, app.CustomStakingKeeper)\n\n\t// validator should have been inactivated\n\tvalidator, _ := app.CustomStakingKeeper.GetValidatorByConsAddr(ctx, sdk.GetConsAddress(val))\n\trequire.Equal(t, stakingtypes.Inactive, validator.GetStatus())\n\n\t// another block missed\n\tctx = ctx.WithBlockHeight(height)\n\tapp.CustomSlashingKeeper.HandleValidatorSignature(ctx, val.Address(), power, false)\n\n\t// validator should be in inactive status yet\n\tvalidator, _ = app.CustomStakingKeeper.GetValidatorByConsAddr(ctx, sdk.GetConsAddress(val))\n\trequire.Equal(t, stakingtypes.Inactive, validator.GetStatus())\n}", "func ValidateControllerRegistrationSpecUpdate(new, old *core.ControllerRegistrationSpec, deletionTimestampSet bool, fldPath *field.Path) field.ErrorList {\n\tallErrs := field.ErrorList{}\n\n\tif deletionTimestampSet && !apiequality.Semantic.DeepEqual(new, old) {\n\t\tallErrs = append(allErrs, apivalidation.ValidateImmutableField(new, old, fldPath)...)\n\t\treturn allErrs\n\t}\n\n\tkindTypeToPrimary := make(map[string]*bool, len(old.Resources))\n\tfor _, resource := range old.Resources {\n\t\tkindTypeToPrimary[resource.Kind+resource.Type] = resource.Primary\n\t}\n\tfor i, resource := range new.Resources {\n\t\tif primary, ok := kindTypeToPrimary[resource.Kind+resource.Type]; ok {\n\t\t\tallErrs = append(allErrs, apivalidation.ValidateImmutableField(resource.Primary, primary, fldPath.Child(\"resources\").Index(i).Child(\"primary\"))...)\n\t\t}\n\t}\n\n\treturn allErrs\n}", "func (m CrossOrderCancelReplaceRequest) HasEncodedSecurityDesc() bool {\n\treturn m.Has(tag.EncodedSecurityDesc)\n}", "func backingSvcIsValid(backing v1alpha1.BackingService) bool {\n\tif !reflect.DeepEqual(backing.Preconfigured, v1alpha1.PreconfiguredBackingService{}) {\n\t\treturn true\n\t} else {\n\t\treturn backing.Name != \"\" && !reflect.DeepEqual(backing.Resources, []v1alpha1.BackingServiceResource{})\n\t}\n}", "func (cs *CredentialSpecResource) Create() error {\n\tvar err error\n\tvar iamCredentials credentials.IAMRoleCredentials\n\n\texecutionCredentials, ok := cs.credentialsManager.GetTaskCredentials(cs.getExecutionCredentialsID())\n\tif ok {\n\t\tiamCredentials = executionCredentials.GetIAMRoleCredentials()\n\t}\n\n\tfor credSpecStr := range cs.credentialSpecContainerMap {\n\t\tcredSpecSplit := strings.SplitAfterN(credSpecStr, \"credentialspec:\", 2)\n\t\tif len(credSpecSplit) != 2 {\n\t\t\tseelog.Errorf(\"Invalid credentialspec: %s\", credSpecStr)\n\t\t\tcontinue\n\t\t}\n\t\tcredSpecValue := credSpecSplit[1]\n\n\t\tif strings.HasPrefix(credSpecValue, \"file://\") {\n\t\t\terr = cs.handleCredentialspecFile(credSpecStr)\n\t\t\tif err != nil {\n\t\t\t\tseelog.Errorf(\"Failed to handle the credentialspec file: %v\", err)\n\t\t\t\tcs.setTerminalReason(err.Error())\n\t\t\t\treturn err\n\t\t\t}\n\t\t\tcontinue\n\t\t}\n\n\t\tparsedARN, err := arn.Parse(credSpecValue)\n\t\tif err != nil {\n\t\t\tcs.setTerminalReason(err.Error())\n\t\t\treturn err\n\t\t}\n\n\t\tparsedARNService := parsedARN.Service\n\t\tif parsedARNService == \"s3\" {\n\t\t\terr = cs.handleS3CredentialspecFile(credSpecStr, credSpecValue, iamCredentials)\n\t\t\tif err != nil {\n\t\t\t\tseelog.Errorf(\"Failed to handle the credentialspec file from s3: %v\", err)\n\t\t\t\tcs.setTerminalReason(err.Error())\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else if parsedARNService == \"ssm\" {\n\t\t\terr = cs.handleSSMCredentialspecFile(credSpecStr, credSpecValue, iamCredentials)\n\t\t\tif err != nil {\n\t\t\t\tseelog.Errorf(\"Failed to handle the credentialspec file from SSM: %v\", err)\n\t\t\t\tcs.setTerminalReason(err.Error())\n\t\t\t\treturn err\n\t\t\t}\n\t\t} else {\n\t\t\terr = errors.New(\"unsupported credentialspec ARN, only s3/ssm ARNs are valid\")\n\t\t\tcs.setTerminalReason(err.Error())\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func TestShouldWaitForIngressGatewaySecretForFileMountedCerts(t *testing.T) {\n\tfetcher := &secretfetcher.SecretFetcher{\n\t\tUseCaClient: false,\n\t}\n\topt := Options{\n\t\tRotationInterval: 100 * time.Millisecond,\n\t\tEvictionDuration: 0,\n\t}\n\tsc := NewSecretCache(fetcher, notifyCb, opt)\n\tif sc.ShouldWaitForIngressGatewaySecret(\"\", \"\", \"\", true) {\n\t\tt.Fatalf(\"Expected not to wait for ingress gateway secret for file mounted certs, but got true\")\n\t}\n}", "func (k *Key) Valid(allowSpecial bool, kc KeyContext) bool {\n\tif !kc.Matches(k.kc) {\n\t\treturn false\n\t}\n\tfor _, t := range k.toks {\n\t\tif t.IsIncomplete() {\n\t\t\treturn false\n\t\t}\n\t\tif !allowSpecial && t.Special() {\n\t\t\treturn false\n\t\t}\n\t\tif t.Kind == \"\" {\n\t\t\treturn false\n\t\t}\n\t\tif t.StringID != \"\" && t.IntID != 0 {\n\t\t\treturn false\n\t\t}\n\t}\n\treturn true\n}", "func (o ArgoCDSpecServerPtrOutput) Insecure() pulumi.BoolPtrOutput {\n\treturn o.ApplyT(func(v *ArgoCDSpecServer) *bool {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Insecure\n\t}).(pulumi.BoolPtrOutput)\n}", "func (m SecurityListRequest) HasSecurityReqID() bool {\n\treturn m.Has(tag.SecurityReqID)\n}", "func CreateComplianceSecretDiagnostic() diagnostics.Diagnostic {\n\ttmpl := template.Must(template.New(\"compliance-secret\").Parse(complianceSecretCreateTemplateStr))\n\n\treturn diagnostics.Diagnostic{\n\t\tName: \"compliance-secret\",\n\t\tTags: diagnostics.Tags{\"compliance\"},\n\t\tGenerate: func(tstCtx diagnostics.TestContext) error {\n\t\t\tbuf := bytes.NewBuffer([]byte{})\n\t\t\tts := time.Now()\n\t\t\tname := \"integration-diagnostic-\" + ts.Format(\"20060102150405\")\n\t\t\tpassword := uuid.Must(uuid.NewV4()).String()\n\t\t\terr := tmpl.Execute(buf, struct {\n\t\t\t\tName string\n\t\t\t\tPassword string\n\t\t\t}{\n\t\t\t\tName: name,\n\t\t\t\tPassword: password,\n\t\t\t})\n\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\treqPath := \"/api/v0/secrets\"\n\t\t\tresp, err := tstCtx.DoLBRequest(\n\t\t\t\treqPath,\n\t\t\t\tlbrequest.WithMethod(\"POST\"),\n\t\t\t\tlbrequest.WithJSONBody(buf.String()),\n\t\t\t)\n\n\t\t\tif resp != nil && resp.StatusCode != 200 {\n\t\t\t\terr = errors.New(\"Status code not 200\")\n\t\t\t}\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"Request POST %s failed\\nBody:\\n%s\", reqPath, buf.String())\n\t\t\t}\n\n\t\t\tdefer func() {\n\t\t\t\t_ = resp.Body.Close()\n\t\t\t}()\n\n\t\t\trespUnmarshalled := make(map[string]interface{})\n\t\t\terr = json.NewDecoder(resp.Body).Decode(&respUnmarshalled)\n\n\t\t\tif err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\n\t\t\tid, ok := respUnmarshalled[\"id\"].(string)\n\n\t\t\tif !ok {\n\t\t\t\treturn errors.New(\"Could not find id in response\")\n\t\t\t}\n\n\t\t\ttstCtx.SetValue(\"compliance-secret\", complianceSecretSave{ID: id})\n\t\t\treturn err\n\t\t},\n\n\t\tVerify: func(tstCtx diagnostics.VerificationTestContext) {\n\t\t\tloaded := complianceSecretSave{}\n\n\t\t\terr := tstCtx.GetValue(\"compliance-secret\", &loaded)\n\t\t\trequire.NoError(tstCtx, err, \"Generated context was not found\")\n\n\t\t\treqPath := fmt.Sprintf(\"/api/v0/secrets/id/%s\", loaded.ID)\n\t\t\tresp, err := tstCtx.DoLBRequest(reqPath)\n\t\t\trequire.NoError(tstCtx, err)\n\t\t\tdefer func() {\n\t\t\t\t_ = resp.Body.Close()\n\t\t\t}()\n\n\t\t\trequire.Equal(tstCtx, 200, resp.StatusCode, \"Failed to GET %s\", reqPath)\n\t\t},\n\n\t\tCleanup: func(tstCtx diagnostics.TestContext) error {\n\t\t\tloaded := complianceSecretSave{}\n\t\t\terr := tstCtx.GetValue(\"compliance-secret\", &loaded)\n\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrap(err, \"Generated context was not found\")\n\t\t\t}\n\n\t\t\treqPath := fmt.Sprintf(\"/api/v0/secrets/id/%s\", loaded.ID)\n\t\t\tresp, err := tstCtx.DoLBRequest(\n\t\t\t\treqPath,\n\t\t\t\tlbrequest.WithMethod(\"DELETE\"),\n\t\t\t)\n\n\t\t\tif err != nil {\n\t\t\t\treturn errors.Wrapf(err, \"Failed to DELETE %s\", reqPath)\n\t\t\t}\n\n\t\t\tdefer func() {\n\t\t\t\t_ = resp.Body.Close()\n\t\t\t}()\n\n\t\t\tif resp.StatusCode != 200 {\n\t\t\t\treturn errors.New(\"Unexpected status code\")\n\t\t\t}\n\n\t\t\treturn nil\n\t\t},\n\t}\n}", "func (m *WireguardSpec) Validate(formats strfmt.Registry) error {\n\treturn nil\n}", "func (req *WidgetRequest) Valid() bool {\n\treq.issues = []string{}\n\tif stringutil.IsWhiteSpace(req.SerialNumber) {\n\t\treq.issues = append(req.issues, \"SerialNumber cannot be blank\")\n\t}\n\tif stringutil.IsWhiteSpace(req.Description) {\n\t\treq.issues = append(req.issues, \"Description cannot be blank\")\n\t}\n\treturn 0 == len(req.issues)\n}", "func MatchesPodSpec(matchConfig MatchConfig, podSpec apiv1.PodSpec) bool {\n\treturn matchesContainers(matchConfig, podSpec.Containers) || matchesContainers(matchConfig, podSpec.InitContainers)\n}", "func (cs *CredentialSpecResource) GetKnownStatus() resourcestatus.ResourceStatus {\n\tcs.lock.RLock()\n\tdefer cs.lock.RUnlock()\n\n\treturn cs.knownStatusUnsafe\n}", "func (o *V1VolumeClaim) GetSpecOk() (*V1VolumeClaimSpec, bool) {\n\tif o == nil || o.Spec == nil {\n\t\treturn nil, false\n\t}\n\treturn o.Spec, true\n}" ]
[ "0.59268194", "0.5744408", "0.5744408", "0.5339567", "0.5226835", "0.5213273", "0.519293", "0.51824284", "0.5161792", "0.512809", "0.5099267", "0.5099267", "0.50246483", "0.50246483", "0.5004235", "0.49910876", "0.4983041", "0.49570647", "0.4919323", "0.49098366", "0.49080744", "0.48815414", "0.48555464", "0.4850875", "0.48205218", "0.4811914", "0.4801103", "0.47976184", "0.4795537", "0.47947422", "0.47946468", "0.4782098", "0.47280014", "0.47230983", "0.4709645", "0.4691092", "0.46860325", "0.46792337", "0.46742338", "0.46723026", "0.46459517", "0.46445718", "0.4641652", "0.46334407", "0.46259207", "0.4597022", "0.45871726", "0.45754257", "0.4564516", "0.45593864", "0.4554398", "0.45502996", "0.45493805", "0.45474672", "0.45415804", "0.45392147", "0.4538358", "0.4531613", "0.45299998", "0.45267278", "0.45253924", "0.4519636", "0.45160154", "0.4510404", "0.45043412", "0.45042366", "0.44919854", "0.4489912", "0.44851393", "0.44823888", "0.44690877", "0.44627506", "0.44626048", "0.44535246", "0.4447107", "0.44438556", "0.44270906", "0.4426777", "0.4410505", "0.4408488", "0.4393925", "0.43934378", "0.43894455", "0.4385954", "0.4384235", "0.4382695", "0.4381684", "0.4381413", "0.43758953", "0.43708158", "0.43702573", "0.43651527", "0.4364925", "0.4363218", "0.43613052", "0.4359776", "0.43596312", "0.43561387", "0.435299", "0.43526217" ]
0.7577245
0
CurrentCertificateNearingExpiry returns a policy function that can be used to check whether an X.509 cert currently issued for a Certificate should be renewed.
func CurrentCertificateNearingExpiry(c clock.Clock) Func { return func(input Input) (string, string, bool) { // Determine if the certificate is nearing expiry solely by looking at // the actual cert, if it exists. We assume that at this point we have // called policy functions that check that input.Secret and // input.Secret.Data exists (SecretDoesNotExist and SecretIsMissingData). x509cert, err := pki.DecodeX509CertificateBytes(input.Secret.Data[corev1.TLSCertKey]) if err != nil { // This case should never happen as it should always be caught by the // secretPublicKeysMatch function beforehand, but handle it just in case. return "InvalidCertificate", fmt.Sprintf("Failed to decode stored certificate: %v", err), true } notBefore := metav1.NewTime(x509cert.NotBefore) notAfter := metav1.NewTime(x509cert.NotAfter) crt := input.Certificate renewalTime := certificates.RenewalTime(notBefore.Time, notAfter.Time, crt.Spec.RenewBefore) renewIn := renewalTime.Time.Sub(c.Now()) if renewIn > 0 { //renewal time is in future, no need to renew return "", "", false } return Renewing, fmt.Sprintf("Renewing certificate as renewal was scheduled at %s", input.Certificate.Status.RenewalTime), true } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func CurrentCertificateHasExpired(c clock.Clock) Func {\n\treturn func(input Input) (string, string, bool) {\n\t\tcertData, ok := input.Secret.Data[corev1.TLSCertKey]\n\t\tif !ok {\n\t\t\treturn MissingData, \"Missing Certificate data\", true\n\t\t}\n\t\t// TODO: replace this with a generic decoder that can handle different\n\t\t// formats such as JKS, P12 etc (i.e. add proper support for keystores)\n\t\tcert, err := pki.DecodeX509CertificateBytes(certData)\n\t\tif err != nil {\n\t\t\t// This case should never happen as it should always be caught by the\n\t\t\t// secretPublicKeysMatch function beforehand, but handle it just in case.\n\t\t\treturn \"InvalidCertificate\", fmt.Sprintf(\"Failed to decode stored certificate: %v\", err), true\n\t\t}\n\n\t\tif c.Now().After(cert.NotAfter) {\n\t\t\treturn Expired, fmt.Sprintf(\"Certificate expired on %s\", cert.NotAfter.Format(time.RFC1123)), true\n\t\t}\n\t\treturn \"\", \"\", false\n\t}\n}", "func (c Certificate) Expired(now time.Time, skew time.Duration) bool {\n\treturn c.IssuedAt.After(now) || c.ExpiresAt.Before(now)\n}", "func goodFor(cert *x509.Certificate) (time.Duration, bool) {\n\t// If we got called with a cert that doesn't exist, just say there's no\n\t// time left, and it needs to be renewed\n\tif cert == nil {\n\t\treturn 0, false\n\t}\n\t// These are all int64's with Seconds since the Epoch, handy for the math\n\tstart, end := cert.NotBefore.Unix(), cert.NotAfter.Unix()\n\tnow := time.Now().UTC().Unix()\n\tif end <= now { // already expired\n\t\treturn 0, false\n\t}\n\tlifespan := end - start // full ttl of cert\n\tduration := end - now // duration remaining\n\tgooddur := (duration * 9) / 10 // 90% of duration\n\tmindur := (lifespan / 10) // 10% of lifespan\n\tif gooddur <= mindur {\n\t\treturn 0, false // almost expired, get a new one\n\t}\n\tif gooddur > 100 { // 100 seconds\n\t\t// add jitter if big enough for it to matter\n\t\tr := rand.New(rand.NewSource(time.Now().UnixNano()))\n\t\t// between 87% and 93%\n\t\tgooddur = gooddur + ((gooddur / 100) * int64(r.Intn(6)-3))\n\t}\n\tsleepFor := time.Duration(gooddur * 1e9) // basically: gooddur*time.Second\n\treturn sleepFor, true\n}", "func MaxAgePolicy(d time.Duration) GCPolicy { return maxAgePolicy(d) }", "func (b *ProviderBasis) IsExpired() bool {\n\tif b.CurrentTime == nil {\n\t\tb.CurrentTime = time.Now\n\t}\n\treturn !b.AlwaysValid && !b.CurrentTime().Before(b.expiration)\n}", "func leaseExpired(grantedAt time.Time) bool {\n\treturn time.Since(grantedAt).Seconds() > (storagerpc.LeaseSeconds + storagerpc.LeaseGuardSeconds)\n}", "func (cc *CurrentConditions) Expired() bool {\n\tnow := time.Now()\n\texpired := now.After(cc.RequestTimestamp.Add(UpdateInterval))\n\treturn expired\n}", "func (certList *CertificateList) ExpiredAt(now time.Time) bool {\n\treturn now.After(certList.TBSCertList.NextUpdate)\n}", "func TimeBasedPolicy(fn func(openedAtUnix int64) bool) PolicyFunc {\n\treturn func(fileState FileState) bool {\n\t\treturn fn(fileState.OpenedAt)\n\t}\n}", "func (c *Cache) IsExpired(key string, current int64) bool {\r\n\treturn (current > c.MagicKeys[key].Expiration)\r\n}", "func filterExpiredCerts(certs ...*x509.Certificate) []*x509.Certificate {\n\tfiveMinutesAgo := time.Now().Add(-5 * time.Minute)\n\n\tvar validCerts []*x509.Certificate\n\tfor _, c := range certs {\n\t\tif c.NotAfter.After(fiveMinutesAgo) {\n\t\t\tvalidCerts = append(validCerts, c)\n\t\t}\n\t}\n\n\treturn validCerts\n}", "func (l Info) IsExpiredWithGracePeriod() bool {\n\treturn l.ExpiresAt.Add(3 * 24 * time.Hour).Before(time.Now())\n}", "func (i *info) expired(now int64) bool {\n\treturn i.TTLStamp <= now\n}", "func isLeaseExpired(lease *db.Lease, context *leaseContext, actualPrincipalSpend float64, principalBudgetAmount float64) (bool, db.LeaseStatusReason) {\n\n\tif context.expireDate >= lease.ExpiresOn {\n\t\treturn true, db.LeaseExpired\n\t} else if context.actualSpend > lease.BudgetAmount {\n\t\treturn true, db.LeaseOverBudget\n\t} else if actualPrincipalSpend > principalBudgetAmount {\n\t\treturn true, db.LeaseOverPrincipalBudget\n\t}\n\n\treturn false, db.LeaseActive\n}", "func (tcertBlock *TCertBlock) isExpired() bool {\n\ttsNow := time.Now()\n\tnotAfter := tcertBlock.GetTCert().GetCertificate().NotAfter\n\tpoolLogger.Debugf(\"#isExpired: %s now: %s deadline: %s \\n \", tsNow.Add(fivemin).After(notAfter), tsNow, notAfter)\n\tif tsNow.Add(fivemin).After(notAfter) {\n\t\treturn true\n\t}\n\treturn false\n}", "func (r *resourceRecord) isCloseToExpiring() bool {\n\telapsed := (r.initialTimeToLive - r.remainingTimeToLive).Seconds()\n\n\t// RFC 6762 section 10 specifies that records should be refreshed when more than 80% of\n\t// their initial time-to-live has elapsed.\n\treturn (elapsed / r.initialTimeToLive.Seconds()) > 0.8\n}", "func (r *RenewalInfoResponse) ShouldRenewAt(now time.Time, willingToSleep time.Duration) *time.Time {\n\t// Explicitly convert all times to UTC.\n\tnow = now.UTC()\n\tstart := r.SuggestedWindow.Start.UTC()\n\tend := r.SuggestedWindow.End.UTC()\n\n\t// Select a uniform random time within the suggested window.\n\twindow := end.Sub(start)\n\trandomDuration := time.Duration(rand.Int63n(int64(window)))\n\trt := start.Add(randomDuration)\n\n\t// If the selected time is in the past, attempt renewal immediately.\n\tif rt.Before(now) {\n\t\treturn &now\n\t}\n\n\t// Otherwise, if the client can schedule itself to attempt renewal at exactly the selected time, do so.\n\twillingToSleepUntil := now.Add(willingToSleep)\n\tif willingToSleepUntil.After(rt) || willingToSleepUntil.Equal(rt) {\n\t\treturn &rt\n\t}\n\n\t// TODO: Otherwise, if the selected time is before the next time that the client would wake up normally, attempt renewal immediately.\n\n\t// Otherwise, sleep until the next normal wake time, re-check ARI, and return to Step 1.\n\treturn nil\n}", "func (d *VirtualMachineController) hasGracePeriodExpired(dom *api.Domain) (hasExpired bool, timeLeft int) {\n\n\thasExpired = false\n\ttimeLeft = 0\n\n\tif dom == nil {\n\t\thasExpired = true\n\t\treturn\n\t}\n\n\tstartTime := int64(0)\n\tif dom.Spec.Metadata.KubeVirt.GracePeriod.DeletionTimestamp != nil {\n\t\tstartTime = dom.Spec.Metadata.KubeVirt.GracePeriod.DeletionTimestamp.UTC().Unix()\n\t}\n\tgracePeriod := dom.Spec.Metadata.KubeVirt.GracePeriod.DeletionGracePeriodSeconds\n\n\t// If gracePeriod == 0, then there will be no startTime set, deletion\n\t// should occur immediately during shutdown.\n\tif gracePeriod == 0 {\n\t\thasExpired = true\n\t\treturn\n\t} else if startTime == 0 {\n\t\t// If gracePeriod > 0, then the shutdown signal needs to be sent\n\t\t// and the gracePeriod start time needs to be set.\n\t\ttimeLeft = -1\n\t\treturn\n\t}\n\n\tnow := time.Now().UTC().Unix()\n\tdiff := now - startTime\n\n\tif diff >= gracePeriod {\n\t\thasExpired = true\n\t\treturn\n\t}\n\n\ttimeLeft = int(gracePeriod - diff)\n\tif timeLeft < 1 {\n\t\ttimeLeft = 1\n\t}\n\treturn\n}", "func (i *Info) expired(now int64) bool {\n\treturn i.TTLStamp <= now\n}", "func CertExpiry(host, port, days string, overrideTLS bool) (bool, bool, error) {\n\tlog.Info(\"Testing SSL expiration on host \", host, \" over port \", port)\n\tvar certExpired bool\n\tvar expireWarning bool\n\n\td := &net.Dialer{\n\t\tTimeout: time.Duration(TimeoutSeconds) * time.Second,\n\t}\n\n\t// InsecureSkipVerify should be false unless checking a self-signed certificate\n\tconn, err := tls.DialWithDialer(d, \"tcp\", host+\":\"+port, &tls.Config{\n\t\tInsecureSkipVerify: overrideTLS,\n\t\tMinVersion: tls.VersionTLS12,\n\t})\n\n\tif err != nil {\n\t\tlog.Warnln([]*x509.Certificate{&x509.Certificate{}}, \"\", err)\n\t\treturn certExpired, expireWarning, err\n\t}\n\tdefer conn.Close()\n\n\t// var cert is assigned the slice of certs that can be pulled from a given host\n\tcert := conn.ConnectionState().PeerCertificates\n\tcurrentTime := time.Now()\n\n\t// convert # of days declared in pod spec from string to uint64, then to uint, to compare against cert expiration info\n\tdaysInt64, _ := strconv.ParseUint(days, 10, 64)\n\tdaysInt := uint(daysInt64)\n\n\t// calculate # of hours until the domain cert (cert[0] from the slice) is invalid, then convery to uint and to # of days\n\tdaysUntilInvalid := (uint(cert[0].NotAfter.Sub(currentTime).Hours())) / uint(24)\n\tlog.Info(\"Certificate for \", host, \" is valid from \", cert[0].NotBefore, \" until \", cert[0].NotAfter)\n\n\t// check that the current date/time is between the certificate's Not Before and Not After window\n\tif currentTime.Before(cert[0].NotBefore) || currentTime.After(cert[0].NotAfter) {\n\t\tcertExpired = true\n\t\tlog.Warn(\"Certificate for domain \", host, \" expired on \", cert[0].NotAfter)\n\t}\n\n\t// check that the # of days in the pod spec is greater than the number of days left until cert expiration\n\tif daysInt >= daysUntilInvalid {\n\t\texpireWarning = true\n\t\tlog.Warn(\"Certificate for domain \", host, \" will expire in \", daysUntilInvalid, \" days\")\n\t}\n\n\tif (daysInt <= daysUntilInvalid) && (currentTime.Before(cert[0].NotAfter) || currentTime.After(cert[0].NotBefore)) {\n\t\tlog.Info(\"Certificate for domain \", host, \" is currently valid and will expire in \", daysUntilInvalid, \" days\")\n\t}\n\treturn certExpired, expireWarning, err\n}", "func (e *EnvProvider) IsExpired() bool {\n\treturn !e.retrieved\n}", "func (a *AssumeRoleProvider) IsExpired() bool {\n\treturn a.expiration.Before(time.Now())\n}", "func TestExpiringCert(t *testing.T) {\n\t// Create cert/key pair\n\tcert, key, err := genCerts(time.Now().Add(360 * time.Hour))\n\tif err != nil {\n\t\tt.Logf(\"Unable to generate test certificates - %s\", err)\n\t\tt.FailNow()\n\t}\n\n\t// Start Listener\n\tl, err := startListener(cert, key)\n\tif err != nil {\n\t\tt.Logf(\"%s\", err)\n\t\tt.FailNow()\n\t}\n\ttime.Sleep(30 * time.Millisecond)\n\tdefer l.Close()\n\n\t// Test if it expires within x days\n\tt.Run(\"ExpiresWithin30Days\", func(t *testing.T) {\n\t\tvar v bool\n\t\tv, err := ExpiresWithinDays(\"127.0.0.1:9000\", 30)\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Unexpected failure when calling ExpiresWithinDays - %s\", err)\n\t\t}\n\t\tif v == false {\n\t\t\tt.Errorf(\"Unexpected result when testing ExpiresWithinDays with a cert that expires in 15 days, expected true and got %+v\", v)\n\t\t}\n\t})\n\n\t// Test if it expires by x date\n\tt.Run(\"ExpiresBeforeDate\", func(t *testing.T) {\n\t\tvar v bool\n\t\tv, err := ExpiresBeforeDate(\"127.0.0.1:9000\", time.Now().Add(720*time.Hour))\n\t\tif err != nil {\n\t\t\tt.Errorf(\"Unexpected failure when calling ExpiresBeforeDate - %s\", err)\n\t\t}\n\t\tif v == false {\n\t\t\tt.Errorf(\"Unexpected result when testing ExpiredsBeforeDate with a cert that expires in 15 days, expected true got %+v\", v)\n\t\t}\n\t})\n}", "func testExpired(ttl time.Duration) cache.DirtyFunc {\n\treturn func(file storage.FileEntry) bool {\n\t\treturn file.LastModified.Before(time.Now().Add(-ttl))\n\t}\n}", "func (mgmt *MonitoringManager) CheckExpiry() {\n\t// Get current time\n\tcurTime := time.Now\n\t// Get new list to store all unexpired clients\n\tunexpired := make([]*Monitor, 0)\n\n\t// Iterate through current list of clients monitoring a facility\n\tfor _,v := range mgmt.MonitorList {\n\t\t// If client's monitoring end time is after the current time, add to the new list\n\t\tif v.End.After(curTime()) {\n\t\t\t// Not expired\n\t\t\tunexpired = append(unexpired, v)\n\t\t}\n\t}\n\n\t// Overwrite the current list with the new list of monitoring clients\n\tmgmt.MonitorList = unexpired\n}", "func certificateExpiry(certificate string) (string, error) {\n\texpiry, err := certificateExpiryTime(certificate)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\treturn expiry.UTC().Format(time.RFC3339), nil\n}", "func (c *CIDOffer) HasExpired() bool {\n\texpiryTime := time.Unix(c.expiry, 0)\n\tnow := time.Now()\n\treturn expiryTime.Before(now)\n}", "func (v value) expired(c *Cache) bool{\n return time.Since(v.time)>c.expire\n}", "func (e *expirationChecker) Expired(ref ChunkEntry, now model.Time) (bool, []model.Interval) {\n\tuserID := unsafeGetString(ref.UserID)\n\tperiod := e.tenantsRetention.RetentionPeriodFor(userID, ref.Labels)\n\treturn now.Sub(ref.Through) > period, nil\n}", "func (cert *Certificate) ExpiresAt() time.Time {\n\treturn cert.Certificate.Leaf.NotAfter.UTC()\n}", "func SigningPolicy(certExpiry time.Duration) *cfconfig.Signing {\n\t// Force the minimum Certificate expiration to be fifteen minutes\n\tif certExpiry < MinNodeCertExpiration {\n\t\tcertExpiry = DefaultNodeCertExpiration\n\t}\n\n\t// Add the backdate\n\tcertExpiry = certExpiry + CertBackdate\n\n\treturn &cfconfig.Signing{\n\t\tDefault: &cfconfig.SigningProfile{\n\t\t\tUsage: []string{\"signing\", \"key encipherment\", \"server auth\", \"client auth\"},\n\t\t\tExpiry: certExpiry,\n\t\t\tBackdate: CertBackdate,\n\t\t\t// Only trust the key components from the CSR. Everything else should\n\t\t\t// come directly from API call params.\n\t\t\tCSRWhitelist: &cfconfig.CSRWhitelist{\n\t\t\t\tPublicKey: true,\n\t\t\t\tPublicKeyAlgorithm: true,\n\t\t\t\tSignatureAlgorithm: true,\n\t\t\t},\n\t\t},\n\t}\n}", "func (s *spec) expiry(current Index) (expiry *Expiry, err error) {\n\tstate, has := s.states[current]\n\tif !has {\n\t\terr = ErrUnknownState{spec: s, Index: current}\n\t\treturn\n\t}\n\tif state.TTL.TTL > 0 {\n\t\texpiry = &state.TTL\n\t}\n\treturn\n}", "func (c *DomainController) expired() error {\n\t// assert domain exists\n\tif err := c.requireDomain(); err != nil {\n\t\tpanic(\"conditions check not allowed on non existing domain\")\n\t}\n\texpireTime := utils.SecondsToTime(c.domain.ValidUntil)\n\t// if expire time is before block time means domain expired\n\tif expireTime.Before(c.ctx.BlockTime()) {\n\t\treturn nil\n\t}\n\n\treturn sdkerrors.Wrapf(types.ErrDomainNotExpired, \"domain %s has not expired\", c.domain.Name)\n}", "func IsExpired(ctx Context, t UnixTime) bool {\n\tblockNow, err := BlockTime(ctx)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"%+v\", err))\n\t}\n\treturn t <= AsUnixTime(blockNow)\n}", "func (l *localOracle) UntilExpired(lockTimeStamp, TTL uint64, opt *oracle.Option) int64 {\n\tnow := time.Now()\n\tif l.hook != nil {\n\t\tnow = l.hook.currentTime\n\t}\n\treturn oracle.ExtractPhysical(lockTimeStamp) + int64(TTL) - oracle.GetPhysical(now)\n}", "func (provider *AdhocProvider) isStale(certificate *tls.Certificate) bool {\n\tttlOffset := provider.TTLOffset\n\tif ttlOffset == 0 {\n\t\tttlOffset = DefaultTTLOffset\n\t}\n\n\texpiresAt := certificate.Leaf.NotAfter.Add(ttlOffset)\n\n\treturn time.Now().After(expiresAt)\n}", "func RenewCert(c *cli.Context, cfg *arc_config.Config) (int, error) {\n\t// check api renew cert uri\n\trenewCertURI, err := RenewCertURI(c)\n\tif err != nil {\n\t\treturn 1, err\n\t}\n\tfmt.Printf(\"Using URI %s \\n\", renewCertURI)\n\n\tnotAfter, err := pki.CertExpirationDate(cfg)\n\tif err != nil {\n\t\treturn 1, err\n\t}\n\n\thoursLeft := pki.CertExpiresIn(notAfter)\n\tdaysLeft := hoursLeft / 24\n\tfmt.Printf(\"Current cert expires on %s (%d days). \\n\", notAfter.Format(\"2006-01-02 15:04:05\"), int(daysLeft))\n\n\t// ask the user to continue\n\tvar s string\n\tfmt.Print(\"Do you want to renew the cert? (y|n): \")\n\t_, err = fmt.Scan(&s)\n\tif err != nil {\n\t\treturn 1, err\n\t}\n\ts = strings.TrimSpace(s)\n\ts = strings.ToLower(s)\n\tif s != \"y\" && s != \"yes\" {\n\t\treturn 0, nil\n\t}\n\n\t// get the new cert\n\terr = pki.RenewCert(cfg, renewCertURI, false)\n\tif err != nil {\n\t\treturn 1, err\n\t}\n\n\tfmt.Println(\"Cert successfully downloaded\")\n\n\treturn 0, nil\n}", "func TestCertificateNotExpired(t *testing.T) {\n\t// given\n\tvar expiredDate = time.Now().Add(time.Hour * 24 * (30 + 1)) // 31 days.\n\n\tvar fakeCerts = []*x509.Certificate{\n\t\t{\n\t\t\tNotAfter: expiredDate,\n\t\t\tSubject: pkix.Name{\n\t\t\t\tCommonName: \"Test cert\",\n\t\t\t},\n\t\t},\n\t}\n\n\t// when\n\tmsg := getCertificateChainMsg(fakeCerts)\n\n\t// then\n\tif msg != \"\" {\n\t\tt.Fatalf(\"Expected empty message was: %s\", msg)\n\t}\n}", "func isExpired(cli *clientv3.Client, ev *clientv3.Event) (bool, error) {\n\tif ev.PrevKv == nil {\n\t\treturn false, nil\n\t}\n\n\tleaseID := clientv3.LeaseID(ev.PrevKv.Lease)\n\tif leaseID == clientv3.NoLease {\n\t\treturn false, nil\n\t}\n\n\tttlResponse, err := cli.TimeToLive(context.Background(), leaseID)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn ttlResponse.TTL == -1, nil\n}", "func (c CachedObject) IsExpired() bool {\r\n\r\n\telapsed := time.Now().Sub(c.CreatedAt.Add(time.Hour * getExpiryTimeInHrs()))\r\n\r\n\tif elapsed > 0.0 {\r\n\t\treturn true\r\n\t}\r\n\r\n\treturn false\r\n}", "func (c *certManager) IssueOrRenewCert(cfg *CertConfig, renewUnder int, verbose bool) (bool, error) {\n\tif !verbose {\n\t\tacme.Logger = log.New(ioutil.Discard, \"\", 0)\n\t}\n\n\tlog.Printf(\"Checking certificate [%s]\", cfg.CertName)\n\tif err := os.MkdirAll(filepath.Dir(c.certFile(cfg.CertName, \"json\")), perms); err != nil {\n\t\treturn false, err\n\t}\n\texisting, err := c.readCertificate(cfg.CertName)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\tvar action = func() (acme.CertificateResource, error) {\n\t\treturn c.client.ObtainCertificate(cfg.Names, true, nil, true)\n\t}\n\n\tif existing == nil {\n\t\tlog.Println(\"No existing cert found. Issuing new...\")\n\t} else {\n\t\tnames, daysLeft, err := getCertInfo(existing.Certificate)\n\t\tif err != nil {\n\t\t\treturn false, err\n\t\t}\n\t\tlog.Printf(\"Found existing cert. %0.2f days remaining.\", daysLeft)\n\t\tnamesOK := dnsNamesEqual(cfg.Names, names)\n\t\tif daysLeft >= float64(renewUnder) && namesOK {\n\t\t\tlog.Println(\"Nothing to do\")\n\t\t\t//nothing to do\n\t\t\treturn false, nil\n\t\t}\n\t\tif !namesOK {\n\t\t\tlog.Println(\"DNS Names don't match expected set. Reissuing.\")\n\t\t} else {\n\t\t\tlog.Println(\"Renewing cert\")\n\t\t\taction = func() (acme.CertificateResource, error) {\n\t\t\t\treturn c.client.RenewCertificate(*existing, true, true)\n\t\t\t}\n\t\t}\n\t}\n\n\tcertResource, err := action()\n\tif err != nil {\n\t\treturn false, err\n\t}\n\tfmt.Printf(\"Obtained certificate for %s\\n\", cfg.CertName)\n\treturn true, c.writeCertificate(cfg.CertName, &certResource)\n}", "func (r *record) isExpired(now time.Time) bool {\n\tif r.Expires == 0 {\n\t\treturn false\n\t}\n\texpiryDateUTC := time.Unix(r.Expires, 0).UTC()\n\treturn now.UTC().After(expiryDateUTC)\n}", "func (exp *ControlleeExpectations) isExpired() bool {\n\treturn clock.RealClock{}.Since(exp.timestamp) > ExpectationsTimeout\n}", "func (t *MongoDBToken) IsExpired() bool {\n\treturn time.Now().UTC().Unix() >= t.Expired.Unix()\n}", "func checkExpiringCerts(chain []*x509.Certificate) (expiringIntermediates []int) {\n\tnow := time.Now()\n\tfor i, cert := range chain {\n\t\tif cert.NotAfter.Sub(now).Hours() < 720 {\n\t\t\texpiringIntermediates = append(expiringIntermediates, i)\n\t\t}\n\t}\n\treturn\n}", "func (p *UserPendingPermissions) Expired(ttl time.Duration, now time.Time) bool {\n\treturn !now.Before(p.UpdatedAt.Add(ttl))\n}", "func (p *SSOCredentialProvider) IsExpired() bool {\n\tt, err := time.Parse(\"2006-01-02T15:04:05UTC\", p.Cache.ExpiresAt)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn t.Before(time.Now())\n}", "func (p *RepoPermissions) Expired(ttl time.Duration, now time.Time) bool {\n\treturn !now.Before(p.UpdatedAt.Add(ttl))\n}", "func (s *StaticProvider) IsExpired() bool {\n\treturn false\n}", "func TestRenewBeforeTTLExpires(t *testing.T) {\n\tttl := 10\n\tc, v, secret := loginHelper(t, fmt.Sprintf(\"%vs\", ttl))\n\tif secret.LeaseDuration < 2 {\n\t\tt.Fatalf(\"expected lease to be at least 2s, but was: %d\", secret.LeaseDuration)\n\t} else if secret.LeaseDuration > 10 {\n\t\tt.Fatalf(\"expected lease to be at most 10s, but was: %d\", secret.LeaseDuration)\n\t}\n\n\t_, err := c.AuthAPIClient.GetAdmins(c.Ctx(), &auth.GetAdminsRequest{})\n\tif err != nil {\n\t\tt.Fatalf(err.Error())\n\t}\n\n\trenewer, err := v.NewRenewer(&vault.RenewerInput{\n\t\tSecret: secret,\n\t\tIncrement: ttl,\n\t})\n\tif err != nil {\n\t\tt.Fatalf(err.Error())\n\t}\n\n\t// Begin a renewer background process, and wait until it fires\n\ttime.Sleep(time.Duration(ttl/2) * time.Second)\n\tgo renewer.Renew()\n\tdefer renewer.Stop()\n\tselect {\n\tcase err := <-renewer.DoneCh():\n\t\tif err != nil {\n\t\t\tt.Fatalf(err.Error())\n\t\t}\n\tcase <-renewer.RenewCh():\n\t}\n\n\t// Make sure that the vault lease was only extended by 10s\n\tleaseInfo, err := v.Logical().Write(\"/sys/leases/lookup\", map[string]interface{}{\n\t\t\"lease_id\": secret.LeaseID,\n\t})\n\tif err != nil {\n\t\tt.Fatal(err.Error())\n\t}\n\tnewDurationStr := leaseInfo.Data[\"ttl\"].(json.Number)\n\tnewDuration, err := newDurationStr.Int64()\n\tif err != nil {\n\t\tt.Fatalf(err.Error())\n\t}\n\tif newDuration < 2 {\n\t\tt.Fatalf(\"expected lease to be at least 2s, but was: %d\", newDuration)\n\t} else if newDuration > 20 {\n\t\tt.Fatalf(\"expected lease to be at most 20s, but was: %d\", newDuration)\n\t}\n\n\t// Make sure that the Pachyderm token was also renewed\n\ttime.Sleep(time.Duration(ttl/2+1) * time.Second) // wait til old lease exires\n\t_, err = c.AuthAPIClient.GetAdmins(c.Ctx(), &auth.GetAdminsRequest{})\n\tif err != nil {\n\t\tt.Fatalf(err.Error())\n\t}\n}", "func (m *VaultCredsProvider) IsExpired() bool {\n\treturn false\n}", "func (b *Object) expired() bool {\n\tif b.expire <= 0 {\n\t\treturn false\n\t}\n\n\treturn time.Now().Unix() >= b.expire\n}", "func (oc *OrdererCapabilities) ExpirationCheck() bool {\n\treturn oc.ExpirationVal\n}", "func (item *item) expired() bool {\n\tif item.ttl <= 0 {\n\t\treturn false\n\t}\n\treturn item.expireAt.Before(time.Now())\n}", "func CheckExpiry(expiryTime time.Time) bool {\n\n\texpired := expiryTime.Before(time.Now())\n\treturn expired\n\n}", "func freshOCSP(resp *ocsp.Response) bool {\n\tnextUpdate := resp.NextUpdate\n\t// If there is an OCSP responder certificate, and it expires before the\n\t// OCSP response, use its expiration date as the end of the OCSP\n\t// response's validity period.\n\tif resp.Certificate != nil && resp.Certificate.NotAfter.Before(nextUpdate) {\n\t\tnextUpdate = resp.Certificate.NotAfter\n\t}\n\t// start checking OCSP staple about halfway through validity period for good measure\n\trefreshTime := resp.ThisUpdate.Add(nextUpdate.Sub(resp.ThisUpdate) / 2)\n\treturn time.Now().Before(refreshTime)\n}", "func TestVerifyExpired(t *testing.T) {\n\toptions := iniVerifyOptions(t)\n\toptions.CurrentTime = time.Date(2020, 01, 01, 01, 01, 01, 0, gmt)\n\n\tp := loadProxy(\"test-samples/RfcProxy.pem\", t)\n\tif e := p.Verify(options); e == nil {\n\t\tt.Error(\"Verification must fail\")\n\t} else {\n\t\tt.Log(e)\n\t}\n}", "func compile(policy *cb.SignaturePolicy, identities []*mb.MSPPrincipal) (func([]msp.Identity, []bool) bool, error) {\n\tif policy == nil {\n\t\treturn nil, fmt.Errorf(\"Empty policy element\")\n\t}\n\n\tswitch t := policy.Type.(type) {\n\tcase *cb.SignaturePolicy_NOutOf_:\n\t\tpolicies := make([]func([]msp.Identity, []bool) bool, len(t.NOutOf.Rules))\n\t\tfor i, policy := range t.NOutOf.Rules {\n\t\t\tcompiledPolicy, err := compile(policy, identities)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, err\n\t\t\t}\n\t\t\tpolicies[i] = compiledPolicy\n\n\t\t}\n\t\treturn func(signedData []msp.Identity, used []bool) bool {\n\t\t\tgrepKey := time.Now().UnixNano()\n\t\t\tcauthdslLogger.Debugf(\"%p gate %d evaluation starts\", signedData, grepKey)\n\t\t\tverified := int32(0)\n\t\t\t_used := make([]bool, len(used))\n\t\t\tfor _, policy := range policies {\n\t\t\t\tcopy(_used, used)\n\t\t\t\tif policy(signedData, _used) {\n\t\t\t\t\tverified++\n\t\t\t\t\tcopy(used, _used)\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tif verified >= t.NOutOf.N {\n\t\t\t\tcauthdslLogger.Debugf(\"%p gate %d evaluation succeeds\", signedData, grepKey)\n\t\t\t} else {\n\t\t\t\tcauthdslLogger.Debugf(\"%p gate %d evaluation fails\", signedData, grepKey)\n\t\t\t}\n\n\t\t\treturn verified >= t.NOutOf.N\n\t\t}, nil\n\tcase *cb.SignaturePolicy_SignedBy:\n\t\tif t.SignedBy < 0 || t.SignedBy >= int32(len(identities)) {\n\t\t\treturn nil, fmt.Errorf(\"identity index out of range, requested %v, but identities length is %d\", t.SignedBy, len(identities))\n\t\t}\n\t\tsignedByID := identities[t.SignedBy]\n\t\treturn func(signedData []msp.Identity, used []bool) bool {\n\t\t\tcauthdslLogger.Debugf(\"%p signed by %d principal evaluation starts (used %v)\", signedData, t.SignedBy, used)\n\t\t\tfor i, sd := range signedData {\n\t\t\t\tif used[i] {\n\t\t\t\t\tcauthdslLogger.Debugf(\"%p skipping identity %d because it has already been used\", signedData, i)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tif cauthdslLogger.IsEnabledFor(zapcore.DebugLevel) {\n\t\t\t\t\t// Unlike most places, this is a huge print statement, and worth checking log level before create garbage\n\t\t\t\t\tcauthdslLogger.Debugf(\"%p processing identity %d - %v\", signedData, i, sd.GetIdentifier())\n\t\t\t\t}\n\t\t\t\terr := sd.SatisfiesPrincipal(signedByID)\n\t\t\t\tif err != nil {\n\t\t\t\t\tcauthdslLogger.Debugf(\"%p identity %d does not satisfy principal: %s\", signedData, i, err)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\tcauthdslLogger.Debugf(\"%p principal evaluation succeeds for identity %d\", signedData, i)\n\t\t\t\tused[i] = true\n\t\t\t\treturn true\n\t\t\t}\n\t\t\tcauthdslLogger.Debugf(\"%p principal evaluation fails\", signedData)\n\t\t\treturn false\n\t\t}, nil\n\tdefault:\n\t\treturn nil, fmt.Errorf(\"Unknown type: %T:%v\", t, t)\n\t}\n}", "func (tcertBlock *TCertBlock) isUpdateExpired(reusedUpdateSecond int) bool {\n\tnotAfter := tcertBlock.GetTCert().GetCertificate().NotAfter\n\ttsNow := time.Now()\n\tif reusedUpdateSecond == 0 {\n\n\t\t//default reusedUpdateSecond is one week or 1/3 of the cert life, the shorter takes\n\t\tnotBefore := tcertBlock.GetTCert().GetCertificate().NotBefore\n\t\ttimeDel := notAfter.Sub(notBefore) / 3\n\n\t\tif timeDel < time.Duration(oneweek) {\n\t\t\treusedUpdateSecond = int(timeDel)\n\t\t} else {\n\t\t\treusedUpdateSecond = int(oneweek)\n\t\t}\n\t}\n\n\tpoolLogger.Debugf(\"#isUpdateExpired now: %s deadline: %s \\n \", tsNow, notAfter)\n\n\treturn tsNow.Add(time.Duration(reusedUpdateSecond)).After(notAfter)\n\n}", "func (c ClaimsWithLeeway) Valid() error {\n\tvErr := new(jwt.ValidationError)\n\tnow := jwt.TimeFunc().Unix()\n\n\t// The claims below are optional, by default, so if they are set to the\n\t// default value in Go, let's not fail the verification for them.\n\tif c.VerifyExpiresAt(now, false) == false {\n\t\tdelta := time.Unix(now, 0).Sub(time.Unix(c.ExpiresAt, 0))\n\t\tvErr.Inner = fmt.Errorf(\"token is expired by %v\", delta)\n\t\tlog.Println(\"token is expired by \" + delta.String())\n\t\tvErr.Errors |= jwt.ValidationErrorExpired\n\t}\n\n\tif c.VerifyIssuedAt(now, false) == false {\n\t\tvErr.Inner = fmt.Errorf(\"Token used before issued\")\n\t\tlog.Println(\"Token used before issued\")\n\t\tvErr.Errors |= jwt.ValidationErrorIssuedAt\n\t}\n\n\tif c.VerifyNotBefore(now, false) == false {\n\t\tvErr.Inner = fmt.Errorf(\"token is not valid yet\")\n\t\tlog.Println(\"token is not valid yet\")\n\t\tvErr.Errors |= jwt.ValidationErrorNotValidYet\n\t}\n\n\tif vErr.Errors == 0 {\n\t\treturn nil\n\t}\n\n\treturn vErr\n}", "func (t *Tree) EffectivePolicy() *Policy {\n\tif t == nil {\n\t\treturn DefaultPolicy\n\t}\n\n\treturn t.effective\n}", "func CheckExpiration(cert *x509.Certificate, epoch time.Time) error {\n\tif cert == nil {\n\t\treturn errors.New(\"certificate is nil\")\n\t}\n\tif cert.NotAfter.Before(epoch) {\n\t\treturn fmt.Errorf(\"certificate expiration time %s is before %s\", formatTime(cert.NotAfter), formatTime(epoch))\n\t}\n\tif cert.NotBefore.After(epoch) {\n\t\treturn fmt.Errorf(\"certificate issued time %s is before %s\", formatTime(cert.NotBefore), formatTime(epoch))\n\t}\n\treturn nil\n}", "func (l *leases) keepRenewedOnce(\n\tctx context.Context, tgt lease, now time.Time,\n) (_ lease, retry bool) {\n\tremaining := tgt.expires.Sub(now)\n\t// We haven't been able to renew before hitting the guard\n\t// duration, so return and allow the lease to be canceled.\n\tif remaining < l.cfg.Guard {\n\t\treturn tgt, false\n\t}\n\t// Wait up to half of the remaining validity time before\n\t// attempting to renew, but rate-limit to the polling interval.\n\tdelay := remaining / 2\n\tif delay < l.cfg.Poll {\n\t\tdelay = l.cfg.Poll\n\t}\n\n\t// Wait until it's time to do something, or we're canceled.\n\tselect {\n\tcase <-ctx.Done():\n\t\treturn tgt, false\n\tcase <-time.After(delay):\n\t}\n\n\tvar ok bool\n\tvar err error\n\ttgt, ok, err = l.renew(ctx, tgt)\n\n\tentry := log.WithFields(log.Fields{\n\t\t\"expires\": tgt.expires, // Include renewed expiration time.\n\t\t\"lease\": tgt.name,\n\t})\n\n\tswitch {\n\tcase errors.Is(err, context.Canceled):\n\t\tentry.Trace(\"context canceled\")\n\t\treturn tgt, false\n\tcase err != nil:\n\t\tentry.WithError(err).Warn(\"could not renew lease\")\n\t\treturn tgt, true\n\tcase !ok:\n\t\tentry.Debug(\"lease was hijacked\")\n\t\treturn tgt, false\n\tdefault:\n\t\tentry.Trace(\"renewed successfully\")\n\t\treturn tgt, true\n\t}\n}", "func isExpired(filename string) bool {\n\tvar t time.Time\n\tlast := timeFromFilename(filename)\n\n\t// exp time for prices dataset is 1 day\n\tif strings.Contains(filename, datasetPrices) {\n\t\tt = last.AddDate(0, 0, 1)\n\t}\n\n\t// exp time for stations dataset is 15 days\n\tif strings.Contains(filename, datasetStations) {\n\t\tt = last.AddDate(0, 0, 15)\n\t}\n\n\treturn t.Before(time.Now())\n}", "func (l *License) Expired() bool {\n\treturn l.Exp.IsZero() == false && time.Now().After(l.Exp)\n}", "func lastWarningThresholdCrossed(host *host.Host) time.Duration {\n\n\t// how long til the host expires\n\ttilExpiration := host.ExpirationTime.Sub(time.Now())\n\n\t// iterate through the thresholds - since they are kept in sorted order,\n\t// the first one crossed will be the most recent one crossed\n\tfor _, threshold := range spawnWarningThresholds {\n\t\tif tilExpiration <= threshold {\n\t\t\treturn threshold\n\t\t}\n\t}\n\n\t// should never be reached\n\treturn time.Duration(0)\n\n}", "func (m *ProviderTerms) Expired() bool {\n\treturn m.ExpiredAt < time.Now()+TermsExpiredDuration\n}", "func domainExpiry(domain string) (time.Time, error) {\n\t// allow InsecureSkipVerify for only checking domains\n\tconn, err := tls.Dial(\"tcp\", domain, &tls.Config{InsecureSkipVerify: true})\n\tif err != nil {\n\t\treturn time.Unix(0, 0), err\n\t}\n\tdefer conn.Close()\n\t//checkedCerts := make(map[string]struct{})\n\t//certs := []string{}\n\tchains := conn.ConnectionState().PeerCertificates\n\treturn chains[0].NotAfter, nil\n}", "func (j *Service) IsExpired(claims Claims) bool {\n\treturn !claims.VerifyExpiresAt(time.Now().Unix(), true)\n}", "func LiveVerifyExpired(obj *models.LiveVerify, appUIVersion string, appVersionCode int) (expired bool) {\n\tif obj == nil {\n\t\treturn true\n\t}\n\t//兼容老版本,老版本不需要做 活体认证的有效期限制\n\tif types.IndonesiaAppRipeVersionLiveVerify > appVersionCode {\n\t\treturn false\n\t}\n\n\tinterval, _ := config.ValidItemInt(types.LiveVerifyInterval)\n\tlogs.Debug(\"[service.account.LiveVerifyExpired] config interval:\", interval,\n\t\t\" obj.Ctime:\", obj.Ctime,\n\t\t\" tools.GetUnixMillis() - obj.Ctime:\", tools.GetUnixMillis()-obj.Ctime)\n\treturn (tools.GetUnixMillis() - obj.Ctime) > int64(interval*60*1000)\n}", "func (o *OIDC) AuthorizeRenew(ctx context.Context, cert *x509.Certificate) error {\n\treturn o.ctl.AuthorizeRenew(ctx, cert)\n}", "func (o WebTestPropertiesValidationRulesOutput) SSLCertRemainingLifetimeCheck() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v WebTestPropertiesValidationRules) *int { return v.SSLCertRemainingLifetimeCheck }).(pulumi.IntPtrOutput)\n}", "func (dc *DelegatedCredential) isExpired(start, now time.Time) bool {\n\tend := start.Add(dc.cred.validTime)\n\treturn !now.Before(end)\n}", "func MaxAgePolicy() *MaxAgePolicyApplyConfiguration {\n\treturn &MaxAgePolicyApplyConfiguration{}\n}", "func (spit *Spit) RemainingExpiration() int {\r\n\treturn int(spit.DateCreatedTime().Add(time.Duration(spit.Exp)*time.Second).Unix()-\r\n\t\ttime.Now().UTC().Unix()) - 1\r\n}", "func (c *mockCredentialsProvider) IsExpired() bool {\n\treturn false\n}", "func (o WebTestPropertiesValidationRulesPtrOutput) SSLCertRemainingLifetimeCheck() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *WebTestPropertiesValidationRules) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SSLCertRemainingLifetimeCheck\n\t}).(pulumi.IntPtrOutput)\n}", "func IsExpired(targetDate time.Time, timeAdded time.Duration) bool {\n\treturn time.Since(targetDate.Add(timeAdded)) > 0\n}", "func expired(token *Token) bool {\n\tif token.Expires.IsZero() && len(token.Access) != 0 {\n\t\treturn false\n\t}\n\treturn token.Expires.Add(-expiryDelta).\n\t\tBefore(time.Now())\n}", "func (le *lessor) expireExists() (l *Lease, ok bool, next bool) {\n\tif le.leaseExpiredNotifier.Len() == 0 {\n\t\treturn nil, false, false\n\t}\n\n\titem := le.leaseExpiredNotifier.Poll()\n\tl = le.leaseMap[item.id]\n\tif l == nil {\n\t\t// lease has expired or been revoked\n\t\t// no need to revoke (nothing is expiry)\n\t\tle.leaseExpiredNotifier.Unregister() // O(log N)\n\t\treturn nil, false, true\n\t}\n\n\tnow := time.Now()\n\tif now.UnixNano() < item.time /* expiration time */ {\n\t\t// Candidate expirations are caught up, reinsert this item\n\t\t// and no need to revoke (nothing is expiry)\n\t\treturn l, false, false\n\t}\n\n\t// ??? expiredLeaseRetryInterval\n\n\t// recheck if revoke is complete after retry interval\n\titem.time = now.Add(le.expiredLeaseRetryInterval).UnixNano()\n\tle.leaseExpiredNotifier.RegisterOrUpdate(item)\n\treturn l, true, false\n}", "func (w *writer) isExpired(now time.Time) bool {\n\tif w.count == 0 {\n\t\treturn false\n\t}\n\tif w.expiryTime == nil {\n\t\treturn false\n\t}\n\treturn w.expiryTime.Before(now)\n}", "func (c *cacheExpirationPolicy) IsExpired(entry *cache.TimestampedEntry) bool {\n\treturn c.clock.Now().After(entry.Obj.(*cacheEntry).expiresAt)\n}", "func (p *Pictures) IsExpired() bool {\n\treturn time.Now().After(p.ExpiresAt)\n}", "func IsExpired(ttl uint32, timestamp uint64) bool {\n\tif ttl == Persistent {\n\t\treturn false\n\t}\n\n\tnow := time.UnixMilli(time.Now().UnixMilli())\n\texpireTime := time.UnixMilli(int64(timestamp))\n\texpireTime = expireTime.Add(time.Duration(ttl) * time.Second)\n\n\treturn expireTime.Before(now)\n}", "func (c *Cache) ExpiryChecker(ctx context.Context) {\n\ttimer := time.NewTicker(checkInterval)\n\tdefer timer.Stop()\n\n\tfor {\n\t\tselect {\n\t\tcase <-timer.C: //timer has talked to us\n\t\t\tc.removeExpiredKeys()\n\n\t\tcase <-ctx.Done():\n\t\t\treturn\n\t\t}\n\t}\n}", "func (l *Lease) Remaining() time.Duration {\n\tl.expiryMu.RLock()\n\tdefer l.expiryMu.RUnlock()\n\tif l.expiry.IsZero() { // expiry为0表示永不过期\n\t\treturn time.Duration(math.MaxInt64)\n\t}\n\n\treturn time.Until(l.expiry)\n}", "func pastActiveDeadline(job *appsv1alpha1.BroadcastJob) bool {\n\tif job.Spec.CompletionPolicy.ActiveDeadlineSeconds == nil || job.Status.StartTime == nil {\n\t\treturn false\n\t}\n\tnow := metav1.Now()\n\tstart := job.Status.StartTime.Time\n\tduration := now.Time.Sub(start)\n\tallowedDuration := time.Duration(*job.Spec.CompletionPolicy.ActiveDeadlineSeconds) * time.Second\n\treturn duration >= allowedDuration\n}", "func TestExpiry(t *testing.T) {\n\t_, privateBytes, err := GenerateKeyPair()\n\trequire.NoError(t, err)\n\tprivateKey, err := utils.ParsePrivateKey(privateBytes)\n\trequire.NoError(t, err)\n\n\tclock := clockwork.NewFakeClockAt(time.Now())\n\n\t// Create a new key that can be used to sign and verify tokens.\n\tkey, err := New(&Config{\n\t\tClock: clock,\n\t\tPrivateKey: privateKey,\n\t\tAlgorithm: defaults.ApplicationTokenAlgorithm,\n\t\tClusterName: \"example.com\",\n\t})\n\trequire.NoError(t, err)\n\n\t// Sign a token with a 1 minute expiration.\n\ttoken, err := key.Sign(SignParams{\n\t\tUsername: \"[email protected]\",\n\t\tRoles: []string{\"foo\", \"bar\"},\n\t\tTraits: wrappers.Traits{\n\t\t\t\"trait1\": []string{\"value-1\", \"value-2\"},\n\t\t},\n\t\tExpires: clock.Now().Add(1 * time.Minute),\n\t\tURI: \"http://127.0.0.1:8080\",\n\t})\n\trequire.NoError(t, err)\n\n\t// Verify that the token is still valid.\n\tclaims, err := key.Verify(VerifyParams{\n\t\tUsername: \"[email protected]\",\n\t\tURI: \"http://127.0.0.1:8080\",\n\t\tRawToken: token,\n\t})\n\trequire.NoError(t, err)\n\trequire.Equal(t, claims.Username, \"[email protected]\")\n\trequire.Equal(t, claims.Roles, []string{\"foo\", \"bar\"})\n\trequire.Equal(t, claims.IssuedAt, josejwt.NewNumericDate(clock.Now()))\n\n\t// Advance time by two minutes and verify the token is no longer valid.\n\tclock.Advance(2 * time.Minute)\n\t_, err = key.Verify(VerifyParams{\n\t\tUsername: \"[email protected]\",\n\t\tURI: \"http://127.0.0.1:8080\",\n\t\tRawToken: token,\n\t})\n\trequire.Error(t, err)\n}", "func (item Item) expired() bool {\n\tif item.Expiration == 0 {\n\t\treturn false\n\t}\n\treturn time.Now().UnixNano() > item.Expiration\n}", "func (o WebTestPropertiesResponseValidationRulesPtrOutput) SSLCertRemainingLifetimeCheck() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *WebTestPropertiesResponseValidationRules) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SSLCertRemainingLifetimeCheck\n\t}).(pulumi.IntPtrOutput)\n}", "func (o WebTestPropertiesResponseValidationRulesOutput) SSLCertRemainingLifetimeCheck() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v WebTestPropertiesResponseValidationRules) *int { return v.SSLCertRemainingLifetimeCheck }).(pulumi.IntPtrOutput)\n}", "func (y *Yaraus) ExpireAt() time.Time {\n\ty.mu.RLock()\n\tdefer y.mu.RUnlock()\n\treturn y.expireAt\n}", "func (c *certManager) RenewExpiringCertificates(before time.Duration) ([]*store.Certificate, error) {\n\tcerts, err := c.store.GetCertificates()\n\tif err != nil {\n\t\treturn nil, logger.Errore(err)\n\t}\n\n\tvar renewedCerts []*store.Certificate\n\tthreshold := time.Now().Add(before)\n\n\tfor _, cert := range certs {\n\t\tif threshold.After(cert.Expires) {\n\t\t\tnewCert, err := c.addCertificate(cert.Subject, cert.AlternativeNames, true)\n\t\t\tif err != nil {\n\t\t\t\treturn nil, logger.Errore(err)\n\t\t\t}\n\t\t\trenewedCerts = append(renewedCerts, newCert)\n\t\t}\n\t}\n\n\treturn renewedCerts, nil\n}", "func (o ServiceCertificateOutput) Expiry() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ServiceCertificate) *string { return v.Expiry }).(pulumi.StringPtrOutput)\n}", "func (c CurationRemark) OlderThanLimit(cutOff time.Time) bool {\n\treturn c.LastCompletionTimestamp.Before(cutOff)\n}", "func (client *Client) GetOnlineCSPICountEventually(cspcName, cspcNamespace string, expectedCSPICount int) int {\n\tvar cspiCount int\n\t// as cspi deletion takes more time now for cleanup of its resources\n\t// for reconciled cspi to come up it can take additional time.\n\tfor i := 0; i < (maxRetry + 100); i++ {\n\t\tcspiList, err := client.GetCSPIList(cspcName, cspcNamespace)\n\t\tExpect(err).To(BeNil())\n\t\tfilteredList := cspiList.Filter(cstor.IsOnline())\n\t\tcspiCount = len(filteredList.Items)\n\t\tif cspiCount == expectedCSPICount {\n\t\t\treturn cspiCount\n\t\t}\n\t\ttime.Sleep(3 * time.Second)\n\t}\n\treturn cspiCount\n}", "func (d *Driver) Expired() bool {\n\tif d.Expiration == 0 {\n\t\treturn false\n\t}\n\treturn time.Now().UnixNano() > d.Expiration\n}", "func (l *Limiter) Allow(ctx context.Context) (func(rate.Op), error) {\n\tvar (\n\t\tdone func(time.Time, rate.Op)\n\t\terr error\n\t\tok bool\n\t)\n\tif done, ok = l.rate.Acquire(); !ok {\n\t\t// NOTE exceed max inflight, use queue\n\t\tif err = l.queue.Push(ctx); err != nil {\n\t\t\tdone(time.Time{}, rate.Ignore)\n\t\t\treturn func(rate.Op) {}, err\n\t\t}\n\t}\n\tstart := time.Now()\n\treturn func(op rate.Op) {\n\t\tdone(start, op)\n\t\tl.queue.Pop()\n\t}, nil\n}", "func newRule(defaultExpiration time.Duration, allowed int, estimated ...int) *singleRule {\n\tif allowed <= 0 {\n\t\tallowed = 1\n\t}\n\tuserEstimated := 0\n\tif len(estimated) > 0 {\n\t\tuserEstimated = estimated[0]\n\t}\n\tif userEstimated <= 0 {\n\t\tuserEstimated = allowed\n\t}\n\tcleanupInterval := defaultExpiration / 100\n\tif cleanupInterval < time.Second*1 {\n\t\tcleanupInterval = time.Second * 1\n\t}\n\tif cleanupInterval > time.Second*60 {\n\t\tcleanupInterval = time.Second * 60\n\t}\n\tvc := createRule(defaultExpiration, cleanupInterval, allowed, userEstimated)\n\tgo vc.deleteExpired()\n\treturn vc\n}", "func (c Choco) Expired() bool {\n\treturn time.Since(c.TimeStamp) > time.Second\n}" ]
[ "0.5933611", "0.5186232", "0.50210214", "0.49988246", "0.4949053", "0.4882632", "0.4854657", "0.47842163", "0.47803733", "0.47186136", "0.46823356", "0.46754807", "0.4669454", "0.4652836", "0.46410698", "0.46317783", "0.46216932", "0.46113858", "0.45738754", "0.4567185", "0.4542057", "0.4541848", "0.45337832", "0.4532199", "0.44756466", "0.44535738", "0.44508383", "0.4448246", "0.4432546", "0.44197634", "0.4419364", "0.4407285", "0.44006845", "0.43987834", "0.43824276", "0.4378349", "0.437828", "0.43703464", "0.43682346", "0.43674", "0.4362335", "0.43602717", "0.4342553", "0.43402088", "0.4319392", "0.4309578", "0.43085536", "0.4307154", "0.43053582", "0.4295379", "0.42888814", "0.42838553", "0.428177", "0.42816722", "0.42790237", "0.4278932", "0.4277543", "0.4262784", "0.4256516", "0.42539287", "0.42491096", "0.42454866", "0.42418087", "0.42411354", "0.42278618", "0.422324", "0.4222561", "0.42104334", "0.42058882", "0.4199122", "0.41954824", "0.4194101", "0.41939363", "0.4192011", "0.41918048", "0.41884568", "0.41684133", "0.4159898", "0.4156376", "0.41523504", "0.41461045", "0.41384962", "0.4134136", "0.41192377", "0.4116994", "0.41094354", "0.41012093", "0.40948224", "0.40880567", "0.40823728", "0.40808946", "0.40796483", "0.40676233", "0.40673065", "0.4058796", "0.4057283", "0.40493074", "0.40451312", "0.40387377", "0.40382856" ]
0.7937122
0
CurrentCertificateHasExpired is used exclusively to check if the current issued certificate has actually expired rather than just nearing expiry.
func CurrentCertificateHasExpired(c clock.Clock) Func { return func(input Input) (string, string, bool) { certData, ok := input.Secret.Data[corev1.TLSCertKey] if !ok { return MissingData, "Missing Certificate data", true } // TODO: replace this with a generic decoder that can handle different // formats such as JKS, P12 etc (i.e. add proper support for keystores) cert, err := pki.DecodeX509CertificateBytes(certData) if err != nil { // This case should never happen as it should always be caught by the // secretPublicKeysMatch function beforehand, but handle it just in case. return "InvalidCertificate", fmt.Sprintf("Failed to decode stored certificate: %v", err), true } if c.Now().After(cert.NotAfter) { return Expired, fmt.Sprintf("Certificate expired on %s", cert.NotAfter.Format(time.RFC1123)), true } return "", "", false } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *CIDOffer) HasExpired() bool {\n\texpiryTime := time.Unix(c.expiry, 0)\n\tnow := time.Now()\n\treturn expiryTime.Before(now)\n}", "func (m *Attachment) HasExpired() bool {\n\tvar validTime = m.SigningTime.Add(time.Duration(m.SigningMinutes) * time.Minute)\n\treturn validTime.Unix() < time.Now().Unix()\n}", "func (b *ProviderBasis) IsExpired() bool {\n\tif b.CurrentTime == nil {\n\t\tb.CurrentTime = time.Now\n\t}\n\treturn !b.AlwaysValid && !b.CurrentTime().Before(b.expiration)\n}", "func (request *AccessToken) HasExpired() bool {\n\tcurrentTime := time.Now().Unix()\n\treturn currentTime > request.ExpiresAt\n}", "func (e *EnvProvider) IsExpired() bool {\n\treturn !e.retrieved\n}", "func (tcertBlock *TCertBlock) isExpired() bool {\n\ttsNow := time.Now()\n\tnotAfter := tcertBlock.GetTCert().GetCertificate().NotAfter\n\tpoolLogger.Debugf(\"#isExpired: %s now: %s deadline: %s \\n \", tsNow.Add(fivemin).After(notAfter), tsNow, notAfter)\n\tif tsNow.Add(fivemin).After(notAfter) {\n\t\treturn true\n\t}\n\treturn false\n}", "func (m *VaultCredsProvider) IsExpired() bool {\n\treturn false\n}", "func (c Certificate) Expired(now time.Time, skew time.Duration) bool {\n\treturn c.IssuedAt.After(now) || c.ExpiresAt.Before(now)\n}", "func (s *subscription) IsExpired() bool {\n\treturn s.ExpiresAt.Before(time.Now())\n}", "func HasExpired(dev *schemas.Developer) bool {\n\t// null time or before now\n\treturn dev.Expiration.Equal(time.Time{}) || dev.Expiration.Before(time.Now())\n}", "func (c CachedObject) IsExpired() bool {\r\n\r\n\telapsed := time.Now().Sub(c.CreatedAt.Add(time.Hour * getExpiryTimeInHrs()))\r\n\r\n\tif elapsed > 0.0 {\r\n\t\treturn true\r\n\t}\r\n\r\n\treturn false\r\n}", "func (c *mockCredentialsProvider) IsExpired() bool {\n\treturn false\n}", "func (j *Service) IsExpired(claims Claims) bool {\n\treturn !claims.VerifyExpiresAt(time.Now().Unix(), true)\n}", "func (i *Info) expired(now int64) bool {\n\treturn i.TTLStamp <= now\n}", "func (i *info) expired(now int64) bool {\n\treturn i.TTLStamp <= now\n}", "func IsExpired(targetDate time.Time, timeAdded time.Duration) bool {\n\treturn time.Since(targetDate.Add(timeAdded)) > 0\n}", "func (t *MongoDBToken) IsExpired() bool {\n\treturn time.Now().UTC().Unix() >= t.Expired.Unix()\n}", "func (c *Cache) IsExpired(key string, current int64) bool {\r\n\treturn (current > c.MagicKeys[key].Expiration)\r\n}", "func (exp *ControlleeExpectations) isExpired() bool {\n\treturn clock.RealClock{}.Since(exp.timestamp) > ExpectationsTimeout\n}", "func (p *SSOCredentialProvider) IsExpired() bool {\n\tt, err := time.Parse(\"2006-01-02T15:04:05UTC\", p.Cache.ExpiresAt)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\treturn t.Before(time.Now())\n}", "func (dc *DelegatedCredential) isExpired(start, now time.Time) bool {\n\tend := start.Add(dc.cred.validTime)\n\treturn !now.Before(end)\n}", "func (t *BcsUser) HasExpired() bool {\n\tif time.Now().After(t.ExpiresAt) {\n\t\treturn true\n\t}\n\treturn false\n}", "func (s *StaticProvider) IsExpired() bool {\n\treturn false\n}", "func (card Card) IsExpired() bool {\n\treturn time.Now().After(card.ExpireAt)\n}", "func (b *Object) expired() bool {\n\tif b.expire <= 0 {\n\t\treturn false\n\t}\n\n\treturn time.Now().Unix() >= b.expire\n}", "func (r *record) isExpired(now time.Time) bool {\n\tif r.Expires == 0 {\n\t\treturn false\n\t}\n\texpiryDateUTC := time.Unix(r.Expires, 0).UTC()\n\treturn now.UTC().After(expiryDateUTC)\n}", "func CurrentCertificateNearingExpiry(c clock.Clock) Func {\n\n\treturn func(input Input) (string, string, bool) {\n\n\t\t// Determine if the certificate is nearing expiry solely by looking at\n\t\t// the actual cert, if it exists. We assume that at this point we have\n\t\t// called policy functions that check that input.Secret and\n\t\t// input.Secret.Data exists (SecretDoesNotExist and SecretIsMissingData).\n\t\tx509cert, err := pki.DecodeX509CertificateBytes(input.Secret.Data[corev1.TLSCertKey])\n\t\tif err != nil {\n\t\t\t// This case should never happen as it should always be caught by the\n\t\t\t// secretPublicKeysMatch function beforehand, but handle it just in case.\n\t\t\treturn \"InvalidCertificate\", fmt.Sprintf(\"Failed to decode stored certificate: %v\", err), true\n\t\t}\n\n\t\tnotBefore := metav1.NewTime(x509cert.NotBefore)\n\t\tnotAfter := metav1.NewTime(x509cert.NotAfter)\n\t\tcrt := input.Certificate\n\t\trenewalTime := certificates.RenewalTime(notBefore.Time, notAfter.Time, crt.Spec.RenewBefore)\n\n\t\trenewIn := renewalTime.Time.Sub(c.Now())\n\t\tif renewIn > 0 {\n\t\t\t//renewal time is in future, no need to renew\n\t\t\treturn \"\", \"\", false\n\t\t}\n\n\t\treturn Renewing, fmt.Sprintf(\"Renewing certificate as renewal was scheduled at %s\", input.Certificate.Status.RenewalTime), true\n\t}\n}", "func (certList *CertificateList) ExpiredAt(now time.Time) bool {\n\treturn now.After(certList.TBSCertList.NextUpdate)\n}", "func (l Info) IsExpired() bool {\n\treturn l.ExpiresAt.Before(time.Now())\n}", "func (a *AssumeRoleProvider) IsExpired() bool {\n\treturn a.expiration.Before(time.Now())\n}", "func HasExpired(cloudEvent map[string]interface{}) bool {\n\te, ok := cloudEvent[ExpirationField]\n\tif ok && e != \"\" {\n\t\texpiration, err := time.Parse(time.RFC3339, fmt.Sprintf(\"%s\", e))\n\t\tif err != nil {\n\t\t\treturn false\n\t\t}\n\n\t\treturn expiration.UTC().Before(time.Now().UTC())\n\t}\n\n\treturn false\n}", "func (item *item) expired() bool {\n\tif item.ttl <= 0 {\n\t\treturn false\n\t}\n\treturn item.expireAt.Before(time.Now())\n}", "func expired(token *Token) bool {\n\tif token.Expires.IsZero() && len(token.Access) != 0 {\n\t\treturn false\n\t}\n\treturn token.Expires.Add(-expiryDelta).\n\t\tBefore(time.Now())\n}", "func (t *Token) IsExpired() bool {\n\tif t.Expiry.IsZero() {\n\t\treturn false\n\t}\n\treturn t.Expiry.Round(0).Add(-expiryDelta).Before(timeNow())\n}", "func (r IABResponse) IsExpired() bool {\n\tswitch {\n\tcase !r.IsValidSubscription():\n\t\treturn false\n\tdefault:\n\t\tnow := time.Now().UnixNano() / int64(time.Millisecond)\n\t\treturn r.SubscriptionPurchase.ExpiryTimeMillis < now\n\t}\n}", "func TestVerifyExpired(t *testing.T) {\n\toptions := iniVerifyOptions(t)\n\toptions.CurrentTime = time.Date(2020, 01, 01, 01, 01, 01, 0, gmt)\n\n\tp := loadProxy(\"test-samples/RfcProxy.pem\", t)\n\tif e := p.Verify(options); e == nil {\n\t\tt.Error(\"Verification must fail\")\n\t} else {\n\t\tt.Log(e)\n\t}\n}", "func (v *VerificationCode) IsExpired() bool {\n\tnow := time.Now().UTC()\n\treturn v.ExpiresAt.Before(now) && v.LongExpiresAt.Before(now)\n}", "func (d *AccessData) IsExpired() bool {\n\treturn d.IsExpiredAt(time.Now())\n}", "func (r *Record) IsExpired() bool {\n\treturn IsExpired(r.H.Meta.TTL, r.H.Meta.Timestamp)\n}", "func IsExpired(ctx Context, t UnixTime) bool {\n\tblockNow, err := BlockTime(ctx)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"%+v\", err))\n\t}\n\treturn t <= AsUnixTime(blockNow)\n}", "func (tv *TimedValue) IsExpired() bool {\n\treturn tv.IsExpiredAt(time.Now())\n}", "func isExpired(cli *clientv3.Client, ev *clientv3.Event) (bool, error) {\n\tif ev.PrevKv == nil {\n\t\treturn false, nil\n\t}\n\n\tleaseID := clientv3.LeaseID(ev.PrevKv.Lease)\n\tif leaseID == clientv3.NoLease {\n\t\treturn false, nil\n\t}\n\n\tttlResponse, err := cli.TimeToLive(context.Background(), leaseID)\n\tif err != nil {\n\t\treturn false, err\n\t}\n\n\treturn ttlResponse.TTL == -1, nil\n}", "func (err *ValidationError) IsExpired() bool { return err.exp }", "func (t *token) IsExpired() bool {\n\tif t == nil {\n\t\treturn true\n\t}\n\treturn t.Expired()\n}", "func IsContextDeadlineExpired(err error) bool {\n\terr = errs.Cause(err)\n\tif err == context.DeadlineExceeded {\n\t\treturn true\n\t}\n\tif ok, err := libCause(err); ok {\n\t\treturn IsContextDeadlineExpired(err)\n\t}\n\treturn false\n}", "func (cc *CurrentConditions) Expired() bool {\n\tnow := time.Now()\n\texpired := now.After(cc.RequestTimestamp.Add(UpdateInterval))\n\treturn expired\n}", "func (item Item) expired() bool {\n\tif item.Expiration == 0 {\n\t\treturn false\n\t}\n\treturn time.Now().UnixNano() > item.Expiration\n}", "func (p *peerAddr) isExpired(timeout time.Duration, curTime time.Time) bool {\n\treturn curTime.Sub(p.lastPing.Value.(time.Time)) >= timeout\n}", "func (p *Pictures) IsExpired() bool {\n\treturn time.Now().After(p.ExpiresAt)\n}", "func (q *queueData) expired() bool {\n\treturn q.ExpireAt < time.Now().Unix()\n}", "func LiveVerifyExpired(obj *models.LiveVerify, appUIVersion string, appVersionCode int) (expired bool) {\n\tif obj == nil {\n\t\treturn true\n\t}\n\t//兼容老版本,老版本不需要做 活体认证的有效期限制\n\tif types.IndonesiaAppRipeVersionLiveVerify > appVersionCode {\n\t\treturn false\n\t}\n\n\tinterval, _ := config.ValidItemInt(types.LiveVerifyInterval)\n\tlogs.Debug(\"[service.account.LiveVerifyExpired] config interval:\", interval,\n\t\t\" obj.Ctime:\", obj.Ctime,\n\t\t\" tools.GetUnixMillis() - obj.Ctime:\", tools.GetUnixMillis()-obj.Ctime)\n\treturn (tools.GetUnixMillis() - obj.Ctime) > int64(interval*60*1000)\n}", "func leaseExpired(grantedAt time.Time) bool {\n\treturn time.Since(grantedAt).Seconds() > (storagerpc.LeaseSeconds + storagerpc.LeaseGuardSeconds)\n}", "func (c *CSRFStore) expired() bool {\n\treturn c.token == nil || time.Now().After(c.token.ExpiresAt)\n\n}", "func (s *Session) IsExpired() bool {\n\treturn s.Expires.Before(time.Now())\n}", "func (upload *Upload) IsExpired() bool {\n\tif upload.ExpireAt != nil {\n\t\tif time.Now().After(*upload.ExpireAt) {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (item Item) isExpired() bool {\n\tif item.Expiration == 0 {\n\t\treturn false\n\t}\n\treturn item.Expiration < time.Now().UnixNano()\n}", "func (s *Static) IsExpired() bool {\n\treturn false\n}", "func isLeaseExpired(lease *db.Lease, context *leaseContext, actualPrincipalSpend float64, principalBudgetAmount float64) (bool, db.LeaseStatusReason) {\n\n\tif context.expireDate >= lease.ExpiresOn {\n\t\treturn true, db.LeaseExpired\n\t} else if context.actualSpend > lease.BudgetAmount {\n\t\treturn true, db.LeaseOverBudget\n\t} else if actualPrincipalSpend > principalBudgetAmount {\n\t\treturn true, db.LeaseOverPrincipalBudget\n\t}\n\n\treturn false, db.LeaseActive\n}", "func (k Key) IsExpired() bool {\n\texpiry := k.Expires()\n\tif expiry.Equal(timeZero) {\n\t\treturn false\n\t}\n\n\treturn expiry.Before(time.Now().UTC())\n}", "func (s *Session) IsExpired() bool {\n\treturn s.ExpiredAt.Before(time.Now())\n}", "func (w *writer) isExpired(now time.Time) bool {\n\tif w.count == 0 {\n\t\treturn false\n\t}\n\tif w.expiryTime == nil {\n\t\treturn false\n\t}\n\treturn w.expiryTime.Before(now)\n}", "func TestCertificateNotExpired(t *testing.T) {\n\t// given\n\tvar expiredDate = time.Now().Add(time.Hour * 24 * (30 + 1)) // 31 days.\n\n\tvar fakeCerts = []*x509.Certificate{\n\t\t{\n\t\t\tNotAfter: expiredDate,\n\t\t\tSubject: pkix.Name{\n\t\t\t\tCommonName: \"Test cert\",\n\t\t\t},\n\t\t},\n\t}\n\n\t// when\n\tmsg := getCertificateChainMsg(fakeCerts)\n\n\t// then\n\tif msg != \"\" {\n\t\tt.Fatalf(\"Expected empty message was: %s\", msg)\n\t}\n}", "func (v value) expired(c *Cache) bool{\n return time.Since(v.time)>c.expire\n}", "func (tb *timerBuilder) IsTimerExpired(td *timerDetails, referenceTime time.Time) bool {\n\t// Cql timestamp is in milli sec resolution, here we do the check in terms of second resolution.\n\texpiry := td.TimerSequenceID.VisibilityTimestamp.Unix()\n\treturn expiry <= referenceTime.Unix()\n}", "func (c *cacheExpirationPolicy) IsExpired(entry *cache.TimestampedEntry) bool {\n\treturn c.clock.Now().After(entry.Obj.(*cacheEntry).expiresAt)\n}", "func (s *CacheService) IsExpired(base string) bool {\n\t// No cache? bail.\n\tif cache[base] == nil || (len(cache[base].Rates) <= 0) {\n\t\treturn true\n\t}\n\n\t// Expired cache? bail.\n\tlastUpdated := cache[base].UpdatedAt\n\tif lastUpdated != nil && lastUpdated.Add(cacheTTL).Before(time.Now()) {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (r *OperationReqReconciler) isExpired(request *userv1.Operationrequest) bool {\n\tif request.Status.Phase != userv1.RequestCompleted && request.CreationTimestamp.Add(r.expirationTime).Before(time.Now()) {\n\t\tr.Logger.Info(\"operation request is expired\", \"name\", request.Name)\n\t\treturn true\n\t}\n\treturn false\n}", "func (p *Permission) IsExpired() (bool, error) {\n\tcreatedAt, err := time.Parse(time.RFC3339, p.CreatedAt)\n\tif err != nil {\n\t\treturn false, fmt.Errorf(\"invalid time format for field \\\"created_at\\\": %s\", p.CreatedAt)\n\t}\n\treturn p.TTL >= 0 && time.Since(createdAt) > p.TTL, nil\n}", "func (task *Task) IsExpired() bool {\n\tswitch task.Schedule.Regularity {\n\tcase apiModels.OneTime, apiModels.Trigger:\n\t\treturn common.ValidTime(time.Now().UTC(), task.RunTimeUTC)\n\tcase apiModels.Recurrent:\n\t\treturn !common.ValidTime(task.Schedule.EndRunTime.UTC(), task.RunTimeUTC)\n\t}\n\treturn true\n}", "func CertificatesExpired(opts Options) error {\n\tif opts.Signer.Secret == nil || opts.CABundle == nil {\n\t\treturn nil\n\t}\n\n\tfor _, cert := range opts.Certificates {\n\t\tif cert.Secret == nil {\n\t\t\treturn nil\n\t\t}\n\t}\n\n\trawCA, err := crypto.GetCAFromBytes(opts.Signer.Secret.Data[corev1.TLSCertKey], opts.Signer.Secret.Data[corev1.TLSPrivateKeyKey])\n\tif err != nil {\n\t\treturn kverrors.Wrap(err, \"failed to get signing CA from secret\")\n\t}\n\n\tcaBundle := opts.CABundle.Data[CAFile]\n\tcaCerts, err := crypto.CertsFromPEM([]byte(caBundle))\n\tif err != nil {\n\t\treturn kverrors.Wrap(err, \"failed to get ca bundle certificates from configmap\")\n\t}\n\n\tvar reasons []string\n\tfor name, cert := range opts.Certificates {\n\t\treason := cert.Rotation.NeedNewCertificate(cert.Secret.Annotations, rawCA, caCerts, opts.Rotation.TargetCertRefresh)\n\t\tif reason != \"\" {\n\t\t\treasons = append(reasons, fmt.Sprintf(\"%s: %s\", name, reason))\n\t\t}\n\t}\n\n\tif len(reasons) == 0 {\n\t\treturn nil\n\t}\n\n\treturn &CertExpiredError{Message: \"certificates expired\", Reasons: reasons}\n}", "func (c Choco) Expired() bool {\n\treturn time.Since(c.TimeStamp) > time.Second\n}", "func (provider *AdhocProvider) isStale(certificate *tls.Certificate) bool {\n\tttlOffset := provider.TTLOffset\n\tif ttlOffset == 0 {\n\t\tttlOffset = DefaultTTLOffset\n\t}\n\n\texpiresAt := certificate.Leaf.NotAfter.Add(ttlOffset)\n\n\treturn time.Now().After(expiresAt)\n}", "func isExpired(expiryBytes []byte) bool {\n\texpiry := &time.Time{}\n\terr := expiry.UnmarshalText(expiryBytes)\n\tif err != nil {\n\t\treturn true\n\t}\n\treturn time.Now().After(*expiry)\n}", "func (t *Token) Expired() bool {\n\treturn time.Now().Unix() >= t.ExpiredAt\n}", "func isTokenExpired(jwtData *JWTData) bool {\n\n\tnowTime := time.Now().Unix()\n\texpireTime := int64(jwtData.Exp)\n\n\tif expireTime < nowTime {\n\t\tlog.Warnf(\"Token is expired!\")\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (o *AccessKey) HasExpirationDate() bool {\n\tif o != nil && o.ExpirationDate != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (i *Item) IsExpired() bool {\n\t//zero means never expire\n\tif i.ExpiresAt.IsZero() {\n\t\treturn false\n\t}\n\treturn time.Now().After(i.ExpiresAt)\n}", "func (i *Item) IsExpired() bool {\n\t//zero means never expire\n\tif i.ExpiresAt.IsZero() {\n\t\treturn false\n\t}\n\treturn time.Now().After(i.ExpiresAt)\n}", "func (t *TombstoneGC) PendingExpiration() bool {\n\tt.lock.Lock()\n\tdefer t.lock.Unlock()\n\treturn len(t.expires) > 0\n}", "func (l Info) IsExpiredWithGracePeriod() bool {\n\treturn l.ExpiresAt.Add(3 * 24 * time.Hour).Before(time.Now())\n}", "func (s *StoreService) IsExpired(hash string) bool {\n\tif _, found := s.store.Get(hash); found {\n\t\treturn false\n\t}\n\treturn true\n}", "func (o *LoyaltySubLedger) HasExpiredPoints() bool {\n\tif o != nil && o.ExpiredPoints != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func CheckExpiration(cert *x509.Certificate, epoch time.Time) error {\n\tif cert == nil {\n\t\treturn errors.New(\"certificate is nil\")\n\t}\n\tif cert.NotAfter.Before(epoch) {\n\t\treturn fmt.Errorf(\"certificate expiration time %s is before %s\", formatTime(cert.NotAfter), formatTime(epoch))\n\t}\n\tif cert.NotBefore.After(epoch) {\n\t\treturn fmt.Errorf(\"certificate issued time %s is before %s\", formatTime(cert.NotBefore), formatTime(epoch))\n\t}\n\treturn nil\n}", "func isExpired(filename string) bool {\n\tvar t time.Time\n\tlast := timeFromFilename(filename)\n\n\t// exp time for prices dataset is 1 day\n\tif strings.Contains(filename, datasetPrices) {\n\t\tt = last.AddDate(0, 0, 1)\n\t}\n\n\t// exp time for stations dataset is 15 days\n\tif strings.Contains(filename, datasetStations) {\n\t\tt = last.AddDate(0, 0, 15)\n\t}\n\n\treturn t.Before(time.Now())\n}", "func IsContextCanceledOrExpired(err error) bool {\n\terr = errs.Cause(err)\n\tif err == context.Canceled || err == context.DeadlineExceeded {\n\t\treturn true\n\t}\n\tif ok, err := libCause(err); ok {\n\t\treturn IsContextCanceledOrExpired(err)\n\t}\n\treturn false\n}", "func IsExpired(ttl uint32, timestamp uint64) bool {\n\tif ttl == Persistent {\n\t\treturn false\n\t}\n\n\tnow := time.UnixMilli(time.Now().UnixMilli())\n\texpireTime := time.UnixMilli(int64(timestamp))\n\texpireTime = expireTime.Add(time.Duration(ttl) * time.Second)\n\n\treturn expireTime.Before(now)\n}", "func (this *SmtpWorker) TimeoutHasExpired(startTime time.Time) bool {\n\treturn int(time.Since(startTime).Seconds()) > smtpconstants.COMMAND_TIMEOUT_SECONDS\n}", "func (t TToken) checkExpired() error {\n\texp, err := t.getExpiry()\n\tif err != nil {\n\t\treturn err\n\t}\n\tif secondsPassed := time.Since(exp).Seconds(); secondsPassed > 30 {\n\t\treturn errors.New(\"token expired more than 30 seconds ago (#) \")\n\t}\n\treturn nil\n}", "func (m *Manager) isTokenExpired(token *Token) bool {\n\tif !m.bearerAuth {\n\t\treturn false\n\t}\n\tunixTime := time.Now().Unix()\n\treturn token.Expires < unixTime\n}", "func (o *OAuth2Config) IsExpired() bool {\n\treturn o.Token == nil || !o.Token.Valid()\n}", "func (l *License) Expired() bool {\n\treturn l.Exp.IsZero() == false && time.Now().After(l.Exp)\n}", "func (d *VirtualMachineController) hasGracePeriodExpired(dom *api.Domain) (hasExpired bool, timeLeft int) {\n\n\thasExpired = false\n\ttimeLeft = 0\n\n\tif dom == nil {\n\t\thasExpired = true\n\t\treturn\n\t}\n\n\tstartTime := int64(0)\n\tif dom.Spec.Metadata.KubeVirt.GracePeriod.DeletionTimestamp != nil {\n\t\tstartTime = dom.Spec.Metadata.KubeVirt.GracePeriod.DeletionTimestamp.UTC().Unix()\n\t}\n\tgracePeriod := dom.Spec.Metadata.KubeVirt.GracePeriod.DeletionGracePeriodSeconds\n\n\t// If gracePeriod == 0, then there will be no startTime set, deletion\n\t// should occur immediately during shutdown.\n\tif gracePeriod == 0 {\n\t\thasExpired = true\n\t\treturn\n\t} else if startTime == 0 {\n\t\t// If gracePeriod > 0, then the shutdown signal needs to be sent\n\t\t// and the gracePeriod start time needs to be set.\n\t\ttimeLeft = -1\n\t\treturn\n\t}\n\n\tnow := time.Now().UTC().Unix()\n\tdiff := now - startTime\n\n\tif diff >= gracePeriod {\n\t\thasExpired = true\n\t\treturn\n\t}\n\n\ttimeLeft = int(gracePeriod - diff)\n\tif timeLeft < 1 {\n\t\ttimeLeft = 1\n\t}\n\treturn\n}", "func (a *ACLToken) IsExpired(t time.Time) bool {\n\n\t// Check the token has an expiration time before potentially modifying the\n\t// supplied time. This allows us to avoid extra work, if it isn't needed.\n\tif !a.HasExpirationTime() {\n\t\treturn false\n\t}\n\n\t// Check and ensure the time location is set to UTC. This is vital for\n\t// consistency with multi-region global tokens.\n\tif t.Location() != time.UTC {\n\t\tt = t.UTC()\n\t}\n\n\treturn a.ExpirationTime.Before(t) || t.IsZero()\n}", "func (s *Store) HasExpirationChanged() bool {\n\ts.lock.RLock()\n\texpirationChanged := s.expirationChanged\n\ts.lock.RUnlock()\n\treturn expirationChanged\n}", "func (d *Result) HasExpiration() bool {\n\treturn d.withExpiration\n}", "func (d *Driver) Expired() bool {\n\tif d.Expiration == 0 {\n\t\treturn false\n\t}\n\treturn time.Now().UnixNano() > d.Expiration\n}", "func (m *ProviderTerms) Expired() bool {\n\treturn m.ExpiredAt < time.Now()+TermsExpiredDuration\n}", "func (o *MuteFindingRequestProperties) HasExpirationDate() bool {\n\treturn o != nil && o.ExpirationDate != nil\n}", "func (o *MuteFindingResponseProperties) HasExpirationDate() bool {\n\treturn o != nil && o.ExpirationDate != nil\n}", "func (l *PersistableLedger) expired(entryTs int64) bool {\n\texpirationTime := time.Now().Unix() - l.expiration\n\treturn entryTs < expirationTime\n}" ]
[ "0.7141972", "0.6887886", "0.6805477", "0.67659885", "0.6743062", "0.67240334", "0.6720572", "0.66714543", "0.663431", "0.66193986", "0.656803", "0.65591764", "0.65590584", "0.65142214", "0.6497867", "0.6490132", "0.6462192", "0.6421913", "0.6413423", "0.6400249", "0.6380333", "0.6363089", "0.6352595", "0.63255405", "0.6323582", "0.6322076", "0.6320266", "0.6316144", "0.6287515", "0.62740207", "0.6258761", "0.6256133", "0.625485", "0.62379843", "0.6220088", "0.6213299", "0.6198158", "0.6194594", "0.6169379", "0.6155921", "0.6151304", "0.61512446", "0.61495334", "0.6135663", "0.6131489", "0.612075", "0.6104499", "0.61044973", "0.60966426", "0.608702", "0.6085452", "0.6072462", "0.6068463", "0.6067045", "0.6044816", "0.60433865", "0.6034574", "0.6023423", "0.5999189", "0.59973985", "0.59855473", "0.5953178", "0.59393144", "0.5911688", "0.5901637", "0.58863914", "0.58805114", "0.5868117", "0.58665", "0.5862869", "0.58406734", "0.5823394", "0.58151", "0.58147293", "0.5808677", "0.5782894", "0.57702494", "0.57702494", "0.5769361", "0.57277113", "0.57241017", "0.57226294", "0.5715314", "0.5711276", "0.56882095", "0.56460625", "0.5634874", "0.56279975", "0.5625839", "0.56188077", "0.56044346", "0.559976", "0.5598165", "0.5592506", "0.55874467", "0.55836314", "0.55618334", "0.5556701", "0.5532034", "0.5521813" ]
0.7846872
0
NewRestHandler create new rest handler
func NewRestHandler(mw interfaces.Middleware, uc usecase.Usecase, validator interfaces.Validator) *RestHandler { return &RestHandler{ mw: mw, uc: uc, validator: validator, } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func NewRestHandler(mw interfaces.Middleware, uc usecase.BotUsecase) *RestHandler {\n\treturn &RestHandler{\n\t\tmw: mw,\n\t\tuc: uc,\n\t}\n}", "func NewRestHandler(mw interfaces.Middleware) *RestHandler {\n\treturn &RestHandler{\n\t\tmw: mw,\n\t}\n}", "func NewRestHandler(opts *RestHandlerOptions) *RestHandler {\n\treturn &RestHandler{\n\t\teventService: opts.EventService,\n\t\tlogger: opts.Logger,\n\t}\n}", "func NewRestHandler(cipher crypt.Cipher, backend proto.Job, secret string) http.Handler {\n\thandler := RestHandler{cipher, backend}\n\n\tmux := mux.NewRouter()\n\tjwtsec := middleware.JWTSecure(secret)\n\tmux.Path(\"/jobs\").Methods(http.MethodPost).Handler(jwtsec(handler.Create))\n\tmux.Path(\"/jobs\").Methods(http.MethodGet).Handler(jwtsec(handler.List))\n\tmux.Path(\"/jobs/{jobid}/activate\").Methods(http.MethodPost).Handler(jwtsec(handler.Activate))\n\tmux.Path(\"/category/{categoryid}\").Methods(http.MethodGet).Handler(http.HandlerFunc(handler.Percentage))\n\treturn mux\n}", "func New(service storage.Service) http.Handler {\n\tr := mux.NewRouter()\n\n\th := handler{service}\n\n\tr.HandleFunc(\"/health\", health)\n\tr.HandleFunc(\"/insfile\", responseJSONWithModules(h.insfileWithModules)).Methods(http.MethodPost)\n\tr.HandleFunc(\"/insfile/text\", responseTextWithModules(h.insfileWithModules)).Methods(http.MethodPost)\n\tr.HandleFunc(\"/insfile/traverse\", responseJSON(h.insfile)).Methods(http.MethodPost)\n\tr.HandleFunc(\"/insfile/traverse/text\", responseText(h.insfile)).Methods(http.MethodPost)\n\n\treturn r\n}", "func NewHandler(chatServer *chat.Server, apiDocPath string) http.Handler {\n\tchatServerInstance = chatServer\n\n\thandler := rest.NewHTTPHandler()\n\tregisterPaths(handler)\n\tif apiDocPath != \"\" {\n\t\trest.ConfigureSwagger(apiDocPath, handler)\n\t}\n\treturn handler\n}", "func (s *Server) newHandler() http.Handler {\n\tr := mux.NewRouter()\n\tr.HandleFunc(\"/register\", s.wrapMiddleware(registerHandler)).Methods(\"POST\")\n\tr.HandleFunc(\"/session/{id}\", s.wrapMiddleware(getHandler)).Methods(\"GET\")\n\tr.HandleFunc(\"/session\", s.wrapMiddleware(createHandler)).Methods(\"POST\")\n\tr.HandleFunc(\"/readiness\", predis.NewReadinessCheck(s.pool))\n\n\treturn r\n}", "func CreateRouter(handlerFunc http.HandlerFunc) *mux.Router {\n router := mux.NewRouter()\n\n config := dots_config.GetSystemConfig()\n prefixPath := config.ClientRestfulApiConfiguration.RestfulApiPath\n\n restfulHandlerFunc := createRestfulHandlerFunc(handlerFunc)\n\n // router.HandleFunc(\"/test\", restfulHandlerFunc).Methods(\"GET\")\n router.HandleFunc(prefixPath + MITIGATION_PATH + \"/cuid={cuid}\", restfulHandlerFunc).Methods(\"GET\")\n router.HandleFunc(prefixPath + MITIGATION_PATH + \"/cuid={cuid}/mid={mid}\", restfulHandlerFunc).Methods(\"GET\")\n router.HandleFunc(prefixPath + MITIGATION_PATH + \"/cuid={cuid}/mid={mid}\", restfulHandlerFunc).Methods(\"PUT\")\n router.HandleFunc(prefixPath + MITIGATION_PATH + \"/cuid={cuid}/mid={mid}\", restfulHandlerFunc).Methods(\"DELETE\")\n\n return router\n}", "func New(o *Options, logger common.Logger, broker *broker.Broker) *Handler {\n\trouter := route.New()\n\n\th := &Handler{\n\t\toptions: o,\n\t\tlogger: logger,\n\t\trouter: router,\n\t\tbroker: broker,\n\t}\n\n\t// Setup HTTP endpoints\n\trouter.Get(\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\thttp.Redirect(w, r, \"/overview\", http.StatusFound)\n\t})\n\trouter.Get(\"/overview\", h.handleOverview)\n\trouter.Get(\"/logs\", func(w http.ResponseWriter, r *http.Request) {\n\t\thttp.Redirect(w, r, \"/logs/*\", http.StatusFound)\n\t})\n\trouter.Get(\"/logs/:pattern\", h.handleLogs)\n\trouter.Get(\"/request\", h.handleRequest)\n\trouter.Post(\"/request\", h.handleRequestPost)\n\n\t// Static files\n\trouter.Get(\"/static/*filepath\", route.FileServe(path.Join(o.AssetsPath, \"static\")))\n\n\t// Prometheus HTTP endpoint\n\trouter.Get(\"/metrics\", promhttp.HandlerFor(prometheus.Gatherers{prometheus.DefaultGatherer, broker.Monitoring.Registry}, promhttp.HandlerOpts{}).ServeHTTP)\n\n\t// cellaserv HTTP API\n\trouter.Get(\"/api/v1/request/:service/:method\", h.apiRequest)\n\trouter.Post(\"/api/v1/request/:service/:method\", h.apiRequest)\n\trouter.Post(\"/api/v1/publish/:event\", h.apiPublish)\n\trouter.Get(\"/api/v1/subscribe/:event\", h.apiSubscribe)\n\t// TODO(halfr): spy\n\n\t// Go debug\n\trouter.Get(\"/debug/*subpath\", handleDebug)\n\trouter.Post(\"/debug/*subpath\", handleDebug)\n\n\treturn h\n}", "func NewHandler(transporteService domain.TransporteService) http.Handler {\n\thandler := &handler{\n\t\ttransporteService: transporteService,\n\t}\n\n\tgin.SetMode(gin.ReleaseMode)\n\n\trouter := gin.New()\n\trouter.Use(handler.recovery())\n\trouter.GET(\"/\", func(c *gin.Context) {\n\t\tc.HTML(http.StatusOK, \"index.html\", gin.H{\n\t\t\t\"title\": \"Home Page\",\n\t\t})\n\t})\n\n\tv1 := router.Group(\"/v1\")\n\n\tv1.GET(\"/calcularTransporte\", handler.PostCalcularTransporte)\n\tv1.GET(\"/veiculos\", handler.GetVeiculos)\n\n\treturn router\n}", "func New(handler http.Handler) backend.CallResourceHandler {\n\treturn &httpResourceHandler{\n\t\thandler: handler,\n\t}\n}", "func New(this *Handler) *Handler {\n\ta := &API{Cfg: this.Cfg, Interactor: this.Interactor}\n\troutes(a).Register()\n\treturn this\n}", "func (g *GRPC) httpRESTHandler(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(next.ServeHTTP)\n\t// TODO(jdoak): GOTTA FIX THIS!!!\n\t/*\n\t\tfunc(w http.ResponseWriter, r *http.Request) {\n\t\t\thandler := g.decompressHandler(\n\t\t\t\tg.compressHandler(\n\t\t\t\t\tnext,\n\t\t\t\t),\n\t\t\t)\n\t\t\thandler.ServeHTTP(w, r)\n\t\t},\n\t*/\n\n}", "func NewUserHandler() *UserHandler {\n h := &UserHandler{\n Router: httprouter.New(),\n Logger: log.New(os.Stderr, \"\", log.LstdFlags),\n }\n h.POST(\"/api/user\", h.handlePostUser)\n return h\n}", "func NewHandler(next http.Handler) Handler {\n\treturn Handler{\n\t\tNext: next,\n\t\tLogger: JSONLogger,\n\t\tSkip: SkipHealthEndpoint,\n\t}\n}", "func CreateHandler(config *Config) http.Handler {\n\trouter := httprouter.New()\n\trouter.RedirectTrailingSlash = false\n\n\trouter.GET(\"/\", indexHandler{config: config}.Handle)\n\n\tfor name, pkg := range config.Packages {\n\t\thandle := packageHandler{\n\t\t\tpkgName: name,\n\t\t\tpkg: pkg,\n\t\t\tconfig: config,\n\t\t}.Handle\n\t\trouter.GET(fmt.Sprintf(\"/%s\", name), handle)\n\t\trouter.GET(fmt.Sprintf(\"/%s/*path\", name), handle)\n\t}\n\n\treturn router\n}", "func NewHandler(s service.Service) http.Handler {\n\tr := mux.NewRouter()\n\t// base handler\n\tbase := alice.New(newSetUserMid(s))\n\t// handler with auth required\n\tauthRequired := base.Append(newAuthRequiredMid)\n\n\th := &handler{s}\n\n\t// r.PathPrefix(\"/images\").Handler(httputil.NewSingleHostReverseProxy(proxyURL))\n\tr.Handle(\"/v1/login\", base.Then(errHandler(h.register))).Methods(http.MethodPost)\n\tr.Handle(\"/v1/me\", authRequired.Then(errHandler(h.me))).Methods(http.MethodGet)\n\tr.Handle(\"/v1/me\", authRequired.Then(errHandler(h.update))).Methods(http.MethodPatch)\n\tr.Handle(\"/v1/me/reacts\", authRequired.Then(errHandler(h.react))).Methods(http.MethodPost)\n\tr.Handle(\"/v1/me/abuses\", authRequired.Then(errHandler(h.reportAbuse))).Methods(http.MethodPost)\n\n\tr.Handle(\"/v1/me/discover-people\", authRequired.Then(errHandler(h.discoverPeople))).Methods(http.MethodGet)\n\n\tr.Handle(\"/v1/me/pictures\", authRequired.Then(errHandler(h.uploadPicture))).Methods(http.MethodPost)\n\tr.Handle(\"/v1/me/pictures\", authRequired.Then(errHandler(h.pictures))).Methods(http.MethodGet)\n\tr.Handle(\"/v1/me/pictures/{id}\", authRequired.Then(errHandler(h.deletePicture))).Methods(http.MethodDelete)\n\tr.Handle(\"/v1/me/pictures/{id}/profile\", authRequired.Then(errHandler(h.setProfilePicture))).Methods(http.MethodPut)\n\n\treturn r\n}", "func NewHandler(adventure Adventure) http.Handler {\n\treturn handler{adventure}\n}", "func New(opts ...Option) *Handler {\n\th := &Handler{}\n\tfor _, option := range opts {\n\t\toption(h)\n\t}\n\tif h.router == nil {\n\t\th.router = &tree{}\n\t}\n\tif h.locales == nil {\n\t\th.locales = newLocaleMatcher([]language.Tag{language.English})\n\t}\n\tif h.decoder == nil {\n\t\th.decoder = NewContentTypeDecoder(map[string]Decoder{\n\t\t\t\"application/json\": &jsonDecoder{},\n\t\t})\n\t}\n\tif h.encoder == nil {\n\t\tencoder := &jsonEncoder{}\n\t\th.encoder = NewAcceptEncoder(map[string]Encoder{\n\t\t\t\"\": encoder,\n\t\t\t\"*/*\": encoder,\n\t\t\t\"application/*\": encoder,\n\t\t\t\"application/json\": encoder,\n\t\t})\n\t}\n\tif h.resolver == nil {\n\t\th.resolver = ResolverFunc(defaultResolver)\n\t}\n\tif h.pool == nil {\n\t\th.pool = &pool{free: make(chan *bytes.Buffer, 1<<6)}\n\t}\n\tif h.log == nil {\n\t\th.log = defaultLogger\n\t}\n\tif h.observer == nil {\n\t\th.observer = &discardObserver{}\n\t}\n\treturn h\n}", "func NewHandler(l logger.Logger) http.Handler {\n\thandlerLogger := l.With(\"module\", moduleNameHTTPHandler)\n\tr := chi.NewRouter()\n\n\tr.Use(logger.ChiRequestLoggerMiddleware(handlerLogger))\n\tr.NotFound(transport.Error(handlerLogger, http.StatusNotFound))\n\tr.MethodNotAllowed(transport.Error(handlerLogger, http.StatusMethodNotAllowed))\n\n\tr.Route(\"/_\", func(r chi.Router) {\n\t\tr.Get(\"/health\", transport.NewHealthHandler())\n\t})\n\n\treturn r\n}", "func NewHandler(intfs *Interfaces) infra.Handler {\n\tf := func(r *infra.Request) *infra.HandlerResult {\n\t\thandler := &handler{\n\t\t\tintfs: intfs,\n\t\t\trequest: r,\n\t\t}\n\t\treturn handler.Handle()\n\t}\n\treturn infra.HandlerFunc(f)\n}", "func (am AppModule) NewHandler() sdk.Handler {\n\treturn NewHandler(am.keeper)\n}", "func (am AppModule) NewHandler() sdk.Handler {\n\treturn NewHandler(am.keeper)\n}", "func (am AppModule) NewHandler() sdk.Handler {\n\treturn NewHandler(am.keeper)\n}", "func MakeHandler() *http.Handler {\n\tapi := rest.NewApi()\n\n\tauthMiddleware, err := CreateAuthMiddleware()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tapi.Use(&rest.IfMiddleware{\n\t\t// Only authenticate non login or register requests\n\t\tCondition: func(request *rest.Request) bool {\n\t\t\treturn (request.URL.Path != variables.APIPathLoginUserServer) && (request.URL.Path != variables.APIPathRegisterUserServer)\n\t\t},\n\t\tIfTrue: authMiddleware,\n\t})\n\tapi.Use(rest.DefaultProdStack...)\n\trouter, err := rest.MakeRouter(\n\t\trest.Post(variables.APIPathLoginUserServer, authMiddleware.LoginHandler),\n\t\trest.Get(variables.APIPathRefreshUserServer, authMiddleware.RefreshHandler),\n\t\trest.Post(variables.APIPathRegisterUserServer, PostRegister),\n\t\trest.Get(variables.APIPathUserServer, GetUser),\n\t\trest.Post(variables.APIPathUserServer, PostUser),\n\t)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tapi.SetApp(router)\n\thandler := api.MakeHandler()\n\treturn &handler\n}", "func newHTTPHandler(c ServiceController, k8sStorage ServiceStorage) *httpHandler {\n\treturn &httpHandler{\n\t\tcontroller: c,\n\t\tk8sStorage: k8sStorage,\n\t}\n}", "func NewHandler(db *DB) http.Handler {\n\tvar r = mux.NewRouter()\n\tr.Handle(\"/\", &IndexHandler{}).Methods(\"GET\")\n\tr.Handle(\"/assets/{filename}\", &AssetsHandler{}).Methods(\"GET\")\n\tr.Handle(\"/subscribe\", NewSubscribeHandler(db)).Methods(\"GET\")\n\treturn r\n}", "func (path *Path) NewHandler() http.HandlerFunc {\r\n\treturn func(w http.ResponseWriter, r *http.Request) {\r\n\t\tdata, err := doRequest(path.Site, path, r)\r\n\t\tpath.Flush(w, data, err)\r\n\t}\r\n}", "func NewHandler(ctx context.Context, f HandlerFunc) http.HandlerFunc {\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tresp := f(r.Context(), w, r)\n\n\t\tif err := resp.Failure(); err != nil {\n\t\t\tlog.From(ctx).Error(\"handling api\", zap.Error(err))\n\t\t}\n\n\t\terr := json.NewEncoder(w).Encode(resp)\n\t\tif err != nil {\n\t\t\tHTTPErr(ctx, w, errors.Wrap(err, \"encoding response\"))\n\t\t}\n\t}\n}", "func NewHandler(c *HandlerConfig) func(http.ResponseWriter, *http.Request) {\n\n\t// pushback receives the push request and writes it into a file\n\t// according to a mapping provided by a json configuration\n\treturn func(w http.ResponseWriter, r *http.Request) {\n\t\tt := r.Header.Get(\"Content-Type\")\n\n\t\tif t != \"binary/octet-stream\" {\n\t\t\tlog.Printf(\"Wrong Content-Type %s\", t)\n\t\t\tw.Write([]byte(fmt.Sprintf(\"%s is not a supported Content-Type\", t)))\n\t\t\treturn\n\t\t}\n\n\t\t// Open test file\n\t\tf, err := os.Create(fmt.Sprintf(\"%s/%s.pushback\", c.Path, \"test\"))\n\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Could not open file %e\", err)\n\t\t}\n\n\t\tn, err := io.Copy(f, r.Body)\n\t\tif err != nil {\n\t\t\tlog.Printf(\"Could only receive %d\", n)\n\t\t}\n\n\t\tw.Write([]byte(\"OK\"))\n\t}\n}", "func createHttpHandler(h http.HandlerFunc, method Method, defaultStatusCode int) HttpHandlerAdapter {\n\n\tadapter := HttpHandlerAdapter {\n\t\thandler: func (rw http.ResponseWriter, r *http.Request) {\n\t\t\tlogger.Infof(\" %s %s 200 OK \", r.Method, r.URL.Path )\n\t\t\th(rw, r)\n\t\t},\n\t\tdefaultStatusCode: defaultStatusCode,\n\t\tmethod: method};\n\n\treturn adapter;\n}", "func NewHandler(site *Site, path *Path) http.HandlerFunc {\r\n\treturn func(w http.ResponseWriter, r *http.Request) {\r\n\t\tdata, err := doRequest(site, path, r)\r\n\t\tif err != nil {\r\n\t\t\tsite.Flush(w, nil, err)\r\n\t\t\treturn\r\n\t\t}\r\n\t\tsite.Flush(w, data, err)\r\n\t}\r\n}", "func Create() http.Handler {\n\trouter := httprouter.New()\n\n\trouter.Handle(\"GET\", \"/\", middle.ResponseHandler(Hello))\n\trouter.Handle(\"POST\", \"/post\", middle.ResponseHandler(Hello))\n\trouter.Handle(\"GET\", \"/error\", middle.ResponseHandler(ErrorRoute))\n\trouter.Handle(\"GET\", \"/user-error\", middle.ResponseHandler(UserErrorRoute))\n\trouter.Handle(\"GET\", \"/multi-error\", middle.ResponseHandler(MultiErrorRoute))\n\trouter.Handle(\"GET\", \"/panic\", middle.ResponseHandler(Panic))\n\trouter.Handle(\"GET\", \"/version\", Version)\n\n\treturn alice.New(\n\t\tmiddle.RecoveryHandler,\n\t\tmiddle.FrameHandler,\n\t\tmiddle.RequestIDHandler,\n\t\tmiddle.RequestPathHandler,\n\t\tmiddle.BodyHandler).\n\t\tThen(router)\n}", "func New(s *service.Service) http.Handler {\n h := &handler{s}\n api := way.NewRouter()\n api.HandleFunc(\"POST\", \"/login\", h.login)\n api.HandleFunc(\"POST\", \"/send_magic_link\", h.sendMagicLink)\n api.HandleFunc(\"GET\", \"/auth_redirect\", h.authRedirect)\n api.HandleFunc(\"GET\", \"/user\", h.authUser)\n api.HandleFunc(\"POST\", \"/users/:username/toggle_follow\", h.toggleFollow)\n api.HandleFunc(\"PUT\", \"/user/avatar\", h.updateAvatar)\n api.HandleFunc(\"POST\", \"/users\", h.createUser)\n api.HandleFunc(\"GET\", \"/users\", h.users)\n api.HandleFunc(\"GET\", \"/users/:username\", h.user)\n api.HandleFunc(\"GET\", \"/users/:username/followers\", h.followers)\n api.HandleFunc(\"GET\", \"/users/:username/posts\", h.posts)\n api.HandleFunc(\"GET\", \"/users/:username/followees\", h.followees)\n\n api.HandleFunc(\"POST\", \"/posts\", h.createPost)\n api.HandleFunc(\"GET\", \"/posts/:post_id\", h.post)\n api.HandleFunc(\"POST\", \"/posts/:post_id/toggle_like\", h.togglePostLike)\n api.HandleFunc(\"POST\", \"/posts/:post_id/comments\", h.createComment)\n api.HandleFunc(\"GET\", \"/posts/:post_id/comments\", h.comments)\n\n api.HandleFunc(\"POST\", \"/comments/:comment_id/toggle_like\", h.toggleCommentLike)\n api.HandleFunc(\"GET\", \"/timeline\", h.timeline)\n api.HandleFunc(\"POST\", \"/posts/:post_id/toggle_subscription\", h.togglePostSubscription)\n\n api.HandleFunc(\"GET\", \"/notifications\", h.notifications)\n api.HandleFunc(\"POST\", \"/notifications/:notification_id/mark_as_read\", h.markNotificationAsRead)\n api.HandleFunc(\"POST\", \"/mark_notifications_as_read\", h.markAllNotificationsAsRead)\n\n fs := http.FileServer(&spaFileSystem{http.Dir(\"public\")})\n r := way.NewRouter()\n r.Handle(\"*\", \"/api...\", http.StripPrefix(\"/api\", h.withAuth(api)))\n r.Handle(\"GET\", \"/...\", fs)\n return r\n}", "func NewHandler(s service) *chi.Mux {\n\tr := chi.NewRouter()\n\n\tcors := cors.New(cors.Options{\n\t\t// AllowedOrigins: []string{\"https://foo.com\"}, // Use this to allow specific origin hosts\n\t\tAllowedOrigins: []string{\"*\"},\n\t\t// AllowOriginFunc: func(r *http.Request, origin string) bool { return true },\n\t\tAllowedMethods: []string{\"GET\", \"POST\", \"OPTIONS\"},\n\t\tAllowedHeaders: []string{\"Accept\", \"Authorization\", \"Content-Type\", \"X-CSRF-Token\"},\n\t\tExposedHeaders: []string{\"Link\"},\n\t\tAllowCredentials: true,\n\t\tMaxAge: 300, // Maximum value not ignored by any of major browsers\n\t})\n\tr.Use(cors.Handler)\n\n\t// Public routes\n\tr.Group(func(r chi.Router) {\n\t\tr.Get(\"/list\", getPluginListHandler(s))\n\t\tr.Get(\"/{plugin}/{format}\", getFeedHandler(s))\n\t})\n\n\tr.Group(func(r chi.Router) {\n\t\tr.Use(s.authenticator)\n\t\tr.Post(\"/{plugin}/refresh\", refreshFeedHandler(s))\n\t})\n\n\treturn r\n}", "func makeRestRoutes(router *gin.Engine, controller *dependency.Dependency) {\n\tapi := router.Group(\"api\")\n\t{\n\t\t// link forward , hitting this route will cause the link to be resolved and forwarded in the browser\n\t\tapi.GET(\"forward/links/:id\", controller.ForwardLink)\n\t\t// get them the json record\n\t\tapi.GET(\"links/:id\", controller.FetchLink)\n\t\t// generate a new link\n\t\tapi.POST(\"links\", controller.ShortenLink)\n\t}\n}", "func MakeHttpHandler(ctx context.Context, endpoints endpoint.Endpoints, logger log.Logger, zipkinTracer *gozipkin.Tracer,) http.Handler {\n\tr := mux.NewRouter()\n\t// 链路追踪\n\tzipkinServer := zipkin.HTTPServerTrace(zipkinTracer, zipkin.Name(\"http-transport\"))\n\toptions := []kithttp.ServerOption{\n\t\tkithttp.ServerErrorLogger(logger),\n\t\tkithttp.ServerErrorEncoder(kithttp.DefaultErrorEncoder),\n\t\tkithttp.ServerErrorEncoder(func(ctx context.Context, err error, w http.ResponseWriter) {\n\t\t\tlogger.Log(fmt.Sprint(ctx.Value(ContextReqUUid)))\n\t\t\tw.WriteHeader(http.StatusOK)\n\t\t\tjson.NewEncoder(w).Encode(err)\n\t\t}),\n\t\tkithttp.ServerBefore(func(ctx context.Context, request *http.Request) context.Context {\n\t\t\tUUID := uuid.NewV5(uuid.Must(uuid.NewV4(),nil), \"req_uuid\").String()\n\t\t\tlogger.Log(\"给请求添加uuid\", zap.Any(\"UUID\", UUID))\n\t\t\tctx = context.WithValue(ctx, ContextReqUUid, UUID)\n\t\t\treturn ctx\n\t\t}),\n\t\tzipkinServer,\n\t}\n\tr.Methods(\"POST\").Path(\"/movie-tags\").Handler(kithttp.NewServer(\n\t\tendpoints.MovieTags,\n\t\tdecodeMoviesTagsRequest, // 请求参数\n\t\tencode.JsonResponse,\n\t\toptions...,\n\t))\n // 暴露具体的 endpoint\n\tr.Methods(\"POST\").Path(\"/movie-list\").Handler(kithttp.NewServer(\n\t\tendpoints.MoviesList,\n\t\tdecodeHotPlayMoviesrRequest, // 请求参数\n\t\tencode.JsonResponse,\n\t\toptions...,\n\t))\n\n\tr.Methods(\"POST\").Path(\"/movie-detail\").Handler(kithttp.NewServer(\n\t\tendpoints.MovieDetail,\n\t\tdecodMovieDetailRequest, // 请求参数\n\t\tencode.JsonResponse,\n\t\toptions...,\n\t))\n\tr.Methods(\"POST\").Path(\"/credits\").Handler(kithttp.NewServer(\n\t\tendpoints.MovieCreditsWithTypes,\n\t\tdecodeMovieCreditsWithTypes, // 请求参数\n\t\tencode.JsonResponse,\n\t\toptions...,\n\t))\n\treturn r\n}", "func (r *Rest) Create(url string) *Rest {\n\tr.tmp = RestEndPoint{}\n\tr.tmp.Url = r.Prefix + url\n\tif r.interceptor != nil {\n\t\tif len(r.tmp.interceptor) == 0 {\n\t\t\tr.tmp.interceptor = r.interceptor\n\t\t} else {\n\t\t\tfor _, inte := range r.interceptor {\n\t\t\t\tr.tmp.interceptor = append(r.tmp.interceptor, inte)\n\t\t\t}\n\t\t}\n\t}\n\treturn r\n}", "func NewHandler() *Handler {\n\th := &Handler{\n\t\tHandler: &http.Handler{},\n\t\tProductHandler: NewProductHandler(),\n\t}\n\th.Handler.ProductHandler = h.ProductHandler.ProductHandler\n\treturn h\n}", "func NewHandler(srv Service) http.Handler {\n\th := handlers{srv}\n\tr := chi.NewRouter()\n\tr.Get(\"/\", h.handleGetDemoSample)\n\tr.Get(\"/error_demo\", h.handleGetErrorDemoSample)\n\treturn r\n}", "func initRestServer(cfg domain.ServerRestConfig, handlers []domain.RestHandler) *rest.Server {\n\trestServer := rest.NewServer(cfg, handlers)\n\trouter := restServer.Router()\n\n\t// Jaeger middleware initialization\n\trouter.Use(tracing.Middleware())\n\n\t// Swagger handler initialization\n\trouter.GET(cfg.SwaggerUrl+\"/*any\", ginSwagger.WrapHandler(swaggerFiles.Handler))\n\n\t// Prometheus handler initialization\n\trouter.GET(cfg.MetricsUrl, metric.PrometheusHandler())\n\n\t// PProf handler initialization\n\tpprof.Register(router)\n\n\treturn restServer\n}", "func (r *HandlerRouteItem) CreateHandler(w http.ResponseWriter, req *http.Request) http.Handler {\r\n\treturn r.handler\r\n}", "func newRoute(registedPath string, handler Handler) *Route {\n\tr := &Route{handler: handler, fullpath: registedPath}\n\tr.processPath()\n\treturn r\n}", "func createRestfulHandlerFunc(handlerFunc http.HandlerFunc) (http.HandlerFunc) {\n return func (w http.ResponseWriter, r *http.Request) {\n var errMessage string\n \n // Check content-type option\n\t\tcontentType := r.Header.Get(string(messages.CONTENT_TYPE))\n\t\tif contentType == \"\" || contentType != CONTENT_TYPE_VALUE {\n log.Warnf(\"dots_client.restfulApiHandler -- unsupported content-type: %+v\", contentType)\n errMessage = fmt.Sprintf(\"%s is an unsupported content-type \\n\", contentType)\n w.WriteHeader(http.StatusBadRequest)\n w.Write([]byte(errMessage))\n return\n }\n \n handlerFunc(w, r)\n }\n}", "func (app AppModule) NewHandler() sdk.Handler {\n\treturn NewHandler(app.msKeeper, app.poaKeeper)\n}", "func RegisterRestEndpoints(endpointInsts map[string]RestEndpointInst) {\n\n\tfor url, endpointInst := range endpointInsts {\n\t\tregistered[url] = endpointInst\n\n\t\tHandleFunc(url, func() func(w http.ResponseWriter, r *http.Request) {\n\t\t\tvar handlerURL = url\n\t\t\tvar handlerInst = endpointInst\n\n\t\t\treturn func(w http.ResponseWriter, r *http.Request) {\n\n\t\t\t\t// Create a new handler instance\n\n\t\t\t\thandler := handlerInst()\n\n\t\t\t\t// Handle request in appropriate method\n\n\t\t\t\tres := strings.TrimSpace(r.URL.Path[len(handlerURL):])\n\n\t\t\t\tif len(res) > 0 && res[len(res)-1] == '/' {\n\t\t\t\t\tres = res[:len(res)-1]\n\t\t\t\t}\n\n\t\t\t\tvar resources []string\n\n\t\t\t\tif res != \"\" {\n\t\t\t\t\tresources = strings.Split(res, \"/\")\n\t\t\t\t}\n\n\t\t\t\tswitch r.Method {\n\t\t\t\tcase \"GET\":\n\t\t\t\t\thandler.HandleGET(w, r, resources)\n\n\t\t\t\tcase \"POST\":\n\t\t\t\t\thandler.HandlePOST(w, r, resources)\n\n\t\t\t\tcase \"PUT\":\n\t\t\t\t\thandler.HandlePUT(w, r, resources)\n\n\t\t\t\tcase \"DELETE\":\n\t\t\t\t\thandler.HandleDELETE(w, r, resources)\n\n\t\t\t\tdefault:\n\t\t\t\t\thttp.Error(w, http.StatusText(http.StatusMethodNotAllowed),\n\t\t\t\t\t\thttp.StatusMethodNotAllowed)\n\t\t\t\t}\n\t\t\t}\n\t\t}())\n\t}\n}", "func CreateRouteHandler(rw http.ResponseWriter, req *http.Request) {\n\tvar r util.Redirect\n\tdecoder := json.NewDecoder(req.Body)\n\terr := decoder.Decode(&r)\n\tif err != nil {\n\t\tutil.WriteErrorResp(rw, http.StatusBadRequest, \"Invalid Request\")\n\t\treturn\n\t}\n\n\tkey, err := getKey()\n\tif err != nil {\n\t\tutil.WriteErrorResp(rw, http.StatusInternalServerError, \"Could not generate id\")\n\t\treturn\n\t}\n\n\terr = client.Set(key, r.Target, 0).Err()\n\tif err != nil {\n\t\tutil.WriteErrorResp(rw, http.StatusInternalServerError, \"failed to write to db\")\n\t\treturn\n\t}\n\n\tr.Key = key\n\tutil.WriterJSONResponse(rw, r)\n}", "func New(s *service.Service) http.Handler {\n\th := &handler{s}\n\n\tapi := chi.NewRouter()\n\n\tapi.Get(\"/\", root)\n\tapi.Post(\"/login\", h.login)\n\tapi.Post(\"/register\", h.registerPatient)\n\n\tapiWithCors := cors.Default().Handler(api)\n\treturn apiWithCors\n}", "func NewHandler(ctx context.Context, endpoint string) (Handler, error) {\n\tif len(endpoint) == 0 {\n\t\treturn nil, errors.New(\"endpoint is empty\")\n\t}\n\n\t// Get the client\n\tcfg := openapi.NewConfiguration()\n\tapiClient := openapi.NewAPIClient(cfg)\n\tif endpoint != \"localhost/cnwan\" {\n\t\tapiClient.ChangeBasePath(strings.Replace(cfg.BasePath, \"localhost/cnwan\", endpoint, 1))\n\t}\n\n\treturn &servicesHandler{\n\t\tclient: apiClient,\n\t\tmainCtx: ctx,\n\t}, nil\n}", "func newRESTEndpointService(hostPortStr string) endpointService {\n\treturn endpointService(\n\t\tnewRESTDiscoveryService(fmt.Sprintf(edsRestEndpointTemplate, hostPortStr)),\n\t)\n}", "func CreateHandler(svc ServiceModel) *http.Server {\n\treturn http.NewServer(\n\t\tmakeCreateEndpoint(svc),\n\t\tdecodeCreateRequest,\n\t\thttp.EncodeJSONResponse,\n\t)\n}", "func (s *Service) RegisterRESTHandler(userDefineHandlers map[string]http.Handler) {\n\thandler, group := SetUpRestHandler(s)\n\tapiutil.RegisterUserDefinedHandlers(userDefineHandlers, &group, handler)\n}", "func New(container *service.Container, nr *newrelic.Application, logger *zap.Logger) *mux.Router {\n\tr := mux.NewRouter().StrictSlash(true)\n\n\tzapMiddlewares := []mux.MiddlewareFunc{\n\t\tzapmw.WithZap(logger),\n\t\tzapmw.Request(zapcore.InfoLevel, \"request\"),\n\t\tzapmw.Recoverer(zapcore.ErrorLevel, \"recover\", zapmw.RecovererDefault),\n\t}\n\n\tr.Use(nrgorilla.Middleware(nr))\n\tr.Use(zapMiddlewares...)\n\n\t// Route => handler\n\tr.Methods(\"GET\").Path(\"/ping\").Handler(handlers.Ping())\n\n\tr.Methods(\"GET\").Path(\"/swagger.yaml\").Handler(handlers.SwaggerFile())\n\tr.Methods(\"GET\").Path(\"/documentation\").Handler(middleware.SwaggerUI(middleware.SwaggerUIOpts{\n\t\tSpecURL: \"/swagger.yaml\",\n\t\tPath: \"documentation\",\n\t}, r.NotFoundHandler))\n\n\tr.Methods(\"PUT\").Path(\"/templates\").Handler(handlers.UpsertTemplates(container.TemplatesService, logger))\n\tr.Methods(\"GET\").Path(\"/templates\").Handler(handlers.IndexTemplates(container.TemplatesService, logger))\n\tr.Methods(\"GET\").Path(\"/templates/{name}\").Handler(handlers.GetTemplates(container.TemplatesService, logger))\n\tr.Methods(\"DELETE\").Path(\"/templates/{name}\").Handler(handlers.DeleteTemplates(container.TemplatesService, logger))\n\tr.Methods(\"POST\").Path(\"/templates/{name}/render\").Handler(handlers.RenderTemplates(container.TemplatesService, logger))\n\tr.Methods(\"PUT\").Path(\"/teams/{teamName}/credentials\").Handler(handlers.UpdateAlertCredentials(container.AlertmanagerService, logger))\n\tr.Methods(\"GET\").Path(\"/teams/{teamName}/credentials\").Handler(handlers.GetAlertCredentials(container.AlertmanagerService, logger))\n\n\tr.Methods(\"PUT\").Path(\"/rules\").Handler(handlers.UpsertRule(container.RulesService, logger))\n\tr.Methods(\"GET\").Path(\"/rules\").Handler(handlers.GetRules(container.RulesService, logger))\n\tr.Methods(\"POST\").Path(\"/history\").Handler(handlers.CreateAlertHistory(container.AlertHistoryService, logger))\n\tr.Methods(\"GET\").Path(\"/history\").Handler(handlers.GetAlertHistory(container.AlertHistoryService, logger))\n\n\tr.Methods(\"POST\").Path(\"/oauth/slack/token\").Handler(handlers.ExchangeCode(container.CodeExchangeService, logger))\n\tr.Methods(\"POST\").Path(\"/notifications\").Handler(handlers.Notify(container.NotifierServices, logger))\n\n\tr.Methods(\"GET\").Path(\"/workspaces/{workspaceName}/channels\").Handler(handlers.GetWorkspaceChannels(container.WorkspaceService, logger))\n\n\t// Handle middlewares for NotFoundHandler and MethodNotAllowedHandler since Mux doesn't apply middlewares to them. Ref: https://github.com/gorilla/mux/issues/416\n\t_, r.NotFoundHandler = newrelic.WrapHandle(nr, \"NotFoundHandler\", applyMiddlewaresToHandler(zapMiddlewares, http.NotFoundHandler()))\n\t_, r.MethodNotAllowedHandler = newrelic.WrapHandle(nr, \"MethodNotAllowedHandler\", applyMiddlewaresToHandler(zapMiddlewares, http.HandlerFunc(func(rw http.ResponseWriter, r *http.Request) {\n\t\trw.WriteHeader(http.StatusMethodNotAllowed)\n\t})))\n\n\treturn r\n}", "func NewRouter() *mux.Router {\n router := mux.NewRouter().StrictSlash(true)\n for _, route := range routes { \n var handler http.Handler\n log.Println(route.Name)\n handler = route.HandlerFunc\n \n router.\n Methods(route.Method).\n Path(route.Pattern).\n Name(route.Name).\n Handler(handler)\n }\n return router\n}", "func New(secret string, rcvr interface{}) *Handler {\n\tif secret == \"\" {\n\t\tpanic(\"webhook: called New with empty secret\")\n\t}\n\treturn &Handler{\n\t\tsecret: secret,\n\t\trcvr: reflect.ValueOf(rcvr),\n\t\tmethod: payloadMethods(reflect.TypeOf(rcvr)),\n\t}\n}", "func NewHandler(entrypoint, network, address string) http.Handler {\n\tconnFactory := gofast.SimpleConnFactory(network, address)\n\tpool := gofast.NewClientPool(\n\t\tgofast.SimpleClientFactory(connFactory),\n\t\t10,\n\t\t60*time.Second,\n\t)\n\th := gofast.NewHandler(\n\t\tgofast.NewFileEndpoint(entrypoint)(gofast.BasicSession),\n\t\tpool.CreateClient,\n\t)\n\treturn h\n}", "func NewHandler() *Handler {\n\th := &Handler{\n\t\tHandler: client.NewHandler(),\n\t}\n\th.Handler.RequestService = &h.RequestService\n\n\treturn h\n}", "func New() *Controller {\n\tvar allHandlers []operation.Handler\n\n\trpService := operation.New()\n\n\thandlers := rpService.GetRESTHandlers()\n\n\tallHandlers = append(allHandlers, handlers...)\n\n\treturn &Controller{handlers: allHandlers}\n}", "func NewHandler(bouncer security.BouncerService,\n\tstatus *portainer.Status,\n\tdemoService *demo.Service,\n\tdataStore dataservices.DataStore,\n\tupgradeService upgrade.Service) *Handler {\n\n\th := &Handler{\n\t\tRouter: mux.NewRouter(),\n\t\tdataStore: dataStore,\n\t\tdemoService: demoService,\n\t\tstatus: status,\n\t\tupgradeService: upgradeService,\n\t}\n\n\trouter := h.PathPrefix(\"/system\").Subrouter()\n\n\tadminRouter := router.PathPrefix(\"/\").Subrouter()\n\tadminRouter.Use(bouncer.AdminAccess)\n\n\tadminRouter.Handle(\"/upgrade\", httperror.LoggerHandler(h.systemUpgrade)).Methods(http.MethodPost)\n\n\tauthenticatedRouter := router.PathPrefix(\"/\").Subrouter()\n\tauthenticatedRouter.Use(bouncer.AuthenticatedAccess)\n\n\tauthenticatedRouter.Handle(\"/version\", http.HandlerFunc(h.version)).Methods(http.MethodGet)\n\tauthenticatedRouter.Handle(\"/nodes\", httperror.LoggerHandler(h.systemNodesCount)).Methods(http.MethodGet)\n\tauthenticatedRouter.Handle(\"/info\", httperror.LoggerHandler(h.systemInfo)).Methods(http.MethodGet)\n\n\tpublicRouter := router.PathPrefix(\"/\").Subrouter()\n\tpublicRouter.Use(bouncer.PublicAccess)\n\n\tpublicRouter.Handle(\"/status\", httperror.LoggerHandler(h.systemStatus)).Methods(http.MethodGet)\n\n\t// Deprecated /status endpoint, will be removed in the future.\n\th.Handle(\"/status\",\n\t\tbouncer.PublicAccess(httperror.LoggerHandler(h.statusInspectDeprecated))).Methods(http.MethodGet)\n\th.Handle(\"/status/version\",\n\t\tbouncer.AuthenticatedAccess(http.HandlerFunc(h.versionDeprecated))).Methods(http.MethodGet)\n\th.Handle(\"/status/nodes\",\n\t\tbouncer.AuthenticatedAccess(httperror.LoggerHandler(h.statusNodesCountDeprecated))).Methods(http.MethodGet)\n\n\treturn h\n}", "func NewRouter() *mux.Router {\n router := mux.NewRouter().StrictSlash(true)\n for _, route := range routes {\n var handler http.Handler\n log.Println(route.Name)\n handler = route.HandlerFunc\n\n router.\n Methods(route.Method).\n Path(route.Pattern).\n Name(route.Name).\n Handler(handler)\n }\n return router\n}", "func (self *CentralBooking) InstallHandlers(router *mux.Router) {\n router.\n Methods(\"POST\").\n Path(\"/register/instance\").\n HandlerFunc(self.RegisterInstance)\n\n // apeing vault\n router.\n Methods(\"GET\").\n Path(\"/sys/health\").\n HandlerFunc(self.CheckHealth)\n}", "func (o *Operation) GetRESTHandlers() []Handler {\n\treturn []Handler{\n\t\tsupport.NewHTTPHandler(keystoresEndpoint, keystoresEndpoint, http.MethodPost, o.createKeystoreHandler),\n\t\tsupport.NewHTTPHandler(keysEndpoint, keysEndpoint, http.MethodPost, o.createKeyHandler),\n\t\tsupport.NewHTTPHandler(capabilityEndpoint, capabilityEndpoint, http.MethodPost, o.updateCapabilityHandler),\n\t\tsupport.NewHTTPHandler(exportEndpoint, exportEndpoint, http.MethodGet, o.exportKeyHandler),\n\t\tsupport.NewHTTPHandler(importEndpoint, importEndpoint, http.MethodPost, o.importKeyHandler),\n\t\tsupport.NewHTTPHandler(signEndpoint, signEndpoint, http.MethodPost, o.signHandler),\n\t\tsupport.NewHTTPHandler(verifyEndpoint, verifyEndpoint, http.MethodPost, o.verifyHandler),\n\t\tsupport.NewHTTPHandler(encryptEndpoint, encryptEndpoint, http.MethodPost, o.encryptHandler),\n\t\tsupport.NewHTTPHandler(decryptEndpoint, decryptEndpoint, http.MethodPost, o.decryptHandler),\n\t\tsupport.NewHTTPHandler(computeMACEndpoint, computeMACEndpoint, http.MethodPost, o.computeMACHandler),\n\t\tsupport.NewHTTPHandler(verifyMACEndpoint, verifyMACEndpoint, http.MethodPost, o.verifyMACHandler),\n\t\tsupport.NewHTTPHandler(wrapEndpoint, wrapEndpoint, http.MethodPost, o.wrapHandler),\n\t\tsupport.NewHTTPHandler(unwrapEndpoint, unwrapEndpoint, http.MethodPost, o.unwrapHandler),\n\t\t// CryptoBox operations\n\t\tsupport.NewHTTPHandler(easyEndpoint, easyEndpoint, http.MethodPost, o.easyHandler),\n\t\tsupport.NewHTTPHandler(easyOpenEndpoint, easyOpenEndpoint, http.MethodPost, o.easyOpenHandler),\n\t\tsupport.NewHTTPHandler(sealOpenEndpoint, sealOpenEndpoint, http.MethodPost, o.sealOpenHandler),\n\t\t// BBS+ operations\n\t\tsupport.NewHTTPHandler(signMultiEndpoint, signMultiEndpoint, http.MethodPost, o.signMultiHandler),\n\t\tsupport.NewHTTPHandler(verifyMultiEndpoint, verifyMultiEndpoint, http.MethodPost, o.verifyMultiHandler),\n\t\tsupport.NewHTTPHandler(deriveProofEndpoint, deriveProofEndpoint, http.MethodPost, o.deriveProofHandler),\n\t\tsupport.NewHTTPHandler(verifyProofEndpoint, verifyProofEndpoint, http.MethodPost, o.verifyProofHandler),\n\t}\n}", "func NewHandler(_ context.Context, svr *server.Server) (http.Handler, apiutil.APIServiceGroup, error) {\n\tautoScalingHandler := http.NewServeMux()\n\trd := render.New(render.Options{\n\t\tIndentJSON: true,\n\t})\n\tautoScalingHandler.Handle(autoScalingPrefix, negroni.New(\n\t\tserverapi.NewRedirector(svr),\n\t\tnegroni.Wrap(NewHTTPHandler(svr, rd))),\n\t)\n\treturn autoScalingHandler, autoscalingServiceGroup, nil\n}", "func NewHandler(config *api.APIConfig, service Service, databaseClient database.Client) Handler {\n\treturn Handler{\n\t\tconfig: config,\n\t\tservice: service,\n\t\tdatabaseClient: databaseClient,\n\t}\n}", "func New() *Handler {\n\tp := &Handler{}\n\tp.reverseProxy = &httputil.ReverseProxy{\n\t\tTransport: &http.Transport{\n\t\t\tDial: func(network, address string) (net.Conn, error) {\n\t\t\t\treturn p.dial(context.TODO(), address)\n\t\t\t},\n\t\t},\n\t\tDirector: func(*http.Request) {\n\t\t\t// Do nothing, the request is self-sufficient\n\t\t},\n\t}\n\treturn p\n}", "func NewHandler(e *echo.Group, us domain.UserService) *Handler {\n\th := &Handler{us}\n\te.POST(\"/user/login\", h.Login)\n\te.POST(\"/user/register\", h.Register)\n\treturn h\n}", "func NewRestAPI[T any](db database.ReadWriter) *RestAPI[T] {\n\treturn &RestAPI[T]{\n\t\tdb: db,\n\t}\n}", "func NewHandler(db *rationl.DB, clientID, secret string) http.Handler {\n\tvar authorizeHandler = newAuthorizeHandler(db, clientID, secret)\n\n\tr := mux.NewRouter()\n\tr.Handle(\"/\", &indexHandler{db})\n\tr.HandleFunc(\"/assets/{filename}\", assetsHandleFunc)\n\n\tr.Handle(\"/authorize\", authorizeHandler)\n\tr.Handle(\"/authorize/callback\", authorizeHandler)\n\n\tr.Handle(\"/investigations\", &InvestigationsHandler{db}).Methods(\"GET\")\n\tr.Handle(\"/investigations\", &CreateInvestigationHandler{db}).Methods(\"POST\")\n\tr.Handle(\"/investigations/new\", &NewInvestigationHandler{db}).Methods(\"GET\")\n\tr.Handle(\"/investigations/{id}\", &InvestigationHandler{db}).Methods(\"GET\")\n\t// r.Handle(\"/investigations/{id}\", &UpdateInvestigationHandler{db}).Methods(\"PATCH\")\n\tr.Handle(\"/investigations/{id}/edit\", &EditInvestigationHandler{db}).Methods(\"GET\")\n\t// r.Handle(\"/investigations/{id}\", &DeleteInvestigationHandler{db}).Methods(\"DELETE\")\n\treturn r\n}", "func (r *Route) handler(h Handler) http.Handler {\n\treturn &handler{h}\n}", "func New(s *service.Service) http.Handler {\n\th := handler{s}\n\tapi := way.NewRouter()\n\tapi.HandleFunc(\"POST\", \"/\", h.twillioHandler)\n\tapi.HandleFunc(\"POST\", \"/passcodes\", h.generatePasswords)\n\tapi.HandleFunc(\"GET\", \"/passcodes\", h.getPasscodes)\n\t// api.HandleFunc(\"POST\", \"/passcode/validate\", h.validatePasscode)\n\treturn api\n}", "func NewRouter(defClient rb.DefinitionManager,\n\tprofileClient rb.ProfileManager,\n\tinstClient app.InstanceManager,\n\tqueryClient app.QueryManager,\n\tconfigClient app.ConfigManager,\n\tconnectionClient connection.ConnectionManager,\n\ttemplateClient rb.ConfigTemplateManager,\n\tsubscriptionClient app.InstanceStatusSubManager,\n\thealthcheckClient healthcheck.InstanceHCManager) *mux.Router {\n\n\trouter := mux.NewRouter()\n\n\t// Setup Instance handler routes\n\tif instClient == nil {\n\t\tinstClient = app.NewInstanceClient()\n\t}\n\tinstHandler := instanceHandler{client: instClient}\n\tinstRouter := router.PathPrefix(\"/v1\").Subrouter()\n\tinstRouter.HandleFunc(\"/instance\", instHandler.createHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/instance\", instHandler.listHandler).Methods(\"GET\")\n\t// Match rb-names, versions or profiles\n\tinstRouter.HandleFunc(\"/instance\", instHandler.listHandler).\n\t\tQueries(\"rb-name\", \"{rb-name}\",\n\t\t\t\"rb-version\", \"{rb-version}\",\n\t\t\t\"profile-name\", \"{profile-name}\").Methods(\"GET\")\n\t//Want to get full Data -> add query param: /install/{instID}?full=true\n\tinstRouter.HandleFunc(\"/instance/{instID}\", instHandler.getHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/upgrade\", instHandler.upgradeHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/query\", instHandler.queryHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/query\", instHandler.queryHandler).\n\t\tQueries(\"ApiVersion\", \"{ApiVersion}\",\n\t\t\t\"Kind\", \"{Kind}\",\n\t\t\t\"Name\", \"{Name}\",\n\t\t\t\"Labels\", \"{Labels}\").Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}\", instHandler.deleteHandler).Methods(\"DELETE\")\n\n\t// Status handler routes\n\tif subscriptionClient == nil {\n\t\tsubscriptionClient = app.NewInstanceStatusSubClient()\n\t\tsubscriptionClient.RestoreWatchers()\n\t}\n\tinstanceStatusSubHandler := instanceStatusSubHandler{client: subscriptionClient}\n\tinstRouter.HandleFunc(\"/instance/{instID}/status\", instHandler.statusHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/status/subscription\", instanceStatusSubHandler.listHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/status/subscription\", instanceStatusSubHandler.createHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/status/subscription/{subID}\", instanceStatusSubHandler.getHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/status/subscription/{subID}\", instanceStatusSubHandler.updateHandler).Methods(\"PUT\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/status/subscription/{subID}\", instanceStatusSubHandler.deleteHandler).Methods(\"DELETE\")\n\n\t// Query handler routes\n\tif queryClient == nil {\n\t\tqueryClient = app.NewQueryClient()\n\t}\n\tqueryHandler := queryHandler{client: queryClient}\n\tqueryRouter := router.PathPrefix(\"/v1\").Subrouter()\n\tqueryRouter.HandleFunc(\"/query\", queryHandler.queryHandler).Methods(\"GET\")\n\tqueryRouter.HandleFunc(\"/query\", queryHandler.queryHandler).\n\t\tQueries(\"Namespace\", \"{Namespace}\",\n\t\t\t\"CloudRegion\", \"{CloudRegion}\",\n\t\t\t\"ApiVersion\", \"{ApiVersion}\",\n\t\t\t\"Kind\", \"{Kind}\",\n\t\t\t\"Name\", \"{Name}\",\n\t\t\t\"Labels\", \"{Labels}\").Methods(\"GET\")\n\n\t//Setup the broker handler here\n\t//Use the base router without any path prefixes\n\tbrokerHandler := brokerInstanceHandler{client: instClient}\n\trouter.HandleFunc(\"/{cloud-owner}/{cloud-region}/infra_workload\", brokerHandler.createHandler).Methods(\"POST\")\n\trouter.HandleFunc(\"/{cloud-owner}/{cloud-region}/infra_workload/{instID}\", brokerHandler.getHandler).Methods(\"GET\")\n\trouter.HandleFunc(\"/{cloud-owner}/{cloud-region}/infra_workload\", brokerHandler.findHandler).Queries(\"name\", \"{name}\").Methods(\"GET\")\n\trouter.HandleFunc(\"/{cloud-owner}/{cloud-region}/infra_workload/{instID}\", brokerHandler.deleteHandler).Methods(\"DELETE\")\n\n\t//Setup the connectivity api handler here\n\tif connectionClient == nil {\n\t\tconnectionClient = connection.NewConnectionClient()\n\t}\n\tconnectionHandler := connectionHandler{client: connectionClient}\n\tinstRouter.HandleFunc(\"/connectivity-info\", connectionHandler.createHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/connectivity-info/{connname}\", connectionHandler.getHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/connectivity-info/{connname}\", connectionHandler.deleteHandler).Methods(\"DELETE\")\n\n\t//Setup resource bundle definition routes\n\tif defClient == nil {\n\t\tdefClient = rb.NewDefinitionClient()\n\t}\n\tdefHandler := rbDefinitionHandler{client: defClient}\n\tresRouter := router.PathPrefix(\"/v1/rb\").Subrouter()\n\tresRouter.HandleFunc(\"/definition\", defHandler.createHandler).Methods(\"POST\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/content\", defHandler.uploadHandler).Methods(\"POST\")\n\tresRouter.HandleFunc(\"/definition/{rbname}\", defHandler.listVersionsHandler).Methods(\"GET\")\n\tresRouter.HandleFunc(\"/definition\", defHandler.listAllHandler).Methods(\"GET\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}\", defHandler.getHandler).Methods(\"GET\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}\", defHandler.updateHandler).Methods(\"PUT\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}\", defHandler.deleteHandler).Methods(\"DELETE\")\n\n\t//Setup resource bundle profile routes\n\tif profileClient == nil {\n\t\tprofileClient = rb.NewProfileClient()\n\t}\n\tprofileHandler := rbProfileHandler{client: profileClient}\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/profile\", profileHandler.createHandler).Methods(\"POST\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/profile\", profileHandler.listHandler).Methods(\"GET\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/profile/{prname}/content\", profileHandler.uploadHandler).Methods(\"POST\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/profile/{prname}\", profileHandler.getHandler).Methods(\"GET\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/profile/{prname}\", profileHandler.updateHandler).Methods(\"PUT\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/profile/{prname}\", profileHandler.deleteHandler).Methods(\"DELETE\")\n\n\t// Config Template\n\tif templateClient == nil {\n\t\ttemplateClient = rb.NewConfigTemplateClient()\n\t}\n\ttemplateHandler := rbTemplateHandler{client: templateClient}\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/config-template\", templateHandler.createHandler).Methods(\"POST\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/config-template\", templateHandler.listHandler).Methods(\"GET\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/config-template/{tname}/content\", templateHandler.uploadHandler).Methods(\"POST\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/config-template/{tname}\", templateHandler.getHandler).Methods(\"GET\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/config-template/{tname}\", templateHandler.updateHandler).Methods(\"PUT\")\n\tresRouter.HandleFunc(\"/definition/{rbname}/{rbversion}/config-template/{tname}\", templateHandler.deleteHandler).Methods(\"DELETE\")\n\n\t// Config value\n\tif configClient == nil {\n\t\tconfigClient = app.NewConfigClient()\n\t}\n\tconfigHandler := rbConfigHandler{client: configClient}\n\tinstRouter.HandleFunc(\"/instance/{instID}/config\", configHandler.createHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config\", configHandler.listHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}\", configHandler.getHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}\", configHandler.updateHandler).Methods(\"PUT\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}\", configHandler.deleteAllHandler).Methods(\"DELETE\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}/delete\", configHandler.deleteHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}/rollback\", configHandler.rollbackHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}/tag\", configHandler.tagListHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}/tag/{tagname}\", configHandler.getTagHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}/version\", configHandler.versionListHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}/version/{cfgversion}\", configHandler.getVersionHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/config/{cfgname}/tagit\", configHandler.tagitHandler).Methods(\"POST\")\n\n\t// Instance Healthcheck API\n\tif healthcheckClient == nil {\n\t\thealthcheckClient = healthcheck.NewHCClient()\n\t}\n\thealthcheckHandler := instanceHCHandler{client: healthcheckClient}\n\tinstRouter.HandleFunc(\"/instance/{instID}/healthcheck\", healthcheckHandler.listHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/healthcheck\", healthcheckHandler.createHandler).Methods(\"POST\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/healthcheck/{hcID}\", healthcheckHandler.getHandler).Methods(\"GET\")\n\tinstRouter.HandleFunc(\"/instance/{instID}/healthcheck/{hcID}\", healthcheckHandler.deleteHandler).Methods(\"DELETE\")\n\n\t// Add healthcheck path\n\tinstRouter.HandleFunc(\"/healthcheck\", healthCheckHandler).Methods(\"GET\")\n\n\treturn router\n}", "func NewHandler() http.Handler {\n\tuserMap = make(map[int]*User)\n\tlastId = 0\n\n\tmux := mux.NewRouter()\n\tmux.HandleFunc(\"/\", indexHandler)\n\tmux.HandleFunc(\"/users\", usersHandler).Methods(http.MethodGet)\n\tmux.HandleFunc(\"/users\", createUserHandler).Methods(http.MethodPost)\n\tmux.HandleFunc(\"/users/{id:[0-9a-z]+}\", getUserInfoHandler).Methods(http.MethodGet)\n\tmux.HandleFunc(\"/users/{id:[0-9a-z]+}\", deleteUserInfoHandler).Methods(http.MethodDelete)\n\tmux.HandleFunc(\"/users/{id:[0-9a-z]+}\", updateUserInfoHandler).Methods(http.MethodPut)\n\treturn mux\n}", "func NewAppHandler(args ServerArgs) http.Handler {\n\tv1 := NewV1Handler(args)\n\tif args.Debug {\n\t\t//args.Debug = false\n\t\tv1 = NewHTTPLogger(\"v1\").Handler(v1)\n\t}\n\n\treturn &App{\n\t\tPublicHandler: http.FileServer(args.Filesystem),\n\t\tIndexHandler: NewIndexHandler(args.GetAsset),\n\t\tV1ApiHandler: v1,\n\t}\n}", "func NewHandler() *Handler {\n\th := &Handler{\n\t\trouter: httprouter.New(),\n\t\tLogger: log15.New(),\n\t}\n\th.router.Handler(\"GET\", status.Path, status.Handler(h.healthStatus))\n\th.router.GET(\"/status\", h.handleGetStatus)\n\th.router.POST(\"/stop\", h.handlePostStop)\n\treturn h\n}", "func newRouter(hdl *handler.AppHandler) *mux.Router {\n\n\t// I should take this as a dependency, but need to do some work with wire\n\trtr := mux.NewRouter()\n\n\t// I should take this as a dependency, but need to do some work with wire\n\tc := alice.New()\n\n\t// add Standard Handler chain and zerolog logger to Context\n\tc = hdl.AddStandardHandlerChain(c)\n\n\t// send Router through PathPrefix method to validate any standard\n\t// subroutes you may want for your APIs. e.g. I always want to be\n\t// sure that every request has \"/api\" as part of it's path prefix\n\t// without having to put it into every handle path in my various\n\t// routing functions\n\trtr = rtr.PathPrefix(\"/api\").Subrouter()\n\n\t// Match only POST requests at /api/v1/movies\n\t// with Content-Type header = application/json\n\trtr.Handle(\"/v1/movies\",\n\t\tc.Append(hdl.AccessTokenHandler).\n\t\t\tThen(http.HandlerFunc(hdl.CreateMovie))).\n\t\tMethods(\"POST\").\n\t\tHeaders(\"Content-Type\", \"application/json\")\n\n\t// Match only PUT requests having an ID at /api/v1/movies/{id}\n\t// with the Content-Type header = application/json\n\trtr.Handle(\"/v1/movies/{id}\",\n\t\tc.Append(hdl.AccessTokenHandler).\n\t\t\tThen(http.HandlerFunc(hdl.UpdateMovie))).\n\t\tMethods(\"PUT\").\n\t\tHeaders(\"Content-Type\", \"application/json\")\n\n\t// Match only DELETE requests having an ID at /api/v1/movies/{id}\n\trtr.Handle(\"/v1/movies/{id}\",\n\t\tc.Append(hdl.AccessTokenHandler).\n\t\t\tThen(http.HandlerFunc(hdl.DeleteMovie))).\n\t\tMethods(\"DELETE\")\n\n\t// Match only GET requests having an ID at /api/v1/movies/{id}\n\trtr.Handle(\"/v1/movies/{id}\",\n\t\tc.Append(hdl.AccessTokenHandler).\n\t\t\tThen(http.HandlerFunc(hdl.FindByID))).\n\t\tMethods(\"GET\")\n\n\t// Match only GET requests /api/v1/movies\n\trtr.Handle(\"/v1/movies\",\n\t\tc.Append(hdl.AccessTokenHandler).\n\t\t\tThen(http.HandlerFunc(hdl.FindAll))).\n\t\tMethods(\"GET\")\n\n\t// Match only GET requests at /api/v1/ping\n\trtr.Handle(\"/v1/ping\",\n\t\tc.Then(http.HandlerFunc(hdl.Ping))).\n\t\tMethods(\"GET\")\n\n\treturn rtr\n}", "func New(cfg config.Proxy, bp httputil.BufferPool, token ntokend.TokenProvider, access service.AccessProvider, role service.RoleProvider, svcCert service.SvcCertProvider) Handler {\n\treturn &handler{\n\t\tproxy: &httputil.ReverseProxy{\n\t\t\tBufferPool: bp,\n\t\t},\n\t\ttoken: token,\n\t\taccess: access,\n\t\trole: role,\n\t\tcfg: cfg,\n\t\tsvcCert: svcCert,\n\t}\n}", "func NewHandler(cfg config.ServiceConfig,\n\trouter *mux.Router,\n\tregistry *prometheus.Registry) {\n\thandler := Handler{\n\t\tcfg: cfg,\n\t\tregistry: registry,\n\t}\n\n\thandler.routesBuilder(router)\n}", "func NewRestAPI() *RestAPI {\n\tinitializeDrivers()\n\n\thandler, err := makeHandler()\n\tif err != nil {\n\t\tlog.Fatal(err.Error())\n\t}\n\n\treturn &RestAPI{\n\t\tHandler: handler,\n\t}\n}", "func (r *Router) handler(h Handler) http.Handler {\n\treturn &handler{h}\n}", "func (r *Router) handler(h Handler) http.Handler {\n\treturn &handler{h}\n}", "func NewHTTPHandler(endpoints endpoint.Endpoints, options map[string][]kithttp.ServerOption) http.Handler {\n\tm := mux.NewRouter()\n\tm = m.PathPrefix(\"/v1/inventorysvc\").Subrouter()\n\n\tmakeCreateMerchantHandler(m, endpoints, options[\"CreateMerchant\"])\n\tmakeListMerchantHandler(m, endpoints, options[\"ListMerchant\"])\n\tmakeCreateProductHandler(m, endpoints, options[\"CreateProduct\"])\n\tmakeListProductHandler(m, endpoints, options[\"ListProduct\"])\n\n\treturn m\n}", "func (routeObj *Routes)NewRouter() *mux.Router {\n log := logger.GetLoggerInstance()\n router := mux.NewRouter().StrictSlash(true)\n routeObj.CreateAllRoutes()\n for _, route := range routeObj.entries {\n var handler http.Handler\n handler = route.HandlerFunc\n router.\n Methods(route.Method).\n Path(route.Pattern).\n Name(route.Name).\n Handler(handler)\n log.Trace(\"Created route for %s\", route.Name)\n }\n routeObj.controller = new(controller)\n return router\n}", "func NewHandler(cfg Config) http.Handler {\n\tr := resolver.New(\n\t\tresolver.Config{\n\t\t\tLogger: cfg.Logger,\n\t\t\tReceiverFactory: cfg.ReceiverFactory,\n\t\t},\n\t)\n\tu := &jobs{cfg.Logger, r}\n\trouter := mux.NewRouter()\n\trouter.Use(\n\t\tfunc(next http.Handler) http.Handler {\n\t\t\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t\t\tctx := newServicesContext(r.Context())\n\t\t\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t\t\t})\n\t\t})\n\troutes := []struct {\n\t\tname string\n\t\thandler http.HandlerFunc\n\t}{\n\t\t{\"sync_services\", u.syncServices},\n\t\t{\"gc\", u.garbageCollector},\n\t}\n\tfor _, route := range routes {\n\t\trouter.Path(\"/\" + route.name).\n\t\t\tMethods(http.MethodPost).\n\t\t\tHandlerFunc(route.handler).\n\t\t\tName(route.name)\n\t}\n\treturn router\n}", "func New(scepAuth scep.Interface) api.RouterHandler {\n\treturn &Handler{scepAuth}\n}", "func makeHandler() (http.Handler, *serializerLog, *mapStorage) {\n\treturn makeHandlerSubPath(\"\")\n}", "func NewHandler() Handler {\n\treturn Handler{}\n}", "func New(config *config.Config, fetcher *fetcher.Fetcher) Handler {\n\treturn Handler{\n\t\tfetcher: fetcher,\n\t\tconfig: config,\n\t}\n}", "func NewHandler(db moira.Database, log moira.Logger, index moira.Searcher, config *api.Config, metricSourceProvider *metricSource.SourceProvider, webConfigContent []byte) http.Handler {\n\tdatabase = db\n\tsearchIndex = index\n\trouter := chi.NewRouter()\n\trouter.Use(render.SetContentType(render.ContentTypeJSON))\n\trouter.Use(moiramiddle.UserContext)\n\trouter.Use(moiramiddle.RequestLogger(log))\n\trouter.Use(middleware.NoCache)\n\n\trouter.NotFound(notFoundHandler)\n\trouter.MethodNotAllowed(methodNotAllowedHandler)\n\n\trouter.Route(\"/api\", func(router chi.Router) {\n\t\trouter.Use(moiramiddle.DatabaseContext(database))\n\t\trouter.Get(\"/config\", getWebConfig(webConfigContent))\n\t\trouter.Route(\"/user\", user)\n\t\trouter.With(moiramiddle.Triggers(config.LocalMetricTTL, config.RemoteMetricTTL)).Route(\"/trigger\", triggers(metricSourceProvider, searchIndex))\n\t\trouter.Route(\"/tag\", tag)\n\t\trouter.Route(\"/pattern\", pattern)\n\t\trouter.Route(\"/event\", event)\n\t\trouter.Route(\"/contact\", contact)\n\t\trouter.Route(\"/subscription\", subscription)\n\t\trouter.Route(\"/notification\", notification)\n\t\trouter.Route(\"/health\", health)\n\t\trouter.Route(\"/teams\", teams)\n\t})\n\tif config.EnableCORS {\n\t\treturn cors.AllowAll().Handler(router)\n\t}\n\treturn router\n}", "func NewHandler(service Service) Handler {\n\treturn Handler{service: service}\n}", "func NewRestfulServer(s *Server, bind string, base string, ready chan<- bool) (*Restful, error) {\n\n\tr := new(Restful)\n\tr.server = s\n\tglog.Infof(\"Rest server base dir %s\", base)\n\tr.basedir = base\n\n\tok, err := io.IsDir(base)\n\tif err != nil {\n\t\tglog.Errorf(\"base dir %s is not valid directory\", err)\n\t\tclose(ready)\n\t\treturn nil, err\n\t}\n\n\tif !ok {\n\t\tglog.Errorf(\"base dir %s is not valid directory\", base)\n\t\tclose(ready)\n\t\treturn nil, fmt.Errorf(\"base dir %s is not valid directory\", base)\n\t}\n\n\t// register all end points\n\trouter := mux.NewRouter().StrictSlash(true)\n\trouter.HandleFunc(ApiIndex, r.HandlerIndex)\n\trouter.HandleFunc(\"/submit/{key}/{val}\", r.SubmitKeyValue)\n\trouter.HandleFunc(ApiCommitted, r.getCommitted)\n\trouter.HandleFunc(ApiGet+\"/{key}\", r.getValue)\n\trouter.HandleFunc(ApiLog, r.getLog)\n\trouter.HandleFunc(ApiShutdown, r.shutdownGrpc)\n\trouter.HandleFunc(ApiLeader, r.leader)\n\trouter.HandleFunc(ApiPeerList, r.HandlerPeerList)\n\trouter.HandleFunc(ApiFlows, r.HandlerFlows)\n\trouter.HandleFunc(\"/size\", r.HandlerCommitedSize)\n\trouter.HandleFunc(ApiFlowIndex, r.HandleFlowIndex)\n\trouter.HandleFunc(ApiRole, r.HandlerRole)\n\trouter.HandleFunc(ApiShutdownNode, r.HandlerNodeShutdown)\n\n\t// template css\n\tss := http.StripPrefix(\"/template/\", http.FileServer(http.Dir(base+\"/pkg/template/\")))\n\trouter.PathPrefix(\"/template/\").Handler(ss)\n\n\tglog.Infof(\"[restful server started]: %s\", bind)\n\n\tr.restServer = &http.Server{\n\t\tHandler: router,\n\t\tAddr: bind,\n\t\tWriteTimeout: 15 * time.Second,\n\t\tReadTimeout: 15 * time.Second,\n\t}\n\n\t//\tvar log = logrus.New()\n\tweblogger.SetFormatter(&weblogger.JSONFormatter{})\n\tlogFile := base + \"/\" + bind + \"restapi.log\"\n\tfile, err := os.OpenFile(logFile, os.O_CREATE|os.O_WRONLY, 0666)\n\tif err == nil {\n\t\tweblogger.SetOutput(file)\n\t} else {\n\t\tglog.Info(\"Failed to log to file, using default stderr\")\n\t}\n\n\tstandardFields := weblogger.Fields{\n\t\t\"hostname\": \"staging-1\",\n\t\t\"appname\": \"foo-app\",\n\t\t\"session\": \"1ce3f6v\",\n\t}\n\n\tweblogger.WithFields(standardFields).WithFields(weblogger.Fields{\n\t\t\"string\": r.restServer.Addr,\n\t\t\"int\": 1,\n\t\t\"float\": 1.1}).Info(\"server started\")\n\n\tr.ready = ready\n\treturn r, nil\n}", "func NewHandler(s *Server) *Handler {\n\th := &Handler{\n\t\tserver: s,\n\t\tmux: pat.New(),\n\t}\n\n\t// Series routes.\n\th.mux.Get(\"/db/:db/series\", http.HandlerFunc(h.serveQuery))\n\th.mux.Post(\"/db/:db/series\", http.HandlerFunc(h.serveWriteSeries))\n\th.mux.Del(\"/db/:db/series/:series\", http.HandlerFunc(h.serveDeleteSeries))\n\th.mux.Get(\"/db\", http.HandlerFunc(h.serveDatabases))\n\th.mux.Post(\"/db\", http.HandlerFunc(h.serveCreateDatabase))\n\th.mux.Del(\"/db/:name\", http.HandlerFunc(h.serveDeleteDatabase))\n\n\t// Cluster admins routes.\n\th.mux.Get(\"/cluster_admins/authenticate\", http.HandlerFunc(h.serveAuthenticateClusterAdmin))\n\th.mux.Get(\"/cluster_admins\", http.HandlerFunc(h.serveClusterAdmins))\n\th.mux.Post(\"/cluster_admins\", http.HandlerFunc(h.serveCreateClusterAdmin))\n\th.mux.Post(\"/cluster_admins/:user\", http.HandlerFunc(h.serveUpdateClusterAdmin))\n\th.mux.Del(\"/cluster_admins/:user\", http.HandlerFunc(h.serveDeleteClusterAdmin))\n\n\t// Database users routes.\n\th.mux.Get(\"/db/:db/authenticate\", http.HandlerFunc(h.serveAuthenticateDBUser))\n\th.mux.Get(\"/db/:db/users\", http.HandlerFunc(h.serveDBUsers))\n\th.mux.Post(\"/db/:db/users\", http.HandlerFunc(h.serveCreateDBUser))\n\th.mux.Get(\"/db/:db/users/:user\", http.HandlerFunc(h.serveDBUser))\n\th.mux.Post(\"/db/:db/users/:user\", http.HandlerFunc(h.serveUpdateDBUser))\n\th.mux.Del(\"/db/:db/users/:user\", http.HandlerFunc(h.serveDeleteDBUser))\n\n\t// Utilities\n\th.mux.Get(\"/ping\", http.HandlerFunc(h.servePing))\n\th.mux.Get(\"/interfaces\", http.HandlerFunc(h.serveInterfaces))\n\n\t// Shard routes.\n\th.mux.Get(\"/cluster/shards\", http.HandlerFunc(h.serveShards))\n\th.mux.Post(\"/cluster/shards\", http.HandlerFunc(h.serveCreateShard))\n\th.mux.Del(\"/cluster/shards/:id\", http.HandlerFunc(h.serveDeleteShard))\n\n\t// Shard space routes.\n\th.mux.Get(\"/cluster/shard_spaces\", http.HandlerFunc(h.serveShardSpaces))\n\th.mux.Post(\"/cluster/shard_spaces/:db\", http.HandlerFunc(h.serveCreateShardSpace))\n\th.mux.Post(\"/cluster/shard_spaces/:db/:name\", http.HandlerFunc(h.serveUpdateShardSpace))\n\th.mux.Del(\"/cluster/shard_spaces/:db/:name\", http.HandlerFunc(h.serveDeleteShardSpace))\n\n\t// Cluster config endpoints\n\th.mux.Get(\"/cluster/servers\", http.HandlerFunc(h.serveServers))\n\th.mux.Del(\"/cluster/servers/:id\", http.HandlerFunc(h.serveDeleteServer))\n\n\treturn h\n}", "func NewRest() *Rest {\n\tb := &Rest{}\n\tb.endpoints = make(map[string]RestEndPoint)\n\tb.interceptor = nil\n\tb.tmp = RestEndPoint{}\n\treturn b\n}", "func NewHandler(router *mux.Router) http.Handler {\n\ts := server{\n\t\tg: NewGetter(),\n\t\trouter: router,\n\t}\n\ts.routes()\n\treturn &s\n}", "func NewHandler(h http.Handler) http.Handler {\n\treturn handler{Handler: h}\n}", "func NewHttpHandler(ctx context.Context, endpoint Endpoints, logger *logrus.Logger) http.Handler {\n\tr := mux.NewRouter()\n\tr.UseEncodedPath()\n\toptions := []httptransport.ServerOption{\n\t\t//httptransport.ServerErrorLogger(logger),\n\t\thttptransport.ServerErrorEncoder(encodeError),\n\t}\n\n\tr.Methods(\"GET\").Path(\"/proxy\").HandlerFunc(proxyHandler)\n\tr.Methods(\"GET\").Path(\"/ping\").HandlerFunc(healthCheckHandler)\n\tr.Methods(\"POST\").Path(\"/fetch/splash\").Handler(httptransport.NewServer(\n\t\tendpoint.SplashFetchEndpoint,\n\t\tDecodeSplashFetcherRequest,\n\t\tEncodeFetcherContent,\n\t\toptions...,\n\t))\n\n\tr.Methods(\"POST\").Path(\"/fetch/base\").Handler(httptransport.NewServer(\n\t\tendpoint.BaseFetchEndpoint,\n\t\tDecodeBaseFetcherRequest,\n\t\tEncodeFetcherContent,\n\t\toptions...,\n\t))\n\n\tr.Methods(\"POST\").Path(\"/response/splash\").Handler(httptransport.NewServer(\n\t\tendpoint.SplashResponseEndpoint,\n\t\tDecodeSplashFetcherRequest,\n\t\tEncodeFetcherResponse,\n\t\toptions...,\n\t))\n\tr.Methods(\"POST\").Path(\"/response/base\").Handler(httptransport.NewServer(\n\t\tendpoint.BaseResponseEndpoint,\n\t\tDecodeBaseFetcherRequest,\n\t\tEncodeFetcherResponse,\n\t\toptions...,\n\t))\n\treturn r\n}", "func newRouter() *mux.Router {\n r := mux.NewRouter()\n r.HandleFunc(\"/api/v1/hello\", handlerHello).Methods(\"GET\")\n r.HandleFunc(\"/ws-echo\", handlerWS)\n return r\n}", "func newHandler(chain *Chain, handler Handler) (ch *ChainHandler) {\n\tch = new(ChainHandler)\n\tch.chain = chain\n\tch.handler = handler\n\treturn\n}", "func MakeHTTPSHandler(s Service) http.Handler {\n\tr := chi.NewRouter()\n\n\t//Obtener personas por su identificador\n\tgetPersonByHandler := kithttp.NewServer(\n\t\tmakeGetPersonByIDEndPoint(s),\n\t\tgetPersonByIDRequestDecoder,\n\t\tkithttp.EncodeJSONResponse,\n\t)\n\tr.Method(http.MethodGet, \"/id/{id}\", getPersonByHandler)\n\n\t//Obtener personas paginadas\n\tgetPersonHandler := kithttp.NewServer(\n\t\tmakeGetPersonsEndPoint(s),\n\t\tgetPersonsRequestDecoder,\n\t\tkithttp.EncodeJSONResponse,\n\t)\n\tr.Method(http.MethodPost, \"/paginated\", getPersonHandler)\n\n\t//Agregar a una persona\n\taddPersonHandler := kithttp.NewServer(\n\t\tmakeAddPersonEndpoint(s),\n\t\taddPersonRequestDecoder,\n\t\tkithttp.EncodeJSONResponse,\n\t)\n\tr.Method(http.MethodPost, \"/insert\", addPersonHandler)\n\n\t//Actualizar personas\n\tupdatePersonHandler := kithttp.NewServer(\n\t\tmakeUpdatePersonEndpoint(s),\n\t\tupdatePersonRequestDecoder,\n\t\tkithttp.EncodeJSONResponse,\n\t)\n\tr.Method(http.MethodPut, \"/update\", updatePersonHandler)\n\n\t//Eliminar PERSONA\n\tdeletePersonHandler := kithttp.NewServer(\n\t\tmakeDeletePersonEndPoint(s),\n\t\tdeletePersonRequestDecoder,\n\t\tkithttp.EncodeJSONResponse,\n\t)\n\tr.Method(http.MethodDelete, \"/delete/{id}\", deletePersonHandler)\n\treturn r\n}", "func (c *Operation) registerHandler() {\n\t// Add more protocol endpoints here to expose them as controller API endpoints\n\tc.handlers = []Handler{\n\t\tsupport.NewHTTPHandler(login, http.MethodGet, c.login),\n\t\tsupport.NewHTTPHandler(settings, http.MethodGet, c.settings),\n\t\tsupport.NewHTTPHandler(getCreditScore, http.MethodGet, c.getCreditScore),\n\t\tsupport.NewHTTPHandler(callback, http.MethodGet, c.callback),\n\t\tsupport.NewHTTPHandler(oidcRedirectPath, http.MethodGet, c.oidcRedirect),\n\n\t\t// issuer rest apis (html decoupled)\n\t\tsupport.NewHTTPHandler(authPath, http.MethodGet, c.auth),\n\t\tsupport.NewHTTPHandler(searchPath, http.MethodGet, c.search),\n\t\tsupport.NewHTTPHandler(verifyDIDAuthPath, http.MethodPost, c.verifyDIDAuthHandler),\n\t\tsupport.NewHTTPHandler(createCredentialPath, http.MethodPost, c.createCredentialHandler),\n\t\tsupport.NewHTTPHandler(generateCredentialPath, http.MethodPost, c.generateCredentialHandler),\n\n\t\t// chapi\n\t\tsupport.NewHTTPHandler(revoke, http.MethodPost, c.revokeVC),\n\t\tsupport.NewHTTPHandler(generate, http.MethodPost, c.generateVC),\n\n\t\t// didcomm\n\t\tsupport.NewHTTPHandler(didcommToken, http.MethodPost, c.didcommTokenHandler),\n\t\tsupport.NewHTTPHandler(didcommCallback, http.MethodGet, c.didcommCallbackHandler),\n\t\tsupport.NewHTTPHandler(didcommCredential, http.MethodPost, c.didcommCredentialHandler),\n\t\tsupport.NewHTTPHandler(didcommAssuranceData, http.MethodPost, c.didcommAssuraceHandler),\n\n\t\tsupport.NewHTTPHandler(didcommInit, http.MethodGet, c.initiateDIDCommConnection),\n\t\tsupport.NewHTTPHandler(didcommUserEndpoint, http.MethodGet, c.getIDHandler),\n\n\t\t// oidc\n\t\tsupport.NewHTTPHandler(oauth2GetRequestPath, http.MethodGet, c.createOIDCRequest),\n\t\tsupport.NewHTTPHandler(oauth2CallbackPath, http.MethodGet, c.handleOIDCCallback),\n\n\t\t// JSON-LD contexts API\n\t\tsupport.NewHTTPHandler(jsonldcontextrest.AddContextPath, http.MethodPost, c.addJSONLDContextHandler),\n\t}\n}", "func CreateHandler(typ string, loader Loader, config jsonconfig.Obj) (http.Handler, error) {\n\tmapLock.Lock()\n\tctor, ok := handlerConstructors[typ]\n\tmapLock.Unlock()\n\tif !ok {\n\t\treturn nil, fmt.Errorf(\"blobserver: Handler type %q not known or loaded\", typ)\n\t}\n\treturn ctor(loader, config)\n}" ]
[ "0.7223239", "0.7136761", "0.69532394", "0.6933166", "0.6649029", "0.6507859", "0.6464878", "0.64074403", "0.63570154", "0.628557", "0.6263668", "0.6247545", "0.61324114", "0.6068617", "0.60212016", "0.60161275", "0.5988077", "0.595724", "0.59482455", "0.59472746", "0.59422386", "0.5942137", "0.5942137", "0.5942137", "0.59378237", "0.59277666", "0.59179044", "0.5917812", "0.59173334", "0.59139645", "0.59095496", "0.58878624", "0.58832437", "0.58730775", "0.5872186", "0.58338344", "0.58254063", "0.58207875", "0.5815243", "0.58145046", "0.5810699", "0.5802047", "0.57972634", "0.57889086", "0.5788859", "0.57872283", "0.5778043", "0.57764894", "0.5757034", "0.5755849", "0.57386917", "0.57240844", "0.5714924", "0.5714119", "0.57113045", "0.5706925", "0.57061726", "0.57005185", "0.5694487", "0.5687103", "0.56847256", "0.56719494", "0.56708235", "0.56699854", "0.5667015", "0.5665652", "0.56624013", "0.5658594", "0.5655132", "0.56538254", "0.5650422", "0.56502193", "0.56451565", "0.56437063", "0.56303746", "0.56161", "0.5614754", "0.5612968", "0.5612128", "0.5612128", "0.56102854", "0.56038105", "0.5597663", "0.5588143", "0.55836684", "0.55814654", "0.5574729", "0.557113", "0.5568364", "0.5567232", "0.5564047", "0.5562134", "0.5555316", "0.5555315", "0.5550802", "0.55504996", "0.5545459", "0.55414635", "0.5539601", "0.5538988" ]
0.71187013
2
Mount handler with root "/" handling version in here
func (h *RestHandler) Mount(root *echo.Group) { v1Root := root.Group(candihelper.V1) account := v1Root.Group("/account", echo.WrapMiddleware(h.mw.HTTPBearerAuth)) account.GET("", h.getAllAccount, echo.WrapMiddleware(h.mw.HTTPPermissionACL("resource.public"))) account.GET("/:id", h.getDetailAccountByID, echo.WrapMiddleware(h.mw.HTTPPermissionACL("resource.public"))) account.POST("", h.createAccount, echo.WrapMiddleware(h.mw.HTTPPermissionACL("resource.public"))) account.PUT("/:id", h.updateAccount, echo.WrapMiddleware(h.mw.HTTPPermissionACL("resource.public"))) account.DELETE("/:id", h.deleteAccount, echo.WrapMiddleware(h.mw.HTTPPermissionACL("resource.public"))) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (mx *Mux) Mount(pattern string, handler interface{}) {\n\t// Provide runtime safety for ensuring a pattern isn't mounted on an existing\n\t// routing pattern.\n\tif mx.tree.findPattern(pattern+\"*\") || mx.tree.findPattern(pattern+\"/*\") {\n\t\tpanic(fmt.Sprintf(\"chi: attempting to Mount() a handler on an existing path, '%s'\", pattern))\n\t}\n\n\t// Assign Sub-Router's with the parent not found & method not allowed handler if not specified.\n\tsubr, ok := handler.(*Mux)\n\tif ok && subr.notFoundHandler == nil && mx.notFoundHandler != nil {\n\t\tsubr.NotFound(mx.notFoundHandler)\n\t}\n\tif ok && subr.methodNotAllowedHandler == nil && mx.methodNotAllowedHandler != nil {\n\t\tsubr.MethodNotAllowed(mx.methodNotAllowedHandler)\n\t}\n\n\thttpHandler := HttpHandler(handler)\n\tvar mh ContextHandler\n\n\tif mux, ok := handler.(Router); ok {\n\t\tmux.SetPrefix(pattern)\n\t\t// Wrap the Sub-router in a handlerFunc to scope the request path for routing.\n\t\tmmx, ok := mux.(*Mux)\n\n\t\tif ok {\n\t\t\tmmx.parent = mx\n\t\t}\n\n\t\tmh = &MountHandler{func(w http.ResponseWriter, r *http.Request, ctx *RouteContext) {\n\t\t\tctx.RoutePath = mx.nextRoutePath(ctx)\n\t\t\thttpHandler.ServeHTTPContext(w, r, ctx)\n\t\t}, mux}\n\t} else {\n\t\t// Wrap the Sub-router in a handlerFunc to scope the request path for routing.\n\t\tmh = &MountHandler{func(w http.ResponseWriter, r *http.Request, ctx *RouteContext) {\n\t\t\tctx.RoutePath = mx.nextRoutePath(ctx)\n\t\t\thttpHandler.ServeHTTPContext(w, r, ctx)\n\t\t}, handler}\n\t}\n\n\tif pattern == \"\" || pattern[len(pattern)-1] != '/' {\n\t\tnotFoundHandler := HttpHandler(func(w http.ResponseWriter, r *http.Request, arg *RouteContext) {\n\t\t\tif nfh := mx.NotFoundHandler(); nfh != nil {\n\t\t\t\tnfh.ServeHTTPContext(w, r, arg)\n\t\t\t}\n\t\t})\n\n\t\tmx.handle(ALL|STUB, pattern, mh)\n\t\tmx.handle(ALL|STUB, pattern+\"/\", notFoundHandler)\n\t\tpattern += \"/\"\n\t}\n\n\tmethod := ALL\n\tsubroutes, _ := handler.(Routes)\n\tif subroutes != nil {\n\t\tmethod |= STUB\n\t}\n\n\tfor _, n := range mx.handle(method, pattern+\"*\", mh) {\n\t\tif subroutes != nil {\n\t\t\tn.subroutes = subroutes\n\t\t}\n\t}\n}", "func Mount(r *mux.Router, path string, handler http.Handler) {\n\tr.PathPrefix(path).Handler(\n\t\thttp.StripPrefix(\n\t\t\tstrings.TrimSuffix(path, \"/\"),\n\t\t\tAddSlashForRoot(handler),\n\t\t),\n\t)\n}", "func Mount(multiplexer multiplexer, pattern string, handler http.Handler) {\n\tpattern = pathkit.Clean(pattern)\n\thandler = http.StripPrefix(pattern, handler)\n\thandler = MountPoint(pattern, handler)\n\tmultiplexer.Handle(pattern, handler)\n\tmultiplexer.Handle(pattern+`/`, handler)\n}", "func (s *server) Mount(path string, handler http.Handler) {\n\ts.Router.Mount(path, handler)\n}", "func (a *App) MountHandler(verb, path string, handler Handler, mw ...Middleware) {\n\thandler = wrapMiddleware(mw, handler)\n\n\t// wrap application level middlewares\n\thandler = wrapMiddleware(a.mw, handler)\n\n\th := func(w http.ResponseWriter, r *http.Request) {\n\t\t// TODO: start tracer span\n\n\t\t// add relevant values the context for propagation\n\t\tv := Values{\n\t\t\tNow: time.Now(),\n\t\t}\n\t\tctx := context.WithValue(r.Context(), KeyValues, &v)\n\n\t\tif err := handler(ctx, w, r); err != nil {\n\t\t\tlog.Println(err)\n\t\t\ta.Shutdown()\n\t\t\treturn\n\t\t}\n\t}\n\ta.HandleFunc(path, h).Methods(verb)\n}", "func Mount(mux *http.ServeMux) {\n\tmux.HandleFunc(\"/\", index) // list all new\n\tmux.Handle(\"/upload/\", http.StripPrefix(\"/upload\", http.FileServer(http.Dir(\"upload\"))))\n\t//mux.HandleFunc(\"/news/\", newsView)\n\tmux.Handle(\"/news/\", http.StripPrefix(\"/news\", http.HandlerFunc(newsView)))\n\tmux.HandleFunc(\"/register\", adminRegister)\n\t// mux.Handle(\"/news/\", http.StripPrefix(\"/news\", http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t// \tid := r.URL.Path[1:]\n\t// \tnewsView(id).ServeHTTP(w, r)\n\t// })))\n\n\tadminMux := http.NewServeMux()\n\tadminMux.HandleFunc(\"/login\", adminLogin)\n\tadminMux.HandleFunc(\"/listTest\", adminListTest)\n\tadminMux.HandleFunc(\"/create\", adminCreate)\n\tadminMux.HandleFunc(\"/createTest\", adminCreateTest)\n\tadminMux.HandleFunc(\"/edit\", adminEdit)\n\n\tadminMux.HandleFunc(\"/editTest\", adminEditTest)\n\tadminMux.HandleFunc(\"/runTest\", adminRunTest)\n\n\tmux.Handle(\"/admin/\", http.StripPrefix(\"/admin\", onlyAdmin(adminMux)))\n\t// /news/:path\n\t// /admin/login/\n\t// /admin/list\n\t// /admin/create\n\t// /admin/edit\n\n}", "func Mount(mux *http.ServeMux) {\n\tmux.Handle(\"/\", fetchUser(http.HandlerFunc(index))) // list all news\n\tmux.Handle(\"/upload/\", http.StripPrefix(\"/upload\", http.FileServer(http.Dir(\"upload\"))))\n\tmux.Handle(\"/news/\", http.StripPrefix(\"/news\", http.HandlerFunc(newsView))) // /news/:id\n\n\tmux.HandleFunc(\"/register\", adminRegister)\n\tmux.HandleFunc(\"/login\", adminLogin)\n\n\tadminMux := http.NewServeMux()\n\tadminMux.HandleFunc(\"/logout\", adminLogout)\n\tadminMux.HandleFunc(\"/list\", adminList)\n\tadminMux.HandleFunc(\"/create\", adminCreate)\n\tadminMux.HandleFunc(\"/edit\", adminEdit)\n\n\tmux.Handle(\"/admin/\", http.StripPrefix(\"/admin\", onlyAdmin(adminMux)))\n}", "func Handler(basepath string, data io.ReadSeeker) http.Handler {\n\tif basepath == \"\" {\n\t\tbasepath = \"/\"\n\t}\n\tas := &assetfs.AssetStore{\n\t\tNames: internal.AssetNames,\n\t\tData: internal.Asset,\n\t\tInfo: internal.AssetInfo,\n\t}\n\tfs, err := assetfs.New(as)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"failed to create static fs: %v\", err))\n\t}\n\tmux := http.NewServeMux()\n\tfsh := http.FileServer(http.FileSystem(fs))\n\tif basepath != \"/\" {\n\t\tfsh = http.StripPrefix(basepath, fsh)\n\t}\n\tp := assetfs.AddPrefix(basepath, BasePath)\n\tf := assetfs.AddPrefix(basepath, SpecFile)\n\tmux.HandleFunc(basepath, func(w http.ResponseWriter, r *http.Request) {\n\t\tif r.URL.Path == basepath {\n\t\t\thttp.Redirect(w, r, p+\"?url=\"+f, http.StatusSeeOther)\n\t\t\treturn\n\t\t}\n\t\tfsh.ServeHTTP(w, r)\n\t})\n\tmux.Handle(f, &handler{modTime: time.Now(), body: data})\n\treturn mux\n}", "func mount(e MountableResource, path string, handler MountHandler) {\n\thandler(e.Group(path))\n}", "func (h *eventServiceHTTPHandler) Mount(group *echo.Group) {\n\tgroup.POST(\"/create\", h.CreateEvent)\n\tgroup.POST(\"/ticket/create\", h.CreateTicket)\n\tgroup.GET(\"/get_info\", h.GetEventInfo)\n}", "func Mount(mux *http.ServeMux) {\n\tmiddleware := http.NewServeMux()\n\tmux.Handle(\"/\", checkPathCorrect(middleware))\n\tmiddleware.Handle(\"/\", http.FileServer(http.Dir(\"./static\")))\n\tmiddleware.HandleFunc(\"/movies\", list)\n\tmiddleware.Handle(\"/movie/\", http.StripPrefix(\"/movie\", http.HandlerFunc(get)))\n\tmiddleware.HandleFunc(\"/favorite\", favorite)\n\tmiddleware.Handle(\"/public/\", http.StripPrefix(\"/public\", http.FileServer(http.Dir(\"static\"))))\n}", "func (h *todoHTTPHandler) Mount(group *echo.Group) {\n group.POST(\"\", h.CreateTodo)\n group.GET(\"\", h.GetAllTodo)\n group.GET(\"/:id\", h.GetTodo)\n group.PUT(\"/:id\", h.UpdateTodo)\n group.DELETE(\"/:id\", h.DeleteTodo)\n group.PUT(\"/status/:id\", h.UpdateStatusTodo)\n}", "func handleMount(source, location, fstype string, flags uintptr, data string) error {\n\tif err := os.MkdirAll(location, os.FileMode(0755)); err != nil {\n\t\treturn err\n\t}\n\treturn syscall.Mount(source, location, fstype, flags, data)\n}", "func SetRootHandler(path string) {\n\troothandlerpath = path\n}", "func (p *Plugin) Mount(rt *h2o.Router) {\n\trt.GET(\"/site/info\", p.getSiteInfo)\n\trt.POST(\"/install\", p.mustDatabaseEmpty, p.postInstall)\n\n\trt.Group(func(r *h2o.Router) {\n\t\tr.GET(\"/users\", p.indexAdminUsers)\n\n\t\tr.GET(\"/locales\", p.getAdminLocales)\n\t\tr.POST(\"/locales\", p.postAdminLocales)\n\t\tr.DELETE(\"/locales/{:code}\", p.deleteAdminLocales)\n\n\t}, \"/admin\", p.Jwt.MustAdminMiddleware)\n\n\trt.Group(func(r *h2o.Router) {\n\t\tr.GET(\"/status\", p.getAdminSiteStatus)\n\t\tr.POST(\"/info\", p.postAdminSiteInfo)\n\t\tr.POST(\"/author\", p.postAdminSiteAuthor)\n\t\tr.GET(\"/seo\", p.getAdminSiteSeo)\n\t\tr.POST(\"/seo\", p.postAdminSiteSeo)\n\t\tr.GET(\"/smtp\", p.getAdminSiteSMTP)\n\t\tr.POST(\"/smtp\", p.postAdminSiteSMTP)\n\t}, \"/site\", p.Jwt.MustAdminMiddleware)\n\n\trt.Crud(\n\t\t\"/notices\",\n\t\t[]h2o.HandlerFunc{p.indexNotices},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createNotice},\n\t\t[]h2o.HandlerFunc{p.showNotice},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateNotice},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyNotice},\n\t)\n\n\trt.Crud(\n\t\t\"/leave-words\",\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.indexLeaveWords},\n\t\t[]h2o.HandlerFunc{p.createLeaveWord},\n\t\tnil,\n\t\tnil,\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyLeaveWord},\n\t)\n\n\trt.Crud(\n\t\t\"/friend-links\",\n\t\t[]h2o.HandlerFunc{p.indexFriendLinks},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createFriendLink},\n\t\t[]h2o.HandlerFunc{p.showFriendLink},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateFriendLink},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyFriendLink},\n\t)\n\n\trt.Crud(\n\t\t\"/links\",\n\t\t[]h2o.HandlerFunc{p.indexLinks},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createLink},\n\t\t[]h2o.HandlerFunc{p.showLink},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateLink},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyLink},\n\t)\n\n\trt.Crud(\n\t\t\"/cards\",\n\t\t[]h2o.HandlerFunc{p.indexCards},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createCard},\n\t\t[]h2o.HandlerFunc{p.showCard},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updateCard},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyCard},\n\t)\n\n\trt.Crud(\n\t\t\"/posts\",\n\t\t[]h2o.HandlerFunc{p.indexPosts},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.createPost},\n\t\t[]h2o.HandlerFunc{p.showPost},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.updatePost},\n\t\t[]h2o.HandlerFunc{p.Jwt.MustAdminMiddleware, p.destroyPost},\n\t)\n}", "func Test_App_Mount(t *testing.T) {\n\tt.Parallel()\n\tmicro := New()\n\tmicro.Get(\"/doe\", func(c *Ctx) error {\n\t\treturn c.SendStatus(StatusOK)\n\t})\n\n\tapp := New()\n\tapp.Mount(\"/john\", micro)\n\tresp, err := app.Test(httptest.NewRequest(MethodGet, \"/john/doe\", http.NoBody))\n\tutils.AssertEqual(t, nil, err, \"app.Test(req)\")\n\tutils.AssertEqual(t, 200, resp.StatusCode, \"Status code\")\n\tutils.AssertEqual(t, uint32(2), app.handlersCount)\n}", "func (h *PetHandler) Mount(r chi.Router, rs Routes) {\n\tif rs.has(PetCreate) {\n\t\tr.Post(\"/\", h.Create)\n\t}\n\tif rs.has(PetRead) {\n\t\tr.Get(\"/{id}\", h.Read)\n\t}\n\tif rs.has(PetUpdate) {\n\t\tr.Patch(\"/{id}\", h.Update)\n\t}\n\tif rs.has(PetDelete) {\n\t\tr.Delete(\"/{id}\", h.Delete)\n\t}\n\tif rs.has(PetList) {\n\t\tr.Get(\"/\", h.List)\n\t}\n\tif rs.has(PetOwner) {\n\t\tr.Get(\"/{id}/owner\", h.Owner)\n\t}\n}", "func AddSlashForRoot(next http.Handler) http.Handler {\n\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\t// @TODO MAYBE ENSURE SUFFIX DOESN'T ALREADY EXIST?\n\t\tif r.URL.Path == \"\" {\n\t\t\tr.URL.Path = \"/\"\n\t\t}\n\n\t\tnext.ServeHTTP(w, r)\n\t})\n}", "func init() {\n\thttp.HandleFunc(\"/savecasedriver\",\thandlerSaveCaseDriver)\n\thttp.HandleFunc(\"/saveimage\",\t\t\t\thandlerSaveImage)\n\thttp.HandleFunc(\"/sendemail\",\t\t\t\thandlerSendEmail)\n\thttp.HandleFunc(\"/serve/\",\t\t\t\t\thandlerServe)\n\thttp.HandleFunc(\"/login\",\t\t\t\t\t\thandlerLogin)\n\n\t// API Versions\n\thttp.HandleFunc(\"/api/1.0/\",\t\t\t\thandlerAPI10) // API version 1.0\n\n http.HandleFunc(\"/\",\t\t\t\t\t\t\t\thandlerRoot)\n}", "func (service *LoanSrvc) Mount(ctx context.Context, mux goahttp.Muxer) goahttp.Server {\n\tendpoints := loan.NewEndpoints(service)\n\tsrv := server.New(endpoints, mux, goahttp.RequestDecoder, goahttp.ResponseEncoder, api.ErrorHandler, nil)\n\tserver.Mount(mux, srv)\n\n\tfor _, m := range srv.Mounts {\n\t\tlog.WithContext(ctx).Infof(\"%q mounted on %s %s\", m.Method, m.Verb, m.Pattern)\n\t}\n\treturn srv\n}", "func RootHandler(w http.ResponseWriter, req *http.Request) {\n\tw.Header().Set(\"Content-type\", \"text/html\")\n\tif err := req.ParseForm(); err != nil {\n\t\thttp.Error(w, fmt.Sprintf(\"error parsing url %s\", err), http.StatusInternalServerError)\n\t}\n\tpath := mux.Vars(req)[\"path\"]\n\tif path == \"\" || path == \"/\" {\n\t\tpath = \"index.tpl\"\n\t}\n\tif !strings.HasSuffix(path, \".tpl\") {\n\t\tpath += \".tpl\"\n\t}\n\tif _, ok := registeredTpl[path]; !ok {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tfmt.Fprintf(w, \"Not found\")\n\t\treturn\n\t}\n\tif err := templates.ExecuteTemplate(w, path, Page{\n\t\tTitle: \"Home\",\n\t}); err != nil {\n\t\tlog.Printf(\"Error executing template: %s\", err)\n\t\thttp.Error(w, fmt.Sprintf(\"error parsing template: %s\", err), http.StatusInternalServerError)\n\t}\n}", "func (h AppServer) Handler (w http.ResponseWriter, r *http.Request) {\n\twasHandled := false\n\turlPath := r.URL.Path\n\tl := len(urlPath)\n\tif l > 0 {\n\t\tif urlPath[l-1:l] != \"/\" {\n\t\t\t// tack on a trailing slash\n\t\t\turlPath = urlPath + \"/\"\n\t\t}\n\t\tfmt.Println(\"appServer handler path=\", urlPath)\n\t\t\n\t\tfor p := range h.Handlers {\n\t\t\tif len(urlPath) >= len(p) &&\turlPath[:len(p)] == p {\n\t\t\t\twasHandled = true\n\t\t\t\tphf := h.Handlers[p]\n\t\t\t\tDispatchMethod(phf, w, r)\n\t\t\t} \n\t\t}\n\t}\n\tif !wasHandled {\n\t\t// not specific handler, assume it's a file\n\t\tif h.FileServerInst != nil {\n\t\t\tDispatchMethod(h.FileServerInst, w, r)\n\t\t} else {\n\t\t\thttp.Error(w, \"File not Found\", http.StatusNotFound)\n\t\t}\n\t}\n\n}", "func RootHandler(w http.ResponseWriter, r *http.Request) {\n\tfilePath := r.URL.Path[len(\"/\"):]\n\n\tsource, err := ioutil.ReadFile(root + filePath)\n\tif err != nil {\n\t\tsource, err = ioutil.ReadFile(root + filePath + \"/index.html\")\n\t\tif err != nil {\n\t\t\tw.WriteHeader(http.StatusNotFound)\n\t\t\tw.Write([]byte(err.Error()))\n\t\t\treturn\n\t\t}\n\t\tfilePath += \"index.html\"\n\t}\n\n\t// Set response headers\n\theaders.SetDefaultHeaders(w)\n\theaders.SetContentTypeHeader(w, filePath)\n\n\t// Send the response\n\tw.Write(source)\n\n\t//TODO: Log more detailed information.\n\tlog.Println(\"(rootHandler) The requested file has been sent: \", root+filePath)\n}", "func (h *RestHandler) Mount(root *echo.Group) {\n\tv1Root := root.Group(helper.V1)\n\n\tcustomer := v1Root.Group(\"/customer\")\n\tcustomer.GET(\"\", h.hello)\n}", "func rootHandler(w http.ResponseWriter, r *http.Request) {\n\tparts := strings.Split(r.URL.Path, \"/\")\n\tif r.URL.Path == \"/igcinfo/api/\" {\n\t\tinfoHandler(w, r)\n\t\treturn\n\t} else if r.URL.Path == \"/igcinfo/api/igc/\" {\n\t\tigcHandler(w, r)\n\t\treturn\n\t} else if id, err := uuid.Parse(parts[4]); strings.HasPrefix(r.URL.Path, \"/igcinfo/api/igc/\") && err == nil && len(parts) < 6 {\n\t\ttrackHandler(w, r, id)\n\t\treturn\n\t} else if id, err := uuid.Parse(parts[4]); strings.HasPrefix(r.URL.Path, \"/igcinfo/api/igc/\") && err == nil && len(parts[5]) > 0 {\n\t\ttrackFieldHandler(w, r, id, parts[5])\n\t\treturn\n\t}\n\n\thttp.NotFound(w, r)\n}", "func Handle(pattern string, handler Handler) { DefaultServeMux.Handle(pattern, handler) }", "func Handle(pattern string, handler Handler) { DefaultServeMux.Handle(pattern, handler) }", "func Handle(pattern string, handler Handler) { DefaultServeMux.Handle(pattern, handler) }", "func catchAllHandler(w http.ResponseWriter, r *http.Request) {\n\tp := path.Clean(\"/\" + r.URL.Path)\n\tif p == \"/\" {\n\t\tp += \"index\"\n\t}\n\tp = filepath.Join(config.Dir, filepath.FromSlash(p))\n\n\tif _, err := os.Stat(p); os.IsNotExist(err) {\n\t\tserveTemplate(w, r)\n\t\treturn\n\t}\n\n\thttp.ServeFile(w, r, p)\n}", "func (d *Daemon) rootHandler(w *rest.ResponseWriter, r *rest.Request) {\n\tw.Header().Set(\"Access-Control-Allow-Origin\", \"*\")\n\tdata, _ := Asset(\"gui/index.html\")\n\tw.Write(data)\n}", "func startHandlers(path string) {\n\n\thttp.HandleFunc(\"/\", handleIndex)\n\thttp.HandleFunc(\"/login\", handleLogin)\n\thttp.HandleFunc(\"/logout\", handleLogout)\n\thttp.HandleFunc(\"/create_ticket\", handleCreateTicket)\n\thttp.HandleFunc(\"/holiday\", handleHoliday)\n\thttp.HandleFunc(\"/ticketSend\", handleTicketSent)\n\thttp.HandleFunc(\"/ticket\", handleTicket)\n\thttp.HandleFunc(\"/updateTicket\", handleUpdateTicket)\n\thttp.HandleFunc(\"/unassignTicket\", handleUnassignTicket)\n\thttp.HandleFunc(\"/assignTicket\", handleAssignTicket)\n\n\t// Map the css, js and img folders to the location specified\n\thttp.Handle(\"/static/\", http.StripPrefix(\"/static/\", http.FileServer(http.Dir(path+\"/static\"))))\n}", "func RootHandler(w http.ResponseWriter, r *http.Request){\n\t// We add this if statement to mention that the root path is explicitly the \"/\"\n\tif r.URL.Path != \"/\"{\n\t\tw.WriteHeader(http.StatusNotFound)\n\t\tw.Write([]byte(\"The Access is not Defined \\n\"))\n\t\treturn\n\t}\n\t// if the path is correct then the code will be written to the server\n\tw.WriteHeader(http.StatusOK)\n\tw.Write([]byte(\"RUNNING API VERSION 1 \\n\"))\n\tfmt.Fprint(w, \"The easiest API \\n\")\n\n}", "func Mount(mux goahttp.Muxer, h *Server) {\n\tMountAddHandler(mux, h.Add)\n\tMountGetHandler(mux, h.Get)\n\tMountUpdateHandler(mux, h.Update)\n\tMountListMineHandler(mux, h.ListMine)\n\tMountListProjectHandler(mux, h.ListProject)\n\tMountPhotoHandler(mux, h.Photo)\n\tMountCORSHandler(mux, h.CORS)\n}", "func root(ctx context.Context, d *aegis.HandlerDependencies, req *aegis.APIGatewayProxyRequest, res *aegis.APIGatewayProxyResponse, params url.Values) error {\n\thost := req.GetHeader(\"Host\")\n\tstage := req.RequestContext.Stage\n\tres.HTML(200, \"<h1>Welcome</h1><p>This is an unprotected route.<br /><a href=\\\"https://\"+host+\"/\"+stage+\"/login\\\">Click here to login.</a></p>\")\n\treturn nil\n}", "func handleRoot(w http.ResponseWriter, r *http.Request) {\n\tfmt.Fprintf(w, \"hello from admission webhook server , you have hit : %q\", html.EscapeString(r.URL.Path))\n}", "func init() {\n\tmux = http.NewServeMux()\n\tmux.Handle(\"/\", handlerFn(\"Hello World!\"))\n\tmux.Handle(\"/foo\", handlerFn(\"foo\"))\n\tmux.Handle(\"/bar\", handlerFn(\"bar\"))\n\tmux.Handle(\"/baz\", handlerFn(\"baz\"))\n}", "func (s *Server) handleRootRequest(w http.ResponseWriter, req *http.Request) {\n if req.URL.Path != \"/\" {\n http.NotFound(w, req)\n return\n }\n\n http.Redirect(w, req, \"/ui/\", http.StatusFound)\n}", "func defaultMount(ctx context.Context, c *Component, r *http.Request) error {\n\treturn nil\n}", "func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\n\tinitService(service)\n\tvar h goa.Handler\n\n\th = ctrl.FileHandler(\"/swagger-ui/*filepath\", \"swagger-ui/dist\")\n\tservice.Mux.Handle(\"GET\", \"/swagger-ui/*filepath\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger-ui/dist\", \"route\", \"GET /swagger-ui/*filepath\")\n\n\th = ctrl.FileHandler(\"/swagger.json\", \"swagger/swagger.json\")\n\tservice.Mux.Handle(\"GET\", \"/swagger.json\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger/swagger.json\", \"route\", \"GET /swagger.json\")\n\n\th = ctrl.FileHandler(\"/swagger-ui/\", \"swagger-ui/dist/index.html\")\n\tservice.Mux.Handle(\"GET\", \"/swagger-ui/\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger-ui/dist/index.html\", \"route\", \"GET /swagger-ui/\")\n}", "func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\n\tinitService(service)\n\tvar h goa.Handler\n\n\th = ctrl.FileHandler(\"/swagger-ui/*filepath\", \"swagger-ui/dist\")\n\tservice.Mux.Handle(\"GET\", \"/swagger-ui/*filepath\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger-ui/dist\", \"route\", \"GET /swagger-ui/*filepath\")\n\n\th = ctrl.FileHandler(\"/swagger.json\", \"swagger/swagger.json\")\n\tservice.Mux.Handle(\"GET\", \"/swagger.json\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger/swagger.json\", \"route\", \"GET /swagger.json\")\n\n\th = ctrl.FileHandler(\"/swagger-ui/\", \"swagger-ui/dist/index.html\")\n\tservice.Mux.Handle(\"GET\", \"/swagger-ui/\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger-ui/dist/index.html\", \"route\", \"GET /swagger-ui/\")\n}", "func (r *Router) Mount(pattern string, router *Router, mws ...Middleware) {\n\trouter.parent = r\n\tr.subrouters = append(r.subrouters, router)\n\n\tvar p string\n\tif pattern == \"/\" {\n\t\tp = r.pattern\n\t} else {\n\t\tp = r.pattern + pattern\n\t}\n\trouter.pattern = p\n\n\thost := r.host\n\tfor i, route := range router.routes {\n\t\trouter.Host(route.Host())\n\t\tfor _, method := range route.Methods() {\n\t\t\trouter.Handle(method, route.Pattern(), route.Handler(method))\n\t\t}\n\n\t\trouter.routes = append(router.routes[:i], router.routes[i+1:]...)\n\t}\n\t// Restore previous host\n\tr.host = host\n}", "func rootHandler(c appengine.Context, w http.ResponseWriter, r *http.Request) *appError {\n\tlogoutURL, err := user.LogoutURL(c, \"/\")\n\tif err != nil {\n\t\tc.Warningf(\"creating logout URL: %v\", err)\n\t\tlogoutURL = \"/\"\n\t}\n\tuploadURL, err := blobstore.UploadURL(c, \"/upload\", nil)\n\tif err != nil {\n\t\treturn appErrorf(err, \"could not create blobstore upload url\")\n\t}\n\tusername := \"none\"\n\tif u := user.Current(c); u != nil {\n\t\tusername = u.String()\n\t}\n\terr = rootTemplate.Execute(w, &rootTemplateData{\n\t\tLogoutURL: logoutURL,\n\t\tUploadURL: uploadURL.String(),\n\t\tUser: username,\n\t})\n\tif err != nil {\n\t\treturn appErrorf(err, \"could not write template\")\n\t}\n\treturn nil\n}", "func init() {\n\thttp.HandleFunc(\"/\", playHtmlHandle)\n\thttp.HandleFunc(\"/runid\", runIdHandle)\n\thttp.HandleFunc(\"/img\", imgHandle)\n\thttp.HandleFunc(\"/clicked\", clickedHandle)\n\thttp.HandleFunc(\"/cheat\", cheatHandle)\n\thttp.HandleFunc(\"/new\", newGameHandle)\n\thttp.HandleFunc(\"/help\", helpHtmlHandle)\n}", "func (e *Viewer) Mount(r chi.Router) {\n\tr.Get(\"/*\", e.serve)\n\tr.Get(\"/swagger.spec\", e.load)\n}", "func (r *Router) Mount(src *Router) *Router {\n\tfor k, _ := range src.Router.Routes {\n\t\troute := r.Route(k.Pattern).Name(k.Noun)\n\t\tfor method, handler := range k.Handlers {\n\t\t\troute.Handle(handler, method)\n\t\t}\n\t}\n\treturn r\n}", "func (handlers *Handlers) rootHandler(w http.ResponseWriter, r *http.Request) {\n\t_, err := w.Write([]byte(\"OK!!\\n\"))\n\tif err != nil {\n\t\tlog.Println(err.Error() + \" Failed to write response bytes in root handler\")\n\t}\n}", "func (h *RestHandler) Mount(root *echo.Group) {\n\tbot := root.Group(\"/bot\")\n\n\tbot.POST(\"/callback\", h.callback)\n\tbot.POST(\"/pushmessage\", h.pushMessage, h.mw.HTTPBasicAuth(false))\n}", "func Serve() {\n\thttp.Handle(\"/\", Handler())\n}", "func (fs *fileServer) Serve(root http.FileSystem) http.Handler {\n return http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n upath := r.URL.Path\n if !strings.HasPrefix(upath, \"/\") {\n upath = \"/\" + upath\n r.URL.Path = upath\n }\n\n name := path.Clean(upath)\n\n\n // INFO: By having html this high up, it converts the document to HTML instread of go or some other rendering method\n //fmt.Fprintf(w, \"<a href=\\\"%s\\\">%s</a>\\n\", \"/search?q=\", htmlReplacer.Replace(\"pow\"))\n\n if file, ok := fs.cache[name]; ok {\n fs.serveContent(w, r, file.name, file.modTime, file.size, bytes.NewReader(file.data))\n // Serves all files as HTML\n //io.Copy(w, bytes.NewReader(file.data))\n return\n }\n\n // redirect .../index.html to .../\n // can't use Redirect() because that would make the path absolute,\n // which would be a problem running under StripPrefix\n if strings.HasSuffix(r.URL.Path, fs.indexPage) {\n localRedirect(w, r, \"./\")\n return\n }\n\n f, err := root.Open(name)\n if err != nil {\n http.NotFound(w, r)\n return\n }\n defer f.Close()\n\n d, err := f.Stat()\n if err != nil {\n http.NotFound(w, r)\n return\n }\n\n // redirect to canonical path: / at end of directory url\n // r.URL.Path always begins with /\n url := r.URL.Path\n if d.IsDir() {\n if url[len(url)-1] != '/' {\n localRedirect(w, r, path.Base(url)+\"/\")\n return\n }\n } else {\n if url[len(url)-1] == '/' {\n localRedirect(w, r, \"../\"+path.Base(url))\n return\n }\n }\n\n \n // use contents of index.html for directory, if present\n if d.IsDir() {\n index := name + fs.indexPage\n ff, err := root.Open(index)\n if err == nil {\n defer ff.Close()\n dd, err := ff.Stat()\n if err == nil {\n d = dd\n f = ff\n }\n }\n }\n\n // Still a directory? (we didn't find an index.html file)\n if d.IsDir() {\n if checkLastModified(w, r, d.ModTime()) {\n return\n }\n dirList(w, f)\n return\n }\n\n fs.serveContent(w, r, d.Name(), d.ModTime(), d.Size(), f)\n\n var buf bytes.Buffer\n io.Copy(&buf, f)\n\n fs.cache[name] = file{\n name: d.Name(),\n modTime: d.ModTime(),\n data: buf.Bytes(),\n }\n })\n}", "func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\n\tinitService(service)\n\tvar h goa.Handler\n\tservice.Mux.Handle(\"OPTIONS\", \"/swagger/*filepath\", ctrl.MuxHandler(\"preflight\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\n\n\th = ctrl.FileHandler(\"/swagger/*filepath\", \"public/swagger/\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swagger/*filepath\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"public/swagger/\", \"route\", \"GET /swagger/*filepath\")\n\n\th = ctrl.FileHandler(\"/swagger/\", \"public/swagger/index.html\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swagger/\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"public/swagger/index.html\", \"route\", \"GET /swagger/\")\n}", "func Mount(m *httpmux.Mux) {\n\tm.Handle(\"/\", methodmux.Get(\n\t\thime.Handler(index),\n\t))\n\tm.Handle(\"/signout\", methodmux.Post(\n\t\thime.Handler(signOut),\n\t))\n\n\tprofile := m.Group(\"/profile\", mustSignedIn)\n\tprofile.Handle(\"/\", methodmux.Get(\n\t\thime.Handler(getProfile),\n\t))\n\tprofile.Handle(\"/edit\", methodmux.GetPost(\n\t\thime.Handler(getProfileEdit),\n\t\thime.Handler(postProfileEdit),\n\t))\n\n\t// course\n\tm.Handle(\"/course/\", prefixhandler.New(\"/course\", courseIDKey{}, newCourseHandler()))\n}", "func rootHandler(w http.ResponseWriter, r *http.Request) {\r\n\t// Parsea la plantilla root.html \r\n\tif t, err := template.ParseFiles(filepath.Join(templates, \"root.html\")); err != nil {\r\n\t\t// Se ha presentado un error\r\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\r\n\t} else {\r\n\t\t// retorna la respuesta al cliente por medio de t.Execute\r\n\t\tt.Execute(w, nil)\r\n\t}\r\n}", "func init() {\n\thttp.HandleFunc(\"/\", errorAdapter(rootHandler))\n}", "func MountAddHandler(mux goahttp.Muxer, h http.Handler) {\n\tf, ok := handleStationOrigin(h).(http.HandlerFunc)\n\tif !ok {\n\t\tf = func(w http.ResponseWriter, r *http.Request) {\n\t\t\th.ServeHTTP(w, r)\n\t\t}\n\t}\n\tmux.Handle(\"POST\", \"/stations\", f)\n}", "func (h *UserHandler) Mount(r chi.Router, rs Routes) {\n\tif rs.has(UserCreate) {\n\t\tr.Post(\"/\", h.Create)\n\t}\n\tif rs.has(UserRead) {\n\t\tr.Get(\"/{id}\", h.Read)\n\t}\n\tif rs.has(UserUpdate) {\n\t\tr.Patch(\"/{id}\", h.Update)\n\t}\n\tif rs.has(UserDelete) {\n\t\tr.Delete(\"/{id}\", h.Delete)\n\t}\n\tif rs.has(UserList) {\n\t\tr.Get(\"/\", h.List)\n\t}\n\tif rs.has(UserPets) {\n\t\tr.Get(\"/{id}/pets\", h.Pets)\n\t}\n}", "func (h *HTTPHandler) Mount(g *echo.Group) {\n\tg.GET(\"/users\", h.FindAllUser)\n}", "func setuphandlers() {\n\thttp.HandleFunc(\"/\", rootHandler)\n\thttp.HandleFunc(\"/status\", statusHandler)\n\thttp.HandleFunc(\"/stats\", statsHandler)\n\thttp.HandleFunc(\"/request\", requestHandler)\n}", "func Handle(path string, v interface{}, options ...RestFunc) {\n\tDefaultServeMux.Handle(path, v, options...)\n}", "func rootHandler(w http.ResponseWriter, r *http.Request) {\n\thttp.Error(w, \"Image Server\", http.StatusNotImplemented)\n}", "func (s *Server) route() {\n\ts.r.HandleFunc(\"/\", s.HandleArticlesGetAll())\n\ts.r.HandleFunc(\"/article/{slug}\", s.HandleArticlesGetOne())\n\ts.r.HandleFunc(\"/talks\", s.HandleTalksGetOne())\n\ts.r.HandleFunc(\"/talks/{slug}\", s.HandleTalksGetOne())\n\ts.r.HandleFunc(\"/about\", s.HandleAboutGet())\n}", "func main() {\n\tlog.SetOutput(os.Stdout)\n\tflag.Parse()\n\tvar router = mux.NewRouter()\n\tvar api = router.PathPrefix(\"/api\").Subrouter()\n\trouter.PathPrefix(\"/static/\").Handler(http.StripPrefix(\"/static/\", http.FileServer(http.Dir(\"./static\"))))\n\tapi.NotFoundHandler = http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t})\n\n\t//api.Use(func(next http.Handler) http.Handler {\n\t//\treturn http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t//\t\tif r.Header.Get(\"x-auth-token\") != \"admin\" {\n\t//\t\t\tw.WriteHeader(http.StatusUnauthorized)\n\t//\t\t\treturn\n\t//\t\t}\n\t//\t\tlog.Println(r.RequestURI)\n\t//\t\tnext.ServeHTTP(w, r)\n\t//\t})\n\t//})\n\n\tvar api1 = api.PathPrefix(\"/v1\").Subrouter()\n\tapi1.HandleFunc(\"/status\", func(w http.ResponseWriter, r *http.Request) {\n\t\tw.WriteHeader(http.StatusOK)\n\t})\n\tapi1.NotFoundHandler = http.HandlerFunc(func(w http.ResponseWriter, r *http.Request) {\n\t\tw.WriteHeader(http.StatusNotFound)\n\t})\n\n\tglueapi.Handlers(api1.PathPrefix(\"/glue/\").Subrouter())\n\ts3api.Handlers(api1.PathPrefix(\"/s3/\").Subrouter())\n\n\tsrv := &http.Server{\n\t\tHandler: router,\n\t\tAddr: \":\" + *port,\n\t\t// Good practice: enforce timeouts for servers you create!\n\t\tWriteTimeout: 15 * time.Second,\n\t\tReadTimeout: 15 * time.Second,\n\t}\n\n\tlog.Fatal(srv.ListenAndServe())\n\t//http.ListenAndServe(\":\"+*port, router)\n}", "func handleWebroot(webroot string) {\n\tdefer func() {\n\t\tif r := recover(); r != nil {\n\t\t\tlog.Fatal(\"when using --webroot, --endpoint must not equal '/': \", r)\n\t\t}\n\t}()\n\n\t_, err := os.Stat(webroot)\n\tif err != nil && webroot != \"\" {\n\t\tlog.Fatal(\"webroot '\", webroot, \"' inaccessible: \", err)\n\t}\n\n\tif webroot != \"\" {\n\t\thandler := responders.PublicResponder(webroot)\n\t\thttp.Handle(\"/\", handler)\n\t}\n}", "func main() {\n\n\t// use custom handler\n\n\txhandler := routes.RegexpHandler{}\n\n\txhandler.HandleRegexp(\"/abc\\\\=.+\", handleAbc)\n\n\txhandler.HandleRegexp(\"/\", handleRoot)\n\n\thttp.Handle(\"/\", xhandler)\n\n\t// use net/http handler\n\n\thttp.HandleFunc(\"/zzz/\", handleSpec)\n\n\thttp.HandleFunc(\"/sublvl\", func(w http.ResponseWriter, rq *http.Request) {\n\t\thandleRoot(w, rq)\n\t})\n\n\tlog.Fatal(http.ListenAndServe(\":8080\", nil))\n}", "func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\n\tinitService(service)\n\tvar h goa.Handler\n\tservice.Mux.Handle(\"OPTIONS\", \"/swagger.json\", ctrl.MuxHandler(\"preflight\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\n\tservice.Mux.Handle(\"OPTIONS\", \"/swaggerui/*filepath\", ctrl.MuxHandler(\"preflight\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\n\n\th = ctrl.FileHandler(\"/swagger.json\", \"swagger/swagger.json\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swagger.json\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger/swagger.json\", \"route\", \"GET /swagger.json\")\n\n\th = ctrl.FileHandler(\"/swaggerui/*filepath\", \"swaggerui\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swaggerui/*filepath\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swaggerui\", \"route\", \"GET /swaggerui/*filepath\")\n\n\th = ctrl.FileHandler(\"/swaggerui/\", \"swaggerui/index.html\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swaggerui/\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swaggerui/index.html\", \"route\", \"GET /swaggerui/\")\n}", "func routeHandler(res http.ResponseWriter, req *http.Request) {\n\tpath := req.URL.Path\n\tswitch path {\n\tcase \"/\": HomeRoute(res, req)\n\tcase \"/test\": TestRoute(res, req)\n\tcase \"/login\": LoginRoute(res, req)\n\tcase \"/register\": RegisterRoute(res, req)\n case \"/logout\": LogoutRoute(res, req)\n\tcase \"/library\": LibraryRoute(res, req)\n\tcase \"/upload\": UploadRoute(res, req)\n\tcase \"/read\": ReadRoute(res, req)\n\tdefault: data,err := ioutil.ReadFile(path[1:])\n \tif err != nil{\n \t\t\tNotFoundRoute(res, req)\n \t} else {\n \t\t\tres.Write(data)\n \t}\n\t}\n}", "func (h *App) ServeHTTP(res http.ResponseWriter, req *http.Request) {\n\tvar head string\n\n\tswitch req.URL.Path {\n\tcase \"/\", \"/orgs\", \"/users\", \"/devices\", \"/sign-in\", \"/groups\", \"/msg\":\n\t\th.IndexHandler.ServeHTTP(res, req)\n\n\tdefault:\n\t\thead, req.URL.Path = ShiftPath(req.URL.Path)\n\t\tswitch head {\n\t\tcase \"public\":\n\t\t\th.PublicHandler.ServeHTTP(res, req)\n\t\tcase \"v1\":\n\t\t\th.V1ApiHandler.ServeHTTP(res, req)\n\t\tdefault:\n\t\t\thttp.Error(res, \"Not Found\", http.StatusNotFound)\n\t\t}\n\t}\n}", "func MountSwaggerController(service *goa.Service, ctrl SwaggerController) {\n\tinitService(service)\n\tvar h goa.Handler\n\tservice.Mux.Handle(\"OPTIONS\", \"/swagger-ui/*filepath\", ctrl.MuxHandler(\"preflight\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\n\tservice.Mux.Handle(\"OPTIONS\", \"/swagger.json\", ctrl.MuxHandler(\"preflight\", handleSwaggerOrigin(cors.HandlePreflight()), nil))\n\n\th = ctrl.FileHandler(\"/swagger-ui/*filepath\", \"swagger-ui/\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swagger-ui/*filepath\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger-ui/\", \"route\", \"GET /swagger-ui/*filepath\")\n\n\th = ctrl.FileHandler(\"/swagger.json\", \"swagger/swagger.json\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swagger.json\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger/swagger.json\", \"route\", \"GET /swagger.json\")\n\n\th = ctrl.FileHandler(\"/swagger-ui/\", \"swagger-ui/index.html\")\n\th = handleSwaggerOrigin(h)\n\tservice.Mux.Handle(\"GET\", \"/swagger-ui/\", ctrl.MuxHandler(\"serve\", h, nil))\n\tservice.LogInfo(\"mount\", \"ctrl\", \"Swagger\", \"files\", \"swagger-ui/index.html\", \"route\", \"GET /swagger-ui/\")\n}", "func HandleRoot(w http.ResponseWriter, r *http.Request) (err error) {\n\tw.Header().Set(\"Content-Type\", \"text/html; charset=utf-8\")\n\tw.WriteHeader(http.StatusOK)\n\tw.Write([]byte(`<html><body><a href=\"/login\">Login using Twitch</a></body></html>`))\n\n\treturn\n}", "func IndexHandler(ctx *Context, w http.ResponseWriter, r *http.Request) (int, error) {\n\tvar host string\n\tif r.Header.Get(\"X-Forwarded-Host\") != \"\" {\n\t\thost = r.Header.Get(\"X-Forwarded-Host\")\n\t} else {\n\t\thost = r.Host\n\t}\n\n\tvar hostConfig *gabs.Container\n\tvar ok bool\n\n\t// Check if host present in Config.\n\tchildren := ctx.Config.ChildrenMap()\n\tif hostConfig, ok = children[host]; !ok {\n\t\treturn 404, fmt.Errorf(\"Shortcut '%s' not found in Config.\", host)\n\t}\n\n\ttokens := tokenize(host + r.URL.Path)\n\n\t// Set up handles on token and Config. We might need to skip ahead if there's a custom schema set.\n\ttokensStart := tokens.Front()\n\tconf := ctx.Config\n\n\tvar path bytes.Buffer\n\tif s := hostConfig.Path(sslKey).Data(); s != nil && s.(bool) {\n\t\tpath.WriteString(httpPrefix)\n\t} else if s := hostConfig.Path(schemaKey).Data(); s != nil && s.(string) != \"\" {\n\t\tpath.WriteString(hostConfig.Path(schemaKey).Data().(string) + \":/\")\n\t\t// move one token ahead to parse expansions correctly.\n\t\tconf = conf.ChildrenMap()[tokensStart.Value.(string)]\n\t\ttokensStart = tokensStart.Next()\n\t} else {\n\t\t// Default to regular https prefix.\n\t\tpath.WriteString(httpsPrefix)\n\t}\n\n\tExpandPath(conf, tokensStart, &path)\n\n\t// send result\n\thttp.Redirect(w, r, path.String(), http.StatusFound)\n\n\treturn 302, nil\n}", "func (srv *Server) ApplyHandlers() {\n\tsrv.router.Handle(\"/*\", http.FileServer(http.Dir(\"./web\")))\n\tsrv.router.Get(\"/socket\", srv.socketHandler)\n}", "func homeHandler(env *Env, w http.ResponseWriter, r *http.Request) error {\n return renderTemplate(w, \"index\", \"base\", nil)\n}", "func Mount(mux goahttp.Muxer, h *Server) {\n\tMountRegisterHandler(mux, h.Register)\n\tMountRetrieveHandler(mux, h.Retrieve)\n\tMountHmrcCallbackHandler(mux, h.HmrcCallback)\n}", "func (mux *ServeMux) handler(host, path string) (h http.Handler, pattern string) {\n\tmux.mu.RLock()\n\tdefer mux.mu.RUnlock()\n\n\t// Host-specific pattern takes precedence over generic ones\n\tif mux.hosts {\n\t\th, pattern = mux.match(host + path)\n\t}\n\tif h == nil {\n\t\th, pattern = mux.match(path)\n\t}\n\tif h == nil {\n\t\th, pattern = http.NotFoundHandler(), \"\"\n\t}\n\treturn\n}", "func (mux *ServeMux) handler(host, path string) (h Handler, pattern string) {\n\tmux.mu.RLock()\n\tdefer mux.mu.RUnlock()\n\n\t// Host-specific pattern takes precedence over generic ones\n\tif mux.hosts {\n\t\th, pattern = mux.match(host + path)\n\t}\n\tif h == nil {\n\t\th, pattern = mux.match(path)\n\t}\n\tif h == nil {\n\t\th, pattern = NotFoundHandler(), \"\"\n\t}\n\treturn\n}", "func (r *Router) handle(c *Ctx) {\n\tvar handler HandlerFunc\n\treq := c.Request()\n\tw := c.Writer()\n\tpath := req.URL.Path\n\tmethod := req.Method\n\tres := r.trie.Match(path)\n\n\tif res.Node == nil {\n\t\t// FixedPathRedirect or TrailingSlashRedirect\n\t\tif res.TSR != \"\" || res.FPR != \"\" {\n\t\t\treq.URL.Path = res.TSR\n\t\t\tif res.FPR != \"\" {\n\t\t\t\treq.URL.Path = res.FPR\n\t\t\t}\n\t\t\tcode := 301\n\t\t\tif method != \"GET\" {\n\t\t\t\tcode = 307\n\t\t\t}\n\t\t\thttp.Redirect(w, req, req.URL.String(), code)\n\t\t\treturn\n\t\t}\n\t\tif r.noRoute == nil {\n\t\t\thttp.Error(w, fmt.Sprintf(`\"%s\" not implemented`, path), 501)\n\t\t\treturn\n\t\t}\n\t\thandler = r.noRoute\n\t} else {\n\t\t// ok := false\n\t\thd := res.Node.GetHandler(method)\n\t\thandler, _ = hd.(HandlerFunc)\n\t\t// handler = r.wrapHandler(hd)\n\t\t// if !ok {\n\t\t// \tpanic(\"handler error\")\n\t\t// }\n\t\tif handler == nil {\n\t\t\t// OPTIONS support\n\t\t\tif method == http.MethodOptions {\n\t\t\t\tw.Header().Set(\"Allow\", res.Node.GetAllow())\n\t\t\t\tw.WriteHeader(204)\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tif r.noMethod == nil {\n\t\t\t\t// If no route handler is returned, it's a 405 error\n\t\t\t\tw.Header().Set(\"Allow\", res.Node.GetAllow())\n\t\t\t\thttp.Error(w, fmt.Sprintf(`\"%s\" not allowed in \"%s\"`, method, path), 405)\n\t\t\t\treturn\n\t\t\t}\n\t\t\thandler = r.noMethod\n\t\t}\n\t}\n\n\tif len(res.Params) != 0 {\n\t\tc.params = res.Params\n\t}\n\tc.handlers = append(c.handlers, handler)\n\tc.Next()\n}", "func MountApplication(mountpoints map[string]http.Handler, host *Host, app *Application) {\n listing := &DirList{Host: host}\n\n\t// Serve the static build files from the mountpoint path.\n\turl := app.PublishUrl()\n\tlog.Printf(\"Serving app %s from %s\", url, app.PublicDirectory())\n fileserver := http.FileServer(http.Dir(app.PublicDirectory()))\n mountpoints[url] = http.StripPrefix(url, PublicHandler{Listing: listing, App: app, FileServer: fileserver})\n}", "func (s *Server) Start() *mux.Router {\n\trouter := mux.NewRouter()\n\n\trouter.HandleFunc(\"/\", s.Config.Index)\n\trouter.HandleFunc(\"/host/{host}\", s.Config.HostSummary)\n\trouter.HandleFunc(\"/group/{group}\", s.Config.GroupSummary)\n\trouter.PathPrefix(\"/\").Handler(http.FileServer(http.Dir(\"./assets/\")))\n\n\treturn router\n}", "func rootHandler(w http.ResponseWriter, r *http.Request) {\n\t//fmt.Fprintf(w, \"<h1>Hello All</h1>\")\n\tt,_ := template.ParseFiles(\"root.html\")\n\tt.Execute(w, nil)\n\n}", "func (s *server) routes() {\n s.router.HandleFunc(\"/\", s.homePageHandler)\n s.router.HandleFunc(\"/signup/\", s.signupHandler)\n s.router.HandleFunc(\"/signin/\", s.signinHandler)\n s.router.HandleFunc(\"/signout/\", s.makeHandler(s.signoutHandler))\n s.router.HandleFunc(\"/view/\", s.makeHandler(s.viewHandler))\n s.router.HandleFunc(\"/save/\", s.makeHandler(s.saveHandler))\n s.router.HandleFunc(\"/edit/\", s.makeHandler(s.editHandler))\n s.router.HandleFunc(\"/delete/\", s.makeHandler(s.deleteHandler))\n\n s.validPath = regexp.MustCompile(\n \"^/(new|view|save|edit|delete|signout)/([0-9]*)$\")\n}", "func (mux *ServeMux) Handler(r *Request) (h Handler, pattern string) {\n\n\t// CONNECT requests are not canonicalized.\n\tif r.Method == \"CONNECT\" {\n\t\t// If r.URL.Path is /tree and its handler is not registered,\n\t\t// the /tree -> /tree/ redirect applies to CONNECT requests\n\t\t// but the path canonicalization does not.\n\t\tif u, ok := mux.redirectToPathSlash(r.URL.Host, r.URL.Path, r.URL); ok {\n\t\t\treturn RedirectHandler(u.String(), StatusMovedPermanently), u.Path\n\t\t}\n\n\t\treturn mux.handler(r.Host, r.URL.Path)\n\t}\n\n\t// All other requests have any port stripped and path cleaned\n\t// before passing to mux.handler.\n\thost := stripHostPort(r.Host)\n\tpath := cleanPath(r.URL.Path)\n\n\t// If the given path is /tree and its handler is not registered,\n\t// redirect for /tree/.\n\tif u, ok := mux.redirectToPathSlash(host, path, r.URL); ok {\n\t\treturn RedirectHandler(u.String(), StatusMovedPermanently), u.Path\n\t}\n\n\tif path != r.URL.Path {\n\t\t_, pattern = mux.handler(host, path)\n\t\tu := &url.URL{Path: path, RawQuery: r.URL.RawQuery}\n\t\treturn RedirectHandler(u.String(), StatusMovedPermanently), pattern\n\t}\n\n\treturn mux.handler(host, r.URL.Path)\n}", "func (h *Host) initHostHandler() {\n\thttp.HandleFunc(h.url+\":\"+h.port+\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\th.reqLog.Access(r)\n\t\tw = h.addHeaders(w)\n\t\tif h.static != \"\" {\n\t\t\tif strings.HasPrefix(r.URL.Path, h.static) {\n\t\t\t\thttp.ServeFile(w, r, h.root+r.URL.Path)\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t\t//get next proxy\n\t\trewriteProxy := h.NextProxy()\n\t\tr.RequestURI = strings.Replace(r.RequestURI, \"/\", \"\", 1)\n\t\tr.URL.Path = strings.Replace(r.URL.Path, \"/\", \"\", 1)\n\t\trewriteProxy.ServeHTTP(w, r)\n\t})\n}", "func InitRouter() {\n\tr := mux.NewRouter()\n\n\tr.HandleFunc(\"/login\", service.Session{}.Login)\n\tr.HandleFunc(\"/logmein\", service.Session{}.LogMeIn).Methods(\"GET\")\n\tr.HandleFunc(\"/logout\", service.Session{}.Logout).Methods(\"GET\")\n\n\tr.HandleFunc(\"/faq\", func(w http.ResponseWriter, r *http.Request) {\n\t\thc := &kinli.HttpContext{W: w, R: r}\n\t\tpage := kinli.NewPage(hc, \"Frequently Asked Questions\", \"\", \"\", nil)\n\t\tkinli.DisplayPage(w, \"faq\", page)\n\t}).Methods(\"GET\")\n\n\tr.HandleFunc(\"/example\", func(w http.ResponseWriter, r *http.Request) {\n\t\thc := &kinli.HttpContext{W: w, R: r}\n\t\tpage := kinli.NewPage(hc, \"Example Form\", \"\", \"\", nil)\n\t\tkinli.DisplayPage(w, \"example\", page)\n\t}).Methods(\"GET\")\n\n\tr.HandleFunc(\"/\", func(w http.ResponseWriter, r *http.Request) {\n\t\thc := &kinli.HttpContext{W: w, R: r}\n\t\tpage := kinli.NewPage(hc, \"hello page\", \"\", \"\", nil)\n\t\tkinli.DisplayPage(w, \"home\", page)\n\t}).Methods(\"GET\")\n\n\tr.HandleFunc(\"/{uid}\", service.FormSubmissionRequest).Methods(\"POST\")\n\n\tr.NotFoundHandler = http.HandlerFunc(notFound)\n\n\tinitStatic(r)\n\n\tsrv := &http.Server{\n\t\tHandler: r,\n\t\tAddr: common.Config.LocalServer,\n\t\tWriteTimeout: 60 * time.Second,\n\t\tReadTimeout: 60 * time.Second,\n\t}\n\tlog.Println(\"Starting server on\", common.Config.LocalServer)\n\tlog.Fatal(srv.ListenAndServe())\n\n}", "func (this *Router) Handle(title, path string, handlers ...HandlerFunc) *Router {\n\tif len(path) < 1 || path[0] != '/' || strings.Contains(path, \"//\") {\n\t\tpanic(\"add router faild, invalid path \" + path)\n\t}\n\tif sepIndex := strings.Index(path[1:], \"/\") + 1; sepIndex > 1 {\n\t\troot := path[:sepIndex]\n\t\tsubpath := path[sepIndex:]\n\t\tvar group *Router = nil\n\t\tfor _, router := range this.children {\n\t\t\tif router.path == root {\n\t\t\t\tgroup = router\n\t\t\t}\n\t\t}\n\t\tif group == nil {\n\t\t\tgroup = this.Group(root)\n\t\t}\n\t\treturn group.Handle(title, subpath, handlers...)\n\t}\n\thandlerChain := append([]HandlerFunc{}, this.handlerChain...)\n\thandlerChain = append(handlerChain, handlers...)\n\trouter := &Router{\n\t\ttitle: title,\n\t\tpath: path,\n\t\trealPath: this.realPath + path,\n\t\thandlerChain: handlerChain,\n\t\tchildren: []*Router{},\n\t}\n\tthis.children = append(this.children, router)\n\tlog.Log(\"DEBUG\", \"add router\", router.realPath)\n\treturn router\n}", "func (_m *RESTHandler) Mount(group *echo.Group) {\n\t_m.Called(group)\n}", "func (osh *SystemHandler) Mount(source string, target string, fsType string, flags uintptr, data string) error {\n\treturn syscall.Mount(source, target, fsType, flags, data)\n}", "func (hr *httpRouter) Handler() http.Handler {\n\n\tc, _ := console.New(console.Options{Color: true})\n\t_ = logger.Register(\"console\", logger.Config{Writer: c})\n\tcLogger, _ := logger.Get(\"console\")\n\tl := log.New(cLogger)\n\n\tfmt.Print(\"Loading Routes...\")\n\t//add files in a directory\n\tro := newHttpRouterExtended(hr)\n\n\tmw := middleware.Chain{}\n\n\t//adding files\n\tfor path, file := range hr.file {\n\t\tro.HandlerFunc(\"GET\", path, mw.Add(l.MW).Handle(\n\t\t\tfunc(w http.ResponseWriter, req *http.Request) {\n\t\t\t\thttp.ServeFile(w, req, hr.file[req.Context().Value(router.PATTERN).(string)])\n\t\t\t}))\n\t\tfmt.Printf(\"\\n\\x1b[32m %#v [GET]%v \\x1b[49m\\x1b[39m \", path, file)\n\t}\n\n\t// adding directories\n\tfor k, path := range hr.dir {\n\t\tfileServer := http.FileServer(http.Dir(path))\n\t\tpattern := k + \"/*filepath\"\n\t\tro.HandlerFunc(\"GET\", pattern, mw.Add(l.MW).Handle(\n\t\t\tfunc(w http.ResponseWriter, req *http.Request) {\n\t\t\t\t//disable directory listing\n\t\t\t\tif strings.HasSuffix(req.URL.Path, \"/\") {\n\t\t\t\t\thttp.NotFound(w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif val, ok := req.Context().Value(router.PARAMS).(map[string][]string)[\"filepath\"]; ok {\n\t\t\t\t\treq.URL.Path = val[0]\n\t\t\t\t\tfileServer.ServeHTTP(w, req)\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\thttp.NotFound(w, req)\n\t\t\t\treturn\n\n\t\t\t}))\n\t\tfmt.Printf(\"\\n\\x1b[32m %#v [GET]%v \\x1b[49m\\x1b[39m \", pattern, http.Dir(path))\n\t}\n\n\t//register all controller routes\n\tfor _, r := range hr.routes {\n\t\tfmt.Printf(\"\\n\\x1b[32m %#v :name \\x1b[49m\\x1b[39m \", r.pattern)\n\t\tfor method, fn := range r.controller.MappingBy(r.pattern) {\n\t\t\tif r.mws != nil {\n\t\t\t\tro.HandlerFunc(strings.ToUpper(method), r.pattern, r.mws.Handle(r.controller.ServeHTTP)) //TODO ????? error no url pattern\n\t\t\t} else {\n\t\t\t\tro.HandlerFunc(strings.ToUpper(method), r.pattern, r.controller.ServeHTTP)\n\t\t\t}\n\t\t\tfmt.Printf(\"\\x1b[32m [%v]%v name \\x1b[49m\\x1b[39m \", method, fn)\n\t\t}\n\t}\n\n\t//Not Found Handler\n\tif hr.notFound != nil {\n\t\tro.NotFound = hr.notFound\n\t}\n\n\treturn ro\n}", "func rootHandler(w http.ResponseWriter, r *http.Request, title string) {\n\tp, err := loadRoot(title)\n\n\tp.Body = template.HTML(blackfriday.MarkdownCommon([]byte(p.Body)))\n\tp.Body = template.HTML(convertWikiMarkup([]byte(p.Body)))\n\n\terr = templates.ExecuteTemplate(w, \"root.html\", p)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusInternalServerError)\n\t}\n}", "func (mux *Mux) Handler(r *Request) Handler {\n\tscheme := r.URL.Scheme\n\thost := r.URL.Hostname()\n\tpath := cleanPath(r.URL.Path)\n\n\t// If the given path is /tree and its handler is not registered,\n\t// redirect for /tree/.\n\tif u, ok := mux.redirectToPathSlash(muxKey{scheme, host, path}, r.URL); ok {\n\t\treturn StatusHandler(StatusPermanentRedirect, u.String())\n\t}\n\n\tif path != r.URL.Path {\n\t\tu := *r.URL\n\t\tu.Path = path\n\t\treturn StatusHandler(StatusPermanentRedirect, u.String())\n\t}\n\n\tmux.mu.RLock()\n\tdefer mux.mu.RUnlock()\n\n\th := mux.match(muxKey{scheme, host, path})\n\tif h == nil {\n\t\t// Try wildcard\n\t\tif wildcard, ok := getWildcard(host); ok {\n\t\t\th = mux.match(muxKey{scheme, wildcard, path})\n\t\t}\n\t}\n\tif h == nil {\n\t\th = NotFoundHandler()\n\t}\n\treturn h\n}", "func MountCORSHandler(mux goahttp.Muxer, h http.Handler) {\n\th = handleStationOrigin(h)\n\tf, ok := h.(http.HandlerFunc)\n\tif !ok {\n\t\tf = func(w http.ResponseWriter, r *http.Request) {\n\t\t\th.ServeHTTP(w, r)\n\t\t}\n\t}\n\tmux.Handle(\"OPTIONS\", \"/stations\", f)\n\tmux.Handle(\"OPTIONS\", \"/stations/{id}\", f)\n\tmux.Handle(\"OPTIONS\", \"/stations/@/{id}\", f)\n\tmux.Handle(\"OPTIONS\", \"/projects/{id}/stations\", f)\n\tmux.Handle(\"OPTIONS\", \"/stations/{id}/photo\", f)\n}", "func (e *Engine) setupFallback() {\n\te.srv.HTTPErrorHandler = func(err error, c echo.Context) {\n\t\tif err != echo.ErrNotFound {\n\t\t\treturn\n\t\t}\n\t\turi := genericPath(c.Request().RequestURI)\n\t\tfs, err := os.Stat(uri)\n\t\tif err != nil {\n\t\t\tc.Error(err)\n\t\t\treturn\n\t\t}\n\n\t\tif fs.IsDir() {\n\t\t\t// if working at pwd, capture all dirs\n\t\t\tif len(e.dirs) == 1 && e.dirs[0] == \".\" {\n\t\t\t\t// check if file excluded\n\t\t\t\tif isExclude(uri) {\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t\tif err := e.doDispatch(c, fs); err != nil {\n\t\t\t\t\tc.Error(err)\n\t\t\t\t}\n\t\t\t}\n\t\t\t// only capture missing dir when serving pwd\n\t\t\t// return for any other dirs\n\t\t\treturn\n\t\t}\n\n\t\t// capture any files here\n\t\terr = e.doServeFile(uri)(c)\n\t\tif err != nil {\n\t\t\tc.Error(err)\n\t\t}\n\t}\n}", "func NewHandler(s *Server) *Handler {\n\th := &Handler{\n\t\tserver: s,\n\t\tmux: pat.New(),\n\t}\n\n\t// Series routes.\n\th.mux.Get(\"/db/:db/series\", http.HandlerFunc(h.serveQuery))\n\th.mux.Post(\"/db/:db/series\", http.HandlerFunc(h.serveWriteSeries))\n\th.mux.Del(\"/db/:db/series/:series\", http.HandlerFunc(h.serveDeleteSeries))\n\th.mux.Get(\"/db\", http.HandlerFunc(h.serveDatabases))\n\th.mux.Post(\"/db\", http.HandlerFunc(h.serveCreateDatabase))\n\th.mux.Del(\"/db/:name\", http.HandlerFunc(h.serveDeleteDatabase))\n\n\t// Cluster admins routes.\n\th.mux.Get(\"/cluster_admins/authenticate\", http.HandlerFunc(h.serveAuthenticateClusterAdmin))\n\th.mux.Get(\"/cluster_admins\", http.HandlerFunc(h.serveClusterAdmins))\n\th.mux.Post(\"/cluster_admins\", http.HandlerFunc(h.serveCreateClusterAdmin))\n\th.mux.Post(\"/cluster_admins/:user\", http.HandlerFunc(h.serveUpdateClusterAdmin))\n\th.mux.Del(\"/cluster_admins/:user\", http.HandlerFunc(h.serveDeleteClusterAdmin))\n\n\t// Database users routes.\n\th.mux.Get(\"/db/:db/authenticate\", http.HandlerFunc(h.serveAuthenticateDBUser))\n\th.mux.Get(\"/db/:db/users\", http.HandlerFunc(h.serveDBUsers))\n\th.mux.Post(\"/db/:db/users\", http.HandlerFunc(h.serveCreateDBUser))\n\th.mux.Get(\"/db/:db/users/:user\", http.HandlerFunc(h.serveDBUser))\n\th.mux.Post(\"/db/:db/users/:user\", http.HandlerFunc(h.serveUpdateDBUser))\n\th.mux.Del(\"/db/:db/users/:user\", http.HandlerFunc(h.serveDeleteDBUser))\n\n\t// Utilities\n\th.mux.Get(\"/ping\", http.HandlerFunc(h.servePing))\n\th.mux.Get(\"/interfaces\", http.HandlerFunc(h.serveInterfaces))\n\n\t// Shard routes.\n\th.mux.Get(\"/cluster/shards\", http.HandlerFunc(h.serveShards))\n\th.mux.Post(\"/cluster/shards\", http.HandlerFunc(h.serveCreateShard))\n\th.mux.Del(\"/cluster/shards/:id\", http.HandlerFunc(h.serveDeleteShard))\n\n\t// Shard space routes.\n\th.mux.Get(\"/cluster/shard_spaces\", http.HandlerFunc(h.serveShardSpaces))\n\th.mux.Post(\"/cluster/shard_spaces/:db\", http.HandlerFunc(h.serveCreateShardSpace))\n\th.mux.Post(\"/cluster/shard_spaces/:db/:name\", http.HandlerFunc(h.serveUpdateShardSpace))\n\th.mux.Del(\"/cluster/shard_spaces/:db/:name\", http.HandlerFunc(h.serveDeleteShardSpace))\n\n\t// Cluster config endpoints\n\th.mux.Get(\"/cluster/servers\", http.HandlerFunc(h.serveServers))\n\th.mux.Del(\"/cluster/servers/:id\", http.HandlerFunc(h.serveDeleteServer))\n\n\treturn h\n}", "func HttpRootHandler(w http.ResponseWriter, r *http.Request) {\n\tcmd := CommandStruct{}\n\n\tbs, err := ioutil.ReadAll(r.Body)\n\tif err != nil {\n\t\treturn\n\t}\n\t_ = json.Unmarshal(bs, &cmd)\n\tswitch cmd.Command {\n\tcase \"check\":\n\t\tfmt.Fprintln(w, \"HTTP-Server is online.\")\n\t\tbreak\n\n\tcase \"disc\":\n\t\tfmt.Fprintln(w, \"HTTP-Server shutting down...\")\n\t\thttpAlive <- 1\n\t\tbreak\n\n\tcase \"lista\":\n\t\tfmt.Fprintln(w, \"List of all available NBD-devices:\")\n\t\tfor key, value := range AvailableList {\n\t\t\tif AvailableList[key] != \"\" {\n\t\t\t\tfmt.Fprintln(w, value)\n\t\t\t}\n\t\t}\n\t\tbreak\n\n\tcase \"listm\":\n\t\tfmt.Fprintln(w, \"List of all mounted NBD-devices:\")\n\t\tfor key, value := range MountedList {\n\t\t\tfmt.Fprintln(w, key+\"\\t\"+value)\n\t\t}\n\t\tbreak\n\n\tcase \"mount\":\n\t\tif strings.Contains(cmd.Device, \"/dev/nbd\") {\n\t\t\tfor i := 0; i < len(AvailableList); i++ {\n\t\t\t\tif AvailableList[i] == cmd.Device {\n\n\t\t\t\t\tLinkedLogins[len(LinkedLogins)+1], err = nethandler.SetupConnection(cmd.Image, cmd.User, cmd.Pass, cmd.Device)\n\t\t\t\t\tif err != nil {\n\t\t\t\t\t\tfmt.Fprintf(w, \"Error: \", err)\n\t\t\t\t\t\tfmt.Fprintf(w, \"\\n\")\n\t\t\t\t\t\treturn\n\t\t\t\t\t}\n\n\t\t\t\t\tAddToMountedList(cmd.Device, cmd.Image)\n\t\t\t\t\tfmt.Fprintf(w, \"Successfully mounted \"+cmd.Image+\" to \"+cmd.Device+\"\\n\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tfor _, value := range AvailableList {\n\t\t\t\tif value != \"\" {\n\t\t\t\t\tAddToMountedList(value, cmd.Image)\n\t\t\t\t\tfmt.Fprintf(w, \"Device \"+cmd.Device+\" is already mounted.\\n\"+cmd.Image+\" has been mounted to \"+value+\" instead.\\n\")\n\t\t\t\t\treturn\n\t\t\t\t}\n\t\t\t}\n\t\t\tfmt.Fprintf(w, \"No more devices available!\\n\")\n\t\t} else {\n\t\t\tfmt.Fprintf(w, \"Specified device not recognised.\\n\")\n\t\t}\n\t\tbreak\n\n\tcase \"unmount\":\n\t\t//TODO Real unmounting of NBD-devices\n\t\tfor key, _ := range AvailableList {\n\t\t\tif AvailableList[key] == \"\" {\n\t\t\t\tdelete(MountedList, cmd.Device)\n\t\t\t\tAvailableList[key] = cmd.Device\n\t\t\t\tfmt.Fprint(w, \"Successfully unmounted \"+cmd.Device)\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tbreak\n\t}\n}", "func HTTPHandler(router *mux.Router) {\n\trouter.Handle(\"/\", ImageHandler())\n\trouter.PathPrefix(\"/unsafe/\").Handler(ImageHandler())\n\trouter.Handle(\"/healthcheck\", HealthcheckHandler())\n}", "func Mount(router *mux.Router) {\n\n}", "func MountPhotoHandler(mux goahttp.Muxer, h http.Handler) {\n\tf, ok := handleStationOrigin(h).(http.HandlerFunc)\n\tif !ok {\n\t\tf = func(w http.ResponseWriter, r *http.Request) {\n\t\t\th.ServeHTTP(w, r)\n\t\t}\n\t}\n\tmux.Handle(\"GET\", \"/stations/{id}/photo\", f)\n}", "func indexHandler(w http.ResponseWriter, r *http.Request) {\n\tif r.URL.Path != \"/\" {\n\t\thttp.NotFound(w, r)\n\t\treturn\n\t}\n\n\thttp.ServeFile(w, r, \"pages/main.html\")\n}", "func pathHandler(w http.ResponseWriter, r *http.Request) {\n\tpath := \"/var/www/ear7h-net/\" + r.URL.Path[1:]\n\t//open file and send\n\tf, err := os.Open(path)\n\tif err != nil {\n\t\tfourOhFour(w, r)\n\t} else {\n\t\thttp.ServeContent(w, r, r.URL.Path, time.Now(), f)\n\t}\n}", "func (r *Router) masterHandler(c *gin.Context) {\n\tklog.V(4).Infof(\"no router for method:%s, url:%s\", c.Request.Method, c.Request.URL.Path)\n\tc.JSON(404, gin.H{\n\t\t\"Method\": c.Request.Method,\n\t\t\"Path\": c.Request.URL.Path,\n\t\t\"error\": \"router not found\"})\n}", "func Serve(root, iface string, port int) (err error) {\n\t// Root the path, and clean it if necessary.\n\n\t// 18/01/2013 It might make sense to move this to a helper routine\n\t// or further up in the stack.\n\tif !path.IsAbs(root) {\n\t\tvar wd string\n\t\twd, err = os.Getwd()\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\troot = path.Join(wd, root)\n\t} else {\n\t\troot = path.Clean(root)\n\t}\n\tl.Printf(\"Starting http server %s:%d\\nRoot path %q\", iface, port, root)\n\trootpath = root\n\n\thttp.HandleFunc(\"/p/\", handle)\n\thttp.ListenAndServe(iface+\":\"+strconv.Itoa(port), nil)\n\treturn\n}", "func (self *CentralBooking) InstallHandlers(router *mux.Router) {\n router.\n Methods(\"POST\").\n Path(\"/register/instance\").\n HandlerFunc(self.RegisterInstance)\n\n // apeing vault\n router.\n Methods(\"GET\").\n Path(\"/sys/health\").\n HandlerFunc(self.CheckHealth)\n}", "func (r *SilencedRouter) Mount(parent *mux.Router) {\n\troutes := ResourceRoute{\n\t\tRouter: parent,\n\t\tPathPrefix: \"/namespaces/{namespace}/{resource:silenced}\",\n\t}\n\n\thandlers := handlers.NewHandlers[*corev2.Silenced](r.store)\n\n\troutes.Del(handlers.DeleteResource)\n\troutes.Get(r.get)\n\troutes.Post(r.create)\n\troutes.Put(r.createOrReplace)\n\troutes.List(r.listr, corev3.SilencedFields)\n\troutes.ListAllNamespaces(r.listr, \"/{resource:silenced}\", corev3.SilencedFields)\n\n\t// Custom routes for listing by subscription and checks for a specific\n\t// namespace, in addition to all namespaces for checks.\n\troutes.Router.HandleFunc(\"/{resource:silenced}/checks/{check}\", listHandler(r.list)).Methods(http.MethodGet)\n\troutes.Router.HandleFunc(routes.PathPrefix+\"/subscriptions/{subscription}\", listHandler(r.list)).Methods(http.MethodGet)\n\troutes.Router.HandleFunc(routes.PathPrefix+\"/checks/{check}\", listHandler(r.list)).Methods(http.MethodGet)\n}" ]
[ "0.702594", "0.6997063", "0.6954332", "0.6820072", "0.65756345", "0.6451301", "0.6383405", "0.6365718", "0.6285303", "0.62754136", "0.6251029", "0.6157279", "0.6116233", "0.60878783", "0.6076186", "0.60656357", "0.6065411", "0.6024053", "0.6022069", "0.60135174", "0.60126704", "0.6005589", "0.6004503", "0.60038495", "0.5991979", "0.59673274", "0.59673274", "0.59673274", "0.5966083", "0.5946025", "0.5934402", "0.593274", "0.59320784", "0.5915197", "0.5913959", "0.5913508", "0.5905052", "0.5902104", "0.5896477", "0.5896477", "0.58728826", "0.5862679", "0.5854989", "0.58452594", "0.5838185", "0.58378357", "0.58346957", "0.582368", "0.5816042", "0.5792327", "0.57894605", "0.5784616", "0.57818025", "0.57736343", "0.5773397", "0.57706916", "0.57699347", "0.5760902", "0.5760156", "0.57574904", "0.5721992", "0.57217157", "0.57118344", "0.57063615", "0.56997764", "0.56936055", "0.5681272", "0.56789786", "0.5678943", "0.5676648", "0.56746376", "0.5670061", "0.5669512", "0.5663938", "0.56606144", "0.5650369", "0.5646647", "0.564429", "0.5644125", "0.56419843", "0.5641957", "0.5628667", "0.56214565", "0.5611402", "0.5610784", "0.56051207", "0.56039834", "0.55896366", "0.5584176", "0.5578367", "0.5568307", "0.556564", "0.5561139", "0.5555659", "0.55513", "0.5548231", "0.5545511", "0.55443543", "0.5543333", "0.5541932", "0.5530312" ]
0.0
-1
SiaPath returns the siapath of a remote directory.
func (rd *RemoteDir) SiaPath() modules.SiaPath { return rd.siapath }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (sd *SiaDir) SiaPath() modules.SiaPath {\n\tsd.mu.Lock()\n\tdefer sd.mu.Unlock()\n\treturn sd.siaPath\n}", "func (rf *RemoteFile) SiaPath() modules.SiaPath {\n\trf.mu.Lock()\n\tdefer rf.mu.Unlock()\n\treturn rf.siaPath\n}", "func (o *Object) remotePath() string {\n\treturn o.fs.slashRootSlash + o.remote\n}", "func newSiaPath(path string) modules.SiaPath {\n\tsp, err := modules.NewSiaPath(path)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\treturn sp\n}", "func (b *Binary) RemotePath() string {\n\treturn b.file.RemotePath()\n}", "func (r *Repository) Path(id digest.Digest) (dir, path string) {\n\tdir = filepath.Join(r.Root, id.Hex()[:2])\n\treturn dir, filepath.Join(dir, id.Hex()[2:])\n}", "func (proxy *remoteDriverProxy) Path(name string) (string, error) {\n\tvar req = remoteVolumePathReq{\n\t\tName: name,\n\t}\n\n\tvar resp remoteVolumePathResp\n\n\tif err := proxy.client.CallService(remoteVolumePathService, &req, &resp, true); err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif resp.Err != \"\" {\n\t\treturn \"\", errors.New(resp.Err)\n\t}\n\n\treturn resp.Mountpoint, nil\n}", "func LoadSiaDir(rootDir string, siaPath modules.SiaPath, deps modules.Dependencies, wal *writeaheadlog.WAL) (sd *SiaDir, err error) {\n\tsd = &SiaDir{\n\t\tdeps: deps,\n\t\tsiaPath: siaPath,\n\t\trootDir: rootDir,\n\t\twal: wal,\n\t}\n\tsd.metadata, err = loadSiaDirMetadata(siaPath.SiaDirMetadataSysPath(rootDir), modules.ProdDependencies)\n\treturn sd, err\n}", "func (r *RepoRef) SkeletonsPath() string {\n\treturn filepath.Join(r.LocalPath(), SkeletonsDir)\n}", "func (*HttpLocalSites) GetPath() string { return \"/api/objects/http/local_site/\" }", "func (server *SingleInstance) Path() string {\n\treturn server.path\n}", "func (d Driver) getSecureDiffPath(id, parent string, canBeRemote bool) string {\n\tvar diffDirName string\n\n\tif parent == \"\" || d.isParent(id, parent) {\n\t\tdiffDirName = \"diff\"\n\t} else {\n\t\tdiffDirName = fmt.Sprintf(\"%s-%s\", \"diff\", parent)\n\t}\n\n\tlocalSecureDiffPath := path.Join(d.dir(id), constSecureBaseDirName, diffDirName)\n\tremoteSecureDiffPath := path.Join(d.options.remoteDir, id, constSecureBaseDirName, diffDirName)\n\tlogrus.Debugf(\"secureoverlay2: getSecureDiffPath %s. localSecureDiffPath %s remoteSecureDiffPath\", localSecureDiffPath, remoteSecureDiffPath)\n\tdiffPath := localSecureDiffPath\n\t// remote only \"wins\" if local does not exist and remote exists\n\tif canBeRemote && d.options.remoteDir != \"\" {\n\t\tif b, _ := exists(localSecureDiffPath); !b {\n\t\t\tif b, _ := exists(remoteSecureDiffPath); b {\n\t\t\t\tdiffPath = remoteSecureDiffPath\n\t\t\t}\n\t\t}\n\t}\n\tlogrus.Debugf(\"secureoverlay2: getSecureDiffPath w. id: %s, parent: %s, canBeRemote: %v returns %s\",\n\t\tid, parent, canBeRemote, diffPath)\n\n\treturn diffPath\n}", "func (d *driver) fullPath(path string) string {\n\treturn _path.Join(\"/ipfs\", d.roothash, path)\n}", "func (o ApplicationStatusOperationStateSyncResultSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (d *Datastore) s3Path(p string) string {\n\treturn path.Join(d.RootDirectory, p)\n}", "func (p *Part) RemotePath(prefix string) string {\n\tfor strings.HasSuffix(prefix, \"/\") {\n\t\tprefix = prefix[:len(prefix)-1]\n\t}\n\treturn fmt.Sprintf(\"%s/%s/%016X_%016X_%016X\", prefix, p.Path, p.FileSize, p.Offset, p.Size)\n}", "func (o ApplicationStatusOperationStateOperationSyncSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (fi *fileInfo) Path() string {\n\treturn fi.fullPath\n}", "func (ss *Sources) workspacePath(id consumerID) string {\n\treturn filepath.Join(ss.RootPath, \"workspace\", id.Namespace, id.Name, id.consumer)\n}", "func (s *Store) Path() string { return s.path }", "func (s *Store) Path() string { return s.path }", "func BattleShiritoriPath(battleID string, userID string) string {\n\treturn fmt.Sprintf(\"/streams/battles/%v/%v\", battleID, userID)\n}", "func (r *Resolver) Path(resource interface{}, id string) (path string, err error) {\n\tprovider := r.Provider\n\tswitch resource.(type) {\n\tcase *Provider:\n\t\tr := Provider{}\n\t\tr.UID = id\n\t\tr.Link()\n\t\tpath = r.SelfLink\n\tcase *Folder:\n\t\tr := Folder{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Datacenter:\n\t\tr := Datacenter{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Cluster:\n\t\tr := Cluster{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Host:\n\t\tr := Host{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Network:\n\t\tr := Network{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Datastore:\n\t\tr := Datastore{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *VM:\n\t\tr := VM{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tcase *Workload:\n\t\tr := Workload{}\n\t\tr.ID = id\n\t\tr.Link(provider)\n\t\tpath = r.SelfLink\n\tdefault:\n\t\terr = liberr.Wrap(\n\t\t\tbase.ResourceNotResolvedError{\n\t\t\t\tObject: resource,\n\t\t\t})\n\t}\n\n\tpath = strings.TrimRight(path, \"/\")\n\n\treturn\n}", "func (o ApplicationStatusOperationStateSyncResultSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusOperationStateSyncResultSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (sds *SiaDirSet) NewSiaDir(siaPath string) (*SiaDirSetEntry, error) {\n\tsds.mu.Lock()\n\tdefer sds.mu.Unlock()\n\t// Check is SiaDir already exists\n\tsiaPath = strings.Trim(siaPath, \"/\")\n\texists, err := sds.exists(siaPath)\n\tif exists {\n\t\treturn nil, ErrPathOverload\n\t}\n\tif !os.IsNotExist(err) && err != nil {\n\t\treturn nil, err\n\t}\n\tsd, err := New(siaPath, sds.rootDir, sds.wal)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tentry := sds.newSiaDirSetEntry(sd)\n\tthreadUID := randomThreadUID()\n\tentry.threadMap[threadUID] = newThreadType()\n\tsds.siaDirMap[siaPath] = entry\n\treturn &SiaDirSetEntry{\n\t\tsiaDirSetEntry: entry,\n\t\tthreadUID: threadUID,\n\t}, nil\n}", "func suriPath() (exists bool) {\n var err error\n path, err := utils.GetKeyValueString(\"suriPath\", \"path\")\n if err != nil {\n logs.Error(\"suriPath Error getting data from main.conf\")\n }\n\n if _, err := os.Stat(path); os.IsNotExist(err) {\n logs.Error(\"Suricata not installed, at least folder /etc/suricata dosn't exist\")\n return false\n }\n return true\n}", "func (o TransferJobTransferSpecAwsS3DataSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecAwsS3DataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (driver *Driver) Path(volumeName, volumeID string) (string, error) {\n\tif volumeName == \"\" && volumeID == \"\" {\n\t\treturn \"\", errors.New(\"Missing volume name or ID\")\n\t}\n\n\tinstances, err := driver.sdm.GetInstance()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tswitch {\n\tcase len(instances) == 0:\n\t\treturn \"\", errors.New(\"No instances\")\n\tcase len(instances) > 1:\n\t\treturn \"\", errors.New(\"Too many instances returned, limit the storagedrivers\")\n\t}\n\n\tvolumes, err := driver.sdm.GetVolume(volumeID, volumeName)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tswitch {\n\tcase len(volumes) == 0:\n\t\treturn \"\", errors.New(\"No volumes returned by name\")\n\tcase len(volumes) > 1:\n\t\treturn \"\", errors.New(\"Multiple volumes returned by name\")\n\t}\n\n\tvolumeAttachment, err := driver.sdm.GetVolumeAttach(volumes[0].VolumeID, instances[0].InstanceID)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif len(volumeAttachment) == 0 {\n\t\treturn \"\", nil\n\t}\n\n\tmounts, err := driver.osdm.GetMounts(volumeAttachment[0].DeviceName, \"\")\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\tif len(mounts) == 0 {\n\t\treturn \"\", nil\n\t}\n\n\treturn mounts[0].Mountpoint, nil\n}", "func (o IopingSpecVolumeVolumeSourceProjectedSourcesSecretItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesSecretItems) string { return v.Path }).(pulumi.StringOutput)\n}", "func (i *Instance) Path() (string, error) {\n\tref, _, _, err := i.GetAsAny(WmiPathKey)\n\treturn ref.(string), err\n}", "func (s *Store) IDPath() string { return filepath.Join(s.path, \"id\") }", "func (d ImagefsDriver) Path(r *volume.PathRequest) (*volume.PathResponse, error) {\n\tgetReq := volume.GetRequest{\n\t\tName: r.Name,\n\t}\n\tret, err := d.Get(&getReq)\n\tvar _ret *volume.PathResponse\n\tif ret != nil {\n\t\t_ret = &volume.PathResponse{\n\t\t\tMountpoint: ret.Volume.Mountpoint,\n\t\t}\n\t}\n\treturn _ret, err\n}", "func getRepoPath() (string, error) {\n\t// Set default base path and directory name\n\tdirectoryName := \".saturn\"\n\n\t// Join the path and directory name, then expand the home path\n\tfullPath, err := homedir.Expand(filepath.Join(\"~\", directoryName))\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Return the shortest lexical representation of the path\n\treturn filepath.Clean(fullPath), nil\n}", "func PantahubS3Path() string {\n\tif GetEnv(EnvPantahubStorageDriver) == \"s3\" {\n\t\treturn GetEnv(EnvPantahubStoragePath)\n\t}\n\n\tbasePath := path.Join(GetEnv(EnvPantahubS3Path), GetEnv(EnvPantahubStoragePath))\n\n\tif basePath == \"\" {\n\t\tbasePath = \".\"\n\t}\n\n\treturn basePath\n}", "func (c *Credentials) path() (string, error) {\n\tusr, err := user.Current()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to get current user\")\n\t}\n\treturn path.Join(usr.HomeDir, *cred), nil\n}", "func (ls *LocationStore) Path() string {\n\treturn fmt.Sprintf(\"env/locations/%s.yml\", ls.name)\n}", "func TestNewSiaDir(t *testing.T) {\n\tif testing.Short() && !build.VLONG {\n\t\tt.SkipNow()\n\t}\n\tt.Parallel()\n\t// Create filesystem.\n\troot := filepath.Join(testDir(t.Name()), \"fs-root\")\n\tfs := newTestFileSystem(root)\n\t// Create dir /sub/foo\n\tsp := newSiaPath(\"sub/foo\")\n\tif err := fs.NewSiaDir(sp, modules.DefaultDirPerm); err != nil {\n\t\tt.Fatal(err)\n\t}\n\t// The whole path should exist.\n\tif _, err := os.Stat(filepath.Join(root, sp.String())); err != nil {\n\t\tt.Fatal(err)\n\t}\n}", "func (s *session) Path() dbus.ObjectPath {\n\treturn s.path\n}", "func (o ApplicationStatusOperationStateOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationStatusOperationStateOperationSyncSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func Path(named Named) (name string) {\n\tif r, ok := named.(namedRepository); ok {\n\t\treturn r.Path()\n\t}\n\t_, path := splitDomain(named.Name())\n\treturn path\n}", "func (c *Client) Path() (string, error) {\n\treturn c.GetProperty(\"path\")\n}", "func (b *Bucket) Path() (string, error) {\n\tconf := b.conf.Viper.ConfigFileUsed()\n\tif conf == \"\" {\n\t\treturn b.cwd, nil\n\t}\n\treturn filepath.Dir(filepath.Dir(conf)), nil\n}", "func (r *RemoteSyslogServer) GetPath() string {\n\treturn fmt.Sprintf(\"/api/objects/remote_syslog/server/%s\", r.Reference)\n}", "func (h *HttpLocalSite) GetPath() string {\n\treturn fmt.Sprintf(\"/api/objects/http/local_site/%s\", h.Reference)\n}", "func (r *ServiceLinkedRole) Path() pulumi.StringOutput {\n\treturn (pulumi.StringOutput)(r.s.State[\"path\"])\n}", "func (k *Key) path() (string, error) {\n\tusr, err := user.Current()\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"failed to get current user\")\n\t}\n\treturn path.Join(usr.HomeDir, *identity), nil\n}", "func (d *DirectHash) getPath() string {\n\tdomain := d.RootDomain\n\tif len(d.ReleaseBranch) > 0 {\n\t\tdomain = string(d.ReleaseBranch) + \".\" + d.RootDomain\n\t}\n\treturn fmt.Sprintf(\"https://%s/\"+strings.Trim(d.Path, \"/\"), domain, d.ReleaseBranch.String())\n}", "func (id ID) Path() string {\n\treturn id.path\n}", "func (p *SeriesPartition) Path() string { return p.path }", "func (o *StorageNetAppCifsShareAllOf) GetPath() string {\n\tif o == nil || o.Path == nil {\n\t\tvar ret string\n\t\treturn ret\n\t}\n\treturn *o.Path\n}", "func (o FioSpecVolumeVolumeSourceProjectedSourcesSecretItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesSecretItems) string { return v.Path }).(pulumi.StringOutput)\n}", "func (m *CgroupfsManager) SandboxCgroupPath(sbParent, sbID string) (cgParent, cgPath string, _ error) {\n\tif strings.HasSuffix(path.Base(sbParent), \".slice\") {\n\t\treturn \"\", \"\", fmt.Errorf(\"cri-o configured with cgroupfs cgroup manager, but received systemd slice as parent: %s\", sbParent)\n\t}\n\n\tif err := verifyCgroupHasEnoughMemory(sbParent, m.memoryPath, m.memoryMaxFile); err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\n\treturn sbParent, filepath.Join(sbParent, containerCgroupPath(sbID)), nil\n}", "func (o ApplicationOperationSyncSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationOperationSyncSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (o IopingSpecVolumeVolumeSourceNfsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceNfs) string { return v.Path }).(pulumi.StringOutput)\n}", "func (o TransferJobTransferSpecAwsS3DataSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *TransferJobTransferSpecAwsS3DataSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (d Driver) getSecureCryptMntPath(id string) string {\n\treturn path.Join(d.dir(id), constSecureBaseDirName, constSecureCryptMntDirName)\n}", "func (n EpisodeDetails) GetSeriesPath(destination string) string {\n\treturn filepath.Join(destination, FileNameCleaner(n.Showtitle))\n}", "func (*CaRsas) GetPath() string { return \"/api/objects/ca/rsa/\" }", "func (i *Image) Path() string {\n\treturn i.p\n}", "func (s *Store) Path() string {\n\treturn s.path\n}", "func (s *Store) Path() string {\n\treturn s.path\n}", "func (cs *ControllerServer) getVolumeSharePath(vol *nfsVolume) string {\n\treturn filepath.Join(string(filepath.Separator), vol.baseDir, vol.subDir)\n}", "func (d *MinioDriver) Path(r volume.Request) volume.Response {\n\td.m.RLock()\n\tdefer d.m.RUnlock()\n\n\tv, exists := d.volumes[r.Name]\n\tif !exists {\n\t\treturn volumeResp(\"\", \"\", nil, capability, newErrVolNotFound(r.Name).Error())\n\t}\n\treturn volumeResp(v.mountpoint, r.Name, nil, capability, \"\")\n}", "func Path(subpath string) (string, error) {\n\tif !isinit {\n\t\treturn \"\", fmt.Errorf(\"pathresolver not initialized\")\n\t}\n\n\tif subpath == \"\" {\n\t\treturn basedirectory, nil\n\t}\n\n\tif filepath.IsAbs(subpath) {\n\t\treturn \"\", fmt.Errorf(\"cannot use absolute path as subpath\")\n\t}\n\n\treturn filepath.Join(basedirectory, subpath), nil\n}", "func (o ApplicationStatusSyncComparedToSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusSyncComparedToSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (*RemoteSyslogServers) GetPath() string { return \"/api/objects/remote_syslog/server/\" }", "func (o IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItems) string { return v.Path }).(pulumi.StringOutput)\n}", "func (o ArgoCDSpecServerIngressOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecServerIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (o IopingSpecVolumeVolumeSourceGlusterfsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceGlusterfs) string { return v.Path }).(pulumi.StringOutput)\n}", "func (r *SysNet) getPath() (string, error) {\n\tvar procPath string\n\n\tswitch r.Path {\n\tcase sysNetPathCore:\n\t\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathCore), r.Property)\n\t\tbreak\n\tcase sysNetPathIPv4:\n\n\t\tif r.Link != \"\" {\n\t\t\tprocPath = path.Join(path.Join(path.Join(path.Join(sysNetPath, sysNetPathIPv4), \"conf\"), r.Link), r.Property)\n\t\t} else {\n\t\t\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathIPv4), r.Property)\n\t\t}\n\t\tbreak\n\tcase sysNetPathIPv6:\n\n\t\tif r.Link != \"\" {\n\t\t\tprocPath = path.Join(path.Join(path.Join(path.Join(sysNetPath, sysNetPathIPv6), \"conf\"), r.Link), r.Property)\n\t\t} else {\n\t\t\tprocPath = path.Join(path.Join(sysNetPath, sysNetPathIPv6), r.Property)\n\t\t}\n\t\tbreak\n\tdefault:\n\t\treturn \"\", errors.New(\"Path not found\")\n\t}\n\n\treturn procPath, nil\n}", "func (o IopingSpecVolumeVolumeSourceProjectedSourcesServiceAccountTokenOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceProjectedSourcesServiceAccountToken) string { return v.Path }).(pulumi.StringOutput)\n}", "func (c *CaRsa) GetPath() string { return fmt.Sprintf(\"/api/objects/ca/rsa/%s\", c.Reference) }", "func (c *Config) StoragePath() string {\n\tif c.options.StoragePath == \"\" {\n\t\tconst dirName = \"\"\n\n\t\tstorageDir := fs.FindDir(fs.StoragePaths)\n\t\tif fs.PathWritable(storageDir) && !c.ReadOnly() {\n\t\t\treturn storageDir\n\t\t}\n\t\t// Use .esp in home directory?\n\t\tif usr, _ := user.Current(); usr.HomeDir != \"\" {\n\t\t\tp := fs.Abs(filepath.Join(usr.HomeDir, fs.HiddenPath, dirName))\n\t\t\tif fs.PathWritable(p) || c.ReadOnly() {\n\t\t\t\treturn p\n\t\t\t}\n\t\t}\n\n\t\t// Fallback directory in case nothing else works.\n\t\tif c.ReadOnly() {\n\t\t\treturn fs.Abs(filepath.Join(fs.HiddenPath, dirName))\n\t\t}\n\n\t\treturn \"\"\n\t}\n\n\treturn fs.Abs(c.options.StoragePath)\n}", "func (m *mountPoint) Path() string {\n\tif m.Volume != nil {\n\t\treturn m.Volume.Path()\n\t}\n\n\treturn m.Source\n}", "func (p Pather) GetPath(svc addr.HostSVC, ps *seg.PathSegment) (*snet.SVCAddr, error) {\n\tif len(ps.ASEntries) == 0 {\n\t\treturn nil, serrors.New(\"empty path\")\n\t}\n\n\tbeta := ps.Info.SegmentID\n\t// The hop fields need to be in reversed order.\n\thopFields := make([]path.HopField, len(ps.ASEntries))\n\tfor i, entry := range ps.ASEntries {\n\t\thopFields[len(hopFields)-1-i] = path.HopField{\n\t\t\tConsIngress: entry.HopEntry.HopField.ConsIngress,\n\t\t\tConsEgress: entry.HopEntry.HopField.ConsEgress,\n\t\t\tExpTime: entry.HopEntry.HopField.ExpTime,\n\t\t\tMac: entry.HopEntry.HopField.MAC,\n\t\t}\n\t\t// the last AS entry is our AS for this we don't need to modify the beta.\n\t\tif i < len(ps.ASEntries)-1 {\n\t\t\tbeta = beta ^ binary.BigEndian.Uint16(entry.HopEntry.HopField.MAC[:2])\n\t\t}\n\t}\n\n\thops := len(hopFields)\n\tdec := scion.Decoded{\n\t\tBase: scion.Base{\n\t\t\tPathMeta: scion.MetaHdr{\n\t\t\t\tCurrHF: 0,\n\t\t\t\tCurrINF: 0,\n\t\t\t\tSegLen: [3]uint8{uint8(hops), 0, 0},\n\t\t\t},\n\t\t\tNumHops: hops,\n\t\t\tNumINF: 1,\n\t\t},\n\t\tInfoFields: []path.InfoField{{\n\t\t\tTimestamp: util.TimeToSecs(ps.Info.Timestamp),\n\t\t\tConsDir: false,\n\t\t\tSegID: beta,\n\t\t}},\n\t\tHopFields: hopFields,\n\t}\n\tpath, err := snetpath.NewSCIONFromDecoded(dec)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"serializing path\", err)\n\t}\n\tifID := dec.HopFields[0].ConsIngress\n\tnextHop := p.NextHopper.UnderlayNextHop(ifID)\n\tif nextHop == nil {\n\t\treturn nil, serrors.New(\"first-hop border router not found\", \"intf_id\", ifID)\n\t}\n\treturn &snet.SVCAddr{\n\t\tIA: ps.FirstIA(),\n\t\tPath: path,\n\t\tNextHop: nextHop,\n\t\tSVC: svc,\n\t}, nil\n\n}", "func (c *Client) Path() string {\n\tc.mu.RLock()\n\tdefer c.mu.RUnlock()\n\treturn c.path\n}", "func (store FileStore) infoPath(id string) string {\n\treturn filepath.Join(store.Path, id+\".info\")\n}", "func (h Hashicorp) GetPath() string {\n\treturn h.FilePath\n}", "func (o TransferJobTransferSpecGcsDataSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecGcsDataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (module *SdkLibrary) apiDistPath(apiScope *apiScope) string {\n\treturn path.Join(\"apistubs\", module.distGroup(), apiScope.name)\n}", "func (o ApplicationOperationSyncSourcePtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ApplicationOperationSyncSource) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (r *ManagerResource) Path() string {\n\treturn r.path\n}", "func Path() string {\n\treturn c.Path\n}", "func (o IopingSpecVolumeVolumeSourceHostPathOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v IopingSpecVolumeVolumeSourceHostPath) string { return v.Path }).(pulumi.StringOutput)\n}", "func (o ArgoCDSpecServerGrpcIngressOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ArgoCDSpecServerGrpcIngress) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func SandboxConfigurationItemPath(id string, item Item) (string, error) {\n\tif id == \"\" {\n\t\treturn \"\", fmt.Errorf(\"Empty sandbox ID\")\n\t}\n\n\titemFile, err := itemToFile(item)\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\treturn filepath.Join(VCStorePrefix, ConfigStoragePath(), id, itemFile), nil\n}", "func (o ArgoCDSpecServerIngressPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *ArgoCDSpecServerIngress) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (i *Index) securePath(path string) (cleaned string, hostpath string) {\n\tpath, err := url.QueryUnescape(path)\n\tif err != nil {\n\t\treturn \"/\", i.Root\n\t}\n\tcleaned = filepath.Clean(path)\n\thostpath = i.Root + cleaned\n\treturn\n}", "func (o LocalCopyResponseOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LocalCopyResponse) string { return v.Path }).(pulumi.StringOutput)\n}", "func (o ApplicationStatusOperationStateSyncResultSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateSyncResultSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func (o IopingSpecVolumeVolumeSourceNfsPtrOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *IopingSpecVolumeVolumeSourceNfs) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn &v.Path\n\t}).(pulumi.StringPtrOutput)\n}", "func (o FioSpecVolumeVolumeSourceProjectedSourcesServiceAccountTokenOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesServiceAccountToken) string { return v.Path }).(pulumi.StringOutput)\n}", "func (p *Plugin) ScopedPath(s string) string {\n\tif p.PluginObj.Config.PropagatedMount != \"\" && strings.HasPrefix(s, p.PluginObj.Config.PropagatedMount) {\n\t\t// re-scope to the propagated mount path on the host\n\t\treturn filepath.Join(filepath.Dir(p.Rootfs), \"propagated-mount\", strings.TrimPrefix(s, p.PluginObj.Config.PropagatedMount))\n\t}\n\treturn filepath.Join(p.Rootfs, s)\n}", "func (o TransferJobTransferSpecAzureBlobStorageDataSourceOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v TransferJobTransferSpecAzureBlobStorageDataSource) *string { return v.Path }).(pulumi.StringPtrOutput)\n}", "func getPath(svc addr.HostSVC, ps *seg.PathSegment,\n\ttopoProv topology.Provider) (*snet.SVCAddr, error) {\n\n\tp, err := legacyPath(ps)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"constructing path from segment\", err)\n\t}\n\tif err := p.Reverse(); err != nil {\n\t\treturn nil, serrors.WrapStr(\"reversing path\", err)\n\t}\n\tif err := p.InitOffsets(); err != nil {\n\t\treturn nil, serrors.WrapStr(\"initializing offsets\", err)\n\t}\n\thopF, err := p.GetHopField(p.HopOff)\n\tif err != nil {\n\t\treturn nil, serrors.WrapStr(\"extracting first hop field\", err)\n\t}\n\ttopo := topoProv.Get()\n\tifID := hopF.ConsIngress\n\tUnderlayNextHop, ok := topo.UnderlayNextHop2(ifID)\n\tif !ok {\n\t\treturn nil, serrors.New(\"first-hop border router not found\", \"intf_id\", ifID)\n\t}\n\treturn &snet.SVCAddr{IA: ps.FirstIA(), Path: p, NextHop: UnderlayNextHop, SVC: svc}, nil\n}", "func (v VirtualSwitch) Path() (string, error) {\n\treturn v.virtualSwitch.Path()\n}", "func (o FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItemsOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v FioSpecVolumeVolumeSourceProjectedSourcesDownwardAPIItems) string { return v.Path }).(pulumi.StringOutput)\n}", "func (o LocalCopyOutput) Path() pulumi.StringOutput {\n\treturn o.ApplyT(func(v LocalCopy) string { return v.Path }).(pulumi.StringOutput)\n}", "func (c *Client) Path() string {\n\treturn Path\n}", "func (o ApplicationStatusOperationStateOperationSyncSourceHelmFileParametersOutput) Path() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v ApplicationStatusOperationStateOperationSyncSourceHelmFileParameters) *string { return v.Path }).(pulumi.StringPtrOutput)\n}" ]
[ "0.74364364", "0.7279818", "0.57709485", "0.56959575", "0.5652686", "0.5584653", "0.55214775", "0.5389453", "0.53671575", "0.53460824", "0.533256", "0.52875644", "0.52684146", "0.5267466", "0.5247035", "0.52464074", "0.51968586", "0.5180792", "0.5175526", "0.5171551", "0.5171551", "0.5149007", "0.5134776", "0.5132581", "0.5127024", "0.5122917", "0.5081935", "0.5072549", "0.5065655", "0.5044717", "0.5043704", "0.5033308", "0.50218445", "0.5018088", "0.50139695", "0.50037605", "0.4998196", "0.49921614", "0.49828467", "0.49802768", "0.49619955", "0.4961612", "0.49522248", "0.4951539", "0.49467763", "0.4939332", "0.49350157", "0.49290487", "0.4928961", "0.49207598", "0.49056935", "0.49028972", "0.49003157", "0.48988467", "0.48890874", "0.4870896", "0.48647636", "0.48580962", "0.485039", "0.4848557", "0.4848557", "0.48399955", "0.4831611", "0.48266563", "0.48237526", "0.4818253", "0.48071283", "0.4803512", "0.48011374", "0.47945023", "0.47937265", "0.4788762", "0.4780677", "0.4777919", "0.47764984", "0.47762063", "0.4754437", "0.4750877", "0.4749953", "0.4744859", "0.47386527", "0.47355336", "0.47309208", "0.4729769", "0.4719194", "0.47187114", "0.47182325", "0.4704324", "0.47033232", "0.47016972", "0.4697198", "0.46923193", "0.46872768", "0.4684552", "0.4679754", "0.46779963", "0.4675425", "0.46706453", "0.4667558", "0.4665334" ]
0.85262775
0
update fs block size to BlkSize
func getBlkSize(p string) { fs := syscall.Statfs_t{} err := syscall.Statfs(p, &fs) if err != nil { log.Fatal("get block size error with:", err) } BlockSize = uint64(fs.Bsize) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (sm3 *SM3) BlockSize() int { return 64 }", "func (s *State) BlockSize() int { return 2 }", "func (s *State) BlockSize() int { return 8 }", "func (d *state) BlockSize() int { return d.rate }", "func (s *State) BlockSize() int { return 1 }", "func (h *Hash) BlockSize() int { return len(h.buf) }", "func (d *digest) BlockSize() int { return 1 }", "func (d *RabinKarp64) BlockSize() int { return 1 }", "func (bdi *Info) BlockSize() int {\n\treturn int(C.spdk_bdev_get_block_size(bdi.ptr()))\n}", "func (c Zfs) UpdateBlock(b *i3barjson.Block) {\n\tb.Color = c.Color\n\tfullTextFmt := fmt.Sprintf(\"%s%%s\", c.Label)\n\n\tzpoolCmd := exec.Command(\"sudo\", \"zpool\", \"status\", c.PoolName)\n\tout, err := zpoolCmd.Output()\n\n\tif err != nil {\n\t\tb.Urgent = true\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\treturn\n\t}\n\n\tzpoolLines := strings.Split(string(out), \"\\n\")\n\tfor _, zpoolLine := range zpoolLines {\n\t\tline := strings.TrimSpace(zpoolLine)\n\t\tif strings.HasPrefix(line, \"state\") {\n\t\t\tsplit := strings.Split(line, \":\")\n\t\t\tstatus := strings.TrimSpace(split[1])\n\n\t\t\tif status == \"ONLINE\" {\n\t\t\t\tb.Urgent = false\n\t\t\t} else {\n\t\t\t\tb.Urgent = true\n\t\t\t}\n\t\t\tb.FullText = fmt.Sprintf(fullTextFmt, status)\n\t\t\treturn\n\t\t}\n\t}\n\n\tb.Urgent = true\n\tb.FullText = fmt.Sprintf(fullTextFmt, \"NOT FOUND\")\n\treturn\n}", "func (h *ihash) BlockSize() int { return h.blockSize }", "func (syncer *MerkleSyncer) updateHeight() {\n\tatomic.AddUint64(&syncer.height, 1)\n}", "func (k Keeper) BlockByteSize(ctx sdk.Ctx) (res int64) {\n\tk.Paramstore.Get(ctx, types.KeyBlockByteSize, &res)\n\treturn\n}", "func (ob *Observer) updateBlock(curHeight, nextHeight int64, curBlockHash string) error {\n\tblock, err := ob.deps.Recorder.Block(nextHeight)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"[Observer.updateBlock]: failed to get block info, height=%d\", nextHeight)\n\t}\n\n\tif curHeight != 0 && block.ParentBlockHash != curBlockHash {\n\t\tif err := ob.DeleteBlock(curHeight); err != nil {\n\t\t\treturn errors.Wrap(err, \"[Observer.updateBlock]: failed to delete a forked block\")\n\t\t}\n\n\t\treturn nil\n\t}\n\n\tif err := ob.RecordBlockAndTxs(block); err != nil {\n\t\treturn errors.Wrap(err, \"[Observer.updateBlock]: failed to save and process block\")\n\t}\n\n\treturn nil\n}", "func (xxh *xxHash) BlockSize() int {\n\treturn 1\n}", "func (d *btrfs) Update(changedConfig map[string]string) error {\n\t// We only care about btrfs.mount_options.\n\tval, ok := changedConfig[\"btrfs.mount_options\"]\n\tif ok {\n\t\t// Custom mount options don't work inside containers\n\t\tif d.state.OS.RunningInUserNS {\n\t\t\treturn nil\n\t\t}\n\n\t\t// Trigger a re-mount.\n\t\td.config[\"btrfs.mount_options\"] = val\n\t\tmntFlags, mntOptions := filesystem.ResolveMountOptions(strings.Split(d.getMountOptions(), \",\"))\n\t\tmntFlags |= unix.MS_REMOUNT\n\n\t\terr := TryMount(\"\", GetPoolMountPath(d.name), \"none\", mntFlags, mntOptions)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\tsize, ok := changedConfig[\"size\"]\n\tif ok {\n\t\t// Figure out loop path\n\t\tloopPath := loopFilePath(d.name)\n\n\t\tif d.config[\"source\"] != loopPath {\n\t\t\treturn fmt.Errorf(\"Cannot resize non-loopback pools\")\n\t\t}\n\n\t\t// Resize loop file\n\t\tf, err := os.OpenFile(loopPath, os.O_RDWR, 0600)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdefer func() { _ = f.Close() }()\n\n\t\tsizeBytes, _ := units.ParseByteSizeString(size)\n\n\t\terr = f.Truncate(sizeBytes)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tloopDevPath, err := loopDeviceSetup(loopPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tdefer func() { _ = loopDeviceAutoDetach(loopDevPath) }()\n\n\t\terr = loopDeviceSetCapacity(loopDevPath)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\t_, err = shared.RunCommand(\"btrfs\", \"filesystem\", \"resize\", \"max\", GetPoolMountPath(d.name))\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (ns *EsIndexer) UpdateLastBlockHeightFromDb() {\n\tbestBlock, err := ns.GetBestBlockFromDb()\n\tif err != nil {\n\t\treturn\n\t}\n\tns.lastBlockHeight = bestBlock.BlockNo\n\tns.lastBlockHash = bestBlock.id\n}", "func (s *StatFS) BlockFiles(size int) int {\n\treturn size / int(s.stat.Bsize)\n}", "func (digest *sm3Digest) BlockSize() int {\n\treturn BlockSize\n}", "func (d Avx512Digest) BlockSize() int { return BlockSize }", "func (nsc *NilConsumerStatsCollector) UpdateCheckpointSize(int) {}", "func (s *Store) UpdateTreeStoreSize(key string, newSize int64) error {\n\treturn s.db.Do(func(tx *sql.Tx) error {\n\t\t_, err := tx.Exec(\"UPDATE aciinfo SET treestoresize = $1 WHERE blobkey == $2\", newSize, key)\n\t\treturn err\n\t})\n}", "func (fs *FileSystem) UpdateFst(file inode) error {\n\tfs.nextFreeInode = fs.nextFreeInode[1:]\n\tfs.nextFreeDataBlock = fs.nextFreeDataBlock[len(file.dataList):]\n\treturn nil\n}", "func (fsys *FS) Statfs(path string, stat *fuse.Statfs_t) (errc int) {\n\tdefer fs.Trace(path, \"\")(\"stat=%+v, errc=%d\", stat, &errc)\n\tconst blockSize = 4096\n\tfsBlocks := uint64(1 << 50)\n\tif runtime.GOOS == \"windows\" {\n\t\tfsBlocks = (1 << 43) - 1\n\t}\n\tstat.Blocks = fsBlocks // Total data blocks in file system.\n\tstat.Bfree = fsBlocks // Free blocks in file system.\n\tstat.Bavail = fsBlocks // Free blocks in file system if you're not root.\n\tstat.Files = 1E9 // Total files in file system.\n\tstat.Ffree = 1E9 // Free files in file system.\n\tstat.Bsize = blockSize // Block size\n\tstat.Namemax = 255 // Maximum file name length?\n\tstat.Frsize = blockSize // Fragment size, smallest addressable data size in the file system.\n\treturn 0\n}", "func getSizeForPath(path string) (uint64, error) {\n\ts := syscall.Statfs_t{}\n\tif err := syscall.Statfs(path, &s); err != nil {\n\t\treturn 0, fmt.Errorf(\"failed to statfs on %s, %+v\", path, err)\n\t}\n\n\treturn s.Blocks * uint64(s.Bsize), nil\n}", "func (b *Block) Size() common.StorageSize {\n\tif size := b.size.Load(); size != nil {\n\t\treturn size.(common.StorageSize)\n\t}\n\tc := writeCounter(0)\n\trlp.Encode(&c, b)\n\tb.size.Store(common.StorageSize(c))\n\treturn common.StorageSize(c)\n}", "func blocklistUpdate() error {\n\tctx, cancel := context.WithTimeout(context.TODO(), ezhttp.DefaultTimeout10s)\n\tdefer cancel()\n\tres, err := ezhttp.Get(\n\t\tctx,\n\t\t\"https://download.dnscrypt.info/blacklists/domains/mybase.txt\",\n\t\tezhttp.Header(\"User-Agent\", \"github.com/function61/function53\"))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn atomicfilewrite.Write(blocklistFilename, func(blocklist io.Writer) error {\n\t\t_, err := io.Copy(blocklist, res.Body)\n\t\treturn err\n\t})\n}", "func (leaf *Node) update(newInfo os.FileInfo, withContent bool) (err error) {\n\tif newInfo == nil {\n\t\tnewInfo, err = os.Stat(leaf.SysPath)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"lib/memfs: Node.update %q: %s\",\n\t\t\t\tleaf.Path, err.Error())\n\t\t}\n\t}\n\n\tif leaf.Mode != newInfo.Mode() {\n\t\tleaf.Mode = newInfo.Mode()\n\t\treturn nil\n\t}\n\n\tleaf.ModTime = newInfo.ModTime()\n\tleaf.Size = newInfo.Size()\n\n\tif !withContent || newInfo.IsDir() {\n\t\treturn nil\n\t}\n\n\treturn leaf.updateContent()\n}", "func (f FormatHeader) BlockSize() uint16 {\n\treturn (f.BitsPerSample / 8) * f.NumChannels\n}", "func (x *gcm) BlockSize() int { return x.blockSize }", "func (h *blockHeader) setSize(v uint32) {\n\tconst mask = 7\n\t*h = (*h)&mask | blockHeader(v<<3)\n}", "func (c *BlockCache) setDbFiles(height int) {\n\tif height <= c.nextBlock {\n\t\tif height < c.firstBlock {\n\t\t\theight = c.firstBlock\n\t\t}\n\t\tindex := height - c.firstBlock\n\t\tif err := c.lengthsFile.Truncate(int64(index * 4)); err != nil {\n\t\t\tLog.Fatal(\"truncate lengths file failed: \", err)\n\t\t}\n\t\tif err := c.blocksFile.Truncate(c.starts[index]); err != nil {\n\t\t\tLog.Fatal(\"truncate blocks file failed: \", err)\n\t\t}\n\t\tc.Sync()\n\t\tc.starts = c.starts[:index+1]\n\t\tc.nextBlock = height\n\t\tc.setLatestHash()\n\t}\n}", "func Diskuse(path string) (cap uint64, used uint64) {\n\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\tcap = fs.Blocks * uint64(fs.Bsize)\n\tfree := fs.Bfree * uint64(fs.Bsize) // yup, I just did that\n\tused = cap - free\n\treturn cap, used\n}", "func (leaf *Node) updateContent() (err error) {\n\tif leaf.Size > MaxFileSize {\n\t\treturn nil\n\t}\n\n\tleaf.V, err = ioutil.ReadFile(leaf.SysPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (fup *folderUpdatePrepper) updateResolutionUsageLockedCache(\n\tctx context.Context, lState *kbfssync.LockState, md *RootMetadata,\n\tbps blockPutState, unmergedChains, mergedChains *crChains,\n\tmostRecentMergedMD ImmutableRootMetadata,\n\trefs, unrefs map[data.BlockPointer]bool) error {\n\tmd.SetRefBytes(0)\n\tmd.SetUnrefBytes(0)\n\tmd.SetMDRefBytes(0)\n\tmd.SetDiskUsage(mostRecentMergedMD.DiskUsage())\n\tmd.SetMDDiskUsage(mostRecentMergedMD.MDDiskUsage())\n\n\tlocalBlocks := make(map[data.BlockPointer]data.Block)\n\tfor _, ptr := range bps.Ptrs() {\n\t\tif block, err := bps.GetBlock(ctx, ptr); err == nil && block != nil {\n\t\t\tlocalBlocks[ptr] = block\n\t\t}\n\t}\n\n\t// Add bytes for every ref'd block.\n\trefPtrsToFetch := make([]data.BlockPointer, 0, len(refs))\n\tvar refSum uint64\n\tfor ptr := range refs {\n\t\tif block, ok := localBlocks[ptr]; ok {\n\t\t\trefSum += uint64(block.GetEncodedSize())\n\t\t} else {\n\t\t\trefPtrsToFetch = append(refPtrsToFetch, ptr)\n\t\t}\n\t\tfup.vlog.CLogf(ctx, libkb.VLog1, \"Ref'ing block %v\", ptr)\n\t}\n\n\t// Look up the total sum of the ref blocks in parallel to get\n\t// their sizes.\n\t//\n\t// TODO: If the blocks weren't already in the cache, this call\n\t// won't cache them, so it's kind of wasting work. Furthermore,\n\t// we might be able to get the encoded size from other sources as\n\t// well (such as its directory entry or its indirect file block)\n\t// if we happened to have come across it before.\n\trefSumFetched, err := fup.blocks.GetCleanEncodedBlocksSizeSum(\n\t\tctx, lState, md.ReadOnly(), refPtrsToFetch, nil, fup.branch(), false)\n\tif err != nil {\n\t\treturn err\n\t}\n\trefSum += refSumFetched\n\n\tfup.vlog.CLogf(ctx, libkb.VLog1, \"Ref'ing a total of %d bytes\", refSum)\n\tmd.AddRefBytes(refSum)\n\tmd.AddDiskUsage(refSum)\n\n\tunrefPtrsToFetch := make([]data.BlockPointer, 0, len(unrefs))\n\tvar unrefSum uint64\n\tfor ptr := range unrefs {\n\t\toriginal, ok := unmergedChains.originals[ptr]\n\t\tif !ok {\n\t\t\toriginal = ptr\n\t\t}\n\t\tif original != ptr || unmergedChains.isCreated(original) {\n\t\t\t// Only unref pointers that weren't created as part of the\n\t\t\t// unmerged branch. Either they existed already or they\n\t\t\t// were created as part of the merged branch.\n\t\t\tcontinue\n\t\t}\n\t\t// Also make sure this wasn't already removed or overwritten\n\t\t// on the merged branch.\n\t\toriginal, ok = mergedChains.originals[ptr]\n\t\tif !ok {\n\t\t\toriginal = ptr\n\t\t}\n\t\tmergedChain, ok := mergedChains.byOriginal[original]\n\t\tif (ok && original != mergedChain.mostRecent && original == ptr) ||\n\t\t\tmergedChains.isDeleted(original) {\n\t\t\tcontinue\n\t\t}\n\n\t\tif info, ok := fup.cachedInfos[ptr]; ok {\n\t\t\tunrefSum += uint64(info.EncodedSize)\n\t\t} else {\n\t\t\tunrefPtrsToFetch = append(unrefPtrsToFetch, ptr)\n\t\t}\n\t}\n\n\t// Look up the unref blocks in parallel to get their sizes. Since\n\t// we don't know whether these are files or directories, just look\n\t// them up generically. Ignore any recoverable errors for unrefs.\n\t// Note that we can't combine these with the above ref fetches\n\t// since they require a different MD. If the merged changes\n\t// didn't change any blocks (in particular, the root block), we\n\t// can assume all the blocks we are unreferencing were live;\n\t// otherwise, we need to check with the server to make sure.\n\tonlyCountIfLive := len(mergedChains.byOriginal) != 0\n\tunrefSumFetched, err := fup.blocks.GetCleanEncodedBlocksSizeSum(\n\t\tctx, lState, mostRecentMergedMD, unrefPtrsToFetch, unrefs,\n\t\tfup.branch(), onlyCountIfLive)\n\tif err != nil {\n\t\treturn err\n\t}\n\tunrefSum += unrefSumFetched\n\n\t// Subtract bytes for every unref'd block that wasn't created in\n\t// the unmerged branch.\n\tfup.vlog.CLogf(ctx, libkb.VLog1, \"Unref'ing a total of %d bytes\", unrefSum)\n\tmd.AddUnrefBytes(unrefSum)\n\tmd.SetDiskUsage(md.DiskUsage() - unrefSum)\n\treturn nil\n}", "func (b *SnailBlock) Size() common.StorageSize {\n\tif size := b.size.Load(); size != nil {\n\t\treturn size.(common.StorageSize)\n\t}\n\tc := writeCounter(0)\n\trlp.Encode(&c, b)\n\tb.size.Store(common.StorageSize(c))\n\treturn common.StorageSize(c)\n}", "func SetBpfBuflen(fd, l int) (int, error) {\n\terr := ioctlPtr(fd, BIOCSBLEN, unsafe.Pointer(&l))\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\treturn l, nil\n}", "func UpdateClusterMemoryCurrentBytes(memoryCount int64) {\n\tmemoryCurrentBytes.Set(float64(memoryCount))\n}", "func (u *uploader) initSize() {\n\tu.totalSize = -1\n\n\tswitch r := u.in.Body.(type) {\n\tcase io.Seeker:\n\t\tn, err := aws.SeekerLen(r)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\tu.totalSize = n\n\n\t\t// Try to adjust partSize if it is too small and account for\n\t\t// integer division truncation.\n\t\tif u.totalSize/u.cfg.PartSize >= int64(u.cfg.MaxUploadParts) {\n\t\t\t// Add one to the part size to account for remainders\n\t\t\t// during the size calculation. e.g odd number of bytes.\n\t\t\tu.cfg.PartSize = (u.totalSize / int64(u.cfg.MaxUploadParts)) + 1\n\t\t}\n\t}\n}", "func (layout Layout) hashesPerBlock() int64 {\n\treturn layout.blockSize / layout.digestSize\n}", "func (m *metricSshcheckSftpStatus) updateCapacity() {\n\tif m.data.Sum().DataPoints().Len() > m.capacity {\n\t\tm.capacity = m.data.Sum().DataPoints().Len()\n\t}\n}", "func (*digest) BlockSize() int {\n\treturn int(BlockSize)\n}", "func (*digest) BlockSize() int {\n\treturn int(BlockSize)\n}", "func newBlockfileMgr(id string, conf *Conf, indexConfig *blkstorage.IndexConfig, indexStore *leveldbhelper.DBHandle) *blockfileMgr {\n\tlogger.Debugf(\"newBlockfileMgr() initializing file-based block storage for ledger: %s \", id)\n\tvar rwMutexs []*sync.RWMutex\n\n\t//Determine the root directory for the blockfile storage, if it does not exist create it\n\trootDir := conf.getLedgerBlockDir(id)\n\t_, err := util.CreateDirIfMissing(rootDir)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Error: %s\", err))\n\t}\n\t// Instantiate the manager, i.e. blockFileMgr structure\n\tmgr := &blockfileMgr{rootDir: rootDir, conf: conf, db: indexStore, rwMutexs: rwMutexs}\n\n\t// cp = checkpointInfo, retrieve from the database the file suffix or number of where blocks were stored.\n\t// It also retrieves the current size of that file and the last block number that was written to that file.\n\t// At init checkpointInfo:latestFileChunkSuffixNum=[0], latestFileChunksize=[0], lastBlockNumber=[0]\n\tcpInfo, err := mgr.loadCurrentInfo()\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not get block file info for current block file from db: %s\", err))\n\t}\n\tif cpInfo == nil {\n\t\tlogger.Info(`Getting block information from block storage`)\n\t\tif cpInfo, err = constructCheckpointInfoFromBlockFiles(rootDir); err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Could not build checkpoint info from block files: %s\", err))\n\t\t}\n\t\tlogger.Debugf(\"Info constructed by scanning the blocks dir = %s\", spew.Sdump(cpInfo))\n\t} else {\n\t\tlogger.Debug(`Synching block information from block storage (if needed)`)\n\t\tsyncCPInfoFromFS(rootDir, cpInfo)\n\t}\n\terr = mgr.saveCurrentInfo(cpInfo, true)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not save next block file info to db: %s\", err))\n\t}\n\n\tmgr.oldestFileChunkSuffixNum = syncOldestFileNum(rootDir)\n\t//If start up is a restart of an existing storage,new the rwMutex for the files\n\tif conf.dumpConf.Enabled {\n\t\tfor i := 0; i <= cpInfo.latestFileChunkSuffixNum; i++ {\n\t\t\trwMutex := new(sync.RWMutex)\n\t\t\tmgr.rwMutexs = append(mgr.rwMutexs, rwMutex)\n\t\t}\n\t}\n\tmgr.dumpMutex = new(sync.Mutex)\n\n\t//Open a writer to the file identified by the number and truncate it to only contain the latest block\n\t// that was completely saved (file system, index, cpinfo, etc)\n\tcurrentFileWriter, err := newBlockfileWriter(deriveBlockfilePath(rootDir, cpInfo.latestFileChunkSuffixNum))\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not open writer to current file: %s\", err))\n\t}\n\t//Truncate the file to remove excess past last block\n\terr = currentFileWriter.truncateFile(cpInfo.latestFileChunksize)\n\tif err != nil {\n\t\tpanic(fmt.Sprintf(\"Could not truncate current file to known size in db: %s\", err))\n\t}\n\n\t// Create a new KeyValue store database handler for the blocks index in the keyvalue database\n\tmgr.index = newBlockIndex(indexConfig, indexStore)\n\n\t// Update the manager with the checkpoint info and the file writer\n\tmgr.cpInfo = cpInfo\n\tmgr.currentFileWriter = currentFileWriter\n\t// Create a checkpoint condition (event) variable, for the goroutine waiting for\n\t// or announcing the occurrence of an event.\n\tmgr.cpInfoCond = sync.NewCond(&sync.Mutex{})\n\n\t// init BlockchainInfo for external API's\n\tbcInfo := &common.BlockchainInfo{\n\t\tHeight: 0,\n\t\tCurrentBlockHash: nil,\n\t\tPreviousBlockHash: nil}\n\n\tif !cpInfo.isChainEmpty {\n\t\t//If start up is a restart of an existing storage, sync the index from block storage and update BlockchainInfo for external API's\n\t\tmgr.syncIndex()\n\t\tlastBlockHeader, err := mgr.retrieveBlockHeaderByNumber(cpInfo.lastBlockNumber)\n\t\tif err != nil {\n\t\t\tpanic(fmt.Sprintf(\"Could not retrieve header of the last block form file: %s\", err))\n\t\t}\n\t\tlastBlockHash := lastBlockHeader.Hash()\n\t\tpreviousBlockHash := lastBlockHeader.PreviousHash\n\t\tbcInfo = &common.BlockchainInfo{\n\t\t\tHeight: cpInfo.lastBlockNumber + 1,\n\t\t\tCurrentBlockHash: lastBlockHash,\n\t\t\tPreviousBlockHash: previousBlockHash}\n\t}\n\tmgr.bcInfo.Store(bcInfo)\n\treturn mgr\n}", "func DirSizeByte(path string) uint64 {\n\tdirSize = 0\n\tfilepath.Walk(path, readSize)\n\treturn dirSize\n}", "func calculateBufferSize(blocks uint64) uint64 {\n\tif nbb := NBufferBlocks; blocks < nbb {\n\t\treturn blocks\n\t} else {\n\t\treturn nbb\n\t}\n}", "func (b *BlockSplitterSimple) SetMaxDirEntriesByBlockSize(\n\tcodec kbfscodec.Codec) error {\n\tdirEnv := os.Getenv(\"KEYBASE_BSPLIT_MAX_DIR_ENTRIES\")\n\tif len(dirEnv) > 0 {\n\t\t// Don't override the environment variable.\n\t\treturn nil\n\t}\n\n\tblock := NewDirBlock().(*DirBlock)\n\tbigName := strings.Repeat(\"a\", MaxNameBytesDefault)\n\t// Make \"typical\" DirEntry, though the max dir entry is a bit\n\t// bigger than this (can contain a variable-length symlink path,\n\t// for example).\n\tde := DirEntry{\n\t\tBlockInfo: BlockInfo{\n\t\t\tBlockPointer: BlockPointer{\n\t\t\t\tDirectType: DirectBlock,\n\t\t\t},\n\t\t},\n\t\tEntryInfo: EntryInfo{\n\t\t\tPrevRevisions: PrevRevisions{\n\t\t\t\t{Revision: 0, Count: 0},\n\t\t\t\t{Revision: 1, Count: 1},\n\t\t\t\t{Revision: 2, Count: 2},\n\t\t\t\t{Revision: 3, Count: 3},\n\t\t\t\t{Revision: 4, Count: 4},\n\t\t\t},\n\t\t},\n\t}\n\tblock.Children[bigName] = de\n\tencodedBlock, err := codec.Encode(block)\n\tif err != nil {\n\t\treturn err\n\t}\n\toneEntrySize := int64(len(encodedBlock))\n\tb.maxDirEntriesPerBlock = int(b.maxSize / oneEntrySize)\n\tif b.maxDirEntriesPerBlock == 0 {\n\t\tb.maxDirEntriesPerBlock = 1\n\t}\n\treturn nil\n}", "func (bdi *Info) CountBlocks() int {\n\treturn int(C.spdk_bdev_get_num_blocks(bdi.ptr()))\n}", "func updateTxSize(tx *transaction.Transaction) (*transaction.Transaction, error) {\n\tbw := io.NewBufBinWriter()\n\ttx.EncodeBinary(bw.BinWriter)\n\tif bw.Err != nil {\n\t\treturn nil, fmt.Errorf(\"encode binary: %w\", bw.Err)\n\t}\n\treturn transaction.NewTransactionFromBytes(tx.Bytes())\n}", "func (c *Container) rootFsSize() (int64, error) {\n\tcontainer, err := c.runtime.store.Container(c.ID())\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\t// Ignore the size of the top layer. The top layer is a mutable RW layer\n\t// and is not considered a part of the rootfs\n\trwLayer, err := c.runtime.store.Layer(container.LayerID)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tlayer, err := c.runtime.store.Layer(rwLayer.Parent)\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\tsize := int64(0)\n\tfor layer.Parent != \"\" {\n\t\tlayerSize, err := c.runtime.store.DiffSize(layer.Parent, layer.ID)\n\t\tif err != nil {\n\t\t\treturn size, errors.Wrapf(err, \"getting diffsize of layer %q and its parent %q\", layer.ID, layer.Parent)\n\t\t}\n\t\tsize += layerSize\n\t\tlayer, err = c.runtime.store.Layer(layer.Parent)\n\t\tif err != nil {\n\t\t\treturn 0, err\n\t\t}\n\t}\n\t// Get the size of the last layer. Has to be outside of the loop\n\t// because the parent of the last layer is \"\", andlstore.Get(\"\")\n\t// will return an error.\n\tlayerSize, err := c.runtime.store.DiffSize(layer.Parent, layer.ID)\n\treturn size + layerSize, err\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterNodePoolLinuxOsConfig) *int { return v.SwapFileSizeMb }).(pulumi.IntPtrOutput)\n}", "func (req *DronaRequest) updateOsize(size int64) {\n\treq.Lock()\n\tdefer req.Unlock()\n\treq.objectSize = size\n}", "func statfsImpl(t *kernel.Task, d *fs.Dirent, addr usermem.Addr) error {\n\tinfo, err := d.Inode.StatFS(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Construct the statfs structure and copy it out.\n\tstatfs := linux.Statfs{\n\t\tType: info.Type,\n\t\t// Treat block size and fragment size as the same, as\n\t\t// most consumers of this structure will expect one\n\t\t// or the other to be filled in.\n\t\tBlockSize: d.Inode.StableAttr.BlockSize,\n\t\tBlocks: info.TotalBlocks,\n\t\t// We don't have the concept of reserved blocks, so\n\t\t// report blocks free the same as available blocks.\n\t\t// This is a normal thing for filesystems, to do, see\n\t\t// udf, hugetlbfs, tmpfs, among others.\n\t\tBlocksFree: info.FreeBlocks,\n\t\tBlocksAvailable: info.FreeBlocks,\n\t\tFiles: info.TotalFiles,\n\t\tFilesFree: info.FreeFiles,\n\t\t// Same as Linux for simple_statfs, see fs/libfs.c.\n\t\tNameLength: linux.NAME_MAX,\n\t\tFragmentSize: d.Inode.StableAttr.BlockSize,\n\t\t// Leave other fields 0 like simple_statfs does.\n\t}\n\t_, err = t.CopyOut(addr, &statfs)\n\treturn err\n}", "func (c Temperature) UpdateBlock(b *i3barjson.Block) {\n\tb.Color = c.Color\n\tfullTextFmt := fmt.Sprintf(\"%s%%s\", c.Label)\n\ttotalTemp := 0\n\tprocs := 0\n\tsysFileDirList, err := ioutil.ReadDir(c.CpuTempPath)\n\tif err != nil {\n\t\tb.Urgent = true\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\treturn\n\t}\n\tif len(sysFileDirList) != 1 {\n\t\tb.Urgent = true\n\t\tmsg := fmt.Sprintf(\n\t\t\t\"in %s, expected 1 file, got %d\",\n\t\t\tc.CpuTempPath,\n\t\t\tlen(sysFileDirList),\n\t\t)\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, msg)\n\t\treturn\n\t}\n\tsysFileDirPath := fmt.Sprintf(\n\t\t\"%s/%s\",\n\t\tc.CpuTempPath,\n\t\tsysFileDirList[0].Name(),\n\t)\n\tsysFileNameFmt := fmt.Sprintf(\"%s/%%s\", sysFileDirPath)\n\tsysFiles, err := ioutil.ReadDir(sysFileDirPath)\n\tif err != nil {\n\t\tb.Urgent = true\n\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\treturn\n\t}\n\tfor _, sysFile := range sysFiles {\n\t\tsysFileName := sysFile.Name()\n\t\tif !strings.HasSuffix(sysFileName, \"input\") {\n\t\t\tcontinue\n\t\t}\n\t\tr, err := os.Open(fmt.Sprintf(sysFileNameFmt, sysFileName))\n\t\tif err != nil {\n\t\t\tb.Urgent = true\n\t\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\t\treturn\n\t\t}\n\t\tdefer r.Close()\n\t\tvar temp int\n\t\t_, err = fmt.Fscanf(r, \"%d\", &temp)\n\t\tif err != nil {\n\t\t\tb.Urgent = true\n\t\t\tb.FullText = fmt.Sprintf(fullTextFmt, err.Error())\n\t\t\treturn\n\t\t}\n\t\ttotalTemp += temp\n\t\tprocs++\n\t}\n\tavgTemp := float64(totalTemp) / float64(procs*1000)\n\tif avgTemp >= c.CritTemp {\n\t\tb.Urgent = true\n\t} else {\n\t\tb.Urgent = false\n\t}\n\tb.FullText = fmt.Sprintf(\"%s%.2f°C\", c.Label, avgTemp)\n}", "func updateRemainingSpace(ctx *downloaderContext) {\n\n\tctx.globalStatus.RemainingSpace = ctx.globalConfig.MaxSpace -\n\t\tctx.globalStatus.UsedSpace - ctx.globalStatus.ReservedSpace\n\n\tlog.Infof(\"RemainingSpace %d, maxspace %d, usedspace %d, reserved %d\\n\",\n\t\tctx.globalStatus.RemainingSpace, ctx.globalConfig.MaxSpace,\n\t\tctx.globalStatus.UsedSpace, ctx.globalStatus.ReservedSpace)\n}", "func (f *FileSystem) Statfs(ctx context.Context, req *fuse.StatfsRequest, resp *fuse.StatfsResponse) error {\n\t// each block size is 4096 bytes by default.\n\tconst unit = uint64(4096)\n\n\tresp.Bsize = uint32(unit)\n\tresp.Blocks = uint64(f.account.Disk.Size) / unit\n\tresp.Bavail = uint64(f.account.Disk.Avail) / unit\n\tresp.Bfree = uint64(f.account.Disk.Avail) / unit\n\n\treturn nil\n}", "func (lf *ListFile) NumBytes() int64 {\n\t// NOTE: here we don't use IsClosed() because\n\t// it uses the mutex; Size() is used in noMutexIterateLines\n\t// which is called after another mutex is locked,\n\t// making IsClosed() wait forever for the mutex unlock.\n\tif lf.isClosed {\n\t\treturn 0\n\t}\n\n\terr := lf.file.Sync()\n\tif err != nil {\n\t\t// TODO: not panic??\n\t\tpanic(err)\n\t}\n\n\tinfo, err := lf.file.Stat()\n\tif err != nil {\n\t\t// TODO: not panic??\n\t\tpanic(err)\n\t}\n\n\treturn info.Size()\n}", "func (b *Block) Size() metric.StorageSize {\n\tif cached := b.cache.size.Load(); cached != nil {\n\t\treturn cached.(metric.StorageSize)\n\t}\n\tvar size metric.StorageSize\n\trlp.Encode(&size, b)\n\tb.cache.size.Store(size)\n\treturn size\n}", "func (pb *PutBlock) Height() uint64 { return pb.height }", "func (n *nodeBlock) sync() (err error) {\n\tmm := false\n\tif n.offset == 0 {\n\t\tif n._super.mmapSizeUsed+nodeBlockSize <= n._super.mmapSize {\n\t\t\tn.offset = int64(n._super.mmapSizeUsed)\n\t\t\tn._super.mmapSizeUsed += nodeBlockSize\n\t\t\tmm = true\n\t\t} else {\n\t\t\tn.offset, err = n._super._fd.Seek(0, os.SEEK_END)\n\t\t\tif err != nil {\n\t\t\t\treturn\n\t\t\t}\n\t\t}\n\t}\n\n\tx := *(*[nodeBlockSize]byte)(unsafe.Pointer(n))\n\n\tif mm {\n\t\tcopy(n._super._mmap[n.offset:], x[:])\n\t} else {\n\t\t_, err = n._super._fd.Seek(n.offset, 0)\n\t\tif err != nil {\n\t\t\treturn\n\t\t}\n\t\t_, err = n._super._fd.Write(x[:])\n\t}\n\n\tif err == nil {\n\t\tn._super._snapshotChPending[n] = x\n\t}\n\treturn\n}", "func (fm *FinalModelStructBytes) FBEAllocationSize(fbeValue *StructBytes) int {\n fbeResult := 0 +\n fm.F1.FBEAllocationSize(fbeValue.F1) +\n fm.F2.FBEAllocationSize(fbeValue.F2) +\n fm.F3.FBEAllocationSize(fbeValue.F3) +\n 0\n return fbeResult\n}", "func (d *Directory) updateDirectorySize(p string) {\n\tpathParams := strings.Split(p, \"/\")\n\td.Size = 0\n\td.lock()\n\tdefer d.unlock()\n\tfor i := 0; i < len(d.INodes); i++ {\n\t\tswitch d.INodes[i].(type) {\n\t\tcase *Directory:\n\t\t\tif d.INodes[i].GetName() == pathParams[1] {\n\t\t\t\tsubPath := strings.Join(pathParams[2:], \"/\")\n\t\t\t\tsubPath = \"/\" + subPath\n\t\t\t\td.INodes[i].(*Directory).updateDirectorySize(subPath)\n\t\t\t}\n\t\t\td.Size += d.INodes[i].GetSize()\n\t\tcase *File:\n\t\t\td.Size += d.INodes[i].GetSize()\n\t\t}\n\t}\n}", "func (cs *ConsensusState) updateHeight(height int64) {\n\tcs.Height = height\n}", "func (g *testGenerator) updateBlockState(oldBlockName string, oldBlockHash chainhash.Hash, newBlockName string, newBlock *wire.MsgBlock) {\n\t// Remove existing entries.\n\tdelete(g.blocks, oldBlockHash)\n\tdelete(g.blocksByName, oldBlockName)\n\n\t// Add new entries.\n\tnewBlockHash := newBlock.BlockHash()\n\tg.blocks[newBlockHash] = newBlock\n\tg.blocksByName[newBlockName] = newBlock\n}", "func (b *Buffer) update() {\n\tb.NumLines = len(b.lines)\n}", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v KubernetesClusterDefaultNodePoolLinuxOsConfig) *int { return v.SwapFileSizeMb }).(pulumi.IntPtrOutput)\n}", "func (p *pbkdf2) getHashLengthBitBlockCount() int {\n\treturn int(math.Ceil(float64(p.size) / float64(p.pseudoRandomFunction.Size())))\n}", "func (x *ecbEncrypter) BlockSize() int { return x.blockSize }", "func (trd *trxDispatcher) updateLastSeenBlock() {\n\t// get the current value\n\tlsb := trd.blkObserver.Load()\n\tlog.Noticef(\"last seen block is #%d\", lsb)\n\n\t// make the change in the database so the progress persists\n\terr := repo.UpdateLastKnownBlock((*hexutil.Uint64)(&lsb))\n\tif err != nil {\n\t\tlog.Errorf(\"could not update last seen block; %s\", err.Error())\n\t}\n}", "func UpdateCgroupDeviceWeight(pid, innerPath, value string) error {\n\tif pid == \"0\" {\n\t\treturn nil\n\t}\n\n\tcgroupPath, err := FindCgroupPath(pid, \"blkio\", innerPath)\n\tif err != nil {\n\t\treturn err\n\t}\n\tpath := filepath.Join(cgroupPath, \"blkio.weight_device\")\n\tif err := ioutil.WriteFile(path, []byte(value), 0600); err != nil {\n\t\treturn fmt.Errorf(\"%s, please check whether current OS support blkio weight device configuration for bfq scheduler\", err)\n\t}\n\treturn nil\n}", "func (req *DronaRequest) updateAsize(size int64) {\n\treq.Lock()\n\tdefer req.Unlock()\n\treq.asize = size\n}", "func (fileSystem *FileSystem) Put(fileName string) {\n\tif (*fileSystem).PfsFile == nil {\n\t\tfmt.Println(\"You must open pfs first\")\n\t\treturn\n\t}\n\n\tdata, err := ioutil.ReadFile(fileName)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t\treturn\n\t}\n\n\tif len(data) > totalDataBlocks*dataBlockSize {\n\t\tfmt.Println(\"File is too big to be inserted in pfs\")\n\t\treturn\n\t}\n\n\tfileSize := uint16(len(data)) // Get the file size\n\n\t// Finding how many blocks is easy enough\n\tnecesaryBlocks := uint16(math.Ceil(float64(fileSize) / float64(dataBlockSize)))\n\tfoundBlocks := uint16(0)\n\tlocation := -1\n\tfor idx, availability := range (*fileSystem).Directory.FreeDataBlockArray {\n\t\tif foundBlocks > 0 && !availability {\n\t\t\tfoundBlocks = 0\n\t\t\tcontinue\n\t\t}\n\n\t\tif availability {\n\t\t\tfoundBlocks++\n\t\t}\n\t\tif foundBlocks == necesaryBlocks {\n\t\t\tlocation = idx\n\t\t\tbreak\n\t\t}\n\t}\n\n\tif location == -1 {\n\t\tfmt.Println(\"Not enough free blocks to store file.\")\n\t\treturn\n\t}\n\n\t// We have found the location of data that should be written\n\tblockID := location - int(necesaryBlocks) + 1\n\toffset := dataAddress + blockID*dataBlockSize\n\t(*fileSystem).PfsFile.Seek(int64(offset), 0)\n\t(*fileSystem).PfsFile.Write(data)\n\n\t// Now we need to find a FCB to store the records of the value\n\tfor idx, fcb := range (*fileSystem).Directory.FCBArray {\n\t\tif !fcb.ContainsValidData {\n\t\t\tblock, _ := NewFCB(fileName, fileSize, uint8(blockID))\n\t\t\t(*fileSystem).Directory.FCBArray[idx] = block\n\t\t\tbreak\n\t\t}\n\t}\n\n\t// Now we also need to set the free data blocks as ocupied\n\tfor i := 0; i < int(necesaryBlocks); i++ {\n\t\t(*fileSystem).Directory.FreeDataBlockArray[blockID+i] = false\n\t}\n\n\t// We know update the directory\n\t(*fileSystem).Directory.Metadata.NumberOfFilesStored++\n}", "func (o KubernetesClusterNodePoolLinuxOsConfigPtrOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterNodePoolLinuxOsConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SwapFileSizeMb\n\t}).(pulumi.IntPtrOutput)\n}", "func UpdateDumpFile(fileInode inode, data string, fileName string, parentInode inode, parentInodeNum int) error {\n\t// open dump file\n\tfile, err := os.OpenFile(DumpFile, os.O_WRONLY, os.ModeAppend)\n\tif err != nil {\n\t\tfmt.Println(\"Error in opening file while loading the filesystem: \", err)\n\t\tos.Exit(1)\n\t}\n\tdefer file.Close()\n\n\toffset := OffsetInodeTable + fileInode.inodeNum\n\t// update inode bitmap\n\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(strconv.Itoa(SetBit))); err != nil {\n\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\treturn err\n\t}\n\n\t// update inode block\n\toffset = OffsetInodeBlock + (fileInode.inodeNum * InodeBlockSize)\n\tinodeInfo := frameInodeString(fileInode)\n\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(inodeInfo)); err != nil {\n\t\tfmt.Println(\"Error: unable to update inode block\", err)\n\t\treturn err\n\t}\n\n\t//update parent inode block\n\tif parentInodeNum != -1 { // skipping it for root directory\n\t\toffset = OffsetInodeBlock + (parentInodeNum * InodeBlockSize)\n\t\tinodeInfo = frameInodeString(parentInode)\n\t\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(inodeInfo)); err != nil {\n\t\t\tfmt.Println(\"Error: unable to update parent inode block\", err)\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// update data bitmap and fill the data block as well ...\n\toffset = OffsetDataTable\n\tlistByte := []byte(data)\n\tcount := 0\n\tfor _, val := range fileInode.dataList {\n\t\t// update inode bitmap\n\t\tif err := disklib.WriteBlockOnDisk(file, offset+val, []byte(strconv.Itoa(SetBit))); err != nil {\n\t\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\t\treturn err\n\t\t}\n\n\t\t// update data block in disk\n\t\tif err := disklib.WriteBlockOnDisk(file, OffsetDataBlock+(val*DataBlockSize), []byte(listByte[count:count+DataBlockSize])); err != nil {\n\t\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\t\treturn err\n\t\t}\n\t\tcount += DataBlockSize\n\t}\n\n\t// put fst in disk\n\toffset = OffsetFST + (fileInode.inodeNum * CellFSTSize)\n\tstr := fmt.Sprintf(\"%3d\", fileInode.inodeNum) + fmt.Sprintf(\"%10s\", fileName)\n\tif err := disklib.WriteBlockOnDisk(file, offset, []byte(str)); err != nil {\n\t\tfmt.Println(\"Error: unable to update inode bitmap\", err)\n\t\treturn err\n\t}\n\treturn nil\n\n}", "func DirSizeB(path string) (int64, error) {\n\tvar size int64\n\terr := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error {\n\t\tif !info.IsDir() {\n\t\t\tsize += info.Size()\n\t\t}\n\t\treturn err\n\t})\n\treturn size, err\n}", "func UpdateCapacity(cap int) {\n\ttype requestInfo struct {\n\t\tNode string `json:\"node\" binding:\"required\"`\n\t\tCapacity int `json:\"capability\" binding:\"required\"`\n\t}\n\tgo func() {\n\t\tlog.InfoF(\"node %s capacity updated to %d\", conf.Get(\"nodeLanHost\").(string), cap)\n\t\turl := fmt.Sprintf(\"http://%s/nodeUpdateCapacity\", conf.Get(\"registerCenterHost\"))\n\t\t_, _, errs := gorequest.New().Post(url).Send(requestInfo{Node: conf.Get(\"nodeLanHost\").(string), Capacity: cap}).End()\n\t\tif errs != nil && len(errs) > 0 {\n\t\t\tlog.InfoF(\"requestGroupDispatchLoop error:%s\", errs[0])\n\t\t}\n\t}()\n}", "func (fs *FS) fsInfo(ctx context.Context, path string) (int64, int64, int64, int64, int64, int64, error) {\n\tstatfs := &unix.Statfs_t{}\n\terr := unix.Statfs(path, statfs)\n\tif err != nil {\n\t\treturn 0, 0, 0, 0, 0, 0, err\n\t}\n\n\t// Available is blocks available * fragment size\n\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\n\n\t// Capacity is total block count * fragment size\n\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\n\n\t// Usage is block being used * fragment size (aka block size).\n\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\n\n\tinodes := int64(statfs.Files)\n\tinodesFree := int64(statfs.Ffree)\n\tinodesUsed := inodes - inodesFree\n\n\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\n}", "func BlockSize() int {\n\treturn config.Record.BlockSize\n}", "func (b *Block) Size() int {\n\tpbb, err := b.ToProto()\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn pbb.Size()\n}", "func (b *Block) Size() int {\n\tpbb, err := b.ToProto()\n\tif err != nil {\n\t\treturn 0\n\t}\n\n\treturn pbb.Size()\n}", "func (s *hashBLAKE2s) BlockLen() int {\n\treturn blake2s.BlockSize\n}", "func (x *ecbDecrypter) BlockSize() int { return x.blockSize }", "func vmUpdateAdditionalBandwidth(db *Database, vm *VirtualMachine) {\n\t// determine how much of the plan bandwidth to add to the user's bandwidth pool for current month\n\tnow := time.Now()\n\tmonthStart := time.Date(now.Year(), now.Month(), 1, 0, 0, 0, 0, time.UTC)\n\tmonthEnd := monthStart.AddDate(0, 1, 0)\n\tvar factor float64\n\n\tif vm.CreatedTime.Before(monthStart) {\n\t\tfactor = float64(now.Sub(monthStart))\n\t} else {\n\t\tfactor = float64(now.Sub(vm.CreatedTime))\n\t}\n\n\tfactor /= float64(monthEnd.Sub(monthStart))\n\tif factor > 1 {\n\t\tfactor = 1\n\t}\n\n\tadditionalBandwidth := int64((factor * float64(vm.Plan.Bandwidth) + 15) * 1024 * 1024 * 1024)\n\trows := db.Query(\"SELECT id FROM region_bandwidth WHERE region = ? AND user_id = ?\", vm.Region, vm.UserId)\n\tif rows.Next() {\n\t\tvar rowId int\n\t\trows.Scan(&rowId)\n\t\trows.Close()\n\t\tdb.Exec(\"UPDATE region_bandwidth SET bandwidth_additional = bandwidth_additional + ? WHERE id = ?\", additionalBandwidth, rowId)\n\t} else {\n\t\tdb.Exec(\"INSERT INTO region_bandwidth (user_id, region, bandwidth_additional) VALUES (?, ?, ?)\", vm.UserId, vm.Region, additionalBandwidth)\n\t}\n}", "func (b *B) SetBytes(n int64) {}", "func (n *TreeNode) AddSize(by uint64) {\n\tn.mutex.RLock()\n\tn.size += by\n\tn.mutex.RUnlock()\n\treturn\n}", "func (cipher *Simon96Cipher) BlockSize() int {\n\treturn 12\n}", "func (fm FinalModelTimestamp) FBEAllocationSize(value time.Time) int { return fm.FBESize() }", "func (db *DB) grow(sz int) error {\n\t// Ignore if the new size is less than available file size.\n\tif sz <= db.filesz {\n\t\treturn nil\n\t}\n\n\t// If the data is smaller than the alloc size then only allocate what's needed.\n\t// Once it goes over the allocation size then allocate in chunks.\n\tif db.datasz <= db.AllocSize {\n\t\tsz = db.datasz\n\t} else {\n\t\tsz += db.AllocSize\n\t}\n\n\t// Truncate and fsync to ensure file size metadata is flushed.\n\t// https://github.com/boltdb/bolt/issues/284\n\tif !db.NoGrowSync && !db.readOnly {\n\t\tif runtime.GOOS != \"windows\" {\n\t\t\tif err := db.file.Truncate(int64(sz)); err != nil {\n\t\t\t\treturn fmt.Errorf(\"file resize error: %s\", err)\n\t\t\t}\n\t\t}\n\t\tif err := db.file.Sync(); err != nil {\n\t\t\treturn fmt.Errorf(\"file sync error: %s\", err)\n\t\t}\n\t\tif db.Mlock {\n\t\t\t// unlock old file and lock new one\n\t\t\tif err := db.mrelock(db.filesz, sz); err != nil {\n\t\t\t\treturn fmt.Errorf(\"mlock/munlock error: %s\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\tdb.filesz = sz\n\treturn nil\n}", "func (b *BaseNode) updateBytes(n Node) {\n\tbw := io.NewBufBinWriter()\n\tbw.Grow(1 + n.Size())\n\tencodeNodeWithType(n, bw.BinWriter)\n\tb.bytes = bw.Bytes()\n\tb.bytesValid = true\n}", "func (b *bpfHandle) SetReadBufferSize(size int) error {\n\tif ok, err := b.ioctlWithInt(ioctlBIOCSBLEN, size); !ok {\n\t\treturn err\n\t}\n\n\tnumData := make([]byte, ioctlIntegerSize)\n\tb.ioctlWithData(ioctlBIOCGBLEN, numData)\n\tif binary.LittleEndian.Uint32(numData) < uint32(size) {\n\t\treturn errors.New(\"unsupported buffer size\")\n\t}\n\n\tb.readBuffer = make([]byte, size)\n\treturn nil\n}", "func (c *BlockCache) Add(height int, block *walletrpc.CompactBlock) error {\n\t// Invariant: m[firstBlock..nextBlock) are valid.\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tif height > c.nextBlock {\n\t\t// Cache has been reset (for example, checksum error)\n\t\treturn nil\n\t}\n\tif height < c.firstBlock {\n\t\t// Should never try to add a block before Sapling activation height\n\t\tLog.Fatal(\"cache.Add height below Sapling: \", height)\n\t\treturn nil\n\t}\n\tif height < c.nextBlock {\n\t\t// Should never try to \"backup\" (call Reorg() instead).\n\t\tLog.Fatal(\"cache.Add height going backwards: \", height)\n\t\treturn nil\n\t}\n\tbheight := int(block.Height)\n\n\tif bheight != height {\n\t\t// This could only happen if zcashd returned the wrong\n\t\t// block (not the height we requested).\n\t\tLog.Fatal(\"cache.Add wrong height: \", bheight, \" expecting: \", height)\n\t\treturn nil\n\t}\n\n\t// Add the new block and its length to the db files.\n\tdata, err := proto.Marshal(block)\n\tif err != nil {\n\t\treturn err\n\t}\n\tb := append(checksum(height, data), data...)\n\tn, err := c.blocksFile.Write(b)\n\tif err != nil {\n\t\tLog.Fatal(\"blocks write failed: \", err)\n\t}\n\tif n != len(b) {\n\t\tLog.Fatal(\"blocks write incorrect length: expected: \", len(b), \"written: \", n)\n\t}\n\tb = make([]byte, 4)\n\tbinary.LittleEndian.PutUint32(b, uint32(len(data)))\n\tn, err = c.lengthsFile.Write(b)\n\tif err != nil {\n\t\tLog.Fatal(\"lengths write failed: \", err)\n\t}\n\tif n != len(b) {\n\t\tLog.Fatal(\"lengths write incorrect length: expected: \", len(b), \"written: \", n)\n\t}\n\n\t// update the in-memory variables\n\toffset := c.starts[len(c.starts)-1]\n\tc.starts = append(c.starts, offset+int64(len(data)+8))\n\n\tif c.latestHash == nil {\n\t\tc.latestHash = make([]byte, len(block.Hash))\n\t}\n\tcopy(c.latestHash, block.Hash)\n\tc.nextBlock++\n\t// Invariant: m[firstBlock..nextBlock) are valid.\n\treturn nil\n}", "func GetFSInfo(path string) (total, available int) {\n\ttotal = -1\n\tavailable = -1\n\tvar buf syscall.Statfs_t\n\n\tif syscall.Statfs(path, &buf) != nil {\n\t\treturn\n\t}\n\n\tif buf.Bsize <= 0 {\n\t\treturn\n\t}\n\n\ttotal = int((uint64(buf.Bsize) * buf.Blocks) / (1000 * 1000))\n\tavailable = int((uint64(buf.Bsize) * buf.Bavail) / (1000 * 1000))\n\n\treturn\n}", "func (p *Peer) UpdateLastBlockHeight(newHeight int32) {\n\tp.statsMtx.Lock()\n\tif newHeight <= p.lastBlock {\n\t\tp.statsMtx.Unlock()\n\t\treturn\n\t}\n\tlog.Tracef(\"Updating last block height of peer %v from %v to %v\",\n\t\tp.addr, p.lastBlock, newHeight)\n\tp.lastBlock = newHeight\n\tp.statsMtx.Unlock()\n}", "func (fm FinalModelEnumUInt32) FBEAllocationSize(value EnumUInt32) int { return fm.FBESize() }", "func convertSize(size uint64) C.ulonglong {\n\treturn C.ulonglong(size * 2) // kB to 512-byte sectors\n}", "func (mio *Mio) getOptimalBlockSz(bufSz int) (bsz, gsz int) {\n if bufSz > maxM0BufSz {\n bufSz = maxM0BufSz\n }\n pver := C.m0_pool_version_find(&C.instance.m0c_pools_common,\n &mio.obj.ob_attr.oa_pver)\n if pver == nil {\n log.Panic(\"cannot find the object's pool version\")\n }\n pa := &pver.pv_attr\n if pa.pa_P < pa.pa_N + pa.pa_K + pa.pa_S {\n log.Panic(\"pool width (%v) is less than the parity group size\" +\n \" (%v + %v + %v == %v), check pool parity configuration\",\n pa.pa_P, pa.pa_N, pa.pa_K, pa.pa_S,\n pa.pa_N + pa.pa_K + pa.pa_S)\n }\n\n usz := int(C.m0_obj_layout_id_to_unit_size(mio.objLid))\n gsz = usz * int(pa.pa_N) // group size in data units only\n\n // bs should be max 4-times pool-width deep counting by 1MB units, or\n // 8-times deep counting by 512K units, 16-times deep by 256K units,\n // and so on. Several units to one target will be aggregated to make\n // fewer network RPCs, disk i/o operations and BE transactions.\n // For unit sizes of 32K or less, the koefficient (k) is 128, which\n // makes it 32K * 128 == 4MB - the maximum amount per target when\n // the performance is still good on LNet (which has max 1MB frames).\n // XXX: it may be different on libfabric, should be re-measured.\n k := C.uint(128 / ((usz + 0x7fff) / 0x8000))\n if k == 0 {\n k = 1\n }\n // P * N / (N + K + S) - number of data units to span the pool-width\n maxBs := int(k * C.uint(usz) * pa.pa_P * pa.pa_N /\n (pa.pa_N + pa.pa_K + pa.pa_S))\n maxBs = roundup(maxBs, gsz) // multiple of group size\n\n if bufSz >= maxBs {\n return maxBs, gsz\n } else if bufSz <= gsz {\n return gsz, gsz\n } else {\n return roundup(bufSz, gsz), gsz\n }\n}", "func (constr Construction) BlockSize() int { return 16 }", "func (o KubernetesClusterDefaultNodePoolLinuxOsConfigPtrOutput) SwapFileSizeMb() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v *KubernetesClusterDefaultNodePoolLinuxOsConfig) *int {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.SwapFileSizeMb\n\t}).(pulumi.IntPtrOutput)\n}", "func (b *Block) Size() int {\n\treturn len(b.wire)\n}", "func (devices *DeviceSet) ResizePool(size int64) error {\n\tdirname := devices.loopbackDir()\n\tdatafilename := path.Join(dirname, \"data\")\n\tif len(devices.dataDevice) > 0 {\n\t\tdatafilename = devices.dataDevice\n\t}\n\tmetadatafilename := path.Join(dirname, \"metadata\")\n\tif len(devices.metadataDevice) > 0 {\n\t\tmetadatafilename = devices.metadataDevice\n\t}\n\n\tdatafile, err := os.OpenFile(datafilename, os.O_RDWR, 0)\n\tif datafile == nil {\n\t\treturn err\n\t}\n\tdefer datafile.Close()\n\n\tfi, err := datafile.Stat()\n\tif fi == nil {\n\t\treturn err\n\t}\n\n\tif fi.Size() > size {\n\t\treturn fmt.Errorf(\"devmapper: Can't shrink file\")\n\t}\n\n\tdataloopback := loopback.FindLoopDeviceFor(datafile)\n\tif dataloopback == nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to find loopback mount for: %s\", datafilename)\n\t}\n\tdefer dataloopback.Close()\n\n\tmetadatafile, err := os.OpenFile(metadatafilename, os.O_RDWR, 0)\n\tif metadatafile == nil {\n\t\treturn err\n\t}\n\tdefer metadatafile.Close()\n\n\tmetadataloopback := loopback.FindLoopDeviceFor(metadatafile)\n\tif metadataloopback == nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to find loopback mount for: %s\", metadatafilename)\n\t}\n\tdefer metadataloopback.Close()\n\n\t// Grow loopback file\n\tif err := datafile.Truncate(size); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to grow loopback file: %s\", err)\n\t}\n\n\t// Reload size for loopback device\n\tif err := loopback.SetCapacity(dataloopback); err != nil {\n\t\treturn fmt.Errorf(\"unable to update loopback capacity: %s\", err)\n\t}\n\n\t// Suspend the pool\n\tif err := devicemapper.SuspendDevice(devices.getPoolName()); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to suspend pool: %s\", err)\n\t}\n\n\t// Reload with the new block sizes\n\tif err := devicemapper.ReloadPool(devices.getPoolName(), dataloopback, metadataloopback, devices.thinpBlockSize); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to reload pool: %s\", err)\n\t}\n\n\t// Resume the pool\n\tif err := devicemapper.ResumeDevice(devices.getPoolName()); err != nil {\n\t\treturn fmt.Errorf(\"devmapper: Unable to resume pool: %s\", err)\n\t}\n\n\treturn nil\n}" ]
[ "0.6398401", "0.6159081", "0.61049414", "0.6102407", "0.60230005", "0.59387565", "0.5909034", "0.58386946", "0.5696438", "0.55819315", "0.55369323", "0.55319744", "0.5520192", "0.5518587", "0.54999447", "0.5494222", "0.54749113", "0.546401", "0.54373175", "0.5408651", "0.5381527", "0.53756344", "0.53732264", "0.53259104", "0.5310815", "0.5305676", "0.5295887", "0.52399707", "0.52352995", "0.5229974", "0.5185518", "0.5182078", "0.5159868", "0.5159796", "0.51534414", "0.5148623", "0.51304775", "0.51213604", "0.51156986", "0.5092056", "0.5065336", "0.50630146", "0.50630146", "0.50522894", "0.505215", "0.50508386", "0.50464", "0.5038033", "0.502839", "0.50281864", "0.50243306", "0.50187206", "0.5018458", "0.49985227", "0.49827752", "0.4981893", "0.49743578", "0.49689808", "0.49616164", "0.4960614", "0.49561852", "0.4955078", "0.49526072", "0.49506283", "0.49401417", "0.49354264", "0.4932481", "0.49295735", "0.49248686", "0.49222845", "0.49207133", "0.49203518", "0.49185267", "0.49168202", "0.49126175", "0.49089667", "0.49028394", "0.49024814", "0.48985076", "0.48985076", "0.48923045", "0.4891238", "0.48876664", "0.4886914", "0.48844817", "0.4883141", "0.48818976", "0.48676148", "0.48655823", "0.48626712", "0.48552784", "0.48511532", "0.48387212", "0.48260468", "0.48243353", "0.4823574", "0.4821207", "0.48165795", "0.48079115", "0.48067385" ]
0.65449804
0
Du stat disk usage like linux du cmd
func Du(paths []string) { for _, path := range paths { // check whether path is a normal dir or file. fd, err := os.Open(path) if err != nil { log.Fatal(err) } fd.Close() getBlkSize(paths[0]) Result = PathStatus{} Result.Name = path filepath.Walk(path, walkFn) fmt.Printf("%+v\n", Result) } }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func DiskUsage(path string) (UsageInfo, error) {\n\tvar usage UsageInfo\n\n\tif path == \"\" {\n\t\treturn usage, fmt.Errorf(\"invalid directory\")\n\t}\n\n\t// First check whether the quota system knows about this directory\n\t// A nil quantity or error means that the path does not support quotas\n\t// or xfs_quota tool is missing and we should use other mechanisms.\n\tstartTime := time.Now()\n\tconsumption, _ := fsquota.GetConsumption(path)\n\tif consumption != nil {\n\t\tusage.Bytes = consumption.Value()\n\t\tdefer servermetrics.CollectVolumeStatCalDuration(\"fsquota\", startTime)\n\t} else {\n\t\tdefer servermetrics.CollectVolumeStatCalDuration(\"du\", startTime)\n\t}\n\n\tinodes, _ := fsquota.GetInodes(path)\n\tif inodes != nil {\n\t\tusage.Inodes = inodes.Value()\n\t}\n\n\tif inodes != nil && consumption != nil {\n\t\treturn usage, nil\n\t}\n\n\ttopLevelStat := &unix.Stat_t{}\n\terr := unix.Stat(path, topLevelStat)\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\t// dedupedInode stores inodes that could be duplicates (nlink > 1)\n\tdedupedInodes := make(map[uint64]struct{})\n\n\terr = filepath.Walk(path, func(path string, info os.FileInfo, err error) error {\n\t\t// ignore files that have been deleted after directory was read\n\t\tif os.IsNotExist(err) {\n\t\t\treturn nil\n\t\t}\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"unable to count inodes for %s: %s\", path, err)\n\t\t}\n\n\t\t// according to the docs, Sys can be nil\n\t\tif info.Sys() == nil {\n\t\t\treturn fmt.Errorf(\"fileinfo Sys is nil\")\n\t\t}\n\n\t\ts, ok := info.Sys().(*syscall.Stat_t)\n\t\tif !ok {\n\t\t\treturn fmt.Errorf(\"unsupported fileinfo; could not convert to stat_t\")\n\t\t}\n\n\t\tif s.Dev != topLevelStat.Dev {\n\t\t\t// don't descend into directories on other devices\n\t\t\treturn filepath.SkipDir\n\t\t}\n\n\t\t// Dedupe hardlinks\n\t\tif s.Nlink > 1 {\n\t\t\tif _, ok := dedupedInodes[s.Ino]; !ok {\n\t\t\t\tdedupedInodes[s.Ino] = struct{}{}\n\t\t\t} else {\n\t\t\t\treturn nil\n\t\t\t}\n\t\t}\n\n\t\tif consumption == nil {\n\t\t\tusage.Bytes += int64(s.Blocks) * int64(512) // blocksize in bytes\n\t\t}\n\n\t\tif inodes == nil {\n\t\t\tusage.Inodes++\n\t\t}\n\n\t\treturn nil\n\t})\n\n\treturn usage, err\n}", "func DiskUsage(path string) (disk DiskStatus, err error) {\n\tfs := syscall.Statfs_t{}\n\terr = syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}", "func DiskUsage(path string) (disk DiskStatus) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}", "func DiskUsage(path string) (disk DiskStatus) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}", "func DiskUsage(path string) (disk DiskStatus) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}", "func DiskUsage(path string) (disk DiskStatus) {\n\n\tif path == \"\" {\n\t\treturn\n\t}\n\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tdisk.All = fs.Blocks * uint64(fs.Bsize)\n\tdisk.Free = fs.Bfree * uint64(fs.Bsize)\n\tdisk.Used = disk.All - disk.Free\n\treturn disk\n}", "func Usage(path string) (*Status, error) {\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\treturn nil, errors.Wrapf(err, \"get disk status for path '%s' failed\", path)\n\t}\n\tvar disk Status\n\tdisk.BytesTotal = fs.Blocks * uint64(fs.Bsize)\n\tdisk.BytesFree = fs.Bfree * uint64(fs.Bsize)\n\tdisk.BytesUsed = disk.BytesTotal - disk.BytesFree\n\n\tdisk.InodesTotal = uint64(fs.Files)\n\tdisk.InodesFree = uint64(fs.Ffree)\n\tdisk.InodesUsed = disk.InodesTotal - disk.InodesFree\n\treturn &disk, nil\n}", "func du() {\n\trf, err := os.Stat(\"/\")\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\trdev := rf.Sys().(*syscall.Stat_t).Dev\n\tf := os.DirFS(\"/\")\n\terr = fs.WalkDir(f, \".\", func(p string, d fs.DirEntry, err error) error {\n\t\tif err != nil {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\ti, err := d.Info()\n\t\tif err != nil {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\tsys := i.Sys()\n\t\tif sys == nil {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\tdev := sys.(*syscall.Stat_t).Dev\n\t\tfmt.Printf(\"p=%v size=%v dev=%v\\n\", p, i.Size(), dev)\n\t\tif dev != rdev {\n\t\t\treturn fs.SkipDir\n\t\t}\n\t\treturn nil\n\t})\n\tif err != nil {\n\t\tlog.Print(err)\n\t}\n}", "func (c *DefaultChecker) DiskPerf() (info DiskPerfInfo, err error) {\n\tvar data []byte\n\n\texe := execCommand(\"dd\", \"bs=1M\", \"count=256\", \"if=/dev/zero\", fmt.Sprintf(\"of=%v\", path.Join(c.WritePath, \"test\")), \"conv=fdatasync\")\n\tdata, err = exe.CombinedOutput()\n\tgohtypes.PanicIfError(\"Not possible to execute the 'dd' command.\", 500, err)\n\n\tout := string(data)\n\tlogrus.Info(out)\n\n\trate := strings.Split(strings.Split(strings.Replace(out, \"\\n\", \"\", -1), \" s, \")[1], \" \")\n\n\tinfo = DiskPerfInfo{\n\t\tWriteSpeed: rate[0],\n\t\tUnit: rate[1],\n\t}\n\treturn\n}", "func DiskUsage(path string) (disk DiskStatus) {\n\th := syscall.MustLoadDLL(\"kernel32.dll\")\n\tc := h.MustFindProc(\"GetDiskFreeSpaceExW\")\n\tlpFreeBytesAvailable := int64(0)\n\tlpTotalNumberOfBytes := int64(0)\n\tlpTotalNumberOfFreeBytes := int64(0)\n\t_, _, err := c.Call(uintptr(unsafe.Pointer(syscall.StringToUTF16Ptr(path[:2]))),\n\t\tuintptr(unsafe.Pointer(&lpFreeBytesAvailable)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfBytes)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfFreeBytes)))\n\tif err != nil {\n\t\tif !strings.Contains(fmt.Sprint(err), \"successfully\") {\n\t\t\tlogger.Error(\"Error during retrieving memory statistic:\", err)\n\t\t\treturn\n\t\t}\n\t}\n\tdisk.All = uint64(lpTotalNumberOfBytes)\n\tdisk.Free = uint64(lpTotalNumberOfFreeBytes)\n\tdisk.Used = disk.All - disk.Free\n\treturn\n}", "func GetDiskUsage(path string) ([]float64, error) {\n\tvar usage []float64 = make([]float64, 6) // Data to return\n\tvar sc syscall.Statfs_t // Filesystem stat\n\tvar err error // Error catching\n\tvar buffer bytes.Buffer // Buffer for json indent\n\tvar content []byte // Json's content\n\n\t// Get filesystem stat\n\terr = syscall.Statfs(path, &sc)\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\t// Convert structure => json\n\tcontent, err = json.Marshal(sc)\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\t// Indent json\n\tjson.Indent(&buffer, content, \"\", \" \")\n\n\t// Set data to return\n\tusage[0] = float64(sc.Bsize) * float64(sc.Blocks) // TotalStorage\n\tusage[1] = float64(sc.Files) // TotalFileNodes\n\tusage[3] = float64(sc.Bsize) * float64(sc.Bfree) // FreeStorage\n\tusage[2] = usage[0] - usage[3] // UsedStorage\n\tusage[5] = float64(sc.Ffree) // FreeFileNodes\n\tusage[4] = usage[1] - usage[5] // UsedFileNodes\n\n\treturn usage, nil\n}", "func DiskUsage(path string) (*DiskStatus, error) {\n\tlpDirectoryName, err := syscall.UTF16PtrFromString(path)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tlpFreeBytesAvailable := int64(0)\n\tlpTotalNumberOfBytes := int64(0)\n\tlpTotalNumberOfFreeBytes := int64(0)\n\t_, _, e := syscall.Syscall6(pGetDiskFreeSpaceEx, 4,\n\t\tuintptr(unsafe.Pointer(lpDirectoryName)),\n\t\tuintptr(unsafe.Pointer(&lpFreeBytesAvailable)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfBytes)),\n\t\tuintptr(unsafe.Pointer(&lpTotalNumberOfFreeBytes)), 0, 0)\n\tif e != 0 {\n\t\treturn nil, errors.New(e.Error())\n\t}\n\tstatus := &DiskStatus{\n\t\tAll: lpTotalNumberOfBytes,\n\t\tFree: lpFreeBytesAvailable,\n\t}\n\tstatus.Used = status.All - status.Free\n\treturn status, nil\n}", "func GetDiskUsage(rw http.ResponseWriter) error {\n\tu, r := disk.Usage(\"/\")\n\tif r != nil {\n\t\treturn r\n\t}\n\n\treturn share.JSONResponse(u, rw)\n}", "func getDiskInfo(path string) (uint64, uint64, error) {\n\ts := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &s)\n\tif err != nil {\n\t\treturn 0,0, err\n\t}\n\treservedBlocks := s.Bfree - s.Bavail\n\n\ttotal := uint64(s.Frsize) * (s.Blocks - reservedBlocks)\n\tfree := uint64(s.Frsize) * s.Bavail\n\t// Check for overflows.\n\t// https://github.com/minio/minio/issues/8035\n\t// XFS can show wrong values at times error out\n\t// in such scenarios.\n\tif free > total {\n\t\treturn 0, 0, fmt.Errorf(\"detected free space (%d) > total disk space (%d), fs corruption at (%s). please run 'fsck'\", free, total, path)\n\t}\n\n\treturn total, free, nil\n}", "func Diskuse(path string) (cap uint64, used uint64) {\n\n\tfs := syscall.Statfs_t{}\n\terr := syscall.Statfs(path, &fs)\n\tif err != nil {\n\t\tfmt.Println(err)\n\t}\n\n\tcap = fs.Blocks * uint64(fs.Bsize)\n\tfree := fs.Bfree * uint64(fs.Bsize) // yup, I just did that\n\tused = cap - free\n\treturn cap, used\n}", "func (d *Datastore) DiskUsage() (uint64, error) {\n\tvar total uint64 = 0\n\n\terr := d.meta.ListKeys(func(key []byte) error {\n\t\tmeta, err := d.meta.GetMetadata(key)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\ttotal += uint64(meta.StorageSize)\n\t\treturn nil\n\t})\n\n\treturn total, err\n}", "func Disk(pt ...bool) ([]*disk.UsageStat, error) {\n\tvar ptBool bool\n\tif len(pt) > 0 {\n\t\tptBool = pt[0]\n\t}\n\n\tvar usage []*disk.UsageStat\n\tparts, err := disk.Partitions(ptBool)\n\n\tfor _, part := range parts {\n\t\tuse, err := disk.Usage(part.Mountpoint)\n\t\tif err != nil {\n\t\t\treturn usage, err\n\t\t}\n\t\tusage = append(usage, use)\n\t\t// printUsage(use)\n\t}\n\n\treturn usage, err\n}", "func DiskStat() DiskInfo {\n\tusage, _ := disk.Usage(\"./\")\n\tpartitions, _ := disk.Partitions(true)\n\tiOCounters, _ := disk.IOCounters()\n\n\tue := usage.String()\n\n\tdiskstat := DiskInfo{\n\t\tUsage: ue,\n\t\tPartition: partitions,\n\t\tIOCounters: iOCounters,\n\t}\n\n\treturn diskstat\n}", "func PrintDiskStats() {\n\tfmt.Println(DiskStats)\n}", "func (p *btrfsPool) Usage() (usage Usage, err error) {\n\tmnt, err := p.Mounted()\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\tvolumes, err := p.Volumes()\n\n\tif err != nil {\n\t\treturn usage, errors.Wrapf(err, \"failed to list pool '%s' volumes\", mnt)\n\t}\n\n\tusage.Size = p.device.Size\n\n\tfor _, volume := range volumes {\n\t\tvol, err := volume.Usage()\n\t\tif err != nil {\n\t\t\treturn Usage{}, errors.Wrapf(err, \"failed to calculate volume '%s' usage\", volume.Path())\n\t\t}\n\n\t\tusage.Used += vol.Used\n\t\tusage.Excl += vol.Excl\n\t}\n\n\treturn\n}", "func (s *SigarChecker) GetDiskUsage(path string) (usePercent float64, avail uint64) {\n\tvar fsu sigar.FileSystemUsage\n\n\terr := fsu.Get(path)\n\n\tutil.Must(err)\n\n\treturn fsu.UsePercent(), fsu.Avail\n}", "func DirSize(dir string, excludes []string) (int, error) {\n\targs := []string{\"-bs\", dir}\n\tfor _, s := range excludes {\n\t\targs = append(args, fmt.Sprintf(\"--exclude=\\\"%s\\\"\", s))\n\t}\n\n\tcmd := exec.Command(\"/usr/bin/du\", args...)\n\tdata, err := cmd.Output()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\n\treturn strconv.Atoi(strings.Split(string(data), \"\\t\")[0])\n}", "func testDiskUsageDoubleCount(dirFunc mkShardFunc, t *testing.T) {\n\ttemp, cleanup := tempdir(t)\n\tdefer cleanup()\n\tdefer checkTemp(t, temp)\n\n\tfs, err := flatfs.CreateOrOpen(temp, dirFunc(2), false)\n\tif err != nil {\n\t\tt.Fatalf(\"New fail: %v\\n\", err)\n\t}\n\tdefer fs.Close()\n\n\tvar count int\n\tvar wg sync.WaitGroup\n\ttestKey := datastore.NewKey(\"TEST\")\n\n\tput := func() {\n\t\tdefer wg.Done()\n\t\tfor i := 0; i < count; i++ {\n\t\t\tv := []byte(\"10bytes---\")\n\t\t\terr := fs.Put(bg, testKey, v)\n\t\t\tif err != nil {\n\t\t\t\tt.Errorf(\"Put fail: %v\\n\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\tdel := func() {\n\t\tdefer wg.Done()\n\t\tfor i := 0; i < count; i++ {\n\t\t\terr := fs.Delete(bg, testKey)\n\t\t\tif err != nil && !strings.Contains(err.Error(), \"key not found\") {\n\t\t\t\tt.Errorf(\"Delete fail: %v\\n\", err)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Add one element and then remove it and check disk usage\n\t// makes sense\n\tcount = 1\n\twg.Add(2)\n\tput()\n\tdu, _ := fs.DiskUsage(bg)\n\tdel()\n\tdu2, _ := fs.DiskUsage(bg)\n\tif du-10 != du2 {\n\t\tt.Error(\"should have deleted exactly 10 bytes:\", du, du2)\n\t}\n\n\t// Add and remove many times at the same time\n\tcount = 200\n\twg.Add(4)\n\tgo put()\n\tgo del()\n\tgo put()\n\tgo del()\n\twg.Wait()\n\n\tdu3, _ := fs.DiskUsage(bg)\n\thas, err := fs.Has(bg, testKey)\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\n\tif has { // put came last\n\t\tif du3 != du {\n\t\t\tt.Error(\"du should be the same as after first put:\", du, du3)\n\t\t}\n\t} else { // delete came last\n\t\tif du3 != du2 {\n\t\t\tt.Error(\"du should be the same as after first delete:\", du2, du3)\n\t\t}\n\t}\n}", "func getDiskSize(path string) (uint64, error) {\n\tdiskInfo := exec.Command(\"qemu-img\", \"info\", \"--output\", \"json\", path)\n\tstdout, err := diskInfo.StdoutPipe()\n\tif err != nil {\n\t\treturn 0, err\n\t}\n\tif err := diskInfo.Start(); err != nil {\n\t\treturn 0, err\n\t}\n\ttmpInfo := struct {\n\t\tVirtualSize uint64 `json:\"virtual-size\"`\n\t\tFilename string `json:\"filename\"`\n\t\tClusterSize int64 `json:\"cluster-size\"`\n\t\tFormat string `json:\"format\"`\n\t\tFormatSpecific struct {\n\t\t\tType string `json:\"type\"`\n\t\t\tData map[string]string `json:\"data\"`\n\t\t}\n\t\tDirtyFlag bool `json:\"dirty-flag\"`\n\t}{}\n\tif err := json.NewDecoder(stdout).Decode(&tmpInfo); err != nil {\n\t\treturn 0, err\n\t}\n\tif err := diskInfo.Wait(); err != nil {\n\t\treturn 0, err\n\t}\n\treturn tmpInfo.VirtualSize, nil\n}", "func Stats(c *libvirt.Connect, uuid string) error {\n\t//Check exists\n\td, err := c.LookupDomainByUUIDString(uuid)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed to lookup: %s\", err)\n\t}\n\n\t//Check is running\n\ts, _, err := d.GetState()\n\tif err != nil {\n\t\treturn fmt.Errorf(\"failed check state: %s\", err)\n\t}\n\tif s != libvirt.DOMAIN_RUNNING {\n\t\treturn fmt.Errorf(\"domain not running: %d\", s)\n\t}\n\n\tmemStats, err := memStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"STAT: %+v\\n\", memStats)\n\tfmt.Printf(\"STAT Used: %+v\\n\", memStats.Available-memStats.Unused)\n\tfmt.Printf(\"STAT Last: %s\\n\", time.Unix(int64(memStats.LastUpdate), 0))\n\n\tcpuStats, total, err := cpuStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"%+v\\n\", cpuStats)\n\tfmt.Printf(\"Total: %+#v\\n\", total)\n\n\tnetStats, err := netStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"NET: %+v\\n\", netStats)\n\n\t_, dTotal, err := diskStats(d)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfmt.Printf(\"DISK: %+v\\n\", dTotal)\n\n\treturn nil\n}", "func Stat(path, index string, raw bool) (value string) {\n\tvar row = map[string]int{\"quota\": 3, \"usage\": 2}\n\n\targs := []string{\"qgroup\", \"show\", \"-re\", config.Agent.LxcPrefix}\n\tif raw {\n\t\targs = []string{\"qgroup\", \"show\", \"-re\", \"--raw\", config.Agent.LxcPrefix}\n\t}\n\tout, err := exec.Command(\"btrfs\", args...).Output()\n\tlog.Check(log.FatalLevel, \"Getting btrfs stats\", err)\n\tind := id(path)\n\tscanner := bufio.NewScanner(bytes.NewReader(out))\n\tfor scanner.Scan() {\n\t\tif line := strings.Fields(scanner.Text()); len(line) > 3 && strings.HasSuffix(line[0], \"/\"+ind) {\n\t\t\tvalue = line[row[index]]\n\t\t}\n\t}\n\treturn value\n}", "func (f *StorageFingerprint) diskFree(path string) (volume string, total, free uint64, err error) {\n\tabsPath, err := filepath.Abs(path)\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to determine absolute path for %s\", path)\n\t}\n\n\t// Use -k to standardize the output values between darwin and linux\n\tvar dfArgs string\n\tif runtime.GOOS == \"linux\" {\n\t\t// df on linux needs the -P option to prevent linebreaks on long filesystem paths\n\t\tdfArgs = \"-kP\"\n\t} else {\n\t\tdfArgs = \"-k\"\n\t}\n\n\tmountOutput, err := exec.Command(\"df\", dfArgs, absPath).Output()\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to determine mount point for %s\", absPath)\n\t}\n\t// Output looks something like:\n\t//\tFilesystem 1024-blocks Used Available Capacity iused ifree %iused Mounted on\n\t//\t/dev/disk1 487385240 423722532 63406708 87% 105994631 15851677 87% /\n\t//\t[0] volume [1] capacity [2] SKIP [3] free\n\tlines := strings.Split(string(mountOutput), \"\\n\")\n\tif len(lines) < 2 {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse `df` output; expected at least 2 lines\")\n\t}\n\tfields := strings.Fields(lines[1])\n\tif len(fields) < 4 {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse `df` output; expected at least 4 columns\")\n\t}\n\tvolume = fields[0]\n\n\ttotal, err = strconv.ParseUint(fields[1], 10, 64)\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse storage.bytestotal size in kilobytes\")\n\t}\n\t// convert to bytes\n\ttotal *= 1024\n\n\tfree, err = strconv.ParseUint(fields[3], 10, 64)\n\tif err != nil {\n\t\treturn \"\", 0, 0, fmt.Errorf(\"failed to parse storage.bytesfree size in kilobytes\")\n\t}\n\t// convert to bytes\n\tfree *= 1024\n\n\treturn volume, total, free, nil\n}", "func GetDiskUsageStats(path string) (uint64, uint64, error) {\n\tvar dirSize, inodeCount uint64\n\n\terr := filepath.Walk(path, func(path string, info os.FileInfo, err error) error {\n\t\tfileStat, error := os.Lstat(path)\n\t\tif error != nil {\n\t\t\tif fileStat.Mode()&os.ModeSymlink != 0 {\n\t\t\t\t// Is a symlink; no error should be returned\n\t\t\t\treturn nil\n\t\t\t}\n\t\t\treturn error\n\t\t}\n\n\t\tdirSize += uint64(info.Size())\n\t\tinodeCount++\n\n\t\treturn nil\n\t})\n\n\tif err != nil {\n\t\treturn 0, 0, err\n\t}\n\n\treturn dirSize, inodeCount, err\n}", "func (v *btrfsVolume) Usage() (usage Usage, err error) {\n\tctx := context.TODO()\n\tinfo, err := v.utils.SubvolumeInfo(ctx, v.Path())\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\tgroups, err := v.utils.QGroupList(ctx, v.Path())\n\tif err != nil {\n\t\treturn usage, err\n\t}\n\n\tgroup, ok := groups[fmt.Sprintf(\"0/%d\", info.ID)]\n\tif !ok {\n\t\t// no qgroup associated with the subvolume id! means no limit, but we also\n\t\t// cannot read the usage.\n\t\treturn\n\t}\n\n\t// used is basically amount of space reserved for this\n\t// volume. We assume that's total usage of the volume\n\tused := group.MaxRfer\n\n\tif used == 0 {\n\t\t// in case no limit is set on the subvolume, we assume\n\t\t// it's size is the size of the files on that volumes\n\t\t// or a special case when the volume is a zdb volume\n\t\tused, err = volumeUsage(v.Path())\n\t\tif err != nil {\n\t\t\treturn usage, errors.Wrap(err, \"failed to get subvolume usage\")\n\t\t}\n\t}\n\n\treturn Usage{Used: used, Size: group.MaxRfer, Excl: group.Excl}, nil\n}", "func (f *Fs) About(ctx context.Context) (*fs.Usage, error) {\n\tvar available, total, free int64\n\troot, e := syscall.UTF16PtrFromString(f.root)\n\tif e != nil {\n\t\treturn nil, fmt.Errorf(\"failed to read disk usage: %w\", e)\n\t}\n\t_, _, e1 := getFreeDiskSpace.Call(\n\t\tuintptr(unsafe.Pointer(root)),\n\t\tuintptr(unsafe.Pointer(&available)), // lpFreeBytesAvailable - for this user\n\t\tuintptr(unsafe.Pointer(&total)), // lpTotalNumberOfBytes\n\t\tuintptr(unsafe.Pointer(&free)), // lpTotalNumberOfFreeBytes\n\t)\n\tif e1 != syscall.Errno(0) {\n\t\treturn nil, fmt.Errorf(\"failed to read disk usage: %w\", e1)\n\t}\n\tusage := &fs.Usage{\n\t\tTotal: fs.NewUsageValue(total), // quota of bytes that can be used\n\t\tUsed: fs.NewUsageValue(total - free), // bytes in use\n\t\tFree: fs.NewUsageValue(available), // bytes which can be uploaded before reaching the quota\n\t}\n\treturn usage, nil\n}", "func print_stats(){\nfmt.Print(\"\\nMemory usage statistics:\\n\")\nfmt.Printf(\"%v names\\n\",len(name_dir))\nfmt.Printf(\"%v replacement texts\\n\",len(text_info))\n}", "func printIostat(v *gocui.View, s stat.Diskstats) error {\n\t// print header\n\t_, err := fmt.Fprintf(v, \"\\033[30;47m Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await r_await w_await %%util\\033[0m\\n\")\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfor i := 0; i < len(s); i++ {\n\t\t// skip devices which never do IOs\n\t\tif s[i].Completed == 0 {\n\t\t\tcontinue\n\t\t}\n\n\t\t// print stats\n\t\t_, err := fmt.Fprintf(v, \"%20s\\t%10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f %10.2f\\n\",\n\t\t\ts[i].Device,\n\t\t\ts[i].Rmerged, s[i].Wmerged, s[i].Rcompleted, s[i].Wcompleted,\n\t\t\ts[i].Rsectors, s[i].Wsectors, s[i].Arqsz, s[i].Tweighted,\n\t\t\ts[i].Await, s[i].Rawait, s[i].Wawait, s[i].Util,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func stat(t *kernel.Task, d *fs.Dirent, dirPath bool, statAddr usermem.Addr) error {\n\tif dirPath && !fs.IsDir(d.Inode.StableAttr) {\n\t\treturn syserror.ENOTDIR\n\t}\n\tuattr, err := d.Inode.UnstableAttr(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\ts := statFromAttrs(t, d.Inode.StableAttr, uattr)\n\treturn s.CopyOut(t, statAddr)\n}", "func statfsImpl(t *kernel.Task, d *fs.Dirent, addr usermem.Addr) error {\n\tinfo, err := d.Inode.StatFS(t)\n\tif err != nil {\n\t\treturn err\n\t}\n\t// Construct the statfs structure and copy it out.\n\tstatfs := linux.Statfs{\n\t\tType: info.Type,\n\t\t// Treat block size and fragment size as the same, as\n\t\t// most consumers of this structure will expect one\n\t\t// or the other to be filled in.\n\t\tBlockSize: d.Inode.StableAttr.BlockSize,\n\t\tBlocks: info.TotalBlocks,\n\t\t// We don't have the concept of reserved blocks, so\n\t\t// report blocks free the same as available blocks.\n\t\t// This is a normal thing for filesystems, to do, see\n\t\t// udf, hugetlbfs, tmpfs, among others.\n\t\tBlocksFree: info.FreeBlocks,\n\t\tBlocksAvailable: info.FreeBlocks,\n\t\tFiles: info.TotalFiles,\n\t\tFilesFree: info.FreeFiles,\n\t\t// Same as Linux for simple_statfs, see fs/libfs.c.\n\t\tNameLength: linux.NAME_MAX,\n\t\tFragmentSize: d.Inode.StableAttr.BlockSize,\n\t\t// Leave other fields 0 like simple_statfs does.\n\t}\n\t_, err = t.CopyOut(addr, &statfs)\n\treturn err\n}", "func TestDiskUsage(t *testing.T) {\n\tt.Parallel()\n\twinners := []parameters{[]string{\"/\", \"99\"}, []string{\"/\", \"98\"}}\n\tlosers := []parameters{[]string{\"/\", \"1\"}, []string{\"/\", \"2\"}}\n\ttestInputs(t, diskUsage, winners, losers)\n}", "func (fsys *FS) stat(node mountlib.Node, stat *fuse.Stat_t) (errc int) {\n\tvar Size uint64\n\tvar Blocks uint64\n\tvar modTime time.Time\n\tvar Mode os.FileMode\n\tswitch x := node.(type) {\n\tcase *mountlib.Dir:\n\t\tmodTime = x.ModTime()\n\t\tMode = mountlib.DirPerms | fuse.S_IFDIR\n\tcase *mountlib.File:\n\t\tvar err error\n\t\tmodTime, Size, Blocks, err = x.Attr(mountlib.NoModTime)\n\t\tif err != nil {\n\t\t\treturn translateError(err)\n\t\t}\n\t\tMode = mountlib.FilePerms | fuse.S_IFREG\n\t}\n\t//stat.Dev = 1\n\tstat.Ino = node.Inode() // FIXME do we need to set the inode number?\n\tstat.Mode = uint32(Mode)\n\tstat.Nlink = 1\n\tstat.Uid = mountlib.UID\n\tstat.Gid = mountlib.GID\n\t//stat.Rdev\n\tstat.Size = int64(Size)\n\tt := fuse.NewTimespec(modTime)\n\tstat.Atim = t\n\tstat.Mtim = t\n\tstat.Ctim = t\n\tstat.Blksize = 512\n\tstat.Blocks = int64(Blocks)\n\tstat.Birthtim = t\n\t// fs.Debugf(nil, \"stat = %+v\", *stat)\n\treturn 0\n}", "func Info(path string) (int64, int64, int64, int64, int64, int64, error) {\n\tstatfs := &unix.Statfs_t{}\n\terr := unix.Statfs(path, statfs)\n\tif err != nil {\n\t\treturn 0, 0, 0, 0, 0, 0, err\n\t}\n\n\t// Available is blocks available * fragment size\n\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\n\n\t// Capacity is total block count * fragment size\n\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\n\n\t// Usage is block being used * fragment size (aka block size).\n\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\n\n\tinodes := int64(statfs.Files)\n\tinodesFree := int64(statfs.Ffree)\n\tinodesUsed := inodes - inodesFree\n\n\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\n}", "func getFSSizeMb(f *framework.Framework, pod *v1.Pod) (int64, error) {\n\tvar output string\n\tvar err error\n\n\tif supervisorCluster {\n\t\tnamespace := getNamespaceToRunTests(f)\n\t\tcmd := []string{\"exec\", pod.Name, \"--namespace=\" + namespace, \"--\", \"/bin/sh\", \"-c\", \"df -Tkm | grep /mnt/volume1\"}\n\t\toutput = framework.RunKubectlOrDie(namespace, cmd...)\n\t\tgomega.Expect(strings.Contains(output, ext4FSType)).NotTo(gomega.BeFalse())\n\t} else {\n\t\toutput, _, err = fvolume.PodExec(f, pod, \"df -T -m | grep /mnt/volume1\")\n\t\tif err != nil {\n\t\t\treturn -1, fmt.Errorf(\"unable to find mount path via `df -T`: %v\", err)\n\t\t}\n\t}\n\n\tarrMountOut := strings.Fields(string(output))\n\tif len(arrMountOut) <= 0 {\n\t\treturn -1, fmt.Errorf(\"error when parsing output of `df -T`. output: %s\", string(output))\n\t}\n\tvar devicePath, strSize string\n\tdevicePath = arrMountOut[0]\n\tif devicePath == \"\" {\n\t\treturn -1, fmt.Errorf(\"error when parsing output of `df -T` to find out devicePath of /mnt/volume1. output: %s\",\n\t\t\tstring(output))\n\t}\n\tstrSize = arrMountOut[2]\n\tif strSize == \"\" {\n\t\treturn -1, fmt.Errorf(\"error when parsing output of `df -T` to find out size of /mnt/volume1: output: %s\",\n\t\t\tstring(output))\n\t}\n\n\tintSizeInMb, err := strconv.ParseInt(strSize, 10, 64)\n\tif err != nil {\n\t\treturn -1, fmt.Errorf(\"failed to parse size %s into int size\", strSize)\n\t}\n\n\treturn intSizeInMb, nil\n}", "func getMemUsage(memStat types.MemoryStats) uint64 {\n\t// Version 1 of the Linux cgroup API uses total_inactive_file\n\tif v, ok := memStat.Stats[\"total_inactive_file\"]; ok && v < memStat.Usage {\n\t\treturn memStat.Usage - v\n\t}\n\n\t// Version 2 of the Linux cgroup API uses inactive_file\n\tif v := memStat.Stats[\"inactive_file\"]; v < memStat.Usage {\n\t\treturn memStat.Usage - v\n\t}\n\n\treturn memStat.Usage\n}", "func DiskUsed() (uint64, error) {\n\tusage, err := Disk()\n\t// for i := 0; i < len(usage); i++ {\n\tif len(usage) > 0 {\n\t\tuseDisk := usage[0].Used\n\t\treturn useDisk, err\n\t}\n\n\treturn 0, err\n}", "func stat(name string) (*info, error) {\n\tf, err := os.Open(filepath.Clean(name))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tvar di info\n\n\tvar buf bytes.Buffer\n\n\t_, err = io.CopyN(&buf, f, int64(binary.Size(di.Header)))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfi, err := f.Stat()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = f.Close()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\terr = binary.Read(&buf, binary.LittleEndian, &di.Header)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tif di.Header.MagicNumber != 0x564d444b { // SPARSE_MAGICNUMBER\n\t\treturn nil, ErrInvalidFormat\n\t}\n\n\tif di.Header.Flags&(1<<16) == 0 { // SPARSEFLAG_COMPRESSED\n\t\t// Needs to be converted, for example:\n\t\t// vmware-vdiskmanager -r src.vmdk -t 5 dst.vmdk\n\t\t// qemu-img convert -O vmdk -o subformat=streamOptimized src.vmdk dst.vmdk\n\t\treturn nil, ErrInvalidFormat\n\t}\n\n\tdi.Capacity = di.Header.Capacity * 512 // VMDK_SECTOR_SIZE\n\tdi.Size = fi.Size()\n\tdi.Name = filepath.Base(name)\n\tdi.ImportName = strings.TrimSuffix(di.Name, \".vmdk\")\n\n\treturn &di, nil\n}", "func (c *Cache) updateStats() error {\n\tvar newUsed int64\n\terr := c.walk(func(osPath string, fi os.FileInfo, name string) error {\n\t\tif !fi.IsDir() {\n\t\t\t// Update the atime with that of the file\n\t\t\tatime := times.Get(fi).AccessTime()\n\t\t\tc.updateStat(name, atime, fi.Size())\n\t\t\tnewUsed += fi.Size()\n\t\t} else {\n\t\t\tc.cacheDir(name)\n\t\t}\n\t\treturn nil\n\t})\n\tc.itemMu.Lock()\n\tc.used = newUsed\n\tc.itemMu.Unlock()\n\treturn err\n}", "func (a *app) printStat() {\n\tfmt.Printf(\"Total capacity %vgb \\n\", a.sq.Limit/1024/1024/1024)\n\tfmt.Printf(\"Usage %vmb \\n\", a.sq.Usage/1024/1024)\n\tfmt.Printf(\"In Drive %vmb \\n\", a.sq.UsageInDrive/1024/1024)\n\tfmt.Printf(\"In Trash %vmb \\n\", a.sq.UsageInDriveTrash/1024/1024)\n\n\tperOfUse := float64(a.sq.UsageInDrive) / float64(a.sq.Limit) * 100\n\tfmt.Println(\"Current percentage of use:\", perOfUse)\n}", "func (s *SystemMetrics) GetDiskIOStats(c chan *DiskStats) {\n\tbefore, err := disk.Get()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\ttime.Sleep(time.Duration(time.Second * 3))\n\n\tnow, err := disk.Get()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\tc <- &DiskStats{\n\t\t// default is always in bytes. hence, convert into the required format.\n\t\tDiskIO: (int(now.Used) - int(before.Used)) / 1000, // in kilo-bytes\n\t\tCached: int(now.Cached / 1000000), // mega-bytes\n\t}\n}", "func DiskTotal() (uint64, error) {\n\tusage, err := Disk()\n\t// for i := 0; i < len(usage); i++ {\n\tif len(usage) > 0 {\n\t\ttotalDisk := usage[0].Total\n\t\treturn totalDisk, err\n\t}\n\n\treturn 0, err\n}", "func GetFSInfo(ch chan metrics.Metric) {\n\tmountedFS, err := disk.Partitions(false)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"collector\": \"filesystem\",\n\t\t\t\"os\": \"linux\",\n\t\t\t\"action\": \"GetFileSystems\",\n\t\t}).Errorf(\"Unable to find mounted filesystems: %+v\", err)\n\t}\n\tfor _, FSs := range mountedFS {\n\t\tfsStats, err := disk.Usage(FSs.Mountpoint)\n\t\tif err != nil {\n\t\t\tlog.WithFields(log.Fields{\n\t\t\t\t\"collector\": \"filesystem\",\n\t\t\t\t\"os\": \"linux\",\n\t\t\t\t\"action\": \"GetFSStats\",\n\t\t\t}).Errorf(\"Unable to get stats from mounted filesystem: %+v\", err)\n\t\t}\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"collector\": \"filesystem\",\n\t\t\t\"os\": \"linux\",\n\t\t}).Debug(fsStats)\n\t\tfsStat := metrics.FileSystem{\n\t\t\tFileSystem: *fsStats,\n\t\t}\n\t\tch <- fsStat\n\t}\n}", "func (i *ImageService) LayerDiskUsage(ctx context.Context) (int64, error) {\n\tvar allLayersSize int64\n\t// TODO(thaJeztah): do we need to take multiple snapshotters into account? See https://github.com/moby/moby/issues/45273\n\tsnapshotter := i.client.SnapshotService(i.snapshotter)\n\tsnapshotter.Walk(ctx, func(ctx context.Context, info snapshots.Info) error {\n\t\tusage, err := snapshotter.Usage(ctx, info.Name)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tallLayersSize += usage.Size\n\t\treturn nil\n\t})\n\treturn allLayersSize, nil\n}", "func DiskPercent() (string, error) {\n\tusage, err := Disk()\n\tif len(usage) > 0 {\n\t\tuseDisk := fmt.Sprintf(\"%.2f\", usage[0].UsedPercent)\n\t\treturn useDisk, err\n\t}\n\n\treturn \"0.00\", err\n}", "func getRemainingDiskSpace(ctxPtr *volumemgrContext) (uint64, error) {\n\n\tvar totalDiskSize uint64\n\n\tpubContentTree := ctxPtr.pubContentTreeStatus\n\titemsContentTree := pubContentTree.GetAll()\n\tfor _, iterContentTreeStatusJSON := range itemsContentTree {\n\t\titerContentTreeStatus := iterContentTreeStatusJSON.(types.ContentTreeStatus)\n\t\tif iterContentTreeStatus.State < types.LOADED {\n\t\t\tlog.Tracef(\"Content tree %s State %d < LOADED\",\n\t\t\t\titerContentTreeStatus.Key(), iterContentTreeStatus.State)\n\t\t\tcontinue\n\t\t}\n\t\ttotalDiskSize += uint64(iterContentTreeStatus.CurrentSize)\n\t}\n\n\tpubVolume := ctxPtr.pubVolumeStatus\n\titemsVolume := pubVolume.GetAll()\n\tfor _, iterVolumeStatusJSON := range itemsVolume {\n\t\titerVolumeStatus := iterVolumeStatusJSON.(types.VolumeStatus)\n\t\t// we start consume space when moving into CREATING_VOLUME state\n\t\tif iterVolumeStatus.State < types.CREATING_VOLUME {\n\t\t\tlog.Tracef(\"Volume %s State %d < CREATING_VOLUME\",\n\t\t\t\titerVolumeStatus.Key(), iterVolumeStatus.State)\n\t\t\tcontinue\n\t\t}\n\t\ttotalDiskSize += volumehandlers.GetVolumeHandler(log, ctxPtr, &iterVolumeStatus).UsageFromStatus()\n\t}\n\tdeviceDiskUsage, err := diskmetrics.PersistUsageStat(log)\n\tif err != nil {\n\t\terr := fmt.Errorf(\"Failed to get diskUsage for /persist. err: %s\", err)\n\t\tlog.Error(err)\n\t\treturn 0, err\n\t}\n\tdeviceDiskSize := deviceDiskUsage.Total\n\tdiskReservedForDom0 := diskmetrics.Dom0DiskReservedSize(log, ctxPtr.globalConfig, deviceDiskSize)\n\tvar allowedDeviceDiskSize uint64\n\tif deviceDiskSize < diskReservedForDom0 {\n\t\terr = fmt.Errorf(\"Total Disk Size(%d) <= diskReservedForDom0(%d)\",\n\t\t\tdeviceDiskSize, diskReservedForDom0)\n\t\tlog.Errorf(\"***getRemainingDiskSpace: err: %s\", err)\n\t\treturn uint64(0), err\n\t}\n\tallowedDeviceDiskSize = deviceDiskSize - diskReservedForDom0\n\tif allowedDeviceDiskSize < totalDiskSize {\n\t\treturn 0, nil\n\t} else {\n\t\treturn allowedDeviceDiskSize - totalDiskSize, nil\n\t}\n}", "func (s *Server) statproc() {\n\tvar (\n\t\tv *volume.Volume\n\t\tolds *stat.Stats\n\t\tnews = new(stat.Stats)\n\t)\n\tfor {\n\t\tolds = s.info.Stats\n\t\t*news = *olds\n\t\ts.info.Stats = news // use news instead, for current display\n\t\tolds.Reset()\n\t\tfor _, v = range s.store.Volumes {\n\t\t\tv.Stats.Calc()\n\t\t\tolds.Merge(v.Stats)\n\t\t}\n\t\tolds.Calc()\n\t\ts.info.Stats = olds\n\t\ttime.Sleep(statDuration)\n\t}\n}", "func ShowDiskStatus(c *gin.Context) {\n\tdiskinfo, err := diskInfo()\n\tif err != nil {\n\t\tlog.Fatal(\"Get Disk information error: \", err.Error())\n\t}\n\tutils.Render(\n\t\tc,\n\t\tgin.H{\n\t\t\t\"title\": \"System Status\",\n\t\t\t\"Name\": \"System base information\",\n\t\t\t\"payload\": diskinfo,\n\t\t},\n\t\t\"stat.html\",\n\t)\n}", "func TestParseDiskUsageBasic(t *testing.T) {\n\tcontent, err := ioutil.ReadFile(\"../test_resources/stdout\")\n\tif err != nil {\n\t\tt.Fatal(err)\n\t}\n\ts := string(content)\n\n\tdiskUsage, _ := parseDiskUsage(&s)\n\tif diskUsage != 1111.0 {\n\t\tt.Errorf(\"json value returned unexpected: got %f want %f\",\n\t\t\tdiskUsage, 1111.0)\n\t}\n}", "func DiskFree() (uint64, error) {\n\tusage, err := Disk()\n\t// for i := 0; i < len(usage); i++ {\n\tif len(usage) > 0 {\n\t\tfreeDisk := usage[0].Free\n\t\treturn freeDisk, err\n\t}\n\n\treturn 0, err\n}", "func (c *CryptohomeBinary) GetAccountDiskUsage(ctx context.Context, username string) ([]byte, error) {\n\treturn c.call(ctx, \"--action=get_account_disk_usage\", \"--user=\"+username)\n}", "func getHDDString() string {\n\tpartitions, _ := disk.Partitions(false)\n\tbiggestDiskSize := uint64(0)\n\tbiggestDiskUsed := uint64(0)\n\tbiggestDiskName := \"\"\n\tfor _, partition := range partitions {\n\t\td, _ := disk.Usage(partition.Mountpoint)\n\t\tif d.Total > biggestDiskSize {\n\t\t\tbiggestDiskName = partition.Mountpoint\n\t\t\tbiggestDiskUsed = d.Used\n\t\t\tbiggestDiskSize = d.Total\n\t\t}\n\t}\n\treturn fmt.Sprintf(\"%v: %v/%v\", biggestDiskName, formatSize(biggestDiskUsed, 1000), formatSize(biggestDiskSize, 1000))\n}", "func GetDiskStats(disks cmn.SimpleKVs) DiskStats {\n\tif len(disks) < largeNumDisks {\n\t\toutput := make(DiskStats, len(disks))\n\n\t\tfor disk := range disks {\n\t\t\tstat, ok := readSingleDiskStat(disk)\n\t\t\tif !ok {\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\toutput[disk] = stat\n\t\t}\n\t\treturn output\n\t}\n\n\treturn readMultipleDiskStats(disks)\n}", "func (fs *FS) fsInfo(ctx context.Context, path string) (int64, int64, int64, int64, int64, int64, error) {\n\tstatfs := &unix.Statfs_t{}\n\terr := unix.Statfs(path, statfs)\n\tif err != nil {\n\t\treturn 0, 0, 0, 0, 0, 0, err\n\t}\n\n\t// Available is blocks available * fragment size\n\tavailable := int64(statfs.Bavail) * int64(statfs.Bsize)\n\n\t// Capacity is total block count * fragment size\n\tcapacity := int64(statfs.Blocks) * int64(statfs.Bsize)\n\n\t// Usage is block being used * fragment size (aka block size).\n\tusage := (int64(statfs.Blocks) - int64(statfs.Bfree)) * int64(statfs.Bsize)\n\n\tinodes := int64(statfs.Files)\n\tinodesFree := int64(statfs.Ffree)\n\tinodesUsed := inodes - inodesFree\n\n\treturn available, capacity, usage, inodes, inodesFree, inodesUsed, nil\n}", "func (c *Container) statInsideMount(containerPath string) (*copier.StatForItem, string, string, error) {\n\tresolvedRoot := \"/\"\n\tresolvedPath := c.pathAbs(containerPath)\n\tvar statInfo *copier.StatForItem\n\n\terr := c.joinMountAndExec(\n\t\tfunc() error {\n\t\t\tvar statErr error\n\t\t\tstatInfo, statErr = secureStat(resolvedRoot, resolvedPath)\n\t\t\treturn statErr\n\t\t},\n\t)\n\n\treturn statInfo, resolvedRoot, resolvedPath, err\n}", "func RunListDisk() {\n\n\t// dir, err := filepath.Abs(filepath.Dir(os.Args[0]))\n\t// if err != nil {\n\t// \tlog.Fatal(err)\n\t// \treturn\n\t// }\n\n\t// lsscsipath := path.Join(dir, \"lsscsi\")\n\t// if _, err := os.Stat(lsscsipath); os.IsNotExist(err) {\n\t// \tlsscsipath = \"lsscsi\"\n\t// }\n\tlsscsipath := \"lsscsi\"\n\tcmd := exec.Command(lsscsipath, \"-s\", \"-g\")\n\tstdout, err := cmd.StdoutPipe()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\tif err := cmd.Start(); err != nil {\n\t\tlog.Fatal(err)\n\t}\n\n\ttimer := time.AfterFunc(10*time.Second, func() {\n\t\tcmd.Process.Kill()\n\t})\n\n\tscanner := bufio.NewScanner(stdout)\n\tvar hddinfo []string\n\tvar hddchanged bool\n\tvar wg sync.WaitGroup\n\tfor scanner.Scan() {\n\t\tss := scanner.Text()\n\t\tfmt.Println(ss)\n\t\thddinfo = append(hddinfo, ss)\n\t\tif !DetectData.MatchKey(ss) {\n\t\t\thddchanged = true\n\t\t}\n\t\tif !DetectData.ContainsKey(ss) {\n\t\t\t//\\s Matches any white-space character.\n\t\t\tr := regexp.MustCompile(`^([\\s\\S]{13})(disk[\\s\\S]{4})([\\s\\S]{9})([\\s\\S]{17})([\\s\\S]{6})([\\s\\S]{11})([\\s\\S]{11})([\\s\\S]+)$`)\n\t\t\tdiskinfos := r.FindStringSubmatch(ss)\n\t\t\tif len(diskinfos) == 9 {\n\t\t\t\tvar dddect = NewSyncDataDetect()\n\t\t\t\tdddect.detectHDD.Locpath = strings.Trim(diskinfos[1], \" \")\n\t\t\t\tdddect.detectHDD.Type = strings.Trim(diskinfos[2], \" \")\n\t\t\t\tdddect.detectHDD.Manufacture = strings.Trim(diskinfos[3], \" \")\n\t\t\t\tdddect.detectHDD.Model = strings.Trim(diskinfos[4], \" \")\n\t\t\t\tdddect.detectHDD.Version = strings.Trim(diskinfos[5], \" \")\n\t\t\t\tdddect.detectHDD.LinuxName = strings.Trim(diskinfos[6], \" \")\n\t\t\t\tdddect.detectHDD.SGLibName = strings.Trim(diskinfos[7], \" \")\n\t\t\t\tdddect.detectHDD.Size = strings.Trim(diskinfos[8], \" \")\n\n\t\t\t\tif strings.Index(dddect.detectHDD.LinuxName, `/dev/`) == -1 {\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\t\t\t\t//hddchanged = true\n\t\t\t\tDetectData.AddValue(ss, dddect)\n\t\t\t\twg.Add(1)\n\t\t\t\tgo dddect.ReadDataFromSmartCtl(&wg)\n\t\t\t}\n\t\t} else {\n\t\t\tif vv, ok := DetectData.Get(ss); ok {\n\t\t\t\tif len(vv.detectHDD.UILabel) == 0 && len(vv.detectHDD.Otherinfo) == 0 {\n\t\t\t\t\twg.Add(1)\n\t\t\t\t\tgo vv.ReadDataFromSmartCtl(&wg)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\ttimer.Stop()\n\tDetectData.RemoveOld(hddinfo)\n\n\ttime.Sleep(4 * time.Second)\n\n\tif hddchanged {\n\t\tfmt.Print(\"changed!\")\n\t\tcclist, err := configxmldata.Conf.GetCardListIndex()\n\t\tif err == nil {\n\t\t\tfor _, i := range cclist {\n\t\t\t\twg.Add(1)\n\t\t\t\tgo SASHDDinfo.RunCardInfo(i, &wg)\n\t\t\t}\n\t\t}\n\t\tfor i := 0; i < 30; i++ {\n\t\t\tif waitTimeout(&wg, 10*time.Second) {\n\t\t\t\tfmt.Println(\"Timed out waiting for wait group\")\n\t\t\t\tMergeCalibration()\n\t\t\t} else {\n\t\t\t\tfmt.Println(\"Wait group finished\")\n\t\t\t\tMergeCalibration()\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t} else {\n\t\twaitTimeout(&wg, 300*time.Second)\n\t}\n\n}", "func (o *OSS) ReturnSize(groupID int64) error {\n\n partLine := partLine()\n\n totalData := map[string]map[string]int{}\n wg := &sync.WaitGroup{}\n ch := make(chan base.BaseInfo, 1000)\n wg.Add(2)\n go register(groupID, ch, wg)\n go fileCalc(groupID, ch, wg, o, totalData)\n\n time.Sleep(2 * time.Second)\n wg.Wait()\n\n for t := range totalData {\n ts := strconv.Itoa(totalData[t][\"totalSize\"])\n\n write.CreateFile(t, partLine + \"\\n\")\n write.CreateFile(t, fmt.Sprintf(\"Total: RecordCount: %d ; FileCount: %d ; FileSize: %s .\\n\",totalData[t][\"RecordCount\"],totalData[t][\"totalCount\"], utils.FormatSize(ts) ))\n }\n return nil\n}", "func umountProc(syncFD int) {\n\tsyncFile := os.NewFile(uintptr(syncFD), \"procfs umount sync FD\")\n\tbuf := make([]byte, 1)\n\tif w, err := syncFile.Write(buf); err != nil || w != 1 {\n\t\tutil.Fatalf(\"unable to write into the proc umounter descriptor: %v\", err)\n\t}\n\tsyncFile.Close()\n\n\tvar waitStatus unix.WaitStatus\n\tif _, err := unix.Wait4(0, &waitStatus, 0, nil); err != nil {\n\t\tutil.Fatalf(\"error waiting for the proc umounter process: %v\", err)\n\t}\n\tif !waitStatus.Exited() || waitStatus.ExitStatus() != 0 {\n\t\tutil.Fatalf(\"the proc umounter process failed: %v\", waitStatus)\n\t}\n\tif err := unix.Access(\"/proc/self\", unix.F_OK); err != unix.ENOENT {\n\t\tutil.Fatalf(\"/proc is still accessible\")\n\t}\n}", "func Space(inputs SpaceInputs) error {\n\tfilesystemPath := inputs.FilesystemPath\n\tdb[\"filter\"] = inputs.Filter\n\tvar err error\n\tif !verify(filesystemPath) {\n\t\terr = errors.New(\"Path is not abosule path\")\n\t\treturn err\n\t}\n\tstart := time.Now()\n\terr = filepath.Walk(filesystemPath, process)\n\troot := getDiskUsage(\"/\")\n\tif err != nil {\n\t\treturn err\n\t}\n\tt := time.Now()\n\telapsed := t.Sub(start)\n\telapsedInMin := elapsed.Minutes()\n\tlargeFileProcessed, _ := db[largeFilenameKey]\n\tlargeSizeProcessed, _ := db[largeSizeKey]\n\ttotalSizeProcessed, _ := db[totalSizeKey]\n\tlargeSize, _ := strconv.ParseUint(largeSizeProcessed, base, bitSize)\n\ttotalSize, _ := strconv.ParseUint(totalSizeProcessed, base, bitSize)\n\tif root.all > 0 {\n\t\ttotalSize = root.all\n\t}\n\tlargeFileSizePercentage := (float64(largeSize) / float64(totalSize)) * 100.00\n\tdata := [][]string{\n\t\t[]string{\"LARGE_FILE_NAME\", largeFileProcessed, \"NA\"},\n\t\t[]string{\"LARGE_FILE_SIZE\", bytefmt.ByteSize(largeSize), strconv.FormatFloat(largeFileSizePercentage, 'f', floatPrec, bitSize)},\n\t\t[]string{\"DISK_TOTAL_SIZE\", bytefmt.ByteSize(root.all), \"NA\"},\n\t\t[]string{\"DISK_USED_PERCENTAGE\", \"--\", strconv.FormatFloat(root.usedPer, 'f', floatPrec, bitSize)},\n\t\t[]string{\"PROCESSING_TIME\", strconv.FormatFloat(elapsedInMin, 'f', floatPrec, bitSize) + \" min(s)\", \"NA\"},\n\t}\n\tviews.Print(data)\n\treturn err\n}", "func (b *Bucket) Statd(path string) (file *File) {\n\treq, err := http.NewRequest(\"HEAD\", b.URL+path+\" d\", nil)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tresp, err := b.run(req, 0)\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\tresp.Body.Close()\n\tif resp.StatusCode == http.StatusOK {\n\t\tfile = &File{\n\t\t\tPath: path,\n\t\t\tLastModified: resp.Header.Get(\"Last-Modified\"),\n\t\t\tETag: resp.Header.Get(\"ETag\"),\n\t\t\tSize: 0,\n\t\t\tIsDir: true,\n\t\t}\n\t}\n\treturn\n}", "func (filterdev *NetworkTap) Stats() (*syscall.BpfStat, error) {\n\tvar s syscall.BpfStat\n\t_, _, err := syscall.Syscall(syscall.SYS_IOCTL, uintptr(filterdev.device.Fd()), syscall.BIOCGSTATS, uintptr(unsafe.Pointer(&s)))\n\tif err != 0 {\n\t\treturn nil, syscall.Errno(err)\n\t}\n\treturn &s, nil\n}", "func Deserialize(p []byte) *structs.DiskStats {\n\tstts := &structs.DiskStats{}\n\tdevF := &flat.Device{}\n\tstatsFlat := flat.GetRootAsDiskStats(p, 0)\n\tstts.Timestamp = statsFlat.Timestamp()\n\tlen := statsFlat.DeviceLength()\n\tstts.Device = make([]structs.Device, len)\n\tfor i := 0; i < len; i++ {\n\t\tvar dev structs.Device\n\t\tif statsFlat.Device(devF, i) {\n\t\t\tdev.Major = devF.Major()\n\t\t\tdev.Minor = devF.Minor()\n\t\t\tdev.Name = string(devF.Name())\n\t\t\tdev.ReadsCompleted = devF.ReadsCompleted()\n\t\t\tdev.ReadsMerged = devF.ReadsMerged()\n\t\t\tdev.ReadSectors = devF.ReadSectors()\n\t\t\tdev.ReadingTime = devF.ReadingTime()\n\t\t\tdev.WritesCompleted = devF.WritesCompleted()\n\t\t\tdev.WritesMerged = devF.WritesMerged()\n\t\t\tdev.WrittenSectors = devF.WrittenSectors()\n\t\t\tdev.WritingTime = devF.WritingTime()\n\t\t\tdev.IOInProgress = devF.IOInProgress()\n\t\t\tdev.IOTime = devF.IOTime()\n\t\t\tdev.WeightedIOTime = devF.WeightedIOTime()\n\t\t}\n\t\tstts.Device[i] = dev\n\t}\n\treturn stts\n}", "func DiskQuota(path string, size ...string) string {\n\tparent := id(path)\n\texec.Command(\"btrfs\", \"qgroup\", \"create\", \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/opt\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/var\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/home\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\texec.Command(\"btrfs\", \"qgroup\", \"assign\", \"0/\"+id(path+\"/rootfs\"), \"1/\"+parent, config.Agent.LxcPrefix+path).Run()\n\n\tif len(size) > 0 && len(size[0]) > 0 {\n\t\tout, err := exec.Command(\"btrfs\", \"qgroup\", \"limit\", size[0]+\"G\", \"1/\"+parent, config.Agent.LxcPrefix+path).CombinedOutput()\n\t\tlog.Check(log.ErrorLevel, \"Limiting BTRFS group 1/\"+parent+\" \"+string(out), err)\n\t\texec.Command(\"btrfs\", \"quota\", \"rescan\", \"-w\", config.Agent.LxcPrefix).Run()\n\t}\n\treturn Stat(path, \"quota\", false)\n}", "func dirSize(path string) (float64, error) {\n\tvar size int64\n\terr := filepath.Walk(path, func(_ string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\tsize += 0\n\t\t} else {\n\t\t\tif !info.IsDir() {\n\t\t\t\tsize += info.Size()\n\t\t\t}\n\t\t}\n\t\treturn nil\n\t})\n\tr, _ := decimal.NewFromFloat(float64(size) / (1024 * 1024 * 1024)).Round(2).Float64()\n\treturn r, err\n}", "func (o ClusterNodeGroupSystemDiskOutput) Count() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v ClusterNodeGroupSystemDisk) *int { return v.Count }).(pulumi.IntPtrOutput)\n}", "func (c *cpusetHandler) Stat(ctr *CgroupControl, m *Metrics) error {\n\treturn nil\n}", "func (a *app) gatherStat() {\n\tabout, err := a.srv.About.Get().Fields(\"storageQuota\").Do()\n\tif err != nil {\n\t\tlog.Fatalf(\"Unable to execute an about request: %v\", err)\n\t}\n\n\ta.sq = about.StorageQuota\n}", "func (s Sysinfo) MemUsedPercent() uint64 {\n\treturn s.MemUsed() * 100 / s.MemTotal()\n}", "func (c *Client) Stat() (count, size int, err error) {\n\tline, err := c.Cmd(\"%s\\r\\n\", STAT)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Number of messages in maildrop\n\tcount, err = strconv.Atoi(strings.Fields(line)[1])\n\tif err != nil {\n\t\treturn\n\t}\n\tif count == 0 {\n\t\treturn\n\t}\n\n\t// Total size of messages in bytes\n\tsize, err = strconv.Atoi(strings.Fields(line)[2])\n\tif err != nil {\n\t\treturn\n\t}\n\tif size == 0 {\n\t\treturn\n\t}\n\treturn\n}", "func diskInfo(y *ec2.Instance) diskInfoType {\n\n\tdeviceMap := map[string]int64{}\n\n\trootdevice := *y.RootDeviceName\n\tfor _, f := range y.BlockDeviceMappings {\n\t\tdeviceMap[aws.StringValue(f.DeviceName)] = f.Ebs.AttachTime.Unix()\n\t}\n\tdd := time.Unix(int64(deviceMap[rootdevice]), 0)\n\tsplitdate := strings.Split(dd.Format(time.RFC850), \" \")[1]\n\tdiskAttached := len(deviceMap)\n\tdItype := diskInfoType{date: splitdate, diskCount: diskAttached}\n\treturn dItype\n\n}", "func (c *cpuMeasure) update() error {\n\tfile, err := os.Open(stat)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer file.Close()\n\n\tvar lineName string\n\tvar n, cpuCount int\n\n\tscanner := bufio.NewScanner(file)\n\tfor scanner.Scan() {\n\t\tline := scanner.Text()\n\n\t\tif strings.Contains(line, \"cpu\") {\n\t\t\tn, err = fmt.Sscanf(line,\n\t\t\t\t\"%s %d %d %d %d %d %d %d %d %d %d\", &lineName,\n\t\t\t\t&c.cores[cpuCount][0], &c.cores[cpuCount][1],\n\t\t\t\t&c.cores[cpuCount][2], &c.cores[cpuCount][3],\n\t\t\t\t&c.cores[cpuCount][4], &c.cores[cpuCount][5],\n\t\t\t\t&c.cores[cpuCount][6], &c.cores[cpuCount][7],\n\t\t\t\t&c.cores[cpuCount][8], &c.cores[cpuCount][9],\n\t\t\t)\n\t\t\tcheckSscanf(lineName, err, n, 11)\n\t\t\tcpuCount++\n\t\t} else if strings.Contains(line, \"ctxt\") {\n\t\t\tn, err = fmt.Sscanf(line, \"ctxt %d\", &c.SwitchCtxt)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"btime\") {\n\t\t\tn, err = fmt.Sscanf(line, \"btime %d\", &c.BootTime)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"processes\") {\n\t\t\tn, err = fmt.Sscanf(line, \"processes %d\", &c.Processes)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"procs_running\") {\n\t\t\tn, err = fmt.Sscanf(line, \"procs_running %d\", &c.ProcsRunning)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t} else if strings.Contains(line, \"procs_blocked\") {\n\t\t\tn, err = fmt.Sscanf(line, \"procs_blocked %d\", &c.ProcsBlocked)\n\t\t\tcheckSscanf(\"ctxt\", err, n, 1)\n\t\t}\n\t}\n\n\treturn nil\n}", "func VolumeScanner(pathToScan string) error {\n\tresult := make(map[string]DirStats)\n\tvar largestDir string\n\tlargestSize := int64(0)\n\ttotalSize := int64(0)\n\tvar cwd string\n\terr := filepath.Walk(pathToScan, func(path string, info os.FileInfo, err error) error {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tif info.IsDir() {\n\t\t\tcwd = path\n\t\t\tresult[cwd] = DirStats{\n\t\t\t\tPath: cwd,\n\t\t\t\tTotalSize: int64(0),\n\t\t\t}\n\t\t\treturn err\n\t\t}\n\t\tdir := filepath.Dir(path)\n\t\tdirStat, _ := result[dir]\n\t\tdirStat.TotalSize += info.Size()\n\t\ttotalSize += info.Size()\n\t\tif dirStat.TotalSize > largestSize {\n\t\t\tlargestSize = dirStat.TotalSize\n\t\t\tlargestDir = dir\n\t\t}\n\t\treturn err\n\n\t})\n\tdata := [][]string{\n\t\t[]string{\"TOTAL_SIZE\", bytefmt.ByteSize(uint64(totalSize)), \"\"},\n\t\t[]string{\"LARGEST_DIR\", largestDir, \"\"},\n\t\t[]string{\"LARGEST_DIR_SIZE\", bytefmt.ByteSize(uint64(largestSize)), \"\"},\n\t}\n\n\tviews.Print(data)\n\treturn err\n\n}", "func (u *User) GetDiskUsage() int {\n\tif u == nil || u.DiskUsage == nil {\n\t\treturn 0\n\t}\n\treturn *u.DiskUsage\n}", "func report(p *rc.Process, wallTime time.Duration) {\n\tsv, err := p.SystemVersion()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\n\tss, err := p.SystemStatus()\n\tif err != nil {\n\t\tlog.Println(err)\n\t\treturn\n\t}\n\n\tproc, err := p.Stop()\n\tif err != nil {\n\t\treturn\n\t}\n\n\trusage, ok := proc.SysUsage().(*syscall.Rusage)\n\tif !ok {\n\t\treturn\n\t}\n\n\tlog.Println(\"Version:\", sv.Version)\n\tlog.Println(\"Alloc:\", ss.Alloc/1024, \"KiB\")\n\tlog.Println(\"Sys:\", ss.Sys/1024, \"KiB\")\n\tlog.Println(\"Goroutines:\", ss.Goroutines)\n\tlog.Println(\"Wall time:\", wallTime)\n\tlog.Println(\"Utime:\", time.Duration(rusage.Utime.Nano()))\n\tlog.Println(\"Stime:\", time.Duration(rusage.Stime.Nano()))\n\tif runtime.GOOS == \"darwin\" {\n\t\t// Darwin reports in bytes, Linux seems to report in KiB even\n\t\t// though the manpage says otherwise.\n\t\trusage.Maxrss /= 1024\n\t}\n\tlog.Println(\"MaxRSS:\", rusage.Maxrss, \"KiB\")\n\n\tfmt.Printf(\"%s,%d,%d,%d,%.02f,%.02f,%.02f,%d\\n\",\n\t\tsv.Version,\n\t\tss.Alloc/1024,\n\t\tss.Sys/1024,\n\t\tss.Goroutines,\n\t\twallTime.Seconds(),\n\t\ttime.Duration(rusage.Utime.Nano()).Seconds(),\n\t\ttime.Duration(rusage.Stime.Nano()).Seconds(),\n\t\trusage.Maxrss)\n}", "func Stat(t *kernel.Task, args arch.SyscallArguments) (uintptr, *kernel.SyscallControl, error) {\n\taddr := args[0].Pointer()\n\tstatAddr := args[1].Pointer()\n\n\tpath, dirPath, err := copyInPath(t, addr, false /* allowEmpty */)\n\tif err != nil {\n\t\treturn 0, nil, err\n\t}\n\n\treturn 0, nil, fileOpOn(t, linux.AT_FDCWD, path, true /* resolve */, func(root *fs.Dirent, d *fs.Dirent, _ uint) error {\n\t\treturn stat(t, d, dirPath, statAddr)\n\t})\n}", "func (rso *RadosStripedObject) Stat() (size uint64, pmtime uint64, err error) {\n\tobj := C.CString(rso.ObjectName)\n\tdefer C.free(unsafe.Pointer(obj))\n\n\tvar c_size C.uint64_t\n\tvar c_time_t C.time_t\n\tret := C.rados_striper_stat(rso.Striper, obj, &c_size, &c_time_t)\n\tif ret < 0 {\n\t\treturn 0, 0, errors.New(\"get stat failed\")\n\t}\n\treturn uint64(c_size), uint64(C.uint64_t(c_time_t)), nil\n}", "func (s *Simple) DiskInfo(req *acomm.Request) (interface{}, *url.URL, error) {\n\tvar args CPUInfoArgs\n\tif err := req.UnmarshalArgs(&args); err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif args.GuestID == \"\" {\n\t\treturn nil, nil, errors.New(\"missing guest_id\")\n\t}\n\n\tresult := &DiskInfoResult{\n\t\t&DiskInfo{\n\t\t\tDevice: \"vda1\",\n\t\t\tSize: 10 * (1024 * 1024 * 1024), // 10 GB in bytes\n\t\t},\n\t}\n\n\treturn result, nil, nil\n}", "func MBUsed() float64 {\n var m runtime.MemStats\n runtime.ReadMemStats(&m)\n return float64(m.TotalAlloc) / BytesPerMBF \n}", "func (s *DataNode) incDiskErrCnt(partitionID uint64, err error, flag uint8) {\n\tif err == nil {\n\t\treturn\n\t}\n\tdp := s.space.Partition(partitionID)\n\tif dp == nil {\n\t\treturn\n\t}\n\td := dp.Disk()\n\tif d == nil {\n\t\treturn\n\t}\n\tif !IsDiskErr(err.Error()) {\n\t\treturn\n\t}\n\tif flag == WriteFlag {\n\t\td.incWriteErrCnt()\n\t} else if flag == ReadFlag {\n\t\td.incReadErrCnt()\n\t}\n}", "func ProcStat(c *gin.Context) {\n\tres := CmdExec(\"cat /proc/stat | head -n 1 | awk '{$1=\\\"\\\";print}'\")\n\tresArray := strings.Split(res[0], \" \")\n\tvar cpu []int64\n\tvar totalcpu, idlecpu int64\n\tfor _, v := range resArray {\n\t\ttemp, err := strconv.ParseInt(v, 10, 64)\n\t\tif err == nil {\n\t\t\tcpu = append(cpu, temp)\n\t\t\ttotalcpu = totalcpu + temp\n\t\t}\n\t}\n\tidlecpu = cpu[3]\n\tc.JSON(http.StatusOK, gin.H{\n\t\t\"totalcpu\": totalcpu,\n\t\t\"idlecpu\": idlecpu,\n\t})\n}", "func dirents(dir string) []os.FileInfo {\n\tentries, err := ioutil.ReadDir(dir)\n\tif err != nil {\n\t\tfmt.Fprintf(os.Stderr, \"du1: %v\\n\", err)\n\t\treturn nil\n\t}\n\treturn entries\n}", "func (s *Store) scrubDisk(diskno int, d Disk) {\n\t// We use 0 as capacity for token bucket limiter, which means we'll wait\n\t// for \"s.config.ScrubRate\" tokens to be refilled every time we call \"Take\",\n\t// which is fine for our usage here.\n\ttb := tokenbucket.New(0, 0)\n\n\tfor {\n\t\t// In case of errors, empty disk, etc., don't spin. PL-1113\n\t\ttime.Sleep(5 * time.Minute)\n\n\t\trate := s.Config().ScrubRate\n\t\tif rate < 1024 {\n\t\t\tcontinue\n\t\t}\n\t\ttb.SetRate(float32(rate), 0)\n\n\t\tdir, err := d.OpenDir()\n\t\tif err == core.ErrDiskRemoved {\n\t\t\treturn\n\t\t}\n\t\tif err != core.NoError {\n\t\t\tlog.Errorf(\"aborting disk scrub for disk %d, failed to open dir, err=%s\", diskno, err)\n\t\t\tcontinue\n\t\t}\n\n\t\tlog.Infof(\"scrub of disk %d starting\", diskno)\n\n\t\tvar scrubbed, ttracts, ok, bad int\n\t\tvar bytes int64\n\t\tstart := time.Now()\n\n\t\tfor {\n\t\t\ttracts, terr := d.ReadDir(dir)\n\t\t\tif terr != core.NoError {\n\t\t\t\tbreak\n\t\t\t}\n\t\t\tttracts += len(tracts)\n\t\t\tfor _, tract := range tracts {\n\t\t\t\t// If we can't lock it someone else is probably scrubbing it by virtue of reading or writing it.\n\t\t\t\tif !s.tryLockTract(tract, READ) {\n\t\t\t\t\tlog.V(5).Infof(\"tract %s is busy, won't scrub this iteration\", tract)\n\t\t\t\t\tcontinue\n\t\t\t\t}\n\n\t\t\t\t// Scrub returns how many bytes it read. We use this to throttle scrubbing to s.config.ScrubRate bytes/sec.\n\t\t\t\tn, err := d.Scrub(tract)\n\t\t\t\ts.unlock(tract, READ)\n\n\t\t\t\tif err == core.ErrDiskRemoved {\n\t\t\t\t\treturn\n\t\t\t\t}\n\n\t\t\t\t// This might sleep so we want to unlock the tract before calling it.\n\t\t\t\ttb.Take(float32(n))\n\n\t\t\t\t// Collect and log some stats.\n\t\t\t\tif s.maybeReportError(tract, err) {\n\t\t\t\t\tbad++\n\t\t\t\t} else {\n\t\t\t\t\tok++\n\t\t\t\t}\n\t\t\t\tscrubbed++\n\t\t\t\tbytes += n\n\t\t\t\tif scrubbed%10 == 0 {\n\t\t\t\t\tlogStats(diskno, start, scrubbed, ok, bad, ttracts, bytes, 2)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t\td.CloseDir(dir)\n\t\tlogStats(diskno, start, scrubbed, ok, bad, ttracts, bytes, 0)\n\t}\n}", "func GetFSInfo(path string) (total, available int) {\n\ttotal = -1\n\tavailable = -1\n\tvar buf syscall.Statfs_t\n\n\tif syscall.Statfs(path, &buf) != nil {\n\t\treturn\n\t}\n\n\tif buf.Bsize <= 0 {\n\t\treturn\n\t}\n\n\ttotal = int((uint64(buf.Bsize) * buf.Blocks) / (1000 * 1000))\n\tavailable = int((uint64(buf.Bsize) * buf.Bavail) / (1000 * 1000))\n\n\treturn\n}", "func GetStat() (uint64, uint64) {\n\treturn filesAmount, dirsAmount\n}", "func (d *dataUsageCache) dui(path string, buckets []BucketInfo) madmin.DataUsageInfo {\n\te := d.find(path)\n\tif e == nil {\n\t\t// No entry found, return empty.\n\t\treturn madmin.DataUsageInfo{}\n\t}\n\tflat := d.flatten(*e)\n\tdui := madmin.DataUsageInfo{\n\t\tLastUpdate: d.Info.LastUpdate,\n\t\tObjectsTotalCount: flat.Objects,\n\t\tObjectsTotalSize: uint64(flat.Size),\n\t\tBucketsCount: uint64(len(e.Children)),\n\t\tBucketsUsage: d.bucketsUsageInfo(buckets),\n\t}\n\tif flat.ReplicationStats != nil {\n\t\tdui.ReplicationPendingSize = flat.ReplicationStats.PendingSize\n\t\tdui.ReplicatedSize = flat.ReplicationStats.ReplicatedSize\n\t\tdui.ReplicationFailedSize = flat.ReplicationStats.FailedSize\n\t\tdui.ReplicationPendingCount = flat.ReplicationStats.PendingCount\n\t\tdui.ReplicationFailedCount = flat.ReplicationStats.FailedCount\n\t\tdui.ReplicaSize = flat.ReplicationStats.ReplicaSize\n\t}\n\treturn dui\n}", "func (s *CPUStat) Usage() float64 {\n\treturn s.All.Usage() * float64(len(s.cpus))\n}", "func (p *xlStorageDiskIDCheck) monitorDiskStatus() {\n\tt := time.NewTicker(5 * time.Second)\n\tdefer t.Stop()\n\tfn := mustGetUUID()\n\tfor range t.C {\n\t\tif len(p.health.tokens) == 0 {\n\t\t\t// Queue is still full, no need to check.\n\t\t\tcontinue\n\t\t}\n\t\terr := p.storage.WriteAll(context.Background(), minioMetaTmpBucket, fn, []byte{10000: 42})\n\t\tif err != nil {\n\t\t\tcontinue\n\t\t}\n\t\tb, err := p.storage.ReadAll(context.Background(), minioMetaTmpBucket, fn)\n\t\tif err != nil || len(b) != 10001 {\n\t\t\tcontinue\n\t\t}\n\t\terr = p.storage.Delete(context.Background(), minioMetaTmpBucket, fn, DeleteOptions{\n\t\t\tRecursive: false,\n\t\t\tForce: false,\n\t\t})\n\t\tif err == nil {\n\t\t\tlogger.Info(\"node(%s): Read/Write/Delete successful, bringing drive %s online. Drive was offline for %s.\", globalLocalNodeName, p.storage.String(),\n\t\t\t\ttime.Since(time.Unix(0, atomic.LoadInt64(&p.health.lastSuccess))))\n\t\t\tatomic.StoreInt32(&p.health.status, diskHealthOK)\n\t\t\treturn\n\t\t}\n\t}\n}", "func (s *CRDBStorage) Stat(key string) (certmagic.KeyInfo, error) {\n\tctx, cancel := context.WithTimeout(context.Background(), defaultQueryTimeout)\n\tdefer cancel()\n\trow := s.DB.QueryRowContext(ctx, `select length(value), updated from certmagic_values where \"key\" = $1`, key)\n\tinfo := certmagic.KeyInfo{\n\t\tKey: key,\n\t\tIsTerminal: true,\n\t}\n\tif err := row.Scan(&info.Size, &info.Modified); err != nil {\n\t\tif errors.Is(err, sql.ErrNoRows) {\n\t\t\treturn certmagic.KeyInfo{}, certmagic.ErrNotExist(err)\n\t\t}\n\t\treturn certmagic.KeyInfo{}, err\n\t}\n\treturn info, nil\n}", "func dirents(dir string) []os.FileInfo {\r\n\tsema <- struct{}{} // acquire token\r\n\tdefer func() { <-sema }() // release token\r\n\r\n\tentries, err := ioutil.ReadDir(dir)\r\n\tif err != nil {\r\n\t\tfmt.Fprintf(os.Stderr, \"du: %v\\n\", err)\r\n\t\treturn nil\r\n\t}\r\n\treturn entries\r\n}", "func (o LocalDiskOutput) DiskCount() pulumi.IntPtrOutput {\n\treturn o.ApplyT(func(v LocalDisk) *int { return v.DiskCount }).(pulumi.IntPtrOutput)\n}", "func (e *BackupEnv) GetDirSize(source string) (int64) {\n directory, _ := os.Open(source);\n var sum int64 = 0;\n defer directory.Close();\n\n objects, _ := directory.Readdir(-1)\n for _, obj := range objects {\n if obj.IsDir() {\n sum += e.GetDirSize(source + \"/\" + obj.Name());\n } else {\n stat, _ := os.Stat(source + \"/\" + obj.Name());\n sum += stat.Size();\n }\n }\n\n return sum;\n}", "func RamUsage() {\r\n v, _ := mem.VirtualMemory()\r\n fmt.Printf(\"RAM{ Total: %v, Free:%v, UsedPercent:%f%%}\\n\", v.Total, v.Free, v.UsedPercent)\r\n}", "func measureStorageDevice(blkDevicePath string) error {\n\tlog.Printf(\"Storage Collector: Measuring block device %s\\n\", blkDevicePath)\n\tfile, err := os.Open(blkDevicePath)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"couldn't open disk=%s err=%v\", blkDevicePath, err)\n\t}\n\n\teventDesc := fmt.Sprintf(\"Storage Collector: Measured %s\", blkDevicePath)\n\treturn tpm.ExtendPCRDebug(pcr, file, eventDesc)\n}", "func printSysstat(v *gocui.View, s stat.Stat) error {\n\tvar err error\n\n\t/* line1: current time and load average */\n\t_, err = fmt.Fprintf(v, \"pgcenter: %s, load average: %.2f, %.2f, %.2f\\n\",\n\t\ttime.Now().Format(\"2006-01-02 15:04:05\"),\n\t\ts.LoadAvg.One, s.LoadAvg.Five, s.LoadAvg.Fifteen)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t/* line2: cpu usage */\n\t_, err = fmt.Fprintf(v, \" %%cpu: \\033[37;1m%4.1f\\033[0m us, \\033[37;1m%4.1f\\033[0m sy, \\033[37;1m%4.1f\\033[0m ni, \\033[37;1m%4.1f\\033[0m id, \\033[37;1m%4.1f\\033[0m wa, \\033[37;1m%4.1f\\033[0m hi, \\033[37;1m%4.1f\\033[0m si, \\033[37;1m%4.1f\\033[0m st\\n\",\n\t\ts.CpuStat.User, s.CpuStat.Sys, s.CpuStat.Nice, s.CpuStat.Idle,\n\t\ts.CpuStat.Iowait, s.CpuStat.Irq, s.CpuStat.Softirq, s.CpuStat.Steal)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t/* line3: memory usage */\n\t_, err = fmt.Fprintf(v, \" MiB mem: \\033[37;1m%6d\\033[0m total, \\033[37;1m%6d\\033[0m free, \\033[37;1m%6d\\033[0m used, \\033[37;1m%8d\\033[0m buff/cached\\n\",\n\t\ts.Meminfo.MemTotal, s.Meminfo.MemFree, s.Meminfo.MemUsed,\n\t\ts.Meminfo.MemCached+s.Meminfo.MemBuffers+s.Meminfo.MemSlab)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t/* line4: swap usage, dirty and writeback */\n\t_, err = fmt.Fprintf(v, \"MiB swap: \\033[37;1m%6d\\033[0m total, \\033[37;1m%6d\\033[0m free, \\033[37;1m%6d\\033[0m used, \\033[37;1m%6d/%d\\033[0m dirty/writeback\\n\",\n\t\ts.Meminfo.SwapTotal, s.Meminfo.SwapFree, s.Meminfo.SwapUsed,\n\t\ts.Meminfo.MemDirty, s.Meminfo.MemWriteback)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *NFSCollector) Describe(ch chan<- *prometheus.Desc) {\n\tch <- c.Up\n\tch <- c.NFSInfo\n\tch <- c.DiskFree\n\tch <- c.NICInfo\n\tch <- c.NICReceive\n\tch <- c.NICSend\n\n\tch <- c.MaintenanceScheduled\n\tch <- c.MaintenanceInfo\n\tch <- c.MaintenanceStartTime\n\tch <- c.MaintenanceEndTime\n}", "func (f *Fs) About(ctx context.Context) (*fs.Usage, error) {\n\topts := rest.Opts{\n\t\tMethod: \"PROPFIND\",\n\t\tPath: \"\",\n\t\tExtraHeaders: map[string]string{\n\t\t\t\"Depth\": \"0\",\n\t\t},\n\t}\n\topts.Body = bytes.NewBuffer([]byte(`<?xml version=\"1.0\" ?>\n<D:propfind xmlns:D=\"DAV:\">\n <D:prop>\n <D:quota-available-bytes/>\n <D:quota-used-bytes/>\n </D:prop>\n</D:propfind>\n`))\n\tvar q api.Quota\n\tvar resp *http.Response\n\tvar err error\n\terr = f.pacer.Call(func() (bool, error) {\n\t\tresp, err = f.srv.CallXML(ctx, &opts, nil, &q)\n\t\treturn f.shouldRetry(ctx, resp, err)\n\t})\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tusage := &fs.Usage{}\n\tif i, err := strconv.ParseInt(q.Used, 10, 64); err == nil && i >= 0 {\n\t\tusage.Used = fs.NewUsageValue(i)\n\t}\n\tif i, err := strconv.ParseInt(q.Available, 10, 64); err == nil && i >= 0 {\n\t\tusage.Free = fs.NewUsageValue(i)\n\t}\n\tif usage.Used != nil && usage.Free != nil {\n\t\tusage.Total = fs.NewUsageValue(*usage.Used + *usage.Free)\n\t}\n\treturn usage, nil\n}", "func updateMetrics(interval time.Duration, diskRoot string) {\n\tlogger := logrus.WithField(\"sync-loop\", \"updateMetrics\")\n\tticker := time.NewTicker(interval)\n\tfor ; true; <-ticker.C {\n\t\tlogger.Info(\"tick\")\n\t\t_, bytesFree, bytesUsed, _, _, _, err := diskutil.GetDiskUsage(diskRoot)\n\t\tif err != nil {\n\t\t\tlogger.WithError(err).Error(\"Failed to get disk metrics\")\n\t\t} else {\n\t\t\tpromMetrics.DiskFree.Set(float64(bytesFree) / 1e9)\n\t\t\tpromMetrics.DiskUsed.Set(float64(bytesUsed) / 1e9)\n\t\t\tpromMetrics.DiskTotal.Set(float64(bytesFree+bytesUsed) / 1e9)\n\t\t}\n\t}\n}" ]
[ "0.7427519", "0.6913226", "0.6908944", "0.6908944", "0.6908944", "0.6850082", "0.6845622", "0.68086916", "0.67880803", "0.6696417", "0.66945654", "0.66048414", "0.6564812", "0.64671034", "0.6415186", "0.63966525", "0.6337184", "0.6207931", "0.6153916", "0.6146029", "0.6145258", "0.6116076", "0.6114989", "0.60801154", "0.6067873", "0.6029154", "0.6003057", "0.5980582", "0.5962313", "0.5911625", "0.5880454", "0.58735424", "0.58672833", "0.58442676", "0.5832619", "0.57554525", "0.5737849", "0.57301164", "0.5717931", "0.571664", "0.571028", "0.56971127", "0.5675387", "0.56729925", "0.5661262", "0.5639872", "0.5628183", "0.56103593", "0.55855405", "0.5575156", "0.5534724", "0.5530005", "0.5512552", "0.54882336", "0.54822147", "0.5462636", "0.5426426", "0.5410591", "0.54026526", "0.5390695", "0.53873783", "0.53803426", "0.53712326", "0.53668964", "0.5360813", "0.53590465", "0.53558", "0.5354744", "0.5354114", "0.535246", "0.5343847", "0.53411627", "0.5308106", "0.52986884", "0.52955246", "0.529139", "0.5272785", "0.52600473", "0.5257732", "0.52522355", "0.5250704", "0.5249574", "0.524948", "0.52467304", "0.52450097", "0.5244901", "0.5224547", "0.52129126", "0.52082425", "0.52064675", "0.52015716", "0.51909375", "0.5186846", "0.5185947", "0.5174491", "0.517434", "0.51715106", "0.51713604", "0.51673776", "0.51664156" ]
0.56129587
47
Max returns the larger of x or y.
func Max(x, y int64) int64 { if x > y { return x } return y }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func Max(x, y int64) int64 {\n if x > y {\n return x\n }\n return y\n}", "func Max(x, y float64) float64 {\n\n\treturn math.Max(x, y)\n}", "func max(x, y int64) int64 {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}", "func Max[T comdef.XintOrFloat](x, y T) T {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Max(x int, y int) int {\n\tif x > y { return x }\n\treturn y\n}", "func Max(x, y int) int {\n if x < y {\n return y\n }\n return x\n}", "func Max(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}", "func Max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func Max(x, y int) int {\n\tif y > x {\n\t\treturn y\n\t}\n\treturn x\n}", "func max(x int, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y float64) float64 {\n\tswitch {\n\tcase math.IsNaN(x) || math.IsNaN(y):\n\t\treturn math.NaN()\n\tcase math.IsInf(x, 1) || math.IsInf(y, 1):\n\t\treturn math.Inf(1)\n\n\tcase x == 0 && x == y:\n\t\tif math.Signbit(x) {\n\t\t\treturn y\n\t\t}\n\t\treturn x\n\t}\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}", "func Max[T constraints.Ordered](x T, y T) T {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func max(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t} else {\n\t\treturn y\n\t}\n}", "func Max(x, y int64) int64 {\n\treturn x ^ ((x ^ y) & ((x - y) >> 63))\n}", "func (a Vec2) Max(b Vec2) Vec2 {\n\tvar r Vec2\n\tif a.X > b.X {\n\t\tr.X = a.X\n\t} else {\n\t\tr.X = b.X\n\t}\n\tif a.Y > b.Y {\n\t\tr.Y = a.Y\n\t} else {\n\t\tr.Y = b.Y\n\t}\n\treturn r\n}", "func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func maxi(x int, y int) int {\n if x >= y {\n return x\n } else {\n return y\n }\n}", "func Max(x, y *big.Int) *big.Int {\n\tif x.Cmp(y) == 1 {\n\t\treturn x\n\t}\n\treturn y\n}", "func Max(arg, arg2 float64) float64 {\n\treturn math.Max(arg, arg2)\n}", "func Max(a uint64, b uint64) uint64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b float64) float64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(x, y int32) int32 {\n\treturn x - (((x - y) >> 31) & (x - y))\n}", "func Max(x, y string) string {\n\tif Compare(x, y) < 0 {\n\t\treturn y\n\t}\n\treturn x\n}", "func MaxFloat64(x, y float64) float64 {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}", "func max(a, b int64) int64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int64) int64 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a int, b int) int {\n\tif a >= b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\r\n\tif a > b {\r\n\t\treturn a\r\n\t}\r\n\treturn b\r\n}", "func Max(a int, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a interface{}, b interface{}) interface{} {\n\tif a == nil {\n\t\treturn b\n\t}\n\tif b == nil {\n\t\treturn a\n\t}\n\tif Cmp(a, b) > 0 {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func (m *Arg) GetMax() (val float64, set bool) {\n\tif m.Max == nil {\n\t\treturn\n\t}\n\n\treturn *m.Max, true\n}", "func maxPoints(max, other point) point {\n\tif other.x > max.x {\n\t\tmax.x = other.x\n\t}\n\tif other.y > max.y {\n\t\tmax.y = other.y\n\t}\n\treturn max\n}", "func Max(a int, b int) int {\n if (b > a) {\n return b;\n }\n\n return a;\n}", "func Max(a, b int) int {\n\treturn neogointernal.Opcode2(\"MAX\", a, b).(int)\n}", "func Max(valueA gcv.Value, valueB gcv.Value) (gcv.Value, error) {\n\tif valueA.Type() == gcv.Complex || valueB.Type() == gcv.Complex {\n\t\treturn nil, errors.New(\"Max is not supported for Complex numbers\")\n\t}\n\treturn gcv.MakeValue(math.Max(valueA.Real(), valueB.Real())), nil\n}", "func Max(i, j int) int {\n\tif i > j {\n\t\treturn i\n\t}\n\treturn j\n}", "func max(a int, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(a, b int) int {\n\treturn int(math.Max(float64(a), float64(b)))\n}", "func Max(a, b int) int {\n\tif a-b > 0 {\n\t\treturn a\n\t}\n\n\treturn b\n}", "func max(a, b int) int {\nif a < b {\nreturn b\n}\nreturn a\n}", "func Max(val, max any) bool { return valueCompare(val, max, \"<=\") }", "func Max(a, b uint32) uint32 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func Max(numbers ...cty.Value) (cty.Value, error) {\n\treturn MaxFunc.Call(numbers)\n}", "func (self *State)Max(a,b any)any{\n self.IncOperations(self.coeff[\"max\"]+self.off[\"max\"])\n return wrap2(a,b,math.Max)\n}", "func Max(a, operand int) int {\n\tif a > operand {\n\t\treturn a\n\t}\n\treturn operand\n}", "func max(first int, second int) int {\n\tif first >= second {\n\t\treturn first\n\t} else {\n\t\treturn second\n\t}\n}", "func max(a, b int) int {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func Max(vals ...float64) float64 {\n\tmax := vals[0]\n\tfor _, v := range vals {\n\t\tif v > max {\n\t\t\tmax = v\n\t\t}\n\t}\n\n\treturn max\n}", "func IntMax(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}", "func (c *Capsule) Max() Coord {\n\treturn c.P1.Max(c.P2).AddScalar(c.Radius)\n}", "func max(a, b int) int {\n\tif b > a {\n\t\treturn b\n\t}\n\treturn a\n}", "func Max(xs ...float64) float64 {\n\tm := xs[0]\n\tfor i := 1; i < len(xs); i++ {\n\t\tif m < xs[i] {\n\t\t\tm = xs[i]\n\t\t}\n\t}\n\n\treturn m\n}", "func max(a, b int32) int32 {\n\tif a >= b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b ImpactAmount) ImpactAmount {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func max(a, b int32) int32 {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func MaxFloat(x, y float64) float64 {\n\treturn math.Max(x, y)\n}", "func max(a, b int32) int32 {\n\tif a < b {\n\t\treturn b\n\t}\n\treturn a\n}", "func MaxI(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func SwapMax[T comdef.XintOrFloat](x, y T) (T, T) {\n\tif x > y {\n\t\treturn x, y\n\t}\n\treturn y, x\n}", "func (g *Graph) Max(x1 Node, x2 Node) Node {\n\treturn g.NewOperator(fn.NewMax(x1, x2), x1, x2)\n}", "func MinMax(x, min, max int) int { return x }", "func MaxInt64(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}", "func MaxInt(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\n\treturn y\n}", "func max(num1, num2 int) int {\nresult int\n\n\tif (num1 > num2){\n\t\tresult = num1\n\t} else{\n\t\tresult = num2\n\t}\nreturn result\n}", "func MaxI64(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func MaxI64(x, y int64) int64 {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func (o *Range) GetMax() gdnative.Real {\n\t//log.Println(\"Calling Range.GetMax()\")\n\n\t// Build out the method's arguments\n\tptrArguments := make([]gdnative.Pointer, 0, 0)\n\n\t// Get the method bind\n\tmethodBind := gdnative.NewMethodBind(\"Range\", \"get_max\")\n\n\t// Call the parent method.\n\t// float\n\tretPtr := gdnative.NewEmptyReal()\n\tgdnative.MethodBindPtrCall(methodBind, o.GetBaseObject(), ptrArguments, retPtr)\n\n\t// If we have a return type, convert it from a pointer into its actual object.\n\tret := gdnative.NewRealFromPointer(retPtr)\n\treturn ret\n}", "func (px *Paxos) Max() int {\n\t// Your code here.\n\treturn px.max\n}", "func MaxInt(x, y int) int {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func MAX(a, b int) int {\n\tif a > b {\n\t\treturn a\n\t}\n\treturn b\n}", "func MaxInt(x, y int) int {\n\tif x < y {\n\t\treturn y\n\t}\n\treturn x\n}", "func MaxLamport(x, y Lamport) Lamport {\n\tif x > y {\n\t\treturn x\n\t}\n\treturn y\n}", "func (v *Vector) Max(m *Vector) {\n\tif m.X > v.X {\n\t\tv.X = m.X\n\t}\n\tif m.Y > v.Y {\n\t\tv.Y = m.Y\n\t}\n\tif m.Z > v.Z {\n\t\tv.Z = m.Z\n\t}\n}" ]
[ "0.8431474", "0.8313492", "0.82121664", "0.8178514", "0.81000376", "0.807142", "0.80567896", "0.80367714", "0.80367714", "0.8026792", "0.79476416", "0.79179174", "0.7908209", "0.7897175", "0.78907746", "0.78907746", "0.78907746", "0.78907746", "0.78907746", "0.78712374", "0.76928234", "0.76893675", "0.7688665", "0.7688665", "0.7688665", "0.76567596", "0.75705755", "0.7567119", "0.75659263", "0.741002", "0.74039495", "0.738016", "0.73081607", "0.72831273", "0.72831273", "0.7208672", "0.7188059", "0.71852016", "0.71818304", "0.71231407", "0.71231407", "0.71231407", "0.71231407", "0.71231407", "0.70794576", "0.70794576", "0.7033021", "0.7031511", "0.7029756", "0.70153874", "0.69883513", "0.6983458", "0.6975979", "0.6967625", "0.69609183", "0.6928662", "0.69091177", "0.6897687", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.68943536", "0.6879462", "0.6868003", "0.6855652", "0.6851298", "0.6848685", "0.6840217", "0.6829919", "0.6829149", "0.68278414", "0.68239653", "0.68156725", "0.6804227", "0.6799002", "0.6787549", "0.67735755", "0.67688906", "0.67611164", "0.67587006", "0.6712741", "0.6702978", "0.6693654", "0.66725016", "0.6670033", "0.6670033", "0.6646335", "0.66434634", "0.6640711", "0.6636861", "0.6622298", "0.66192377", "0.6599641" ]
0.8420511
1
Now() should replace usage of time.Now(). If chrono has been paused, Now() returns the time when it was paused. If there is any skew (due to forwarding or reversing), this is always added to the end time.
func (c *chrono) Now() time.Time { var now time.Time if c.paused { now = c.pausedAt } else { now = time.Now() } return now.Add(c.skew) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (u *Util) Now() time.Time {\n\tif controlDuration != 0 {\n\t\treturn time.Now().Add(controlDuration).UTC()\n\t}\n\treturn time.Now().UTC()\n}", "func (c *Clock) Now() time.Time {\n\treturn time.Now().UTC().Truncate(time.Second)\n}", "func (realClocker) Now() time.Time { return time.Now() }", "func (c *Clock) Now() time.Time { return time.Now() }", "func (c *StoppedClock) Now() time.Time {\n\treturn c.t\n}", "func Now() time.Time { return now() }", "func Now() time.Time { return time.Now().UTC() }", "func Now() time.Time { return time.Now().UTC() }", "func (c *RunningClock) Now() time.Time {\n\treturn time.Now()\n}", "func (c *StoppedClock) Now() time.Time {\n\treturn c.time\n}", "func (c *FakeClock) Now() time.Time {\n\tc.steps++\n\treturn c.Time(c.steps)\n}", "func (c stubClocker) Now() time.Time { return c.t }", "func (c *Clock) Now() time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tstep := c.step\n\tif c.skipStep {\n\t\tstep = 0\n\t\tc.skipStep = false\n\t}\n\tc.advanceLocked(rt, step)\n\n\treturn c.present\n}", "func (p *PredefinedFake) Now() time.Time {\n\tadjustedTime := p.Base.Add(p.Delays[p.Next])\n\tp.Next++\n\treturn adjustedTime\n}", "func Now() time.Time {\n\treturn Clock.Now()\n}", "func now() time.Duration { return time.Since(startTime) }", "func (clock *Clock) Now() time.Time {\n\tvar now time.Time\n\tclock.do(func(s *state) {\n\t\tnow = s.t\n\t\ts.t = s.t.Add(clock.step)\n\t})\n\treturn now\n}", "func Now() time.Time { return time.Now() }", "func Now() time.Time {\n\tif IsTest() {\n\t\treturn now\n\t}\n\n\treturn time.Now()\n}", "func (p PT) Now() int64 {\n\tif p.Seconds {\n\t\treturn time.Now().Unix()\n\t}\n\treturn time.Now().UnixNano()\n}", "func Now() time.Time {\n\treturn CurrentClock().Now()\n}", "func (Clock) Now() time.Time {\n\treturn time.Now()\n}", "func (t *Time) Now() time.Time {\n\treturn t.current\n}", "func (fc *fakeClock) Now() time.Time {\n\treturn fc.time\n}", "func (defaultClock) Now() time.Time {\n\treturn time.Now()\n}", "func (c *CumulativeClock) Now() time.Time {\n\treturn c.current\n}", "func (d *dummyClock) Now() time.Time {\n\treturn time.Time{}\n}", "func (t *DefaultClock) Now() time.Time {\n\treturn time.Now()\n}", "func Now() time.Time {\n\treturn time.Unix(0, time.Now().UnixNano()/1e6*1e6)\n}", "func Now() Time {\n\treturn DefaultScheduler.Now()\n}", "func Now() time.Time {\n\treturn now()\n}", "func Now() time.Time {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tvar now time.Time\n\tif testMode {\n\t\tnow = testNow()\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.In(localtz.Get())\n}", "func (RealClock) Now() time.Time {\n\treturn time.Now()\n}", "func (t *Timer) Now() time.Duration {\n\treturn nowFunc().Sub(t.Zero)\n}", "func (c *MockClock) Now() time.Time {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\treturn c.now\n}", "func Now() time.Time {\n\treturn Work.Now()\n}", "func (wc WallClock) Now() time.Time {\n\treturn time.Now()\n}", "func (tc *TestClock) Now() time.Time {\n\ttc.l.RLock()\n\tdefer tc.l.RUnlock()\n\treturn tc.t\n}", "func (a *IncrementingFakeTimeSource) Now() time.Time {\n\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\n\ta.NextIncrement++\n\n\treturn adjustedTime\n}", "func now() time.Time {\n\treturn time.Now().UTC()\n}", "func Now() time.Time {\n\treturn c.Now()\n}", "func (c *Context) Now() time.Time {\n\treturn c.currentTime\n}", "func Now(ctx Context) time.Time {\n\treturn getWorkflowEnvironment(ctx).Now()\n}", "func Now() Time {\n\treturn Time(time.Now().UnixNano() / 1000)\n}", "func Now() time.Time {\n\treturn time.Now().In(_defaultLocation)\n}", "func (s *Scheduler) now() Time {\n\treturn s.provider.Now().Add(s.timeOffset)\n}", "func nowTime() time.Time {\n\treturn time.Now().UTC()\n}", "func Now() Time {\n\treturn Time{format(time.Now())}\n}", "func Now() Time {\n\treturn NewTime(time.Now())\n}", "func (c *webCtx) Now() time.Time {\n\treturn c.now\n}", "func (f FakeTimeSource) Now() time.Time {\n\treturn f.FakeTime\n}", "func now() int64 {\n\treturn time.Nanoseconds()\n}", "func (c *ClockVal) Now() {\n\tc.CurrentTime = NowTime()\n}", "func (m *Mock) Now() time.Time {\n\tm.Lock()\n\tdefer m.Unlock()\n\treturn m.now\n}", "func Now() *time.Time {\n\tt := time.Now()\n\treturn &t\n}", "func (t *TimeTravelCtx) now() time.Time {\n\tt.mutex.RLock()\n\tdefer t.mutex.RUnlock()\n\treturn t.ts\n}", "func Now() Time {\n\t/*\n\t\tft := windows.Filetime{}\n\t\twindows.GetSystemTimeAsFileTime(&ft)\n\t\treturn Time(ft.Nanoseconds() / nanosecondsScale)\n\t*/\n\t// On Windows, it turns out to be much faster to simply call time.Now().Unix(), so do it\n\treturn Time(time.Now().Unix())\n}", "func Now() Time {\n\tvar t Time\n\tt.FromNSec(uint64(gotime.Now().UnixNano()))\n\treturn t\n}", "func TimeNow() time.Time {\n\treturn time.Now().UTC()\n}", "func (m *Mock) Now() time.Time {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\treturn m.now\n}", "func (m *timeSource) Now() mstime.Time {\n\treturn mstime.Now()\n}", "func (f *FakeTimeSource) Now() time.Time {\n\tf.mu.RLock()\n\tdefer f.mu.RUnlock()\n\treturn f.now\n}", "func GetCurrentTime() uint64 {\n\treturn ComposeTSByTime(time.Now(), 0)\n}", "func (f *FixedTimeSource) Now() time.Time {\n\treturn f.fakeTime\n}", "func TestClock_Now(t *testing.T) {\n\tnow := raft.NewClock().Now()\n\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\n\t\tt.Fatalf(\"clock time is different than wall time: exp=%v, got=%v\", exp, now)\n\t}\n}", "func (s systemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func (s *Scheduler) now() time.Time {\n\tif s.nowTimeFn == nil {\n\t\treturn time.Now()\n\t} else {\n\t\treturn s.nowTimeFn()\n\t}\n}", "func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func TimeNow() Time {\n\treturn Time{\n\t\ttime.Now(),\n\t}\n}", "func (ns *Namespace) Now() _time.Time {\n\treturn _time.Now()\n}", "func Now() time.Time {\n\treturn time.Date(int(Year.Status().(uint16)), time.Month(Month.Status().(uint8)), int(Day.Status().(uint16)), int(Hour.Status().(uint8)), int(Minute.Status().(uint8)), int(Second.Status().(uint8)), 0, time.FixedZone(\"custom\", int(Tz.Status().(int8)) * 3600))\n}", "func stoppableNow() time.Time {\n\tstopMux.RLock()\n\tdefer stopMux.RUnlock()\n\n\tif stoppedAt != nil {\n\t\treturn *stoppedAt\n\t}\n\treturn time.Now().Add(-stoppedFor)\n}", "func (s *Scheduler) Now() Time {\n\ts.lock.RLock()\n\tdefer s.lock.RUnlock()\n\treturn s.now()\n}", "func (t Time) GetNow() time.Time {\n\treturn now()\n}", "func CurrentTime() Time {\n\tif atomic.LoadUint32(&isRunning) != 0 {\n\t\treturn currentTime\n\t}\n\n\treturn Time(time.Now().UnixNano() / 1e6)\n}", "func Now() Timespec {\n\treturn NewTimespec(time.Now())\n}", "func Now() DateTime {\n\treturn DateTimeFromTime(time.Now())\n}", "func Now(upToSecond ...bool) *TimeStamp {\n\treturn TimeFrom(time.Now(), upToSecond...)\n}", "func (em *eventManager) Now() time.Time {\n\tem.mu.Lock()\n\tdefer em.mu.Unlock()\n\treturn em.now\n}", "func Now() time.Time {\n\treturn time.Now().In(LOCATION)\n}", "func (rs *requestContext) Now() time.Time {\n\treturn rs.now\n}", "func GetNow() time.Time {\n\treturn time.Now().UTC()\n}", "func GetNow() time.Time {\n\treturn time.Now().UTC()\n}", "func NowTime() time.Time {\n\treturn ExtractTimeFromDatetime(time.Now())\n}", "func CurrentTime() time.Time {\n\treturn time.Now()\n}", "func ExampleNow() {\n\tt := gtime.Now()\n\tfmt.Println(t)\n\n\t// May Output:\n\t// 2021-11-06 13:41:08\n}", "func (tr *TextRegion) TimeNow() {\n\ttr.Time.Now()\n}", "func Now() int64 {\n\treturn time.Now().Unix()\n}", "func Now() int64 {\n\treturn time.Now().Unix()\n}", "func (t *TimeService) Now(request *NowRequest) (*NowResponse, error) {\n\trsp := &NowResponse{}\n\treturn rsp, t.client.Call(\"time\", \"Now\", request, rsp)\n}", "func GetCurrentTime() typeutil.Timestamp {\n\treturn ComposeTSByTime(time.Now(), 0)\n}", "func Now(ctx context.Context) time.Time {\n\tif ts := ctx.Value(ContextKey); ts != nil {\n\t\tswitch v := ts.(type) {\n\t\tcase NowProvider:\n\t\t\treturn v()\n\t\tcase time.Time:\n\t\t\treturn v\n\t\tdefault:\n\t\t\tpanic(fmt.Sprintf(\"Unknown value for ContextKey: %v\", v))\n\t\t}\n\t}\n\treturn time.Now()\n}", "func (t *TOTP) Now() string {\n\treturn t.At(currentTimestamp())\n}", "func dbNow() time.Time {\n\treturn roundTime(time.Now())\n}", "func Now() Date {\n\tn := time.Now()\n\treturn Of(n)\n}", "func tNow() uint64 {\n\treturn uint64(time.Now().In(time.UTC).UnixNano())\n}", "func (c *Cron) now() time.Time {\n\treturn time.Now().In(c.location)\n}", "func makeFakeNow() func() time.Time {\n\tvar t time.Time\n\treturn func() time.Time {\n\t\tt = t.Add(time.Second)\n\t\treturn t\n\t}\n}", "func defaultNow() time.Time {\n\treturn time.Now().UTC()\n}" ]
[ "0.7707415", "0.7624919", "0.7593016", "0.7560435", "0.7556393", "0.75166434", "0.750841", "0.750841", "0.75061864", "0.75025976", "0.74946547", "0.7469169", "0.74512213", "0.7446948", "0.74375767", "0.7418802", "0.7408312", "0.7363236", "0.73555875", "0.73452246", "0.7335264", "0.7324511", "0.73211706", "0.72900397", "0.727462", "0.72717696", "0.7267732", "0.7267188", "0.7235727", "0.71980613", "0.7178802", "0.7170116", "0.71680427", "0.7163473", "0.7153584", "0.71471274", "0.7145478", "0.71379447", "0.71215934", "0.706913", "0.70586264", "0.69819075", "0.6975091", "0.6967875", "0.69553274", "0.69470847", "0.692535", "0.6902087", "0.68836117", "0.68544614", "0.68344927", "0.6816418", "0.6807244", "0.6792128", "0.6790532", "0.67905074", "0.6772249", "0.67709064", "0.6756", "0.67420495", "0.6730117", "0.67076683", "0.6687628", "0.66865784", "0.66746855", "0.66665804", "0.6655333", "0.6634851", "0.6634851", "0.66326106", "0.65709186", "0.65596914", "0.65103865", "0.64965206", "0.6493475", "0.6489606", "0.6461745", "0.6433064", "0.6424062", "0.64114916", "0.6406821", "0.63955444", "0.6391871", "0.6391871", "0.6343702", "0.63428", "0.6324995", "0.63193285", "0.62775713", "0.62775713", "0.62687683", "0.62623066", "0.6259163", "0.623397", "0.6224072", "0.6217173", "0.6191839", "0.6184208", "0.61661226", "0.61654395" ]
0.84008926
0
Forwards time of chrono with skew time. This can be used in both running and paused mode.
func (c *chrono) Forward(skew time.Duration) { c.skew = skew }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}", "func Timeskew(cctx *testcontext.Context, name, offset string, pods ...string) (Teardown, error) {\n\ttc := chaos.TimeChaos{}\n\ttc.Name = name\n\ttc.Namespace = cctx.Namespace\n\n\ttc.Spec.Mode = chaos.AllMode\n\ttc.Spec.Selector = selectPods(pods)\n\ttc.Spec.TimeOffset = offset\n\n\tif err := cctx.Generic.Create(cctx, &tc); err != nil {\n\t\treturn nil, err\n\t}\n\treturn func(ctx context.Context) error {\n\t\treturn cctx.Generic.Delete(ctx, &tc)\n\t}, nil\n}", "func (c *chrono) Reset() {\n\tc.skew = 0\n}", "func (Order) Clockwise() Winding { return Clockwise }", "func (c *Clock) AdvanceTo(t time.Time) {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\tc.realTime = rt\n\tc.present = t\n\tc.events.AdvanceTo(c.present)\n}", "func AdvanceTime(by Duration) (newOffset Duration) {\n\treturn DefaultScheduler.AdvanceTime(by)\n}", "func UpTime() func() int64 {\n\tvar startTime int64 = timeStamp()\n\treturn func() int64 {\n\t\treturn timeStamp() - startTime\n\t}\n}", "func (c *MockClock) AdvanceTo(t time.Time) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tc.advanceTo(t)\n}", "func (c stubClocker) Now() time.Time { return c.t }", "func (c *Client) ClockSkew() time.Duration {\n\treturn c.clockSkew\n}", "func (c *Clock) Now() time.Time {\n\treturn time.Now().UTC().Truncate(time.Second)\n}", "func (s *Scheduler) AdvanceTime(by Duration) (newOffset Duration) {\n\ts.lock.Lock()\n\tdefer s.lock.Unlock()\n\n\ts.timeOffset += by\n\n\tif s.scheduled.Timer != nil {\n\t\ts.scheduled.Timer.Reset(s.scheduled.ID.Due().Sub(s.now()))\n\t}\n\treturn s.timeOffset\n}", "func (tt *TimeLatencyTracking) ComputeSkew() (offset int64, latency int64, _ error) {\n\tif tt.PollerSendTimestamp == 0 || tt.PollerRecvTimestamp == 0 || tt.ServerRecvTimestamp == 0 || tt.ServerRespTimestamp == 0 {\n\t\treturn 0, 0, errors.New(\"Unable to compute with any unset timestamp\")\n\t}\n\n\t// Variable aliases for the timeline\n\tT1 := tt.PollerSendTimestamp\n\tT2 := tt.ServerRecvTimestamp\n\tT3 := tt.ServerRespTimestamp\n\tT4 := tt.PollerRecvTimestamp\n\n\toffset = ((T2 - T1) + (T3 - T4)) / 2\n\tlatency = ((T4 - T1) + (T3 - T2))\n\n\treturn\n}", "func AdvanceBy(duration time.Duration) {\n\tAdvanceTo(Now().Add(duration))\n}", "func (Order) CounterClockwise() Winding { return CounterClockwise }", "func now() time.Duration { return time.Since(startTime) }", "func monotonicTimeSince(start time.Time) (end time.Time) {\n\treturn start.Add(time.Since(start))\n}", "func (p *PredefinedFake) Now() time.Time {\n\tadjustedTime := p.Base.Add(p.Delays[p.Next])\n\tp.Next++\n\treturn adjustedTime\n}", "func AdvanceTo(newTime time.Time) {\n\tnow := Now()\n\tfound := false\n\tfor _, s := range sortedSchedulers() {\n\t\tnextTick := s.tickAfter(now)\n\t\tif nextTick.After(now) && !nextTick.After(newTime) {\n\t\t\tfound = true\n\t\t\tnowInTest.Store(nextTick)\n\t\t\ts.Lock()\n\t\t\tif s.paused {\n\t\t\t\ts.fireOnResume = true\n\t\t\t} else {\n\t\t\t\ts.notifyFn()\n\t\t\t}\n\t\t\ts.Unlock()\n\t\t}\n\t}\n\tif !found {\n\t\tnowInTest.Store(newTime)\n\t\treturn\n\t}\n\tif Now() != newTime {\n\t\tAdvanceTo(newTime)\n\t}\n}", "func (c *FakeClock) Now() time.Time {\n\tc.steps++\n\treturn c.Time(c.steps)\n}", "func (u *Util) Now() time.Time {\n\tif controlDuration != 0 {\n\t\treturn time.Now().Add(controlDuration).UTC()\n\t}\n\treturn time.Now().UTC()\n}", "func sleep(delay uint) {\n\tfmt.Println(\"Time Now is:\",time.Now());\n\t//Time.After returns current time\n\tdelayTime := <-time.After(time.Second * time.Duration(delay))\n\tfmt.Println(\"Timestamp after delay:\",delay,\" by time.After is:\", delayTime);\n}", "func (realClocker) Now() time.Time { return time.Now() }", "func main() {\n\n\t//t1 := time.Now()\nt1 := time.Now()\n\nfmt.Printf(\"%T\\n\", t1)\n\nfmt.Println(t1.Format(time.StampMilli))\n\ntime.Sleep(33*time.Millisecond)\n\n\tt2 := time.Now()\nfmt.Println(t2.Format(time.StampMilli))\n\nduration := t2.Sub(t1)\nfmt.Println(duration.Seconds())\n\n}", "func TimeSinceSmoothStep(d0 time.Duration, p0 float64, d1 time.Duration, p1 float64) TimePriority {\n\tx0 := float64(d0)\n\tx1 := float64(d1)\n\treturn func(t time.Time) float64 {\n\t\tx := float64(time.Since(t))\n\t\treturn smoothstep(x, x0, p0, x1, p1)\n\t}\n}", "func (c *MockClock) Set(t time.Time) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tif t.After(c.now) {\n\t\tc.advanceTo(t)\n\t} else {\n\t\tc.now = t // move back in time\n\t}\n}", "func (ms *MVCCStats) Forward(nowNanos int64) {\n\tif ms.LastUpdateNanos >= nowNanos {\n\t\treturn\n\t}\n\tms.AgeTo(nowNanos)\n}", "func (c *StoppedClock) Now() time.Time {\n\treturn c.t\n}", "func Until(gameTime time.Time, fromTime time.Time) uint16 {\n\tfmt.Println(gameTime)\n\treturn uint16(math.RoundToEven(gameTime.Sub(fromTime).Seconds()))\n}", "func (c *Clock) Now() time.Time { return time.Now() }", "func SawTooth(x, period float64) float64 {\n\tx += period / 2\n\tt := x / period\n\treturn period*(t-math.Floor(t)) - period/2\n}", "func (c *MockClock) advanceTo(end time.Time) {\n\tfor {\n\t\tt := c.nextWakeupTime()\n\t\tif t == (time.Time{}) || t.After(end) {\n\t\t\tbreak\n\t\t}\n\t\tc.advanceToNextWakeup()\n\t}\n\tc.now = end\n}", "func toDelay(apiTime int64) time.Duration {\n\treturn time.Now().Sub(time.Unix(apiTime, 10))\n}", "func roundUp(t time.Time, d time.Duration) time.Time {\n\tr := t.Round(d)\n\tif t.After(r) {\n\t\tr = r.Add(d)\n\t}\n\treturn r\n}", "func (c *StoppedClock) Now() time.Time {\n\treturn c.time\n}", "func Time(then time.Time) string {\n\treturn RelTime(then, time.Now(), \"ago\", \"from now\")\n}", "func (c *CumulativeClock) Now() time.Time {\n\treturn c.current\n}", "func (a *IncrementingFakeTimeSource) Now() time.Time {\n\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\n\ta.NextIncrement++\n\n\treturn adjustedTime\n}", "func (o BucketAutoclassOutput) ToggleTime() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v BucketAutoclass) *string { return v.ToggleTime }).(pulumi.StringPtrOutput)\n}", "func (c FakeClock) Time(step int) time.Time {\n\t// We truncate to microsecond precision because Postgres' timestamptz type\n\t// doesn't handle nanoseconds.\n\treturn c.epoch.Add(time.Duration(step) * c.step).UTC().Truncate(time.Microsecond)\n}", "func stoppableNow() time.Time {\n\tstopMux.RLock()\n\tdefer stopMux.RUnlock()\n\n\tif stoppedAt != nil {\n\t\treturn *stoppedAt\n\t}\n\treturn time.Now().Add(-stoppedFor)\n}", "func AnounanceTime(){\n\tfor {\n\t\tfmt.Println(time.Now())\n\t\ttime.Sleep(5*time.Second)\n\t}\n}", "func Now(ctx Context) time.Time {\n\treturn getWorkflowEnvironment(ctx).Now()\n}", "func (t *Time) Sleep(s time.Duration) {\n\tt.current = t.current.Add(s)\n}", "func makeFakeNow() func() time.Time {\n\tvar t time.Time\n\treturn func() time.Time {\n\t\tt = t.Add(time.Second)\n\t\treturn t\n\t}\n}", "func (t *TimeLine) forward(num, denom uint32, runCallbacks bool) {\n\tend := t.cursor + t.Ticks(num, denom)\n\tif runCallbacks {\n\t\tt.lastDelta = t.runCallbacks(t.cursor, end)\n\t}\n\tt.cursor = end\n}", "func Since(t Time) Duration {\n\treturn Now().Sub(t)\n}", "func (t *Time) Since(position Position) Distance {\n\treturn Distance(t.Current() - position)\n}", "func (c *chrono) Pause() {\n\tc.pausedAt = c.Now()\n\tc.paused = true\n}", "func (wc WallClock) Now() time.Time {\n\treturn time.Now()\n}", "func Tomorrow() time.Time {\n\treturn time.Now().AddDate(0, 0, 1)\n}", "func Now() time.Time { return time.Now().UTC() }", "func Now() time.Time { return time.Now().UTC() }", "func Time() time.Time {\n\tnow := time.Now().UTC()\n\tdif := time.Duration(rand.Int())\n\tstart := now.Add(dif * -1)\n\tend := now.Add(dif)\n\treturn TimeSpan(start, end)\n}", "func (fc *fakeClock) Now() time.Time {\n\treturn fc.time\n}", "func (c *MockClock) advanceToNextWakeup() {\n\tif len(c.wakeups) < 1 {\n\t\treturn\n\t}\n\tw := heap.Pop(&c.wakeups).(*wakeup)\n\tc.logf(\"MockClock: Advancing time from now=%s to next wakeup time %s.\",\n\t\ttsStr(c.now), tsStr(w.t))\n\tc.now = w.t\n\tselect {\n\tcase w.c <- w.t:\n\tdefault:\n\t}\n\t// give things a chance to run\n\truntime.Gosched()\n\tc.logf(\"MockClock: Advanced time, now=%s.\", tsStr(c.now))\n}", "func (c *MockClock) Advance(delta time.Duration) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tend := c.now.Add(delta)\n\tc.advanceTo(end)\n}", "func (t *Time) Sleep(distance Distance) {\n\ttime.Sleep(time.Duration(distance))\n}", "func (o BucketAutoclassPtrOutput) ToggleTime() pulumi.StringPtrOutput {\n\treturn o.ApplyT(func(v *BucketAutoclass) *string {\n\t\tif v == nil {\n\t\t\treturn nil\n\t\t}\n\t\treturn v.ToggleTime\n\t}).(pulumi.StringPtrOutput)\n}", "func Thursday(t time.Time) time.Time {\n\treturn t.AddDate(0, 0, -Weekday(t)+4)\n}", "func (s *Stopwatch) Handover(s2 *Stopwatch) time.Duration {\n\ts2.t = time.Now()\n\td := s2.t.Sub(s.t)\n\ts.acc += d\n\treturn d\n}", "func (c *Clock) Since(t time.Time) time.Duration {\n\treturn c.Now().Sub(t)\n}", "func roundup(t time.Time, d time.Duration) time.Time {\n\trounded := t.Round(d)\n\tif rounded.Before(t) {\n\t\trounded = rounded.Add(d)\n\t}\n\treturn rounded\n}", "func (s *Scheduler) now() Time {\n\treturn s.provider.Now().Add(s.timeOffset)\n}", "func (ship *Ship) fwdThrustersOff() {\n\tship.fwdThrusters = false\n\tif !ship.isThrusting() {\n\t\tstopRcsSound()\n\t}\n}", "func (c *MockClock) Since(t time.Time) time.Duration {\n\treturn c.Now().Sub(t)\n}", "func (c *RunningClock) Now() time.Time {\n\treturn time.Now()\n}", "func Until(t Time) Duration {}", "func (c *chrono) Continue() {\n\tc.paused = false\n}", "func (t Time) After(u Time) bool {}", "func TestClock_Now(t *testing.T) {\n\tnow := raft.NewClock().Now()\n\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\n\t\tt.Fatalf(\"clock time is different than wall time: exp=%v, got=%v\", exp, now)\n\t}\n}", "func loop() {\n\n\tdelta := 0 * time.Nanosecond\n\n\tlast := time.Now()\n\n\tfor true {\n\t\tcur := time.Now()\n\t\tdelta += cur.Sub(last)\n\t\tlast = cur\n\n\t\tfor delta >= 15*time.Millisecond {\n\t\t\tdelta -= time.Millisecond\n\t\t\t//fmt.Println(\"Up\")\n\t\t}\n\t\t//fmt.Println(\"Re\")\n\t}\n\n}", "func FakeClockNowWithExtraTime(d time.Duration) {\n\tclock.Now = func() time.Time {\n\t\treturn originalMockedTime.Add(d)\n\t}\n}", "func (clock *Clock) Now() time.Time {\n\tvar now time.Time\n\tclock.do(func(s *state) {\n\t\tnow = s.t\n\t\ts.t = s.t.Add(clock.step)\n\t})\n\treturn now\n}", "func (ft *Time) Next() {\n\ta := rand.Float64()\n\n\t// Ensure first time doesn't have any variance to respect the start time parameter\n\tif ft.firstVal {\n\t\tft.firstVal = false\n\t\tft.v = ft.ts\n\n\t\tif ft.keepStats {\n\t\t\tft.Stats.Add(ft.v)\n\t\t}\n\n\t\treturn\n\t}\n\n\tft.ts = ft.ts.Add(time.Duration(ft.increment) * time.Millisecond)\n\ttmp := (float64(ft.variance) * a) - float64(int64(float64(ft.variance)*a))\n\ttmp2 := float64(-1)\n\n\tif ft.direction < 0 {\n\t\ttmp2 = float64(-1)\n\t} else if ft.direction > 0 {\n\t\ttmp2 = float64(1)\n\t} else if tmp > 0.5 {\n\t\ttmp2 = float64(1)\n\t}\n\n\tc := int64(round(float64(ft.variance)*a, 0.0000000005) * tmp2)\n\tft.v = ft.ts.Add(time.Duration(c) * time.Millisecond)\n\n\tif ft.keepStats {\n\t\tft.Stats.Add(ft.v)\n\t}\n}", "func (c *ClockVal) Now() {\n\tc.CurrentTime = NowTime()\n}", "func (p PT) Now() int64 {\n\tif p.Seconds {\n\t\treturn time.Now().Unix()\n\t}\n\treturn time.Now().UnixNano()\n}", "func (c *Clock) Now() time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tstep := c.step\n\tif c.skipStep {\n\t\tstep = 0\n\t\tc.skipStep = false\n\t}\n\tc.advanceLocked(rt, step)\n\n\treturn c.present\n}", "func OutputContinueTo(t time.Duration) {\n\tdeltaDur := t - outputToDur\n\tdeltaTz := Tz(masterFreq * float64((deltaDur)/time.Second))\n\t//debug.Printf(\"mix.OutputContinueTo(%+v) deltaDur:%+v nowTz:%+v deltaTz:%+v begin...\", t, deltaDur, nowTz, deltaTz)\n\tApiOutputNext(deltaTz)\n\toutputToDur = t\n\t//debug.Printf(\"mix.OutputContinueTo(%+v) ...done! nowTz:%+v outputToDur:%+v\", t, nowTz, outputToDur)\n}", "func (o BucketAutoclassResponseOutput) ToggleTime() pulumi.StringOutput {\n\treturn o.ApplyT(func(v BucketAutoclassResponse) string { return v.ToggleTime }).(pulumi.StringOutput)\n}", "func Now() time.Time { return now() }", "func (d *MonotonicRandomWalkDistribution) Advance() {\n\td.Step.Advance()\n\td.State += math.Abs(d.Step.Get())\n}", "func TestCap(t *testing.T) {\n\tbacko := NewBacko(milliseconds(100), 2, 0, milliseconds(600))\n\n\tassert.Equal(t, milliseconds(100), backo.Duration(0))\n\tassert.Equal(t, milliseconds(200), backo.Duration(1))\n\tassert.Equal(t, milliseconds(400), backo.Duration(2))\n\tassert.Equal(t, milliseconds(600), backo.Duration(3))\n}", "func (tl *TimeLord) warp() {\n\ttl.guard = monkey.Patch(time.Now, func() time.Time {\n\t\ttl.guard.Unpatch()\n\t\tdefer tl.guard.Restore()\n\n\t\treturn time.Now().Add(tl.offset)\n\t})\n}", "func (ct *ClockTimestamper) Stamp(height abi.ChainEpoch) uint64 {\n\tstartTime := ct.c.StartTimeOfEpoch(height)\n\n\treturn uint64(startTime.Unix())\n}", "func (fc *FakeClock) Sleep(d time.Duration) {\n\t<-fc.After(d)\n}", "func (t *TimeTravelCtx) now() time.Time {\n\tt.mutex.RLock()\n\tdefer t.mutex.RUnlock()\n\treturn t.ts\n}", "func sinceDate(t time.Time) string { return time.Since(t).Truncate(time.Second).String() }", "func StubNow(stub func() time.Time) { now = stub }", "func NewTimeOver(side Colour) Outcome { return Outcome{Winner: side.Invert(), Reason: timeOver} }", "func (ship *Ship) cwThrustersOff() {\n\tship.cwThrusters = false\n\tif !ship.isThrusting() {\n\t\tstopRcsSound()\n\t}\n}", "func TimeShrinker(v interface{}) gopter.Shrink {\n\tt := v.(time.Time)\n\tsec := t.Unix()\n\tnsec := int64(t.Nanosecond())\n\tsecShrink := int64Shrink{\n\t\toriginal: sec,\n\t\thalf: sec,\n\t}\n\tnsecShrink := int64Shrink{\n\t\toriginal: nsec,\n\t\thalf: nsec,\n\t}\n\treturn gopter.Shrink(secShrink.Next).Map(func(v int64) time.Time {\n\t\treturn time.Unix(v, nsec)\n\t}).Interleave(gopter.Shrink(nsecShrink.Next).Map(func(v int64) time.Time {\n\t\treturn time.Unix(sec, v)\n\t}))\n}", "func StandardFor(t time.Time) times.Object {\n\treturn Standard(t.Second(), t.Minute(), t.Hour(), t.Day())\n}", "func TimeFrom(t time.Time, upToSecond ...bool) *TimeStamp {\n\tts := &TimeStamp{t, t.Unix(), 0}\n\tif len(upToSecond) > 0 && upToSecond[0] == true {\n\t\tts.Time = time.Unix(ts.UnixSecond, 0).Local()\n\t\tts.UnixNanoSecond = ts.UnixSecond * 1e9\n\t} else {\n\t\tts.UnixNanoSecond = t.UnixNano()\n\t}\n\treturn ts\n}", "func TestMonotonic(t *testing.T) {\n\told := Now()\n\tfor i := 0; i < 50; i++ {\n\t\tnext := Now()\n\t\tif next.After(old) {\n\t\t\tt.Error(\"Whitening insufficient\")\n\t\t}\n\t\ttime.Sleep(time.Duration(whitenerMask)/time.Nanosecond + 1)\n\t\tnext = Now()\n\t\tif !next.After(old) {\n\t\t\tt.Error(\"Not monotonically increasing on whitened nano-second scale\")\n\t\t}\n\t\told = next\n\t}\n}", "func (c *Clock) Advance(d time.Duration) time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\n\tc.advanceLocked(rt, d)\n\treturn c.present\n}", "func (l *limiter) floor(mark time.Time, now time.Time) time.Time {\n\tif t := now.Add(-l.quantum); !mark.After(t) {\n\t\treturn t\n\t}\n\treturn mark\n}", "func (d *MonotonicUpDownRandomWalkDistribution) Advance() {\n\td.Step.Advance()\n\td.State += d.Step.Get() * float64(d.direction)\n\tif d.State < d.Min {\n\t\td.State = d.Min\n\t\td.direction = 1\n\t} else if d.State > d.Max {\n\t\td.State = d.Max\n\t\td.direction = -1\n\t}\n}", "func (c *T) SkipNow()", "func GetCurrentSegment() time.Time {\n\treturn time.Now().UTC().Add(-(SegmentDur / 2)).Round(SegmentDur)\n}" ]
[ "0.6377099", "0.57420695", "0.55626804", "0.52779603", "0.5123773", "0.50687426", "0.5051111", "0.5007728", "0.49764827", "0.49639702", "0.49189162", "0.48556083", "0.4847172", "0.48447493", "0.48446742", "0.48409605", "0.4827538", "0.48221955", "0.47696277", "0.4712712", "0.46684983", "0.46447614", "0.4639994", "0.46390235", "0.46322784", "0.46279216", "0.46159396", "0.46143475", "0.45857015", "0.458469", "0.4575363", "0.45659685", "0.4560727", "0.4547705", "0.45358622", "0.45318428", "0.4515254", "0.4500201", "0.44983536", "0.44778347", "0.44768745", "0.44742745", "0.4463998", "0.44606817", "0.44397876", "0.4438386", "0.44350022", "0.44242248", "0.44209474", "0.44130993", "0.4408065", "0.4403237", "0.4403237", "0.4397341", "0.43957078", "0.43931156", "0.43916333", "0.43915167", "0.43913218", "0.4389962", "0.43754026", "0.4374432", "0.43721148", "0.43689024", "0.43624708", "0.4346447", "0.43423003", "0.43350083", "0.43214065", "0.43194172", "0.4317958", "0.43114153", "0.43071178", "0.43022218", "0.4280368", "0.4261439", "0.4258817", "0.4252349", "0.42464542", "0.42438734", "0.42428762", "0.42419526", "0.42364377", "0.42238113", "0.42218247", "0.42206946", "0.42198256", "0.42174518", "0.42142338", "0.42142045", "0.42106998", "0.42076206", "0.42072263", "0.42063943", "0.42023113", "0.42003888", "0.41998807", "0.41986006", "0.41964772", "0.41958177" ]
0.6801785
0
Resets any previous set clock skew.
func (c *chrono) Reset() { c.skew = 0 }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (sw *Stopwatch) Reset() {\n\tsw.t = time.Now()\n}", "func (rw *RandW) Reset() {\n\trw.r = rand.New(rand.NewSource(uint64(time.Now().UnixNano())))\n}", "func Reset() {\n\tmutex.Lock()\n\taverages = make(map[string]*currentAverage)\n\tmutex.Unlock()\n}", "func (s *Stopwatch) Reset() {\n\t*s = Stopwatch{}\n}", "func (d *RabinKarp64) Reset() {\n\td.tables = nil\n\td.value = 0\n\td.window = d.window[:0]\n\td.oldest = 0\n\td.updateTables()\n}", "func ResetNow() { now = time.Now }", "func (rd *ratedisp) reset() {\n\trd.start = time.Now()\n\trd.cnt = 0\n\trd.size = 0\n}", "func (e *Timing) Reset() {\n\te.Min = 0\n\te.Max = 0\n\te.Value = 0\n\te.Values = make(float64Slice, 0)\n\te.Count = 0\n}", "func resetTimeNow() {\n\ttimeNow = time.Now\n}", "func (t *TimeLine) Reset() {\n\tt.cursor = 0\n\tt.lastDelta = 0\n}", "func (s *Stopwatch) Reset() {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.isRunning() {\n\t\tfmt.Printf(\"WARNING: Stopwatch.Reset() isRunning is true\\n\")\n\t}\n\ts.refTime = time.Time{}\n\ts.elapsedTime = 0\n}", "func (ts *TextState) Reset() {\n\tts.Tm = transform.IdentityMatrix()\n\tts.Tlm = transform.IdentityMatrix()\n}", "func (c *TimeAvgAggregator) Reset(w Window) {\n\tc.integral = 0\n\tif c.initialized {\n\t\tc.startTime = w.StartTime\n\t\tc.startValue = c.endValue\n\t\tc.endTime = w.EndTime\n\t}\n}", "func Timeskew(cctx *testcontext.Context, name, offset string, pods ...string) (Teardown, error) {\n\ttc := chaos.TimeChaos{}\n\ttc.Name = name\n\ttc.Namespace = cctx.Namespace\n\n\ttc.Spec.Mode = chaos.AllMode\n\ttc.Spec.Selector = selectPods(pods)\n\ttc.Spec.TimeOffset = offset\n\n\tif err := cctx.Generic.Create(cctx, &tc); err != nil {\n\t\treturn nil, err\n\t}\n\treturn func(ctx context.Context) error {\n\t\treturn cctx.Generic.Delete(ctx, &tc)\n\t}, nil\n}", "func (tw *TimingsWrapper) Reset() {\n\ttw.timings.Reset()\n}", "func (lp *Loadpoint) resetMeasuredPhases() {\n\tlp.Lock()\n\tlp.measuredPhases = 0\n\tlp.Unlock()\n\n\tlp.publish(phasesActive, lp.activePhases())\n}", "func (tw *MultiTimingsWrapper) Reset() {\n\ttw.timings.Reset()\n}", "func (s *Stopwatch) Reset(offset time.Duration, active bool) {\n\tnow := time.Now()\n\ts.Lock()\n\tdefer s.Unlock()\n\ts.start = now.Add(-offset)\n\tif active {\n\t\ts.stop = time.Time{}\n\t} else {\n\t\ts.stop = now\n\t}\n\ts.mark = 0\n\ts.laps = nil\n}", "func (t *Transform) Reset() {\n\tt.access.Lock()\n\tt.parent = nil\n\tt.built = nil\n\tt.localToWorld = nil\n\tt.worldToLocal = nil\n\tt.quat = nil\n\tt.pos = lmath.Vec3Zero\n\tt.rot = lmath.Vec3Zero\n\tt.scale = lmath.Vec3One\n\tt.shear = lmath.Vec3Zero\n\tt.access.Unlock()\n}", "func ResetBoard() {\n\tfor i := 0; i < (Scale * Scale); i++ {\n\t\tBoard[i].Fill = false\n\t\tBoard[i].Slogic = strconv.Itoa(i + 1)\n\t}\n}", "func (p *TimePanel) Reset() {\n}", "func (w *Week) reset() {\n\tw.days = nil\n}", "func (mock *Serf) Reset() {\n\tlockSerfBootstrap.Lock()\n\tmock.calls.Bootstrap = nil\n\tlockSerfBootstrap.Unlock()\n\tlockSerfCluster.Lock()\n\tmock.calls.Cluster = nil\n\tlockSerfCluster.Unlock()\n\tlockSerfID.Lock()\n\tmock.calls.ID = nil\n\tlockSerfID.Unlock()\n\tlockSerfJoin.Lock()\n\tmock.calls.Join = nil\n\tlockSerfJoin.Unlock()\n\tlockSerfMember.Lock()\n\tmock.calls.Member = nil\n\tlockSerfMember.Unlock()\n\tlockSerfShutdown.Lock()\n\tmock.calls.Shutdown = nil\n\tlockSerfShutdown.Unlock()\n}", "func Reset() {\n\tC.yices_reset()\n}", "func (s *Spinner) Reset() {\n\ts.step = 0\n\ts.createFrames()\n}", "func (c *Clock) Reset(last UUID) error {\n\tif err := last.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\t// Reset clock internals to match uuid standards.\n\tc.id = last.ID\n\tc.tickT = last.Type\n\tc.origin = last.Origin\n\n\t// reset last tick of clock.\n\tc.last = &last\n\n\treturn nil\n}", "func (a *MovAvg) Reset() {\n\ta.r, a.w = 0, 0\n\tfor i := range a.sum {\n\t\ta.sum[i] = 0\n\t\ta.v[i] = 0\n\t}\n}", "func reseed() {\n\trand.Seed(time.Now().UTC().UnixNano())\n}", "func (c *CTR) Reset() {\n\tc.blocks = 0\n\tc.ks = nil\n}", "func (s *Surface) ResetTransform() {\n\ts.Ctx.Call(\"resetTransform\")\n}", "func (u *Util) ResetControlDuration() {\n\tcontrolDuration = 0\n}", "func (t Time) ResetNow() time.Time {\n\tnow = defaultNow\n\treturn now()\n}", "func (tr *trooper) reset() {\n\ttr.trash()\n\ttr.addCenter()\n\tfor cnt, b := range tr.bits {\n\t\tb.reset(tr.ipos[cnt])\n\t}\n\ttr.healthChanged(tr.health())\n}", "func (context *context) ResetTimings() {\n\tcontext.model.ctx.Whisper_reset_timings()\n}", "func reset() {\n\tterm.Sync()\n}", "func (self *Weights) reset() {\n\tif len(self.Scale)>0 {\n\t\tself.Scale = self.Scale[:0]\n\t}\n}", "func (wpr *Wrapper) Reset() {\n\twpr.O = 0\n\twpr.L = 0\n}", "func (tr *trooper) resetEnergy() {\n\ttr.teleportEnergy = tr.temax\n\ttr.cloakEnergy = 1000\n}", "func (c *Context) Reset() {\n\tc.sp = &storagePointer{0, 0, 0}\n\tfor _, storage := range c.storages {\n\t\tstorage.Disk.Format()\n\t}\n}", "func (c *CumulativeClock) Reset() {\n\tc.Set(c.start)\n}", "func UnfreezeClock(t *testing.T) {\n\tif t == nil {\n\t\tpanic(\"nice try\")\n\t}\n\tc = &DefaultClock{}\n}", "func (sc *Scavenger) Reset() {\n\tsc.mu.Lock()\n\tsc.entries = nil\n\tsc.mu.Unlock()\n}", "func (t *TriDense) Reset() {\n\t// N and Stride must be zeroed in unison.\n\tt.mat.N, t.mat.Stride = 0, 0\n\t// Defensively zero Uplo to ensure\n\t// it is set correctly later.\n\tt.mat.Uplo = 0\n\tt.mat.Data = t.mat.Data[:0]\n}", "func (gta *GlobalTSOAllocator) Reset() {\n\ttsoAllocatorRole.WithLabelValues(gta.timestampOracle.dcLocation).Set(0)\n\tgta.timestampOracle.ResetTimestamp()\n}", "func Unset() {\n\tmock = time.Time{}\n}", "func (b *blockEnc) reset(prev *blockEnc) {\n\tb.extraLits = 0\n\tb.literals = b.literals[:0]\n\tb.size = 0\n\tb.sequences = b.sequences[:0]\n\tb.output = b.output[:0]\n\tb.last = false\n\tif prev != nil {\n\t\tb.recentOffsets = prev.prevRecentOffsets\n\t}\n\tb.dictLitEnc = nil\n}", "func (c *Canvas) Reset() {\n\tfor y := 0; uint8(y) < canvasHeight; y++ {\n\t\tfor x := 0; uint8(x) < canvasWidth; x++ {\n\t\t\t(*c)[y][x] = 0\n\t\t}\n\t}\n}", "func (f *factory) ResetSecretsLocation() {\n\tf.secretLocation = nil\n}", "func Reset() {\n\tstats.Reset()\n}", "func (e *Zero) Reset() {}", "func (ssys *Ankiddie) FullReset() {\n\tssys.m.Lock()\n\tdefer ssys.m.Unlock()\n\tfor _, env := range ssys.envs {\n\t\tenv.cancel()\n\t}\n\tssys.envs = make(map[uint]*Environment)\n}", "func (s *Suite) Reset() {\n\tfor _, set := range s.sets {\n\t\tset.Reset()\n\t}\n}", "func (t Time) ResetTime() Time {\n\treturn Date(t.Year(), t.Month(), t.Day(), 0, 0, 0, 0, time.UTC).In(t.Location())\n}", "func (c *Cipher) Reset() {\n\tfor i := range c.state {\n\t\tc.state[i] = 0\n\t}\n\tfor i := range c.buf {\n\t\tc.buf[i] = 0\n\t}\n}", "func (r *randList) Reset() {\n\tr.offset = 0\n\tr.perm = rand.Perm(len(r.list))\n}", "func (m *Mouse) Reset(monWidth, monHeight uint32) {\n\tm.PosX = monWidth / 2\n\tm.PosY = monHeight / 2\n\tm.LeftBtnUp()\n\tm.RightBtnUp()\n\tm.Wheel.ScrollVal = 5\n\tm.Sensitivity(1)\n\tm.WriteJSON()\n}", "func (sl *StagesLatency) ResetStatistics() {\n\tsl.first = duplicateSlice(sl.last)\n\tsl.FirstCollected = sl.LastCollected\n\n\tsl.calculate()\n}", "func Reset() {\n\tstopMux.Lock()\n\tstoppedAt = nil\n\tstoppedFor = 0\n\tstopMux.Unlock()\n}", "func (r *Raft) ResetTimer(){\n\t//fmt.Println(\"Election TImer Reset\")\n\tif r.Id==0 {\n \tElectionTimer.Reset(time.Millisecond*10000) \t\n\t}else if r.Id==1 {\n \tElectionTimer.Reset(time.Millisecond*3000)\n }else if r.Id==2 {\n \tElectionTimer.Reset(time.Millisecond*12000)\n\t}else if r.Id==3 {\n \tElectionTimer.Reset(time.Millisecond*14000)\n }else if r.Id==4 {\n \tElectionTimer.Reset(time.Millisecond*16000)\n\t}else {\n\tElectionTimer.Reset(time.Millisecond*18000)\n\t}\n\n}", "func (w *denseRankWindow) Reset(context.Context) {\n\tw.denseRank = 0\n\tw.peerRes = nil\n}", "func (t *Timer) Reset() {\n\tt.goalTime = math.MaxFloat64\n\tt.startTime = time.Now()\n}", "func (s *SimPDF) ResetMargins() {\n\ts.AddMargins(s.Margin)\n\ts.PDF.SetLeftMargin(s.Margin.Left)\n\ts.PDF.SetRightMargin(s.Margin.Right)\n}", "func (mu *MuHash) Reset() {\n\tmu.numerator.SetToOne()\n\tmu.denominator.SetToOne()\n}", "func resetStart() {\n\tstart = time.Now()\n}", "func (imd *IMDraw) Reset() {\n\timd.points = imd.points[:0]\n\timd.Color = pixel.Alpha(1)\n\timd.Picture = pixel.ZV\n\timd.Intensity = 0\n\timd.Precision = 64\n\timd.EndShape = NoEndShape\n}", "func (st *SlimTrie) Reset() {\n\tst.Children.Array32.Reset()\n\tst.Steps.Array32.Reset()\n\tst.Leaves.Array32.Reset()\n}", "func (t *MCTS) Reset() {\n\tt.Lock()\n\tdefer t.Unlock()\n\n\tt.freelist = t.freelist[:0]\n\tt.freeables = t.freeables[:0]\n\tfor i := range t.nodes {\n\t\tt.nodes[i].move = -1\n\t\tt.nodes[i].visits = 0\n\t\tt.nodes[i].status = 0\n\t\tt.nodes[i].psa = 0\n\t\tt.nodes[i].hasChildren = false\n\t\tt.nodes[i].qsa = 0\n\t\tt.freelist = append(t.freelist, t.nodes[i].id)\n\t}\n\n\tfor i := range t.children {\n\t\tt.children[i] = t.children[i][:0]\n\t}\n\n\tt.nodes = t.nodes[:0]\n\tt.policies = nil\n\truntime.GC()\n}", "func (s *sum64_128) Reset() {\n\ts.h1, s.h2, s.k1, s.k2, s.length, s.offset = 0, 0, 0, 0, 0, 0\n}", "func (m *ZoneproductMutation) ResetZonestock() {\n\tm.zonestock = nil\n\tm.removedzonestock = nil\n}", "func (c *canvasRenderer) ResetTransform() {\n\tif c.currentLayer == c.topLayer {\n\t\tc.currentLayer.Transform = sprec.IdentityMat4()\n\t} else {\n\t\tc.currentLayer.Transform = c.currentLayer.previous.Transform\n\t}\n}", "func (a *Agent) Reset() {\n\ta.Sketch.Reset()\n\ta.Buf = nil // TODO: pool\n}", "func (m *IntervalMutation) ResetTrades() {\n\tm.trades = nil\n\tm.clearedtrades = false\n\tm.removedtrades = nil\n}", "func (es *ExponentialSleeper) Reset() {\n\tes.delay = es.initDelay\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (mb *MetricsBuilder) Reset(options ...metricBuilderOption) {\n\tmb.startTime = pcommon.NewTimestampFromTime(time.Now())\n\tfor _, op := range options {\n\t\top(mb)\n\t}\n}", "func (d *Detector) Reset() {\n\tC.fvad_reset(d.fvad)\n}", "func (g *Game) Reset() {\n\tg.prepareGame()\n\n\tg.GreedsReleased = false\n\tg.TimeElapsed = 0\n\tg.KillsCount = 0\n\tg.Over = false\n}", "func (t *Timer) Reset() {\n\tt.currentTime = t.getCurrentTimeMs()\n\tt.lastTime = t.currentTime\n\tt.tick = 0\n}", "func (g *Float32s) Reset() {\r\n\t// No state\r\n}", "func (c *SwitchTicker) Reset() {\n\tatomic.StoreInt64(&c.failCount, 0)\n}", "func (h *handlerImpl) reset(adjMatrix adjacencyMatrix) {\n\tfor from := range adjMatrix {\n\t\tfor to := range adjMatrix[from] {\n\t\t\tadjMatrix[from][to].disabled = false\n\t\t}\n\t}\n}", "func (el *Election) Reset(validators *pos.Validators, frameToDecide idx.Frame) {\n\tel.validators = validators\n\tel.frameToDecide = frameToDecide\n\tel.votes = make(map[voteID]voteValue)\n\tel.decidedRoots = make(map[idx.StakerID]voteValue)\n}", "func (sm3 *SM3) Reset() {\n\t// Reset digest\n\tsm3.digest[0] = 0x7380166f\n\tsm3.digest[1] = 0x4914b2b9\n\tsm3.digest[2] = 0x172442d7\n\tsm3.digest[3] = 0xda8a0600\n\tsm3.digest[4] = 0xa96f30bc\n\tsm3.digest[5] = 0x163138aa\n\tsm3.digest[6] = 0xe38dee4d\n\tsm3.digest[7] = 0xb0fb0e4e\n\n\tsm3.length = 0 // Reset numberic states\n\tsm3.unhandleMsg = []byte{}\n}", "func (s *Snowball) Reset() {\n\ts.Preferred = nil\n\ts.Last = nil\n\n\ts.Counts = make(map[[blake2b.Size256]byte]uint16)\n\ts.Count = 0\n\n\ts.Done = false\n}", "func (r *Robot) Reset() {\n\t*r = Robot{}\n}", "func (v *mandelbrotViewer) reset() {\n\tv.redraw = true\n\tv.maxIterations = defaultIterations\n\tv.rMin = rMin\n\tv.rMax = rMax\n\tv.iMin = iMin\n\tv.iMax = iMax\n\tv.zoom = zoom\n}", "func (d *state) Reset() {\n\t// Zero the permutation's state.\n\tfor i := range d.a {\n\t\td.a[i] = 0\n\t}\n\td.state = spongeAbsorbing\n\td.buf = d.storage[:0]\n}", "func ResetScreen() {\n\tClearScreen()\n\tResetAttrs()\n\tCursorHome()\n}", "func (g *Game) resetGame() {\n\tg.board = make([][]int, boardHeight)\n\tfor y := 0; y < boardHeight; y++ {\n\t\tg.board[y] = make([]int, boardWidth)\n\t\tfor x := 0; x < boardWidth; x++ {\n\t\t\tg.board[y][x] = 0\n\t\t}\n\t}\n\n\tg.state = gameIntro\n\tg.dots = 1\n\tg.d = 0\n\tg.x = 5\n\tg.y = 5\n\tg.direction = 3\n\tg.dotLocation = []int{0, 0}\n\tg.prevLocations = [][]int{{0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}, {0, 0}}\n\tg.dx = []int{1, 0, 0, 0}\n\tg.dy = []int{1, 0, 0, 0}\n\tg.dotX = -500\n\tg.dotY = -500\n\n\tg.fallingTimer = time.NewTimer(time.Duration(1000000 * time.Second))\n\tg.fallingTimer.Stop()\n\n}", "func (vm *BFVM) Reset() {\n\tvm.position = 0\n\n\tfor _, c := range vm.tape {\n\t\tc.Zero()\n\t}\n}", "func (m *ram) Reset() (err error) {\n\tfor i := 0; i < len(m.b); i++ {\n\t\tm.b[i] = 0\n\t}\n\treturn\n}", "func (rf *Raft) reset(term int) {\n\trf.term = term\n\tfor idx := range rf.votes {\n\t\trf.votes[idx] = -1\n\t}\n\trf.lastHeartBeat = time.Now()\n\trf.lastElection = time.Now()\n\trf.vote = -1\n}", "func (w *Window) Reset() {\n\tw.highest = 0\n\t// this is fine because higher blocks are cleared during Check()\n\tw.blocks[0] = 0\n}", "func Reset() string {\n\treturn csi(\"m\")\n}" ]
[ "0.61890984", "0.60074925", "0.5998274", "0.5847359", "0.5753193", "0.5735723", "0.5669589", "0.56392", "0.5606564", "0.5587392", "0.5555873", "0.55551696", "0.55442333", "0.550156", "0.54839456", "0.54712147", "0.5442829", "0.5439494", "0.5414043", "0.5409556", "0.5399395", "0.5383782", "0.5365399", "0.53583115", "0.5349352", "0.5347378", "0.53031236", "0.52984345", "0.5296025", "0.52884495", "0.5281995", "0.52719116", "0.525522", "0.5254747", "0.5231971", "0.521111", "0.5178473", "0.51774335", "0.516984", "0.51593214", "0.5156711", "0.51559037", "0.51389563", "0.5127372", "0.51192427", "0.5111568", "0.5101283", "0.5090413", "0.50880367", "0.5079992", "0.5079832", "0.507684", "0.50753546", "0.5073108", "0.50718445", "0.5061409", "0.5044037", "0.5034725", "0.5030166", "0.5022912", "0.5019557", "0.5017229", "0.5016622", "0.5016191", "0.50030184", "0.49928075", "0.4990893", "0.4980057", "0.4979305", "0.49771127", "0.49754313", "0.49633843", "0.4963343", "0.49567586", "0.49567586", "0.49567586", "0.49567586", "0.49567586", "0.49567586", "0.49567586", "0.49567586", "0.49556893", "0.49528968", "0.4952801", "0.49465436", "0.4938712", "0.49375966", "0.49341542", "0.49333543", "0.492955", "0.49293807", "0.49221754", "0.4922105", "0.492001", "0.49161577", "0.4913444", "0.49099994", "0.49099702", "0.49066544", "0.48975143" ]
0.7780016
0
Pause "Stops" time by recording current time and shortcircuit Now() to return this time instead of the actual time (plus skew).
func (c *chrono) Pause() { c.pausedAt = c.Now() c.paused = true }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}", "func stoppableNow() time.Time {\n\tstopMux.RLock()\n\tdefer stopMux.RUnlock()\n\n\tif stoppedAt != nil {\n\t\treturn *stoppedAt\n\t}\n\treturn time.Now().Add(-stoppedFor)\n}", "func (c *StoppedClock) Now() time.Time {\n\treturn c.t\n}", "func (c *StoppedClock) Now() time.Time {\n\treturn c.time\n}", "func (c *Clock) Now() time.Time {\n\treturn time.Now().UTC().Truncate(time.Second)\n}", "func Stop() {\n\tstopMux.Lock()\n\tif stoppedAt != nil {\n\t\tpanic(\"Time was already stopped\")\n\t}\n\tnow := Now()\n\tstoppedAt = &now\n\tstopMux.Unlock()\n}", "func (c *Clock) Now() time.Time { return time.Now() }", "func (u *Util) Now() time.Time {\n\tif controlDuration != 0 {\n\t\treturn time.Now().Add(controlDuration).UTC()\n\t}\n\treturn time.Now().UTC()\n}", "func (c stubClocker) Now() time.Time { return c.t }", "func (realClocker) Now() time.Time { return time.Now() }", "func (t *Time) Now() time.Time {\n\treturn t.current\n}", "func (c *ClockVal) Now() {\n\tc.CurrentTime = NowTime()\n}", "func Now() time.Time { return now() }", "func (c *FakeClock) Now() time.Time {\n\tc.steps++\n\treturn c.Time(c.steps)\n}", "func (c *RunningClock) Now() time.Time {\n\treturn time.Now()\n}", "func (d *dummyClock) Now() time.Time {\n\treturn time.Time{}\n}", "func Now() time.Time { return time.Now().UTC() }", "func Now() time.Time { return time.Now().UTC() }", "func now() time.Duration { return time.Since(startTime) }", "func (clock *Clock) Now() time.Time {\n\tvar now time.Time\n\tclock.do(func(s *state) {\n\t\tnow = s.t\n\t\ts.t = s.t.Add(clock.step)\n\t})\n\treturn now\n}", "func (p *PredefinedFake) Now() time.Time {\n\tadjustedTime := p.Base.Add(p.Delays[p.Next])\n\tp.Next++\n\treturn adjustedTime\n}", "func Now() time.Time { return time.Now() }", "func Now() time.Time {\n\treturn CurrentClock().Now()\n}", "func (Clock) Now() time.Time {\n\treturn time.Now()\n}", "func (c *Clock) Now() time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\n\tstep := c.step\n\tif c.skipStep {\n\t\tstep = 0\n\t\tc.skipStep = false\n\t}\n\tc.advanceLocked(rt, step)\n\n\treturn c.present\n}", "func (p PT) Now() int64 {\n\tif p.Seconds {\n\t\treturn time.Now().Unix()\n\t}\n\treturn time.Now().UnixNano()\n}", "func (s *Stopwatch) Stop() {\n\ts.Lock()\n\tdefer s.Unlock()\n\tif s.active() {\n\t\ts.stop = time.Now()\n\t}\n}", "func (t *DefaultClock) Now() time.Time {\n\treturn time.Now()\n}", "func (defaultClock) Now() time.Time {\n\treturn time.Now()\n}", "func Now() time.Time {\n\treturn Clock.Now()\n}", "func (fc *fakeClock) Now() time.Time {\n\treturn fc.time\n}", "func (tc *TestClock) Now() time.Time {\n\ttc.l.RLock()\n\tdefer tc.l.RUnlock()\n\treturn tc.t\n}", "func Now() Time {\n\treturn DefaultScheduler.Now()\n}", "func (c *CumulativeClock) Now() time.Time {\n\treturn c.current\n}", "func Now() Time {\n\treturn Time{format(time.Now())}\n}", "func Now(upToSecond ...bool) *TimeStamp {\n\treturn TimeFrom(time.Now(), upToSecond...)\n}", "func InstrumentNow() {\n\tNow = stoppableNow\n}", "func (c *MockClock) Now() time.Time {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\treturn c.now\n}", "func Now() *time.Time {\n\tt := time.Now()\n\treturn &t\n}", "func (t *Timer) Now() time.Duration {\n\treturn nowFunc().Sub(t.Zero)\n}", "func Now() Time {\n\treturn NewTime(time.Now())\n}", "func Now() time.Time {\n\treturn Work.Now()\n}", "func Now() time.Time {\n\treturn now()\n}", "func (tr *TextRegion) TimeNow() {\n\ttr.Time.Now()\n}", "func CurrentTime() Time {\n\tif atomic.LoadUint32(&isRunning) != 0 {\n\t\treturn currentTime\n\t}\n\n\treturn Time(time.Now().UnixNano() / 1e6)\n}", "func Now() time.Time {\n\tif IsTest() {\n\t\treturn now\n\t}\n\n\treturn time.Now()\n}", "func Now() time.Time {\n\treturn c.Now()\n}", "func (c *Context) Now() time.Time {\n\treturn c.currentTime\n}", "func (t Time) ResetNow() time.Time {\n\tnow = defaultNow\n\treturn now()\n}", "func Now() time.Time {\n\treturn time.Unix(0, time.Now().UnixNano()/1e6*1e6)\n}", "func (a *IncrementingFakeTimeSource) Now() time.Time {\n\tadjustedTime := a.BaseTime.Add(a.Increments[a.NextIncrement])\n\ta.NextIncrement++\n\n\treturn adjustedTime\n}", "func Now() Time {\n\treturn Time(time.Now().UnixNano() / 1000)\n}", "func (RealClock) Now() time.Time {\n\treturn time.Now()\n}", "func nowTime() time.Time {\n\treturn time.Now().UTC()\n}", "func resetTimeNow() {\n\ttimeNow = time.Now\n}", "func (wc WallClock) Now() time.Time {\n\treturn time.Now()\n}", "func Now() time.Time {\n\treturn time.Now().In(_defaultLocation)\n}", "func GetCurrentTime() uint64 {\n\treturn ComposeTSByTime(time.Now(), 0)\n}", "func Resume() {\n\tstopMux.Lock()\n\tif stoppedAt == nil {\n\t\tpanic(\"Time was not stopped\")\n\t}\n\tstoppedFor += time.Since(*stoppedAt)\n\tstoppedAt = nil\n\tstopMux.Unlock()\n}", "func ResetNow() { now = time.Now }", "func TimeNow() Time {\n\treturn Time{\n\t\ttime.Now(),\n\t}\n}", "func (s *Stopwatch) Stop() {\n\ts.Lock()\n\tdefer s.Unlock()\n\n\tif s.isRunning() {\n\t\ts.elapsedTime += time.Since(s.refTime)\n\t\ts.refTime = time.Time{}\n\t} else {\n\t\tfmt.Printf(\"WARNING: Stopwatch.Stop() isRunning is false\\n\")\n\t}\n}", "func Now() Time {\n\t/*\n\t\tft := windows.Filetime{}\n\t\twindows.GetSystemTimeAsFileTime(&ft)\n\t\treturn Time(ft.Nanoseconds() / nanosecondsScale)\n\t*/\n\t// On Windows, it turns out to be much faster to simply call time.Now().Unix(), so do it\n\treturn Time(time.Now().Unix())\n}", "func CurrentTime() time.Time {\n\treturn time.Now()\n}", "func Now() Time {\n\tvar t Time\n\tt.FromNSec(uint64(gotime.Now().UnixNano()))\n\treturn t\n}", "func now() time.Time {\n\treturn time.Now().UTC()\n}", "func Now() time.Time {\n\treturn time.Date(int(Year.Status().(uint16)), time.Month(Month.Status().(uint8)), int(Day.Status().(uint16)), int(Hour.Status().(uint8)), int(Minute.Status().(uint8)), int(Second.Status().(uint8)), 0, time.FixedZone(\"custom\", int(Tz.Status().(int8)) * 3600))\n}", "func GetCurrentTime() typeutil.Timestamp {\n\treturn ComposeTSByTime(time.Now(), 0)\n}", "func (s *Scheduler) now() Time {\n\treturn s.provider.Now().Add(s.timeOffset)\n}", "func (t *Tracker) Stop() {\n\tt.Finish = time.Now()\n\tt.Duration = time.Since(t.Run)\n}", "func (sit *SyncIntervalTimer) Pause() {\n\t<-sit.sync\n}", "func TimeNow() time.Time {\n\treturn time.Now().UTC()\n}", "func (m *timeSource) Now() mstime.Time {\n\treturn mstime.Now()\n}", "func (f FakeTimeSource) Now() time.Time {\n\treturn f.FakeTime\n}", "func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func (s SystemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func Now() time.Time {\n\tmu.Lock()\n\tdefer mu.Unlock()\n\tvar now time.Time\n\tif testMode {\n\t\tnow = testNow()\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.In(localtz.Get())\n}", "func (s systemTimeSource) Now() time.Time {\n\treturn time.Now()\n}", "func (t Time) GetNow() time.Time {\n\treturn now()\n}", "func SetNow(n time.Time) {\n\tnow = n\n}", "func Now() time.Time {\n\treturn time.Now().In(LOCATION)\n}", "func TestClock_Now(t *testing.T) {\n\tnow := raft.NewClock().Now()\n\tif exp := time.Now(); exp.Sub(now) > 1*time.Second {\n\t\tt.Fatalf(\"clock time is different than wall time: exp=%v, got=%v\", exp, now)\n\t}\n}", "func Now(ctx Context) time.Time {\n\treturn getWorkflowEnvironment(ctx).Now()\n}", "func SetTime(ct int) {\n\tcurrenttime = ct\n}", "func RestoreNow() {\n\tNow = time.Now\n}", "func (tr *TimedRun) Stop(t *Timer) {\n\tstop := currentTimeFunc()\n\ttr.cl.Lock()\n\tdefer tr.cl.Unlock()\n\tif _, ok := tr.categories[t.category]; !ok {\n\t\ttr.categories[t.category] = 0\n\t}\n\ttr.categories[t.category] += stop.Sub(t.startTime)\n}", "func (m *Mock) Now() time.Time {\n\tm.Lock()\n\tdefer m.Unlock()\n\treturn m.now\n}", "func (sw *Stopwatch) Reset() {\n\tsw.t = time.Now()\n}", "func (c *Timer) Pause() {\n\tc.ticker.Stop()\n\tc.passed += time.Now().Sub(c.lastTick)\n\tc.lastTick = time.Now()\n\tc.options.OnPaused()\n}", "func makeFakeNow() func() time.Time {\n\tvar t time.Time\n\treturn func() time.Time {\n\t\tt = t.Add(time.Second)\n\t\treturn t\n\t}\n}", "func (j *ScheduledJob) Pause() {\n\tj.rec.NextRun = time.Time{}\n\tj.markDirty(\"next_run\")\n}", "func (gores *Gores) CurrentTime() int64 {\n\ttimestamp := time.Now().Unix()\n\treturn timestamp\n}", "func newFakeTime() {\n\tfakeCurrentTime = fakeTime().Add(time.Hour * 24 * 2)\n}", "func Time() time.Time {\n\tnow := time.Now().UTC()\n\tdif := time.Duration(rand.Int())\n\tstart := now.Add(dif * -1)\n\tend := now.Add(dif)\n\treturn TimeSpan(start, end)\n}", "func NowTime() time.Time {\n\treturn ExtractTimeFromDatetime(time.Now())\n}", "func (f *FakeTimeSource) Now() time.Time {\n\tf.mu.RLock()\n\tdefer f.mu.RUnlock()\n\treturn f.now\n}", "func (tx *tx) SetNow(now time.Time) { tx.now = now }", "func StopCurrent() error {\n\tsession := toggl.OpenSession(viper.GetString(\"token\"))\n\taccount, err := session.GetAccount()\n\tif err != nil {\n\t\treturn err\n\t}\n\tvar timeEntry *toggl.TimeEntry\n\ttimeEntry, _ = getCurrentTimeEntry(account)\n\tif nil != timeEntry {\n\t\tsession.StopTimeEntry(*timeEntry)\n\t}\n\n\treturn nil\n}", "func Now() Timespec {\n\treturn NewTimespec(time.Now())\n}", "func (m *Mock) Now() time.Time {\n\tm.mu.Lock()\n\tdefer m.mu.Unlock()\n\treturn m.now\n}" ]
[ "0.7088481", "0.6919379", "0.67659324", "0.67019147", "0.6559164", "0.6402291", "0.63203806", "0.6272591", "0.6245372", "0.61367035", "0.61324173", "0.61246", "0.6115938", "0.60733736", "0.6040197", "0.60223395", "0.6020786", "0.6020786", "0.60146177", "0.6012215", "0.59966373", "0.5959776", "0.594986", "0.59412545", "0.5932818", "0.5914386", "0.5879223", "0.5872968", "0.58612317", "0.5852448", "0.58428824", "0.582453", "0.5794403", "0.579429", "0.5771547", "0.57653993", "0.57504296", "0.57231873", "0.56907177", "0.56901085", "0.56872255", "0.5681397", "0.56751794", "0.56665623", "0.5663568", "0.5654582", "0.5639455", "0.5635226", "0.5632688", "0.56285256", "0.5625493", "0.56074226", "0.5604195", "0.55978066", "0.5587286", "0.5576184", "0.5575942", "0.5544697", "0.5544119", "0.55364", "0.55070937", "0.5504631", "0.547006", "0.5465696", "0.54551595", "0.54522413", "0.5430223", "0.5428441", "0.54235643", "0.54144454", "0.5408864", "0.539088", "0.5390319", "0.5384731", "0.5373761", "0.5373761", "0.537357", "0.5373175", "0.5362128", "0.5352885", "0.5338776", "0.5330385", "0.5323998", "0.53135836", "0.5288172", "0.5284446", "0.52778125", "0.5271241", "0.52665097", "0.5262797", "0.5260879", "0.5251372", "0.5250993", "0.5248289", "0.52457446", "0.52367693", "0.523104", "0.5229294", "0.5227222", "0.5227182" ]
0.5683553
41
Continues time after having been paused. This has no effect if clock is already running.
func (c *chrono) Continue() { c.paused = false }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (c *chrono) Pause() {\n\tc.pausedAt = c.Now()\n\tc.paused = true\n}", "func (c *Timer) Pause() {\n\tc.ticker.Stop()\n\tc.passed += time.Now().Sub(c.lastTick)\n\tc.lastTick = time.Now()\n\tc.options.OnPaused()\n}", "func (sit *SyncIntervalTimer) Pause() {\n\t<-sit.sync\n}", "func (j *ScheduledJob) Pause() {\n\tj.rec.NextRun = time.Time{}\n\tj.markDirty(\"next_run\")\n}", "func (e *Emulator) Pause() {}", "func (c *Context) PAUSE() {\n\tc.addinstruction(x86.PAUSE())\n}", "func (clk *Clock) Continue() {\n\tclk.midiOut <- midiContinue\n}", "func (timer *WallclockTimer) Resume() error {\n\ttimer.command <- \"resume\"\n\treturn nil\n}", "func Resume() {\n\tstopMux.Lock()\n\tif stoppedAt == nil {\n\t\tpanic(\"Time was not stopped\")\n\t}\n\tstoppedFor += time.Since(*stoppedAt)\n\tstoppedAt = nil\n\tstopMux.Unlock()\n}", "func pause() {\n\ttime.Sleep(500 * time.Millisecond)\n}", "func PAUSE() { ctx.PAUSE() }", "func (s *sched) pause() {\n\told := atomic.LoadPointer(&s.timer)\n\t// if old is nil then there is someone who tries to stop the timer.\n\tif old != nil {\n\t\t(*time.Timer)(old).Stop()\n\t}\n}", "func Pause() {\n\tatomic.AddUint64(&sched0.pausing, 1)\n\tsched0.pause()\n}", "func (p *Peer) pause() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Pause()\n}", "func (p *Peer) pause() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Pause()\n}", "func (sit *SyncIntervalTimer) Resume() {\n\tsit.sync <- struct{}{}\n}", "func pause() tea.Msg {\n\ttime.Sleep(time.Millisecond * 600)\n\treturn DoneMsg{}\n}", "func Clock(ctx context.Context) {\n\tfor {\n\t\tselect {\n\t\tcase <- ctx.Done():\n\t\t\tfmt.Printf(\"\\r[ # ]\")\n\t\t\treturn\n\t\tdefault:\n\t\t\tfor _, c := range `-\\|/` {\n\t\t\t\tfmt.Printf(\"\\r[ %c ]\", c)\n\t\t\t\ttime.Sleep(200 * time.Millisecond)\n\t\t\t}\n\t\t}\n\t}\n}", "func (p *Pacer) Pause() {\n\tp.pause <- struct{}{} // block this channel\n\tp.paused <- struct{}{} // set flag to indicate paused\n}", "func (p *Peer) resume() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Reset()\n}", "func (p *Peer) resume() {\n\tp.mutex.Lock()\n\tdefer p.mutex.Unlock()\n\tp.heartbeatTimer.Reset()\n}", "func (p *ProgressMeter) Pause() {\n\tif atomic.CompareAndSwapInt32(&p.started, 1, 0) {\n\t\tp.finished <- true\n\t}\n}", "func (doDebugger Debugger) Pause() (err error) {\n\tb := debugger.Pause()\n\treturn b.Do(doDebugger.ctxWithExecutor)\n}", "func (cpu *Mos6502) Clock() {\n\t// if current instruction complete, read and execute next instruction\n\tif cpu.cycles == 0 {\n\t\topcode := cpu.read(cpu.pc)\n\t\tinstruction := cpu.lookup[opcode]\n\n\t\tcpu.setStatusFlag(U, true)\n\t\tcpu.pc++\n\n\t\tcpu.cycles = instruction.cycles\n\t\tadditionalCycleAddr := instruction.setAddressMode()\n\t\tadditionalCycleOp := instruction.performOp()\n\t\tcpu.cycles += additionalCycleAddr & additionalCycleOp\n\n\t\tcpu.setStatusFlag(U, true)\n\t}\n\n\tcpu.cycles--\n}", "func (t *task) Pause(_ context.Context) error {\n\tif t.ctr.ociSpec.Windows.HyperV == nil {\n\t\treturn cerrdefs.ErrNotImplemented\n\t}\n\n\tt.ctr.mu.Lock()\n\tdefer t.ctr.mu.Unlock()\n\n\tif err := t.assertIsCurrentTask(); err != nil {\n\t\treturn err\n\t}\n\tif t.ctr.hcsContainer == nil {\n\t\treturn errdefs.NotFound(errors.WithStack(fmt.Errorf(\"container %q not found\", t.ctr.id)))\n\t}\n\tif err := t.ctr.hcsContainer.Pause(); err != nil {\n\t\treturn err\n\t}\n\n\tt.ctr.isPaused = true\n\n\tt.ctr.client.eventQ.Append(t.ctr.id, func() {\n\t\terr := t.ctr.client.backend.ProcessEvent(t.ctr.id, libcontainerdtypes.EventPaused, libcontainerdtypes.EventInfo{\n\t\t\tContainerID: t.ctr.id,\n\t\t\tProcessID: t.id,\n\t\t})\n\t\tt.ctr.client.logger.WithFields(log.Fields{\n\t\t\t\"container\": t.ctr.id,\n\t\t\t\"event\": libcontainerdtypes.EventPaused,\n\t\t}).Info(\"sending event\")\n\t\tif err != nil {\n\t\t\tt.ctr.client.logger.WithError(err).WithFields(log.Fields{\n\t\t\t\t\"container\": t.ctr.id,\n\t\t\t\t\"event\": libcontainerdtypes.EventPaused,\n\t\t\t}).Error(\"failed to process event\")\n\t\t}\n\t})\n\n\treturn nil\n}", "func (helper *testHelper) Pause(secs int) {}", "func (o *OMXPlayer) Pause() error {\n\tif o.status == statusPaused {\n\t\treturn nil\n\t}\n\n\t// TODO: test this properly by mocking dbus package.\n\tif o.testing {\n\t\to.status = statusPaused\n\t\treturn nil\n\t}\n\n\t_, err := o.dbusSend(\"Pause\", dbus.FlagNoAutoStart)\n\tif err != nil {\n\t\treturn err\n\t}\n\to.status = statusPaused\n\treturn nil\n}", "func (c *Clock) AdvanceTo(t time.Time) {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\tc.realTime = rt\n\tc.present = t\n\tc.events.AdvanceTo(c.present)\n}", "func (s *Scheduler) Pause() error {\n\ts.mu.Lock()\n\tdefer s.mu.Unlock()\n\ts.paused = true\n\treturn nil\n}", "func (c *Clock) Advance(d time.Duration) time.Time {\n\tc.init()\n\trt := c.maybeGetRealTime()\n\n\tc.mu.Lock()\n\tdefer c.mu.Unlock()\n\tc.skipStep = true\n\n\tc.advanceLocked(rt, d)\n\treturn c.present\n}", "func (e *Enumeration) Pause() {\n\te.pause <- struct{}{}\n}", "func (e *Enumeration) Pause() {\n\te.pause <- struct{}{}\n}", "func (ft *fakeTimer) advance(d time.Duration) {\n\tft.lock.Lock()\n\tdefer ft.lock.Unlock()\n\n\tft.now = ft.now.Add(d)\n\tif ft.active && !ft.now.Before(ft.timeout) {\n\t\tft.active = false\n\t\tft.c <- ft.timeout\n\t}\n}", "func (timer *Timer) Pause() error {\n\tch := make(chan error)\n\ttimer.pause <- ch\n\treturn <-ch\n}", "func (c *Switch) Pause() {\n\tc.lock.RLock()\n\tif cp, ok := c.subRenderables[c.curRenderable].(CanPause); ok {\n\t\tcp.Pause()\n\t}\n\tc.lock.RUnlock()\n}", "func (_Cakevault *CakevaultTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Cakevault.contract.Transact(opts, \"pause\")\n}", "func (w *Walker) Pause() {\n\tw.pauseRequested = true\n}", "func (p *Pacer) Resume() {\n\t<-p.paused // clear flag to indicate paused\n\t<-p.pause // unblock this channel\n}", "func (p *Playlist) Pause() {\n\tp.ch <- \"pause\"\n}", "func (rep *Reporter) Continue() {\n\tlock.Lock()\n\tdefer lock.Unlock()\n\tif rep == nil {\n\t\treturn\n\t}\n\n\tif !rep.metricsCh.IsPaused() {\n\t\treturn\n\t}\n\n\trep.metricsCh.Continue()\n}", "func AdvanceTo(newTime time.Time) {\n\tnow := Now()\n\tfound := false\n\tfor _, s := range sortedSchedulers() {\n\t\tnextTick := s.tickAfter(now)\n\t\tif nextTick.After(now) && !nextTick.After(newTime) {\n\t\t\tfound = true\n\t\t\tnowInTest.Store(nextTick)\n\t\t\ts.Lock()\n\t\t\tif s.paused {\n\t\t\t\ts.fireOnResume = true\n\t\t\t} else {\n\t\t\t\ts.notifyFn()\n\t\t\t}\n\t\t\ts.Unlock()\n\t\t}\n\t}\n\tif !found {\n\t\tnowInTest.Store(newTime)\n\t\treturn\n\t}\n\tif Now() != newTime {\n\t\tAdvanceTo(newTime)\n\t}\n}", "func (j Jibi) Pause() {\n\tj.RunCommand(CmdPause, nil)\n}", "func (c *StoppedClock) SetTime(time time.Time) {\n\tc.time = time\n}", "func main() {\n var working float64\n var relax float64\n var switchact bool = false\n var switch2relax bool = true\n var switch2working bool = false\n\n fmt.Println(\"Type any key for start the Workday\")\n fmt.Println(\"The counter starts with relaxing type any key for start working\")\n fmt.Println(\"You start workint at \" ,time.Now())\n ch := make(chan string)\n go func(ch chan string) {\n // disable input buffering\n exec.Command(\"stty\", \"-F\", \"/dev/tty\", \"cbreak\", \"min\", \"1\").Run()\n // do not display entered characters on the screen\n exec.Command(\"stty\", \"-F\", \"/dev/tty\", \"-echo\").Run()\n var b []byte = make([]byte, 1)\n //Routine for check for standard input with no wait for a key\n for {\n os.Stdin.Read(b)\n ch <- string(b)\n }\n }(ch)\n\n //The intention with this fork was do a Thiker and with this controll the flow of time\n for {\n select {\n \n //In case of a signal by key lanch this case, change status of working to relax\n //and show time elapsed\n case stdin, _ := <-ch:\n fmt.Println(\"\\n \\n Keys pressed:\", stdin)\n if switch2relax == true {\n fmt.Println(\"\\n Now you are relaxing\\n \\n\")\n }else {\n fmt.Println(\"\\n Now you are working \\n \\n\")\n }\n fmt.Printf(\"working for %f Hours \\n\", ((working/60.0)/60.0))\n fmt.Printf(\"Relaxing for %f Hours \\n\", (relax/60.0)/60.0)\n\n switch2relax = !switch2relax\n switch2working = !switch2working\n switchact = true\n \n //By default the program is count time, if the routin is not counting time\n //working count time relaxing, this script has no pause\n default:\n if switchact == true {\n if switch2relax == true {\n working=1.0+working\n if ((working/60.0)/60.0) == 8.00000 {\n exec.Command(\"mplayer\", \"stopwork.mp3\").Run()\n }\n //fmt.Println(\"Working for \", working)\n }else{\n relax=1.0+relax\n //fmt.Println(\"relaxing for \", relax)\n }\n }\n }\n time.Sleep(time.Second * 1)\n }\n}", "func (_Cakevault *CakevaultSession) Pause() (*types.Transaction, error) {\n\treturn _Cakevault.Contract.Pause(&_Cakevault.TransactOpts)\n}", "func (p *Profile) Pause() error {\n\tif !p.started {\n\t\treturn errors.New(\"mxnet profile was not started\")\n\t}\n\tif p.stopped == true || p.paused == true {\n\t\treturn nil\n\t}\n\tdefer func() {\n\t\tp.paused = true\n\t}()\n\tp.lastPauseTime = time.Now()\n\tsuccess, err := C.MXProfilePause(C.int(1))\n\tif err != nil {\n\t\treturn err\n\t}\n\tif success != 0 {\n\t\treturn GetLastError()\n\t}\n\n\treturn nil\n}", "func (m *Machine) Pause() error {\n\tm.State = driver.Paused\n\tfmt.Printf(\"Pause %s: %s\\n\", m.Name, m.State)\n\treturn nil\n}", "func (t Track) Pause(bool) {\n\tpanic(\"implement me\")\n}", "func (_Cakevault *CakevaultTransactorSession) Pause() (*types.Transaction, error) {\n\treturn _Cakevault.Contract.Pause(&_Cakevault.TransactOpts)\n}", "func (c *chrono) Now() time.Time {\n\tvar now time.Time\n\tif c.paused {\n\t\tnow = c.pausedAt\n\t} else {\n\t\tnow = time.Now()\n\t}\n\treturn now.Add(c.skew)\n}", "func (s *SlaveHealthChecker) Pause() {\n\ts.Lock()\n\tdefer s.Unlock()\n\ts.paused = true\n}", "func (_OracleMgr *OracleMgrTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _OracleMgr.contract.Transact(opts, \"pause\")\n}", "func (_ERC20Pausable *ERC20PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _ERC20Pausable.contract.Transact(opts, \"pause\")\n}", "func Pause(msg string) {\n\tfmt.Print(msg, \"[Press Enter to Continue]: \")\n\tvar s string\n\t_, err := fmt.Scan(&s)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n}", "func (n *BaseNode) Continue() {\n\tn.continueFunc(func() {\n\t\tif n.cancelPause != nil {\n\t\t\tn.cancelPause()\n\t\t\tn.cancelPause = nil\n\t\t}\n\t})\n}", "func (d *Device) Pause() {\n\tdefer d.mu.Unlock()\n\td.mu.Lock()\n\td.paused = true\n}", "func (_DelegateProfile *DelegateProfileTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _DelegateProfile.contract.Transact(opts, \"pause\")\n}", "func UpdateTime(dt int) {\n\tcurrenttime = currenttime + dt\n\tif currenttime >= 2400 {\n\t\tcurrenttime = currenttime - 2400\n\t\tdaycount = daycount + 1\n\t}\n}", "func (a *Agent) PauseSync() {\n\t// Do this outside of lock as it has it's own locking\n\ta.sync.Pause()\n\n\t// Coordinate local state watchers\n\ta.syncMu.Lock()\n\tdefer a.syncMu.Unlock()\n\tif a.syncCh == nil {\n\t\ta.syncCh = make(chan struct{})\n\t}\n}", "func (c *MockClock) Set(t time.Time) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tif t.After(c.now) {\n\t\tc.advanceTo(t)\n\t} else {\n\t\tc.now = t // move back in time\n\t}\n}", "func (_MonsterAccessControl *MonsterAccessControlTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _MonsterAccessControl.contract.Transact(opts, \"pause\")\n}", "func (p *Proxy) Pause() {\n\tp.configMutex.Lock()\n\tif p.state == proxyActive {\n\t\tp.state = proxyPaused\n\t\tp.connMutex.Unlock()\n\t}\n\tp.configMutex.Unlock()\n}", "func (s *ExperimentSpec) Pause() bool {\n\tif s.ManualOverride != nil && s.ManualOverride.Action == ActionPause {\n\t\treturn true\n\t}\n\treturn false\n}", "func (c *MockClock) advanceToNextWakeup() {\n\tif len(c.wakeups) < 1 {\n\t\treturn\n\t}\n\tw := heap.Pop(&c.wakeups).(*wakeup)\n\tc.logf(\"MockClock: Advancing time from now=%s to next wakeup time %s.\",\n\t\ttsStr(c.now), tsStr(w.t))\n\tc.now = w.t\n\tselect {\n\tcase w.c <- w.t:\n\tdefault:\n\t}\n\t// give things a chance to run\n\truntime.Gosched()\n\tc.logf(\"MockClock: Advanced time, now=%s.\", tsStr(c.now))\n}", "func (c *MockClock) advanceTo(end time.Time) {\n\tfor {\n\t\tt := c.nextWakeupTime()\n\t\tif t == (time.Time{}) || t.After(end) {\n\t\t\tbreak\n\t\t}\n\t\tc.advanceToNextWakeup()\n\t}\n\tc.now = end\n}", "func SetTime(ct int) {\n\tcurrenttime = ct\n}", "func (cg *CgroupImpl) Pause() error {\n\treturn cg.setFreezeState(1)\n}", "func (l *LightningLoader) Pause() {\n\tl.Lock()\n\tdefer l.Unlock()\n\tif l.isClosed() {\n\t\tl.logger.Warn(\"try to pause, but already closed\")\n\t\treturn\n\t}\n\tif l.cancel != nil {\n\t\tl.cancel()\n\t}\n\tl.core.Stop()\n}", "func AnounanceTime(){\n\tfor {\n\t\tfmt.Println(time.Now())\n\t\ttime.Sleep(5*time.Second)\n\t}\n}", "func (_Token *TokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Token.contract.Transact(opts, \"pause\")\n}", "func (_ChpRegistry *ChpRegistryTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _ChpRegistry.contract.Transact(opts, \"pause\")\n}", "func (_Cakevault *CakevaultCaller) Paused(opts *bind.CallOpts) (bool, error) {\n\tvar out []interface{}\n\terr := _Cakevault.contract.Call(opts, &out, \"paused\")\n\n\tif err != nil {\n\t\treturn *new(bool), err\n\t}\n\n\tout0 := *abi.ConvertType(out[0], new(bool)).(*bool)\n\n\treturn out0, err\n\n}", "func (_ElvToken *ElvTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _ElvToken.contract.Transact(opts, \"pause\")\n}", "func (t Time) Clock() (hour, min, sec int) {}", "func (_MonsterOwnership *MonsterOwnershipTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _MonsterOwnership.contract.Transact(opts, \"pause\")\n}", "func (c *Compound) Pause() {\n\tc.lock.RLock()\n\tif cp, ok := c.subRenderables[c.curRenderable].(CanPause); ok {\n\t\tcp.Pause()\n\t}\n\tc.lock.RUnlock()\n}", "func Pause() {\n\tNamedLoggers.Range(func(key, value interface{}) bool {\n\t\tlogger := value.(*Logger)\n\t\tlogger.Pause()\n\t\tlogger.Flush()\n\t\treturn true\n\t})\n}", "func (d *Debugger) Continue() {\n\tif d.BreakpointActive {\n\t\td.Cont <- true\n\t}\n}", "func (_LifToken *LifTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _LifToken.contract.Transact(opts, \"pause\")\n}", "func (_Pausable *PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Pausable.contract.Transact(opts, \"pause\")\n}", "func (_Pausable *PausableTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Pausable.contract.Transact(opts, \"pause\")\n}", "func (c *MockClock) Advance(delta time.Duration) {\n\tc.mutex.Lock()\n\tdefer c.mutex.Unlock()\n\n\tend := c.now.Add(delta)\n\tc.advanceTo(end)\n}", "func (cr *Playback) Pause() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tcurrentParty, err := session.CurrentParty(c)\n\t\tif err != nil {\n\t\t\tif err != sql.ErrNoRows {\n\t\t\t\tc.Error(e.Internal.CausedBy(err))\n\t\t\t\tc.Abort()\n\t\t\t\treturn\n\t\t\t}\n\n\t\t\tc.Error(e.NotFound.WithMessage(\"Host has not started a party\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\t// It is a bad request to pause when no track is playing\n\t\tif !currentParty.CurrentTrack.Valid {\n\t\t\tc.Error(e.BadRequest.WithDetail(\"Host is not playing music\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\t// Pause the current track\n\t\tvar currentTrack models.PlayingTrack\n\t\terr = currentParty.CurrentTrack.Unmarshal(currentTrack)\n\t\tif err != nil {\n\t\t\tc.Error(e.Internal.CausedBy(err))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\telapsedParam, exists := c.GetQuery(\"elapsed\")\n\t\tif !exists {\n\t\t\tc.Error(e.BadRequest.WithDetail(\"Must provide elapsed duration to pause playback\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\t\telapsed, err := strconv.ParseUint(elapsedParam, 10, 64)\n\t\tif err != nil {\n\t\t\tc.Error(\n\t\t\t\te.BadRequest.\n\t\t\t\t\tWithDetail(\"Must provide valid elapsed duration to pause playback\").\n\t\t\t\t\tCausedBy(err),\n\t\t\t)\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\t// It is a bad request to pause a track that is already paused\n\t\tif currentTrack.Paused {\n\t\t\tc.Error(e.BadRequest.WithDetail(\"Current track is already paused\"))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\tcurrentTrack.Paused = true\n\t\tcurrentTrack.Elapsed = uint(elapsed)\n\t\terr = cr.updateTrackAndBroadcast(currentParty, &currentTrack)\n\t\tif err != nil {\n\t\t\tc.Error(e.Internal.CausedBy(err))\n\t\t\tc.Abort()\n\t\t\treturn\n\t\t}\n\n\t\tc.JSON(http.StatusOK, models.EmptyRespose)\n\t}\n}", "func (_MonsterAccessControl *MonsterAccessControlTransactorSession) Pause() (*types.Transaction, error) {\n\treturn _MonsterAccessControl.Contract.Pause(&_MonsterAccessControl.TransactOpts)\n}", "func (t *Time) Sleep(s time.Duration) {\n\tt.current = t.current.Add(s)\n}", "func main() {\n\tp := fmt.Println\n\t\n\tp(\"------- Alarm Clock - Main initialized -------\")\n\tvar wg sync.WaitGroup\n\t\n\n\t//start := time.Now()\n\tcurrent := time.Now()\n\tfmt.Print(\"Klockan är: \", current.Hour(),\":\", current.Minute(),\".\",current.Second(),\"\\n\")\n\tp(\"---------------------------------\")\n\t//diff := current.Sub(start)\n\t//fmt.Println(\"DIFF:\",diff)\n\tp(\"START TIME:\",current)\n\t\n\t//add one to waitgroup object for each new goroutine started\n\twg.Add(1)\n\tgo Remind(\"Dags att äta\",3*time.Second)\n\twg.Add(1)\n\tgo Remind(\"Dags att arbeta\",8*time.Second)\n\twg.Add(1)\n\tgo Remind(\"Dags att sova\",24*time.Second)\n\t\n\twg.Wait()\n\t//p(currTime.Format(\"15:04:05 (2006-01-02)\"))\n\t//p(\"TIME DIFF: \",currTime.Sub(startTime))\n\t\n\n}", "func (_DelegateProfile *DelegateProfileTransactorSession) Pause() (*types.Transaction, error) {\n\treturn _DelegateProfile.Contract.Pause(&_DelegateProfile.TransactOpts)\n}", "func (c *Client) Pause() error {\n\treturn nil\n}", "func TestProgressResume(t *testing.T) {\n\tp := &Progress{\n\t\tNext: 2,\n\t\tPaused: true,\n\t}\n\tp.maybeDecrTo(1, 1)\n\tif p.Paused {\n\t\tt.Errorf(\"paused= %v, want false\", p.Paused)\n\t}\n\tp.Paused = true\n\tp.maybeUpdate(2)\n\tif p.Paused {\n\t\tt.Errorf(\"paused= %v, want false\", p.Paused)\n\t}\n}", "func (n *BaseNode) Pause() {\n\tn.pauseFunc(func() {\n\t\tn.ctxPause, n.cancelPause = context.WithCancel(n.ctx)\n\t})\n}", "func (container *container) Pause() error {\r\n\treturn convertSystemError(container.system.Pause(context.Background()), container)\r\n}", "func NextPause() func() (pause time.Duration) {\n rando := rand.New(rand.NewSource(time.Now().UnixNano()))\n return func() (pause time.Duration) {\n pause = time.Duration((rando.Intn(500)) * int(time.Millisecond))\n return\n }\n}", "func (_PausableToken *PausableTokenTransactor) Pause(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _PausableToken.contract.Transact(opts, \"pause\")\n}", "func (o RunnerOutput) Paused() pulumi.BoolOutput {\n\treturn o.ApplyT(func(v *Runner) pulumi.BoolOutput { return v.Paused }).(pulumi.BoolOutput)\n}", "func (t *SpotifyDevice) Pause() {\n\tif t.client == nil {\n\t\treturn\n\t}\n\tlog.Println(\"Pausing Spotify device\")\n\tif err := t.client.PauseOpt(t.playOpts); err != nil {\n\t\tlog.Println(err)\n\t}\n}", "func (_MonsterAccessControl *MonsterAccessControlSession) Pause() (*types.Transaction, error) {\n\treturn _MonsterAccessControl.Contract.Pause(&_MonsterAccessControl.TransactOpts)\n}", "func (st *LevelCompleteState) OnPause(world w.World) {}", "func (s *SlaveHealthChecker) Continue(slaveUPID *upid.UPID) {\n\ts.Lock()\n\tdefer s.Unlock()\n\ts.paused = false\n\ts.slaveUPID = slaveUPID\n}", "func (_DelegateProfile *DelegateProfileSession) Pause() (*types.Transaction, error) {\n\treturn _DelegateProfile.Contract.Pause(&_DelegateProfile.TransactOpts)\n}", "func (fgs *FakeGraphSync) Pause(ctx context.Context, requestID graphsync.RequestID) error {\n\tfgs.pauses <- requestID\n\treturn fgs.ReturnedPauseError\n}" ]
[ "0.67877674", "0.653856", "0.6279754", "0.6165107", "0.6063943", "0.6050643", "0.6011128", "0.6008339", "0.5997136", "0.59503627", "0.59455216", "0.5916581", "0.58829933", "0.5872751", "0.5872751", "0.58583057", "0.57535905", "0.57200044", "0.56405187", "0.5613111", "0.5613111", "0.56104696", "0.56087905", "0.55646914", "0.5563314", "0.550896", "0.55066174", "0.54984736", "0.5495456", "0.5481841", "0.5450739", "0.5450739", "0.5447858", "0.54471993", "0.54377466", "0.5433592", "0.5430805", "0.53919566", "0.53906846", "0.5390298", "0.53871286", "0.53265625", "0.53138626", "0.52900684", "0.5288507", "0.52870095", "0.5281264", "0.52765656", "0.5261484", "0.52584726", "0.5239021", "0.5234105", "0.5219808", "0.5209068", "0.52085245", "0.51810974", "0.51807606", "0.51799476", "0.5178544", "0.51756287", "0.5173574", "0.5166772", "0.5154566", "0.51509416", "0.5147486", "0.51215285", "0.510734", "0.51034176", "0.5097587", "0.50900215", "0.50880927", "0.5082865", "0.50586", "0.505711", "0.5044188", "0.5040813", "0.50352293", "0.50329983", "0.5029847", "0.50236815", "0.50236815", "0.5004772", "0.49994218", "0.49925694", "0.49838725", "0.49754855", "0.49745432", "0.49652076", "0.49650443", "0.4963122", "0.49604392", "0.49596387", "0.49585953", "0.49580398", "0.49558106", "0.49533692", "0.49461812", "0.49450165", "0.49448425", "0.49348" ]
0.7427754
0
ValidConfig will return true if there are values set for each Property of the Azure config object
func (a *Azure) ValidConfig() bool { return a.SubscriptionID != "" && a.ClientID != "" && a.ClientSecret != "" && a.TenantID != "" }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (g *Google) ValidConfig() bool {\n\treturn g.AccountFileJSON != \"\" && g.ProjectID != \"\" && g.Zone != \"\"\n}", "func (c *Config) Valid() error {\n\tif len(c.Servers) == 0 {\n\t\treturn fmt.Errorf(\"No servers in config\")\n\t}\n\tfor _, s := range c.Servers {\n\t\terr := s.Valid()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Config) IsValid() bool {\n\tif len(c.GithubAccessToken) != 40 {\n\t\treturn false\n\t}\n\tif len(c.GitName) == 0 {\n\t\treturn false\n\t}\n\tif len(c.GitEmail) == 0 {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (config *Config) IsValid() bool {\n\tif config.Basic.Enabled && config.OAuth2.Enabled {\n\t\treturn false\n\t}\n\n\tif config.Basic.Enabled && config.X509.Enabled {\n\t\treturn false\n\t}\n\n\tif config.OAuth2.Enabled && config.X509.Enabled {\n\t\treturn false\n\t}\n\n\tif !config.Basic.Enabled && !config.OAuth2.Enabled && !config.X509.Enabled {\n\t\treturn false\n\t}\n\n\treturn true\n}", "func (sc *StartupConfig) Validate() (bool, error) {\n\tif sc.Meta == nil {\n\t\treturn false, fmt.Errorf(\"meta object is nil\")\n\t}\n\n\tif sc.Server == nil {\n\t\t//return false, fmt.Errorf(\"Server is nil\")\n\t}\n\n\tif sc.Database == nil {\n\t\treturn false, fmt.Errorf(\"database object is nil\")\n\t}\n\n\tif sc.Session == nil {\n\t\treturn false, fmt.Errorf(\"session object is nil\")\n\t}\n\n\tif sc.Crypto == nil {\n\t\treturn false, fmt.Errorf(\"crypto object is nil\")\n\t}\n\n\tif sc.Secrets == nil {\n\t\treturn false, fmt.Errorf(\"secrets object is nil\")\n\t}\n\n\tif sc.Capsul == nil {\n\t\treturn false, fmt.Errorf(\"capsul object is nil\")\n\t}\n\n\tif sc.CustomCapsul == nil {\n\t\treturn false, fmt.Errorf(\"custom capsul object is nil\")\n\t}\n\n\tif sc.Create == nil {\n\t\treturn false, fmt.Errorf(\"create object is nil\")\n\t}\n\n\treturn true, nil\n}", "func (o *Config) IsValid() *AppError {\n\tif err := o.ServiceSettings.isValid(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (c *configuration) validateConfig() error {\n\tif c.Provider == \"\" {\n\t\treturn errors.New(msgConfigNoProvider)\n\t}\n\n\tif len(c.Servers) == 0 {\n\t\treturn errors.New(msgConfigNoServers)\n\t}\n\n\tfor i, srv := range c.Servers {\n\t\tif srv.FabricIface == \"\" {\n\t\t\treturn errors.Errorf(\n\t\t\t\tmsgConfigServerNoIface+\" for I/O service %d\", i)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c Config) IsValid() bool {\n\tif c.Port == \"\" || c.DSN == \"\" {\n\t\treturn false\n\t}\n\treturn true\n}", "func (c *Config) IsValid() error {\n\tif len(c.NetworkServices) == 0 {\n\t\treturn errors.New(\"no network services are specified\")\n\t}\n\tif c.Name == \"\" {\n\t\treturn errors.New(\"no cleint name specified\")\n\t}\n\tif c.ConnectTo.String() == \"\" {\n\t\treturn errors.New(\"no NSMGr ConnectTO URL are sepecified\")\n\t}\n\treturn nil\n}", "func (c *Config) IsValid() bool {\n\tif c.DiffType == \"schema\" || c.DiffType == \"data\" {\n\t\treturn c.Source != \"\" && c.Target != \"\"\n\t} else if c.DiffType == \"md\" || c.DiffType == \"wiki\" || c.DiffType == \"sql\" {\n\t\treturn c.Source != \"\"\n\t}\n\treturn false\n}", "func (c Config) Validate() error {\n\tif len(c.Project) == 0 {\n\t\treturn errors.New(\"missing project in configuration\")\n\t}\n\tif len(c.Bucket) == 0 {\n\t\treturn errors.New(\"missing bucket in configuration\")\n\t}\n\tif len(c.LastMigrationObjectName) == 0 {\n\t\treturn errors.New(\"missing state name in configuration\")\n\t}\n\treturn nil\n}", "func (c Config) IsValid() bool {\n\treturn c.RandomTemperature > 0 && c.NumSimulation > 0\n}", "func (c *Config) Validate() (err error) {\n\tcon := *c\n\tfor i, v := range con {\n\t\tk, err := registry.OpenKey(v.GetScope(), v.Path, registry.ALL_ACCESS)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tdefer k.Close()\n\t\tfor n, p := range v.Properties {\n\t\t\tval := &con[i].Properties[n].PrevValue\n\t\t\tswitch p.Type {\n\t\t\tcase \"DWord\", \"QWord\":\n\t\t\t\tif s, _, err := k.GetIntegerValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tcase \"String\":\n\t\t\t\tif s, _, err := k.GetStringValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tcase \"Strings\":\n\t\t\t\tif s, _, err := k.GetStringsValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tcase \"Binary\":\n\t\t\t\tif s, _, err := k.GetBinaryValue(p.Name); err == nil {\n\t\t\t\t\t*val = s\n\t\t\t\t} else {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\tdefault:\n\t\t\t\tvar buf []byte\n\t\t\t\tif _, _, err := k.GetValue(p.Name, buf); err != nil {\n\t\t\t\t\treturn err\n\t\t\t\t}\n\t\t\t\treturn fmt.Errorf(\"%s of %s path in %s scope returned code %d.\") // TODO: Convert const int representation of value types to explicitly match what the user should type into their JSON config.\n\t\t\t}\n\t\t}\n\t}\n\treturn nil\n}", "func (c AuthConfig) Validate() []error {\n\tvar errs []error\n\n\tif len(c.JwksURI) == 0 {\n\t\terrs = append(errs, errors.Errorf(\"AuthConfig requires a non-empty JwksURI config value\"))\n\t}\n\n\treturn errs\n}", "func ValidateConfig(configMap *corev1.ConfigMap) (*kle.Config, error) {\n\tconfig, err := kle.NewConfigFromMap(configMap.Data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tfor component := range config.EnabledComponents {\n\t\tif !validComponents.Has(component) {\n\t\t\treturn nil, fmt.Errorf(\"invalid enabledComponent %q: valid values are %q\", component, validComponents.List())\n\t\t}\n\t}\n\n\treturn config, nil\n}", "func (c *Config) IsValid() bool {\n\t_, latErr := strconv.ParseFloat(c.Latitude, 10)\n\t_, lonErr := strconv.ParseFloat(c.Longitude, 10)\n\treturn c.APIKey != \"\" && c.BaseUrl != \"\" && latErr == nil && lonErr == nil\n}", "func (c *config) Validate() []error {\n\tvar errs []error\n\tif value, err := validateProvider(c.Provider); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.Provider = value\n\t}\n\tif value, err := validateAccountName(c.AccountName); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.AccountName = value\n\t}\n\tif value, err := validateAccountSecret(c.AccountSecret); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.AccountSecret = value\n\t}\n\tif value, err := validateDNSContent(c.DNSContent); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.DNSContent = value\n\t}\n\tif value, err := validateDockerLabel(c.DockerLabel); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.DockerLabel = value\n\t}\n\tif value, err := validateStore(c.Store); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.Store = value\n\t}\n\tif value, err := validateDataDirectory(c.DataDirectory); err != nil {\n\t\terrs = append(errs, err)\n\t} else {\n\t\tc.DataDirectory = value\n\t}\n\treturn errs\n}", "func (m *Manager) IsDBConfigValid(config config.DatabaseConfigs) error {\n\tm.lock.RLock()\n\tdefer m.lock.RUnlock()\n\n\t// Only count the length of enabled databases\n\tvar length int\n\tfor _, c := range config {\n\t\tif c.Enabled {\n\t\t\tlength++\n\t\t}\n\t}\n\n\treturn nil\n}", "func (sc *StoreConfig) Valid() bool {\n\treturn sc.Clock != nil && sc.Transport != nil &&\n\t\tsc.RaftTickInterval != 0 && sc.RaftHeartbeatIntervalTicks > 0 &&\n\t\tsc.RaftElectionTimeoutTicks > 0 && sc.ScanInterval >= 0 &&\n\t\tsc.AmbientCtx.Tracer != nil\n}", "func (c *Config) Validate() error {\n\tvar errs error\n\tif c.Tenant.RemoteServiceAPI == \"\" {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.remote_service_api is required\"))\n\t}\n\tif c.Tenant.InternalAPI == \"\" && c.Analytics.FluentdEndpoint == \"\" && c.Analytics.CredentialsJSON == nil {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.internal_api or tenant.analytics.fluentd_endpoint is required if no service account\"))\n\t}\n\tif c.Tenant.OrgName == \"\" {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.org_name is required\"))\n\t}\n\tif c.Tenant.EnvName == \"\" {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"tenant.env_name is required\"))\n\t}\n\tif (c.Global.TLS.CertFile != \"\" || c.Global.TLS.KeyFile != \"\") &&\n\t\t(c.Global.TLS.CertFile == \"\" || c.Global.TLS.KeyFile == \"\") {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"global.tls.cert_file and global.tls.key_file are both required if either are present\"))\n\t}\n\tif (c.Analytics.TLS.CAFile != \"\" || c.Analytics.TLS.CertFile != \"\" || c.Analytics.TLS.KeyFile != \"\") &&\n\t\t(c.Analytics.TLS.CAFile == \"\" || c.Analytics.TLS.CertFile == \"\" || c.Analytics.TLS.KeyFile == \"\") {\n\t\terrs = multierror.Append(errs, fmt.Errorf(\"all analytics.tls options are required if any are present\"))\n\t}\n\treturn errs\n}", "func (c Config) Validate() error {\n\t// Check DbConfig\n\tif err := c.Db.Validate(); err != nil {\n\t\treturn fmt.Errorf(\"error verifying db config: %s\", err.Error())\n\t}\n\n\t// Check AuthConfig\n\tif err := c.Auth.Validate(); err != nil {\n\t\treturn fmt.Errorf(\"error verifying auth config: %s\", err.Error())\n\t}\n\n\t// All good\n\treturn nil\n}", "func (c *Config) Validate() error {\n\n\tif err := c.Data.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\t//if err := c.HintedHandoff.Validate(); err != nil {\n\t//\treturn err\n\t//}\n\tfor _, graphite := range c.GraphiteInputs {\n\t\tif err := graphite.Validate(); err != nil {\n\t\t\treturn fmt.Errorf(\"invalid graphite config: %v\", err)\n\t\t}\n\t}\n\n\tif err := c.Monitor.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.ContinuousQuery.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.Retention.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.Precreator.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tif err := c.Subscriber.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tfor _, collectd := range c.CollectdInputs {\n\t\tif err := collectd.Validate(); err != nil {\n\t\t\treturn fmt.Errorf(\"invalid collectd config: %v\", err)\n\t\t}\n\t}\n\n\tif err := c.TLS.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func IsConfigPresent(resources []ResourceConfig, resourceConfig ResourceConfig) bool {\n\tfor _, resource := range resources {\n\t\tif resource.Name == resourceConfig.Name && resource.Source == resourceConfig.Source {\n\t\t\tif reflect.DeepEqual(resource.Config, resourceConfig.Config) {\n\t\t\t\treturn true\n\t\t\t}\n\t\t}\n\t}\n\treturn false\n}", "func (config *Config) Validate() error {\n\tif len(config.NsqLookupdAddress) == 0 {\n\t\treturn fmt.Errorf(\"parameter NsqLookupdAddress missing\")\n\t}\n\tif len(config.NsqdAddress) == 0 {\n\t\treturn fmt.Errorf(\"parameter NsqdAddress missing\")\n\t}\n\tif len(config.BotName) == 0 {\n\t\treturn fmt.Errorf(\"parameter BotName missing\")\n\t}\n\tif len(config.BambooUrl) == 0 {\n\t\treturn fmt.Errorf(\"parameter BambooUrl missing\")\n\t}\n\tif len(config.BambooUsername) == 0 {\n\t\treturn fmt.Errorf(\"parameter BambooUsername missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter AuthUrl missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter AuthApplicationName missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter AuthApplicationPassword missing\")\n\t}\n\tif len(config.BambooPassword) == 0 {\n\t\treturn fmt.Errorf(\"parameter BambooPassword missing\")\n\t}\n\tif len(config.Prefix) == 0 {\n\t\treturn fmt.Errorf(\"parameter Prefix missing\")\n\t}\n\treturn nil\n}", "func (o *OAuth2Config) IsValid() bool {\n\treturn o.TokenURL != \"\" && o.AuthURL != \"\" && len(o.Scopes) != 0\n}", "func validateConfig(config ProvisionerConfig) error {\n\tswitch config.CloudProvider {\n\tcase constants.AWS, constants.Azure, constants.GCE, constants.Ops:\n\tdefault:\n\t\treturn trace.BadParameter(\"unknown cloud provider %s\", config.CloudProvider)\n\t}\n\n\terr := validator.New().Struct(&config)\n\tif err == nil {\n\t\treturn nil\n\t}\n\n\tvar errs []error\n\tif validationErrors, ok := err.(validator.ValidationErrors); ok {\n\t\tfor _, fieldError := range validationErrors {\n\t\t\terrs = append(errs,\n\t\t\t\ttrace.BadParameter(` * %s=\"%v\" fails \"%s\"`,\n\t\t\t\t\tfieldError.Field(), fieldError.Value(), fieldError.Tag()))\n\t\t}\n\t}\n\treturn trace.NewAggregate(errs...)\n}", "func (_ runtimeConfigValidator) ValidateConfig(config *stiapi.Config) []validation.ValidationError {\n\treturn validation.ValidateConfig(config)\n}", "func (cfg *Config) Validate() error {\n\tchecks := []struct {\n\t\tbad bool\n\t\terrMsg string\n\t}{\n\t\t{cfg.AuthorizeURL == \"\", \"no authorizeURL specified\"},\n\t\t{cfg.TokenURL == \"\", \"no tokenURL specified\"},\n\t\t{cfg.ClientID == \"\", \"no clientID specified\"},\n\t\t{cfg.ClientSecret == \"\" && !cfg.AllowEmptyClientSecret, \"no clientSecret specified\"},\n\t\t{cfg.RedirectURL == \"\", \"no redirectURL specified\"},\n\t\t{cfg.SessionSecurityKey == \"\", \"no SessionSecurityKey specified\"},\n\t\t{cfg.APIServerURL == \"\", \"no apiServerURL specified\"},\n\t}\n\n\tfor _, check := range checks {\n\t\tif check.bad {\n\t\t\treturn fmt.Errorf(\"invalid config: %s\", check.errMsg)\n\t\t}\n\t}\n\treturn nil\n}", "func (c Config) Validate() error {\n\treturn validation.ValidateStruct(&c,\n\t\tvalidation.Field(&c.AppMode, validation.Required),\n\t\tvalidation.Field(&c.AppName, validation.Required),\n\t\tvalidation.Field(&c.DBType, validation.Required),\n\t\tvalidation.Field(&c.DSN, validation.Required),\n\t\tvalidation.Field(&c.JWTSigningKey, validation.Required),\n\t\tvalidation.Field(&c.JWTExpiration, validation.Required),\n\t\tvalidation.Field(&c.MailSmtphost, validation.Required),\n\t\tvalidation.Field(&c.MailSmtpport, validation.Required),\n\t\tvalidation.Field(&c.MailUsername, validation.Required),\n\t\tvalidation.Field(&c.MailPassword, validation.Required),\n\t\tvalidation.Field(&c.AppFqdn, validation.Required),\n\t\tvalidation.Field(&c.HttpEntrypoint, validation.Required),\n\t\tvalidation.Field(&c.WebservName, validation.Required),\n\t\tvalidation.Field(&c.GoogleCredentialFile, validation.Required),\n\t\tvalidation.Field(&c.GoogleRedirectPath, validation.Required),\n\t\tvalidation.Field(&c.AppSecretKey, validation.Required),\n\t\tvalidation.Field(&c.BizName, validation.Required),\n\t\tvalidation.Field(&c.BizShortname, validation.Required),\n\t\tvalidation.Field(&c.BizEmail, validation.Required),\n\t\tvalidation.Field(&c.BizPhone),\n\t\tvalidation.Field(&c.BizPhone2),\n\t\tvalidation.Field(&c.BizLogo, validation.Required),\n\t)\n}", "func (c Config) Validate() (err error) {\n\tvar fi os.FileInfo\n\n\t// validate key fingerprint\n\t_, err = HexStringToFingerprint(c.MasterKeyFingerprint)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// validate TLSCert\n\tif len(c.TLSCert) == 0 {\n\t\treturn errors.New(\"Missing config param: TLSCert\")\n\t}\n\tfi, err = os.Stat(c.TLSCert)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in TLSCert '%s': %s\", c.TLSCert, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in TLSCert '%s': expected file path, got directory\", c.TLSCert)\n\t}\n\n\t// validate TLSKey\n\tif len(c.TLSKey) == 0 {\n\t\treturn errors.New(\"Missing config param: TLSKey\")\n\t}\n\tfi, err = os.Stat(c.TLSKey)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in TLSKey '%s': %s\", c.TLSKey, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in TLSKey '%s': expected file path, got directory\", c.TLSKey)\n\t}\n\n\t// validate SecRing\n\tif len(c.SecRing) == 0 {\n\t\treturn errors.New(\"Missing config param: SecRing\")\n\t}\n\tfi, err = os.Stat(c.SecRing)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in SecRing '%s': %s\", c.SecRing, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in SecRing '%s': expected file path, got directory\", c.SecRing)\n\t}\n\n\t// validate ProdSupportPubRing\n\tif len(c.ProdSupportPubRing) == 0 {\n\t\treturn errors.New(\"Missing config param: ProdSupportPubRing\")\n\t}\n\tfi, err = os.Stat(c.ProdSupportPubRing)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in ProdSupportPubRing '%s': %s\", c.ProdSupportPubRing, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in ProdSupportPubRing '%s': expected file path, got directory\", c.ProdSupportPubRing)\n\t}\n\n\t// validate DataRoot\n\tif len(c.DataRoot) == 0 {\n\t\treturn errors.New(\"Missing config param: DataRoot\")\n\t}\n\tfi, err = os.Stat(c.DataRoot)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.DataRoot, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in DataRoot '%s': %s\", c.DataRoot, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in DataRoot '%s': expected directory, got file path\", c.DataRoot)\n\t\t}\n\t}\n\n\t// validate ProdSupportDir\n\tif len(c.ProdSupportDir) == 0 {\n\t\treturn errors.New(\"Missing config param: ProdSupportDir\")\n\t}\n\tfi, err = os.Stat(c.ProdSupportDir)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.ProdSupportDir, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in ProdSupportDir '%s': %s\", c.ProdSupportDir, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in ProdSupportDir '%s': expected directory, got file path\", c.ProdSupportDir)\n\t\t}\n\t}\n\n\t// validate KeyRoot\n\tif len(c.KeyRoot) == 0 {\n\t\treturn errors.New(\"Missing config param: KeyRoot\")\n\t}\n\tfi, err = os.Stat(c.KeyRoot)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.KeyRoot, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in KeyRoot '%s': %s\", c.KeyRoot, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in KeyRoot '%s': expected directory, got file path\", c.KeyRoot)\n\t\t}\n\t}\n\n\t// validate MetaRoot\n\tif len(c.MetaRoot) == 0 {\n\t\treturn errors.New(\"Missing config param: MetaRoot\")\n\t}\n\tfi, err = os.Stat(c.MetaRoot)\n\tif err != nil {\n\t\t// doesn't exist... can we create it?\n\t\tif err = os.MkdirAll(c.MetaRoot, 0744); err != nil {\n\t\t\treturn fmt.Errorf(\"Config error in MetaRoot '%s': %s\", c.MetaRoot, err)\n\t\t}\n\t} else {\n\t\tif !fi.IsDir() {\n\t\t\treturn fmt.Errorf(\"Config error in MetaRoot '%s': expected directory, got file path\", c.MetaRoot)\n\t\t}\n\t}\n\n\t// validate HTTPLog\n\tif len(c.HTTPLog) > 0 {\n\t\tfi, err = os.Stat(filepath.Dir(c.HTTPLog))\n\t\tif err != nil {\n\t\t\t// doesn't exist... can we create it?\n\t\t\tif err = os.MkdirAll(filepath.Dir(c.HTTPLog), 0744); err != nil {\n\t\t\t\treturn fmt.Errorf(\"Config error in HTTPLog '%s': %s\", c.HTTPLog, err)\n\t\t\t}\n\t\t}\n\t}\n\n\t// validate HtpasswdFile\n\tif len(c.HtpasswdFile) == 0 {\n\t\treturn errors.New(\"Missing config param: HtpasswdFile\")\n\t}\n\tfi, err = os.Stat(c.HtpasswdFile)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"Config error in HtpasswdFile '%s': %s\", c.HtpasswdFile, err)\n\t}\n\tif fi.IsDir() {\n\t\treturn fmt.Errorf(\"Config error in HtpasswdFile '%s': expected file path, got directory\", c.HtpasswdFile)\n\t}\n\n\tif len(c.MasterKeyPassphrase) == 0 {\n\t\tlog.Println(\"no passphrase specified for secure keyring\")\n\t}\n\n\treturn nil\n}", "func (c *Config) Validate(log log.Logger) error {\n\tif err := c.SchemaConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid schema config\")\n\t}\n\tif err := c.StorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid storage config\")\n\t}\n\tif err := c.QueryRange.Validate(log); err != nil {\n\t\treturn errors.Wrap(err, \"invalid queryrange config\")\n\t}\n\tif err := c.TableManager.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid tablemanager config\")\n\t}\n\treturn nil\n}", "func (c *Config) Validate() error {\n\tif c.ServerConf.RestAPIPort < 0 || c.ServerConf.RestAPIPort > 65535 {\n\t\treturn errors.New(\"invalid restful port: \" + strconv.Itoa(c.ServerConf.RestAPIPort))\n\t}\n\n\t// remove \"/\"\n\tc.ZeekConf.LogDir = processDirName(c.ZeekConf.LogDir)\n\tc.ZeekConf.PcapDir = processDirName(c.ZeekConf.PcapDir)\n\tc.ZeekConf.ExtractedFileDir = processDirName(c.ZeekConf.ExtractedFileDir)\n\tc.SeaweedfsConf.ExtractedFileDir = processDirName(c.SeaweedfsConf.ExtractedFileDir)\n\n\tif c.SeaweedfsConf.Retries < 1 {\n\t\treturn errors.New(\"seaweedfs post retry times < 1\")\n\t}\n\treturn nil\n}", "func (config *Config) validateConfig() error {\n\tfor _, element := range config.AllowedEntries {\n\t\tif element.Width <= 0 && element.Height <= 0 {\n\t\t\treturn fmt.Errorf(\"The width and height of the configuration element with name \\\"%s\\\" are invalid.\", element.Name)\n\t\t}\n\n\t\tif element.Name == \"\" {\n\t\t\treturn fmt.Errorf(\"Name must be set\")\n\t\t}\n\n\t\tif element.Type == \"\" {\n\t\t\treturn fmt.Errorf(\"Type must be set\")\n\t\t}\n\n\t\ttypes := paint.GetAvailableTypes()\n\t\tif _, found := types[element.Type]; !found {\n\t\t\treturn fmt.Errorf(\"Type must be either %s, %s or %s at element \\\"%s\\\"\", paint.TypeCrop, paint.TypeResize, paint.TypeFit, element.Name)\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c *Config) Validate() error {\n\tvar errs []error\n\n\tif len(c.Hosts) == 0 {\n\t\terrs = append(errs, fmt.Errorf(\"missing hosts\"))\n\t}\n\tif c.Port == \"\" {\n\t\terrs = append(errs, fmt.Errorf(\"missing port\"))\n\t}\n\n\treturn apierrors.NewAggregate(errs)\n}", "func (cfg config) Validate() {\n\tif v, exists := cfg[\"tls_min_version\"]; exists {\n\t\tvar t tlsMinVersionOption\n\t\terr := t.Set(fmt.Sprintf(\"%v\", v))\n\t\tif err == nil {\n\t\t\tnewVal := fmt.Sprintf(\"%v\", t.Get())\n\t\t\tif newVal != \"0\" {\n\t\t\t\tcfg[\"tls_min_version\"] = newVal\n\t\t\t} else {\n\t\t\t\tdelete(cfg, \"tls_min_version\")\n\t\t\t}\n\t\t} else {\n\t\t\tlogFatal(\"failed parsing tls_min_version %+v\", v)\n\t\t}\n\t}\n\tif v, exists := cfg[\"log_level\"]; exists {\n\t\tvar t lg.LogLevel\n\t\terr := t.Set(fmt.Sprintf(\"%v\", v))\n\t\tif err == nil {\n\t\t\tcfg[\"log_level\"] = t\n\t\t} else {\n\t\t\tlogFatal(\"failed parsing log_level %+v\", v)\n\t\t}\n\t}\n}", "func (c Config) Validate() error {\n\tncNames := map[string]bool{}\n\tfor _, nc := range c.NotificationConfig {\n\t\tif nc.Name == nil {\n\t\t\treturn fmt.Errorf(\"Missing name in notification config: %s\", proto.MarshalTextString(nc))\n\t\t}\n\t\tfor _, pdc := range nc.PagerdutyConfig {\n\t\t\tif pdc.ServiceKey == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing service key in PagerDuty notification config: %s\", proto.MarshalTextString(pdc))\n\t\t\t}\n\t\t}\n\t\tfor _, ec := range nc.EmailConfig {\n\t\t\tif ec.Email == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing email address in email notification config: %s\", proto.MarshalTextString(ec))\n\t\t\t}\n\t\t}\n\t\tfor _, ec := range nc.PushoverConfig {\n\t\t\tif ec.Token == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing token in Pushover notification config: %s\", proto.MarshalTextString(ec))\n\t\t\t}\n\t\t\tif ec.UserKey == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing user key in Pushover notification config: %s\", proto.MarshalTextString(ec))\n\t\t\t}\n\t\t}\n\t\tfor _, hcc := range nc.HipchatConfig {\n\t\t\tif hcc.AuthToken == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing token in HipChat config: %s\", proto.MarshalTextString(hcc))\n\t\t\t}\n\t\t\tif hcc.RoomId == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing room in HipChat config: %s\", proto.MarshalTextString(hcc))\n\t\t\t}\n\t\t}\n\t\tfor _, sc := range nc.SlackConfig {\n\t\t\tif sc.WebhookUrl == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing webhook URL in Slack config: %s\", proto.MarshalTextString(sc))\n\t\t\t}\n\t\t}\n\t\tfor _, fc := range nc.FlowdockConfig {\n\t\t\tif fc.ApiToken == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing API token in Flowdock config: %s\", proto.MarshalTextString(fc))\n\t\t\t}\n\t\t\tif fc.FromAddress == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing from_address Flowdock config: %s\", proto.MarshalTextString(fc))\n\t\t\t}\n\t\t}\n\n\t\tif _, ok := ncNames[nc.GetName()]; ok {\n\t\t\treturn fmt.Errorf(\"Notification config name not unique: %s\", nc.GetName())\n\t\t}\n\n\t\tncNames[nc.GetName()] = true\n\t}\n\n\tfor _, a := range c.AggregationRule {\n\t\tfor _, f := range a.Filter {\n\t\t\tif f.NameRe == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing name pattern (name_re) in filter definition: %s\", proto.MarshalTextString(f))\n\t\t\t}\n\t\t\tif f.ValueRe == nil {\n\t\t\t\treturn fmt.Errorf(\"Missing value pattern (value_re) in filter definition: %s\", proto.MarshalTextString(f))\n\t\t\t}\n\t\t}\n\n\t\tif _, ok := ncNames[a.GetNotificationConfigName()]; !ok {\n\t\t\treturn fmt.Errorf(\"No such notification config: %s\", a.GetNotificationConfigName())\n\t\t}\n\t}\n\n\treturn nil\n}", "func (cfg Config) Validate() error {\n\tvar errs []string\n\tif cfg.Releaser == nil {\n\t\terrs = append(errs, \"releaser not supplied\")\n\t}\n\tif cfg.History == nil {\n\t\terrs = append(errs, \"history DB not supplied\")\n\t}\n\tif len(errs) > 0 {\n\t\treturn errors.New(\"invalid: \" + strings.Join(errs, \"; \"))\n\t}\n\treturn nil\n}", "func (c *Config) Validate() error {\n\tnames := make(map[string]struct{}, len(c.Configs))\n\tfor idx, c := range c.Configs {\n\t\tif c.Name == \"\" {\n\t\t\treturn fmt.Errorf(\"tempo config at index %d is missing a name\", idx)\n\t\t}\n\t\tif _, exist := names[c.Name]; exist {\n\t\t\treturn fmt.Errorf(\"found multiple tempo configs with name %s\", c.Name)\n\t\t}\n\t\tnames[c.Name] = struct{}{}\n\t}\n\n\treturn nil\n}", "func validateConfig(cfg YamlConfig) error {\n\n\t// If Directory is nil, try to retrieve from env var, if not then error\n\tif cfg.Directory == \"\" {\n\t\treturn errors.New(\"Did not find logs directory in YAML configuration\")\n\t}\n\n\t// If Directory resolved, check if directory exists, if not then error\n\tif _, err := os.Stat(cfg.Directory); os.IsNotExist(err) {\n\t\treturn errors.New(fmt.Sprintf(\"Resolved logs directory %s does not exist, exiting\", cfg.Directory))\n\t}\n\n\t// If Name or LogPattern missing, error\n\tif cfg.ConfigName == \"\" {\n\t\treturn errors.New(\"YAML configuration missing required 'ConfigName' key, exiting\")\n\t} else if cfg.LogPattern == \"\" {\n\t\treturn errors.New(\"YAML configuration missing required 'LogPattern' key, exiting\")\n\t}\n\n\tconnectorsConfigs := getConnectorsConfigs(cfg)\n\n\tfor _, connCfg := range connectorsConfigs {\n\t\t// Assert connectors have valid common fields values\n\t\terr := validateConnectorsCommonFields(connCfg)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\t// Assert connectors have valid S3 connector fields\n\t\terr = connCfg.validate()\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (m *Config) Validate() error {\n\treturn m.validate(false)\n}", "func (cc *Config) Validate() error {\n\tif len(cc.BrokerList) == 0 {\n\t\treturn ErrConfigNoBrokers\n\t}\n\n\tif cc.ReadTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidReadTimeout\n\t}\n\n\tif cc.WriteTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidWriteTimeout\n\t}\n\n\tif cc.ConnectTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidConnectTimeout\n\t}\n\n\tif cc.KeepAliveTimeout < time.Millisecond {\n\t\treturn ErrConfigInvalidKeepAliveTimeout\n\t}\n\n\tif cc.FetchSize < 1 {\n\t\treturn ErrConfigInvalidFetchSize\n\t}\n\n\tif cc.MetadataRetries < 0 {\n\t\treturn ErrConfigInvalidMetadataRetries\n\t}\n\n\tif cc.MetadataBackoff < time.Millisecond {\n\t\treturn ErrConfigInvalidMetadataBackoff\n\t}\n\n\tif cc.MetadataTTL < time.Millisecond {\n\t\treturn ErrConfigInvalidMetadataTTL\n\t}\n\n\tif cc.CommitOffsetRetries < 0 {\n\t\treturn ErrConfigInvalidCommitOffsetRetries\n\t}\n\n\tif cc.CommitOffsetBackoff < time.Millisecond {\n\t\treturn ErrConfigInvalidCommitOffsetBackoff\n\t}\n\n\tif cc.ConsumerMetadataRetries < 0 {\n\t\treturn ErrConfigInvalidConsumerMetadataRetries\n\t}\n\n\tif cc.ConsumerMetadataBackoff < time.Millisecond {\n\t\treturn ErrConfigInvalidConsumerMetadataBackoff\n\t}\n\n\tif cc.ClientID == \"\" {\n\t\treturn ErrConfigEmptyClientID\n\t}\n\n\treturn nil\n}", "func (m *TunnelConfig) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.validateIPAddress(formats); err != nil {\n\t\t// prop\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateSharedSecret(formats); err != nil {\n\t\t// prop\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (mgc Config) Validate() error {\n\tif mgc.User == \"\" {\n\t\treturn errors.New(\"Config.User is required\")\n\t}\n\tif mgc.Password == \"\" {\n\t\treturn errors.New(\"Config.Password is required\")\n\t}\n\tif mgc.AuthDB == \"\" {\n\t\treturn errors.New(\"Config.AuthDB is required\")\n\t}\n\tif mgc.Host == \"\" {\n\t\treturn errors.New(\"Config.Host is required\")\n\t}\n\tif mgc.DB == \"\" {\n\t\treturn errors.New(\"Config.DB is required\")\n\t}\n\treturn nil\n}", "func (m *EncryptionAtRestConfig) Validate(formats strfmt.Registry) error {\n\tvar res []error\n\n\tif err := m.validateKmsConfigUUID(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateOpType(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif err := m.validateType(formats); err != nil {\n\t\tres = append(res, err)\n\t}\n\n\tif len(res) > 0 {\n\t\treturn errors.CompositeValidationError(res...)\n\t}\n\treturn nil\n}", "func (config Configuration) Valid() error {\n\tif config.Issuer == \"\" {\n\t\treturn errors.New(\"authproxy/oidc: issuer is empty\")\n\t}\n\tif config.AuthorizationEndpoint == \"\" {\n\t\treturn errors.New(\"authproxy/oidc: authorization_endpoint is empty\")\n\t}\n\tif config.JWKSURI == \"\" {\n\t\treturn errors.New(\"authproxy/oidc: jwks_uri is empty\")\n\t}\n\tif len(config.ResponseTypesSupported) == 0 {\n\t\treturn errors.New(\"authproxy/oidc: response_types_supported is empty\")\n\t}\n\tif len(config.SubjectTypesSupported) == 0 {\n\t\treturn errors.New(\"authproxy/oidc: subject_types_supported is empty\")\n\t}\n\tif len(config.IDTokenSigningAlgValuesSupported) == 0 {\n\t\treturn errors.New(\"authproxy/oidc: id_token_signing_alg_values_supported is empty\")\n\t}\n\n\treturn nil\n}", "func (c *Config) Validate() error {\n\tvar err error\n\tif c.Endpoint == \"\" {\n\t\terr = multierr.Append(err, errors.New(\"no manager endpoint was specified\"))\n\t\treturn err\n\t}\n\n\tres, err := url.Parse(c.Endpoint)\n\tif err != nil {\n\t\terr = multierr.Append(err, fmt.Errorf(\"unable to parse url %s: %w\", c.Endpoint, err))\n\t\treturn err\n\t}\n\n\tif res.Scheme != \"http\" && res.Scheme != \"https\" {\n\t\terr = multierr.Append(err, errors.New(\"url scheme must be http or https\"))\n\t}\n\n\tif c.Username == \"\" {\n\t\terr = multierr.Append(err, errors.New(\"username not provided and is required\"))\n\t}\n\n\tif c.Password == \"\" {\n\t\terr = multierr.Append(err, errors.New(\"password not provided and is required\"))\n\t}\n\treturn err\n}", "func (cfg *Config) Validate() error {\n\tif !cfg.LogDataEnabled && !cfg.ProfilingDataEnabled {\n\t\treturn errors.New(`either \"log_data_enabled\" or \"profiling_data_enabled\" has to be true`)\n\t}\n\tif cfg.HTTPClientSettings.Endpoint == \"\" {\n\t\treturn errors.New(`requires a non-empty \"endpoint\"`)\n\t}\n\t_, err := cfg.getURL()\n\tif err != nil {\n\t\treturn fmt.Errorf(`invalid \"endpoint\": %w`, err)\n\t}\n\tif cfg.Token == \"\" {\n\t\treturn errors.New(`requires a non-empty \"token\"`)\n\t}\n\n\tif cfg.MaxContentLengthLogs > maxContentLengthLogsLimit {\n\t\treturn fmt.Errorf(`requires \"max_content_length_logs\" <= %d`, maxContentLengthLogsLimit)\n\t}\n\n\tif cfg.MaxContentLengthMetrics > maxContentLengthMetricsLimit {\n\t\treturn fmt.Errorf(`requires \"max_content_length_metrics\" <= %d`, maxContentLengthMetricsLimit)\n\t}\n\n\tif cfg.MaxContentLengthTraces > maxContentLengthTracesLimit {\n\t\treturn fmt.Errorf(`requires \"max_content_length_traces\" <= %d`, maxContentLengthTracesLimit)\n\t}\n\n\tif cfg.MaxEventSize > maxMaxEventSize {\n\t\treturn fmt.Errorf(`requires \"max_event_size\" <= %d`, maxMaxEventSize)\n\t}\n\n\tif err := cfg.QueueSettings.Validate(); err != nil {\n\t\treturn fmt.Errorf(\"sending_queue settings has invalid configuration: %w\", err)\n\t}\n\treturn nil\n}", "func (c *Config) Validate() bool {\n\tswitch c.Ops {\n\tcase 0, 1, 2, 4, 8:\n\t\treturn true\n\t}\n\treturn false\n}", "func ValidateConfig() {\n\tconfig := GetConfig()\n\n\t// list of required keys\n\t// if these do not exist, prevent running the app\n\trequiredKeys := []string{\n\t\t\"app.core.root_directory\",\n\t\t\"app.core.log_level\",\n\t\t\"app.core.rest.url\",\n\t\t\"app.core.rest.user\",\n\t\t\"app.core.rest.xor_key\",\n\t\t\"app.core.rest.password\",\n\t\t\"app.core.rest.masked\",\n\t\t\"app.core.db.initialised\",\n\t\t\"app.core.db.path\",\n\t}\n\n\tfor _, key := range requiredKeys {\n\t\tif !config.IsSet(key) {\n\t\t\tlog.WithFields(log.Fields{\"error\": fmt.Sprintf(\"The key %s is required, but could not be found! See the sample file for reference!\", key)}).Error(\"Invalid config file detected!\")\n\t\t\tos.Exit(1)\n\t\t}\n\t}\n}", "func (cfg *Config) Validate() error {\n\tif cfg.AssumeRole.STSRegion == \"\" && cfg.Region != \"\" {\n\t\tcfg.AssumeRole.STSRegion = cfg.Region\n\t}\n\n\tcredsProvider, err := getCredsProviderFromConfig(cfg)\n\tif err != nil {\n\t\treturn fmt.Errorf(\"could not retrieve credential provider: %w\", err)\n\t}\n\tif credsProvider == nil {\n\t\treturn fmt.Errorf(\"credsProvider cannot be nil\")\n\t}\n\tcfg.credsProvider = credsProvider\n\n\treturn nil\n}", "func (c *Config) Validate() error {\n\tvar errs multierror.Errors\n\n\tfor _, ht := range c.HashTypes {\n\t\tif !ht.IsValid() {\n\t\t\terrs = append(errs, errors.Errorf(\"invalid hash_types value '%v'\", ht))\n\t\t}\n\t}\n\n\tvar err error\n\n\tc.MaxFileSizeBytes, err = humanize.ParseBytes(c.MaxFileSize)\n\tif err != nil {\n\t\terrs = append(errs, errors.Wrap(err, \"invalid max_file_size value\"))\n\t} else if c.MaxFileSizeBytes <= 0 {\n\t\terrs = append(errs, errors.Errorf(\"max_file_size value (%v) must be positive\", c.MaxFileSize))\n\t}\n\n\tc.ScanRateBytesPerSec, err = humanize.ParseBytes(c.ScanRatePerSec)\n\tif err != nil {\n\t\terrs = append(errs, errors.Wrap(err, \"invalid scan_rate_per_sec value\"))\n\t}\n\n\treturn errs.Err()\n}", "func (c Config) Validate() error {\n\tif c.ApplicationID != \"\" && c.AlgoliaAPIKey == \"\" {\n\t\treturn errors.New(\"API key must not be empty if indexer is enabled\")\n\t}\n\tif c.ApplicationID != \"\" && c.AlgoliaSearchKey == \"\" {\n\t\treturn errors.New(\"Search key must not be empty if indexer is enabled\")\n\t}\n\tif c.ApplicationID != \"\" && c.IndexName == \"\" {\n\t\treturn errors.New(\"Index name must not be empty if indexer is enabled\")\n\t}\n\treturn nil\n}", "func (cfg *NetworkServiceConfig) IsValid() error {\n\tif cfg.Mechanism == \"\" {\n\t\treturn errors.New(\"invalid mechanism specified\")\n\t}\n\tswitch cfg.Mechanism {\n\tcase memif.MECHANISM:\n\t\t// Verify folder for memif file exists and writable.\n\t\t//TODO: Add support of this validation.\n\tcase kernel.MECHANISM:\n\t\t// Verify interface name\n\t\tif len(cfg.Path) > 1 {\n\t\t\treturn errors.New(\"invalid client interface name specified\")\n\t\t}\n\t\tif len(cfg.Path[0]) > 15 {\n\t\t\treturn errors.New(\"interface part cannot exceed 15 characters\")\n\t\t}\n\t}\n\treturn nil\n}", "func (c *Config) Validate() error {\n\treturn nil\n}", "func (c *Config) Validate() error {\n\treturn nil\n}", "func (c *Config) Validate() error {\n\treturn nil\n}", "func (c *Config) Validate() error {\n\t// TODO(felix): complete validates\n\treturn nil\n}", "func (cfg *Config) Validate() error {\n\tif err := cfg.CassandraStorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Cassandra Storage config\")\n\t}\n\tif err := cfg.GCPStorageConfig.Validate(util_log.Logger); err != nil {\n\t\treturn errors.Wrap(err, \"invalid GCP Storage Storage config\")\n\t}\n\tif err := cfg.Swift.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Swift Storage config\")\n\t}\n\tif err := cfg.IndexQueriesCacheConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Index Queries Cache config\")\n\t}\n\tif err := cfg.AzureStorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid Azure Storage config\")\n\t}\n\tif err := cfg.AWSStorageConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid AWS Storage config\")\n\t}\n\tif err := cfg.BoltDBShipperConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid boltdb-shipper config\")\n\t}\n\tif err := cfg.TSDBShipperConfig.Validate(); err != nil {\n\t\treturn errors.Wrap(err, \"invalid tsdb config\")\n\t}\n\n\treturn cfg.NamedStores.validate()\n}", "func (cfg Config) Validate() error {\n\treturn validation.ValidateStruct(\n\t\t&cfg,\n\t\tvalidation.Field(&cfg.NodeID, validation.Required),\n\t\tvalidation.Field(&cfg.ListenAddr, validation.Required, is.Host),\n\t\tvalidation.Field(&cfg.DataDir, validation.Required),\n\t\tvalidation.Field(&cfg.CompactionEnabled, validation.Required),\n\t\tvalidation.Field(&cfg.Peers),\n\t)\n}", "func (cfg *Config) Validate() error {\n\tif cfg.HTTPServerSettings == nil && cfg.GRPCServerSettings == nil {\n\t\treturn errAtLeastOneProtocol\n\t}\n\n\tif cfg.Source.File != \"\" && cfg.Source.Remote != nil {\n\t\treturn errTooManySources\n\t}\n\n\tif cfg.Source.File == \"\" && cfg.Source.Remote == nil {\n\t\treturn errNoSources\n\t}\n\n\treturn nil\n}", "func (cfg *Config) Validate() error {\n\tif cfg.HTTPServerSettings == nil && cfg.GRPCServerSettings == nil {\n\t\treturn errAtLeastOneProtocol\n\t}\n\n\tif cfg.Source.File != \"\" && cfg.Source.Remote != nil {\n\t\treturn errTooManySources\n\t}\n\n\tif cfg.Source.File == \"\" && cfg.Source.Remote == nil {\n\t\treturn errNoSources\n\t}\n\n\treturn nil\n}", "func ValidateConfig(c Config) error {\n\tif c.Stack.ID == \"\" {\n\t\treturn errors.New(\"stack.id is required\")\n\t}\n\n\tif c.Stack.BuildImage == \"\" {\n\t\treturn errors.New(\"stack.build-image is required\")\n\t}\n\n\tif c.Stack.RunImage == \"\" {\n\t\treturn errors.New(\"stack.run-image is required\")\n\t}\n\n\treturn nil\n}", "func ValidateConfig(Validations Validations, config *types.Config) error {\n\tfor _, Service := range config.Services {\n\n\t\terr := ValidateVolumes(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = ValidateSecrets(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = ValidateNetworks(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\terr = ValidateResources(Validations, Service)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfmt.Println(Service.NetworkMode)\n\n\t}\n\n\treturn nil\n}", "func validateConfigValue(value interface{}) error {\n\tvar err error\n\tswitch v := reflect.ValueOf(value); v.Kind() {\n\tcase reflect.Slice:\n\t\tlog.Infof(\">> checking config: found slice %+v\\n\", v.Slice(0, v.Cap()))\n\t\tfor i := 0; i < v.Len(); i++ {\n\t\t\terr = validateConfigValue(v.Index(i).Interface())\n\t\t}\n\tcase reflect.Map:\n\t\tlog.Infof(\">> checking config: found map %+v\\n\", v.MapKeys())\n\t\tfor _, key := range v.MapKeys() {\n\t\t\terr = validateConfigValue(v.MapIndex(key).Interface())\n\t\t}\n\tcase reflect.String:\n\t\tlog.Infof(\">> checking config: found string \\n\")\n\tdefault:\n\t\terr = fmt.Errorf(\"tfs doesn't handle %s type in the configuration file\", v.Kind())\n\t}\n\treturn err\n}", "func (c *hostNameFormatConfig) IsValid(name string) bool {\n\tfor _, validator := range c.validators {\n\t\tif !validator.IsValid(name) {\n\t\t\treturn false\n\t\t}\n\t}\n\n\treturn true\n}", "func validateConfig(appConfig AppConfig) error {\n\n\treturn nil\n}", "func (c Config) Validate() error {\n\tif c.CollectorEndpoint == \"\" && c.AgentEndpoint == \"\" {\n\t\treturn xerrors.New(\"jaeger: either collector endpoint or agent endpoint must be configured\")\n\t}\n\tif c.ServiceName == \"\" {\n\t\treturn xerrors.New(\"jaeger: service name must not be blank\")\n\t}\n\n\treturn nil\n}", "func (c *Config) valid() error {\n\tif c.Score == nil {\n\t\treturn errors.New(\"Expected Score to not be nil\")\n\t}\n\tif c.Sampler == nil {\n\t\treturn errors.New(\"Expected Sampler to not be nil\")\n\t}\n\treturn nil\n}", "func (c *Config) Validate() error {\n\tvar allErrs error\n\n\tif c.Endpoint == \"\" {\n\t\treturn multierr.Append(allErrs, errEmptyEndpoint)\n\t}\n\n\thost, portStr, err := net.SplitHostPort(c.Endpoint)\n\tif err != nil {\n\t\treturn multierr.Append(allErrs, fmt.Errorf(\"%w: %s\", errBadEndpoint, err.Error()))\n\t}\n\n\tif host == \"\" {\n\t\tallErrs = multierr.Append(allErrs, errBadEndpoint)\n\t}\n\n\tport, err := strconv.ParseInt(portStr, 10, 32)\n\tif err != nil {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: %s\", errBadPort, err.Error()))\n\t}\n\n\tif port < 0 || port > 65535 {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: %d\", errBadPort, port))\n\t}\n\n\tif c.Username != \"\" && c.Password == \"\" {\n\t\tallErrs = multierr.Append(allErrs, errEmptyPassword)\n\t}\n\n\tif c.Password != \"\" && c.Username == \"\" {\n\t\tallErrs = multierr.Append(allErrs, errEmptyUsername)\n\t}\n\tif c.Timeout.Milliseconds() < 0 {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: must be positive\", errNegativeTimeout))\n\t}\n\n\tif c.TLS != nil {\n\t\t_, err := c.TLS.LoadTLSConfig()\n\t\tif err != nil {\n\t\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: %s\", errFailedTLSLoad, err.Error()))\n\t\t}\n\t}\n\n\tif c.TLS != nil && c.TLSName == \"\" {\n\t\tallErrs = multierr.Append(allErrs, fmt.Errorf(\"%w: when using TLS\", errEmptyEndpointTLSName))\n\t}\n\n\treturn allErrs\n}", "func (c CopyConfig) Validate() error {\n\tif err := c.GitilesConfig.Validate(); err != nil {\n\t\treturn skerr.Wrap(err)\n\t}\n\tif len(c.Copies) == 0 {\n\t\treturn skerr.Fmt(\"Copies are required\")\n\t}\n\tfor _, copy := range c.Copies {\n\t\tif err := copy.Validate(); err != nil {\n\t\t\treturn skerr.Wrap(err)\n\t\t}\n\t}\n\treturn nil\n}", "func (o *IamLdapConfigParamsAllOf) HasConfigHost() bool {\n\tif o != nil && o.ConfigHost != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (d *common) validateConfig(info *api.NetworkACLPut) error {\n\terr := d.validateConfigMap(info.Config, nil)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Normalise rules before validation for duplicate detection.\n\tfor i := range info.Ingress {\n\t\tinfo.Ingress[i].Normalise()\n\t}\n\n\tfor i := range info.Egress {\n\t\tinfo.Egress[i].Normalise()\n\t}\n\n\t// Validate each ingress rule.\n\tfor i, ingressRule := range info.Ingress {\n\t\terr := d.validateRule(ruleDirectionIngress, ingressRule)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Invalid ingress rule %d: %w\", i, err)\n\t\t}\n\n\t\t// Check for duplicates.\n\t\tfor ri, r := range info.Ingress {\n\t\t\tif ri == i {\n\t\t\t\tcontinue // Skip ourselves.\n\t\t\t}\n\n\t\t\tif r == ingressRule {\n\t\t\t\treturn fmt.Errorf(\"Duplicate of ingress rule %d\", i)\n\t\t\t}\n\t\t}\n\t}\n\n\t// Validate each egress rule.\n\tfor i, egressRule := range info.Egress {\n\t\terr := d.validateRule(ruleDirectionEgress, egressRule)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"Invalid egress rule %d: %w\", i, err)\n\t\t}\n\n\t\t// Check for duplicates.\n\t\tfor ri, r := range info.Egress {\n\t\t\tif ri == i {\n\t\t\t\tcontinue // Skip ourselves.\n\t\t\t}\n\n\t\t\tif r == egressRule {\n\t\t\t\treturn fmt.Errorf(\"Duplicate of egress rule %d\", i)\n\t\t\t}\n\t\t}\n\t}\n\n\treturn nil\n}", "func ValidateConfig(config *Config) error {\n\tif config.Gcal.ClientID == \"\" {\n\t\treturn errors.New(\"config validattion error: gcal.client_id is missing\")\n\t}\n\tif config.Gcal.ClientSecret == \"\" {\n\t\treturn errors.New(\"config validattion error: gcal.client_secret is missing\")\n\t}\n\n\treturn nil\n}", "func (config appConfig) Validate() error {\n\treturn validator.New().Struct(&config)\n}", "func (tc *TestConfig) IsValid() bool {\n\treturn tc.APIConfig != nil && tc.err == nil\n}", "func Validate(cfg *Config) error {\n\tif cfg == nil {\n\t\treturn errors.New(\"config needs to be defined\")\n\t}\n\n\tvar result *multierror.Error\n\tif cfg.HostProvider == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"a host provider needs to be provided\"))\n\t}\n\tif cfg.BaseClusterCloudprovider == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"the cloudprovider of the hostcluster needs to be defined\"))\n\t}\n\tif cfg.Shoots.DefaultTest == nil {\n\t\tresult = multierror.Append(result, errors.New(\"a default test needs to be defined\"))\n\t}\n\n\tif cfg.Shoots.Namespace == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"the shoot project namespace has to be defined\"))\n\t}\n\n\tif cfg.Gardener.Version == \"\" && cfg.Gardener.Commit == \"\" {\n\t\tresult = multierror.Append(result, errors.New(\"a gardener version or commit has to be defined\"))\n\t}\n\n\tif len(cfg.GardenerExtensions) == 0 {\n\t\tresult = multierror.Append(result, errors.New(\"the gardener extensions have to be defined\"))\n\t}\n\n\treturn util.ReturnMultiError(result)\n}", "func (o *RemoteEnvironmentConfigListDto) HasValues() bool {\n\tif o != nil && o.Values != nil {\n\t\treturn true\n\t}\n\n\treturn false\n}", "func (e ExternalService) validateConfig() error {\n\tsl := gojsonschema.NewSchemaLoader()\n\tsc, err := sl.Compile(gojsonschema.NewStringLoader(e.schema()))\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to compile schema for external service of kind %q\", e.Kind)\n\t}\n\n\tnormalized, err := jsonc.Parse(e.Config)\n\tif err != nil {\n\t\treturn errors.Wrapf(err, \"failed to normalize JSON\")\n\t}\n\n\tres, err := sc.Validate(gojsonschema.NewBytesLoader(normalized))\n\tif err != nil {\n\t\treturn errors.Wrap(err, \"failed to validate config against schema\")\n\t}\n\n\terrs := new(multierror.Error)\n\tfor _, err := range res.Errors() {\n\t\terrs = multierror.Append(errs, errors.New(err.String()))\n\t}\n\n\treturn errs.ErrorOrNil()\n}", "func (cm ConfigMessage) IsValid() errors.Error {\n\tif cm.MspID == \"\" {\n\t\treturn errors.New(errors.InvalidConfigMessage, \"MSPID cannot be empty\")\n\t}\n\n\tif len(cm.Peers) == 0 && len(cm.Apps) == 0 {\n\t\treturn errors.New(errors.InvalidConfigMessage, \"Either peers or apps should be set\")\n\t}\n\n\tif len(cm.Peers) > 0 {\n\t\tfor _, config := range cm.Peers {\n\t\t\tif err := config.IsValid(); err != nil {\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\n\t}\n\n\treturn nil\n}", "func (c *config) checkValidConfig() hcl.Diagnostics {\n\tvar diagnostics hcl.Diagnostics\n\n\tdiagnostics = append(diagnostics, c.checkNotEmptyWorkers()...)\n\tdiagnostics = append(diagnostics, c.checkWorkerPoolNamesUnique()...)\n\tdiagnostics = append(diagnostics, c.checkReservationIDs()...)\n\tdiagnostics = append(diagnostics, c.validateOSVersion()...)\n\tdiagnostics = append(diagnostics, c.checkCPUManagerPolicy()...)\n\n\tif c.ConntrackMaxPerCore < 0 {\n\t\tdiagnostics = append(diagnostics, &hcl.Diagnostic{\n\t\t\tSeverity: hcl.DiagError,\n\t\t\tSummary: \"conntrack_max_per_core can't be negative value\",\n\t\t\tDetail: fmt.Sprintf(\"'conntrack_max_per_core' value is %d\", c.ConntrackMaxPerCore),\n\t\t})\n\t}\n\n\tif c.OIDC != nil {\n\t\t_, diags := c.OIDC.ToKubeAPIServerFlags(c.clusterDomain())\n\t\tdiagnostics = append(diagnostics, diags...)\n\t}\n\n\tif _, diags := c.resolveNodePrivateCIDRs(); diags != nil {\n\t\tdiagnostics = append(diagnostics, diags...)\n\t}\n\n\treturn diagnostics\n}", "func (pc PeerConfig) IsValid() errors.Error {\n\tif pc.PeerID == \"\" {\n\t\treturn errors.New(errors.InvalidPeerConfig, \"PeerID cannot be empty\")\n\t}\n\tif len(pc.App) == 0 {\n\t\treturn errors.New(errors.InvalidPeerConfig, \"App cannot be empty\")\n\t}\n\t//App is required\n\tfor _, appConfig := range pc.App {\n\t\tif err := appConfig.IsValid(); err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn nil\n}", "func (c *Config) Validate() error {\n\tfor _, ci := range c.Converters {\n\t\tif ci.Pattern == \"\" {\n\t\t\treturn errors.New(\"[pattern] is required for a converter\")\n\t\t}\n\t\tif ci.MetricName == \"\" {\n\t\t\treturn errors.New(\"[metricName] is required for a converter\")\n\t\t}\n\t}\n\n\treturn nil\n}", "func (kce *KafkaConfigExecutor) Validate(config *gateways.ConfigContext) error {\n\tkafkaConfig, err := parseConfig(config.Data.Config)\n\tif err != nil {\n\t\treturn gateways.ErrConfigParseFailed\n\t}\n\tif kafkaConfig == nil {\n\t\treturn fmt.Errorf(\"%+v, configuration must be non empty\", gateways.ErrInvalidConfig)\n\t}\n\tif kafkaConfig.URL == \"\" {\n\t\treturn fmt.Errorf(\"%+v, url must be specified\", gateways.ErrInvalidConfig)\n\t}\n\tif kafkaConfig.Topic == \"\" {\n\t\treturn fmt.Errorf(\"%+v, topic must be specified\", gateways.ErrInvalidConfig)\n\t}\n\tif kafkaConfig.Partition == \"\" {\n\t\treturn fmt.Errorf(\"%+v, partition must be specified\", gateways.ErrInvalidConfig)\n\t}\n\treturn nil\n}", "func (config *Config) valid() error {\n\t// check minlevel validation\n\tif \"\" != config.MinLevel && !LevelFromString(config.MinLevel).valid() {\n\t\treturn ErrConfigBadAttributes\n\t}\n\n\t// check filters len\n\tif len(config.Filters) < 1 {\n\t\treturn ErrConfigFiltersNotFound\n\t}\n\n\t// check filter one by one\n\tfor _, filter := range config.Filters {\n\t\tif \"\" == filter.Levels {\n\t\t\treturn ErrConfigLevelsNotFound\n\t\t}\n\n\t\tif (file{}) != filter.File {\n\t\t\t// seem not needed now\n\t\t\t//if \"\" == filter.File.Path {\n\t\t\t//return ErrConfigFilePathNotFound\n\t\t\t//}\n\t\t} else if (rotateFile{}) != filter.RotateFile {\n\t\t\tif \"\" == filter.RotateFile.Path {\n\t\t\t\treturn ErrConfigFilePathNotFound\n\t\t\t}\n\n\t\t\tif \"\" == filter.RotateFile.Type {\n\t\t\t\treturn ErrConfigFileRotateTypeNotFound\n\t\t\t}\n\t\t} else if (socket{}) != filter.Socket {\n\t\t\tif \"\" == filter.Socket.Address {\n\t\t\t\treturn ErrConfigSocketAddressNotFound\n\t\t\t}\n\n\t\t\tif \"\" == filter.Socket.Network {\n\t\t\t\treturn ErrConfigSocketNetworkNotFound\n\t\t\t}\n\t\t} else {\n\t\t\treturn ErrConfigMissingFilterType\n\t\t}\n\t}\n\n\treturn nil\n}", "func (config *Config) Validate() error {\n\n\tif _, err := os.Stat(filepath.Join(config.KirdPath, config.KernelFile)); os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"kernel '%s' not found\", filepath.Join(config.KirdPath, config.KernelFile))\n\t}\n\tif _, err := os.Stat(filepath.Join(config.KirdPath, config.InitrdFile)); os.IsNotExist(err) {\n\t\treturn fmt.Errorf(\"initrd '%s' not found\", filepath.Join(config.KirdPath, config.InitrdFile))\n\t}\n\n\t// Ensure all the MappedVirtualDisks exist on the host\n\tfor _, mvd := range config.MappedVirtualDisks {\n\t\tif _, err := os.Stat(mvd.HostPath); err != nil {\n\t\t\treturn fmt.Errorf(\"mapped virtual disk '%s' not found\", mvd.HostPath)\n\t\t}\n\t\tif mvd.ContainerPath == \"\" {\n\t\t\treturn fmt.Errorf(\"mapped virtual disk '%s' requested without a container path\", mvd.HostPath)\n\t\t}\n\t}\n\n\treturn nil\n}", "func ValidateConfig(fldPath *field.Path, config tmv1beta1.ConfigElement) field.ErrorList {\n\tvar allErrs field.ErrorList\n\tif config.Name == \"\" {\n\t\tallErrs = append(allErrs, field.Required(fldPath.Child(\"name\"), \"Required value\"))\n\t}\n\n\t// configmaps should either have a value or a value from defined\n\tif len(config.Value) == 0 && config.ValueFrom == nil {\n\t\tallErrs = append(allErrs, field.Required(fldPath.Child(\"value/valueFrom\"), \"A config must consist of a value or a reference to a value\"))\n\t}\n\n\t// if a valuefrom is defined then a configmap or a secret reference should be defined\n\tif config.ValueFrom != nil {\n\t\tallErrs = append(allErrs, strconf.Validate(fldPath.Child(\"valueFrom\"), config.ValueFrom)...)\n\t}\n\n\tif config.Type != tmv1beta1.ConfigTypeEnv && config.Type != tmv1beta1.ConfigTypeFile {\n\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"type\"), config.Type, \"unknown config type\"))\n\t\treturn allErrs\n\t}\n\n\tif config.Type == tmv1beta1.ConfigTypeEnv {\n\t\tif errs := validation.IsEnvVarName(config.Name); len(errs) != 0 {\n\t\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"name\"), config.Name, strings.Join(errs, \":\")))\n\t\t}\n\t\tif errs := validation.IsCIdentifier(config.Name); len(errs) != 0 {\n\t\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"name\"), config.Name, strings.Join(errs, \":\")))\n\t\t}\n\t}\n\n\tif config.Type == tmv1beta1.ConfigTypeFile {\n\t\tif config.Path == \"\" {\n\t\t\tallErrs = append(allErrs, field.Required(fldPath.Child(\"path\"), fmt.Sprintf(\"path is required for configtype %q\", tmv1beta1.ConfigTypeFile)))\n\t\t}\n\t\t// check if value is base64 encoded\n\t\tif config.Value != \"\" {\n\t\t\tif _, err := base64.StdEncoding.DecodeString(config.Value); err != nil {\n\t\t\t\tallErrs = append(allErrs, field.Invalid(fldPath.Child(\"value\"), config.Value, \"Value must be base64 encoded\"))\n\t\t\t}\n\t\t}\n\t}\n\treturn allErrs\n}", "func (c DbConfig) Validate() error {\n\t// Holds empty config fields\n\tempty := []string{}\n\n\t// Host\n\tif len(c.Host) == 0 {\n\t\tempty = append(empty, \"Host\")\n\t}\n\n\t// User\n\tif len(c.User) == 0 {\n\t\tempty = append(empty, \"User\")\n\t}\n\n\t// Password\n\tif len(c.Password) == 0 {\n\t\tempty = append(empty, \"Password\")\n\t}\n\n\t// Db\n\tif len(c.Db) == 0 {\n\t\tempty = append(empty, \"Db\")\n\t}\n\n\t// If any empty\n\tif len(empty) > 0 {\n\t\treturn fmt.Errorf(\"the db config fields: %s, were empty\",\n\t\t\tstrings.Join(empty, \",\"))\n\t}\n\n\t// All good\n\treturn nil\n}", "func (scc *ServiceCustomConfig) Validate() error {\n\n\tif len(scc.MyserviceInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for service not configured\")\n\t}\n\tif scc.MyserviceInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for service not configured\")\n\t}\n\n\tif len(scc.CommandClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Core Command client not configured\")\n\t}\n\tif scc.CommandClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Core Command client not configured\")\n\t}\n\n\tif len(scc.NotificationClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Core Command client not configured\")\n\t}\n\tif scc.NotificationClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Core Command client not configured\")\n\t}\n\n\tif len(scc.SchedulerClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Scheduler client not configured\")\n\t}\n\tif scc.SchedulerClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Scheduler client not configured\")\n\t}\n\n\tif len(scc.RuleEngineClientInfo.Host) == 0 {\n\t\treturn errors.New(\"host setting for Rule Engine client not configured\")\n\t}\n\tif scc.RuleEngineClientInfo.Port == 0 {\n\t\treturn errors.New(\"port setting for Rule Engine client not configured\")\n\t}\n\n\treturn nil\n}", "func (c *Config) validate() error {\n\tif err := validation.ValidateStruct(c); err != nil {\n\t\treturn err\n\t}\n\n\tif c.EnableBuiltInFiltering != nil && !*c.EnableBuiltInFiltering {\n\t\treturn errors.New(\"enableBuiltInFiltering must be true or unset, false is no longer supported\")\n\t}\n\n\tif _, err := url.Parse(c.IngestURL); err != nil {\n\t\treturn fmt.Errorf(\"%s is not a valid ingest URL: %v\", c.IngestURL, err)\n\t}\n\n\tif _, err := url.Parse(c.APIURL); err != nil {\n\t\treturn fmt.Errorf(\"%s is not a valid API URL: %v\", c.APIURL, err)\n\t}\n\n\tif _, err := url.Parse(c.EventEndpointURL); err != nil {\n\t\treturn fmt.Errorf(\"%s is not a valid event endpoint URL: %v\", c.EventEndpointURL, err)\n\t}\n\n\tif c.TraceEndpointURL != \"\" {\n\t\tif _, err := url.Parse(c.TraceEndpointURL); err != nil {\n\t\t\treturn fmt.Errorf(\"%s is not a valid trace endpoint URL: %v\", c.TraceEndpointURL, err)\n\t\t}\n\t}\n\n\tif err := c.Collectd.Validate(); err != nil {\n\t\treturn err\n\t}\n\n\tfor i := range c.Monitors {\n\t\tif err := c.Monitors[i].Validate(); err != nil {\n\t\t\treturn fmt.Errorf(\"monitor config for type '%s' is invalid: %v\", c.Monitors[i].Type, err)\n\t\t}\n\t}\n\n\treturn c.Writer.Validate()\n}", "func (cfg *HTTPConfig) Valid() error {\n\t// TODO: more validation, for example, the file existence\n\tif cfg.Schema != \"https\" && cfg.Schema != \"http\" {\n\t\treturn fmt.Errorf(\"invalid schema '%s', only 'http' and 'https' are supported \", cfg.Schema)\n\t}\n\n\treturn nil\n}", "func (v ConfigValueFilter) IsValid() bool {\n\tfor _, existing := range allowedConfigValueFilterEnumValues {\n\t\tif existing == v {\n\t\t\treturn true\n\t\t}\n\t}\n\treturn false\n}", "func (c *DataGeneratorConfig) Validate() error {\n\terr := c.BaseConfig.Validate()\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif c.InitialScale == 0 {\n\t\tc.InitialScale = c.BaseConfig.Scale\n\t}\n\n\tif c.LogInterval == 0 {\n\t\treturn fmt.Errorf(errLogIntervalZero)\n\t}\n\n\terr = utils.ValidateGroups(c.InterleavedGroupID, c.InterleavedNumGroups)\n\n\tif c.Use == UseCaseDevopsGeneric && c.MaxMetricCountPerHost < 1 {\n\t\treturn fmt.Errorf(errMaxMetricCountValue)\n\t}\n\n\treturn err\n}", "func (v *Config) Validate() error {\n\tctx := v.ctx\n\n\tif v.Log.Level == \"info\" {\n\t\tWarn.Println(ctx, \"info level hurts performance\")\n\t}\n\n\tif len(v.Stat.Disks) > 0 {\n\t\tWarn.Println(ctx, \"stat disks not support\")\n\t}\n\n\tif v.Workers < 0 || v.Workers > 64 {\n\t\treturn fmt.Errorf(\"workers must in [0, 64], actual is %v\", v.Workers)\n\t}\n\tif v.Listen <= 0 || v.Listen > 65535 {\n\t\treturn fmt.Errorf(\"listen must in (0, 65535], actual is %v\", v.Listen)\n\t}\n\tif v.ChunkSize < 128 || v.ChunkSize > 65535 {\n\t\treturn fmt.Errorf(\"chunk_size must in [128, 65535], actual is %v\", v.ChunkSize)\n\t}\n\n\tif v.Go.GcInterval < 0 || v.Go.GcInterval > 24*3600 {\n\t\treturn fmt.Errorf(\"go gc_interval must in [0, 24*3600], actual is %v\", v.Go.GcInterval)\n\t}\n\n\tif v.Log.Level != \"info\" && v.Log.Level != \"trace\" && v.Log.Level != \"warn\" && v.Log.Level != \"error\" {\n\t\treturn fmt.Errorf(\"log.leve must be info/trace/warn/error, actual is %v\", v.Log.Level)\n\t}\n\tif v.Log.Tank != \"console\" && v.Log.Tank != \"file\" {\n\t\treturn fmt.Errorf(\"log.tank must be console/file, actual is %v\", v.Log.Tank)\n\t}\n\tif v.Log.Tank == \"file\" && len(v.Log.File) == 0 {\n\t\treturn errors.New(\"log.file must not be empty for file tank\")\n\t}\n\n\tfor i, p := range v.Vhosts {\n\t\tif p.Name == \"\" {\n\t\t\treturn fmt.Errorf(\"the %v vhost is empty\", i)\n\t\t}\n\t}\n\n\treturn nil\n}", "func ConfigOK(config Config) bool {\n\treturn config.Location != \"\" && config.User != \"\" && config.password != \"\"\n}", "func (m *LolClubsClubsConfig) Validate(formats strfmt.Registry) error {\n\treturn nil\n}", "func (c Config) Validate() (bool, error) {\n\t// First do some sanity checks\n\thostPorts := map[string]interface{}{}\n\n\tfor container, portForwards := range c.Forwards {\n\t\tfor _, portForward := range portForwards {\n\n\t\t\t// Make sure that port lists were actually provided\n\t\t\tif len(portForward.Ports) == 0 {\n\t\t\t\treturn false, fmt.Errorf(\"No ports provided for container %s\", container)\n\t\t\t}\n\t\t\tfor hPort := range portForward.Ports {\n\t\t\t\t_, err := strconv.Atoi(hPort)\n\t\t\t\tif err != nil {\n\t\t\t\t\treturn false, fmt.Errorf(\"Invalid port %s provided for container %s\", hPort, container)\n\t\t\t\t}\n\n\t\t\t\t// Can only forward a port from the host to one container, check to ensure no duplicate host ports\n\t\t\t\tfullPort := portForward.Protocol + \":\" + hPort\n\t\t\t\t_, ok := hostPorts[fullPort]\n\t\t\t\tif ok {\n\t\t\t\t\treturn false, fmt.Errorf(\"Port %s has already been mapped\", fullPort)\n\t\t\t\t}\n\t\t\t\thostPorts[fullPort] = nil\n\t\t\t\tportForward.Name = container\n\t\t\t}\n\t\t}\n\t}\n\treturn true, nil\n}", "func (c *Config) Validate() (err error) {\n\t// Handle queue size\n\tif c.QueueSize <= 0 {\n\t\tc.QueueSize = defaultQueueSize\n\t}\n\n\t// Handle the number of workers\n\tif c.Workers <= 0 {\n\t\tc.Workers = runtime.NumCPU()\n\t}\n\n\t// Handle the addr\n\tif c.Addr == \"\" {\n\t\tc.Addr = defaultAddr\n\t}\n\n\t// Handle the metrics addr\n\tif c.MetricsAddr == \"\" {\n\t\tc.MetricsAddr = defaultMetricsAddr\n\t}\n\n\t// Handle the log level\n\tif c.LogLevel == \"\" {\n\t\tc.LogLevel = \"info\"\n\t} else {\n\t\tc.LogLevel = strings.ToLower(c.LogLevel)\n\t\tif _, ok := logLevels[c.LogLevel]; !ok {\n\t\t\treturn Errorf(ErrInvalidConfig, \"%q is an invalid log level, use trace, debug, info, caution, status, warn, or silent\", c.LogLevel)\n\t\t}\n\t}\n\tc.setLogLevel()\n\n\t// Handle the caution threshold\n\tif c.CautionThreshold == 0 {\n\t\tc.CautionThreshold = out.DefaultCautionThreshold\n\t}\n\tc.setCautionThreshold()\n\n\treturn nil\n}", "func validateConfig(args ArgCollection) error {\n\texpectedKeys := []string{}\n\tconfigFileRequired := false\n\tif args.ProviderFlag == \"custom\" {\n\t\tconfigFileRequired = true\n\t\texpectedKeys = append(expectedKeys, []string{\"providers.custom.insecure\", \"providers.custom.endpoint_format\", \"providers.custom.regions\", \"providers.custom.address_style\"}...)\n\t}\n\tif args.WriteToDB {\n\t\tconfigFileRequired = true\n\t\texpectedKeys = append(expectedKeys, []string{\"db.uri\"}...)\n\t}\n\tif args.UseMq {\n\t\tconfigFileRequired = true\n\t\texpectedKeys = append(expectedKeys, []string{\"mq.queue_name\", \"mq.uri\"}...)\n\t}\n\t// User didn't give any arguments that require the config file\n\tif !configFileRequired {\n\t\treturn nil\n\t}\n\n\t// Try to find and read config file\n\tif err := viper.ReadInConfig(); err != nil {\n\t\tif _, ok := err.(viper.ConfigFileNotFoundError); ok {\n\t\t\treturn errors.New(\"config file not found\")\n\t\t} else {\n\t\t\treturn err\n\t\t}\n\t}\n\n\t// Verify all expected keys are in the config file\n\tfor _, k := range expectedKeys {\n\t\tif !viper.IsSet(k) {\n\t\t\treturn fmt.Errorf(\"config file missing key: %s\", k)\n\t\t}\n\t}\n\treturn nil\n}", "func (c configuration) Validate() error {\n\tvar errs error\n\n\terrs = errors.Append(errs, c.Auth.Validate())\n\terrs = errors.Append(errs, c.Config.Validate())\n\n\tif c.Environment == \"\" {\n\t\terrs = errors.Append(errs, errors.New(\"environment is required\"))\n\t}\n\n\t// TODO: this config is only used here, so the validation is here too. Either the config or the validation should be moved somewhere else.\n\tif c.Distribution.PKE.Amazon.GlobalRegion == \"\" {\n\t\terrs = errors.Append(errs, errors.New(\"pke amazon global region is required\"))\n\t}\n\n\treturn errs\n}", "func (config StorageConfigStruct) Validate() error {\n\treturn nil\n}" ]
[ "0.6982893", "0.67954654", "0.67327917", "0.65675837", "0.6560407", "0.6475244", "0.64741606", "0.6416732", "0.6389593", "0.63582075", "0.6342166", "0.6282044", "0.6264434", "0.62190115", "0.6190914", "0.61886424", "0.61752874", "0.61643165", "0.61314374", "0.6065363", "0.6061954", "0.6032592", "0.6018692", "0.601484", "0.6006803", "0.6006371", "0.6003116", "0.6001448", "0.5997107", "0.59835184", "0.59798545", "0.5973025", "0.59688985", "0.5968867", "0.5968517", "0.59621924", "0.5957821", "0.59565663", "0.59427094", "0.5931806", "0.5930262", "0.5929543", "0.5921107", "0.59152794", "0.5913778", "0.59125125", "0.59114504", "0.590905", "0.5904003", "0.5900556", "0.5898378", "0.58954227", "0.58811885", "0.5879261", "0.5879261", "0.5879261", "0.5873312", "0.58590055", "0.58581686", "0.584782", "0.584782", "0.58375776", "0.58356726", "0.58304083", "0.58110565", "0.58109397", "0.5809816", "0.580944", "0.58093315", "0.5800911", "0.5796492", "0.5794249", "0.5785868", "0.57725745", "0.5772552", "0.5771563", "0.57696015", "0.5768524", "0.57637054", "0.57627934", "0.57596225", "0.575839", "0.5757816", "0.575475", "0.57406676", "0.573856", "0.57382", "0.57305086", "0.57250303", "0.5724823", "0.5706314", "0.570441", "0.5687705", "0.5682664", "0.5672296", "0.56670696", "0.56621116", "0.566108", "0.56584996", "0.56527084" ]
0.80435836
0
MAIN FUNCTIONS Handle Transaction Create
func (main *Main) Create(e echo.Context) (err error) { // get request and validate req := new(request.Create) e.Bind(req) if err = e.Validate(req); err != nil { return rest.ConstructErrorResponse(e, exception.NewInputValidationFailed(err.Error())) } // map req to input data reqData := input.NewNewTransactionCreate( map[string]interface{}{ "Name": req.Name, "Qty": req.Qty, "Price": req.Price, "Weight": req.Weight, "Images": req.Images, "Description": req.Description, }, ) //insert data to db transaction, exc := TransactionModel.Create(reqData) if exc != nil { return rest.ConstructErrorResponse(e, exc) } data := map[string]contract.Model{ "created_transaction": transaction, } return rest.ConstructSuccessResponse(e, data) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func TransactionCreate(c *gin.Context) {\n\tvar t models.Transaction\n\tbuffer, err := ioutil.ReadAll(c.Request.Body)\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusNotAcceptable, err)\n\t}\n\n\terr2 := jsonapi.Unmarshal(buffer, &t)\n\n\tif err2 != nil {\n\t\tparseFail := appError.JSONParseFailure\n\t\tparseFail.Detail = err2.Error()\n\t\tc.AbortWithError(http.StatusMethodNotAllowed, err2).\n\t\t\tSetMeta(parseFail)\n\t\treturn\n\t}\n\n\tt.CreatorID = c.Keys[\"CurrentUserID\"].(uint)\n\n\t// Validate our new transaction\n\tisValid, errApp := t.Validate()\n\n\tif isValid == false {\n\t\tc.AbortWithError(errApp.Status, errApp).\n\t\t\tSetMeta(errApp)\n\t\treturn\n\t}\n\n\tdatabase.DBCon.Create(&t)\n\n\tdatabase.DBCon.First(&t.Recipient, t.RecipientID)\n\tdatabase.DBCon.First(&t.Sender, t.SenderID)\n\tdatabase.DBCon.First(&t.Creator, t.CreatorID)\n\n\tdata, err := jsonapi.Marshal(&t)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tc.Data(http.StatusCreated, \"application/vnd.api+json\", data)\n}", "func CreateTransaction(resp http.ResponseWriter, req *http.Request) {\n\tresp.Header().Add(\"Content-Type\", \"application/json\")\n\tvar transaction domain.Transaction\n\tdecoder := json.NewDecoder(req.Body)\n\n\terr := decoder.Decode(&transaction)\n\n\tif err != nil {\n\t\thttp.Error(resp, \"Bad Request\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tapiErr := service.CreateTransaction(transaction)\n\n\tif apiErr != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tencoder := json.NewEncoder(resp)\n\t\tencoder.Encode(apiErr)\n\t}\n\n\tresp.WriteHeader(http.StatusOK)\n\n\tjsonValue, _ := json.Marshal(utils.AppMsgs{http.StatusOK, \"Transaction was recorded\"})\n\n\tresp.Write(jsonValue)\n\n}", "func (client *Client) CreateTransaction(txn *CreateTransaction) (_ *Response, err error) {\n\tpath := \"/transaction\"\n\turi := fmt.Sprintf(\"%s%s\", client.apiBaseURL, path)\n\n\tb, err := json.Marshal(txn)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treq, err := http.NewRequest(http.MethodPost, uri, bytes.NewBuffer(b))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tresp, err := client.performRequest(req, string(b))\n\treturn resp, err\n}", "func (db *Datastore) Create(txn transaction.Transaction) error {\n\tfmt.Println(`Creating txn with data:`, txn)\n\t// your DB operations to transactions ...\n\treturn nil\n}", "func createTransaction(request Request) (Transaction, error) {\n\tamount, err := strconv.ParseFloat(request.Load_amount[1:], 64)\n\ttransaction := Transaction{\n\t\tId: TransactionId(request.Id),\n\t\tCustomer_id: CustomerId(request.Customer_id),\n\t\tAmount: amount,\n\t\tTime: request.Time,\n\t}\n\n\treturn transaction, err\n}", "func (db *MySQLDB) CreateTransaction(ctx context.Context, rec *TransactionRecord) (int, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"CreateTransaction\")\n\n\t// start db transaction\n\ttx, err := db.instance.BeginTx(ctx, nil)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.BeginTx got %s\", err.Error())\n\t\treturn 0, err\n\t}\n\n\t// create transaction record\n\ttrans, err := tx.ExecContext(ctx, \"INSERT INTO transactions(user_id, date, grand_total, discount, reason) VALUES(?,?,?,?,?)\", rec.UserID, rec.Date, 0, rec.Discount, rec.Reason)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\ttID, err := trans.LastInsertId()\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\tgrandTotal := 0\n\n\t//loop tx detail\n\tfor i := 0; i < len(rec.TransactionDetail); i++ {\n\t\tdetail := rec.TransactionDetail[i]\n\n\t\t//get product stock from products table\n\t\tp, err := db.GetProductByID(ctx, detail.ProductID)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, err\n\t\t}\n\n\t\t//check qty\n\t\tif p.Qty-detail.Qty < 0 {\n\t\t\tfLog.Errorf(\"product qty is not enough\")\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, fmt.Errorf(\"product qty is not enough\")\n\t\t}\n\n\t\tqty := p.Qty - detail.Qty\n\t\tsubTotal := p.Price * detail.Qty\n\t\tgrandTotal = grandTotal + subTotal\n\n\t\t//update qty from products table\n\t\t_, err = tx.ExecContext(ctx, \"UPDATE products SET qty=? WHERE id=?\", qty, detail.ProductID)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, err\n\t\t}\n\n\t\t//insert transaction detail\n\t\t_, err = tx.ExecContext(ctx, \"INSERT INTO transaction_detail(transaction_id, product_id, price, qty, sub_total) VALUES(?,?,?,?,?)\", tID, detail.ProductID, p.Price, detail.Qty, subTotal)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\t\terrRollback := tx.Rollback()\n\t\t\tif errRollback != nil {\n\t\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\t\treturn 0, errRollback\n\t\t\t}\n\t\t\treturn 0, err\n\t\t}\n\t}\n\n\tgrandTotal = grandTotal - rec.Discount\n\n\t// update transaction grand total\n\t_, err = tx.ExecContext(ctx, \"UPDATE transactions SET grand_total=? WHERE id=?\", grandTotal, tID)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\t// commit transaction\n\terr = tx.Commit()\n\tif err != nil {\n\t\tfLog.Errorf(\"db.tx.ExecContext got %s\", err.Error())\n\t\terrRollback := tx.Rollback()\n\t\tif errRollback != nil {\n\t\t\tfLog.Errorf(\"error rollback, got %s\", err.Error())\n\t\t\treturn 0, errRollback\n\t\t}\n\t\treturn 0, err\n\t}\n\n\treturn int(tID), nil\n}", "func (facade *BranchManagerFacade) createTransaction(srcAccountId, destAccountId string, amount float32) *Transaction {\n\n\tvar transaction = facade.transaction.create(srcAccountId, destAccountId, amount)\n\n\treturn transaction\n}", "func CreateTransaction(transaction *models.Transaction) error {\n\tsession, collection := service.Connect(collectionName)\n\tdefer session.Close()\n\n\tif transaction.ID == \"\" {\n\t\ttransaction.ID = bson.NewObjectId()\n\t}\n\n\terr := collection.Insert(transaction)\n\n\treturn err\n}", "func (repository *transactionRepository) CreateTransaction(transaction model.Transaction) (err error) {\n\ttx, err := repository.dbAdapter.Begin()\n\n\tif err != nil {\n\t\treturn\n\t}\n\n\t_, err = tx.Exec(\n\t\t\"INSERT INTO transactions(id, account_id, operation_type, amount) VALUES(?,?,?,?)\",\n\t\ttransaction.GetID().GetValue(),\n\t\ttransaction.GetAccount().GetID().GetValue(),\n\t\ttransaction.GetOperationType().GetValue(),\n\t\ttransaction.GetAmountValueByOperationType(),\n\t)\n\n\tif err != nil {\n\t\ttx.Rollback()\n\n\t\treturn\n\t}\n\n\t_, err = tx.Exec(\n\t\t\"UPDATE accounts set credit_limit = ? where id = ?\",\n\t\ttransaction.GetAccount().GetAvailableCreditLimit().GetValue(),\n\t\ttransaction.GetAccount().GetID().GetValue(),\n\t)\n\n\tif err != nil {\n\t\ttx.Rollback()\n\n\t\treturn\n\t}\n\n\terr = tx.Commit()\n\n\treturn\n}", "func CreateTransaction(from uint, to uint, amount int) {\n\tdb := helpers.ConnectDB().Debug()\n\tdefer db.Close()\n\n\ttransaction := &interfaces.Transaction{From: from, To: to, Amount: amount}\n\tdb.Create(&transaction)\n}", "func (a API) CreateRawTransaction(cmd *btcjson.CreateRawTransactionCmd) (e error) {\n\tRPCHandlers[\"createrawtransaction\"].Call <-API{a.Ch, cmd, nil}\n\treturn\n}", "func CreateTransactions(\n\trouter *mux.Router,\n\tpk cryptography.PrivateKey,\n\tuser users.User,\n\tserverSDK sdks.Servers,\n\troutePrefix string,\n\tsignedTrsBufferSize int,\n\tatomicTrsBufferSize int,\n\taggregatedTrsBufferSize int,\n\ttrsAggregationDelay time.Duration,\n) *Transactions {\n\n\t//channels:\n\tnewSignedTrs := make(chan signed_transactions.Transaction, signedTrsBufferSize)\n\tnewAtomicSignedTrs := make(chan signed_transactions.AtomicTransaction, atomicTrsBufferSize)\n\tnewAggregatedTrs := make(chan aggregated_transactions.Transactions, aggregatedTrsBufferSize)\n\n\t//factories:\n\tmetaDataBuilderFactory := concrete_metadata.CreateBuilderFactory()\n\thtBuilderFactory := concrete_hashtrees.CreateHashTreeBuilderFactory()\n\tpublicKeyBuilderFactory := concrete_cryptography.CreatePublicKeyBuilderFactory()\n\tsigBuilderFactory := concrete_cryptography.CreateSignatureBuilderFactory(publicKeyBuilderFactory)\n\tuserSigBuilderFactory := concrete_users.CreateSignatureBuilderFactory(sigBuilderFactory, htBuilderFactory, metaDataBuilderFactory)\n\n\t//transactions and blocks factories:\n\tblockChainMetaDataBuilderFactory := concrete_blockchain_metadata.CreateBuilderFactory()\n\tsignedTrsBuilderFactory := concrete_signed_transactions.CreateTransactionBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tsignedTransBuilderFactory := concrete_signed_transactions.CreateTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tsignedAtomicTransBuilderFactory := concrete_signed_transactions.CreateAtomicTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tatomicSignedTrsBuilderFactory := concrete_signed_transactions.CreateAtomicTransactionBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\tsignedAggregatedTrsBuilderFactory := concrete_aggregated_transactions.CreateTransactionsBuilderFactory(htBuilderFactory, blockChainMetaDataBuilderFactory)\n\n\t//create the leader SDK:\n\tleadSDK := concrete_sdks.CreateLeaders(userSigBuilderFactory, routePrefix, pk, user)\n\n\t//create the transaction API:\n\ttransactionsAPI := apis.CreateTransactions(\n\t\troutePrefix,\n\t\trouter,\n\t\tsignedTrsBuilderFactory,\n\t\tatomicSignedTrsBuilderFactory,\n\t\tnewSignedTrs,\n\t\tnewAtomicSignedTrs,\n\t)\n\n\t//create the transaction agent:\n\ttrsAgent := agents.CreatePushTransactionsToLeaders(\n\t\tsignedAggregatedTrsBuilderFactory,\n\t\tsignedTransBuilderFactory,\n\t\tsignedAtomicTransBuilderFactory,\n\t\ttrsAggregationDelay,\n\t\tnewSignedTrs,\n\t\tnewAtomicSignedTrs,\n\t\tnewAggregatedTrs,\n\t)\n\n\tout := Transactions{\n\t\tapi: transactionsAPI,\n\t\tagent: trsAgent,\n\t\tleadSDK: leadSDK,\n\t\tservSDK: serverSDK,\n\t\tnewAggregatedTrs: newAggregatedTrs,\n\t}\n\n\treturn &out\n}", "func (u *User) CreateTransaction(nodeID, data string, idempotencyKey ...string) (map[string]interface{}, error) {\n\tlog.info(\"========== CREATE TRANSACTION ==========\")\n\turl := buildURL(path[\"users\"], u.UserID, path[\"nodes\"], nodeID, path[\"transactions\"])\n\n\treturn u.do(\"POST\", url, data, idempotencyKey)\n}", "func (f formatter) CreateTransaction(t bill.Transaction, ins, outs []string) (string, error) {\n\titem := transactionItem{\n\t\tRaw: t.Raw,\n\t\tAt: t.At.Format(\"2006-01-02\"),\n\t\tPayee: t.Payee,\n\t\tTitle: t.Title,\n\t\tPayeeAccounts: ins,\n\t\tAccounts: outs,\n\t\tAmount: t.Amount,\n\t\tCurrency: t.Currency,\n\t}\n\tif item.PayeeAccounts == nil || len(item.PayeeAccounts) == 0 {\n\t\titem.PayeeAccounts = []string{\"TODO\"}\n\t}\n\tif item.Accounts == nil || len(item.Accounts) == 0 {\n\t\titem.Accounts = []string{\"TODO\"}\n\t}\n\n\tvar tplBytes bytes.Buffer\n\ttpl, err := template.New(\"transaction\").Parse(transactionTpl)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"template error, should be a bug: %w\", err)\n\t}\n\n\terr = tpl.Execute(&tplBytes, item)\n\tif err != nil {\n\t\treturn \"\", fmt.Errorf(\"template exec error, should be a bug: %w\", err)\n\t}\n\n\treturn tplBytes.String(), nil\n}", "func (transaction *Transaction) create(srcAccountId, destAccountId string, amount float32) *Transaction {\n\n\tfmt.Println(\"creating transaction\")\n\ttransaction.srcAccountId = srcAccountId\n\ttransaction.destAccountId = destAccountId\n\ttransaction.amount = amount\n\treturn transaction\n}", "func createTransaction(\n\tctx context.Context,\n\tdb storage.Database,\n\tappserviceID string,\n) (\n\ttransactionJSON []byte,\n\ttxnID, maxID int,\n\teventsRemaining bool,\n\terr error,\n) {\n\t// Retrieve the latest events from the DB (will return old events if they weren't successfully sent)\n\ttxnID, maxID, events, eventsRemaining, err := db.GetEventsWithAppServiceID(ctx, appserviceID, transactionBatchSize)\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\n\t\t\t\"appservice\": appserviceID,\n\t\t}).WithError(err).Fatalf(\"appservice worker unable to read queued events from DB\")\n\n\t\treturn\n\t}\n\n\t// Check if these events do not already have a transaction ID\n\tif txnID == -1 {\n\t\t// If not, grab next available ID from the DB\n\t\ttxnID, err = db.GetLatestTxnID(ctx)\n\t\tif err != nil {\n\t\t\treturn nil, 0, 0, false, err\n\t\t}\n\n\t\t// Mark new events with current transactionID\n\t\tif err = db.UpdateTxnIDForEvents(ctx, appserviceID, maxID, txnID); err != nil {\n\t\t\treturn nil, 0, 0, false, err\n\t\t}\n\t}\n\n\tvar ev []*gomatrixserverlib.HeaderedEvent\n\tfor i := range events {\n\t\tev = append(ev, &events[i])\n\t}\n\n\t// Create a transaction and store the events inside\n\ttransaction := gomatrixserverlib.ApplicationServiceTransaction{\n\t\tEvents: gomatrixserverlib.HeaderedToClientEvents(ev, gomatrixserverlib.FormatAll),\n\t}\n\n\ttransactionJSON, err = json.Marshal(transaction)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (tm *TransactionMutex) Create(transactionType TransactionType, fileUUID string) {\n\ttm.mu.Lock()\n\tdefer tm.mu.Unlock()\n\tnewTransaction := Transaction{\n\t\tUUID: NewUUID(),\n\t\tCreationTimestamp: time.Now().Unix(),\n\t\tType: transactionType,\n\t\tTargetFileUUID: fileUUID,\n\t\tVersion: config.Version,\n\t}\n\ttm.Transactions = append(tm.Transactions, newTransaction)\n}", "func (tr *Transactor) CreateTransaction(ctx context.Context, ipfsHash string, tags []Tag, w WalletSigner, amount string, data []byte, target string) (*Transaction, error) {\n\tlastTx, err := tr.Client.TxAnchor(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tprice, err := tr.Client.GetReward(ctx, data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Non encoded transaction fields\n\ttx := NewTransaction(\n\t\ttags,\n\t\tipfsHash,\n\t\tlastTx,\n\t\tw.PubKeyModulus(),\n\t\tamount,\n\t\ttarget,\n\t\tdata,\n\t\tprice,\n\t)\n\n\treturn tx, nil\n}", "func CreateTransaction(session gocqlx.Session, transaction *Transaction) (*Transaction, error) {\n\tuuid, _ := gocql.RandomUUID()\n\ttransaction.ID = uuid\n\ttransaction.CreatedAt = time.Now()\n\n\tlog.Print(transaction)\n\n\tq := session.Query(transactionTable.Insert()).BindStruct(transaction)\n\tif err := q.ExecRelease(); err != nil {\n\t\treturn transaction, err\n\t}\n\treturn transaction, nil\n}", "func (u *createTrans) CreateTransaction(transaction *models.Transaction) error {\n\n\tacc, err := u.accountServ.Get(transaction.Account.ID)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tnewCredit := acc.AvailableCreditLimit + transaction.Amount\n\tif newCredit < 0 {\n\t\treturn models.ErrAmountBeyondCreditLimit\n\t}\n\n\terr = u.tranServ.Create(transaction)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tacc.AvailableCreditLimit = newCredit\n\treturn u.accountServ.UpdateCreditLimit(acc)\n}", "func (s *TransactionService) Create(walletID string, tp *TransactionParams) (*Transaction, error) {\n\tu := fmt.Sprintf(\"/kms/wallets/%s/transactions/\", walletID)\n\ttransaction := &Transaction{}\n\tp := &Params{}\n\tp.SetAuthProvider(s.auth)\n\terr := s.client.Call(http.MethodPost, u, tp, transaction, p)\n\treturn transaction, err\n}", "func (s *Service) CreateTx(tx *sql.Tx, headerID uint, ms Models) error {\n\treturn s.storage.CreateTx(tx, headerID, ms)\n}", "func addTransaction(db meddler.DB, lender *User, debtor *User, amount int, expense *Expense) (*Transaction, error) {\n trans := new(Transaction)\n trans.LenderId = lender.Id\n trans.DebtorId = debtor.Id\n trans.Amount = amount\n trans.Date = expense.Date\n trans.ExpenseId = expense.Id\n\n err := meddler.Insert(db, \"transactions\", trans)\n if err != nil {\n return nil, err\n }\n\n lender.UpdateBalance(db, amount)\n debtor.UpdateBalance(db, -amount)\n\n return trans, nil\n}", "func (c Client) createTransaction(merchantID string, body io.Reader) (*TransactionID, error) {\n\tpath := fmt.Sprintf(\"/merchants/%s/transactions\", merchantID)\n\treq, err := http.NewRequest(\"POST\", c.getURL(path), body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tvar marshalled map[string]*TransactionID\n\treturn marshalled[\"transaction\"], c.executeRequestAndMarshal(req, &marshalled)\n}", "func (h *Harness) CreateTransaction(\n\ttargetOutputs []*wire.TxOut,\n\tfeeRate amt.Amount, change bool,\n) (*wire.MsgTx, error) {\n\treturn h.wallet.CreateTransaction(targetOutputs, feeRate, change)\n}", "func CreateTransaction(from, to, amount string) (txeB64 string) {\n\n\ttx, err := build.Transaction(\n\t\tbuild.SourceAccount{AddressOrSeed: from},\n\t\tbuild.PublicNetwork,\n\t\t//b.AutoSequence{SequenceProvider: horizon.DefaultPublicNetClient}, ???\n\t\tbuild.Payment(\n\t\t\tbuild.Destination{AddressOrSeed: to},\n\t\t\tbuild.NativeAmount{Amount: amount},\n\t\t),\n\t)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\ttxe, err := tx.Sign(from)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\ttxeB64, err = txe.Base64()\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\n\treturn\n}", "func Transaction(c *gin.Context) {\n\n\tt_type,_ := strconv.Atoi(c.PostForm(\"transaction_type\")) // 1 : sales , 2 : missing products (hilang)\n\tstatus := 1\n\tmessage := \"Success\"\n var responseTransaction ResponseTransaction\n\tvar newstocks int\n\tvar products Products\n\tvar products_arr []Products\n\tvar stock_ins_arr []Stock_Ins\n\tvar stock_outs Stock_Outs\n\tvar stock_ins Stock_Ins\n\tvar note string\n\ttransaction_id := \"\"\n\tsellPrice,_ := strconv.Atoi(c.PostForm(\"sell_price\"))\n\tvar buyPrice int\n\tqtY,_ := strconv.Atoi(c.PostForm(\"qty\"))\n\tcurrentdatetime := time.Now().Format(\"2006-01-02 15:04:05\")\n\tdb := InitDb() //db intiate\n\t//get data products\n\tdb.Where(\"sku = ?\", c.PostForm(\"sku\")).First(&products).Limit(1).Scan(&products_arr)\n\n\t//check if the sku is exist?\n\tif(len(products_arr) > 0) {\n\t\ttx := db.Begin()\n\n\t\t/**\n\t * Identify product is gone / transaction by sales\n\t */\n\n\t\tif (t_type == 1) {\n\n\t\t\ttransaction_id = generateTransactionID()\n\n\t\t\t//get data products\n\t\t\tdb.Where(\"sku = ?\", c.PostForm(\"sku\")).First(&stock_ins).Limit(1).Scan(&stock_ins_arr)\n\n\t\t\t// get the data stock after transaction\n\t\t\tfor i,element := range stock_ins_arr{\n\t\t\t\tif (i == 0) {\n\t\t\t\t\tbuyPrice = element.Buy_Price\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tnote = \"Pesanan \"+transaction_id\n\t\t\ttransactions := Transactions{Id:transaction_id,Buy_Price:buyPrice,Sell_Price:sellPrice,Qty:qtY,Sku:c.PostForm(\"sku\"),Created_Date:currentdatetime}\n\t\t\tif err := tx.Create(&transactions).Error; err != nil {\n\t\t\t\ttx.Rollback()\n\t\t\t\tstatus = 0\n\t\t\t\tmessage = \"failed to insert data transaction\"\n\t\t\t}\n\n\n\t\t} else if (t_type == 2) {\n\n\t\t\tnote = \"Barang Hilang\"\n\n\t\t}\n\t\t//insert data to stock_outs\n\t\tstock_outs = Stock_Outs{Sku:c.PostForm(\"sku\"),Created_Date:currentdatetime,Qty:qtY,Note:note,Transaction_Id:transaction_id}\n\t\tif err := tx.Create(&stock_outs).Error; err != nil {\n\t\t\ttx.Rollback()\n\t\t\tstatus = 0\n\t\t\tmessage = \"failed to insert data stocks_outs\"\n\t\t}\n\n\t\t// get the data stock after transaction\n\t\tfor i,element := range products_arr{\n\t\t\tif (i == 0) {\n\t\t\t\tnewstocks = element.Stocks - qtY\n\t\t\t}\n\t\t}\n\n\t\t//update product stocks in table products\n\t\tif err := tx.Model(&products).Where(\"sku = ?\", c.PostForm(\"sku\")).Update(\"stocks\", newstocks).Error; err != nil {\n\t\t\ttx.Rollback()\n\t\t\tstatus = 0\n\t\t\tmessage = \"failed to update data products\"\n\t\t}\n\n\n\t\t//transaction commit\n\t\ttx.Commit()\n\t}else{\n\t\tstatus = 0\n\t\tmessage = \"SKU Not found!\"\n\t}\n\n\tif status == 1{\n\t\tresponseTransaction = ResponseTransaction{Status:status,Message:message,Data:DataTransaction{Sku:c.PostForm(\"sku\"),Buy_Price:buyPrice,Sell_Price:sellPrice,Created_Date:currentdatetime,Product_name:c.PostForm(\"product_name\"),Stocks:newstocks,Transaction_Id:transaction_id}}\n\t}else{\n\t\tresponseTransaction = ResponseTransaction{Status:status,Message:message}\n\t}\n\n\t// Close connection database\n\tdefer db.Close()\n\tc.JSON(200, responseTransaction)\n}", "func (c *Constructor) createTransaction(\n\tctx context.Context,\n\tintent []*types.Operation,\n) (*types.TransactionIdentifier, string, error) {\n\tmetadataRequest, err := c.helper.Preprocess(\n\t\tctx,\n\t\tc.network,\n\t\tintent,\n\t\tnil,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to preprocess\", err)\n\t}\n\n\trequiredMetadata, err := c.helper.Metadata(\n\t\tctx,\n\t\tc.network,\n\t\tmetadataRequest,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to construct metadata\", err)\n\t}\n\n\tunsignedTransaction, payloads, err := c.helper.Payloads(\n\t\tctx,\n\t\tc.network,\n\t\tintent,\n\t\trequiredMetadata,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to construct payloads\", err)\n\t}\n\n\tparsedOps, signers, _, err := c.helper.Parse(\n\t\tctx,\n\t\tc.network,\n\t\tfalse,\n\t\tunsignedTransaction,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to parse unsigned transaction\", err)\n\t}\n\n\tif len(signers) != 0 {\n\t\treturn nil, \"\", fmt.Errorf(\n\t\t\t\"signers should be empty in unsigned transaction but found %d\",\n\t\t\tlen(signers),\n\t\t)\n\t}\n\n\tif err := c.parser.ExpectedOperations(intent, parsedOps, false, false); err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unsigned parsed ops do not match intent\", err)\n\t}\n\n\tsignatures, err := c.helper.Sign(ctx, payloads)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to sign payloads\", err)\n\t}\n\n\tnetworkTransaction, err := c.helper.Combine(\n\t\tctx,\n\t\tc.network,\n\t\tunsignedTransaction,\n\t\tsignatures,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to combine signatures\", err)\n\t}\n\n\tsignedParsedOps, signers, _, err := c.helper.Parse(\n\t\tctx,\n\t\tc.network,\n\t\ttrue,\n\t\tnetworkTransaction,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to parse signed transaction\", err)\n\t}\n\n\tif err := c.parser.ExpectedOperations(intent, signedParsedOps, false, false); err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: signed parsed ops do not match intent\", err)\n\t}\n\n\tif err := parser.ExpectedSigners(payloads, signers); err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: signed transactions signers do not match intent\", err)\n\t}\n\n\ttransactionIdentifier, err := c.helper.Hash(\n\t\tctx,\n\t\tc.network,\n\t\tnetworkTransaction,\n\t)\n\tif err != nil {\n\t\treturn nil, \"\", fmt.Errorf(\"%w: unable to get transaction hash\", err)\n\t}\n\n\treturn transactionIdentifier, networkTransaction, nil\n}", "func (service ServiceTx) Create(edgeJob *portainer.EdgeJob) error {\n\treturn service.CreateWithID(portainer.EdgeJobID(service.GetNextIdentifier()), edgeJob)\n}", "func (c *TransactionClient) Create() *TransactionCreate {\n\tmutation := newTransactionMutation(c.config, OpCreate)\n\treturn &TransactionCreate{config: c.config, hooks: c.Hooks(), mutation: mutation}\n}", "func (gw *Gateway) CreateTransaction(params wallet.CreateTransactionParams) (*coin.Transaction, []wallet.UxBalance, error) {\n\tif !gw.Config.EnableWalletAPI {\n\t\treturn nil, nil, wallet.ErrWalletAPIDisabled\n\t}\n\n\tvar txn *coin.Transaction\n\tvar inputs []wallet.UxBalance\n\tvar err error\n\tgw.strand(\"CreateTransaction\", func() {\n\t\ttxn, inputs, err = gw.v.CreateTransaction(params)\n\t})\n\treturn txn, inputs, err\n}", "func (p *PsqlInvoiceHeader) CreateTx(tx *sql.Tx, m *invoiceheader.Model) error {\n\tstmt, err := tx.Prepare(psqlCreateInvoiceHeader)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdefer stmt.Close()\n\n\treturn stmt.QueryRow(m.Client).Scan(&m.ID, &m.CreatedAt)\n}", "func (c Client) CreateTransaction(merchantID string, dto TransactionDTO) (*TransactionID, error) {\n\tb, err := json.Marshal(dto)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\treturn c.createTransaction(merchantID, bytes.NewBuffer(b))\n}", "func (t *explainTablet) CreateTransaction(ctx context.Context, target *querypb.Target, dtid string, participants []*querypb.Target) (err error) {\n\tt.mu.Lock()\n\tt.currentTime = t.vte.batchTime.Wait()\n\tt.mu.Unlock()\n\treturn t.tsv.CreateTransaction(ctx, target, dtid, participants)\n}", "func (u *User) CreateDummyTransaction(nodeID string, queryParams ...string) (map[string]interface{}, error) {\n\tlog.info(\"========== CREATE DUMMY TRANSACTION ==========\")\n\turl := buildURL(path[\"users\"], u.UserID, path[\"nodes\"], nodeID) + \"/dummy-tran\"\n\n\treturn u.do(\"GET\", url, \"\", queryParams)\n}", "func (tx *Tx) Create(ls Lease) error {\n\ttx.leases = append(tx.leases, ls)\n\ttx.ops = append(tx.ops, Op{\n\t\tType: Create,\n\t\tLease: ls,\n\t})\n\tsort.Sort(tx.leases)\n\treturn nil\n}", "func (m *Mysql) CreateTX(tx *sql.Tx, paramAudit *paramAuditDom.ParamsAudit) error {\n\tstmt, err := tx.Prepare(mysqlInsert)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\tnow := utils.Now()\n\terr = mysql.ExecAffectingOneRow(\n\t\tstmt,\n\t\tparamAudit.PrevParam.ParamID,\n\t\tparamAudit.PrevParam.Name,\n\t\tparamAudit.PrevParam.Value,\n\t\tparamAudit.PrevParam.Active,\n\t\tparamAudit.Param.Name,\n\t\tparamAudit.Param.Value,\n\t\tparamAudit.Param.Active,\n\t\tparamAudit.SetUserID,\n\t\tnow[\"date\"], now[\"time\"], now[\"unix\"],\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (nos *NewOrderServiceImpl) ProcessTransaction(req *models.NewOrder) (*models.NewOrderOutput, error) {\n\tlog.Printf(\"Starting the New Order Transaction for row: c=%d w=%d d=%d n=%d\", req.CustomerID, req.WarehouseID, req.DistrictID, req.UniqueItems)\n\n\tresult, err := nos.execute(req)\n\tif err != nil {\n\t\treturn nil, fmt.Errorf(\"error occured while executing the new order transaction. Err: %v\", err)\n\t}\n\n\tlog.Printf(\"Completed the New Order Transaction for row: c=%d w=%d d=%d n=%d\", req.CustomerID, req.WarehouseID, req.DistrictID, req.UniqueItems)\n\treturn result, nil\n}", "func (c *Jrpc) CreateRawTransaction(in *pty.ReqCreatePrivacyTx, result *interface{}) error {\n\treply, err := c.cli.CreateRawTransaction(context.Background(), in)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t*result = hex.EncodeToString(types.Encode(reply))\n\treturn err\n}", "func (b *Store) CreateTx(ctx context.Context, tx *sql.Tx, userID, username, password string) error {\n\terr := permission.LimitCheckAny(ctx, permission.System, permission.Admin, permission.MatchUser(userID))\n\tif err != nil {\n\t\treturn err\n\t}\n\n\terr = validate.Many(\n\t\tvalidate.UUID(\"UserID\", userID),\n\t\tvalidate.UserName(\"UserName\", username),\n\t\tvalidate.Text(\"Password\", password, 8, 200),\n\t)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\thashedPassword, err := bcrypt.GenerateFromPassword([]byte(password), passCost)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = tx.StmtContext(ctx, b.insert).ExecContext(ctx, userID, username, string(hashedPassword))\n\treturn err\n}", "func (s *transactionStore) Create(ctx context.Context, transaction *configapi.Transaction) error {\n\tif transaction.ID == \"\" {\n\t\ttransaction.ID = newTransactionID()\n\t}\n\tif transaction.Version != 0 {\n\t\treturn errors.NewInvalid(\"not a new object\")\n\t}\n\tif transaction.Revision != 0 {\n\t\treturn errors.NewInvalid(\"not a new object\")\n\t}\n\ttransaction.Revision = 1\n\ttransaction.Created = time.Now()\n\ttransaction.Updated = time.Now()\n\n\t// Append a new entry to the transaction log.\n\tentry, err := s.transactions.Append(ctx, transaction.ID, transaction)\n\tif err != nil {\n\t\treturn errors.FromAtomix(err)\n\t}\n\ttransaction.Index = configapi.Index(entry.Index)\n\ttransaction.Version = uint64(entry.Version)\n\treturn nil\n}", "func (c *Constructor) CreateTransactions(\n\tctx context.Context,\n\tclearBroadcasts bool,\n) error {\n\t// Before starting loop, delete any pending broadcasts if configuration\n\t// indicates to do so.\n\tif clearBroadcasts {\n\t\tbroadcasts, err := c.helper.ClearBroadcasts(ctx)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to clear broadcasts\", err)\n\t\t}\n\n\t\tlog.Printf(\n\t\t\t\"Cleared pending %d broadcasts: %s\\n\",\n\t\t\tlen(broadcasts),\n\t\t\ttypes.PrettyPrintStruct(broadcasts),\n\t\t)\n\t}\n\n\tfor ctx.Err() == nil {\n\t\tsender, balance, coinIdentifier, err := c.findSender(ctx)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to find sender\", err)\n\t\t}\n\n\t\t// Determine Action\n\t\tscenarioCtx, scenarioOps, err := c.generateScenario(\n\t\t\tctx,\n\t\t\tsender,\n\t\t\tbalance,\n\t\t\tcoinIdentifier,\n\t\t)\n\t\tif errors.Is(err, ErrInsufficientFunds) {\n\t\t\tbroadcasts, err := c.helper.AllBroadcasts(ctx)\n\t\t\tif err != nil {\n\t\t\t\treturn fmt.Errorf(\"%w: unable to get broadcasts\", err)\n\t\t\t}\n\n\t\t\tif len(broadcasts) > 0 {\n\t\t\t\t// we will wait for in-flight to process\n\t\t\t\ttime.Sleep(defaultSleepTime * time.Second)\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tif err := c.generateNewAndRequest(ctx); err != nil {\n\t\t\t\treturn fmt.Errorf(\"%w: unable to generate new address\", err)\n\t\t\t}\n\n\t\t\tcontinue\n\t\t} else if err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to generate intent\", err)\n\t\t}\n\n\t\tintent, err := scenario.PopulateScenario(ctx, scenarioCtx, scenarioOps)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to populate scenario\", err)\n\t\t}\n\n\t\t// Create transaction\n\t\ttransactionIdentifier, networkTransaction, err := c.createTransaction(ctx, intent)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\n\t\t\t\t\"%w: unable to create transaction with operations %s\",\n\t\t\t\terr,\n\t\t\t\ttypes.PrettyPrintStruct(intent),\n\t\t\t)\n\t\t}\n\n\t\tlogger.LogScenario(scenarioCtx, transactionIdentifier, c.currency)\n\n\t\t// Broadcast Transaction\n\t\terr = c.helper.Broadcast(\n\t\t\tctx,\n\t\t\tsender,\n\t\t\tintent,\n\t\t\ttransactionIdentifier,\n\t\t\tnetworkTransaction,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to enqueue transaction for broadcast\", err)\n\t\t}\n\n\t\tif err := c.handler.TransactionCreated(ctx, sender, transactionIdentifier); err != nil {\n\t\t\treturn fmt.Errorf(\"%w: unable to handle transaction creation\", err)\n\t\t}\n\t}\n\n\treturn ctx.Err()\n}", "func Transaction(rt *Runtime, c chan goengage.Fundraise) (err error) {\n\trt.Log.Println(\"Transaction: start\")\n\tfor true {\n\t\tr, ok := <-c\n\t\tif !ok {\n\t\t\tbreak\n\t\t}\n\n\t\trt.Log.Printf(\"%v Transaction\\n\", r.ActivityID)\n\t\tif rt.GoodYear(r.ActivityDate) {\n\t\t\tif len(r.Transactions) != 0 {\n\t\t\t\tfor _, c := range r.Transactions {\n\t\t\t\t\tc.ActivityID = r.ActivityID\n\t\t\t\t\trt.DB.Create(&c)\n\t\t\t\t}\n\t\t\t}\n\t\t}\n\t}\n\trt.Log.Println(\"Transaction: end\")\n\treturn nil\n}", "func (a *ChainAdaptor) CreateUtxoTransaction(req *proto.CreateUtxoTransactionRequest) (*proto.CreateUtxoTransactionReply, error) {\n\tvinNum := len(req.Vins)\n\tvar totalAmountIn, totalAmountOut int64\n\n\tif vinNum == 0 {\n\t\terr := fmt.Errorf(\"no Vin in req:%v\", req)\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\t// check the Fee\n\tfee, ok := big.NewInt(0).SetString(req.Fee, 0)\n\tif !ok {\n\t\terr := errors.New(\"CreateTransaction, fail to get fee\")\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\tfor _, in := range req.Vins {\n\t\ttotalAmountIn += in.Amount\n\t}\n\n\tfor _, out := range req.Vouts {\n\t\ttotalAmountOut += out.Amount\n\t}\n\n\tif totalAmountIn != totalAmountOut+fee.Int64() {\n\t\terr := errors.New(\"CreateTransaction, total amount in != total amount out + fee\")\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\trawTx, err := a.createRawTx(req.Vins, req.Vouts)\n\tif err != nil {\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\tbuf := bytes.NewBuffer(make([]byte, 0, rawTx.SerializeSize()))\n\terr = rawTx.Serialize(buf)\n\tif err != nil {\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\t// build the pkScript and Generate signhash for each Vin,\n\tsignHashes, err := a.calcSignHashes(req.Vins, req.Vouts)\n\tif err != nil {\n\t\treturn &proto.CreateUtxoTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\tlog.Info(\"CreateTransaction\", \"usigned tx\", hex.EncodeToString(buf.Bytes()))\n\n\treturn &proto.CreateUtxoTransactionReply{\n\t\tCode: proto.ReturnCode_SUCCESS,\n\t\tTxData: buf.Bytes(),\n\t\tSignHashes: signHashes,\n\t}, nil\n}", "func (c *RPC) CreateTransaction(recvAddr string, amount uint64) (*coin.Transaction, error) {\n\t// TODO -- this can support sending to multiple receivers at once,\n\t// which would be necessary if the exchange was busy\n\tsendAmount := cli.SendAmount{\n\t\tAddr: recvAddr,\n\t\tCoins: amount,\n\t}\n\n\tif err := validateSendAmount(sendAmount); err != nil {\n\t\treturn nil, err\n\t}\n\n\ttxn, err := cli.CreateRawTxFromWallet(c.rpcClient, c.walletFile, c.changeAddr, []cli.SendAmount{sendAmount})\n\tif err != nil {\n\t\treturn nil, RPCError{err}\n\t}\n\n\treturn txn, nil\n}", "func Db_create(db_name string) string{\n\t// Open the my.db data_backup file in your current directory.\n\t// It will be created if it doesn't exist.\n\tdb, err := bolt.Open(db_name, 0600, nil)\n\t//defer db.Close()\n\tif err != nil {\n\t\tlog.Fatal(err)\n\t}\n\ttx, err := db.Begin(true)\n\tif err != nil {\n\t\tfmt.Println(\"Error : \",err)\n\t\tdb.Close()\n\t\treturn \"ERROR,DB_COMMIT\"\n\t}\n\tfmt.Printf(\"Created DB with ID %d\", tx.ID())\n\tdb.Close()\n\treturn \"OK,DB_COMMIT\"\n}", "func (t *SimpleChaincode) create_event(stub shim.ChaincodeStubInterface, args []string) ([]byte, error) {\n\tvar tEvent TransactionEvent\n\t\n\ttranID \t\t\t:= \"\\\"TranID\\\":\\\"\"+args[0]+\"\\\", \"\n\tsenderName \t\t:= \"\\\"SenderName\\\":\\\"\"+args[1]+\"\\\", \"\n\tsenderCountry := \"\\\"SenderCountry\\\":\\\"\"+args[2]+\"\\\", \"\n\treceiverName \t:= \"\\\"ReceiverName\\\":\\\"\"+args[3]+\"\\\", \"\n\treceiverCountry := \"\\\"ReceiverCountry\\\":\\\"\"+args[4]+\"\\\", \"\n\tamount \t\t\t:= \"\\\"Amount\\\":\\\"\"+args[5]+\"\\\", \"\n\n // Concatenates the variables to create the total JSON object\n\tevent_json := \"{\"+tranID+senderName+senderCountry+receiverName+receiverCountry+amount+\"}\" \t\t\n\t// Convert the JSON defined above into a TransactionEvent object for go\n\terr := json.Unmarshal([]byte(event_json), &tEvent)\t\t\t\t\t\t\t\t\t\t\n\tif err != nil { \n\t\treturn nil, errors.New(\"Invalid JSON object\") \n\t}\n\n\tbytes, err := json.Marshal(tEvent)\n\tif err != nil { \n\t\treturn nil, errors.New(\"Error converting transaction event\") \n\t}\n\n\t// Save new tran event record\n\terr = stub.PutState(tEvent.TranID, bytes)\n\tif err != nil { \n\t\tfmt.Printf(\"create_event: Error storing transaction event: %s\", err); \n\t\treturn nil, errors.New(\"Error storing transaction event\") \n\t}\n\n\t// Update tranIDs with newly created ID and store it in chain.\n\tbytes, err = stub.GetState(\"tranIDs\")\n\tif err != nil { \n\t\treturn nil, errors.New(\"Unable to get tranIDs\") \n\t}\n\n\tvar tranHld TRAN_Holder\n\terr = json.Unmarshal(bytes, &tranHld)\n\tif err != nil {\t\n\t\treturn nil, errors.New(\"Corrupt TRAN_Holder record\") \n\t}\n\n\ttranHld.TranIDs = append(tranHld.TranIDs, args[0])\n\tbytes, err = json.Marshal(tranHld)\n\n\terr = stub.PutState(\"tranIDs\", bytes)\n\tif err != nil { \n\t\tfmt.Printf(\"create_event: Error storing TranIDs: %s\", err); \n\t\treturn nil, errors.New(\"Error storing TranIDs\") \n\t}\n\n\treturn nil, nil \n}", "func (t *Procure2Pay) CreatePurchaseOrder(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n var objpurchaseOrder purchaseOrder\n\tvar objitem item\n\tvar err error\n\t\n\tfmt.Println(\"Entering CreatePurchaseOrder\")\n\n\tif len(args) < 1 {\n\t\tfmt.Println(\"Invalid number of args\")\n\t\treturn shim.Error(err.Error())\n\t}\n\n\tfmt.Println(\"Args [0] is : %v\\n\", args[0])\n\n\t//unmarshal customerInfo data from UI to \"customerInfo\" struct\n\terr = json.Unmarshal([]byte(args[0]), &objpurchaseOrder)\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to unmarshal CreatePurchaseOrder input purchaseOrder: %s\\n\", err)\n\t\treturn shim.Error(err.Error())\n\t\t}\n\n\tfmt.Println(\"purchase order object PO ID variable value is : %s\\n\", objpurchaseOrder.POID)\n\tfmt.Println(\"purchase order object PO ID variable value is : %s\\n\", objpurchaseOrder.Quantity)\n\n\t// Data insertion for Couch DB starts here \n\ttransJSONasBytes, err := json.Marshal(objpurchaseOrder)\n\terr = stub.PutState(objpurchaseOrder.POID, transJSONasBytes)\n\t// Data insertion for Couch DB ends here\n\n\t//unmarshal LoanTransactions data from UI to \"LoanTransactions\" struct\n\terr = json.Unmarshal([]byte(args[0]), &objitem)\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to unmarshal CreatePurchaseOrder input purchaseOrder: %s\\n\", err)\n\t\treturn shim.Error(err.Error())\n\t\t}\n\n\tfmt.Println(\"item object Item ID variable value is : %s\\n\", objitem.ItemID)\n\n\t// Data insertion for Couch DB starts here \n\ttransJSONasBytesLoan, err := json.Marshal(objitem)\n\terr = stub.PutState(objitem.ItemID, transJSONasBytesLoan)\n\t// Data insertion for Couch DB ends here\n\n\tfmt.Println(\"Create Purchase Order Successfully Done\")\n\n\tif err != nil {\n\t\tfmt.Printf(\"\\nUnable to make transevent inputs : %v \", err)\n\t\treturn shim.Error(err.Error())\n\t\t//return nil,nil\n\t}\n\treturn shim.Success(nil)\n}", "func (r RepresentativeRepo) CreateTx(tx *sql.Tx, repr d.Representative) error {\n\terr := r.UserRepo.CreateTx(tx, repr.User)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tconst insertQuery = `INSERT INTO \"Representative\"(representative_id, job_title, ref_user, ref_company)\n\tVALUES ($1, $2, $3, $4);`\n\t_, err = tx.Exec(insertQuery,\n\t\trepr.ID,\n\t\trepr.JobTitle,\n\t\trepr.User.ID,\n\t\trepr.CompanyID,\n\t)\n\n\tif err != nil {\n\t\t_ = tx.Rollback()\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (p *PgDb) CreateEmpTx(r *model.Emp, tx *sqlx.Tx, isValidate bool) (*model.Emp, error) {\n\tcnt := \"postgres (p *PgDb) CreateEmpTx\" // Имя текущего метода для логирования\n\t//mylog.PrintfDebug(\"[DEBUG] %v - START, param: '%+v'\", cnt, r)\n\n\t// Проверяем определен ли контекст транзакции\n\tif tx == nil {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - tx *sqlx.Tx is NULL\", cnt)\n\t\tlog.Printf(errM)\n\t\treturn nil, errors.New(errM)\n\t}\n\n\t//=====================================================================\n\t// Добавить валидацию входной структуры\n\t//=====================================================================\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\t// Если запускаем с проверками\n\tif isValidate {\n\t\t{ //Если Dept NULL или НЕ существует, то ошибка\n\t\t\tif !r.Deptno.Valid {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - r.Deptno is NULL\", cnt)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.New(errM)\n\t\t\t}\n\t\t\tdeptno := int(r.Deptno.Int64)\n\t\t\t// Запрос в транзакции\n\t\t\tdeptExists, err := p.DeptExistsTx(deptno, tx)\n\t\t\tif err != nil {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.DeptExistsTx(deptno, tx), args = '%v'\", cnt, deptno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t\t}\n\t\t\tif !deptExists {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - dept '%v' does not exist\", cnt, deptno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.New(errM)\n\t\t\t}\n\t\t\t//mylog.PrintfDebug(\"[DEBUG] %v - dept %v exists\", cnt, deptno)\n\t\t}\n\t\t{ //Если Emp существует, то игнорируем\n\t\t\texists, err := p.EmpExistsTx(r.Empno, tx)\n\t\t\tif err != nil {\n\t\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.EmpExistsTx(r.Empno, tx), args = '%v'\", cnt, r.Empno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t\t}\n\t\t\t// Если запись существует, то ни чего не делаем, возвращем, что пришло на вход\n\t\t\tif exists {\n\t\t\t\terrM := fmt.Sprintf(\"[WARN] %v - WARN - emp '%v' already exist - nothing to do\", cnt, r.Empno)\n\t\t\t\tlog.Printf(errM)\n\t\t\t\treturn nil, nil\n\t\t\t}\n\t\t\t//mylog.PrintfDebug(\"[DEBUG] %v - emp '%v' does not exist - can be created\", cnt, r.Empno)\n\t\t}\n\t}\n\t// =====================================================================\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\tstmText := sqlInsertEmpText\n\t// =====================================================================\n\n\t//Выполняем команду\n\tres, err := tx.NamedExec(stmText, r)\n\tif err != nil {\n\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - tx.NamedExec(stmText, r), args = '%+v'\", cnt, r)\n\t\tlog.Printf(errM)\n\t\treturn nil, errors.Wrap(err, errM)\n\t}\n\n\t{ // Необязательная часть - можно удалить в последствии\n\t\t// Проверим количество обработанных строк\n\t\trowCount, err := res.RowsAffected()\n\t\t_ = rowCount\n\t\t//mylog.PrintfDebug(\"[DEBUG] %v -- process %v rows\", cnt, rowCount)\n\t\tif err != nil {\n\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - res.RowsAffected()\", cnt)\n\t\t\tlog.Printf(errM)\n\t\t\treturn nil, errors.Wrap(err, errM)\n\t\t}\n\t}\n\n\t// =====================================================================\n\t// Переменная часть кода\n\t// =====================================================================\n\t// считаем данные обновленные данные - в БД могли быть тригера, которые поменяли данные\n\t// если запустили без проверок, то можно не возвращать результат - он будет запрошен уровнем выше\n\tif isValidate {\n\t\tv, err := p.GetEmpTx(r.Empno, tx)\n\t\tif err != nil {\n\t\t\terrM := fmt.Sprintf(\"[ERROR] %v - ERROR - p.GetEmpTx(r.Empno, tx), args = '%v'\", cnt, r.Empno)\n\t\t\tlog.Printf(errM)\n\t\t\treturn nil, errors.WithMessage(err, errM)\n\t\t}\n\t\tr = v\n\t}\n\t// =====================================================================\n\n\t//mylog.PrintfDebug(\"[DEBUG] %v - SUCCESS\", cnt)\n\n\treturn r, nil\n}", "func (gt GameType) CreateTx(action string, message json.RawMessage) (*types.Transaction, error) {\n\ttlog.Debug(\"Game.CreateTx\", \"action\", action)\n\tif action == ActionCreateGame {\n\t\tvar param GamePreCreateTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreCreateTx(&param)\n\t} else if action == ActionMatchGame {\n\t\tvar param GamePreMatchTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreMatchTx(&param)\n\t} else if action == ActionCancelGame {\n\t\tvar param GamePreCancelTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreCancelTx(&param)\n\t} else if action == ActionCloseGame {\n\t\tvar param GamePreCloseTx\n\t\terr := json.Unmarshal(message, &param)\n\t\tif err != nil {\n\t\t\ttlog.Error(\"CreateTx\", \"Error\", err)\n\t\t\treturn nil, types.ErrInvalidParam\n\t\t}\n\n\t\treturn CreateRawGamePreCloseTx(&param)\n\t}\n\treturn nil, types.ErrNotSupport\n}", "func Create(reqDto *ReqCreateDto, custDto *ReqCustomerDto) (statusCode int, code string, respDto *RespBase, err error) {\r\n\treqDto.ServiceType = \"TMS_CREATE_ORDER\"\r\n\tbizData, err := xml.Marshal(reqDto.BizData)\r\n\tif err != nil {\r\n\t\tcode = E02\r\n\t\treturn\r\n\t}\r\n\tsignParam := string(bizData) + reqDto.PartnerKey\r\n\treqDto.Sign, err = sign.GetMD5Hash(signParam, true)\r\n\tif err != nil {\r\n\t\tcode = E02\r\n\t\treturn\r\n\t}\r\n\treqMap := make(map[string]string, 0)\r\n\treqMap[\"serviceType\"] = reqDto.ServiceType\r\n\treqMap[\"partnerID\"] = reqDto.PartnerID\r\n\treqMap[\"bizData\"] = string(bizData)\r\n\treqMap[\"sign\"] = reqDto.Sign\r\n\r\n\tdata := base.JoinMapString(reqMap)\r\n\r\n\treq := httpreq.New(http.MethodPost, custDto.Url, data, func(httpReq *httpreq.HttpReq) error {\r\n\t\thttpReq.ReqDataType = httpreq.FormType\r\n\t\thttpReq.RespDataType = httpreq.XmlType\r\n\t\treturn nil\r\n\t})\r\n\tstatusCode, err = req.Call(&respDto)\r\n\tif err != nil {\r\n\t\tcode = E01\r\n\t\treturn\r\n\t}\r\n\tif statusCode != http.StatusOK {\r\n\t\tcode = E01\r\n\t\terr = fmt.Errorf(\"http status exp:200,act:%v\", statusCode)\r\n\t\treturn\r\n\t}\r\n\tif respDto.Result != true {\r\n\t\tcode = E03\r\n\t\terr = fmt.Errorf(\"%v-%v\", respDto.ErrorCode, respDto.ErrorDescription)\r\n\t\treturn\r\n\t}\r\n\tcode = SUC\r\n\treturn\r\n}", "func (t *Procure2Pay) CreateItems(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\n\tvar objitem item\t\n\tvar err error\n\n\tfmt.Println(\"Entering CreateItems\")\n\n\tif (len(args) < 1) {\n\t\tfmt.Println(\"Invalid number of args\")\n\t\treturn shim.Error(err.Error())\n\t\t//return nil, errors.New(\"Expected atleast one arguments for initiate Transaction\")\n\t}\n\n\tfmt.Println(\"Args [0] is : %v\\n\",args[0])\n\t\n\t//unmarshal item data from UI to \"item\" struct\n\terr = json.Unmarshal([]byte(args[0]), &objitem)\n\tif err != nil {\n\t\tfmt.Printf(\"Unable to unmarshal CreateItem input item: %s\\n\", err)\n\t\treturn shim.Error(err.Error())\n\t\t//return nil, nil\n\t}\n\n\tfmt.Println(\"item object ItemID variable value is : %s\\n\",objitem.ItemID);\n\t\n\t\t// Data insertion for Couch DB starts here \n\t\ttransJSONasBytes, err := json.Marshal(objitem)\n\t\terr = stub.PutState(objitem.ItemID, transJSONasBytes)\n\t\t// Data insertion for Couch DB ends here \n\n\t\tfmt.Println(\"Create items Successfully Done\")\t\n\t\n\t\tif err != nil {\n\t\t\t\tfmt.Printf(\"\\nUnable to make transevent inputs : %v \", err)\n\t\t\t\treturn shim.Error(err.Error())\n\t\t\t\t//return nil,nil\n\t\t\t}\n\treturn shim.Success(nil)\n\t//return nil, nil\n}", "func Transaction(next http.Handler) http.Handler {\n\tfn := func(w http.ResponseWriter, r *http.Request) {\n\t\tt, ctx := orm.NewTransaction(r.Context())\n\t\tdefer func() {\n\t\t\tif rec := recover(); rec != nil {\n\t\t\t\tt.Rollback()\n\t\t\t\t// Panic to let recoverer handle 500\n\t\t\t\tpanic(rec)\n\t\t\t} else {\n\t\t\t\terr := t.Commit()\n\t\t\t\tif err != nil {\n\t\t\t\t\tpanic(err)\n\t\t\t\t}\n\t\t\t}\n\t\t}()\n\t\tnext.ServeHTTP(w, r.WithContext(ctx))\n\t}\n\treturn http.HandlerFunc(fn)\n}", "func (ingest *Ingestion) Transaction(\n\tid int64,\n\ttx *core.Transaction,\n\tfee *core.TransactionFee,\n) error {\n\n\tsql := ingest.transactionInsertBuilder(id, tx, fee)\n\t_, err := ingest.DB.Exec(sql)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func MakeResTransaction(creator utils.Addr,key utils.Key,\n inps []TrInput,out TrOutput,\n jobblock,jobtrans,hashsol string,\n evaluation float64,isMin bool)*ResTransaction{\n tr := new(ResTransaction)\n tr.Timestamp = time.Now()\n tr.Output = out\n tr.Inputs = inps\n tr.Creator = creator\n tr.JobBlock = jobblock\n tr.JobTrans = jobtrans\n tr.HashSol = hashsol\n tr.Evaluation = evaluation\n tr.IsMin = isMin\n tr.Hash = tr.GetHash()\n tr.Signature = fmt.Sprintf(\"%x\",utils.GetSignatureFromHash(tr.Hash,key))\n return tr\n}", "func (c CreateTransactionService) Execute(newTransaction entities.Transaction) (*entities.Transaction, error) {\n\ttransactType := newTransaction.Type\n\tif transactType != \"income\" && transactType != \"outcome\" {\n\t\treturn nil, errors.New(\"Cannot create transaction type different fom income or outcome\")\n\t}\n\n\tif transactType == \"outcome\" && !c.checksValidBalance(newTransaction) {\n\t\treturn nil, errors.New(\"Cannot create transaction with invalid balance\")\n\t}\n\n\tcreatedTransaction, err := c.Repo.Create(&newTransaction)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\treturn createdTransaction, nil\n}", "func TransactionUpdate(c *gin.Context) {\n\tvar t models.Transaction\n\tvar newT models.Transaction\n\n\tif database.DBCon.First(&t, c.Param(\"id\")).RecordNotFound() {\n\t\tc.AbortWithError(http.StatusNotFound, appError.RecordNotFound).\n\t\t\tSetMeta(appError.RecordNotFound)\n\t\treturn\n\t}\n\n\t// Ensure current user is creator of transaction\n\tif t.CreatorID != c.Keys[\"CurrentUserID\"].(uint) {\n\t\tc.AbortWithError(appError.InsufficientPermission.Status, appError.InsufficientPermission).\n\t\t\tSetMeta(appError.InsufficientPermission)\n\t\treturn\n\t}\n\n\tbuffer, err := ioutil.ReadAll(c.Request.Body)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusNotAcceptable, err)\n\t}\n\n\terr2 := jsonapi.Unmarshal(buffer, &newT)\n\n\tif err2 != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tt.Type = newT.Type\n\tt.Amount = newT.Amount\n\tt.Memo = newT.Memo\n\tt.RecipientID = newT.RecipientID\n\tt.SenderID = newT.SenderID\n\n\t// Validate our new transaction\n\tisValid, errApp := t.Validate()\n\n\tif isValid == false {\n\t\tc.AbortWithError(errApp.Status, errApp).\n\t\t\tSetMeta(errApp)\n\t\treturn\n\t}\n\n\tdatabase.DBCon.Save(&t)\n\n\tdatabase.DBCon.First(&t.Recipient, t.RecipientID)\n\tdatabase.DBCon.First(&t.Sender, t.SenderID)\n\tdatabase.DBCon.First(&t.Creator, t.CreatorID)\n\n\tdata, err := jsonapi.Marshal(&t)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tc.Data(http.StatusOK, \"application/vnd.api+json\", data)\n}", "func (t *Transaction) createID() (string, error) {\n\n\t// Strip ID of txn\n\ttn := &Transaction{\n\t\tID: nil,\n\t\tVersion: t.Version,\n\t\tInputs: t.Inputs,\n\t\tOutputs: t.Outputs,\n\t\tOperation: t.Operation,\n\t\tAsset: t.Asset,\n\t\tMetadata: t.Metadata,\n\t}\n\t// Serialize transaction - encoding/json follows RFC7159 and BDB marshalling\n\tdbytes, err := tn.JSON()\n\tif err != nil {\n\t\treturn \"\", err\n\t}\n\n\t// Return hash of serialized txn object\n\th := sha3.Sum256(dbytes)\n\treturn hex.EncodeToString(h[:]), nil\n}", "func (_Content *ContentTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _Content.contract.Transact(opts, \"runCreate\")\n}", "func (tm *TransactionManager) CreateContract(sk string, data []byte, gasPrice uint64, nonce uint64, gasLimit uint64) (string, error) {\n\treturn tm.SendTx(sk, \"\", nil, data, gasPrice, nonce, gasLimit)\n}", "func (_LvRecording *LvRecordingTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _LvRecording.contract.Transact(opts, \"runCreate\")\n}", "func (bi *Blockchainidentifier) SaveTransaction(w http.ResponseWriter, r *http.Request) {\n\tlog.Println(r.RemoteAddr + \" POST /transactions/new\")\n\n\tvar t transaction\n\n\terr := json.NewDecoder(r.Body).Decode(&t)\n\tif err != nil {\n\t\thttp.Error(w, \"ERROR: \"+err.Error(), 500)\n\t\treturn\n\t}\n\n\tif bi.isValid(t) == false {\n\t\thttp.Error(w, \"ERROR: Missing values in transaction\", 400)\n\t\treturn\n\t}\n\n\tt.Timestamp = time.Now().UTC().Format(\"2006-01-02 15:04:05\")\n\n\tnewblockindex := bi.newTransaction(t)\n\n\tresponseMessage := map[string]string{\n\t\t\"message\": \"Transaction will be added in Block#\" + strconv.Itoa(newblockindex),\n\t}\n\n\tw.Header().Set(\"Content-Type\", \"application/json; charset=utf-8\")\n\tw.WriteHeader(http.StatusCreated)\n\tjson.NewEncoder(w).Encode(responseMessage)\n}", "func NewTransaction(p *requestParams) {\n\tw, r, c, u := p.w, p.r, p.c, p.u\n\n\td := json.NewDecoder(r.Body)\n\tvar request TransactionRequest\n\tif err := d.Decode(&request); err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tif len(request.Amounts) != len(request.Accounts) {\n\t\thttp.Error(w, \"Amounts and accounts of different lengths\", http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tdate, err := time.Parse(dateStringFormat, request.Date)\n\tif err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\tuserKey := userKey(c, u)\n\ttransactionId := uuid.NewRandom().String()\n\taccountKeys := make([]*datastore.Key, len(request.Accounts))\n\tsplitKeys := make([]*datastore.Key, len(request.Accounts))\n\tsplits := make([]*transaction.Split, len(request.Accounts))\n\n\tfor i := range request.Accounts {\n\t\taccountKeys[i] = datastore.NewKey(c, \"Account\", \"\", request.Accounts[i], userKey)\n\t\tsplitKeys[i] = datastore.NewKey(c, \"Split\", transactionId, 0, accountKeys[i])\n\t\tsplits[i] = &transaction.Split{\n\t\t\tAmount: request.Amounts[i],\n\t\t\tAccount: request.Accounts[i],\n\t\t\tMemo: request.Memo,\n\t\t\tDate: date,\n\t\t}\n\t}\n\n\tx := transaction.NewTransaction()\n\tx.AddSplits(splits)\n\n\tif err := x.ValidateAmount(); err != nil {\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\terr = datastore.RunInTransaction(c, func(c appengine.Context) error {\n\t\taccounts := make([]transaction.Account, len(accountKeys))\n\t\tif err := datastore.GetMulti(c, accountKeys, accounts); err != nil {\n\t\t\treturn err\n\t\t}\n\t\tfor i := range accounts {\n\t\t\tx.AddAccount(&accounts[i], accountKeys[i].IntID())\n\t\t}\n\n\t\tif err := x.Commit(); err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\tputStatus := make(chan error)\n\n\t\tgo func() {\n\t\t\t_, err := datastore.PutMulti(c, accountKeys, accounts)\n\t\t\tputStatus <- err\n\t\t}()\n\t\tgo func() {\n\t\t\t_, err := datastore.PutMulti(c, splitKeys, splits)\n\t\t\tputStatus <- err\n\t\t}()\n\n\t\terr := <-putStatus\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn <-putStatus\n\t}, nil)\n\tif err != nil {\n\t\t// TODO(cjc25): This might not be a 400: if e.g. datastore failed it should\n\t\t// be a 500. Interpret err and return the right thing.\n\t\thttp.Error(w, err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n}", "func (s *dbTestSuite) CreateTestTx() (*sql.Tx, func()) {\n\ttx, err := s.db.DB().Begin()\n\ts.Nil(err)\n\tcommit := func() {\n\t\ts.Nil(tx.Commit())\n\t}\n\n\treturn tx, commit\n}", "func (p *KVServer) NewTransaction(req bool, resp *NewTransactionResp) error {\n\ttxID := nextTransactionID\n\tnextTransactionID = nextTransactionID + 10\n\tmutex.Lock()\n\tkvStore := keyValueStore\n\tmutex.Unlock()\n\t*resp = NewTransactionResp{txID, kvStore}\n\treturn nil\n}", "func CreateServiceSubscriptionTransaction(db *sql.DB, params url.Values) (data []byte, err error) {\n\tvar (\n\t\tresult sql.Result\n\t\tresponse = SQLResponse{}\n\t\ttx *sql.Tx\n\t\tserviceTypeStr string\n\t\tserviceType int\n\t\tserviceName string\n\t\tdescription string\n\t\torganizationName string\n\t\tactiveUntil string\n\t\tamountPaidStr string\n\t\tamountPaid int\n\t\tprocessedTimeStamp string\n\t)\n\n\tif tx, err = db.Begin(); err != nil {\n\t\treturn\n\t}\n\tif amountPaidStr, err = common.GetRequiredParam(params, \"amountPaid\"); err != nil {\n\t\treturn\n\t}\n\tif amountPaid, err = strconv.Atoi(amountPaidStr); err != nil {\n\t\treturn\n\t}\n\tif processedTimeStamp, err = common.GetRequiredParam(params, \"processedTimestamp\"); err != nil {\n\t\treturn\n\t}\n\tif serviceTypeStr, err = common.GetRequiredParam(params, \"serviceType\"); err != nil {\n\t\treturn\n\t}\n\tif serviceType, err = strconv.Atoi(serviceTypeStr); err != nil {\n\t\treturn\n\t}\n\tif serviceName, err = common.GetRequiredParam(params, \"serviceName\"); err != nil {\n\t\treturn\n\t}\n\tif description, err = common.GetRequiredParam(params, \"description\"); err != nil {\n\t\treturn\n\t}\n\tif organizationName, err = common.GetRequiredParam(params, \"organizationName\"); err != nil {\n\t\treturn\n\t}\n\tif activeUntil, err = common.GetRequiredParam(params, \"activeUntil\"); err != nil {\n\t\treturn\n\t}\n\tif result, err = tx.Exec(\"INSERT INTO ServiceSubscriptionTransaction (amountPaid, processedTimestamp,type,serviceName,description, organizationName, activeUntil) VALUES(?,?,?,?,?,?,?);\",\n\t\tamountPaid, processedTimeStamp, serviceType, serviceName, description, organizationName, activeUntil); err != nil {\n\t\ttx.Rollback()\n\t\treturn\n\t}\n\tif err = tx.Commit(); err != nil {\n\t\treturn\n\t}\n\tif response.AffectedRows, err = result.RowsAffected(); err != nil {\n\t\treturn\n\t}\n\tdata, err = json.Marshal(response)\n\treturn\n}", "func (_LvRecordableStream *LvRecordableStreamTransactor) RunCreate(opts *bind.TransactOpts) (*types.Transaction, error) {\n\treturn _LvRecordableStream.contract.Transact(opts, \"runCreate\")\n}", "func (m *MySQLModelAudit) CreateTx(tx *sql.Tx, modelAudit *modelaudit.Model) error {\n\tstmt, err := tx.Prepare(mysqlCreateModelAudit)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\tnow := utils.Now()\n\treturn mysql.ExecAffectingOneRow(\n\t\tstmt,\n\t\tmodelAudit.PrevModel.ModelID,\n\t\tmodelAudit.PrevModel.Mjml,\n\t\tmodelAudit.PrevModel.Html,\n\t\tmodelAudit.PrevModel.Variables,\n\t\tmodelAudit.Model.Mjml,\n\t\tmodelAudit.Model.Html,\n\t\tmodelAudit.Model.Variables,\n\t\tmodelAudit.SetUserID,\n\t\tnow[\"date\"], now[\"time\"], now[\"unix\"],\n\t)\n}", "func (w *WalletChaincode) create(stub shim.ChaincodeStubInterface, args []string) pb.Response {\n\tif len(args) < 6 {\n\t\treturn shim.Error(\"Incorrect number of arguments. Expecting at least 2\")\n\t}\n\taddress := args[2]\n\taccountKey := buildAccountKey(address)\n\tif err := stub.PutState(accountKey, []byte(address)); err != nil {\n\t\treturn shim.Error(fmt.Sprintf(\"Error putting data for key [%s]: %s\", accountKey, err))\n\t}\n\tfmt.Println(\"create an account: \", accountKey)\n\n\twalletKey := buildWalletKey(args[0], args[1], address)\n\tif err := stub.PutState(walletKey, []byte(args[5])); err != nil {\n\t\treturn shim.Error(fmt.Sprintf(\"Error putting data for key [%s]: %s\", walletKey, err))\n\t}\n\tfmt.Println(\"create a wallet: \", walletKey)\n\n\t// seqBytes, err := stub.GetState(SEQUENCE)\n\t// if err != nil {\n\t// \treturn shim.Error(\"Failed to get state\")\n\t// }\n\t// if seqBytes == nil {\n\t// \treturn shim.Error(\"Entity not found\")\n\t// }\n\t// seq, _ := strconv.ParseInt(string(seqBytes), 10, 64)\n\tseq := atomic.AddUint64(&w.Sequence, 1)\n\tsequenceKey := buildSequenceKey(seq)\n\tjsonTx := \"{\\\"sequence\\\":\\\"\" + strconv.FormatUint(seq, 10) + \"\\\",\\\"txid\\\":\\\"\" + string(stub.GetTxID()) + \"\\\"}\"\n\tif err := stub.PutState(sequenceKey, []byte(jsonTx)); err != nil {\n\t\treturn shim.Error(fmt.Sprintf(\"Error putting data for key [%s]: %s\", walletKey, err))\n\t}\n\n\tfmt.Println(\"create success: \", stub.GetTxID())\n\treturn shim.Success([]byte(fmt.Sprintf(\"{\\\"wallet\\\":\\\"%s\\\", \\\"txid\\\":\\\"%s\\\"}\", walletKey, stub.GetTxID())))\n}", "func (tx *Transaction) Create(entity interface{}) error {\n\treturn createFunc(tx.Query, entity)\n}", "func Create(t *contract.SQLTable) error {\n\tsqlstr := t.GetCreateSQL()\n\tif t.Database == \"\" {\n\t\tt.Database = \"default\"\n\t}\n\tdb, err := dblist.Get(t.Database)\n\tif err != nil {\n\t\treturn err\n\t}\n\t_, err = db.Exec(sqlstr)\n\treturn err\n}", "func (s *PaymentStorage) Create(\n\tctx context.Context,\n\tid aggregate.ID,\n\tversion transaction.Version,\n\torganisationID transaction.OrganisationID,\n\tattributes interface{},\n) error {\n\tlogger := log.FromContext(ctx)\n\n\tquery := `INSERT INTO %[1]s (id, version, organisation_id, attributes) VALUES ($1, $2, $3, $4)`\n\tquery = fmt.Sprintf(query, s.table)\n\n\tif logger != nil {\n\t\tlogger.Debugf(\"exec in transaction sql %s, values %+v\", query, []interface{}{\n\t\t\tid,\n\t\t\tversion,\n\t\t\torganisationID,\n\t\t\tattributes,\n\t\t})\n\t}\n\n\treturn execInTransaction(s.db, func(tx *sqlx.Tx) error {\n\t\t_, err := tx.ExecContext(ctx, query, id, version, organisationID, attributes)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\n\t\treturn nil\n\t})\n}", "func (_TableManager *TableManagerTransactor) CreateTable(opts *bind.TransactOpts, path string, tableInfo TableInfo) (*types.Transaction, *types.Receipt, error) {\n\treturn _TableManager.contract.Transact(opts, \"createTable\", path, tableInfo)\n}", "func (s *SmartContract) CreateContract(ctx contractapi.TransactionContextInterface, id string, sellerID string, consumerID string, created string, contract string, signature string) error {\n\tctc := Contract{\n\t\tContext: \"http://wuldid.ddns.net\",\n\t\tSellerID: sellerID,\n\t\tConsumerID: consumerID,\n\t\tCreated: created,\n\t\tContract: contract,\n\t\tSignature: signature,\n\t}\n\n\texists, err := s.DidExists(ctx, id)\n\tif err != nil {\n\t\tfmt.Errorf(\"Unexpected error!! : %q\", err)\n\t}\n\tif !exists {\n\t\tctcJSON, _ := json.Marshal(ctc)\n\t\treturn ctx.GetStub().PutState(id, ctcJSON)\n\t} else {\n\t\treturn fmt.Errorf(\"Don't exsit did!\")\n\t}\n\n}", "func (client *Client) CreateTransactionType(transactionType string, customIndexes []CustomIndexStructure) (_ *Response, err error) {\n\tpath := \"/transaction-type\"\n\turi := fmt.Sprintf(\"%s%s\", client.apiBaseURL, path)\n\tvar params TransactionType\n\tparams.Version = \"1\"\n\tparams.Type = transactionType\n\tparams.CustomIndexes = customIndexes\n\n\tb, err := json.Marshal(params)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tresp, err := client.httpClient.Post(uri, \"content/json\", bytes.NewBuffer(b))\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tdefer func() {\n\t\t_ = resp.Body.Close()\n\t}()\n\tvar chainResp Response\n\trespBytes, err := ioutil.ReadAll(resp.Body)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\tif err := json.Unmarshal(respBytes, &chainResp.Response); err != nil {\n\t\treturn nil, err\n\t}\n\tchainResp.Status = resp.StatusCode\n\tif 200 <= resp.StatusCode && resp.StatusCode < 300 {\n\t\tchainResp.OK = true\n\t}\n\treturn &chainResp, err\n}", "func (r Virtual_Guest) CreateArchiveTransaction(groupName *string, blockDevices []datatypes.Virtual_Guest_Block_Device, note *string) (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\tparams := []interface{}{\n\t\tgroupName,\n\t\tblockDevices,\n\t\tnote,\n\t}\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"createArchiveTransaction\", params, &r.Options, &resp)\n\treturn\n}", "func (t *TxCreate) CreatePaymentTx() (string, string, error) {\n\tsender := t.paymentSender\n\treceiver := account.AccountTypeAnonymous\n\ttargetAction := action.ActionTypePayment\n\tt.logger.Debug(\"account\",\n\t\tzap.String(\"sender\", sender.String()),\n\t\tzap.String(\"receiver\", receiver.String()),\n\t)\n\n\t// get payment data from payment_request\n\tuserPayments, totalAmount, paymentRequestIds, err := t.createUserPayment()\n\tif err != nil {\n\t\treturn \"\", \"\", err\n\t}\n\tif len(userPayments) == 0 {\n\t\tt.logger.Debug(\"no data in userPayments\")\n\t\t// no data\n\t\treturn \"\", \"\", nil\n\t}\n\n\t// check sender's total balance\n\t// GetOneUnAllocated\n\taddrItem, err := t.addrRepo.GetOneUnAllocated(sender)\n\tif err != nil {\n\t\treturn \"\", \"\", errors.Wrap(err, \"fail to call addrRepo.GetAll(account.AccountTypeClient)\")\n\t}\n\tsenderBalance, err := t.xrp.GetBalance(addrItem.WalletAddress)\n\tif err != nil {\n\t\treturn \"\", \"\", errors.Wrap(err, \"fail to call xrp.GetAccountInfo()\")\n\t}\n\n\tif senderBalance <= totalAmount {\n\t\treturn \"\", \"\", errors.New(\"sender balance is insufficient to send\")\n\t}\n\n\t// create raw transaction each address\n\tserializedTxs := make([]string, 0, len(userPayments))\n\ttxDetailItems := make([]*models.XRPDetailTX, 0, len(userPayments))\n\tvar sequence uint64\n\tfor _, userPayment := range userPayments {\n\t\t// call CreateRawTransaction\n\t\tinstructions := &pb.Instructions{\n\t\t\tMaxLedgerVersionOffset: xrp.MaxLedgerVersionOffset,\n\t\t}\n\t\tif sequence != 0 {\n\t\t\tinstructions.Sequence = sequence\n\t\t}\n\t\ttxJSON, rawTxString, err := t.xrp.CreateRawTransaction(addrItem.WalletAddress, userPayment.receiverAddr, userPayment.floatAmount, instructions)\n\t\tif err != nil {\n\t\t\tt.logger.Warn(\"fail to call xrp.CreateRawTransaction()\", zap.Error(err))\n\t\t\tcontinue\n\t\t}\n\t\tt.logger.Debug(\"txJSON\", zap.Any(\"txJSON\", txJSON))\n\t\tgrok.Value(txJSON)\n\n\t\t// sequence for next rawTransaction\n\t\tsequence = txJSON.Sequence + 1\n\n\t\t// generate UUID to trace transaction because unsignedTx is not unique\n\t\tuid := uuid.NewV4().String()\n\n\t\tserializedTxs = append(serializedTxs, fmt.Sprintf(\"%s,%s\", uid, rawTxString))\n\n\t\t// create insert data for eth_detail_tx\n\t\ttxDetailItem := &models.XRPDetailTX{\n\t\t\tUUID: uid,\n\t\t\tCurrentTXType: tx.TxTypeUnsigned.Int8(),\n\t\t\tSenderAccount: sender.String(),\n\t\t\tSenderAddress: addrItem.WalletAddress,\n\t\t\tReceiverAccount: receiver.String(),\n\t\t\tReceiverAddress: userPayment.receiverAddr,\n\t\t\tAmount: txJSON.Amount,\n\t\t\tXRPTXType: txJSON.TransactionType,\n\t\t\tFee: txJSON.Fee,\n\t\t\tFlags: txJSON.Flags,\n\t\t\tLastLedgerSequence: txJSON.LastLedgerSequence,\n\t\t\tSequence: txJSON.Sequence,\n\t\t\t// SigningPubkey: txJSON.SigningPubKey,\n\t\t\t// TXNSignature: txJSON.TxnSignature,\n\t\t\t// Hash: txJSON.Hash,\n\t\t}\n\t\ttxDetailItems = append(txDetailItems, txDetailItem)\n\t}\n\n\treturn t.afterTxCreation(targetAction, sender, serializedTxs, txDetailItems, paymentRequestIds)\n}", "func (allocateIncome *AllocateIncome) Create() (err error) {\n\t//插入数据\n\tstmt, err := Db.Prepare(\"INSERT INTO allocateIncome(UDID, pool_id,mining_id,income,fee,date,status) values(?,?,?,?,?,?,?)\")\n\tif err != nil {\n\t\treturn\n\t}\n\tres, err := stmt.Exec(allocateIncome.UDID, allocateIncome.PoolID, allocateIncome.MiningId, allocateIncome.Income, allocateIncome.Fee, time.Now().Unix(), allocateIncome.Status)\n\tif err != nil {\n\t\treturn\n\t}\n\n\tid, err := res.LastInsertId()\n\tlog.Info(id)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func (w *Wallet) txCreator() {\n\tquit := w.quitChan()\nout:\n\tfor {\n\t\tselect {\n\t\tcase txr := <-w.createTxRequests:\n\t\t\theldUnlock, err := w.holdUnlock()\n\t\t\tif err != nil {\n\t\t\t\ttxr.resp <- createTxResponse{nil, err}\n\t\t\t\tcontinue\n\t\t\t}\n\t\t\ttx, err := w.txToOutputs(txr.req)\n\t\t\theldUnlock.release()\n\t\t\ttxr.resp <- createTxResponse{tx, err}\n\t\tcase <-quit:\n\t\t\tbreak out\n\t\t}\n\t}\n\tw.wg.Done()\n}", "func (w *Wallet) CreateSimpleTx(r CreateTxReq) (*txauthor.AuthoredTx, er.R) {\n\treq := createTxRequest{\n\t\treq: r,\n\t\tresp: make(chan createTxResponse),\n\t}\n\tw.createTxRequests <- req\n\tresp := <-req.resp\n\treturn resp.tx, resp.err\n}", "func (db *Database) NewTransaction() *Transaction{\r\n\treturn &Transaction{db,make(map[string]*record)}\r\n}", "func (p *psqlInvoiceItem) CreateTx(tx *sql.Tx, headerID uint, ms invoiceItem.Models) error {\n\tstmt, err := tx.Prepare(psqlCreateInvoiceItem)\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\tfor _, item := range ms {\n\t\terr = stmt.QueryRow(headerID, item.ProductID).Scan(&item.ID, &item.CreatedAt)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\treturn nil\n\t// return stmt.QueryRow(m.).Scan(&m.ID, &m.CreatedAt)\n}", "func (tr *Transactor) CreateTransactionArweave(ctx context.Context, tags []Tag, w WalletSigner, amount string, data []byte, target string) (*Transaction, error) {\n\tlastTx, err := tr.Client.TxAnchor(ctx)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\tprice, err := tr.Client.GetReward(ctx, data)\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\n\t// Non encoded transaction fields\n\ttx := NewTransactionArweave(\n\t\ttags,\n\t\tlastTx,\n\t\tw.PubKeyModulus(),\n\t\tamount,\n\t\ttarget,\n\t\tdata,\n\t\tprice,\n\t)\n\n\treturn tx, nil\n}", "func create(\n\ttrx storage.Transaction,\n\ttxId merkle.Digest,\n\townerData OwnerData,\n\towner *account.Account,\n) {\n\t// increment the count for owner\n\tnKey := owner.Bytes()\n\tcount := trx.Get(storage.Pool.OwnerNextCount, nKey)\n\tif nil == count {\n\t\tcount = []byte{0, 0, 0, 0, 0, 0, 0, 0}\n\t} else if uint64ByteSize != len(count) {\n\t\tlogger.Panic(\"OwnerNextCount database corrupt\")\n\t}\n\tnewCount := make([]byte, uint64ByteSize)\n\tbinary.BigEndian.PutUint64(newCount, binary.BigEndian.Uint64(count)+1)\n\ttrx.Put(storage.Pool.OwnerNextCount, nKey, newCount, []byte{})\n\n\t// write to the owner list\n\toKey := append(owner.Bytes(), count...)\n\ttrx.Put(storage.Pool.OwnerList, oKey, txId[:], []byte{})\n\n\t// write new index record\n\tdKey := append(owner.Bytes(), txId[:]...)\n\ttrx.Put(storage.Pool.OwnerTxIndex, dKey, count, []byte{})\n\n\t// save owner data record\n\ttrx.Put(storage.Pool.OwnerData, txId[:], ownerData.Pack(), []byte{})\n}", "func CreateTestContractCreationTransaction(\n\tsigner hmytypes.Signer, shard uint32, nonce, gasLimit uint64, gasPrice, amount *big.Int, data []byte,\n) (*hmytypes.Transaction, error) {\n\tfromKey, err := crypto.GenerateKey()\n\tif err != nil {\n\t\treturn nil, err\n\t}\n\ttx := hmytypes.NewContractCreation(nonce, shard, amount, gasLimit, gasPrice, data)\n\treturn hmytypes.SignTx(tx, signer, fromKey)\n}", "func Create(responseWriter http.ResponseWriter, request *http.Request) {\n\tfmt.Println(\"[ CreateOrder ]\")\n\tbody, _ := json.Marshal(request.Body)\n\tfmt.Println(\"[ CreateOrder ] Body=\" + string(body))\n\t//params := mux.Vars(request)\n\tvar orderEntity OrderEntity\n\t_ = json.NewDecoder(request.Body).Decode(&orderEntity)\n\n\tvar result OrderEntity = Insert(orderEntity)\n\n\tWriteMessages(result, Topic.TOPIC_SUCCESS)\n\n\tjson.NewEncoder(responseWriter).Encode(result)\n}", "func (s *Service) CreateClientTx(tx *gorm.DB, clientID, secret, redirectURI string) (*models.OauthClient, error) {\n\treturn s.createClientCommon(tx, clientID, secret, redirectURI)\n}", "func (dt DefaultTransactor) Create(ctx context.Context, m *movie.Movie) error {\n\ttx, err := dt.datastorer.BeginTx(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Prepare the sql statement using bind variables\n\tstmt, err := tx.PrepareContext(ctx, `\n\tselect o_create_timestamp,\n\t\t o_update_timestamp\n\t from demo.create_movie (\n\t\tp_id => $1,\n\t\tp_extl_id => $2,\n\t\tp_title => $3,\n\t\tp_rated => $4,\n\t\tp_released => $5,\n\t\tp_run_time => $6,\n\t\tp_director => $7,\n\t\tp_writer => $8,\n\t\tp_create_client_id => $9,\n\t\tp_create_username => $10)`)\n\n\tif err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\tdefer stmt.Close()\n\n\t// At some point, I will add a whole client flow, but for now\n\t// faking a client uuid....\n\tfakeClientID := uuid.New()\n\n\t// Execute stored function that returns the create_date timestamp,\n\t// hence the use of QueryContext instead of Exec\n\trows, err := stmt.QueryContext(ctx,\n\t\tm.ID, //$1\n\t\tm.ExternalID, //$2\n\t\tm.Title, //$3\n\t\tm.Rated, //$4\n\t\tm.Released, //$5\n\t\tm.RunTime, //$6\n\t\tm.Director, //$7\n\t\tm.Writer, //$8\n\t\tfakeClientID, //$9\n\t\tm.CreateUser.Email) //$10\n\n\tif err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\tdefer rows.Close()\n\n\t// Iterate through the returned record(s)\n\tfor rows.Next() {\n\t\tif err := rows.Scan(&m.CreateTime, &m.UpdateTime); err != nil {\n\t\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t\t}\n\t}\n\n\t// If any error was encountered while iterating through rows.Next above\n\t// it will be returned here\n\tif err := rows.Err(); err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\n\t// Commit the Transaction\n\tif err := dt.datastorer.CommitTx(tx); err != nil {\n\t\treturn errs.E(errs.Database, dt.datastorer.RollbackTx(tx, err))\n\t}\n\n\treturn nil\n}", "func createOrder(writer http.ResponseWriter, request *http.Request) {\n\ttransactionId := request.Header.Get(\"transactionId\")\n\torder := <-orderService.GetOrder(transactionId)\n\tif order.Id.Value == \"\" {\n\t\tlog.Printf(\"Creating order for transactionId :%s......\", transactionId)\n\t\torderHandler.CreateOrder(transactionId, command.CreateOrderCommand{Id: transactionId})\n\t}\n\trenderResponse(writer, []byte(transactionId))\n}", "func createSignedTransaction(fromAddress common.Address, toAddress *common.Address, amount *big.Int, gasLimit uint64, data []byte) (signedTx *types.Transaction, err error) {\n\t// Create the transaction\n\ttx, err := createTransaction(fromAddress, toAddress, amount, gasLimit, data)\n\tif err != nil {\n\t\treturn\n\t}\n\n\t// Sign the transaction\n\tsignedTx, err = signTransaction(fromAddress, tx)\n\tif err != nil {\n\t\terr = fmt.Errorf(\"Failed to sign transaction: %v\", err)\n\t\treturn\n\t}\n\n\t// Increment the nonce for the next transaction\n\tnextNonce(fromAddress)\n\n\treturn\n}", "func (r *Renter) createContractTransaction(terms modules.ContractTerms, merkleRoot crypto.Hash) (txn consensus.Transaction, id string, err error) {\n\t// Get the payout as set by the missed proofs, and the client fund as determined by the terms.\n\tvar payout consensus.Currency\n\tfor _, output := range terms.MissedProofOutputs {\n\t\tpayout = payout.Add(output.Value)\n\t}\n\n\t// Get the cost to the client as per the terms in the contract.\n\tsizeCurrency := consensus.NewCurrency64(terms.FileSize)\n\tdurationCurrency := consensus.NewCurrency64(uint64(terms.Duration))\n\tclientCost := terms.Price.Mul(sizeCurrency).Mul(durationCurrency)\n\n\t// Fill out the contract.\n\tcontract := consensus.FileContract{\n\t\tFileMerkleRoot: merkleRoot,\n\t\tFileSize: terms.FileSize,\n\t\tStart: terms.DurationStart + terms.Duration,\n\t\tExpiration: terms.DurationStart + terms.Duration + terms.WindowSize,\n\t\tPayout: payout,\n\t\tValidProofOutputs: terms.ValidProofOutputs,\n\t\tMissedProofOutputs: terms.MissedProofOutputs,\n\t}\n\n\t// Create the transaction.\n\tid, err = r.wallet.RegisterTransaction(txn)\n\tif err != nil {\n\t\treturn\n\t}\n\t_, err = r.wallet.FundTransaction(id, clientCost)\n\tif err != nil {\n\t\treturn\n\t}\n\ttxn, _, err = r.wallet.AddFileContract(id, contract)\n\tif err != nil {\n\t\treturn\n\t}\n\n\treturn\n}", "func TransactionHandler(w http.ResponseWriter, r *http.Request) {\n\taction := r.URL.Path[len(\"/api/transactions\"):]\n\n\tlog.Println(\"Handling method\", r.Method, \"with action\", action)\n\n\tswitch r.Method {\n\tcase \"POST\":\n\t\tswitch action {\n\t\tcase \"\": // Create new transaction\n\t\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorForm, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar t TransactionResponse\n\n\t\t\terr = json.Unmarshal(body, &t)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpass := []byte(t.Password)\n\n\t\t\ttx, e := createAndBroadcastTx(t.Recipient, *big.NewInt(t.Amount), pass)\n\n\t\t\tvar jsonResponse TransactionResponse\n\t\t\tif e == nil {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseOk, Hash: hex.EncodeToString(tx.Hash())}\n\t\t\t} else {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseFailed, ErrorText: e.Error()}\n\t\t\t}\n\n\t\t\tres, err := json.Marshal(jsonResponse)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfmt.Fprintf(w, string(res))\n\t\tdefault:\n\t\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\t\t}\n\tcase \"GET\":\n\t\tswitch action {\n\t\tcase \"\":\n\t\t\tvar txs []TransactionJson\n\t\t\tfor _, tx := range Config.DeserializedTxs {\n\t\t\t\ttxs = append(txs, EncodeToFriendlyStruct(tx))\n\n\t\t\t}\n\n\t\t\tif len(txs) == 0 {\n\t\t\t\tfmt.Fprintf(w, string(\"[]\"))\n\t\t\t} else {\n\n\t\t\t\tres, err := json.Marshal(txs)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Println(\"Nope\", err.Error())\n\t\t\t\t} else {\n\t\t\t\t\tfmt.Fprintf(w, string(res))\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\tdefault:\n\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\n\t}\n}", "func (p *PsqlInvoiceItem) CreateTx(tx *sql.Tx, headerID uint, ms invoiceitem.Models) error {\n\tstmt, err := tx.Prepare(psqlCreateInvoiceItem)\n\n\tif err != nil {\n\t\treturn err\n\t}\n\tdefer stmt.Close()\n\n\tfor _, item := range ms {\n\t\terr = stmt.QueryRow(headerID, item.ProductID).Scan(\n\t\t\t&item.ID,\n\t\t\t&item.CreatedAt,\n\t\t)\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t}\n\n\treturn err\n\n}", "func (a *ChainAdaptor) CreateUtxoSignedTransaction(req *proto.CreateUtxoSignedTransactionRequest) (*proto.CreateSignedTransactionReply, error) {\n\tr := bytes.NewReader(req.TxData)\n\tvar msgTx wire.MsgTx\n\terr := msgTx.Deserialize(r)\n\tif err != nil {\n\t\tlog.Error(\"CreateSignedTransaction msgTx.Deserialize\", \"err\", err)\n\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\tif len(req.Signatures) != len(msgTx.TxIn) {\n\t\tlog.Error(\"CreateSignedTransaction invalid params\", \"err\", \"Signature number mismatch Txin number\")\n\t\terr = errors.New(\"Signature number != Txin number\")\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\tif len(req.PublicKeys) != len(msgTx.TxIn) {\n\t\tlog.Error(\"CreateSignedTransaction invalid params\", \"err\", \"Pubkey number mismatch Txin number\")\n\t\terr = errors.New(\"Pubkey number != Txin number\")\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\t// assemble signatures\n\tfor i, in := range msgTx.TxIn {\n\t\tbtcecPub, err2 := btcec.ParsePubKey(req.PublicKeys[i], btcec.S256())\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction ParsePubKey\", \"err\", err2)\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tvar pkData []byte\n\t\tif btcec.IsCompressedPubKey(req.PublicKeys[i]) {\n\t\t\tpkData = btcecPub.SerializeCompressed()\n\t\t} else {\n\t\t\tpkData = btcecPub.SerializeUncompressed()\n\t\t}\n\n\t\t// verify transaction\n\t\tpreTx, err2 := a.getClient().GetRawTransactionVerbose(&in.PreviousOutPoint.Hash)\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction GetRawTransactionVerbose\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tlog.Info(\"CreateSignedTransaction \", \"from address\", preTx.Vout[in.PreviousOutPoint.Index].ScriptPubKey.Addresses[0])\n\n\t\tfromAddress, err2 := btcutil.DecodeAddress(preTx.Vout[in.PreviousOutPoint.Index].ScriptPubKey.Addresses[0], a.getClient().GetNetwork())\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction DecodeAddress\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tfromPkScript, err2 := txscript.PayToAddrScript(fromAddress)\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction PayToAddrScript\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\t// creat sigscript and verify\n\t\tif len(req.Signatures[i]) < 64 {\n\t\t\terr2 = errors.New(\"Invalid signature length\")\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\t\tr := new(big.Int).SetBytes(req.Signatures[i][0:32])\n\t\ts := new(big.Int).SetBytes(req.Signatures[i][32:64])\n\n\t\tbtcecSig := &btcec.Signature{\n\t\t\tR: r,\n\t\t\tS: s,\n\t\t}\n\t\tsig := append(btcecSig.Serialize(), byte(txscript.SigHashAll))\n\t\tsigScript, err2 := txscript.NewScriptBuilder().AddData(sig).AddData(pkData).Script()\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction NewScriptBuilder\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\n\t\tmsgTx.TxIn[i].SignatureScript = sigScript\n\t\tamount := btcToSatoshi(preTx.Vout[in.PreviousOutPoint.Index].Value).Int64()\n\t\tlog.Info(\"CreateSignedTransaction \", \"amount\", preTx.Vout[in.PreviousOutPoint.Index].Value, \"int amount\", amount)\n\n\t\tvm, err2 := txscript.NewEngine(fromPkScript, &msgTx, i, txscript.StandardVerifyFlags, nil, nil, amount)\n\t\tif err2 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction NewEngine\", \"err\", err2)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err2.Error(),\n\t\t\t}, err2\n\t\t}\n\t\tif err3 := vm.Execute(); err3 != nil {\n\t\t\tlog.Error(\"CreateSignedTransaction NewEngine Execute\", \"err\", err3)\n\n\t\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\t\tMsg: err3.Error(),\n\t\t\t}, err3\n\t\t}\n\n\t}\n\n\t// serialize tx\n\tbuf := bytes.NewBuffer(make([]byte, 0, msgTx.SerializeSize()))\n\n\terr = msgTx.Serialize(buf)\n\tif err != nil {\n\t\tlog.Error(\"CreateSignedTransaction tx Serialize\", \"err\", err)\n\n\t\treturn &proto.CreateSignedTransactionReply{\n\t\t\tCode: proto.ReturnCode_ERROR,\n\t\t\tMsg: err.Error(),\n\t\t}, err\n\t}\n\n\thash := msgTx.TxHash()\n\treturn &proto.CreateSignedTransactionReply{\n\t\tCode: proto.ReturnCode_SUCCESS,\n\t\tSignedTxData: buf.Bytes(),\n\t\tHash: (&hash).CloneBytes(),\n\t}, nil\n}", "func (cs *CustStoreSqlite) AddTransaction(\n id string,\n customerId string,\n loadAmountCents int64,\n time time.Time,\n accepted bool,\n) error {\n // Check transaction id\n isDuplicate, err := cs.isDuplicate(id, customerId)\n if err != nil {\n return err\n }\n if isDuplicate {\n return custstore.DuplicateError\n }\n // Add to transactions\n err = cs.addTransaction(id, customerId, loadAmountCents, time, accepted)\n return err\n}", "func CreateContractObject(args []string) (SalesContractObject, error) {\n\t// S001 LHTMO bosch\n\tvar err error\n\tvar myContract SalesContractObject\n\n\t// Check there are 3 Arguments provided as per the the struct\n\tif len(args) != 8 {\n\t\tfmt.Println(\"CreateContractObject(): Incorrect number of arguments. Expecting 8 \")\n\t\treturn myContract, errors.New(\"CreateContractObject(): Incorrect number of arguments. Expecting 8 \")\n\t}\n\n\t// Validate Serialno is an integer\n\n\tstage, err := strconv.Atoi(args[1])\n\tif err != nil {\n\t\tfmt.Println(\"CreateAssetObject(): Stage should be an integer create failed! \")\n\t\treturn myContract, errors.New(\"CreateAssetbject(): Stage should be an integer create failed. \")\n\t}\n\tif stage != 0 {\n\t\tfmt.Println(\"CreateAssetObject(): Stage should be set as open \")\n\t\treturn myContract, errors.New(\"CreateAssetbject(): Stage should be set as open\")\n\t}\n\n\tmyContract = SalesContractObject{args[0], STATE_OPEN, args[2], args[3], args[4], args[5], args[6], time.Now().Format(\"20060102150405\")}\n\n\tfmt.Println(\"CreateContractObject(): Contract Object created: \", myContract.Contractid, myContract.Stage, myContract.Buyer, myContract.Transporter, myContract.Seller, myContract.AssetID, myContract.DocumentID, time.Now().Format(\"20060102150405\"))\n\treturn myContract, nil\n}", "func (a *ApiDB) CreateContract(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Add(\"Content-Type\", \"application/json\")\n\tp := MODELS.CREATE_UPDATE_CONTRACT_REQUEST{}\n\terr := json.NewDecoder(r.Body).Decode(&p)\n\tif err != nil {\n\t\tio.WriteString(w, `{\"message\": \"wrong format!\"}`)\n\t\treturn\n\t}\n\n\tresult := BUSINESS.CreateContract(a.Db, p)\n\tif result {\n\t\tio.WriteString(w, ` { \"status\": 200,\n \"message\": \"Create contract success\",\n \"data\": {\n \"status\": 1\n }\n}\n`)\n\t} else {\n\t\tio.WriteString(w, `{ \"message\": \"Can’t create contract\"}`)\n\t}\n}", "func (d Database) CreateRoot() error {\n\tdb, connectErr := d.Connect()\n\tdefer db.Close()\n\tutil.Handle(\"Error creating a DB connection: \", connectErr)\n\tvar count int\n\terr := db.QueryRow(\"SELECT COUNT(*) FROM \"+d.Cf.GetTableName()+\" ORDER BY $1 DESC\", \"tx_time\").Scan(&count)\n\tswitch {\n\tcase err != nil:\n\t\tutil.Handle(\"There was a problem counting database transactions: \", err)\n\t\treturn err\n\tdefault:\n\t\t// fmt.Printf(\"Found %v transactions in the db.\", count)\n\t\tif count == 0 {\n\t\t\ttxTime := \"1603203489229912200\"\n\t\t\ttxType := \"1\"\n\t\t\ttxSubg := \"0\"\n\t\t\ttxPrnt := \"0\"\n\t\t\ttxData := \"8d3729b91a13878508c564fbf410ae4f33fcb4cfdb99677f4b23d4c4adb447650964b4fe9da16299831b9cc17aaabd5b8d81fb05460be92af99d128584101a30\" // ?\n\t\t\ttxPrev := \"c66f4851618cd53104d4a395212958abf88d96962c0c298a0c7a7c1242fac5c2ee616c8c4f140a2e199558ead6d18ae263b2311b590b0d7bf3777be5b3623d9c\" // RockSteady was here\n\t\t\thash := sha512.Sum512([]byte(txTime + txType + txData + txPrev))\n\t\t\ttxHash := hex.EncodeToString(hash[:])\n\t\t\ttxMile := true\n\t\t\ttxLead := false\n\t\t\ttxEpoc := txHash\n\t\t\ttx := db.MustBegin()\n\t\t\ttx.MustExec(\"INSERT INTO \"+d.Cf.GetTableName()+\" (tx_time, tx_type, tx_hash, tx_data, tx_prev, tx_epoc, tx_subg, tx_prnt, tx_mile, tx_lead, tx_height ) VALUES ($1, $2, $3, $4, $5, $6, $7, $8, $9, $10, $11)\", txTime, txType, txHash, txData, txPrev, txEpoc, txSubg, txPrnt, txMile, txLead, 0)\n\t\t\ttx.Commit()\n\t\t\treturn nil\n\t\t} else if count > 0 {\n\t\t\treturn errors.New(\"Root tx already present. \")\n\t\t}\n\t}\n\treturn nil\n}", "func CreateTransfer(w http.ResponseWriter, r *http.Request) {\r\n\tparams := mux.Vars(r)\r\n\tvar v = params[\"id\"] //We use a simpler variable as the two 'Transfer' functions are fairly long\r\n\t\r\n\tfor _, item := range certif {\r\n\t\t//Let's first check the user has the right to transfer this certificate\r\n if item.Id == v && item.Ownerid==clientnum { \r\n\t\t\tcheckk:=\"ok\"\r\n\t\t\t\r\n\t\t\t//We browse the transfers list to verify there is no transfer for this certificate\r\n\t\t\tfor _, itemm := range trans {\r\n\t\t\t\tif itemm.Id == v {\r\n\t\t\t\t\tcheckk=\"not ok\"\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t\t//If so, we inform the user\r\n\t\t\tif checkk==\"not ok\" {\r\n\t\t\t\tfmt.Fprintf(w, \"There is already a transfer for this certificate which is pending.\")\r\n\t\t\t} else {\r\n\t\t\t\tvar transcertif Transfer //Creation of a new transfer\r\n\t\t\t\t\r\n\t\t\t\t//We collect the receiving party's email address \r\n\t\t\t\t_ = json.NewDecoder(r.Body).Decode(&transcertif) \r\n\t\t\t\t\r\n\t\t\t\ttranscertif.Id=v //Certificate ID stays unchanged \r\n\t\t\t\ttranscertif.Status=\"Pending\" //The transfer status is now changed until the answer of the receiving party\r\n\t\t\t\t\r\n\t\t\t\t//We verify if the email address of the receiver is correct and is not the current user's \r\n\t\t\t\tcheck:=\"not ok\" \r\n\t\t\t\tfor _, ite := range owner {\r\n\t\t\t\t\tif ite.Email==transcertif.To && ite.Userid!=clientnum {\r\n\t\t\t\t\t\tcheck=\"ok\"\r\n\t\t\t\t\t}\r\n\t\t\t\t}\r\n\t\t\t\t\r\n\t\t\t\tif check==\"ok\" { //Then, if the email is valid, the transfer proposal is sent\r\n\t\t\t\t\t\ttrans=append(trans, transcertif) //The transfer is added to the transfers list\r\n\t\t\t\t\t\t//The 'transfer' status of the certificate is changed\r\n\t\t\t\t\t\tfor index, itp := range certif {\r\n\t\t\t\t\t\t\tif itp.Id==v {\r\n\t\t\t\t\t\t\t\tcertif[index].Transfer=\"Transfer pending\"\r\n\t\t\t\t\t\t\t}\r\n\t\t\t\t\t\t}\r\n\t\t\t\t\t\tfmt.Fprintf(w, \"Transfer proposal sent.\")\r\n\t\t\t\t\t\t\r\n\t\t\t\t}\telse { //If the email is not valid, the user is informed\r\n\t\t\t\t\t\tfmt.Fprintf(w, \"This email is not valid.\")\r\n\t\t\t\t}\r\n\t\t\t}\r\n\t\t} else if item.Id == v && item.Ownerid!=clientnum {\r\n\t\t\tfmt.Fprintf(w, \"You are not allowed to transfer this certificate.\")\r\n\t\t}\r\n\t}\r\n}" ]
[ "0.7724118", "0.7264912", "0.70775414", "0.6888156", "0.68655086", "0.68548596", "0.6757293", "0.6735288", "0.67024076", "0.6690598", "0.66825473", "0.66563475", "0.66508013", "0.6650783", "0.6646745", "0.6620966", "0.651692", "0.6492073", "0.64902925", "0.64711386", "0.64431155", "0.6413492", "0.6349267", "0.6324566", "0.6317189", "0.6317171", "0.63057953", "0.6300514", "0.629059", "0.62549233", "0.624068", "0.62293965", "0.6223501", "0.6218654", "0.6213387", "0.6183986", "0.6165097", "0.61645687", "0.6162605", "0.6137689", "0.61369556", "0.6129068", "0.60602653", "0.60508245", "0.6050218", "0.60404915", "0.6036547", "0.60299987", "0.60229063", "0.6019304", "0.60180694", "0.6013867", "0.6009765", "0.60041904", "0.59954953", "0.599169", "0.59661", "0.59639084", "0.594549", "0.59435207", "0.59420186", "0.5926969", "0.5926498", "0.5924816", "0.5915466", "0.590202", "0.589787", "0.5877139", "0.5876875", "0.58694655", "0.5852218", "0.58480793", "0.5842476", "0.5840367", "0.5835763", "0.58305955", "0.5823804", "0.58232665", "0.57938266", "0.57832116", "0.57825506", "0.57759595", "0.57736665", "0.57582563", "0.5751235", "0.575008", "0.57492316", "0.5744091", "0.574083", "0.57363623", "0.57355344", "0.57302046", "0.57297397", "0.57266575", "0.5724789", "0.57241297", "0.5717956", "0.5708074", "0.57025945", "0.57023287" ]
0.6715117
8
Handle Transaction Detail Show
func (main *Main) GetDetail(e echo.Context) (err error) { // get path parameter transactionCode := e.Param("code") // get transaction details transactionDetail, exc := TransactionModel.Get("code", transactionCode) if exc != nil { return rest.ConstructErrorResponse(e, exc) } // prepare data data := map[string]contract.Model{ "transaction_detail": output.NewTransactionDetail(transactionDetail), } return rest.ConstructSuccessResponse(e, data) }
{ "objective": { "self": [], "paired": [], "triplet": [ [ "query", "document", "negatives" ] ] } }
[ "func (tc *TransactionsController) Show(c *gin.Context) {\n\thash := common.HexToHash(c.Param(\"TxHash\"))\n\tif tx, err := tc.App.GetStore().FindTxByAttempt(hash); err == orm.ErrorNotFound {\n\t\tpublicError(c, http.StatusNotFound, errors.New(\"Transaction not found\"))\n\t} else if err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else if doc, err := jsonapi.Marshal(presenters.NewTx(tx)); err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err)\n\t} else {\n\t\tc.Data(http.StatusOK, MediaType, doc)\n\t}\n}", "func ViewTransaction(rw http.ResponseWriter, r *http.Request) {\n\t// get the token\n\treqToken := r.Header.Get(\"Authorization\")\n\n\t// get the claims\n\tclaims, isNotValid := GetClaims(reqToken, rw)\n\tif isNotValid {\n\t\treturn\n\t}\n\n\tdt, err := db.GetUserTransaction(claims.Roll)\n\tif err != nil {\n\t\trw.WriteHeader(http.StatusInternalServerError)\n\t\trw.Write(Rsp(err.Error(), \"Server Error\"))\n\t\treturn\n\t}\n\trw.WriteHeader(http.StatusOK)\n\tres := c.RespData{\n\t\tMessage: \"All data\",\n\t\tData: dt,\n\t}\n\tjson.NewEncoder(rw).Encode(res)\n}", "func (c *PurchaseController) Show(ctx *app.ShowPurchaseContext) error {\n\n\tsession := Database.Session.Copy()\n\tdefer session.Close()\n\n\tresult := app.Purchase{}\n\n\terr := session.DB(\"services-pos\").C(\"Purchase\").FindId(bson.ObjectIdHex(ctx.TransactionID)).One(&result)\n\n\tif err != nil {\n\t\treturn ctx.NotFound()\n\t}\n\n\tresult.TransactionID = ctx.TransactionID\n\tresult.Href = app.PurchaseHref(ctx.TransactionID)\n\n\treturn ctx.OK(&result)\n}", "func (g *Goods) Detail(c Context) {\n\t// TODO\n\tc.String(http.StatusOK, \"get goods detail\")\n}", "func (this *FamilyAccount) showDetails() {\n\tfmt.Println(\"------My Income and Expense Detail-------\")\n\tif this.flag {\n\t\t//因为我们用的是FamilyAccount结构体里传过来的字段,所以不能直接yongflag, 要用this. , 表示调用这个方法的结构体变量里面的字段\n\t\tfmt.Println(this.details)\n\t} else {\n\t\tfmt.Println(\"No current income and expenditure details!\")\n\t}\n}", "func (r *BTCRPC) GetTransactionDetail(txhash string) ([]byte, error) {\n\tvar (\n\t\ttx []byte\n\t\terr error\n\t)\n\n\terr = r.Client.Call(\"getrawtransaction\", jsonrpc.Params{txhash, 1}, &tx)\n\treturn tx, err\n}", "func (h *Handler) show(c echo.Context) (e error) {\n\tctx := c.(*cuxs.Context)\n\n\tvar id int64\n\tvar as *model.SalesReturn\n\tif id, e = common.Decrypt(ctx.Param(\"id\")); e == nil {\n\t\tif as, e = ShowSalesReturn(\"id\", id); e == nil {\n\t\t\tctx.Data(as)\n\t\t} else {\n\t\t\te = echo.ErrNotFound\n\t\t}\n\t}\n\treturn ctx.Serve(e)\n}", "func (p *Proxy) handleShowTxnz(session *driver.Session, query string, node sqlparser.Statement) (*sqltypes.Result, error) {\n\treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// privilegePlug := spanner.plugins.PlugPrivilege()\n\t// if !privilegePlug.IsSuperPriv(session.User()) {\n\t// \treturn nil, sqldb.NewSQLErrorf(sqldb.ER_SPECIFIC_ACCESS_DENIED_ERROR, \"Access denied; lacking super privilege for the operation\")\n\t// }\n\n\t// qr := &sqltypes.Result{}\n\t// qr.Fields = []*querypb.Field{\n\t// \t{Name: \"TxnID\", Type: querypb.Type_INT64},\n\t// \t{Name: \"Start\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"Duration\", Type: querypb.Type_INT32},\n\t// \t{Name: \"XaState\", Type: querypb.Type_VARCHAR},\n\t// \t{Name: \"TxnState\", Type: querypb.Type_VARCHAR},\n\t// }\n\n\t// rows := spanner.scatter.Txnz().GetTxnzRows()\n\t// for _, row := range rows {\n\t// \trow := []sqltypes.Value{\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT64, []byte(fmt.Sprintf(\"%v\", uint64(row.TxnID)))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.Start.Format(\"20060102150405.000\"))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_INT32, []byte(fmt.Sprintf(\"%v\", row.Duration))),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.XaState)),\n\t// \t\tsqltypes.MakeTrusted(querypb.Type_VARCHAR, []byte(row.State)),\n\t// \t}\n\t// \tqr.Rows = append(qr.Rows, row)\n\t// }\n\t// return qr, nil\n}", "func (showTxCommand ShowTransactionCommand) Run(ctx context.Context, wallet walletcore.Wallet) error {\n\ttransaction, err := wallet.GetTransaction(showTxCommand.Args.TxHash)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tbasicOutput := \"Hash\\t%s\\n\" +\n\t\t\"Confirmations\\t%d\\n\" +\n\t\t\"Included in block\\t%d\\n\" +\n\t\t\"Type\\t%s\\n\" +\n\t\t\"Amount %s\\t%s\\n\" +\n\t\t\"Date\\t%s\\n\" +\n\t\t\"Size\\t%s\\n\" +\n\t\t\"Fee\\t%s\\n\" +\n\t\t\"Rate\\t%s/kB\\n\"\n\n\ttxDirection := strings.ToLower(transaction.Direction.String())\n\ttxSize := fmt.Sprintf(\"%.1f kB\", float64(transaction.Size)/1000)\n\tbasicOutput = fmt.Sprintf(basicOutput,\n\t\ttransaction.Hash,\n\t\ttransaction.Confirmations,\n\t\ttransaction.BlockHeight,\n\t\ttransaction.Type,\n\t\ttxDirection, transaction.Amount,\n\t\ttransaction.FormattedTime,\n\t\ttxSize,\n\t\ttransaction.Fee,\n\t\ttransaction.FeeRate)\n\n\tif showTxCommand.Detailed {\n\t\tdetailedOutput := strings.Builder{}\n\t\tdetailedOutput.WriteString(\"General Info\\n\")\n\t\tdetailedOutput.WriteString(basicOutput)\n\t\tdetailedOutput.WriteString(\"\\nInputs\\n\")\n\t\tfor _, input := range transaction.Inputs {\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t%s\\n\", dcrutil.Amount(input.AmountIn).String(), input.PreviousOutpoint))\n\t\t}\n\t\tdetailedOutput.WriteString(\"\\nOutputs\\n\")\n\t\tfor _, out := range transaction.Outputs {\n\t\t\tif len(out.Addresses) == 0 {\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\\t (no address)\\n\", dcrutil.Amount(out.Value).String()))\n\t\t\t\tcontinue\n\t\t\t}\n\n\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"%s\", dcrutil.Amount(out.Value).String()))\n\t\t\tfor _, address := range out.Addresses {\n\t\t\t\taccountName := address.AccountName\n\t\t\t\tif !address.IsMine {\n\t\t\t\t\taccountName = \"external\"\n\t\t\t\t}\n\t\t\t\tdetailedOutput.WriteString(fmt.Sprintf(\"\\t%s (%s)\\n\", address.Address, accountName))\n\t\t\t}\n\t\t}\n\t\ttermio.PrintStringResult(strings.TrimRight(detailedOutput.String(), \" \\n\\r\"))\n\t} else {\n\t\ttermio.PrintStringResult(basicOutput)\n\t}\n\treturn nil\n}", "func viewAnyOrderGet(c *gin.Context) { //admin also have the same view , later combine those two func TBD\n\tOrdID := c.Request.URL.Query()[\"ordid\"][0] // Getting Order ID passed with URL\n\t_, usrName := session.SessinStatus(c, \"user_session_cookie\")\n\tfmt.Println(\"Wnat to see the order details of order number \", OrdID)\n\toK, itemsList, date, status, PayMode, amt := db.GetSingleOredrDetails(OrdID)\n\tif !oK {\n\t\tfmt.Println(\"Something went wrong while picking Single Order Deatils ..Please have a look\")\n\t}\n\tfmt.Println(oK, itemsList, date, status, PayMode, amt)\n\t//\t\tsubTotalToFloat, _ := strconv.ParseFloat(singleCartItem.SubTotal, 64)\n\t//\t\tTotalAmt = TotalAmt + subTotalToFloat\n\t//\tTotalAmtInPaisa := TotalAmt * 100 // This is required while initate for payment in Razorpay\n\n\t//\tTotalAmtString := fmt.Sprintf(\"%.2f\", TotalAmt)\n\n\tc.HTML(\n\t\thttp.StatusOK,\n\t\t\"view_particular_order.html\",\n\t\tgin.H{\"title\": \"OrderDetail\",\n\t\t\t\"ItemsOrdered\": itemsList,\n\t\t\t\"OrdID\": OrdID,\n\t\t\t\"date\": date,\n\t\t\t\"PayMode\": PayMode,\n\t\t\t\"amt\": amt,\n\t\t\t\"OrdStatus\": status,\n\t\t\t\"usrName\": usrName,\n\n\t\t\t// \"TotalAmt\": TotalAmtString,\n\t\t\t// \"TotalAmtInPaisa\": TotalAmtInPaisa,\n\t\t},\n\t)\n\n}", "func (s *Service) GetExplorerTransaction(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tid := r.FormValue(\"id\")\n\n\tdata := &Data{}\n\tdefer func() {\n\t\tif err := json.NewEncoder(w).Encode(data.TX); err != nil {\n\t\t\tutils.Logger().Warn().Err(err).Msg(\"cannot JSON-encode TX\")\n\t\t}\n\t}()\n\tif id == \"\" {\n\t\tutils.Logger().Warn().Msg(\"invalid id parameter\")\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\treturn\n\t}\n\tdb := s.Storage.GetDB()\n\tbytes, err := db.Get([]byte(GetTXKey(id)))\n\tif err != nil {\n\t\tutils.Logger().Warn().Err(err).Str(\"id\", id).Msg(\"cannot read TX\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\ttx := new(Transaction)\n\tif rlp.DecodeBytes(bytes, tx) != nil {\n\t\tutils.Logger().Warn().Str(\"id\", id).Msg(\"cannot convert data from DB\")\n\t\tw.WriteHeader(http.StatusInternalServerError)\n\t\treturn\n\t}\n\tdata.TX = *tx\n}", "func (ctl *SaleCounterProductController) Detail() {\n\t//获取信息一样,直接调用Edit\n\tctl.Edit()\n\tctl.Data[\"Readonly\"] = true\n\tctl.Data[\"Action\"] = \"detail\"\n}", "func orderAdminApproveGet(c *gin.Context) {\n\tOrdID := c.Request.URL.Query()[\"ordid\"][0] // Getting Order ID passed with URL\n\t_, usrName := session.SessinStatus(c, \"user_session_cookie\")\n\tfmt.Println(\"Wnat to see the order details of order number \", OrdID)\n\toK, itemsList, date, status, PayMode, amt := db.GetSingleOredrDetails(OrdID)\n\tif !oK {\n\t\tfmt.Println(\"Something went wrong while picking Single Order Deatils ..Please have a look\")\n\t}\n\tfmt.Println(oK, itemsList, date, status, PayMode, amt)\n\tc.HTML(\n\t\thttp.StatusOK,\n\t\t\"order_adminview.html\",\n\t\tgin.H{\"title\": \"OrderDetail\",\n\t\t\t\"ItemsOrdered\": itemsList,\n\t\t\t\"OrdID\": OrdID,\n\t\t\t\"date\": date,\n\t\t\t\"PayMode\": PayMode,\n\t\t\t\"amt\": amt,\n\t\t\t\"OrdStatus\": status,\n\t\t\t\"usrName\": usrName,\n\n\t\t\t// \"TotalAmt\": TotalAmtString,\n\t\t\t// \"TotalAmtInPaisa\": TotalAmtInPaisa,\n\t\t},\n\t)\n}", "func (c *TradeController) Show(ctx *app.ShowTradeContext) error {\n\t// TradeController_Show: start_implement\n\n\t// Put your logic here\n\tt, ok := tradeRegistry[ctx.TradeID]\n\tif !ok {\n\t\treturn ctx.NotFound()\n\t}\n\tres := &app.GoaTrade{\n\t\tTradeID: t.TradeID,\n\t\tContractID: t.ContractID,\n\t\tCounterpartyID: t.CounterpartyID,\n\t}\n\treturn ctx.OK(res)\n\t// TradeController_Show: end_implement\n}", "func (action *TransactionShowAction) JSON() {\n\taction.Do(\n\t\taction.EnsureHistoryFreshness,\n\t\taction.loadParams,\n\t\taction.checkAllowed,\n\t\taction.loadRecord,\n\t\taction.loadResource,\n\t\tfunc() { hal.Render(action.W, action.Resource) },\n\t)\n}", "func view(ctx context.Context, db transactor, fn func(*bolt.Tx) error) error {\n\ttx, ok := ctx.Value(transactionKey{}).(*bolt.Tx)\n\tif !ok {\n\t\treturn db.View(fn)\n\t}\n\treturn fn(tx)\n}", "func (db *DB) View(ctx context.Context, fn func(*TransactionManager) error) error {\n\ttx := db.BeginTransaction(false)\n\tdefer tx.Discard()\n\treturn fn(tx)\n}", "func displayTransactions(scid string, option int, ID string) {\n\t\t\n\tscKeys:= []string{\"numberOfOwners\", \"txCount\"}\n\tresult:= getKeysFromDaemon(scid, scKeys)\n\tif result == \"\" {return}\n\n\n\t//Response ok, extract keys from JSON\n\t\n\n\ttxCount := gjson.Get(result, \"txs.#.sc_keys.txCount\")\n\ttxCountArray:= txCount.Array()[0]\n\ttxCountInt:= txCountArray.Int()\n\t//fmt.Printf(\"Tx Count: %d\\n\", txCountInt)\n\n\t//Make a slice of keys so we can request in RPC call\n\tx:= int(txCountInt) //txCount in wallet smart contract is always 1 ahead of actual number of transactions\t\n\tx4:= x * 4\t\n\tkeySlice:= make([]string, x4) \n\t\n\tfor i:=0; i<x; i++ {\n\t\tz:= strconv.Itoa(i)\n\t\tkeySlice[i] = \"txIndex_\" + z\n\t\tkeySlice[i+x] = \"recipient_\" + z\n\t\tkeySlice[i+(x*2)] = \"amount_\" + z\n\t\tkeySlice[i+(x*3)] = \"sent_\" + z\n\t}\n\t\t\n\t//fmt.Println(keySlice)\n\tdisplayData(scid, keySlice, x, option, ID)\n\n\n}", "func (s *Store) Show(c *gin.Context) {\n\n}", "func (h *Reports) Transactions(ctx context.Context, w http.ResponseWriter, r *http.Request, params map[string]string) error {\n\n\tvar data = make(map[string]interface{})\n\tvar total float64\n\n\tclaims, err := auth.ClaimsFromContext(ctx)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tfields := []datatable.DisplayField{\n\t\t{Field: \"id\", Title: \"ID\", Visible: false, Searchable: true, Orderable: true, Filterable: false},\n\t\t{Field: \"amount\", Title: \"Amount\", Visible: true, Searchable: false, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Quantity\"},\n\t\t{Field: \"created_at\", Title: \"Date\", Visible: true, Searchable: true, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Date\"},\n\t\t{Field: \"customer_name\", Title: \"Customer\", Visible: true, Searchable: true, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Account\"},\n\t\t{Field: \"account\", Title: \"Account Number\", Visible: true, Searchable: true, Orderable: true, Filterable: true, FilterPlaceholder: \"filter Account\"},\n\t\t{Field: \"sales_rep_id\", Title: \"Recorded By\", Visible: true, Searchable: true, Orderable: false, Filterable: true, FilterPlaceholder: \"filter Recorder\"},\n\t}\n\n\tmapFunc := func(q transaction.TxReportResponse, cols []datatable.DisplayField) (resp []datatable.ColumnValue, err error) {\n\t\tfor i := 0; i < len(cols); i++ {\n\t\t\tcol := cols[i]\n\t\t\tvar v datatable.ColumnValue\n\t\t\tswitch col.Field {\n\t\t\tcase \"id\":\n\t\t\t\tv.Value = fmt.Sprintf(\"%s\", q.ID)\n\t\t\tcase \"amount\":\n\t\t\t\tv.Value = fmt.Sprintf(\"%f\", q.Amount)\n\t\t\t\tp := message.NewPrinter(language.English)\n\t\t\t\tv.Formatted = p.Sprintf(\"<a href='%s'>%.2f</a>\", urlCustomersTransactionsView(q.CustomerID, q.AccountID, q.ID), q.Amount)\n\t\t\tcase \"created_at\":\n\t\t\t\tdate := web.NewTimeResponse(ctx, time.Unix(q.CreatedAt, 0))\n\t\t\t\tv.Value = date.LocalDate\n\t\t\t\tv.Formatted = date.LocalDate\n\t\t\tcase \"narration\":\n\t\t\t\tvalues := strings.Split(q.Narration, \":\")\n\t\t\t\tif len(values) > 1 {\n\t\t\t\t\tif values[0] == \"sale\" {\n\t\t\t\t\t\tv.Value = values[1]\n\t\t\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlSalesView(values[2]), v.Value)\n\t\t\t\t\t}\n\t\t\t\t} else {\n\t\t\t\t\tv.Value = q.Narration\n\t\t\t\t\tv.Formatted = q.Narration\n\t\t\t\t}\n\t\t\tcase \"payment_method\":\n\t\t\t\tv.Value = q.PaymentMethod\n\t\t\t\tv.Formatted = q.PaymentMethod\n\t\t\tcase \"customer_name\":\n\t\t\t\tv.Value = q.CustomerName\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlCustomersView(q.CustomerID), v.Value)\n\t\t\tcase \"account\":\n\t\t\t\tv.Value = q.AccountNumber\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlCustomersAccountsView(q.CustomerID, q.AccountID), v.Value)\n\t\t\tcase \"sales_rep_id\":\n\t\t\t\tv.Value = q.SalesRepID\n\t\t\t\tv.Formatted = fmt.Sprintf(\"<a href='%s'>%s</a>\", urlUsersView(q.SalesRepID), q.SalesRep)\n\t\t\tdefault:\n\t\t\t\treturn resp, errors.Errorf(\"Failed to map value for %s.\", col.Field)\n\t\t\t}\n\t\t\tresp = append(resp, v)\n\t\t}\n\n\t\treturn resp, nil\n\t}\n\n\tvar txWhere = []string{\"tx_type = 'deposit'\"}\n\tvar txArgs []interface{}\n\t// todo sales rep filtering\n\tif v := r.URL.Query().Get(\"sales_rep_id\"); v != \"\" {\n\t\ttxWhere = append(txWhere, \"sales_rep_id = $1\")\n\t\ttxArgs = append(txArgs, v)\n\t\tdata[\"salesRepID\"] = v\n\t}\n\n\tif v := r.URL.Query().Get(\"payment_method\"); v != \"\" {\n\t\ttxWhere = append(txWhere, fmt.Sprintf(\"payment_method = $%d\", len(txArgs)+1))\n\t\ttxArgs = append(txArgs, v)\n\t\tdata[\"paymentMethod\"] = v\n\t}\n\n\tvar date = time.Now()\n\tif v := r.URL.Query().Get(\"start_date\"); v != \"\" {\n\t\tdate, err = time.Parse(\"01/02/2006\", v)\n\t\tif err != nil {\n\t\t\tdate = time.Now()\n\t\t\treturn err\n\t\t}\n\t}\n\tdate = date.Truncate(time.Millisecond)\n\tdate = now.New(date).BeginningOfDay().Add(-1 * time.Hour)\n\ttxWhere = append(txWhere, fmt.Sprintf(\"created_at >= $%d\", len(txArgs)+1))\n\ttxArgs = append(txArgs, date.UTC().Unix())\n\tdata[\"startDate\"] = date.Format(\"01/02/2006\")\n\n\tdate = time.Now()\n\tif v := r.URL.Query().Get(\"end_date\"); v != \"\" {\n\t\tdate, err = time.Parse(\"01/02/2006\", v)\n\t\tif err != nil {\n\t\t\tdate = time.Now()\n\t\t\treturn err\n\t\t}\n\n\t}\n\tdate = date.Truncate(time.Millisecond)\n\tdate = now.New(date).EndOfDay().Add(-1 * time.Hour)\n\ttxWhere = append(txWhere, fmt.Sprintf(\"created_at <= $%d\", len(txArgs)+1))\n\ttxArgs = append(txArgs, date.Unix())\n\tdata[\"endDate\"] = date.Format(\"01/02/2006\")\n\n\tloadFunc := func(ctx context.Context, sorting string, fields []datatable.DisplayField) (resp [][]datatable.ColumnValue, err error) {\n\n\t\tvar order []string\n\t\tif len(sorting) > 0 {\n\t\t\torder = strings.Split(sorting, \",\")\n\t\t}\n\n\t\tfor i := range txWhere {\n\t\t\ttxWhere[i] = \"tx.\" + txWhere[i]\n\t\t}\n\t\tres, err := h.TransactionRepo.TxReport(ctx, claims, transaction.FindRequest{\n\t\t\tOrder: order, Where: strings.Join(txWhere, \" AND \"), Args: txArgs,\n\t\t})\n\t\tif err != nil {\n\t\t\treturn resp, err\n\t\t}\n\n\t\tfor _, a := range res {\n\t\t\tl, err := mapFunc(a, fields)\n\t\t\tif err != nil {\n\t\t\t\treturn resp, errors.Wrapf(err, \"Failed to map brand for display.\")\n\t\t\t}\n\n\t\t\tresp = append(resp, l)\n\t\t}\n\n\t\treturn resp, nil\n\t}\n\n\tdt, err := datatable.New(ctx, w, r, h.Redis, fields, loadFunc)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tif dt.HasCache() {\n\t\treturn nil\n\t}\n\n\tif ok, err := dt.Render(); ok {\n\t\tif err != nil {\n\t\t\treturn err\n\t\t}\n\t\treturn nil\n\t}\n\n\tusers, err := h.UserRepos.Find(ctx, claims, user.UserFindRequest{\n\t\tOrder: []string{\"first_name\", \"last_name\"},\n\t})\n\tif err != nil {\n\t\treturn err\n\t}\n\n\ttotal, err = h.TransactionRepo.DepositAmountByWhere(ctx, strings.Join(txWhere, \" and \"), txArgs)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\tdata[\"paymentMethods\"] = transaction.PaymentMethods\n\tdata[\"users\"] = users\n\tdata[\"total\"] = total\n\tdata[\"datatable\"] = dt.Response()\n\n\treturn h.Renderer.Render(ctx, w, r, TmplLayoutBase, \"report-transactions.gohtml\", web.MIMETextHTMLCharsetUTF8, http.StatusOK, data)\n}", "func viewandedititemGet(c *gin.Context) {\n\tIsSectionActive, _ := session.SessinStatus(c, \"admin_session_cookie\")\n\tif !IsSectionActive {\n\t\tfmt.Println(\"No Active Sessions found \")\n\t\t// c.HTML(http.StatusOK, \"admin_login.html\", []string{\"a\", \"b\", \"c\"})\n\t\tc.HTML(\n\t\t\thttp.StatusOK,\n\t\t\t\"admin_login.html\",\n\t\t\tgin.H{\"title\": \"success login\",\n\t\t\t\t\"diplay\": \"none\",\n\t\t\t},\n\t\t)\n\t} else {\n\t\titemID := c.Request.URL.Query()[\"itemid\"][0] // Getting Order ID passed with URL\n\t\tfmt.Println(\"Initiating to View/Edit item ,having ID\", itemID)\n\t\t//populateCategoryItems(c, itemID)\n\t\t//GetItemDetails(itemID string) (itemDesc string, itemRate float64, unit string,itmID,itmStock int,itmBuyRate float64) {\n\t\t//Don't Confuse above function will redirect\n\t\t//to edit page, usual practice is giving here\n\t\t//but we achived this by modifying the existing\n\t\t//code so it happend so..\n\t\titmDesc, itmSelRate, itmUnit, itmID, itmStock, itmBuyPrice := db.GetItemDetails(itemID)\n\t\tc.HTML(\n\t\t\thttp.StatusOK,\n\t\t\t\"edit_item.html\", gin.H{\n\t\t\t\t\"delWarning\": \"none\",\n\t\t\t\t\"updateSucess\": \"none\",\n\t\t\t\t\"title\": \"Edit Item\",\n\t\t\t\t\"itmID\": itmID,\n\t\t\t\t\"itmDesc\": itmDesc,\n\t\t\t\t\"itmUnit\": itmUnit,\n\t\t\t\t\"itmBuyPrice\": itmBuyPrice,\n\t\t\t\t\"itmSelRate\": itmSelRate,\n\t\t\t\t\"itmStock\": itmStock,\n\t\t\t})\n\t}\n}", "func (db *DB) View(fn func(*TransactionManager) error) error {\n\ttx := db.BeginTransaction(false)\n\tdefer tx.Discard()\n\treturn fn(tx)\n}", "func transactionSummary(ctx *quorumContext, w http.ResponseWriter, r *http.Request) (int, error) {\n\tif ctx.conn == nil {\n\t\tfmt.Fprintf(w, \"Cannot get transaction before deploying contract\\n\")\n\t\treturn 400, nil\n\t}\n\n\t// parse hash from request URL\n\tkeys := r.URL.Query()[\"hash\"]\n\tif len(keys) < 1 {\n\t\tfmt.Fprintf(w, \"Invalid parameter, require 'hash'\")\n\t\treturn 400, nil\n\t}\n\tlog.Println(\"Hash supplied :\", keys[0])\n\thash := common.HexToHash(keys[0])\n\n\tbasic_context := context.Background()\n\ttx, pending, err := ctx.conn.TransactionByHash(basic_context, hash)\n\t//common.HexToHash(\"0x378674bebd1430d9ce63adc792c573da56e69b8d6c97174c93a43c5991ae0d61\"))\n\tif err != nil {\n\t\tfmt.Fprintf(w, \"Failed to get transaction details: %v\", err)\n\t\treturn 500, err\n\t}\n\tfmt.Fprintf(w, \"Transaction pending? %v; details: %v\\n\",\n\t\tpending, tx.String())\n\treturn 200, nil\n}", "func (self* userRestAPI) transactions(w http.ResponseWriter, r *http.Request) {\n\n\t// Read arguments\n\tband,number,err := self.extractBandAndNumber(r)\n\tif err != nil {\n\t\tlogError(err)\n\t\thttp.Error(w, fmt.Sprintf(\"\\nFailed to parse arguments '%s'\\n\",err), http.StatusInternalServerError)\n\t\treturn\n\t}\n\n\t// Retrieve transactions for specified traveller\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\thistory,err := self.engine.TransactionsAsJSON(band,number)\n\tif err != nil {\n\t\tlogError(err)\n\t\thttp.Error(w, fmt.Sprintf(\"\\nFailed to retrieve transaction history with error '%s'\\n\",err), http.StatusInternalServerError)\n\t\treturn\n\t}\n\tio.WriteString(w,history)\n\n}", "func (c *PlanController) PlanDetails() {\n\tplanViewModel := viewmodels.Plan{}\n\tr := c.Ctx.Request\n\tw := c.Ctx.ResponseWriter\n\tsessionValues, sessionStatus := SessionForPlan(w,r)\n\tplanViewModel.SessionFlag = sessionStatus\n\tplanViewModel.CompanyPlan = sessionValues.CompanyPlan\n\tplanViewModel.CompanyTeamName =sessionValues.CompanyTeamName\n\tc.Data[\"vm\"] = planViewModel\n\tc.TplName = \"template/plan.html\"\n}", "func (action TransactionShowAction) ServeHTTPC(c web.C, w http.ResponseWriter, r *http.Request) {\n\tap := &action.Action\n\tap.Prepare(c, w, r)\n\taction.Log = action.Log.WithField(\"action\", \"TransactionShowAction\")\n\tap.Execute(&action)\n}", "func (api *API) Get(tid string) (*pagarme.Response, *pagarme.Transaction, error) {\n\tresp, err := api.Config.Do(http.MethodGet, \"/transactions/\"+tid, nil)\n\tif err != nil {\n\t\treturn nil, nil, err\n\t}\n\tif werr := www.ExtractError(resp); werr != nil {\n\t\treturn werr, nil, nil\n\t}\n\tresult := &pagarme.Transaction{}\n\tif err := www.Unmarshal(resp, result); err != nil {\n\t\tapi.Config.Logger.Error(\"could not unmarshal transaction [Get]: \" + err.Error())\n\t\treturn nil, nil, err\n\t}\n\n\treturn www.Ok(), result, nil\n}", "func (tb *transactionBuilder) View() (types.Transaction, []types.Transaction) {\n\treturn tb.transaction, tb.parents\n}", "func (tb *transactionBuilder) View() (types.Transaction, []types.Transaction) {\n\treturn tb.transaction, tb.parents\n}", "func ShowDeposit2ChannelMsgDetail(Status string) {\n\tswitch Status {\n\tcase \"200 OK\":\n\t\tlog.Println(\"Success Deposit\")\n\tcase \"400 Bad Request\":\n\t\tlog.Println(\"The provided json is in some way malformed!\")\n\tcase \"402 Payment required\":\n\t\tlog.Println(\"Insufficient balance to do a deposit\")\n\tcase \"408 Request Timeout\":\n\t\tlog.Println(\"The deposit event was not read in time by the ethereum node\")\n\tcase \"409 Conflict\":\n\t\tlog.Println(\"Provided channel does not exist\")\n\tcase \"500 Server Error\":\n\t\tlog.Println(\"Internal Raiden node error\")\n\tcase \"504 TimeOut\":\n\t\tlog.Println(\"No response,timeout\")\n\tdefault:\n\t\tfmt.Printf(\"Unknown error,Deposit Failure! %s\\n\", Status)\n\t}\n}", "func (account *Account) ShowTx(P string) bool {\n\tfor i := 0; i < account.WalletIndex; i++ {\n\t\tif P == account.Wallet[i].P {\n\t\t\tJSON, _ := json.MarshalIndent(account.Wallet[i], \"\", \" \")\n\n\t\t\tfmt.Println(string(JSON))\n\t\t\treturn true\n\n\t\t}\n\t}\n\treturn false\n}", "func (s *searcher) Transaction(resp http.ResponseWriter, req *http.Request) {\n\tsearchTerms := mux.Vars(req)\n\n\ttransactionID := searchTerms[\"transaction_id\"]\n\tif len(transactionID) == 0 {\n\t\tresp.WriteHeader(http.StatusBadRequest)\n\t\tresp.Write([]byte(\"transaction ID is empty\"))\n\t\treturn\n\t}\n\n\tif len(transactionID) != 64 {\n\t\tresp.WriteHeader(http.StatusBadRequest)\n\t\tresp.Write([]byte(\"transaction ID is not 64 characters\"))\n\t\treturn\n\t}\n\n\tfileName, transactionIndex, err := s.searchIndex.GetTransactionPathByID(transactionID)\n\tif err != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tresp.Write([]byte(fmt.Sprintf(\"error finding transaction: %s\", err.Error())))\n\t\treturn\n\t}\n\n\ttransactions, err := s.searchIndex.GetTransactionsFromSingleFile(fileName, []int{transactionIndex})\n\tif err != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tresp.Write([]byte(fmt.Sprintf(\"error finding transaction: %s\", err.Error())))\n\t\treturn\n\t}\n\n\tresultBytes, err := json.Marshal(transactions)\n\tif err != nil {\n\t\tresp.WriteHeader(http.StatusInternalServerError)\n\t\tresp.Write([]byte(fmt.Sprintf(\"error marshallig transaction to json: %s\", err.Error())))\n\t\treturn\n\t}\n\n\tresp.WriteHeader(http.StatusOK)\n\tresp.Write(resultBytes)\n}", "func (db *MySQLDB) GetTransactionDetailByTransactionID(ctx context.Context, transactionID int) ([]*TransactionDetailRecord, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"GetTransactionByTransactionID\")\n\ttransaction := &TransactionRecord{}\n\n\trow := db.instance.QueryRowContext(ctx, \"SELECT id FROM transactions WHERE id = ?\", transactionID)\n\terr := row.Scan(&transaction.ID)\n\tif err != nil {\n\t\tfLog.Errorf(\"row.Scan got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\tq := fmt.Sprintf(\"SELECT td.transaction_id, td.product_id, td.qty, td.sub_total, p.sku FROM transaction_detail td INNER JOIN products p ON td.product_id = p.id WHERE transaction_id = %v\", transactionID)\n\trows, err := db.instance.QueryContext(ctx, q)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.QueryContext got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\ttDetail := make([]*TransactionDetailRecord, 0)\n\tfor rows.Next() {\n\t\ttD := &TransactionDetailRecord{}\n\t\terr := rows.Scan(&tD.TransactionID, &tD.ProductID, &tD.Qty, &tD.SubTotal, tD.SKU)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"rows.Scan got %s\", err.Error())\n\t\t} else {\n\t\t\ttDetail = append(tDetail, tD)\n\t\t}\n\t}\n\n\treturn tDetail, nil\n}", "func (res BackTestResult) Show() TradestatPort {\n\t//\tp := NewPortfolio()\n\tmds := bean.NewRPCMDSConnC(\"tcp\", res.dbhost+\":\"+res.dbport)\n\tratesbook := make(ReferenceRateBook)\n\n\t// FIXME: think about how to show multi pair result\n\tvar stat TradestatPort\n\tif len(res.pairs) > 0 {\n\t\tp := res.pairs[0]\n\t\ttxn, _ := mds.GetTransactions(p, res.start, res.end)\n\t\tratesbook[p] = RefRatesFromTxn(txn)\n\t\t//\t\tsnapts.Print()\n\t\t//\t\tperfts.Print()\n\n\t\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\n\t\tstat.Print()\n\t}\n\treturn stat\n}", "func (s *Store) TxDetails(ns walletdb.ReadBucket, txHash *chainhash.Hash) (*TxDetails, error) {\n\t// First, check whether there exists an unmined transaction with this\n\t// hash. Use it if found.\n\tv := existsRawUnmined(ns, txHash[:])\n\tif v != nil {\n\t\treturn s.unminedTxDetails(ns, txHash, v)\n\t}\n\n\t// Otherwise, if there exists a mined transaction with this matching\n\t// hash, skip over to the newest and begin fetching all details.\n\tk, v := latestTxRecord(ns, txHash)\n\tif v == nil {\n\t\t// not found\n\t\treturn nil, nil\n\t}\n\treturn s.minedTxDetails(ns, txHash, k, v)\n}", "func (dao *InfoDao) Transaction(ctx context.Context, f func(ctx context.Context, tx *gdb.TX) error) (err error) {\n\treturn dao.Ctx(ctx).Transaction(ctx, f)\n}", "func Show(c *gin.Context) {\r\n\tpost := getById(c)\r\n\tc.JSON(http.StatusOK, gin.H{\r\n\t\t\"messege\": \"\",\r\n\t\t\"data\": post,\r\n\t})\r\n}", "func tableInfoFlowShow(meta *halproto.TableMetadata) {\n\tfMeta := meta.GetFlowMeta()\n\tinsStr := fmt.Sprintf(\"%d[%d]\", fMeta.GetNumInserts(), fMeta.GetNumInsertFailures())\n\tupdStr := fmt.Sprintf(\"%d[%d]\", fMeta.GetNumUpdates(), fMeta.GetNumUpdateFailures())\n\tdelStr := fmt.Sprintf(\"%d[%d]\", fMeta.GetNumDeletes(), fMeta.GetNumDeleteFailures())\n\tfmt.Printf(\"%-30s%-10d%-10s%-10d%-10d%-10d%-10d%-10s%-10s%-10s\\n\",\n\t\tmeta.GetTableName(),\n\t\tmeta.GetTableId(),\n\t\tutils.TableKindToStr(meta.GetKind()),\n\t\tfMeta.GetCapacity(),\n\t\tfMeta.GetCollCapacity(),\n\t\tfMeta.GetHashUsage(),\n\t\tfMeta.GetCollUsage(),\n\t\tinsStr,\n\t\tupdStr,\n\t\tdelStr)\n}", "func (c *BalanceClient) RetrieveTransaction(id string) (*BalanceTransaction, error) {\n\tbalanceTransaction := BalanceTransaction{}\n\terr := c.client.get(\"/balance/history/\"+id, nil, &balanceTransaction)\n\treturn &balanceTransaction, err\n}", "func (s *Server) handleTransaction(client string, req *pb.Command) (err error) {\n\t// Get the transfer from the original command, will panic if nil\n\ttransfer := req.GetTransfer()\n\tmsg := fmt.Sprintf(\"starting transaction of %0.2f from %s to %s\", transfer.Amount, transfer.Account, transfer.Beneficiary)\n\ts.updates.Broadcast(req.Id, msg, pb.MessageCategory_LEDGER)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Handle Demo UI errors before the account lookup\n\tif transfer.OriginatingVasp != \"\" && transfer.OriginatingVasp != s.vasp.Name {\n\t\tlog.Info().Str(\"requested\", transfer.OriginatingVasp).Str(\"local\", s.vasp.Name).Msg(\"requested originator does not match local VASP\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrWrongVASP, \"message sent to the wrong originator VASP\"),\n\t\t)\n\t}\n\n\t// Lookup the account associated with the transfer originator\n\tvar account Account\n\tif err = LookupAccount(s.db, transfer.Account).First(&account).Error; err != nil {\n\t\tif errors.Is(err, gorm.ErrRecordNotFound) {\n\t\t\tlog.Info().Str(\"account\", transfer.Account).Msg(\"not found\")\n\t\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\t\tpb.Errorf(pb.ErrNotFound, \"account not found\"),\n\t\t\t)\n\t\t}\n\t\treturn fmt.Errorf(\"could not fetch account: %s\", err)\n\t}\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"account %04d accessed successfully\", account.ID), pb.MessageCategory_LEDGER)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Lookup the wallet of the beneficiary\n\tvar beneficiary Wallet\n\tif err = LookupBeneficiary(s.db, transfer.Beneficiary).First(&beneficiary).Error; err != nil {\n\t\tif errors.Is(err, gorm.ErrRecordNotFound) {\n\t\t\tlog.Info().Str(\"beneficiary\", transfer.Beneficiary).Msg(\"not found\")\n\t\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\t\tpb.Errorf(pb.ErrNotFound, \"beneficiary wallet not found\"),\n\t\t\t)\n\t\t}\n\t\treturn fmt.Errorf(\"could not fetch beneficiary wallet: %s\", err)\n\t}\n\n\tif transfer.CheckBeneficiary {\n\t\tif transfer.BeneficiaryVasp != beneficiary.Provider.Name {\n\t\t\tlog.Info().\n\t\t\t\tStr(\"expected\", transfer.BeneficiaryVasp).\n\t\t\t\tStr(\"actual\", beneficiary.Provider.Name).\n\t\t\t\tMsg(\"check beneficiary failed\")\n\t\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\t\tpb.Errorf(pb.ErrWrongVASP, \"beneficiary wallet does not match beneficiary vasp\"),\n\t\t\t)\n\t\t}\n\t}\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"wallet %s provided by %s\", beneficiary.Address, beneficiary.Provider.Name), pb.MessageCategory_BLOCKCHAIN)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// TODO: lookup peer from cache rather than always doing a directory service lookup\n\tvar peer *peers.Peer\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"search for %s in directory service\", beneficiary.Provider.Name), pb.MessageCategory_TRISADS)\n\tif peer, err = s.peers.Search(beneficiary.Provider.Name); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not search peer from directory service\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not search peer from directory service\"),\n\t\t)\n\t}\n\tinfo := peer.Info()\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"identified TRISA remote peer %s at %s via directory service\", info.ID, info.Endpoint), pb.MessageCategory_TRISADS)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\tvar signKey *rsa.PublicKey\n\ts.updates.Broadcast(req.Id, \"exchanging peer signing keys\", pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\tif signKey, err = peer.ExchangeKeys(true); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not exchange keys with remote peer\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not exchange keyrs with remote peer\"),\n\t\t)\n\t}\n\n\t// Prepare the transaction\n\t// Save the pending transaction and increment the accounts pending field\n\txfer := Transaction{\n\t\tEnvelope: uuid.New().String(),\n\t\tAccount: account,\n\t\tAmount: decimal.NewFromFloat32(transfer.Amount),\n\t\tDebit: true,\n\t\tCompleted: false,\n\t}\n\n\tif err = s.db.Save(&xfer).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not save transaction\"),\n\t\t)\n\t}\n\n\t// Save the pending transaction on the account\n\t// TODO: remove pending transactions\n\taccount.Pending++\n\tif err = s.db.Save(&account).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save originator account\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not save originator account\"),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, \"ready to execute transaction\", pb.MessageCategory_BLOCKCHAIN)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Create an identity and transaction payload for TRISA exchange\n\ttransaction := &generic.Transaction{\n\t\tTxid: fmt.Sprintf(\"%d\", xfer.ID),\n\t\tOriginator: account.WalletAddress,\n\t\tBeneficiary: beneficiary.Address,\n\t\tAmount: float64(transfer.Amount),\n\t\tNetwork: \"TestNet\",\n\t\tTimestamp: xfer.Timestamp.Format(time.RFC3339),\n\t}\n\tidentity := &ivms101.IdentityPayload{\n\t\tOriginator: &ivms101.Originator{},\n\t\tOriginatingVasp: &ivms101.OriginatingVasp{},\n\t}\n\tif identity.OriginatingVasp.OriginatingVasp, err = s.vasp.LoadIdentity(); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not load originator vasp\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not load originator vasp\"),\n\t\t)\n\t}\n\n\tidentity.Originator = &ivms101.Originator{\n\t\tOriginatorPersons: make([]*ivms101.Person, 0, 1),\n\t\tAccountNumbers: []string{account.WalletAddress},\n\t}\n\tvar originator *ivms101.Person\n\tif originator, err = account.LoadIdentity(); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not load originator identity\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not load originator identity\"),\n\t\t)\n\t}\n\tidentity.Originator.OriginatorPersons = append(identity.Originator.OriginatorPersons, originator)\n\n\tpayload := &protocol.Payload{}\n\tif payload.Transaction, err = anypb.New(transaction); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not serialize transaction payload\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not serialize transaction payload\"),\n\t\t)\n\t}\n\tif payload.Identity, err = anypb.New(identity); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not serialize identity payload\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not serialize identity payload\"),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, \"transaction and identity payload constructed\", pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Secure the envelope with the remote beneficiary's signing keys\n\tvar envelope *protocol.SecureEnvelope\n\tif envelope, err = handler.New(xfer.Envelope, payload, nil).Seal(signKey); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not create or sign secure envelope\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not create or sign secure envelope\"),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"secure envelope %s sealed: encrypted with AES-GCM and RSA - sending ...\", envelope.Id), pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Conduct the TRISA transaction, handle errors and send back to user\n\tif envelope, err = peer.Transfer(envelope); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not perform TRISA exchange\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"received %s information exchange reply from %s\", envelope.Id, peer.String()), pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Open the response envelope with local private keys\n\tvar opened *handler.Envelope\n\tif opened, err = handler.Open(envelope, s.trisa.sign); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not unseal TRISA response\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\t// Verify the contents of the response\n\tpayload = opened.Payload\n\tif payload.Identity.TypeUrl != \"type.googleapis.com/ivms101.IdentityPayload\" {\n\t\tlog.Warn().Str(\"type\", payload.Identity.TypeUrl).Msg(\"unsupported identity type\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"unsupported identity type\", payload.Identity.TypeUrl),\n\t\t)\n\t}\n\n\tif payload.Transaction.TypeUrl != \"type.googleapis.com/trisa.data.generic.v1beta1.Transaction\" {\n\t\tlog.Warn().Str(\"type\", payload.Transaction.TypeUrl).Msg(\"unsupported transaction type\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"unsupported transaction type\", payload.Transaction.TypeUrl),\n\t\t)\n\t}\n\n\tidentity = &ivms101.IdentityPayload{}\n\ttransaction = &generic.Transaction{}\n\tif err = payload.Identity.UnmarshalTo(identity); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not unmarshal identity\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\tif err = payload.Transaction.UnmarshalTo(transaction); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not unmarshal transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\ts.updates.Broadcast(req.Id, \"successfully decrypted and parsed secure envelope\", pb.MessageCategory_TRISAP2P)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\t// Update the completed transaction and save to disk\n\txfer.Beneficiary = Identity{\n\t\tWalletAddress: transaction.Beneficiary,\n\t}\n\txfer.Completed = true\n\txfer.Timestamp, _ = time.Parse(time.RFC3339, transaction.Timestamp)\n\n\t// Serialize the identity information as JSON data\n\tvar data []byte\n\tif data, err = json.Marshal(identity); err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, \"could not marshal IVMS 101 identity\"),\n\t\t)\n\t}\n\txfer.Identity = string(data)\n\n\tif err = s.db.Save(&xfer).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\t// Save the pending transaction on the account\n\t// TODO: remove pending transactions\n\taccount.Pending--\n\taccount.Completed++\n\taccount.Balance.Sub(xfer.Amount)\n\tif err = s.db.Save(&account).Error; err != nil {\n\t\tlog.Error().Err(err).Msg(\"could not save transaction\")\n\t\treturn s.updates.SendTransferError(client, req.Id,\n\t\t\tpb.Errorf(pb.ErrInternal, err.Error()),\n\t\t)\n\t}\n\n\tmsg = fmt.Sprintf(\"transaction %04d complete: %s transfered from %s to %s\", xfer.ID, xfer.Amount.String(), xfer.Originator.WalletAddress, xfer.Beneficiary.WalletAddress)\n\ts.updates.Broadcast(req.Id, msg, pb.MessageCategory_BLOCKCHAIN)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\ts.updates.Broadcast(req.Id, fmt.Sprintf(\"%04d new account balance: %s\", account.ID, account.Balance), pb.MessageCategory_LEDGER)\n\ttime.Sleep(time.Duration(rand.Int63n(1000)) * time.Millisecond)\n\n\trep := &pb.Message{\n\t\tType: pb.RPC_TRANSFER,\n\t\tId: req.Id,\n\t\tTimestamp: time.Now().Format(time.RFC3339),\n\t\tCategory: pb.MessageCategory_LEDGER,\n\t\tReply: &pb.Message_Transfer{Transfer: &pb.TransferReply{\n\t\t\tTransaction: xfer.Proto(),\n\t\t}},\n\t}\n\n\treturn s.updates.Send(client, rep)\n}", "func (eobj emp)details(){\n\tfmt.Println(\"Employee details are \",eobj.id,\" \",eobj.name,\" \",eobj.age)\n}", "func (c *TrackerController) Show(ctx *app.ShowTrackerContext) error {\n\treturn application.Transactional(c.db, func(appl application.Application) error {\n\t\tt, err := appl.Trackers().Load(ctx.Context, ctx.ID)\n\t\tif err != nil {\n\t\t\tswitch err.(type) {\n\t\t\tcase remoteworkitem.NotFoundError:\n\t\t\t\tlog.Printf(\"not found, id=%s\", ctx.ID)\n\t\t\t\treturn goa.ErrNotFound(err.Error())\n\t\t\tdefault:\n\t\t\t\treturn err\n\t\t\t}\n\t\t}\n\t\treturn ctx.OK(t)\n\t})\n}", "func (srv *Server) walletTransactionHandler(w http.ResponseWriter, req *http.Request, ps httprouter.Params) {\n\t// Parse the id from the url.\n\tvar id types.TransactionID\n\tjsonID := \"\\\"\" + ps.ByName(\"id\") + \"\\\"\"\n\terr := id.UnmarshalJSON([]byte(jsonID))\n\tif err != nil {\n\t\twriteError(w, \"error after call to /wallet/history: \"+err.Error(), http.StatusBadRequest)\n\t\treturn\n\t}\n\n\ttxn, ok := srv.wallet.Transaction(id)\n\tif !ok {\n\t\twriteError(w, \"error when calling /wallet/transaction/$(id): transaction not found\", http.StatusBadRequest)\n\t\treturn\n\t}\n\twriteJSON(w, WalletTransactionGETid{\n\t\tTransaction: txn,\n\t})\n}", "func GetTransactionHandler(w http.ResponseWriter, r *http.Request) {\n\tvars := mux.Vars(r)\n\t// retrieve the parameters\n\tparam := make(map[string]uint64)\n\tfor _, key := range []string{\"blockId\", \"txId\"} {\n\t\tparam[key], _ = strconv.ParseUint(vars[\"blockId\"], 10, 64)\n\t}\n\n\ttmp := atomic.LoadUint64(&lastBlock)\n\tif param[\"blockId\"] > tmp {\n\t\tw.WriteHeader(http.StatusBadRequest)\n\t\terr := fmt.Errorf(\"requested id %d latest %d\", param[\"blockId\"], lastBlock)\n\t\tlog.Println(err.Error())\n\t\t_, _ = w.Write([]byte(err.Error()))\n\t\treturn\n\t}\n\t// retuning anything in the body regardless of any error code\n\t// it may contain\n\t_, _, body, _ := dataCollection.GetTransaction(param[\"blockId\"], param[\"txId\"], config.DefaultRequestsTimeout)\n\twriteResponse(body, &w)\n}", "func Show(w http.ResponseWriter, r *http.Request) {\n\tc := flight.Context(w, r)\n\n\titem, _, err := summary.ByID(c.DB, c.Param(\"id\"))\n\tif err != nil {\n\t\tc.FlashErrorGeneric(err)\n\t\tc.Redirect(uri)\n\t\treturn\n\t}\n\n\tv := c.View.New(\"summary/show\")\n\tv.Vars[\"item\"] = item\n\tv.Render(w, r)\n}", "func show(req events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {\n\t// Get the `employeeid` query string parameter from the request and\n\t// validate it.\n\temployeeid := req.QueryStringParameters[\"employeeid\"]\n\tif !employeeidRegexp.MatchString(employeeid) {\n\t\treturn clientError(http.StatusBadRequest)\n\t} //fetch a specific employee record from dynamodb in this case\n\n\t// Fetch the employee record from the database based on the employeeid value.\n\temp, err := getItem(employeeid)\n\tif err != nil {\n\t\treturn serverError(err)\n\t}\n\tif emp == nil {\n\t\treturn clientError(http.StatusNotFound)\n\t}\n\n\t// The APIGatewayProxyResponse.Body field needs to be a string, so\n\t// we marshal the employee record into JSON.\n\tjs, err := json.Marshal(emp)\n\tif err != nil {\n\t\treturn serverError(err)\n\t}\n\n\t// Return a response with a 200 OK status and the JSON employee record\n\t// as the body.\n\treturn events.APIGatewayProxyResponse{\n\t\tStatusCode: http.StatusOK,\n\t\tBody: string(js),\n\t}, nil\n}", "func (_VinTracker *VinTrackerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _VinTracker.Contract.VinTrackerTransactor.contract.Transact(opts, method, params...)\n}", "func (db *MySQLDB) GetTransactionByTransactionID(ctx context.Context, transactionID int) (*TransactionRecord, error) {\n\tfLog := mysqlLog.WithField(\"func\", \"GetTransactionByTransactionID\")\n\ttransaction := &TransactionRecord{}\n\n\trow := db.instance.QueryRowContext(ctx, \"SELECT id, user_id, date, grand_total, discount, reason FROM transactions WHERE id = ?\", transactionID)\n\terr := row.Scan(&transaction.ID, &transaction.UserID, &transaction.Date, &transaction.GrandTotal, &transaction.Discount, &transaction.Reason)\n\tif err != nil {\n\t\tfLog.Errorf(\"row.Scan got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\tq := fmt.Sprintf(\"SELECT td.transaction_id, td.product_id, td.qty, td.sub_total, p.sku FROM transaction_detail td INNER JOIN products p ON td.product_id = p.id WHERE transaction_id = %v\", transactionID)\n\trows, err := db.instance.QueryContext(ctx, q)\n\tif err != nil {\n\t\tfLog.Errorf(\"db.instance.QueryContext got %s\", err.Error())\n\t\treturn nil, err\n\t}\n\n\ttDetail := make([]*TransactionDetailRecord, 0)\n\tfor rows.Next() {\n\t\ttD := &TransactionDetailRecord{}\n\t\terr := rows.Scan(&tD.TransactionID, &tD.ProductID, &tD.Qty, &tD.SubTotal, &tD.SKU)\n\t\tif err != nil {\n\t\t\tfLog.Errorf(\"rows.Scan got %s\", err.Error())\n\t\t} else {\n\t\t\ttDetail = append(tDetail, tD)\n\t\t}\n\t}\n\n\ttransaction.TransactionDetail = tDetail\n\n\treturn transaction, nil\n}", "func (_DetailedERC20 *DetailedERC20Raw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.DetailedERC20Transactor.contract.Transact(opts, method, params...)\n}", "func (_DetailedERC20 *DetailedERC20Raw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.DetailedERC20Transactor.contract.Transact(opts, method, params...)\n}", "func TransactionIndex(c *gin.Context) {\n\trelatedObjectID := c.Query(\"relatedObjectId\")\n\trelatedObjectType := c.Query(\"relatedObjectType\")\n\tisSettledQuery := c.Query(\"isSettled\")\n\tstatusQuery := c.Query(\"status\")\n\tcurUserID := c.Keys[\"CurrentUserID\"]\n\n\tvar transactions []models.Transaction\n\n\tquery := database.DBCon\n\n\tisSettled, err := strconv.ParseBool(isSettledQuery)\n\tif isSettledQuery != \"\" && err == nil {\n\t\tquery = query.Where(\"is_settled = ?\", isSettled)\n\t}\n\n\t// TODO: Check that statusQuery is a valid status\n\tif statusQuery != \"\" {\n\t\tquery = query.Where(\"status = ?\", statusQuery)\n\t}\n\n\tif relatedObjectID != \"\" && relatedObjectType != \"\" {\n\t\tquery.\n\t\t\tWhere(\"related_object_id = ? AND related_object_type = ?\", relatedObjectID, relatedObjectType).\n\t\t\tOrder(\"created_at desc\").\n\t\t\tFind(&transactions)\n\t} else {\n\t\tquery.\n\t\t\tWhere(\"creator_id = ?\", curUserID).\n\t\t\tFind(&transactions)\n\t}\n\n\t// Get creator and relatedUser\n\t// TODO: n + 1 query problem here, so we'll figure this out later\n\tfor i := range transactions {\n\t\tdatabase.DBCon.First(&transactions[i].Recipient, transactions[i].RecipientID)\n\t\tdatabase.DBCon.First(&transactions[i].Sender, transactions[i].SenderID)\n\t\tdatabase.DBCon.First(&transactions[i].Creator, transactions[i].CreatorID)\n\t}\n\n\tdata, err := jsonapi.Marshal(transactions)\n\n\tif err != nil {\n\t\tc.AbortWithError(http.StatusInternalServerError, err).\n\t\t\tSetMeta(appError.JSONParseFailure)\n\t\treturn\n\t}\n\n\tc.Data(http.StatusOK, \"application/vnd.api+json\", data)\n}", "func displayTransferDetails(xmlMessage string){\n // Replace all &quot; with single quote\n strings.ReplaceAll(xmlMessage, \"&quot;\", \"'\")\n // Create an parsed XML document\n doc, err := xmlquery.Parse(strings.NewReader(xmlMessage))\n if err != nil {\n panic(err)\n }\n\n // Get required 'transaction' element from Xml message\n transaction := xmlquery.FindOne(doc, \"//transaction\")\n if transaction != nil {\n transferId := transaction.SelectAttr(\"ID\")\n if action := transaction.SelectElement(\"action\"); action != nil {\n if strings.EqualFold(action.InnerText(),\"completed\") {\n // Process transfer complete Xml message\n var supplementMsg string\n status := transaction.SelectElement(\"status\")\n if status != nil {\n supplementMsg = status.SelectElement(\"supplement\").InnerText()\n fmt.Printf(\"\\n[%s] TransferID: %s Status: %s\\n \\tSupplement: %s\\n\",\n action.SelectAttr(\"time\"),\n strings.ToUpper(transferId),\n action.InnerText(),\n supplementMsg)\n }\n\n destAgent := transaction.SelectElement(\"destinationAgent\")\n statistics := transaction.SelectElement(\"statistics\")\n // Retrieve statistics\n var actualStartTimeText = \"\"\n var retryCount string\n var numFileFailures string\n var numFileWarnings string\n if statistics != nil {\n actualStartTime := statistics.SelectElement(\"actualStartTime\")\n if actualStartTime != nil {\n actualStartTimeText = actualStartTime.InnerText()\n }\n if statistics.SelectElement(\"retryCount\") != nil {\n retryCount = statistics.SelectElement(\"retryCount\").InnerText()\n }\n if statistics.SelectElement(\"numFileFailures\") != nil {\n numFileFailures = statistics.SelectElement(\"numFileFailures\").InnerText()\n }\n if statistics.SelectElement(\"numFileWarnings\") != nil {\n numFileWarnings = statistics.SelectElement(\"numFileWarnings\").InnerText()\n }\n }\n var elapsedTime time.Duration\n if actualStartTimeText != \"\" {\n startTime := getFormattedTime(actualStartTimeText)\n completePublishTIme := getFormattedTime(action.SelectAttr(\"time\"))\n elapsedTime = completePublishTIme.Sub(startTime)\n }\n\n fmt.Printf(\"\\tDestination Agent: %s\\n\\tStart time: %s\\n\\tCompletion Time: %s\\n\\tElapsed time: %s\\n\\tRetry Count: %s\\n\\tFailures:%s\\n\\tWarnings:%s\\n\\n\",\n destAgent.SelectAttr(\"agent\"),\n actualStartTimeText,\n action.SelectAttr(\"time\"),\n elapsedTime,\n retryCount,\n numFileFailures,\n numFileWarnings)\n } else if strings.EqualFold(action.InnerText(),\"progress\") {\n // Process transfer progress Xml message\n destAgent := transaction.SelectElement(\"destinationAgent\")\n progressPublishTimeText := action.SelectAttr(\"time\")\n fmt.Printf(\"\\n[%s] %s Status: %s Destination: %s \\n\", progressPublishTimeText,\n strings.ToUpper(transferId),\n action.InnerText(),\n destAgent.SelectAttr(\"agent\"))\n transferSet := transaction.SelectElement(\"transferSet\")\n startTimeText := transferSet.SelectAttr(\"startTime\")\n //startTime := getFormattedTime(startTimeText)\n //progressPublishTime := getFormattedTime(progressPublishTimeText)\n //elapsedTime := progressPublishTime.Sub(startTime)\n fmt.Printf(\"\\tStart time: %s\\n\\tTotal items in transfer request: %s\\n\\tBytes sent: %s\\n\",\n startTimeText,\n transferSet.SelectAttr(\"total\"),\n transferSet.SelectAttr(\"bytesSent\"))\n\n // Loop through all items in the progress message and display details.\n items := transferSet.SelectElements(\"item\")\n for i := 0 ; i < len(items); i++ {\n status := items[i].SelectElement(\"status\")\n resultCode := status.SelectAttr(\"resultCode\")\n var sourceName string\n var sourceSize = \"-1\"\n queueSource := items[i].SelectElement(\"source/queue\")\n if queueSource != nil {\n sourceName = queueSource.InnerText()\n } else {\n fileName := items[i].SelectElement(\"source/file\")\n if fileName != nil {\n sourceName = fileName.InnerText()\n sourceSize = fileName.SelectAttr(\"size\")\n }\n }\n\n var destinationName string\n queueDest := items[i].SelectElement(\"destination/queue\")\n var destinationSize = \"-1\"\n if queueDest != nil {\n destinationName = queueDest.InnerText()\n } else {\n fileName := items[i].SelectElement(\"destination/file\")\n if fileName != nil {\n destinationName = fileName.InnerText()\n destinationSize = fileName.SelectAttr(\"size\")\n }\n }\n\n // Display details of each item\n fmt.Printf(\"\\tItem # %d\\n\\t\\tSource: %s\\tSize: %s bytes\\n\\t\\tDestination: %s\\tSize: %s bytes\\n\",\n i+1,\n sourceName, sourceSize,\n destinationName, destinationSize)\n // Process result code and append any supplement\n if resultCode != \"0\" {\n supplement := status.SelectElement(\"supplement\")\n if supplement != nil {\n fmt.Printf(\"\\t\\tResult code %s Supplement %s\\n\", resultCode, supplement.InnerText())\n } else {\n fmt.Printf(\"\\t\\tResult code %s\\n\", resultCode)\n }\n } else {\n fmt.Printf(\"\\t\\tResult code %s\\n\", resultCode)\n }\n }\n } else if strings.EqualFold(action.InnerText(),\"started\") {\n // Process transfer started Xml message\n destAgent := transaction.SelectElement(\"destinationAgent\")\n destinationAgentName := destAgent.SelectAttr(\"agent\")\n transferSet := transaction.SelectElement(\"transferSet\")\n startTime := \"\"\n if transferSet != nil {\n startTime = transferSet.SelectAttr(\"startTime\")\n } else {\n startTime = action.SelectAttr(\"time\")\n }\n fmt.Printf(\"[%s] TransferID: %s Status: %s Destination: %s\\n\",\n startTime,\n strings.ToUpper(transferId),\n action.InnerText(),\n destinationAgentName)\n }\n }\n }\n}", "func Show(w http.ResponseWriter, r *http.Request) {\r\n\tdb := dbConn()\r\n\tnId := r.URL.Query().Get(\"id\")\r\n\tselDB, err := db.Query(\"SELECT * FROM Pegawai WHERE id=?\", nId)\r\n\tif err != nil {\r\n\t\tpanic(err.Error())\r\n\t}\r\n\temp := Pegawai{}\r\n\tfor selDB.Next() {\r\n\t\t//buat variabel untuk menampung data\r\n\t\t//sesuaikan sama nama kolom database (huruf kecil)\r\n\t\tvar id int\r\n\t\tvar nama, alamat, jabatan string\r\n\r\n\t\terr = selDB.Scan(&id, &nama, &alamat, &jabatan)\r\n\t\tif err != nil {\r\n\t\t\tpanic(err.Error())\r\n\t\t}\r\n\r\n\t\t//kanan nama var struct - kiri nama kolom database yang diinisialisasikan diatas\r\n\t\temp.Id = id\r\n\t\temp.Nama = nama\r\n\t\temp.Alamat = alamat\r\n\t\temp.Jabatan = jabatan\r\n\t}\r\n\ttmpl.ExecuteTemplate(w, \"Show\", emp)\r\n\tdefer db.Close()\r\n}", "func (t *Transaction) String() string {\n\treturn t.From + \" -> \" + t.To + \" : \" + strconv.Itoa(t.Amount)\n}", "func (T transaction) String() string {\n\treturn fmt.Sprintf(\"{\\n\\t\\t\\tsender:%v,\\n\\t\\t\\treceiver:%v,\\n\\t\\t\\tamount:%v\\n\\t\\t}\", T.sender, T.receiver, T.amount)\n}", "func (t *Transaction) String() string {\n\tswitch t.Type {\n\tcase TIN, TEP:\n\t\treturn fmt.Sprintf(\"%22s {%d} %s %.3f %s [%s]\", t.Datetime, t.Index, t.Category.FullName, t.Amount, t.AmountCurrency, t.Type.Name)\n\tcase TOB:\n\t\treturn fmt.Sprintf(\"%22s {%d} Opening new Account\", t.Datetime, t.Index)\n\tcase TBJ:\n\t\treturn fmt.Sprintf(\"%22s {%d} Update balance by %.3f %s\", t.Datetime, t.Index, t.Amount, t.AmountCurrency)\n\tcase TMT:\n\t\treturn fmt.Sprintf(\"%22s {%d} Move %.3f %s to '%s' (%.3f %s)\", t.Datetime, t.Index, t.Amount, t.AmountCurrency, t.AccountTo, t.AmountTo, t.AmountToCurrency)\n\tdefault:\n\t\treturn fmt.Sprintf(\"%22s {%d} [%s] TODO: DIDN'T IMPLEMENT THIS TYPE YET\", t.Datetime, t.Index, t.Type.Name)\n\t}\n}", "func (_CommitteeManager *CommitteeManagerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _CommitteeManager.Contract.CommitteeManagerTransactor.contract.Transact(opts, method, params...)\n}", "func ShowTxStatusTracker(stdout io.Writer, hash string, rpcClient types2.Client) error {\n\ts := spinner.New(spinner.CharSets[14], 100*time.Millisecond)\n\ts.Writer = stdout\n\ts.Prefix = \" \"\n\ts.Start()\n\tlastStatus := \"\"\n\n\tvar err error\n\tvar resp *api2.ResultTx\n\tattempts := 0\n\tfor {\n\t\tattempts += 1\n\t\tif attempts == 3 {\n\t\t\treturn err\n\t\t}\n\n\t\ttime.Sleep(1 * time.Second)\n\t\tresp, err = api.GetTransaction(hash, rpcClient)\n\t\tif err != nil {\n\t\t\ts.Stop()\n\t\t\tcontinue\n\t\t}\n\n\t\tif lastStatus == resp.Status {\n\t\t\tcontinue\n\t\t}\n\n\t\tlastStatus = resp.Status\n\t\tif resp.Status == types3.TxStatusInMempool {\n\t\t\ts.Suffix = colorfmt.YellowStringf(\" In mempool\")\n\t\t} else if resp.Status == types3.TxStatusInPushpool {\n\t\t\ts.Suffix = colorfmt.YellowStringf(\" In pushpool\")\n\t\t} else {\n\t\t\ts.FinalMSG = colorfmt.GreenString(\" Confirmed!\\n\")\n\t\t\ts.Stop()\n\t\t\tbreak\n\t\t}\n\t}\n\treturn nil\n}", "func (res BackTestResult) Show() TradestatPort {\n\t//\tp := NewPortfolio()\n\t// mds := bean.NewRPCMDSConnC(\"tcp\", res.dbhost+\":\"+res.dbport)\n\tratesbook := make(ReferenceRateBook)\n\n\t// FIXME: think about how to show multi pair result\n\tvar stat TradestatPort\n\tif len(res.pairs) > 0 {\n\t\tp := res.pairs[0]\n\t\t//\t\ttxn, _ := mds.GetTransactions2(NameFcoin, p, res.start, res.end)\n\t\tratesbook[p] = RefRatesFromTxn(res.Txn)\n\t\t//\t\tsnapts.Print()\n\t\t//\t\tperfts.Print()\n\n\t\tstat = *Tradestat(p.Base, res.Txn, NewPortfolio(), ratesbook)\n\t\tssTS := GenerateSnapshotTS(res.Txn, NewPortfolio())\n\t\tmaxPos := 0.0\n\t\tmaxNeg := 0.0\n\t\tfor _, ss := range ssTS {\n\t\t\tmaxPos = math.Max(maxPos, ss.Port.Balance(p.Coin))\n\t\t\tmaxNeg = math.Min(maxNeg, ss.Port.Balance(p.Coin))\n\t\t}\n\t\tfmt.Println(\"max position\", maxPos, maxNeg)\n\t\tstat.Print()\n\t}\n\treturn stat\n}", "func (ok OKHandler) DeliverTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.DeliverResult, err error) {\n\treturn sdk.DeliverResult{Log: ok.Log}, nil\n}", "func Index(txn *cheshire.Txn) {\n\t//create a context map to be passed to the template\n\tcontext := make(map[string]interface{})\n\tcontext[\"services\"] = Servs.RouterTables()\n\tcheshire.RenderInLayout(txn, \"/index.html\", \"/template.html\", context)\n}", "func (c *dummyWavesMDLrpcclient) GetTransaction(txid string) (*model.Transactions, error) {\n\ttransaction, _, err := client.NewTransactionsService(c.MainNET).GetTransactionsInfoID(txid)\n\treturn transaction, err\n}", "func (_VinTracker *VinTrackerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _VinTracker.Contract.contract.Transact(opts, method, params...)\n}", "func (action *LedgerShowAction) JSON() {\n\tquery := action.Query()\n\n\tif action.Err != nil {\n\t\treturn\n\t}\n\n\taction.Err = db.Get(action.Ctx, query, &action.Record)\n\n\tif action.Err != nil {\n\t\treturn\n\t}\n\n\thal.Render(action.W, NewLedgerResource(action.Record))\n}", "func (_CommitteeManager *CommitteeManagerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _CommitteeManager.Contract.contract.Transact(opts, method, params...)\n}", "func (t *Transaction) String() string {\n\tvar builder strings.Builder\n\tbuilder.WriteString(\"Transaction(\")\n\tbuilder.WriteString(fmt.Sprintf(\"id=%v\", t.ID))\n\tbuilder.WriteString(\", sequenceInDay=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.SequenceInDay))\n\tbuilder.WriteString(\", status=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Status))\n\tbuilder.WriteString(\", executedAmount=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.ExecutedAmount))\n\tbuilder.WriteString(\", executedCurrencyCode=\")\n\tbuilder.WriteString(t.ExecutedCurrencyCode)\n\tbuilder.WriteString(\", exchangeRate=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.ExchangeRate))\n\tbuilder.WriteString(\", originatingAmount=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.OriginatingAmount))\n\tbuilder.WriteString(\", originatingCurrencyCode=\")\n\tbuilder.WriteString(t.OriginatingCurrencyCode)\n\tbuilder.WriteString(\", direction=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Direction))\n\tbuilder.WriteString(\", runningBalance=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.RunningBalance))\n\tbuilder.WriteString(\", createdDate=\")\n\tbuilder.WriteString(t.CreatedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", postedDate=\")\n\tbuilder.WriteString(t.PostedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", executedDate=\")\n\tbuilder.WriteString(t.ExecutedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", updatedDate=\")\n\tbuilder.WriteString(t.UpdatedDate.Format(time.ANSIC))\n\tbuilder.WriteString(\", description=\")\n\tbuilder.WriteString(t.Description)\n\tbuilder.WriteString(\", memo=\")\n\tbuilder.WriteString(t.Memo)\n\tbuilder.WriteString(\", group=\")\n\tbuilder.WriteString(t.Group)\n\tbuilder.WriteString(\", type=\")\n\tbuilder.WriteString(t.Type)\n\tbuilder.WriteString(\", mainCategory=\")\n\tbuilder.WriteString(t.MainCategory)\n\tbuilder.WriteString(\", subCategory=\")\n\tbuilder.WriteString(t.SubCategory)\n\tbuilder.WriteString(\", checkNumber=\")\n\tbuilder.WriteString(t.CheckNumber)\n\tbuilder.WriteString(\", latitude=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Latitude))\n\tbuilder.WriteString(\", longitude=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Longitude))\n\tbuilder.WriteString(\", merchantCode=\")\n\tbuilder.WriteString(t.MerchantCode)\n\tbuilder.WriteString(\", reversal=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Reversal))\n\tbuilder.WriteString(\", reversalFor=\")\n\tbuilder.WriteString(t.ReversalFor)\n\tbuilder.WriteString(\", reversed=\")\n\tbuilder.WriteString(fmt.Sprintf(\"%v\", t.Reversed))\n\tbuilder.WriteString(\", reversedBy=\")\n\tbuilder.WriteString(t.ReversedBy)\n\tbuilder.WriteString(\", url=\")\n\tbuilder.WriteString(t.URL)\n\tbuilder.WriteByte(')')\n\treturn builder.String()\n}", "func (controller *AccountController) DoTransaction(ctx *gin.Context) {\n\tsourceID, ok := ctx.GetPostForm(\"sourceid\")\n\tif !ok {\n\t\tlog.WithFields(log.Fields{\"URL\": ctx.Request.URL.String()}).Warn(\"No SourceID found in postform\")\n\n\t\terrResp, _ := restapi.NewErrorResponse(\"No sourceID given\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(errResp))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\ttargetID, ok := ctx.GetPostForm(\"targetid\")\n\tif !ok {\n\t\tlog.WithFields(log.Fields{\"URL\": ctx.Request.URL.String()}).Warn(\"No TargetID found in postform\")\n\n\t\terrResp, _ := restapi.NewErrorResponse(\"No targetID given\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(errResp))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tamount, err := strconv.Atoi(ctx.PostForm(\"amount\"))\n\tif err != nil {\n\t\tlog.WithFields(log.Fields{\"URL\": ctx.Request.URL.String()}).Warn(\"No int amount found in postform\")\n\n\t\tresponse, _ := restapi.NewErrorResponse(\"No valid diff value\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tinfo, err := authStuff.GetLoginInfoFromCtx(ctx)\n\tif err != nil {\n\t\tresponse, _ := restapi.NewErrorResponse(err.Error()).Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n\n\tif err := controller.service.Transaction(sourceID, targetID, info.Name, amount); err == nil {\n\t\tresponse, _ := restapi.NewOkResponse(\"\").Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Next()\n\t} else {\n\t\tlog.WithFields(log.Fields{\"user\": info.Name}).WithError(err).Error(\"Transaction Error\")\n\n\t\tresponse, _ := restapi.NewErrorResponse(err.Error()).Marshal()\n\t\tfmt.Fprint(ctx.Writer, string(response))\n\t\tctx.Abort()\n\t\treturn\n\t}\n}", "func (_TableManager *TableManagerRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _TableManager.Contract.TableManagerTransactor.contract.Transact(opts, method, params...)\n}", "func (_TableManager *TableManagerTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, *types.Receipt, error) {\n\treturn _TableManager.Contract.contract.Transact(opts, method, params...)\n}", "func (_ArbGasInfo *ArbGasInfoRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _ArbGasInfo.Contract.ArbGasInfoTransactor.contract.Transact(opts, method, params...)\n}", "func (_DetailedERC20 *DetailedERC20TransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.contract.Transact(opts, method, params...)\n}", "func (_DetailedERC20 *DetailedERC20TransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _DetailedERC20.Contract.contract.Transact(opts, method, params...)\n}", "func TransactionScreen(w fyne.Window) fyne.CanvasObject {\n\treturn widget.NewTabContainer(\n\t\twidget.NewTabItem(res.GetLocalString(\"Transfer\"), makeTransferTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Move\"), makeMoveTransTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Vote\"), makeVoteTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Unvote\"), makeUnvoteTab(w)),\n\t\twidget.NewTabItem(res.GetLocalString(\"Register Miner\"), makeMinerTab(w)),\n\t)\n}", "func ShowTask(w http.ResponseWriter, r *http.Request, repo *tasks.TaskRepository) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\tparams := mux.Vars(r)\n\ttaskID, err := strconv.Atoi(params[\"id\"])\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\ttask, err := repo.GetTask(taskID)\n\tif err != nil {\n\t\tpanic(err)\n\t}\n\tjson.NewEncoder(w).Encode(apiIndexTask(task))\n}", "func (api *DeprecatedApiService) getTransactionIntro(params map[string]string) map[string]string {\n\tresult := make(map[string]string)\n\n\t_, is_only_check_exist := params[\"only_check_exist\"]\n\n\ttrsid, ok1 := params[\"id\"]\n\tif !ok1 {\n\t\tresult[\"err\"] = \"param id must.\"\n\t\treturn result\n\t}\n\n\tvar trshx []byte\n\tif txhx, e := hex.DecodeString(trsid); e == nil && len(txhx) == 32 {\n\t\ttrshx = txhx\n\t} else {\n\t\tresult[\"err\"] = \"transaction hash error.\"\n\t\treturn result\n\t}\n\n\t// Query transaction\n\tblkhei, trsresbytes, err := api.blockchain.GetChainEngineKernel().StateRead().ReadTransactionBytesByHash(trshx)\n\tif err != nil {\n\t\tresult[\"err\"] = err.Error()\n\t\treturn result\n\t}\n\tif trsresbytes == nil {\n\t\tresult[\"err\"] = \"transaction not fond.\"\n\t\treturn result\n\t}\n\n\t// Whether to just judge whether it exists\n\tif is_only_check_exist && len(trsresbytes) > 0 {\n\t\tresult[\"ret\"] = \"0\"\n\t\tresult[\"exist\"] = \"yes\"\n\t\treturn result\n\t}\n\n\ttrsres, _, err := transactions.ParseTransaction(trsresbytes, 0)\n\tif err != nil {\n\t\tresult[\"err\"] = err.Error()\n\t\treturn result\n\t}\n\n\t// Resolve actions\n\tvar allactions = trsres.GetActionList()\n\tvar actions_ary []string\n\tvar actions_strings = \"\"\n\tfor _, act := range allactions {\n\t\tvar kind = act.Kind()\n\t\tactstr := fmt.Sprintf(`{\"k\":%d`, kind)\n\t\tif kind == 1 {\n\t\t\tacc := act.(*actions.Action_1_SimpleToTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"to\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 13 {\n\t\t\tacc := act.(*actions.Action_13_FromTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 14 {\n\t\t\tacc := act.(*actions.Action_14_FromToTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"to\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 2 {\n\t\t\tacc := act.(*actions.Action_2_OpenPaymentChannel)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"left_addr\":\"%s\",\"left_amt\":\"%s\",\"right_addr\":\"%s\",\"right_amt\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t\tacc.LeftAddress.ToReadable(),\n\t\t\t\tacc.LeftAmount.ToFinString(),\n\t\t\t\tacc.RightAddress.ToReadable(),\n\t\t\t\tacc.RightAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 3 {\n\t\t\tacc := act.(*actions.Action_3_ClosePaymentChannel)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 4 {\n\t\t\tacc := act.(*actions.Action_4_DiamondCreate)\n\t\t\tactstr += fmt.Sprintf(`,\"number\":\"%d\",\"name\":\"%s\",\"address\":\"%s\"`,\n\t\t\t\tacc.Number,\n\t\t\t\tacc.Diamond,\n\t\t\t\tacc.Address.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 5 {\n\t\t\tacc := act.(*actions.Action_5_DiamondTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"count\":1,\"names\":\"%s\",\"from\":\"%s\",\"to\":\"%s\"`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Diamond,\n\t\t\t\ttrsres.GetAddress().ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 6 {\n\t\t\tacc := act.(*actions.Action_6_OutfeeQuantityDiamondTransfer)\n\t\t\tdmds := make([]string, len(acc.DiamondList.Diamonds))\n\t\t\tfor i, v := range acc.DiamondList.Diamonds {\n\t\t\t\tdmds[i] = string(v)\n\t\t\t}\n\t\t\tactstr += fmt.Sprintf(`,\"count\":%d,\"names\":\"%s\",\"from\":\"%s\",\"to\":\"%s\"`,\n\t\t\t\tacc.DiamondList.Count,\n\t\t\t\tstrings.Join(dmds, \",\"),\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 7 {\n\t\t\tacc := act.(*actions.Action_7_SatoshiGenesis)\n\t\t\tactstr += fmt.Sprintf(`,\"trs_no\":%d,\"btc_num\":%d,\"hac_subsidy\":%d,\"address\":\"%s\",\"lockbls_id\":\"%s\"`,\n\t\t\t\tacc.TransferNo,\n\t\t\t\tacc.BitcoinQuantity,\n\t\t\t\tacc.AdditionalTotalHacAmount,\n\t\t\t\tacc.OriginAddress.ToReadable(),\n\t\t\t\thex.EncodeToString(actions.GainLockblsIdByBtcMove(uint32(acc.TransferNo))),\n\t\t\t)\n\t\t} else if kind == 8 {\n\t\t\tacc := act.(*actions.Action_8_SimpleSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"to\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 9 {\n\t\t\tacc := act.(*actions.Action_9_LockblsCreate)\n\t\t\tactstr += fmt.Sprintf(`,\"lockbls_id\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.LockblsId),\n\t\t\t\tacc.TotalStockAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 10 {\n\t\t\tacc := act.(*actions.Action_10_LockblsRelease)\n\t\t\tactstr += fmt.Sprintf(`,\"lockbls_id\":\"%s\",\"amount\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.LockblsId),\n\t\t\t\tacc.ReleaseAmount.ToFinString(),\n\t\t\t)\n\t\t} else if kind == 11 {\n\t\t\tacc := act.(*actions.Action_11_FromToSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"to\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.ToAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 28 {\n\t\t\tacc := act.(*actions.Action_28_FromSatoshiTransfer)\n\t\t\tactstr += fmt.Sprintf(`,\"from\":\"%s\",\"amount\":%d`,\n\t\t\t\tacc.FromAddress.ToReadable(),\n\t\t\t\tacc.Amount,\n\t\t\t)\n\t\t} else if kind == 12 {\n\t\t\tacc := act.(*actions.Action_12_ClosePaymentChannelBySetupAmount)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 21 {\n\t\t\tacc := act.(*actions.Action_21_ClosePaymentChannelBySetupOnlyLeftAmount)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t} else if kind == 22 {\n\t\t\tacc := act.(*actions.Action_22_UnilateralClosePaymentChannelByNothing)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\"\":\"%s\",\"bill_number\"\":0`,\n\t\t\t\thex.EncodeToString(acc.ChannelId), acc.AssertCloseAddress.ToReadable(),\n\t\t\t)\n\t\t} else if kind == 23 {\n\t\t\tacc := act.(*actions.Action_23_UnilateralCloseOrRespondChallengePaymentChannelByRealtimeReconciliation)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"%s\",\"bill_number\":%d`,\n\t\t\t\thex.EncodeToString(acc.Reconciliation.GetChannelId()), acc.AssertAddress.ToReadable(), acc.Reconciliation.GetAutoNumber(),\n\t\t\t)\n\t\t} else if kind == 24 {\n\t\t\tacc := act.(*actions.Action_24_UnilateralCloseOrRespondChallengePaymentChannelByChannelChainTransferBody)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"%s\",\"bill_number\":%d`,\n\t\t\t\thex.EncodeToString(acc.ChannelChainTransferTargetProveBody.GetChannelId()), acc.AssertAddress.ToReadable(), acc.ChannelChainTransferTargetProveBody.GetAutoNumber(),\n\t\t\t)\n\t\t} else if kind == 27 {\n\t\t\tacc := act.(*actions.Action_27_ClosePaymentChannelByClaimDistribution)\n\t\t\tactstr += fmt.Sprintf(`,\"channel_id\":\"%s\",\"assert_address\":\"any\",\"bill_number\"\":\"closed\"`,\n\t\t\t\thex.EncodeToString(acc.ChannelId),\n\t\t\t)\n\t\t}\n\t\tactstr += \"}\"\n\t\tactions_ary = append(actions_ary, actstr)\n\t}\n\tactions_strings = strings.Join(actions_ary, \",\")\n\n\t// Transaction return data\n\ttxaddr := fields.Address(trsres.GetAddress())\n\tvar txfee = trsres.GetFee()\n\tvar txfeeminergot = trsres.GetFeeOfMinerRealReceived()\n\tresult[\"jsondata\"] = fmt.Sprintf(\n\t\t`{\"block\":{\"height\":%d,\"timestamp\":%d},\"type\":%d,\"address\":\"%s\",\"fee\":\"%s\",\"feeminergot\":\"%s\",\"timestamp\":%d,\"actioncount\":%d,\"actions\":[%s]`,\n\t\tblkhei,\n\t\ttrsres.GetTimestamp(),\n\t\ttrsres.Type(),\n\t\ttxaddr.ToReadable(), // Primary address\n\t\ttxfee.ToFinString(),\n\t\ttxfeeminergot.ToFinString(),\n\t\ttrsres.GetTimestamp(),\n\t\tlen(allactions),\n\t\tactions_strings,\n\t)\n\n\tif _, ok := params[\"txbodyhex\"]; ok {\n\t\tresult[\"jsondata\"] += fmt.Sprintf(`,\"txbodyhex\":\"%s\"`,\n\t\t\thex.EncodeToString(trsresbytes))\n\t}\n\n\t// Wrap up and return\n\tresult[\"jsondata\"] += \"}\"\n\treturn result\n}", "func TransactionHandler(w http.ResponseWriter, r *http.Request) {\n\taction := r.URL.Path[len(\"/api/transactions\"):]\n\n\tlog.Println(\"Handling method\", r.Method, \"with action\", action)\n\n\tswitch r.Method {\n\tcase \"POST\":\n\t\tswitch action {\n\t\tcase \"\": // Create new transaction\n\t\t\tbody, err := ioutil.ReadAll(r.Body)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorForm, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tvar t TransactionResponse\n\n\t\t\terr = json.Unmarshal(body, &t)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tpass := []byte(t.Password)\n\n\t\t\ttx, e := createAndBroadcastTx(t.Recipient, *big.NewInt(t.Amount), pass)\n\n\t\t\tvar jsonResponse TransactionResponse\n\t\t\tif e == nil {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseOk, Hash: hex.EncodeToString(tx.Hash())}\n\t\t\t} else {\n\t\t\t\tjsonResponse = TransactionResponse{Amount: t.Amount, Recipient: t.Recipient, Status: ResponseFailed, ErrorText: e.Error()}\n\t\t\t}\n\n\t\t\tres, err := json.Marshal(jsonResponse)\n\t\t\tif err != nil {\n\t\t\t\tcreateJsonErrorResponse(w, r, http.StatusInternalServerError, ErrorJson, err.Error())\n\t\t\t\treturn\n\t\t\t}\n\t\t\tfmt.Fprintf(w, string(res))\n\t\tdefault:\n\t\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\t\t}\n\tcase \"GET\":\n\t\tswitch action {\n\t\tcase \"\":\n\t\t\tvar txs []TransactionJson\n\t\t\tfor _, tx := range Config.DeserializedTxs {\n\t\t\t\ttxs = append(txs, EncodeToFriendlyStruct(tx))\n\n\t\t\t}\n\n\t\t\tif len(txs) == 0 {\n\t\t\t\tfmt.Fprintf(w, string(\"[]\"))\n\t\t\t} else {\n\n\t\t\t\tres, err := json.Marshal(txs)\n\t\t\t\tif err != nil {\n\t\t\t\t\tfmt.Println(\"Nope\", err.Error())\n\t\t\t\t} else {\n\t\t\t\t\tfmt.Fprintf(w, string(res))\n\t\t\t\t}\n\t\t\t}\n\n\t\t}\n\tdefault:\n\t\tcreateJsonErrorResponse(w, r, http.StatusNotFound, Error404, fmt.Sprint(\"No action: \", r.Method, action))\n\n\t}\n}", "func (action *TransactionIndexAction) JSON() {\n\taction.Do(\n\t\taction.EnsureHistoryFreshness,\n\t\taction.loadParams,\n\t\taction.checkAllowed,\n\t\taction.ValidateCursorWithinHistory,\n\t\taction.loadRecords,\n\t\taction.loadPage,\n\t\tfunc() {\n\t\t\thal.Render(action.W, action.Page)\n\t\t},\n\t)\n}", "func (EchoHandler) DeliverTx(ctx sdk.Context, store state.SimpleDB, tx sdk.Tx) (res sdk.DeliverResult, err error) {\n\tdata, err := data.ToWire(tx)\n\treturn sdk.DeliverResult{Data: data}, err\n}", "func (s *Service) HandleShow(w http.ResponseWriter, r *http.Request) {\n\tsubs, err := s.subscriptionRepository.FindOne(\n\t\tr.Context(), s.getResourceID(r), s.getSubscriptionID(r),\n\t)\n\tif err != nil {\n\t\tstatus := http.StatusInternalServerError\n\t\tif errRepo, ok := err.(flare.SubscriptionRepositoryError); ok && errRepo.NotFound() {\n\t\t\tstatus = http.StatusNotFound\n\t\t}\n\n\t\ts.writer.Error(w, \"error during subscription search\", err, status)\n\t\treturn\n\t}\n\n\ts.writer.Response(w, transformSubscription(subs), http.StatusOK, nil)\n}", "func (me TSAFPTTransactionID) String() string { return xsdt.String(me).String() }", "func (u *UseCase) ShowOrder(w http.ResponseWriter, r *http.Request) {\n\n\tmerchant := r.Header.Get(\"merchant_id\")\n\tuuid := mux.Vars(r)[\"order_id\"]\n\n\torder, err := cache.ShowOrder(merchant, uuid)\n\tif err == nil && order != nil {\n\t\trespondWithJSON(w, http.StatusOK, order)\n\t\treturn\n\t}\n\n\tvar dbOrders models.OrderPg\n\tu.DB.Conn.Table(\"orders\").Where(\"uuid = ?\", uuid).First(&dbOrders)\n\n\trespondWithJSON(w, http.StatusOK, dbOrders.Payload)\n}", "func tableInfoShow(meta *halproto.TableMetadata) {\n\tswitch meta.GetKind() {\n\tcase halproto.TableKind_TABLE_INDEX:\n\t\ttableInfoIndexShow(meta)\n\tcase halproto.TableKind_TABLE_TCAM:\n\t\ttableInfoTcamShow(meta)\n\tcase halproto.TableKind_TABLE_HASH:\n\t\ttableInfoHashShow(meta)\n\tcase halproto.TableKind_TABLE_FLOW:\n\t\ttableInfoFlowShow(meta)\n\tcase halproto.TableKind_TABLE_MET:\n\t\ttableInfoMetShow(meta)\n\tdefault:\n\t\tfmt.Printf(\"Invalid table type: %d\", meta.GetKind())\n\t}\n}", "func (r Virtual_Guest) GetActiveTransaction() (resp datatypes.Provisioning_Version1_Transaction, err error) {\n\terr = r.Session.DoRequest(\"SoftLayer_Virtual_Guest\", \"getActiveTransaction\", nil, &r.Options, &resp)\n\treturn\n}", "func (t *Trade) View(c echo.Context, id int) (*gorsk.Trade, error) {\n\treturn t.tdb.View(t.db, id)\n}", "func GetTransaction() gin.HandlerFunc {\n\treturn func(c *gin.Context) {\n\t\tsugar, _ := item.New(\"Sugar\", map[string]float64{\"Kabras\": 110, \"Mumias\": 110}, \"kg(s)\")\n\t\tpurchase, message, err := transaction.New(sugar, map[string]float64{\"Nzoia\": 150}, 3)\n\t\tc.JSON(\n\t\t\thttp.StatusOK,\n\t\t\tGetResponse{GetData{purchase}, message, responseerr.GetStrErr(err)},\n\t\t)\n\t}\n}", "func (as *AddrServer) HandleTxGet(w http.ResponseWriter, r *http.Request) {\n\tw.Header().Set(\"Content-Type\", \"application/json\")\n\ttxid := mux.Vars(r)[\"txid\"]\n\n\t// paginate through transactions\n\ttxns, err := as.GetRawTransaction(txid)\n\tif err != nil {\n\t\tw.WriteHeader(400)\n\t\tw.Write(NewPostError(\"error fetching all transactions for address\", err))\n\t\treturn\n\t}\n\tout, _ := json.Marshal(txns.Result)\n\tw.Write(out)\n}", "func (t *trip) Show(ctx context.Context, id int32) (*model.Trip, error) {\n\trow := t.conn.QueryRow(ctx, `SELECT \n\t\t\t\ttrips.id, trips.dates, trips.price, origin.name, destination.name\n\t\t\t\tFROM trips \n\t\t\t\tINNER JOIN cities AS origin ON trips.origin_id = origin.id\n\t\t\t\tINNER JOIN cities AS destination ON trips.destination_id = destination.id\n\t\t\t\tWHERE trips.id = $1`, id)\n\n\tvar trip model.Trip\n\n\tvar origin, destination model.City\n\n\tif err := row.Scan(&trip.ID, &trip.Dates, &trip.Price, &origin.Name, &destination.Name); err != nil {\n\t\treturn nil, err\n\t}\n\n\ttrip.Origin = &origin\n\ttrip.Destination = &destination\n\n\treturn &trip, nil\n}", "func (core *coreService) TraceTransaction(ctx context.Context, actHash string, config *logger.Config) ([]byte, *action.Receipt, *logger.StructLogger, error) {\n\tactInfo, err := core.Action(util.Remove0xPrefix(actHash), false)\n\tif err != nil {\n\t\treturn nil, nil, nil, err\n\t}\n\tact, err := (&action.Deserializer{}).SetEvmNetworkID(core.EVMNetworkID()).ActionToSealedEnvelope(actInfo.Action)\n\tif err != nil {\n\t\treturn nil, nil, nil, err\n\t}\n\tsc, ok := act.Action().(*action.Execution)\n\tif !ok {\n\t\treturn nil, nil, nil, errors.New(\"the type of action is not supported\")\n\t}\n\ttraces := logger.NewStructLogger(config)\n\tctx = protocol.WithVMConfigCtx(ctx, vm.Config{\n\t\tDebug: true,\n\t\tTracer: traces,\n\t\tNoBaseFee: true,\n\t})\n\taddr, _ := address.FromString(address.ZeroAddress)\n\tretval, receipt, err := core.SimulateExecution(ctx, addr, sc)\n\treturn retval, receipt, traces, err\n}", "func (c *BulkDeletesController) Show(ctx *gin.Context) {\n\tid := ctx.Param(\"taskID\")\n\ttask := models.BulkDeleteRunTask{}\n\n\tif err := c.App.GetStore().One(\"ID\", id, &task); err == storm.ErrNotFound {\n\t\tctx.AbortWithError(404, errors.New(\"Bulk delete task not found\"))\n\t} else if err != nil {\n\t\tctx.AbortWithError(500, err)\n\t} else if doc, err := jsonapi.Marshal(&task); err != nil {\n\t\tctx.AbortWithError(500, err)\n\t} else {\n\t\tctx.Data(200, MediaType, doc)\n\t}\n}", "func (gw *Gateway) GetTransaction(txid cipher.SHA256) (*visor.Transaction, error) {\n\tvar txn *visor.Transaction\n\tvar err error\n\n\tgw.strand(\"GetTransaction\", func() {\n\t\ttxn, err = gw.v.GetTransaction(txid)\n\t})\n\n\treturn txn, err\n}", "func (_ArbGasInfo *ArbGasInfoTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _ArbGasInfo.Contract.contract.Transact(opts, method, params...)\n}", "func (t *Transaction) DisplayInline() string {\n\treturn fmt.Sprintf(\"%s - - Operation: %s - Status: %s - Side: %s - Amount(cts): %d\", t.ID, t.OperationType, t.Status, t.Side, t.AmountCents)\n}", "func tableInfoHashShow(meta *halproto.TableMetadata) {\n\thMeta := meta.GetHashMeta()\n\tinsStr := fmt.Sprintf(\"%d[%d]\", hMeta.GetNumInserts(), hMeta.GetNumInsertFailures())\n\tupdStr := fmt.Sprintf(\"%d[%d]\", hMeta.GetNumUpdates(), hMeta.GetNumUpdateFailures())\n\tdelStr := fmt.Sprintf(\"%d[%d]\", hMeta.GetNumDeletes(), hMeta.GetNumDeleteFailures())\n\tfmt.Printf(\"%-30s%-10d%-10s%-10d%-10d%-10d%-10d%-10s%-10s%-10s\\n\",\n\t\tmeta.GetTableName(),\n\t\tmeta.GetTableId(),\n\t\tutils.TableKindToStr(meta.GetKind()),\n\t\thMeta.GetCapacity(),\n\t\thMeta.GetOtcamCapacity(),\n\t\thMeta.GetHashUsage(),\n\t\thMeta.GetOtcamUsage(),\n\t\tinsStr,\n\t\tupdStr,\n\t\tdelStr)\n}", "func Details(attrs []htmlgo.Attribute, children ...HTML) HTML {\n\treturn &htmlgo.Tree{Tag: \"details\", Attributes: attrs, Children: children}\n}", "func (db *DB) View(fn func(*Tx) error) error {\n\tt, err := db.Begin(false)\n\tif err != nil {\n\t\treturn err\n\t}\n\n\t// Make sure the transaction rolls back in the event of a panic.\n\tdefer func() {\n\t\tif t.db != nil {\n\t\t\tt.rollback()\n\t\t}\n\t}()\n\n\t// Mark as a managed tx so that the inner function cannot manually rollback.\n\tt.managed = true\n\n\t// If an error is returned from the function then pass it through.\n\terr = fn(t)\n\tt.managed = false\n\tif err != nil {\n\t\t_ = t.Rollback()\n\t\treturn err\n\t}\n\n\tif err := t.Rollback(); err != nil {\n\t\treturn err\n\t}\n\n\treturn nil\n}", "func (HTTPOperation) GetDetails() (string, string, string) {\n\treturn \"update\", \"updated\", \"vaccine availability \" + id\n}", "func (tc *TransactionsController) Index(c *gin.Context, size, page, offset int) {\n\ttxs, count, err := tc.App.GetStore().Transactions(offset, size)\n\tptxs := make([]presenters.Tx, len(txs))\n\tfor i, tx := range txs {\n\t\tptxs[i] = presenters.NewTx(&tx)\n\t}\n\tpaginatedResponse(c, \"Transactions\", size, page, ptxs, count, err)\n}", "func (_Onesplitaudit *OnesplitauditTransactorRaw) Transact(opts *bind.TransactOpts, method string, params ...interface{}) (*types.Transaction, error) {\n\treturn _Onesplitaudit.Contract.contract.Transact(opts, method, params...)\n}", "func tableInfoIndexShow(meta *halproto.TableMetadata) {\n\tiMeta := meta.GetIndexMeta()\n\tinsStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumInserts(), iMeta.GetNumInsertFailures())\n\tupdStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumUpdates(), iMeta.GetNumUpdateFailures())\n\tdelStr := fmt.Sprintf(\"%d[%d]\", iMeta.GetNumDeletes(), iMeta.GetNumDeleteFailures())\n\tfmt.Printf(\"%-30s%-10d%-10s%-10d%-10s%-10d%-10s%-10s%-10s%-10s\\n\",\n\t\tmeta.GetTableName(),\n\t\tmeta.GetTableId(),\n\t\tutils.TableKindToStr(meta.GetKind()),\n\t\tiMeta.GetCapacity(),\n\t\t\"-\",\n\t\tiMeta.GetUsage(),\n\t\t\"-\",\n\t\tinsStr,\n\t\tupdStr,\n\t\tdelStr)\n}", "func (resolver *ResolverTODO) Show(params graphql.ResolveParams) (interface{}, error) {\n\ttodo, err := resolver.Db.Show(params.Args[\"id\"].(string))\n\n\tif err != nil {\n\t\treturn nil, err\n\t} else if todo.ID == \"\" {\n\t\treturn nil, errors.New(\"todo not found\")\n\t}\n\n\treturn todo, nil\n}" ]
[ "0.7065399", "0.6323545", "0.6089757", "0.58380103", "0.5829757", "0.5806046", "0.5743745", "0.5695702", "0.5692406", "0.5600741", "0.55515116", "0.55509126", "0.55220157", "0.54999244", "0.53384566", "0.53074944", "0.52982825", "0.52785516", "0.5276625", "0.52471644", "0.5238717", "0.52065337", "0.51978594", "0.51897407", "0.51788306", "0.5154507", "0.5149336", "0.5148842", "0.5148842", "0.5124913", "0.5117847", "0.5116096", "0.50937843", "0.5069765", "0.5049436", "0.50435567", "0.5006893", "0.49923867", "0.4981022", "0.49754423", "0.4921016", "0.4898289", "0.4889206", "0.48864287", "0.48829022", "0.48819882", "0.486109", "0.48377797", "0.48292157", "0.48292157", "0.4825918", "0.48157176", "0.48137054", "0.48055518", "0.48034588", "0.480199", "0.47995147", "0.47936705", "0.47911152", "0.47875905", "0.47867686", "0.47867665", "0.47850013", "0.4778883", "0.47752193", "0.47638208", "0.47636342", "0.47625333", "0.47538182", "0.4751926", "0.47489476", "0.47489476", "0.47414246", "0.47372147", "0.4732912", "0.47310746", "0.47151774", "0.4705625", "0.47027025", "0.46943593", "0.46935827", "0.4692121", "0.4691574", "0.46858248", "0.4676262", "0.46733868", "0.46709293", "0.46631366", "0.46594134", "0.46525756", "0.46511626", "0.46491662", "0.46372992", "0.46351305", "0.4628336", "0.46106213", "0.46100897", "0.46069738", "0.46058023", "0.4605689" ]
0.59414774
3