{ // 获取包含Hugging Face文本的span元素 const spans = link.querySelectorAll('span.whitespace-nowrap, span.hidden.whitespace-nowrap'); spans.forEach(span => { if (span.textContent && span.textContent.trim().match(/Hugging\s*Face/i)) { span.textContent = 'AI快站'; } }); }); // 替换logo图片的alt属性 document.querySelectorAll('img[alt*="Hugging"], img[alt*="Face"]').forEach(img => { if (img.alt.match(/Hugging\s*Face/i)) { img.alt = 'AI快站 logo'; } }); } // 替换导航栏中的链接 function replaceNavigationLinks() { // 已替换标记,防止重复运行 if (window._navLinksReplaced) { return; } // 已经替换过的链接集合,防止重复替换 const replacedLinks = new Set(); // 只在导航栏区域查找和替换链接 const headerArea = document.querySelector('header') || document.querySelector('nav'); if (!headerArea) { return; } // 在导航区域内查找链接 const navLinks = headerArea.querySelectorAll('a'); navLinks.forEach(link => { // 如果已经替换过,跳过 if (replacedLinks.has(link)) return; const linkText = link.textContent.trim(); const linkHref = link.getAttribute('href') || ''; // 替换Spaces链接 - 仅替换一次 if ( (linkHref.includes('/spaces') || linkHref === '/spaces' || linkText === 'Spaces' || linkText.match(/^s*Spacess*$/i)) && linkText !== 'PDF TO Markdown' && linkText !== 'PDF TO Markdown' ) { link.textContent = 'PDF TO Markdown'; link.href = 'https://fast360.xyz'; link.setAttribute('target', '_blank'); link.setAttribute('rel', 'noopener noreferrer'); replacedLinks.add(link); } // 删除Posts链接 else if ( (linkHref.includes('/posts') || linkHref === '/posts' || linkText === 'Posts' || linkText.match(/^s*Postss*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } // 替换Docs链接 - 仅替换一次 else if ( (linkHref.includes('/docs') || linkHref === '/docs' || linkText === 'Docs' || linkText.match(/^s*Docss*$/i)) && linkText !== 'Voice Cloning' ) { link.textContent = 'Voice Cloning'; link.href = 'https://vibevoice.info/'; replacedLinks.add(link); } // 删除Enterprise链接 else if ( (linkHref.includes('/enterprise') || linkHref === '/enterprise' || linkText === 'Enterprise' || linkText.match(/^s*Enterprises*$/i)) ) { if (link.parentNode) { link.parentNode.removeChild(link); } replacedLinks.add(link); } }); // 查找可能嵌套的Spaces和Posts文本 const textNodes = []; function findTextNodes(element) { if (element.nodeType === Node.TEXT_NODE) { const text = element.textContent.trim(); if (text === 'Spaces' || text === 'Posts' || text === 'Enterprise') { textNodes.push(element); } } else { for (const child of element.childNodes) { findTextNodes(child); } } } // 只在导航区域内查找文本节点 findTextNodes(headerArea); // 替换找到的文本节点 textNodes.forEach(node => { const text = node.textContent.trim(); if (text === 'Spaces') { node.textContent = node.textContent.replace(/Spaces/g, 'PDF TO Markdown'); } else if (text === 'Posts') { // 删除Posts文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } else if (text === 'Enterprise') { // 删除Enterprise文本节点 if (node.parentNode) { node.parentNode.removeChild(node); } } }); // 标记已替换完成 window._navLinksReplaced = true; } // 替换代码区域中的域名 function replaceCodeDomains() { // 特别处理span.hljs-string和span.njs-string元素 document.querySelectorAll('span.hljs-string, span.njs-string, span[class*="hljs-string"], span[class*="njs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换hljs-string类的span中的域名(移除多余的转义符号) document.querySelectorAll('span.hljs-string, span[class*="hljs-string"]').forEach(span => { if (span.textContent && span.textContent.includes('huggingface.co')) { span.textContent = span.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 替换pre和code标签中包含git clone命令的域名 document.querySelectorAll('pre, code').forEach(element => { if (element.textContent && element.textContent.includes('git clone')) { const text = element.innerHTML; if (text.includes('huggingface.co')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 处理特定的命令行示例 document.querySelectorAll('pre, code').forEach(element => { const text = element.innerHTML; if (text.includes('huggingface.co')) { // 针对git clone命令的专门处理 if (text.includes('git clone') || text.includes('GIT_LFS_SKIP_SMUDGE=1')) { element.innerHTML = text.replace(/huggingface.co/g, 'aifasthub.com'); } } }); // 特别处理模型下载页面上的代码片段 document.querySelectorAll('.flex.border-t, .svelte_hydrator, .inline-block').forEach(container => { const content = container.innerHTML; if (content && content.includes('huggingface.co')) { container.innerHTML = content.replace(/huggingface.co/g, 'aifasthub.com'); } }); // 特别处理模型仓库克隆对话框中的代码片段 try { // 查找包含"Clone this model repository"标题的对话框 const cloneDialog = document.querySelector('.svelte_hydration_boundary, [data-target="MainHeader"]'); if (cloneDialog) { // 查找对话框中所有的代码片段和命令示例 const codeElements = cloneDialog.querySelectorAll('pre, code, span'); codeElements.forEach(element => { if (element.textContent && element.textContent.includes('huggingface.co')) { if (element.innerHTML.includes('huggingface.co')) { element.innerHTML = element.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { element.textContent = element.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); } // 更精确地定位克隆命令中的域名 document.querySelectorAll('[data-target]').forEach(container => { const codeBlocks = container.querySelectorAll('pre, code, span.hljs-string'); codeBlocks.forEach(block => { if (block.textContent && block.textContent.includes('huggingface.co')) { if (block.innerHTML.includes('huggingface.co')) { block.innerHTML = block.innerHTML.replace(/huggingface.co/g, 'aifasthub.com'); } else { block.textContent = block.textContent.replace(/huggingface.co/g, 'aifasthub.com'); } } }); }); } catch (e) { // 错误处理但不打印日志 } } // 当DOM加载完成后执行替换 if (document.readyState === 'loading') { document.addEventListener('DOMContentLoaded', () => { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); }); } else { replaceHeaderBranding(); replaceNavigationLinks(); replaceCodeDomains(); // 只在必要时执行替换 - 3秒后再次检查 setTimeout(() => { if (!window._navLinksReplaced) { console.log('[Client] 3秒后重新检查导航链接'); replaceNavigationLinks(); } }, 3000); } // 增加一个MutationObserver来处理可能的动态元素加载 const observer = new MutationObserver(mutations => { // 检查是否导航区域有变化 const hasNavChanges = mutations.some(mutation => { // 检查是否存在header或nav元素变化 return Array.from(mutation.addedNodes).some(node => { if (node.nodeType === Node.ELEMENT_NODE) { // 检查是否是导航元素或其子元素 if (node.tagName === 'HEADER' || node.tagName === 'NAV' || node.querySelector('header, nav')) { return true; } // 检查是否在导航元素内部 let parent = node.parentElement; while (parent) { if (parent.tagName === 'HEADER' || parent.tagName === 'NAV') { return true; } parent = parent.parentElement; } } return false; }); }); // 只在导航区域有变化时执行替换 if (hasNavChanges) { // 重置替换状态,允许再次替换 window._navLinksReplaced = false; replaceHeaderBranding(); replaceNavigationLinks(); } }); // 开始观察document.body的变化,包括子节点 if (document.body) { observer.observe(document.body, { childList: true, subtree: true }); } else { document.addEventListener('DOMContentLoaded', () => { observer.observe(document.body, { childList: true, subtree: true }); }); } })(); ' % content.encode('utf-8')\n content = content.decode(\"utf-8\")\n else:\n content = '%s' % content\n parser.feed(content)\n parser.close()\n content = parser.body\n if ('IPYNB_USE_META_SUMMARY' in self.settings.keys() and self.settings['IPYNB_USE_META_SUMMARY'] is False) or 'IPYNB_USE_META_SUMMARY' not in self.settings.keys():\n metadata['summary'] = parser.summary\n\n content = fix_css(content, info)\n return content, metadata\n\n\nclass MyHTMLParser(HTMLReader._HTMLParser):\n \"\"\"\n Custom Pelican `HTMLReader._HTMLParser` to create the summary of the content\n based on settings['SUMMARY_MAX_LENGTH'].\n\n Summary is stoped if founds any div containing ipython notebook code cells.\n This is needed in order to generate valid HTML for the summary,\n a simple string split will break the html generating errors on the theme.\n The downside is that the summary length is not exactly the specified, it stops at\n completed div/p/li/etc tags.\n \"\"\"\n def __init__(self, settings, filename):\n HTMLReader._HTMLParser.__init__(self, settings, filename)\n self.settings = settings\n self.filename = filename\n self.wordcount = 0\n self.summary = None\n\n self.stop_tags = [('div', ('class', 'input')), ('div', ('class', 'output')), ('h2', ('id', 'Header-2'))]\n if 'IPYNB_STOP_SUMMARY_TAGS' in self.settings.keys():\n self.stop_tags = self.settings['IPYNB_STOP_SUMMARY_TAGS']\n if 'IPYNB_EXTEND_STOP_SUMMARY_TAGS' in self.settings.keys():\n self.stop_tags.extend(self.settings['IPYNB_EXTEND_STOP_SUMMARY_TAGS'])\n\n def handle_starttag(self, tag, attrs):\n HTMLReader._HTMLParser.handle_starttag(self, tag, attrs)\n\n if self.wordcount < self.settings['SUMMARY_MAX_LENGTH']:\n mask = [stoptag[0] == tag and (stoptag[1] is None or stoptag[1] in attrs) for stoptag in self.stop_tags]\n if any(mask):\n self.summary = self._data_buffer\n self.wordcount = self.settings['SUMMARY_MAX_LENGTH']\n\n def handle_endtag(self, tag):\n HTMLReader._HTMLParser.handle_endtag(self, tag)\n\n if self.wordcount < self.settings['SUMMARY_MAX_LENGTH']:\n self.wordcount = len(strip_tags(self._data_buffer).split(' '))\n if self.wordcount >= self.settings['SUMMARY_MAX_LENGTH']:\n self.summary = self._data_buffer\n\n\ndef strip_tags(html):\n \"\"\"\n Strip html tags from html content (str)\n Useful for summary creation\n \"\"\"\n s = HTMLTagStripper()\n s.feed(html)\n return s.get_data()\n\n\nclass HTMLTagStripper(HTMLParser):\n \"\"\"\n Custom HTML Parser to strip HTML tags\n Useful for summary creation\n \"\"\"\n def __init__(self):\n HTMLParser.__init__(self)\n self.reset()\n self.fed = []\n\n def handle_data(self, html):\n self.fed.append(html)\n\n def get_data(self):\n return ''.join(self.fed)\n"},"license":{"kind":"string","value":"agpl-3.0"},"hash":{"kind":"number","value":-8813531780843853000,"string":"-8,813,531,780,843,853,000"},"line_mean":{"kind":"number","value":34.9696969697,"string":"34.969697"},"line_max":{"kind":"number","value":175,"string":"175"},"alpha_frac":{"kind":"number","value":0.6094355518,"string":"0.609436"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.9752176825184193,"string":"3.975218"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42159,"cells":{"repo_name":{"kind":"string","value":"mtlynch/ndt-e2e-clientworker"},"path":{"kind":"string","value":"client_wrapper/install_selenium_extensions.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"3193"},"content":{"kind":"string","value":"import argparse\nimport os\nimport platform\nimport urllib\nimport tempfile\n\nimport names\n\ndriver_urls = {\n 'chrome_os_x': {\n 'url':\n 'http://chromedriver.storage.googleapis.com/2.21/chromedriver_mac32.zip',\n 'file_name': 'chromedriver_mac32.zip'\n },\n 'chrome_ubuntu': {\n 'url':\n 'http://chromedriver.storage.googleapis.com/2.21/chromedriver_linux64.zip',\n 'file_name': 'chromedriver_linux64.zip'\n },\n 'chrome_windows_10': {\n 'url':\n 'http://chromedriver.storage.googleapis.com/2.21/chromedriver_win32.zip',\n 'file_name': 'chromedriver_win32.zip'\n },\n 'edge_windows_10': {\n 'url':\n 'https://download.microsoft.com/download/8/D/0/8D0D08CF-790D-4586-B726-C6469A9ED49C/MicrosoftWebDriver.msi',\n 'file_name': 'MicrosoftWebDriver.msi'\n },\n 'safari_os_x': {\n 'url':\n 'http://selenium-release.storage.googleapis.com/2.48/SafariDriver.safariextz',\n 'file_name': 'SafariDriver.safariextz',\n }\n}\n\n\ndef _download_chrome_drivers():\n \"\"\"Downloads Chrome drivers for Selenium.\"\"\"\n # Mac OS X\n if platform.system() == 'Darwin':\n remote_file = driver_urls['chrome_os_x']\n elif platform.system() == 'Linux':\n remote_file = driver_urls['chrome_ubuntu']\n elif platform.system() == 'Windows':\n remote_file = driver_urls['chrome_windows_10']\n else:\n raise ValueError('Unsupported OS specified: %s' % (platform.system()))\n _download_temp_file(remote_file['url'], remote_file['file_name'])\n\n\ndef _download_temp_file(url, file_name):\n \"\"\"Downloads file into temp directory.\n\n Args:\n url: A string representing the URL the file is to be downloaded from.\n file_name: A string representing the name of the file to be downloaded.\n \"\"\"\n temp_dir = tempfile.mkdtemp()\n download_path = os.path.join(temp_dir, file_name)\n print('File downloading to %s' % download_path)\n urllib.URLopener().retrieve(url, download_path)\n\n\ndef _download_edge_drivers():\n \"\"\"Downloads Edge drivers for Selenium.\"\"\"\n remote_file = driver_urls['edge_windows_10']\n _download_temp_file(remote_file['url'], remote_file['file_name'])\n\n\ndef _download_safari_drivers():\n \"\"\"Downloads Safari drivers for Selenium.\"\"\"\n remote_file = driver_urls['safari_os_x']\n _download_temp_file(remote_file['url'], remote_file['file_name'])\n\n\ndef main(args):\n if args.browser == names.CHROME:\n _download_chrome_drivers()\n elif args.browser == names.EDGE:\n _download_edge_drivers()\n elif args.browser == names.SAFARI:\n _download_safari_drivers()\n elif args.browser == names.FIREFOX:\n pass\n else:\n raise ValueError('Unsupported browser specified: %s' % (args.browser))\n\n\nif __name__ == '__main__':\n parser = argparse.ArgumentParser(\n prog='NDT E2E Testing Client Selenium Extension Installer',\n formatter_class=argparse.ArgumentDefaultsHelpFormatter)\n parser.add_argument('--browser',\n help='Browser to run under (for browser-based client)',\n choices=('chrome', 'firefox', 'safari', 'edge'))\n main(parser.parse_args())\n"},"license":{"kind":"string","value":"apache-2.0"},"hash":{"kind":"number","value":7890214524247196000,"string":"7,890,214,524,247,196,000"},"line_mean":{"kind":"number","value":31.9175257732,"string":"31.917526"},"line_max":{"kind":"number","value":116,"string":"116"},"alpha_frac":{"kind":"number","value":0.6348261823,"string":"0.634826"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.5242825607064017,"string":"3.524283"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42160,"cells":{"repo_name":{"kind":"string","value":"cwacek/python-jsonschema-objects"},"path":{"kind":"string","value":"test/test_regression_126.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1829"},"content":{"kind":"string","value":"import pytest\nimport python_jsonschema_objects as pjs\nimport collections\n\n\n@pytest.fixture\ndef schema():\n return {\n \"$schema\": \"http://json-schema.org/draft-04/schema#\",\n \"title\": \"Test\",\n \"definitions\": {\n \"MyEnum1\": {\"type\": \"string\", \"enum\": [\"E_A\", \"E_B\"]},\n \"MyEnum2\": {\"type\": \"string\", \"enum\": [\"F_A\", \"F_B\", \"F_C\", \"F_D\"]},\n \"MyInt\": {\n \"default\": \"0\",\n \"type\": \"integer\",\n \"minimum\": 0,\n \"maximum\": 4294967295,\n },\n \"MyObj1\": {\n \"type\": \"object\",\n \"properties\": {\n \"e1\": {\"$ref\": \"#/definitions/MyEnum1\"},\n \"e2\": {\"$ref\": \"#/definitions/MyEnum2\"},\n \"i1\": {\"$ref\": \"#/definitions/MyInt\"},\n },\n \"required\": [\"e1\", \"e2\", \"i1\"],\n },\n \"MyArray\": {\n \"type\": \"array\",\n \"items\": {\"$ref\": \"#/definitions/MyObj1\"},\n \"minItems\": 0,\n \"uniqueItems\": True,\n },\n \"MyMsg1\": {\n \"type\": \"object\",\n \"properties\": {\"a1\": {\"$ref\": \"#/definitions/MyArray\"}},\n },\n \"MyMsg2\": {\"type\": \"object\", \"properties\": {\"s1\": {\"type\": \"string\"}}},\n },\n \"type\": \"object\",\n \"oneOf\": [{\"$ref\": \"#/definitions/MyMsg1\"}, {\"$ref\": \"#/definitions/MyMsg2\"}],\n }\n\n\ndef test_regression_126(schema):\n builder = pjs.ObjectBuilder(schema)\n ns = builder.build_classes(standardize_names=False)\n\n Obj1 = ns.MyObj1\n Array1 = ns.MyArray\n Msg1 = ns.MyMsg1\n o1 = Obj1(e1=\"E_A\", e2=\"F_C\", i1=2600)\n o2 = Obj1(e1=\"E_B\", e2=\"F_D\", i1=2500)\n objs = Array1([o1, o2])\n msg = Msg1(a1=objs)\n\n print(msg.serialize())\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":-1888342341075785700,"string":"-1,888,342,341,075,785,700"},"line_mean":{"kind":"number","value":30.5344827586,"string":"30.534483"},"line_max":{"kind":"number","value":86,"string":"86"},"alpha_frac":{"kind":"number","value":0.4286495353,"string":"0.42865"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.3621323529411766,"string":"3.362132"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42161,"cells":{"repo_name":{"kind":"string","value":"Krakn/learning"},"path":{"kind":"string","value":"src/python/advent_of_code/2017/05/a_maze_of_twisty_trampolines_all_alike.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"3322"},"content":{"kind":"string","value":"#!/usr/bin/env python3\n\n'''\n--- Day 5: A Maze of Twisty Trampolines, All Alike ---\n'''\n\n\ndef load_input(filename):\n '''\n Parse input file, returning an array of maze offsets.\n '''\n maze = list()\n with open(filename, 'r') as file_input:\n for line in file_input.readlines():\n maze.append(int(line.strip()))\n return maze\n\n\ndef part1(maze):\n '''\n --- Part 1 ---\n\n An urgent interrupt arrives from the CPU: it's trapped in a maze of jump\n instructions, and it would like assistance from any programs with spare\n cycles to help find the exit.\n\n The message includes a list of the offsets for each jump. Jumps are\n relative: -1 moves to the previous instruction, and 2 skips the next one.\n Start at the first instruction in the list. The goal is to follow the jumps\n until one leads outside the list.\n\n In addition, these instructions are a little strange; after each jump, the\n offset of that instruction increases by 1. So, if you come across an offset\n of 3, you would move three instructions forward, but change it to a 4 for\n the next time it is encountered.\n\n For example, consider the following list of jump offsets:\n\n 0 3 0 1 -3 Positive jumps (\"forward\") move downward; negative jumps move\n upward. For legibility in this example, these offset values\n will be written all on one line, with the current instruction\n marked in parentheses. The following steps would be taken\n before an exit is found:\n\n (0) 3 0 1 -3 - Before we have taken any steps.\n (1) 3 0 1 -3 - Jump with offset 0 (that is, don't jump at all).\n Fortunately, the instruction is then incremented\n to 1.\n 2 (3) 0 1 -3 - Step forward because of the instruction we just modified.\n The first instruction is incremented again, now to 2.\n 2 4 0 1 (-3) - Jump all the way to the end; leave a 4 behind.\n 2 (4) 0 1 -2 - Go back to where we just were; increment -3 to -2.\n 2 5 0 1 -2 - Jump 4 steps forward, escaping the maze. In this\n example, the exit is reached in 5 steps.\n\n How many steps does it take to reach the exit?\n '''\n index = 0\n steps = 0\n while index >= 0 and index < len(maze):\n maze[index] += 1\n index = index + maze[index] - 1\n steps += 1\n return steps\n\n\ndef part2(maze):\n '''\n --- Part Two ---\n\n Now, the jumps are even stranger: after each jump, if the offset was three\n or more, instead decrease it by 1. Otherwise, increase it by 1 as before.\n\n Using this rule with the above example, the process now takes 10 steps,\n and the offset values after finding the exit are left as 2 3 2 3 -1.\n\n How many steps does it now take to reach the exit?\n '''\n index = 0\n steps = 0\n while index >= 0 and index < len(maze):\n if maze[index] >= 3:\n maze[index] -= 1\n index = index + maze[index] + 1\n else:\n maze[index] += 1\n index = index + maze[index] - 1\n steps += 1\n return steps\n\n\nif __name__ == \"__main__\":\n MAZE1 = load_input('input.txt')\n MAZE2 = load_input('input.txt')\n print(\"Part 1:\", part1(MAZE1))\n print(\"Part 2:\", part2(MAZE2))\n"},"license":{"kind":"string","value":"isc"},"hash":{"kind":"number","value":-1629784482675308500,"string":"-1,629,784,482,675,308,500"},"line_mean":{"kind":"number","value":33.6041666667,"string":"33.604167"},"line_max":{"kind":"number","value":79,"string":"79"},"alpha_frac":{"kind":"number","value":0.6095725467,"string":"0.609573"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.8052691867124855,"string":"3.805269"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42162,"cells":{"repo_name":{"kind":"string","value":"winterbird-code/adbb"},"path":{"kind":"string","value":"adbb/__init__.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"2124"},"content":{"kind":"string","value":"#!/usr/bin/env python\n#\n# This file is part of adbb.\n#\n# adbb is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# adbb is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with adbb. If not, see .\n\nimport multiprocessing\nimport logging\nimport logging.handlers\nimport sys\n\nimport adbb.db\nfrom adbb.link import AniDBLink\n\nfrom adbb.animeobjs import Anime, AnimeTitle, Episode, File\n\nfrom adbb.anames import get_titles\n\nanidb_client_name = \"adbb\"\nanidb_client_version = 2\nanidb_api_version = 3\n\nlog = None\n_anidb = None\n_sessionmaker = None\n\n\ndef init(\n anidb_user,\n anidb_pwd,\n sql_db_url,\n debug=False,\n loglevel='info',\n logger=None,\n outgoing_udp_port=9876):\n\n if logger is None:\n logger = logging.getLogger(__name__)\n logger.setLevel(loglevel.upper())\n if debug:\n logger.setLevel(logging.DEBUG)\n lh = logging.StreamHandler()\n lh.setFormatter(logging.Formatter(\n '%(asctime)s %(levelname)s %(filename)s:%(lineno)d - %(message)s'))\n logger.addHandler(lh)\n lh = logging.handlers.SysLogHandler(address='/dev/log')\n lh.setFormatter(logging.Formatter(\n 'adbb %(filename)s/%(funcName)s:%(lineno)d - %(message)s'))\n logger.addHandler(lh)\n\n global log, _anidb, _sessionmaker\n log = logger\n _sessionmaker = adbb.db.init_db(sql_db_url)\n _anidb = adbb.link.AniDBLink(\n anidb_user,\n anidb_pwd,\n myport=outgoing_udp_port)\n\n\ndef get_session():\n return _sessionmaker()\n\n\ndef close_session(session):\n session.close()\n\n\ndef close():\n global _anidb\n _anidb.stop()\n"},"license":{"kind":"string","value":"gpl-3.0"},"hash":{"kind":"number","value":-7247581232844061000,"string":"-7,247,581,232,844,061,000"},"line_mean":{"kind":"number","value":25.2222222222,"string":"25.222222"},"line_max":{"kind":"number","value":83,"string":"83"},"alpha_frac":{"kind":"number","value":0.6638418079,"string":"0.663842"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.504950495049505,"string":"3.50495"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42163,"cells":{"repo_name":{"kind":"string","value":"jcfr/mystic"},"path":{"kind":"string","value":"examples/TEST_ffitPP2_b.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1429"},"content":{"kind":"string","value":"#!/usr/bin/env python\n#\n# Author: Mike McKerns (mmckerns @caltech and @uqfoundation)\n# Copyright (c) 1997-2015 California Institute of Technology.\n# License: 3-clause BSD. The full license text is available at:\n# - http://trac.mystic.cacr.caltech.edu/project/mystic/browser/mystic/LICENSE\n\"\"\"\nTesting the polynomial fitting problem of [1] using scipy's Nelder-Mead algorithm.\n\nReference:\n\n[1] Storn, R. and Price, K. Differential Evolution - A Simple and Efficient\nHeuristic for Global Optimization over Continuous Spaces. Journal of Global\nOptimization 11: 341-359, 1997.\n\"\"\"\n\nfrom test_ffit import Chebyshev8, plot_solution, print_solution\nfrom TEST_ffitPP_b import ChebyshevCost\n\nif __name__ == '__main__':\n import random\n from mystic.solvers import fmin\n #from mystic._scipyoptimize import fmin\n from mystic.tools import random_seed\n random_seed(123)\n\n import pp\n import sys\n\n if len(sys.argv) > 1:\n tunnelport = sys.argv[1]\n ppservers = (\"localhost:%s\" % tunnelport,)\n else:\n ppservers = ()\n\n myserver = pp.Server(ppservers=ppservers)\n\n trials = []\n for trial in range(8):\n x = tuple([random.uniform(-100,100) + Chebyshev8[i] for i in range(9)])\n trials.append(x)\n\n results = [myserver.submit(fmin,(ChebyshevCost,x),(),()) for x in trials]\n\n for solution in results:\n print_solution(solution())\n\n #plot_solution(solution)\n\n# end of file\n"},"license":{"kind":"string","value":"bsd-3-clause"},"hash":{"kind":"number","value":4351184427107404300,"string":"4,351,184,427,107,404,300"},"line_mean":{"kind":"number","value":27.58,"string":"27.58"},"line_max":{"kind":"number","value":82,"string":"82"},"alpha_frac":{"kind":"number","value":0.6871938418,"string":"0.687194"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.2775229357798166,"string":"3.277523"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42164,"cells":{"repo_name":{"kind":"string","value":"Aloomaio/googleads-python-lib"},"path":{"kind":"string","value":"examples/ad_manager/v201805/creative_service/create_creative_from_template.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"3666"},"content":{"kind":"string","value":"#!/usr/bin/env python\n#\n# Copyright 2015 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"This code example creates a new template creative for a given advertiser.\n\nTo determine which companies are advertisers, run get_advertisers.py.\nTo determine which creative templates exist, run\nget_all_creative_templates.py.\n\nThe LoadFromStorage method is pulling credentials and properties from a\n\"googleads.yaml\" file. By default, it looks for this file in your home\ndirectory. For more information, see the \"Caching authentication information\"\nsection of our README.\n\n\"\"\"\n\n\nimport os\nimport uuid\n\n# Import appropriate modules from the client library.\nfrom googleads import ad_manager\n\n# Set id of the advertiser (company) that the creative will be assigned to.\nADVERTISER_ID = 'INSERT_ADVERTISER_COMPANY_ID_HERE'\n\n\ndef main(client, advertiser_id):\n # Initialize appropriate service.\n creative_service = client.GetService('CreativeService', version='v201805')\n\n # Use the image banner with optional third party tracking template.\n creative_template_id = '10000680'\n\n # Create image asset.\n file_name = 'image%s.jpg' % uuid.uuid4()\n image_data = open(os.path.join(os.path.split(__file__)[0], '..', '..', 'data',\n 'medium_rectangle.jpg'), 'r').read()\n size = {\n 'width': '300',\n 'height': '250'\n }\n asset = {\n 'xsi_type': 'CreativeAsset',\n 'fileName': file_name,\n 'assetByteArray': image_data,\n 'size': size\n }\n\n # Create creative from templates.\n creative = {\n 'xsi_type': 'TemplateCreative',\n 'name': 'Template Creative #%s' % uuid.uuid4(),\n 'advertiserId': advertiser_id,\n 'size': size,\n 'creativeTemplateId': creative_template_id,\n 'creativeTemplateVariableValues': [\n {\n 'xsi_type': 'AssetCreativeTemplateVariableValue',\n 'uniqueName': 'Imagefile',\n 'asset': asset\n },\n {\n 'xsi_type': 'LongCreativeTemplateVariableValue',\n 'uniqueName': 'Imagewidth',\n 'value': '300'\n },\n {\n 'xsi_type': 'LongCreativeTemplateVariableValue',\n 'uniqueName': 'Imageheight',\n 'value': '250'\n },\n {\n 'xsi_type': 'UrlCreativeTemplateVariableValue',\n 'uniqueName': 'ClickthroughURL',\n 'value': 'www.google.com'\n },\n {\n 'xsi_type': 'StringCreativeTemplateVariableValue',\n 'uniqueName': 'Targetwindow',\n 'value': '_blank'\n }\n ]\n }\n\n # Call service to create the creative.\n creative = creative_service.createCreatives([creative])[0]\n\n # Display results.\n print ('Template creative with id \"%s\", name \"%s\", and type \"%s\" was '\n 'created and can be previewed at %s.'\n % (creative['id'], creative['name'],\n ad_manager.AdManagerClassType(creative), creative['previewUrl']))\n\nif __name__ == '__main__':\n # Initialize client object.\n ad_manager_client = ad_manager.AdManagerClient.LoadFromStorage()\n main(ad_manager_client, ADVERTISER_ID)\n"},"license":{"kind":"string","value":"apache-2.0"},"hash":{"kind":"number","value":1060630061611670300,"string":"1,060,630,061,611,670,300"},"line_mean":{"kind":"number","value":32.027027027,"string":"32.027027"},"line_max":{"kind":"number","value":80,"string":"80"},"alpha_frac":{"kind":"number","value":0.6382978723,"string":"0.638298"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.887592788971368,"string":"3.887593"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42165,"cells":{"repo_name":{"kind":"string","value":"eduble/panteda"},"path":{"kind":"string","value":"operators/map/heatmap.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"4193"},"content":{"kind":"string","value":"#!/usr/bin/env python3\nimport numpy as np\nimport numpy.random\nfrom time import time\n\n# web mercator projection functions\n# ---------------------------------\ndef linear_lat(lat, atanh = np.arctanh, sin = np.sin, radians = np.radians):\n return atanh(sin(radians(lat)))\n\ndef inv_linear_lat(ll, asin = np.arcsin, tanh = np.tanh, degrees = np.degrees):\n return degrees(asin(tanh(ll)))\n\ndef lng_to_x(w, lng_min, lng_max, lng):\n return (lng - lng_min) * (w / (lng_max - lng_min))\n\ndef lat_to_y(h, lat_min, lat_max, lat):\n return (linear_lat(lat) - linear_lat(lat_min)) * (h / (linear_lat(lat_max) - linear_lat(lat_min)))\n\ndef x_to_lng(w, lng_min, lng_max, x):\n return x * ((lng_max - lng_min)/w) + lng_min\n\ndef y_to_lat(h, lat_min, lat_max, y):\n return inv_linear_lat(y * ((linear_lat(lat_max) - linear_lat(lat_min))/h) + linear_lat(lat_min))\n\n# heatmap data generation\n# -----------------------\nclass HeatMap:\n def __init__(self, lnglat, width, height, westlng, eastlng, southlat, northlat):\n # compute pixel bounds of the map\n x = np.append(np.arange(0, width, 5), width)\n y = np.append(np.arange(0, height, 5), height)\n # project pixel bounds coordinates (x, y -> lng, lat)\n edgelng = x_to_lng(width, westlng, eastlng, x)\n centerlng = x_to_lng(width, westlng, eastlng, (x[1:] + x[:-1])/2)\n edgelat = y_to_lat(height, southlat, northlat, y)\n centerlat = y_to_lat(height, southlat, northlat, (y[1:] + y[:-1])/2)\n # prepare computation parameters\n self.bins = edgelng, edgelat\n self.range = (westlng, eastlng), (southlat, northlat)\n self.iterator = lnglat.chunks()\n self.heatmap = None\n # prepare compression parameters\n scalelat = (edgelat[1:] - edgelat[:-1]).min() / 2\n self.approx_centerlat = numpy.rint((centerlat - centerlat[0]) / scalelat)\n scalelng = edgelng[1] - edgelng[0] # longitude is linear\n self.approx_centerlng = numpy.rint((centerlng - centerlng[0]) / scalelng)\n self.scales = dict(lat=scalelat, lng=scalelng)\n self.offsets = dict(lat=centerlat[0], lng=centerlng[0])\n # stream status parameters\n self.done = False\n def compute(self, time_credit):\n # make histogram:\n # - create a pixel grid\n # - given a tuple (lng, lat) increment the corresponding pixel\n deadline = time() + time_credit\n deadline_reached = False\n for chunk in self.iterator:\n lng, lat = chunk.columns\n chunk_heatmap = np.histogram2d(lng, lat, bins=self.bins, range=self.range)[0]\n if self.heatmap is None:\n self.heatmap = chunk_heatmap.T\n else:\n self.heatmap += chunk_heatmap.T\n if time() > deadline:\n deadline_reached = True\n break\n if not deadline_reached:\n # we left the loop because of the end of iteration\n self.done = True\n # get sparse matrix representation: (lat, lng, intensity) tuples.\n # in order to lower network usage, we will transfer this data in a\n # compressed form: lng & lat values will be transfered as integers\n # together with a scaling factor and an offset to be applied.\n def compressed_form(self):\n # count number of points\n count = int(self.heatmap.sum())\n if count == 0:\n # if no points, return empty data\n data = dict(lat = [], lng = [], val = [])\n else:\n # apply threshold and\n # compute approximated sparse matrix data\n nonzero_xy = ((self.heatmap / self.heatmap.max()) > 0.05).nonzero()\n nonzero_x = nonzero_xy[1]\n nonzero_y = nonzero_xy[0]\n data = dict(\n lat = self.approx_centerlat[nonzero_y].astype(int).tolist(),\n lng = self.approx_centerlng[nonzero_x].astype(int).tolist(),\n val = self.heatmap[nonzero_xy].astype(int).tolist()\n )\n return dict(\n data = data,\n scales = self.scales,\n offsets = self.offsets,\n count = count,\n done = self.done\n )\n"},"license":{"kind":"string","value":"gpl-3.0"},"hash":{"kind":"number","value":-2101006383756166100,"string":"-2,101,006,383,756,166,100"},"line_mean":{"kind":"number","value":41.7857142857,"string":"41.785714"},"line_max":{"kind":"number","value":102,"string":"102"},"alpha_frac":{"kind":"number","value":0.5771523969,"string":"0.577152"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.4941666666666666,"string":"3.494167"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42166,"cells":{"repo_name":{"kind":"string","value":"mohitreddy1996/Gender-Detection-from-Signature"},"path":{"kind":"string","value":"src/train_test/random_forests.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1140"},"content":{"kind":"string","value":"from sklearn.metrics import precision_recall_fscore_support\n\nimport pandas as pd\nimport numpy as np\nfrom sklearn.ensemble import RandomForestClassifier\nfrom sklearn.preprocessing import MinMaxScaler, normalize\n\ndf = pd.read_csv('../../Dataset/dataset.csv', delimiter='\\t')\n\ndataset = df.values\n\nmask = np.random.rand(len(df)) < .80\n\ntrain = df[mask]\ntest = df[~mask]\n\nX = pd.DataFrame()\nY = pd.DataFrame()\n\nX = train.ix[:, 2:len(train.columns) - 1]\nY = train.ix[:, len(train.columns) - 1: len(train.columns)]\n\nX_Test = pd.DataFrame()\nY_Test = pd.DataFrame()\n\n# After Normalising\nX_standard = normalize(X)\nprint X_standard.shape\n\n\nX_Test = test.ix[:, 2:len(test.columns) - 1]\nY_Test = test.ix[:, len(test.columns) - 1: len(test.columns)]\n\nX_Test_standard = normalize(X_Test)\nprint X_Test_standard.shape\n\nprint \"Training Data Set Size : \", str(len(X))\nprint \"Testing Data Set Size : \", str(len(X_Test))\n\n# tune parameters here.\nrf = RandomForestClassifier(n_estimators=150, max_features=20)\n\nrf.fit(X_standard, Y)\n# predict\nY_Result = rf.predict(X_Test_standard)\n\nprint precision_recall_fscore_support(Y_Test, Y_Result, average='micro')\n\n\n\n\n\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":5244072690348756000,"string":"5,244,072,690,348,756,000"},"line_mean":{"kind":"number","value":20.9230769231,"string":"20.923077"},"line_max":{"kind":"number","value":72,"string":"72"},"alpha_frac":{"kind":"number","value":0.7114035088,"string":"0.711404"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":2.961038961038961,"string":"2.961039"},"config_test":{"kind":"bool","value":true,"string":"true"},"has_no_keywords":{"kind":"bool","value":true,"string":"true"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42167,"cells":{"repo_name":{"kind":"string","value":"aquaya/ivrhub"},"path":{"kind":"string","value":"ivrhub/models.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"3129"},"content":{"kind":"string","value":"''' mongoengine models\n'''\nfrom mongoengine import *\n\n\nclass User(Document):\n ''' some are admins some are not\n '''\n admin_rights = BooleanField(required=True)\n api_id = StringField()\n api_key = StringField()\n email = EmailField(required=True, unique=True, max_length=254)\n email_confirmation_code = StringField(required=True)\n email_confirmed = BooleanField(required=True)\n forgot_password_code = StringField()\n last_login_time = DateTimeField(required=True)\n name = StringField()\n organizations = ListField(ReferenceField('Organization'))\n password_hash = StringField(required=True)\n registration_time = DateTimeField(required=True)\n verified = BooleanField(required=True)\n\n\nclass Organization(Document):\n ''' people join orgs\n '''\n description = StringField(default='')\n # url-safe version of the name\n label = StringField(unique=True, required=True)\n location = StringField(default='')\n name = StringField(unique=True, required=True)\n\n\nclass Form(Document):\n ''' the heart of the system\n '''\n # unique code for requesting this form via sms or a call\n calling_code = StringField()\n creation_time = DateTimeField()\n creator = ReferenceField(User)\n description = StringField(default = '')\n # url-safe version of the name\n label = StringField(unique_with='organization')\n language = StringField(default = '')\n name = StringField(unique_with='organization')\n organization = ReferenceField(Organization)\n # have to store questions here as well so we know the order\n questions = ListField(ReferenceField('Question'))\n\n\nclass Question(Document):\n ''' connected to forms\n '''\n audio_filename = StringField()\n audio_url = StringField()\n creation_time = DateTimeField()\n description = StringField()\n form = ReferenceField(Form)\n # url-safe version of the name\n label = StringField(unique_with='form')\n name = StringField(unique_with='form')\n # 'text_prompt', 'audio_file' or 'audio_url'\n prompt_type = StringField(default='text_prompt')\n # 'keypad' or 'voice' or 'no response'\n response_type = StringField(default='keypad')\n s3_key = StringField()\n s3_url = StringField()\n text_prompt = StringField()\n text_prompt_language = StringField(default='en')\n\n\nclass Response(Document):\n ''' individual response to a form\n '''\n call_sid = StringField()\n completion_time = DateTimeField()\n form = ReferenceField(Form)\n # whether this was a 'call' or 'ringback' or 'scheduled call'\n initiated_using = StringField()\n initiation_time = DateTimeField()\n # track the progress of the response\n last_question_asked = ReferenceField(Question)\n # any notes about the response as a whole\n notes = StringField()\n respondent_phone_number = StringField()\n\n\nclass Answer(Document):\n ''' connected to questions and responses\n '''\n audio_url = StringField()\n keypad_input = StringField()\n # any notes on this answer (like a transcription)\n notes = StringField()\n question = ReferenceField(Question)\n response = ReferenceField(Response)\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":7868206963524489000,"string":"7,868,206,963,524,489,000"},"line_mean":{"kind":"number","value":31.59375,"string":"31.59375"},"line_max":{"kind":"number","value":66,"string":"66"},"alpha_frac":{"kind":"number","value":0.6868008949,"string":"0.686801"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":4.298076923076923,"string":"4.298077"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42168,"cells":{"repo_name":{"kind":"string","value":"tzangms/PyConTW"},"path":{"kind":"string","value":"pycon_project/biblion/views.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"3501"},"content":{"kind":"string","value":"from datetime import datetime\n\nfrom django.core.urlresolvers import reverse\nfrom django.http import HttpResponse, Http404\nfrom django.shortcuts import render_to_response, get_object_or_404\nfrom django.template import RequestContext\nfrom django.template.loader import render_to_string\nfrom django.utils import simplejson as json\n\nfrom django.contrib.sites.models import Site\n\nfrom biblion.exceptions import InvalidSection\nfrom biblion.models import Post, FeedHit\nfrom biblion.settings import ALL_SECTION_NAME\n\n\ndef blog_index(request):\n \n posts = Post.objects.current()\n posts = posts.filter(language=request.LANGUAGE_CODE)\n \n return render_to_response(\"biblion/blog_list.html\", {\n \"posts\": posts,\n }, context_instance=RequestContext(request))\n\n\ndef blog_section_list(request, section):\n \n try:\n posts = Post.objects.section(section)\n except InvalidSection:\n raise Http404()\n \n return render_to_response(\"biblion/blog_section_list.html\", {\n \"section_slug\": section,\n \"section_name\": dict(Post.SECTION_CHOICES)[Post.section_idx(section)],\n \"posts\": posts,\n }, context_instance=RequestContext(request))\n\n\ndef blog_post_detail(request, **kwargs):\n \n if \"post_pk\" in kwargs:\n if request.user.is_authenticated() and request.user.is_staff:\n queryset = Post.objects.all()\n post = get_object_or_404(queryset, pk=kwargs[\"post_pk\"])\n else:\n raise Http404()\n else:\n queryset = Post.objects.current()\n queryset = queryset.filter(\n published__year = int(kwargs[\"year\"]),\n published__month = int(kwargs[\"month\"]),\n published__day = int(kwargs[\"day\"]),\n )\n post = get_object_or_404(queryset, slug=kwargs[\"slug\"])\n post.inc_views()\n \n return render_to_response(\"biblion/blog_post.html\", {\n \"post\": post,\n }, context_instance=RequestContext(request))\n\n\ndef serialize_request(request):\n data = {\n \"path\": request.path,\n \"META\": {\n \"QUERY_STRING\": request.META.get(\"QUERY_STRING\"),\n \"REMOTE_ADDR\": request.META.get(\"REMOTE_ADDR\"),\n }\n }\n for key in request.META:\n if key.startswith(\"HTTP\"):\n data[\"META\"][key] = request.META[key]\n return json.dumps(data)\n\n\ndef blog_feed(request, section=None):\n \n try:\n posts = Post.objects.section(section)\n except InvalidSection:\n raise Http404()\n \n if section is None:\n section = ALL_SECTION_NAME\n \n current_site = Site.objects.get_current()\n \n feed_title = \"%s Blog: %s\" % (current_site.name, section[0].upper() + section[1:])\n \n blog_url = \"http://%s%s\" % (current_site.domain, reverse(\"blog\"))\n \n url_name, kwargs = \"blog_feed\", {\"section\": section}\n feed_url = \"http://%s%s\" % (current_site.domain, reverse(url_name, kwargs=kwargs))\n \n if posts:\n feed_updated = posts[0].published\n else:\n feed_updated = datetime(2009, 8, 1, 0, 0, 0)\n \n # create a feed hit\n hit = FeedHit()\n hit.request_data = serialize_request(request)\n hit.save()\n \n atom = render_to_string(\"biblion/atom_feed.xml\", {\n \"feed_id\": feed_url,\n \"feed_title\": feed_title,\n \"blog_url\": blog_url,\n \"feed_url\": feed_url,\n \"feed_updated\": feed_updated,\n \"entries\": posts,\n \"current_site\": current_site,\n })\n return HttpResponse(atom, mimetype=\"application/atom+xml\")\n"},"license":{"kind":"string","value":"bsd-3-clause"},"hash":{"kind":"number","value":-8178591767137141000,"string":"-8,178,591,767,137,141,000"},"line_mean":{"kind":"number","value":29.1810344828,"string":"29.181034"},"line_max":{"kind":"number","value":86,"string":"86"},"alpha_frac":{"kind":"number","value":0.6235361325,"string":"0.623536"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.822052401746725,"string":"3.822052"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42169,"cells":{"repo_name":{"kind":"string","value":"jseabold/statsmodels"},"path":{"kind":"string","value":"statsmodels/sandbox/distributions/sppatch.py"},"copies":{"kind":"string","value":"5"},"size":{"kind":"string","value":"24020"},"content":{"kind":"string","value":"'''patching scipy to fit distributions and expect method\n\nThis adds new methods to estimate continuous distribution parameters with some\nfixed/frozen parameters. It also contains functions that calculate the expected\nvalue of a function for any continuous or discrete distribution\n\nIt temporarily also contains Bootstrap and Monte Carlo function for testing the\ndistribution fit, but these are neither general nor verified.\n\nAuthor: josef-pktd\nLicense: Simplified BSD\n'''\nfrom statsmodels.compat.python import lmap\nimport numpy as np\nfrom scipy import stats, optimize, integrate\n\n\n########## patching scipy\n\n#vonmises does not define finite bounds, because it is intended for circular\n#support which does not define a proper pdf on the real line\n\nstats.distributions.vonmises.a = -np.pi\nstats.distributions.vonmises.b = np.pi\n\n#the next 3 functions are for fit with some fixed parameters\n#As they are written, they do not work as functions, only as methods\n\ndef _fitstart(self, x):\n '''example method, method of moment estimator as starting values\n\n Parameters\n ----------\n x : ndarray\n data for which the parameters are estimated\n\n Returns\n -------\n est : tuple\n preliminary estimates used as starting value for fitting, not\n necessarily a consistent estimator\n\n Notes\n -----\n This needs to be written and attached to each individual distribution\n\n This example was written for the gamma distribution, but not verified\n with literature\n\n '''\n loc = np.min([x.min(),0])\n a = 4/stats.skew(x)**2\n scale = np.std(x) / np.sqrt(a)\n return (a, loc, scale)\n\ndef _fitstart_beta(self, x, fixed=None):\n '''method of moment estimator as starting values for beta distribution\n\n Parameters\n ----------\n x : ndarray\n data for which the parameters are estimated\n fixed : None or array_like\n sequence of numbers and np.nan to indicate fixed parameters and parameters\n to estimate\n\n Returns\n -------\n est : tuple\n preliminary estimates used as starting value for fitting, not\n necessarily a consistent estimator\n\n Notes\n -----\n This needs to be written and attached to each individual distribution\n\n References\n ----------\n for method of moment estimator for known loc and scale\n https://en.wikipedia.org/wiki/Beta_distribution#Parameter_estimation\n http://www.itl.nist.gov/div898/handbook/eda/section3/eda366h.htm\n NIST reference also includes reference to MLE in\n Johnson, Kotz, and Balakrishan, Volume II, pages 221-235\n\n '''\n #todo: separate out this part to be used for other compact support distributions\n # e.g. rdist, vonmises, and truncnorm\n # but this might not work because it might still be distribution specific\n a, b = x.min(), x.max()\n eps = (a-b)*0.01\n if fixed is None:\n #this part not checked with books\n loc = a - eps\n scale = (a - b) * (1 + 2*eps)\n else:\n if np.isnan(fixed[-2]):\n #estimate loc\n loc = a - eps\n else:\n loc = fixed[-2]\n if np.isnan(fixed[-1]):\n #estimate scale\n scale = (b + eps) - loc\n else:\n scale = fixed[-1]\n\n #method of moment for known loc scale:\n scale = float(scale)\n xtrans = (x - loc)/scale\n xm = xtrans.mean()\n xv = xtrans.var()\n tmp = (xm*(1-xm)/xv - 1)\n p = xm * tmp\n q = (1 - xm) * tmp\n\n return (p, q, loc, scale) #check return type and should fixed be returned ?\n\ndef _fitstart_poisson(self, x, fixed=None):\n '''maximum likelihood estimator as starting values for Poisson distribution\n\n Parameters\n ----------\n x : ndarray\n data for which the parameters are estimated\n fixed : None or array_like\n sequence of numbers and np.nan to indicate fixed parameters and parameters\n to estimate\n\n Returns\n -------\n est : tuple\n preliminary estimates used as starting value for fitting, not\n necessarily a consistent estimator\n\n Notes\n -----\n This needs to be written and attached to each individual distribution\n\n References\n ----------\n MLE :\n https://en.wikipedia.org/wiki/Poisson_distribution#Maximum_likelihood\n\n '''\n #todo: separate out this part to be used for other compact support distributions\n # e.g. rdist, vonmises, and truncnorm\n # but this might not work because it might still be distribution specific\n a = x.min()\n eps = 0 # is this robust ?\n if fixed is None:\n #this part not checked with books\n loc = a - eps\n else:\n if np.isnan(fixed[-1]):\n #estimate loc\n loc = a - eps\n else:\n loc = fixed[-1]\n\n #MLE for standard (unshifted, if loc=0) Poisson distribution\n\n xtrans = (x - loc)\n lambd = xtrans.mean()\n #second derivative d loglike/ dlambd Not used\n #dlldlambd = 1/lambd # check\n\n return (lambd, loc) #check return type and should fixed be returned ?\n\n\ndef nnlf_fr(self, thetash, x, frmask):\n # new frozen version\n # - sum (log pdf(x, theta),axis=0)\n # where theta are the parameters (including loc and scale)\n #\n try:\n if frmask is not None:\n theta = frmask.copy()\n theta[np.isnan(frmask)] = thetash\n else:\n theta = thetash\n loc = theta[-2]\n scale = theta[-1]\n args = tuple(theta[:-2])\n except IndexError:\n raise ValueError(\"Not enough input arguments.\")\n if not self._argcheck(*args) or scale <= 0:\n return np.inf\n x = np.array((x-loc) / scale)\n cond0 = (x <= self.a) | (x >= self.b)\n if (np.any(cond0)):\n return np.inf\n else:\n N = len(x)\n #raise ValueError\n return self._nnlf(x, *args) + N*np.log(scale)\n\ndef fit_fr(self, data, *args, **kwds):\n '''estimate distribution parameters by MLE taking some parameters as fixed\n\n Parameters\n ----------\n data : ndarray, 1d\n data for which the distribution parameters are estimated,\n args : list ? check\n starting values for optimization\n kwds :\n\n - 'frozen' : array_like\n values for frozen distribution parameters and, for elements with\n np.nan, the corresponding parameter will be estimated\n\n Returns\n -------\n argest : ndarray\n estimated parameters\n\n\n Examples\n --------\n generate random sample\n >>> np.random.seed(12345)\n >>> x = stats.gamma.rvs(2.5, loc=0, scale=1.2, size=200)\n\n estimate all parameters\n >>> stats.gamma.fit(x)\n array([ 2.0243194 , 0.20395655, 1.44411371])\n >>> stats.gamma.fit_fr(x, frozen=[np.nan, np.nan, np.nan])\n array([ 2.0243194 , 0.20395655, 1.44411371])\n\n keep loc fixed, estimate shape and scale parameters\n >>> stats.gamma.fit_fr(x, frozen=[np.nan, 0.0, np.nan])\n array([ 2.45603985, 1.27333105])\n\n keep loc and scale fixed, estimate shape parameter\n >>> stats.gamma.fit_fr(x, frozen=[np.nan, 0.0, 1.0])\n array([ 3.00048828])\n >>> stats.gamma.fit_fr(x, frozen=[np.nan, 0.0, 1.2])\n array([ 2.57792969])\n\n estimate only scale parameter for fixed shape and loc\n >>> stats.gamma.fit_fr(x, frozen=[2.5, 0.0, np.nan])\n array([ 1.25087891])\n\n Notes\n -----\n self is an instance of a distribution class. This can be attached to\n scipy.stats.distributions.rv_continuous\n\n *Todo*\n\n * check if docstring is correct\n * more input checking, args is list ? might also apply to current fit method\n\n '''\n loc0, scale0 = lmap(kwds.get, ['loc', 'scale'],[0.0, 1.0])\n Narg = len(args)\n\n if Narg == 0 and hasattr(self, '_fitstart'):\n x0 = self._fitstart(data)\n elif Narg > self.numargs:\n raise ValueError(\"Too many input arguments.\")\n else:\n args += (1.0,)*(self.numargs-Narg)\n # location and scale are at the end\n x0 = args + (loc0, scale0)\n\n if 'frozen' in kwds:\n frmask = np.array(kwds['frozen'])\n if len(frmask) != self.numargs+2:\n raise ValueError(\"Incorrect number of frozen arguments.\")\n else:\n # keep starting values for not frozen parameters\n for n in range(len(frmask)):\n # Troubleshooting ex_generic_mle_tdist\n if isinstance(frmask[n], np.ndarray) and frmask[n].size == 1:\n frmask[n] = frmask[n].item()\n\n # If there were array elements, then frmask will be object-dtype,\n # in which case np.isnan will raise TypeError\n frmask = frmask.astype(np.float64)\n x0 = np.array(x0)[np.isnan(frmask)]\n else:\n frmask = None\n\n #print(x0\n #print(frmask\n return optimize.fmin(self.nnlf_fr, x0,\n args=(np.ravel(data), frmask), disp=0)\n\n\n#The next two functions/methods calculate expected value of an arbitrary\n#function, however for the continuous functions intquad is use, which might\n#require continuouity or smoothness in the function.\n\n\n#TODO: add option for Monte Carlo integration\n\ndef expect(self, fn=None, args=(), loc=0, scale=1, lb=None, ub=None, conditional=False):\n '''calculate expected value of a function with respect to the distribution\n\n location and scale only tested on a few examples\n\n Parameters\n ----------\n all parameters are keyword parameters\n fn : function (default: identity mapping)\n Function for which integral is calculated. Takes only one argument.\n args : tuple\n argument (parameters) of the distribution\n lb, ub : numbers\n lower and upper bound for integration, default is set to the support\n of the distribution\n conditional : bool (False)\n If true then the integral is corrected by the conditional probability\n of the integration interval. The return value is the expectation\n of the function, conditional on being in the given interval.\n\n Returns\n -------\n expected value : float\n\n Notes\n -----\n This function has not been checked for it's behavior when the integral is\n not finite. The integration behavior is inherited from scipy.integrate.quad.\n\n '''\n if fn is None:\n def fun(x, *args):\n return x*self.pdf(x, loc=loc, scale=scale, *args)\n else:\n def fun(x, *args):\n return fn(x)*self.pdf(x, loc=loc, scale=scale, *args)\n if lb is None:\n lb = loc + self.a * scale #(self.a - loc)/(1.0*scale)\n if ub is None:\n ub = loc + self.b * scale #(self.b - loc)/(1.0*scale)\n if conditional:\n invfac = (self.sf(lb, loc=loc, scale=scale, *args)\n - self.sf(ub, loc=loc, scale=scale, *args))\n else:\n invfac = 1.0\n return integrate.quad(fun, lb, ub,\n args=args)[0]/invfac\n\n\ndef expect_v2(self, fn=None, args=(), loc=0, scale=1, lb=None, ub=None, conditional=False):\n '''calculate expected value of a function with respect to the distribution\n\n location and scale only tested on a few examples\n\n Parameters\n ----------\n all parameters are keyword parameters\n fn : function (default: identity mapping)\n Function for which integral is calculated. Takes only one argument.\n args : tuple\n argument (parameters) of the distribution\n lb, ub : numbers\n lower and upper bound for integration, default is set using\n quantiles of the distribution, see Notes\n conditional : bool (False)\n If true then the integral is corrected by the conditional probability\n of the integration interval. The return value is the expectation\n of the function, conditional on being in the given interval.\n\n Returns\n -------\n expected value : float\n\n Notes\n -----\n This function has not been checked for it's behavior when the integral is\n not finite. The integration behavior is inherited from scipy.integrate.quad.\n\n The default limits are lb = self.ppf(1e-9, *args), ub = self.ppf(1-1e-9, *args)\n\n For some heavy tailed distributions, 'alpha', 'cauchy', 'halfcauchy',\n 'levy', 'levy_l', and for 'ncf', the default limits are not set correctly\n even when the expectation of the function is finite. In this case, the\n integration limits, lb and ub, should be chosen by the user. For example,\n for the ncf distribution, ub=1000 works in the examples.\n\n There are also problems with numerical integration in some other cases,\n for example if the distribution is very concentrated and the default limits\n are too large.\n\n '''\n #changes: 20100809\n #correction and refactoring how loc and scale are handled\n #uses now _pdf\n #needs more testing for distribution with bound support, e.g. genpareto\n\n if fn is None:\n def fun(x, *args):\n return (loc + x*scale)*self._pdf(x, *args)\n else:\n def fun(x, *args):\n return fn(loc + x*scale)*self._pdf(x, *args)\n if lb is None:\n #lb = self.a\n try:\n lb = self.ppf(1e-9, *args) #1e-14 quad fails for pareto\n except ValueError:\n lb = self.a\n else:\n lb = max(self.a, (lb - loc)/(1.0*scale)) #transform to standardized\n if ub is None:\n #ub = self.b\n try:\n ub = self.ppf(1-1e-9, *args)\n except ValueError:\n ub = self.b\n else:\n ub = min(self.b, (ub - loc)/(1.0*scale))\n if conditional:\n invfac = self._sf(lb,*args) - self._sf(ub,*args)\n else:\n invfac = 1.0\n return integrate.quad(fun, lb, ub,\n args=args, limit=500)[0]/invfac\n\n### for discrete distributions\n\n#TODO: check that for a distribution with finite support the calculations are\n# done with one array summation (np.dot)\n\n#based on _drv2_moment(self, n, *args), but streamlined\ndef expect_discrete(self, fn=None, args=(), loc=0, lb=None, ub=None,\n conditional=False):\n '''calculate expected value of a function with respect to the distribution\n for discrete distribution\n\n Parameters\n ----------\n (self : distribution instance as defined in scipy stats)\n fn : function (default: identity mapping)\n Function for which integral is calculated. Takes only one argument.\n args : tuple\n argument (parameters) of the distribution\n optional keyword parameters\n lb, ub : numbers\n lower and upper bound for integration, default is set to the support\n of the distribution, lb and ub are inclusive (ul<=k<=ub)\n conditional : bool (False)\n If true then the expectation is corrected by the conditional\n probability of the integration interval. The return value is the\n expectation of the function, conditional on being in the given\n interval (k such that ul<=k<=ub).\n\n Returns\n -------\n expected value : float\n\n Notes\n -----\n * function is not vectorized\n * accuracy: uses self.moment_tol as stopping criterium\n for heavy tailed distribution e.g. zipf(4), accuracy for\n mean, variance in example is only 1e-5,\n increasing precision (moment_tol) makes zipf very slow\n * suppnmin=100 internal parameter for minimum number of points to evaluate\n could be added as keyword parameter, to evaluate functions with\n non-monotonic shapes, points include integers in (-suppnmin, suppnmin)\n * uses maxcount=1000 limits the number of points that are evaluated\n to break loop for infinite sums\n (a maximum of suppnmin+1000 positive plus suppnmin+1000 negative integers\n are evaluated)\n\n\n '''\n\n #moment_tol = 1e-12 # increase compared to self.moment_tol,\n # too slow for only small gain in precision for zipf\n\n #avoid endless loop with unbound integral, eg. var of zipf(2)\n maxcount = 1000\n suppnmin = 100 #minimum number of points to evaluate (+ and -)\n\n if fn is None:\n def fun(x):\n #loc and args from outer scope\n return (x+loc)*self._pmf(x, *args)\n else:\n def fun(x):\n #loc and args from outer scope\n return fn(x+loc)*self._pmf(x, *args)\n # used pmf because _pmf does not check support in randint\n # and there might be problems(?) with correct self.a, self.b at this stage\n # maybe not anymore, seems to work now with _pmf\n\n self._argcheck(*args) # (re)generate scalar self.a and self.b\n if lb is None:\n lb = (self.a)\n else:\n lb = lb - loc\n\n if ub is None:\n ub = (self.b)\n else:\n ub = ub - loc\n if conditional:\n invfac = self.sf(lb,*args) - self.sf(ub+1,*args)\n else:\n invfac = 1.0\n\n tot = 0.0\n low, upp = self._ppf(0.001, *args), self._ppf(0.999, *args)\n low = max(min(-suppnmin, low), lb)\n upp = min(max(suppnmin, upp), ub)\n supp = np.arange(low, upp+1, self.inc) #check limits\n #print('low, upp', low, upp\n tot = np.sum(fun(supp))\n diff = 1e100\n pos = upp + self.inc\n count = 0\n\n #handle cases with infinite support\n\n while (pos <= ub) and (diff > self.moment_tol) and count <= maxcount:\n diff = fun(pos)\n tot += diff\n pos += self.inc\n count += 1\n\n if self.a < 0: #handle case when self.a = -inf\n diff = 1e100\n pos = low - self.inc\n while (pos >= lb) and (diff > self.moment_tol) and count <= maxcount:\n diff = fun(pos)\n tot += diff\n pos -= self.inc\n count += 1\n if count > maxcount:\n # replace with proper warning\n print('sum did not converge')\n return tot/invfac\n\nstats.distributions.rv_continuous.fit_fr = fit_fr\nstats.distributions.rv_continuous.nnlf_fr = nnlf_fr\nstats.distributions.rv_continuous.expect = expect\nstats.distributions.rv_discrete.expect = expect_discrete\nstats.distributions.beta_gen._fitstart = _fitstart_beta #not tried out yet\nstats.distributions.poisson_gen._fitstart = _fitstart_poisson #not tried out yet\n\n########## end patching scipy\n\n\ndef distfitbootstrap(sample, distr, nrepl=100):\n '''run bootstrap for estimation of distribution parameters\n\n hard coded: only one shape parameter is allowed and estimated,\n loc=0 and scale=1 are fixed in the estimation\n\n Parameters\n ----------\n sample : ndarray\n original sample data for bootstrap\n distr : distribution instance with fit_fr method\n nrepl : int\n number of bootstrap replications\n\n Returns\n -------\n res : array (nrepl,)\n parameter estimates for all bootstrap replications\n\n '''\n nobs = len(sample)\n res = np.zeros(nrepl)\n for ii in range(nrepl):\n rvsind = np.random.randint(nobs, size=nobs)\n x = sample[rvsind]\n res[ii] = distr.fit_fr(x, frozen=[np.nan, 0.0, 1.0])\n return res\n\ndef distfitmc(sample, distr, nrepl=100, distkwds={}):\n '''run Monte Carlo for estimation of distribution parameters\n\n hard coded: only one shape parameter is allowed and estimated,\n loc=0 and scale=1 are fixed in the estimation\n\n Parameters\n ----------\n sample : ndarray\n original sample data, in Monte Carlo only used to get nobs,\n distr : distribution instance with fit_fr method\n nrepl : int\n number of Monte Carlo replications\n\n Returns\n -------\n res : array (nrepl,)\n parameter estimates for all Monte Carlo replications\n\n '''\n arg = distkwds.pop('arg')\n nobs = len(sample)\n res = np.zeros(nrepl)\n for ii in range(nrepl):\n x = distr.rvs(arg, size=nobs, **distkwds)\n res[ii] = distr.fit_fr(x, frozen=[np.nan, 0.0, 1.0])\n return res\n\n\ndef printresults(sample, arg, bres, kind='bootstrap'):\n '''calculate and print(Bootstrap or Monte Carlo result\n\n Parameters\n ----------\n sample : ndarray\n original sample data\n arg : float (for general case will be array)\n bres : ndarray\n parameter estimates from Bootstrap or Monte Carlo run\n kind : {'bootstrap', 'montecarlo'}\n output is printed for Mootstrap (default) or Monte Carlo\n\n Returns\n -------\n None, currently only printing\n\n Notes\n -----\n still a bit a mess because it is used for both Bootstrap and Monte Carlo\n\n made correction:\n reference point for bootstrap is estimated parameter\n\n not clear:\n I'm not doing any ddof adjustment in estimation of variance, do we\n need ddof>0 ?\n\n todo: return results and string instead of printing\n\n '''\n print('true parameter value')\n print(arg)\n print('MLE estimate of parameters using sample (nobs=%d)'% (nobs))\n argest = distr.fit_fr(sample, frozen=[np.nan, 0.0, 1.0])\n print(argest)\n if kind == 'bootstrap':\n #bootstrap compares to estimate from sample\n argorig = arg\n arg = argest\n\n print('%s distribution of parameter estimate (nrepl=%d)'% (kind, nrepl))\n print('mean = %f, bias=%f' % (bres.mean(0), bres.mean(0)-arg))\n print('median', np.median(bres, axis=0))\n print('var and std', bres.var(0), np.sqrt(bres.var(0)))\n bmse = ((bres - arg)**2).mean(0)\n print('mse, rmse', bmse, np.sqrt(bmse))\n bressorted = np.sort(bres)\n print('%s confidence interval (90%% coverage)' % kind)\n print(bressorted[np.floor(nrepl*0.05)], bressorted[np.floor(nrepl*0.95)])\n print('%s confidence interval (90%% coverage) normal approximation' % kind)\n print(stats.norm.ppf(0.05, loc=bres.mean(), scale=bres.std()),)\n print(stats.norm.isf(0.05, loc=bres.mean(), scale=bres.std()))\n print('Kolmogorov-Smirnov test for normality of %s distribution' % kind)\n print(' - estimated parameters, p-values not really correct')\n print(stats.kstest(bres, 'norm', (bres.mean(), bres.std())))\n\n\nif __name__ == '__main__':\n\n examplecases = ['largenumber', 'bootstrap', 'montecarlo'][:]\n\n if 'largenumber' in examplecases:\n\n print('\\nDistribution: vonmises')\n\n for nobs in [200]:#[20000, 1000, 100]:\n x = stats.vonmises.rvs(1.23, loc=0, scale=1, size=nobs)\n print('\\nnobs:', nobs)\n print('true parameter')\n print('1.23, loc=0, scale=1')\n print('unconstrained')\n print(stats.vonmises.fit(x))\n print(stats.vonmises.fit_fr(x, frozen=[np.nan, np.nan, np.nan]))\n print('with fixed loc and scale')\n print(stats.vonmises.fit_fr(x, frozen=[np.nan, 0.0, 1.0]))\n\n print('\\nDistribution: gamma')\n distr = stats.gamma\n arg, loc, scale = 2.5, 0., 20.\n\n for nobs in [200]:#[20000, 1000, 100]:\n x = distr.rvs(arg, loc=loc, scale=scale, size=nobs)\n print('\\nnobs:', nobs)\n print('true parameter')\n print('%f, loc=%f, scale=%f' % (arg, loc, scale))\n print('unconstrained')\n print(distr.fit(x))\n print(distr.fit_fr(x, frozen=[np.nan, np.nan, np.nan]))\n print('with fixed loc and scale')\n print(distr.fit_fr(x, frozen=[np.nan, 0.0, 1.0]))\n print('with fixed loc')\n print(distr.fit_fr(x, frozen=[np.nan, 0.0, np.nan]))\n\n\n ex = ['gamma', 'vonmises'][0]\n\n if ex == 'gamma':\n distr = stats.gamma\n arg, loc, scale = 2.5, 0., 1\n elif ex == 'vonmises':\n distr = stats.vonmises\n arg, loc, scale = 1.5, 0., 1\n else:\n raise ValueError('wrong example')\n\n nobs = 100\n nrepl = 1000\n\n sample = distr.rvs(arg, loc=loc, scale=scale, size=nobs)\n\n print('\\nDistribution:', distr)\n if 'bootstrap' in examplecases:\n print('\\nBootstrap')\n bres = distfitbootstrap(sample, distr, nrepl=nrepl )\n printresults(sample, arg, bres)\n\n if 'montecarlo' in examplecases:\n print('\\nMonteCarlo')\n mcres = distfitmc(sample, distr, nrepl=nrepl,\n distkwds=dict(arg=arg, loc=loc, scale=scale))\n printresults(sample, arg, mcres, kind='montecarlo')\n"},"license":{"kind":"string","value":"bsd-3-clause"},"hash":{"kind":"number","value":-3253983564312369000,"string":"-3,253,983,564,312,369,000"},"line_mean":{"kind":"number","value":32.0398899587,"string":"32.03989"},"line_max":{"kind":"number","value":91,"string":"91"},"alpha_frac":{"kind":"number","value":0.6175270608,"string":"0.617527"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.8278884462151392,"string":"3.827888"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42170,"cells":{"repo_name":{"kind":"string","value":"codefisher/mozbutton_sdk"},"path":{"kind":"string","value":"builder/restartless_button.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"28578"},"content":{"kind":"string","value":"import os\nimport re\nimport json\nimport codecs\nimport lxml.etree as ET\nfrom copy import deepcopy\nfrom collections import namedtuple, defaultdict\ntry:\n from PIL import Image\nexcept ImportError:\n pass\n\nfrom builder.ext_button import Button, Option, ChromeString, ChromeFile\n\ntry:\n basestring\nexcept NameError:\n basestring = str # py3\n\nKeys = namedtuple(\"Keys\", ['command', 'button'])\nExtraUI = namedtuple(\"ExtraUI\", [\"parent\", \"parent_id\", \"index\", \"code\", \"after\"])\n\nclass RestartlessButton(Button):\n\n def __init__(self, *args, **kwargs):\n super(RestartlessButton, self).__init__(*args, **kwargs)\n self._ui_ids = set()\n self._included_js_files = []\n self._bootstrap_globals = []\n self._bootstrap_startup = []\n self._bootstrap_shutdown = []\n\n for folder, button, files in self._info:\n if \"bootstrap\" in files:\n for file_name in os.listdir(os.path.join(folder, \"bootstrap\")):\n if file_name[0] != \".\":\n with open(os.path.join(folder, \"bootstrap\", file_name), \"r\") as js_fp:\n data = js_fp.read()\n if file_name == \"global.js\":\n self._bootstrap_globals.append(data)\n elif file_name == \"startup.js\":\n self._bootstrap_startup.append(data)\n elif file_name == \"shutdown.js\":\n self._bootstrap_shutdown.append(data)\n\n def get_files(self):\n for file_name, data in self.get_jsm_files().items():\n yield (file_name + \".jsm\", data)\n\n def locale_files(self, button_locales, *args, **kwargs):\n dtd_data = button_locales.get_dtd_data(self.get_locale_strings(),\n self, untranslated=False, format_type=\"properties\")\n for locale, data in dtd_data.items():\n yield locale, \"button_labels.properties\", data\n locales_inuse = dtd_data.keys()\n key_strings = button_locales.get_string_data(self.get_key_strings(),\n self, format_type=\"properties\")\n for locale, data in self.locale_file_filter(key_strings, locales_inuse):\n yield locale, \"keys.properties\", data\n for locale, file_name, data in super(RestartlessButton, self).locale_files(button_locales, locales_inuse):\n yield locale, file_name, data\n\n def jsm_keyboard_shortcuts(self, file_name):\n if not self._settings.get(\"use_keyboard_shortcuts\"):\n return\n for button in self._button_keys.keys():\n func = self._button_commands.get(file_name, {}).get(button)\n if func is not None:\n yield Keys(self._patch_call(func), button)\n\n def option_data(self):\n scripts = []\n if self._settings.get(\"use_keyboard_shortcuts\"):\n scripts.append(\"key-option.js\")\n with open(self.find_file(\"key-option.xul\"), \"r\") as key_option_file:\n key_option_template = key_option_file.read()\n for button in self._button_keys.keys():\n xul = self.format_string(key_option_template,\n button=button,\n menu_label=button + \".label\")\n applications = \" \".join(self._button_applications[button])\n self._button_options[button + \"-key-item\"].append(\n Option(\"tb-key-shortcut.option.title:lightning.png:\" + applications, xul))\n self._button_applications[\n button + \"-key-item\"] = self._button_applications[button]\n files, javascript = super(RestartlessButton, self).option_data()\n return files, javascript + scripts\n\n def get_pref_list(self):\n settings = super(RestartlessButton, self).get_pref_list()\n pref_root = self._settings.get(\"pref_root\")\n if self._settings.get('use_keyboard_shortcuts'):\n for button in self._button_keys.keys():\n settings.append((\"{}key-disabled.{}\".format(pref_root, button), 'false'))\n properties = self.pref_locale_file(\"'chrome://{chrome_name}/locale/{prefix}keys.properties'\")\n settings.append((\"{}key.{}\".format(pref_root, button), properties))\n settings.append((\"{}modifier.{}\".format(pref_root, button), properties))\n return settings\n\n def get_js_files(self):\n js_files = super(RestartlessButton, self).get_js_files()\n if self._settings.get(\"use_keyboard_shortcuts\"):\n with open(self.find_file(\"key-option.js\")) as key_option_fp:\n js_files[\"key-option\"] = self.string_subs(key_option_fp.read())\n self._included_js_files = js_files.keys()\n return js_files\n\n def get_chrome_strings(self):\n for chrome_string in super(RestartlessButton, self).get_chrome_strings():\n yield chrome_string\n yield ChromeString(file_name='bootstrap.js', data=self.create_bootstrap())\n defaults = self.get_defaults()\n if defaults:\n yield ChromeString(file_name=os.path.join(\"chrome\", \"content\", \"defaultprefs.js\"), data=defaults)\n\n def get_chrome_files(self):\n for chrome_file in super(RestartlessButton, self).get_chrome_files():\n yield chrome_file\n yield ChromeFile(file_name=os.path.join(\"chrome\", \"content\", \"customizable.jsm\"), path=self.find_file('customizable.jsm'))\n\n def create_bootstrap(self):\n chrome_name = self._settings.get(\"chrome_name\")\n loaders = []\n resource = \"\"\n if self.resource_files:\n resource = \"createResource('{0}', 'chrome://{0}/content/resources/');\".format(chrome_name)\n window_modules = defaultdict(list)\n for file_name in self._button_files:\n for overlay in self._settings.get(\"files_to_window\").get(file_name, ()):\n window_modules[overlay].append(file_name)\n\n for overlay, modules in window_modules.items():\n mods = \"\\n\\t\\t\".join([\"modules.push('chrome://{0}/content/{1}.jsm');\".format(chrome_name, file_name) for file_name in modules])\n loaders.append(\"(uri == '{0}') {{\\n\\t\\t{1}\\n\\t}}\".format(overlay, mods))\n if self._settings.get(\"show_updated_prompt\"):\n install_template = self.env.get_template('bootstrap.js')\n install = install_template.render(**self._settings)\n else:\n install = \"\"\n template = self.env.get_template('bootstrap.js')\n return template.render(\n resource=resource, install=install,\n globals=self.string_subs(\"\\n\".join(self._bootstrap_globals)),\n startup=self.string_subs(\"\\n\".join(self._bootstrap_startup)),\n shutdown=self.string_subs(\"\\n\".join(self._bootstrap_shutdown)),\n loaders = \"if\" + \" else if\".join(loaders),\n **self._settings)\n\n def _jsm_create_menu(self, file_name, buttons):\n if not self._settings.get('menuitems'):\n return ''\n statements = []\n data = self.create_menu_dom(file_name, buttons)\n in_submenu = [menuitem for menuitem in data if menuitem.parent_id is None]\n in_menu = [menuitem for menuitem in data if menuitem.parent_id is not None]\n num = 0\n template = self.env.get_template('menu.js')\n if in_submenu:\n menu_id, menu_label, locations = self._settings.get(\"menu_meta\")\n if isinstance(locations, basestring):\n locations = [locations]\n for i, location in enumerate(locations):\n menu_id_num = \"{0}_{1}\".format(menu_id, i) if i else menu_id\n meta = self._settings.get(\"file_to_menu\").get(location, {}).get(file_name)\n if meta:\n menu_name, insert_after = meta\n statements.append(template.render(**{\n \"menu_name\": menu_name,\n \"menu_id\": menu_id_num,\n \"label\": menu_label,\n \"class\": \"menu-iconic\",\n \"menu_label\": menu_label,\n \"insert_after\": insert_after,\n \"menuitems_sorted\": self._settings.get(\"menuitems_sorted\")\n }))\n num += 3\n for item, _, _ in in_submenu:\n item_statements, count, _ = self._create_dom(\n item, top=\"menupopup_2\", count=num, doc=\"document\")\n num = count + 1\n statements.extend(item_statements)\n for item, menu_name, insert_after in in_menu:\n statements.append(\"var menupopup_{0} = document.getElementById('{1}');\".format(num, menu_name))\n var_name = \"menupopup_%s\" % num\n num += 1\n item.attrib[\"insertafter\"] = insert_after\n item_statements, count, _ = self._create_dom(item, top=var_name, count=num)\n num = count + 1\n statements.extend(item_statements)\n return \"\\n\\t\".join(statements)\n \n def _dom_string_lookup(self, value):\n result = []\n items = re.findall(r'&.+?;|[^&;]+', value)\n for item in items:\n if item == \"&brandShortName;\":\n result.append(\"Cc['@mozilla.org/xre/app-info;1'].createInstance(Ci.nsIXULAppInfo).name\")\n elif item[0] == '&' and item[-1] == ';':\n result.append(\"buttonStrings.get('%s')\" % item[1:-1])\n else:\n result.append(\"'%s'\" % item)\n return ' + '.join(result)\n\n def _create_dom(self, root, top=None, count=0, doc='document', child_parent=None, rename=None, append_children=True):\n num = count\n if rename == None:\n rename = {}\n children = []\n statements = [\n \"var %s_%s = %s.createElement('%s');\" % (root.tag, num, doc, rename.get(root.tag, root.tag)),\n ]\n javascript_object = self._settings.get(\"javascript_object\")\n for key, value in sorted(root.attrib.items(), key=self._attr_key):\n if key == 'id':\n statements.append(\"%s_%s.id = '%s';\" % (root.tag, num, value))\n elif key in ('label', 'tooltiptext') or (root.tag == 'key' and key in ('key', 'keycode', 'modifiers')):\n statements.append(\"%s_%s.setAttribute('%s', %s);\" % ((root.tag, num, key, self._dom_string_lookup(value))))\n elif key == \"class\":\n for val in value.split():\n statements.append('%s_%s.classList.add(\"%s\");' % (root.tag, num, val))\n elif key[0:2] == 'on':\n if key == 'oncommand' and root.tag == 'key':\n # we do this because key elements without a oncommand are optimized away\n # but we can't call our function, because that might not exist \n # in the window scope, so the event listener has to be used\n statements.append(\"%s_%s.setAttribute('oncommand', 'void(0);');\" % (root.tag, num))\n statements.append(\"%s_%s.addEventListener('%s', function(event) {\\n\\t\\t\\t\\t%s\\n\\t\\t\\t}, false);\" % (root.tag, num, key[2:], self._patch_call(value)))\n elif key == \"insertafter\":\n pass\n elif key == \"showamenu\":\n statements.append(\"{}_{}.addEventListener('DOMMenuItemActive', {}.menuLoaderEvent, false);\".format(root.tag, num, javascript_object))\n statements.append(\"%s_%s._handelMenuLoaders = true;\" % (root.tag, num))\n statements.append(\"%s_%s.setAttribute('%s', '%s');\" % ((root.tag, num, key, value)))\n elif key == \"toolbarname\":\n # this is just for our custom toolbars which are named \"Toolbar Buttons 1\" and the like\n name, sep, other = value.partition(' ')\n other = \" + '%s%s'\" % (sep, other) if sep else \"\"\n value = \"buttonStrings.get('%s')%s\" % (name, other)\n statements.append(\"%s_%s.setAttribute('%s', %s);\" % ((root.tag, num, key, value)))\n elif key == \"type\" and value == \"menu-button\" and 'id' in root.attrib:\n statements.append('''if(extensionPrefs.getPrefType('menupopup.hide.{0}') == extensionPrefs.PREF_INVALID || !extensionPrefs.getBoolPref('menupopup.hide.{0}')) {{\\n\\t\\t\\t\\t{1}_{2}.setAttribute(\"{3}\", \"{4}\");\\n\\t\\t\\t}}'''.format(root.attrib['id'], root.tag, num, key, value))\n else:\n statements.append('%s_%s.setAttribute(\"%s\", \"%s\");' % ((root.tag, num, key, value)))\n for node in root:\n sub_nodes, count, _ = self._create_dom(node, '%s_%s' % (root.tag, num), count+1, doc=doc, rename=rename, child_parent=(child_parent if top == None else None))\n if append_children:\n statements.extend(sub_nodes)\n else:\n children = sub_nodes\n if not top:\n statements.append('return %s_%s;' % (root.tag, num))\n else:\n if \"insertafter\" in root.attrib:\n statements.append(\"%s.insertBefore(%s_%s, %s.getElementById('%s').nextSibling);\" % (top, root.tag, num, doc, root.attrib.get(\"insertafter\")))\n else:\n statements.append('%s.appendChild(%s_%s);' % (top if not child_parent else child_parent, root.tag, num))\n return statements, count, children\n \n def _attr_key(self, attr):\n order = ('id', 'defaultarea', 'type', 'label', 'tooltiptext', 'command', 'onclick', 'oncommand')\n if attr[0].lower() in order:\n return order.index(attr[0].lower())\n return 100\n \n def _create_dom_button(self, button_id, root, file_name, count, toolbar_ids):\n add_to_main_toolbar = self._settings.get(\"add_to_main_toolbar\")\n if 'viewid' in root.attrib:\n self._ui_ids.add(root.attrib[\"viewid\"])\n statements, _, children = self._create_dom(root, child_parent=\"popupset\", append_children=False)\n children[0] = \"\"\"var popupset = document.getElementById('PanelUI-multiView');\n\t\t\t\tif(popupset) {\n\t\t\t\t\tvar menupopup_1 = document.createElement('panelview');\n\t\t\t\t} else {\n\t\t\t\t\tvar menupopup_1 = document.createElement('menupopup');\n\t\t\t\t\tpopupset = document.documentElement;\n\t\t\t\t}\"\"\"\n data = {\n \"type\": \"'view'\",\n \"onBeforeCreated\": 'function (document) {\\n\\t\\t\\t\\tvar window = document.defaultView;\\n\\t\\t\\t\\t%s\\n\\t\\t\\t}' % \"\\n\\t\\t\\t\\t\".join(children),\n }\n elif 'usepanelview' in root.attrib:\n self._ui_ids.add(\"{0}-panel-view\".format(root.attrib[\"id\"]))\n root.attrib[\"onclick\"] = \"\"\"if(event.target != event.currentTarget || ('button' in event && event.button != 0)) {{\n\t\t\t\t\treturn;\n\t\t\t\t}}\n\t\t\t\tvar item = event.target;\n\t\t\t\tif(item.nodeName == 'key') {{\n\t\t\t\t\titem = document.getElementById('{0}');\n\t\t\t\t}}\n\t\t\t\tif(item.getAttribute('cui-areatype') == 'menu-panel') {{\n\t\t\t\t\tvar win = item.ownerDocument.defaultView;\n\t\t\t\t\tevent.preventDefault();\n\t\t\t\t\tevent.stopPropagation();\n\t\t\t\t\titem.ownerDocument.getElementById('{0}-panel-view').ownerButton = item;\n\t\t\t\t\twin.PanelUI.showSubView('{0}-panel-view', item, CustomizableUI.AREA_PANEL);\n\t\t\t\t}}\"\"\".format(root.attrib[\"id\"])\n if 'type' not in root.attrib:\n popup_opener = \"\"\" else {\n\t\t\t\t\titem.firstChild.openPopup(item, \"after_start\");\n\t\t\t\t}\"\"\"\n if 'oncommand' not in root.attrib:\n root.attrib[\"oncommand\"] = root.attrib[\"onclick\"] + popup_opener\n else:\n root.attrib[\"onclick\"] += popup_opener\n statements, _, _ = self._create_dom(root)\n root_clone = deepcopy(root)\n popup = root_clone[0]\n if root.attrib['usepanelview'] == 'button-menu':\n del root_clone.attrib[\"type\"]\n popup.insert(0, ET.Element(\"menuseparator\"))\n popup.insert(0, ET.Element(\"menuitem\", root_clone.attrib))\n for node in popup:\n node.attrib['class'] = 'subviewbutton'\n if 'onpopupshowing' in popup.attrib:\n popup.attrib['onViewShowing'] = popup.attrib['onpopupshowing']\n del popup.attrib['onpopupshowing']\n if 'onpopuphiding' in popup.attrib:\n popup.attrib['onViewHiding'] = popup.attrib['onpopuphiding']\n del popup.attrib['onpopuphiding']\n _, _, children = self._create_dom(root_clone, child_parent=\"popupset\", rename={'menuitem': 'toolbarbutton'}, append_children=False)\n children.pop(0)\n data = {\n \"type\": \"'custom'\",\n \"onBuild\": '''function (document) {\n\t\t\t\tvar window = document.defaultView;\n\t\t\t\tvar popupset = document.getElementById('PanelUI-multiView');\n\t\t\t\tif(popupset) {\n\t\t\t\t\tvar menupopup_1 = document.createElement('panelview');\n\t\t\t\t\t%s\n\t\t\t\t\tmenupopup_1.id = \"%s-panel-view\";\n\t\t\t\t}\n\t\t\t\t%s\n\t\t}''' % (\"\\n\\t\\t\\t\\t\\t\".join(children), root.attrib['id'], \"\\n\\t\\t\\t\\t\".join(statements))\n }\n else:\n statements, _, _ = self._create_dom(root)\n data = {\n \"type\": \"'custom'\",\n \"onBuild\": 'function (document) {\\n\\t\\t\\t\\tvar window = document.defaultView;\\n\\t\\t\\t\\t%s\\n\\t\\t\\t}' % \"\\n\\t\\t\\t\\t\".join(statements)\n }\n self._apply_toolbox(file_name, data)\n toolbar_max_count = self._settings.get(\"buttons_per_toolbar\")\n if add_to_main_toolbar and button_id in add_to_main_toolbar:\n data['defaultArea'] = \"'%s'\" % self._settings.get('file_to_main_toolbar').get(file_name)\n elif self._settings.get(\"put_button_on_toolbar\"):\n toolbar_index = count // toolbar_max_count\n if len(toolbar_ids) > toolbar_index:\n data['defaultArea'] = \"'%s'\" % toolbar_ids[toolbar_index]\n for key, value in root.attrib.items():\n if key in ('label', 'tooltiptext'):\n data[key] = self._dom_string_lookup(value)\n elif key == \"id\":\n data[key] = \"'%s'\" % value\n elif key == 'oncommand':\n self._button_commands[file_name][button_id] = value\n elif key == 'viewid':\n data[\"viewId\"] = \"'%s'\" % value\n elif key == 'onviewshowing':\n data[\"onViewShowing\"] = \"function(event){\\n\\t\\t\\t\\t%s\\n\\t\\t\\t}\" % self._patch_call(value)\n elif key == 'onviewhideing':\n data[\"onViewHiding\"] = \"function(event){\\n\\t\\t\\t\\t%s\\n\\t\\t\\t}\" % self._patch_call(value)\n for js_file in self._get_js_file_list(file_name):\n if self._button_js_setup.get(js_file, {}).get(button_id):\n data[\"onCreated\"] = \"function(aNode){\\n\\t\\t\\tvar document = aNode.ownerDocument;\\n\\t\\t\\t%s\\n\\t\\t}\" % self._button_js_setup[js_file][button_id]\n items = sorted(data.items(), key=self._attr_key)\n return \"CustomizableUI.createWidget({\\n\\t\\t\\t%s\\n\\t\\t});\" % \",\\n\\t\\t\\t\".join(\"%s: %s\" % (key, value) for key, value in items)\n\n def _apply_toolbox(self, file_name, data):\n toolbox_info = self._settings.get(\"file_to_toolbar_box2\").get(file_name)\n if toolbox_info:\n window_file, toolbox_id = toolbox_info\n data[\"toolbox\"] = \"'%s'\" % toolbox_id\n if window_file:\n data[\"window\"] = \"'%s'\" % window_file\n\n\n def _patch_call(self, value):\n data = []\n if re.search(r'\\bthis\\b', value):\n value = re.sub(r'\\bthis\\b', 'aThis', value)\n data.append(\"var aThis = event.currentTarget;\")\n if re.search(r'\\bdocument\\b', value):\n data.append(\"var document = event.target.ownerDocument;\")\n if re.search(r'\\bwindow\\b', value):\n data.append(\"var window = event.target.ownerDocument.defaultView;\")\n data.append(value)\n return \"\\n\\t\\t\\t\\t\".join(data)\n\n def _create_jsm_button(self, button_id, root, file_name, count, toolbar_ids):\n toolbar_max_count = self._settings.get(\"buttons_per_toolbar\")\n add_to_main_toolbar = self._settings.get(\"add_to_main_toolbar\")\n data = {}\n attr = root.attrib\n self._apply_toolbox(file_name, data)\n if add_to_main_toolbar and button_id in add_to_main_toolbar:\n data['defaultArea'] = \"'%s'\" % self._settings.get('file_to_main_toolbar').get(file_name)\n elif self._settings.get(\"put_button_on_toolbar\"):\n toolbar_index = count // toolbar_max_count\n if len(toolbar_ids) > toolbar_index:\n data['defaultArea'] = \"'%s'\" % toolbar_ids[toolbar_index]\n for key, value in attr.items():\n if key in ('label', 'tooltiptext'):\n data[key] = self._dom_string_lookup(value)\n elif key == \"id\":\n data[key] = \"'%s'\" % value\n elif key in ('onclick', 'oncommand'):\n if key == 'oncommand':\n self._button_commands[file_name][button_id] = value\n key = 'onCommand' if key == 'oncommand' else 'onClick'\n data[key] = \"function(event) {\\n\\t\\t\\t\\t%s\\n\\t\\t\\t}\" % self._patch_call(value)\n for js_file in self._get_js_file_list(file_name):\n if self._button_js_setup.get(js_file, {}).get(button_id):\n data[\"onCreated\"] = \"function(aNode) {\\n\\t\\t\\t\\tvar document = aNode.ownerDocument;\\n\\t\\t\\t\\t%s\\n\\t\\t\\t}\" % self._button_js_setup[js_file][button_id]\n items = sorted(data.items(), key=self._attr_key)\n result = \"CustomizableUI.createWidget({\\n\\t\\t\\t%s\\n\\t\\t});\" % \",\\n\\t\\t\\t\".join(\"%s: %s\" % (key, value) for (key, value) in items)\n return result\n\n def get_jsm_files(self):\n result = {}\n simple_attrs = {'label', 'tooltiptext', 'id', 'oncommand', 'onclick', 'key', 'class'}\n button_hash, toolbar_template = self._get_toolbar_info()\n template = self.env.get_template('button.jsm')\n javascript_object = self._settings.get(\"javascript_object\")\n\n for file_name, values in self._button_xul.items():\n jsm_buttons = []\n js_includes = [js_file for js_file in self._get_js_file_list(file_name)\n if js_file != \"loader\" and js_file in self._included_js_files]\n toolbars, toolbar_ids = self._create_jsm_toolbar(button_hash, toolbar_template, file_name, values)\n count = 0\n modules = set()\n for button_id, xul in values.items():\n root = ET.fromstring(xul.replace('&', '&amp;'))\n modules.update(self._modules[button_id])\n attr = root.attrib\n if not len(root) and not set(attr.keys()).difference(simple_attrs) and (not \"class\" in attr or attr[\"class\"] == \"toolbarbutton-1 chromeclass-toolbar-additional\"):\n jsm_buttons.append(self._create_jsm_button(button_id, root, file_name, count, toolbar_ids))\n else:\n jsm_buttons.append(self._create_dom_button(button_id, root, file_name, count, toolbar_ids))\n count += 1\n default_mods = {\n \"resource://gre/modules/Services.jsm\",\n \"resource:///modules/CustomizableUI.jsm\",\n \"resource://services-common/stringbundle.js\"\n }\n modules_import = \"\\n\".join(\"try { Cu.import('%s'); } catch(e) {}\" % mod for mod in modules if mod and mod not in default_mods)\n if self._settings.get(\"menu_meta\"):\n menu_id, menu_label, _ = self._settings.get(\"menu_meta\")\n else:\n menu_id, menu_label = \"\", \"\"\n end = set()\n menu = self._jsm_create_menu(file_name, values)\n for js_file in set(self._get_js_file_list(file_name) + [file_name]):\n if self._button_js_setup.get(js_file, {}):\n end.update(self._button_js_setup[js_file].values())\n if (self._settings.get(\"menuitems\") and menu) or self._settings.get('location_placement'):\n end.add(javascript_object + \".setUpMenuShower(document);\")\n extra_ui = self.create_extra_ui(file_name, values)\n result[file_name] = template.render(\n modules=modules_import,\n locale_file_prefix=self._settings.get(\"locale_file_prefix\"),\n scripts=js_includes,\n button_ids=json.dumps(list(values.keys())),\n toolbar_ids=json.dumps(toolbar_ids),\n toolbars=toolbars,\n menu_id=menu_id,\n ui_ids=json.dumps(list(self._ui_ids)),\n toolbox=self._settings.get(\"file_to_toolbar_box\").get(file_name, ('', ''))[1],\n menu=menu,\n keys=list(self.jsm_keyboard_shortcuts(file_name)),\n end=\"\\n\\t\".join(end),\n buttons=jsm_buttons,\n extra_ui=extra_ui,\n javascript_object=self._settings.get(\"javascript_object\"),\n pref_root=self._settings.get(\"pref_root\"),\n chrome_name=self._settings.get(\"chrome_name\")\n )\n return result\n\n def create_extra_ui(self, file_name, values):\n location = self._settings.get(\"location_placement\")\n result = []\n if location and file_name in self._settings.get(\"file_to_location\", {}).get(location):\n for index, (button_id, xul) in enumerate(values.items()):\n parent, parent_id, after, attrib = self._settings.get(\"file_to_location\").get(location).get(file_name)\n root = ET.fromstring(xul.replace('&', '&amp;'))\n root.attrib[\"insertafter\"] = after\n root.attrib[\"id\"] += \"-extra-ui\"\n self._ui_ids.add(root.attrib[\"id\"])\n if attrib:\n for name, value in attrib.items():\n if value is None:\n del root.attrib[name]\n else:\n root.attrib[name] = value\n parent_var = \"{}_{}\".format(parent, index)\n statements, _, _ = self._create_dom(root, top=parent_var)\n result.append(ExtraUI(parent, parent_id, index, \"\\n\\t\\t\".join(statements), after))\n return result\n \n def _create_jsm_toolbar(self, button_hash, toolbar_template, file_name, values):\n toolbar_ids = []\n toolbars = []\n if file_name in self._settings.get(\"extra_toolbars_disabled\"):\n return '', []\n count = 0\n max_count = self._settings.get(\"buttons_per_toolbar\")\n buttons = list(values.keys())\n for box_setting, include_setting in [(\"file_to_toolbar_box\", \"include_toolbars\"),\n (\"file_to_bottom_box\", \"include_satusbars\")]:\n toolbar_node, toolbar_box = self._settings.get(box_setting).get(file_name, ('', ''))\n data = {\n \"defaultset\": \"\",\n \"persist\": \"collapsed,hidden\",\n \"context\": \"toolbar-context-menu\",\n \"class\": \"toolbar-buttons-toolbar chromeclass-toolbar\",\n \"mode\": \"icons\",\n \"iconsize\": \"small\",\n \"customizable\": \"true\",\n }\n if self._settings.get(include_setting) and toolbar_box:\n number = self.toolbar_count(include_setting, values, max_count)\n for i in range(number):\n if self._settings.get(\"put_button_on_toolbar\"):\n data[\"defaultset\"] = \",\".join(buttons[i * max_count:(i + 1) * max_count])\n button_hash.update(bytes(i))\n hash = button_hash.hexdigest()[:6]\n label_number = \"\" if (number + count) == 1 else \" %s\" % (i + count + 1)\n toolbar_ids.append(\"tb-toolbar-%s\" % hash)\n if include_setting != \"include_toolbars\":\n data[\"toolboxid\"] = toolbar_box\n data[\"id\"] = \"tb-toolbar-%s\" % hash\n toolbarname = self._dom_string_lookup(\"&tb-toolbar-buttons-toggle-toolbar.name;%s\" % label_number)\n values[\"tb-toolbar-buttons-toggle-toolbar-%s\" % hash] = toolbar_template.replace(\"{{hash}}\", hash).replace(\"{{ number }}\", label_number)\n toolbars.append(\"\"\"createToolbar(document, '%s', %s, %s)\"\"\" % (toolbar_box, json.dumps(data), toolbarname))\n count += number\n return \"\\n\\t\\t\".join(toolbars), toolbar_ids\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":4315887822240294400,"string":"4,315,887,822,240,294,400"},"line_mean":{"kind":"number","value":52.217877095,"string":"52.217877"},"line_max":{"kind":"number","value":288,"string":"288"},"alpha_frac":{"kind":"number","value":0.5496885716,"string":"0.549689"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.7622432859399684,"string":"3.762243"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42171,"cells":{"repo_name":{"kind":"string","value":"soybean217/lora-python"},"path":{"kind":"string","value":"UServer/admin_server/admin_http_api/api/api_group.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"3730"},"content":{"kind":"string","value":"import json\nfrom wtforms import ValidationError\n\nfrom userver.object.application import Application\nfrom . import api, root\nfrom flask import request, Response\nfrom userver.object.group import Group\nfrom binascii import hexlify\nfrom utils.errors import KeyDuplicateError, PatchError\nfrom .decorators import group_filter_valid, group_exists\nfrom .forms import get_formdata_from_json_or_form\nfrom .forms.form_group import AddGroupForm, PatchGroup, device_operate\nfrom ..http_auth import auth\n\n\n@api.route(root + 'groups', methods=['GET'])\n@auth.auth_required\n@group_filter_valid\ndef group_list(user=None, app=None):\n if request.method == 'GET':\n if app is not None:\n groups = Group.objects.all(app_eui=app.app_eui)\n elif user is not None:\n groups = []\n apps = Application.query.filter_by(user_id=user.id)\n for app in apps:\n groups += Group.objects.all(app.app_eui)\n else:\n groups = Group.objects.all()\n groups = [group.obj_to_dict() for group in groups]\n groups_json = json.dumps(groups)\n return Response(status=200, response=groups_json)\n # elif request.method == 'POST':\n # formdata = get_formdata_from_json_or_form(request)\n # add_group = AddGroupForm(formdata)\n # try:\n # if add_group.validate():\n # if len(add_group['appskey'].data) != 0:\n # group = Group(add_group['app_eui'].data, add_group['name'].data, add_group['addr'].data, add_group['nwkskey'].data, appskey=add_group['appskey'].data)\n # else:\n # group = Group(add_group['app_eui'].data, add_group['name'].data, add_group['addr'].data, add_group['nwkskey'].data)\n # group.save()\n # return Response(status=201, response=json.dumps(group.obj_to_dict()))\n # else:\n # return Response(status=406, response=json.dumps({'errors': add_group.errors,\n # 'succeed': False}))\n # except KeyDuplicateError as error:\n # return Response(status=403, response=json.dumps({\"error\": str(error),\n # \"succeed\": False}))\n\n\n@api.route(root + 'groups/', methods=['GET'])\n@auth.auth_required\n@group_exists\ndef group_index(group):\n if request.method == 'GET':\n group_json = json.dumps(group.obj_to_dict())\n return group_json, 200\n # elif request.method == 'PATCH':\n # try:\n # formdata = get_formdata_from_json_or_form(request)\n # PatchGroup.patch(group, formdata)\n # return Response(status=200, response=json.dumps(group.obj_to_dict()))\n # except (AssertionError, ValidationError, PatchError) as e:\n # return json.dumps({\"error\": str(e)}), 406\n # elif request.method == 'POST':\n # POST Down Msg\n # pass\n # elif request.method == 'DELETE':\n # try:\n # group.delete()\n # return json.dumps({'errors': \"Group: %s deleted.\" % hexlify(group.id).decode(),\n # 'succeed': False}), 200\n # except Exception as e:\n # return json.dumps({'errors': \"Fail to delete group: %s.\\n%s\" % (hexlify(group.id).decode(), str(e)),\n # 'succeed': False}), 400\n # elif request.method == 'POST':\n # formdata = get_formdata_from_json_or_form(request)\n # error = device_operate(group, formdata)\n # if error is None or len(error) == 0:\n # return json.dumps({'success': True}), 200\n # else:\n # return json.dumps({'error': str(error)}), 406\n #\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":-6227232879890345000,"string":"-6,227,232,879,890,345,000"},"line_mean":{"kind":"number","value":43.4047619048,"string":"43.404762"},"line_max":{"kind":"number","value":172,"string":"172"},"alpha_frac":{"kind":"number","value":0.5742627346,"string":"0.574263"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.7188434695912265,"string":"3.718843"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42172,"cells":{"repo_name":{"kind":"string","value":"spino327/sdr_testbed"},"path":{"kind":"string","value":"DistributedTestbed/SlaveRX.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"6293"},"content":{"kind":"string","value":"'''\nCopyright (c) 2011, Universidad Industrial de Santander, Colombia\nUniversity of Delaware\nAll rights reserved.\n\n@author: Sergio Pino\n@author: Henry Arguello\nWebsite: http://www.eecis.udel.edu/\nemails : sergiop@udel.edu - henarfu@udel.edu\nDate : Feb, 2011\n'''\n\nimport socket\nimport time\nimport sys\nfrom receiver.RXApp import RXApp\nfrom util.PropertyReader import readProperties\nfrom util import Utils\n\nclass SlaveRX(object):\n '''\n SlaveRX is responsible of control the RX USRP node.\n '''\n\n def __init__(self, host, port, path):\n '''\n Constructor\n @param host: refers to the local host address\n @param port: port for the server to listen\n @param path: File system path where the data will be stored\n '''\n \n # server\n self.server = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n self.server.bind((host, port))\n self.server.listen(1)\n self.path = path\n self.app = None\n \n def setRXProperties(self, lo_off, fc, dec, gain, sync):\n '''\n Set the USRP RX properties\n \n @param lo_off: local oscillator offset (int)\n @param fc: Center frequency (float)\n @param dec: Decimation factor (int)\n @param gain: Gain of the receiver in dB (int)\n @param sync: True if the Hardware will use the GPSDO (boolean) \n '''\n\n self.lo_off = lo_off\n self.fc = fc\n self.dec = dec\n self.gain = gain\n self.sync = sync\n \n def launch(self):\n '''\n calls startup\n '''\n print(\"i: launch SlaveRX\")\n \n while True:\n \n sc, addr = self.server.accept()\n sc.settimeout(10*60)\n print(\"\\n\\ni: SlaveRX Connection from \" + str(addr) + \", time \" + time.strftime(\"%d-%m-%y/%H:%M:%S\"))\n \n tic = time.time()\n \n try:\n self.__startup__(sc, addr)\n \n except Exception, e:\n print(\"e: \" + str(e))\n \n sc.close()\n print(\"i: SlaveRX Connection closed, duration: \" + str(time.time() - tic) + \" [seg]\\n\\n\")\n \n print(\"i: SlaveRX end launch\")\n \n def record(self, prefix, at, signame):\n \"\"\"\n @param prefix: prefix path folder where the signals are stored, e.g. /home/folder/\n @param at: attenuation factor \n @param signame: filename of the signal \n \n Start recording\n \"\"\"\n # creating the folder\n folder = self.path + prefix\n folder = folder if (folder.endswith(\"/\")) else folder + \"/\"\n Utils.ensure_dir(folder)\n \n # signal file\n filename = folder + signame + \"_at\" + str(at) +\"_G\" + str(self.gain) + \".dat\"\n \n print(\"i: record filename = \" + filename)\n \n self.app = RXApp(self.fc, self.dec, self.gain, \"addr=192.168.10.2\", self.sync, filename, self.lo_off)\n self.app.launch()\n \n def __startup__(self, sc, addr):\n '''\n Responsible for starting the application; for creating and showing\n the initial GUI.\n '''\n print(\"i: startup\")\n \n msg = sc.recv(1024)\n if msg == \"start\":\n sc.send(\"ok\")\n print(\"i: start ok\")\n \n msg = sc.recv(1024)\n print(\"i: msg = \" + msg)\n\n while msg != \"finish\":\n \n tic = time.time()\n \n if msg.find(\"startRec\") >= 0:\n # message \"startRec:/prefix_path/:at:signame:\"\n print(\"i: startRec received\")\n values = msg.split(\":\")\n \n prefix = values[1]\n at = float(values[2])\n signame = values[3]\n \n self.record(prefix, at, signame)\n sc.send(\"ok\")\n \n elif msg.find(\"stopRec\") >= 0:\n print(\"i: stopRec received\")\n \n if self.app.stopApp():\n print(\"i: stopRec successful\")\n sc.send(\"ok\")\n else:\n print(\"i: stopRec failed\")\n sc.send(\"error\")\n \n else:\n print(\"i: ending\")\n break\n \n print(\"i: cmd duration: \" + str(time.time() - tic) + \" [seg]\\n\")\n \n msg = sc.recv(1024)\n\n else:\n print(\"e: not start\")\n sc.send(\"error\")\n \n if msg == \"finish\":\n print(\"i: finish cmd received\")\n \n sc.close()\n print(\"i: end startup\")\n \n def __exit__(self):\n '''\n This method runs on the event dispatching thread.\n '''\n print \"somebody call me!\"\n self.__exit__()\n \n \nif __name__ == '__main__':\n '''\n Creates an instance of the specified {@code Application}\n subclass, sets the {@code ApplicationContext} {@code\n application} property, and then calls the new {@code\n Application's} {@code startup} method. The {@code launch} method is\n typically called from the Application's {@code main}:\n '''\n \n # Reading the properties\n confFile = \"confRX.txt\"\n if(len(sys.argv) > 1):\n arg = sys.argv[1]\n confFile = arg if len(arg) > 0 else confFile \n else:\n print(\"working with default config file path\")\n \n properties = readProperties(confFile)\n \n print(\"Properties:\")\n for p in properties:\n print(\"\\t\" + p + \" : \" + properties[p])\n \n path = properties[\"rxpath\"] \n path = path if (path.endswith(\"/\")) else path+\"/\"\n sync = True if properties[\"sync\"] == \"True\" else False\n \n app = SlaveRX(properties[\"rxip\"],\n int(properties[\"rxport\"]),\n path)\n \n app.setRXProperties(int(properties[\"lo_off\"]), \n float(properties[\"fc\"]), \n int(properties[\"dec\"]), \n int(properties[\"gain\"]), \n sync)\n \n app.launch()\n exit()"},"license":{"kind":"string","value":"apache-2.0"},"hash":{"kind":"number","value":-2031581930941409300,"string":"-2,031,581,930,941,409,300"},"line_mean":{"kind":"number","value":29.2596153846,"string":"29.259615"},"line_max":{"kind":"number","value":113,"string":"113"},"alpha_frac":{"kind":"number","value":0.4829175274,"string":"0.482918"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":4.301435406698564,"string":"4.301435"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42173,"cells":{"repo_name":{"kind":"string","value":"alanc10n/py-rau"},"path":{"kind":"string","value":"pyrau/rau.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1747"},"content":{"kind":"string","value":"import argparse\nfrom redis import StrictRedis\nfrom pyrau.commands import Command\n\n\ndef delete(args, command):\n \"\"\" Execute the delete command \"\"\"\n command.delete(args.pattern)\n\n\ndef keys(args, command):\n \"\"\" Execute the keys command \"\"\"\n details = args.details | args.sorted\n command.keys(args.pattern, details, args.sorted)\n\n\ndef parse_args():\n parser = argparse.ArgumentParser()\n parser.add_argument('-H', '--host', default='localhost', help='Host')\n parser.add_argument('-p', '--port', default=6379, type=int, help='Port')\n parser.add_argument('-b', '--batch_size', default=20,\n type=int, help='Batch size for pipeline operations')\n\n subparsers = parser.add_subparsers(help='Commands')\n del_parser = subparsers.add_parser('delete', help='Delete key(s)')\n del_parser.add_argument('pattern', type=str, help='Key pattern')\n del_parser.set_defaults(func=delete)\n\n key_parser = subparsers.add_parser('keys', help=\"List keys\")\n key_parser.add_argument('-p', '--pattern', help='Key pattern',\n default=None)\n key_parser.add_argument('-d', '--details',\n help='Include details for key(s)',\n action='store_true')\n key_parser.add_argument('-s', '--sorted',\n help='Sort result by size, implies --details',\n action='store_true')\n key_parser.set_defaults(func=keys)\n\n args = parser.parse_args()\n return args\n\n\ndef main():\n args = parse_args()\n redis = StrictRedis(host=args.host, port=args.port)\n command = Command(redis)\n command.batch_size = args.batch_size\n args.func(args, command)\n\n\nif __name__ == '__main__':\n main()\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":-40468994944548810,"string":"-40,468,994,944,548,810"},"line_mean":{"kind":"number","value":31.9622641509,"string":"31.962264"},"line_max":{"kind":"number","value":76,"string":"76"},"alpha_frac":{"kind":"number","value":0.6067544362,"string":"0.606754"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.943566591422122,"string":"3.943567"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42174,"cells":{"repo_name":{"kind":"string","value":"TerryRen/TrPython"},"path":{"kind":"string","value":"NetLib/SuperCaptcha.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"9743"},"content":{"kind":"string","value":"#python 2.7\n#coding=utf-8\n\n__author__ = \"Terry.Ren\"\n\n#try:\n# import Image\n#except ImportError:\n# from PIL import Image\n\nfrom PIL import Image\nfrom PIL import ImageDraw\nimport ImageEnhance\nimport os\nimport urllib\nimport StringIO\nimport uuid\nimport pytesseract #open source\n\nclass Captcha(object): \n def __init__(self, isDebug = False):\n self.__isDebug = isDebug\n self.__currentStepId = 1\n self.__tempFileList = []\n\n def __BuildTempFileFullName(self, localDir, extName):\n fname = str(uuid.uuid1()) + \"_\" + str(self.__currentStepId) + \".\" + extName\n fname = os.path.join(localDir,fname)\n self.__currentStepId += 1\n self.__tempFileList.append(fname)\n\n return fname\n\n '''\n Store remote image to local dir\n '''\n def __StoreImage2LocalDir(self, imageUrl , localDir , extName):\n response = urllib.urlopen(imageUrl)\n tempFileFullName = self.__BuildTempFileFullName(localDir, extName)\n with open(tempFileFullName, 'wb') as f: \n f.write(response.read()) \n \n return tempFileFullName\n\n def Clearup(self):\n for filename in self.__tempFileList:\n if os.path.isfile(filename):\n os.remove(filename)\n \n '''\n image enhance\n '''\n def __imageEnhance(self, image):\n enhancer = ImageEnhance.Contrast(image) \n image_enhancer = enhancer.enhance(4)\n\n return image_enhancer\n\n '''\n two value\n '''\n def __twoValue(self, image):\n img = image.convert('RGBA') # convert to RGBA\n pix = img.load() #read pix\n\n for x in range(img.size[0]): #remove [top-bottom] border\n pix[x, 0] = pix[x, img.size[1] - 1] = (255, 255, 255, 255)\n\n for y in range(img.size[1]): #remove [left-right] border\n pix[0, y] = pix[img.size[0] - 1, y] = (255, 255, 255, 255)\n\n for y in range(img.size[1]): # two value: R=95,G=95,B=95\n for x in range(img.size[0]):\n if pix[x, y][0] < 95 or pix[x, y][1] < 95 or pix[x, y][2] < 95:\n pix[x, y] = (0, 0, 0, 255)\n else:\n pix[x, y] = (255, 255, 255, 255)\n return img\n\n '''\n Get Captcha Code from on-line web site\n '''\n def GetOnlineCaptchaCode(self, imageUrl, isStoreOriginalImage = False, localDir = '', extName = 'jpg'):\n\n if isStoreOriginalImage == True:\n if not os.path.isdir(localDir):\n raise ValueError(\"please validate the argument GetOnlineCaptchaCode.localDir...\")\n \n localFileName = self.__StoreImage2LocalDir(imageUrl , localDir , extName)\n img = Image.open(localFileName)\n else:\n imgBuf = StringIO.StringIO(urllib.urlopen(imageUrl).read())\n img = Image.open(imgBuf)\n \n print img.format, img.size, img.mode\n\n # image Enhance\n img = self.__imageEnhance(img)\n if self.__isDebug: \n img.save(self.__BuildTempFileFullName(localDir, extName))\n\n img = self.__twoValue(img) \n \n tempFileFullName = self.__BuildTempFileFullName(localDir, extName) \n img.save(tempFileFullName) # must use local file via tesseract-orc\n\n text = pytesseract.image_to_string(Image.open(tempFileFullName))\n \n return text \n\n '''\n Get Captcha Code from local\n '''\n def GetLocalCaptchaCode(self, imagePath, extName = 'jpg'):\n\n localDir = os.path.dirname(imagePath)\n img = Image.open(imagePath)\n \n print img.format, img.size, img.mode\n\n # image Enhance\n img = self.__imageEnhance(img)\n if self.__isDebug: \n img.save(self.__BuildTempFileFullName(localDir, extName))\n\n img = img.convert('RGBA') # convert to RGBA\n pix = img.load() #read pix\n\n for x in range(img.size[0]): #remove [top-bottom] border\n pix[x, 0] = pix[x, img.size[1] - 1] = (255, 255, 255, 255)\n\n for y in range(img.size[1]): #remove [left-right] border\n pix[0, y] = pix[img.size[0] - 1, y] = (255, 255, 255, 255)\n\n for y in range(img.size[1]): # two value: R=95,G=95,B=95\n for x in range(img.size[0]):\n if pix[x, y][0] < 90 or pix[x, y][1] < 90 or pix[x, y][2] < 90:\n pix[x, y] = (0, 0, 0, 255)\n else:\n pix[x, y] = (255, 255, 255, 255)\n \n tempFileFullName = self.__BuildTempFileFullName(localDir, extName) \n img.save(tempFileFullName) # must use local file via tesseract-orc\n\n text = pytesseract.image_to_string(Image.open(tempFileFullName))\n \n return text \n\n def TestImage(self):\n data = [(1,0),(0,1)]\n size = (2,2)\n image = Image.new(\"1\",size) \n draw = ImageDraw.Draw(image) \n \n for x in xrange(0,size[0]): \n for y in xrange(0,size[1]): \n draw.point((x,y),data[x][y]) \n \n image.save(\"D:\\\\GitHub\\\\TrPython\\\\NetLib\\\\Test\\\\1.gif\") \n\n\n\nclass SmartCaptcha(object):\n def __init__(self, isDebug = False):\n self.__isDebug = isDebug\n self.__currentStepId = 1\n self.__tempFileList = []\n\n def __BuildTempFileFullName(self, localDir, extName):\n fname = str(uuid.uuid1()) + \"_\" + str(self.__currentStepId) + \".\" + extName\n fname = os.path.join(localDir,fname)\n self.__currentStepId += 1\n self.__tempFileList.append(fname)\n\n return fname\n\n '''\n Store remote image to local dir\n '''\n def __StoreImage2LocalDir(self, imageUrl , localDir , extName):\n response = urllib.urlopen(imageUrl)\n tempFileFullName = self.__BuildTempFileFullName(localDir, extName)\n with open(tempFileFullName, 'wb') as f: \n f.write(response.read()) \n \n return tempFileFullName\n\n def Clearup(self):\n for filename in self.__tempFileList:\n if os.path.isfile(filename):\n os.remove(filename)\n \n '''\n image enhance\n '''\n def __imageEnhance(self, image):\n enhancer = ImageEnhance.Contrast(image) \n image_enhancer = enhancer.enhance(4)\n\n return image_enhancer\n\n '''\n two value\n '''\n def __twoValue(self, image):\n img = image.convert('RGBA') # convert to RGBA\n pix = img.load() #read pix\n\n for x in range(img.size[0]): #remove [top-bottom] border\n pix[x, 0] = pix[x, img.size[1] - 1] = (255, 255, 255, 255)\n\n for y in range(img.size[1]): #remove [left-right] border\n pix[0, y] = pix[img.size[0] - 1, y] = (255, 255, 255, 255)\n\n for y in range(img.size[1]): # two value: R=100,G=100,B=120\n for x in range(img.size[0]):\n if pix[x, y][0] < 100 and pix[x, y][1] < 100:\n pix[x, y] = (0, 0, 0, 255)\n else:\n pix[x, y] = (255, 255, 255, 255)\n return img\n\n\n def __getEffectivePoint(self, pix, x , y):\n point, sx , sy = 0, x-1, y-1\n #print sx+3 , sy +3 ,x , y\n for i in xrange(3):\n for j in xrange(3):\n if sx+i == x and sy+j == y:\n continue\n if pix[sx+i,sy+j] == pix[x,y]:\n point += 1\n\n return point;\n\n '''\n 1111111\n 1011101\n 1011101\n 1111111\n '''\n def __clearNoise(self, img, effectivePoint ,processCount):\n \n for ct in xrange(0, processCount):\n pix = img.load() #read pix\n for x in xrange(1,img.size[0] - 1):\n for y in xrange(1, img.size[1] - 1):\n point = self.__getEffectivePoint(pix , x , y)\n if point < effectivePoint:\n pix[x, y] = (255, 255, 255, 255) # set to Noise\n \n return img\n\n '''\n Get Captcha Code from local\n '''\n def GetLocalCaptchaCode(self, imagePath, extName = 'jpg'):\n\n localDir = os.path.dirname(imagePath)\n img = Image.open(imagePath)\n \n print img.format, img.size, img.mode\n\n # image Enhance\n img = self.__imageEnhance(img)\n if self.__isDebug: \n img.save(self.__BuildTempFileFullName(localDir, extName))\n # two value\n img = self.__twoValue(img) \n if self.__isDebug:\n img.save(self.__BuildTempFileFullName(localDir, extName))\n # clear Noise\n img = self.__clearNoise(img, 3 , 1)\n\n # orc\n tempFileFullName = self.__BuildTempFileFullName(localDir, extName) \n img.save(tempFileFullName) # must use local file via tesseract-orc\n\n text = pytesseract.image_to_string(Image.open(tempFileFullName))\n \n return text \n\nif __name__ == \"__main__\":\n print '[unit test]'\n\n #validate1 = Captcha()\n #print validate1.GetOnlineCaptchaCode(\"http://202.119.81.113:8080/verifycode.servlet\")\n\n #validate2 = Captcha(True)\n #print validate2.GetOnlineCaptchaCode(\"http://202.119.81.113:8080/verifycode.servlet\",True,\"D:\\\\GitHub\\\\TrPython\\\\NetLib\\\\Test\")\n #validate2.Clearup()\n\n #validate3 = Captcha(True)\n #print validate3.GetLocalCaptchaCode(\"D:\\\\GitHub\\\\TrPython\\\\NetLib\\\\Test\\\\1400.gif\",\"gif\")\n #validate3.TestImage()\n\n validate4 = SmartCaptcha(True)\n print validate4.GetLocalCaptchaCode(\"D:\\\\GitHub\\\\TrPython\\\\NetLib\\\\Test\\\\xxf2.jpg\",\"jpg\")\n #print validate4.GetLocalCaptchaCode(\"D:\\\\GitHub\\\\TrPython\\\\NetLib\\\\Test\\\\queaa.jpg\",\"jpg\")\n\n print pytesseract.image_to_string(Image.open(\"D:\\\\GitHub\\\\TrPython\\\\NetLib\\\\Test\\\\xxf2.jpg\"))\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n"},"license":{"kind":"string","value":"apache-2.0"},"hash":{"kind":"number","value":4605390416347193000,"string":"4,605,390,416,347,193,000"},"line_mean":{"kind":"number","value":29.6006289308,"string":"29.600629"},"line_max":{"kind":"number","value":132,"string":"132"},"alpha_frac":{"kind":"number","value":0.5500976261,"string":"0.550098"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.5104617604617605,"string":"3.510462"},"config_test":{"kind":"bool","value":true,"string":"true"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42175,"cells":{"repo_name":{"kind":"string","value":"weggert/calendar_sync"},"path":{"kind":"string","value":"calendar_sync.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"6753"},"content":{"kind":"string","value":"#!/usr/bin/python\nimport fileinput\nimport os\n\n\nclass CalendarManager:\n\n def __init__(self, calendar_name, dry_run, include_descriptions):\n self.calendar_name = calendar_name\n self.dry_run = dry_run\n self.include_descriptions = include_descriptions\n\n def clear_calendar(self):\n command = \"\"\"\nosascript -e 'tell application \"Calendar\" to tell calendar \"%s\"\n set eventList to every event\n repeat with e in eventList\n delete e\n end repeat\nend tell'\n\"\"\"\n command = command % self.calendar_name\n\n if not self.dry_run:\n os.system(command)\n print 'Calendar cleared'\n \n def create_calendar_event(self, summary, start_date, end_date, all_day, location, description):\n if not self.include_descriptions:\n description = ''\n\n properties = 'start date:theStartDate, end date:theEndDate, summary:\"%s\", description:\"%s\", location:\"%s\"'\\\n % (summary, description, location)\n if all_day is True:\n properties += ', allday event:true'\n command = \"\"\"\nosascript -e 'set theStartDate to date \"%s\"\nset theEndDate to date \"%s\"\ntell application \"Calendar\" to tell calendar \"%s\"\n set theEvent to make new event with properties {%s}\nend tell'\n\"\"\"\n command = command % (start_date, end_date, self.calendar_name, properties)\n\n if not self.dry_run:\n os.system(command)\n self.print_summary(summary, start_date, end_date, all_day, location, description)\n\n @staticmethod\n def print_summary(summary, start_date, end_date, all_day, location, description):\n print 'Summary: ' + summary\n print ' Start: ' + start_date\n print ' End: ' + end_date\n print ' All Day: ' + str(all_day)\n print ' Location: ' + location\n print ' Description: ' + description\n print ''\n\n\nclass CalendarSummaryProcessor:\n\n class LineType:\n EventStart, Summary, Location, Date, Time, Where, Notes, Status, Other = range(9)\n\n def __init__(self):\n pass\n\n def __init__(self, calendar_name, dry_run, include_descriptions):\n self.calendar_manager = CalendarManager(\n calendar_name=calendar_name,\n dry_run=dry_run,\n include_descriptions=include_descriptions)\n self.reset()\n self.processing_event = False\n self.first_description_line = True\n self.last_description_line_was_blank = False\n self.summary = ''\n self.date = ''\n self.time = ''\n self.location = ''\n self.description = ''\n\n def reset(self):\n self.processing_event = False\n self.first_description_line = True\n self.last_description_line_was_blank = False\n self.summary = ''\n self.date = ''\n self.time = ''\n self.location = ''\n self.description = ''\n\n def process_summary(self):\n self.calendar_manager.clear_calendar()\n\n for input_line in fileinput.input():\n line_type = self.get_line_type(input_line)\n\n if line_type is self.LineType.EventStart:\n if self.processing_event:\n if self.summary != 'Remote'\\\n and self.summary != 'IP Video - Daily Scrum'\\\n and self.summary != 'Cloud Team Scrum':\n start_date, end_date, all_day = self.get_start_end_dates(self.date, self.time)\n self.calendar_manager.create_calendar_event(\n self.summary, start_date, end_date, all_day, self.location, self.description)\n self.reset()\n\n if line_type is self.LineType.Summary:\n self.summary = self.sanitize_line(input_line.strip()[9:])\n self.processing_event = True\n\n if line_type is self.LineType.Date:\n self.date = input_line.strip()[6:]\n\n if line_type is self.LineType.Time:\n self.time = input_line.strip()[6:]\n\n if line_type is self.LineType.Location:\n self.location = self.sanitize_line(input_line.strip()[10:])\n self.processing_event = True\n\n if line_type is self.LineType.Other:\n description_line = self.sanitize_line(input_line.strip())\n if len(description_line) > 0:\n self.description = self.description + description_line + '\\n'\n self.last_description_line_was_blank = False\n else:\n if not self.first_description_line and not self.last_description_line_was_blank:\n self.description += '\\n'\n self.last_description_line_was_blank = True\n self.first_description_line = False\n\n if self.processing_event:\n start_date, end_date, all_day = self.get_start_end_dates(self.date, self.time)\n self.calendar_manager.create_calendar_event(\n self.summary, start_date, end_date, all_day, self.location, self.description)\n\n @staticmethod\n def get_start_end_dates(date, time):\n dates = date.split(\" to \")\n times = time.split(\" to \")\n start_date = dates[0] + ' ' + times[0]\n end_date = dates[1] + ' ' + times[1]\n all_day = False\n if times[0] == '12:00:00 AM' and times[1] == \"12:00:00 AM\" and dates[0] != dates[1]:\n all_day = True\n return start_date, end_date, all_day\n\n def get_line_type(self, input_line):\n if input_line.startswith('EVENT'):\n return self.LineType.EventStart\n\n if input_line.startswith('Summary:'):\n return self.LineType.Summary\n\n if input_line.startswith('Date:'):\n return self.LineType.Date\n\n if input_line.startswith('Time:'):\n return self.LineType.Time\n\n if input_line.startswith('Location:'):\n return self.LineType.Location\n\n if input_line.startswith('Where'):\n return self.LineType.Where\n\n if input_line.startswith('Notes'):\n return self.LineType.Notes\n\n if input_line.startswith('Status'):\n return self.LineType.Status\n\n return self.LineType.Other\n\n def process_named_line(self, input_line):\n colon_position = input_line.find(':')\n return self.sanitize_line(input_line[colon_position+1:].strip())\n\n @staticmethod\n def sanitize_line(input_line):\n return input_line.replace(\"'\", \"\").replace('\"', '').replace('*~*~*~*~*~*~*~*~*~*', '').strip()\n\n\nCalendarSummaryProcessor(calendar_name='Work Calendar',\n dry_run=False,\n include_descriptions=True).process_summary()"},"license":{"kind":"string","value":"apache-2.0"},"hash":{"kind":"number","value":-4109685570745103000,"string":"-4,109,685,570,745,103,000"},"line_mean":{"kind":"number","value":34.9255319149,"string":"34.925532"},"line_max":{"kind":"number","value":115,"string":"115"},"alpha_frac":{"kind":"number","value":0.5770768547,"string":"0.577077"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":4.08529945553539,"string":"4.085299"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42176,"cells":{"repo_name":{"kind":"string","value":"allenai/allennlp"},"path":{"kind":"string","value":"allennlp/modules/text_field_embedders/basic_text_field_embedder.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"5232"},"content":{"kind":"string","value":"from typing import Dict\nimport inspect\n\nimport torch\nfrom overrides import overrides\n\nfrom allennlp.common.checks import ConfigurationError\nfrom allennlp.data import TextFieldTensors\nfrom allennlp.modules.text_field_embedders.text_field_embedder import TextFieldEmbedder\nfrom allennlp.modules.time_distributed import TimeDistributed\nfrom allennlp.modules.token_embedders.token_embedder import TokenEmbedder\nfrom allennlp.modules.token_embedders import EmptyEmbedder\n\n\n@TextFieldEmbedder.register(\"basic\")\nclass BasicTextFieldEmbedder(TextFieldEmbedder):\n \"\"\"\n This is a `TextFieldEmbedder` that wraps a collection of\n [`TokenEmbedder`](../token_embedders/token_embedder.md) objects. Each\n `TokenEmbedder` embeds or encodes the representation output from one\n [`allennlp.data.TokenIndexer`](../../data/token_indexers/token_indexer.md). As the data produced by a\n [`allennlp.data.fields.TextField`](../../data/fields/text_field.md) is a dictionary mapping names to these\n representations, we take `TokenEmbedders` with corresponding names. Each `TokenEmbedders`\n embeds its input, and the result is concatenated in an arbitrary (but consistent) order.\n\n Registered as a `TextFieldEmbedder` with name \"basic\", which is also the default.\n\n # Parameters\n\n token_embedders : `Dict[str, TokenEmbedder]`, required.\n A dictionary mapping token embedder names to implementations.\n These names should match the corresponding indexer used to generate\n the tensor passed to the TokenEmbedder.\n \"\"\"\n\n def __init__(self, token_embedders: Dict[str, TokenEmbedder]) -> None:\n super().__init__()\n # NOTE(mattg): I'd prefer to just use ModuleDict(token_embedders) here, but that changes\n # weight locations in torch state dictionaries and invalidates all prior models, just for a\n # cosmetic change in the code.\n self._token_embedders = token_embedders\n for key, embedder in token_embedders.items():\n name = \"token_embedder_%s\" % key\n self.add_module(name, embedder)\n self._ordered_embedder_keys = sorted(self._token_embedders.keys())\n\n @overrides\n def get_output_dim(self) -> int:\n output_dim = 0\n for embedder in self._token_embedders.values():\n output_dim += embedder.get_output_dim()\n return output_dim\n\n def forward(\n self, text_field_input: TextFieldTensors, num_wrapping_dims: int = 0, **kwargs\n ) -> torch.Tensor:\n if sorted(self._token_embedders.keys()) != sorted(text_field_input.keys()):\n message = \"Mismatched token keys: %s and %s\" % (\n str(self._token_embedders.keys()),\n str(text_field_input.keys()),\n )\n embedder_keys = set(self._token_embedders.keys())\n input_keys = set(text_field_input.keys())\n if embedder_keys > input_keys and all(\n isinstance(embedder, EmptyEmbedder)\n for name, embedder in self._token_embedders.items()\n if name in embedder_keys - input_keys\n ):\n # Allow extra embedders that are only in the token embedders (but not input) and are empty to pass\n # config check\n pass\n else:\n raise ConfigurationError(message)\n\n embedded_representations = []\n for key in self._ordered_embedder_keys:\n # Note: need to use getattr here so that the pytorch voodoo\n # with submodules works with multiple GPUs.\n embedder = getattr(self, \"token_embedder_{}\".format(key))\n if isinstance(embedder, EmptyEmbedder):\n # Skip empty embedders\n continue\n forward_params = inspect.signature(embedder.forward).parameters\n forward_params_values = {}\n missing_tensor_args = set()\n for param in forward_params.keys():\n if param in kwargs:\n forward_params_values[param] = kwargs[param]\n else:\n missing_tensor_args.add(param)\n\n for _ in range(num_wrapping_dims):\n embedder = TimeDistributed(embedder)\n\n tensors: Dict[str, torch.Tensor] = text_field_input[key]\n if len(tensors) == 1 and len(missing_tensor_args) == 1:\n # If there's only one tensor argument to the embedder, and we just have one tensor to\n # embed, we can just pass in that tensor, without requiring a name match.\n token_vectors = embedder(list(tensors.values())[0], **forward_params_values)\n else:\n # If there are multiple tensor arguments, we have to require matching names from the\n # TokenIndexer. I don't think there's an easy way around that.\n token_vectors = embedder(**tensors, **forward_params_values)\n if token_vectors is not None:\n # To handle some very rare use cases, we allow the return value of the embedder to\n # be None; we just skip it in that case.\n embedded_representations.append(token_vectors)\n return torch.cat(embedded_representations, dim=-1)\n"},"license":{"kind":"string","value":"apache-2.0"},"hash":{"kind":"number","value":8035300947094428000,"string":"8,035,300,947,094,428,000"},"line_mean":{"kind":"number","value":47.4444444444,"string":"47.444444"},"line_max":{"kind":"number","value":114,"string":"114"},"alpha_frac":{"kind":"number","value":0.6406727829,"string":"0.640673"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":4.253658536585366,"string":"4.253659"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42177,"cells":{"repo_name":{"kind":"string","value":"perlygatekeeper/glowing-robot"},"path":{"kind":"string","value":"google_test/free_the_bunny_prisoners/solution_5_fails.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1090"},"content":{"kind":"string","value":"import itertools\n\ndef solution(bunnies,keys_required):\n answer = []\n for i in range(bunnies):\n answer.append([])\n# if keys_required > bunnies:\n# return None\n if keys_required == 0:\n return [[0]]\n elif keys_required == 1:\n key = 0\n for group in range(bunnies):\n answer[group].append(key)\n elif bunnies == keys_required:\n key = 0\n for group in range(bunnies):\n answer[group].append(key)\n key += 1\n else:\n key = 0\n for item in itertools.combinations(range(bunnies), keys_required):\n for group in item:\n answer[group].append(key)\n key += 1\n return answer\n\nfor num_buns in range(1,10):\n for num_required in range(10):\n key_dist = solution(num_buns,num_required)\n print(\"-\" * 60)\n print(\"Answer for {0:d} bunnies, requiring {1:d}\".format(num_buns,num_required))\n if ( len(key_dist[0]) * len(key_dist) ) < 25:\n print(key_dist)\n else:\n for bun in key_dist:\n print(bun)\n"},"license":{"kind":"string","value":"artistic-2.0"},"hash":{"kind":"number","value":-8434445743401300000,"string":"-8,434,445,743,401,300,000"},"line_mean":{"kind":"number","value":28.4594594595,"string":"28.459459"},"line_max":{"kind":"number","value":88,"string":"88"},"alpha_frac":{"kind":"number","value":0.5385321101,"string":"0.538532"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.449367088607595,"string":"3.449367"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42178,"cells":{"repo_name":{"kind":"string","value":"mrjmad/nagademon_2014"},"path":{"kind":"string","value":"nagademon2014/maingame/models/history_elements.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"6460"},"content":{"kind":"string","value":"# -*- coding: utf-8 -*-\n\nfrom __future__ import (print_function, division, absolute_import, unicode_literals)\nfrom django.utils.encoding import python_2_unicode_compatible\n\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.conf import settings\nfrom django.db import models\n\nUSER_MODEL = settings.AUTH_USER_MODEL\n\n\n@python_2_unicode_compatible\nclass Character(models.Model):\n short_name = models.CharField(_(\"NPC's short Name\"), max_length=20, unique=True)\n first_name = models.CharField(\"Firstname of Character\", max_length=50)\n last_name = models.CharField(\"Lastname of Character\", max_length=50)\n gender = models.PositiveSmallIntegerField(u\"Gender of Character\")\n description = models.TextField(\"Description\")\n\n def __str__(self):\n return u\"%s %s\" % (self.first_name, self.last_name)\n\n class Meta:\n abstract = True\n\n\n@python_2_unicode_compatible\nclass PlayerCharacter(Character):\n\n def __str__(self):\n return u\"PC : %s %s\" % (self.first_name, self.last_name)\n\n\n@python_2_unicode_compatible\nclass NPCharacter(Character):\n\n def __str__(self):\n return u\"NPC : %s %s\" % (self.first_name, self.last_name)\n\n\nclass PlaceManager(models.Manager):\n def get_by_natural_key(self, short_name):\n return self.get(short_name=short_name)\n\n\n@python_2_unicode_compatible\nclass Place(models.Model):\n objects = PlaceManager()\n begin_sound = models.CharField(_(\"Begin's Sound\"), max_length=200, blank=True, null=True)\n ambiance_sound = models.CharField(_(\"Ambiance's Sound\"), max_length=200, blank=True, null=True)\n short_name = models.CharField(_(\"Place's short Name\"), max_length=20, unique=True)\n name = models.CharField(\"Scene's Name\", max_length=200)\n filename = models.CharField(\"Scene's Filename\", max_length=80)\n text = models.TextField(\"Scene's Text\")\n\n def __str__(self):\n return self.name\n\n def natural_key(self):\n return self.short_name,\n\n\n@python_2_unicode_compatible\nclass Scene(models.Model):\n short_name = models.CharField(_(\"Scene's short Name\"), max_length=20, unique=True)\n name = models.CharField(\"Scene's Name\", max_length=200)\n filename = models.CharField(\"Scene's Filename\", max_length=80)\n begin_sound = models.CharField(_(\"Begin's Sound\"), max_length=200, blank=True, null=True)\n ambiance_sound = models.CharField(_(\"Ambiance's Sound\"), max_length=200, blank=True, null=True)\n synopsis = models.TextField(\"Scene's synopsis, only for authors\")\n final = models.BooleanField(\"Final Round ?\", default=False)\n place = models.ForeignKey(Place, verbose_name=\"Scene's Place\",\n blank=True, null=True)\n is_active = models.BooleanField(_(\"Is active ?\"), default=True)\n order = models.PositiveIntegerField(_(\"Scene's Order\"), default=0)\n need_a_trigger = models.BooleanField(_(\"Activable only by a trigger\"), default=False)\n\n def __str__(self):\n return self.name\n\n\n@python_2_unicode_compatible\nclass PartScene(models.Model):\n text = models.CharField(\"Scene's Text\", max_length=400)\n for_scene = models.ForeignKey(Scene, verbose_name=\"Scene\")\n limited_to_player = models.ForeignKey(PlayerCharacter, blank=True, null=True)\n parent = models.ForeignKey('self', blank=True, null=True)\n active = models.BooleanField(default=True)\n\n def __str__(self):\n return \"Text %s |for scene :%s\" % (self.text, self.for_scene)\n\n\n@python_2_unicode_compatible\nclass Choice1PartSceneto1Scene(models.Model):\n text = models.CharField(\"Choice's Text\", max_length=400)\n for_part_scene = models.ForeignKey(PartScene, verbose_name=\"Current Part Scene\",\n related_name=\"current_choices_set\")\n next_scene = models.ForeignKey(Scene, verbose_name=\"Next Scene\",\n related_name=\"leading_choices_set\",\n null=True, blank=True)\n next_part_scene = models.ForeignKey(PartScene, verbose_name=\"Next Part Scene\",\n related_name=\"leading_choices_set\",\n null=True, blank=True)\n\n def __str__(self):\n return \"%s |for scene %s , part scene id :%s\" % (self.text,\n self.for_part_scene.for_scene,\n self.for_part_scene.id)\n\n\n@python_2_unicode_compatible\nclass Quest(models.Model):\n short_name = models.CharField(_(\"Quest's short Name\"), max_length=20, unique=True)\n title = models.CharField(\"Quest's Title\", max_length=140)\n text = models.TextField(\"Quest's Text\")\n time_frame = models.PositiveIntegerField(_(\"Maximum Time (in minutes) for validate the Quest\"), default=0)\n given_by = models.ForeignKey(NPCharacter, verbose_name=_('Given by'))\n scene = models.ForeignKey(Scene, verbose_name=_(\"Scene who Quest is activable\"),\n related_name=_(\"quests_for_scene\"))\n scene_after = models.ForeignKey(Scene, verbose_name=_(\"Scene after the End's Quest\"),\n related_name=_(\"finished_quests_for_scene\"))\n apparition_function = models.CharField(_(\"Name of Apparition's Function\"), max_length=120, blank=True, null=True)\n validation_function = models.CharField(_(\"Name of Validation's Function\"), max_length=120)\n\n def __str__(self):\n return \"%s | for scene :%s, by NPC %s in time %s\" % (self.title, self.scene, self.given_by,\n self.timedelta)\n\n\nclass ObjectType(models.Model):\n name = models.CharField(u\"Type Object Name\", max_length=200)\n description = models.TextField(\"Type's Description\", blank=True, null=True)\n short_name = models.CharField(_(\"Type Object's short Name\"), max_length=20, unique=True)\n\n\nclass OneObject(models.Model):\n name = models.CharField(_(\"Type Object Name\"), max_length=200)\n type = models.ForeignKey(ObjectType, verbose_name=_(\"Object's Type\"))\n description = models.TextField(\"Object's Description\", blank=True, null=True)\n initial_place = models.ForeignKey(Place, verbose_name=_(\"Object's Initial place\"),\n related_name=_(\"initial_objects_set\"), blank=True, null=True)\n stored_in = models.ForeignKey(Place, related_name=_(\"objects_stored_set\"),\n verbose_name=_(\"Where the object is stored\"), blank=True, null=True)\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":5791698670006022000,"string":"5,791,698,670,006,022,000"},"line_mean":{"kind":"number","value":43.5517241379,"string":"43.551724"},"line_max":{"kind":"number","value":117,"string":"117"},"alpha_frac":{"kind":"number","value":0.6452012384,"string":"0.645201"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.8752249550089983,"string":"3.875225"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42179,"cells":{"repo_name":{"kind":"string","value":"vitale232/ves"},"path":{"kind":"string","value":"ves/VESinverse_vectorized.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"12839"},"content":{"kind":"string","value":"# -*- coding: utf-8 -*-\r\n\"\"\"\r\nCreated on Thu Jan 28 16:32:48 2016\r\n\r\n@author: jclark\r\n\r\nthis code uses the Ghosh method to determine the apparent resistivities\r\nfor a layered earth model. Either schlumberger or Wenner configurations\r\ncan be used\r\n\"\"\"\r\n\r\n\r\nimport numpy as np\r\nimport random\r\nimport matplotlib\r\nmatplotlib.use('Qt5Agg')\r\nimport matplotlib.pyplot as plt\r\nplt.style.use('bmh')\r\nimport sys\r\n\r\n# Schlumberger filter\r\nfltr1 = [0., .00046256, -.0010907, .0017122, -.0020687,\r\n .0043048, -.0021236, .015995, .017065, .098105, .21918, .64722,\r\n 1.1415, .47819, -3.515, 2.7743, -1.201, .4544, -.19427, .097364,\r\n -.054099, .031729, -.019109, .011656, -.0071544, .0044042,\r\n -.002715, .0016749, -.0010335, .00040124]\r\n\r\n#Wenner Filter\r\nfltr2 = [0., .000238935, .00011557, .00017034, .00024935,\r\n .00036665, .00053753, .0007896, .0011584, .0017008, .0024959,\r\n .003664, .0053773, .007893, .011583, .016998, .024934, .036558,\r\n .053507, .078121, .11319, .16192, .22363, .28821, .30276, .15523,\r\n -.32026, -.53557, .51787, -.196, .054394, -.015747, .0053941,\r\n -.0021446, .000665125]\r\nprint(len(fltr1))\r\nprint(len(fltr2))\r\n\r\n#I know there must be a better method to assign lists. And probably numpy\r\n#arrays would be best. But my Python wasn't up to it. If the last letter\r\n#is an 'l' that means it is a log10 of the value\r\n\r\n# 65 is completely arbitrary\r\np = [0] * 20 # earth layer parameters?\r\nr = [0] * 65 # apparent resistivty?\r\nrl = [0] * 65 # np.log(r) ?\r\nt = [0] * 50 #\r\nb = [0] * 65 #\r\nasav = [0] * 65 # voltage spacing in meters?\r\nasavl = [0] * 65 # np.log(asav)\r\nadatl = [0] * 65 # interpolated voltage spacing ( np.log(10) / 6 )?\r\nrdatl = [0] * 65 # np.log()\r\n# adat = [0] * 65 # voltage spacing input\r\n# rdat = [0] * 65 # apparent res input\r\npkeep = [0] * 65 # earth parameters after applying equations?\r\nrkeep = [0] * 65 # r after applying equations?\r\nrkeepl = [0] * 65 # np.log()!\r\npltanswer = [0] * 65\r\npltanswerl = [0] * 65\r\npltanswerkeep = [0] * 65\r\npltanswerkeepl = [0] * 65\r\n\r\nrl = [0] * 65\r\nsmall = [0] * 65\r\nxlarge = [0] * 65\r\n\r\nx=[0] * 100\r\ny = [0] * 100\r\ny2 = [0] * 100\r\nu = [0] * 5000\r\nnew_x = [0] * 1000\r\nnew_y = [0] * 1000\r\nndat = 13\r\n#hard coded data input - spacing and apparent resistivities measured\r\n#in teh field\r\nadat = [0., 0.55, 0.95, 1.5, 2.5, 3., 4.5, 5.5, 9., 12., 20., 30., 70.]\r\nrdat = [0., 125., 110., 95., 40., 24., 15., 10.5, 8., 6., 6.5, 11., 25.]\r\none30 = 1.e30 # What's the purpose of this and should it be user input?\r\nrms = one30 # Just a starting value for rmserror?\r\nerrmin = 1.e10 # Should this be user input?\r\n\r\n# INPUT\r\narray_spacing = 'wenner' # 1 is for shchlumberger and 2 is for Wenner\r\nnLayers = 3 #number of layers\r\nn = 2 * nLayers - 1 # What does n represent? number of parameters\r\n\r\n\r\nspac = 0.2 # smallest electrode spacing - should this come from the input file?\r\nm = 20 # number of points where resistivity is calculated\r\n\r\nspac = np.log(spac)\r\ndelx = np.log(10.0) / 6. # I take it this is the sample interval on the log scale?\r\n\r\n\r\n\r\n# this is where the range in parameters should be input from a GUI\r\n# I'm hard coding this in for now\r\n\r\n#enter thickenss range for each layer and then resistivity range.\r\n#for 3 layers small[1] and small[2] are low end of thickness range\r\n# small[3], small[4] and small[5] are the low end of resistivities\r\n\r\n# I think I have it coded up that these are getting grabbed from the rectangles currently.\r\n# Is that the best way to go?\r\nsmall[1] = 1.\r\nsmall[2] = 10.\r\nsmall[3] = 20.\r\nsmall[4] = 2.\r\nsmall[5] = 500.\r\n\r\nxlarge[1] = 5\r\nxlarge[2] = 75.\r\nxlarge[3] = 200.\r\nxlarge[4] = 100\r\nxlarge[5] = 3000.\r\n\r\n\r\n\r\n\r\n\r\niter_ = 10000 #number of iterations for the Monte Carlo guesses. to be input on GUI\r\n# Is 10000 the most reasonable default, or should I play with it?\r\n\r\ndef readData(adat, rdat, ndat, return_indexed=False):\r\n #normally this is where the data would be read from the csv file\r\n # but now I'm just hard coding it in as global lists\r\n\r\n for i in range(1, ndat):\r\n adatl[i] = np.log10(adat[i])\r\n rdatl[i] = np.log10(rdat[i])\r\n\r\n if return_indexed:\r\n return adatl[:ndat], rdatl[:ndat]\r\n else:\r\n return adatl, rdatl\r\n\r\n<<<<<<< HEAD\n=======\n\r\ndef error(): # simple rms error calc\r\n sumerror = 0.\r\n #pltanswer = [0]*64\r\n spline(m, one30, one30, asavl, rl, y2) # So this calculates the predicted fit?\r\n # and essentially operates on the list in place?\r\n for i in range(1, ndat): # So you always skip the value 0? due to -inf returns?\r\n ans = splint(m, adatl[i], asavl, rl, y2) # Then this calulates error?\r\n sumerror = sumerror + (rdatl[i] - ans) * (rdatl[i] - ans)\r\n #print(i,sum1,rdat[i],rdatl[i],ans)\r\n pltanswerl[i] = ans\r\n pltanswer[i] = np.power(10, ans)\r\n rms = np.sqrt(sumerror / (ndat - 1))\r\n\r\n # check the spline routine\r\n# for i in range(1,m+1,1):\r\n# anstest = splint(m, asavl[i],asavl,rl,y2)\r\n# print( asavl[i], rl[i], anstest)\r\n #print(' rms = ', rms)\r\n# if you erally want to get a good idea of all perdictions from Montecarlo\r\n# perform the following plot (caution - change iter to a smaller number)\r\n #plt.loglog(adat[1:ndat],pltanswer[1:ndat])\r\n return rms\r\n\r\n>>>>>>> 60497dd... ?s\ndef transf(y, i):\r\n # these lines apparently find the computer precision ep\r\n ep = 1.0\r\n ep = ep / 2.0\r\n fctr = ep + 1.\r\n while fctr > 1.:\r\n ep = ep / 2.0\r\n fctr = ep + 1.\r\n\r\n u = 1. / np.exp(y) # y = spac - 19. * delx - 0.13069\r\n t[1] = p[n]\r\n for j in range(2, nLayers + 1, 1):\r\n pwr = -2. * u * p[nLayers + 1 - j]\r\n if pwr < np.log(2. * ep):\r\n pwr = np.log(2. * ep)\r\n a = np.exp(pwr)\r\n b = (1. - a) / (1. + a)\r\n rs = p[n + 1 - j]\r\n tpr = b * rs\r\n t[j] = (tpr + t[j - 1]) / (1. + tpr * t[j - 1] / (rs * rs))\r\n r[i] = t[nLayers]\r\n return\r\n\r\ndef filters(b, k):\r\n for i in range(1, m + 1):\r\n re = 0.\r\n for j in range(1, k + 1):\r\n re = re + b[j] * r[i + k - j] # include ranges of thickness, res . push button for rmse error, observed data\r\n # surf thicknes .2 - 100\r\n # res 2-3000 # could use huge ranges at cost of time\r\n r[i] = re\r\n return\r\n\r\ndef rmsfit():\r\n if array_spacing.lower() == 'wenner':\r\n y = spac - 19. * delx - 0.13069\r\n mum1 = m + 28\r\n for i in range(1, mum1 + 1):\r\n transf(y, i)\r\n y = y + delx\r\n filters(fltr1, 29)\r\n\r\n elif array_spacing.lower() == 'schlumberger':\r\n s = np.log(2.)\r\n y = spac - 10.8792495 * delx\r\n mum2 = m + 33\r\n for i in range(1, mum2 + 1):\r\n transf(y, i)\r\n a = r[i]\r\n y1 = y + s\r\n transf(y1, i)\r\n r[i] = 2. * a - r[i]\r\n y = y + delx\r\n filters(fltr2, 34)\r\n else:\r\n print(\"\\nType of survey not indicated.\")\r\n raise SystemExit('Exiting.\\n\\n Take better care next time.')\r\n\r\n x = spac\r\n #print(\"A-Spacing App. Resistivity\")\r\n for i in range(1, m + 1):\r\n a = np.exp(x)\r\n asav[i] = a\r\n asavl[i] = np.log10(a)\r\n rl[i] = np.log10(r[i])\r\n x = x + delx\r\n #print(\"%7.2f %9.3f \" % ( asav[i], r[i]))\r\n\r\n rms = error()\r\n\r\n return rms\r\n\r\ndef error(): # simple rms error calc\r\n sumerror = 0.\r\n #pltanswer = [0]*64\r\n spline(m, one30, one30, asavl, rl, y2) # So this calculates the predicted fit?\r\n # and essentially operates on the list in place?\r\n for i in range(1, ndat): # So you always skip the value 0? due to -inf returns?\r\n ans = splint(m, adatl[i], asavl, rl, y2) # Then this calulates error?\r\n sumerror = sumerror + (rdatl[i] - ans) * (rdatl[i] - ans)\r\n #print(i,sum1,rdat[i],rdatl[i],ans)\r\n pltanswerl[i] = ans\r\n pltanswer[i] = np.power(10, ans)\r\n rms = np.sqrt(sumerror / (ndat - 1))\r\n\r\n # check the spline routine\r\n# for i in range(1,m+1,1):\r\n# anstest = splint(m, asavl[i],asavl,rl,y2)\r\n# print( asavl[i], rl[i], anstest)\r\n #print(' rms = ', rms)\r\n# if you erally want to get a good idea of all perdictions from Montecarlo\r\n# perform the following plot (caution - change iter to a smaller number)\r\n #plt.loglog(adat[1:ndat],pltanswer[1:ndat])\r\n return rms\r\n# my code to do a spline fit to predicted data at the nice spacing of Ghosh\r\n# use splint to determine the spline interpolated prediction at the\r\n# spacing where the measured resistivity was taken - to compare observation\r\n# to prediction\r\ndef spline(n, yp1, ypn, x=[] ,y=[] ,y2=[]):\r\n \"\"\"Still struggling to understand the general operation of this function.\"\"\"\r\n u = [0] * 1000\r\n one29 = 0.99e30\r\n #print(x,y)\r\n if yp1 > one29:\r\n y2[0] = 0.\r\n u[0] = 0.\r\n else:\r\n y2[0] = -0.5\r\n u[0] = (3. / (x[1] - x[0])) * ((y[1] - y[0]) / (x[1] - x[0]) - yp1)\r\n\r\n for i in range(1, n):\r\n #print(i,x[i])\r\n sig = (x[i] - x[i-1]) / (x[i+1] - x[i-1])\r\n p=sig * y2[i - 1] + 2.\r\n y2[i] = (sig-1.) / p\r\n u[i] = (((6. * ((y[i+1] - y[i]) / (x[i+1] - x[i]) - (y[i] - y[i-1]) /\r\n x[i] - x[i-1])) / (x[i + 1] - x[i - 1]) - sig * u[i - 1]) / p)\r\n\r\n if ypn > one29:\r\n qn = 0.\r\n un = 0.\r\n else:\r\n qn = 0.5\r\n un = (3. / (x[n] - x[n - 1])) * (ypn - (y[n] - y[n - 1]) / (x[n] - x[n - 1]))\r\n\r\n y2[n] = (un - qn * u[n - 1]) / (qn * y2[n - 1] + 1.)\r\n for k in range(n-1, -1, -1):\r\n y2[k] = y2[k] * y2[k + 1] + u[k]\r\n\r\n return\r\n\r\ndef splint(n, x ,xa=[], ya=[], y2a=[]): # Is this function the T function?\r\n \"\"\"Still struggling to understand the general operation of this function.\"\"\"\r\n klo = 0\r\n khi = n\r\n while khi - klo > 1:\r\n k = int((khi + klo) // 2)\r\n if xa[k] > x:\r\n khi = k\r\n else:\r\n klo = k\r\n h = xa[khi] - xa[klo]\r\n if abs(h) < 1e-20:\r\n print(\" bad xa input\")\r\n #print(x,xa[khi],xa[klo])\r\n a = (xa[khi] - x) / h\r\n b = (x - xa[klo]) / h\r\n y = (a * ya[klo] + b * ya[khi] + ((a * a * a - a) * y2a[klo] +\r\n (b * b * b - b) * y2a[khi]) * (h * h) /6.)\r\n #print(\"x= \", x,\"y= \", y, \" ya= \", ya[khi],\" y2a= \", y2a[khi], \" h= \",h)\r\n\r\n return y\r\n\r\n\r\n#main here\r\nif __name__ == '__main__':\r\n\r\n adatl, rdatl = readData(adat, rdat, ndat, return_indexed=False)\r\n\r\n print(adat[1:ndat],rdat[1:ndat])\r\n print('log stufffff')\r\n\r\n print(adatl[1:ndat], rdatl[1:ndat]) # is this to skip 0?\r\n\r\n#enter thickenss range for each layer and then resistivity range.\r\n#for 3 layers small[1] and small[2] are low end of thickness range\r\n# small[3], small[4] and small[5] are the low end of resistivities\r\n\r\n for iloop in range(1, int(iter_/2) + 1):\r\n #print( ' iloop is ', iloop)\r\n for i in range(1, n + 1): # number of parameters + 1\r\n randNumber = random.random() # IS this just to add noise to the model?\r\n # #print(randNumber, ' random')\r\n # print(xlarge)\r\n # print(small)\r\n # s = input('')\r\n # print('xlarge[i]: {}, small[i]: {}'.format(xlarge[i], small[i]))\r\n p[i] = (xlarge[i] - small[i]) * randNumber + small[i]\r\n # print(p)\r\n print('\\n')\r\n print(p)\r\n # s = input('')\r\n rms = rmsfit()\r\n\r\n if rms < errmin:\r\n print('rms ', rms, ' errmin ', errmin)\r\n for i in range(1, n + 1):\r\n pkeep[i] = p[i]\r\n for i in range(1, m + 1):\r\n rkeep[i] = r[i]\r\n rkeepl[i] = rl[i]\r\n for i in range(1, ndat + 1):\r\n pltanswerkeepl[i] = pltanswerl[i]\r\n pltanswerkeep[i] = pltanswer[i]\r\n errmin = rms\r\n\r\n#output the best fitting earth model\r\n print(' Layer ', ' Thickness ', ' Res_ohm-m ')\r\n for i in range(1,nLayers,1):\r\n print(i, pkeep[i], pkeep[nLayers+i-1])\r\n\r\n print( nLayers, ' Infinite ', pkeep[n])\r\n for i in range(1,m+1, 1):\r\n asavl[i] = np.log10(asav[i])\r\n\r\n#output the error of fit\r\n print( ' RMS error ', errmin)\r\n print( ' Spacing', ' Res_pred ', ' Log10_spacing ', ' Log10_Res_pred ')\r\n for i in range(1,m+1,1):\r\n #print(asav[i], rkeep[i], asavl[i], rkeepl[i])\r\n print(\"%7.2f %9.3f %9.3f %9.3f\" % ( asav[i], rkeep[i],\r\n asavl[i], rkeepl[i]))\r\n\r\n print('plot a lot')\r\n plt.loglog(asav[1:m],rkeep[1:m],'-') # resistivity prediction curve\r\n plt.loglog(adat[1:ndat],pltanswerkeep[1:ndat], 'ro') # predicted data red dots\r\n s=7\r\n plt.loglog(adat[1:ndat],rdat[1:ndat],'bo',markersize=s) #original data blue dots\r\n plt.show()\r\n plt.grid(True)\r\n sys.exit(0)\r\n\r\n"},"license":{"kind":"string","value":"lgpl-3.0"},"hash":{"kind":"number","value":-1006991285408766500,"string":"-1,006,991,285,408,766,500"},"line_mean":{"kind":"number","value":31.4427083333,"string":"31.442708"},"line_max":{"kind":"number","value":120,"string":"120"},"alpha_frac":{"kind":"number","value":0.5328296596,"string":"0.53283"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":2.7052254530130635,"string":"2.705225"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42180,"cells":{"repo_name":{"kind":"string","value":"mikoim/funstuff"},"path":{"kind":"string","value":"null/crawler/tt2db.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1709"},"content":{"kind":"string","value":"# -*- coding: utf-8 -*-\n\nimport urllib.request\nimport time\nimport pymongo\nimport http.client\nimport re\n\ndef httpWrapper(url):\n try:\n data_raw = urllib.request.urlopen(url).read().decode('utf-8')\n except:\n return \"NULL\"\n\n return data_raw\n\ndef getGirlName(data_raw):\n matches = re.findall('名前[ ]+?/[ ]+?(.+?)(|\\n)*( |)*(|\\n)*( |)*(\\(|
)', data_raw)\n\n for match in matches[0]:\n return match.replace(' ', '')\n\n return\n\ndef getGrilPhotos(data_raw):\n matches = re.findall('(photos/.+?.jpg)', data_raw)\n\n if len(matches) == 0:\n matches = re.findall('', data_raw)\n\n return matches\n\ndef getLastModTime(path):\n conn = http.client.HTTPConnection(\"twintail-japan.com\")\n conn.request(\"HEAD\", path)\n res = conn.getresponse()\n return int(time.mktime(time.strptime(res.getheaders()[2][1], '%a, %d %b %Y %H:%M:%S %Z')) * 1000)\n\nconn = pymongo.Connection()\ndb = conn.tw2db\ncol = db.tm\n\nfor x in range(1, 3):\n baseUrl = \"http://twintail-japan.com/sailor/contents/%d.html\" % x\n data_raw = httpWrapper(baseUrl)\n\n if data_raw != \"NULL\":\n name = getGirlName(data_raw)\n\n for photo in getGrilPhotos(data_raw):\n dbtml = {'author' : '', 'time' : '', 'title' : '', 'via' : '', 'src' : '', 'message' : ''}\n dbtml['author'] = name\n dbtml['title'] = name + \" @ セーラ服とツインテール\"\n dbtml['via'] = baseUrl\n dbtml['message'] = \"\"\n dbtml['time'] = getLastModTime(\"/sailor/contents/%d.html\" % x)\n dbtml['src'] = 'http://twintail-japan.com/sailor/contents/%s' % (photo)\n\n col.insert(dbtml)\n\n print(x)"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":-7780927202663628000,"string":"-7,780,927,202,663,628,000"},"line_mean":{"kind":"number","value":26.606557377,"string":"26.606557"},"line_max":{"kind":"number","value":102,"string":"102"},"alpha_frac":{"kind":"number","value":0.5626856803,"string":"0.562686"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":2.9787610619469027,"string":"2.978761"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42181,"cells":{"repo_name":{"kind":"string","value":"digifant/eMonitor"},"path":{"kind":"string","value":"tools/update-osm-data.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"10402"},"content":{"kind":"string","value":"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nimport logging\nimport logging.handlers\nimport traceback\nimport os\nimport time\nfrom optparse import OptionParser\nimport MySQLdb\nimport codecs\nimport requests\nimport sys\nimport pdb\nimport argparse\nfrom pprint import pprint\n\ndef osmWebUrl (lat,lng):\n return \"http://www.openstreetmap.org/?&mlat=%s&mlon=%s&zoom=17\" % (lat,lng)\n\ndef str2bool(v):\n if v.lower() in ('yes', 'true', 't', 'y', '1', 'j', 'ja'):\n return True\n elif v.lower() in ('no', 'false', 'f', 'n', '0', 'nein'):\n return False\n else:\n raise argparse.ArgumentTypeError('Boolean value expected.')\n\ndef prompt(query):\n sys.stdout.write('%s [y/n]: ' % query)\n val = raw_input()\n try:\n ret = str2bool(val)\n except ValueError:\n sys.stdout.write('Please answer with a y/n\\n')\n return prompt(query)\n return ret\n\n# returns None if not found!\ndef queryOsmNominatin(street, streetno, city ):\n url = 'http://nominatim.openstreetmap.org/search'\n params = 'format=json&city={}&street={}'.format(city, street)\n #params = 'format=json&city=%s&street=%s' % (city, address)\n if streetno != '':\n params += ' {}'.format(streetno)\n params = params.replace (' ', '+')\n params = params.replace ('<', '&lt;')\n params = params.replace ('>', '&gt;')\n logging.debug (\"OSM nominatim query: %s?%s\" % (url,params))\n headers = {\n 'User-Agent': 'OSMSyncForFireFighterStreetDbOfOurTown',\n 'From': 'bofhnospam@koffeinbetrieben.de'\n }\n r = requests.get('{}?{}'.format(url, params), timeout=3, headers=headers)\n #logging.debug(\"osm nomination result: %s\" % pprint(r.json()))\n #import pdb; pdb.set_trace()\n _position = None\n try:\n _position = {'lat':r.json()[0]['lat'], 'lng':r.json()[0]['lon'], 'osm_id':r.json()[0]['osm_id'].decode('iso-8859-1').encode('utf8') }\n except IndexError:\n logging.error (\"street %s not found! (housenumber=%s)\" % (street, streetno))\n #logging.debug (_position)\n return _position\n\n\ndef updateMysqlStreets (db, user, passwd, command):\n # Open database connection\n db = MySQLdb.connect(\"localhost\",user,passwd,db )\n\n # prepare a cursor object using cursor() method\n cursor = db.cursor()\n\n # execute SQL query using execute() method.\n cursor.execute(\"SELECT VERSION()\")\n\n # Fetch a single row using fetchone() method.\n data = cursor.fetchone()\n\n print \"Database version : %s \" % data\n\n not_found = {}\n\n if command == \"update_position\":\n sql = \"SELECT * FROM streets\"\n try:\n cursor.execute(sql)\n results = cursor.fetchall()\n for row in results:\n print (\"Street DB %s lat=%s lng=%s\" % (row[1].decode('iso-8859-1').encode('utf8'), row[5], row[6]) )\n if ( row[0] > 0 ):\n _position = queryOsmNominatin (street=row[1].decode('iso-8859-1').encode('utf8'), streetno='', city='Kleinblittersdorf')\n #No heavy uses (an absolute maximum of 1 request per second).\n #http://wiki.openstreetmap.org/wiki/Nominatim_usage_policy\n time.sleep (1)\n if _position != None:\n if row[9] == int(_position['osm_id']):\n sql = 'update streets set lat=%s, lng=%s where id = %s' % (float(_position['lat']), float(_position['lng']), int(row[0]))\n logging.debug (\"sql query %s\" % sql)\n try:\n cursor.execute(sql)\n db.commit()\n logging.info (\"street %s updated lat and lng to (%s,%s)\" % (row[1].decode('iso-8859-1').encode('utf8'), float(_position['lat']), float(_position['lng'])))\n except:\n db.rollback()\n logging.error (\"SQL Error %s\" % traceback.format_exc())\n else:\n logging.fatal (\"OSMID stimmt nicht überein! %s vs %s\" % (row[9], _position['osm_id'] ))\n else:\n logging.fatal (\"OSM nominatin Query failed!\")\n not_found[row[0]] = row[1].decode('iso-8859-1').encode('utf8')\n\n except:\n logging.error (\"DB Error %s\" % traceback.format_exc() )\n\n # disconnect from server\n db.close()\n\n logging.info (\"Sync finished\")\n if len(not_found) > 0:\n logging.error (\"didnt found %s streets:\" % len(not_found))\n for k in not_found.keys():\n logging.error (\"not found: id=%s streetname=%s\" % (k, not_found[k]))\n\n\ndef verifyMysqlStreets (db, user, passwd, command, street=-1):\n # Open database connection\n db = MySQLdb.connect(\"localhost\",user,passwd,db )\n\n # prepare a cursor object using cursor() method\n cursor = db.cursor()\n\n # execute SQL query using execute() method.\n cursor.execute(\"SELECT VERSION()\")\n\n # Fetch a single row using fetchone() method.\n data = cursor.fetchone()\n\n print \"Database version : %s \" % data\n\n not_found = {}\n\n if command == \"verify_streets\":\n sql = \"SELECT * FROM streets\"\n if street > 0:\n sql = sql + \" where id=%i\" % street\n\n try:\n cursor.execute(sql)\n results = cursor.fetchall()\n for row in results:\n print (\"Street %s lat=%s lng=%s url=%s\" % (row[1].decode('iso-8859-1').encode('utf8'), row[5], row[6], osmWebUrl(row[5],row[6]) ) )\n if ( row[0] > 0 ):\n _position = queryOsmNominatin (street=row[1].decode('iso-8859-1').encode('utf8'), streetno='', city='Kleinblittersdorf')\n if _position != None:\n sql = 'update streets set lat=%s, lng=%s, osmid=%s where id = %s' % (float(_position['lat']), float(_position['lng']), int(_position['osm_id']), int(row[0]))\n logging.debug (\"sql query %s\" % sql)\n if row[9] == int(_position['osm_id']):\n logging.info (\"osmid=%s db lat=%s db lng=%s OsmNominatim lat=%s lng=%s new url=%s\" % (row[9], row[5], row[6], float(_position['lat']), float(_position['lng']), osmWebUrl(float(_position['lat']),float(_position['lng'])) ) )\n if round(float(row[5]),4) != round(float(_position['lat']),4) or round(float(row[6]),4) != round(float(_position['lng']),4):\n logging.info (\"%i NO MATCH\" % row[9])\n if options.ask_fix and prompt (\"Fix?\"):\n try:\n cursor.execute(sql)\n db.commit()\n logging.info (\"street %s updated lat, lng, osmid to (%s,%s,%s)\" % (row[1].decode('iso-8859-1').encode('utf8'), float(_position['lat']), float(_position['lng']), (_position['osm_id'])))\n except:\n db.rollback()\n logging.error (\"SQL Error %s\" % traceback.format_exc())\n else:\n logging.info (\"%i MATCH\" % row[9])\n else:\n logging.fatal (\"OSMID stimmt nicht überein! %s vs %s url=%s\" % (row[9], _position['osm_id'], osmWebUrl(float(_position['lat']),float(_position['lng']))))\n if options.ask_fix and prompt (\"Fix?\"):\n try:\n cursor.execute(sql)\n db.commit()\n logging.info (\"street %s updated lat, lng, osmid to (%s,%s,%s)\" % (row[1].decode('iso-8859-1').encode('utf8'), float(_position['lat']), float(_position['lng']), (_position['osm_id'])))\n except:\n db.rollback()\n logging.error (\"SQL Error %s\" % traceback.format_exc())\n else:\n logging.fatal (\"OSM nominatin Query failed!\")\n not_found[row[0]] = row[1].decode('iso-8859-1').encode('utf8')\n\n #No heavy uses (an absolute maximum of 1 request per second).\n #http://wiki.openstreetmap.org/wiki/Nominatim_usage_policy\n time.sleep (1)\n\n\n except:\n logging.error (\"DB Error %s\" % traceback.format_exc() )\n\n # disconnect from server\n db.close()\n\n logging.info (\"verify finished\")\n\n\n\nif __name__ == '__main__':\n parser = OptionParser()\n parser.add_option(\"-d\", \"--database\", dest=\"database\", help=\"mysql database name\", default=\"emonitor\")\n parser.add_option(\"-u\", \"--user\", dest=\"user\", help=\"mysql user\", default='emonitor')\n parser.add_option(\"-p\", \"--passwd\", dest=\"passwd\", help=\"mysql password\", default='emonitor')\n parser.add_option(\"--update-streets-position\", dest=\"update_streets_position\", help=\"update positions for all streets\", action=\"store_true\", default=False)\n parser.add_option(\"--verify-street-position\", dest=\"verify_street_position\", help=\"verify positions for given street\", type=int, default=-1)\n parser.add_option(\"-v\", \"--verify-all-streets-position\", dest=\"verify_all_streets_position\", help=\"verify positions for given street\", action=\"store_true\", default=False)\n parser.add_option(\"-a\", \"--ask-fix\", dest=\"ask_fix\", help=\"ask for fixing\", action=\"store_true\", default=False)\n\n (options, args) = parser.parse_args()\n\n #logging.basicConfig(filename='screenshot-and-telegram.log', level=logging.DEBUG)\n logging.basicConfig(level=logging.DEBUG)\n\n if options.update_streets_position:\n updateMysqlStreets (db=options.database, user=options.user, passwd=options.passwd, command=\"update_position\")\n\n if options.verify_street_position > 0:\n verifyMysqlStreets (db=options.database, user=options.user, passwd=options.passwd, command=\"verify_streets\", street=int(options.verify_street_position))\n\n if options.verify_all_streets_position:\n verifyMysqlStreets (db=options.database, user=options.user, passwd=options.passwd, command=\"verify_streets\")\n\n #queryOsmNominatin(street=\"Rexrothstraße\", streetno='', city='Kleinblittersdorf')\n"},"license":{"kind":"string","value":"bsd-3-clause"},"hash":{"kind":"number","value":-7600303221806536000,"string":"-7,600,303,221,806,536,000"},"line_mean":{"kind":"number","value":45.4241071429,"string":"45.424107"},"line_max":{"kind":"number","value":250,"string":"250"},"alpha_frac":{"kind":"number","value":0.5443792672,"string":"0.544379"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.7814545454545456,"string":"3.781455"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42182,"cells":{"repo_name":{"kind":"string","value":"suma12/asterix"},"path":{"kind":"string","value":"asterix/APDU.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"31348"},"content":{"kind":"string","value":"\"\"\" asterix/APDU.py\n\n__author__ = \"Petr Tobiska\"\n\nAuthor: Petr Tobiska, mailto:petr.tobiska@gmail.com\n\nThis file is part of asterix, a framework for communication with smartcards\n based on pyscard. This file implements handfull APDU commands.\n\nasterix is free software; you can redistribute it and/or modify\nit under the terms of the GNU Lesser General Public License as published by\nthe Free Software Foundation; either version 2.1 of the License, or\n(at your option) any later version.\n\nasterix is distributed in the hope that it will be useful,\nbut WITHOUT ANY WARRANTY; without even the implied warranty of\nMERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\nGNU Lesser General Public License for more details.\n\nYou should have received a copy of the GNU Lesser General Public License\nalong with pyscard; if not, write to the Free Software\nFoundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA\n\nReferences:\n[GP CS] GlobalPlatform Card Specification, Version 2.2.1, Jan 2011\n[GP AmD] Secure Channel Protocol 03, Card Specification v 2.2 - Amendment D,\n Version 1.1, Sep 2009\n\n\"\"\"\n\nimport re\nimport hashlib\nimport random\nfrom struct import pack, unpack\nfrom binascii import hexlify, unhexlify\n# PyCrypto\nfrom Crypto.Cipher import DES, DES3, AES\n# ECSDA\nfrom ecdsa import ecdsa, ellipticcurve\n# pyscard\nfrom smartcard.ATR import ATR\n# asterix\nfrom formutil import s2l, l2s, derLen, derLV, s2int, int2s, s2ECP, chunks,\\\n split2TLV, findTLValue, swapNibbles\nfrom GAF import GAF\nfrom applet import DESsign\nfrom SCP03 import CMAC\nfrom mycard import ISOException, resetCard\n__all__ = ('calcKCV', 'putKey', 'storeDataPutKey',\n 'push2B_DGI', 'X963keyDerivation', 'Push3scenario',\n 'selectApplet', 'openLogCh', 'closeLogCh',\n 'getStatus', 'getExtCardRes', 'getData',\n 'selectFile', 'readBinary', 'readRecord',\n 'updateBinary', 'updateRecord',\n 'verifyPin', 'changePin', 'disablePin', 'enablePin', 'unblockPin',\n 'selectUSIM', 'cardInfo', 'KeyType')\n\nINS_VERIFY_PIN = 0x20\nINS_CHANGE_PIN = 0x24\nINS_DISABLE_PIN = 0x26\nINS_ENABLE_PIN = 0x28\nINS_UNBLOCK_PIN = 0x2C\nINS_MANAGE_LOGCH = 0x70\nINS_SELECT = 0xA4\nINS_READBIN = 0xB0\nINS_READREC = 0xB2\nINS_GETDATA = 0xCA \nINS_UPDBIN = 0xD6\nINS_UPDREC = 0xDC\nINS_PUTKEY = 0xD8\nINS_STOREDATA = 0xE2\nINS_GETSTATUS = 0xF2\n\n\nclass KeyType:\n \"\"\"Key types as defined in [GP CS] Tab 11.16\"\"\"\n # subset of currently supported keys\n DES_IMPLICIT = 0x80\n TDES_CBC = 0x82\n DES_ECB = 0x83\n DES_CBC = 0x84\n AES = 0x88\n\n\ndef calcKCV(keyValue, zAES=False):\n \"\"\"Calculate KCV for symmetric key.\nkeyValue - key values as string (DES, 3DES2k, 3DES3k, AES)\nzAES - True if key is AES (i.e. encrypt block of '01' instead of '00')\n\nReturn 3B-long string.\"\"\"\n if zAES:\n assert len(keyValue) in (16, 24, 32), \"Wrong length of AES key\"\n block = '\\x01'*16\n tkey = AES.new(keyValue, AES.MODE_ECB)\n else:\n assert len(keyValue) in (8, 16, 24), \"Wrong length of (3)DES key\"\n block = '\\x00'*8\n if len(keyValue) == 8:\n tkey = DES.new(keyValue, DES.MODE_ECB)\n else:\n tkey = DES3.new(keyValue, DES.MODE_ECB)\n return tkey.encrypt(block)[:3]\n\n\ndef putKey(oldKeyVersion, newKeyVersion, keyId, keyComponents,\n zMoreCmd=False, zMultiKey=False, keyDEK=None,\n lenMAC=8):\n \"\"\"Build APDU for PUT KEY command.\noldKeyVersion - key version to be replaced. If zero, new key is created.\nnewKeyVersion - key version of key being put\nkeyId - id of the 1st key being put\nkeyComponents - list of key components being put.\n Each componet is a tuple of key type (u8) and value (string).\nzMoreCmd - P1.b8, signals if there is more commands\nzMultiKey - P2.b8, signals if more than one component being put\nkeyDEK - KIK or DEK key. keyDEK.encrypt(data) called to encrypt\n (including padding) key component value if not None.\n If has attribute zAES and keyDEK.zAES evaluates as True, it is\n considered as AES key and [GP AmD] 7.2 formatting is used.\nlenMAC - length of CMAC for AES.\n Applicable if AES key with key id=0x02 (KID) and\n key version 0x01-0x0F or 0x11 is being put with AES keyDEK\n (see ETSI 102.226 rel 9+, 8.2.1.5)\n\nReturns APDU built (as list of u8).\n\nSee [GP CS] 11.8 and [GP AmD] 7.2 for reference.\nSee [GP CS] Tab 11.16 for coding of key type.\nCurrently only Format1 supported.\n\"\"\"\n # sanity check\n assert 0 <= oldKeyVersion < 0x80\n assert 0 < newKeyVersion < 0x80\n assert 0 < keyId < 0x80\n assert len(keyComponents) > 0\n assert lenMAC in (4, 8)\n\n P1 = (zMoreCmd and 0x80 or 0) | oldKeyVersion\n P2 = (zMultiKey and 0x80 or 0) | keyId\n\n data = chr(newKeyVersion)\n for kc in keyComponents:\n keyType, keyVal = kc[:2] # ignore eventual keyUsage and keyAccess\n assert 0 <= keyType < 0xFF\n if keyDEK:\n encValue = keyDEK.encrypt(keyVal)\n # for AES as keyDEK, prepend length of component\n if 'zAES' in dir(keyDEK) and keyDEK.zAES:\n encValue = derLen(keyVal) + encValue\n # see ETSI 102.226 rel 9+, 8.2.1.5\n if keyType == KeyType.AES and keyId == 2 and \\\n newKeyVersion in range(0x01, 0x10) + [0x11]:\n encValue += chr(lenMAC)\n else:\n encValue = keyVal\n # calculate KCV\n if keyType in (KeyType.DES_IMPLICIT, KeyType.TDES_CBC,\n KeyType.DES_ECB, KeyType.DES_CBC, KeyType.AES):\n kcv = calcKCV(keyVal, keyType == KeyType.AES)\n else:\n kcv = ''\n\n data += chr(keyType) + derLen(encValue) + encValue + derLen(kcv) + kcv\n keyId += 1\n\n apdu = [0x80, INS_PUTKEY, P1, P2, len(data)] + s2l(data)\n return apdu\n\n\ndef push2B_DGI(keyVer, keys, keyCASDenc):\n \"\"\" Create DGI 00A6 and 8010 for Push2B scenario\nkeyVer - key verions (u8)\nkeys - ((keytype, keyvalue)); 1 or 3 sym. keys\nkeyCASDenc - a method to call for encryption 8010 content\nReturn DGIs built (as list of strings).\"\"\"\n # DGI tag on 2B (GP Card Spec 2.2.1, 11.1.12)\n # DGI length coding as in GP Systems Scripting Language Spec. v1.1.0, an. B\n # i.e. on 1B for x < 255, FF for x >=255\n KAT = GAF(\"\"\" -- Control Reference Template (KAT)\n -- see GP 2.2.1 AmA 4.4\n 00A6 #[\n A6 #(\n 90 #(04) -- scenario identifier: Push#2B\n 95 #($keyUsage)\n 80 #($keyType)\n 81 #($keyLen)\n 83 #($keyVer)\n -- 45 #($SDIN) -- optional Security Domain Image Number\n )] \"\"\")\n assert len(keys) in (1, 3), \"One or three sym. keys expected\"\n keyUsage = len(keys) == 1 and '\\x5C' or '\\x10' # Tab. 13\n keyType = keys[0][0]\n assert all([k[0] == keyType for k in keys]), \"Key types differ\"\n # remap keyType to '80' as required by GP UICC config 10.3.1\n if keyType in (KeyType.TDES_CBC, KeyType.DES_ECB, KeyType.DES_CBC):\n keyType = KeyType.DES_IMPLICIT\n lens = [len(k[1]) for k in keys]\n l = max(lens)\n assert l == min(lens), \"Key lengths differ\"\n dgi00A6 = KAT.eval(keyUsage=keyUsage, keyType=chr(keyType),\n keyLen=chr(l), keyVer=chr(keyVer))\n data = keyCASDenc(''.join([k[1] for k in keys]))\n dgi8010 = pack(\">H\", 0x8010) + chr(len(data)) + data\n return (dgi00A6, dgi8010)\n\n\ndef storeDataPutKeyDGI(keyVer, keyComponents, keyId=1, keyDEK=None):\n \"\"\"Build DGI for Store Data for Put Key.\nkeyVer - key version of key being created\nkeyComponents - list of key components being put.\n Each componet is a tuple of key type (u8), value (string)\n and optionally Key Usage Qualifier and Key Access\n (u8, defaults 0x18, 0x14 or 0x48 for key UQ, 0x00 for key ac.)\nkeyId - id of the 1st key being created (optional, u8, default 1)\nkeyDEK - KIK or DEK key. keyDEK.encrypt(data) called to encrypt\n (including padding) key component value if not None.\n If has attribute zAES and keyDEK.zAES evaluates as True, it is\n considered as AES key and [GP AmD] 7.2 formatting is used.\n\nReturns DGIs built (as list of string).\nSee GP 2.2.1 AmA 4.10.2 for reference.\n\"\"\"\n # sanity check\n assert 0 < keyVer and keyVer < 0x80\n assert 0 < keyId and keyId < 0x80\n assert len(keyComponents) > 0\n\n KeyUQ = (None, 0x38, 0x34, 0xC8) # see GP 2.2.1, 11.1.9\n templ = \"\"\" B9 #(95#($keyUQ) 96#($keyAc) 80#($keyType) 81#($keyLen)\n 82#($keyId) 83#($keyVer) 84#($KCV))\"\"\"\n d = {'keyVer': chr(keyVer)}\n B9 = ''\n dgi8113 = []\n for kc in keyComponents:\n assert len(kc) in (2, 4), \"wrong keyComponent\" + kc.__str__()\n if len(kc) == 2:\n keyType, keyVal = kc\n keyUQ = 1 <= keyId <= 3 and KeyUQ[keyId] or 0xFF\n keyAc = 0x00\n else:\n keyType, keyVal, keyUQ, keyAc = kc\n d['keyLen'] = chr(len(keyVal))\n assert 0 <= keyType < 0xFF\n if keyType in (KeyType.DES_IMPLICIT, KeyType.TDES_CBC,\n KeyType.DES_ECB, KeyType.DES_CBC, KeyType.AES):\n d['KCV'] = calcKCV(keyVal, keyType == KeyType.AES)\n else:\n d['KCV'] = ''\n d['keyId'] = chr(keyId)\n for k in ('keyType', 'keyUQ', 'keyAc', 'keyId'):\n d[k] = chr(locals()[k])\n tlv = GAF(templ).eval(**d)\n if keyDEK:\n encValue = keyDEK.encrypt(keyVal)\n else:\n encValue = keyVal\n B9 += tlv\n dgi8113.append(pack(\">HB\", 0x8113, len(encValue)) + encValue)\n keyId += 1\n return(pack(\">HB\", 0x00B9, len(B9)) + B9, dgi8113)\n\n\ndef storeDataPutKey(keyVer, keyComponents, keyId=1, keyDEK=None):\n \"\"\"Build APDU for Store Data for Put Key.\nkeyVer, keyComponents, keyId and keyDEK as in storeDataPutKeyDGI.\nReturn APDU a u8 list.\"\"\"\n dgi00B9, dgi8113 = storeDataPutKeyDGI(keyVer, keyComponents,\n keyId, keyDEK)\n data = dgi00B9 + ''.join(dgi8113)\n assert len(data) < 256, \"Longer Put Key not implemented\"\n P1 = 0x88\n P2 = 0\n apdu = [0x80, INS_STOREDATA, P1, P2, len(data)] + s2l(data)\n return apdu\n\n# ###### Scenario 3 stuff\n# Preloaded ECC Curve Parameters, GP 2.2.1 AmE 4.5\n# N.B., all have cofactor = 1\nECC_Curves = {\n 0x00: ecdsa.generator_256, # NIST P-256\n 0x01: ecdsa.generator_384, # NIST P-384\n 0x02: ecdsa.generator_521, # NIST P-521\n # 0x03: brainpoolP256r1,\n # 0x04: brainpoolP256t1,\n # 0x05: brainpoolP384r1,\n # 0x06: brainpoolP384t1,\n # 0x07: brainpoolP512r1,\n # 0x08: brainpoolP512t1,\n}\n\n# tag definition\nT_IIN = 0x42\nT_SDIN = T_CIN = 0x45\nT_keyType = 0x80\nT_keyLen = 0x81\nT_keyID = 0x82\nT_keyVer = 0x83\nT_DR = 0x85\nT_HostID = 0x84\nT_receipt = 0x86\nT_scenarioID = 0x90\nT_seqCounter = 0x91\nT_keyUsage = 0x95\nT_keyAcc = 0x96\nT_CRT = 0xA6\n\n\ndef X963keyDerivation(sharedSecret, bytelen, sharedInfo='',\n h = hashlib.sha256):\n \"\"\" X9.63 Key Derivation Function as deifned in TR-03111 4.3.3\nbytelen - expected length of Key Data\nsharedSecret, sharedInfo - strings\nh - function to create HASH object (default hashlib.sha256)\nReturn Key Data (string)\nReference: TR-03111: BSI TR-03111 Elliptic Curve Cryptography, Version 2.0\n https://www.bsi.bund.de/SharedDocs/Downloads/EN/BSI/Publications/TechGuidelines/TR03111/BSI-TR-03111_pdf.html\"\"\"\n keyData = ''\n l = h().digest_size\n j = (bytelen - 1)/l + 1\n for i in xrange(1, 1+j):\n keyData += h(sharedSecret + pack(\">L\", i) + sharedInfo).digest()\n return keyData[:bytelen]\n\n\ndef DESMAC(key, data):\n \"\"\" Calculate MAC single DES with final 3DES\"\"\"\n return DESsign(key).calc(data)\n\nktDES = KeyType.DES_IMPLICIT\nktAES = KeyType.AES\n\n\nclass Push3scenario:\n \"\"\" Implementation of Global Platform Push #3 scenario (ECKA)\"\"\"\n def __init__(self, keyParRef, pkCASD, **kw):\n \"\"\" Constructor\nkeyParRef - Key Parameter Reference\npkCASD - PK.CASD.ECKA (tuple long x, long y)\noptional **kw: IIN, CIN (as strings)\"\"\"\n assert keyParRef in ECC_Curves, \\\n \"Unknown Key param reference 0x%02X\" % keyParRef\n self.keyParRef = keyParRef\n self.generator = ECC_Curves[keyParRef]\n self.curve = self.generator.curve()\n self.bytelen = len(int2s(self.curve.p()))\n assert self.bytelen in (32, 48, 64, 66) # currently allowed keys\n pkCASDxy = s2ECP(pkCASD)\n assert self.curve.contains_point(*pkCASDxy),\\\n \"PK.CASD.ECKA not on the curve\"\n self.pkCASD = ellipticcurve.Point(self.curve, *pkCASDxy)\n for k in ('IIN', 'CIN'):\n if k in kw:\n assert isinstance(kw[k], str)\n self.__dict__[k] = kw[k]\n\n def makeDGI(self, keyVer, privkey=None,\n keys=([(KeyType.AES, 16)]*3),\n zDelete=False, zDR=False, zID=False, **kw):\n \"\"\" Prepare data for Push #3 scenario and generate keys.\nkeyVer - key version to create\nprivkey - eSK.AP.ECKA (secret multiplier as string)\n randomly generated if None\nkeys - [(keyType, keyLen)] to generate\nzDelete, zDR, zID - bits 1-3 of Parameters of scenario, (GP AmE, Tab. 4-17)\noptional **kw: keyId, seqCounter, SDIN, HostID\nReturn \"\"\"\n if privkey is None:\n secexp = random.randrange(2, self.generator.order())\n else:\n secexp = s2int(privkey)\n assert 1 < secexp < self.generator.order(), \"Wrong eSK.AP.ECKA\"\n print \"eSK.AP.ECKA = %X\" % secexp\n pubkey = self.generator * secexp\n dgi7F49 = pack(\">HBB\", 0x7F49, 2*self.bytelen+1, 4) + \\\n int2s(pubkey.x(), self.bytelen * 8) + \\\n int2s(pubkey.y(), self.bytelen * 8)\n # calculate Shared Secret, suppose that cofactor is 1\n S_AB = secexp * self.pkCASD\n self.sharedSecret = int2s(S_AB.x(), self.bytelen * 8)\n print \"Shared Secret =\", hexlify(self.sharedSecret).upper()\n # build DGI 00A6\n if zID:\n assert hasattr(self, 'IIN'), \"Missing IIN while CardId requested\"\n assert hasattr(self, 'CIN'), \"Missing cIN while CardId requested\"\n assert 'HostID' in kw and isinstance(kw['HostID'], str)\n self.HostCardID = ''.join([derLV(v) for v in\n (kw['HostID'], self.IIN, self.CIN)])\n else:\n self.HostCardID = ''\n self.zDR = zDR\n scenarioPar = (zDelete and 1 or 0) +\\\n (zDR and 2 or 0) +\\\n (zID and 4 or 0)\n assert all([k[0] in (KeyType.DES_IMPLICIT, KeyType.AES) for k in keys])\n ktl1 = keys[0]\n zDifKey = any([keys[i] != ktl1 for i in xrange(1, len(keys))])\n tA6value = pack(\"BBBB\", T_scenarioID, 2, 3, scenarioPar)\n if zDifKey:\n self.receiptAlgo = CMAC\n self.keyLens = [16] + [k[1] for k in keys]\n self.keyDesc = ''\n if 'keyId' in kw:\n tA6value += pack(\"BBB\", T_keyID, 1, kw['keyId'])\n tA6value += pack(\"BBB\", T_keyVer, 1, keyVer)\n # default keyUsage from GP 2.2.1 AmE tab. 4-16 for ENC, MAC, DEK\n for k, keyUsage in zip(keys, (0x38, 0x34, 0xC8)):\n if len(k) > 2:\n keyUsage = k[2]\n tB9value = pack(\"BBB\", T_keyUsage, 1, keyUsage)\n if len(k) >= 4: # optional key Access as fourth elem. of key\n tB9value += pack(\"BBB\", T_keyAcc, 1, k[3])\n tB9value += pack(\"BBB\", T_keyType, 1, k[0])\n tB9value += pack(\"BBB\", T_keyLen, 1, k[1])\n self.keyDesc += pack(\"BBB\", keyUsage, *k[:2])\n tA6value += '\\xB9' + derLV(tB9value)\n else:\n assert len(keys) in (1, 3), \\\n \"One or three secure ch. keys expected.\"\n self.keyLens = [ktl1[1]] * (1 + len(keys))\n self.receiptAlgo = ktl1[0] == KeyType.AES and CMAC or DESMAC\n keyUsage = len(keys) == 1 and 0x5C or 0x10\n self.keyDesc = pack(\"BBB\", keyUsage, *ktl1[:2])\n tA6value += pack(\"BBB\", T_keyUsage, 1, keyUsage)\n if len(ktl1) == 4:\n tA6value += pack(\"BBB\", T_keyAcc, 1, ktl1[3])\n tA6value += pack(\"BBB\", T_keyType, 1, ktl1[0])\n tA6value += pack(\"BBB\", T_keyLen, 1, ktl1[1])\n if 'keyId' in kw:\n tA6value += pack(\"BBB\", T_keyID, 1, kw['keyId'])\n tA6value += pack(\"BBB\", T_keyVer, 1, keyVer)\n if 'seqCounter' in kw:\n tA6value += chr(T_seqCounter) + derLV(kw['seqCounter'])\n if 'SDIN' in kw:\n tA6value += chr(T_SDIN) + derLV(kw['SDIN'])\n if zID:\n tA6value += chr(T_HostID) + derLV(kw['HostID'])\n self.tA6 = chr(T_CRT) + derLV(tA6value)\n dgi00A6 = pack(\">HB\", 0x00A6, len(self.tA6)) + self.tA6\n return (dgi00A6, dgi7F49)\n\n def generKeys(self, respData):\n \"\"\" Verify receipt and generate symmetric keys.\nrespData - response to Store Data (string)\nReturn generated keys (tuple of strings)\"\"\"\n try:\n data2rec = self.tA6\n except KeyError:\n print \"Run makeDGI first\"\n return\n respTLV = split2TLV(respData)\n if self.zDR:\n lenDR = (self.bytelen // 32) * 16 # map to 16, 24 or 32\n DR = respTLV[0][1]\n assert len(respTLV) == 2 and \\\n respTLV[0][0] == T_DR and len(DR) == lenDR\n data2rec += pack(\"BB\", T_DR, lenDR) + DR\n else:\n assert len(respTLV) == 1\n assert respTLV[-1][0] == T_receipt\n receipt = respTLV[-1][1]\n\n sharedInfo = self.keyDesc\n if self.zDR:\n sharedInfo += DR\n if hasattr(self, 'HostCardID'):\n sharedInfo += self.HostCardID\n print \"Shared Info =\", hexlify(sharedInfo).upper()\n\n keyData = X963keyDerivation(self.sharedSecret, sum(self.keyLens),\n sharedInfo)\n keyDataIt = chunks(keyData, self.keyLens)\n receiptKey = keyDataIt.next()\n print \"Receipt Key =\", hexlify(receiptKey).upper()\n expReceipt = self.receiptAlgo(receiptKey, data2rec)\n assert receipt == expReceipt, \"Receipt verification failed\"\n return [k for k in keyDataIt if k] # skip empty rest\n\n\ndef selectApplet(c, AID, logCh=0):\n \"\"\" Select applet on a given logical channel or\nopen new log. channel if logCh is None. \"\"\"\n if logCh is None:\n logCh = openLogCh(c)\n # select the Applet on the given logical channel\n apdu = [logCh, INS_SELECT, 4, 0, len(AID)] + s2l(AID)\n resp, sw1, sw2 = c.transmit(apdu)\n if sw1 == 0x6C and len(AID) == 0:\n apdu = [logCh, INS_SELECT, 4, 0, sw2]\n resp, sw1, sw2 = c.transmit(apdu)\n if(sw1 == 0x61):\n apdu = [logCh, 0xC0, 0, 0, sw2]\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n respdata = l2s(resp)\n # close channel\n return (respdata, logCh)\n\n\ndef openLogCh(c):\n \"\"\" Manage channel to open logical channel. \"\"\"\n apdu = [0, INS_MANAGE_LOGCH, 0, 0, 1]\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n return resp[0]\n\n\ndef closeLogCh(c, logCh):\n apdu = [0, INS_MANAGE_LOGCH, 0x80, logCh, 0]\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n\n\nclass GetStatusData:\n \"\"\"Represent and interpret data from Get status for Packages and Modules\"\"\"\n def __init__(self, respdataPM, respdataApp):\n ind = 0\n self.packages = []\n while len(respdataPM) > ind:\n length = respdataPM[ind]\n pack_aid = l2s(respdataPM[ind+1: ind+1+length])\n ind += length + 1\n lcs = respdataPM[ind]\n priv = respdataPM[ind+1]\n nmod = respdataPM[ind+2]\n ind += 3\n mods = []\n for i in xrange(nmod):\n length = respdataPM[ind]\n mods.append(l2s(respdataPM[ind+1: ind+1+length]))\n ind += length + 1\n self.packages.append({'pack_aid': pack_aid,\n 'lcs': lcs,\n 'priv': priv,\n 'modules': mods})\n ind = 0\n self.insts = []\n while len(respdataApp) > ind:\n length = respdataApp[ind]\n app_aid = l2s(respdataApp[ind+1: ind+1+length])\n ind += length + 1\n lcs = respdataApp[ind]\n priv = respdataApp[ind+1]\n ind += 2\n self.insts.append({'app_aid': app_aid,\n 'lcs': lcs,\n 'priv': priv})\n\n def __str__(self):\n res = ''\n for p in self.packages:\n res += \"Package AID: %s %02X %02X\\n\" % \\\n (hexlify(p['pack_aid']).upper().ljust(32),\n p['lcs'], p['priv'])\n for m in p['modules']:\n res += \" module %s\\n\" % hexlify(m).upper().ljust(32)\n for p in self.insts:\n res += \"Insts AID : %s %02X %02X\\n\" % \\\n (hexlify(p['app_aid']).upper().ljust(32),\n p['lcs'], p['priv'])\n return res\n\n\ndef getStatus(sc, AID_pref=''):\n \"\"\" Issue GET STATUS apdu for packages and modules, and instances. \"\"\"\n res = {}\n for P1 in (0x10, 0x40):\n apdu = [0x80, INS_GETSTATUS, P1, 0, 2+len(AID_pref), 0x4F,\n len(AID_pref)] + s2l(AID_pref)\n respdata, sw1, sw2 = sc.transmit(apdu)\n sw = (sw1 << 8) + sw2\n while sw == 0x6310:\n apdu = [0x80, INS_GETSTATUS, P1, 1, 2+len(AID_pref), 0x4F,\n len(AID_pref)] + s2l(AID_pref)\n resp, sw1, sw2 = sc.transmit(apdu)\n respdata += resp\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n res[P1] = respdata\n return GetStatusData(res[0x10], res[0x40])\n\n\ndef getData(c, tag):\n P1 = tag >> 8\n P2 = tag & 0xFF\n apdu = [0x80, INS_GETDATA, P1, P2, 0]\n resp, sw1, sw2 = c.transmit(apdu)\n if sw1 == 0x6C:\n apdu[4] = sw2\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n return l2s(resp)\n\n\ndef getExtCardRes(c):\n \"\"\" Issue GET DATA with tag FF21 in order to receive Extended\nCard Resources (GP 2.2.1, 11.3 & ETSI TS 102.226, 8.2.1.7).\nReturns [num. of install applets, free NVM, free RAM]\"\"\"\n # CLA = 0x00: return only value\n # CLA = 0x80: return TLV, i.e. 0xFF21 #(value)\n apdu = [0x80, INS_GETDATA, 0xFF, 0x21, 0]\n resp, sw1, sw2 = c.transmit(apdu)\n if sw1 == 0x6C:\n apdu[4] = sw2\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n payload = l2s(resp)\n result = [s2int(findTLValue(payload, (0xFF21, tag))) for\n tag in (0x81, 0x82, 0x83)]\n return result\n\n\ndef selectFile(c, path, logCh=0):\n \"\"\" Select file by path from MF or MF for empty path \"\"\"\n if len(path) > 0:\n apdu = [logCh, INS_SELECT, 8, 4, len(path)] + s2l(path)\n else:\n apdu = [logCh, INS_SELECT, 0, 4, 2, 0x3F, 0x00]\n resp, sw1, sw2 = c.transmit(apdu)\n if sw1 == 0x61:\n resp, sw1, sw2 = c.transmit([0, 0xC0, 0, 0, sw2])\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n return l2s(resp)\n\n\ndef readBinary(c, le, logCh=0, offset=0):\n \"\"\"Read Binary on currently selected EF\"\"\"\n P1 = (offset >> 8) & 0x7F\n P2 = offset & 0xFF\n apdu = [logCh, INS_READBIN, P1, P2, le]\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n return l2s(resp)\n\n\ndef readRecord(c, recNum, logCh=0):\n \"\"\" Read record from currently selected EF\"\"\"\n apdu = [logCh, INS_READREC, recNum, 4, 0]\n resp, sw1, sw2 = c.transmit(apdu)\n if sw1 == 0x6C:\n apdu[4] = sw2\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n return l2s(resp)\n\ndef updateBinary(c, data, logCh=0, offset=0):\n \"\"\"Update binary on currently selected EF\"\"\"\n assert len(data) < 0x100\n P1 = (offset >> 8) & 0x7F\n P2 = offset & 0xFF\n apdu = [logCh, INS_UPDBIN, P1, P2, len(data)] + s2l(data)\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n\n\ndef updateRecord(c, recNum, data, logCh=0):\n \"\"\" Update record from currently selected EF\"\"\"\n assert len(data) < 0x100\n apdu = [logCh, INS_UPDREC, recNum, 4, len(data)] + s2l(data)\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n\ndef verifyPin(c, pin=None, P2=0x01, logCh=0):\n \"\"\"Verify PIN\npin - value (str, 4-8bytes). If None, just get number of tries.\nP2 - PIN identification (0x01: PIN1 (default), 0x81: PIN2, etc.)\nlogCh - logical channel (default 0)\nReturn number of remaing tries or True if verification succesfull.\n\"\"\"\n lc = 0 if pin is None else 8\n apdu = [logCh, INS_VERIFY_PIN, 0, P2, lc]\n if pin is not None:\n assert 4 <= len(pin) <= 8\n pin += '\\xFF' * (8 - len(pin))\n apdu += s2l(pin)\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw == 0x6983: # PIN blocked\n return 0\n if 0x63C0 <= sw <= 0x63CA: # remaining tries\n return sw - 0x63C0\n if sw != 0x9000:\n raise ISOException(sw)\n return True # pin verified\n\ndef changePin(c, oldPin, newPin, P2=0x01, logCh=0):\n \"\"\"Change PIN\noldPin - old PIN value (str, 4-8bytes)\nnewPin - new PIN value (str, 4-8bytes)\nP2 - PIN identification (0x01: PIN1 (default), 0x81: PIN2, etc.)\nlogCh - logical channel (default 0)\n\"\"\"\n assert 4 <= len(oldPin) <= 8\n oldPin += '\\xFF' * (8 - len(oldPin))\n assert 4 <= len(newPin) <= 8\n newPin += '\\xFF' * (8 - len(newPin))\n apdu = [logCh, INS_CHANGE_PIN, 0, P2, 0x10] + s2l(oldPin) + s2l(newPin)\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n\ndef disablePin(c, pin, P2=0x01, logCh=0):\n \"\"\"Disable PIN\npin - PIN value (str, 4-8bytes)\nP2 - PIN identification (0x01: PIN1 (default), 0x81: PIN2, etc.)\nlogCh - logical channel (default 0)\n\"\"\"\n assert 4 <= len(pin) <= 8\n pin += '\\xFF' * (8 - len(pin))\n apdu = [logCh, INS_DISABLE_PIN, 0, P2, 8] + s2l(pin)\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n \ndef enablePin(c, pin, P2=0x01, logCh=0):\n \"\"\"Enable PIN\npin - PIN value (str, 4-8bytes)\nP2 - PIN identification (0x01: PIN1 (default), 0x81: PIN2, etc.)\nlogCh - logical channel (default 0)\n\"\"\"\n assert 4 <= len(pin) <= 8\n pin += '\\xFF' * (8 - len(pin))\n apdu = [logCh, INS_ENABLE_PIN, 0, P2, 8] + s2l(pin)\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n \ndef unblockPin(c, puk, newPin, P2=0x01, logCh=0):\n \"\"\"unblock PIN\npuk - new PIN value (str, 4-8bytes)\nnewPin - PIN value (str, 4-8bytes)\nP2 - PIN identification (0x01: PIN1 (default), 0x81: PIN2, etc.)\nlogCh - logical channel (default 0)\n\"\"\"\n assert len(puk) == 8\n assert 4 <= len(newPin) <= 8\n newPin += '\\xFF' * (8 - len(newPin))\n apdu = [logCh, INS_UNBLOCK_PIN, 0, P2, 0x10] + s2l(puk) + s2l(newPin)\n resp, sw1, sw2 = c.transmit(apdu)\n sw = (sw1 << 8) + sw2\n if sw != 0x9000:\n raise ISOException(sw)\n\ndef selectUSIM(c, logCh=0):\n \"\"\"Select USIM, return AID\nRead EF_DIR, USIM = first application with AID of USIM (3GPP TS 31.110)\"\"\"\n # read EF_DIR\n infoDIR = selectFile(c, unhexlify('2F00'), logCh)\n # see ETSI 102.221 11.1.1.4.3 for coding\n fileDesc = findTLValue(infoDIR, (0x62, 0x82))\n assert len(fileDesc) == 5 and \\\n fileDesc[:2] == '\\x42\\x21' # linear EF\n recLen, nRec = unpack(\">HB\", fileDesc[2:5])\n aids = []\n for recNum in xrange(1, nRec+1):\n try:\n r = readRecord(c, recNum)\n if r == '\\xFF' * len(r):\n continue\n aid = findTLValue(r, (0x61, 0x4F))\n aids.append(aid)\n except ISOException:\n break\n # search for USIM\n for aid in aids:\n if aid[:7] == unhexlify('A0000000871002'):\n infoUSIM = selectApplet(c, aid, logCh)\n return aid\n return None\n\ndef cardInfo(c, USIMpin=None, logCh=0):\n \"\"\"Deselect, read EF_DIR, EF_ICCID\"\"\"\n resetCard(c)\n histBytes = l2s(ATR(c.getATR()).getHistoricalBytes())\n infoMF = selectFile(c, '', logCh)\n # read EF_ICCID\n infoICCID = selectFile(c, unhexlify('2FE2'), logCh)\n fileSize = s2int(findTLValue(infoICCID, (0x62, 0x80)))\n assert fileSize == 10, \"Wrong size of EF_ICCID\"\n iccid = swapNibbles(readBinary(c, fileSize))\n # read EF_DIR\n infoDIR = selectFile(c, unhexlify('2F00'), logCh)\n # see ETSI 102.221 11.1.1.4.3 for coding\n fileDesc = findTLValue(infoDIR, (0x62, 0x82))\n assert len(fileDesc) == 5 and \\\n fileDesc[:2] == '\\x42\\x21' # linear EF\n recLen, nRec = unpack(\">HB\", fileDesc[2:5])\n dirDO = []\n for recNum in xrange(1, nRec+1):\n try:\n r = readRecord(c, recNum)\n if r == '\\xFF' * len(r):\n continue\n aid = findTLValue(r, (0x61, 0x4F))\n label = findTLValue(r, (0x61, 0x50))\n dirDO.append({'AID': aid, 'label': label})\n except ISOException:\n break\n # select USIM and try to read IMSI\n aids = [DO['AID'] for DO in dirDO\n if DO['AID'][:7] == unhexlify('A0000000871002')]\n if len(aids) >= 1:\n aid_usim = aids[0] # choose the first AID found\n else:\n aid_usim = None\n if aid_usim:\n infoUSIM = selectApplet(c, aid_usim, logCh)\n if USIMpin is not None:\n verifyPin(c, USIMpin, logCh=logCh)\n infoIMSI = selectFile(c, unhexlify('7FFF6F07'), logCh)\n try:\n bimsi = readBinary(c, 9, logCh)\n digits = reduce(lambda d, n: d + [ord(n) & 0x0F, ord(n) >> 4],\n bimsi[1:1+ord(bimsi[0])], [])\n digits.pop(0) # remove first nibble 8 or 9\n while digits[-1] == 0x0F:\n digits.pop() # remove trailing F\n imsi = ''.join([chr(ord('0')+i) for i in digits])\n except ISOException:\n imsi = None\n else:\n imsi = None\n # select default applet and get tags 45 and 42\n selectApplet(c, '', logCh)\n try:\n iin = findTLValue(getData(c, T_IIN), (T_IIN,))\n except ISOException:\n iin = None\n try:\n cin = findTLValue(getData(c, T_CIN), (T_CIN,))\n except ISOException:\n cin = None\n return histBytes, iccid, dirDO, imsi, iin, cin\n"},"license":{"kind":"string","value":"lgpl-2.1"},"hash":{"kind":"number","value":-262022843162776060,"string":"-262,022,843,162,776,060"},"line_mean":{"kind":"number","value":35.6214953271,"string":"35.621495"},"line_max":{"kind":"number","value":115,"string":"115"},"alpha_frac":{"kind":"number","value":0.5717111139,"string":"0.571711"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":2.99150682317015,"string":"2.991507"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42183,"cells":{"repo_name":{"kind":"string","value":"xerond/lucia"},"path":{"kind":"string","value":"ledEditor/cfilegen.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"2560"},"content":{"kind":"string","value":"from effectgroup import EffectGroup\nfrom effectdescriptions import EffectDescriptions\nfrom myutils import Utils\n\ndef generateFile(fileName,ledCount,effectGroups):\n\tf = open(fileName,'w')\n\tf.write(\"#ifndef H_SONG_INSTRUCTIONS\\n#define H_SONG_INSTRUCTIONS\\n#include \\\"avr/pgmspace.h\\\"\\n#include \\\"song_instructions.h\\\"\\nconst char song_instructions[] PROGMEM = {\")\n\tlastTime = 0\n\tfor curEffectGroup in effectGroups:\n\t\twriteBuffer = \"\"\n\t\tnewTime = curEffectGroup.getTimeAs10msCount()\n\t\ttD = newTime - lastTime\n\t\tlastTime = newTime\n\t\twriteBuffer += \"0xff,\\n\"\n\t\twriteBuffer += Utils.short_to_hex(tD) + \"\\n\"\n\t\tfor ledIndex in range (0,ledCount):\n\t\t\tledEffect = curEffectGroup.getLedEffect(ledIndex)\n\t\t\ttempLedBytes = generateLedEffectBytes(ledIndex,ledEffect)\n\t\t\tif tempLedBytes <> \"\":\n\t\t\t\twriteBuffer += \"\\t\" + generateLedEffectBytes(ledIndex,ledEffect) + \"\\n\"\n\t\twriteBuffer += \"0xff,\\n\"\n\t\tf.write(writeBuffer)\n\tf.write(\"0x00,};\\n#endif\")\n\n\n#generates a string for led effect\ndef generateLedEffectBytes(ledNumber,ledEffect):\n\teffectNumber = ledEffect[EffectGroup.INDEX_EFFECT_NUMBER]\n\t\n\t#get the real effect number \n\t#TODO we are accessing a global here, eek!\n\tprint \"Effect num is: \" + str(effectNumber)\n\trealEffectNumber = EffectDescriptions.quickEffectLookup[effectNumber]['realId']\n\n\n\teffectData = ledEffect[EffectGroup.INDEX_EFFECT_DATA]\n\t\n\t#if effect number is < 0, ignore it\n\tif effectNumber < 0:\n\t\treturn \"\"\n\n\treturnStr = Utils.byte_to_hex(ledNumber) + Utils.byte_to_hex(realEffectNumber)\n\t\n\t#get the effect description\n\teffectDescr = EffectDescriptions.quickEffectLookup[effectNumber]\n\t\n\t#Depending on the data, time to output the values accordingly\n\treqAttributes = effectDescr['reqAttributes']\n\tattribCount = len(reqAttributes)\n\tfor i in range (0,attribCount):\n\t\tcurAttrib = reqAttributes[i]\n\t\tattribType = curAttrib[EffectDescriptions.INDEX_TYPE]\n\t\tcurData = effectData[i]\n\t\tif(attribType == EffectDescriptions.VAR_COLOR):\n\t\t\treturnStr += Utils.short_to_hex(curData[0])\n\t\t\treturnStr += Utils.short_to_hex(curData[1])\n\t\t\treturnStr += Utils.short_to_hex(curData[2])\n\t\telif(attribType == EffectDescriptions.VAR_BYTE):\n\t\t\treturnStr += Utils.byte_to_hex(int(curData))\n\t\telif(attribType == EffectDescriptions.VAR_WORD):\n\t\t\treturnStr += Utils.short_to_hex(int(curData))\n\t\telif(attribType == EffectDescriptions.VAR_DWORD):\n\t\t\treturnStr += Utils.dword_to_hex(int(curData))\n\t\telif(attribType == EffectDescriptions.VAR_HIDDEN_BYTE):\n\t\t\treturnStr += Utils.short_to_hex(int(curData))\n\t\t\t\n\t\telse:\n\t\t\tprint \"ERROR! COULD NOT DECODE EFFECT!\"\n\treturn returnStr\t\n\n\t\t\n\t\n\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":6627170382357033000,"string":"6,627,170,382,357,033,000"},"line_mean":{"kind":"number","value":34.0684931507,"string":"34.068493"},"line_max":{"kind":"number","value":175,"string":"175"},"alpha_frac":{"kind":"number","value":0.74375,"string":"0.74375"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.073229291716687,"string":"3.073229"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42184,"cells":{"repo_name":{"kind":"string","value":"notepadqq/NotepadqqApi_Python"},"path":{"kind":"string","value":"notepadqq_api/notepadqq_api.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"3531"},"content":{"kind":"string","value":"import asyncio\nimport sys\n\nfrom notepadqq_api.message_channel import MessageChannel\nfrom notepadqq_api.message_interpreter import MessageInterpreter\nfrom notepadqq_api.stubs import Stubs\n\nclass NotepadqqApi():\n \"\"\"Provides access to the Notepadqq Api.\"\"\"\n\n _NQQ_STUB_ID = 1\n\n def __init__(self, socket_path=None, extension_id=None):\n \"\"\"Construct a new Api object that can be used to invoke Notepadqq\n methods and to receive its events.\n If not provided, socket_path and extension_id are respectively\n sys.argv[1] and sys.argv[2]\n \"\"\"\n if socket_path is None:\n try:\n socket_path = sys.argv[1]\n except IndexError:\n raise ValueError(\"Socket path not provided\") \n \n if extension_id is None:\n try:\n extension_id = sys.argv[2]\n except IndexError:\n raise ValueError(\"Extension id not provided\")\n \n self._socket_path = socket_path\n self._extension_id = extension_id\n\n self._message_channel = MessageChannel(self._socket_path)\n self._message_interpreter = MessageInterpreter(self._message_channel)\n\n self._nqq = Stubs.Notepadqq(self._message_interpreter, self._NQQ_STUB_ID)\n\n def run_event_loop(self, started_callback=None):\n \"\"\"Start the event loop. If started_callback is provided, it will\n be called as soon as the connection with Notepadqq is ready.\n \"\"\"\n if started_callback is not None:\n self.notepadqq.on('currentExtensionStarted', started_callback)\n\n loop = asyncio.get_event_loop()\n loop.run_until_complete(self._message_channel.start(loop, self._on_new_message))\n\n @property\n def extension_id(self):\n \"\"\"The id assigned to this extension by Notepadqq\"\"\"\n return self._extension_id\n\n @property\n def notepadqq(self):\n \"\"\"Get an instance of the main Notepadqq object\"\"\"\n return self._nqq\n\n def on_window_created(self, callback):\n \"\"\"Execute a callback for every new window.\n This is preferable to the \"newWindow\" event of Notepadqq, because it\n could happen that the extension isn't ready soon enough to receive\n the \"newWindow\" event for the first window. This method, instead,\n ensures that the passed callback will be called once and only once\n for each current or future window.\n \"\"\"\n captured_windows = []\n\n # Invoke the callback for every currently open window\n for window in self.notepadqq.windows():\n if window not in captured_windows:\n captured_windows.append(window)\n callback(window)\n\n # Each time a new window gets opened, invoke the callback.\n # When Notepadqq is starting and initializing all the extensions,\n # we might not be fast enough to receive this event: this is why\n # we manually invoked the callback for every currently open window.\n def on_new_window(window):\n if window not in captured_windows:\n callback(window)\n\n self.notepadqq.on('newWindow', on_new_window)\n\n def for_each_window(self, f):\n \"\"\"Decorator alternative for self.on_window_created(f)\"\"\"\n self.on_window_created(f)\n return f\n\n def _on_new_message(self, msg):\n # Called whenever a new message is received from the channel\n self._message_interpreter.process_message(msg)\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":-2928871828242222600,"string":"-2,928,871,828,242,222,600"},"line_mean":{"kind":"number","value":37.3804347826,"string":"37.380435"},"line_max":{"kind":"number","value":88,"string":"88"},"alpha_frac":{"kind":"number","value":0.6329651657,"string":"0.632965"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":4.3700495049504955,"string":"4.37005"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42185,"cells":{"repo_name":{"kind":"string","value":"rackerlabs/deuce-valere"},"path":{"kind":"string","value":"deucevalere/common/validation.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1337"},"content":{"kind":"string","value":"\"\"\"\nDeuce Valere - Common - Validation\n\"\"\"\nimport datetime\n\nfrom deuceclient.api import *\nfrom deuceclient.auth.base import AuthenticationBase\nfrom deuceclient.client.deuce import DeuceClient\nfrom deuceclient.common.validation import *\nfrom deuceclient.common.validation_instance import *\nfrom stoplight import Rule, ValidationFailed, validation_function\n\n\n@validation_function\ndef val_authenticator_instance(value):\n if not isinstance(value, AuthenticationBase):\n raise ValidationFailed('authenticator must be derived from '\n 'deuceclient.auth.base.AuthenticationBase')\n\n\n@validation_function\ndef val_deuceclient_instance(value):\n if not isinstance(value, DeuceClient):\n raise ValidationFailed('invalid Deuce Client instance')\n\n\n@validation_function\ndef val_expire_age(value):\n if not isinstance(value, datetime.timedelta):\n raise ValidationFailed('must be type datetime.timedelta')\n\n\ndef _abort(error_code):\n abort_errors = {\n 100: TypeError\n }\n raise abort_errors[error_code]\n\nAuthEngineRule = Rule(val_authenticator_instance(), lambda: _abort(100))\nClientRule = Rule(val_deuceclient_instance(), lambda: _abort(100))\n\nExpireAgeRule = Rule(val_expire_age(), lambda: _abort(100))\nExpireAgeRuleNoneOkay = Rule(val_expire_age(none_ok=True), lambda: _abort(100))\n"},"license":{"kind":"string","value":"apache-2.0"},"hash":{"kind":"number","value":-5001324934594096000,"string":"-5,001,324,934,594,096,000"},"line_mean":{"kind":"number","value":30.0930232558,"string":"30.093023"},"line_max":{"kind":"number","value":79,"string":"79"},"alpha_frac":{"kind":"number","value":0.7434554974,"string":"0.743455"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.8419540229885056,"string":"3.841954"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42186,"cells":{"repo_name":{"kind":"string","value":"hemebond/kapua"},"path":{"kind":"string","value":"courses/views.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"4832"},"content":{"kind":"string","value":"# Copyright 2011 James O'Neill\n#\n# This file is part of Kapua.\n#\n# Kapua is free software: you can redistribute it and/or modify\n# it under the terms of the GNU General Public License as published by\n# the Free Software Foundation, either version 3 of the License, or\n# (at your option) any later version.\n#\n# Kapua is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with Kapua. If not, see .\n\n\nfrom django.contrib.auth.decorators import login_required\nfrom django.utils.decorators import method_decorator\nfrom django.views.generic import ListView, DetailView, UpdateView, \\\n FormView, CreateView\nfrom django.views.generic.detail import SingleObjectMixin\nfrom django.http import HttpResponseRedirect\nfrom django.shortcuts import redirect\n\nfrom .models import Course, Page\nfrom .forms import CourseForm, PageForm\n\n\nclass CourseList(ListView):\n\tmodel = Course\n\n\nclass CourseAdd(CreateView):\n\ttemplate_name = \"courses/course_edit.html\"\n\tform_class = CourseForm\n\tcontext_object_name = \"course\"\n\n\t@method_decorator(login_required)\n\tdef dispatch(self, *args, **kwargs):\n\t\treturn super(CourseAdd, self).dispatch(*args, **kwargs)\n\n\nclass CourseDetail(DetailView):\n\ttemplate_name = \"courses/course_detail.html\"\n\tmodel = Course\n\tcontext_object_name = \"course\"\n\n\tdef get(self, request, *args, **kwargs):\n\t\tself.object = self.get_object()\n\n\t\tif self.object.pages.exists():\n\t\t\treturn redirect('kapua-page-detail', self.object.pages.get(level=0).pk)\n\n\t\tcontext = self.get_context_data(object=self.object)\n\t\treturn self.render_to_response(context)\n\n\nclass CourseEdit(UpdateView):\n\ttemplate_name = \"courses/course_edit.html\"\n\tform_class = CourseForm\n\tmodel = Course\n\n\nclass PageAdd(SingleObjectMixin, FormView):\n\tmodel = Course\n\ttemplate_name = \"courses/page_edit.html\"\n\tform_class = PageForm\n\n\t@method_decorator(login_required)\n\tdef dispatch(self, *args, **kwargs):\n\t\treturn super(PageAdd, self).dispatch(*args, **kwargs)\n\n\tdef get_form(self, form_class):\n\t\tself.object = self.get_object()\n\t\treturn super(PageAdd, self).get_form(form_class)\n\n\tdef get_form_kwargs(self):\n\t\t\"\"\"\n\t\tReturns the keyword arguments for instantiating the form.\n\t\t\"\"\"\n\t\tform_kwargs = super(PageAdd, self).get_form_kwargs()\n\n\t\tform_kwargs.update({\n\t\t\t'valid_targets': self.object.pages.filter(level__gt=0)\n\t\t})\n\n\t\treturn form_kwargs\n\n\tdef form_valid(self, form):\n\t\tposition = form.cleaned_data.get('position', 'last-child')\n\t\ttarget = form.cleaned_data.get('target', None)\n\t\tcourse = self.object\n\n\t\tpage = form.save(commit=False)\n\t\tpage.course = course\n\n\t\tif not target:\n\t\t\tif course.pages.exists():\n\t\t\t\ttarget = course.pages.get(level=0)\n\t\t\t\tposition = 'last-child'\n\n\t\tif target:\n\t\t\tpage.insert_at(\n\t\t\t\ttarget=target,\n\t\t\t\tposition=position,\n\t\t\t\tsave=True,\n\t\t\t)\n\t\t\tself.success_url = page.get_absolute_url()\n\t\telse:\n\t\t\tpage.save()\n\t\t\tself.success_url = course.get_absolute_url()\n\n\t\treturn super(PageAdd, self).form_valid(form)\n\n\tdef get_context_data(self, *args, **kwargs):\n\t\tcontext = super(PageAdd, self).get_context_data(*args, **kwargs)\n\n\t\tif context['form'].errors:\n\t\t\tcontext['error_message'] = context['form'].errors\n\n\t\treturn context\n\n\nclass PageDetail(DetailView):\n\ttemplate_name = \"courses/page_detail.html\"\n\tcontext_object_name = \"page\"\n\tmodel = Page\n\n\tdef get_context_data(self, **kwargs):\n\t\t# Call the base implementation first to get a context\n\t\tcontext = super(PageDetail, self).get_context_data(**kwargs)\n\t\tcontext['course'] = self.object.course\n\t\tpages = context['course'].pages.all()\n\n\t\tfor index, page in enumerate(pages):\n\t\t\tif page.pk == self.object.pk:\n\t\t\t\tif index > 0:\n\t\t\t\t\tcontext['previous_page'] = pages[index - 1]\n\n\t\t\t\tif index < (len(pages) - 1):\n\t\t\t\t\tcontext['next_page'] = pages[index + 1]\n\n\t\t\t\tbreak\n\n\t\t# Remove the root page\n\t\tcontext['pages'] = pages.filter(level__gt=0)\n\n\t\t# This gets the ancestors of the current page but exluces the\n\t\t# root page\n\t\tcontext['breadcrumbs'] = pages.filter(\n\t\t\tlft__lt=self.object.lft,\n\t\t\trght__gt=self.object.rght\n\t\t).exclude(\n\t\t\tlevel=0\n\t\t)\n\n\t\treturn context\n\n\nclass PageEdit(UpdateView):\n\ttemplate_name = \"courses/page_edit.html\"\n\tform_class = PageForm\n\tmodel = Page\n\n\t@method_decorator(login_required)\n\tdef dispatch(self, *args, **kwargs):\n\t\treturn super(PageEdit, self).dispatch(*args, **kwargs)\n\n\tdef form_valid(self, form):\n\t\tself.object = form.save()\n\n\t\ttarget = form.cleaned_data.get('target')\n\t\tif target:\n\t\t\tposition = form.cleaned_data.get('position')\n\t\t\tself.object.move_to(\n\t\t\t\ttarget=target,\n\t\t\t\tposition=position\n\t\t\t)\n\n\t\treturn redirect('kapua-page-detail', self.object.pk)\n"},"license":{"kind":"string","value":"gpl-3.0"},"hash":{"kind":"number","value":8351613602219685000,"string":"8,351,613,602,219,685,000"},"line_mean":{"kind":"number","value":25.5494505495,"string":"25.549451"},"line_max":{"kind":"number","value":74,"string":"74"},"alpha_frac":{"kind":"number","value":0.7104718543,"string":"0.710472"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":3.264864864864865,"string":"3.264865"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42187,"cells":{"repo_name":{"kind":"string","value":"Pikl/PiklBot"},"path":{"kind":"string","value":"cogs/pikl.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1281"},"content":{"kind":"string","value":"import discord\nfrom discord.ext import commands\n\nclass Pikl:\n\t\"\"\"Super pikly commands.\"\"\"\n\n\tdef __init__(self, bot):\n\t\tself.bot = bot\n\n\t@commands.command(hidden=False)\n\tasync def helloworld(self):\n\t\t\"\"\"Hello, world!\"\"\"\n\t\tawait self.bot.say(\"Hello, world!\")\n\n\t@commands.command(hidden=False)\t\n\tasync def postraidembed(self):\n\t\t\"\"\"Posts an embedded message with a bunch of raid info\"\"\"\n\t\tembed = discord.Embed(colour=discord.Colour(0x2ecc40), description=\"Some helpful information to aid and review [Dawn] raids.\\n\")\n\n\t\tembed.set_image(url=\"https://cdn.discordapp.com/attachments/350137990959464459/354412417381433354/unknown.png\")\n\t\tembed.set_thumbnail(url=\"https://wiki.guildwars2.com/images/5/5e/Legendary_Insight.png\")\n\t\tembed.set_author(name=\"Dawn Raid Information\", icon_url=\"http://raid.pikly.uk/images/dawn-logo.png\")\n\t\tembed.set_footer(text=\": 'Stack on Pikl'\", icon_url=\"http://raid.pikly.uk/images/dawn-logo.png\")\n\n\t\tembed.add_field(name=\"Raid Logs & Videos\", value=\"https://raid.pikly.uk/\", inline=True)\n\t\tembed.add_field(name=\"Raid Class Spreadsheet\", value=\"[Spreadsheet here](https://docs.google.com/spreadsheets/d/1zm46Jb8UBIoYP1_mewoOvLKopx_Sks9hYGm8OeWaQI8/edit?usp=sharing)\", inline=True)\n\n\t\tawait self.bot.say(embed=embed)\n\ndef setup(bot):\n\tbot.add_cog(Pikl(bot))"},"license":{"kind":"string","value":"gpl-3.0"},"hash":{"kind":"number","value":-9634640449993438,"string":"-9,634,640,449,993,438"},"line_mean":{"kind":"number","value":40.3548387097,"string":"40.354839"},"line_max":{"kind":"number","value":191,"string":"191"},"alpha_frac":{"kind":"number","value":0.737704918,"string":"0.737705"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":2.6911764705882355,"string":"2.691176"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42188,"cells":{"repo_name":{"kind":"string","value":"jtwaleson/decrypt"},"path":{"kind":"string","value":"decrypt/decrypt.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"1745"},"content":{"kind":"string","value":"#!/usr/bin/env python\nimport curses\nimport time\nimport fileinput\nimport random\nimport string\n\nscreen = curses.initscr()\nlines = []\nchance = 0.1\nconfirmed_per_line = []\n\n\ndef main():\n curses.noecho()\n try:\n curses.curs_set(0)\n except:\n pass\n screen.keypad(1)\n try:\n for line in fileinput.input():\n confirmed_per_line.append([])\n lines.append(line.rstrip())\n iterate()\n fileinput.close()\n while iterate(increase=True):\n pass\n time.sleep(2)\n except KeyboardInterrupt:\n pass\n finally:\n curses.endwin()\n for line in lines:\n print(line)\n\n\ndef iterate(increase=False):\n global chance, confirmed_per_line, lines\n still_random = 0\n if increase:\n chance += 0.01\n screen.erase()\n (y, x) = screen.getmaxyx()\n final_line = len(lines)\n if final_line > y:\n first_line = final_line - y\n else:\n first_line = 0\n for line_num in range(first_line, final_line):\n line = lines[line_num]\n for col in range(min(x, len(line))):\n try:\n if col not in confirmed_per_line[line_num]:\n still_random += 1\n if random.random() < chance:\n confirmed_per_line[line_num].append(col)\n screen.addch(line_num - first_line,\n col,\n random.choice(string.punctuation),\n curses.A_REVERSE)\n else:\n screen.addstr(line_num - first_line, col, line[col])\n except:\n pass\n\n screen.refresh()\n time.sleep(0.1)\n return still_random > 0\n\n"},"license":{"kind":"string","value":"mit"},"hash":{"kind":"number","value":3421878440241431600,"string":"3,421,878,440,241,431,600"},"line_mean":{"kind":"number","value":23.9285714286,"string":"23.928571"},"line_max":{"kind":"number","value":72,"string":"72"},"alpha_frac":{"kind":"number","value":0.5146131805,"string":"0.514613"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":4.058139534883721,"string":"4.05814"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42189,"cells":{"repo_name":{"kind":"string","value":"vdrhtc/Measurement-automation"},"path":{"kind":"string","value":"drivers/pyspcm.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"7735"},"content":{"kind":"string","value":"import os\nimport platform\nimport sys \nfrom ctypes import *\n\n# load registers for easier access\nfrom drivers.py_header.regs import *\n\n# load registers for easier access\nfrom drivers.py_header.spcerr import *\n\nSPCM_DIR_PCTOCARD = 0\nSPCM_DIR_CARDTOPC = 1\n\nSPCM_BUF_DATA = 1000 # main data buffer for acquired or generated samples\nSPCM_BUF_ABA = 2000 # buffer for ABA data, holds the A-DATA (slow samples)\nSPCM_BUF_TIMESTAMP = 3000 # buffer for timestamps\n\n\n# determine bit width of os\noPlatform = platform.architecture()\nif (oPlatform[0] == '64bit'):\n bIs64Bit = 1\nelse:\n bIs64Bit = 0\n\n# define pointer aliases\nint8 = c_int8\nint16 = c_int16\nint32 = c_int32\nint64 = c_int64\n\nptr8 = POINTER (int8)\nptr16 = POINTER (int16)\nptr32 = POINTER (int32)\nptr64 = POINTER (int64)\n\nuint8 = c_uint8\nuint16 = c_uint16\nuint32 = c_uint32\nuint64 = c_uint64\n\nuptr8 = POINTER (uint8)\nuptr16 = POINTER (uint16)\nuptr32 = POINTER (uint32)\nuptr64 = POINTER (uint64)\n\n\n# Windows\nif os.name == 'nt':\n #sys.stdout.write(\"Python Version: {0} on Windows\\n\\n\".format (\n # platform.python_version()))\n\n # define card handle type\n if (bIs64Bit):\n # for unknown reasons c_void_p gets messed up on Win7/64bit, but this works:\n drv_handle = POINTER(c_uint64)\n else:\n drv_handle = c_void_p\n\n # Load DLL into memory.\n # use windll because all driver access functions use _stdcall calling convention under windows\n if (bIs64Bit == 1):\n spcmDll = windll.LoadLibrary (\"c:\\\\windows\\\\system32\\\\spcm_win64.dll\")\n else:\n spcmDll = windll.LoadLibrary (\"c:\\\\windows\\\\system32\\\\spcm_win32.dll\")\n\n # load spcm_hOpen\n if (bIs64Bit):\n spcm_hOpen = getattr (spcmDll, \"spcm_hOpen\")\n else:\n spcm_hOpen = getattr (spcmDll, \"_spcm_hOpen@4\")\n spcm_hOpen.argtype = [c_char_p]\n spcm_hOpen.restype = drv_handle \n\n # load spcm_vClose\n if (bIs64Bit):\n spcm_vClose = getattr (spcmDll, \"spcm_vClose\")\n else:\n spcm_vClose = getattr (spcmDll, \"_spcm_vClose@4\")\n spcm_vClose.argtype = [drv_handle]\n spcm_vClose.restype = None\n\n # load spcm_dwGetErrorInfo\n if (bIs64Bit):\n spcm_dwGetErrorInfo_i32 = getattr (spcmDll, \"spcm_dwGetErrorInfo_i32\")\n else:\n spcm_dwGetErrorInfo_i32 = getattr (spcmDll, \"_spcm_dwGetErrorInfo_i32@16\")\n spcm_dwGetErrorInfo_i32.argtype = [drv_handle, uptr32, ptr32, c_char_p]\n spcm_dwGetErrorInfo_i32.restype = uint32\n\n # load spcm_dwGetParam_i32\n if (bIs64Bit):\n spcm_dwGetParam_i32 = getattr (spcmDll, \"spcm_dwGetParam_i32\")\n else:\n spcm_dwGetParam_i32 = getattr (spcmDll, \"_spcm_dwGetParam_i32@12\")\n spcm_dwGetParam_i32.argtype = [drv_handle, int32, ptr32]\n spcm_dwGetParam_i32.restype = uint32\n\n # load spcm_dwGetParam_i64\n if (bIs64Bit):\n spcm_dwGetParam_i64 = getattr (spcmDll, \"spcm_dwGetParam_i64\")\n else:\n spcm_dwGetParam_i64 = getattr (spcmDll, \"_spcm_dwGetParam_i64@12\")\n spcm_dwGetParam_i64.argtype = [drv_handle, int32, ptr64]\n spcm_dwGetParam_i64.restype = uint32\n\n # load spcm_dwSetParam_i32\n if (bIs64Bit):\n spcm_dwSetParam_i32 = getattr (spcmDll, \"spcm_dwSetParam_i32\")\n else:\n spcm_dwSetParam_i32 = getattr (spcmDll, \"_spcm_dwSetParam_i32@12\")\n spcm_dwSetParam_i32.argtype = [drv_handle, int32, int32]\n spcm_dwSetParam_i32.restype = uint32\n\n # load spcm_dwSetParam_i64\n if (bIs64Bit):\n spcm_dwSetParam_i64 = getattr (spcmDll, \"spcm_dwSetParam_i64\")\n else:\n spcm_dwSetParam_i64 = getattr (spcmDll, \"_spcm_dwSetParam_i64@16\")\n spcm_dwSetParam_i64.argtype = [drv_handle, int32, int64]\n spcm_dwSetParam_i64.restype = uint32\n\n # load spcm_dwSetParam_i64m\n if (bIs64Bit):\n spcm_dwSetParam_i64m = getattr (spcmDll, \"spcm_dwSetParam_i64m\")\n else:\n spcm_dwSetParam_i64m = getattr (spcmDll, \"_spcm_dwSetParam_i64m@16\")\n spcm_dwSetParam_i64m.argtype = [drv_handle, int32, int32, int32]\n spcm_dwSetParam_i64m.restype = uint32\n\n # load spcm_dwDefTransfer_i64\n if (bIs64Bit):\n spcm_dwDefTransfer_i64 = getattr (spcmDll, \"spcm_dwDefTransfer_i64\")\n else:\n spcm_dwDefTransfer_i64 = getattr (spcmDll, \"_spcm_dwDefTransfer_i64@36\")\n spcm_dwDefTransfer_i64.argtype = [drv_handle, uint32, uint32, uint32, c_void_p, uint64, uint64]\n spcm_dwDefTransfer_i64.restype = uint32\n\n # load spcm_dwInvalidateBuf\n if (bIs64Bit):\n spcm_dwInvalidateBuf = getattr (spcmDll, \"spcm_dwInvalidateBuf\")\n else:\n spcm_dwInvalidateBuf = getattr (spcmDll, \"_spcm_dwInvalidateBuf@8\")\n spcm_dwInvalidateBuf.argtype = [drv_handle, uint32]\n spcm_dwInvalidateBuf.restype = uint32\n\n # load spcm_dwGetContBuf_i64\n if (bIs64Bit):\n spcm_dwGetContBuf_i64 = getattr (spcmDll, \"spcm_dwGetContBuf_i64\")\n else:\n spcm_dwGetContBuf_i64 = getattr (spcmDll, \"_spcm_dwGetContBuf_i64@16\")\n spcm_dwGetContBuf_i64.argtype = [drv_handle, uint32, POINTER(c_void_p), uptr64]\n spcm_dwGetContBuf_i64.restype = uint32\n\n\nelif os.name == 'posix':\n sys.stdout.write(\"Python Version: {0} on Linux\\n\\n\".format (platform.python_version()))\n\n # define card handle type\n if (bIs64Bit):\n drv_handle = POINTER(c_uint64)\n else:\n drv_handle = c_void_p\n\n # Load DLL into memory.\n # use cdll because all driver access functions use cdecl calling convention under linux \n spcmDll = cdll.LoadLibrary (\"libspcm_linux.so\")\n\n # load spcm_hOpen\n spcm_hOpen = getattr (spcmDll, \"spcm_hOpen\")\n spcm_hOpen.argtype = [c_char_p]\n spcm_hOpen.restype = drv_handle \n\n # load spcm_vClose\n spcm_vClose = getattr (spcmDll, \"spcm_vClose\")\n spcm_vClose.argtype = [drv_handle]\n spcm_vClose.restype = None\n\n # load spcm_dwGetErrorInfo\n spcm_dwGetErrorInfo_i32 = getattr (spcmDll, \"spcm_dwGetErrorInfo_i32\")\n spcm_dwGetErrorInfo_i32.argtype = [drv_handle, uptr32, ptr32, c_char_p]\n spcm_dwGetErrorInfo_i32.restype = uint32\n\n # load spcm_dwGetParam_i32\n spcm_dwGetParam_i32 = getattr (spcmDll, \"spcm_dwGetParam_i32\")\n spcm_dwGetParam_i32.argtype = [drv_handle, int32, ptr32]\n spcm_dwGetParam_i32.restype = uint32\n\n # load spcm_dwGetParam_i64\n spcm_dwGetParam_i64 = getattr (spcmDll, \"spcm_dwGetParam_i64\")\n spcm_dwGetParam_i64.argtype = [drv_handle, int32, ptr64]\n spcm_dwGetParam_i64.restype = uint32\n\n # load spcm_dwSetParam_i32\n spcm_dwSetParam_i32 = getattr (spcmDll, \"spcm_dwSetParam_i32\")\n spcm_dwSetParam_i32.argtype = [drv_handle, int32, int32]\n spcm_dwSetParam_i32.restype = uint32\n\n # load spcm_dwSetParam_i64\n spcm_dwSetParam_i64 = getattr (spcmDll, \"spcm_dwSetParam_i64\")\n spcm_dwSetParam_i64.argtype = [drv_handle, int32, int64]\n spcm_dwSetParam_i64.restype = uint32\n\n # load spcm_dwSetParam_i64m\n spcm_dwSetParam_i64m = getattr (spcmDll, \"spcm_dwSetParam_i64m\")\n spcm_dwSetParam_i64m.argtype = [drv_handle, int32, int32, int32]\n spcm_dwSetParam_i64m.restype = uint32\n\n # load spcm_dwDefTransfer_i64\n spcm_dwDefTransfer_i64 = getattr (spcmDll, \"spcm_dwDefTransfer_i64\")\n spcm_dwDefTransfer_i64.argtype = [drv_handle, uint32, uint32, uint32, c_void_p, uint64, uint64]\n spcm_dwDefTransfer_i64.restype = uint32\n\n # load spcm_dwInvalidateBuf\n spcm_dwInvalidateBuf = getattr (spcmDll, \"spcm_dwInvalidateBuf\")\n spcm_dwInvalidateBuf.argtype = [drv_handle, uint32]\n spcm_dwInvalidateBuf.restype = uint32\n\n # load spcm_dwGetContBuf_i64\n spcm_dwGetContBuf_i64 = getattr (spcmDll, \"spcm_dwGetContBuf_i64\")\n spcm_dwGetContBuf_i64.argtype = [drv_handle, uint32, POINTER(c_void_p), uptr64]\n spcm_dwGetContBuf_i64.restype = uint32\n\nelse:\n raise Exception ('Operating system not supported by pySpcm')\n"},"license":{"kind":"string","value":"gpl-3.0"},"hash":{"kind":"number","value":5788646918922448000,"string":"5,788,646,918,922,448,000"},"line_mean":{"kind":"number","value":33.2256637168,"string":"33.225664"},"line_max":{"kind":"number","value":99,"string":"99"},"alpha_frac":{"kind":"number","value":0.6828700711,"string":"0.68287"},"autogenerated":{"kind":"bool","value":false,"string":"false"},"ratio":{"kind":"number","value":2.7092819614711035,"string":"2.709282"},"config_test":{"kind":"bool","value":false,"string":"false"},"has_no_keywords":{"kind":"bool","value":false,"string":"false"},"few_assignments":{"kind":"bool","value":false,"string":"false"}}},{"rowIdx":42190,"cells":{"repo_name":{"kind":"string","value":"Dev-Cloud-Platform/Dev-Cloud"},"path":{"kind":"string","value":"dev_cloud/web_service/urls/user/environment.py"},"copies":{"kind":"string","value":"1"},"size":{"kind":"string","value":"5340"},"content":{"kind":"string","value":"# -*- coding: utf-8 -*-\n# @COPYRIGHT_begin\n#\n# Copyright [2015] Michał Szczygieł, M4GiK Software\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http://www.apache.org/licenses/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# @COPYRIGHT_end\nfrom django.conf.urls import patterns, url, include\nfrom core.utils.decorators import user_permission, vm_permission\nfrom web_service.views.user.enviroment import wizard_setup, generate_dependencies, customize_environment, \\\n define_environment, summary, validation_process, validation_process_ip, validation_process_resources, \\\n validation_process_ip_pre, view_environment, environments_list, get_vm_status, destroy_vm, refresh_vm_tasks, \\\n show_vnc, get_cpu_load, get_ssh_key, view_predefined, customize_predefined_environment, \\\n define_predefined_environment\n\nmain_patterns = patterns('web_service.views.user.enviroment',\n url(r'^app/create/environment/$', user_permission(wizard_setup),\n name='personalized_environment'),\n url(r'^app/create/environment/technology/(?P\\w+)/$',\n user_permission(generate_dependencies),\n name='generate_dependencies'),\n url(\n r'^app/create/environment/customize/(?P\\w+)/(?P[\\w\\-]+)/(?P\\w+)/$',\n user_permission(customize_environment), name='customize_environment'),\n url(r'^app/create/environment/define/(?P\\w+)/(?P\\w+)/$',\n user_permission(define_environment), name='define_environment'),\n url(r'^app/create/environment/summary/$', user_permission(summary), name='summary'),\n url(r'^app/create/environment/validation_process/(?P