commit
stringlengths
40
40
old_file
stringlengths
4
106
new_file
stringlengths
4
106
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
2.95k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.31k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
diff
stringlengths
49
3.61k
e409e8d77eec8e53978512da56bf52f768d46c1a
create-application.py
create-application.py
import subprocess, os, sys, optparse fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) capath = os.path.abspath( os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py") ) skeletonpath = os.path.abspath( os.path.join(fullpath, "unify", "application", "skeleton") ) subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:])
import os, sys, shutil print("Unify create skeleton") print("(C) 2012 Sebastian Fastner, Mainz, Germany") print() if (len(sys.argv) != 2): print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0])) exit(1) NAMESPACE = sys.argv[1] UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) SKELETONPATH = os.path.abspath( os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify") ) TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE)) REPLACEMENTS = { "NAMESPACE" : NAMESPACE, "UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH) } if os.path.exists(TARGETPATH): print("Path %s exists. Aborting." % TARGETPATH) exit(2) shutil.copytree(SKELETONPATH, TARGETPATH) def patch_line(line): for key in REPLACEMENTS: check = "${" + key + "}" line = line.replace(check, REPLACEMENTS[key]) return line def handle_file(directory, filename): outfile_name = os.path.join(directory, filename.replace(".tmpl", "")) infile_name = os.path.join(directory, filename) with open(outfile_name, "w") as outfile: with open(infile_name) as infile: for line in infile: outfile.write(patch_line(line)) os.remove(infile_name) def handle_dir(directory): shutil.move(directory, directory[:-6] + NAMESPACE) for root, dirs, files in os.walk(TARGETPATH,topdown=False): for file in files: if ".tmpl." in file: handle_file(root, file) if root.endswith("custom"): handle_dir(root) print("Creat application skeleton in %s ... done" % TARGETPATH)
Add variable replacements to create application
Add variable replacements to create application
Python
mit
unify/unify,unify/unify,unify/unify,unify/unify,unify/unify,unify/unify
- import subprocess, os, sys, optparse + import os, sys, shutil + print("Unify create skeleton") + print("(C) 2012 Sebastian Fastner, Mainz, Germany") + print() + + if (len(sys.argv) != 2): + print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0])) + exit(1) + + NAMESPACE = sys.argv[1] + - fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) + UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) - capath = os.path.abspath( - os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py") + SKELETONPATH = os.path.abspath( + os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify") ) + TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE)) - skeletonpath = os.path.abspath( - os.path.join(fullpath, "unify", "application", "skeleton") - ) - subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:]) + REPLACEMENTS = { + "NAMESPACE" : NAMESPACE, + "UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH) + } + if os.path.exists(TARGETPATH): + print("Path %s exists. Aborting." % TARGETPATH) + exit(2) + + shutil.copytree(SKELETONPATH, TARGETPATH) + + def patch_line(line): + for key in REPLACEMENTS: + check = "${" + key + "}" + line = line.replace(check, REPLACEMENTS[key]) + return line + + def handle_file(directory, filename): + outfile_name = os.path.join(directory, filename.replace(".tmpl", "")) + infile_name = os.path.join(directory, filename) + with open(outfile_name, "w") as outfile: + with open(infile_name) as infile: + for line in infile: + outfile.write(patch_line(line)) + os.remove(infile_name) + + def handle_dir(directory): + shutil.move(directory, directory[:-6] + NAMESPACE) + + for root, dirs, files in os.walk(TARGETPATH,topdown=False): + for file in files: + if ".tmpl." in file: + handle_file(root, file) + + if root.endswith("custom"): + handle_dir(root) + + print("Creat application skeleton in %s ... done" % TARGETPATH)
Add variable replacements to create application
## Code Before: import subprocess, os, sys, optparse fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) capath = os.path.abspath( os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py") ) skeletonpath = os.path.abspath( os.path.join(fullpath, "unify", "application", "skeleton") ) subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:]) ## Instruction: Add variable replacements to create application ## Code After: import os, sys, shutil print("Unify create skeleton") print("(C) 2012 Sebastian Fastner, Mainz, Germany") print() if (len(sys.argv) != 2): print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0])) exit(1) NAMESPACE = sys.argv[1] UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) SKELETONPATH = os.path.abspath( os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify") ) TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE)) REPLACEMENTS = { "NAMESPACE" : NAMESPACE, "UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH) } if os.path.exists(TARGETPATH): print("Path %s exists. Aborting." % TARGETPATH) exit(2) shutil.copytree(SKELETONPATH, TARGETPATH) def patch_line(line): for key in REPLACEMENTS: check = "${" + key + "}" line = line.replace(check, REPLACEMENTS[key]) return line def handle_file(directory, filename): outfile_name = os.path.join(directory, filename.replace(".tmpl", "")) infile_name = os.path.join(directory, filename) with open(outfile_name, "w") as outfile: with open(infile_name) as infile: for line in infile: outfile.write(patch_line(line)) os.remove(infile_name) def handle_dir(directory): shutil.move(directory, directory[:-6] + NAMESPACE) for root, dirs, files in os.walk(TARGETPATH,topdown=False): for file in files: if ".tmpl." in file: handle_file(root, file) if root.endswith("custom"): handle_dir(root) print("Creat application skeleton in %s ... done" % TARGETPATH)
- import subprocess, os, sys, optparse + import os, sys, shutil + print("Unify create skeleton") + print("(C) 2012 Sebastian Fastner, Mainz, Germany") + print() + + if (len(sys.argv) != 2): + print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0])) + exit(1) + + NAMESPACE = sys.argv[1] + - fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) ? ^^^^^^^^ + UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0])) ? ^^^^^^^^^ - capath = os.path.abspath( - os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py") + SKELETONPATH = os.path.abspath( + os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify") ) + TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE)) - skeletonpath = os.path.abspath( - os.path.join(fullpath, "unify", "application", "skeleton") - ) - subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:]) + REPLACEMENTS = { + "NAMESPACE" : NAMESPACE, + "UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH) + } + + if os.path.exists(TARGETPATH): + print("Path %s exists. Aborting." % TARGETPATH) + exit(2) + + shutil.copytree(SKELETONPATH, TARGETPATH) + + def patch_line(line): + for key in REPLACEMENTS: + check = "${" + key + "}" + line = line.replace(check, REPLACEMENTS[key]) + return line + + def handle_file(directory, filename): + outfile_name = os.path.join(directory, filename.replace(".tmpl", "")) + infile_name = os.path.join(directory, filename) + with open(outfile_name, "w") as outfile: + with open(infile_name) as infile: + for line in infile: + outfile.write(patch_line(line)) + os.remove(infile_name) + + def handle_dir(directory): + shutil.move(directory, directory[:-6] + NAMESPACE) + + for root, dirs, files in os.walk(TARGETPATH,topdown=False): + for file in files: + if ".tmpl." in file: + handle_file(root, file) + + if root.endswith("custom"): + handle_dir(root) + + print("Creat application skeleton in %s ... done" % TARGETPATH)
994606d2641115f8af59657204d3d64f540bbfbd
data_structures/linked_list.py
data_structures/linked_list.py
class Node(object): def __init__(self, val, next=None): self.val = val self.next = next def __repr__(self): return '{val}'.format(val=self.val) class LinkedList(object): def __init__(self, values=None, head=None): self.head = head self.length = 0 def __repr__(self): pass def __len__(self): pass def __iter__(self): pass def size(self): pass def search(self): pass def display(self): pass def remove(self): pass def insert(self): pass def pop(self): pass
class Node(object): def __init__(self, val, next=None): self.val = val self.next = next def __repr__(self): return '{val}'.format(val=self.val) class LinkedList(object): def __init__(self, iterable=()): self._current = None self.head = None self.length = 0 for val in reversed(iterable): self.insert(val) def __repr__(self): '''Print string representation of Linked List.''' node = self.head output = '' for node in self: output += '{!r}'.format(node.val) return '({})'.format(output.rstrip(' ,')) def __len__(self): return self.length def __iter__(self): if self.head is not None: self._current = self.head return self def next(self): if self._current is None: raise StopIteration node = self._current self._current = self._current.next return node def insert(self): pass def size(self): pass def search(self): pass def display(self): pass def remove(self): pass def pop(self): pass
Update magic methods, and reorg args.
Update magic methods, and reorg args.
Python
mit
sjschmidt44/python_data_structures
class Node(object): def __init__(self, val, next=None): self.val = val self.next = next def __repr__(self): return '{val}'.format(val=self.val) class LinkedList(object): - def __init__(self, values=None, head=None): + def __init__(self, iterable=()): + self._current = None - self.head = head + self.head = None self.length = 0 + for val in reversed(iterable): + self.insert(val) def __repr__(self): - pass + '''Print string representation of Linked List.''' + node = self.head + output = '' + for node in self: + output += '{!r}'.format(node.val) + return '({})'.format(output.rstrip(' ,')) def __len__(self): - pass + return self.length def __iter__(self): + if self.head is not None: + self._current = self.head + return self + + def next(self): + if self._current is None: + raise StopIteration + node = self._current + self._current = self._current.next + return node + + def insert(self): pass def size(self): pass def search(self): pass def display(self): pass def remove(self): pass - def insert(self): - pass - def pop(self): pass
Update magic methods, and reorg args.
## Code Before: class Node(object): def __init__(self, val, next=None): self.val = val self.next = next def __repr__(self): return '{val}'.format(val=self.val) class LinkedList(object): def __init__(self, values=None, head=None): self.head = head self.length = 0 def __repr__(self): pass def __len__(self): pass def __iter__(self): pass def size(self): pass def search(self): pass def display(self): pass def remove(self): pass def insert(self): pass def pop(self): pass ## Instruction: Update magic methods, and reorg args. ## Code After: class Node(object): def __init__(self, val, next=None): self.val = val self.next = next def __repr__(self): return '{val}'.format(val=self.val) class LinkedList(object): def __init__(self, iterable=()): self._current = None self.head = None self.length = 0 for val in reversed(iterable): self.insert(val) def __repr__(self): '''Print string representation of Linked List.''' node = self.head output = '' for node in self: output += '{!r}'.format(node.val) return '({})'.format(output.rstrip(' ,')) def __len__(self): return self.length def __iter__(self): if self.head is not None: self._current = self.head return self def next(self): if self._current is None: raise StopIteration node = self._current self._current = self._current.next return node def insert(self): pass def size(self): pass def search(self): pass def display(self): pass def remove(self): pass def pop(self): pass
class Node(object): def __init__(self, val, next=None): self.val = val self.next = next def __repr__(self): return '{val}'.format(val=self.val) class LinkedList(object): - def __init__(self, values=None, head=None): + def __init__(self, iterable=()): + self._current = None - self.head = head ? ^ -- + self.head = None ? ^^^ self.length = 0 + for val in reversed(iterable): + self.insert(val) def __repr__(self): - pass + '''Print string representation of Linked List.''' + node = self.head + output = '' + for node in self: + output += '{!r}'.format(node.val) + return '({})'.format(output.rstrip(' ,')) def __len__(self): - pass + return self.length def __iter__(self): + if self.head is not None: + self._current = self.head + return self + + def next(self): + if self._current is None: + raise StopIteration + node = self._current + self._current = self._current.next + return node + + def insert(self): pass def size(self): pass def search(self): pass def display(self): pass def remove(self): pass - def insert(self): - pass - def pop(self): pass
26598254cd48a716527eb4689ad96551c5a39790
ksp_login/__init__.py
ksp_login/__init__.py
__version__ = '0.6.0' __version_info__ = __version__.split('.') from django.utils.translation import ugettext_lazy as _ def __activate_social_auth_monkeypatch(): from social_core.backends.base import BaseAuth from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth) from social_core.backends.livejournal import LiveJournalOpenId from social_core.backends.yahoo import YahooOpenId from social_core.backends.google import GoogleOpenId from social_core.backends.yandex import YandexOpenId BaseAuth.REQUIRED_FIELD_NAME = None BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity') LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user' LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username') # Reset to None in those OpenID backends where nothing is required. GoogleOpenId.REQUIRED_FIELD_NAME = None GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YahooOpenId.REQUIRED_FIELD_NAME = None YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YandexOpenId.REQUIRED_FIELD_NAME = None YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None __activate_social_auth_monkeypatch()
__version__ = '0.6.0' __version_info__ = tuple(map(int, __version__.split('.'))) from django.utils.translation import ugettext_lazy as _ def __activate_social_auth_monkeypatch(): from social_core.backends.base import BaseAuth from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth) from social_core.backends.livejournal import LiveJournalOpenId from social_core.backends.yahoo import YahooOpenId from social_core.backends.google import GoogleOpenId from social_core.backends.yandex import YandexOpenId BaseAuth.REQUIRED_FIELD_NAME = None BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity') LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user' LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username') # Reset to None in those OpenID backends where nothing is required. GoogleOpenId.REQUIRED_FIELD_NAME = None GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YahooOpenId.REQUIRED_FIELD_NAME = None YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YandexOpenId.REQUIRED_FIELD_NAME = None YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None __activate_social_auth_monkeypatch()
Make version info tuple of ints.
Make version info tuple of ints.
Python
bsd-3-clause
koniiiik/ksp_login,koniiiik/ksp_login,koniiiik/ksp_login
__version__ = '0.6.0' - __version_info__ = __version__.split('.') + __version_info__ = tuple(map(int, __version__.split('.'))) from django.utils.translation import ugettext_lazy as _ def __activate_social_auth_monkeypatch(): from social_core.backends.base import BaseAuth from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth) from social_core.backends.livejournal import LiveJournalOpenId from social_core.backends.yahoo import YahooOpenId from social_core.backends.google import GoogleOpenId from social_core.backends.yandex import YandexOpenId BaseAuth.REQUIRED_FIELD_NAME = None BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity') LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user' LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username') # Reset to None in those OpenID backends where nothing is required. GoogleOpenId.REQUIRED_FIELD_NAME = None GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YahooOpenId.REQUIRED_FIELD_NAME = None YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YandexOpenId.REQUIRED_FIELD_NAME = None YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None + __activate_social_auth_monkeypatch()
Make version info tuple of ints.
## Code Before: __version__ = '0.6.0' __version_info__ = __version__.split('.') from django.utils.translation import ugettext_lazy as _ def __activate_social_auth_monkeypatch(): from social_core.backends.base import BaseAuth from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth) from social_core.backends.livejournal import LiveJournalOpenId from social_core.backends.yahoo import YahooOpenId from social_core.backends.google import GoogleOpenId from social_core.backends.yandex import YandexOpenId BaseAuth.REQUIRED_FIELD_NAME = None BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity') LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user' LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username') # Reset to None in those OpenID backends where nothing is required. GoogleOpenId.REQUIRED_FIELD_NAME = None GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YahooOpenId.REQUIRED_FIELD_NAME = None YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YandexOpenId.REQUIRED_FIELD_NAME = None YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None __activate_social_auth_monkeypatch() ## Instruction: Make version info tuple of ints. ## Code After: __version__ = '0.6.0' __version_info__ = tuple(map(int, __version__.split('.'))) from django.utils.translation import ugettext_lazy as _ def __activate_social_auth_monkeypatch(): from social_core.backends.base import BaseAuth from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth) from social_core.backends.livejournal import LiveJournalOpenId from social_core.backends.yahoo import YahooOpenId from social_core.backends.google import GoogleOpenId from social_core.backends.yandex import YandexOpenId BaseAuth.REQUIRED_FIELD_NAME = None BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity') LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user' LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username') # Reset to None in those OpenID backends where nothing is required. GoogleOpenId.REQUIRED_FIELD_NAME = None GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YahooOpenId.REQUIRED_FIELD_NAME = None YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YandexOpenId.REQUIRED_FIELD_NAME = None YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None __activate_social_auth_monkeypatch()
__version__ = '0.6.0' - __version_info__ = __version__.split('.') + __version_info__ = tuple(map(int, __version__.split('.'))) ? +++++++++++++++ ++ from django.utils.translation import ugettext_lazy as _ def __activate_social_auth_monkeypatch(): from social_core.backends.base import BaseAuth from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth) from social_core.backends.livejournal import LiveJournalOpenId from social_core.backends.yahoo import YahooOpenId from social_core.backends.google import GoogleOpenId from social_core.backends.yandex import YandexOpenId BaseAuth.REQUIRED_FIELD_NAME = None BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity') LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user' LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username') # Reset to None in those OpenID backends where nothing is required. GoogleOpenId.REQUIRED_FIELD_NAME = None GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YahooOpenId.REQUIRED_FIELD_NAME = None YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None YandexOpenId.REQUIRED_FIELD_NAME = None YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None + __activate_social_auth_monkeypatch()
ae55577e4cea64a0052eb0c219641435c9c0210c
samples/model-builder/init_sample.py
samples/model-builder/init_sample.py
from typing import Optional from google.auth import credentials as auth_credentials from google.cloud import aiplatform # [START aiplatform_sdk_init_sample] def init_sample( project: Optional[str] = None, location: Optional[str] = None, experiment: Optional[str] = None, staging_bucket: Optional[str] = None, credentials: Optional[auth_credentials.Credentials] = None, encryption_spec_key_name: Optional[str] = None, ): aiplatform.init( project=project, location=location, experiment=experiment, staging_bucket=staging_bucket, credentials=credentials, encryption_spec_key_name=encryption_spec_key_name, ) # [END aiplatform_sdk_init_sample]
from typing import Optional from google.auth import credentials as auth_credentials # [START aiplatform_sdk_init_sample] def init_sample( project: Optional[str] = None, location: Optional[str] = None, experiment: Optional[str] = None, staging_bucket: Optional[str] = None, credentials: Optional[auth_credentials.Credentials] = None, encryption_spec_key_name: Optional[str] = None, ): from google.cloud import aiplatform aiplatform.init( project=project, location=location, experiment=experiment, staging_bucket=staging_bucket, credentials=credentials, encryption_spec_key_name=encryption_spec_key_name, ) # [END aiplatform_sdk_init_sample]
Update init sample to import inside of function.
chore: Update init sample to import inside of function. PiperOrigin-RevId: 485079470
Python
apache-2.0
googleapis/python-aiplatform,googleapis/python-aiplatform
from typing import Optional from google.auth import credentials as auth_credentials - from google.cloud import aiplatform # [START aiplatform_sdk_init_sample] def init_sample( project: Optional[str] = None, location: Optional[str] = None, experiment: Optional[str] = None, staging_bucket: Optional[str] = None, credentials: Optional[auth_credentials.Credentials] = None, encryption_spec_key_name: Optional[str] = None, ): + + from google.cloud import aiplatform + aiplatform.init( project=project, location=location, experiment=experiment, staging_bucket=staging_bucket, credentials=credentials, encryption_spec_key_name=encryption_spec_key_name, ) # [END aiplatform_sdk_init_sample]
Update init sample to import inside of function.
## Code Before: from typing import Optional from google.auth import credentials as auth_credentials from google.cloud import aiplatform # [START aiplatform_sdk_init_sample] def init_sample( project: Optional[str] = None, location: Optional[str] = None, experiment: Optional[str] = None, staging_bucket: Optional[str] = None, credentials: Optional[auth_credentials.Credentials] = None, encryption_spec_key_name: Optional[str] = None, ): aiplatform.init( project=project, location=location, experiment=experiment, staging_bucket=staging_bucket, credentials=credentials, encryption_spec_key_name=encryption_spec_key_name, ) # [END aiplatform_sdk_init_sample] ## Instruction: Update init sample to import inside of function. ## Code After: from typing import Optional from google.auth import credentials as auth_credentials # [START aiplatform_sdk_init_sample] def init_sample( project: Optional[str] = None, location: Optional[str] = None, experiment: Optional[str] = None, staging_bucket: Optional[str] = None, credentials: Optional[auth_credentials.Credentials] = None, encryption_spec_key_name: Optional[str] = None, ): from google.cloud import aiplatform aiplatform.init( project=project, location=location, experiment=experiment, staging_bucket=staging_bucket, credentials=credentials, encryption_spec_key_name=encryption_spec_key_name, ) # [END aiplatform_sdk_init_sample]
from typing import Optional from google.auth import credentials as auth_credentials - from google.cloud import aiplatform # [START aiplatform_sdk_init_sample] def init_sample( project: Optional[str] = None, location: Optional[str] = None, experiment: Optional[str] = None, staging_bucket: Optional[str] = None, credentials: Optional[auth_credentials.Credentials] = None, encryption_spec_key_name: Optional[str] = None, ): + + from google.cloud import aiplatform + aiplatform.init( project=project, location=location, experiment=experiment, staging_bucket=staging_bucket, credentials=credentials, encryption_spec_key_name=encryption_spec_key_name, ) # [END aiplatform_sdk_init_sample]
5d3d47e0fae9ddb9f445972e5186429163aabf40
statirator/core/management/commands/init.py
statirator/core/management/commands/init.py
import os from optparse import make_option from django.core.management.base import BaseCommand class Command(BaseCommand): help = "Init the static site project" args = '[directory]' option_list = ( make_option( '--title', '-t', dest='title', default='Default site', help='Site title [Default: "%defaults"]'), make_option( '--domain', '-d', dest='domain', default='example.com', help='Domain name [Default: "%default"]'), make_option( '--languages', '-l', dest='languages', default=['he', 'en'], action='append', help='Supported languages. [Default: "%default"]') ) + BaseCommand.option_list def handle(self, directory, **options): from django.conf.global_settings import LANGUAGES extra = { 'build': 'build', 'default_lang': options['languages'][0], 'languages': [l for l in LANGUAGES if l[0] in options["languages"]], 'extensions': ('py', ), 'files': (), 'template': os.path.abspath( os.path.join( os.path.dirname(__file__), os.pardir, os.pardir, os.pardir, 'project_template')), } extra.update(options) from django.core.management import call_command call_command('startproject', 'conf', directory, **extra)
import os import logging from django.core.management.base import BaseCommand from optparse import make_option class Command(BaseCommand): help = "Init the static site project" args = '[directory]' option_list = ( make_option( '--title', '-t', dest='title', default='Default site', help='Site title [Default: "%defaults"]'), make_option( '--domain', '-d', dest='domain', default='example.com', help='Domain name [Default: "%default"]'), make_option( '--languages', '-l', dest='languages', default=['he', 'en'], action='append', help='Supported languages. [Default: "%default"]') ) + BaseCommand.option_list def handle(self, directory, **options): logging.info("Initializing project structure in %s", directory) os.makedirs(directory) from django.conf.global_settings import LANGUAGES extra = { 'build': 'build', 'default_lang': options['languages'][0], 'languages': [l for l in LANGUAGES if l[0] in options["languages"]], 'extensions': ('py', ), 'files': (), 'template': os.path.abspath( os.path.join( os.path.dirname(__file__), os.pardir, os.pardir, os.pardir, 'project_template')), } extra.update(options) from django.core.management import call_command call_command('startproject', 'conf', directory, **extra)
Create the directory before calling the startprojcet command
Create the directory before calling the startprojcet command
Python
mit
MeirKriheli/statirator,MeirKriheli/statirator,MeirKriheli/statirator
import os + import logging + from django.core.management.base import BaseCommand from optparse import make_option - from django.core.management.base import BaseCommand class Command(BaseCommand): help = "Init the static site project" args = '[directory]' option_list = ( make_option( '--title', '-t', dest='title', default='Default site', help='Site title [Default: "%defaults"]'), make_option( '--domain', '-d', dest='domain', default='example.com', help='Domain name [Default: "%default"]'), make_option( '--languages', '-l', dest='languages', default=['he', 'en'], action='append', help='Supported languages. [Default: "%default"]') ) + BaseCommand.option_list def handle(self, directory, **options): + logging.info("Initializing project structure in %s", directory) + os.makedirs(directory) + from django.conf.global_settings import LANGUAGES extra = { 'build': 'build', 'default_lang': options['languages'][0], 'languages': [l for l in LANGUAGES if l[0] in options["languages"]], 'extensions': ('py', ), 'files': (), 'template': os.path.abspath( os.path.join( os.path.dirname(__file__), os.pardir, os.pardir, os.pardir, 'project_template')), } extra.update(options) from django.core.management import call_command call_command('startproject', 'conf', directory, **extra)
Create the directory before calling the startprojcet command
## Code Before: import os from optparse import make_option from django.core.management.base import BaseCommand class Command(BaseCommand): help = "Init the static site project" args = '[directory]' option_list = ( make_option( '--title', '-t', dest='title', default='Default site', help='Site title [Default: "%defaults"]'), make_option( '--domain', '-d', dest='domain', default='example.com', help='Domain name [Default: "%default"]'), make_option( '--languages', '-l', dest='languages', default=['he', 'en'], action='append', help='Supported languages. [Default: "%default"]') ) + BaseCommand.option_list def handle(self, directory, **options): from django.conf.global_settings import LANGUAGES extra = { 'build': 'build', 'default_lang': options['languages'][0], 'languages': [l for l in LANGUAGES if l[0] in options["languages"]], 'extensions': ('py', ), 'files': (), 'template': os.path.abspath( os.path.join( os.path.dirname(__file__), os.pardir, os.pardir, os.pardir, 'project_template')), } extra.update(options) from django.core.management import call_command call_command('startproject', 'conf', directory, **extra) ## Instruction: Create the directory before calling the startprojcet command ## Code After: import os import logging from django.core.management.base import BaseCommand from optparse import make_option class Command(BaseCommand): help = "Init the static site project" args = '[directory]' option_list = ( make_option( '--title', '-t', dest='title', default='Default site', help='Site title [Default: "%defaults"]'), make_option( '--domain', '-d', dest='domain', default='example.com', help='Domain name [Default: "%default"]'), make_option( '--languages', '-l', dest='languages', default=['he', 'en'], action='append', help='Supported languages. [Default: "%default"]') ) + BaseCommand.option_list def handle(self, directory, **options): logging.info("Initializing project structure in %s", directory) os.makedirs(directory) from django.conf.global_settings import LANGUAGES extra = { 'build': 'build', 'default_lang': options['languages'][0], 'languages': [l for l in LANGUAGES if l[0] in options["languages"]], 'extensions': ('py', ), 'files': (), 'template': os.path.abspath( os.path.join( os.path.dirname(__file__), os.pardir, os.pardir, os.pardir, 'project_template')), } extra.update(options) from django.core.management import call_command call_command('startproject', 'conf', directory, **extra)
import os + import logging + from django.core.management.base import BaseCommand from optparse import make_option - from django.core.management.base import BaseCommand class Command(BaseCommand): help = "Init the static site project" args = '[directory]' option_list = ( make_option( '--title', '-t', dest='title', default='Default site', help='Site title [Default: "%defaults"]'), make_option( '--domain', '-d', dest='domain', default='example.com', help='Domain name [Default: "%default"]'), make_option( '--languages', '-l', dest='languages', default=['he', 'en'], action='append', help='Supported languages. [Default: "%default"]') ) + BaseCommand.option_list def handle(self, directory, **options): + logging.info("Initializing project structure in %s", directory) + os.makedirs(directory) + from django.conf.global_settings import LANGUAGES extra = { 'build': 'build', 'default_lang': options['languages'][0], 'languages': [l for l in LANGUAGES if l[0] in options["languages"]], 'extensions': ('py', ), 'files': (), 'template': os.path.abspath( os.path.join( os.path.dirname(__file__), os.pardir, os.pardir, os.pardir, 'project_template')), } extra.update(options) from django.core.management import call_command call_command('startproject', 'conf', directory, **extra)
1e5d549b6fdf62c1016451f9dfe566c9546b2f38
bcbio/bed/__init__.py
bcbio/bed/__init__.py
import pybedtools as bt import six def concat(bed_files, catted=None): """ recursively concat a set of BED files, returning a sorted bedtools object of the result """ if len(bed_files) == 0: if catted: return catted.sort() else: return catted if not catted: bed_files = list(bed_files) catted = bt.BedTool(bed_files.pop()) else: catted = catted.cat(bed_files.pop(), postmerge=False, force_truncate=False) return concat(bed_files, catted) def merge(bedfiles): """ given a BED file or list of BED files merge them an return a bedtools object """ if isinstance(bedfiles, list): catted = concat(bedfiles) else: catted = concat([bedfiles]) if catted: return concat(bedfiles).sort().merge() else: return catted
import pybedtools as bt import six def concat(bed_files, catted=None): """ recursively concat a set of BED files, returning a sorted bedtools object of the result """ bed_files = [x for x in bed_files if x] if len(bed_files) == 0: if catted: # move to a .bed extension for downstream tools if not already sorted_bed = catted.sort() if not sorted_bed.fn.endswith(".bed"): return sorted_bed.moveto(sorted_bed.fn + ".bed") else: return sorted_bed else: return catted if not catted: bed_files = list(bed_files) catted = bt.BedTool(bed_files.pop()) else: catted = catted.cat(bed_files.pop(), postmerge=False, force_truncate=False) return concat(bed_files, catted) def merge(bedfiles): """ given a BED file or list of BED files merge them an return a bedtools object """ if isinstance(bedfiles, list): catted = concat(bedfiles) else: catted = concat([bedfiles]) if catted: return concat(bedfiles).sort().merge() else: return catted
Move the file to have an extension of .bed.
Move the file to have an extension of .bed. A lot of tools detect what type of file it is by the extension, so this lets us pass on the BedTool.fn as the filename and not break things.
Python
mit
guillermo-carrasco/bcbio-nextgen,lbeltrame/bcbio-nextgen,gifford-lab/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,vladsaveliev/bcbio-nextgen,brainstorm/bcbio-nextgen,mjafin/bcbio-nextgen,lbeltrame/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,brainstorm/bcbio-nextgen,fw1121/bcbio-nextgen,verdurin/bcbio-nextgen,lpantano/bcbio-nextgen,brainstorm/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,a113n/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,chapmanb/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,chapmanb/bcbio-nextgen,biocyberman/bcbio-nextgen,chapmanb/bcbio-nextgen,vladsaveliev/bcbio-nextgen,gifford-lab/bcbio-nextgen,fw1121/bcbio-nextgen,hjanime/bcbio-nextgen,biocyberman/bcbio-nextgen,lpantano/bcbio-nextgen,guillermo-carrasco/bcbio-nextgen,lbeltrame/bcbio-nextgen,guillermo-carrasco/bcbio-nextgen,mjafin/bcbio-nextgen,a113n/bcbio-nextgen,lpantano/bcbio-nextgen,fw1121/bcbio-nextgen,gifford-lab/bcbio-nextgen,verdurin/bcbio-nextgen,biocyberman/bcbio-nextgen,vladsaveliev/bcbio-nextgen,a113n/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,mjafin/bcbio-nextgen,hjanime/bcbio-nextgen,hjanime/bcbio-nextgen,verdurin/bcbio-nextgen
import pybedtools as bt import six def concat(bed_files, catted=None): """ recursively concat a set of BED files, returning a sorted bedtools object of the result """ + bed_files = [x for x in bed_files if x] if len(bed_files) == 0: if catted: + # move to a .bed extension for downstream tools if not already - return catted.sort() + sorted_bed = catted.sort() + if not sorted_bed.fn.endswith(".bed"): + return sorted_bed.moveto(sorted_bed.fn + ".bed") + else: + return sorted_bed else: return catted if not catted: bed_files = list(bed_files) catted = bt.BedTool(bed_files.pop()) else: catted = catted.cat(bed_files.pop(), postmerge=False, force_truncate=False) return concat(bed_files, catted) def merge(bedfiles): """ given a BED file or list of BED files merge them an return a bedtools object """ if isinstance(bedfiles, list): catted = concat(bedfiles) else: catted = concat([bedfiles]) if catted: return concat(bedfiles).sort().merge() else: return catted
Move the file to have an extension of .bed.
## Code Before: import pybedtools as bt import six def concat(bed_files, catted=None): """ recursively concat a set of BED files, returning a sorted bedtools object of the result """ if len(bed_files) == 0: if catted: return catted.sort() else: return catted if not catted: bed_files = list(bed_files) catted = bt.BedTool(bed_files.pop()) else: catted = catted.cat(bed_files.pop(), postmerge=False, force_truncate=False) return concat(bed_files, catted) def merge(bedfiles): """ given a BED file or list of BED files merge them an return a bedtools object """ if isinstance(bedfiles, list): catted = concat(bedfiles) else: catted = concat([bedfiles]) if catted: return concat(bedfiles).sort().merge() else: return catted ## Instruction: Move the file to have an extension of .bed. ## Code After: import pybedtools as bt import six def concat(bed_files, catted=None): """ recursively concat a set of BED files, returning a sorted bedtools object of the result """ bed_files = [x for x in bed_files if x] if len(bed_files) == 0: if catted: # move to a .bed extension for downstream tools if not already sorted_bed = catted.sort() if not sorted_bed.fn.endswith(".bed"): return sorted_bed.moveto(sorted_bed.fn + ".bed") else: return sorted_bed else: return catted if not catted: bed_files = list(bed_files) catted = bt.BedTool(bed_files.pop()) else: catted = catted.cat(bed_files.pop(), postmerge=False, force_truncate=False) return concat(bed_files, catted) def merge(bedfiles): """ given a BED file or list of BED files merge them an return a bedtools object """ if isinstance(bedfiles, list): catted = concat(bedfiles) else: catted = concat([bedfiles]) if catted: return concat(bedfiles).sort().merge() else: return catted
import pybedtools as bt import six def concat(bed_files, catted=None): """ recursively concat a set of BED files, returning a sorted bedtools object of the result """ + bed_files = [x for x in bed_files if x] if len(bed_files) == 0: if catted: + # move to a .bed extension for downstream tools if not already - return catted.sort() ? ^^^^ + sorted_bed = catted.sort() ? ++ + ^^^^^^^ + if not sorted_bed.fn.endswith(".bed"): + return sorted_bed.moveto(sorted_bed.fn + ".bed") + else: + return sorted_bed else: return catted if not catted: bed_files = list(bed_files) catted = bt.BedTool(bed_files.pop()) else: catted = catted.cat(bed_files.pop(), postmerge=False, force_truncate=False) return concat(bed_files, catted) def merge(bedfiles): """ given a BED file or list of BED files merge them an return a bedtools object """ if isinstance(bedfiles, list): catted = concat(bedfiles) else: catted = concat([bedfiles]) if catted: return concat(bedfiles).sort().merge() else: return catted
025c3f6b73c97fdb58b1a492efcb6efe44cfdab0
twisted/plugins/caldav.py
twisted/plugins/caldav.py
from zope.interface import implements from twisted.plugin import IPlugin from twisted.application.service import IServiceMaker from twisted.python import reflect def serviceMakerProperty(propname): def getProperty(self): return getattr(reflect.namedClass(self.serviceMakerClass), propname) return property(getProperty) class TAP(object): implements(IPlugin, IServiceMaker) def __init__(self, serviceMakerClass): self.serviceMakerClass = serviceMakerClass self._serviceMaker = None options = serviceMakerProperty("options") tapname = serviceMakerProperty("tapname") description = serviceMakerProperty("description") def makeService(self, options): if self._serviceMaker is None: self._serviceMaker = reflect.namedClass(self.serviceMakerClass)() return self._serviceMaker.makeService(options) TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker") CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker") CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
from zope.interface import implements from twisted.plugin import IPlugin from twisted.application.service import IServiceMaker from twisted.python import reflect from twisted.internet.protocol import Factory Factory.noisy = False def serviceMakerProperty(propname): def getProperty(self): return getattr(reflect.namedClass(self.serviceMakerClass), propname) return property(getProperty) class TAP(object): implements(IPlugin, IServiceMaker) def __init__(self, serviceMakerClass): self.serviceMakerClass = serviceMakerClass self._serviceMaker = None options = serviceMakerProperty("options") tapname = serviceMakerProperty("tapname") description = serviceMakerProperty("description") def makeService(self, options): if self._serviceMaker is None: self._serviceMaker = reflect.namedClass(self.serviceMakerClass)() return self._serviceMaker.makeService(options) TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker") CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker") CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
Set Factory.noisy to False by default
Set Factory.noisy to False by default git-svn-id: 81e381228600e5752b80483efd2b45b26c451ea2@3933 e27351fd-9f3e-4f54-a53b-843176b1656c
Python
apache-2.0
trevor/calendarserver,trevor/calendarserver,trevor/calendarserver
from zope.interface import implements from twisted.plugin import IPlugin from twisted.application.service import IServiceMaker from twisted.python import reflect + + from twisted.internet.protocol import Factory + Factory.noisy = False + def serviceMakerProperty(propname): def getProperty(self): return getattr(reflect.namedClass(self.serviceMakerClass), propname) return property(getProperty) class TAP(object): implements(IPlugin, IServiceMaker) + def __init__(self, serviceMakerClass): self.serviceMakerClass = serviceMakerClass self._serviceMaker = None - options = serviceMakerProperty("options") + options = serviceMakerProperty("options") - tapname = serviceMakerProperty("tapname") + tapname = serviceMakerProperty("tapname") description = serviceMakerProperty("description") def makeService(self, options): if self._serviceMaker is None: self._serviceMaker = reflect.namedClass(self.serviceMakerClass)() return self._serviceMaker.makeService(options) TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker") CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker") CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
Set Factory.noisy to False by default
## Code Before: from zope.interface import implements from twisted.plugin import IPlugin from twisted.application.service import IServiceMaker from twisted.python import reflect def serviceMakerProperty(propname): def getProperty(self): return getattr(reflect.namedClass(self.serviceMakerClass), propname) return property(getProperty) class TAP(object): implements(IPlugin, IServiceMaker) def __init__(self, serviceMakerClass): self.serviceMakerClass = serviceMakerClass self._serviceMaker = None options = serviceMakerProperty("options") tapname = serviceMakerProperty("tapname") description = serviceMakerProperty("description") def makeService(self, options): if self._serviceMaker is None: self._serviceMaker = reflect.namedClass(self.serviceMakerClass)() return self._serviceMaker.makeService(options) TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker") CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker") CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker") ## Instruction: Set Factory.noisy to False by default ## Code After: from zope.interface import implements from twisted.plugin import IPlugin from twisted.application.service import IServiceMaker from twisted.python import reflect from twisted.internet.protocol import Factory Factory.noisy = False def serviceMakerProperty(propname): def getProperty(self): return getattr(reflect.namedClass(self.serviceMakerClass), propname) return property(getProperty) class TAP(object): implements(IPlugin, IServiceMaker) def __init__(self, serviceMakerClass): self.serviceMakerClass = serviceMakerClass self._serviceMaker = None options = serviceMakerProperty("options") tapname = serviceMakerProperty("tapname") description = serviceMakerProperty("description") def makeService(self, options): if self._serviceMaker is None: self._serviceMaker = reflect.namedClass(self.serviceMakerClass)() return self._serviceMaker.makeService(options) TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker") CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker") CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
from zope.interface import implements from twisted.plugin import IPlugin from twisted.application.service import IServiceMaker from twisted.python import reflect + + from twisted.internet.protocol import Factory + Factory.noisy = False + def serviceMakerProperty(propname): def getProperty(self): return getattr(reflect.namedClass(self.serviceMakerClass), propname) return property(getProperty) class TAP(object): implements(IPlugin, IServiceMaker) + def __init__(self, serviceMakerClass): self.serviceMakerClass = serviceMakerClass self._serviceMaker = None - options = serviceMakerProperty("options") + options = serviceMakerProperty("options") ? ++++ - tapname = serviceMakerProperty("tapname") + tapname = serviceMakerProperty("tapname") ? ++++ description = serviceMakerProperty("description") def makeService(self, options): if self._serviceMaker is None: self._serviceMaker = reflect.namedClass(self.serviceMakerClass)() return self._serviceMaker.makeService(options) TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker") CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker") CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
30a4281f2602bd6b9d90d89375785a2645854a0d
enthought/enable2/pyglet_backend/pyglet_app.py
enthought/enable2/pyglet_backend/pyglet_app.py
from enthought.enable.pyglet_backend.pyglet_app import *
__all__ = ["get_app", "PygletApp"] from enthought.enable.pyglet_backend.pyglet_app import * # Import the objects which are not declared in __all__, # but are still defined in the real module, such that people # can import them explicitly when needed, just as they could # with the real module. # # It is unlikely that someone will import these objects, since # they start with '_'. However, the proxy's job is to mimic the # behavior of the real module as closely as possible. # The proxy's job is not to define or change the API. # from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp
Improve the proxy module which maps to a module which uses __all__.
Improve the proxy module which maps to a module which uses __all__. The notes I made in the code apply to all proxy modules which map to a module which uses __all__.
Python
bsd-3-clause
tommy-u/enable,tommy-u/enable,tommy-u/enable,tommy-u/enable
+ + __all__ = ["get_app", "PygletApp"] + from enthought.enable.pyglet_backend.pyglet_app import * + + # Import the objects which are not declared in __all__, + # but are still defined in the real module, such that people + # can import them explicitly when needed, just as they could + # with the real module. + # + # It is unlikely that someone will import these objects, since + # they start with '_'. However, the proxy's job is to mimic the + # behavior of the real module as closely as possible. + # The proxy's job is not to define or change the API. + # + from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp + +
Improve the proxy module which maps to a module which uses __all__.
## Code Before: from enthought.enable.pyglet_backend.pyglet_app import * ## Instruction: Improve the proxy module which maps to a module which uses __all__. ## Code After: __all__ = ["get_app", "PygletApp"] from enthought.enable.pyglet_backend.pyglet_app import * # Import the objects which are not declared in __all__, # but are still defined in the real module, such that people # can import them explicitly when needed, just as they could # with the real module. # # It is unlikely that someone will import these objects, since # they start with '_'. However, the proxy's job is to mimic the # behavior of the real module as closely as possible. # The proxy's job is not to define or change the API. # from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp
+ + __all__ = ["get_app", "PygletApp"] + from enthought.enable.pyglet_backend.pyglet_app import * + + + # Import the objects which are not declared in __all__, + # but are still defined in the real module, such that people + # can import them explicitly when needed, just as they could + # with the real module. + # + # It is unlikely that someone will import these objects, since + # they start with '_'. However, the proxy's job is to mimic the + # behavior of the real module as closely as possible. + # The proxy's job is not to define or change the API. + # + from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp +
850fba4b07e4c444aa8640c6f4c3816f8a3259ea
website_medical_patient_species/controllers/main.py
website_medical_patient_species/controllers/main.py
from openerp import http from openerp.http import request from openerp.addons.website_medical.controllers.main import ( WebsiteMedical ) class WebsiteMedical(WebsiteMedical): def _inject_medical_detail_vals(self, patient_id=0, **kwargs): vals = super(WebsiteMedical, self)._inject_medical_detail_vals( patient_id, **kwargs ) species_ids = request.env['medical.patient.species'].search([]) vals.update({ 'species': species_ids, }) return vals
from openerp.http import request from openerp.addons.website_medical.controllers.main import ( WebsiteMedical ) class WebsiteMedical(WebsiteMedical): def _inject_medical_detail_vals(self, patient_id=0, **kwargs): vals = super(WebsiteMedical, self)._inject_medical_detail_vals( patient_id, **kwargs ) species_ids = request.env['medical.patient.species'].search([]) vals.update({ 'species': species_ids, }) return vals
Fix lint * Remove stray import to fix lint
[FIX] website_medical_patient_species: Fix lint * Remove stray import to fix lint
Python
agpl-3.0
laslabs/vertical-medical,laslabs/vertical-medical
- from openerp import http from openerp.http import request from openerp.addons.website_medical.controllers.main import ( WebsiteMedical ) class WebsiteMedical(WebsiteMedical): def _inject_medical_detail_vals(self, patient_id=0, **kwargs): vals = super(WebsiteMedical, self)._inject_medical_detail_vals( patient_id, **kwargs ) species_ids = request.env['medical.patient.species'].search([]) vals.update({ 'species': species_ids, }) return vals
Fix lint * Remove stray import to fix lint
## Code Before: from openerp import http from openerp.http import request from openerp.addons.website_medical.controllers.main import ( WebsiteMedical ) class WebsiteMedical(WebsiteMedical): def _inject_medical_detail_vals(self, patient_id=0, **kwargs): vals = super(WebsiteMedical, self)._inject_medical_detail_vals( patient_id, **kwargs ) species_ids = request.env['medical.patient.species'].search([]) vals.update({ 'species': species_ids, }) return vals ## Instruction: Fix lint * Remove stray import to fix lint ## Code After: from openerp.http import request from openerp.addons.website_medical.controllers.main import ( WebsiteMedical ) class WebsiteMedical(WebsiteMedical): def _inject_medical_detail_vals(self, patient_id=0, **kwargs): vals = super(WebsiteMedical, self)._inject_medical_detail_vals( patient_id, **kwargs ) species_ids = request.env['medical.patient.species'].search([]) vals.update({ 'species': species_ids, }) return vals
- from openerp import http from openerp.http import request from openerp.addons.website_medical.controllers.main import ( WebsiteMedical ) class WebsiteMedical(WebsiteMedical): def _inject_medical_detail_vals(self, patient_id=0, **kwargs): vals = super(WebsiteMedical, self)._inject_medical_detail_vals( patient_id, **kwargs ) species_ids = request.env['medical.patient.species'].search([]) vals.update({ 'species': species_ids, }) return vals
093c9065de9e0e08f248bbb84696bf30309bd536
examples/parallel/timer.py
examples/parallel/timer.py
import rx import concurrent.futures import time seconds = [5, 1, 2, 4, 3] def sleep(t): time.sleep(t) return t def output(result): print '%d seconds' % result with concurrent.futures.ProcessPoolExecutor(5) as executor: rx.Observable.from_(seconds).flat_map( lambda s: executor.submit(sleep, s) ).subscribe(output) # 1 seconds # 2 seconds # 3 seconds # 4 seconds # 5 seconds
from __future__ import print_function import rx import concurrent.futures import time seconds = [5, 1, 2, 4, 3] def sleep(t): time.sleep(t) return t def output(result): print('%d seconds' % result) with concurrent.futures.ProcessPoolExecutor(5) as executor: rx.Observable.from_(seconds).flat_map( lambda s: executor.submit(sleep, s) ).subscribe(output) # 1 seconds # 2 seconds # 3 seconds # 4 seconds # 5 seconds
Fix parallel example for Python 3
Fix parallel example for Python 3
Python
mit
dbrattli/RxPY,ReactiveX/RxPY,ReactiveX/RxPY
+ from __future__ import print_function + import rx import concurrent.futures import time seconds = [5, 1, 2, 4, 3] def sleep(t): time.sleep(t) return t def output(result): - print '%d seconds' % result + print('%d seconds' % result) with concurrent.futures.ProcessPoolExecutor(5) as executor: rx.Observable.from_(seconds).flat_map( lambda s: executor.submit(sleep, s) ).subscribe(output) # 1 seconds # 2 seconds # 3 seconds # 4 seconds # 5 seconds
Fix parallel example for Python 3
## Code Before: import rx import concurrent.futures import time seconds = [5, 1, 2, 4, 3] def sleep(t): time.sleep(t) return t def output(result): print '%d seconds' % result with concurrent.futures.ProcessPoolExecutor(5) as executor: rx.Observable.from_(seconds).flat_map( lambda s: executor.submit(sleep, s) ).subscribe(output) # 1 seconds # 2 seconds # 3 seconds # 4 seconds # 5 seconds ## Instruction: Fix parallel example for Python 3 ## Code After: from __future__ import print_function import rx import concurrent.futures import time seconds = [5, 1, 2, 4, 3] def sleep(t): time.sleep(t) return t def output(result): print('%d seconds' % result) with concurrent.futures.ProcessPoolExecutor(5) as executor: rx.Observable.from_(seconds).flat_map( lambda s: executor.submit(sleep, s) ).subscribe(output) # 1 seconds # 2 seconds # 3 seconds # 4 seconds # 5 seconds
+ from __future__ import print_function + import rx import concurrent.futures import time seconds = [5, 1, 2, 4, 3] def sleep(t): time.sleep(t) return t def output(result): - print '%d seconds' % result ? ^ + print('%d seconds' % result) ? ^ + with concurrent.futures.ProcessPoolExecutor(5) as executor: rx.Observable.from_(seconds).flat_map( lambda s: executor.submit(sleep, s) ).subscribe(output) # 1 seconds # 2 seconds # 3 seconds # 4 seconds # 5 seconds
c32118b2157e6c2cfd435461ee23edfa79aa917e
api/__init__.py
api/__init__.py
import ConfigParser from peewee import * config = ConfigParser.RawConfigParser() config.read('server.conf') database = SqliteDatabase('gallery.db') from collection import CollectionModel from album import AlbumModel from user import UserModel, UserResource from photo import PhotoModel database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True) if UserModel.select().count() == 0: UserModel.create( admin = True, name = 'Admin', username = 'Admin', password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy' ) from boto.s3.connection import S3Connection s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey')) if s3.lookup(config.get('S3', 'Bucket')) is None: s3.create_bucket(config.get('S3', 'Bucket')) from flask import Flask from flask.ext.restful import Api app = Flask(__name__) api = Api(app)
import ConfigParser from peewee import * config = ConfigParser.RawConfigParser() config.read('server.conf') database = SqliteDatabase('gallery.db', threadlocals=True) from collection import CollectionModel from album import AlbumModel from user import UserModel, UsersResource from photo import PhotoModel database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True) if UserModel.select().count() == 0: UserModel.create( admin = True, name = 'Admin', username = 'Admin', password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy' ) from boto.s3.connection import S3Connection s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey')) if s3.lookup(config.get('S3', 'Bucket')) is None: s3.create_bucket(config.get('S3', 'Bucket')) from flask import Flask from flask.ext.restful import Api app = Flask(__name__) api = Api(app) api.add_resource(UsersResource, '/users/')
Set local threads to true for peewee
Set local threads to true for peewee
Python
unlicense
karousel/karousel
import ConfigParser from peewee import * config = ConfigParser.RawConfigParser() config.read('server.conf') - database = SqliteDatabase('gallery.db') + database = SqliteDatabase('gallery.db', threadlocals=True) from collection import CollectionModel from album import AlbumModel - from user import UserModel, UserResource + from user import UserModel, UsersResource from photo import PhotoModel database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True) if UserModel.select().count() == 0: UserModel.create( admin = True, name = 'Admin', username = 'Admin', password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy' ) from boto.s3.connection import S3Connection s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey')) if s3.lookup(config.get('S3', 'Bucket')) is None: s3.create_bucket(config.get('S3', 'Bucket')) from flask import Flask from flask.ext.restful import Api app = Flask(__name__) api = Api(app) + api.add_resource(UsersResource, '/users/') +
Set local threads to true for peewee
## Code Before: import ConfigParser from peewee import * config = ConfigParser.RawConfigParser() config.read('server.conf') database = SqliteDatabase('gallery.db') from collection import CollectionModel from album import AlbumModel from user import UserModel, UserResource from photo import PhotoModel database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True) if UserModel.select().count() == 0: UserModel.create( admin = True, name = 'Admin', username = 'Admin', password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy' ) from boto.s3.connection import S3Connection s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey')) if s3.lookup(config.get('S3', 'Bucket')) is None: s3.create_bucket(config.get('S3', 'Bucket')) from flask import Flask from flask.ext.restful import Api app = Flask(__name__) api = Api(app) ## Instruction: Set local threads to true for peewee ## Code After: import ConfigParser from peewee import * config = ConfigParser.RawConfigParser() config.read('server.conf') database = SqliteDatabase('gallery.db', threadlocals=True) from collection import CollectionModel from album import AlbumModel from user import UserModel, UsersResource from photo import PhotoModel database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True) if UserModel.select().count() == 0: UserModel.create( admin = True, name = 'Admin', username = 'Admin', password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy' ) from boto.s3.connection import S3Connection s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey')) if s3.lookup(config.get('S3', 'Bucket')) is None: s3.create_bucket(config.get('S3', 'Bucket')) from flask import Flask from flask.ext.restful import Api app = Flask(__name__) api = Api(app) api.add_resource(UsersResource, '/users/')
import ConfigParser from peewee import * config = ConfigParser.RawConfigParser() config.read('server.conf') - database = SqliteDatabase('gallery.db') + database = SqliteDatabase('gallery.db', threadlocals=True) ? +++++++++++++++++++ from collection import CollectionModel from album import AlbumModel - from user import UserModel, UserResource + from user import UserModel, UsersResource ? + from photo import PhotoModel database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True) if UserModel.select().count() == 0: UserModel.create( admin = True, name = 'Admin', username = 'Admin', password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy' ) from boto.s3.connection import S3Connection s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey')) if s3.lookup(config.get('S3', 'Bucket')) is None: s3.create_bucket(config.get('S3', 'Bucket')) from flask import Flask from flask.ext.restful import Api app = Flask(__name__) api = Api(app) + + api.add_resource(UsersResource, '/users/')
88fc0f980f0efa403ab5ce7d6775bce008b284fc
_setup_database.py
_setup_database.py
import argparse from setup.create_teams import migrate_teams from setup.create_divisions import create_divisions from setup.create_players import migrate_players from setup.create_player_seasons import create_player_seasons from setup.create_player_seasons import create_player_data from utils import prepare_logging prepare_logging(log_types=['file', 'screen']) if __name__ == '__main__': parser = argparse.ArgumentParser( description='Setup script for NHL database creation.') parser.add_argument( 'steps', metavar='setup_steps', help='Setup steps to execute.', choices=['a', 't', 'd', 'p', 'ps', 'pd']) args = parser.parse_args() setup_steps = args.steps # migrating teams from json file to database if setup_steps in ['t', 'a']: migrate_teams(simulation=True) # creating divisions from division configuration file if setup_steps in ['d', 'a']: create_divisions(simulation=True) # migrating players from json file to database if setup_steps in ['p', 'a']: migrate_players(simulation=True) # retrieving player season statistics for all players in database if setup_steps in ['ps', 'a']: create_player_seasons(simulation=False) # retrieving individual player data for all players in database if setup_steps in ['pd', 'a']: create_player_data(simulation=False)
import argparse from setup.create_teams import migrate_teams from setup.create_divisions import create_divisions from setup.create_players import migrate_players from setup.create_player_seasons import create_player_seasons from setup.create_player_seasons import create_player_data from setup.create_player_seasons import create_player_contracts from utils import prepare_logging prepare_logging(log_types=['file', 'screen']) if __name__ == '__main__': parser = argparse.ArgumentParser( description='Setup script for NHL database creation.') parser.add_argument( 'steps', metavar='setup_steps', help='Setup steps to execute.', choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd']) args = parser.parse_args() setup_steps = args.steps # migrating teams from json file to database if setup_steps in ['t', 'a']: migrate_teams(simulation=True) # creating divisions from division configuration file if setup_steps in ['d', 'a']: create_divisions(simulation=True) # migrating players from json file to database if setup_steps in ['p', 'a']: migrate_players(simulation=True) # retrieving player season statistics for all players in database if setup_steps in ['ps', 'a']: create_player_seasons(simulation=False) # retrieving individual player data for all players in database if setup_steps in ['pd', 'a']: create_player_data(simulation=False) if setup_steps in ['c']: create_player_contracts(simulation=False)
Add contract retrieval option to database setup script
Add contract retrieval option to database setup script
Python
mit
leaffan/pynhldb
import argparse from setup.create_teams import migrate_teams from setup.create_divisions import create_divisions from setup.create_players import migrate_players from setup.create_player_seasons import create_player_seasons from setup.create_player_seasons import create_player_data + from setup.create_player_seasons import create_player_contracts from utils import prepare_logging prepare_logging(log_types=['file', 'screen']) if __name__ == '__main__': parser = argparse.ArgumentParser( description='Setup script for NHL database creation.') parser.add_argument( 'steps', metavar='setup_steps', help='Setup steps to execute.', - choices=['a', 't', 'd', 'p', 'ps', 'pd']) + choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd']) args = parser.parse_args() setup_steps = args.steps # migrating teams from json file to database if setup_steps in ['t', 'a']: migrate_teams(simulation=True) # creating divisions from division configuration file if setup_steps in ['d', 'a']: create_divisions(simulation=True) # migrating players from json file to database if setup_steps in ['p', 'a']: migrate_players(simulation=True) # retrieving player season statistics for all players in database if setup_steps in ['ps', 'a']: create_player_seasons(simulation=False) # retrieving individual player data for all players in database if setup_steps in ['pd', 'a']: create_player_data(simulation=False) + if setup_steps in ['c']: + create_player_contracts(simulation=False) +
Add contract retrieval option to database setup script
## Code Before: import argparse from setup.create_teams import migrate_teams from setup.create_divisions import create_divisions from setup.create_players import migrate_players from setup.create_player_seasons import create_player_seasons from setup.create_player_seasons import create_player_data from utils import prepare_logging prepare_logging(log_types=['file', 'screen']) if __name__ == '__main__': parser = argparse.ArgumentParser( description='Setup script for NHL database creation.') parser.add_argument( 'steps', metavar='setup_steps', help='Setup steps to execute.', choices=['a', 't', 'd', 'p', 'ps', 'pd']) args = parser.parse_args() setup_steps = args.steps # migrating teams from json file to database if setup_steps in ['t', 'a']: migrate_teams(simulation=True) # creating divisions from division configuration file if setup_steps in ['d', 'a']: create_divisions(simulation=True) # migrating players from json file to database if setup_steps in ['p', 'a']: migrate_players(simulation=True) # retrieving player season statistics for all players in database if setup_steps in ['ps', 'a']: create_player_seasons(simulation=False) # retrieving individual player data for all players in database if setup_steps in ['pd', 'a']: create_player_data(simulation=False) ## Instruction: Add contract retrieval option to database setup script ## Code After: import argparse from setup.create_teams import migrate_teams from setup.create_divisions import create_divisions from setup.create_players import migrate_players from setup.create_player_seasons import create_player_seasons from setup.create_player_seasons import create_player_data from setup.create_player_seasons import create_player_contracts from utils import prepare_logging prepare_logging(log_types=['file', 'screen']) if __name__ == '__main__': parser = argparse.ArgumentParser( description='Setup script for NHL database creation.') parser.add_argument( 'steps', metavar='setup_steps', help='Setup steps to execute.', choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd']) args = parser.parse_args() setup_steps = args.steps # migrating teams from json file to database if setup_steps in ['t', 'a']: migrate_teams(simulation=True) # creating divisions from division configuration file if setup_steps in ['d', 'a']: create_divisions(simulation=True) # migrating players from json file to database if setup_steps in ['p', 'a']: migrate_players(simulation=True) # retrieving player season statistics for all players in database if setup_steps in ['ps', 'a']: create_player_seasons(simulation=False) # retrieving individual player data for all players in database if setup_steps in ['pd', 'a']: create_player_data(simulation=False) if setup_steps in ['c']: create_player_contracts(simulation=False)
import argparse from setup.create_teams import migrate_teams from setup.create_divisions import create_divisions from setup.create_players import migrate_players from setup.create_player_seasons import create_player_seasons from setup.create_player_seasons import create_player_data + from setup.create_player_seasons import create_player_contracts from utils import prepare_logging prepare_logging(log_types=['file', 'screen']) if __name__ == '__main__': parser = argparse.ArgumentParser( description='Setup script for NHL database creation.') parser.add_argument( 'steps', metavar='setup_steps', help='Setup steps to execute.', - choices=['a', 't', 'd', 'p', 'ps', 'pd']) + choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd']) ? +++++ args = parser.parse_args() setup_steps = args.steps # migrating teams from json file to database if setup_steps in ['t', 'a']: migrate_teams(simulation=True) # creating divisions from division configuration file if setup_steps in ['d', 'a']: create_divisions(simulation=True) # migrating players from json file to database if setup_steps in ['p', 'a']: migrate_players(simulation=True) # retrieving player season statistics for all players in database if setup_steps in ['ps', 'a']: create_player_seasons(simulation=False) # retrieving individual player data for all players in database if setup_steps in ['pd', 'a']: create_player_data(simulation=False) + + if setup_steps in ['c']: + create_player_contracts(simulation=False)
ed5dcd72b661878913be224d641c5595c73ef049
tests/test_auditory.py
tests/test_auditory.py
from __future__ import division, print_function import pytest import numpy as np from pambox import auditory as aud import scipy.io as sio from numpy.testing import assert_allclose def test_lowpass_filtering_of_envelope(): mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat", squeeze_me=True) envelope = mat['unfiltered_env'] target = mat['lp_filtered_env'] filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.) assert_allclose(filtered_envelope, target, atol=1e-7)
from __future__ import division, print_function import pytest import numpy as np from pambox import auditory as aud import scipy.io as sio from numpy.testing import assert_allclose def test_lowpass_filtering_of_envelope(): mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat", squeeze_me=True) envelope = mat['unfiltered_env'] target = mat['lp_filtered_env'] filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.) assert_allclose(filtered_envelope, target, atol=1e-7) def test_erb(): bw = aud.erbbw(1000) assert_allclose(bw, 132.63, rtol=1e-4)
Test of the erb calculation
Test of the erb calculation
Python
bsd-3-clause
achabotl/pambox
from __future__ import division, print_function import pytest import numpy as np from pambox import auditory as aud import scipy.io as sio from numpy.testing import assert_allclose def test_lowpass_filtering_of_envelope(): mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat", squeeze_me=True) envelope = mat['unfiltered_env'] target = mat['lp_filtered_env'] filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.) assert_allclose(filtered_envelope, target, atol=1e-7) + def test_erb(): + bw = aud.erbbw(1000) + assert_allclose(bw, 132.63, rtol=1e-4)
Test of the erb calculation
## Code Before: from __future__ import division, print_function import pytest import numpy as np from pambox import auditory as aud import scipy.io as sio from numpy.testing import assert_allclose def test_lowpass_filtering_of_envelope(): mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat", squeeze_me=True) envelope = mat['unfiltered_env'] target = mat['lp_filtered_env'] filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.) assert_allclose(filtered_envelope, target, atol=1e-7) ## Instruction: Test of the erb calculation ## Code After: from __future__ import division, print_function import pytest import numpy as np from pambox import auditory as aud import scipy.io as sio from numpy.testing import assert_allclose def test_lowpass_filtering_of_envelope(): mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat", squeeze_me=True) envelope = mat['unfiltered_env'] target = mat['lp_filtered_env'] filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.) assert_allclose(filtered_envelope, target, atol=1e-7) def test_erb(): bw = aud.erbbw(1000) assert_allclose(bw, 132.63, rtol=1e-4)
from __future__ import division, print_function import pytest import numpy as np from pambox import auditory as aud import scipy.io as sio from numpy.testing import assert_allclose def test_lowpass_filtering_of_envelope(): mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat", squeeze_me=True) envelope = mat['unfiltered_env'] target = mat['lp_filtered_env'] filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.) assert_allclose(filtered_envelope, target, atol=1e-7) + def test_erb(): + bw = aud.erbbw(1000) + assert_allclose(bw, 132.63, rtol=1e-4)
64042be2b6febf64d601adaa6f85a542ae9b876d
sunpy/instr/iris/iris.py
sunpy/instr/iris/iris.py
import sunpy.io import sunpy.time import sunpy.map __all__ = ['SJI_to_cube'] def SJI_to_cube(filename, start=0, stop=None): """ Read a SJI file and return a MapCube ..warning:: This function is a very early beta and is not stable. Further work is on going to improve SunPy IRIS support. Parameters ---------- filename: string File to read start: Temporal axis index to create MapCube from stop: Temporal index to stop MapCube at Returns ------- iris_cube: sunpy.map.MapCube A map cube of the SJI sequence """ hdus = sunpy.io.read_file(filename) #Get the time delta time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS']) splits = time_range.split(hdus[0][0].shape[0]) if not stop: stop = len(splits) headers = [hdus[0][1]]*(stop-start) datas = hdus[0][0][start:stop] #Make the cube: iris_cube = sunpy.map.Map(zip(datas,headers),cube=True) #Set the date/time for i,m in enumerate(iris_cube): m.meta['DATE-OBS'] = splits[i].center().isoformat() return iris_cube
import sunpy.io import sunpy.time import sunpy.map __all__ = ['SJI_to_cube'] def SJI_to_cube(filename, start=0, stop=None, hdu=0): """ Read a SJI file and return a MapCube ..warning:: This function is a very early beta and is not stable. Further work is on going to improve SunPy IRIS support. Parameters ---------- filename: string File to read start: Temporal axis index to create MapCube from stop: Temporal index to stop MapCube at hdu: Choose hdu index Returns ------- iris_cube: sunpy.map.MapCube A map cube of the SJI sequence """ hdus = sunpy.io.read_file(filename) #Get the time delta time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS']) splits = time_range.split(hdus[hdu][0].shape[0]) if not stop: stop = len(splits) headers = [hdus[hdu][1]]*(stop-start) datas = hdus[hdu][0][start:stop] #Make the cube: iris_cube = sunpy.map.Map(zip(datas,headers),cube=True) #Set the date/time for i,m in enumerate(iris_cube): m.meta['DATE-OBS'] = splits[i].center().isoformat() return iris_cube
Change hdu[0] to hdu for optional indexing
Change hdu[0] to hdu for optional indexing
Python
bsd-2-clause
Alex-Ian-Hamilton/sunpy,dpshelio/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy
import sunpy.io import sunpy.time import sunpy.map __all__ = ['SJI_to_cube'] - def SJI_to_cube(filename, start=0, stop=None): + def SJI_to_cube(filename, start=0, stop=None, hdu=0): """ Read a SJI file and return a MapCube ..warning:: This function is a very early beta and is not stable. Further work is on going to improve SunPy IRIS support. Parameters ---------- filename: string File to read start: Temporal axis index to create MapCube from stop: Temporal index to stop MapCube at + hdu: + Choose hdu index + Returns ------- iris_cube: sunpy.map.MapCube A map cube of the SJI sequence """ hdus = sunpy.io.read_file(filename) #Get the time delta - time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS']) + time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS']) - splits = time_range.split(hdus[0][0].shape[0]) + splits = time_range.split(hdus[hdu][0].shape[0]) if not stop: stop = len(splits) - headers = [hdus[0][1]]*(stop-start) + headers = [hdus[hdu][1]]*(stop-start) - datas = hdus[0][0][start:stop] + datas = hdus[hdu][0][start:stop] #Make the cube: iris_cube = sunpy.map.Map(zip(datas,headers),cube=True) #Set the date/time for i,m in enumerate(iris_cube): m.meta['DATE-OBS'] = splits[i].center().isoformat() return iris_cube +
Change hdu[0] to hdu for optional indexing
## Code Before: import sunpy.io import sunpy.time import sunpy.map __all__ = ['SJI_to_cube'] def SJI_to_cube(filename, start=0, stop=None): """ Read a SJI file and return a MapCube ..warning:: This function is a very early beta and is not stable. Further work is on going to improve SunPy IRIS support. Parameters ---------- filename: string File to read start: Temporal axis index to create MapCube from stop: Temporal index to stop MapCube at Returns ------- iris_cube: sunpy.map.MapCube A map cube of the SJI sequence """ hdus = sunpy.io.read_file(filename) #Get the time delta time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS']) splits = time_range.split(hdus[0][0].shape[0]) if not stop: stop = len(splits) headers = [hdus[0][1]]*(stop-start) datas = hdus[0][0][start:stop] #Make the cube: iris_cube = sunpy.map.Map(zip(datas,headers),cube=True) #Set the date/time for i,m in enumerate(iris_cube): m.meta['DATE-OBS'] = splits[i].center().isoformat() return iris_cube ## Instruction: Change hdu[0] to hdu for optional indexing ## Code After: import sunpy.io import sunpy.time import sunpy.map __all__ = ['SJI_to_cube'] def SJI_to_cube(filename, start=0, stop=None, hdu=0): """ Read a SJI file and return a MapCube ..warning:: This function is a very early beta and is not stable. Further work is on going to improve SunPy IRIS support. Parameters ---------- filename: string File to read start: Temporal axis index to create MapCube from stop: Temporal index to stop MapCube at hdu: Choose hdu index Returns ------- iris_cube: sunpy.map.MapCube A map cube of the SJI sequence """ hdus = sunpy.io.read_file(filename) #Get the time delta time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS']) splits = time_range.split(hdus[hdu][0].shape[0]) if not stop: stop = len(splits) headers = [hdus[hdu][1]]*(stop-start) datas = hdus[hdu][0][start:stop] #Make the cube: iris_cube = sunpy.map.Map(zip(datas,headers),cube=True) #Set the date/time for i,m in enumerate(iris_cube): m.meta['DATE-OBS'] = splits[i].center().isoformat() return iris_cube
import sunpy.io import sunpy.time import sunpy.map __all__ = ['SJI_to_cube'] - def SJI_to_cube(filename, start=0, stop=None): + def SJI_to_cube(filename, start=0, stop=None, hdu=0): ? +++++++ """ Read a SJI file and return a MapCube ..warning:: This function is a very early beta and is not stable. Further work is on going to improve SunPy IRIS support. Parameters ---------- filename: string File to read start: Temporal axis index to create MapCube from stop: Temporal index to stop MapCube at + hdu: + Choose hdu index + Returns ------- iris_cube: sunpy.map.MapCube A map cube of the SJI sequence """ hdus = sunpy.io.read_file(filename) #Get the time delta - time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS']) ? ^ ^ + time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS']) ? ^^^ ^^^ - splits = time_range.split(hdus[0][0].shape[0]) ? ^ + splits = time_range.split(hdus[hdu][0].shape[0]) ? ^^^ if not stop: stop = len(splits) - headers = [hdus[0][1]]*(stop-start) ? ^ + headers = [hdus[hdu][1]]*(stop-start) ? ^^^ - datas = hdus[0][0][start:stop] ? ^ + datas = hdus[hdu][0][start:stop] ? ^^^ #Make the cube: iris_cube = sunpy.map.Map(zip(datas,headers),cube=True) #Set the date/time for i,m in enumerate(iris_cube): m.meta['DATE-OBS'] = splits[i].center().isoformat() return iris_cube
43f67067c470386b6b24080642cc845ec1655f58
utils/networking.py
utils/networking.py
import fcntl import socket import struct from contextlib import contextmanager @contextmanager def use_interface(ifname): """ :type ifname: str """ ip = _ip_address_for_interface(ifname.encode('ascii')) original_socket = socket.socket def rebound_socket(*args, **kwargs): sock = original_socket(*args, **kwargs) sock.bind((ip, 0)) return sock socket.socket = rebound_socket yield socket.socket = original_socket def _ip_address_for_interface(ifname): """ :type ifname: bytes :rtype: str """ sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) return socket.inet_ntoa(fcntl.ioctl( sock.fileno(), 0x8915, # SIOCGIFADDR struct.pack('256s', ifname[:15]) )[20:24])
import fcntl import socket import struct from contextlib import contextmanager @contextmanager def use_interface(ifname): """ :type ifname: str """ ip = _ip_address_for_interface(ifname) original_socket = socket.socket def rebound_socket(*args, **kwargs): sock = original_socket(*args, **kwargs) sock.bind((ip, 0)) return sock socket.socket = rebound_socket yield socket.socket = original_socket def _ip_address_for_interface(ifname): """ :type ifname: str :rtype: str """ ifname = ifname.encode('ascii') sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) return socket.inet_ntoa(fcntl.ioctl( sock.fileno(), 0x8915, # SIOCGIFADDR struct.pack('256s', ifname[:15]) )[20:24])
Make _ip_address_for_interface easier to use
Make _ip_address_for_interface easier to use
Python
apache-2.0
OPWEN/opwen-webapp,ascoderu/opwen-webapp,ascoderu/opwen-webapp,OPWEN/opwen-webapp,OPWEN/opwen-webapp,ascoderu/opwen-cloudserver,ascoderu/opwen-cloudserver,ascoderu/opwen-webapp
import fcntl import socket import struct from contextlib import contextmanager @contextmanager def use_interface(ifname): """ :type ifname: str """ - ip = _ip_address_for_interface(ifname.encode('ascii')) + ip = _ip_address_for_interface(ifname) original_socket = socket.socket def rebound_socket(*args, **kwargs): sock = original_socket(*args, **kwargs) sock.bind((ip, 0)) return sock socket.socket = rebound_socket yield socket.socket = original_socket def _ip_address_for_interface(ifname): """ - :type ifname: bytes + :type ifname: str :rtype: str """ + ifname = ifname.encode('ascii') sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) return socket.inet_ntoa(fcntl.ioctl( sock.fileno(), 0x8915, # SIOCGIFADDR struct.pack('256s', ifname[:15]) )[20:24])
Make _ip_address_for_interface easier to use
## Code Before: import fcntl import socket import struct from contextlib import contextmanager @contextmanager def use_interface(ifname): """ :type ifname: str """ ip = _ip_address_for_interface(ifname.encode('ascii')) original_socket = socket.socket def rebound_socket(*args, **kwargs): sock = original_socket(*args, **kwargs) sock.bind((ip, 0)) return sock socket.socket = rebound_socket yield socket.socket = original_socket def _ip_address_for_interface(ifname): """ :type ifname: bytes :rtype: str """ sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) return socket.inet_ntoa(fcntl.ioctl( sock.fileno(), 0x8915, # SIOCGIFADDR struct.pack('256s', ifname[:15]) )[20:24]) ## Instruction: Make _ip_address_for_interface easier to use ## Code After: import fcntl import socket import struct from contextlib import contextmanager @contextmanager def use_interface(ifname): """ :type ifname: str """ ip = _ip_address_for_interface(ifname) original_socket = socket.socket def rebound_socket(*args, **kwargs): sock = original_socket(*args, **kwargs) sock.bind((ip, 0)) return sock socket.socket = rebound_socket yield socket.socket = original_socket def _ip_address_for_interface(ifname): """ :type ifname: str :rtype: str """ ifname = ifname.encode('ascii') sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) return socket.inet_ntoa(fcntl.ioctl( sock.fileno(), 0x8915, # SIOCGIFADDR struct.pack('256s', ifname[:15]) )[20:24])
import fcntl import socket import struct from contextlib import contextmanager @contextmanager def use_interface(ifname): """ :type ifname: str """ - ip = _ip_address_for_interface(ifname.encode('ascii')) ? --------------- - + ip = _ip_address_for_interface(ifname) original_socket = socket.socket def rebound_socket(*args, **kwargs): sock = original_socket(*args, **kwargs) sock.bind((ip, 0)) return sock socket.socket = rebound_socket yield socket.socket = original_socket def _ip_address_for_interface(ifname): """ - :type ifname: bytes ? ^^ ^^ + :type ifname: str ? ^ ^ :rtype: str """ + ifname = ifname.encode('ascii') sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM) return socket.inet_ntoa(fcntl.ioctl( sock.fileno(), 0x8915, # SIOCGIFADDR struct.pack('256s', ifname[:15]) )[20:24])
ded80de3c276b57cd36d94ab393937289f772a25
django_prometheus/db/backends/postgresql/base.py
django_prometheus/db/backends/postgresql/base.py
import django import psycopg2.extensions from django_prometheus.db.common import DatabaseWrapperMixin, \ ExportingCursorWrapper if django.VERSION >= (1, 9): from django.db.backends.postgresql import base else: from django.db.backends.postgresql_psycopg2 import base class DatabaseFeatures(base.DatabaseFeatures): """Our database has the exact same features as the base one.""" pass class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper): def get_connection_params(self): conn_params = super(DatabaseWrapper, self).get_connection_params() conn_params['cursor_factory'] = ExportingCursorWrapper( psycopg2.extensions.cursor, self.alias, self.vendor, ) return conn_params def create_cursor(self, name=None): # cursor_factory is a kwarg to connect() so restore create_cursor()'s # default behavior return base.DatabaseWrapper.create_cursor(self, name=name)
import django import psycopg2.extensions from django_prometheus.db.common import DatabaseWrapperMixin, \ ExportingCursorWrapper if django.VERSION >= (1, 9): from django.db.backends.postgresql import base else: from django.db.backends.postgresql_psycopg2 import base class DatabaseFeatures(base.DatabaseFeatures): """Our database has the exact same features as the base one.""" pass class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper): def get_connection_params(self): conn_params = super(DatabaseWrapper, self).get_connection_params() conn_params['cursor_factory'] = ExportingCursorWrapper( psycopg2.extensions.cursor, self.alias, self.vendor, ) return conn_params def create_cursor(self, name=None): # cursor_factory is a kwarg to connect() so restore create_cursor()'s # default behavior if django.VERSION >= (1, 11, 0): return base.DatabaseWrapper.create_cursor(self, name=name) else: return base.DatabaseWrapper.create_cursor(self)
Fix backwards compatibility for postgresql backend on Django 1.10 and earlier
Fix backwards compatibility for postgresql backend on Django 1.10 and earlier
Python
apache-2.0
korfuri/django-prometheus,obytes/django-prometheus,korfuri/django-prometheus,obytes/django-prometheus
import django import psycopg2.extensions from django_prometheus.db.common import DatabaseWrapperMixin, \ ExportingCursorWrapper if django.VERSION >= (1, 9): from django.db.backends.postgresql import base else: from django.db.backends.postgresql_psycopg2 import base class DatabaseFeatures(base.DatabaseFeatures): """Our database has the exact same features as the base one.""" pass class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper): def get_connection_params(self): conn_params = super(DatabaseWrapper, self).get_connection_params() conn_params['cursor_factory'] = ExportingCursorWrapper( psycopg2.extensions.cursor, self.alias, self.vendor, ) return conn_params def create_cursor(self, name=None): # cursor_factory is a kwarg to connect() so restore create_cursor()'s # default behavior + if django.VERSION >= (1, 11, 0): - return base.DatabaseWrapper.create_cursor(self, name=name) + return base.DatabaseWrapper.create_cursor(self, name=name) + else: + return base.DatabaseWrapper.create_cursor(self)
Fix backwards compatibility for postgresql backend on Django 1.10 and earlier
## Code Before: import django import psycopg2.extensions from django_prometheus.db.common import DatabaseWrapperMixin, \ ExportingCursorWrapper if django.VERSION >= (1, 9): from django.db.backends.postgresql import base else: from django.db.backends.postgresql_psycopg2 import base class DatabaseFeatures(base.DatabaseFeatures): """Our database has the exact same features as the base one.""" pass class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper): def get_connection_params(self): conn_params = super(DatabaseWrapper, self).get_connection_params() conn_params['cursor_factory'] = ExportingCursorWrapper( psycopg2.extensions.cursor, self.alias, self.vendor, ) return conn_params def create_cursor(self, name=None): # cursor_factory is a kwarg to connect() so restore create_cursor()'s # default behavior return base.DatabaseWrapper.create_cursor(self, name=name) ## Instruction: Fix backwards compatibility for postgresql backend on Django 1.10 and earlier ## Code After: import django import psycopg2.extensions from django_prometheus.db.common import DatabaseWrapperMixin, \ ExportingCursorWrapper if django.VERSION >= (1, 9): from django.db.backends.postgresql import base else: from django.db.backends.postgresql_psycopg2 import base class DatabaseFeatures(base.DatabaseFeatures): """Our database has the exact same features as the base one.""" pass class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper): def get_connection_params(self): conn_params = super(DatabaseWrapper, self).get_connection_params() conn_params['cursor_factory'] = ExportingCursorWrapper( psycopg2.extensions.cursor, self.alias, self.vendor, ) return conn_params def create_cursor(self, name=None): # cursor_factory is a kwarg to connect() so restore create_cursor()'s # default behavior if django.VERSION >= (1, 11, 0): return base.DatabaseWrapper.create_cursor(self, name=name) else: return base.DatabaseWrapper.create_cursor(self)
import django import psycopg2.extensions from django_prometheus.db.common import DatabaseWrapperMixin, \ ExportingCursorWrapper if django.VERSION >= (1, 9): from django.db.backends.postgresql import base else: from django.db.backends.postgresql_psycopg2 import base class DatabaseFeatures(base.DatabaseFeatures): """Our database has the exact same features as the base one.""" pass class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper): def get_connection_params(self): conn_params = super(DatabaseWrapper, self).get_connection_params() conn_params['cursor_factory'] = ExportingCursorWrapper( psycopg2.extensions.cursor, self.alias, self.vendor, ) return conn_params def create_cursor(self, name=None): # cursor_factory is a kwarg to connect() so restore create_cursor()'s # default behavior + if django.VERSION >= (1, 11, 0): - return base.DatabaseWrapper.create_cursor(self, name=name) + return base.DatabaseWrapper.create_cursor(self, name=name) ? ++++ + else: + return base.DatabaseWrapper.create_cursor(self)
20e8ef6bd68100a70b9d50013630ff71d8b7ec94
changes/artifacts/__init__.py
changes/artifacts/__init__.py
from __future__ import absolute_import, print_function from .manager import Manager from .coverage import CoverageHandler from .xunit import XunitHandler manager = Manager() manager.register(CoverageHandler, ['coverage.xml']) manager.register(XunitHandler, ['xunit.xml', 'junit.xml'])
from __future__ import absolute_import, print_function from .manager import Manager from .coverage import CoverageHandler from .xunit import XunitHandler manager = Manager() manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml']) manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml'])
Support wildcard matches on coverage/junit results
Support wildcard matches on coverage/junit results
Python
apache-2.0
dropbox/changes,bowlofstew/changes,wfxiang08/changes,dropbox/changes,wfxiang08/changes,bowlofstew/changes,bowlofstew/changes,dropbox/changes,bowlofstew/changes,wfxiang08/changes,wfxiang08/changes,dropbox/changes
from __future__ import absolute_import, print_function from .manager import Manager from .coverage import CoverageHandler from .xunit import XunitHandler manager = Manager() - manager.register(CoverageHandler, ['coverage.xml']) + manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml']) - manager.register(XunitHandler, ['xunit.xml', 'junit.xml']) + manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml'])
Support wildcard matches on coverage/junit results
## Code Before: from __future__ import absolute_import, print_function from .manager import Manager from .coverage import CoverageHandler from .xunit import XunitHandler manager = Manager() manager.register(CoverageHandler, ['coverage.xml']) manager.register(XunitHandler, ['xunit.xml', 'junit.xml']) ## Instruction: Support wildcard matches on coverage/junit results ## Code After: from __future__ import absolute_import, print_function from .manager import Manager from .coverage import CoverageHandler from .xunit import XunitHandler manager = Manager() manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml']) manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml'])
from __future__ import absolute_import, print_function from .manager import Manager from .coverage import CoverageHandler from .xunit import XunitHandler manager = Manager() - manager.register(CoverageHandler, ['coverage.xml']) + manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml']) ? ++++++++++++++++++ - manager.register(XunitHandler, ['xunit.xml', 'junit.xml']) + manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml']) ? ++++++++++++++++++++++++++++++
799d6738bd189fa202f45c10e7b5361f71f14c57
bin/request_domain.py
bin/request_domain.py
import requests, sys, json otp = sys.argv[1] domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain' domain_txt_path = '/opt/cloudfleet/data/config/domain.txt' print('retrieving domain for blimp: ' + domain_req_url) r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp}) if r.status_code == 200: print('Success: %s' % r.text) result_dict = r.json() if "domain" in result_dict: with open(domain_txt_path, 'w') as domain_txt_file: domain_txt_file.write(result_dict['domain']) quit(0) else: quit(1) else: print('Error: %s' % r.text) quit(1)
import requests, sys, json otp = sys.argv[1] domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain' domain_txt_path = '/opt/cloudfleet/data/config/domain.txt' print('retrieving domain for blimp: ' + domain_req_url) r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp}) if r.status_code == 200: print('Success: %s' % r.text) result_dict = r.json() if "domain" in result_dict: with open(domain_txt_path, 'w') as domain_txt_file: domain_txt_file.write(result_dict['domain']) quit(0) else: quit(1) else: print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text) quit(1)
Clarify error if otp is wrong
Clarify error if otp is wrong
Python
agpl-3.0
cloudfleet/blimp-engineroom,cloudfleet/blimp-engineroom
import requests, sys, json otp = sys.argv[1] domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain' domain_txt_path = '/opt/cloudfleet/data/config/domain.txt' print('retrieving domain for blimp: ' + domain_req_url) r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp}) if r.status_code == 200: print('Success: %s' % r.text) result_dict = r.json() if "domain" in result_dict: with open(domain_txt_path, 'w') as domain_txt_file: domain_txt_file.write(result_dict['domain']) quit(0) else: quit(1) else: - print('Error: %s' % r.text) + print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text) quit(1)
Clarify error if otp is wrong
## Code Before: import requests, sys, json otp = sys.argv[1] domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain' domain_txt_path = '/opt/cloudfleet/data/config/domain.txt' print('retrieving domain for blimp: ' + domain_req_url) r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp}) if r.status_code == 200: print('Success: %s' % r.text) result_dict = r.json() if "domain" in result_dict: with open(domain_txt_path, 'w') as domain_txt_file: domain_txt_file.write(result_dict['domain']) quit(0) else: quit(1) else: print('Error: %s' % r.text) quit(1) ## Instruction: Clarify error if otp is wrong ## Code After: import requests, sys, json otp = sys.argv[1] domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain' domain_txt_path = '/opt/cloudfleet/data/config/domain.txt' print('retrieving domain for blimp: ' + domain_req_url) r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp}) if r.status_code == 200: print('Success: %s' % r.text) result_dict = r.json() if "domain" in result_dict: with open(domain_txt_path, 'w') as domain_txt_file: domain_txt_file.write(result_dict['domain']) quit(0) else: quit(1) else: print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text) quit(1)
import requests, sys, json otp = sys.argv[1] domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain' domain_txt_path = '/opt/cloudfleet/data/config/domain.txt' print('retrieving domain for blimp: ' + domain_req_url) r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp}) if r.status_code == 200: print('Success: %s' % r.text) result_dict = r.json() if "domain" in result_dict: with open(domain_txt_path, 'w') as domain_txt_file: domain_txt_file.write(result_dict['domain']) quit(0) else: quit(1) else: - print('Error: %s' % r.text) + print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text) quit(1)
34ac848cc19477f032a78a4ccbc782d2694d1969
bluebottle/votes/models.py
bluebottle/votes/models.py
from django.db import models from django.conf import settings from django.utils.translation import ugettext as _ from django_extensions.db.fields import CreationDateTimeField class Vote(models.Model): """ Mixin for generating an invoice reference. """ created = CreationDateTimeField(_('created')) project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL) voter = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('voter'), blank=True, null=True) def __unicode__(self): return "{0} -> {1}".format(self.voter, self.project) class Meta: unique_together = (('project', 'voter'), )
from django.db import models from django.conf import settings from django.utils.translation import ugettext as _ from django_extensions.db.fields import CreationDateTimeField class Vote(models.Model): """ Mixin for generating an invoice reference. """ created = CreationDateTimeField(_('created')) project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL) voter = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('voter'), blank=True, null=True) def __unicode__(self): return "{0} -> {1}".format(self.voter, self.project) class Meta: unique_together = (('project', 'voter'), ) ordering = ('-created', )
Sort votes by created desc
Sort votes by created desc BB-4430 #resolve
Python
bsd-3-clause
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle
from django.db import models from django.conf import settings from django.utils.translation import ugettext as _ from django_extensions.db.fields import CreationDateTimeField class Vote(models.Model): """ Mixin for generating an invoice reference. """ created = CreationDateTimeField(_('created')) project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL) voter = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('voter'), blank=True, null=True) def __unicode__(self): return "{0} -> {1}".format(self.voter, self.project) class Meta: unique_together = (('project', 'voter'), ) + ordering = ('-created', )
Sort votes by created desc
## Code Before: from django.db import models from django.conf import settings from django.utils.translation import ugettext as _ from django_extensions.db.fields import CreationDateTimeField class Vote(models.Model): """ Mixin for generating an invoice reference. """ created = CreationDateTimeField(_('created')) project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL) voter = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('voter'), blank=True, null=True) def __unicode__(self): return "{0} -> {1}".format(self.voter, self.project) class Meta: unique_together = (('project', 'voter'), ) ## Instruction: Sort votes by created desc ## Code After: from django.db import models from django.conf import settings from django.utils.translation import ugettext as _ from django_extensions.db.fields import CreationDateTimeField class Vote(models.Model): """ Mixin for generating an invoice reference. """ created = CreationDateTimeField(_('created')) project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL) voter = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('voter'), blank=True, null=True) def __unicode__(self): return "{0} -> {1}".format(self.voter, self.project) class Meta: unique_together = (('project', 'voter'), ) ordering = ('-created', )
from django.db import models from django.conf import settings from django.utils.translation import ugettext as _ from django_extensions.db.fields import CreationDateTimeField class Vote(models.Model): """ Mixin for generating an invoice reference. """ created = CreationDateTimeField(_('created')) project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL) voter = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('voter'), blank=True, null=True) def __unicode__(self): return "{0} -> {1}".format(self.voter, self.project) class Meta: unique_together = (('project', 'voter'), ) + ordering = ('-created', )
5978eedb3147bc0f124335d9e408d6c4895de3a7
__init__.py
__init__.py
import os import sys import marshal if sys.hexversion < 0x03030000: raise ImportError('python >= 3.3 required') if sys.implementation.cache_tag is None: raise ImportError('python implementation does not use bytecode') PY_TAG = sys.implementation.cache_tag PY_VERSION = sys.hexversion BUNDLE_DIR = os.path.join(__path__[0], 'bundle') BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle') if not os.path.exists(BUNDLE_FILE): raise ImportError('python implementation {!r} not supported'.format(PY_TAG)) with open(BUNDLE_FILE, 'rb') as _fd: for _c in marshal.load(_fd): eval(_c) del _c del _fd
import os import sys import marshal if not hasattr(sys, 'implementation'): raise ImportError('python >= 3.3 required') if sys.implementation.cache_tag is None: raise ImportError('python implementation does not use bytecode') PY_TAG = sys.implementation.cache_tag PY_VERSION = sys.hexversion BUNDLE_DIR = os.path.join(__path__[0], 'bundle') BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle') if not os.path.exists(BUNDLE_FILE): raise ImportError('python implementation {!r} not supported'.format(PY_TAG)) with open(BUNDLE_FILE, 'rb') as _fd: for _c in marshal.load(_fd): eval(_c) del _c del _fd
Use a different way of ensuring 3.3+.
Use a different way of ensuring 3.3+.
Python
mit
pyos/dg
import os import sys import marshal - if sys.hexversion < 0x03030000: + if not hasattr(sys, 'implementation'): raise ImportError('python >= 3.3 required') if sys.implementation.cache_tag is None: raise ImportError('python implementation does not use bytecode') PY_TAG = sys.implementation.cache_tag PY_VERSION = sys.hexversion BUNDLE_DIR = os.path.join(__path__[0], 'bundle') BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle') if not os.path.exists(BUNDLE_FILE): raise ImportError('python implementation {!r} not supported'.format(PY_TAG)) with open(BUNDLE_FILE, 'rb') as _fd: for _c in marshal.load(_fd): eval(_c) del _c del _fd
Use a different way of ensuring 3.3+.
## Code Before: import os import sys import marshal if sys.hexversion < 0x03030000: raise ImportError('python >= 3.3 required') if sys.implementation.cache_tag is None: raise ImportError('python implementation does not use bytecode') PY_TAG = sys.implementation.cache_tag PY_VERSION = sys.hexversion BUNDLE_DIR = os.path.join(__path__[0], 'bundle') BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle') if not os.path.exists(BUNDLE_FILE): raise ImportError('python implementation {!r} not supported'.format(PY_TAG)) with open(BUNDLE_FILE, 'rb') as _fd: for _c in marshal.load(_fd): eval(_c) del _c del _fd ## Instruction: Use a different way of ensuring 3.3+. ## Code After: import os import sys import marshal if not hasattr(sys, 'implementation'): raise ImportError('python >= 3.3 required') if sys.implementation.cache_tag is None: raise ImportError('python implementation does not use bytecode') PY_TAG = sys.implementation.cache_tag PY_VERSION = sys.hexversion BUNDLE_DIR = os.path.join(__path__[0], 'bundle') BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle') if not os.path.exists(BUNDLE_FILE): raise ImportError('python implementation {!r} not supported'.format(PY_TAG)) with open(BUNDLE_FILE, 'rb') as _fd: for _c in marshal.load(_fd): eval(_c) del _c del _fd
import os import sys import marshal - if sys.hexversion < 0x03030000: + if not hasattr(sys, 'implementation'): raise ImportError('python >= 3.3 required') if sys.implementation.cache_tag is None: raise ImportError('python implementation does not use bytecode') PY_TAG = sys.implementation.cache_tag PY_VERSION = sys.hexversion BUNDLE_DIR = os.path.join(__path__[0], 'bundle') BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle') if not os.path.exists(BUNDLE_FILE): raise ImportError('python implementation {!r} not supported'.format(PY_TAG)) with open(BUNDLE_FILE, 'rb') as _fd: for _c in marshal.load(_fd): eval(_c) del _c del _fd
fa2fb3387912474eff2b6c2a14d6304fcf5cd1f8
erasmus/cogs/bible/testing_server_preferences_group.py
erasmus/cogs/bible/testing_server_preferences_group.py
from __future__ import annotations from typing import TYPE_CHECKING from botus_receptus.app_commands import test_guilds_only from discord import app_commands from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup if TYPE_CHECKING: from ...erasmus import Erasmus from ...l10n import GroupLocalizer from .types import ParentCog @app_commands.default_permissions(administrator=True) @app_commands.guild_only() @test_guilds_only class TestingServerPreferencesGroup( app_commands.Group, name='test-server-prefs', description='Testing group' ): bot: Erasmus localizer: GroupLocalizer daily_bread = DailyBreadPreferencesGroup() def initialize_from_parent(self, parent: ParentCog, /) -> None: self.bot = parent.bot self.localizer = parent.localizer.for_group(self) self.daily_bread.initialize_from_parent(self)
from __future__ import annotations from typing import TYPE_CHECKING from botus_receptus.app_commands import test_guilds_only from discord import app_commands from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup if TYPE_CHECKING: from ...erasmus import Erasmus from ...l10n import GroupLocalizer from .types import ParentCog @app_commands.default_permissions(administrator=True) @app_commands.guild_only() @test_guilds_only class TestingServerPreferencesGroup( app_commands.Group, name='test-server-prefs', description='Testing group' ): bot: Erasmus localizer: GroupLocalizer daily_bread = DailyBreadPreferencesGroup() def initialize_from_parent(self, parent: ParentCog, /) -> None: self.bot = parent.bot self.localizer = parent.localizer.for_group('serverprefs') self.daily_bread.initialize_from_parent(self)
Use serverprefs localizer for TestingServerPreferencesGroup
Use serverprefs localizer for TestingServerPreferencesGroup
Python
bsd-3-clause
bryanforbes/Erasmus
from __future__ import annotations from typing import TYPE_CHECKING from botus_receptus.app_commands import test_guilds_only from discord import app_commands from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup if TYPE_CHECKING: from ...erasmus import Erasmus from ...l10n import GroupLocalizer from .types import ParentCog @app_commands.default_permissions(administrator=True) @app_commands.guild_only() @test_guilds_only class TestingServerPreferencesGroup( app_commands.Group, name='test-server-prefs', description='Testing group' ): bot: Erasmus localizer: GroupLocalizer daily_bread = DailyBreadPreferencesGroup() def initialize_from_parent(self, parent: ParentCog, /) -> None: self.bot = parent.bot - self.localizer = parent.localizer.for_group(self) + self.localizer = parent.localizer.for_group('serverprefs') self.daily_bread.initialize_from_parent(self)
Use serverprefs localizer for TestingServerPreferencesGroup
## Code Before: from __future__ import annotations from typing import TYPE_CHECKING from botus_receptus.app_commands import test_guilds_only from discord import app_commands from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup if TYPE_CHECKING: from ...erasmus import Erasmus from ...l10n import GroupLocalizer from .types import ParentCog @app_commands.default_permissions(administrator=True) @app_commands.guild_only() @test_guilds_only class TestingServerPreferencesGroup( app_commands.Group, name='test-server-prefs', description='Testing group' ): bot: Erasmus localizer: GroupLocalizer daily_bread = DailyBreadPreferencesGroup() def initialize_from_parent(self, parent: ParentCog, /) -> None: self.bot = parent.bot self.localizer = parent.localizer.for_group(self) self.daily_bread.initialize_from_parent(self) ## Instruction: Use serverprefs localizer for TestingServerPreferencesGroup ## Code After: from __future__ import annotations from typing import TYPE_CHECKING from botus_receptus.app_commands import test_guilds_only from discord import app_commands from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup if TYPE_CHECKING: from ...erasmus import Erasmus from ...l10n import GroupLocalizer from .types import ParentCog @app_commands.default_permissions(administrator=True) @app_commands.guild_only() @test_guilds_only class TestingServerPreferencesGroup( app_commands.Group, name='test-server-prefs', description='Testing group' ): bot: Erasmus localizer: GroupLocalizer daily_bread = DailyBreadPreferencesGroup() def initialize_from_parent(self, parent: ParentCog, /) -> None: self.bot = parent.bot self.localizer = parent.localizer.for_group('serverprefs') self.daily_bread.initialize_from_parent(self)
from __future__ import annotations from typing import TYPE_CHECKING from botus_receptus.app_commands import test_guilds_only from discord import app_commands from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup if TYPE_CHECKING: from ...erasmus import Erasmus from ...l10n import GroupLocalizer from .types import ParentCog @app_commands.default_permissions(administrator=True) @app_commands.guild_only() @test_guilds_only class TestingServerPreferencesGroup( app_commands.Group, name='test-server-prefs', description='Testing group' ): bot: Erasmus localizer: GroupLocalizer daily_bread = DailyBreadPreferencesGroup() def initialize_from_parent(self, parent: ParentCog, /) -> None: self.bot = parent.bot - self.localizer = parent.localizer.for_group(self) ? ^ + self.localizer = parent.localizer.for_group('serverprefs') ? + ^^^^^^^ ++ self.daily_bread.initialize_from_parent(self)
8c26cb08dd08b7e34352e51b06ecb9129ac201a1
stagecraft/libs/schemas/schemas.py
stagecraft/libs/schemas/schemas.py
from django.conf import settings from json import loads as json_loads from os import path def get_schema(): schema_root = path.join( settings.BASE_DIR, 'stagecraft/apps/datasets/schemas/timestamp.json' ) with open(schema_root) as f: json_f = json_loads(f.read()) return json_f
from django.conf import settings from json import loads as json_loads from os import path def get_schema(): schema_root = path.join( settings.BASE_DIR, 'stagecraft/apps/datasets/schemas/timestamp.json' ) with open(schema_root) as f: schema = json_loads(f.read()) return schema
Make the schema return object a bit more obvious and descriptive
Make the schema return object a bit more obvious and descriptive
Python
mit
alphagov/stagecraft,alphagov/stagecraft,alphagov/stagecraft,alphagov/stagecraft
from django.conf import settings from json import loads as json_loads from os import path def get_schema(): schema_root = path.join( settings.BASE_DIR, 'stagecraft/apps/datasets/schemas/timestamp.json' ) with open(schema_root) as f: - json_f = json_loads(f.read()) + schema = json_loads(f.read()) - return json_f + return schema
Make the schema return object a bit more obvious and descriptive
## Code Before: from django.conf import settings from json import loads as json_loads from os import path def get_schema(): schema_root = path.join( settings.BASE_DIR, 'stagecraft/apps/datasets/schemas/timestamp.json' ) with open(schema_root) as f: json_f = json_loads(f.read()) return json_f ## Instruction: Make the schema return object a bit more obvious and descriptive ## Code After: from django.conf import settings from json import loads as json_loads from os import path def get_schema(): schema_root = path.join( settings.BASE_DIR, 'stagecraft/apps/datasets/schemas/timestamp.json' ) with open(schema_root) as f: schema = json_loads(f.read()) return schema
from django.conf import settings from json import loads as json_loads from os import path def get_schema(): schema_root = path.join( settings.BASE_DIR, 'stagecraft/apps/datasets/schemas/timestamp.json' ) with open(schema_root) as f: - json_f = json_loads(f.read()) ? - ^^^^ + schema = json_loads(f.read()) ? ^^^^^ - return json_f + return schema
1de4a0edd0f3c43b53e3a91c10d23155889791c6
tca/chat/tests.py
tca/chat/tests.py
from django.test import TestCase # Create your tests here.
from django.test import TestCase from django.core.urlresolvers import reverse from urllib import urlencode import json class ViewTestCaseMixin(object): """A mixin providing some convenience methods for testing views. Expects that a ``view_name`` property exists on the class which mixes it in. """ def get_view_url(self, *args, **kwargs): return reverse(self.view_name, args=args, kwargs=kwargs) def build_url(self, base_url, query_dict=None): url_template = "{base_url}?{query_string}" if query_dict is None: return base_url return url_template.format( base_url=base_url, query_string=urlencode(query_dict) ) def get(self, parameters=None, *args, **kwargs): """ Sends a GET request to the view-under-test and returns the response :param parameters: The query string parameters of the GET request """ base_url = self.get_view_url(*args, **kwargs) return self.client.get(self.build_url(base_url, parameters)) def post(self, body=None, content_type='application/json', *args, **kwargs): """ Sends a POST request to the view-under-test and returns the response :param body: The content to be included in the body of the request """ base_url = self.get_view_url(*args, **kwargs) if body is None: body = '' return self.client.post( self.build_url(base_url), body, content_type=content_type) def post_json(self, json_payload, *args, **kwargs): """ Sends a POST request to the view-under-test and returns the response. The body of the POST request is formed by serializing the ``json_payload`` object to JSON. """ payload = json.dumps(json_payload) return self.post( body=payload, content_type='application/json', *args, **kwargs)
Add a helper mixin for view test cases
Add a helper mixin for view test cases The mixin defines some helper methods which are useful when testing views (REST endpoints).
Python
bsd-3-clause
mlalic/TumCampusAppBackend,mlalic/TumCampusAppBackend
from django.test import TestCase - # Create your tests here. + from django.core.urlresolvers import reverse + from urllib import urlencode + import json + + + class ViewTestCaseMixin(object): + """A mixin providing some convenience methods for testing views. + + Expects that a ``view_name`` property exists on the class which + mixes it in. + """ + + def get_view_url(self, *args, **kwargs): + return reverse(self.view_name, args=args, kwargs=kwargs) + + def build_url(self, base_url, query_dict=None): + url_template = "{base_url}?{query_string}" + + if query_dict is None: + return base_url + + return url_template.format( + base_url=base_url, + query_string=urlencode(query_dict) + ) + + def get(self, parameters=None, *args, **kwargs): + """ + Sends a GET request to the view-under-test and returns the response + + :param parameters: The query string parameters of the GET request + """ + base_url = self.get_view_url(*args, **kwargs) + + return self.client.get(self.build_url(base_url, parameters)) + + def post(self, body=None, content_type='application/json', *args, **kwargs): + """ + Sends a POST request to the view-under-test and returns the response + + :param body: The content to be included in the body of the request + """ + base_url = self.get_view_url(*args, **kwargs) + + if body is None: + body = '' + + return self.client.post( + self.build_url(base_url), + body, + content_type=content_type) + + def post_json(self, json_payload, *args, **kwargs): + """ + Sends a POST request to the view-under-test and returns the response. + The body of the POST request is formed by serializing the + ``json_payload`` object to JSON. + """ + payload = json.dumps(json_payload) + + return self.post( + body=payload, + content_type='application/json', + *args, **kwargs) +
Add a helper mixin for view test cases
## Code Before: from django.test import TestCase # Create your tests here. ## Instruction: Add a helper mixin for view test cases ## Code After: from django.test import TestCase from django.core.urlresolvers import reverse from urllib import urlencode import json class ViewTestCaseMixin(object): """A mixin providing some convenience methods for testing views. Expects that a ``view_name`` property exists on the class which mixes it in. """ def get_view_url(self, *args, **kwargs): return reverse(self.view_name, args=args, kwargs=kwargs) def build_url(self, base_url, query_dict=None): url_template = "{base_url}?{query_string}" if query_dict is None: return base_url return url_template.format( base_url=base_url, query_string=urlencode(query_dict) ) def get(self, parameters=None, *args, **kwargs): """ Sends a GET request to the view-under-test and returns the response :param parameters: The query string parameters of the GET request """ base_url = self.get_view_url(*args, **kwargs) return self.client.get(self.build_url(base_url, parameters)) def post(self, body=None, content_type='application/json', *args, **kwargs): """ Sends a POST request to the view-under-test and returns the response :param body: The content to be included in the body of the request """ base_url = self.get_view_url(*args, **kwargs) if body is None: body = '' return self.client.post( self.build_url(base_url), body, content_type=content_type) def post_json(self, json_payload, *args, **kwargs): """ Sends a POST request to the view-under-test and returns the response. The body of the POST request is formed by serializing the ``json_payload`` object to JSON. """ payload = json.dumps(json_payload) return self.post( body=payload, content_type='application/json', *args, **kwargs)
from django.test import TestCase - # Create your tests here. + from django.core.urlresolvers import reverse + from urllib import urlencode + + import json + + + class ViewTestCaseMixin(object): + """A mixin providing some convenience methods for testing views. + + Expects that a ``view_name`` property exists on the class which + mixes it in. + """ + + def get_view_url(self, *args, **kwargs): + return reverse(self.view_name, args=args, kwargs=kwargs) + + def build_url(self, base_url, query_dict=None): + url_template = "{base_url}?{query_string}" + + if query_dict is None: + return base_url + + return url_template.format( + base_url=base_url, + query_string=urlencode(query_dict) + ) + + def get(self, parameters=None, *args, **kwargs): + """ + Sends a GET request to the view-under-test and returns the response + + :param parameters: The query string parameters of the GET request + """ + base_url = self.get_view_url(*args, **kwargs) + + return self.client.get(self.build_url(base_url, parameters)) + + def post(self, body=None, content_type='application/json', *args, **kwargs): + """ + Sends a POST request to the view-under-test and returns the response + + :param body: The content to be included in the body of the request + """ + base_url = self.get_view_url(*args, **kwargs) + + if body is None: + body = '' + + return self.client.post( + self.build_url(base_url), + body, + content_type=content_type) + + def post_json(self, json_payload, *args, **kwargs): + """ + Sends a POST request to the view-under-test and returns the response. + The body of the POST request is formed by serializing the + ``json_payload`` object to JSON. + """ + payload = json.dumps(json_payload) + + return self.post( + body=payload, + content_type='application/json', + *args, **kwargs)
fce1b1bdb5a39bbe57b750cd453a9697b8447d6b
chat.py
chat.py
import re from redis import Redis import json from datetime import datetime def is_valid_chatroom(chatroom): return re.match('[A-Za-z_\\d]+$', chatroom) is not None def get_redis(): return Redis() def get_conversation(chatroom): if chatroom is None or len(chatroom) == 0: return None # if chatroom doesn't exist create it! storage = get_redis() return [ json.loads(m) for m in storage.lrange('notifexample:' + chatroom, 0, -1) ] def send_message(chatroom, user_id, name, message): if '<script>' in message: message += '-- Not this time DefConFags' storage = get_redis() now = datetime.now() created_on = now.strftime('%Y-%m-%d %H:%M:%S') storage.rpush( 'notifexample:' + chatroom, json.dumps({ 'author': name, 'userID': user_id, 'message': message, 'createdOn': created_on }) )
import re from redis import Redis import json from datetime import datetime def is_valid_chatroom(chatroom): return re.match('[A-Za-z_\\d]+$', chatroom) is not None def get_redis(): return Redis() def get_conversation(chatroom): if chatroom is None or len(chatroom) == 0: return None storage = get_redis() return [ json.loads(m) for m in storage.lrange('notifexample:' + chatroom, 0, -1) ] def send_message(chatroom, user_id, name, message): if '<script>' in message: message += '-- Not this time DefConFags' storage = get_redis() now = datetime.now() created_on = now.strftime('%Y-%m-%d %H:%M:%S') # if chatroom doesn't exist create it! storage.rpush( 'notifexample:' + chatroom, json.dumps({ 'author': name, 'userID': user_id, 'message': message, 'createdOn': created_on }) )
Correct position of comment :)
Correct position of comment :)
Python
bsd-3-clause
arturosevilla/notification-server-example,arturosevilla/notification-server-example
import re from redis import Redis import json from datetime import datetime def is_valid_chatroom(chatroom): return re.match('[A-Za-z_\\d]+$', chatroom) is not None def get_redis(): return Redis() def get_conversation(chatroom): if chatroom is None or len(chatroom) == 0: return None - # if chatroom doesn't exist create it! storage = get_redis() return [ json.loads(m) for m in storage.lrange('notifexample:' + chatroom, 0, -1) ] def send_message(chatroom, user_id, name, message): if '<script>' in message: message += '-- Not this time DefConFags' storage = get_redis() now = datetime.now() created_on = now.strftime('%Y-%m-%d %H:%M:%S') + # if chatroom doesn't exist create it! storage.rpush( 'notifexample:' + chatroom, json.dumps({ 'author': name, 'userID': user_id, 'message': message, 'createdOn': created_on }) )
Correct position of comment :)
## Code Before: import re from redis import Redis import json from datetime import datetime def is_valid_chatroom(chatroom): return re.match('[A-Za-z_\\d]+$', chatroom) is not None def get_redis(): return Redis() def get_conversation(chatroom): if chatroom is None or len(chatroom) == 0: return None # if chatroom doesn't exist create it! storage = get_redis() return [ json.loads(m) for m in storage.lrange('notifexample:' + chatroom, 0, -1) ] def send_message(chatroom, user_id, name, message): if '<script>' in message: message += '-- Not this time DefConFags' storage = get_redis() now = datetime.now() created_on = now.strftime('%Y-%m-%d %H:%M:%S') storage.rpush( 'notifexample:' + chatroom, json.dumps({ 'author': name, 'userID': user_id, 'message': message, 'createdOn': created_on }) ) ## Instruction: Correct position of comment :) ## Code After: import re from redis import Redis import json from datetime import datetime def is_valid_chatroom(chatroom): return re.match('[A-Za-z_\\d]+$', chatroom) is not None def get_redis(): return Redis() def get_conversation(chatroom): if chatroom is None or len(chatroom) == 0: return None storage = get_redis() return [ json.loads(m) for m in storage.lrange('notifexample:' + chatroom, 0, -1) ] def send_message(chatroom, user_id, name, message): if '<script>' in message: message += '-- Not this time DefConFags' storage = get_redis() now = datetime.now() created_on = now.strftime('%Y-%m-%d %H:%M:%S') # if chatroom doesn't exist create it! storage.rpush( 'notifexample:' + chatroom, json.dumps({ 'author': name, 'userID': user_id, 'message': message, 'createdOn': created_on }) )
import re from redis import Redis import json from datetime import datetime def is_valid_chatroom(chatroom): return re.match('[A-Za-z_\\d]+$', chatroom) is not None def get_redis(): return Redis() def get_conversation(chatroom): if chatroom is None or len(chatroom) == 0: return None - # if chatroom doesn't exist create it! storage = get_redis() return [ json.loads(m) for m in storage.lrange('notifexample:' + chatroom, 0, -1) ] def send_message(chatroom, user_id, name, message): if '<script>' in message: message += '-- Not this time DefConFags' storage = get_redis() now = datetime.now() created_on = now.strftime('%Y-%m-%d %H:%M:%S') + # if chatroom doesn't exist create it! storage.rpush( 'notifexample:' + chatroom, json.dumps({ 'author': name, 'userID': user_id, 'message': message, 'createdOn': created_on }) )
175c72d97d073a64714cebef05bd37f0221f94fa
test_octave_kernel.py
test_octave_kernel.py
"""Example use of jupyter_kernel_test, with tests for IPython.""" import sys import unittest import jupyter_kernel_test as jkt class OctaveKernelTests(jkt.KernelTests): kernel_name = "octave" language_name = "octave" code_hello_world = "disp('hello, world')" code_display_data = [ {'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'}, {'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'} ] if sys.platform == 'darwin' else [] completion_samples = [ { 'text': 'one', 'matches': {'ones', 'onenormest'}, }, ] code_page_something = "ones?" if __name__ == '__main__': unittest.main()
"""Example use of jupyter_kernel_test, with tests for IPython.""" import sys import unittest import jupyter_kernel_test as jkt class OctaveKernelTests(jkt.KernelTests): kernel_name = "octave" language_name = "octave" code_hello_world = "disp('hello, world')" code_display_data = [ {'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'}, {'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'} ] if sys.platform == 'darwin' else [] completion_samples = [ { 'text': 'acos', 'matches': {'acos', 'acosd', 'acosh'}, }, ] code_page_something = "ones?" if __name__ == '__main__': unittest.main()
Fix tests with Octave 5.
Fix tests with Octave 5.
Python
bsd-3-clause
Calysto/octave_kernel,Calysto/octave_kernel
"""Example use of jupyter_kernel_test, with tests for IPython.""" import sys import unittest import jupyter_kernel_test as jkt class OctaveKernelTests(jkt.KernelTests): kernel_name = "octave" language_name = "octave" code_hello_world = "disp('hello, world')" code_display_data = [ {'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'}, {'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'} ] if sys.platform == 'darwin' else [] completion_samples = [ { - 'text': 'one', + 'text': 'acos', - 'matches': {'ones', 'onenormest'}, + 'matches': {'acos', 'acosd', 'acosh'}, }, ] code_page_something = "ones?" if __name__ == '__main__': unittest.main()
Fix tests with Octave 5.
## Code Before: """Example use of jupyter_kernel_test, with tests for IPython.""" import sys import unittest import jupyter_kernel_test as jkt class OctaveKernelTests(jkt.KernelTests): kernel_name = "octave" language_name = "octave" code_hello_world = "disp('hello, world')" code_display_data = [ {'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'}, {'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'} ] if sys.platform == 'darwin' else [] completion_samples = [ { 'text': 'one', 'matches': {'ones', 'onenormest'}, }, ] code_page_something = "ones?" if __name__ == '__main__': unittest.main() ## Instruction: Fix tests with Octave 5. ## Code After: """Example use of jupyter_kernel_test, with tests for IPython.""" import sys import unittest import jupyter_kernel_test as jkt class OctaveKernelTests(jkt.KernelTests): kernel_name = "octave" language_name = "octave" code_hello_world = "disp('hello, world')" code_display_data = [ {'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'}, {'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'} ] if sys.platform == 'darwin' else [] completion_samples = [ { 'text': 'acos', 'matches': {'acos', 'acosd', 'acosh'}, }, ] code_page_something = "ones?" if __name__ == '__main__': unittest.main()
"""Example use of jupyter_kernel_test, with tests for IPython.""" import sys import unittest import jupyter_kernel_test as jkt class OctaveKernelTests(jkt.KernelTests): kernel_name = "octave" language_name = "octave" code_hello_world = "disp('hello, world')" code_display_data = [ {'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'}, {'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'} ] if sys.platform == 'darwin' else [] completion_samples = [ { - 'text': 'one', ? ^^ + 'text': 'acos', ? ++ ^ - 'matches': {'ones', 'onenormest'}, ? -- ^^^ --- ^ + 'matches': {'acos', 'acosd', 'acosh'}, ? ++ ++ ^^^^^^^^ ^ }, ] code_page_something = "ones?" if __name__ == '__main__': unittest.main()
4d1ab55f2bbe8041421002a91dc4f58783913591
services/search_indexes.py
services/search_indexes.py
from aldryn_search.utils import get_index_base from .models import Service class ServiceIndex(get_index_base()): haystack_use_for_indexing = True index_title = True def get_title(self, obj): # XXX what about language? concatenate all available languages? return obj.name_en def get_index_queryset(self, language): # XXX exclude objects with blank name for the selected language, not simply for EN return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='') def get_model(self): return Service def get_search_data(self, service, language, request): # XXX return data for the selected language, not simply for EN return ' '.join(( service.provider.name_en, service.name_en, service.area_of_service.name_en, service.description_en, service.additional_info_en, service.type.name_en, ))
from aldryn_search.utils import get_index_base from .models import Service class ServiceIndex(get_index_base()): haystack_use_for_indexing = True index_title = True def get_title(self, obj): return obj.name def get_index_queryset(self, language): # For this language's index, don't include services with no name # provided in this language. return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{ 'name_%s' % language: '' }) def get_model(self): return Service def get_search_data(self, service, language, request): description = getattr(service, 'description_%s' % language, '') additional_info = getattr(service, 'additional_info_%s' % language, '') return ' '.join(( service.provider.name, service.name, service.area_of_service.name, description, additional_info, service.type.name, ))
Implement language-specific aspects of indexing
Implement language-specific aspects of indexing
Python
bsd-3-clause
theirc/ServiceInfo,theirc/ServiceInfo,theirc/ServiceInfo,theirc/ServiceInfo
from aldryn_search.utils import get_index_base from .models import Service class ServiceIndex(get_index_base()): haystack_use_for_indexing = True index_title = True def get_title(self, obj): - # XXX what about language? concatenate all available languages? - return obj.name_en + return obj.name def get_index_queryset(self, language): - # XXX exclude objects with blank name for the selected language, not simply for EN + # For this language's index, don't include services with no name + # provided in this language. - return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='') + return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{ + 'name_%s' % language: '' + }) def get_model(self): return Service def get_search_data(self, service, language, request): - # XXX return data for the selected language, not simply for EN + description = getattr(service, 'description_%s' % language, '') + additional_info = getattr(service, 'additional_info_%s' % language, '') return ' '.join(( - service.provider.name_en, + service.provider.name, - service.name_en, + service.name, - service.area_of_service.name_en, + service.area_of_service.name, - service.description_en, + description, - service.additional_info_en, + additional_info, - service.type.name_en, + service.type.name, ))
Implement language-specific aspects of indexing
## Code Before: from aldryn_search.utils import get_index_base from .models import Service class ServiceIndex(get_index_base()): haystack_use_for_indexing = True index_title = True def get_title(self, obj): # XXX what about language? concatenate all available languages? return obj.name_en def get_index_queryset(self, language): # XXX exclude objects with blank name for the selected language, not simply for EN return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='') def get_model(self): return Service def get_search_data(self, service, language, request): # XXX return data for the selected language, not simply for EN return ' '.join(( service.provider.name_en, service.name_en, service.area_of_service.name_en, service.description_en, service.additional_info_en, service.type.name_en, )) ## Instruction: Implement language-specific aspects of indexing ## Code After: from aldryn_search.utils import get_index_base from .models import Service class ServiceIndex(get_index_base()): haystack_use_for_indexing = True index_title = True def get_title(self, obj): return obj.name def get_index_queryset(self, language): # For this language's index, don't include services with no name # provided in this language. return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{ 'name_%s' % language: '' }) def get_model(self): return Service def get_search_data(self, service, language, request): description = getattr(service, 'description_%s' % language, '') additional_info = getattr(service, 'additional_info_%s' % language, '') return ' '.join(( service.provider.name, service.name, service.area_of_service.name, description, additional_info, service.type.name, ))
from aldryn_search.utils import get_index_base from .models import Service class ServiceIndex(get_index_base()): haystack_use_for_indexing = True index_title = True def get_title(self, obj): - # XXX what about language? concatenate all available languages? - return obj.name_en ? --- + return obj.name def get_index_queryset(self, language): - # XXX exclude objects with blank name for the selected language, not simply for EN + # For this language's index, don't include services with no name + # provided in this language. - return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='') ? ^^^^^^^^^^^ + return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{ ? ^^^ + 'name_%s' % language: '' + }) def get_model(self): return Service def get_search_data(self, service, language, request): - # XXX return data for the selected language, not simply for EN + description = getattr(service, 'description_%s' % language, '') + additional_info = getattr(service, 'additional_info_%s' % language, '') return ' '.join(( - service.provider.name_en, ? --- + service.provider.name, - service.name_en, ? --- + service.name, - service.area_of_service.name_en, ? --- + service.area_of_service.name, - service.description_en, ? -------- --- + description, - service.additional_info_en, ? -------- --- + additional_info, - service.type.name_en, ? --- + service.type.name, ))
609864faf36b9a82db9fd63d28b5a0da7a22c4f5
eforge/__init__.py
eforge/__init__.py
from eforge.menu import ItemOrder EFORGE_PLUGIN = { 'name': 'EForge Core', 'credit': 'Copyright &copy; 2010 Element43 and contributors', 'provides': { 'mnu': [('project-page', ItemOrder(000, 'Summary'))], }, } VERSION = (0, 5, 0, 'beta 1') def get_version(): return '%d.%d.%d %s' % VERSION
from eforge.menu import ItemOrder EFORGE_PLUGIN = { 'name': 'EForge Core', 'credit': 'Copyright &copy; 2010 Element43 and contributors', 'provides': { 'mnu': [('project-page', ItemOrder(000, 'Summary'))], }, } VERSION = (0, 5, 99, '(git master)') def get_version(): return '%d.%d.%d %s' % VERSION
Change master version information to 0.5.99 (git master)
Change master version information to 0.5.99 (git master) Todo: We should probably add the smarts to EForge to grab the git revision for master, at least if Dulwich is installed :-)
Python
isc
oshepherd/eforge,oshepherd/eforge,oshepherd/eforge
from eforge.menu import ItemOrder EFORGE_PLUGIN = { 'name': 'EForge Core', 'credit': 'Copyright &copy; 2010 Element43 and contributors', 'provides': { 'mnu': [('project-page', ItemOrder(000, 'Summary'))], }, } - VERSION = (0, 5, 0, 'beta 1') + VERSION = (0, 5, 99, '(git master)') def get_version(): return '%d.%d.%d %s' % VERSION
Change master version information to 0.5.99 (git master)
## Code Before: from eforge.menu import ItemOrder EFORGE_PLUGIN = { 'name': 'EForge Core', 'credit': 'Copyright &copy; 2010 Element43 and contributors', 'provides': { 'mnu': [('project-page', ItemOrder(000, 'Summary'))], }, } VERSION = (0, 5, 0, 'beta 1') def get_version(): return '%d.%d.%d %s' % VERSION ## Instruction: Change master version information to 0.5.99 (git master) ## Code After: from eforge.menu import ItemOrder EFORGE_PLUGIN = { 'name': 'EForge Core', 'credit': 'Copyright &copy; 2010 Element43 and contributors', 'provides': { 'mnu': [('project-page', ItemOrder(000, 'Summary'))], }, } VERSION = (0, 5, 99, '(git master)') def get_version(): return '%d.%d.%d %s' % VERSION
from eforge.menu import ItemOrder EFORGE_PLUGIN = { 'name': 'EForge Core', 'credit': 'Copyright &copy; 2010 Element43 and contributors', 'provides': { 'mnu': [('project-page', ItemOrder(000, 'Summary'))], }, } - VERSION = (0, 5, 0, 'beta 1') + VERSION = (0, 5, 99, '(git master)') def get_version(): return '%d.%d.%d %s' % VERSION
959b5fd80a2eeb4ddb56dea07edd16c1aeabc4ff
userprofile/admin.py
userprofile/admin.py
from django.contrib import admin from .models import Profile, Skill, DutyTime, Group admin.site.register(Profile) admin.site.register(Skill) admin.site.register(DutyTime) admin.site.register(Group)
from django.contrib import admin from .models import Profile, Skill, DutyTime, Group class ProfileAdmin(admin.ModelAdmin): list_filter = ( ('tos_accepted', admin.BooleanFieldListFilter), ) admin.site.register(Profile, ProfileAdmin) admin.site.register(Skill) admin.site.register(DutyTime) admin.site.register(Group)
Add filtering option to see profiles that have not accepted new tos
Add filtering option to see profiles that have not accepted new tos
Python
mit
hackerspace-ntnu/website,hackerspace-ntnu/website,hackerspace-ntnu/website
from django.contrib import admin from .models import Profile, Skill, DutyTime, Group + + class ProfileAdmin(admin.ModelAdmin): + list_filter = ( + ('tos_accepted', admin.BooleanFieldListFilter), + ) + + - admin.site.register(Profile) + admin.site.register(Profile, ProfileAdmin) admin.site.register(Skill) admin.site.register(DutyTime) admin.site.register(Group)
Add filtering option to see profiles that have not accepted new tos
## Code Before: from django.contrib import admin from .models import Profile, Skill, DutyTime, Group admin.site.register(Profile) admin.site.register(Skill) admin.site.register(DutyTime) admin.site.register(Group) ## Instruction: Add filtering option to see profiles that have not accepted new tos ## Code After: from django.contrib import admin from .models import Profile, Skill, DutyTime, Group class ProfileAdmin(admin.ModelAdmin): list_filter = ( ('tos_accepted', admin.BooleanFieldListFilter), ) admin.site.register(Profile, ProfileAdmin) admin.site.register(Skill) admin.site.register(DutyTime) admin.site.register(Group)
from django.contrib import admin from .models import Profile, Skill, DutyTime, Group + + class ProfileAdmin(admin.ModelAdmin): + list_filter = ( + ('tos_accepted', admin.BooleanFieldListFilter), + ) + + - admin.site.register(Profile) + admin.site.register(Profile, ProfileAdmin) ? ++++++++++++++ admin.site.register(Skill) admin.site.register(DutyTime) admin.site.register(Group)
6fbe58692005e5c8b7a9c4f4e98984ae86d347a2
pinax/messages/context_processors.py
pinax/messages/context_processors.py
from .models import Thread def user_messages(request): c = {} if request.user.is_authenticated(): c["inbox_count"] = Thread.inbox(request.user).count() return c
from .models import Thread def user_messages(request): c = {} if request.user.is_authenticated(): c["inbox_threads"] = Thread.inbox(request.user) c["unread_threads"] = Thread.unread(request.user) return c
Return querysets in context processor to be more useful
Return querysets in context processor to be more useful
Python
mit
eldarion/user_messages,pinax/pinax-messages,pinax/pinax-messages,arthur-wsw/pinax-messages,eldarion/user_messages,arthur-wsw/pinax-messages
from .models import Thread def user_messages(request): c = {} if request.user.is_authenticated(): - c["inbox_count"] = Thread.inbox(request.user).count() + c["inbox_threads"] = Thread.inbox(request.user) + c["unread_threads"] = Thread.unread(request.user) return c
Return querysets in context processor to be more useful
## Code Before: from .models import Thread def user_messages(request): c = {} if request.user.is_authenticated(): c["inbox_count"] = Thread.inbox(request.user).count() return c ## Instruction: Return querysets in context processor to be more useful ## Code After: from .models import Thread def user_messages(request): c = {} if request.user.is_authenticated(): c["inbox_threads"] = Thread.inbox(request.user) c["unread_threads"] = Thread.unread(request.user) return c
from .models import Thread def user_messages(request): c = {} if request.user.is_authenticated(): - c["inbox_count"] = Thread.inbox(request.user).count() ? ---- -------- + c["inbox_threads"] = Thread.inbox(request.user) ? ++++++ + c["unread_threads"] = Thread.unread(request.user) return c
aeb68225cc9c999b51b1733bffaf684280044c97
salt/utils/yamldumper.py
salt/utils/yamldumper.py
''' salt.utils.yamldumper ~~~~~~~~~~~~~~~~~~~~~ ''' from __future__ import absolute_import try: from yaml import CDumper as Dumper from yaml import CSafeDumper as SafeDumper except ImportError: from yaml import Dumper from yaml import SafeDumper from salt.utils.odict import OrderedDict class OrderedDumper(Dumper): ''' A YAML dumper that represents python OrderedDict as simple YAML map. ''' class SafeOrderedDumper(SafeDumper): ''' A YAML safe dumper that represents python OrderedDict as simple YAML map. ''' def represent_ordereddict(dumper, data): return dumper.represent_dict(data.items()) OrderedDumper.add_representer(OrderedDict, represent_ordereddict) SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
''' salt.utils.yamldumper ~~~~~~~~~~~~~~~~~~~~~ ''' # pylint: disable=W0232 # class has no __init__ method from __future__ import absolute_import try: from yaml import CDumper as Dumper from yaml import CSafeDumper as SafeDumper except ImportError: from yaml import Dumper from yaml import SafeDumper from salt.utils.odict import OrderedDict class OrderedDumper(Dumper): ''' A YAML dumper that represents python OrderedDict as simple YAML map. ''' class SafeOrderedDumper(SafeDumper): ''' A YAML safe dumper that represents python OrderedDict as simple YAML map. ''' def represent_ordereddict(dumper, data): return dumper.represent_dict(data.items()) OrderedDumper.add_representer(OrderedDict, represent_ordereddict) SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
Disable W0232, no `__init__` method.
Disable W0232, no `__init__` method.
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
''' salt.utils.yamldumper ~~~~~~~~~~~~~~~~~~~~~ ''' + # pylint: disable=W0232 + # class has no __init__ method from __future__ import absolute_import try: from yaml import CDumper as Dumper from yaml import CSafeDumper as SafeDumper except ImportError: from yaml import Dumper from yaml import SafeDumper from salt.utils.odict import OrderedDict class OrderedDumper(Dumper): ''' A YAML dumper that represents python OrderedDict as simple YAML map. ''' class SafeOrderedDumper(SafeDumper): ''' A YAML safe dumper that represents python OrderedDict as simple YAML map. ''' def represent_ordereddict(dumper, data): return dumper.represent_dict(data.items()) OrderedDumper.add_representer(OrderedDict, represent_ordereddict) SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
Disable W0232, no `__init__` method.
## Code Before: ''' salt.utils.yamldumper ~~~~~~~~~~~~~~~~~~~~~ ''' from __future__ import absolute_import try: from yaml import CDumper as Dumper from yaml import CSafeDumper as SafeDumper except ImportError: from yaml import Dumper from yaml import SafeDumper from salt.utils.odict import OrderedDict class OrderedDumper(Dumper): ''' A YAML dumper that represents python OrderedDict as simple YAML map. ''' class SafeOrderedDumper(SafeDumper): ''' A YAML safe dumper that represents python OrderedDict as simple YAML map. ''' def represent_ordereddict(dumper, data): return dumper.represent_dict(data.items()) OrderedDumper.add_representer(OrderedDict, represent_ordereddict) SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict) ## Instruction: Disable W0232, no `__init__` method. ## Code After: ''' salt.utils.yamldumper ~~~~~~~~~~~~~~~~~~~~~ ''' # pylint: disable=W0232 # class has no __init__ method from __future__ import absolute_import try: from yaml import CDumper as Dumper from yaml import CSafeDumper as SafeDumper except ImportError: from yaml import Dumper from yaml import SafeDumper from salt.utils.odict import OrderedDict class OrderedDumper(Dumper): ''' A YAML dumper that represents python OrderedDict as simple YAML map. ''' class SafeOrderedDumper(SafeDumper): ''' A YAML safe dumper that represents python OrderedDict as simple YAML map. ''' def represent_ordereddict(dumper, data): return dumper.represent_dict(data.items()) OrderedDumper.add_representer(OrderedDict, represent_ordereddict) SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
''' salt.utils.yamldumper ~~~~~~~~~~~~~~~~~~~~~ ''' + # pylint: disable=W0232 + # class has no __init__ method from __future__ import absolute_import try: from yaml import CDumper as Dumper from yaml import CSafeDumper as SafeDumper except ImportError: from yaml import Dumper from yaml import SafeDumper from salt.utils.odict import OrderedDict class OrderedDumper(Dumper): ''' A YAML dumper that represents python OrderedDict as simple YAML map. ''' class SafeOrderedDumper(SafeDumper): ''' A YAML safe dumper that represents python OrderedDict as simple YAML map. ''' def represent_ordereddict(dumper, data): return dumper.represent_dict(data.items()) OrderedDumper.add_representer(OrderedDict, represent_ordereddict) SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
a7867806a6bd3abfd6bf2bcac6c490965be000e2
tests/test_completeness.py
tests/test_completeness.py
import unittest as unittest from syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
import unittest as unittest from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
Fix an issue in the tests
Fix an issue in the tests
Python
mit
ksons/jscodegen.py
import unittest as unittest - from syntax import Syntax + from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main() +
Fix an issue in the tests
## Code Before: import unittest as unittest from syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main() ## Instruction: Fix an issue in the tests ## Code After: import unittest as unittest from jscodegen.syntax import Syntax from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
import unittest as unittest - from syntax import Syntax + from jscodegen.syntax import Syntax ? ++++++++++ from jscodegen import CodeGenerator def add_cases(generator): def class_decorator(cls): """Add tests to `cls` generated by `generator()`.""" for f, token in generator(): test = lambda self, i=token, f=f: f(self, i) test.__name__ = "test %s" % token.name setattr(cls, test.__name__, test) return cls return class_decorator def _test_tokens(): def t(self, to): c = CodeGenerator({}) func_name = to.name.lower() try: getattr(c, func_name) self.assertTrue(True, func_name) except AttributeError: self.fail("Not implemented: %s" % func_name) for token in Syntax: yield t, token class TestCase(unittest.TestCase): pass TestCase = add_cases(_test_tokens)(TestCase) if __name__=="__main__": unittest.main()
7bd82f6feb1a34dd7b855cfe2f421232229e19db
pages/search_indexes.py
pages/search_indexes.py
"""Django haystack `SearchIndex` module.""" from pages.models import Page from haystack.indexes import SearchIndex, CharField, DateTimeField from haystack import site class PageIndex(SearchIndex): """Search index for pages content.""" text = CharField(document=True, use_template=True) publication_date = DateTimeField(model_attr='publication_date') def get_queryset(self): """Used when the entire index for model is updated.""" return Page.objects.published() site.register(Page, PageIndex)
"""Django haystack `SearchIndex` module.""" from pages.models import Page from haystack.indexes import SearchIndex, CharField, DateTimeField from haystack import site class PageIndex(SearchIndex): """Search index for pages content.""" text = CharField(document=True, use_template=True) title = CharField(model_attr='title') publication_date = DateTimeField(model_attr='publication_date') def get_queryset(self): """Used when the entire index for model is updated.""" return Page.objects.published() site.register(Page, PageIndex)
Add a title attribute to the SearchIndex for pages.
Add a title attribute to the SearchIndex for pages. This is useful when displaying a list of search results because we can display the title of the result without hitting the database to actually pull the page.
Python
bsd-3-clause
remik/django-page-cms,pombredanne/django-page-cms-1,oliciv/django-page-cms,remik/django-page-cms,remik/django-page-cms,batiste/django-page-cms,oliciv/django-page-cms,batiste/django-page-cms,akaihola/django-page-cms,remik/django-page-cms,akaihola/django-page-cms,pombredanne/django-page-cms-1,oliciv/django-page-cms,batiste/django-page-cms,akaihola/django-page-cms,pombredanne/django-page-cms-1
"""Django haystack `SearchIndex` module.""" from pages.models import Page from haystack.indexes import SearchIndex, CharField, DateTimeField from haystack import site class PageIndex(SearchIndex): """Search index for pages content.""" text = CharField(document=True, use_template=True) + title = CharField(model_attr='title') publication_date = DateTimeField(model_attr='publication_date') def get_queryset(self): """Used when the entire index for model is updated.""" return Page.objects.published() site.register(Page, PageIndex)
Add a title attribute to the SearchIndex for pages.
## Code Before: """Django haystack `SearchIndex` module.""" from pages.models import Page from haystack.indexes import SearchIndex, CharField, DateTimeField from haystack import site class PageIndex(SearchIndex): """Search index for pages content.""" text = CharField(document=True, use_template=True) publication_date = DateTimeField(model_attr='publication_date') def get_queryset(self): """Used when the entire index for model is updated.""" return Page.objects.published() site.register(Page, PageIndex) ## Instruction: Add a title attribute to the SearchIndex for pages. ## Code After: """Django haystack `SearchIndex` module.""" from pages.models import Page from haystack.indexes import SearchIndex, CharField, DateTimeField from haystack import site class PageIndex(SearchIndex): """Search index for pages content.""" text = CharField(document=True, use_template=True) title = CharField(model_attr='title') publication_date = DateTimeField(model_attr='publication_date') def get_queryset(self): """Used when the entire index for model is updated.""" return Page.objects.published() site.register(Page, PageIndex)
"""Django haystack `SearchIndex` module.""" from pages.models import Page from haystack.indexes import SearchIndex, CharField, DateTimeField from haystack import site class PageIndex(SearchIndex): """Search index for pages content.""" text = CharField(document=True, use_template=True) + title = CharField(model_attr='title') publication_date = DateTimeField(model_attr='publication_date') def get_queryset(self): """Used when the entire index for model is updated.""" return Page.objects.published() site.register(Page, PageIndex)
b87073e7c7d4387b6608142de7fd6216a1d093b9
setup.py
setup.py
from distutils.core import setup setup(name = "bitey", description="Bitcode Import Tool", long_description = """ Bitey allows LLVM bitcode to be directly imported into Python as an high performance extension module without the need for writing wrappers. """, license="""BSD""", version = "0.0", author = "David Beazley", author_email = "[email protected]", maintainer = "David Beazley", maintainer_email = "[email protected]", url = "https://github.com/dabeaz/bitey/", packages = ['bitey'], classifiers = [ 'Programming Language :: Python :: 2', ] )
from distutils.core import setup setup(name = "bitey", description="Bitcode Import Tool", long_description = """ Bitey allows LLVM bitcode to be directly imported into Python as an high performance extension module without the need for writing wrappers. """, license="""BSD""", version = "0.0", author = "David Beazley", author_email = "[email protected]", maintainer = "David Beazley", maintainer_email = "[email protected]", url = "https://github.com/dabeaz/bitey/", packages = ['bitey'], classifiers = [ 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
Add trove classifier for Python 3
Add trove classifier for Python 3
Python
bsd-3-clause
dabeaz/bitey,dabeaz/bitey
from distutils.core import setup setup(name = "bitey", description="Bitcode Import Tool", long_description = """ Bitey allows LLVM bitcode to be directly imported into Python as an high performance extension module without the need for writing wrappers. """, license="""BSD""", version = "0.0", author = "David Beazley", author_email = "[email protected]", maintainer = "David Beazley", maintainer_email = "[email protected]", url = "https://github.com/dabeaz/bitey/", packages = ['bitey'], classifiers = [ 'Programming Language :: Python :: 2', + 'Programming Language :: Python :: 3', ] )
Add trove classifier for Python 3
## Code Before: from distutils.core import setup setup(name = "bitey", description="Bitcode Import Tool", long_description = """ Bitey allows LLVM bitcode to be directly imported into Python as an high performance extension module without the need for writing wrappers. """, license="""BSD""", version = "0.0", author = "David Beazley", author_email = "[email protected]", maintainer = "David Beazley", maintainer_email = "[email protected]", url = "https://github.com/dabeaz/bitey/", packages = ['bitey'], classifiers = [ 'Programming Language :: Python :: 2', ] ) ## Instruction: Add trove classifier for Python 3 ## Code After: from distutils.core import setup setup(name = "bitey", description="Bitcode Import Tool", long_description = """ Bitey allows LLVM bitcode to be directly imported into Python as an high performance extension module without the need for writing wrappers. """, license="""BSD""", version = "0.0", author = "David Beazley", author_email = "[email protected]", maintainer = "David Beazley", maintainer_email = "[email protected]", url = "https://github.com/dabeaz/bitey/", packages = ['bitey'], classifiers = [ 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ] )
from distutils.core import setup setup(name = "bitey", description="Bitcode Import Tool", long_description = """ Bitey allows LLVM bitcode to be directly imported into Python as an high performance extension module without the need for writing wrappers. """, license="""BSD""", version = "0.0", author = "David Beazley", author_email = "[email protected]", maintainer = "David Beazley", maintainer_email = "[email protected]", url = "https://github.com/dabeaz/bitey/", packages = ['bitey'], classifiers = [ 'Programming Language :: Python :: 2', + 'Programming Language :: Python :: 3', ] )
e5fd0b527877f5fab1d1a2e76ce32062a4a8d697
bika/lims/browser/batch/samples.py
bika/lims/browser/batch/samples.py
from bika.lims.browser.sample import SamplesView as _SV from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName from zope.interface import implements from Products.CMFPlone.utils import safe_unicode import plone class SamplesView(_SV): def __init__(self, context, request): super(SamplesView, self).__init__(context, request) self.view_url = self.context.absolute_url() + "/samples" if 'path' in self.contentFilter: del(self.contentFilter['path']) def contentsMethod(self, contentFilter): tool = getToolByName(self.context, self.catalog) state = [x for x in self.review_states if x['id'] == self.review_state][0] for k, v in state['contentFilter'].items(): self.contentFilter[k] = v tool_samples = tool(contentFilter) samples = {} for sample in (p.getObject() for p in tool_samples): for ar in sample.getAnalysisRequests(): if ar.getBatch().UID() == self.context.UID(): samples[sample.getId()] = sample return samples.values()
from bika.lims.browser.sample import SamplesView as _SV from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName from zope.interface import implements from Products.CMFPlone.utils import safe_unicode import plone class SamplesView(_SV): def __init__(self, context, request): super(SamplesView, self).__init__(context, request) self.view_url = self.context.absolute_url() + "/samples" if 'path' in self.contentFilter: del(self.contentFilter['path']) def contentsMethod(self, contentFilter): tool = getToolByName(self.context, self.catalog) state = [x for x in self.review_states if x['id'] == self.review_state][0] for k, v in state['contentFilter'].items(): self.contentFilter[k] = v tool_samples = tool(contentFilter) samples = {} for sample in (p.getObject() for p in tool_samples): for ar in sample.getAnalysisRequests(): batch = ar.getBatch() if batch and ar.getBatch().UID() == self.context.UID(): samples[sample.getId()] = sample return samples.values()
Fix exception - batch is not required field of AR
Fix exception - batch is not required field of AR
Python
agpl-3.0
DeBortoliWines/Bika-LIMS,anneline/Bika-LIMS,labsanmartin/Bika-LIMS,rockfruit/bika.lims,anneline/Bika-LIMS,labsanmartin/Bika-LIMS,veroc/Bika-LIMS,anneline/Bika-LIMS,DeBortoliWines/Bika-LIMS,veroc/Bika-LIMS,veroc/Bika-LIMS,labsanmartin/Bika-LIMS,DeBortoliWines/Bika-LIMS,rockfruit/bika.lims
from bika.lims.browser.sample import SamplesView as _SV from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName from zope.interface import implements from Products.CMFPlone.utils import safe_unicode import plone class SamplesView(_SV): def __init__(self, context, request): super(SamplesView, self).__init__(context, request) self.view_url = self.context.absolute_url() + "/samples" if 'path' in self.contentFilter: del(self.contentFilter['path']) def contentsMethod(self, contentFilter): tool = getToolByName(self.context, self.catalog) state = [x for x in self.review_states if x['id'] == self.review_state][0] for k, v in state['contentFilter'].items(): self.contentFilter[k] = v tool_samples = tool(contentFilter) samples = {} for sample in (p.getObject() for p in tool_samples): for ar in sample.getAnalysisRequests(): + batch = ar.getBatch() - if ar.getBatch().UID() == self.context.UID(): + if batch and ar.getBatch().UID() == self.context.UID(): - samples[sample.getId()] = sample + samples[sample.getId()] = sample return samples.values()
Fix exception - batch is not required field of AR
## Code Before: from bika.lims.browser.sample import SamplesView as _SV from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName from zope.interface import implements from Products.CMFPlone.utils import safe_unicode import plone class SamplesView(_SV): def __init__(self, context, request): super(SamplesView, self).__init__(context, request) self.view_url = self.context.absolute_url() + "/samples" if 'path' in self.contentFilter: del(self.contentFilter['path']) def contentsMethod(self, contentFilter): tool = getToolByName(self.context, self.catalog) state = [x for x in self.review_states if x['id'] == self.review_state][0] for k, v in state['contentFilter'].items(): self.contentFilter[k] = v tool_samples = tool(contentFilter) samples = {} for sample in (p.getObject() for p in tool_samples): for ar in sample.getAnalysisRequests(): if ar.getBatch().UID() == self.context.UID(): samples[sample.getId()] = sample return samples.values() ## Instruction: Fix exception - batch is not required field of AR ## Code After: from bika.lims.browser.sample import SamplesView as _SV from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName from zope.interface import implements from Products.CMFPlone.utils import safe_unicode import plone class SamplesView(_SV): def __init__(self, context, request): super(SamplesView, self).__init__(context, request) self.view_url = self.context.absolute_url() + "/samples" if 'path' in self.contentFilter: del(self.contentFilter['path']) def contentsMethod(self, contentFilter): tool = getToolByName(self.context, self.catalog) state = [x for x in self.review_states if x['id'] == self.review_state][0] for k, v in state['contentFilter'].items(): self.contentFilter[k] = v tool_samples = tool(contentFilter) samples = {} for sample in (p.getObject() for p in tool_samples): for ar in sample.getAnalysisRequests(): batch = ar.getBatch() if batch and ar.getBatch().UID() == self.context.UID(): samples[sample.getId()] = sample return samples.values()
from bika.lims.browser.sample import SamplesView as _SV from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName from zope.interface import implements from Products.CMFPlone.utils import safe_unicode import plone class SamplesView(_SV): def __init__(self, context, request): super(SamplesView, self).__init__(context, request) self.view_url = self.context.absolute_url() + "/samples" if 'path' in self.contentFilter: del(self.contentFilter['path']) def contentsMethod(self, contentFilter): tool = getToolByName(self.context, self.catalog) state = [x for x in self.review_states if x['id'] == self.review_state][0] for k, v in state['contentFilter'].items(): self.contentFilter[k] = v tool_samples = tool(contentFilter) samples = {} for sample in (p.getObject() for p in tool_samples): for ar in sample.getAnalysisRequests(): + batch = ar.getBatch() - if ar.getBatch().UID() == self.context.UID(): + if batch and ar.getBatch().UID() == self.context.UID(): ? ++++++++++ - samples[sample.getId()] = sample + samples[sample.getId()] = sample ? ++++ return samples.values()
0241e253c68ca6862a3da26d29a649f65c27ae36
demos/chatroom/experiment.py
demos/chatroom/experiment.py
"""Coordination chatroom game.""" import dallinger as dlgr from dallinger.config import get_config try: unicode = unicode except NameError: # Python 3 unicode = str config = get_config() def extra_settings(): config.register('network', unicode) config.register('n', int) class CoordinationChatroom(dlgr.experiments.Experiment): """Define the structure of the experiment.""" def __init__(self, session): """Initialize the experiment.""" super(CoordinationChatroom, self).__init__(session) self.experiment_repeats = 1 self.num_participants = config.get('n') self.initial_recruitment_size = self.num_participants self.quorum = self.num_participants self.config = config if not self.config.ready: self.config.load_config() self.setup() def create_network(self): """Create a new network by reading the configuration file.""" class_ = getattr( dlgr.networks, self.config.get('network') ) return class_(max_size=self.num_participants) def info_post_request(self, node, info): """Run when a request to create an info is complete.""" for agent in node.neighbors(): node.transmit(what=info, to_whom=agent) def create_node(self, participant, network): """Create a node for a participant.""" return dlgr.nodes.Agent(network=network, participant=participant)
"""Coordination chatroom game.""" import dallinger as dlgr from dallinger.compat import unicode from dallinger.config import get_config config = get_config() def extra_settings(): config.register('network', unicode) config.register('n', int) class CoordinationChatroom(dlgr.experiments.Experiment): """Define the structure of the experiment.""" def __init__(self, session): """Initialize the experiment.""" super(CoordinationChatroom, self).__init__(session) self.experiment_repeats = 1 self.num_participants = config.get('n') self.initial_recruitment_size = self.num_participants self.quorum = self.num_participants self.config = config if not self.config.ready: self.config.load_config() self.setup() def create_network(self): """Create a new network by reading the configuration file.""" class_ = getattr( dlgr.networks, self.config.get('network') ) return class_(max_size=self.num_participants) def info_post_request(self, node, info): """Run when a request to create an info is complete.""" for agent in node.neighbors(): node.transmit(what=info, to_whom=agent) def create_node(self, participant, network): """Create a node for a participant.""" return dlgr.nodes.Agent(network=network, participant=participant)
Use compat for unicode import
Use compat for unicode import
Python
mit
Dallinger/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger
"""Coordination chatroom game.""" import dallinger as dlgr + from dallinger.compat import unicode from dallinger.config import get_config - try: - unicode = unicode - except NameError: # Python 3 - unicode = str config = get_config() def extra_settings(): config.register('network', unicode) config.register('n', int) class CoordinationChatroom(dlgr.experiments.Experiment): """Define the structure of the experiment.""" def __init__(self, session): """Initialize the experiment.""" super(CoordinationChatroom, self).__init__(session) self.experiment_repeats = 1 self.num_participants = config.get('n') self.initial_recruitment_size = self.num_participants self.quorum = self.num_participants self.config = config if not self.config.ready: self.config.load_config() self.setup() def create_network(self): """Create a new network by reading the configuration file.""" class_ = getattr( dlgr.networks, self.config.get('network') ) return class_(max_size=self.num_participants) def info_post_request(self, node, info): """Run when a request to create an info is complete.""" for agent in node.neighbors(): node.transmit(what=info, to_whom=agent) def create_node(self, participant, network): """Create a node for a participant.""" return dlgr.nodes.Agent(network=network, participant=participant)
Use compat for unicode import
## Code Before: """Coordination chatroom game.""" import dallinger as dlgr from dallinger.config import get_config try: unicode = unicode except NameError: # Python 3 unicode = str config = get_config() def extra_settings(): config.register('network', unicode) config.register('n', int) class CoordinationChatroom(dlgr.experiments.Experiment): """Define the structure of the experiment.""" def __init__(self, session): """Initialize the experiment.""" super(CoordinationChatroom, self).__init__(session) self.experiment_repeats = 1 self.num_participants = config.get('n') self.initial_recruitment_size = self.num_participants self.quorum = self.num_participants self.config = config if not self.config.ready: self.config.load_config() self.setup() def create_network(self): """Create a new network by reading the configuration file.""" class_ = getattr( dlgr.networks, self.config.get('network') ) return class_(max_size=self.num_participants) def info_post_request(self, node, info): """Run when a request to create an info is complete.""" for agent in node.neighbors(): node.transmit(what=info, to_whom=agent) def create_node(self, participant, network): """Create a node for a participant.""" return dlgr.nodes.Agent(network=network, participant=participant) ## Instruction: Use compat for unicode import ## Code After: """Coordination chatroom game.""" import dallinger as dlgr from dallinger.compat import unicode from dallinger.config import get_config config = get_config() def extra_settings(): config.register('network', unicode) config.register('n', int) class CoordinationChatroom(dlgr.experiments.Experiment): """Define the structure of the experiment.""" def __init__(self, session): """Initialize the experiment.""" super(CoordinationChatroom, self).__init__(session) self.experiment_repeats = 1 self.num_participants = config.get('n') self.initial_recruitment_size = self.num_participants self.quorum = self.num_participants self.config = config if not self.config.ready: self.config.load_config() self.setup() def create_network(self): """Create a new network by reading the configuration file.""" class_ = getattr( dlgr.networks, self.config.get('network') ) return class_(max_size=self.num_participants) def info_post_request(self, node, info): """Run when a request to create an info is complete.""" for agent in node.neighbors(): node.transmit(what=info, to_whom=agent) def create_node(self, participant, network): """Create a node for a participant.""" return dlgr.nodes.Agent(network=network, participant=participant)
"""Coordination chatroom game.""" import dallinger as dlgr + from dallinger.compat import unicode from dallinger.config import get_config - try: - unicode = unicode - except NameError: # Python 3 - unicode = str config = get_config() def extra_settings(): config.register('network', unicode) config.register('n', int) class CoordinationChatroom(dlgr.experiments.Experiment): """Define the structure of the experiment.""" def __init__(self, session): """Initialize the experiment.""" super(CoordinationChatroom, self).__init__(session) self.experiment_repeats = 1 self.num_participants = config.get('n') self.initial_recruitment_size = self.num_participants self.quorum = self.num_participants self.config = config if not self.config.ready: self.config.load_config() self.setup() def create_network(self): """Create a new network by reading the configuration file.""" class_ = getattr( dlgr.networks, self.config.get('network') ) return class_(max_size=self.num_participants) def info_post_request(self, node, info): """Run when a request to create an info is complete.""" for agent in node.neighbors(): node.transmit(what=info, to_whom=agent) def create_node(self, participant, network): """Create a node for a participant.""" return dlgr.nodes.Agent(network=network, participant=participant)
727ec507284776f3eec91b644cd5bb112bdb0af1
july/people/forms.py
july/people/forms.py
from django import forms class EditUserForm(forms.Form): about_me = forms.CharField(widget=forms.Textarea, required=False) url = forms.CharField(max_length=255, required=False) facebook_url = forms.CharField(max_length=255, required=False) email = forms.EmailField(max_length=255) def __init__(self, *args, **kwargs): user = kwargs.pop('user', None) super(EditUserForm, self).__init__(*args, **kwargs) if user: self.fields['about_me'].initial=user.about_me self.fields['url'].initial=user.url self.fields['facebook_url'].initial=user.facebook_url self.fields['email'].initial=user.email
from django import forms class EditUserForm(forms.Form): about_me = forms.CharField(widget=forms.Textarea, required=False) url = forms.CharField(max_length=255, required=False) facebook_url = forms.CharField(max_length=255, required=False) email = forms.EmailField(max_length=255) def __init__(self, *args, **kwargs): user = kwargs.pop('user', None) super(EditUserForm, self).__init__(*args, **kwargs) if user: self.fields['about_me'].initial=getattr(user, 'about_me', None) self.fields['url'].initial=getattr(user, 'url', None) self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None) self.fields['email'].initial=user.email
Use getattr for expando props
Use getattr for expando props
Python
mit
julython/julython.org,julython/julython.org,ChimeraCoder/GOctober,julython/julython.org,ChimeraCoder/GOctober,julython/julython.org,ChimeraCoder/GOctober
from django import forms class EditUserForm(forms.Form): about_me = forms.CharField(widget=forms.Textarea, required=False) url = forms.CharField(max_length=255, required=False) facebook_url = forms.CharField(max_length=255, required=False) email = forms.EmailField(max_length=255) def __init__(self, *args, **kwargs): user = kwargs.pop('user', None) super(EditUserForm, self).__init__(*args, **kwargs) if user: - self.fields['about_me'].initial=user.about_me + self.fields['about_me'].initial=getattr(user, 'about_me', None) - self.fields['url'].initial=user.url + self.fields['url'].initial=getattr(user, 'url', None) - self.fields['facebook_url'].initial=user.facebook_url + self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None) self.fields['email'].initial=user.email
Use getattr for expando props
## Code Before: from django import forms class EditUserForm(forms.Form): about_me = forms.CharField(widget=forms.Textarea, required=False) url = forms.CharField(max_length=255, required=False) facebook_url = forms.CharField(max_length=255, required=False) email = forms.EmailField(max_length=255) def __init__(self, *args, **kwargs): user = kwargs.pop('user', None) super(EditUserForm, self).__init__(*args, **kwargs) if user: self.fields['about_me'].initial=user.about_me self.fields['url'].initial=user.url self.fields['facebook_url'].initial=user.facebook_url self.fields['email'].initial=user.email ## Instruction: Use getattr for expando props ## Code After: from django import forms class EditUserForm(forms.Form): about_me = forms.CharField(widget=forms.Textarea, required=False) url = forms.CharField(max_length=255, required=False) facebook_url = forms.CharField(max_length=255, required=False) email = forms.EmailField(max_length=255) def __init__(self, *args, **kwargs): user = kwargs.pop('user', None) super(EditUserForm, self).__init__(*args, **kwargs) if user: self.fields['about_me'].initial=getattr(user, 'about_me', None) self.fields['url'].initial=getattr(user, 'url', None) self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None) self.fields['email'].initial=user.email
from django import forms class EditUserForm(forms.Form): about_me = forms.CharField(widget=forms.Textarea, required=False) url = forms.CharField(max_length=255, required=False) facebook_url = forms.CharField(max_length=255, required=False) email = forms.EmailField(max_length=255) def __init__(self, *args, **kwargs): user = kwargs.pop('user', None) super(EditUserForm, self).__init__(*args, **kwargs) if user: - self.fields['about_me'].initial=user.about_me ? ^ + self.fields['about_me'].initial=getattr(user, 'about_me', None) ? ++++++++ ^^^ ++++++++ - self.fields['url'].initial=user.url ? ^ + self.fields['url'].initial=getattr(user, 'url', None) ? ++++++++ ^^^ ++++++++ - self.fields['facebook_url'].initial=user.facebook_url ? ^ + self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None) ? ++++++++ ^^^ ++++++++ self.fields['email'].initial=user.email
41fe44e99361d9006a8b196e9b886ffdb3e8e460
functional_tests/test_evexml.py
functional_tests/test_evexml.py
from django.contrib.staticfiles.testing import StaticLiveServerTestCase from django.test import tag from django.shortcuts import reverse from selenium import webdriver from selenium.webdriver.common.keys import Keys MAX_WAIT = 10 @tag('functional') class SubmissionTest(StaticLiveServerTestCase): """Tests for users who are submitting xml api key. """ @classmethod def setUpClass(cls): super(SubmissionTest, cls).setUpClass() cls.browser = webdriver.Chrome() cls.browser.maximize_window() cls.browser.implicitly_wait(MAX_WAIT) super(SubmissionTest, cls).setUpClass() @classmethod def tearDownClass(cls): cls.browser.refresh() cls.browser.quit() super(SubmissionTest, cls).tearDownClass() def tearDown(self): self.browser.refresh() def test_user_can_see_apikey_form(self): """A user should be able to see the form for submitting api keys. """ # They browse to the eve api keys page. url = self.live_server_url + reverse('eveapi_submit') self.browser.get(self.live_server_url) # They see input boxes for keyID and vCode. keyid_input = self.browser.find_element_by_name('keyID') vcode_input = self.browser.find_element_by_name('vCode')
from django.contrib.staticfiles.testing import StaticLiveServerTestCase from django.test import tag from django.shortcuts import reverse from selenium import webdriver from selenium.webdriver.common.keys import Keys MAX_WAIT = 10 @tag('functional') class SubmissionTest(StaticLiveServerTestCase): """Tests for users who are submitting xml api key. """ @classmethod def setUpClass(cls): super(SubmissionTest, cls).setUpClass() cls.browser = webdriver.Chrome() cls.browser.maximize_window() cls.browser.implicitly_wait(MAX_WAIT) super(SubmissionTest, cls).setUpClass() @classmethod def tearDownClass(cls): cls.browser.refresh() cls.browser.quit() super(SubmissionTest, cls).tearDownClass() def tearDown(self): self.browser.refresh() def test_user_can_see_apikey_form(self): """A user should be able to see the form for submitting api keys. """ # They browse to the eve api keys page. url = self.live_server_url + reverse('eveapi_add') self.browser.get(url) # They see input boxes for keyID and vCode. keyid_input = self.browser.find_element_by_name('keyID') vcode_input = self.browser.find_element_by_name('vCode')
Make test get correct url
Make test get correct url
Python
mit
randomic/aniauth-tdd,randomic/aniauth-tdd
from django.contrib.staticfiles.testing import StaticLiveServerTestCase from django.test import tag from django.shortcuts import reverse from selenium import webdriver from selenium.webdriver.common.keys import Keys MAX_WAIT = 10 @tag('functional') class SubmissionTest(StaticLiveServerTestCase): """Tests for users who are submitting xml api key. """ @classmethod def setUpClass(cls): super(SubmissionTest, cls).setUpClass() cls.browser = webdriver.Chrome() cls.browser.maximize_window() cls.browser.implicitly_wait(MAX_WAIT) super(SubmissionTest, cls).setUpClass() @classmethod def tearDownClass(cls): cls.browser.refresh() cls.browser.quit() super(SubmissionTest, cls).tearDownClass() def tearDown(self): self.browser.refresh() def test_user_can_see_apikey_form(self): """A user should be able to see the form for submitting api keys. """ # They browse to the eve api keys page. - url = self.live_server_url + reverse('eveapi_submit') + url = self.live_server_url + reverse('eveapi_add') - self.browser.get(self.live_server_url) + self.browser.get(url) # They see input boxes for keyID and vCode. keyid_input = self.browser.find_element_by_name('keyID') vcode_input = self.browser.find_element_by_name('vCode')
Make test get correct url
## Code Before: from django.contrib.staticfiles.testing import StaticLiveServerTestCase from django.test import tag from django.shortcuts import reverse from selenium import webdriver from selenium.webdriver.common.keys import Keys MAX_WAIT = 10 @tag('functional') class SubmissionTest(StaticLiveServerTestCase): """Tests for users who are submitting xml api key. """ @classmethod def setUpClass(cls): super(SubmissionTest, cls).setUpClass() cls.browser = webdriver.Chrome() cls.browser.maximize_window() cls.browser.implicitly_wait(MAX_WAIT) super(SubmissionTest, cls).setUpClass() @classmethod def tearDownClass(cls): cls.browser.refresh() cls.browser.quit() super(SubmissionTest, cls).tearDownClass() def tearDown(self): self.browser.refresh() def test_user_can_see_apikey_form(self): """A user should be able to see the form for submitting api keys. """ # They browse to the eve api keys page. url = self.live_server_url + reverse('eveapi_submit') self.browser.get(self.live_server_url) # They see input boxes for keyID and vCode. keyid_input = self.browser.find_element_by_name('keyID') vcode_input = self.browser.find_element_by_name('vCode') ## Instruction: Make test get correct url ## Code After: from django.contrib.staticfiles.testing import StaticLiveServerTestCase from django.test import tag from django.shortcuts import reverse from selenium import webdriver from selenium.webdriver.common.keys import Keys MAX_WAIT = 10 @tag('functional') class SubmissionTest(StaticLiveServerTestCase): """Tests for users who are submitting xml api key. """ @classmethod def setUpClass(cls): super(SubmissionTest, cls).setUpClass() cls.browser = webdriver.Chrome() cls.browser.maximize_window() cls.browser.implicitly_wait(MAX_WAIT) super(SubmissionTest, cls).setUpClass() @classmethod def tearDownClass(cls): cls.browser.refresh() cls.browser.quit() super(SubmissionTest, cls).tearDownClass() def tearDown(self): self.browser.refresh() def test_user_can_see_apikey_form(self): """A user should be able to see the form for submitting api keys. """ # They browse to the eve api keys page. url = self.live_server_url + reverse('eveapi_add') self.browser.get(url) # They see input boxes for keyID and vCode. keyid_input = self.browser.find_element_by_name('keyID') vcode_input = self.browser.find_element_by_name('vCode')
from django.contrib.staticfiles.testing import StaticLiveServerTestCase from django.test import tag from django.shortcuts import reverse from selenium import webdriver from selenium.webdriver.common.keys import Keys MAX_WAIT = 10 @tag('functional') class SubmissionTest(StaticLiveServerTestCase): """Tests for users who are submitting xml api key. """ @classmethod def setUpClass(cls): super(SubmissionTest, cls).setUpClass() cls.browser = webdriver.Chrome() cls.browser.maximize_window() cls.browser.implicitly_wait(MAX_WAIT) super(SubmissionTest, cls).setUpClass() @classmethod def tearDownClass(cls): cls.browser.refresh() cls.browser.quit() super(SubmissionTest, cls).tearDownClass() def tearDown(self): self.browser.refresh() def test_user_can_see_apikey_form(self): """A user should be able to see the form for submitting api keys. """ # They browse to the eve api keys page. - url = self.live_server_url + reverse('eveapi_submit') ? ^^^^^^ + url = self.live_server_url + reverse('eveapi_add') ? ^^^ - self.browser.get(self.live_server_url) ? ----------------- + self.browser.get(url) # They see input boxes for keyID and vCode. keyid_input = self.browser.find_element_by_name('keyID') vcode_input = self.browser.find_element_by_name('vCode')
596613c964311104098e64eeb349216bc7cd0023
saleor/demo/views.py
saleor/demo/views.py
from django.conf import settings from django.shortcuts import render from ..graphql.views import API_PATH, GraphQLView EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API! # # Type queries into this side of the screen, and you will see # intelligent typeaheads aware of the current GraphQL type schema # and live syntax and validation errors highlighted within the text. # # Here is an example query to fetch a list of products: # { products(first: 5, channel: "%(channel_slug)s") { edges { node { id name description } } } } """ % { "channel_slug": settings.DEFAULT_CHANNEL_SLUG } class DemoGraphQLView(GraphQLView): def render_playground(self, request): ctx = { "query": EXAMPLE_QUERY, "api_url": request.build_absolute_uri(str(API_PATH)), } return render(request, "graphql/playground.html", ctx)
from django.conf import settings from django.shortcuts import render from ..graphql.views import GraphQLView EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API! # # Type queries into this side of the screen, and you will see # intelligent typeaheads aware of the current GraphQL type schema # and live syntax and validation errors highlighted within the text. # # Here is an example query to fetch a list of products: # { products(first: 5, channel: "%(channel_slug)s") { edges { node { id name description } } } } """ % { "channel_slug": settings.DEFAULT_CHANNEL_SLUG } class DemoGraphQLView(GraphQLView): def render_playground(self, request): pwa_origin = settings.PWA_ORIGINS[0] ctx = { "query": EXAMPLE_QUERY, "api_url": f"https://{pwa_origin}/graphql/", } return render(request, "graphql/playground.html", ctx)
Fix playground CSP for demo if deployed under proxied domain
Fix playground CSP for demo if deployed under proxied domain
Python
bsd-3-clause
mociepka/saleor,mociepka/saleor,mociepka/saleor
from django.conf import settings from django.shortcuts import render - from ..graphql.views import API_PATH, GraphQLView + from ..graphql.views import GraphQLView EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API! # # Type queries into this side of the screen, and you will see # intelligent typeaheads aware of the current GraphQL type schema # and live syntax and validation errors highlighted within the text. # # Here is an example query to fetch a list of products: # { products(first: 5, channel: "%(channel_slug)s") { edges { node { id name description } } } } """ % { "channel_slug": settings.DEFAULT_CHANNEL_SLUG } class DemoGraphQLView(GraphQLView): def render_playground(self, request): + pwa_origin = settings.PWA_ORIGINS[0] ctx = { "query": EXAMPLE_QUERY, - "api_url": request.build_absolute_uri(str(API_PATH)), + "api_url": f"https://{pwa_origin}/graphql/", } return render(request, "graphql/playground.html", ctx)
Fix playground CSP for demo if deployed under proxied domain
## Code Before: from django.conf import settings from django.shortcuts import render from ..graphql.views import API_PATH, GraphQLView EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API! # # Type queries into this side of the screen, and you will see # intelligent typeaheads aware of the current GraphQL type schema # and live syntax and validation errors highlighted within the text. # # Here is an example query to fetch a list of products: # { products(first: 5, channel: "%(channel_slug)s") { edges { node { id name description } } } } """ % { "channel_slug": settings.DEFAULT_CHANNEL_SLUG } class DemoGraphQLView(GraphQLView): def render_playground(self, request): ctx = { "query": EXAMPLE_QUERY, "api_url": request.build_absolute_uri(str(API_PATH)), } return render(request, "graphql/playground.html", ctx) ## Instruction: Fix playground CSP for demo if deployed under proxied domain ## Code After: from django.conf import settings from django.shortcuts import render from ..graphql.views import GraphQLView EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API! # # Type queries into this side of the screen, and you will see # intelligent typeaheads aware of the current GraphQL type schema # and live syntax and validation errors highlighted within the text. # # Here is an example query to fetch a list of products: # { products(first: 5, channel: "%(channel_slug)s") { edges { node { id name description } } } } """ % { "channel_slug": settings.DEFAULT_CHANNEL_SLUG } class DemoGraphQLView(GraphQLView): def render_playground(self, request): pwa_origin = settings.PWA_ORIGINS[0] ctx = { "query": EXAMPLE_QUERY, "api_url": f"https://{pwa_origin}/graphql/", } return render(request, "graphql/playground.html", ctx)
from django.conf import settings from django.shortcuts import render - from ..graphql.views import API_PATH, GraphQLView ? ---------- + from ..graphql.views import GraphQLView EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API! # # Type queries into this side of the screen, and you will see # intelligent typeaheads aware of the current GraphQL type schema # and live syntax and validation errors highlighted within the text. # # Here is an example query to fetch a list of products: # { products(first: 5, channel: "%(channel_slug)s") { edges { node { id name description } } } } """ % { "channel_slug": settings.DEFAULT_CHANNEL_SLUG } class DemoGraphQLView(GraphQLView): def render_playground(self, request): + pwa_origin = settings.PWA_ORIGINS[0] ctx = { "query": EXAMPLE_QUERY, - "api_url": request.build_absolute_uri(str(API_PATH)), + "api_url": f"https://{pwa_origin}/graphql/", } return render(request, "graphql/playground.html", ctx)
d08e8144b90d3fe89fd449d31bdb655d62f3a749
serfclient/connection.py
serfclient/connection.py
import socket import sys class SerfConnectionError(Exception): pass class SerfConnection(object): """ Manages RPC communication to and from a Serf agent. """ def __init__(self, host='localhost', port=7373): self.host, self.port = host, port self._socket = None def __repr__(self): return "%(class)s<host=%(host)s,port=%(port)s>" % { 'class': self.__class__.__name__, 'host': self.host, 'port': self.port, } def handshake(self): """ Sets up the connection with the Serf agent and does the initial handshake. """ if self._socket: return try: self._socket = socket.create_connection((self.host, self.port)) return True except socket.error: e = sys.exc_info()[1] raise SerfConnectionError(self._error_message(e)) def _error_message(self, exception): return "Error %s connecting %s:%s. %s." % \ (exception.args[0], self.host, self.port, exception.args[1])
import socket import sys class SerfConnectionError(Exception): pass class SerfConnection(object): """ Manages RPC communication to and from a Serf agent. """ def __init__(self, host='localhost', port=7373): self.host, self.port = host, port self._socket = None def __repr__(self): return "%(class)s<host=%(host)s,port=%(port)s>" % { 'class': self.__class__.__name__, 'host': self.host, 'port': self.port, } def handshake(self): """ Sets up the connection with the Serf agent and does the initial handshake. """ if self._socket: return True else: self._socket = self._connect() return True def _connect(self): try: return socket.create_connection((self.host, self.port), 3) except socket.error: e = sys.exc_info()[1] raise SerfConnectionError(self._error_message(e)) def _error_message(self, exception): return "Error %s connecting %s:%s. %s." % \ (exception.args[0], self.host, self.port, exception.args[1])
Move all 'connect' logic into a private method
Move all 'connect' logic into a private method
Python
mit
charleswhchan/serfclient-py,KushalP/serfclient-py
import socket import sys class SerfConnectionError(Exception): pass class SerfConnection(object): """ Manages RPC communication to and from a Serf agent. """ def __init__(self, host='localhost', port=7373): self.host, self.port = host, port self._socket = None def __repr__(self): return "%(class)s<host=%(host)s,port=%(port)s>" % { 'class': self.__class__.__name__, 'host': self.host, 'port': self.port, } def handshake(self): """ Sets up the connection with the Serf agent and does the initial handshake. """ if self._socket: - return + return True + else: + self._socket = self._connect() + return True + + def _connect(self): try: - self._socket = socket.create_connection((self.host, self.port)) + return socket.create_connection((self.host, self.port), 3) - return True except socket.error: e = sys.exc_info()[1] raise SerfConnectionError(self._error_message(e)) def _error_message(self, exception): return "Error %s connecting %s:%s. %s." % \ (exception.args[0], self.host, self.port, exception.args[1])
Move all 'connect' logic into a private method
## Code Before: import socket import sys class SerfConnectionError(Exception): pass class SerfConnection(object): """ Manages RPC communication to and from a Serf agent. """ def __init__(self, host='localhost', port=7373): self.host, self.port = host, port self._socket = None def __repr__(self): return "%(class)s<host=%(host)s,port=%(port)s>" % { 'class': self.__class__.__name__, 'host': self.host, 'port': self.port, } def handshake(self): """ Sets up the connection with the Serf agent and does the initial handshake. """ if self._socket: return try: self._socket = socket.create_connection((self.host, self.port)) return True except socket.error: e = sys.exc_info()[1] raise SerfConnectionError(self._error_message(e)) def _error_message(self, exception): return "Error %s connecting %s:%s. %s." % \ (exception.args[0], self.host, self.port, exception.args[1]) ## Instruction: Move all 'connect' logic into a private method ## Code After: import socket import sys class SerfConnectionError(Exception): pass class SerfConnection(object): """ Manages RPC communication to and from a Serf agent. """ def __init__(self, host='localhost', port=7373): self.host, self.port = host, port self._socket = None def __repr__(self): return "%(class)s<host=%(host)s,port=%(port)s>" % { 'class': self.__class__.__name__, 'host': self.host, 'port': self.port, } def handshake(self): """ Sets up the connection with the Serf agent and does the initial handshake. """ if self._socket: return True else: self._socket = self._connect() return True def _connect(self): try: return socket.create_connection((self.host, self.port), 3) except socket.error: e = sys.exc_info()[1] raise SerfConnectionError(self._error_message(e)) def _error_message(self, exception): return "Error %s connecting %s:%s. %s." % \ (exception.args[0], self.host, self.port, exception.args[1])
import socket import sys class SerfConnectionError(Exception): pass class SerfConnection(object): """ Manages RPC communication to and from a Serf agent. """ def __init__(self, host='localhost', port=7373): self.host, self.port = host, port self._socket = None def __repr__(self): return "%(class)s<host=%(host)s,port=%(port)s>" % { 'class': self.__class__.__name__, 'host': self.host, 'port': self.port, } def handshake(self): """ Sets up the connection with the Serf agent and does the initial handshake. """ if self._socket: - return + return True ? +++++ + else: + self._socket = self._connect() + return True + + def _connect(self): try: - self._socket = socket.create_connection((self.host, self.port)) ? ^^^^^^^^^^ ^^ + return socket.create_connection((self.host, self.port), 3) ? ^ ^^^ +++ - return True except socket.error: e = sys.exc_info()[1] raise SerfConnectionError(self._error_message(e)) def _error_message(self, exception): return "Error %s connecting %s:%s. %s." % \ (exception.args[0], self.host, self.port, exception.args[1])
85769162560d83a58ccc92f818559ddd3dce2a09
pages/index.py
pages/index.py
import web from modules.base import renderer from modules.login import loginInstance from modules.courses import Course #Index page class IndexPage: #Simply display the page def GET(self): if loginInstance.isLoggedIn(): userInput = web.input(); if "logoff" in userInput: loginInstance.disconnect(); return renderer.index(False) else: courses = Course.GetAllCoursesIds() return renderer.main(courses) else: return renderer.index(False) #Try to log in def POST(self): userInput = web.input(); if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password): return renderer.main() else: return renderer.index(True)
import web from modules.base import renderer from modules.login import loginInstance from modules.courses import Course #Index page class IndexPage: #Simply display the page def GET(self): if loginInstance.isLoggedIn(): userInput = web.input(); if "logoff" in userInput: loginInstance.disconnect(); return renderer.index(False) else: return renderer.main(Course.GetAllCoursesIds()) else: return renderer.index(False) #Try to log in def POST(self): userInput = web.input(); if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password): return renderer.main(Course.GetAllCoursesIds()) else: return renderer.index(True)
Fix another bug in the authentication
Fix another bug in the authentication
Python
agpl-3.0
layus/INGInious,GuillaumeDerval/INGInious,GuillaumeDerval/INGInious,layus/INGInious,layus/INGInious,GuillaumeDerval/INGInious,GuillaumeDerval/INGInious,layus/INGInious
import web from modules.base import renderer from modules.login import loginInstance from modules.courses import Course #Index page class IndexPage: #Simply display the page def GET(self): if loginInstance.isLoggedIn(): userInput = web.input(); if "logoff" in userInput: loginInstance.disconnect(); return renderer.index(False) else: - courses = Course.GetAllCoursesIds() - return renderer.main(courses) + return renderer.main(Course.GetAllCoursesIds()) else: return renderer.index(False) #Try to log in def POST(self): userInput = web.input(); if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password): - return renderer.main() + return renderer.main(Course.GetAllCoursesIds()) else: return renderer.index(True)
Fix another bug in the authentication
## Code Before: import web from modules.base import renderer from modules.login import loginInstance from modules.courses import Course #Index page class IndexPage: #Simply display the page def GET(self): if loginInstance.isLoggedIn(): userInput = web.input(); if "logoff" in userInput: loginInstance.disconnect(); return renderer.index(False) else: courses = Course.GetAllCoursesIds() return renderer.main(courses) else: return renderer.index(False) #Try to log in def POST(self): userInput = web.input(); if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password): return renderer.main() else: return renderer.index(True) ## Instruction: Fix another bug in the authentication ## Code After: import web from modules.base import renderer from modules.login import loginInstance from modules.courses import Course #Index page class IndexPage: #Simply display the page def GET(self): if loginInstance.isLoggedIn(): userInput = web.input(); if "logoff" in userInput: loginInstance.disconnect(); return renderer.index(False) else: return renderer.main(Course.GetAllCoursesIds()) else: return renderer.index(False) #Try to log in def POST(self): userInput = web.input(); if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password): return renderer.main(Course.GetAllCoursesIds()) else: return renderer.index(True)
import web from modules.base import renderer from modules.login import loginInstance from modules.courses import Course #Index page class IndexPage: #Simply display the page def GET(self): if loginInstance.isLoggedIn(): userInput = web.input(); if "logoff" in userInput: loginInstance.disconnect(); return renderer.index(False) else: - courses = Course.GetAllCoursesIds() - return renderer.main(courses) ? ^ + return renderer.main(Course.GetAllCoursesIds()) ? ^^^^^^^^^^^^^^ ++++ + else: return renderer.index(False) #Try to log in def POST(self): userInput = web.input(); if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password): - return renderer.main() + return renderer.main(Course.GetAllCoursesIds()) else: return renderer.index(True)
a4e9198194e08b99e11802dd260bd5c203179211
Mollie/API/Object/Customer.py
Mollie/API/Object/Customer.py
from .Base import Base class Customer(Base): @property def id(self): return self.getProperty('id') @property def name(self): return self.getProperty('name') @property def email(self): return self.getProperty('email') @property def locale(self): return self.getProperty('locale') @property def metadata(self): if 'metadata' not in self: return None return self.getProperty('metadata') @property def mode(self): return self.getProperty('mode') @property def resource(self): return self.getProperty('resource') @property def createdAt(self): return self.getProperty('createdAt')
from .Base import Base class Customer(Base): @property def id(self): return self.getProperty('id') @property def name(self): return self.getProperty('name') @property def email(self): return self.getProperty('email') @property def locale(self): return self.getProperty('locale') @property def metadata(self): return self.getProperty('metadata') @property def mode(self): return self.getProperty('mode') @property def resource(self): return self.getProperty('resource') @property def createdAt(self): return self.getProperty('createdAt')
Remove forgotten if in customer.py
Remove forgotten if in customer.py
Python
bsd-2-clause
mollie/mollie-api-python
from .Base import Base class Customer(Base): @property def id(self): return self.getProperty('id') @property def name(self): return self.getProperty('name') @property def email(self): return self.getProperty('email') @property def locale(self): return self.getProperty('locale') @property def metadata(self): - if 'metadata' not in self: - return None return self.getProperty('metadata') @property def mode(self): return self.getProperty('mode') @property def resource(self): return self.getProperty('resource') @property def createdAt(self): return self.getProperty('createdAt')
Remove forgotten if in customer.py
## Code Before: from .Base import Base class Customer(Base): @property def id(self): return self.getProperty('id') @property def name(self): return self.getProperty('name') @property def email(self): return self.getProperty('email') @property def locale(self): return self.getProperty('locale') @property def metadata(self): if 'metadata' not in self: return None return self.getProperty('metadata') @property def mode(self): return self.getProperty('mode') @property def resource(self): return self.getProperty('resource') @property def createdAt(self): return self.getProperty('createdAt') ## Instruction: Remove forgotten if in customer.py ## Code After: from .Base import Base class Customer(Base): @property def id(self): return self.getProperty('id') @property def name(self): return self.getProperty('name') @property def email(self): return self.getProperty('email') @property def locale(self): return self.getProperty('locale') @property def metadata(self): return self.getProperty('metadata') @property def mode(self): return self.getProperty('mode') @property def resource(self): return self.getProperty('resource') @property def createdAt(self): return self.getProperty('createdAt')
from .Base import Base class Customer(Base): @property def id(self): return self.getProperty('id') @property def name(self): return self.getProperty('name') @property def email(self): return self.getProperty('email') @property def locale(self): return self.getProperty('locale') @property def metadata(self): - if 'metadata' not in self: - return None return self.getProperty('metadata') @property def mode(self): return self.getProperty('mode') @property def resource(self): return self.getProperty('resource') @property def createdAt(self): return self.getProperty('createdAt')
c1e1c9d63d5334140aa71c025a90e9500b299307
functional_tests.py
functional_tests.py
from selenium import webdriver browser = webdriver.Firefox() browser.get('http://localhost:8000') assert 'Django' in browser.title
from selenium import webdriver import unittest class NewVisitorTest(unittest.TestCase): def setUp(self): self.browser = webdriver.Firefox() def tearDown(self): self.browser.quit() def test_can_start_a_list_and_retrieve_it_later(self): # Rey has heard about a cool new online to-do app. He goes # to check out its homepage self.browser.get('http://localhost:8000') # He notices the page title and header mention to-do lists self.assertIn('To-Do', self.browser.title) self.fail('Finish the test!') # He is invited to enter a to-do item straight away # He types "Make coffee" into a text box (Rey can't work when he is sleepy) # There is still a text box inviting him to add another item. # He enters "Go to the toilet" (Rey is very methodical) # The page updates again, and now shows both items on his list # Rey wonders whether the site will remember his list. Then he sees that # the site has generated a unique URL for him # He visits that URL - his to-do list is still there # Satisfied, he goes back to sleep if __name__ == '__main__': unittest.main(warnings='ignore')
Update tests to be human-readable
Update tests to be human-readable
Python
apache-2.0
rocity/the-testing-goat,rocity/the-testing-goat
from selenium import webdriver + import unittest + class NewVisitorTest(unittest.TestCase): + def setUp(self): - browser = webdriver.Firefox() + self.browser = webdriver.Firefox() - browser.get('http://localhost:8000') - assert 'Django' in browser.title + def tearDown(self): + self.browser.quit() + + def test_can_start_a_list_and_retrieve_it_later(self): + + # Rey has heard about a cool new online to-do app. He goes + # to check out its homepage + self.browser.get('http://localhost:8000') + + # He notices the page title and header mention to-do lists + self.assertIn('To-Do', self.browser.title) + self.fail('Finish the test!') + + # He is invited to enter a to-do item straight away + + # He types "Make coffee" into a text box (Rey can't work when he is sleepy) + + # There is still a text box inviting him to add another item. + # He enters "Go to the toilet" (Rey is very methodical) + + # The page updates again, and now shows both items on his list + + # Rey wonders whether the site will remember his list. Then he sees that + # the site has generated a unique URL for him + + # He visits that URL - his to-do list is still there + + # Satisfied, he goes back to sleep + + if __name__ == '__main__': + unittest.main(warnings='ignore')
Update tests to be human-readable
## Code Before: from selenium import webdriver browser = webdriver.Firefox() browser.get('http://localhost:8000') assert 'Django' in browser.title ## Instruction: Update tests to be human-readable ## Code After: from selenium import webdriver import unittest class NewVisitorTest(unittest.TestCase): def setUp(self): self.browser = webdriver.Firefox() def tearDown(self): self.browser.quit() def test_can_start_a_list_and_retrieve_it_later(self): # Rey has heard about a cool new online to-do app. He goes # to check out its homepage self.browser.get('http://localhost:8000') # He notices the page title and header mention to-do lists self.assertIn('To-Do', self.browser.title) self.fail('Finish the test!') # He is invited to enter a to-do item straight away # He types "Make coffee" into a text box (Rey can't work when he is sleepy) # There is still a text box inviting him to add another item. # He enters "Go to the toilet" (Rey is very methodical) # The page updates again, and now shows both items on his list # Rey wonders whether the site will remember his list. Then he sees that # the site has generated a unique URL for him # He visits that URL - his to-do list is still there # Satisfied, he goes back to sleep if __name__ == '__main__': unittest.main(warnings='ignore')
from selenium import webdriver + import unittest + class NewVisitorTest(unittest.TestCase): + def setUp(self): - browser = webdriver.Firefox() + self.browser = webdriver.Firefox() ? +++++++++++++ - browser.get('http://localhost:8000') - assert 'Django' in browser.title + def tearDown(self): + self.browser.quit() + + def test_can_start_a_list_and_retrieve_it_later(self): + + # Rey has heard about a cool new online to-do app. He goes + # to check out its homepage + self.browser.get('http://localhost:8000') + + # He notices the page title and header mention to-do lists + self.assertIn('To-Do', self.browser.title) + self.fail('Finish the test!') + + # He is invited to enter a to-do item straight away + + # He types "Make coffee" into a text box (Rey can't work when he is sleepy) + + # There is still a text box inviting him to add another item. + # He enters "Go to the toilet" (Rey is very methodical) + + # The page updates again, and now shows both items on his list + + # Rey wonders whether the site will remember his list. Then he sees that + # the site has generated a unique URL for him + + # He visits that URL - his to-do list is still there + + # Satisfied, he goes back to sleep + + if __name__ == '__main__': + unittest.main(warnings='ignore')
14bd2c0732b5871ac43991a237a8f12a334e982d
sirius/LI_V00/__init__.py
sirius/LI_V00/__init__.py
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = accelerator.create_accelerator # -- default accelerator values for LI_V00 -- energy = _lattice._energy single_bunch_charge = _lattice._single_bunch_charge multi_bunch_charge = _lattice._multi_bunch_charge pulse_duration_interval = _lattice._pulse_duration_interval default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'LI_V00' family_data = _lattice._family_data emittance = _lattice._emittance
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = accelerator.create_accelerator # -- default accelerator values for LI_V00 -- energy = _lattice._energy single_bunch_charge = _lattice._single_bunch_charge multi_bunch_charge = _lattice._multi_bunch_charge pulse_duration_interval = _lattice._pulse_duration_interval default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'LI_V00' family_data = _lattice._family_data emittance = _lattice._emittance global_coupling = 1.0 # "round" beam
Add parameters of initial beam distribution at LI
Add parameters of initial beam distribution at LI
Python
mit
lnls-fac/sirius
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = accelerator.create_accelerator # -- default accelerator values for LI_V00 -- energy = _lattice._energy single_bunch_charge = _lattice._single_bunch_charge multi_bunch_charge = _lattice._multi_bunch_charge pulse_duration_interval = _lattice._pulse_duration_interval default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'LI_V00' family_data = _lattice._family_data - emittance = _lattice._emittance + emittance = _lattice._emittance + global_coupling = 1.0 # "round" beam
Add parameters of initial beam distribution at LI
## Code Before: from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = accelerator.create_accelerator # -- default accelerator values for LI_V00 -- energy = _lattice._energy single_bunch_charge = _lattice._single_bunch_charge multi_bunch_charge = _lattice._multi_bunch_charge pulse_duration_interval = _lattice._pulse_duration_interval default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'LI_V00' family_data = _lattice._family_data emittance = _lattice._emittance ## Instruction: Add parameters of initial beam distribution at LI ## Code After: from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = accelerator.create_accelerator # -- default accelerator values for LI_V00 -- energy = _lattice._energy single_bunch_charge = _lattice._single_bunch_charge multi_bunch_charge = _lattice._multi_bunch_charge pulse_duration_interval = _lattice._pulse_duration_interval default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'LI_V00' family_data = _lattice._family_data emittance = _lattice._emittance global_coupling = 1.0 # "round" beam
from . import lattice as _lattice from . import accelerator as _accelerator from . import record_names create_accelerator = accelerator.create_accelerator # -- default accelerator values for LI_V00 -- energy = _lattice._energy single_bunch_charge = _lattice._single_bunch_charge multi_bunch_charge = _lattice._multi_bunch_charge pulse_duration_interval = _lattice._pulse_duration_interval default_optics_mode = _lattice._default_optics_mode.label lattice_version = 'LI_V00' family_data = _lattice._family_data - emittance = _lattice._emittance + emittance = _lattice._emittance ? +++++++++++ + global_coupling = 1.0 # "round" beam
5edddcc85b0e21bb576b71db63d082c8ace5cf70
examples/boilerplates/samples/google_test.py
examples/boilerplates/samples/google_test.py
''' Google.com testing example ''' from seleniumbase import BaseCase from google_objects import HomePage, ResultsPage class GoogleTests(BaseCase): def test_google_dot_com(self): self.open('http://www.google.com') self.assert_element(HomePage.search_button) self.assert_element(HomePage.feeling_lucky_button) self.update_text(HomePage.search_box, "github\n") self.assert_text("github.com", ResultsPage.search_results) self.assert_element(ResultsPage.google_logo) self.click_link_text("Images") self.assert_element('img[alt="Image result for github"]')
''' Google.com testing example ''' from seleniumbase import BaseCase from google_objects import HomePage, ResultsPage class GoogleTests(BaseCase): def test_google_dot_com(self): self.open('http://www.google.com') self.assert_element(HomePage.search_button) self.assert_element(HomePage.feeling_lucky_button) self.update_text(HomePage.search_box, "github\n") self.assert_text("github.com", ResultsPage.search_results) self.click_link_text("Images") self.assert_element('img[alt="Image result for github"]')
Update Google boilerplate test. (Logo frequently changes)
Update Google boilerplate test. (Logo frequently changes)
Python
mit
seleniumbase/SeleniumBase,mdmintz/seleniumspot,mdmintz/SeleniumBase,mdmintz/seleniumspot,mdmintz/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase
''' Google.com testing example ''' from seleniumbase import BaseCase from google_objects import HomePage, ResultsPage class GoogleTests(BaseCase): def test_google_dot_com(self): self.open('http://www.google.com') self.assert_element(HomePage.search_button) self.assert_element(HomePage.feeling_lucky_button) self.update_text(HomePage.search_box, "github\n") self.assert_text("github.com", ResultsPage.search_results) - self.assert_element(ResultsPage.google_logo) self.click_link_text("Images") self.assert_element('img[alt="Image result for github"]')
Update Google boilerplate test. (Logo frequently changes)
## Code Before: ''' Google.com testing example ''' from seleniumbase import BaseCase from google_objects import HomePage, ResultsPage class GoogleTests(BaseCase): def test_google_dot_com(self): self.open('http://www.google.com') self.assert_element(HomePage.search_button) self.assert_element(HomePage.feeling_lucky_button) self.update_text(HomePage.search_box, "github\n") self.assert_text("github.com", ResultsPage.search_results) self.assert_element(ResultsPage.google_logo) self.click_link_text("Images") self.assert_element('img[alt="Image result for github"]') ## Instruction: Update Google boilerplate test. (Logo frequently changes) ## Code After: ''' Google.com testing example ''' from seleniumbase import BaseCase from google_objects import HomePage, ResultsPage class GoogleTests(BaseCase): def test_google_dot_com(self): self.open('http://www.google.com') self.assert_element(HomePage.search_button) self.assert_element(HomePage.feeling_lucky_button) self.update_text(HomePage.search_box, "github\n") self.assert_text("github.com", ResultsPage.search_results) self.click_link_text("Images") self.assert_element('img[alt="Image result for github"]')
''' Google.com testing example ''' from seleniumbase import BaseCase from google_objects import HomePage, ResultsPage class GoogleTests(BaseCase): def test_google_dot_com(self): self.open('http://www.google.com') self.assert_element(HomePage.search_button) self.assert_element(HomePage.feeling_lucky_button) self.update_text(HomePage.search_box, "github\n") self.assert_text("github.com", ResultsPage.search_results) - self.assert_element(ResultsPage.google_logo) self.click_link_text("Images") self.assert_element('img[alt="Image result for github"]')
3a7428723c66010dec1d246beb63be371428d3fe
qipipe/staging/staging_helpers.py
qipipe/staging/staging_helpers.py
"""Pipeline utility functions.""" import re _SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})' """The subject/session/series regexp pattern.""" def match_session_hierarchy(path): """ Matches the subject, session and series names from the given input path. @param path: the path to match @return: the matching (subject, session, series) tuple, or None if no match """ match = re.search(_SSS_REGEX, path) if match: return match.groups()
"""Pipeline utility functions.""" import re from .staging_error import StagingError _SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})' """The subject/session/series regexp pattern.""" def match_series_hierarchy(path): """ Matches the subject, session and series names from the given input path. @param path: the path to match @return: the matching (subject, session, series) tuple, or None if no match """ match = re.search(_SSS_REGEX, path) if match: return match.groups() else: raise StagingError("The path %s does match the subject/session/series pattern" % path)
Raise error if no match.
Raise error if no match.
Python
bsd-2-clause
ohsu-qin/qipipe
"""Pipeline utility functions.""" import re + from .staging_error import StagingError _SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})' """The subject/session/series regexp pattern.""" - def match_session_hierarchy(path): + def match_series_hierarchy(path): """ Matches the subject, session and series names from the given input path. @param path: the path to match @return: the matching (subject, session, series) tuple, or None if no match """ match = re.search(_SSS_REGEX, path) if match: return match.groups() + else: + raise StagingError("The path %s does match the subject/session/series pattern" % path)
Raise error if no match.
## Code Before: """Pipeline utility functions.""" import re _SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})' """The subject/session/series regexp pattern.""" def match_session_hierarchy(path): """ Matches the subject, session and series names from the given input path. @param path: the path to match @return: the matching (subject, session, series) tuple, or None if no match """ match = re.search(_SSS_REGEX, path) if match: return match.groups() ## Instruction: Raise error if no match. ## Code After: """Pipeline utility functions.""" import re from .staging_error import StagingError _SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})' """The subject/session/series regexp pattern.""" def match_series_hierarchy(path): """ Matches the subject, session and series names from the given input path. @param path: the path to match @return: the matching (subject, session, series) tuple, or None if no match """ match = re.search(_SSS_REGEX, path) if match: return match.groups() else: raise StagingError("The path %s does match the subject/session/series pattern" % path)
"""Pipeline utility functions.""" import re + from .staging_error import StagingError _SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})' """The subject/session/series regexp pattern.""" - def match_session_hierarchy(path): ? ---- + def match_series_hierarchy(path): ? +++ """ Matches the subject, session and series names from the given input path. @param path: the path to match @return: the matching (subject, session, series) tuple, or None if no match """ match = re.search(_SSS_REGEX, path) if match: return match.groups() + else: + raise StagingError("The path %s does match the subject/session/series pattern" % path)
49f332149ae8a9a3b5faf82bc20b46dfaeb0a3ad
indra/sources/ctd/api.py
indra/sources/ctd/api.py
import pandas from .processor import CTDChemicalDiseaseProcessor, \ CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor base_url = 'http://ctdbase.org/reports/' urls = { 'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz', 'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz', 'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz', } processors = { 'chemical_gene': CTDChemicalGeneProcessor, 'chemical_disease': CTDChemicalDiseaseProcessor, 'gene_disease': CTDGeneDiseaseProcessor, } def process_from_web(subset): if subset not in urls: raise ValueError('%s is not a valid CTD subset.') df = pandas.read_csv(urls[subset], sep='\t', comment='#', header=None) return process_dataframe(df) def process_tsv(fname, subset): df = pandas.read_csv(fname, sep='\t', comment='#', header=None) return process_dataframe(df, subset) def process_dataframe(df, subset): if subset not in processors: raise ValueError('%s is not a valid CTD subset.') cp = processors[subset](df) cp.extract_statements() return cp
import pandas from .processor import CTDChemicalDiseaseProcessor, \ CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor base_url = 'http://ctdbase.org/reports/' urls = { 'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz', 'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz', 'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz', } processors = { 'chemical_gene': CTDChemicalGeneProcessor, 'chemical_disease': CTDChemicalDiseaseProcessor, 'gene_disease': CTDGeneDiseaseProcessor, } def process_from_web(subset, url=None): if subset not in urls: raise ValueError('%s is not a valid CTD subset.' % subset) url = url if url else urls[subset] return _process_url_or_file(url, subset) def process_tsv(fname, subset): return _process_url_or_file(fname, subset) def _process_url_or_file(path, subset): df = pandas.read_csv(path, sep='\t', comment='#', header=None, dtype=str, keep_default_na=False) return process_dataframe(df, subset) def process_dataframe(df, subset): if subset not in processors: raise ValueError('%s is not a valid CTD subset.' % subset) cp = processors[subset](df) cp.extract_statements() return cp
Refactor API to have single pandas load
Refactor API to have single pandas load
Python
bsd-2-clause
sorgerlab/indra,bgyori/indra,johnbachman/indra,bgyori/indra,sorgerlab/belpy,sorgerlab/belpy,johnbachman/belpy,bgyori/indra,johnbachman/indra,sorgerlab/belpy,johnbachman/belpy,johnbachman/indra,johnbachman/belpy,sorgerlab/indra,sorgerlab/indra
import pandas from .processor import CTDChemicalDiseaseProcessor, \ CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor base_url = 'http://ctdbase.org/reports/' urls = { 'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz', 'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz', 'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz', } processors = { 'chemical_gene': CTDChemicalGeneProcessor, 'chemical_disease': CTDChemicalDiseaseProcessor, 'gene_disease': CTDGeneDiseaseProcessor, } - def process_from_web(subset): + def process_from_web(subset, url=None): if subset not in urls: - raise ValueError('%s is not a valid CTD subset.') + raise ValueError('%s is not a valid CTD subset.' % subset) + url = url if url else urls[subset] + return _process_url_or_file(url, subset) - df = pandas.read_csv(urls[subset], sep='\t', comment='#', - header=None) - return process_dataframe(df) def process_tsv(fname, subset): + return _process_url_or_file(fname, subset) + + + def _process_url_or_file(path, subset): - df = pandas.read_csv(fname, sep='\t', comment='#', header=None) + df = pandas.read_csv(path, sep='\t', comment='#', + header=None, dtype=str, keep_default_na=False) return process_dataframe(df, subset) def process_dataframe(df, subset): if subset not in processors: - raise ValueError('%s is not a valid CTD subset.') + raise ValueError('%s is not a valid CTD subset.' % subset) cp = processors[subset](df) cp.extract_statements() return cp
Refactor API to have single pandas load
## Code Before: import pandas from .processor import CTDChemicalDiseaseProcessor, \ CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor base_url = 'http://ctdbase.org/reports/' urls = { 'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz', 'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz', 'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz', } processors = { 'chemical_gene': CTDChemicalGeneProcessor, 'chemical_disease': CTDChemicalDiseaseProcessor, 'gene_disease': CTDGeneDiseaseProcessor, } def process_from_web(subset): if subset not in urls: raise ValueError('%s is not a valid CTD subset.') df = pandas.read_csv(urls[subset], sep='\t', comment='#', header=None) return process_dataframe(df) def process_tsv(fname, subset): df = pandas.read_csv(fname, sep='\t', comment='#', header=None) return process_dataframe(df, subset) def process_dataframe(df, subset): if subset not in processors: raise ValueError('%s is not a valid CTD subset.') cp = processors[subset](df) cp.extract_statements() return cp ## Instruction: Refactor API to have single pandas load ## Code After: import pandas from .processor import CTDChemicalDiseaseProcessor, \ CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor base_url = 'http://ctdbase.org/reports/' urls = { 'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz', 'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz', 'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz', } processors = { 'chemical_gene': CTDChemicalGeneProcessor, 'chemical_disease': CTDChemicalDiseaseProcessor, 'gene_disease': CTDGeneDiseaseProcessor, } def process_from_web(subset, url=None): if subset not in urls: raise ValueError('%s is not a valid CTD subset.' % subset) url = url if url else urls[subset] return _process_url_or_file(url, subset) def process_tsv(fname, subset): return _process_url_or_file(fname, subset) def _process_url_or_file(path, subset): df = pandas.read_csv(path, sep='\t', comment='#', header=None, dtype=str, keep_default_na=False) return process_dataframe(df, subset) def process_dataframe(df, subset): if subset not in processors: raise ValueError('%s is not a valid CTD subset.' % subset) cp = processors[subset](df) cp.extract_statements() return cp
import pandas from .processor import CTDChemicalDiseaseProcessor, \ CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor base_url = 'http://ctdbase.org/reports/' urls = { 'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz', 'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz', 'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz', } processors = { 'chemical_gene': CTDChemicalGeneProcessor, 'chemical_disease': CTDChemicalDiseaseProcessor, 'gene_disease': CTDGeneDiseaseProcessor, } - def process_from_web(subset): + def process_from_web(subset, url=None): ? ++++++++++ if subset not in urls: - raise ValueError('%s is not a valid CTD subset.') + raise ValueError('%s is not a valid CTD subset.' % subset) ? +++++++++ + url = url if url else urls[subset] + return _process_url_or_file(url, subset) - df = pandas.read_csv(urls[subset], sep='\t', comment='#', - header=None) - return process_dataframe(df) def process_tsv(fname, subset): + return _process_url_or_file(fname, subset) + + + def _process_url_or_file(path, subset): - df = pandas.read_csv(fname, sep='\t', comment='#', header=None) ? ^^ ^^ ------------- + df = pandas.read_csv(path, sep='\t', comment='#', ? ^ ^^ + header=None, dtype=str, keep_default_na=False) return process_dataframe(df, subset) def process_dataframe(df, subset): if subset not in processors: - raise ValueError('%s is not a valid CTD subset.') + raise ValueError('%s is not a valid CTD subset.' % subset) ? +++++++++ cp = processors[subset](df) cp.extract_statements() return cp
48132de52d573f7f650ab693c1ad0b6007ebfaef
cybox/test/common/vocab_test.py
cybox/test/common/vocab_test.py
import unittest from cybox.common.vocabs import VocabString import cybox.test from cybox.utils import normalize_to_xml class TestVocabString(unittest.TestCase): def test_plain(self): a = VocabString("test_value") self.assertTrue(a.is_plain()) def test_round_trip(self): attr_dict = { 'value': "test_value", 'vocab_name': "test_a", 'vocab_reference': "test_b", 'condition': "test_d", 'apply_condition': "test_0", 'bit_mask': "test_1", 'pattern_type': "test_e", 'regex_syntax': "test_f", 'has_changed': "test_j", 'trend': "test_k", } attr_obj = VocabString.object_from_dict(attr_dict) attr_dict2 = VocabString.dict_from_object(attr_obj) cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type']) if __name__ == "__main__": unittest.main()
import unittest from cybox.common.vocabs import VocabString import cybox.test from cybox.utils import normalize_to_xml class TestVocabString(unittest.TestCase): def test_plain(self): a = VocabString("test_value") self.assertTrue(a.is_plain()) def test_round_trip(self): vocab_dict = { 'value': "test_value", 'vocab_name': "test_a", 'vocab_reference': "test_b", 'condition': "test_d", # Leave out apply_condition since value is not a list. 'bit_mask': "test_1", 'pattern_type': "test_e", 'regex_syntax': "test_f", 'has_changed': "test_j", 'trend': "test_k", } vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) def test_round_trip_list(self): vocab_dict = { 'value': ['Value1', 'Value2', 'Value3'], 'condition': "Equals", 'apply_condition': "ALL", } vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) if __name__ == "__main__": unittest.main()
Clean up controlled vocab tests
Clean up controlled vocab tests
Python
bsd-3-clause
CybOXProject/python-cybox
import unittest from cybox.common.vocabs import VocabString import cybox.test from cybox.utils import normalize_to_xml class TestVocabString(unittest.TestCase): def test_plain(self): a = VocabString("test_value") self.assertTrue(a.is_plain()) def test_round_trip(self): - attr_dict = { + vocab_dict = { 'value': "test_value", 'vocab_name': "test_a", 'vocab_reference': "test_b", 'condition': "test_d", - 'apply_condition': "test_0", + # Leave out apply_condition since value is not a list. 'bit_mask': "test_1", 'pattern_type': "test_e", 'regex_syntax': "test_f", 'has_changed': "test_j", 'trend': "test_k", - } + } + vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) - attr_obj = VocabString.object_from_dict(attr_dict) - attr_dict2 = VocabString.dict_from_object(attr_obj) - cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type']) + cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) + + def test_round_trip_list(self): + vocab_dict = { + 'value': ['Value1', 'Value2', 'Value3'], + 'condition': "Equals", + 'apply_condition': "ALL", + } + + vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) + cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) if __name__ == "__main__": unittest.main()
Clean up controlled vocab tests
## Code Before: import unittest from cybox.common.vocabs import VocabString import cybox.test from cybox.utils import normalize_to_xml class TestVocabString(unittest.TestCase): def test_plain(self): a = VocabString("test_value") self.assertTrue(a.is_plain()) def test_round_trip(self): attr_dict = { 'value': "test_value", 'vocab_name': "test_a", 'vocab_reference': "test_b", 'condition': "test_d", 'apply_condition': "test_0", 'bit_mask': "test_1", 'pattern_type': "test_e", 'regex_syntax': "test_f", 'has_changed': "test_j", 'trend': "test_k", } attr_obj = VocabString.object_from_dict(attr_dict) attr_dict2 = VocabString.dict_from_object(attr_obj) cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type']) if __name__ == "__main__": unittest.main() ## Instruction: Clean up controlled vocab tests ## Code After: import unittest from cybox.common.vocabs import VocabString import cybox.test from cybox.utils import normalize_to_xml class TestVocabString(unittest.TestCase): def test_plain(self): a = VocabString("test_value") self.assertTrue(a.is_plain()) def test_round_trip(self): vocab_dict = { 'value': "test_value", 'vocab_name': "test_a", 'vocab_reference': "test_b", 'condition': "test_d", # Leave out apply_condition since value is not a list. 'bit_mask': "test_1", 'pattern_type': "test_e", 'regex_syntax': "test_f", 'has_changed': "test_j", 'trend': "test_k", } vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) def test_round_trip_list(self): vocab_dict = { 'value': ['Value1', 'Value2', 'Value3'], 'condition': "Equals", 'apply_condition': "ALL", } vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) if __name__ == "__main__": unittest.main()
import unittest from cybox.common.vocabs import VocabString import cybox.test from cybox.utils import normalize_to_xml class TestVocabString(unittest.TestCase): def test_plain(self): a = VocabString("test_value") self.assertTrue(a.is_plain()) def test_round_trip(self): - attr_dict = { ? ^^^ + vocab_dict = { ? +++ ^ 'value': "test_value", 'vocab_name': "test_a", 'vocab_reference': "test_b", 'condition': "test_d", - 'apply_condition': "test_0", + # Leave out apply_condition since value is not a list. 'bit_mask': "test_1", 'pattern_type': "test_e", 'regex_syntax': "test_f", 'has_changed': "test_j", 'trend': "test_k", - } + } ? + + vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) - attr_obj = VocabString.object_from_dict(attr_dict) - attr_dict2 = VocabString.dict_from_object(attr_obj) - cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type']) ? ^^^ ^^^ + cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) ? +++ ^ +++ ^ + + def test_round_trip_list(self): + vocab_dict = { + 'value': ['Value1', 'Value2', 'Value3'], + 'condition': "Equals", + 'apply_condition': "ALL", + } + + vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict) + cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type']) if __name__ == "__main__": unittest.main()
5c41066e9c93c417253cbde325a18079c1c69d1a
scipy/sparse/linalg/isolve/__init__.py
scipy/sparse/linalg/isolve/__init__.py
"Iterative Solvers for Sparse Linear Systems" #from info import __doc__ from iterative import * from minres import minres from lgmres import lgmres from lsqr import lsqr __all__ = filter(lambda s:not s.startswith('_'),dir()) from numpy.testing import Tester test = Tester().test bench = Tester().bench
"Iterative Solvers for Sparse Linear Systems" #from info import __doc__ from iterative import * from minres import minres from lgmres import lgmres from lsqr import lsqr from lsmr import lsmr __all__ = filter(lambda s:not s.startswith('_'),dir()) from numpy.testing import Tester test = Tester().test bench = Tester().bench
Add lsmr to isolve module.
ENH: Add lsmr to isolve module.
Python
bsd-3-clause
bkendzior/scipy,mikebenfield/scipy,anielsen001/scipy,nonhermitian/scipy,aeklant/scipy,apbard/scipy,jjhelmus/scipy,giorgiop/scipy,sonnyhu/scipy,befelix/scipy,jor-/scipy,gfyoung/scipy,piyush0609/scipy,pbrod/scipy,Shaswat27/scipy,pschella/scipy,chatcannon/scipy,Newman101/scipy,ilayn/scipy,Newman101/scipy,mtrbean/scipy,Shaswat27/scipy,newemailjdm/scipy,behzadnouri/scipy,woodscn/scipy,sonnyhu/scipy,jonycgn/scipy,behzadnouri/scipy,lukauskas/scipy,rgommers/scipy,efiring/scipy,newemailjdm/scipy,Eric89GXL/scipy,WillieMaddox/scipy,andim/scipy,ogrisel/scipy,ogrisel/scipy,FRidh/scipy,anielsen001/scipy,pnedunuri/scipy,mortonjt/scipy,gdooper/scipy,haudren/scipy,gertingold/scipy,lhilt/scipy,Srisai85/scipy,zerothi/scipy,pizzathief/scipy,futurulus/scipy,rgommers/scipy,mortonjt/scipy,piyush0609/scipy,piyush0609/scipy,piyush0609/scipy,Eric89GXL/scipy,matthewalbani/scipy,ilayn/scipy,fredrikw/scipy,pyramania/scipy,Kamp9/scipy,Gillu13/scipy,jonycgn/scipy,chatcannon/scipy,nvoron23/scipy,fernand/scipy,witcxc/scipy,Newman101/scipy,efiring/scipy,Newman101/scipy,argriffing/scipy,sargas/scipy,gertingold/scipy,raoulbq/scipy,minhlongdo/scipy,ChanderG/scipy,jjhelmus/scipy,maniteja123/scipy,felipebetancur/scipy,Gillu13/scipy,anntzer/scipy,maciejkula/scipy,jakevdp/scipy,Kamp9/scipy,trankmichael/scipy,woodscn/scipy,jjhelmus/scipy,zerothi/scipy,matthewalbani/scipy,josephcslater/scipy,apbard/scipy,mtrbean/scipy,richardotis/scipy,mortonjt/scipy,mortada/scipy,teoliphant/scipy,jonycgn/scipy,ilayn/scipy,ndchorley/scipy,Srisai85/scipy,zerothi/scipy,sonnyhu/scipy,anntzer/scipy,vigna/scipy,gef756/scipy,cpaulik/scipy,WarrenWeckesser/scipy,ndchorley/scipy,pnedunuri/scipy,jamestwebber/scipy,anielsen001/scipy,FRidh/scipy,vanpact/scipy,Newman101/scipy,pbrod/scipy,raoulbq/scipy,vigna/scipy,jakevdp/scipy,jsilter/scipy,sauliusl/scipy,e-q/scipy,aarchiba/scipy,piyush0609/scipy,cpaulik/scipy,gdooper/scipy,andim/scipy,teoliphant/scipy,hainm/scipy,behzadnouri/scipy,Stefan-Endres/scipy,petebachant/scipy,Shaswat27/scipy,gef756/scipy,jakevdp/scipy,kalvdans/scipy,ales-erjavec/scipy,kleskjr/scipy,mdhaber/scipy,mgaitan/scipy,perimosocordiae/scipy,pizzathief/scipy,aeklant/scipy,niknow/scipy,dch312/scipy,WarrenWeckesser/scipy,jjhelmus/scipy,mhogg/scipy,sargas/scipy,dominicelse/scipy,sriki18/scipy,jor-/scipy,petebachant/scipy,woodscn/scipy,vanpact/scipy,nonhermitian/scipy,futurulus/scipy,dominicelse/scipy,scipy/scipy,mdhaber/scipy,kalvdans/scipy,jsilter/scipy,ortylp/scipy,perimosocordiae/scipy,gfyoung/scipy,fernand/scipy,anntzer/scipy,aman-iitj/scipy,sauliusl/scipy,mikebenfield/scipy,matthew-brett/scipy,mgaitan/scipy,lukauskas/scipy,rmcgibbo/scipy,haudren/scipy,surhudm/scipy,sriki18/scipy,jonycgn/scipy,person142/scipy,befelix/scipy,jor-/scipy,efiring/scipy,vberaudi/scipy,pyramania/scipy,josephcslater/scipy,grlee77/scipy,zerothi/scipy,jsilter/scipy,larsmans/scipy,efiring/scipy,ales-erjavec/scipy,nvoron23/scipy,felipebetancur/scipy,jseabold/scipy,dch312/scipy,Gillu13/scipy,mortada/scipy,gef756/scipy,trankmichael/scipy,ortylp/scipy,Gillu13/scipy,ogrisel/scipy,nmayorov/scipy,e-q/scipy,witcxc/scipy,jseabold/scipy,vhaasteren/scipy,maniteja123/scipy,vigna/scipy,gfyoung/scipy,lhilt/scipy,giorgiop/scipy,grlee77/scipy,haudren/scipy,Eric89GXL/scipy,jamestwebber/scipy,raoulbq/scipy,zxsted/scipy,andim/scipy,mortada/scipy,befelix/scipy,richardotis/scipy,tylerjereddy/scipy,larsmans/scipy,jonycgn/scipy,niknow/scipy,jakevdp/scipy,nonhermitian/scipy,aman-iitj/scipy,teoliphant/scipy,lhilt/scipy,mortonjt/scipy,vanpact/scipy,felipebetancur/scipy,person142/scipy,woodscn/scipy,sriki18/scipy,Srisai85/scipy,FRidh/scipy,mhogg/scipy,behzadnouri/scipy,njwilson23/scipy,mingwpy/scipy,ilayn/scipy,ChanderG/scipy,Kamp9/scipy,minhlongdo/scipy,tylerjereddy/scipy,kleskjr/scipy,jseabold/scipy,aarchiba/scipy,josephcslater/scipy,andyfaff/scipy,dch312/scipy,mingwpy/scipy,raoulbq/scipy,Dapid/scipy,endolith/scipy,scipy/scipy,Srisai85/scipy,zxsted/scipy,fernand/scipy,pbrod/scipy,Stefan-Endres/scipy,rmcgibbo/scipy,zaxliu/scipy,rgommers/scipy,pschella/scipy,Eric89GXL/scipy,bkendzior/scipy,endolith/scipy,hainm/scipy,jonycgn/scipy,witcxc/scipy,zxsted/scipy,vberaudi/scipy,perimosocordiae/scipy,juliantaylor/scipy,grlee77/scipy,sriki18/scipy,Shaswat27/scipy,sonnyhu/scipy,bkendzior/scipy,nmayorov/scipy,Dapid/scipy,ales-erjavec/scipy,WillieMaddox/scipy,arokem/scipy,kleskjr/scipy,newemailjdm/scipy,mgaitan/scipy,aman-iitj/scipy,lukauskas/scipy,maciejkula/scipy,Dapid/scipy,niknow/scipy,ortylp/scipy,giorgiop/scipy,newemailjdm/scipy,WarrenWeckesser/scipy,lukauskas/scipy,mortada/scipy,jamestwebber/scipy,WillieMaddox/scipy,juliantaylor/scipy,juliantaylor/scipy,newemailjdm/scipy,nvoron23/scipy,FRidh/scipy,ilayn/scipy,Dapid/scipy,e-q/scipy,niknow/scipy,WarrenWeckesser/scipy,ndchorley/scipy,mgaitan/scipy,pizzathief/scipy,zaxliu/scipy,richardotis/scipy,pizzathief/scipy,gfyoung/scipy,matthew-brett/scipy,pbrod/scipy,nmayorov/scipy,mikebenfield/scipy,sauliusl/scipy,aarchiba/scipy,gertingold/scipy,zxsted/scipy,sriki18/scipy,zerothi/scipy,jamestwebber/scipy,surhudm/scipy,giorgiop/scipy,matthew-brett/scipy,scipy/scipy,lukauskas/scipy,pschella/scipy,felipebetancur/scipy,perimosocordiae/scipy,maniteja123/scipy,haudren/scipy,futurulus/scipy,ogrisel/scipy,surhudm/scipy,woodscn/scipy,jor-/scipy,niknow/scipy,perimosocordiae/scipy,mhogg/scipy,rmcgibbo/scipy,ales-erjavec/scipy,grlee77/scipy,mdhaber/scipy,fredrikw/scipy,pschella/scipy,Gillu13/scipy,vhaasteren/scipy,mtrbean/scipy,aeklant/scipy,dominicelse/scipy,giorgiop/scipy,vhaasteren/scipy,ales-erjavec/scipy,nvoron23/scipy,kalvdans/scipy,jakevdp/scipy,dominicelse/scipy,vberaudi/scipy,pizzathief/scipy,ChanderG/scipy,matthewalbani/scipy,fernand/scipy,gef756/scipy,ChanderG/scipy,FRidh/scipy,maniteja123/scipy,grlee77/scipy,nonhermitian/scipy,hainm/scipy,felipebetancur/scipy,zaxliu/scipy,vanpact/scipy,Newman101/scipy,cpaulik/scipy,ndchorley/scipy,nmayorov/scipy,gdooper/scipy,trankmichael/scipy,FRidh/scipy,richardotis/scipy,Gillu13/scipy,anntzer/scipy,lhilt/scipy,vigna/scipy,dch312/scipy,newemailjdm/scipy,maciejkula/scipy,efiring/scipy,tylerjereddy/scipy,gdooper/scipy,pyramania/scipy,mtrbean/scipy,gertingold/scipy,nonhermitian/scipy,apbard/scipy,bkendzior/scipy,giorgiop/scipy,Kamp9/scipy,piyush0609/scipy,WarrenWeckesser/scipy,fernand/scipy,ndchorley/scipy,Shaswat27/scipy,cpaulik/scipy,jor-/scipy,matthewalbani/scipy,pbrod/scipy,dch312/scipy,Kamp9/scipy,fredrikw/scipy,Shaswat27/scipy,cpaulik/scipy,larsmans/scipy,ortylp/scipy,mortonjt/scipy,njwilson23/scipy,ChanderG/scipy,mingwpy/scipy,hainm/scipy,scipy/scipy,vhaasteren/scipy,Dapid/scipy,haudren/scipy,Stefan-Endres/scipy,zxsted/scipy,sriki18/scipy,sauliusl/scipy,endolith/scipy,mingwpy/scipy,aeklant/scipy,argriffing/scipy,pnedunuri/scipy,futurulus/scipy,matthew-brett/scipy,fredrikw/scipy,juliantaylor/scipy,futurulus/scipy,teoliphant/scipy,scipy/scipy,petebachant/scipy,Eric89GXL/scipy,mortada/scipy,argriffing/scipy,behzadnouri/scipy,kalvdans/scipy,nmayorov/scipy,mtrbean/scipy,Stefan-Endres/scipy,befelix/scipy,hainm/scipy,surhudm/scipy,sargas/scipy,anielsen001/scipy,rmcgibbo/scipy,petebachant/scipy,mikebenfield/scipy,raoulbq/scipy,vhaasteren/scipy,person142/scipy,efiring/scipy,richardotis/scipy,jsilter/scipy,zaxliu/scipy,kleskjr/scipy,gertingold/scipy,petebachant/scipy,aarchiba/scipy,ales-erjavec/scipy,Srisai85/scipy,maciejkula/scipy,juliantaylor/scipy,arokem/scipy,zerothi/scipy,pyramania/scipy,endolith/scipy,e-q/scipy,rgommers/scipy,chatcannon/scipy,aman-iitj/scipy,nvoron23/scipy,pyramania/scipy,trankmichael/scipy,mortonjt/scipy,Stefan-Endres/scipy,chatcannon/scipy,trankmichael/scipy,apbard/scipy,ortylp/scipy,anntzer/scipy,jamestwebber/scipy,njwilson23/scipy,pnedunuri/scipy,argriffing/scipy,mingwpy/scipy,andim/scipy,Srisai85/scipy,zaxliu/scipy,person142/scipy,mhogg/scipy,aman-iitj/scipy,WillieMaddox/scipy,maniteja123/scipy,endolith/scipy,bkendzior/scipy,larsmans/scipy,kleskjr/scipy,sauliusl/scipy,Kamp9/scipy,cpaulik/scipy,richardotis/scipy,jseabold/scipy,dominicelse/scipy,surhudm/scipy,tylerjereddy/scipy,mgaitan/scipy,befelix/scipy,e-q/scipy,Dapid/scipy,ChanderG/scipy,perimosocordiae/scipy,witcxc/scipy,argriffing/scipy,Eric89GXL/scipy,rgommers/scipy,apbard/scipy,andyfaff/scipy,hainm/scipy,vhaasteren/scipy,woodscn/scipy,fernand/scipy,mgaitan/scipy,argriffing/scipy,tylerjereddy/scipy,andyfaff/scipy,arokem/scipy,behzadnouri/scipy,minhlongdo/scipy,felipebetancur/scipy,fredrikw/scipy,rmcgibbo/scipy,ndchorley/scipy,andyfaff/scipy,Stefan-Endres/scipy,vanpact/scipy,gef756/scipy,chatcannon/scipy,anielsen001/scipy,sauliusl/scipy,gfyoung/scipy,raoulbq/scipy,aman-iitj/scipy,mortada/scipy,zxsted/scipy,zaxliu/scipy,kalvdans/scipy,pnedunuri/scipy,minhlongdo/scipy,arokem/scipy,aeklant/scipy,sonnyhu/scipy,fredrikw/scipy,mingwpy/scipy,witcxc/scipy,njwilson23/scipy,minhlongdo/scipy,matthew-brett/scipy,larsmans/scipy,arokem/scipy,vberaudi/scipy,pnedunuri/scipy,andyfaff/scipy,mdhaber/scipy,gdooper/scipy,vigna/scipy,anntzer/scipy,andim/scipy,vanpact/scipy,pschella/scipy,matthewalbani/scipy,sargas/scipy,mdhaber/scipy,sonnyhu/scipy,njwilson23/scipy,endolith/scipy,maciejkula/scipy,jseabold/scipy,ortylp/scipy,sargas/scipy,teoliphant/scipy,njwilson23/scipy,josephcslater/scipy,gef756/scipy,andim/scipy,scipy/scipy,andyfaff/scipy,mikebenfield/scipy,WarrenWeckesser/scipy,pbrod/scipy,jsilter/scipy,niknow/scipy,jjhelmus/scipy,vberaudi/scipy,trankmichael/scipy,haudren/scipy,mhogg/scipy,lukauskas/scipy,nvoron23/scipy,mdhaber/scipy,mtrbean/scipy,larsmans/scipy,ogrisel/scipy,ilayn/scipy,mhogg/scipy,person142/scipy,rmcgibbo/scipy,lhilt/scipy,josephcslater/scipy,maniteja123/scipy,petebachant/scipy,anielsen001/scipy,vberaudi/scipy,jseabold/scipy,WillieMaddox/scipy,aarchiba/scipy,WillieMaddox/scipy,surhudm/scipy,chatcannon/scipy,minhlongdo/scipy,kleskjr/scipy,futurulus/scipy
"Iterative Solvers for Sparse Linear Systems" #from info import __doc__ from iterative import * from minres import minres from lgmres import lgmres from lsqr import lsqr + from lsmr import lsmr __all__ = filter(lambda s:not s.startswith('_'),dir()) from numpy.testing import Tester test = Tester().test bench = Tester().bench
Add lsmr to isolve module.
## Code Before: "Iterative Solvers for Sparse Linear Systems" #from info import __doc__ from iterative import * from minres import minres from lgmres import lgmres from lsqr import lsqr __all__ = filter(lambda s:not s.startswith('_'),dir()) from numpy.testing import Tester test = Tester().test bench = Tester().bench ## Instruction: Add lsmr to isolve module. ## Code After: "Iterative Solvers for Sparse Linear Systems" #from info import __doc__ from iterative import * from minres import minres from lgmres import lgmres from lsqr import lsqr from lsmr import lsmr __all__ = filter(lambda s:not s.startswith('_'),dir()) from numpy.testing import Tester test = Tester().test bench = Tester().bench
"Iterative Solvers for Sparse Linear Systems" #from info import __doc__ from iterative import * from minres import minres from lgmres import lgmres from lsqr import lsqr + from lsmr import lsmr __all__ = filter(lambda s:not s.startswith('_'),dir()) from numpy.testing import Tester test = Tester().test bench = Tester().bench
0f2950fcb44efc9b629242743574af503e8230d4
tip/algorithms/sorting/mergesort.py
tip/algorithms/sorting/mergesort.py
def merge(a, b): if len(a) * len(b) == 0: return a + b v = (a[0] < b[0] and a or b).pop(0) return [v] + merge(a, b) def mergesort(list): if len(list) < 2: return list m = len(list) / 2 return merge(mergesort(list[:m]), mergesort(list[m:]))
def merge(a, b): if len(a) * len(b) == 0: return a + b v = (a[0] < b[0] and a or b).pop(0) return [v] + merge(a, b) def mergesort(list): if len(list) < 2: return list m = len(list) / 2 return merge(mergesort(list[:int(m)]), mergesort(list[int(m):]))
Fix slices for Python 3
Fix slices for Python 3
Python
unlicense
davidgasquez/tip
def merge(a, b): if len(a) * len(b) == 0: return a + b v = (a[0] < b[0] and a or b).pop(0) return [v] + merge(a, b) def mergesort(list): if len(list) < 2: return list m = len(list) / 2 - return merge(mergesort(list[:m]), mergesort(list[m:])) + return merge(mergesort(list[:int(m)]), mergesort(list[int(m):]))
Fix slices for Python 3
## Code Before: def merge(a, b): if len(a) * len(b) == 0: return a + b v = (a[0] < b[0] and a or b).pop(0) return [v] + merge(a, b) def mergesort(list): if len(list) < 2: return list m = len(list) / 2 return merge(mergesort(list[:m]), mergesort(list[m:])) ## Instruction: Fix slices for Python 3 ## Code After: def merge(a, b): if len(a) * len(b) == 0: return a + b v = (a[0] < b[0] and a or b).pop(0) return [v] + merge(a, b) def mergesort(list): if len(list) < 2: return list m = len(list) / 2 return merge(mergesort(list[:int(m)]), mergesort(list[int(m):]))
def merge(a, b): if len(a) * len(b) == 0: return a + b v = (a[0] < b[0] and a or b).pop(0) return [v] + merge(a, b) def mergesort(list): if len(list) < 2: return list m = len(list) / 2 - return merge(mergesort(list[:m]), mergesort(list[m:])) + return merge(mergesort(list[:int(m)]), mergesort(list[int(m):])) ? ++++ + ++++ +
945aba9548b92f57fc25f9996bfa9c3811e64deb
server/resources.py
server/resources.py
from flask import request from flask_restful import Resource, Api, abort, reqparse from .models import db, Comment, Lecture api = Api() class CommentListResource(Resource): def get(self, lecture_id): db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not db_lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) db_comments = Comment.query.filter(Comment.lecture_id == lecture_id) comments = [ {'id': c.id, 'content': c.content} for c in db_comments ] return { 'comments': comments } def post(self, lecture_id): db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all() if not db_lectures: abort(404, message="Lecture {} does not exist".format(lecture_id)) parser = reqparse.RequestParser() parser.add_argument('data', help='Text content of comment') args = parser.parse_args() if not args.data: abort(400, message="Comment has no data parameter") content = args.data lecture = db_lectures[0] comment = Comment(content, lecture) db.session.add(comment) db.session.commit() return { 'id': comment.id } api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
from flask import request from flask_restful import Resource, Api, abort, reqparse from .models import db, Comment, Lecture api = Api() class CommentListResource(Resource): def get(self, lecture_id): db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not db_lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) db_comments = Comment.query.filter(Comment.lecture_id == lecture_id) comments = [ {'id': c.id, 'content': c.content} for c in db_comments ] return { 'comments': comments } def post(self, lecture_id): lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) parser = reqparse.RequestParser() parser.add_argument('data', help='Text content of comment') args = parser.parse_args() if not args.data: abort(400, message="Comment has no data parameter") content = args.data comment = Comment(content, lecture) db.session.add(comment) db.session.commit() return { 'id': comment.id } api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
Change single Lecture query to use first() in stead of all()
Change single Lecture query to use first() in stead of all()
Python
mit
MACSIFS/IFS,MACSIFS/IFS,MACSIFS/IFS,MACSIFS/IFS
from flask import request from flask_restful import Resource, Api, abort, reqparse from .models import db, Comment, Lecture api = Api() class CommentListResource(Resource): def get(self, lecture_id): db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not db_lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) db_comments = Comment.query.filter(Comment.lecture_id == lecture_id) comments = [ {'id': c.id, 'content': c.content} for c in db_comments ] return { 'comments': comments } def post(self, lecture_id): - db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all() + lecture = Lecture.query.filter(Lecture.id == lecture_id).first() - if not db_lectures: + if not lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) parser = reqparse.RequestParser() parser.add_argument('data', help='Text content of comment') args = parser.parse_args() if not args.data: abort(400, message="Comment has no data parameter") content = args.data - lecture = db_lectures[0] comment = Comment(content, lecture) db.session.add(comment) db.session.commit() return { 'id': comment.id } api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
Change single Lecture query to use first() in stead of all()
## Code Before: from flask import request from flask_restful import Resource, Api, abort, reqparse from .models import db, Comment, Lecture api = Api() class CommentListResource(Resource): def get(self, lecture_id): db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not db_lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) db_comments = Comment.query.filter(Comment.lecture_id == lecture_id) comments = [ {'id': c.id, 'content': c.content} for c in db_comments ] return { 'comments': comments } def post(self, lecture_id): db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all() if not db_lectures: abort(404, message="Lecture {} does not exist".format(lecture_id)) parser = reqparse.RequestParser() parser.add_argument('data', help='Text content of comment') args = parser.parse_args() if not args.data: abort(400, message="Comment has no data parameter") content = args.data lecture = db_lectures[0] comment = Comment(content, lecture) db.session.add(comment) db.session.commit() return { 'id': comment.id } api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments') ## Instruction: Change single Lecture query to use first() in stead of all() ## Code After: from flask import request from flask_restful import Resource, Api, abort, reqparse from .models import db, Comment, Lecture api = Api() class CommentListResource(Resource): def get(self, lecture_id): db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not db_lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) db_comments = Comment.query.filter(Comment.lecture_id == lecture_id) comments = [ {'id': c.id, 'content': c.content} for c in db_comments ] return { 'comments': comments } def post(self, lecture_id): lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) parser = reqparse.RequestParser() parser.add_argument('data', help='Text content of comment') args = parser.parse_args() if not args.data: abort(400, message="Comment has no data parameter") content = args.data comment = Comment(content, lecture) db.session.add(comment) db.session.commit() return { 'id': comment.id } api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
from flask import request from flask_restful import Resource, Api, abort, reqparse from .models import db, Comment, Lecture api = Api() class CommentListResource(Resource): def get(self, lecture_id): db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first() if not db_lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) db_comments = Comment.query.filter(Comment.lecture_id == lecture_id) comments = [ {'id': c.id, 'content': c.content} for c in db_comments ] return { 'comments': comments } def post(self, lecture_id): - db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all() ? --- - ^^^ + lecture = Lecture.query.filter(Lecture.id == lecture_id).first() ? ^^^^^ - if not db_lectures: ? --- - + if not lecture: abort(404, message="Lecture {} does not exist".format(lecture_id)) parser = reqparse.RequestParser() parser.add_argument('data', help='Text content of comment') args = parser.parse_args() if not args.data: abort(400, message="Comment has no data parameter") content = args.data - lecture = db_lectures[0] comment = Comment(content, lecture) db.session.add(comment) db.session.commit() return { 'id': comment.id } api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
58e2059c37d7464e7ab7a1681ea8f465e9378940
ukpostcode/__init__.py
ukpostcode/__init__.py
from __future__ import unicode_literals __version__ = '0.1.0' def formater(postcode): ''' Format a UK Post Code to the official form. Expects a valid postcode. ''' postcode = postcode.upper().replace(' ', '').strip() inward = postcode[-3:] outward = postcode[:-3] return '%s %s' % (outward, inward)
from __future__ import unicode_literals __version__ = '0.1.0' def formater(postcode): ''' Format a UK Post Code to the official form. Expects a valid postcode. ''' postcode = postcode.upper().replace(' ', '').strip() inward = postcode[-3:] outward = postcode[:-3] return '%s %s' % (outward, inward) def validate(postcode): ''' Returns True if the postcode is valid. False otherwise. ''' raise NotImplementedError()
Prepare to code the validator
Prepare to code the validator
Python
apache-2.0
alanjds/pyukpostcode
from __future__ import unicode_literals __version__ = '0.1.0' def formater(postcode): ''' Format a UK Post Code to the official form. Expects a valid postcode. ''' postcode = postcode.upper().replace(' ', '').strip() inward = postcode[-3:] outward = postcode[:-3] return '%s %s' % (outward, inward) + + def validate(postcode): + ''' + Returns True if the postcode is valid. False otherwise. + ''' + raise NotImplementedError() +
Prepare to code the validator
## Code Before: from __future__ import unicode_literals __version__ = '0.1.0' def formater(postcode): ''' Format a UK Post Code to the official form. Expects a valid postcode. ''' postcode = postcode.upper().replace(' ', '').strip() inward = postcode[-3:] outward = postcode[:-3] return '%s %s' % (outward, inward) ## Instruction: Prepare to code the validator ## Code After: from __future__ import unicode_literals __version__ = '0.1.0' def formater(postcode): ''' Format a UK Post Code to the official form. Expects a valid postcode. ''' postcode = postcode.upper().replace(' ', '').strip() inward = postcode[-3:] outward = postcode[:-3] return '%s %s' % (outward, inward) def validate(postcode): ''' Returns True if the postcode is valid. False otherwise. ''' raise NotImplementedError()
from __future__ import unicode_literals __version__ = '0.1.0' def formater(postcode): ''' Format a UK Post Code to the official form. Expects a valid postcode. ''' postcode = postcode.upper().replace(' ', '').strip() inward = postcode[-3:] outward = postcode[:-3] return '%s %s' % (outward, inward) + + + def validate(postcode): + ''' + Returns True if the postcode is valid. False otherwise. + ''' + raise NotImplementedError()
2c02816c05f3863ef76b3a412ac5bad9eecfafdd
testrepository/tests/test_setup.py
testrepository/tests/test_setup.py
"""Tests for setup.py.""" import doctest import os import subprocess import sys from testtools import ( TestCase, ) from testtools.matchers import ( DocTestMatches, ) class TestCanSetup(TestCase): def test_bdist(self): # Single smoke test to make sure we can build a package. path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py') proc = subprocess.Popen([sys.executable, path, 'bdist'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) output, _ = proc.communicate() self.assertEqual(0, proc.returncode) self.assertThat(output, DocTestMatches("...running bdist...", doctest.ELLIPSIS))
"""Tests for setup.py.""" import doctest import os import subprocess import sys from testtools import ( TestCase, ) from testtools.matchers import ( DocTestMatches, ) class TestCanSetup(TestCase): def test_bdist(self): # Single smoke test to make sure we can build a package. path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py') proc = subprocess.Popen([sys.executable, path, 'bdist'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, universal_newlines=True) output, _ = proc.communicate() self.assertEqual(0, proc.returncode) self.assertThat(output, DocTestMatches("""... running install_scripts ... adding '...testr' ...""", doctest.ELLIPSIS))
Make setup.py smoke test more specific again as requested in review
Make setup.py smoke test more specific again as requested in review
Python
apache-2.0
masayukig/stestr,masayukig/stestr,mtreinish/stestr,mtreinish/stestr
"""Tests for setup.py.""" import doctest import os import subprocess import sys from testtools import ( TestCase, ) from testtools.matchers import ( DocTestMatches, ) class TestCanSetup(TestCase): def test_bdist(self): # Single smoke test to make sure we can build a package. path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py') proc = subprocess.Popen([sys.executable, path, 'bdist'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, - stderr=subprocess.STDOUT) + stderr=subprocess.STDOUT, universal_newlines=True) output, _ = proc.communicate() self.assertEqual(0, proc.returncode) - self.assertThat(output, - DocTestMatches("...running bdist...", doctest.ELLIPSIS)) + self.assertThat(output, DocTestMatches("""... + running install_scripts + ... + adding '...testr' + ...""", doctest.ELLIPSIS))
Make setup.py smoke test more specific again as requested in review
## Code Before: """Tests for setup.py.""" import doctest import os import subprocess import sys from testtools import ( TestCase, ) from testtools.matchers import ( DocTestMatches, ) class TestCanSetup(TestCase): def test_bdist(self): # Single smoke test to make sure we can build a package. path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py') proc = subprocess.Popen([sys.executable, path, 'bdist'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) output, _ = proc.communicate() self.assertEqual(0, proc.returncode) self.assertThat(output, DocTestMatches("...running bdist...", doctest.ELLIPSIS)) ## Instruction: Make setup.py smoke test more specific again as requested in review ## Code After: """Tests for setup.py.""" import doctest import os import subprocess import sys from testtools import ( TestCase, ) from testtools.matchers import ( DocTestMatches, ) class TestCanSetup(TestCase): def test_bdist(self): # Single smoke test to make sure we can build a package. path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py') proc = subprocess.Popen([sys.executable, path, 'bdist'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, universal_newlines=True) output, _ = proc.communicate() self.assertEqual(0, proc.returncode) self.assertThat(output, DocTestMatches("""... running install_scripts ... adding '...testr' ...""", doctest.ELLIPSIS))
"""Tests for setup.py.""" import doctest import os import subprocess import sys from testtools import ( TestCase, ) from testtools.matchers import ( DocTestMatches, ) class TestCanSetup(TestCase): def test_bdist(self): # Single smoke test to make sure we can build a package. path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py') proc = subprocess.Popen([sys.executable, path, 'bdist'], stdin=subprocess.PIPE, stdout=subprocess.PIPE, - stderr=subprocess.STDOUT) + stderr=subprocess.STDOUT, universal_newlines=True) output, _ = proc.communicate() self.assertEqual(0, proc.returncode) - self.assertThat(output, - DocTestMatches("...running bdist...", doctest.ELLIPSIS)) + self.assertThat(output, DocTestMatches("""... + running install_scripts + ... + adding '...testr' + ...""", doctest.ELLIPSIS))
87707340ac82f852937dae546380b5d5327f5bc7
txlege84/core/views.py
txlege84/core/views.py
from django.views.generic import ListView from bills.mixins import AllSubjectsMixin from core.mixins import ConveneTimeMixin from legislators.mixins import AllLegislatorsMixin, ChambersMixin from explainers.models import Explainer from topics.models import Topic class LandingView(AllSubjectsMixin, AllLegislatorsMixin, ChambersMixin, ConveneTimeMixin, ListView): model = Topic template_name = 'landing.html' def get_context_data(self, **kwargs): context = super(LandingView, self).get_context_data(**kwargs) context['explainer_list'] = Explainer.objects.all().published() return context
from django.views.generic import ListView from core.mixins import ConveneTimeMixin from explainers.models import Explainer from topics.models import Topic class LandingView(ConveneTimeMixin, ListView): model = Topic template_name = 'landing.html' def get_context_data(self, **kwargs): context = super(LandingView, self).get_context_data(**kwargs) context['explainer_list'] = Explainer.objects.all().published() return context
Remove unneeded mixins from LandingView
Remove unneeded mixins from LandingView
Python
mit
texastribune/txlege84,texastribune/txlege84,texastribune/txlege84,texastribune/txlege84
from django.views.generic import ListView - from bills.mixins import AllSubjectsMixin from core.mixins import ConveneTimeMixin - from legislators.mixins import AllLegislatorsMixin, ChambersMixin from explainers.models import Explainer from topics.models import Topic + class LandingView(ConveneTimeMixin, ListView): - class LandingView(AllSubjectsMixin, AllLegislatorsMixin, - ChambersMixin, ConveneTimeMixin, ListView): model = Topic template_name = 'landing.html' def get_context_data(self, **kwargs): context = super(LandingView, self).get_context_data(**kwargs) context['explainer_list'] = Explainer.objects.all().published() return context
Remove unneeded mixins from LandingView
## Code Before: from django.views.generic import ListView from bills.mixins import AllSubjectsMixin from core.mixins import ConveneTimeMixin from legislators.mixins import AllLegislatorsMixin, ChambersMixin from explainers.models import Explainer from topics.models import Topic class LandingView(AllSubjectsMixin, AllLegislatorsMixin, ChambersMixin, ConveneTimeMixin, ListView): model = Topic template_name = 'landing.html' def get_context_data(self, **kwargs): context = super(LandingView, self).get_context_data(**kwargs) context['explainer_list'] = Explainer.objects.all().published() return context ## Instruction: Remove unneeded mixins from LandingView ## Code After: from django.views.generic import ListView from core.mixins import ConveneTimeMixin from explainers.models import Explainer from topics.models import Topic class LandingView(ConveneTimeMixin, ListView): model = Topic template_name = 'landing.html' def get_context_data(self, **kwargs): context = super(LandingView, self).get_context_data(**kwargs) context['explainer_list'] = Explainer.objects.all().published() return context
from django.views.generic import ListView - from bills.mixins import AllSubjectsMixin from core.mixins import ConveneTimeMixin - from legislators.mixins import AllLegislatorsMixin, ChambersMixin from explainers.models import Explainer from topics.models import Topic + class LandingView(ConveneTimeMixin, ListView): - class LandingView(AllSubjectsMixin, AllLegislatorsMixin, - ChambersMixin, ConveneTimeMixin, ListView): model = Topic template_name = 'landing.html' def get_context_data(self, **kwargs): context = super(LandingView, self).get_context_data(**kwargs) context['explainer_list'] = Explainer.objects.all().published() return context
0e835c6381374c5b00b7387057d056d679f635c4
zproject/legacy_urls.py
zproject/legacy_urls.py
from django.conf.urls import url import zerver.views import zerver.views.streams import zerver.views.auth import zerver.views.tutorial import zerver.views.report # Future endpoints should add to urls.py, which includes these legacy urls legacy_urls = [ # These are json format views used by the web client. They require a logged in browser. # We should remove this endpoint and all code related to it. # It returns a 404 if the stream doesn't exist, which is confusing # for devs, and I don't think we need to go to the server # any more to find out about subscriptions, since they are already # pushed to us via the event system. url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists), ]
from django.urls import path import zerver.views import zerver.views.streams import zerver.views.auth import zerver.views.tutorial import zerver.views.report # Future endpoints should add to urls.py, which includes these legacy urls legacy_urls = [ # These are json format views used by the web client. They require a logged in browser. # We should remove this endpoint and all code related to it. # It returns a 404 if the stream doesn't exist, which is confusing # for devs, and I don't think we need to go to the server # any more to find out about subscriptions, since they are already # pushed to us via the event system. path('json/subscriptions/exists', zerver.views.streams.json_stream_exists), ]
Migrate legacy urls to use modern django pattern.
urls: Migrate legacy urls to use modern django pattern.
Python
apache-2.0
shubhamdhama/zulip,punchagan/zulip,kou/zulip,showell/zulip,hackerkid/zulip,timabbott/zulip,eeshangarg/zulip,kou/zulip,andersk/zulip,zulip/zulip,synicalsyntax/zulip,zulip/zulip,andersk/zulip,shubhamdhama/zulip,showell/zulip,kou/zulip,hackerkid/zulip,shubhamdhama/zulip,andersk/zulip,eeshangarg/zulip,brainwane/zulip,shubhamdhama/zulip,hackerkid/zulip,punchagan/zulip,showell/zulip,brainwane/zulip,rht/zulip,rht/zulip,shubhamdhama/zulip,hackerkid/zulip,eeshangarg/zulip,hackerkid/zulip,brainwane/zulip,brainwane/zulip,punchagan/zulip,punchagan/zulip,punchagan/zulip,synicalsyntax/zulip,timabbott/zulip,zulip/zulip,rht/zulip,eeshangarg/zulip,brainwane/zulip,punchagan/zulip,timabbott/zulip,timabbott/zulip,synicalsyntax/zulip,synicalsyntax/zulip,hackerkid/zulip,synicalsyntax/zulip,showell/zulip,timabbott/zulip,rht/zulip,kou/zulip,rht/zulip,synicalsyntax/zulip,timabbott/zulip,rht/zulip,eeshangarg/zulip,zulip/zulip,zulip/zulip,eeshangarg/zulip,zulip/zulip,punchagan/zulip,hackerkid/zulip,kou/zulip,andersk/zulip,rht/zulip,andersk/zulip,kou/zulip,shubhamdhama/zulip,andersk/zulip,showell/zulip,timabbott/zulip,synicalsyntax/zulip,eeshangarg/zulip,showell/zulip,kou/zulip,brainwane/zulip,brainwane/zulip,andersk/zulip,zulip/zulip,shubhamdhama/zulip,showell/zulip
- from django.conf.urls import url + from django.urls import path import zerver.views import zerver.views.streams import zerver.views.auth import zerver.views.tutorial import zerver.views.report # Future endpoints should add to urls.py, which includes these legacy urls legacy_urls = [ # These are json format views used by the web client. They require a logged in browser. # We should remove this endpoint and all code related to it. # It returns a 404 if the stream doesn't exist, which is confusing # for devs, and I don't think we need to go to the server # any more to find out about subscriptions, since they are already # pushed to us via the event system. - url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists), + path('json/subscriptions/exists', zerver.views.streams.json_stream_exists), ]
Migrate legacy urls to use modern django pattern.
## Code Before: from django.conf.urls import url import zerver.views import zerver.views.streams import zerver.views.auth import zerver.views.tutorial import zerver.views.report # Future endpoints should add to urls.py, which includes these legacy urls legacy_urls = [ # These are json format views used by the web client. They require a logged in browser. # We should remove this endpoint and all code related to it. # It returns a 404 if the stream doesn't exist, which is confusing # for devs, and I don't think we need to go to the server # any more to find out about subscriptions, since they are already # pushed to us via the event system. url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists), ] ## Instruction: Migrate legacy urls to use modern django pattern. ## Code After: from django.urls import path import zerver.views import zerver.views.streams import zerver.views.auth import zerver.views.tutorial import zerver.views.report # Future endpoints should add to urls.py, which includes these legacy urls legacy_urls = [ # These are json format views used by the web client. They require a logged in browser. # We should remove this endpoint and all code related to it. # It returns a 404 if the stream doesn't exist, which is confusing # for devs, and I don't think we need to go to the server # any more to find out about subscriptions, since they are already # pushed to us via the event system. path('json/subscriptions/exists', zerver.views.streams.json_stream_exists), ]
- from django.conf.urls import url ? ----- ^^^ + from django.urls import path ? ^^^^ import zerver.views import zerver.views.streams import zerver.views.auth import zerver.views.tutorial import zerver.views.report # Future endpoints should add to urls.py, which includes these legacy urls legacy_urls = [ # These are json format views used by the web client. They require a logged in browser. # We should remove this endpoint and all code related to it. # It returns a 404 if the stream doesn't exist, which is confusing # for devs, and I don't think we need to go to the server # any more to find out about subscriptions, since they are already # pushed to us via the event system. - url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists), ? ^^^ - - - + path('json/subscriptions/exists', zerver.views.streams.json_stream_exists), ? ^^^^ ]
65c5474936dca27023e45c1644fa2a9492e9a420
tests/convergence_tests/run_convergence_tests_lspr.py
tests/convergence_tests/run_convergence_tests_lspr.py
import os import time import subprocess import datetime from check_for_meshes import check_mesh # tests to run tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py'] # specify CUDA device to use CUDA_DEVICE = '0' ENV = os.environ.copy() ENV['CUDA_DEVICE'] = CUDA_DEVICE mesh_file = '' folder_name = 'lspr_convergence_test_meshes' rename_folder = 'geometry_lspr' size = '~3MB' check_mesh(mesh_file, folder_name, rename_folder, size) tic = time.time() for test in tests: subprocess.call(['python', '{}'.format(test)]) toc = time.time() print("Total runtime for convergence tests: ") print(str(datetime.timedelta(seconds=(toc - tic))))
import os import time import subprocess import datetime from check_for_meshes import check_mesh # tests to run tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py'] # specify CUDA device to use CUDA_DEVICE = '0' ENV = os.environ.copy() ENV['CUDA_DEVICE'] = CUDA_DEVICE mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip' folder_name = 'lspr_convergence_test_meshes' rename_folder = 'geometry_lspr' size = '~3MB' check_mesh(mesh_file, folder_name, rename_folder, size) tic = time.time() for test in tests: subprocess.call(['python', '{}'.format(test)]) toc = time.time() print("Total runtime for convergence tests: ") print(str(datetime.timedelta(seconds=(toc - tic))))
Add path to convergence test lspr zip file
Add path to convergence test lspr zip file
Python
bsd-3-clause
barbagroup/pygbe,barbagroup/pygbe,barbagroup/pygbe
import os import time import subprocess import datetime from check_for_meshes import check_mesh # tests to run tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py'] # specify CUDA device to use CUDA_DEVICE = '0' ENV = os.environ.copy() ENV['CUDA_DEVICE'] = CUDA_DEVICE - mesh_file = '' + mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip' folder_name = 'lspr_convergence_test_meshes' rename_folder = 'geometry_lspr' size = '~3MB' check_mesh(mesh_file, folder_name, rename_folder, size) tic = time.time() for test in tests: subprocess.call(['python', '{}'.format(test)]) toc = time.time() print("Total runtime for convergence tests: ") print(str(datetime.timedelta(seconds=(toc - tic))))
Add path to convergence test lspr zip file
## Code Before: import os import time import subprocess import datetime from check_for_meshes import check_mesh # tests to run tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py'] # specify CUDA device to use CUDA_DEVICE = '0' ENV = os.environ.copy() ENV['CUDA_DEVICE'] = CUDA_DEVICE mesh_file = '' folder_name = 'lspr_convergence_test_meshes' rename_folder = 'geometry_lspr' size = '~3MB' check_mesh(mesh_file, folder_name, rename_folder, size) tic = time.time() for test in tests: subprocess.call(['python', '{}'.format(test)]) toc = time.time() print("Total runtime for convergence tests: ") print(str(datetime.timedelta(seconds=(toc - tic)))) ## Instruction: Add path to convergence test lspr zip file ## Code After: import os import time import subprocess import datetime from check_for_meshes import check_mesh # tests to run tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py'] # specify CUDA device to use CUDA_DEVICE = '0' ENV = os.environ.copy() ENV['CUDA_DEVICE'] = CUDA_DEVICE mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip' folder_name = 'lspr_convergence_test_meshes' rename_folder = 'geometry_lspr' size = '~3MB' check_mesh(mesh_file, folder_name, rename_folder, size) tic = time.time() for test in tests: subprocess.call(['python', '{}'.format(test)]) toc = time.time() print("Total runtime for convergence tests: ") print(str(datetime.timedelta(seconds=(toc - tic))))
import os import time import subprocess import datetime from check_for_meshes import check_mesh # tests to run tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py'] # specify CUDA device to use CUDA_DEVICE = '0' ENV = os.environ.copy() ENV['CUDA_DEVICE'] = CUDA_DEVICE - mesh_file = '' + mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip' folder_name = 'lspr_convergence_test_meshes' rename_folder = 'geometry_lspr' size = '~3MB' check_mesh(mesh_file, folder_name, rename_folder, size) tic = time.time() for test in tests: subprocess.call(['python', '{}'.format(test)]) toc = time.time() print("Total runtime for convergence tests: ") print(str(datetime.timedelta(seconds=(toc - tic))))
bad2fea8a3a8e7a7d1da9ee83ec48657824eaa07
tests/test_filesize.py
tests/test_filesize.py
from rich import filesize def test_traditional(): assert filesize.decimal(0) == "0 bytes" assert filesize.decimal(1) == "1 byte" assert filesize.decimal(2) == "2 bytes" assert filesize.decimal(1000) == "1.0 kB" assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB" def test_pick_unit_and_suffix(): units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"] assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes") assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
from rich import filesize def test_traditional(): assert filesize.decimal(0) == "0 bytes" assert filesize.decimal(1) == "1 byte" assert filesize.decimal(2) == "2 bytes" assert filesize.decimal(1000) == "1.0 kB" assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB" assert filesize.decimal(0, precision=2) == "0 bytes" assert filesize.decimal(1111, precision=0) == "1 kB" assert filesize.decimal(1111, precision=1) == "1.1 kB" assert filesize.decimal(1111, precision=2) == "1.11 kB" assert filesize.decimal(1111, separator="") == "1.1kB" def test_pick_unit_and_suffix(): units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"] assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes") assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
Add some tests for new decimal() params
Add some tests for new decimal() params
Python
mit
willmcgugan/rich
from rich import filesize def test_traditional(): assert filesize.decimal(0) == "0 bytes" assert filesize.decimal(1) == "1 byte" assert filesize.decimal(2) == "2 bytes" assert filesize.decimal(1000) == "1.0 kB" assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB" + assert filesize.decimal(0, precision=2) == "0 bytes" + assert filesize.decimal(1111, precision=0) == "1 kB" + assert filesize.decimal(1111, precision=1) == "1.1 kB" + assert filesize.decimal(1111, precision=2) == "1.11 kB" + assert filesize.decimal(1111, separator="") == "1.1kB" def test_pick_unit_and_suffix(): units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"] assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes") assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
Add some tests for new decimal() params
## Code Before: from rich import filesize def test_traditional(): assert filesize.decimal(0) == "0 bytes" assert filesize.decimal(1) == "1 byte" assert filesize.decimal(2) == "2 bytes" assert filesize.decimal(1000) == "1.0 kB" assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB" def test_pick_unit_and_suffix(): units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"] assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes") assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB") ## Instruction: Add some tests for new decimal() params ## Code After: from rich import filesize def test_traditional(): assert filesize.decimal(0) == "0 bytes" assert filesize.decimal(1) == "1 byte" assert filesize.decimal(2) == "2 bytes" assert filesize.decimal(1000) == "1.0 kB" assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB" assert filesize.decimal(0, precision=2) == "0 bytes" assert filesize.decimal(1111, precision=0) == "1 kB" assert filesize.decimal(1111, precision=1) == "1.1 kB" assert filesize.decimal(1111, precision=2) == "1.11 kB" assert filesize.decimal(1111, separator="") == "1.1kB" def test_pick_unit_and_suffix(): units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"] assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes") assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
from rich import filesize def test_traditional(): assert filesize.decimal(0) == "0 bytes" assert filesize.decimal(1) == "1 byte" assert filesize.decimal(2) == "2 bytes" assert filesize.decimal(1000) == "1.0 kB" assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB" + assert filesize.decimal(0, precision=2) == "0 bytes" + assert filesize.decimal(1111, precision=0) == "1 kB" + assert filesize.decimal(1111, precision=1) == "1.1 kB" + assert filesize.decimal(1111, precision=2) == "1.11 kB" + assert filesize.decimal(1111, separator="") == "1.1kB" def test_pick_unit_and_suffix(): units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"] assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes") assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
3e20365624f02b70d8332ba7ff7da23961337f86
quickstart/python/understand/example-3/create_joke_samples.6.x.py
quickstart/python/understand/example-3/create_joke_samples.6.x.py
from twilio.rest import Client # Your Account Sid and Auth Token from twilio.com/console account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX' auth_token = 'your_auth_token' client = Client(account_sid, auth_token) phrases = [ 'Tell me a joke', 'Tell me a joke', 'I\'d like to hear a joke', 'Do you know any good jokes?', 'Joke', 'Tell joke', 'Tell me something funny', 'Make me laugh', 'I want to hear a joke', 'Can I hear a joke?', 'I like jokes', 'I\'d like to hear a punny joke' ] # Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list for phrase in phrases: sample = client.preview.understand \ .assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .tasks('tell-a-joke') \ .samples \ .create(language='en-us', tagged_text=phrase) print(sample.sid)
from twilio.rest import Client # Your Account Sid and Auth Token from twilio.com/console account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX' auth_token = 'your_auth_token' client = Client(account_sid, auth_token) phrases = [ 'Tell me a joke', 'Tell me a joke', 'I\'d like to hear a joke', 'Do you know any good jokes?', 'Joke', 'Tell joke', 'Tell me something funny', 'Make me laugh', 'I want to hear a joke', 'Can I hear a joke?', 'I like jokes', 'I\'d like to hear a punny joke' ] # Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list # Replace 'UDXXX...' with the SID for the task you just created. for phrase in phrases: sample = client.preview.understand \ .assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .samples \ .create(language='en-us', tagged_text=phrase) print(sample.sid)
Update samples creation for intent rename
Update samples creation for intent rename Update intent --> task, code comment
Python
mit
TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets
from twilio.rest import Client # Your Account Sid and Auth Token from twilio.com/console account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX' auth_token = 'your_auth_token' client = Client(account_sid, auth_token) phrases = [ 'Tell me a joke', 'Tell me a joke', 'I\'d like to hear a joke', 'Do you know any good jokes?', 'Joke', 'Tell joke', 'Tell me something funny', 'Make me laugh', 'I want to hear a joke', 'Can I hear a joke?', 'I like jokes', 'I\'d like to hear a punny joke' ] # Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list + # Replace 'UDXXX...' with the SID for the task you just created. for phrase in phrases: sample = client.preview.understand \ .assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ - .tasks('tell-a-joke') \ + .tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .samples \ .create(language='en-us', tagged_text=phrase) print(sample.sid)
Update samples creation for intent rename
## Code Before: from twilio.rest import Client # Your Account Sid and Auth Token from twilio.com/console account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX' auth_token = 'your_auth_token' client = Client(account_sid, auth_token) phrases = [ 'Tell me a joke', 'Tell me a joke', 'I\'d like to hear a joke', 'Do you know any good jokes?', 'Joke', 'Tell joke', 'Tell me something funny', 'Make me laugh', 'I want to hear a joke', 'Can I hear a joke?', 'I like jokes', 'I\'d like to hear a punny joke' ] # Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list for phrase in phrases: sample = client.preview.understand \ .assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .tasks('tell-a-joke') \ .samples \ .create(language='en-us', tagged_text=phrase) print(sample.sid) ## Instruction: Update samples creation for intent rename ## Code After: from twilio.rest import Client # Your Account Sid and Auth Token from twilio.com/console account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX' auth_token = 'your_auth_token' client = Client(account_sid, auth_token) phrases = [ 'Tell me a joke', 'Tell me a joke', 'I\'d like to hear a joke', 'Do you know any good jokes?', 'Joke', 'Tell joke', 'Tell me something funny', 'Make me laugh', 'I want to hear a joke', 'Can I hear a joke?', 'I like jokes', 'I\'d like to hear a punny joke' ] # Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list # Replace 'UDXXX...' with the SID for the task you just created. for phrase in phrases: sample = client.preview.understand \ .assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .samples \ .create(language='en-us', tagged_text=phrase) print(sample.sid)
from twilio.rest import Client # Your Account Sid and Auth Token from twilio.com/console account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX' auth_token = 'your_auth_token' client = Client(account_sid, auth_token) phrases = [ 'Tell me a joke', 'Tell me a joke', 'I\'d like to hear a joke', 'Do you know any good jokes?', 'Joke', 'Tell joke', 'Tell me something funny', 'Make me laugh', 'I want to hear a joke', 'Can I hear a joke?', 'I like jokes', 'I\'d like to hear a punny joke' ] # Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list + # Replace 'UDXXX...' with the SID for the task you just created. for phrase in phrases: sample = client.preview.understand \ .assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ - .tasks('tell-a-joke') \ + .tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \ .samples \ .create(language='en-us', tagged_text=phrase) print(sample.sid)
d33a624fa6aedb93ae43ba1d2c0f6a76d90ff4a6
foldermd5sums.py
foldermd5sums.py
import json import os import sys import hashlib def get_md5sums(directory): md5sums = [] for filename in os.listdir(directory): md5 = hashlib.md5() with open(os.path.join(directory, filename), 'rb') as fp: for chunk in iter(lambda: fp.read(128 * md5.block_size), b''): md5.update(chunk) md5hash = md5.hexdigest() md5sums.append((filename, md5hash)) return md5sums if __name__ == '__main__': if len(sys.argv) < 3: print('Usage: ' + sys.argv[0] + ' input_directory output.json') sys.exit(1) directory = sys.argv[1] if not os.path.exists(directory): print('Directory does not exist!') sys.exit(1) output_json = sys.argv[2] md5sums = get_md5sums(directory) with open(output_json, 'w') as fp: json.dump(md5sums, fp, indent=0)
import json import os import sys import hashlib def get_relative_filepaths(base_directory): """ Return a list of file paths without the base_directory prefix""" file_list = [] for root, subFolders, files in os.walk('Data'): relative_path="/".join(root.split('/')[1:]) for file in files: file_list.append(os.path.join(relative_path,file)) return file_list def get_md5sums(base_directory): md5sums = [] for filename in get_relative_filepaths(base_directory): md5 = hashlib.md5() full_filepath = os.path.join(base_directory, filename) with open(full_filepath, 'rb') as fp: for chunk in iter(lambda: fp.read(128 * md5.block_size), b''): md5.update(chunk) md5hash = md5.hexdigest() md5sums.append((filename, md5hash)) return md5sums if __name__ == '__main__': if len(sys.argv) < 3: print('Usage: ' + sys.argv[0] + ' input_directory output.json') sys.exit(1) directory = sys.argv[1] if not os.path.exists(directory): print('Directory does not exist!') sys.exit(1) output_json = sys.argv[2] md5sums = get_md5sums(directory) with open(output_json, 'w') as fp: json.dump(md5sums, fp, indent=0)
Allow directory of files to be indexed
ENH: Allow directory of files to be indexed In the Data directory, there may be sub-directories of files that need to be kept separate, but all of them need to be indexed.
Python
apache-2.0
zivy/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,zivy/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks,zivy/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks
import json import os import sys import hashlib + def get_relative_filepaths(base_directory): + """ Return a list of file paths without the base_directory prefix""" + file_list = [] + for root, subFolders, files in os.walk('Data'): + relative_path="/".join(root.split('/')[1:]) + for file in files: + file_list.append(os.path.join(relative_path,file)) + return file_list + + - def get_md5sums(directory): + def get_md5sums(base_directory): md5sums = [] - for filename in os.listdir(directory): + for filename in get_relative_filepaths(base_directory): md5 = hashlib.md5() - with open(os.path.join(directory, filename), 'rb') as fp: + full_filepath = os.path.join(base_directory, filename) + with open(full_filepath, 'rb') as fp: for chunk in iter(lambda: fp.read(128 * md5.block_size), b''): md5.update(chunk) md5hash = md5.hexdigest() md5sums.append((filename, md5hash)) - return md5sums if __name__ == '__main__': if len(sys.argv) < 3: print('Usage: ' + sys.argv[0] + ' input_directory output.json') sys.exit(1) directory = sys.argv[1] if not os.path.exists(directory): print('Directory does not exist!') sys.exit(1) output_json = sys.argv[2] md5sums = get_md5sums(directory) with open(output_json, 'w') as fp: json.dump(md5sums, fp, indent=0)
Allow directory of files to be indexed
## Code Before: import json import os import sys import hashlib def get_md5sums(directory): md5sums = [] for filename in os.listdir(directory): md5 = hashlib.md5() with open(os.path.join(directory, filename), 'rb') as fp: for chunk in iter(lambda: fp.read(128 * md5.block_size), b''): md5.update(chunk) md5hash = md5.hexdigest() md5sums.append((filename, md5hash)) return md5sums if __name__ == '__main__': if len(sys.argv) < 3: print('Usage: ' + sys.argv[0] + ' input_directory output.json') sys.exit(1) directory = sys.argv[1] if not os.path.exists(directory): print('Directory does not exist!') sys.exit(1) output_json = sys.argv[2] md5sums = get_md5sums(directory) with open(output_json, 'w') as fp: json.dump(md5sums, fp, indent=0) ## Instruction: Allow directory of files to be indexed ## Code After: import json import os import sys import hashlib def get_relative_filepaths(base_directory): """ Return a list of file paths without the base_directory prefix""" file_list = [] for root, subFolders, files in os.walk('Data'): relative_path="/".join(root.split('/')[1:]) for file in files: file_list.append(os.path.join(relative_path,file)) return file_list def get_md5sums(base_directory): md5sums = [] for filename in get_relative_filepaths(base_directory): md5 = hashlib.md5() full_filepath = os.path.join(base_directory, filename) with open(full_filepath, 'rb') as fp: for chunk in iter(lambda: fp.read(128 * md5.block_size), b''): md5.update(chunk) md5hash = md5.hexdigest() md5sums.append((filename, md5hash)) return md5sums if __name__ == '__main__': if len(sys.argv) < 3: print('Usage: ' + sys.argv[0] + ' input_directory output.json') sys.exit(1) directory = sys.argv[1] if not os.path.exists(directory): print('Directory does not exist!') sys.exit(1) output_json = sys.argv[2] md5sums = get_md5sums(directory) with open(output_json, 'w') as fp: json.dump(md5sums, fp, indent=0)
import json import os import sys import hashlib + def get_relative_filepaths(base_directory): + """ Return a list of file paths without the base_directory prefix""" + file_list = [] + for root, subFolders, files in os.walk('Data'): + relative_path="/".join(root.split('/')[1:]) + for file in files: + file_list.append(os.path.join(relative_path,file)) + return file_list + + - def get_md5sums(directory): + def get_md5sums(base_directory): ? +++++ md5sums = [] - for filename in os.listdir(directory): + for filename in get_relative_filepaths(base_directory): md5 = hashlib.md5() - with open(os.path.join(directory, filename), 'rb') as fp: + full_filepath = os.path.join(base_directory, filename) + with open(full_filepath, 'rb') as fp: for chunk in iter(lambda: fp.read(128 * md5.block_size), b''): md5.update(chunk) md5hash = md5.hexdigest() md5sums.append((filename, md5hash)) - return md5sums if __name__ == '__main__': if len(sys.argv) < 3: print('Usage: ' + sys.argv[0] + ' input_directory output.json') sys.exit(1) directory = sys.argv[1] if not os.path.exists(directory): print('Directory does not exist!') sys.exit(1) output_json = sys.argv[2] md5sums = get_md5sums(directory) with open(output_json, 'w') as fp: json.dump(md5sums, fp, indent=0)
c6d7f2b1214e86f09431ab1d8e5c312f7a87081d
pttrack/views.py
pttrack/views.py
from django.shortcuts import render from django.http import HttpResponse, HttpResponseRedirect # Create your views here.from django.http import HttpResponse def index(request): return HttpResponse("Hello, world. You're at the BIG TABLE.") def clindate(request, clindate): (year, month, day) = clindate.split("-") return HttpResponse("Clinic date %s" % year+" "+month+" "+day) def patient(request, pt_uuid): return HttpResponse("You're looking at patient %s" % pt_uuid) def intake(request): from . import forms as myforms # if this is a POST request we need to process the form data if request.method == 'POST': # create a form instance and populate it with data from the request: form = myforms.PatientForm(request.POST) # check whether it's valid: if form.is_valid(): # process the data in form.cleaned_data as required # ... # redirect to a new URL: return HttpResponseRedirect('/') # if a GET (or any other method) we'll create a blank form else: form = myforms.PatientForm() return render(request, 'pttrack/intake.html', {'form': form})
from django.shortcuts import render from django.http import HttpResponse, HttpResponseRedirect from django.core.urlresolvers import reverse from . import models as mymodels # Create your views here.from django.http import HttpResponse def index(request): return HttpResponse("Hello, world. You're at the BIG TABLE.") def clindate(request, clindate): (year, month, day) = clindate.split("-") return HttpResponse("Clinic date %s" % year+" "+month+" "+day) def patient(request, pt_uuid): return HttpResponse("You're looking at patient %s" % pt_uuid) def intake(request): from . import forms as myforms # if this is a POST request we need to process the form data if request.method == 'POST': # create a form instance and populate it with data from the request: form = myforms.PatientForm(request.POST) # check whether it's valid: if form.is_valid(): p = mymodels.Patient(**form.cleaned_data) p.save() # redirect to a new URL: return HttpResponseRedirect(reverse(patient, args=(p.id,))) # if a GET (or any other method) we'll create a blank form else: form = myforms.PatientForm() return render(request, 'pttrack/intake.html', {'form': form})
Set up redirect at db saves for new patients.
Set up redirect at db saves for new patients.
Python
mit
SaturdayNeighborhoodHealthClinic/clintools,SaturdayNeighborhoodHealthClinic/clintools,SaturdayNeighborhoodHealthClinic/clintools
from django.shortcuts import render from django.http import HttpResponse, HttpResponseRedirect + from django.core.urlresolvers import reverse + + from . import models as mymodels # Create your views here.from django.http import HttpResponse def index(request): return HttpResponse("Hello, world. You're at the BIG TABLE.") def clindate(request, clindate): (year, month, day) = clindate.split("-") return HttpResponse("Clinic date %s" % year+" "+month+" "+day) def patient(request, pt_uuid): return HttpResponse("You're looking at patient %s" % pt_uuid) def intake(request): from . import forms as myforms # if this is a POST request we need to process the form data if request.method == 'POST': # create a form instance and populate it with data from the request: form = myforms.PatientForm(request.POST) # check whether it's valid: if form.is_valid(): - # process the data in form.cleaned_data as required - # ... + p = mymodels.Patient(**form.cleaned_data) + p.save() + # redirect to a new URL: - return HttpResponseRedirect('/') + return HttpResponseRedirect(reverse(patient, args=(p.id,))) # if a GET (or any other method) we'll create a blank form else: form = myforms.PatientForm() return render(request, 'pttrack/intake.html', {'form': form})
Set up redirect at db saves for new patients.
## Code Before: from django.shortcuts import render from django.http import HttpResponse, HttpResponseRedirect # Create your views here.from django.http import HttpResponse def index(request): return HttpResponse("Hello, world. You're at the BIG TABLE.") def clindate(request, clindate): (year, month, day) = clindate.split("-") return HttpResponse("Clinic date %s" % year+" "+month+" "+day) def patient(request, pt_uuid): return HttpResponse("You're looking at patient %s" % pt_uuid) def intake(request): from . import forms as myforms # if this is a POST request we need to process the form data if request.method == 'POST': # create a form instance and populate it with data from the request: form = myforms.PatientForm(request.POST) # check whether it's valid: if form.is_valid(): # process the data in form.cleaned_data as required # ... # redirect to a new URL: return HttpResponseRedirect('/') # if a GET (or any other method) we'll create a blank form else: form = myforms.PatientForm() return render(request, 'pttrack/intake.html', {'form': form}) ## Instruction: Set up redirect at db saves for new patients. ## Code After: from django.shortcuts import render from django.http import HttpResponse, HttpResponseRedirect from django.core.urlresolvers import reverse from . import models as mymodels # Create your views here.from django.http import HttpResponse def index(request): return HttpResponse("Hello, world. You're at the BIG TABLE.") def clindate(request, clindate): (year, month, day) = clindate.split("-") return HttpResponse("Clinic date %s" % year+" "+month+" "+day) def patient(request, pt_uuid): return HttpResponse("You're looking at patient %s" % pt_uuid) def intake(request): from . import forms as myforms # if this is a POST request we need to process the form data if request.method == 'POST': # create a form instance and populate it with data from the request: form = myforms.PatientForm(request.POST) # check whether it's valid: if form.is_valid(): p = mymodels.Patient(**form.cleaned_data) p.save() # redirect to a new URL: return HttpResponseRedirect(reverse(patient, args=(p.id,))) # if a GET (or any other method) we'll create a blank form else: form = myforms.PatientForm() return render(request, 'pttrack/intake.html', {'form': form})
from django.shortcuts import render from django.http import HttpResponse, HttpResponseRedirect + from django.core.urlresolvers import reverse + + from . import models as mymodels # Create your views here.from django.http import HttpResponse def index(request): return HttpResponse("Hello, world. You're at the BIG TABLE.") def clindate(request, clindate): (year, month, day) = clindate.split("-") return HttpResponse("Clinic date %s" % year+" "+month+" "+day) def patient(request, pt_uuid): return HttpResponse("You're looking at patient %s" % pt_uuid) def intake(request): from . import forms as myforms # if this is a POST request we need to process the form data if request.method == 'POST': # create a form instance and populate it with data from the request: form = myforms.PatientForm(request.POST) # check whether it's valid: if form.is_valid(): - # process the data in form.cleaned_data as required - # ... + p = mymodels.Patient(**form.cleaned_data) + p.save() + # redirect to a new URL: - return HttpResponseRedirect('/') + return HttpResponseRedirect(reverse(patient, args=(p.id,))) # if a GET (or any other method) we'll create a blank form else: form = myforms.PatientForm() return render(request, 'pttrack/intake.html', {'form': form})
0654d962918327e5143fb9250ad344de26e284eb
electrumx_server.py
electrumx_server.py
'''Script to kick off the server.''' import logging import traceback from server.env import Env from server.controller import Controller def main(): '''Set up logging and run the server.''' logging.basicConfig(level=logging.INFO, format='%(asctime)s %(levelname)-9s %(message)-100s ' '%(name)s [%(filename)s:%(lineno)d]') logging.info('ElectrumX server starting') try: controller = Controller(Env()) controller.run() except Exception: traceback.print_exc() logging.critical('ElectrumX server terminated abnormally') else: logging.info('ElectrumX server terminated normally') if __name__ == '__main__': main()
'''Script to kick off the server.''' import logging import traceback from server.env import Env from server.controller import Controller def main(): '''Set up logging and run the server.''' logging.basicConfig(level=logging.INFO, format='%(asctime)s %(levelname)-7s %(message)-100s ' '[%(filename)s:%(lineno)d]') logging.info('ElectrumX server starting') try: controller = Controller(Env()) controller.run() except Exception: traceback.print_exc() logging.critical('ElectrumX server terminated abnormally') else: logging.info('ElectrumX server terminated normally') if __name__ == '__main__': main()
Remove logger name from logs
Remove logger name from logs
Python
mit
thelazier/electrumx,shsmith/electrumx,shsmith/electrumx,erasmospunk/electrumx,erasmospunk/electrumx,thelazier/electrumx
'''Script to kick off the server.''' import logging import traceback from server.env import Env from server.controller import Controller def main(): '''Set up logging and run the server.''' logging.basicConfig(level=logging.INFO, - format='%(asctime)s %(levelname)-9s %(message)-100s ' + format='%(asctime)s %(levelname)-7s %(message)-100s ' - '%(name)s [%(filename)s:%(lineno)d]') + '[%(filename)s:%(lineno)d]') logging.info('ElectrumX server starting') try: controller = Controller(Env()) controller.run() except Exception: traceback.print_exc() logging.critical('ElectrumX server terminated abnormally') else: logging.info('ElectrumX server terminated normally') if __name__ == '__main__': main()
Remove logger name from logs
## Code Before: '''Script to kick off the server.''' import logging import traceback from server.env import Env from server.controller import Controller def main(): '''Set up logging and run the server.''' logging.basicConfig(level=logging.INFO, format='%(asctime)s %(levelname)-9s %(message)-100s ' '%(name)s [%(filename)s:%(lineno)d]') logging.info('ElectrumX server starting') try: controller = Controller(Env()) controller.run() except Exception: traceback.print_exc() logging.critical('ElectrumX server terminated abnormally') else: logging.info('ElectrumX server terminated normally') if __name__ == '__main__': main() ## Instruction: Remove logger name from logs ## Code After: '''Script to kick off the server.''' import logging import traceback from server.env import Env from server.controller import Controller def main(): '''Set up logging and run the server.''' logging.basicConfig(level=logging.INFO, format='%(asctime)s %(levelname)-7s %(message)-100s ' '[%(filename)s:%(lineno)d]') logging.info('ElectrumX server starting') try: controller = Controller(Env()) controller.run() except Exception: traceback.print_exc() logging.critical('ElectrumX server terminated abnormally') else: logging.info('ElectrumX server terminated normally') if __name__ == '__main__': main()
'''Script to kick off the server.''' import logging import traceback from server.env import Env from server.controller import Controller def main(): '''Set up logging and run the server.''' logging.basicConfig(level=logging.INFO, - format='%(asctime)s %(levelname)-9s %(message)-100s ' ? ^ + format='%(asctime)s %(levelname)-7s %(message)-100s ' ? ^ - '%(name)s [%(filename)s:%(lineno)d]') ? --------- + '[%(filename)s:%(lineno)d]') logging.info('ElectrumX server starting') try: controller = Controller(Env()) controller.run() except Exception: traceback.print_exc() logging.critical('ElectrumX server terminated abnormally') else: logging.info('ElectrumX server terminated normally') if __name__ == '__main__': main()
a3f5e1338cc84c60b867fc04175253f7ab460912
relay_api/api/backend.py
relay_api/api/backend.py
import json from relay_api.core.relay import relay from relay_api.conf.config import relays def init_relays(): for r in relays: relays[r]["object"] = relay(relays[r]["gpio"]) relays[r]["state"] = relays[r]["object"].get_state() def get_all_relays(): relays_dict = __get_relay_dict() return json.dumps(relays_dict) def get_relay(relay_name): if relay_name not in relays: return None relay_dict = __get_relay_dict(relay_name) return json.dumps(relay_dict) def __get_relay_dict(relay_name=None): if relay_name: relay_dict = dict.copy(relays["relay_name"]) del(relay_dict["object"]) return relay_dict relays_dict = dict.copy(relays) for r in relays_dict: del(relays_dict[r]["object"]) return relays_dict
import json from relay_api.core.relay import relay from relay_api.conf.config import relays def init_relays(): for r in relays: relays[r]["object"] = relay(relays[r]["gpio"]) relays[r]["state"] = relays[r]["object"].get_state() def get_all_relays(): relays_dict = __get_relay_dict() return json.dumps(relays_dict, indent=4) def get_relay(relay_name): if relay_name not in relays: return None relay_dict = __get_relay_dict(relay_name) return json.dumps(relay_dict, indent=4) def __get_relay_dict(relay_name=None): if relay_name: relay_dict = dict.copy(relays["relay_name"]) del(relay_dict["object"]) return relay_dict relays_dict = dict.copy(relays) for r in relays_dict: del(relays_dict[r]["object"]) return relays_dict
Add indent in json to improve debugging
Add indent in json to improve debugging
Python
mit
pahumadad/raspi-relay-api
import json from relay_api.core.relay import relay from relay_api.conf.config import relays def init_relays(): for r in relays: relays[r]["object"] = relay(relays[r]["gpio"]) relays[r]["state"] = relays[r]["object"].get_state() def get_all_relays(): relays_dict = __get_relay_dict() - return json.dumps(relays_dict) + return json.dumps(relays_dict, indent=4) def get_relay(relay_name): if relay_name not in relays: return None relay_dict = __get_relay_dict(relay_name) - return json.dumps(relay_dict) + return json.dumps(relay_dict, indent=4) def __get_relay_dict(relay_name=None): if relay_name: relay_dict = dict.copy(relays["relay_name"]) del(relay_dict["object"]) return relay_dict relays_dict = dict.copy(relays) for r in relays_dict: del(relays_dict[r]["object"]) return relays_dict
Add indent in json to improve debugging
## Code Before: import json from relay_api.core.relay import relay from relay_api.conf.config import relays def init_relays(): for r in relays: relays[r]["object"] = relay(relays[r]["gpio"]) relays[r]["state"] = relays[r]["object"].get_state() def get_all_relays(): relays_dict = __get_relay_dict() return json.dumps(relays_dict) def get_relay(relay_name): if relay_name not in relays: return None relay_dict = __get_relay_dict(relay_name) return json.dumps(relay_dict) def __get_relay_dict(relay_name=None): if relay_name: relay_dict = dict.copy(relays["relay_name"]) del(relay_dict["object"]) return relay_dict relays_dict = dict.copy(relays) for r in relays_dict: del(relays_dict[r]["object"]) return relays_dict ## Instruction: Add indent in json to improve debugging ## Code After: import json from relay_api.core.relay import relay from relay_api.conf.config import relays def init_relays(): for r in relays: relays[r]["object"] = relay(relays[r]["gpio"]) relays[r]["state"] = relays[r]["object"].get_state() def get_all_relays(): relays_dict = __get_relay_dict() return json.dumps(relays_dict, indent=4) def get_relay(relay_name): if relay_name not in relays: return None relay_dict = __get_relay_dict(relay_name) return json.dumps(relay_dict, indent=4) def __get_relay_dict(relay_name=None): if relay_name: relay_dict = dict.copy(relays["relay_name"]) del(relay_dict["object"]) return relay_dict relays_dict = dict.copy(relays) for r in relays_dict: del(relays_dict[r]["object"]) return relays_dict
import json from relay_api.core.relay import relay from relay_api.conf.config import relays def init_relays(): for r in relays: relays[r]["object"] = relay(relays[r]["gpio"]) relays[r]["state"] = relays[r]["object"].get_state() def get_all_relays(): relays_dict = __get_relay_dict() - return json.dumps(relays_dict) + return json.dumps(relays_dict, indent=4) ? ++++++++++ def get_relay(relay_name): if relay_name not in relays: return None relay_dict = __get_relay_dict(relay_name) - return json.dumps(relay_dict) + return json.dumps(relay_dict, indent=4) ? ++++++++++ def __get_relay_dict(relay_name=None): if relay_name: relay_dict = dict.copy(relays["relay_name"]) del(relay_dict["object"]) return relay_dict relays_dict = dict.copy(relays) for r in relays_dict: del(relays_dict[r]["object"]) return relays_dict
00ddeefdcdacb811f5e665a91139e165d7217f84
week1/poc_2048_merge_template.py
week1/poc_2048_merge_template.py
def merge(line): """ Function that merges a single row or column in 2048. """ l = len(line) s1 = [0]*l j = 0 for i in range(l): if l[i] != 0: s1[j] = l[i] return [] a = [2,0,2,4] print merge(a)
def merge(line): """ Function that merges a single row or column in 2048. """ l = len(line) s1 = [0]*l j = 0 for i in range(l): if line[i] != 0: s1[j] = line[i] j += 1 return s1 a = [2,0,2,4] print (merge(a))
Modify the correct merge 1 fct
Modify the correct merge 1 fct
Python
mit
Crescent-Saturn/Principles-of-Computing
def merge(line): """ Function that merges a single row or column in 2048. """ l = len(line) s1 = [0]*l j = 0 for i in range(l): - if l[i] != 0: + if line[i] != 0: - s1[j] = l[i] - + s1[j] = line[i] + j += 1 - return [] + return s1 a = [2,0,2,4] - print merge(a) + print (merge(a))
Modify the correct merge 1 fct
## Code Before: def merge(line): """ Function that merges a single row or column in 2048. """ l = len(line) s1 = [0]*l j = 0 for i in range(l): if l[i] != 0: s1[j] = l[i] return [] a = [2,0,2,4] print merge(a) ## Instruction: Modify the correct merge 1 fct ## Code After: def merge(line): """ Function that merges a single row or column in 2048. """ l = len(line) s1 = [0]*l j = 0 for i in range(l): if line[i] != 0: s1[j] = line[i] j += 1 return s1 a = [2,0,2,4] print (merge(a))
def merge(line): """ Function that merges a single row or column in 2048. """ l = len(line) s1 = [0]*l j = 0 for i in range(l): - if l[i] != 0: ? ^ + if line[i] != 0: ? ^^^^ +++ - s1[j] = l[i] - + s1[j] = line[i] + j += 1 - return [] ? ^^ + return s1 ? ^^ a = [2,0,2,4] - print merge(a) + print (merge(a)) ? + +
a0903bb9fd988662269e9f2ef7e38acd877a63d5
src/nodeconductor_saltstack/saltstack/handlers.py
src/nodeconductor_saltstack/saltstack/handlers.py
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
Add more details to event logs for property CRUD
Add more details to event logs for property CRUD
Python
mit
opennode/nodeconductor-saltstack
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( - '%s {property_name} has been created.' % instance.get_type_display_name(), + '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( - '%s {property_name} has been updated.' % instance.get_type_display_name(), + '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( - '%s {property_name} has been deleted.' % instance.get_type_display_name(), + '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
Add more details to event logs for property CRUD
## Code Before: from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, }) ## Instruction: Add more details to event logs for property CRUD ## Code After: from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
from __future__ import unicode_literals import logging from .log import event_logger logger = logging.getLogger(__name__) def log_saltstack_property_created(sender, instance, created=False, **kwargs): if created: event_logger.saltstack_property.info( - '%s {property_name} has been created.' % instance.get_type_display_name(), + '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(), ? ++++++++++++++++++++++ event_type='saltstack_property_creation_succeeded', event_context={ 'property': instance, }) else: event_logger.saltstack_property.info( - '%s {property_name} has been updated.' % instance.get_type_display_name(), + '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(), ? ++++++++++++++++++++++ event_type='saltstack_property_update_succeeded', event_context={ 'property': instance, }) def log_saltstack_property_deleted(sender, instance, **kwargs): event_logger.saltstack_property.info( - '%s {property_name} has been deleted.' % instance.get_type_display_name(), + '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(), ? ++++++++++++++++++++++++ event_type='saltstack_property_deletion_succeeded', event_context={ 'property': instance, })
3222fab1b026250d9aee863d068137b03c13a05b
tests/test_check_dependencies.py
tests/test_check_dependencies.py
from check_dependencies import CheckDependencies def test_default(): CheckDependencies(None) def test_hydrotrend(): CheckDependencies("hydrotrend")
from check_dependencies import CheckDependencies def test_default(): CheckDependencies(None) def test_hydrotrend(): CheckDependencies("hydrotrend") def test_cem(): CheckDependencies("cem")
Add dependency check test for CEM
Add dependency check test for CEM
Python
mit
csdms/rpm_models
from check_dependencies import CheckDependencies def test_default(): CheckDependencies(None) def test_hydrotrend(): CheckDependencies("hydrotrend") + def test_cem(): + CheckDependencies("cem") +
Add dependency check test for CEM
## Code Before: from check_dependencies import CheckDependencies def test_default(): CheckDependencies(None) def test_hydrotrend(): CheckDependencies("hydrotrend") ## Instruction: Add dependency check test for CEM ## Code After: from check_dependencies import CheckDependencies def test_default(): CheckDependencies(None) def test_hydrotrend(): CheckDependencies("hydrotrend") def test_cem(): CheckDependencies("cem")
from check_dependencies import CheckDependencies def test_default(): CheckDependencies(None) def test_hydrotrend(): CheckDependencies("hydrotrend") + + def test_cem(): + CheckDependencies("cem")
1599d4ed14fb3d7c7e551c9f6ce3f86d9df17cbd
mammoth/writers/html.py
mammoth/writers/html.py
from __future__ import unicode_literals from .abc import Writer import cgi class HtmlWriter(Writer): def __init__(self): self._fragments = [] def text(self, text): self._fragments.append(_escape_html(text)) def start(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1}>".format(name, attribute_string)) def end(self, name): self._fragments.append("</{0}>".format(name)) def self_closing(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1} />".format(name, attribute_string)) def append(self, html): self._fragments.append(html) def as_string(self): return "".join(self._fragments) def _escape_html(text): return cgi.escape(text, quote=True) def _generate_attribute_string(attributes): if attributes is None: return "" else: return "".join( ' {0}="{1}"'.format(key, _escape_html(attributes[key])) for key in sorted(attributes) )
from __future__ import unicode_literals from xml.sax.saxutils import escape from .abc import Writer class HtmlWriter(Writer): def __init__(self): self._fragments = [] def text(self, text): self._fragments.append(_escape_html(text)) def start(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1}>".format(name, attribute_string)) def end(self, name): self._fragments.append("</{0}>".format(name)) def self_closing(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1} />".format(name, attribute_string)) def append(self, html): self._fragments.append(html) def as_string(self): return "".join(self._fragments) def _escape_html(text): return escape(text, {'"': "&quot;"}) def _generate_attribute_string(attributes): if attributes is None: return "" else: return "".join( ' {0}="{1}"'.format(key, _escape_html(attributes[key])) for key in sorted(attributes) )
Use xml.sax.saxutils.escape instead of deprecated cgi.escape
Use xml.sax.saxutils.escape instead of deprecated cgi.escape ``` /usr/local/lib/python3.6/dist-packages/mammoth/writers/html.py:34: DeprecationWarning: cgi.escape is deprecated, use html.escape instead return cgi.escape(text, quote=True) ```
Python
bsd-2-clause
mwilliamson/python-mammoth
from __future__ import unicode_literals + from xml.sax.saxutils import escape from .abc import Writer - - import cgi class HtmlWriter(Writer): def __init__(self): self._fragments = [] def text(self, text): self._fragments.append(_escape_html(text)) def start(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1}>".format(name, attribute_string)) def end(self, name): self._fragments.append("</{0}>".format(name)) def self_closing(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1} />".format(name, attribute_string)) def append(self, html): self._fragments.append(html) def as_string(self): return "".join(self._fragments) def _escape_html(text): - return cgi.escape(text, quote=True) + return escape(text, {'"': "&quot;"}) def _generate_attribute_string(attributes): if attributes is None: return "" else: return "".join( ' {0}="{1}"'.format(key, _escape_html(attributes[key])) for key in sorted(attributes) )
Use xml.sax.saxutils.escape instead of deprecated cgi.escape
## Code Before: from __future__ import unicode_literals from .abc import Writer import cgi class HtmlWriter(Writer): def __init__(self): self._fragments = [] def text(self, text): self._fragments.append(_escape_html(text)) def start(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1}>".format(name, attribute_string)) def end(self, name): self._fragments.append("</{0}>".format(name)) def self_closing(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1} />".format(name, attribute_string)) def append(self, html): self._fragments.append(html) def as_string(self): return "".join(self._fragments) def _escape_html(text): return cgi.escape(text, quote=True) def _generate_attribute_string(attributes): if attributes is None: return "" else: return "".join( ' {0}="{1}"'.format(key, _escape_html(attributes[key])) for key in sorted(attributes) ) ## Instruction: Use xml.sax.saxutils.escape instead of deprecated cgi.escape ## Code After: from __future__ import unicode_literals from xml.sax.saxutils import escape from .abc import Writer class HtmlWriter(Writer): def __init__(self): self._fragments = [] def text(self, text): self._fragments.append(_escape_html(text)) def start(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1}>".format(name, attribute_string)) def end(self, name): self._fragments.append("</{0}>".format(name)) def self_closing(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1} />".format(name, attribute_string)) def append(self, html): self._fragments.append(html) def as_string(self): return "".join(self._fragments) def _escape_html(text): return escape(text, {'"': "&quot;"}) def _generate_attribute_string(attributes): if attributes is None: return "" else: return "".join( ' {0}="{1}"'.format(key, _escape_html(attributes[key])) for key in sorted(attributes) )
from __future__ import unicode_literals + from xml.sax.saxutils import escape from .abc import Writer - - import cgi class HtmlWriter(Writer): def __init__(self): self._fragments = [] def text(self, text): self._fragments.append(_escape_html(text)) def start(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1}>".format(name, attribute_string)) def end(self, name): self._fragments.append("</{0}>".format(name)) def self_closing(self, name, attributes=None): attribute_string = _generate_attribute_string(attributes) self._fragments.append("<{0}{1} />".format(name, attribute_string)) def append(self, html): self._fragments.append(html) def as_string(self): return "".join(self._fragments) def _escape_html(text): - return cgi.escape(text, quote=True) + return escape(text, {'"': "&quot;"}) def _generate_attribute_string(attributes): if attributes is None: return "" else: return "".join( ' {0}="{1}"'.format(key, _escape_html(attributes[key])) for key in sorted(attributes) )
d88429d072f79c38d65ccaf3519495905f12f03f
calaccess_website/management/commands/updatedownloadswebsite.py
calaccess_website/management/commands/updatedownloadswebsite.py
import logging from django.core.management import call_command from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand logger = logging.getLogger(__name__) class Command(updatecommand): """ Update to the latest CAL-ACCESS snapshot and bake static website pages. """ help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages' def add_arguments(self, parser): """ Adds custom arguments specific to this command. """ super(Command, self).add_arguments(parser) parser.add_argument( "--publish", action="store_true", dest="publish", default=False, help="Publish baked content" ) def handle(self, *args, **options): """ Make it happen. """ super(Command, self).handle(*args, **options) self.header('Creating latest file links') call_command('createlatestlinks') self.header('Baking downloads-website content') call_command('build') if options['publish']: self.header('Publishing baked content to S3 bucket.') call_command('publish') self.success("Done!")
import logging from django.core.management import call_command from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand logger = logging.getLogger(__name__) class Command(updatecommand): """ Update to the latest CAL-ACCESS snapshot and bake static website pages. """ help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages' def add_arguments(self, parser): """ Adds custom arguments specific to this command. """ super(Command, self).add_arguments(parser) parser.add_argument( "--publish", action="store_true", dest="publish", default=False, help="Publish baked content" ) def handle(self, *args, **options): """ Make it happen. """ super(Command, self).handle(*args, **options) call_command('processcalaccessdata') self.header('Creating latest file links') call_command('createlatestlinks') self.header('Baking downloads-website content') call_command('build') if options['publish']: self.header('Publishing baked content to S3 bucket.') call_command('publish') self.success("Done!")
Add processing cmd to update routine
Add processing cmd to update routine
Python
mit
california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website
import logging from django.core.management import call_command from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand logger = logging.getLogger(__name__) class Command(updatecommand): """ Update to the latest CAL-ACCESS snapshot and bake static website pages. """ help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages' def add_arguments(self, parser): """ Adds custom arguments specific to this command. """ super(Command, self).add_arguments(parser) parser.add_argument( "--publish", action="store_true", dest="publish", default=False, help="Publish baked content" ) def handle(self, *args, **options): """ Make it happen. """ super(Command, self).handle(*args, **options) + call_command('processcalaccessdata') self.header('Creating latest file links') call_command('createlatestlinks') self.header('Baking downloads-website content') call_command('build') if options['publish']: self.header('Publishing baked content to S3 bucket.') call_command('publish') self.success("Done!")
Add processing cmd to update routine
## Code Before: import logging from django.core.management import call_command from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand logger = logging.getLogger(__name__) class Command(updatecommand): """ Update to the latest CAL-ACCESS snapshot and bake static website pages. """ help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages' def add_arguments(self, parser): """ Adds custom arguments specific to this command. """ super(Command, self).add_arguments(parser) parser.add_argument( "--publish", action="store_true", dest="publish", default=False, help="Publish baked content" ) def handle(self, *args, **options): """ Make it happen. """ super(Command, self).handle(*args, **options) self.header('Creating latest file links') call_command('createlatestlinks') self.header('Baking downloads-website content') call_command('build') if options['publish']: self.header('Publishing baked content to S3 bucket.') call_command('publish') self.success("Done!") ## Instruction: Add processing cmd to update routine ## Code After: import logging from django.core.management import call_command from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand logger = logging.getLogger(__name__) class Command(updatecommand): """ Update to the latest CAL-ACCESS snapshot and bake static website pages. """ help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages' def add_arguments(self, parser): """ Adds custom arguments specific to this command. """ super(Command, self).add_arguments(parser) parser.add_argument( "--publish", action="store_true", dest="publish", default=False, help="Publish baked content" ) def handle(self, *args, **options): """ Make it happen. """ super(Command, self).handle(*args, **options) call_command('processcalaccessdata') self.header('Creating latest file links') call_command('createlatestlinks') self.header('Baking downloads-website content') call_command('build') if options['publish']: self.header('Publishing baked content to S3 bucket.') call_command('publish') self.success("Done!")
import logging from django.core.management import call_command from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand logger = logging.getLogger(__name__) class Command(updatecommand): """ Update to the latest CAL-ACCESS snapshot and bake static website pages. """ help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages' def add_arguments(self, parser): """ Adds custom arguments specific to this command. """ super(Command, self).add_arguments(parser) parser.add_argument( "--publish", action="store_true", dest="publish", default=False, help="Publish baked content" ) def handle(self, *args, **options): """ Make it happen. """ super(Command, self).handle(*args, **options) + call_command('processcalaccessdata') self.header('Creating latest file links') call_command('createlatestlinks') self.header('Baking downloads-website content') call_command('build') if options['publish']: self.header('Publishing baked content to S3 bucket.') call_command('publish') self.success("Done!")
7a00ff49799afc50da74a748d07c52fef57ebc84
setup.py
setup.py
import tungsten from distutils.core import setup setup( name='Tungsten', version=tungsten.__version__, author='Seena Burns', packages={'tungsten': 'tungsten'}, license=open('LICENSE.txt').read(), description='Wolfram Alpha API built for Python.', long_description=open('README.md').read(), install_requires=[ "requests", ], classifiers=( 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Natural Language :: English', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7' ), )
import tungsten from distutils.core import setup setup( name='Tungsten', version=tungsten.__version__, author='Seena Burns', author_email='[email protected]', url='https://github.com/seenaburns/Tungsten', packages={'tungsten': 'tungsten'}, license=open('LICENSE.txt').read(), description='Wolfram Alpha API built for Python.', long_description=open('README.md').read(), install_requires=[ "requests", ], classifiers=( 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Natural Language :: English', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7' ), )
Add url / author email for PyPI regs
Add url / author email for PyPI regs
Python
bsd-3-clause
seenaburns/Tungsten
import tungsten from distutils.core import setup setup( name='Tungsten', version=tungsten.__version__, author='Seena Burns', + author_email='[email protected]', + url='https://github.com/seenaburns/Tungsten', packages={'tungsten': 'tungsten'}, license=open('LICENSE.txt').read(), description='Wolfram Alpha API built for Python.', long_description=open('README.md').read(), install_requires=[ "requests", ], classifiers=( 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Natural Language :: English', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7' ), )
Add url / author email for PyPI regs
## Code Before: import tungsten from distutils.core import setup setup( name='Tungsten', version=tungsten.__version__, author='Seena Burns', packages={'tungsten': 'tungsten'}, license=open('LICENSE.txt').read(), description='Wolfram Alpha API built for Python.', long_description=open('README.md').read(), install_requires=[ "requests", ], classifiers=( 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Natural Language :: English', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7' ), ) ## Instruction: Add url / author email for PyPI regs ## Code After: import tungsten from distutils.core import setup setup( name='Tungsten', version=tungsten.__version__, author='Seena Burns', author_email='[email protected]', url='https://github.com/seenaburns/Tungsten', packages={'tungsten': 'tungsten'}, license=open('LICENSE.txt').read(), description='Wolfram Alpha API built for Python.', long_description=open('README.md').read(), install_requires=[ "requests", ], classifiers=( 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Natural Language :: English', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7' ), )
import tungsten from distutils.core import setup setup( name='Tungsten', version=tungsten.__version__, author='Seena Burns', + author_email='[email protected]', + url='https://github.com/seenaburns/Tungsten', packages={'tungsten': 'tungsten'}, license=open('LICENSE.txt').read(), description='Wolfram Alpha API built for Python.', long_description=open('README.md').read(), install_requires=[ "requests", ], classifiers=( 'Development Status :: 3 - Alpha', 'Intended Audience :: Developers', 'Natural Language :: English', 'License :: OSI Approved :: BSD License', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7' ), )
446d36cbbf79083b9d41ea5b152c5a845560eb4b
whats_fresh/whats_fresh_api/tests/views/test_stories.py
whats_fresh/whats_fresh_api/tests/views/test_stories.py
from django.test import TestCase from django.test.client import Client from django.core.urlresolvers import reverse from whats_fresh_api.models import * from django.contrib.gis.db import models import json class StoriesTestCase(TestCase): fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json'] def setUp(self): self.expected_json = """ { "error": { "error_status": false, "error_name": null, "error_text": null, "error_level": null }, "story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..." }""" def test_url_endpoint(self): url = reverse('story-details', kwargs={'id': '1'}) self.assertEqual(url, '/stories/1') def test_json_equals(self): c = Client() response = c.get(reverse('story-details', kwargs={'id': '1'})).content parsed_answer = json.loads(response) expected_answer = json.loads(self.expected_json) self.assertTrue(parsed_answer == expected_answer)
from django.test import TestCase from django.test.client import Client from django.core.urlresolvers import reverse from whats_fresh_api.models import * from django.contrib.gis.db import models import json class StoriesTestCase(TestCase): fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json'] def setUp(self): self.expected_json = """" { <<<<<<< HEAD "error": { "error_status": false, "error_name": null, "error_text": null, "error_level": null }, "story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..." }""" def test_url_endpoint(self): url = reverse('story-details', kwargs={'id': '1'}) self.assertEqual(url, '/stories/1') def test_json_equals(self): c = Client() response = c.get(reverse('story-details', kwargs={'id': '1'})).content parsed_answer = json.loads(response) expected_answer = json.loads(self.expected_json) self.assertTrue(parsed_answer == expected_answer)
Add error field to expected JSON
Add error field to expected JSON
Python
apache-2.0
iCHAIT/whats-fresh-api,iCHAIT/whats-fresh-api,osu-cass/whats-fresh-api,iCHAIT/whats-fresh-api,iCHAIT/whats-fresh-api,osu-cass/whats-fresh-api,osu-cass/whats-fresh-api,osu-cass/whats-fresh-api
from django.test import TestCase from django.test.client import Client from django.core.urlresolvers import reverse from whats_fresh_api.models import * from django.contrib.gis.db import models import json class StoriesTestCase(TestCase): fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json'] def setUp(self): - self.expected_json = """ + self.expected_json = """" { + <<<<<<< HEAD "error": { "error_status": false, "error_name": null, "error_text": null, "error_level": null }, "story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..." }""" + def test_url_endpoint(self): url = reverse('story-details', kwargs={'id': '1'}) self.assertEqual(url, '/stories/1') def test_json_equals(self): c = Client() response = c.get(reverse('story-details', kwargs={'id': '1'})).content parsed_answer = json.loads(response) expected_answer = json.loads(self.expected_json) self.assertTrue(parsed_answer == expected_answer)
Add error field to expected JSON
## Code Before: from django.test import TestCase from django.test.client import Client from django.core.urlresolvers import reverse from whats_fresh_api.models import * from django.contrib.gis.db import models import json class StoriesTestCase(TestCase): fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json'] def setUp(self): self.expected_json = """ { "error": { "error_status": false, "error_name": null, "error_text": null, "error_level": null }, "story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..." }""" def test_url_endpoint(self): url = reverse('story-details', kwargs={'id': '1'}) self.assertEqual(url, '/stories/1') def test_json_equals(self): c = Client() response = c.get(reverse('story-details', kwargs={'id': '1'})).content parsed_answer = json.loads(response) expected_answer = json.loads(self.expected_json) self.assertTrue(parsed_answer == expected_answer) ## Instruction: Add error field to expected JSON ## Code After: from django.test import TestCase from django.test.client import Client from django.core.urlresolvers import reverse from whats_fresh_api.models import * from django.contrib.gis.db import models import json class StoriesTestCase(TestCase): fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json'] def setUp(self): self.expected_json = """" { <<<<<<< HEAD "error": { "error_status": false, "error_name": null, "error_text": null, "error_level": null }, "story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..." }""" def test_url_endpoint(self): url = reverse('story-details', kwargs={'id': '1'}) self.assertEqual(url, '/stories/1') def test_json_equals(self): c = Client() response = c.get(reverse('story-details', kwargs={'id': '1'})).content parsed_answer = json.loads(response) expected_answer = json.loads(self.expected_json) self.assertTrue(parsed_answer == expected_answer)
from django.test import TestCase from django.test.client import Client from django.core.urlresolvers import reverse from whats_fresh_api.models import * from django.contrib.gis.db import models import json class StoriesTestCase(TestCase): fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json'] def setUp(self): - self.expected_json = """ + self.expected_json = """" ? + { + <<<<<<< HEAD "error": { "error_status": false, "error_name": null, "error_text": null, "error_level": null }, "story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..." }""" + def test_url_endpoint(self): url = reverse('story-details', kwargs={'id': '1'}) self.assertEqual(url, '/stories/1') def test_json_equals(self): c = Client() response = c.get(reverse('story-details', kwargs={'id': '1'})).content parsed_answer = json.loads(response) expected_answer = json.loads(self.expected_json) self.assertTrue(parsed_answer == expected_answer)
fba4fdf426b0a29ca06deb67587c2bd804adb017
tbgxmlutils/xmlutils.py
tbgxmlutils/xmlutils.py
from xml.dom import minidom import xml.etree.ElementTree as ET import xmltodict def add(k, parent=None, txt=None, attrs=None): if parent is None: handle = ET.Element(k) else: handle = ET.SubElement(parent, k) if txt: handle.text = unicode(txt) try: for k, v in attrs.iteritems(): handle.attrib[k] = v except AttributeError: pass return handle def etree2xml(e, encoding='UTF-8'): return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e) def pretty(xml=None, fn=None): if fn is not None: xml = minidom.parse(fn) elif not isinstance(xml, minidom.Document): xml = minidom.parseString(xml) return xml.toprettyxml(indent=' ') def xml_fn_to_json(fn): fh = open(fn, 'r') json = xmltodict.parse(fh.read()) return json
from xml.dom import minidom import lxml.etree as ET import xmltodict def add(k, parent=None, txt=None, attrs=None): if parent is None: handle = ET.Element(k) else: handle = ET.SubElement(parent, k) if txt: handle.text = unicode(txt) try: for k, v in attrs.iteritems(): handle.attrib[k] = v except AttributeError: pass return handle def etree2xml(e, encoding='UTF-8'): return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e) def pretty(xml=None, fn=None): if fn is not None: xml = minidom.parse(fn) elif not isinstance(xml, minidom.Document): xml = minidom.parseString(xml) return xml.toprettyxml(indent=' ') def xml_fn_to_json(fn): fh = open(fn, 'r') json = xmltodict.parse(fh.read()) return json
Use lxml instead of elementtree.
Use lxml instead of elementtree.
Python
mit
Schwarzschild/TBGXMLUtils
from xml.dom import minidom - import xml.etree.ElementTree as ET + import lxml.etree as ET import xmltodict def add(k, parent=None, txt=None, attrs=None): if parent is None: handle = ET.Element(k) else: handle = ET.SubElement(parent, k) if txt: handle.text = unicode(txt) try: for k, v in attrs.iteritems(): handle.attrib[k] = v except AttributeError: pass return handle def etree2xml(e, encoding='UTF-8'): return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e) def pretty(xml=None, fn=None): if fn is not None: xml = minidom.parse(fn) elif not isinstance(xml, minidom.Document): xml = minidom.parseString(xml) return xml.toprettyxml(indent=' ') def xml_fn_to_json(fn): fh = open(fn, 'r') json = xmltodict.parse(fh.read()) return json
Use lxml instead of elementtree.
## Code Before: from xml.dom import minidom import xml.etree.ElementTree as ET import xmltodict def add(k, parent=None, txt=None, attrs=None): if parent is None: handle = ET.Element(k) else: handle = ET.SubElement(parent, k) if txt: handle.text = unicode(txt) try: for k, v in attrs.iteritems(): handle.attrib[k] = v except AttributeError: pass return handle def etree2xml(e, encoding='UTF-8'): return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e) def pretty(xml=None, fn=None): if fn is not None: xml = minidom.parse(fn) elif not isinstance(xml, minidom.Document): xml = minidom.parseString(xml) return xml.toprettyxml(indent=' ') def xml_fn_to_json(fn): fh = open(fn, 'r') json = xmltodict.parse(fh.read()) return json ## Instruction: Use lxml instead of elementtree. ## Code After: from xml.dom import minidom import lxml.etree as ET import xmltodict def add(k, parent=None, txt=None, attrs=None): if parent is None: handle = ET.Element(k) else: handle = ET.SubElement(parent, k) if txt: handle.text = unicode(txt) try: for k, v in attrs.iteritems(): handle.attrib[k] = v except AttributeError: pass return handle def etree2xml(e, encoding='UTF-8'): return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e) def pretty(xml=None, fn=None): if fn is not None: xml = minidom.parse(fn) elif not isinstance(xml, minidom.Document): xml = minidom.parseString(xml) return xml.toprettyxml(indent=' ') def xml_fn_to_json(fn): fh = open(fn, 'r') json = xmltodict.parse(fh.read()) return json
from xml.dom import minidom - import xml.etree.ElementTree as ET ? ------------ + import lxml.etree as ET ? + import xmltodict def add(k, parent=None, txt=None, attrs=None): if parent is None: handle = ET.Element(k) else: handle = ET.SubElement(parent, k) if txt: handle.text = unicode(txt) try: for k, v in attrs.iteritems(): handle.attrib[k] = v except AttributeError: pass return handle def etree2xml(e, encoding='UTF-8'): return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e) def pretty(xml=None, fn=None): if fn is not None: xml = minidom.parse(fn) elif not isinstance(xml, minidom.Document): xml = minidom.parseString(xml) return xml.toprettyxml(indent=' ') def xml_fn_to_json(fn): fh = open(fn, 'r') json = xmltodict.parse(fh.read()) return json
ab5aac0c9b0e075901c4cd8dd5d134e79f0e0110
brasileirao/spiders/results_spider.py
brasileirao/spiders/results_spider.py
import scrapy import scrapy.selector from brasileirao.items import BrasileiraoItem import hashlib class ResultsSpider(scrapy.Spider): name = "results" start_urls = [ 'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/', ] def parse(self, response): actual_round = 0 for rodada in response.css('.rodadas .confrontos li'): actual_round += 1 for game in rodada.css(".confronto"): home_team = game.css(".partida .time1") away_team = game.css(".partida .time2") item = BrasileiraoItem() item['rodada'] = actual_round item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8') item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8') item['home_score'] = home_team.css(".gols::text").extract_first() item['away_score'] = away_team.css(".gols::text").extract_first() item['date'] = game.css(".info-partida time::attr(datetime)").extract_first() id = item['home_team'] + item['away_team'] item['id'] = hashlib.md5(id).hexdigest() yield item
import scrapy import scrapy.selector from brasileirao.items import BrasileiraoItem import hashlib class ResultsSpider(scrapy.Spider): name = "results" start_urls = [ 'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/', ] def parse(self, response): actual_round = 0 for rodada in response.css('.rodadas .confrontos li'): actual_round += 1 for game in rodada.css(".confronto"): home_team = game.css(".partida .time1") away_team = game.css(".partida .time2") item = BrasileiraoItem() item['rodada'] = actual_round item['home_team'] = home_team.css("abbr::attr(title)").extract_first() item['away_team'] = away_team.css("abbr::attr(title)").extract_first() item['home_score'] = home_team.css(".gols::text").extract_first() item['away_score'] = away_team.css(".gols::text").extract_first() item['date'] = game.css(".info-partida time::attr(datetime)").extract_first() id = item['home_team'] + item['away_team'] item['id'] = hashlib.md5(id).hexdigest() yield item
Set utf-8 as default encoding.
Set utf-8 as default encoding.
Python
mit
pghilardi/live-football-client
+ import scrapy import scrapy.selector from brasileirao.items import BrasileiraoItem import hashlib class ResultsSpider(scrapy.Spider): name = "results" start_urls = [ 'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/', ] def parse(self, response): actual_round = 0 for rodada in response.css('.rodadas .confrontos li'): actual_round += 1 for game in rodada.css(".confronto"): home_team = game.css(".partida .time1") away_team = game.css(".partida .time2") item = BrasileiraoItem() item['rodada'] = actual_round - item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8') + item['home_team'] = home_team.css("abbr::attr(title)").extract_first() - item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8') + item['away_team'] = away_team.css("abbr::attr(title)").extract_first() item['home_score'] = home_team.css(".gols::text").extract_first() item['away_score'] = away_team.css(".gols::text").extract_first() item['date'] = game.css(".info-partida time::attr(datetime)").extract_first() id = item['home_team'] + item['away_team'] item['id'] = hashlib.md5(id).hexdigest() yield item
Set utf-8 as default encoding.
## Code Before: import scrapy import scrapy.selector from brasileirao.items import BrasileiraoItem import hashlib class ResultsSpider(scrapy.Spider): name = "results" start_urls = [ 'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/', ] def parse(self, response): actual_round = 0 for rodada in response.css('.rodadas .confrontos li'): actual_round += 1 for game in rodada.css(".confronto"): home_team = game.css(".partida .time1") away_team = game.css(".partida .time2") item = BrasileiraoItem() item['rodada'] = actual_round item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8') item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8') item['home_score'] = home_team.css(".gols::text").extract_first() item['away_score'] = away_team.css(".gols::text").extract_first() item['date'] = game.css(".info-partida time::attr(datetime)").extract_first() id = item['home_team'] + item['away_team'] item['id'] = hashlib.md5(id).hexdigest() yield item ## Instruction: Set utf-8 as default encoding. ## Code After: import scrapy import scrapy.selector from brasileirao.items import BrasileiraoItem import hashlib class ResultsSpider(scrapy.Spider): name = "results" start_urls = [ 'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/', ] def parse(self, response): actual_round = 0 for rodada in response.css('.rodadas .confrontos li'): actual_round += 1 for game in rodada.css(".confronto"): home_team = game.css(".partida .time1") away_team = game.css(".partida .time2") item = BrasileiraoItem() item['rodada'] = actual_round item['home_team'] = home_team.css("abbr::attr(title)").extract_first() item['away_team'] = away_team.css("abbr::attr(title)").extract_first() item['home_score'] = home_team.css(".gols::text").extract_first() item['away_score'] = away_team.css(".gols::text").extract_first() item['date'] = game.css(".info-partida time::attr(datetime)").extract_first() id = item['home_team'] + item['away_team'] item['id'] = hashlib.md5(id).hexdigest() yield item
+ import scrapy import scrapy.selector from brasileirao.items import BrasileiraoItem import hashlib class ResultsSpider(scrapy.Spider): name = "results" start_urls = [ 'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/', ] def parse(self, response): actual_round = 0 for rodada in response.css('.rodadas .confrontos li'): actual_round += 1 for game in rodada.css(".confronto"): home_team = game.css(".partida .time1") away_team = game.css(".partida .time2") item = BrasileiraoItem() item['rodada'] = actual_round - item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8') ? --------------- + item['home_team'] = home_team.css("abbr::attr(title)").extract_first() - item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8') ? --------------- + item['away_team'] = away_team.css("abbr::attr(title)").extract_first() item['home_score'] = home_team.css(".gols::text").extract_first() item['away_score'] = away_team.css(".gols::text").extract_first() item['date'] = game.css(".info-partida time::attr(datetime)").extract_first() id = item['home_team'] + item['away_team'] item['id'] = hashlib.md5(id).hexdigest() yield item
55e0c877dbe1a073534c9cf445ffe58715160b8e
metadata/RomsLite/hooks/post-stage.py
metadata/RomsLite/hooks/post-stage.py
import os import shutil from wmt.utils.hook import find_simulation_input_file def execute(env): """Perform post-stage tasks for running a component. Parameters ---------- env : dict A dict of component parameter values from WMT. """ for name in ('river_forcing_file', 'waves_forcing_file'): src = find_simulation_input_file(env[name]) shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
import os import shutil from wmt.utils.hook import find_simulation_input_file _DEFAULT_FILES = { 'river_forcing_file': 'river.nc', 'waves_forcing_file': 'waves.nc', } def execute(env): """Perform post-stage tasks for running a component. Parameters ---------- env : dict A dict of component parameter values from WMT. """ for name in _DEFAULT_FILES: if env[name] != _DEFAULT_FILES[name]: try: os.remove(os.path.join(os.curdir, 'Forcing', env[name])) except OSError: pass src = find_simulation_input_file(env[name]) shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
Remove default forcing files if not being used.
Remove default forcing files if not being used.
Python
mit
csdms/wmt-metadata
import os import shutil from wmt.utils.hook import find_simulation_input_file + + _DEFAULT_FILES = { + 'river_forcing_file': 'river.nc', + 'waves_forcing_file': 'waves.nc', + } def execute(env): """Perform post-stage tasks for running a component. Parameters ---------- env : dict A dict of component parameter values from WMT. """ - for name in ('river_forcing_file', 'waves_forcing_file'): + for name in _DEFAULT_FILES: + if env[name] != _DEFAULT_FILES[name]: + try: + os.remove(os.path.join(os.curdir, 'Forcing', env[name])) + except OSError: + pass + src = find_simulation_input_file(env[name]) shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
Remove default forcing files if not being used.
## Code Before: import os import shutil from wmt.utils.hook import find_simulation_input_file def execute(env): """Perform post-stage tasks for running a component. Parameters ---------- env : dict A dict of component parameter values from WMT. """ for name in ('river_forcing_file', 'waves_forcing_file'): src = find_simulation_input_file(env[name]) shutil.copy(src, os.path.join(os.curdir, 'Forcing')) ## Instruction: Remove default forcing files if not being used. ## Code After: import os import shutil from wmt.utils.hook import find_simulation_input_file _DEFAULT_FILES = { 'river_forcing_file': 'river.nc', 'waves_forcing_file': 'waves.nc', } def execute(env): """Perform post-stage tasks for running a component. Parameters ---------- env : dict A dict of component parameter values from WMT. """ for name in _DEFAULT_FILES: if env[name] != _DEFAULT_FILES[name]: try: os.remove(os.path.join(os.curdir, 'Forcing', env[name])) except OSError: pass src = find_simulation_input_file(env[name]) shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
import os import shutil from wmt.utils.hook import find_simulation_input_file + + _DEFAULT_FILES = { + 'river_forcing_file': 'river.nc', + 'waves_forcing_file': 'waves.nc', + } def execute(env): """Perform post-stage tasks for running a component. Parameters ---------- env : dict A dict of component parameter values from WMT. """ - for name in ('river_forcing_file', 'waves_forcing_file'): + for name in _DEFAULT_FILES: + if env[name] != _DEFAULT_FILES[name]: + try: + os.remove(os.path.join(os.curdir, 'Forcing', env[name])) + except OSError: + pass + src = find_simulation_input_file(env[name]) shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
db19dfa17261c3d04de0202b2809ba8abb70326b
tests/unit/test_moxstubout.py
tests/unit/test_moxstubout.py
from oslotest import base from oslotest import moxstubout class TestMoxStubout(base.BaseTestCase): def _stubable(self): pass def test_basic_stubout(self): f = self.useFixture(moxstubout.MoxStubout()) before = TestMoxStubout._stubable f.mox.StubOutWithMock(TestMoxStubout, '_stubable') after = TestMoxStubout._stubable self.assertNotEqual(before, after) f.cleanUp() after2 = TestMoxStubout._stubable self.assertEqual(before, after2)
from oslotest import base from oslotest import moxstubout class TestMoxStubout(base.BaseTestCase): def _stubable(self): pass def test_basic_stubout(self): f = self.useFixture(moxstubout.MoxStubout()) before = TestMoxStubout._stubable f.mox.StubOutWithMock(TestMoxStubout, '_stubable') after = TestMoxStubout._stubable self.assertNotEqual(before, after) f.cleanUp() after2 = TestMoxStubout._stubable self.assertEqual(before, after2) f._clear_cleanups()
Fix build break with Fixtures 1.3
Fix build break with Fixtures 1.3 Our explicit call to cleanUp messes things up in latest fixture, so we need to call _clear_cleanups to stop the test from breaking Change-Id: I8ce2309a94736b47fb347f37ab4027857e19c8a8
Python
apache-2.0
openstack/oslotest,openstack/oslotest
from oslotest import base from oslotest import moxstubout class TestMoxStubout(base.BaseTestCase): def _stubable(self): pass def test_basic_stubout(self): f = self.useFixture(moxstubout.MoxStubout()) before = TestMoxStubout._stubable f.mox.StubOutWithMock(TestMoxStubout, '_stubable') after = TestMoxStubout._stubable self.assertNotEqual(before, after) f.cleanUp() after2 = TestMoxStubout._stubable self.assertEqual(before, after2) + f._clear_cleanups()
Fix build break with Fixtures 1.3
## Code Before: from oslotest import base from oslotest import moxstubout class TestMoxStubout(base.BaseTestCase): def _stubable(self): pass def test_basic_stubout(self): f = self.useFixture(moxstubout.MoxStubout()) before = TestMoxStubout._stubable f.mox.StubOutWithMock(TestMoxStubout, '_stubable') after = TestMoxStubout._stubable self.assertNotEqual(before, after) f.cleanUp() after2 = TestMoxStubout._stubable self.assertEqual(before, after2) ## Instruction: Fix build break with Fixtures 1.3 ## Code After: from oslotest import base from oslotest import moxstubout class TestMoxStubout(base.BaseTestCase): def _stubable(self): pass def test_basic_stubout(self): f = self.useFixture(moxstubout.MoxStubout()) before = TestMoxStubout._stubable f.mox.StubOutWithMock(TestMoxStubout, '_stubable') after = TestMoxStubout._stubable self.assertNotEqual(before, after) f.cleanUp() after2 = TestMoxStubout._stubable self.assertEqual(before, after2) f._clear_cleanups()
from oslotest import base from oslotest import moxstubout class TestMoxStubout(base.BaseTestCase): def _stubable(self): pass def test_basic_stubout(self): f = self.useFixture(moxstubout.MoxStubout()) before = TestMoxStubout._stubable f.mox.StubOutWithMock(TestMoxStubout, '_stubable') after = TestMoxStubout._stubable self.assertNotEqual(before, after) f.cleanUp() after2 = TestMoxStubout._stubable self.assertEqual(before, after2) + f._clear_cleanups()
1f958dc4439fbe435b1d0381d15860708f1f9745
constance/__init__.py
constance/__init__.py
from .base import Config __version__ = '1.0a1' try: from django.apps import AppConfig # noqa except ImportError: config = Config() else: default_app_config = 'constance.apps.ConstanceConfig'
from .base import Config from django.utils.functional import SimpleLazyObject __version__ = '1.0a1' try: from django.apps import AppConfig # noqa except ImportError: config = SimpleLazyObject(Config) else: default_app_config = 'constance.apps.ConstanceConfig'
Make the config object lazy for old Djangos.
Make the config object lazy for old Djangos. This should prevent import time side effects from instantiating the config object directly there.
Python
bsd-3-clause
gmflanagan/waterboy,vinnyrose/django-constance,jerzyk/django-constance,metalpriest/django-constance,thorgate/django-constance,jmerdich/django-constance,jazzband/django-constance,jonzlin95/django-constance,winzard/django-constance,metalpriest/django-constance,jazzband/django-constance,django-leonardo/django-constance,pombredanne/django-constance,APSL/django-constance,michaelkuty/django-constance,dmugtasimov/django-constance,thorgate/django-constance,Andrey86/django-constance,askabelin/django-constance,pombredanne/django-constance,jezdez/django-constance,Andrey86/django-constance,michaelkuty/django-constance,APSL/django-constance,jezdez/django-constance,jerzyk/django-constance,askabelin/django-constance,django-leonardo/django-constance,jmerdich/django-constance,jonzlin95/django-constance,jazzband/django-constance,dmugtasimov/django-constance,winzard/django-constance,vinnyrose/django-constance
from .base import Config + from django.utils.functional import SimpleLazyObject __version__ = '1.0a1' + try: from django.apps import AppConfig # noqa except ImportError: - config = Config() + config = SimpleLazyObject(Config) else: default_app_config = 'constance.apps.ConstanceConfig'
Make the config object lazy for old Djangos.
## Code Before: from .base import Config __version__ = '1.0a1' try: from django.apps import AppConfig # noqa except ImportError: config = Config() else: default_app_config = 'constance.apps.ConstanceConfig' ## Instruction: Make the config object lazy for old Djangos. ## Code After: from .base import Config from django.utils.functional import SimpleLazyObject __version__ = '1.0a1' try: from django.apps import AppConfig # noqa except ImportError: config = SimpleLazyObject(Config) else: default_app_config = 'constance.apps.ConstanceConfig'
from .base import Config + from django.utils.functional import SimpleLazyObject __version__ = '1.0a1' + try: from django.apps import AppConfig # noqa except ImportError: - config = Config() + config = SimpleLazyObject(Config) else: default_app_config = 'constance.apps.ConstanceConfig'
ca74738e9241230fd0cc843aa9b76f67494d02eb
python/intermediate/create_inter_python_data.py
python/intermediate/create_inter_python_data.py
"""Create the data for the Software Carpentry Intermediate Python lectures""" import numpy as np import pandas as pd np.random.seed(26) years = np.arange(1960, 2011) temps = np.random.uniform(70, 90, len(years)) rainfalls = np.random.uniform(100, 300, len(years)) noise = 2 * np.random.randn(len(years)) mosquitos = 0.5 * temps + 0.7 * rainfalls + noise data = zip(years, temps, rainfalls, mosquitos) df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f') df_short = df[-10:] df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f')
"""Create the data for the Software Carpentry Intermediate Python lectures""" import numpy as np import pandas as pd np.random.seed(26) datasets = {'A1': [0, 0.5, 0.7, 10], 'A2': [0, 0.5, 0.7, 50], 'A3': [0, 0.5, 0.3, 50], 'B1': [3, 0.7, 0.2, 50], 'B2': [3, 0.7, 0.7, 50]} def make_data(intercept, tempslope, rainfallslope, numyears): years = np.arange(2010 - numyears, 2011) temps = np.random.uniform(70, 90, len(years)) rainfalls = np.random.uniform(100, 300, len(years)) noise = 2 * np.random.randn(len(years)) mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise return zip(years, temps, rainfalls, mosquitos) def export_data(data, filename): df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) df.to_csv(filename, index=False, float_format='%.0f') for site in datasets: data = make_data(*datasets[site]) if site == 'A1': #create a shorter dataset for first example data = data[-10:] export_data(data, '%s_mosquito_data.csv' % site)
Allow creation of multiple example data files for Inter Python
Allow creation of multiple example data files for Inter Python Generalizes the script for creating data files to allow for the easy generation of larger numbers of data files.
Python
bsd-2-clause
selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest
"""Create the data for the Software Carpentry Intermediate Python lectures""" import numpy as np import pandas as pd np.random.seed(26) - years = np.arange(1960, 2011) - temps = np.random.uniform(70, 90, len(years)) - rainfalls = np.random.uniform(100, 300, len(years)) - noise = 2 * np.random.randn(len(years)) - mosquitos = 0.5 * temps + 0.7 * rainfalls + noise + datasets = {'A1': [0, 0.5, 0.7, 10], + 'A2': [0, 0.5, 0.7, 50], + 'A3': [0, 0.5, 0.3, 50], + 'B1': [3, 0.7, 0.2, 50], + 'B2': [3, 0.7, 0.7, 50]} + def make_data(intercept, tempslope, rainfallslope, numyears): + years = np.arange(2010 - numyears, 2011) + temps = np.random.uniform(70, 90, len(years)) + rainfalls = np.random.uniform(100, 300, len(years)) + noise = 2 * np.random.randn(len(years)) + mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise - data = zip(years, temps, rainfalls, mosquitos) + return zip(years, temps, rainfalls, mosquitos) - df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) - df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f') - df_short = df[-10:] - df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f') + def export_data(data, filename): + df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) + df.to_csv(filename, index=False, float_format='%.0f') + + for site in datasets: + data = make_data(*datasets[site]) + if site == 'A1': + #create a shorter dataset for first example + data = data[-10:] + export_data(data, '%s_mosquito_data.csv' % site) +
Allow creation of multiple example data files for Inter Python
## Code Before: """Create the data for the Software Carpentry Intermediate Python lectures""" import numpy as np import pandas as pd np.random.seed(26) years = np.arange(1960, 2011) temps = np.random.uniform(70, 90, len(years)) rainfalls = np.random.uniform(100, 300, len(years)) noise = 2 * np.random.randn(len(years)) mosquitos = 0.5 * temps + 0.7 * rainfalls + noise data = zip(years, temps, rainfalls, mosquitos) df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f') df_short = df[-10:] df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f') ## Instruction: Allow creation of multiple example data files for Inter Python ## Code After: """Create the data for the Software Carpentry Intermediate Python lectures""" import numpy as np import pandas as pd np.random.seed(26) datasets = {'A1': [0, 0.5, 0.7, 10], 'A2': [0, 0.5, 0.7, 50], 'A3': [0, 0.5, 0.3, 50], 'B1': [3, 0.7, 0.2, 50], 'B2': [3, 0.7, 0.7, 50]} def make_data(intercept, tempslope, rainfallslope, numyears): years = np.arange(2010 - numyears, 2011) temps = np.random.uniform(70, 90, len(years)) rainfalls = np.random.uniform(100, 300, len(years)) noise = 2 * np.random.randn(len(years)) mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise return zip(years, temps, rainfalls, mosquitos) def export_data(data, filename): df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) df.to_csv(filename, index=False, float_format='%.0f') for site in datasets: data = make_data(*datasets[site]) if site == 'A1': #create a shorter dataset for first example data = data[-10:] export_data(data, '%s_mosquito_data.csv' % site)
"""Create the data for the Software Carpentry Intermediate Python lectures""" import numpy as np import pandas as pd np.random.seed(26) - years = np.arange(1960, 2011) - temps = np.random.uniform(70, 90, len(years)) - rainfalls = np.random.uniform(100, 300, len(years)) - noise = 2 * np.random.randn(len(years)) - mosquitos = 0.5 * temps + 0.7 * rainfalls + noise + datasets = {'A1': [0, 0.5, 0.7, 10], + 'A2': [0, 0.5, 0.7, 50], + 'A3': [0, 0.5, 0.3, 50], + 'B1': [3, 0.7, 0.2, 50], + 'B2': [3, 0.7, 0.7, 50]} + def make_data(intercept, tempslope, rainfallslope, numyears): + years = np.arange(2010 - numyears, 2011) + temps = np.random.uniform(70, 90, len(years)) + rainfalls = np.random.uniform(100, 300, len(years)) + noise = 2 * np.random.randn(len(years)) + mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise - data = zip(years, temps, rainfalls, mosquitos) ? ^^ ^^^ + return zip(years, temps, rainfalls, mosquitos) ? ^^^^^^ ^^^ + + def export_data(data, filename): - df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) + df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos']) ? ++++ - df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f') - df_short = df[-10:] - df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f') + df.to_csv(filename, index=False, float_format='%.0f') + + for site in datasets: + data = make_data(*datasets[site]) + if site == 'A1': + #create a shorter dataset for first example + data = data[-10:] + export_data(data, '%s_mosquito_data.csv' % site)
e3a95d00444fb981d7aaf6d3beffca8796a8891f
mycroft/frontends/tts/mimic_tts.py
mycroft/frontends/tts/mimic_tts.py
from subprocess import call from mycroft.frontends.tts.tts_plugin import TtsPlugin class MimicTts(TtsPlugin): def read(self, text): call(['mimic', '-t', text, '-voice', self.config['voice']])
from subprocess import call from os.path import isdir from mycroft.frontends.tts.tts_plugin import TtsPlugin from mycroft.util.git_repo import GitRepo class MimicTts(TtsPlugin): def __init__(self, rt): super().__init__(rt) if not isdir(self.rt.paths.mimic_exe): self.download_mimic() def download_mimic(self): repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master') repo.try_pull() repo.run_inside('./dependencies.sh --prefix="/usr/local"') repo.run_inside('./autogen.sh') repo.run_inside('./configure.sh --prefix="/usr/local"') repo.run_inside('make -j2') def read(self, text): call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']])
Add download and compile step to mimic
Add download and compile step to mimic
Python
apache-2.0
MatthewScholefield/mycroft-simple,MatthewScholefield/mycroft-simple
from subprocess import call + from os.path import isdir + from mycroft.frontends.tts.tts_plugin import TtsPlugin + from mycroft.util.git_repo import GitRepo class MimicTts(TtsPlugin): + def __init__(self, rt): + super().__init__(rt) + + if not isdir(self.rt.paths.mimic_exe): + self.download_mimic() + + def download_mimic(self): + repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master') + repo.try_pull() + repo.run_inside('./dependencies.sh --prefix="/usr/local"') + repo.run_inside('./autogen.sh') + repo.run_inside('./configure.sh --prefix="/usr/local"') + repo.run_inside('make -j2') + def read(self, text): - call(['mimic', '-t', text, '-voice', self.config['voice']]) + call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']])
Add download and compile step to mimic
## Code Before: from subprocess import call from mycroft.frontends.tts.tts_plugin import TtsPlugin class MimicTts(TtsPlugin): def read(self, text): call(['mimic', '-t', text, '-voice', self.config['voice']]) ## Instruction: Add download and compile step to mimic ## Code After: from subprocess import call from os.path import isdir from mycroft.frontends.tts.tts_plugin import TtsPlugin from mycroft.util.git_repo import GitRepo class MimicTts(TtsPlugin): def __init__(self, rt): super().__init__(rt) if not isdir(self.rt.paths.mimic_exe): self.download_mimic() def download_mimic(self): repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master') repo.try_pull() repo.run_inside('./dependencies.sh --prefix="/usr/local"') repo.run_inside('./autogen.sh') repo.run_inside('./configure.sh --prefix="/usr/local"') repo.run_inside('make -j2') def read(self, text): call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']])
from subprocess import call + from os.path import isdir + from mycroft.frontends.tts.tts_plugin import TtsPlugin + from mycroft.util.git_repo import GitRepo class MimicTts(TtsPlugin): + def __init__(self, rt): + super().__init__(rt) + + if not isdir(self.rt.paths.mimic_exe): + self.download_mimic() + + def download_mimic(self): + repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master') + repo.try_pull() + repo.run_inside('./dependencies.sh --prefix="/usr/local"') + repo.run_inside('./autogen.sh') + repo.run_inside('./configure.sh --prefix="/usr/local"') + repo.run_inside('make -j2') + def read(self, text): - call(['mimic', '-t', text, '-voice', self.config['voice']]) ? ^ ^ + call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']]) ? ^^^^^^^^^^^^^^ ^^^^
721f837cbfa0de8804def607908a9744b0d099a8
asl/vendor/__init__.py
asl/vendor/__init__.py
import sys import os _vendor_initialized = False def append_paths(path, vendor_modules): new_path = [] for v in vendor_modules: new_path.append(path + os.sep + v) sys.path = new_path + sys.path def do_init(): global _vendor_initialized if _vendor_initialized: return _vendor_initialized = True path = "." if os.path.exists('./asl/asl/vendor'): path = os.path.abspath('./asl/asl/vendor') else: for p in sys.path: if os.path.exists(p + '/vendor/'): path = os.path.abspath(p + '/vendor/') break if os.path.exists(p + '/asl/vendor/'): path = os.path.abspath(p + '/asl/vendor/') break vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson'] append_paths(path, vendor_modules) do_init()
import sys import os _vendor_initialized = False def append_paths(path, vendor_modules): new_path = [] for v in vendor_modules: new_path.append(path + os.sep + v) sys.path = new_path + sys.path def do_init(): global _vendor_initialized if _vendor_initialized: return _vendor_initialized = True path = os.path.dirname(__file__) vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson'] append_paths(path, vendor_modules) do_init()
Fix of vendor directory search.
Fix of vendor directory search.
Python
mit
AtteqCom/zsl,AtteqCom/zsl
import sys import os _vendor_initialized = False def append_paths(path, vendor_modules): new_path = [] for v in vendor_modules: new_path.append(path + os.sep + v) sys.path = new_path + sys.path def do_init(): global _vendor_initialized if _vendor_initialized: return _vendor_initialized = True + path = os.path.dirname(__file__) - path = "." - if os.path.exists('./asl/asl/vendor'): - path = os.path.abspath('./asl/asl/vendor') - else: - for p in sys.path: - if os.path.exists(p + '/vendor/'): - path = os.path.abspath(p + '/vendor/') - break - if os.path.exists(p + '/asl/vendor/'): - path = os.path.abspath(p + '/asl/vendor/') - break - vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson'] append_paths(path, vendor_modules) do_init()
Fix of vendor directory search.
## Code Before: import sys import os _vendor_initialized = False def append_paths(path, vendor_modules): new_path = [] for v in vendor_modules: new_path.append(path + os.sep + v) sys.path = new_path + sys.path def do_init(): global _vendor_initialized if _vendor_initialized: return _vendor_initialized = True path = "." if os.path.exists('./asl/asl/vendor'): path = os.path.abspath('./asl/asl/vendor') else: for p in sys.path: if os.path.exists(p + '/vendor/'): path = os.path.abspath(p + '/vendor/') break if os.path.exists(p + '/asl/vendor/'): path = os.path.abspath(p + '/asl/vendor/') break vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson'] append_paths(path, vendor_modules) do_init() ## Instruction: Fix of vendor directory search. ## Code After: import sys import os _vendor_initialized = False def append_paths(path, vendor_modules): new_path = [] for v in vendor_modules: new_path.append(path + os.sep + v) sys.path = new_path + sys.path def do_init(): global _vendor_initialized if _vendor_initialized: return _vendor_initialized = True path = os.path.dirname(__file__) vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson'] append_paths(path, vendor_modules) do_init()
import sys import os _vendor_initialized = False def append_paths(path, vendor_modules): new_path = [] for v in vendor_modules: new_path.append(path + os.sep + v) sys.path = new_path + sys.path def do_init(): global _vendor_initialized if _vendor_initialized: return _vendor_initialized = True + path = os.path.dirname(__file__) - path = "." - if os.path.exists('./asl/asl/vendor'): - path = os.path.abspath('./asl/asl/vendor') - else: - for p in sys.path: - if os.path.exists(p + '/vendor/'): - path = os.path.abspath(p + '/vendor/') - break - if os.path.exists(p + '/asl/vendor/'): - path = os.path.abspath(p + '/asl/vendor/') - break - vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson'] append_paths(path, vendor_modules) do_init()
bd70ef56d95958b8f105bdff31b675d66c40bca8
serfnode/handler/supervisor.py
serfnode/handler/supervisor.py
import os import subprocess import docker_utils import jinja2 env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs')) def supervisor_install(block, **kwargs): """Update supervisor with `block` config. - `block` is the name to a .conf template file (in directory `/programs`) - `kwargs` are the key/values to use in the template """ conf_filename = '{}.conf'.format(kwargs['target']) template = env.get_template(block) kwargs.update({ 'DOCKER': docker_utils.DOCKER, 'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET, 'DOCKER_RUN': docker_utils.DOCKER_RUN}) conf = template.render(kwargs) with open(os.path.join( '/etc/supervisor/conf.d', conf_filename), 'w') as f: f.write(conf) def supervisor_exec(*args): return subprocess.check_output( ['supervisorctl'] + list(args)) def supervisor_update(): supervisor_exec('reread') supervisor_exec('update') def start(block, **kwargs): supervisor_install(block, **kwargs) supervisor_update() supervisor_exec('start', '{}:*'.format(kwargs['target'])) def stop(block): supervisor_exec('stop', '{}:*'.format(block))
import os import subprocess import docker_utils import docker import jinja2 env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs')) def supervisor_install(block, **kwargs): """Update supervisor with `block` config. - `block` is the name to a .conf template file (in directory `/programs`) - `kwargs` are the key/values to use in the template """ conf_filename = '{}.conf'.format(kwargs['target']) template = env.get_template(block) kwargs.update({ 'DOCKER': docker_utils.DOCKER, 'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET, 'DOCKER_RUN': docker_utils.DOCKER_RUN}) conf = template.render(kwargs) with open(os.path.join( '/etc/supervisor/conf.d', conf_filename), 'w') as f: f.write(conf) def supervisor_exec(*args): return subprocess.check_output( ['supervisorctl'] + list(args)) def supervisor_update(): supervisor_exec('reread') supervisor_exec('update') def start(block, **kwargs): supervisor_install(block, **kwargs) supervisor_update() supervisor_exec('start', '{}:*'.format(kwargs['target'])) def start_docker(target, name, cmdline): start('app.conf', target=target, ARGS='--cidfile=/app --name={} {}'.format(name, cmdline), NAME=name) def stop(block): supervisor_exec('stop', '{}:*'.format(block))
Add convenience function to start docker
Add convenience function to start docker Mainly to be used from supervisor.
Python
mit
waltermoreira/serfnode,waltermoreira/serfnode,waltermoreira/serfnode
import os import subprocess import docker_utils + import docker import jinja2 env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs')) def supervisor_install(block, **kwargs): """Update supervisor with `block` config. - `block` is the name to a .conf template file (in directory `/programs`) - `kwargs` are the key/values to use in the template """ conf_filename = '{}.conf'.format(kwargs['target']) template = env.get_template(block) kwargs.update({ 'DOCKER': docker_utils.DOCKER, 'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET, 'DOCKER_RUN': docker_utils.DOCKER_RUN}) conf = template.render(kwargs) with open(os.path.join( '/etc/supervisor/conf.d', conf_filename), 'w') as f: f.write(conf) def supervisor_exec(*args): return subprocess.check_output( ['supervisorctl'] + list(args)) def supervisor_update(): supervisor_exec('reread') supervisor_exec('update') def start(block, **kwargs): supervisor_install(block, **kwargs) supervisor_update() supervisor_exec('start', '{}:*'.format(kwargs['target'])) + def start_docker(target, name, cmdline): + start('app.conf', target=target, + ARGS='--cidfile=/app --name={} {}'.format(name, cmdline), + NAME=name) + + def stop(block): supervisor_exec('stop', '{}:*'.format(block)) +
Add convenience function to start docker
## Code Before: import os import subprocess import docker_utils import jinja2 env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs')) def supervisor_install(block, **kwargs): """Update supervisor with `block` config. - `block` is the name to a .conf template file (in directory `/programs`) - `kwargs` are the key/values to use in the template """ conf_filename = '{}.conf'.format(kwargs['target']) template = env.get_template(block) kwargs.update({ 'DOCKER': docker_utils.DOCKER, 'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET, 'DOCKER_RUN': docker_utils.DOCKER_RUN}) conf = template.render(kwargs) with open(os.path.join( '/etc/supervisor/conf.d', conf_filename), 'w') as f: f.write(conf) def supervisor_exec(*args): return subprocess.check_output( ['supervisorctl'] + list(args)) def supervisor_update(): supervisor_exec('reread') supervisor_exec('update') def start(block, **kwargs): supervisor_install(block, **kwargs) supervisor_update() supervisor_exec('start', '{}:*'.format(kwargs['target'])) def stop(block): supervisor_exec('stop', '{}:*'.format(block)) ## Instruction: Add convenience function to start docker ## Code After: import os import subprocess import docker_utils import docker import jinja2 env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs')) def supervisor_install(block, **kwargs): """Update supervisor with `block` config. - `block` is the name to a .conf template file (in directory `/programs`) - `kwargs` are the key/values to use in the template """ conf_filename = '{}.conf'.format(kwargs['target']) template = env.get_template(block) kwargs.update({ 'DOCKER': docker_utils.DOCKER, 'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET, 'DOCKER_RUN': docker_utils.DOCKER_RUN}) conf = template.render(kwargs) with open(os.path.join( '/etc/supervisor/conf.d', conf_filename), 'w') as f: f.write(conf) def supervisor_exec(*args): return subprocess.check_output( ['supervisorctl'] + list(args)) def supervisor_update(): supervisor_exec('reread') supervisor_exec('update') def start(block, **kwargs): supervisor_install(block, **kwargs) supervisor_update() supervisor_exec('start', '{}:*'.format(kwargs['target'])) def start_docker(target, name, cmdline): start('app.conf', target=target, ARGS='--cidfile=/app --name={} {}'.format(name, cmdline), NAME=name) def stop(block): supervisor_exec('stop', '{}:*'.format(block))
import os import subprocess import docker_utils + import docker import jinja2 env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs')) def supervisor_install(block, **kwargs): """Update supervisor with `block` config. - `block` is the name to a .conf template file (in directory `/programs`) - `kwargs` are the key/values to use in the template """ conf_filename = '{}.conf'.format(kwargs['target']) template = env.get_template(block) kwargs.update({ 'DOCKER': docker_utils.DOCKER, 'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET, 'DOCKER_RUN': docker_utils.DOCKER_RUN}) conf = template.render(kwargs) with open(os.path.join( '/etc/supervisor/conf.d', conf_filename), 'w') as f: f.write(conf) def supervisor_exec(*args): return subprocess.check_output( ['supervisorctl'] + list(args)) def supervisor_update(): supervisor_exec('reread') supervisor_exec('update') def start(block, **kwargs): supervisor_install(block, **kwargs) supervisor_update() supervisor_exec('start', '{}:*'.format(kwargs['target'])) + def start_docker(target, name, cmdline): + start('app.conf', target=target, + ARGS='--cidfile=/app --name={} {}'.format(name, cmdline), + NAME=name) + + def stop(block): supervisor_exec('stop', '{}:*'.format(block)) +
fce10cb35be29ba265f2ed189198703c718ad479
quantecon/__init__.py
quantecon/__init__.py
from . import models as models from .compute_fp import compute_fixed_point from .discrete_rv import DiscreteRV from .ecdf import ECDF from .estspec import smooth, periodogram, ar_periodogram from .graph_tools import DiGraph from .gridtools import cartesian, mlinspace from .gth_solve import gth_solve from .kalman import Kalman from .lae import LAE from .arma import ARMA from .lqcontrol import LQ from .lqnash import nnash from .lss import LinearStateSpace from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path from .quadsums import var_quadratic_sum, m_quadratic_sum from .markov import random_markov_chain, random_stochastic_matrix from .rank_nullspace import rank_est, nullspace from .robustlq import RBLQ from .tauchen import approx_markov from . import quad as quad from .util import searchsorted, random_probvec, random_sample_without_replacement #-Module Imports-# import util.random as random #Add Version Attribute from .version import version as __version__
from . import models as models from .compute_fp import compute_fixed_point from .discrete_rv import DiscreteRV from .ecdf import ECDF from .estspec import smooth, periodogram, ar_periodogram from .graph_tools import DiGraph from .gridtools import cartesian, mlinspace from .gth_solve import gth_solve from .kalman import Kalman from .lae import LAE from .arma import ARMA from .lqcontrol import LQ from .lqnash import nnash from .lss import LinearStateSpace from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path from .quadsums import var_quadratic_sum, m_quadratic_sum from .markov import random_markov_chain, random_stochastic_matrix from .rank_nullspace import rank_est, nullspace from .robustlq import RBLQ from .tauchen import approx_markov from . import quad as quad from .util import searchsorted, random_probvec, random_sample_without_replacement #-Module Imports-# from .util import random #Add Version Attribute from .version import version as __version__
Fix for python 3 relative import statement
Fix for python 3 relative import statement
Python
bsd-3-clause
andybrnr/QuantEcon.py,oyamad/QuantEcon.py,agutieda/QuantEcon.py,jviada/QuantEcon.py,QuantEcon/QuantEcon.py,QuantEcon/QuantEcon.py,gxxjjj/QuantEcon.py,agutieda/QuantEcon.py,gxxjjj/QuantEcon.py,oyamad/QuantEcon.py,jviada/QuantEcon.py,andybrnr/QuantEcon.py
from . import models as models from .compute_fp import compute_fixed_point from .discrete_rv import DiscreteRV from .ecdf import ECDF from .estspec import smooth, periodogram, ar_periodogram from .graph_tools import DiGraph from .gridtools import cartesian, mlinspace from .gth_solve import gth_solve from .kalman import Kalman from .lae import LAE from .arma import ARMA from .lqcontrol import LQ from .lqnash import nnash from .lss import LinearStateSpace from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path from .quadsums import var_quadratic_sum, m_quadratic_sum from .markov import random_markov_chain, random_stochastic_matrix from .rank_nullspace import rank_est, nullspace from .robustlq import RBLQ from .tauchen import approx_markov from . import quad as quad from .util import searchsorted, random_probvec, random_sample_without_replacement #-Module Imports-# - import util.random as random + from .util import random #Add Version Attribute from .version import version as __version__
Fix for python 3 relative import statement
## Code Before: from . import models as models from .compute_fp import compute_fixed_point from .discrete_rv import DiscreteRV from .ecdf import ECDF from .estspec import smooth, periodogram, ar_periodogram from .graph_tools import DiGraph from .gridtools import cartesian, mlinspace from .gth_solve import gth_solve from .kalman import Kalman from .lae import LAE from .arma import ARMA from .lqcontrol import LQ from .lqnash import nnash from .lss import LinearStateSpace from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path from .quadsums import var_quadratic_sum, m_quadratic_sum from .markov import random_markov_chain, random_stochastic_matrix from .rank_nullspace import rank_est, nullspace from .robustlq import RBLQ from .tauchen import approx_markov from . import quad as quad from .util import searchsorted, random_probvec, random_sample_without_replacement #-Module Imports-# import util.random as random #Add Version Attribute from .version import version as __version__ ## Instruction: Fix for python 3 relative import statement ## Code After: from . import models as models from .compute_fp import compute_fixed_point from .discrete_rv import DiscreteRV from .ecdf import ECDF from .estspec import smooth, periodogram, ar_periodogram from .graph_tools import DiGraph from .gridtools import cartesian, mlinspace from .gth_solve import gth_solve from .kalman import Kalman from .lae import LAE from .arma import ARMA from .lqcontrol import LQ from .lqnash import nnash from .lss import LinearStateSpace from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path from .quadsums import var_quadratic_sum, m_quadratic_sum from .markov import random_markov_chain, random_stochastic_matrix from .rank_nullspace import rank_est, nullspace from .robustlq import RBLQ from .tauchen import approx_markov from . import quad as quad from .util import searchsorted, random_probvec, random_sample_without_replacement #-Module Imports-# from .util import random #Add Version Attribute from .version import version as __version__
from . import models as models from .compute_fp import compute_fixed_point from .discrete_rv import DiscreteRV from .ecdf import ECDF from .estspec import smooth, periodogram, ar_periodogram from .graph_tools import DiGraph from .gridtools import cartesian, mlinspace from .gth_solve import gth_solve from .kalman import Kalman from .lae import LAE from .arma import ARMA from .lqcontrol import LQ from .lqnash import nnash from .lss import LinearStateSpace from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path from .quadsums import var_quadratic_sum, m_quadratic_sum from .markov import random_markov_chain, random_stochastic_matrix from .rank_nullspace import rank_est, nullspace from .robustlq import RBLQ from .tauchen import approx_markov from . import quad as quad from .util import searchsorted, random_probvec, random_sample_without_replacement #-Module Imports-# - import util.random as random + from .util import random #Add Version Attribute from .version import version as __version__
898028dea2e04d52c32854752bda34d331c7696f
ynr/apps/candidatebot/management/commands/candidatebot_import_email_from_csv.py
ynr/apps/candidatebot/management/commands/candidatebot_import_email_from_csv.py
from __future__ import unicode_literals import csv from django.core.management.base import BaseCommand from candidatebot.helpers import CandidateBot from popolo.models import Person class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument( 'filename', help='Path to the file with the email addresses' ) parser.add_argument( '--source', help='Source of the data. The source CSV column takes precedence' ) def handle(self, **options): with open(options['filename'], 'r') as fh: reader = csv.DictReader(fh) for row in reader: source = row.get('source', options.get('source')) if not row['democlub_id']: continue if not source: raise ValueError("A source is required") try: bot = CandidateBot(row['democlub_id']) bot.add_email(row['email']) bot.save(source) # print(person) except Person.DoesNotExist: print("Person ID {} not found".format( row['democlub_id'])) # print(row)
from __future__ import unicode_literals import csv from django.core.management.base import BaseCommand from candidatebot.helpers import CandidateBot from popolo.models import Person class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument( 'filename', help='Path to the file with the email addresses' ) parser.add_argument( '--source', help='Source of the data. The source CSV column takes precedence' ) def handle(self, **options): with open(options['filename'], 'r') as fh: reader = csv.DictReader(fh) for row in reader: source = row.get('source', options.get('source')) if not row['democlub_id']: continue if not source: raise ValueError("A source is required") try: bot = CandidateBot(row['democlub_id']) try: bot.add_email(row['email']) bot.save(source) except ValueError: #Email exists, move on pass except Person.DoesNotExist: print("Person ID {} not found".format( row['democlub_id'])) # print(row)
Move on if email exists
Move on if email exists
Python
agpl-3.0
DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative
from __future__ import unicode_literals import csv from django.core.management.base import BaseCommand from candidatebot.helpers import CandidateBot from popolo.models import Person class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument( 'filename', help='Path to the file with the email addresses' ) parser.add_argument( '--source', help='Source of the data. The source CSV column takes precedence' ) def handle(self, **options): with open(options['filename'], 'r') as fh: reader = csv.DictReader(fh) for row in reader: source = row.get('source', options.get('source')) if not row['democlub_id']: continue if not source: raise ValueError("A source is required") try: bot = CandidateBot(row['democlub_id']) + try: - bot.add_email(row['email']) + bot.add_email(row['email']) - bot.save(source) + bot.save(source) - # print(person) + except ValueError: + #Email exists, move on + pass except Person.DoesNotExist: print("Person ID {} not found".format( row['democlub_id'])) # print(row)
Move on if email exists
## Code Before: from __future__ import unicode_literals import csv from django.core.management.base import BaseCommand from candidatebot.helpers import CandidateBot from popolo.models import Person class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument( 'filename', help='Path to the file with the email addresses' ) parser.add_argument( '--source', help='Source of the data. The source CSV column takes precedence' ) def handle(self, **options): with open(options['filename'], 'r') as fh: reader = csv.DictReader(fh) for row in reader: source = row.get('source', options.get('source')) if not row['democlub_id']: continue if not source: raise ValueError("A source is required") try: bot = CandidateBot(row['democlub_id']) bot.add_email(row['email']) bot.save(source) # print(person) except Person.DoesNotExist: print("Person ID {} not found".format( row['democlub_id'])) # print(row) ## Instruction: Move on if email exists ## Code After: from __future__ import unicode_literals import csv from django.core.management.base import BaseCommand from candidatebot.helpers import CandidateBot from popolo.models import Person class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument( 'filename', help='Path to the file with the email addresses' ) parser.add_argument( '--source', help='Source of the data. The source CSV column takes precedence' ) def handle(self, **options): with open(options['filename'], 'r') as fh: reader = csv.DictReader(fh) for row in reader: source = row.get('source', options.get('source')) if not row['democlub_id']: continue if not source: raise ValueError("A source is required") try: bot = CandidateBot(row['democlub_id']) try: bot.add_email(row['email']) bot.save(source) except ValueError: #Email exists, move on pass except Person.DoesNotExist: print("Person ID {} not found".format( row['democlub_id'])) # print(row)
from __future__ import unicode_literals import csv from django.core.management.base import BaseCommand from candidatebot.helpers import CandidateBot from popolo.models import Person class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument( 'filename', help='Path to the file with the email addresses' ) parser.add_argument( '--source', help='Source of the data. The source CSV column takes precedence' ) def handle(self, **options): with open(options['filename'], 'r') as fh: reader = csv.DictReader(fh) for row in reader: source = row.get('source', options.get('source')) if not row['democlub_id']: continue if not source: raise ValueError("A source is required") try: bot = CandidateBot(row['democlub_id']) + try: - bot.add_email(row['email']) + bot.add_email(row['email']) ? ++++ - bot.save(source) + bot.save(source) ? ++++ - # print(person) + except ValueError: + #Email exists, move on + pass except Person.DoesNotExist: print("Person ID {} not found".format( row['democlub_id'])) # print(row)
6c2354a1e56477eb983b0adbcc2d15223c158184
foodsaving/subscriptions/consumers.py
foodsaving/subscriptions/consumers.py
from channels.auth import channel_session_user_from_http, channel_session_user from django.utils import timezone from foodsaving.subscriptions.models import ChannelSubscription @channel_session_user_from_http def ws_connect(message): """The user has connected! Register their channel subscription.""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel) message.reply_channel.send({"accept": True}) @channel_session_user def ws_message(message): """They sent us a websocket message! We just update the ChannelSubscription lastseen time..""" user = message.user if not user.is_anonymous(): reply_channel = message.reply_channel.name ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now()) message.reply_channel.send({"accept": True}) @channel_session_user def ws_disconnect(message): """The user has disconnected so we remove all their ChannelSubscriptions""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete() message.reply_channel.send({"accept": True})
from channels.auth import channel_session_user_from_http, channel_session_user from django.utils import timezone from foodsaving.subscriptions.models import ChannelSubscription @channel_session_user_from_http def ws_connect(message): """The user has connected! Register their channel subscription.""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel) message.reply_channel.send({"accept": True}) @channel_session_user def ws_message(message): """They sent us a websocket message! We just update the ChannelSubscription lastseen time..""" user = message.user if not user.is_anonymous(): reply_channel = message.reply_channel.name ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now()) @channel_session_user def ws_disconnect(message): """The user has disconnected so we remove all their ChannelSubscriptions""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
Remove redundant ws accept replies
Remove redundant ws accept replies It's only relevent on connection
Python
agpl-3.0
yunity/yunity-core,yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/foodsaving-backend
from channels.auth import channel_session_user_from_http, channel_session_user from django.utils import timezone from foodsaving.subscriptions.models import ChannelSubscription @channel_session_user_from_http def ws_connect(message): """The user has connected! Register their channel subscription.""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel) message.reply_channel.send({"accept": True}) @channel_session_user def ws_message(message): """They sent us a websocket message! We just update the ChannelSubscription lastseen time..""" user = message.user if not user.is_anonymous(): reply_channel = message.reply_channel.name ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now()) - message.reply_channel.send({"accept": True}) @channel_session_user def ws_disconnect(message): """The user has disconnected so we remove all their ChannelSubscriptions""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete() - message.reply_channel.send({"accept": True})
Remove redundant ws accept replies
## Code Before: from channels.auth import channel_session_user_from_http, channel_session_user from django.utils import timezone from foodsaving.subscriptions.models import ChannelSubscription @channel_session_user_from_http def ws_connect(message): """The user has connected! Register their channel subscription.""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel) message.reply_channel.send({"accept": True}) @channel_session_user def ws_message(message): """They sent us a websocket message! We just update the ChannelSubscription lastseen time..""" user = message.user if not user.is_anonymous(): reply_channel = message.reply_channel.name ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now()) message.reply_channel.send({"accept": True}) @channel_session_user def ws_disconnect(message): """The user has disconnected so we remove all their ChannelSubscriptions""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete() message.reply_channel.send({"accept": True}) ## Instruction: Remove redundant ws accept replies ## Code After: from channels.auth import channel_session_user_from_http, channel_session_user from django.utils import timezone from foodsaving.subscriptions.models import ChannelSubscription @channel_session_user_from_http def ws_connect(message): """The user has connected! Register their channel subscription.""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel) message.reply_channel.send({"accept": True}) @channel_session_user def ws_message(message): """They sent us a websocket message! We just update the ChannelSubscription lastseen time..""" user = message.user if not user.is_anonymous(): reply_channel = message.reply_channel.name ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now()) @channel_session_user def ws_disconnect(message): """The user has disconnected so we remove all their ChannelSubscriptions""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
from channels.auth import channel_session_user_from_http, channel_session_user from django.utils import timezone from foodsaving.subscriptions.models import ChannelSubscription @channel_session_user_from_http def ws_connect(message): """The user has connected! Register their channel subscription.""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel) message.reply_channel.send({"accept": True}) @channel_session_user def ws_message(message): """They sent us a websocket message! We just update the ChannelSubscription lastseen time..""" user = message.user if not user.is_anonymous(): reply_channel = message.reply_channel.name ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now()) - message.reply_channel.send({"accept": True}) @channel_session_user def ws_disconnect(message): """The user has disconnected so we remove all their ChannelSubscriptions""" user = message.user if not user.is_anonymous(): ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete() - message.reply_channel.send({"accept": True})
f890663daa329e3f22d0f619ed6acf9365308c7c
apps/ignite/views.py
apps/ignite/views.py
from django.shortcuts import get_object_or_404 import jingo from challenges.models import Submission, Category from projects.models import Project def splash(request, project, slug, template_name='challenges/show.html'): """Show an individual project challenge.""" project = get_object_or_404(Project, slug=project) challenge = get_object_or_404(project.challenge_set, slug=slug) entries = Submission.objects.filter( phase__challenge=challenge ).exclude( is_draft=True ).extra( order_by="?" ) return jingo.render(request, 'ignite/splash.html', { 'challenge': challenge, 'project': project, 'phases': list(enumerate(challenge.phases.all(), start=1)), 'entries': entries[:10], 'categories': Category.objects.get_active_categories(), })
from django.shortcuts import get_object_or_404 import jingo from challenges.models import Submission, Category from projects.models import Project def splash(request, project, slug, template_name='challenges/show.html'): """Show an individual project challenge.""" project = get_object_or_404(Project, slug=project) challenge = get_object_or_404(project.challenge_set, slug=slug) entries = (Submission.objects.visible() .filter(phase__challenge=challenge) .order_by("?")) return jingo.render(request, 'ignite/splash.html', { 'challenge': challenge, 'project': project, 'phases': list(enumerate(challenge.phases.all(), start=1)), 'entries': entries[:10], 'categories': Category.objects.get_active_categories(), })
Update splash view to use visible() method.
Update splash view to use visible() method.
Python
bsd-3-clause
mozilla/mozilla-ignite,mozilla/mozilla-ignite,mozilla/mozilla-ignite,mozilla/mozilla-ignite
from django.shortcuts import get_object_or_404 import jingo from challenges.models import Submission, Category from projects.models import Project def splash(request, project, slug, template_name='challenges/show.html'): """Show an individual project challenge.""" project = get_object_or_404(Project, slug=project) challenge = get_object_or_404(project.challenge_set, slug=slug) - entries = Submission.objects.filter( + entries = (Submission.objects.visible() + .filter(phase__challenge=challenge) + .order_by("?")) - phase__challenge=challenge - ).exclude( - is_draft=True - ).extra( - order_by="?" - ) + return jingo.render(request, 'ignite/splash.html', { 'challenge': challenge, 'project': project, 'phases': list(enumerate(challenge.phases.all(), start=1)), 'entries': entries[:10], 'categories': Category.objects.get_active_categories(), })
Update splash view to use visible() method.
## Code Before: from django.shortcuts import get_object_or_404 import jingo from challenges.models import Submission, Category from projects.models import Project def splash(request, project, slug, template_name='challenges/show.html'): """Show an individual project challenge.""" project = get_object_or_404(Project, slug=project) challenge = get_object_or_404(project.challenge_set, slug=slug) entries = Submission.objects.filter( phase__challenge=challenge ).exclude( is_draft=True ).extra( order_by="?" ) return jingo.render(request, 'ignite/splash.html', { 'challenge': challenge, 'project': project, 'phases': list(enumerate(challenge.phases.all(), start=1)), 'entries': entries[:10], 'categories': Category.objects.get_active_categories(), }) ## Instruction: Update splash view to use visible() method. ## Code After: from django.shortcuts import get_object_or_404 import jingo from challenges.models import Submission, Category from projects.models import Project def splash(request, project, slug, template_name='challenges/show.html'): """Show an individual project challenge.""" project = get_object_or_404(Project, slug=project) challenge = get_object_or_404(project.challenge_set, slug=slug) entries = (Submission.objects.visible() .filter(phase__challenge=challenge) .order_by("?")) return jingo.render(request, 'ignite/splash.html', { 'challenge': challenge, 'project': project, 'phases': list(enumerate(challenge.phases.all(), start=1)), 'entries': entries[:10], 'categories': Category.objects.get_active_categories(), })
from django.shortcuts import get_object_or_404 import jingo from challenges.models import Submission, Category from projects.models import Project def splash(request, project, slug, template_name='challenges/show.html'): """Show an individual project challenge.""" project = get_object_or_404(Project, slug=project) challenge = get_object_or_404(project.challenge_set, slug=slug) - entries = Submission.objects.filter( ? ^ - - + entries = (Submission.objects.visible() ? + ^ +++ + + .filter(phase__challenge=challenge) + .order_by("?")) - phase__challenge=challenge - ).exclude( - is_draft=True - ).extra( - order_by="?" - ) ? - + return jingo.render(request, 'ignite/splash.html', { 'challenge': challenge, 'project': project, 'phases': list(enumerate(challenge.phases.all(), start=1)), 'entries': entries[:10], 'categories': Category.objects.get_active_categories(), })
15013c51f602786265b59c1d4a7e894eae090d90
tests/test_normalize.py
tests/test_normalize.py
from hypothesis import assume, given from utils import isclose, vectors @given(v=vectors()) def test_normalize_length(v): """v.normalize().length == 1 and v == v.length * v.normalize()""" assume(v) assert isclose(v.normalize().length, 1) assert v.isclose(v.length * v.normalize())
from hypothesis import assume, given from utils import isclose, vectors @given(v=vectors()) def test_normalize_length(v): """v.normalize().length == 1 and v == v.length * v.normalize()""" assume(v) assert isclose(v.normalize().length, 1) assert v.isclose(v.length * v.normalize()) @given(v=vectors()) def test_normalize_angle(v): """Normalization preserves direction.""" assume(v) assert angle_isclose(v.normalize().angle(v), 0)
Test that direction is preserved
tests/normalize: Test that direction is preserved
Python
artistic-2.0
ppb/ppb-vector,ppb/ppb-vector
from hypothesis import assume, given from utils import isclose, vectors @given(v=vectors()) def test_normalize_length(v): """v.normalize().length == 1 and v == v.length * v.normalize()""" assume(v) assert isclose(v.normalize().length, 1) assert v.isclose(v.length * v.normalize()) + + @given(v=vectors()) + def test_normalize_angle(v): + """Normalization preserves direction.""" + assume(v) + assert angle_isclose(v.normalize().angle(v), 0) +
Test that direction is preserved
## Code Before: from hypothesis import assume, given from utils import isclose, vectors @given(v=vectors()) def test_normalize_length(v): """v.normalize().length == 1 and v == v.length * v.normalize()""" assume(v) assert isclose(v.normalize().length, 1) assert v.isclose(v.length * v.normalize()) ## Instruction: Test that direction is preserved ## Code After: from hypothesis import assume, given from utils import isclose, vectors @given(v=vectors()) def test_normalize_length(v): """v.normalize().length == 1 and v == v.length * v.normalize()""" assume(v) assert isclose(v.normalize().length, 1) assert v.isclose(v.length * v.normalize()) @given(v=vectors()) def test_normalize_angle(v): """Normalization preserves direction.""" assume(v) assert angle_isclose(v.normalize().angle(v), 0)
from hypothesis import assume, given from utils import isclose, vectors @given(v=vectors()) def test_normalize_length(v): """v.normalize().length == 1 and v == v.length * v.normalize()""" assume(v) assert isclose(v.normalize().length, 1) assert v.isclose(v.length * v.normalize()) + + + @given(v=vectors()) + def test_normalize_angle(v): + """Normalization preserves direction.""" + assume(v) + assert angle_isclose(v.normalize().angle(v), 0)
c1fbc761e10e06effa49ede1f8dbc04189999bd5
niftynet/layer/post_processing.py
niftynet/layer/post_processing.py
from __future__ import absolute_import, print_function import tensorflow as tf from niftynet.utilities.util_common import look_up_operations from niftynet.layer.base_layer import Layer SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"} class PostProcessingLayer(Layer): """ This layer operation converts the raw network outputs into final inference results. """ def __init__(self, func='', num_classes=0, name='post_processing'): super(PostProcessingLayer, self).__init__(name=name) self.func = look_up_operations(func.upper(), SUPPORTED_OPS) self.num_classes = num_classes def num_output_channels(self): assert self._op._variables_created if self.func == "SOFTMAX": return self.num_classes else: return 1 def layer_op(self, inputs): if self.func == "SOFTMAX": output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32) elif self.func == "ARGMAX": output_tensor = tf.to_int64(tf.argmax(inputs, -1)) output_tensor = tf.expand_dims(output_tensor, axis=-1) elif self.func == "IDENTITY": output_tensor = tf.cast(inputs, tf.float32) return output_tensor
from __future__ import absolute_import, print_function import tensorflow as tf from niftynet.utilities.util_common import look_up_operations from niftynet.layer.base_layer import Layer SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"} class PostProcessingLayer(Layer): """ This layer operation converts the raw network outputs into final inference results. """ def __init__(self, func='', num_classes=0, name='post_processing'): super(PostProcessingLayer, self).__init__(name=name) self.func = look_up_operations(func.upper(), SUPPORTED_OPS) self.num_classes = num_classes def num_output_channels(self): assert self._op._variables_created if self.func == "SOFTMAX": return self.num_classes else: return 1 def layer_op(self, inputs): if self.func == "SOFTMAX": output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32) elif self.func == "ARGMAX": output_tensor = tf.argmax(inputs, -1,output_type=tf.int32) output_tensor = tf.expand_dims(output_tensor, axis=-1) elif self.func == "IDENTITY": output_tensor = tf.cast(inputs, tf.float32) return output_tensor
Change label output to int32 for compatibility with some viewers
Change label output to int32 for compatibility with some viewers
Python
apache-2.0
NifTK/NiftyNet,NifTK/NiftyNet,NifTK/NiftyNet,NifTK/NiftyNet
from __future__ import absolute_import, print_function import tensorflow as tf from niftynet.utilities.util_common import look_up_operations from niftynet.layer.base_layer import Layer SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"} class PostProcessingLayer(Layer): """ This layer operation converts the raw network outputs into final inference results. """ def __init__(self, func='', num_classes=0, name='post_processing'): super(PostProcessingLayer, self).__init__(name=name) self.func = look_up_operations(func.upper(), SUPPORTED_OPS) self.num_classes = num_classes def num_output_channels(self): assert self._op._variables_created if self.func == "SOFTMAX": return self.num_classes else: return 1 def layer_op(self, inputs): if self.func == "SOFTMAX": output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32) elif self.func == "ARGMAX": - output_tensor = tf.to_int64(tf.argmax(inputs, -1)) + output_tensor = tf.argmax(inputs, -1,output_type=tf.int32) output_tensor = tf.expand_dims(output_tensor, axis=-1) elif self.func == "IDENTITY": output_tensor = tf.cast(inputs, tf.float32) return output_tensor
Change label output to int32 for compatibility with some viewers
## Code Before: from __future__ import absolute_import, print_function import tensorflow as tf from niftynet.utilities.util_common import look_up_operations from niftynet.layer.base_layer import Layer SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"} class PostProcessingLayer(Layer): """ This layer operation converts the raw network outputs into final inference results. """ def __init__(self, func='', num_classes=0, name='post_processing'): super(PostProcessingLayer, self).__init__(name=name) self.func = look_up_operations(func.upper(), SUPPORTED_OPS) self.num_classes = num_classes def num_output_channels(self): assert self._op._variables_created if self.func == "SOFTMAX": return self.num_classes else: return 1 def layer_op(self, inputs): if self.func == "SOFTMAX": output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32) elif self.func == "ARGMAX": output_tensor = tf.to_int64(tf.argmax(inputs, -1)) output_tensor = tf.expand_dims(output_tensor, axis=-1) elif self.func == "IDENTITY": output_tensor = tf.cast(inputs, tf.float32) return output_tensor ## Instruction: Change label output to int32 for compatibility with some viewers ## Code After: from __future__ import absolute_import, print_function import tensorflow as tf from niftynet.utilities.util_common import look_up_operations from niftynet.layer.base_layer import Layer SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"} class PostProcessingLayer(Layer): """ This layer operation converts the raw network outputs into final inference results. """ def __init__(self, func='', num_classes=0, name='post_processing'): super(PostProcessingLayer, self).__init__(name=name) self.func = look_up_operations(func.upper(), SUPPORTED_OPS) self.num_classes = num_classes def num_output_channels(self): assert self._op._variables_created if self.func == "SOFTMAX": return self.num_classes else: return 1 def layer_op(self, inputs): if self.func == "SOFTMAX": output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32) elif self.func == "ARGMAX": output_tensor = tf.argmax(inputs, -1,output_type=tf.int32) output_tensor = tf.expand_dims(output_tensor, axis=-1) elif self.func == "IDENTITY": output_tensor = tf.cast(inputs, tf.float32) return output_tensor
from __future__ import absolute_import, print_function import tensorflow as tf from niftynet.utilities.util_common import look_up_operations from niftynet.layer.base_layer import Layer SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"} class PostProcessingLayer(Layer): """ This layer operation converts the raw network outputs into final inference results. """ def __init__(self, func='', num_classes=0, name='post_processing'): super(PostProcessingLayer, self).__init__(name=name) self.func = look_up_operations(func.upper(), SUPPORTED_OPS) self.num_classes = num_classes def num_output_channels(self): assert self._op._variables_created if self.func == "SOFTMAX": return self.num_classes else: return 1 def layer_op(self, inputs): if self.func == "SOFTMAX": output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32) elif self.func == "ARGMAX": - output_tensor = tf.to_int64(tf.argmax(inputs, -1)) + output_tensor = tf.argmax(inputs, -1,output_type=tf.int32) output_tensor = tf.expand_dims(output_tensor, axis=-1) elif self.func == "IDENTITY": output_tensor = tf.cast(inputs, tf.float32) return output_tensor
4d1b96792f73777adaa0a79341901ca82f57839b
use/functional.py
use/functional.py
def pipe(*functions): def closure(x): for fn in functions: if not out: out = fn(x) else: out = fn(out) return out return closure
import collections import functools def pipe(*functions): def closure(x): for fn in functions: if not out: out = fn(x) else: out = fn(out) return out return closure class memoize(object): '''Decorator. Caches a function's return value each time it is called. If called later with the same arguments, the cached value is returned (not reevaluated). ''' def __init__(self, func): self.func = func self.cache = {} def __call__(self, *args): if not isinstance(args, collections.Hashable): # uncacheable. a list, for instance. # better to not cache than blow up. return self.func(*args) if args in self.cache: return self.cache[args] else: value = self.func(*args) self.cache[args] = value return value def __repr__(self): '''Return the function's docstring.''' return self.func.__doc__ def __get__(self, obj, objtype): '''Support instance methods.''' return functools.partial(self.__call__, obj)
Add a simple memoize function
Add a simple memoize function
Python
mit
log0ymxm/corgi
+ import collections + import functools + + def pipe(*functions): def closure(x): for fn in functions: if not out: out = fn(x) else: out = fn(out) return out return closure + + class memoize(object): + '''Decorator. Caches a function's return value each time it is called. + If called later with the same arguments, the cached value is returned + (not reevaluated). + ''' + + def __init__(self, func): + self.func = func + self.cache = {} + + def __call__(self, *args): + if not isinstance(args, collections.Hashable): + # uncacheable. a list, for instance. + # better to not cache than blow up. + return self.func(*args) + if args in self.cache: + return self.cache[args] + else: + value = self.func(*args) + self.cache[args] = value + return value + + def __repr__(self): + '''Return the function's docstring.''' + return self.func.__doc__ + + def __get__(self, obj, objtype): + '''Support instance methods.''' + return functools.partial(self.__call__, obj) +
Add a simple memoize function
## Code Before: def pipe(*functions): def closure(x): for fn in functions: if not out: out = fn(x) else: out = fn(out) return out return closure ## Instruction: Add a simple memoize function ## Code After: import collections import functools def pipe(*functions): def closure(x): for fn in functions: if not out: out = fn(x) else: out = fn(out) return out return closure class memoize(object): '''Decorator. Caches a function's return value each time it is called. If called later with the same arguments, the cached value is returned (not reevaluated). ''' def __init__(self, func): self.func = func self.cache = {} def __call__(self, *args): if not isinstance(args, collections.Hashable): # uncacheable. a list, for instance. # better to not cache than blow up. return self.func(*args) if args in self.cache: return self.cache[args] else: value = self.func(*args) self.cache[args] = value return value def __repr__(self): '''Return the function's docstring.''' return self.func.__doc__ def __get__(self, obj, objtype): '''Support instance methods.''' return functools.partial(self.__call__, obj)
+ import collections + import functools + + def pipe(*functions): def closure(x): for fn in functions: if not out: out = fn(x) else: out = fn(out) return out return closure + + + class memoize(object): + '''Decorator. Caches a function's return value each time it is called. + If called later with the same arguments, the cached value is returned + (not reevaluated). + ''' + + def __init__(self, func): + self.func = func + self.cache = {} + + def __call__(self, *args): + if not isinstance(args, collections.Hashable): + # uncacheable. a list, for instance. + # better to not cache than blow up. + return self.func(*args) + if args in self.cache: + return self.cache[args] + else: + value = self.func(*args) + self.cache[args] = value + return value + + def __repr__(self): + '''Return the function's docstring.''' + return self.func.__doc__ + + def __get__(self, obj, objtype): + '''Support instance methods.''' + return functools.partial(self.__call__, obj)
ddd45afa0708682bb11d606e03e38aed111d7b9c
fireplace/cards/game/all.py
fireplace/cards/game/all.py
from ..utils import * # The Coin class GAME_005: play = ManaThisTurn(CONTROLLER, 1)
from ..utils import * # The Coin class GAME_005: play = ManaThisTurn(CONTROLLER, 1) # Big Banana class TB_006: play = Buff(TARGET, "TB_006e") # Deviate Banana class TB_007: play = Buff(TARGET, "TB_007e") # Rotten Banana class TB_008: play = Hit(TARGET, 1)
Implement Big Banana, Deviate Banana, Rotten Banana
Implement Big Banana, Deviate Banana, Rotten Banana
Python
agpl-3.0
liujimj/fireplace,Ragowit/fireplace,butozerca/fireplace,butozerca/fireplace,smallnamespace/fireplace,amw2104/fireplace,smallnamespace/fireplace,beheh/fireplace,NightKev/fireplace,Meerkov/fireplace,Meerkov/fireplace,liujimj/fireplace,oftc-ftw/fireplace,Ragowit/fireplace,amw2104/fireplace,jleclanche/fireplace,oftc-ftw/fireplace
from ..utils import * # The Coin class GAME_005: play = ManaThisTurn(CONTROLLER, 1) + + # Big Banana + class TB_006: + play = Buff(TARGET, "TB_006e") + + + # Deviate Banana + class TB_007: + play = Buff(TARGET, "TB_007e") + + + # Rotten Banana + class TB_008: + play = Hit(TARGET, 1) +
Implement Big Banana, Deviate Banana, Rotten Banana
## Code Before: from ..utils import * # The Coin class GAME_005: play = ManaThisTurn(CONTROLLER, 1) ## Instruction: Implement Big Banana, Deviate Banana, Rotten Banana ## Code After: from ..utils import * # The Coin class GAME_005: play = ManaThisTurn(CONTROLLER, 1) # Big Banana class TB_006: play = Buff(TARGET, "TB_006e") # Deviate Banana class TB_007: play = Buff(TARGET, "TB_007e") # Rotten Banana class TB_008: play = Hit(TARGET, 1)
from ..utils import * # The Coin class GAME_005: play = ManaThisTurn(CONTROLLER, 1) + + + # Big Banana + class TB_006: + play = Buff(TARGET, "TB_006e") + + + # Deviate Banana + class TB_007: + play = Buff(TARGET, "TB_007e") + + + # Rotten Banana + class TB_008: + play = Hit(TARGET, 1)
0b41bdf6897bb070fc3d90aa5d90228e744dee60
sunpy/util/map_manager.py
sunpy/util/map_manager.py
import weakref import sunpy class MapManager(weakref.WeakSet): """Weak referenced set of maps created using functions decorated by manage_maps.""" def __repr__(self): return str(self.data) def manage_maps(fn): """Maps returned by functions decorated with manage_maps (eg. sunpy.make_map) will be registered in the sunpy.map_manager list.""" def fn_manage_maps(*args, **kwargs): ret = fn(*args, **kwargs) sunpy.map_manager.add(ret) return ret return fn_manage_maps
import weakref import sunpy class MapManager(weakref.WeakSet): """Weak referenced set of maps created using functions decorated by manage_maps.""" pass def manage_maps(fn): """Maps returned by functions decorated with manage_maps (eg. sunpy.make_map) will be registered in the sunpy.map_manager list.""" def fn_manage_maps(*args, **kwargs): ret = fn(*args, **kwargs) sunpy.map_manager.add(ret) return ret return fn_manage_maps
Remove manager repr (user should not need to view contents)
Remove manager repr (user should not need to view contents)
Python
bsd-2-clause
mjm159/sunpy,dpshelio/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy,Alex-Ian-Hamilton/sunpy,mjm159/sunpy
import weakref import sunpy class MapManager(weakref.WeakSet): """Weak referenced set of maps created using functions decorated by manage_maps.""" + pass - - def __repr__(self): - return str(self.data) def manage_maps(fn): """Maps returned by functions decorated with manage_maps (eg. sunpy.make_map) will be registered in the sunpy.map_manager list.""" def fn_manage_maps(*args, **kwargs): ret = fn(*args, **kwargs) sunpy.map_manager.add(ret) return ret return fn_manage_maps
Remove manager repr (user should not need to view contents)
## Code Before: import weakref import sunpy class MapManager(weakref.WeakSet): """Weak referenced set of maps created using functions decorated by manage_maps.""" def __repr__(self): return str(self.data) def manage_maps(fn): """Maps returned by functions decorated with manage_maps (eg. sunpy.make_map) will be registered in the sunpy.map_manager list.""" def fn_manage_maps(*args, **kwargs): ret = fn(*args, **kwargs) sunpy.map_manager.add(ret) return ret return fn_manage_maps ## Instruction: Remove manager repr (user should not need to view contents) ## Code After: import weakref import sunpy class MapManager(weakref.WeakSet): """Weak referenced set of maps created using functions decorated by manage_maps.""" pass def manage_maps(fn): """Maps returned by functions decorated with manage_maps (eg. sunpy.make_map) will be registered in the sunpy.map_manager list.""" def fn_manage_maps(*args, **kwargs): ret = fn(*args, **kwargs) sunpy.map_manager.add(ret) return ret return fn_manage_maps
import weakref import sunpy class MapManager(weakref.WeakSet): """Weak referenced set of maps created using functions decorated by manage_maps.""" + pass - - def __repr__(self): - return str(self.data) def manage_maps(fn): """Maps returned by functions decorated with manage_maps (eg. sunpy.make_map) will be registered in the sunpy.map_manager list.""" def fn_manage_maps(*args, **kwargs): ret = fn(*args, **kwargs) sunpy.map_manager.add(ret) return ret return fn_manage_maps
b3413818bf651c13cef047132813fb26a185cd33
indra/tests/test_reading_files.py
indra/tests/test_reading_files.py
from os import path from indra.tools.reading.read_files import read_files, get_readers from nose.plugins.attrib import attr @attr('slow', 'nonpublic') def test_read_files(): "Test that the system can read files." # Create the test files. example_files = [] # Get txt content abstract_txt = ("This is a paper that contains the phrase: MEK " "phosphorylates ERK.") with open('test_abstract.txt', 'w') as f: f.write(abstract_txt) example_files.append('test_abstract.txt') # Get nxml content pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)), 'pmc_cont_example.nxml') if path.exists(pmc_test_fpath): example_files.append(pmc_test_fpath) assert len(example_files), "No content available to test." # Now read them. readers = get_readers() outputs = read_files(example_files, readers) N_out = len(outputs) N_exp = 2*len(example_files) assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
from os import path from indra.tools.reading.read_files import read_files, get_reader_classes from nose.plugins.attrib import attr from indra.tools.reading.readers import EmptyReader @attr('slow', 'nonpublic', 'notravis') def test_read_files(): "Test that the system can read files." # Create the test files. example_files = [] # Get txt content abstract_txt = ("This is a paper that contains the phrase: MEK " "phosphorylates ERK.") with open('test_abstract.txt', 'w') as f: f.write(abstract_txt) example_files.append('test_abstract.txt') # Get nxml content pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)), 'pmc_cont_example.nxml') if path.exists(pmc_test_fpath): example_files.append(pmc_test_fpath) assert len(example_files), "No content available to test." # Now read them. reader_classes = get_reader_classes() readers = [] for rc in reader_classes: readers.append(rc()) outputs = read_files(example_files, readers) N_out = len(outputs) proper_readers = [r for r in readers if not isinstance(r, EmptyReader)] N_exp = len(proper_readers)*len(example_files) assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
Fix the reading files test.
Fix the reading files test.
Python
bsd-2-clause
johnbachman/belpy,pvtodorov/indra,sorgerlab/indra,bgyori/indra,pvtodorov/indra,sorgerlab/belpy,johnbachman/indra,sorgerlab/indra,sorgerlab/belpy,sorgerlab/belpy,bgyori/indra,pvtodorov/indra,johnbachman/belpy,johnbachman/indra,johnbachman/indra,bgyori/indra,sorgerlab/indra,pvtodorov/indra,johnbachman/belpy
from os import path - from indra.tools.reading.read_files import read_files, get_readers + from indra.tools.reading.read_files import read_files, get_reader_classes from nose.plugins.attrib import attr + from indra.tools.reading.readers import EmptyReader + - @attr('slow', 'nonpublic') + @attr('slow', 'nonpublic', 'notravis') def test_read_files(): "Test that the system can read files." # Create the test files. example_files = [] # Get txt content abstract_txt = ("This is a paper that contains the phrase: MEK " "phosphorylates ERK.") with open('test_abstract.txt', 'w') as f: f.write(abstract_txt) example_files.append('test_abstract.txt') # Get nxml content pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)), 'pmc_cont_example.nxml') if path.exists(pmc_test_fpath): example_files.append(pmc_test_fpath) assert len(example_files), "No content available to test." # Now read them. - readers = get_readers() + reader_classes = get_reader_classes() + readers = [] + for rc in reader_classes: + readers.append(rc()) outputs = read_files(example_files, readers) N_out = len(outputs) + proper_readers = [r for r in readers if not isinstance(r, EmptyReader)] - N_exp = 2*len(example_files) + N_exp = len(proper_readers)*len(example_files) assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
Fix the reading files test.
## Code Before: from os import path from indra.tools.reading.read_files import read_files, get_readers from nose.plugins.attrib import attr @attr('slow', 'nonpublic') def test_read_files(): "Test that the system can read files." # Create the test files. example_files = [] # Get txt content abstract_txt = ("This is a paper that contains the phrase: MEK " "phosphorylates ERK.") with open('test_abstract.txt', 'w') as f: f.write(abstract_txt) example_files.append('test_abstract.txt') # Get nxml content pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)), 'pmc_cont_example.nxml') if path.exists(pmc_test_fpath): example_files.append(pmc_test_fpath) assert len(example_files), "No content available to test." # Now read them. readers = get_readers() outputs = read_files(example_files, readers) N_out = len(outputs) N_exp = 2*len(example_files) assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out) ## Instruction: Fix the reading files test. ## Code After: from os import path from indra.tools.reading.read_files import read_files, get_reader_classes from nose.plugins.attrib import attr from indra.tools.reading.readers import EmptyReader @attr('slow', 'nonpublic', 'notravis') def test_read_files(): "Test that the system can read files." # Create the test files. example_files = [] # Get txt content abstract_txt = ("This is a paper that contains the phrase: MEK " "phosphorylates ERK.") with open('test_abstract.txt', 'w') as f: f.write(abstract_txt) example_files.append('test_abstract.txt') # Get nxml content pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)), 'pmc_cont_example.nxml') if path.exists(pmc_test_fpath): example_files.append(pmc_test_fpath) assert len(example_files), "No content available to test." # Now read them. reader_classes = get_reader_classes() readers = [] for rc in reader_classes: readers.append(rc()) outputs = read_files(example_files, readers) N_out = len(outputs) proper_readers = [r for r in readers if not isinstance(r, EmptyReader)] N_exp = len(proper_readers)*len(example_files) assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
from os import path - from indra.tools.reading.read_files import read_files, get_readers + from indra.tools.reading.read_files import read_files, get_reader_classes ? ++++ +++ from nose.plugins.attrib import attr + from indra.tools.reading.readers import EmptyReader + - @attr('slow', 'nonpublic') + @attr('slow', 'nonpublic', 'notravis') ? ++++++++++++ def test_read_files(): "Test that the system can read files." # Create the test files. example_files = [] # Get txt content abstract_txt = ("This is a paper that contains the phrase: MEK " "phosphorylates ERK.") with open('test_abstract.txt', 'w') as f: f.write(abstract_txt) example_files.append('test_abstract.txt') # Get nxml content pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)), 'pmc_cont_example.nxml') if path.exists(pmc_test_fpath): example_files.append(pmc_test_fpath) assert len(example_files), "No content available to test." # Now read them. - readers = get_readers() + reader_classes = get_reader_classes() ? +++++++ +++++++ + readers = [] + for rc in reader_classes: + readers.append(rc()) outputs = read_files(example_files, readers) N_out = len(outputs) + proper_readers = [r for r in readers if not isinstance(r, EmptyReader)] - N_exp = 2*len(example_files) ? ^ + N_exp = len(proper_readers)*len(example_files) ? ^^^^^^^^^^^^^^^^^^^ assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
9fe573614e2f3ca9a6e738afb7f1af84b541092c
invertedindex.py
invertedindex.py
class InvertedIndex: def __init__(self): self.index = dict() def add_mail(self, mail): for key in ["simple_terms_body", "complexe_terms_body"]: for terms in mail[key]: if terms in self.index.keys(): self.index[terms].append((mail["name"], mail[key][terms])) else: self.index[terms] = list() self.index[terms].append((mail["name"], mail[key][terms]))
class InvertedIndex: def __init__(self): self.index = dict() def add_mail(self, mail): for key in ["simple_terms_body", "complexe_terms_body"]: for terms in mail[key]: if terms in self.index.keys(): self.index[terms].append((mail["name"], mail[key][terms])) else: self.index[terms] = list() self.index[terms].append((mail["name"], mail[key][terms])) def terms(self): for terms in self.index.keys(): yield terms def get_terms(self): return self.index.keys() def file_counter(self, terms): for val in self.index[terms]: yield val def get_file_counter(self, terms): return self.index.values() def file(self, terms): for val in file_counter(terms): yield val[0] def counter(self, terms): for val in file_counter(terms): yield val[1]
Add some access function to inverted index
Add some access function to inverted index
Python
mit
Nedgang/adt_project
class InvertedIndex: def __init__(self): self.index = dict() def add_mail(self, mail): for key in ["simple_terms_body", "complexe_terms_body"]: for terms in mail[key]: if terms in self.index.keys(): self.index[terms].append((mail["name"], mail[key][terms])) else: self.index[terms] = list() self.index[terms].append((mail["name"], mail[key][terms])) + def terms(self): + for terms in self.index.keys(): + yield terms + def get_terms(self): + return self.index.keys() + + def file_counter(self, terms): + for val in self.index[terms]: + yield val + + def get_file_counter(self, terms): + return self.index.values() + + def file(self, terms): + for val in file_counter(terms): + yield val[0] + + def counter(self, terms): + for val in file_counter(terms): + yield val[1] +
Add some access function to inverted index
## Code Before: class InvertedIndex: def __init__(self): self.index = dict() def add_mail(self, mail): for key in ["simple_terms_body", "complexe_terms_body"]: for terms in mail[key]: if terms in self.index.keys(): self.index[terms].append((mail["name"], mail[key][terms])) else: self.index[terms] = list() self.index[terms].append((mail["name"], mail[key][terms])) ## Instruction: Add some access function to inverted index ## Code After: class InvertedIndex: def __init__(self): self.index = dict() def add_mail(self, mail): for key in ["simple_terms_body", "complexe_terms_body"]: for terms in mail[key]: if terms in self.index.keys(): self.index[terms].append((mail["name"], mail[key][terms])) else: self.index[terms] = list() self.index[terms].append((mail["name"], mail[key][terms])) def terms(self): for terms in self.index.keys(): yield terms def get_terms(self): return self.index.keys() def file_counter(self, terms): for val in self.index[terms]: yield val def get_file_counter(self, terms): return self.index.values() def file(self, terms): for val in file_counter(terms): yield val[0] def counter(self, terms): for val in file_counter(terms): yield val[1]
class InvertedIndex: def __init__(self): self.index = dict() def add_mail(self, mail): for key in ["simple_terms_body", "complexe_terms_body"]: for terms in mail[key]: if terms in self.index.keys(): self.index[terms].append((mail["name"], mail[key][terms])) else: self.index[terms] = list() self.index[terms].append((mail["name"], mail[key][terms])) + def terms(self): + for terms in self.index.keys(): + yield terms + + def get_terms(self): + return self.index.keys() + + def file_counter(self, terms): + for val in self.index[terms]: + yield val + + def get_file_counter(self, terms): + return self.index.values() + + def file(self, terms): + for val in file_counter(terms): + yield val[0] + + def counter(self, terms): + for val in file_counter(terms): + yield val[1]
debe3a250a04986583589b1192cb6111b8b6c228
pydelhiconf/uix/screens/screenabout.py
pydelhiconf/uix/screens/screenabout.py
from kivy.uix.screenmanager import Screen from kivy.lang import Builder from kivy.factory import Factory from functools import partial import webbrowser class ScreenAbout(Screen): Builder.load_string(''' <ScreenAbout> spacing: dp(9) name: 'ScreenAbout' ScrollView id: scroll ScrollGrid AsyncImage id: imgbt allow_stretch: True size_hint_y: None height: dp(200) BackLabel id: comm_desc ''') def on_pre_enter(self): self.ids.scroll.opacity = 0 def on_enter(self, onsuccess=False): from network import get_data about = get_data('about', onsuccess=onsuccess) if not about: return about = about.get('0.0.1')[0] imbt = self.ids.imgbt imbt.source = about['logo'] imbt.on_released = partial(webbrowser.open, about['website']) self.ids.comm_desc.text = about['about'] Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
from kivy.uix.screenmanager import Screen from kivy.lang import Builder from kivy.factory import Factory from functools import partial import webbrowser class ScreenAbout(Screen): Builder.load_string(''' <ScreenAbout> spacing: dp(9) name: 'ScreenAbout' ScrollView id: scroll ScrollGrid AsyncImage id: imgbt allow_stretch: True size_hint_y: None height: dp(200) BackLabel id: comm_desc FloatLayout size_hint_y: None height: dp(45) ActiveButton id: but text: "Visit our website" size_hint: None, None width: dp(200) center_x: comm_desc.center_x top: comm_desc.y - dp(10) ''') def on_pre_enter(self): self.ids.scroll.opacity = 0 def on_enter(self, onsuccess=False): from network import get_data about = get_data('about', onsuccess=onsuccess) if not about: return about = about.get('0.0.1')[0] imbt = self.ids.imgbt imbt.source = about['logo'] self.ids.but.on_released = partial(webbrowser.open, about['website']) self.ids.comm_desc.text = about['about'] Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
Add button that links to website
Add button that links to website
Python
agpl-3.0
pydelhi/pydelhi_mobile,shivan1b/pydelhi_mobile,samukasmk/pythonbrasil_mobile,akshayaurora/PyDelhiMobile
from kivy.uix.screenmanager import Screen from kivy.lang import Builder from kivy.factory import Factory from functools import partial import webbrowser class ScreenAbout(Screen): Builder.load_string(''' <ScreenAbout> spacing: dp(9) name: 'ScreenAbout' ScrollView id: scroll ScrollGrid AsyncImage id: imgbt allow_stretch: True size_hint_y: None height: dp(200) BackLabel id: comm_desc + FloatLayout + size_hint_y: None + height: dp(45) + ActiveButton + id: but + text: "Visit our website" + size_hint: None, None + width: dp(200) + center_x: comm_desc.center_x + top: comm_desc.y - dp(10) ''') def on_pre_enter(self): self.ids.scroll.opacity = 0 def on_enter(self, onsuccess=False): from network import get_data about = get_data('about', onsuccess=onsuccess) if not about: return about = about.get('0.0.1')[0] imbt = self.ids.imgbt imbt.source = about['logo'] - imbt.on_released = partial(webbrowser.open, about['website']) + self.ids.but.on_released = partial(webbrowser.open, about['website']) self.ids.comm_desc.text = about['about'] Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
Add button that links to website
## Code Before: from kivy.uix.screenmanager import Screen from kivy.lang import Builder from kivy.factory import Factory from functools import partial import webbrowser class ScreenAbout(Screen): Builder.load_string(''' <ScreenAbout> spacing: dp(9) name: 'ScreenAbout' ScrollView id: scroll ScrollGrid AsyncImage id: imgbt allow_stretch: True size_hint_y: None height: dp(200) BackLabel id: comm_desc ''') def on_pre_enter(self): self.ids.scroll.opacity = 0 def on_enter(self, onsuccess=False): from network import get_data about = get_data('about', onsuccess=onsuccess) if not about: return about = about.get('0.0.1')[0] imbt = self.ids.imgbt imbt.source = about['logo'] imbt.on_released = partial(webbrowser.open, about['website']) self.ids.comm_desc.text = about['about'] Factory.Animation(opacity=1, d=.5).start(self.ids.scroll) ## Instruction: Add button that links to website ## Code After: from kivy.uix.screenmanager import Screen from kivy.lang import Builder from kivy.factory import Factory from functools import partial import webbrowser class ScreenAbout(Screen): Builder.load_string(''' <ScreenAbout> spacing: dp(9) name: 'ScreenAbout' ScrollView id: scroll ScrollGrid AsyncImage id: imgbt allow_stretch: True size_hint_y: None height: dp(200) BackLabel id: comm_desc FloatLayout size_hint_y: None height: dp(45) ActiveButton id: but text: "Visit our website" size_hint: None, None width: dp(200) center_x: comm_desc.center_x top: comm_desc.y - dp(10) ''') def on_pre_enter(self): self.ids.scroll.opacity = 0 def on_enter(self, onsuccess=False): from network import get_data about = get_data('about', onsuccess=onsuccess) if not about: return about = about.get('0.0.1')[0] imbt = self.ids.imgbt imbt.source = about['logo'] self.ids.but.on_released = partial(webbrowser.open, about['website']) self.ids.comm_desc.text = about['about'] Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
from kivy.uix.screenmanager import Screen from kivy.lang import Builder from kivy.factory import Factory from functools import partial import webbrowser class ScreenAbout(Screen): Builder.load_string(''' <ScreenAbout> spacing: dp(9) name: 'ScreenAbout' ScrollView id: scroll ScrollGrid AsyncImage id: imgbt allow_stretch: True size_hint_y: None height: dp(200) BackLabel id: comm_desc + FloatLayout + size_hint_y: None + height: dp(45) + ActiveButton + id: but + text: "Visit our website" + size_hint: None, None + width: dp(200) + center_x: comm_desc.center_x + top: comm_desc.y - dp(10) ''') def on_pre_enter(self): self.ids.scroll.opacity = 0 def on_enter(self, onsuccess=False): from network import get_data about = get_data('about', onsuccess=onsuccess) if not about: return about = about.get('0.0.1')[0] imbt = self.ids.imgbt imbt.source = about['logo'] - imbt.on_released = partial(webbrowser.open, about['website']) ? ^ + self.ids.but.on_released = partial(webbrowser.open, about['website']) ? +++++ ^^^ + self.ids.comm_desc.text = about['about'] Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
460a2430fbd8832f3fada1a74b754d71a27ac282
mockingjay/matcher.py
mockingjay/matcher.py
import abc import re class Matcher(object): __metaclass__ = abc.ABCMeta @abc.abstractmethod def assert_request_matched(self, request): """ Assert that the request matched the spec in this matcher object. """ class HeaderMatcher(Matcher): """ Matcher for the request's header. :param key: the name of the header :param value: the value of the header """ def __init__(self, key, value): self.key = key self.value = value def assert_request_matched(self, request): assert request.headers.get(self.key) == self.value class BodyMatcher(Matcher): """ Matcher for the request body. :param body: can either be a string or a :class:`_sre.SRE_Pattern`: object """ def __init__(self, body): self.body = body def assert_request_matched(self, request): if isinstance(self.body, re._pattern_type): assert self.body.search(request.body) is not None else: assert request.body == self.body
import abc import re class StringOrPattern(object): """ A decorator object that wraps a string or a regex pattern so that it can be compared against another string either literally or using the pattern. """ def __init__(self, subject): self.subject = subject def __eq__(self, other_str): if isinstance(self.subject, re._pattern_type): return self.subject.search(other_str) is not None else: return self.subject == other_str def __hash__(self): return self.subject.__hash__() class Matcher(object): __metaclass__ = abc.ABCMeta @abc.abstractmethod def assert_request_matched(self, request): """ Assert that the request matched the spec in this matcher object. """ class HeaderMatcher(Matcher): """ Matcher for the request's header. :param key: the name of the header :param value: the value of the header """ def __init__(self, key, value): self.key = key self.value = StringOrPattern(value) def assert_request_matched(self, request): assert request.headers.get(self.key) == self.value class BodyMatcher(Matcher): """ Matcher for the request body. :param body: can either be a string or a :class:`_sre.SRE_Pattern`: object """ def __init__(self, body): self.body = StringOrPattern(body) def assert_request_matched(self, request): assert request.body == self.body
Allow all values to be compared with either literally or with a pattern
Allow all values to be compared with either literally or with a pattern
Python
bsd-3-clause
kevinjqiu/mockingjay
import abc import re + + + class StringOrPattern(object): + """ + A decorator object that wraps a string or a regex pattern so that it can + be compared against another string either literally or using the pattern. + """ + def __init__(self, subject): + self.subject = subject + + def __eq__(self, other_str): + if isinstance(self.subject, re._pattern_type): + return self.subject.search(other_str) is not None + else: + return self.subject == other_str + + def __hash__(self): + return self.subject.__hash__() class Matcher(object): __metaclass__ = abc.ABCMeta @abc.abstractmethod def assert_request_matched(self, request): """ Assert that the request matched the spec in this matcher object. """ class HeaderMatcher(Matcher): """ Matcher for the request's header. :param key: the name of the header :param value: the value of the header """ def __init__(self, key, value): self.key = key - self.value = value + self.value = StringOrPattern(value) def assert_request_matched(self, request): assert request.headers.get(self.key) == self.value class BodyMatcher(Matcher): """ Matcher for the request body. :param body: can either be a string or a :class:`_sre.SRE_Pattern`: object """ def __init__(self, body): - self.body = body + self.body = StringOrPattern(body) def assert_request_matched(self, request): - if isinstance(self.body, re._pattern_type): - assert self.body.search(request.body) is not None - else: - assert request.body == self.body + assert request.body == self.body
Allow all values to be compared with either literally or with a pattern
## Code Before: import abc import re class Matcher(object): __metaclass__ = abc.ABCMeta @abc.abstractmethod def assert_request_matched(self, request): """ Assert that the request matched the spec in this matcher object. """ class HeaderMatcher(Matcher): """ Matcher for the request's header. :param key: the name of the header :param value: the value of the header """ def __init__(self, key, value): self.key = key self.value = value def assert_request_matched(self, request): assert request.headers.get(self.key) == self.value class BodyMatcher(Matcher): """ Matcher for the request body. :param body: can either be a string or a :class:`_sre.SRE_Pattern`: object """ def __init__(self, body): self.body = body def assert_request_matched(self, request): if isinstance(self.body, re._pattern_type): assert self.body.search(request.body) is not None else: assert request.body == self.body ## Instruction: Allow all values to be compared with either literally or with a pattern ## Code After: import abc import re class StringOrPattern(object): """ A decorator object that wraps a string or a regex pattern so that it can be compared against another string either literally or using the pattern. """ def __init__(self, subject): self.subject = subject def __eq__(self, other_str): if isinstance(self.subject, re._pattern_type): return self.subject.search(other_str) is not None else: return self.subject == other_str def __hash__(self): return self.subject.__hash__() class Matcher(object): __metaclass__ = abc.ABCMeta @abc.abstractmethod def assert_request_matched(self, request): """ Assert that the request matched the spec in this matcher object. """ class HeaderMatcher(Matcher): """ Matcher for the request's header. :param key: the name of the header :param value: the value of the header """ def __init__(self, key, value): self.key = key self.value = StringOrPattern(value) def assert_request_matched(self, request): assert request.headers.get(self.key) == self.value class BodyMatcher(Matcher): """ Matcher for the request body. :param body: can either be a string or a :class:`_sre.SRE_Pattern`: object """ def __init__(self, body): self.body = StringOrPattern(body) def assert_request_matched(self, request): assert request.body == self.body
import abc import re + + + class StringOrPattern(object): + """ + A decorator object that wraps a string or a regex pattern so that it can + be compared against another string either literally or using the pattern. + """ + def __init__(self, subject): + self.subject = subject + + def __eq__(self, other_str): + if isinstance(self.subject, re._pattern_type): + return self.subject.search(other_str) is not None + else: + return self.subject == other_str + + def __hash__(self): + return self.subject.__hash__() class Matcher(object): __metaclass__ = abc.ABCMeta @abc.abstractmethod def assert_request_matched(self, request): """ Assert that the request matched the spec in this matcher object. """ class HeaderMatcher(Matcher): """ Matcher for the request's header. :param key: the name of the header :param value: the value of the header """ def __init__(self, key, value): self.key = key - self.value = value + self.value = StringOrPattern(value) ? ++++++++++++++++ + def assert_request_matched(self, request): assert request.headers.get(self.key) == self.value class BodyMatcher(Matcher): """ Matcher for the request body. :param body: can either be a string or a :class:`_sre.SRE_Pattern`: object """ def __init__(self, body): - self.body = body + self.body = StringOrPattern(body) def assert_request_matched(self, request): - if isinstance(self.body, re._pattern_type): - assert self.body.search(request.body) is not None - else: - assert request.body == self.body ? ---- + assert request.body == self.body
95186f684328d5b84611f405d47d474c53cad619
cat.py
cat.py
import io import aiohttp from discord.ext import commands import yaml class Cat: def __init__(self, bot): self.bot = bot with open('config.yaml') as file: data = yaml.load(file) self.key = data.get('cat_key', '') self.url = 'http://thecatapi.com/api/images/get' self.params = {'api_key': self.key, 'type': 'png,jpg'} @commands.command() async def cat(self, ctx): s = self.bot.session async with ctx.typing(), s.get(self.url, params=self.params) as resp: image = io.BytesIO(await resp.content.read()) ext = resp.headers['Content-Type'].partition('/')[2] await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}') def setup(bot): bot.add_cog(Cat(bot))
import io import aiohttp import discord from discord.ext import commands from lxml import etree import yaml class Cat: def __init__(self, bot): self.bot = bot with open('config.yaml') as file: data = yaml.load(file) self.key = data.get('cat_key', '') self.url = 'http://thecatapi.com/api/images/get' self.params = {'api_key': self.key, 'type': 'png,jpg', 'format': 'xml', } @commands.command() async def cat(self, ctx): session = self.bot.session async with ctx.typing(): async with session.get(self.url, params=self.params) as resp: root = etree.fromstring(await resp.text()) url = root.find('.//url').text embed = discord.Embed() embed.set_image(url=url) await ctx.send(embed=embed) def setup(bot): bot.add_cog(Cat(bot))
Send image in embed because aiohttp doesn't know how to parse links
Send image in embed because aiohttp doesn't know how to parse links
Python
mit
BeatButton/beattie,BeatButton/beattie-bot
import io import aiohttp + import discord from discord.ext import commands + from lxml import etree import yaml class Cat: def __init__(self, bot): self.bot = bot with open('config.yaml') as file: data = yaml.load(file) self.key = data.get('cat_key', '') self.url = 'http://thecatapi.com/api/images/get' self.params = {'api_key': self.key, - 'type': 'png,jpg'} + 'type': 'png,jpg', + 'format': 'xml', + } @commands.command() async def cat(self, ctx): - s = self.bot.session + session = self.bot.session + async with ctx.typing(): - async with ctx.typing(), s.get(self.url, params=self.params) as resp: + async with session.get(self.url, params=self.params) as resp: - image = io.BytesIO(await resp.content.read()) - ext = resp.headers['Content-Type'].partition('/')[2] - await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}') + root = etree.fromstring(await resp.text()) + url = root.find('.//url').text + embed = discord.Embed() + embed.set_image(url=url) + await ctx.send(embed=embed) + def setup(bot): bot.add_cog(Cat(bot))
Send image in embed because aiohttp doesn't know how to parse links
## Code Before: import io import aiohttp from discord.ext import commands import yaml class Cat: def __init__(self, bot): self.bot = bot with open('config.yaml') as file: data = yaml.load(file) self.key = data.get('cat_key', '') self.url = 'http://thecatapi.com/api/images/get' self.params = {'api_key': self.key, 'type': 'png,jpg'} @commands.command() async def cat(self, ctx): s = self.bot.session async with ctx.typing(), s.get(self.url, params=self.params) as resp: image = io.BytesIO(await resp.content.read()) ext = resp.headers['Content-Type'].partition('/')[2] await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}') def setup(bot): bot.add_cog(Cat(bot)) ## Instruction: Send image in embed because aiohttp doesn't know how to parse links ## Code After: import io import aiohttp import discord from discord.ext import commands from lxml import etree import yaml class Cat: def __init__(self, bot): self.bot = bot with open('config.yaml') as file: data = yaml.load(file) self.key = data.get('cat_key', '') self.url = 'http://thecatapi.com/api/images/get' self.params = {'api_key': self.key, 'type': 'png,jpg', 'format': 'xml', } @commands.command() async def cat(self, ctx): session = self.bot.session async with ctx.typing(): async with session.get(self.url, params=self.params) as resp: root = etree.fromstring(await resp.text()) url = root.find('.//url').text embed = discord.Embed() embed.set_image(url=url) await ctx.send(embed=embed) def setup(bot): bot.add_cog(Cat(bot))
import io import aiohttp + import discord from discord.ext import commands + from lxml import etree import yaml class Cat: def __init__(self, bot): self.bot = bot with open('config.yaml') as file: data = yaml.load(file) self.key = data.get('cat_key', '') self.url = 'http://thecatapi.com/api/images/get' self.params = {'api_key': self.key, - 'type': 'png,jpg'} ? ^ + 'type': 'png,jpg', ? ^ + 'format': 'xml', + } @commands.command() async def cat(self, ctx): - s = self.bot.session + session = self.bot.session ? ++++++ + async with ctx.typing(): - async with ctx.typing(), s.get(self.url, params=self.params) as resp: ? ^^^^^^^ ------ + async with session.get(self.url, params=self.params) as resp: ? ++++ ^^^^ + - image = io.BytesIO(await resp.content.read()) - ext = resp.headers['Content-Type'].partition('/')[2] - await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}') + root = etree.fromstring(await resp.text()) + url = root.find('.//url').text + embed = discord.Embed() + embed.set_image(url=url) + await ctx.send(embed=embed) + def setup(bot): bot.add_cog(Cat(bot))
0db43d894bfb419a7f4b538f755af47fc0b653cb
tests/unit/test_sharpspring.py
tests/unit/test_sharpspring.py
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post") def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "[email protected]", "companyName": "Test Company", } sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called()
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class MockResponse: def __init__(self, json_data, status_code): self.json_data = json_data self.status_code = status_code def raise_for_status(self): pass def json(self): return self.json_data def mocked_requests_post_success(*args, **kwargs): return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "[email protected]", "companyName": "Test Company", } result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() self.assertTrue(result["result"]["creates"][0])
Add mock response to sharpspring test
Add mock response to sharpspring test
Python
apache-2.0
Code4SA/pmg-cms-2,Code4SA/pmg-cms-2,Code4SA/pmg-cms-2
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase + class MockResponse: + def __init__(self, json_data, status_code): + self.json_data = json_data + self.status_code = status_code + + def raise_for_status(self): + pass + + def json(self): + return self.json_data + + + def mocked_requests_post_success(*args, **kwargs): + + return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) + + class TestSharpspring(PMGTestCase): - @patch("pmg.sharpspring.requests.post") + @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "[email protected]", "companyName": "Test Company", } - sharpspring.call("createLeads", {"objects": [details]}) + result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() + self.assertTrue(result["result"]["creates"][0])
Add mock response to sharpspring test
## Code Before: from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post") def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "[email protected]", "companyName": "Test Company", } sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() ## Instruction: Add mock response to sharpspring test ## Code After: from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class MockResponse: def __init__(self, json_data, status_code): self.json_data = json_data self.status_code = status_code def raise_for_status(self): pass def json(self): return self.json_data def mocked_requests_post_success(*args, **kwargs): return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "[email protected]", "companyName": "Test Company", } result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() self.assertTrue(result["result"]["creates"][0])
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase + class MockResponse: + def __init__(self, json_data, status_code): + self.json_data = json_data + self.status_code = status_code + + def raise_for_status(self): + pass + + def json(self): + return self.json_data + + + def mocked_requests_post_success(*args, **kwargs): + + return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) + + class TestSharpspring(PMGTestCase): - @patch("pmg.sharpspring.requests.post") + @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "[email protected]", "companyName": "Test Company", } - sharpspring.call("createLeads", {"objects": [details]}) + result = sharpspring.call("createLeads", {"objects": [details]}) ? +++++++++ post_mock.assert_called() + self.assertTrue(result["result"]["creates"][0])
3055fa16010a1b855142c2e5b866d76daee17c8f
markdown_gen/test/attributes_test.py
markdown_gen/test/attributes_test.py
import unittest import markdown_gen.MardownGen as md class AttributesTests(unittest.TestCase): def test_italic(self): expected = "*italic text*" self.assertEqual(expected, md.gen_italic("italic text")) expected = "_italic text alternative_" self.assertEqual(expected, md.gen_italic("italic text alternative", True)) def test_bold(self): expected = "**bold text**" self.assertEqual(expected, md.gen_bold("bold text")) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_monspace(self): expected = "`monospace`" self.assertEqual(expected, md.gen_monospace("monospace")) def test_strikethrough(self): expected = "~~strikethrough~~" self.assertEqual(expected, md.gen_strikethrough("strikethrough")) if __name__ == '__main__': unittest.main()
import unittest import markdown_gen.MardownGen as md class AttributesTests(unittest.TestCase): def test_italic(self): expected = "*italic text*" self.assertEqual(expected, md.gen_italic("italic text")) expected = "_italic text alternative_" self.assertEqual(expected, md.gen_italic("italic text alternative", True)) def test_bold(self): expected = "**bold text**" self.assertEqual(expected, md.gen_bold("bold text")) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_bold_and_italic(self): expected = "***bold and italic text***" self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text"))) self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text"))) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_monspace(self): expected = "`monospace`" self.assertEqual(expected, md.gen_monospace("monospace")) def test_strikethrough(self): expected = "~~strikethrough~~" self.assertEqual(expected, md.gen_strikethrough("strikethrough")) if __name__ == '__main__': unittest.main()
Add test for bold and italic text
Add test for bold and italic text
Python
epl-1.0
LukasWoodtli/PyMarkdownGen
import unittest import markdown_gen.MardownGen as md class AttributesTests(unittest.TestCase): def test_italic(self): expected = "*italic text*" self.assertEqual(expected, md.gen_italic("italic text")) expected = "_italic text alternative_" self.assertEqual(expected, md.gen_italic("italic text alternative", True)) def test_bold(self): expected = "**bold text**" self.assertEqual(expected, md.gen_bold("bold text")) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) + + def test_bold_and_italic(self): + expected = "***bold and italic text***" + self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text"))) + self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text"))) + + expected = "__bold text alternative__" + self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_monspace(self): expected = "`monospace`" self.assertEqual(expected, md.gen_monospace("monospace")) def test_strikethrough(self): expected = "~~strikethrough~~" self.assertEqual(expected, md.gen_strikethrough("strikethrough")) if __name__ == '__main__': unittest.main()
Add test for bold and italic text
## Code Before: import unittest import markdown_gen.MardownGen as md class AttributesTests(unittest.TestCase): def test_italic(self): expected = "*italic text*" self.assertEqual(expected, md.gen_italic("italic text")) expected = "_italic text alternative_" self.assertEqual(expected, md.gen_italic("italic text alternative", True)) def test_bold(self): expected = "**bold text**" self.assertEqual(expected, md.gen_bold("bold text")) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_monspace(self): expected = "`monospace`" self.assertEqual(expected, md.gen_monospace("monospace")) def test_strikethrough(self): expected = "~~strikethrough~~" self.assertEqual(expected, md.gen_strikethrough("strikethrough")) if __name__ == '__main__': unittest.main() ## Instruction: Add test for bold and italic text ## Code After: import unittest import markdown_gen.MardownGen as md class AttributesTests(unittest.TestCase): def test_italic(self): expected = "*italic text*" self.assertEqual(expected, md.gen_italic("italic text")) expected = "_italic text alternative_" self.assertEqual(expected, md.gen_italic("italic text alternative", True)) def test_bold(self): expected = "**bold text**" self.assertEqual(expected, md.gen_bold("bold text")) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_bold_and_italic(self): expected = "***bold and italic text***" self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text"))) self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text"))) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_monspace(self): expected = "`monospace`" self.assertEqual(expected, md.gen_monospace("monospace")) def test_strikethrough(self): expected = "~~strikethrough~~" self.assertEqual(expected, md.gen_strikethrough("strikethrough")) if __name__ == '__main__': unittest.main()
import unittest import markdown_gen.MardownGen as md class AttributesTests(unittest.TestCase): def test_italic(self): expected = "*italic text*" self.assertEqual(expected, md.gen_italic("italic text")) expected = "_italic text alternative_" self.assertEqual(expected, md.gen_italic("italic text alternative", True)) def test_bold(self): expected = "**bold text**" self.assertEqual(expected, md.gen_bold("bold text")) expected = "__bold text alternative__" self.assertEqual(expected, md.gen_bold("bold text alternative", True)) + + def test_bold_and_italic(self): + expected = "***bold and italic text***" + self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text"))) + self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text"))) + + expected = "__bold text alternative__" + self.assertEqual(expected, md.gen_bold("bold text alternative", True)) def test_monspace(self): expected = "`monospace`" self.assertEqual(expected, md.gen_monospace("monospace")) def test_strikethrough(self): expected = "~~strikethrough~~" self.assertEqual(expected, md.gen_strikethrough("strikethrough")) if __name__ == '__main__': unittest.main()
0b82a5c10a9e728f6f5424429a70fd2951c9b5c5
pythonmisp/__init__.py
pythonmisp/__init__.py
from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute
from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError
Add MispTransportError in package import
Add MispTransportError in package import
Python
apache-2.0
nbareil/python-misp
- from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute + from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError
Add MispTransportError in package import
## Code Before: from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute ## Instruction: Add MispTransportError in package import ## Code After: from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError
- from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute + from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError ? ++++++++++++++++++++
110c362e3e8436700707c2306d115b3b2476a79d
core/models.py
core/models.py
from os import makedirs from os.path import join, exists from urllib import urlretrieve from django.conf import settings from social_auth.signals import socialauth_registered def create_profile(sender, user, response, details, **kwargs): try: # twitter photo_url = response["profile_image_url"] photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1)) except KeyError: # facebook photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"] path = join(settings.MEDIA_ROOT, "photos") if not exists(path): makedirs(path) urlretrieve(photo_url, join(path, str(user.id))) socialauth_registered.connect(create_profile, sender=None)
from os import makedirs from os.path import join, exists from urllib import urlretrieve from django.conf import settings from django.contrib.auth.models import User from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from social_auth.signals import socialauth_registered class Account(models.Model): """ A user's account balance. """ user = models.OneToOneField(User) balance = models.IntegerField(default=5000) @receiver(post_save, sender=User) def user_saved(sender, **kwargs): """ Create an initial account balance for new users. """ Account.objects.get_or_create(user=kwargs["instance"]) @receiver(socialauth_registered, sender=None) def avatar(sender, user, response, details, **kwargs): """ Download the user's Twitter or Facebook avatar once they've authenticated via either service. """ try: # twitter photo_url = response["profile_image_url"] photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1)) except KeyError: # facebook photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"] path = join(settings.MEDIA_ROOT, "photos") if not exists(path): makedirs(path) urlretrieve(photo_url, join(path, str(user.id)))
Add initial account balance for users.
Add initial account balance for users.
Python
bsd-2-clause
stephenmcd/gamblor,stephenmcd/gamblor
from os import makedirs from os.path import join, exists from urllib import urlretrieve from django.conf import settings + from django.contrib.auth.models import User + from django.db import models + from django.db.models.signals import post_save + from django.dispatch import receiver from social_auth.signals import socialauth_registered + class Account(models.Model): + """ + A user's account balance. + """ + user = models.OneToOneField(User) + balance = models.IntegerField(default=5000) + + + @receiver(post_save, sender=User) + def user_saved(sender, **kwargs): + """ + Create an initial account balance for new users. + """ + Account.objects.get_or_create(user=kwargs["instance"]) + + + @receiver(socialauth_registered, sender=None) - def create_profile(sender, user, response, details, **kwargs): + def avatar(sender, user, response, details, **kwargs): + """ + Download the user's Twitter or Facebook avatar once they've + authenticated via either service. + """ try: # twitter photo_url = response["profile_image_url"] photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1)) except KeyError: # facebook photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"] path = join(settings.MEDIA_ROOT, "photos") if not exists(path): makedirs(path) urlretrieve(photo_url, join(path, str(user.id))) - socialauth_registered.connect(create_profile, sender=None) -
Add initial account balance for users.
## Code Before: from os import makedirs from os.path import join, exists from urllib import urlretrieve from django.conf import settings from social_auth.signals import socialauth_registered def create_profile(sender, user, response, details, **kwargs): try: # twitter photo_url = response["profile_image_url"] photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1)) except KeyError: # facebook photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"] path = join(settings.MEDIA_ROOT, "photos") if not exists(path): makedirs(path) urlretrieve(photo_url, join(path, str(user.id))) socialauth_registered.connect(create_profile, sender=None) ## Instruction: Add initial account balance for users. ## Code After: from os import makedirs from os.path import join, exists from urllib import urlretrieve from django.conf import settings from django.contrib.auth.models import User from django.db import models from django.db.models.signals import post_save from django.dispatch import receiver from social_auth.signals import socialauth_registered class Account(models.Model): """ A user's account balance. """ user = models.OneToOneField(User) balance = models.IntegerField(default=5000) @receiver(post_save, sender=User) def user_saved(sender, **kwargs): """ Create an initial account balance for new users. """ Account.objects.get_or_create(user=kwargs["instance"]) @receiver(socialauth_registered, sender=None) def avatar(sender, user, response, details, **kwargs): """ Download the user's Twitter or Facebook avatar once they've authenticated via either service. """ try: # twitter photo_url = response["profile_image_url"] photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1)) except KeyError: # facebook photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"] path = join(settings.MEDIA_ROOT, "photos") if not exists(path): makedirs(path) urlretrieve(photo_url, join(path, str(user.id)))
from os import makedirs from os.path import join, exists from urllib import urlretrieve from django.conf import settings + from django.contrib.auth.models import User + from django.db import models + from django.db.models.signals import post_save + from django.dispatch import receiver from social_auth.signals import socialauth_registered + class Account(models.Model): + """ + A user's account balance. + """ + user = models.OneToOneField(User) + balance = models.IntegerField(default=5000) + + + @receiver(post_save, sender=User) + def user_saved(sender, **kwargs): + """ + Create an initial account balance for new users. + """ + Account.objects.get_or_create(user=kwargs["instance"]) + + + @receiver(socialauth_registered, sender=None) - def create_profile(sender, user, response, details, **kwargs): ? ^^^ ^^^ ----- + def avatar(sender, user, response, details, **kwargs): ? ^^ ^ + """ + Download the user's Twitter or Facebook avatar once they've + authenticated via either service. + """ try: # twitter photo_url = response["profile_image_url"] photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1)) except KeyError: # facebook photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"] path = join(settings.MEDIA_ROOT, "photos") if not exists(path): makedirs(path) urlretrieve(photo_url, join(path, str(user.id))) - - socialauth_registered.connect(create_profile, sender=None)
7198133cf9d24f3d29d300366951b7eac8b2547f
alburnum/maas/viscera/users.py
alburnum/maas/viscera/users.py
"""Objects for users.""" __all__ = [ "User", "Users", ] from . import ( check, Object, ObjectField, ObjectSet, ObjectType, ) class UsersType(ObjectType): """Metaclass for `Users`.""" def __iter__(cls): return map(cls._object, cls._handler.read()) def create(cls, username, email, password, *, is_superuser=False): data = cls._handler.create( username=username, email=email, password=password, is_superuser='1' if is_superuser else '0') return cls._object(data) class Users(ObjectSet, metaclass=UsersType): """The set of users.""" @classmethod def read(cls): return cls(cls) class User(Object): """A user.""" username = ObjectField.Checked( "username", check(str), check(str)) email = ObjectField.Checked( "email", check(str), check(str)) is_admin = ObjectField.Checked( "is_superuser", check(bool), check(bool))
"""Objects for users.""" __all__ = [ "User", "Users", ] from . import ( check, Object, ObjectField, ObjectSet, ObjectType, ) class UsersType(ObjectType): """Metaclass for `Users`.""" def __iter__(cls): return map(cls._object, cls._handler.read()) def create(cls, username, password, *, email=None, is_superuser=False): if email is None: email = "%[email protected]" % username data = cls._handler.create( username=username, email=email, password=password, is_superuser='1' if is_superuser else '0') return cls._object(data) class Users(ObjectSet, metaclass=UsersType): """The set of users.""" @classmethod def read(cls): return cls(cls) class User(Object): """A user.""" username = ObjectField.Checked( "username", check(str), check(str)) email = ObjectField.Checked( "email", check(str), check(str)) is_superuser = ObjectField.Checked( "is_superuser", check(bool), check(bool))
Change to is_superuser and make email optional.
Change to is_superuser and make email optional.
Python
agpl-3.0
blakerouse/python-libmaas,alburnum/alburnum-maas-client
"""Objects for users.""" __all__ = [ "User", "Users", ] from . import ( check, Object, ObjectField, ObjectSet, ObjectType, ) class UsersType(ObjectType): """Metaclass for `Users`.""" def __iter__(cls): return map(cls._object, cls._handler.read()) - def create(cls, username, email, password, *, is_superuser=False): + def create(cls, username, password, *, email=None, is_superuser=False): + if email is None: + email = "%[email protected]" % username data = cls._handler.create( username=username, email=email, password=password, is_superuser='1' if is_superuser else '0') return cls._object(data) class Users(ObjectSet, metaclass=UsersType): """The set of users.""" @classmethod def read(cls): return cls(cls) class User(Object): """A user.""" username = ObjectField.Checked( "username", check(str), check(str)) email = ObjectField.Checked( "email", check(str), check(str)) - is_admin = ObjectField.Checked( + is_superuser = ObjectField.Checked( "is_superuser", check(bool), check(bool))
Change to is_superuser and make email optional.
## Code Before: """Objects for users.""" __all__ = [ "User", "Users", ] from . import ( check, Object, ObjectField, ObjectSet, ObjectType, ) class UsersType(ObjectType): """Metaclass for `Users`.""" def __iter__(cls): return map(cls._object, cls._handler.read()) def create(cls, username, email, password, *, is_superuser=False): data = cls._handler.create( username=username, email=email, password=password, is_superuser='1' if is_superuser else '0') return cls._object(data) class Users(ObjectSet, metaclass=UsersType): """The set of users.""" @classmethod def read(cls): return cls(cls) class User(Object): """A user.""" username = ObjectField.Checked( "username", check(str), check(str)) email = ObjectField.Checked( "email", check(str), check(str)) is_admin = ObjectField.Checked( "is_superuser", check(bool), check(bool)) ## Instruction: Change to is_superuser and make email optional. ## Code After: """Objects for users.""" __all__ = [ "User", "Users", ] from . import ( check, Object, ObjectField, ObjectSet, ObjectType, ) class UsersType(ObjectType): """Metaclass for `Users`.""" def __iter__(cls): return map(cls._object, cls._handler.read()) def create(cls, username, password, *, email=None, is_superuser=False): if email is None: email = "%[email protected]" % username data = cls._handler.create( username=username, email=email, password=password, is_superuser='1' if is_superuser else '0') return cls._object(data) class Users(ObjectSet, metaclass=UsersType): """The set of users.""" @classmethod def read(cls): return cls(cls) class User(Object): """A user.""" username = ObjectField.Checked( "username", check(str), check(str)) email = ObjectField.Checked( "email", check(str), check(str)) is_superuser = ObjectField.Checked( "is_superuser", check(bool), check(bool))
"""Objects for users.""" __all__ = [ "User", "Users", ] from . import ( check, Object, ObjectField, ObjectSet, ObjectType, ) class UsersType(ObjectType): """Metaclass for `Users`.""" def __iter__(cls): return map(cls._object, cls._handler.read()) - def create(cls, username, email, password, *, is_superuser=False): ? ------- + def create(cls, username, password, *, email=None, is_superuser=False): ? ++++++++++++ + if email is None: + email = "%[email protected]" % username data = cls._handler.create( username=username, email=email, password=password, is_superuser='1' if is_superuser else '0') return cls._object(data) class Users(ObjectSet, metaclass=UsersType): """The set of users.""" @classmethod def read(cls): return cls(cls) class User(Object): """A user.""" username = ObjectField.Checked( "username", check(str), check(str)) email = ObjectField.Checked( "email", check(str), check(str)) - is_admin = ObjectField.Checked( ? ^^^^^ + is_superuser = ObjectField.Checked( ? ^^^^^^^^^ "is_superuser", check(bool), check(bool))
cc92b1770acdc5a34eb32c596c0b2ece6bf32b0f
qiprofile_rest/server/settings.py
qiprofile_rest/server/settings.py
import os # The run environment default is production. # Modify this by setting the NODE_ENV environment variable. env = os.getenv('NODE_ENV') or 'production' # The MongoDB database. if env == 'production': MONGO_DBNAME = 'qiprofile' else: MONGO_DBNAME = 'qiprofile_test' # Even though the domain is defined by the Eve MongoEngine # adapter, a DOMAIN setting is required by Eve. This setting # is only used to avoid an Eve complaint about a missing domain. DOMAIN = {'eve-mongoengine': {}}
"""This ``settings`` file specifies the Eve configuration.""" import os # The run environment default is production. # Modify this by setting the NODE_ENV environment variable. env = os.getenv('NODE_ENV') or 'production' # The MongoDB database. if env == 'production': MONGO_DBNAME = 'qiprofile' else: MONGO_DBNAME = 'qiprofile_test' # The MongoDB host default is localhost, but can be reset # by the MONGO_HOST environment variable. host = os.getenv('MONGO_HOST') if host: MONGO_HOST = host # Even though the domain is defined by the Eve MongoEngine # adapter, a DOMAIN setting is required by Eve. This setting # is only used to avoid an Eve complaint about a missing domain. DOMAIN = {'eve-mongoengine': {}}
Allow MONGO_HOST env var override.
Allow MONGO_HOST env var override.
Python
bsd-2-clause
ohsu-qin/qirest,ohsu-qin/qiprofile-rest
+ """This ``settings`` file specifies the Eve configuration.""" + import os # The run environment default is production. # Modify this by setting the NODE_ENV environment variable. env = os.getenv('NODE_ENV') or 'production' # The MongoDB database. if env == 'production': - MONGO_DBNAME = 'qiprofile' + MONGO_DBNAME = 'qiprofile' else: - MONGO_DBNAME = 'qiprofile_test' + MONGO_DBNAME = 'qiprofile_test' + + # The MongoDB host default is localhost, but can be reset + # by the MONGO_HOST environment variable. + host = os.getenv('MONGO_HOST') + if host: + MONGO_HOST = host # Even though the domain is defined by the Eve MongoEngine # adapter, a DOMAIN setting is required by Eve. This setting # is only used to avoid an Eve complaint about a missing domain. DOMAIN = {'eve-mongoengine': {}}
Allow MONGO_HOST env var override.
## Code Before: import os # The run environment default is production. # Modify this by setting the NODE_ENV environment variable. env = os.getenv('NODE_ENV') or 'production' # The MongoDB database. if env == 'production': MONGO_DBNAME = 'qiprofile' else: MONGO_DBNAME = 'qiprofile_test' # Even though the domain is defined by the Eve MongoEngine # adapter, a DOMAIN setting is required by Eve. This setting # is only used to avoid an Eve complaint about a missing domain. DOMAIN = {'eve-mongoengine': {}} ## Instruction: Allow MONGO_HOST env var override. ## Code After: """This ``settings`` file specifies the Eve configuration.""" import os # The run environment default is production. # Modify this by setting the NODE_ENV environment variable. env = os.getenv('NODE_ENV') or 'production' # The MongoDB database. if env == 'production': MONGO_DBNAME = 'qiprofile' else: MONGO_DBNAME = 'qiprofile_test' # The MongoDB host default is localhost, but can be reset # by the MONGO_HOST environment variable. host = os.getenv('MONGO_HOST') if host: MONGO_HOST = host # Even though the domain is defined by the Eve MongoEngine # adapter, a DOMAIN setting is required by Eve. This setting # is only used to avoid an Eve complaint about a missing domain. DOMAIN = {'eve-mongoengine': {}}
+ """This ``settings`` file specifies the Eve configuration.""" + import os # The run environment default is production. # Modify this by setting the NODE_ENV environment variable. env = os.getenv('NODE_ENV') or 'production' # The MongoDB database. if env == 'production': - MONGO_DBNAME = 'qiprofile' + MONGO_DBNAME = 'qiprofile' ? ++ else: - MONGO_DBNAME = 'qiprofile_test' + MONGO_DBNAME = 'qiprofile_test' ? ++ + + # The MongoDB host default is localhost, but can be reset + # by the MONGO_HOST environment variable. + host = os.getenv('MONGO_HOST') + if host: + MONGO_HOST = host # Even though the domain is defined by the Eve MongoEngine # adapter, a DOMAIN setting is required by Eve. This setting # is only used to avoid an Eve complaint about a missing domain. DOMAIN = {'eve-mongoengine': {}}
a28a29af31b1ea604ed97544b2d84a39c9ba3e7b
automation/src/rabird/automation/selenium/webelement.py
automation/src/rabird/automation/selenium/webelement.py
''' @date 2014-11-16 @author Hong-she Liang <[email protected]> ''' # Import the global selenium unit, not our selenium . global_selenium = __import__('selenium') import types import time def set_attribute(self, name, value): value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'" value = value.replace("\n", r"\n") value = value.replace("\r", r"\r") value = value.replace("\t", r"\t") script = "arguments[0].setAttribute('%s', '%s')" % (name, value) self._parent.execute_script(script, self) def force_focus(self): global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform() def force_click(self): self._parent.execute_script("arguments[0].click();", self);
''' @date 2014-11-16 @author Hong-she Liang <[email protected]> ''' # Import the global selenium unit, not our selenium . global_selenium = __import__('selenium') import types import time def set_attribute(self, name, value): value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'" value = value.replace("\n", r"\n") value = value.replace("\r", r"\r") value = value.replace("\t", r"\t") script = "arguments[0].setAttribute('%s', '%s');" % (name, value) self._parent.execute_script(script, self) def force_focus(self): self._parent.execute_script("arguments[0].focus();", self); def force_click(self): self._parent.execute_script("arguments[0].click();", self);
Use javascript to force focus on an element, because the action chains seems take no effect!
Use javascript to force focus on an element, because the action chains seems take no effect!
Python
apache-2.0
starofrainnight/rabird.core,starofrainnight/rabird.auto
''' @date 2014-11-16 @author Hong-she Liang <[email protected]> ''' # Import the global selenium unit, not our selenium . global_selenium = __import__('selenium') import types import time def set_attribute(self, name, value): value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'" value = value.replace("\n", r"\n") value = value.replace("\r", r"\r") value = value.replace("\t", r"\t") - script = "arguments[0].setAttribute('%s', '%s')" % (name, value) + script = "arguments[0].setAttribute('%s', '%s');" % (name, value) self._parent.execute_script(script, self) def force_focus(self): - global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform() + self._parent.execute_script("arguments[0].focus();", self); def force_click(self): self._parent.execute_script("arguments[0].click();", self);
Use javascript to force focus on an element, because the action chains seems take no effect!
## Code Before: ''' @date 2014-11-16 @author Hong-she Liang <[email protected]> ''' # Import the global selenium unit, not our selenium . global_selenium = __import__('selenium') import types import time def set_attribute(self, name, value): value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'" value = value.replace("\n", r"\n") value = value.replace("\r", r"\r") value = value.replace("\t", r"\t") script = "arguments[0].setAttribute('%s', '%s')" % (name, value) self._parent.execute_script(script, self) def force_focus(self): global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform() def force_click(self): self._parent.execute_script("arguments[0].click();", self); ## Instruction: Use javascript to force focus on an element, because the action chains seems take no effect! ## Code After: ''' @date 2014-11-16 @author Hong-she Liang <[email protected]> ''' # Import the global selenium unit, not our selenium . global_selenium = __import__('selenium') import types import time def set_attribute(self, name, value): value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'" value = value.replace("\n", r"\n") value = value.replace("\r", r"\r") value = value.replace("\t", r"\t") script = "arguments[0].setAttribute('%s', '%s');" % (name, value) self._parent.execute_script(script, self) def force_focus(self): self._parent.execute_script("arguments[0].focus();", self); def force_click(self): self._parent.execute_script("arguments[0].click();", self);
''' @date 2014-11-16 @author Hong-she Liang <[email protected]> ''' # Import the global selenium unit, not our selenium . global_selenium = __import__('selenium') import types import time def set_attribute(self, name, value): value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'" value = value.replace("\n", r"\n") value = value.replace("\r", r"\r") value = value.replace("\t", r"\t") - script = "arguments[0].setAttribute('%s', '%s')" % (name, value) + script = "arguments[0].setAttribute('%s', '%s');" % (name, value) ? + self._parent.execute_script(script, self) def force_focus(self): - global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform() + self._parent.execute_script("arguments[0].focus();", self); def force_click(self): self._parent.execute_script("arguments[0].click();", self);
53a2d8781e3e5d8e5879d4ef7c62752483323cf9
bfg9000/shell/__init__.py
bfg9000/shell/__init__.py
import os import subprocess from ..platform_name import platform_name if platform_name() == 'windows': from .windows import * else: from .posix import * class shell_list(list): """A special subclass of list used to mark that this command line uses special shell characters.""" pass def execute(args, shell=False, env=None, quiet=False): if quiet: devnull = open(os.devnull, 'wb') try: return subprocess.check_output( args, universal_newlines=True, shell=shell, env=env ) except: if quiet: devnull.close() raise
import os import subprocess from ..platform_name import platform_name if platform_name() == 'windows': from .windows import * else: from .posix import * class shell_list(list): """A special subclass of list used to mark that this command line uses special shell characters.""" pass def execute(args, shell=False, env=None, quiet=False): stderr = None if quiet: stderr = open(os.devnull, 'wb') try: return subprocess.check_output( args, universal_newlines=True, shell=shell, env=env, stderr=stderr ) except: if quiet: stderr.close() raise
Fix "quiet" mode for shell.execute()
Fix "quiet" mode for shell.execute()
Python
bsd-3-clause
jimporter/bfg9000,jimporter/bfg9000,jimporter/bfg9000,jimporter/bfg9000
import os import subprocess from ..platform_name import platform_name if platform_name() == 'windows': from .windows import * else: from .posix import * class shell_list(list): """A special subclass of list used to mark that this command line uses special shell characters.""" pass def execute(args, shell=False, env=None, quiet=False): + stderr = None if quiet: - devnull = open(os.devnull, 'wb') + stderr = open(os.devnull, 'wb') try: return subprocess.check_output( - args, universal_newlines=True, shell=shell, env=env + args, universal_newlines=True, shell=shell, env=env, stderr=stderr ) except: if quiet: - devnull.close() + stderr.close() raise
Fix "quiet" mode for shell.execute()
## Code Before: import os import subprocess from ..platform_name import platform_name if platform_name() == 'windows': from .windows import * else: from .posix import * class shell_list(list): """A special subclass of list used to mark that this command line uses special shell characters.""" pass def execute(args, shell=False, env=None, quiet=False): if quiet: devnull = open(os.devnull, 'wb') try: return subprocess.check_output( args, universal_newlines=True, shell=shell, env=env ) except: if quiet: devnull.close() raise ## Instruction: Fix "quiet" mode for shell.execute() ## Code After: import os import subprocess from ..platform_name import platform_name if platform_name() == 'windows': from .windows import * else: from .posix import * class shell_list(list): """A special subclass of list used to mark that this command line uses special shell characters.""" pass def execute(args, shell=False, env=None, quiet=False): stderr = None if quiet: stderr = open(os.devnull, 'wb') try: return subprocess.check_output( args, universal_newlines=True, shell=shell, env=env, stderr=stderr ) except: if quiet: stderr.close() raise
import os import subprocess from ..platform_name import platform_name if platform_name() == 'windows': from .windows import * else: from .posix import * class shell_list(list): """A special subclass of list used to mark that this command line uses special shell characters.""" pass def execute(args, shell=False, env=None, quiet=False): + stderr = None if quiet: - devnull = open(os.devnull, 'wb') ? ^^^^^ + stderr = open(os.devnull, 'wb') ? ++ ^^ try: return subprocess.check_output( - args, universal_newlines=True, shell=shell, env=env + args, universal_newlines=True, shell=shell, env=env, stderr=stderr ? +++++++++++++++ ) except: if quiet: - devnull.close() ? ^^^^^ + stderr.close() ? ++ ^^ raise
70c98a42326471d3ed615def61954905673c5972
typhon/nonlte/__init__.py
typhon/nonlte/__init__.py
from .version import __version__ try: __ATRASU_SETUP__ except: __ATRASU_SETUP__ = False if not __ATRASU_SETUP__: from . import spectra from . import setup_atmosphere from . import const from . import nonltecalc from . import mathmatics from . import rtc
try: __ATRASU_SETUP__ except: __ATRASU_SETUP__ = False if not __ATRASU_SETUP__: from . import spectra from . import setup_atmosphere from . import const from . import nonltecalc from . import mathmatics from . import rtc
Remove import of removed version module.
Remove import of removed version module.
Python
mit
atmtools/typhon,atmtools/typhon
- - from .version import __version__ try: __ATRASU_SETUP__ except: __ATRASU_SETUP__ = False if not __ATRASU_SETUP__: from . import spectra from . import setup_atmosphere from . import const from . import nonltecalc from . import mathmatics from . import rtc
Remove import of removed version module.
## Code Before: from .version import __version__ try: __ATRASU_SETUP__ except: __ATRASU_SETUP__ = False if not __ATRASU_SETUP__: from . import spectra from . import setup_atmosphere from . import const from . import nonltecalc from . import mathmatics from . import rtc ## Instruction: Remove import of removed version module. ## Code After: try: __ATRASU_SETUP__ except: __ATRASU_SETUP__ = False if not __ATRASU_SETUP__: from . import spectra from . import setup_atmosphere from . import const from . import nonltecalc from . import mathmatics from . import rtc
- - from .version import __version__ try: __ATRASU_SETUP__ except: __ATRASU_SETUP__ = False if not __ATRASU_SETUP__: from . import spectra from . import setup_atmosphere from . import const from . import nonltecalc from . import mathmatics from . import rtc