commit
stringlengths 40
40
| old_file
stringlengths 4
106
| new_file
stringlengths 4
106
| old_contents
stringlengths 10
2.94k
| new_contents
stringlengths 21
2.95k
| subject
stringlengths 16
444
| message
stringlengths 17
2.63k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 7
43k
| ndiff
stringlengths 52
3.31k
| instruction
stringlengths 16
444
| content
stringlengths 133
4.32k
| diff
stringlengths 49
3.61k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
e409e8d77eec8e53978512da56bf52f768d46c1a
|
create-application.py
|
create-application.py
|
import subprocess, os, sys, optparse
fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
capath = os.path.abspath(
os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py")
)
skeletonpath = os.path.abspath(
os.path.join(fullpath, "unify", "application", "skeleton")
)
subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:])
|
import os, sys, shutil
print("Unify create skeleton")
print("(C) 2012 Sebastian Fastner, Mainz, Germany")
print()
if (len(sys.argv) != 2):
print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0]))
exit(1)
NAMESPACE = sys.argv[1]
UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
SKELETONPATH = os.path.abspath(
os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify")
)
TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE))
REPLACEMENTS = {
"NAMESPACE" : NAMESPACE,
"UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH)
}
if os.path.exists(TARGETPATH):
print("Path %s exists. Aborting." % TARGETPATH)
exit(2)
shutil.copytree(SKELETONPATH, TARGETPATH)
def patch_line(line):
for key in REPLACEMENTS:
check = "${" + key + "}"
line = line.replace(check, REPLACEMENTS[key])
return line
def handle_file(directory, filename):
outfile_name = os.path.join(directory, filename.replace(".tmpl", ""))
infile_name = os.path.join(directory, filename)
with open(outfile_name, "w") as outfile:
with open(infile_name) as infile:
for line in infile:
outfile.write(patch_line(line))
os.remove(infile_name)
def handle_dir(directory):
shutil.move(directory, directory[:-6] + NAMESPACE)
for root, dirs, files in os.walk(TARGETPATH,topdown=False):
for file in files:
if ".tmpl." in file:
handle_file(root, file)
if root.endswith("custom"):
handle_dir(root)
print("Creat application skeleton in %s ... done" % TARGETPATH)
|
Add variable replacements to create application
|
Add variable replacements to create application
|
Python
|
mit
|
unify/unify,unify/unify,unify/unify,unify/unify,unify/unify,unify/unify
|
- import subprocess, os, sys, optparse
+ import os, sys, shutil
+ print("Unify create skeleton")
+ print("(C) 2012 Sebastian Fastner, Mainz, Germany")
+ print()
+
+ if (len(sys.argv) != 2):
+ print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0]))
+ exit(1)
+
+ NAMESPACE = sys.argv[1]
+
- fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
+ UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
- capath = os.path.abspath(
- os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py")
+ SKELETONPATH = os.path.abspath(
+ os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify")
)
+ TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE))
- skeletonpath = os.path.abspath(
- os.path.join(fullpath, "unify", "application", "skeleton")
- )
- subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:])
+ REPLACEMENTS = {
+ "NAMESPACE" : NAMESPACE,
+ "UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH)
+ }
+ if os.path.exists(TARGETPATH):
+ print("Path %s exists. Aborting." % TARGETPATH)
+ exit(2)
+
+ shutil.copytree(SKELETONPATH, TARGETPATH)
+
+ def patch_line(line):
+ for key in REPLACEMENTS:
+ check = "${" + key + "}"
+ line = line.replace(check, REPLACEMENTS[key])
+ return line
+
+ def handle_file(directory, filename):
+ outfile_name = os.path.join(directory, filename.replace(".tmpl", ""))
+ infile_name = os.path.join(directory, filename)
+ with open(outfile_name, "w") as outfile:
+ with open(infile_name) as infile:
+ for line in infile:
+ outfile.write(patch_line(line))
+ os.remove(infile_name)
+
+ def handle_dir(directory):
+ shutil.move(directory, directory[:-6] + NAMESPACE)
+
+ for root, dirs, files in os.walk(TARGETPATH,topdown=False):
+ for file in files:
+ if ".tmpl." in file:
+ handle_file(root, file)
+
+ if root.endswith("custom"):
+ handle_dir(root)
+
+ print("Creat application skeleton in %s ... done" % TARGETPATH)
|
Add variable replacements to create application
|
## Code Before:
import subprocess, os, sys, optparse
fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
capath = os.path.abspath(
os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py")
)
skeletonpath = os.path.abspath(
os.path.join(fullpath, "unify", "application", "skeleton")
)
subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:])
## Instruction:
Add variable replacements to create application
## Code After:
import os, sys, shutil
print("Unify create skeleton")
print("(C) 2012 Sebastian Fastner, Mainz, Germany")
print()
if (len(sys.argv) != 2):
print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0]))
exit(1)
NAMESPACE = sys.argv[1]
UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
SKELETONPATH = os.path.abspath(
os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify")
)
TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE))
REPLACEMENTS = {
"NAMESPACE" : NAMESPACE,
"UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH)
}
if os.path.exists(TARGETPATH):
print("Path %s exists. Aborting." % TARGETPATH)
exit(2)
shutil.copytree(SKELETONPATH, TARGETPATH)
def patch_line(line):
for key in REPLACEMENTS:
check = "${" + key + "}"
line = line.replace(check, REPLACEMENTS[key])
return line
def handle_file(directory, filename):
outfile_name = os.path.join(directory, filename.replace(".tmpl", ""))
infile_name = os.path.join(directory, filename)
with open(outfile_name, "w") as outfile:
with open(infile_name) as infile:
for line in infile:
outfile.write(patch_line(line))
os.remove(infile_name)
def handle_dir(directory):
shutil.move(directory, directory[:-6] + NAMESPACE)
for root, dirs, files in os.walk(TARGETPATH,topdown=False):
for file in files:
if ".tmpl." in file:
handle_file(root, file)
if root.endswith("custom"):
handle_dir(root)
print("Creat application skeleton in %s ... done" % TARGETPATH)
|
- import subprocess, os, sys, optparse
+ import os, sys, shutil
+ print("Unify create skeleton")
+ print("(C) 2012 Sebastian Fastner, Mainz, Germany")
+ print()
+
+ if (len(sys.argv) != 2):
+ print("Syntax: %s <namespace>" % os.path.basename(sys.argv[0]))
+ exit(1)
+
+ NAMESPACE = sys.argv[1]
+
- fullpath = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
? ^^^^^^^^
+ UNIFYPATH = os.path.join(os.getcwd(), os.path.dirname(sys.argv[0]))
? ^^^^^^^^^
- capath = os.path.abspath(
- os.path.join(fullpath, "qooxdoo", "tool", "bin", "create-application.py")
+ SKELETONPATH = os.path.abspath(
+ os.path.join(UNIFYPATH, "unify", "application", "skeleton", "unify")
)
+ TARGETPATH = os.path.abspath(os.path.join(os.getcwd(), NAMESPACE))
- skeletonpath = os.path.abspath(
- os.path.join(fullpath, "unify", "application", "skeleton")
- )
- subprocess.call(["python", capath, "-p", skeletonpath, "-t", "unify"] + sys.argv[1:])
+ REPLACEMENTS = {
+ "NAMESPACE" : NAMESPACE,
+ "UNIFYPATH" : os.path.relpath(UNIFYPATH, TARGETPATH)
+ }
+
+ if os.path.exists(TARGETPATH):
+ print("Path %s exists. Aborting." % TARGETPATH)
+ exit(2)
+
+ shutil.copytree(SKELETONPATH, TARGETPATH)
+
+ def patch_line(line):
+ for key in REPLACEMENTS:
+ check = "${" + key + "}"
+ line = line.replace(check, REPLACEMENTS[key])
+ return line
+
+ def handle_file(directory, filename):
+ outfile_name = os.path.join(directory, filename.replace(".tmpl", ""))
+ infile_name = os.path.join(directory, filename)
+ with open(outfile_name, "w") as outfile:
+ with open(infile_name) as infile:
+ for line in infile:
+ outfile.write(patch_line(line))
+ os.remove(infile_name)
+
+ def handle_dir(directory):
+ shutil.move(directory, directory[:-6] + NAMESPACE)
+
+ for root, dirs, files in os.walk(TARGETPATH,topdown=False):
+ for file in files:
+ if ".tmpl." in file:
+ handle_file(root, file)
+
+ if root.endswith("custom"):
+ handle_dir(root)
+
+ print("Creat application skeleton in %s ... done" % TARGETPATH)
|
994606d2641115f8af59657204d3d64f540bbfbd
|
data_structures/linked_list.py
|
data_structures/linked_list.py
|
class Node(object):
def __init__(self, val, next=None):
self.val = val
self.next = next
def __repr__(self):
return '{val}'.format(val=self.val)
class LinkedList(object):
def __init__(self, values=None, head=None):
self.head = head
self.length = 0
def __repr__(self):
pass
def __len__(self):
pass
def __iter__(self):
pass
def size(self):
pass
def search(self):
pass
def display(self):
pass
def remove(self):
pass
def insert(self):
pass
def pop(self):
pass
|
class Node(object):
def __init__(self, val, next=None):
self.val = val
self.next = next
def __repr__(self):
return '{val}'.format(val=self.val)
class LinkedList(object):
def __init__(self, iterable=()):
self._current = None
self.head = None
self.length = 0
for val in reversed(iterable):
self.insert(val)
def __repr__(self):
'''Print string representation of Linked List.'''
node = self.head
output = ''
for node in self:
output += '{!r}'.format(node.val)
return '({})'.format(output.rstrip(' ,'))
def __len__(self):
return self.length
def __iter__(self):
if self.head is not None:
self._current = self.head
return self
def next(self):
if self._current is None:
raise StopIteration
node = self._current
self._current = self._current.next
return node
def insert(self):
pass
def size(self):
pass
def search(self):
pass
def display(self):
pass
def remove(self):
pass
def pop(self):
pass
|
Update magic methods, and reorg args.
|
Update magic methods, and reorg args.
|
Python
|
mit
|
sjschmidt44/python_data_structures
|
class Node(object):
def __init__(self, val, next=None):
self.val = val
self.next = next
def __repr__(self):
return '{val}'.format(val=self.val)
class LinkedList(object):
- def __init__(self, values=None, head=None):
+ def __init__(self, iterable=()):
+ self._current = None
- self.head = head
+ self.head = None
self.length = 0
+ for val in reversed(iterable):
+ self.insert(val)
def __repr__(self):
- pass
+ '''Print string representation of Linked List.'''
+ node = self.head
+ output = ''
+ for node in self:
+ output += '{!r}'.format(node.val)
+ return '({})'.format(output.rstrip(' ,'))
def __len__(self):
- pass
+ return self.length
def __iter__(self):
+ if self.head is not None:
+ self._current = self.head
+ return self
+
+ def next(self):
+ if self._current is None:
+ raise StopIteration
+ node = self._current
+ self._current = self._current.next
+ return node
+
+ def insert(self):
pass
def size(self):
pass
def search(self):
pass
def display(self):
pass
def remove(self):
pass
- def insert(self):
- pass
-
def pop(self):
pass
|
Update magic methods, and reorg args.
|
## Code Before:
class Node(object):
def __init__(self, val, next=None):
self.val = val
self.next = next
def __repr__(self):
return '{val}'.format(val=self.val)
class LinkedList(object):
def __init__(self, values=None, head=None):
self.head = head
self.length = 0
def __repr__(self):
pass
def __len__(self):
pass
def __iter__(self):
pass
def size(self):
pass
def search(self):
pass
def display(self):
pass
def remove(self):
pass
def insert(self):
pass
def pop(self):
pass
## Instruction:
Update magic methods, and reorg args.
## Code After:
class Node(object):
def __init__(self, val, next=None):
self.val = val
self.next = next
def __repr__(self):
return '{val}'.format(val=self.val)
class LinkedList(object):
def __init__(self, iterable=()):
self._current = None
self.head = None
self.length = 0
for val in reversed(iterable):
self.insert(val)
def __repr__(self):
'''Print string representation of Linked List.'''
node = self.head
output = ''
for node in self:
output += '{!r}'.format(node.val)
return '({})'.format(output.rstrip(' ,'))
def __len__(self):
return self.length
def __iter__(self):
if self.head is not None:
self._current = self.head
return self
def next(self):
if self._current is None:
raise StopIteration
node = self._current
self._current = self._current.next
return node
def insert(self):
pass
def size(self):
pass
def search(self):
pass
def display(self):
pass
def remove(self):
pass
def pop(self):
pass
|
class Node(object):
def __init__(self, val, next=None):
self.val = val
self.next = next
def __repr__(self):
return '{val}'.format(val=self.val)
class LinkedList(object):
- def __init__(self, values=None, head=None):
+ def __init__(self, iterable=()):
+ self._current = None
- self.head = head
? ^ --
+ self.head = None
? ^^^
self.length = 0
+ for val in reversed(iterable):
+ self.insert(val)
def __repr__(self):
- pass
+ '''Print string representation of Linked List.'''
+ node = self.head
+ output = ''
+ for node in self:
+ output += '{!r}'.format(node.val)
+ return '({})'.format(output.rstrip(' ,'))
def __len__(self):
- pass
+ return self.length
def __iter__(self):
+ if self.head is not None:
+ self._current = self.head
+ return self
+
+ def next(self):
+ if self._current is None:
+ raise StopIteration
+ node = self._current
+ self._current = self._current.next
+ return node
+
+ def insert(self):
pass
def size(self):
pass
def search(self):
pass
def display(self):
pass
def remove(self):
pass
- def insert(self):
- pass
-
def pop(self):
pass
|
26598254cd48a716527eb4689ad96551c5a39790
|
ksp_login/__init__.py
|
ksp_login/__init__.py
|
__version__ = '0.6.0'
__version_info__ = __version__.split('.')
from django.utils.translation import ugettext_lazy as _
def __activate_social_auth_monkeypatch():
from social_core.backends.base import BaseAuth
from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth)
from social_core.backends.livejournal import LiveJournalOpenId
from social_core.backends.yahoo import YahooOpenId
from social_core.backends.google import GoogleOpenId
from social_core.backends.yandex import YandexOpenId
BaseAuth.REQUIRED_FIELD_NAME = None
BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None
OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD
OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity')
LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user'
LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username')
# Reset to None in those OpenID backends where nothing is required.
GoogleOpenId.REQUIRED_FIELD_NAME = None
GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YahooOpenId.REQUIRED_FIELD_NAME = None
YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YandexOpenId.REQUIRED_FIELD_NAME = None
YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
__activate_social_auth_monkeypatch()
|
__version__ = '0.6.0'
__version_info__ = tuple(map(int, __version__.split('.')))
from django.utils.translation import ugettext_lazy as _
def __activate_social_auth_monkeypatch():
from social_core.backends.base import BaseAuth
from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth)
from social_core.backends.livejournal import LiveJournalOpenId
from social_core.backends.yahoo import YahooOpenId
from social_core.backends.google import GoogleOpenId
from social_core.backends.yandex import YandexOpenId
BaseAuth.REQUIRED_FIELD_NAME = None
BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None
OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD
OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity')
LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user'
LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username')
# Reset to None in those OpenID backends where nothing is required.
GoogleOpenId.REQUIRED_FIELD_NAME = None
GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YahooOpenId.REQUIRED_FIELD_NAME = None
YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YandexOpenId.REQUIRED_FIELD_NAME = None
YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
__activate_social_auth_monkeypatch()
|
Make version info tuple of ints.
|
Make version info tuple of ints.
|
Python
|
bsd-3-clause
|
koniiiik/ksp_login,koniiiik/ksp_login,koniiiik/ksp_login
|
__version__ = '0.6.0'
- __version_info__ = __version__.split('.')
+ __version_info__ = tuple(map(int, __version__.split('.')))
from django.utils.translation import ugettext_lazy as _
def __activate_social_auth_monkeypatch():
from social_core.backends.base import BaseAuth
from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth)
from social_core.backends.livejournal import LiveJournalOpenId
from social_core.backends.yahoo import YahooOpenId
from social_core.backends.google import GoogleOpenId
from social_core.backends.yandex import YandexOpenId
BaseAuth.REQUIRED_FIELD_NAME = None
BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None
OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD
OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity')
LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user'
LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username')
# Reset to None in those OpenID backends where nothing is required.
GoogleOpenId.REQUIRED_FIELD_NAME = None
GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YahooOpenId.REQUIRED_FIELD_NAME = None
YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YandexOpenId.REQUIRED_FIELD_NAME = None
YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
+
__activate_social_auth_monkeypatch()
|
Make version info tuple of ints.
|
## Code Before:
__version__ = '0.6.0'
__version_info__ = __version__.split('.')
from django.utils.translation import ugettext_lazy as _
def __activate_social_auth_monkeypatch():
from social_core.backends.base import BaseAuth
from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth)
from social_core.backends.livejournal import LiveJournalOpenId
from social_core.backends.yahoo import YahooOpenId
from social_core.backends.google import GoogleOpenId
from social_core.backends.yandex import YandexOpenId
BaseAuth.REQUIRED_FIELD_NAME = None
BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None
OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD
OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity')
LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user'
LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username')
# Reset to None in those OpenID backends where nothing is required.
GoogleOpenId.REQUIRED_FIELD_NAME = None
GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YahooOpenId.REQUIRED_FIELD_NAME = None
YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YandexOpenId.REQUIRED_FIELD_NAME = None
YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
__activate_social_auth_monkeypatch()
## Instruction:
Make version info tuple of ints.
## Code After:
__version__ = '0.6.0'
__version_info__ = tuple(map(int, __version__.split('.')))
from django.utils.translation import ugettext_lazy as _
def __activate_social_auth_monkeypatch():
from social_core.backends.base import BaseAuth
from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth)
from social_core.backends.livejournal import LiveJournalOpenId
from social_core.backends.yahoo import YahooOpenId
from social_core.backends.google import GoogleOpenId
from social_core.backends.yandex import YandexOpenId
BaseAuth.REQUIRED_FIELD_NAME = None
BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None
OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD
OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity')
LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user'
LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username')
# Reset to None in those OpenID backends where nothing is required.
GoogleOpenId.REQUIRED_FIELD_NAME = None
GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YahooOpenId.REQUIRED_FIELD_NAME = None
YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YandexOpenId.REQUIRED_FIELD_NAME = None
YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
__activate_social_auth_monkeypatch()
|
__version__ = '0.6.0'
- __version_info__ = __version__.split('.')
+ __version_info__ = tuple(map(int, __version__.split('.')))
? +++++++++++++++ ++
from django.utils.translation import ugettext_lazy as _
def __activate_social_auth_monkeypatch():
from social_core.backends.base import BaseAuth
from social_core.backends.open_id import (OPENID_ID_FIELD, OpenIdAuth)
from social_core.backends.livejournal import LiveJournalOpenId
from social_core.backends.yahoo import YahooOpenId
from social_core.backends.google import GoogleOpenId
from social_core.backends.yandex import YandexOpenId
BaseAuth.REQUIRED_FIELD_NAME = None
BaseAuth.REQUIRED_FIELD_VERBOSE_NAME = None
OpenIdAuth.REQUIRED_FIELD_NAME = OPENID_ID_FIELD
OpenIdAuth.REQUIRED_FIELD_VERBOSE_NAME = _('OpenID identity')
LiveJournalOpenId.REQUIRED_FIELD_NAME = 'openid_lj_user'
LiveJournalOpenId.REQUIRED_FIELD_VERBOSE_NAME = _('LiveJournal username')
# Reset to None in those OpenID backends where nothing is required.
GoogleOpenId.REQUIRED_FIELD_NAME = None
GoogleOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YahooOpenId.REQUIRED_FIELD_NAME = None
YahooOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
YandexOpenId.REQUIRED_FIELD_NAME = None
YandexOpenId.REQUIRED_FIELD_VERBOSE_NAME = None
+
__activate_social_auth_monkeypatch()
|
ae55577e4cea64a0052eb0c219641435c9c0210c
|
samples/model-builder/init_sample.py
|
samples/model-builder/init_sample.py
|
from typing import Optional
from google.auth import credentials as auth_credentials
from google.cloud import aiplatform
# [START aiplatform_sdk_init_sample]
def init_sample(
project: Optional[str] = None,
location: Optional[str] = None,
experiment: Optional[str] = None,
staging_bucket: Optional[str] = None,
credentials: Optional[auth_credentials.Credentials] = None,
encryption_spec_key_name: Optional[str] = None,
):
aiplatform.init(
project=project,
location=location,
experiment=experiment,
staging_bucket=staging_bucket,
credentials=credentials,
encryption_spec_key_name=encryption_spec_key_name,
)
# [END aiplatform_sdk_init_sample]
|
from typing import Optional
from google.auth import credentials as auth_credentials
# [START aiplatform_sdk_init_sample]
def init_sample(
project: Optional[str] = None,
location: Optional[str] = None,
experiment: Optional[str] = None,
staging_bucket: Optional[str] = None,
credentials: Optional[auth_credentials.Credentials] = None,
encryption_spec_key_name: Optional[str] = None,
):
from google.cloud import aiplatform
aiplatform.init(
project=project,
location=location,
experiment=experiment,
staging_bucket=staging_bucket,
credentials=credentials,
encryption_spec_key_name=encryption_spec_key_name,
)
# [END aiplatform_sdk_init_sample]
|
Update init sample to import inside of function.
|
chore: Update init sample to import inside of function.
PiperOrigin-RevId: 485079470
|
Python
|
apache-2.0
|
googleapis/python-aiplatform,googleapis/python-aiplatform
|
from typing import Optional
from google.auth import credentials as auth_credentials
- from google.cloud import aiplatform
# [START aiplatform_sdk_init_sample]
def init_sample(
project: Optional[str] = None,
location: Optional[str] = None,
experiment: Optional[str] = None,
staging_bucket: Optional[str] = None,
credentials: Optional[auth_credentials.Credentials] = None,
encryption_spec_key_name: Optional[str] = None,
):
+
+ from google.cloud import aiplatform
+
aiplatform.init(
project=project,
location=location,
experiment=experiment,
staging_bucket=staging_bucket,
credentials=credentials,
encryption_spec_key_name=encryption_spec_key_name,
)
# [END aiplatform_sdk_init_sample]
|
Update init sample to import inside of function.
|
## Code Before:
from typing import Optional
from google.auth import credentials as auth_credentials
from google.cloud import aiplatform
# [START aiplatform_sdk_init_sample]
def init_sample(
project: Optional[str] = None,
location: Optional[str] = None,
experiment: Optional[str] = None,
staging_bucket: Optional[str] = None,
credentials: Optional[auth_credentials.Credentials] = None,
encryption_spec_key_name: Optional[str] = None,
):
aiplatform.init(
project=project,
location=location,
experiment=experiment,
staging_bucket=staging_bucket,
credentials=credentials,
encryption_spec_key_name=encryption_spec_key_name,
)
# [END aiplatform_sdk_init_sample]
## Instruction:
Update init sample to import inside of function.
## Code After:
from typing import Optional
from google.auth import credentials as auth_credentials
# [START aiplatform_sdk_init_sample]
def init_sample(
project: Optional[str] = None,
location: Optional[str] = None,
experiment: Optional[str] = None,
staging_bucket: Optional[str] = None,
credentials: Optional[auth_credentials.Credentials] = None,
encryption_spec_key_name: Optional[str] = None,
):
from google.cloud import aiplatform
aiplatform.init(
project=project,
location=location,
experiment=experiment,
staging_bucket=staging_bucket,
credentials=credentials,
encryption_spec_key_name=encryption_spec_key_name,
)
# [END aiplatform_sdk_init_sample]
|
from typing import Optional
from google.auth import credentials as auth_credentials
- from google.cloud import aiplatform
# [START aiplatform_sdk_init_sample]
def init_sample(
project: Optional[str] = None,
location: Optional[str] = None,
experiment: Optional[str] = None,
staging_bucket: Optional[str] = None,
credentials: Optional[auth_credentials.Credentials] = None,
encryption_spec_key_name: Optional[str] = None,
):
+
+ from google.cloud import aiplatform
+
aiplatform.init(
project=project,
location=location,
experiment=experiment,
staging_bucket=staging_bucket,
credentials=credentials,
encryption_spec_key_name=encryption_spec_key_name,
)
# [END aiplatform_sdk_init_sample]
|
5d3d47e0fae9ddb9f445972e5186429163aabf40
|
statirator/core/management/commands/init.py
|
statirator/core/management/commands/init.py
|
import os
from optparse import make_option
from django.core.management.base import BaseCommand
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
|
import os
import logging
from django.core.management.base import BaseCommand
from optparse import make_option
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
logging.info("Initializing project structure in %s", directory)
os.makedirs(directory)
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
|
Create the directory before calling the startprojcet command
|
Create the directory before calling the startprojcet command
|
Python
|
mit
|
MeirKriheli/statirator,MeirKriheli/statirator,MeirKriheli/statirator
|
import os
+ import logging
+ from django.core.management.base import BaseCommand
from optparse import make_option
- from django.core.management.base import BaseCommand
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
+ logging.info("Initializing project structure in %s", directory)
+ os.makedirs(directory)
+
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
|
Create the directory before calling the startprojcet command
|
## Code Before:
import os
from optparse import make_option
from django.core.management.base import BaseCommand
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
## Instruction:
Create the directory before calling the startprojcet command
## Code After:
import os
import logging
from django.core.management.base import BaseCommand
from optparse import make_option
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
logging.info("Initializing project structure in %s", directory)
os.makedirs(directory)
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
|
import os
+ import logging
+ from django.core.management.base import BaseCommand
from optparse import make_option
- from django.core.management.base import BaseCommand
class Command(BaseCommand):
help = "Init the static site project"
args = '[directory]'
option_list = (
make_option(
'--title', '-t', dest='title', default='Default site',
help='Site title [Default: "%defaults"]'),
make_option(
'--domain', '-d', dest='domain', default='example.com',
help='Domain name [Default: "%default"]'),
make_option(
'--languages', '-l', dest='languages', default=['he', 'en'],
action='append', help='Supported languages. [Default: "%default"]')
) + BaseCommand.option_list
def handle(self, directory, **options):
+ logging.info("Initializing project structure in %s", directory)
+ os.makedirs(directory)
+
from django.conf.global_settings import LANGUAGES
extra = {
'build': 'build',
'default_lang': options['languages'][0],
'languages': [l for l in LANGUAGES if l[0] in options["languages"]],
'extensions': ('py', ),
'files': (),
'template': os.path.abspath(
os.path.join(
os.path.dirname(__file__),
os.pardir, os.pardir, os.pardir, 'project_template')),
}
extra.update(options)
from django.core.management import call_command
call_command('startproject', 'conf', directory, **extra)
|
1e5d549b6fdf62c1016451f9dfe566c9546b2f38
|
bcbio/bed/__init__.py
|
bcbio/bed/__init__.py
|
import pybedtools as bt
import six
def concat(bed_files, catted=None):
"""
recursively concat a set of BED files, returning a
sorted bedtools object of the result
"""
if len(bed_files) == 0:
if catted:
return catted.sort()
else:
return catted
if not catted:
bed_files = list(bed_files)
catted = bt.BedTool(bed_files.pop())
else:
catted = catted.cat(bed_files.pop(), postmerge=False,
force_truncate=False)
return concat(bed_files, catted)
def merge(bedfiles):
"""
given a BED file or list of BED files merge them an return a bedtools object
"""
if isinstance(bedfiles, list):
catted = concat(bedfiles)
else:
catted = concat([bedfiles])
if catted:
return concat(bedfiles).sort().merge()
else:
return catted
|
import pybedtools as bt
import six
def concat(bed_files, catted=None):
"""
recursively concat a set of BED files, returning a
sorted bedtools object of the result
"""
bed_files = [x for x in bed_files if x]
if len(bed_files) == 0:
if catted:
# move to a .bed extension for downstream tools if not already
sorted_bed = catted.sort()
if not sorted_bed.fn.endswith(".bed"):
return sorted_bed.moveto(sorted_bed.fn + ".bed")
else:
return sorted_bed
else:
return catted
if not catted:
bed_files = list(bed_files)
catted = bt.BedTool(bed_files.pop())
else:
catted = catted.cat(bed_files.pop(), postmerge=False,
force_truncate=False)
return concat(bed_files, catted)
def merge(bedfiles):
"""
given a BED file or list of BED files merge them an return a bedtools object
"""
if isinstance(bedfiles, list):
catted = concat(bedfiles)
else:
catted = concat([bedfiles])
if catted:
return concat(bedfiles).sort().merge()
else:
return catted
|
Move the file to have an extension of .bed.
|
Move the file to have an extension of .bed.
A lot of tools detect what type of file it is by the extension,
so this lets us pass on the BedTool.fn as the filename and
not break things.
|
Python
|
mit
|
guillermo-carrasco/bcbio-nextgen,lbeltrame/bcbio-nextgen,gifford-lab/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,vladsaveliev/bcbio-nextgen,brainstorm/bcbio-nextgen,mjafin/bcbio-nextgen,lbeltrame/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,brainstorm/bcbio-nextgen,fw1121/bcbio-nextgen,verdurin/bcbio-nextgen,lpantano/bcbio-nextgen,brainstorm/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,a113n/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,chapmanb/bcbio-nextgen,elkingtonmcb/bcbio-nextgen,chapmanb/bcbio-nextgen,biocyberman/bcbio-nextgen,chapmanb/bcbio-nextgen,vladsaveliev/bcbio-nextgen,gifford-lab/bcbio-nextgen,fw1121/bcbio-nextgen,hjanime/bcbio-nextgen,biocyberman/bcbio-nextgen,lpantano/bcbio-nextgen,guillermo-carrasco/bcbio-nextgen,lbeltrame/bcbio-nextgen,guillermo-carrasco/bcbio-nextgen,mjafin/bcbio-nextgen,a113n/bcbio-nextgen,lpantano/bcbio-nextgen,fw1121/bcbio-nextgen,gifford-lab/bcbio-nextgen,verdurin/bcbio-nextgen,biocyberman/bcbio-nextgen,vladsaveliev/bcbio-nextgen,a113n/bcbio-nextgen,Cyberbio-Lab/bcbio-nextgen,mjafin/bcbio-nextgen,hjanime/bcbio-nextgen,hjanime/bcbio-nextgen,verdurin/bcbio-nextgen
|
import pybedtools as bt
import six
def concat(bed_files, catted=None):
"""
recursively concat a set of BED files, returning a
sorted bedtools object of the result
"""
+ bed_files = [x for x in bed_files if x]
if len(bed_files) == 0:
if catted:
+ # move to a .bed extension for downstream tools if not already
- return catted.sort()
+ sorted_bed = catted.sort()
+ if not sorted_bed.fn.endswith(".bed"):
+ return sorted_bed.moveto(sorted_bed.fn + ".bed")
+ else:
+ return sorted_bed
else:
return catted
if not catted:
bed_files = list(bed_files)
catted = bt.BedTool(bed_files.pop())
else:
catted = catted.cat(bed_files.pop(), postmerge=False,
force_truncate=False)
return concat(bed_files, catted)
def merge(bedfiles):
"""
given a BED file or list of BED files merge them an return a bedtools object
"""
if isinstance(bedfiles, list):
catted = concat(bedfiles)
else:
catted = concat([bedfiles])
if catted:
return concat(bedfiles).sort().merge()
else:
return catted
|
Move the file to have an extension of .bed.
|
## Code Before:
import pybedtools as bt
import six
def concat(bed_files, catted=None):
"""
recursively concat a set of BED files, returning a
sorted bedtools object of the result
"""
if len(bed_files) == 0:
if catted:
return catted.sort()
else:
return catted
if not catted:
bed_files = list(bed_files)
catted = bt.BedTool(bed_files.pop())
else:
catted = catted.cat(bed_files.pop(), postmerge=False,
force_truncate=False)
return concat(bed_files, catted)
def merge(bedfiles):
"""
given a BED file or list of BED files merge them an return a bedtools object
"""
if isinstance(bedfiles, list):
catted = concat(bedfiles)
else:
catted = concat([bedfiles])
if catted:
return concat(bedfiles).sort().merge()
else:
return catted
## Instruction:
Move the file to have an extension of .bed.
## Code After:
import pybedtools as bt
import six
def concat(bed_files, catted=None):
"""
recursively concat a set of BED files, returning a
sorted bedtools object of the result
"""
bed_files = [x for x in bed_files if x]
if len(bed_files) == 0:
if catted:
# move to a .bed extension for downstream tools if not already
sorted_bed = catted.sort()
if not sorted_bed.fn.endswith(".bed"):
return sorted_bed.moveto(sorted_bed.fn + ".bed")
else:
return sorted_bed
else:
return catted
if not catted:
bed_files = list(bed_files)
catted = bt.BedTool(bed_files.pop())
else:
catted = catted.cat(bed_files.pop(), postmerge=False,
force_truncate=False)
return concat(bed_files, catted)
def merge(bedfiles):
"""
given a BED file or list of BED files merge them an return a bedtools object
"""
if isinstance(bedfiles, list):
catted = concat(bedfiles)
else:
catted = concat([bedfiles])
if catted:
return concat(bedfiles).sort().merge()
else:
return catted
|
import pybedtools as bt
import six
def concat(bed_files, catted=None):
"""
recursively concat a set of BED files, returning a
sorted bedtools object of the result
"""
+ bed_files = [x for x in bed_files if x]
if len(bed_files) == 0:
if catted:
+ # move to a .bed extension for downstream tools if not already
- return catted.sort()
? ^^^^
+ sorted_bed = catted.sort()
? ++ + ^^^^^^^
+ if not sorted_bed.fn.endswith(".bed"):
+ return sorted_bed.moveto(sorted_bed.fn + ".bed")
+ else:
+ return sorted_bed
else:
return catted
if not catted:
bed_files = list(bed_files)
catted = bt.BedTool(bed_files.pop())
else:
catted = catted.cat(bed_files.pop(), postmerge=False,
force_truncate=False)
return concat(bed_files, catted)
def merge(bedfiles):
"""
given a BED file or list of BED files merge them an return a bedtools object
"""
if isinstance(bedfiles, list):
catted = concat(bedfiles)
else:
catted = concat([bedfiles])
if catted:
return concat(bedfiles).sort().merge()
else:
return catted
|
025c3f6b73c97fdb58b1a492efcb6efe44cfdab0
|
twisted/plugins/caldav.py
|
twisted/plugins/caldav.py
|
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
|
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
from twisted.internet.protocol import Factory
Factory.noisy = False
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
|
Set Factory.noisy to False by default
|
Set Factory.noisy to False by default
git-svn-id: 81e381228600e5752b80483efd2b45b26c451ea2@3933 e27351fd-9f3e-4f54-a53b-843176b1656c
|
Python
|
apache-2.0
|
trevor/calendarserver,trevor/calendarserver,trevor/calendarserver
|
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
+
+ from twisted.internet.protocol import Factory
+ Factory.noisy = False
+
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
+
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
- options = serviceMakerProperty("options")
+ options = serviceMakerProperty("options")
- tapname = serviceMakerProperty("tapname")
+ tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
|
Set Factory.noisy to False by default
|
## Code Before:
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
## Instruction:
Set Factory.noisy to False by default
## Code After:
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
from twisted.internet.protocol import Factory
Factory.noisy = False
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
options = serviceMakerProperty("options")
tapname = serviceMakerProperty("tapname")
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
|
from zope.interface import implements
from twisted.plugin import IPlugin
from twisted.application.service import IServiceMaker
from twisted.python import reflect
+
+ from twisted.internet.protocol import Factory
+ Factory.noisy = False
+
def serviceMakerProperty(propname):
def getProperty(self):
return getattr(reflect.namedClass(self.serviceMakerClass), propname)
return property(getProperty)
class TAP(object):
implements(IPlugin, IServiceMaker)
+
def __init__(self, serviceMakerClass):
self.serviceMakerClass = serviceMakerClass
self._serviceMaker = None
- options = serviceMakerProperty("options")
+ options = serviceMakerProperty("options")
? ++++
- tapname = serviceMakerProperty("tapname")
+ tapname = serviceMakerProperty("tapname")
? ++++
description = serviceMakerProperty("description")
def makeService(self, options):
if self._serviceMaker is None:
self._serviceMaker = reflect.namedClass(self.serviceMakerClass)()
return self._serviceMaker.makeService(options)
TwistedCalDAV = TAP("calendarserver.tap.caldav.CalDAVServiceMaker")
CalDAVNotifier = TAP("twistedcaldav.notify.NotificationServiceMaker")
CalDAVMailGateway = TAP("twistedcaldav.mail.MailGatewayServiceMaker")
|
30a4281f2602bd6b9d90d89375785a2645854a0d
|
enthought/enable2/pyglet_backend/pyglet_app.py
|
enthought/enable2/pyglet_backend/pyglet_app.py
|
from enthought.enable.pyglet_backend.pyglet_app import *
|
__all__ = ["get_app", "PygletApp"]
from enthought.enable.pyglet_backend.pyglet_app import *
# Import the objects which are not declared in __all__,
# but are still defined in the real module, such that people
# can import them explicitly when needed, just as they could
# with the real module.
#
# It is unlikely that someone will import these objects, since
# they start with '_'. However, the proxy's job is to mimic the
# behavior of the real module as closely as possible.
# The proxy's job is not to define or change the API.
#
from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp
|
Improve the proxy module which maps to a module which uses __all__.
|
Improve the proxy module which maps to a module which uses __all__.
The notes I made in the code apply to all proxy modules which map
to a module which uses __all__.
|
Python
|
bsd-3-clause
|
tommy-u/enable,tommy-u/enable,tommy-u/enable,tommy-u/enable
|
+
+ __all__ = ["get_app", "PygletApp"]
+
from enthought.enable.pyglet_backend.pyglet_app import *
+
+ # Import the objects which are not declared in __all__,
+ # but are still defined in the real module, such that people
+ # can import them explicitly when needed, just as they could
+ # with the real module.
+ #
+ # It is unlikely that someone will import these objects, since
+ # they start with '_'. However, the proxy's job is to mimic the
+ # behavior of the real module as closely as possible.
+ # The proxy's job is not to define or change the API.
+ #
+ from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp
+
+
|
Improve the proxy module which maps to a module which uses __all__.
|
## Code Before:
from enthought.enable.pyglet_backend.pyglet_app import *
## Instruction:
Improve the proxy module which maps to a module which uses __all__.
## Code After:
__all__ = ["get_app", "PygletApp"]
from enthought.enable.pyglet_backend.pyglet_app import *
# Import the objects which are not declared in __all__,
# but are still defined in the real module, such that people
# can import them explicitly when needed, just as they could
# with the real module.
#
# It is unlikely that someone will import these objects, since
# they start with '_'. However, the proxy's job is to mimic the
# behavior of the real module as closely as possible.
# The proxy's job is not to define or change the API.
#
from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp
|
+
+ __all__ = ["get_app", "PygletApp"]
+
from enthought.enable.pyglet_backend.pyglet_app import *
+
+
+ # Import the objects which are not declared in __all__,
+ # but are still defined in the real module, such that people
+ # can import them explicitly when needed, just as they could
+ # with the real module.
+ #
+ # It is unlikely that someone will import these objects, since
+ # they start with '_'. However, the proxy's job is to mimic the
+ # behavior of the real module as closely as possible.
+ # The proxy's job is not to define or change the API.
+ #
+ from enthought.enable.pyglet_backend.pyglet_app import _CurrentApp, _PygletApp
+
|
850fba4b07e4c444aa8640c6f4c3816f8a3259ea
|
website_medical_patient_species/controllers/main.py
|
website_medical_patient_species/controllers/main.py
|
from openerp import http
from openerp.http import request
from openerp.addons.website_medical.controllers.main import (
WebsiteMedical
)
class WebsiteMedical(WebsiteMedical):
def _inject_medical_detail_vals(self, patient_id=0, **kwargs):
vals = super(WebsiteMedical, self)._inject_medical_detail_vals(
patient_id,
**kwargs
)
species_ids = request.env['medical.patient.species'].search([])
vals.update({
'species': species_ids,
})
return vals
|
from openerp.http import request
from openerp.addons.website_medical.controllers.main import (
WebsiteMedical
)
class WebsiteMedical(WebsiteMedical):
def _inject_medical_detail_vals(self, patient_id=0, **kwargs):
vals = super(WebsiteMedical, self)._inject_medical_detail_vals(
patient_id,
**kwargs
)
species_ids = request.env['medical.patient.species'].search([])
vals.update({
'species': species_ids,
})
return vals
|
Fix lint * Remove stray import to fix lint
|
[FIX] website_medical_patient_species: Fix lint
* Remove stray import to fix lint
|
Python
|
agpl-3.0
|
laslabs/vertical-medical,laslabs/vertical-medical
|
- from openerp import http
from openerp.http import request
from openerp.addons.website_medical.controllers.main import (
WebsiteMedical
)
class WebsiteMedical(WebsiteMedical):
def _inject_medical_detail_vals(self, patient_id=0, **kwargs):
vals = super(WebsiteMedical, self)._inject_medical_detail_vals(
patient_id,
**kwargs
)
species_ids = request.env['medical.patient.species'].search([])
vals.update({
'species': species_ids,
})
return vals
|
Fix lint * Remove stray import to fix lint
|
## Code Before:
from openerp import http
from openerp.http import request
from openerp.addons.website_medical.controllers.main import (
WebsiteMedical
)
class WebsiteMedical(WebsiteMedical):
def _inject_medical_detail_vals(self, patient_id=0, **kwargs):
vals = super(WebsiteMedical, self)._inject_medical_detail_vals(
patient_id,
**kwargs
)
species_ids = request.env['medical.patient.species'].search([])
vals.update({
'species': species_ids,
})
return vals
## Instruction:
Fix lint * Remove stray import to fix lint
## Code After:
from openerp.http import request
from openerp.addons.website_medical.controllers.main import (
WebsiteMedical
)
class WebsiteMedical(WebsiteMedical):
def _inject_medical_detail_vals(self, patient_id=0, **kwargs):
vals = super(WebsiteMedical, self)._inject_medical_detail_vals(
patient_id,
**kwargs
)
species_ids = request.env['medical.patient.species'].search([])
vals.update({
'species': species_ids,
})
return vals
|
- from openerp import http
from openerp.http import request
from openerp.addons.website_medical.controllers.main import (
WebsiteMedical
)
class WebsiteMedical(WebsiteMedical):
def _inject_medical_detail_vals(self, patient_id=0, **kwargs):
vals = super(WebsiteMedical, self)._inject_medical_detail_vals(
patient_id,
**kwargs
)
species_ids = request.env['medical.patient.species'].search([])
vals.update({
'species': species_ids,
})
return vals
|
093c9065de9e0e08f248bbb84696bf30309bd536
|
examples/parallel/timer.py
|
examples/parallel/timer.py
|
import rx
import concurrent.futures
import time
seconds = [5, 1, 2, 4, 3]
def sleep(t):
time.sleep(t)
return t
def output(result):
print '%d seconds' % result
with concurrent.futures.ProcessPoolExecutor(5) as executor:
rx.Observable.from_(seconds).flat_map(
lambda s: executor.submit(sleep, s)
).subscribe(output)
# 1 seconds
# 2 seconds
# 3 seconds
# 4 seconds
# 5 seconds
|
from __future__ import print_function
import rx
import concurrent.futures
import time
seconds = [5, 1, 2, 4, 3]
def sleep(t):
time.sleep(t)
return t
def output(result):
print('%d seconds' % result)
with concurrent.futures.ProcessPoolExecutor(5) as executor:
rx.Observable.from_(seconds).flat_map(
lambda s: executor.submit(sleep, s)
).subscribe(output)
# 1 seconds
# 2 seconds
# 3 seconds
# 4 seconds
# 5 seconds
|
Fix parallel example for Python 3
|
Fix parallel example for Python 3
|
Python
|
mit
|
dbrattli/RxPY,ReactiveX/RxPY,ReactiveX/RxPY
|
+ from __future__ import print_function
+
import rx
import concurrent.futures
import time
seconds = [5, 1, 2, 4, 3]
def sleep(t):
time.sleep(t)
return t
def output(result):
- print '%d seconds' % result
+ print('%d seconds' % result)
with concurrent.futures.ProcessPoolExecutor(5) as executor:
rx.Observable.from_(seconds).flat_map(
lambda s: executor.submit(sleep, s)
).subscribe(output)
# 1 seconds
# 2 seconds
# 3 seconds
# 4 seconds
# 5 seconds
|
Fix parallel example for Python 3
|
## Code Before:
import rx
import concurrent.futures
import time
seconds = [5, 1, 2, 4, 3]
def sleep(t):
time.sleep(t)
return t
def output(result):
print '%d seconds' % result
with concurrent.futures.ProcessPoolExecutor(5) as executor:
rx.Observable.from_(seconds).flat_map(
lambda s: executor.submit(sleep, s)
).subscribe(output)
# 1 seconds
# 2 seconds
# 3 seconds
# 4 seconds
# 5 seconds
## Instruction:
Fix parallel example for Python 3
## Code After:
from __future__ import print_function
import rx
import concurrent.futures
import time
seconds = [5, 1, 2, 4, 3]
def sleep(t):
time.sleep(t)
return t
def output(result):
print('%d seconds' % result)
with concurrent.futures.ProcessPoolExecutor(5) as executor:
rx.Observable.from_(seconds).flat_map(
lambda s: executor.submit(sleep, s)
).subscribe(output)
# 1 seconds
# 2 seconds
# 3 seconds
# 4 seconds
# 5 seconds
|
+ from __future__ import print_function
+
import rx
import concurrent.futures
import time
seconds = [5, 1, 2, 4, 3]
def sleep(t):
time.sleep(t)
return t
def output(result):
- print '%d seconds' % result
? ^
+ print('%d seconds' % result)
? ^ +
with concurrent.futures.ProcessPoolExecutor(5) as executor:
rx.Observable.from_(seconds).flat_map(
lambda s: executor.submit(sleep, s)
).subscribe(output)
# 1 seconds
# 2 seconds
# 3 seconds
# 4 seconds
# 5 seconds
|
c32118b2157e6c2cfd435461ee23edfa79aa917e
|
api/__init__.py
|
api/__init__.py
|
import ConfigParser
from peewee import *
config = ConfigParser.RawConfigParser()
config.read('server.conf')
database = SqliteDatabase('gallery.db')
from collection import CollectionModel
from album import AlbumModel
from user import UserModel, UserResource
from photo import PhotoModel
database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True)
if UserModel.select().count() == 0:
UserModel.create(
admin = True,
name = 'Admin',
username = 'Admin',
password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy'
)
from boto.s3.connection import S3Connection
s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey'))
if s3.lookup(config.get('S3', 'Bucket')) is None:
s3.create_bucket(config.get('S3', 'Bucket'))
from flask import Flask
from flask.ext.restful import Api
app = Flask(__name__)
api = Api(app)
|
import ConfigParser
from peewee import *
config = ConfigParser.RawConfigParser()
config.read('server.conf')
database = SqliteDatabase('gallery.db', threadlocals=True)
from collection import CollectionModel
from album import AlbumModel
from user import UserModel, UsersResource
from photo import PhotoModel
database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True)
if UserModel.select().count() == 0:
UserModel.create(
admin = True,
name = 'Admin',
username = 'Admin',
password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy'
)
from boto.s3.connection import S3Connection
s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey'))
if s3.lookup(config.get('S3', 'Bucket')) is None:
s3.create_bucket(config.get('S3', 'Bucket'))
from flask import Flask
from flask.ext.restful import Api
app = Flask(__name__)
api = Api(app)
api.add_resource(UsersResource, '/users/')
|
Set local threads to true for peewee
|
Set local threads to true for peewee
|
Python
|
unlicense
|
karousel/karousel
|
import ConfigParser
from peewee import *
config = ConfigParser.RawConfigParser()
config.read('server.conf')
- database = SqliteDatabase('gallery.db')
+ database = SqliteDatabase('gallery.db', threadlocals=True)
from collection import CollectionModel
from album import AlbumModel
- from user import UserModel, UserResource
+ from user import UserModel, UsersResource
from photo import PhotoModel
database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True)
if UserModel.select().count() == 0:
UserModel.create(
admin = True,
name = 'Admin',
username = 'Admin',
password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy'
)
from boto.s3.connection import S3Connection
s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey'))
if s3.lookup(config.get('S3', 'Bucket')) is None:
s3.create_bucket(config.get('S3', 'Bucket'))
from flask import Flask
from flask.ext.restful import Api
app = Flask(__name__)
api = Api(app)
+ api.add_resource(UsersResource, '/users/')
+
|
Set local threads to true for peewee
|
## Code Before:
import ConfigParser
from peewee import *
config = ConfigParser.RawConfigParser()
config.read('server.conf')
database = SqliteDatabase('gallery.db')
from collection import CollectionModel
from album import AlbumModel
from user import UserModel, UserResource
from photo import PhotoModel
database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True)
if UserModel.select().count() == 0:
UserModel.create(
admin = True,
name = 'Admin',
username = 'Admin',
password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy'
)
from boto.s3.connection import S3Connection
s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey'))
if s3.lookup(config.get('S3', 'Bucket')) is None:
s3.create_bucket(config.get('S3', 'Bucket'))
from flask import Flask
from flask.ext.restful import Api
app = Flask(__name__)
api = Api(app)
## Instruction:
Set local threads to true for peewee
## Code After:
import ConfigParser
from peewee import *
config = ConfigParser.RawConfigParser()
config.read('server.conf')
database = SqliteDatabase('gallery.db', threadlocals=True)
from collection import CollectionModel
from album import AlbumModel
from user import UserModel, UsersResource
from photo import PhotoModel
database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True)
if UserModel.select().count() == 0:
UserModel.create(
admin = True,
name = 'Admin',
username = 'Admin',
password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy'
)
from boto.s3.connection import S3Connection
s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey'))
if s3.lookup(config.get('S3', 'Bucket')) is None:
s3.create_bucket(config.get('S3', 'Bucket'))
from flask import Flask
from flask.ext.restful import Api
app = Flask(__name__)
api = Api(app)
api.add_resource(UsersResource, '/users/')
|
import ConfigParser
from peewee import *
config = ConfigParser.RawConfigParser()
config.read('server.conf')
- database = SqliteDatabase('gallery.db')
+ database = SqliteDatabase('gallery.db', threadlocals=True)
? +++++++++++++++++++
from collection import CollectionModel
from album import AlbumModel
- from user import UserModel, UserResource
+ from user import UserModel, UsersResource
? +
from photo import PhotoModel
database.create_tables([PhotoModel, AlbumModel, UserModel, CollectionModel], True)
if UserModel.select().count() == 0:
UserModel.create(
admin = True,
name = 'Admin',
username = 'Admin',
password = '$2a$12$pMtKl1b7h1sFKbMdBvPqbuza1tJN2ZNNAFMEs1RQmwqYTbBwrrKpy'
)
from boto.s3.connection import S3Connection
s3 = S3Connection(config.get('S3', 'AccessKey'), config.get('S3', 'SecretKey'))
if s3.lookup(config.get('S3', 'Bucket')) is None:
s3.create_bucket(config.get('S3', 'Bucket'))
from flask import Flask
from flask.ext.restful import Api
app = Flask(__name__)
api = Api(app)
+
+ api.add_resource(UsersResource, '/users/')
|
88fc0f980f0efa403ab5ce7d6775bce008b284fc
|
_setup_database.py
|
_setup_database.py
|
import argparse
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
from setup.create_player_seasons import create_player_seasons
from setup.create_player_seasons import create_player_data
from utils import prepare_logging
prepare_logging(log_types=['file', 'screen'])
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Setup script for NHL database creation.')
parser.add_argument(
'steps', metavar='setup_steps', help='Setup steps to execute.',
choices=['a', 't', 'd', 'p', 'ps', 'pd'])
args = parser.parse_args()
setup_steps = args.steps
# migrating teams from json file to database
if setup_steps in ['t', 'a']:
migrate_teams(simulation=True)
# creating divisions from division configuration file
if setup_steps in ['d', 'a']:
create_divisions(simulation=True)
# migrating players from json file to database
if setup_steps in ['p', 'a']:
migrate_players(simulation=True)
# retrieving player season statistics for all players in database
if setup_steps in ['ps', 'a']:
create_player_seasons(simulation=False)
# retrieving individual player data for all players in database
if setup_steps in ['pd', 'a']:
create_player_data(simulation=False)
|
import argparse
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
from setup.create_player_seasons import create_player_seasons
from setup.create_player_seasons import create_player_data
from setup.create_player_seasons import create_player_contracts
from utils import prepare_logging
prepare_logging(log_types=['file', 'screen'])
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Setup script for NHL database creation.')
parser.add_argument(
'steps', metavar='setup_steps', help='Setup steps to execute.',
choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd'])
args = parser.parse_args()
setup_steps = args.steps
# migrating teams from json file to database
if setup_steps in ['t', 'a']:
migrate_teams(simulation=True)
# creating divisions from division configuration file
if setup_steps in ['d', 'a']:
create_divisions(simulation=True)
# migrating players from json file to database
if setup_steps in ['p', 'a']:
migrate_players(simulation=True)
# retrieving player season statistics for all players in database
if setup_steps in ['ps', 'a']:
create_player_seasons(simulation=False)
# retrieving individual player data for all players in database
if setup_steps in ['pd', 'a']:
create_player_data(simulation=False)
if setup_steps in ['c']:
create_player_contracts(simulation=False)
|
Add contract retrieval option to database setup script
|
Add contract retrieval option to database setup script
|
Python
|
mit
|
leaffan/pynhldb
|
import argparse
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
from setup.create_player_seasons import create_player_seasons
from setup.create_player_seasons import create_player_data
+ from setup.create_player_seasons import create_player_contracts
from utils import prepare_logging
prepare_logging(log_types=['file', 'screen'])
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Setup script for NHL database creation.')
parser.add_argument(
'steps', metavar='setup_steps', help='Setup steps to execute.',
- choices=['a', 't', 'd', 'p', 'ps', 'pd'])
+ choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd'])
args = parser.parse_args()
setup_steps = args.steps
# migrating teams from json file to database
if setup_steps in ['t', 'a']:
migrate_teams(simulation=True)
# creating divisions from division configuration file
if setup_steps in ['d', 'a']:
create_divisions(simulation=True)
# migrating players from json file to database
if setup_steps in ['p', 'a']:
migrate_players(simulation=True)
# retrieving player season statistics for all players in database
if setup_steps in ['ps', 'a']:
create_player_seasons(simulation=False)
# retrieving individual player data for all players in database
if setup_steps in ['pd', 'a']:
create_player_data(simulation=False)
+ if setup_steps in ['c']:
+ create_player_contracts(simulation=False)
+
|
Add contract retrieval option to database setup script
|
## Code Before:
import argparse
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
from setup.create_player_seasons import create_player_seasons
from setup.create_player_seasons import create_player_data
from utils import prepare_logging
prepare_logging(log_types=['file', 'screen'])
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Setup script for NHL database creation.')
parser.add_argument(
'steps', metavar='setup_steps', help='Setup steps to execute.',
choices=['a', 't', 'd', 'p', 'ps', 'pd'])
args = parser.parse_args()
setup_steps = args.steps
# migrating teams from json file to database
if setup_steps in ['t', 'a']:
migrate_teams(simulation=True)
# creating divisions from division configuration file
if setup_steps in ['d', 'a']:
create_divisions(simulation=True)
# migrating players from json file to database
if setup_steps in ['p', 'a']:
migrate_players(simulation=True)
# retrieving player season statistics for all players in database
if setup_steps in ['ps', 'a']:
create_player_seasons(simulation=False)
# retrieving individual player data for all players in database
if setup_steps in ['pd', 'a']:
create_player_data(simulation=False)
## Instruction:
Add contract retrieval option to database setup script
## Code After:
import argparse
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
from setup.create_player_seasons import create_player_seasons
from setup.create_player_seasons import create_player_data
from setup.create_player_seasons import create_player_contracts
from utils import prepare_logging
prepare_logging(log_types=['file', 'screen'])
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Setup script for NHL database creation.')
parser.add_argument(
'steps', metavar='setup_steps', help='Setup steps to execute.',
choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd'])
args = parser.parse_args()
setup_steps = args.steps
# migrating teams from json file to database
if setup_steps in ['t', 'a']:
migrate_teams(simulation=True)
# creating divisions from division configuration file
if setup_steps in ['d', 'a']:
create_divisions(simulation=True)
# migrating players from json file to database
if setup_steps in ['p', 'a']:
migrate_players(simulation=True)
# retrieving player season statistics for all players in database
if setup_steps in ['ps', 'a']:
create_player_seasons(simulation=False)
# retrieving individual player data for all players in database
if setup_steps in ['pd', 'a']:
create_player_data(simulation=False)
if setup_steps in ['c']:
create_player_contracts(simulation=False)
|
import argparse
from setup.create_teams import migrate_teams
from setup.create_divisions import create_divisions
from setup.create_players import migrate_players
from setup.create_player_seasons import create_player_seasons
from setup.create_player_seasons import create_player_data
+ from setup.create_player_seasons import create_player_contracts
from utils import prepare_logging
prepare_logging(log_types=['file', 'screen'])
if __name__ == '__main__':
parser = argparse.ArgumentParser(
description='Setup script for NHL database creation.')
parser.add_argument(
'steps', metavar='setup_steps', help='Setup steps to execute.',
- choices=['a', 't', 'd', 'p', 'ps', 'pd'])
+ choices=['a', 'c', 't', 'd', 'p', 'ps', 'pd'])
? +++++
args = parser.parse_args()
setup_steps = args.steps
# migrating teams from json file to database
if setup_steps in ['t', 'a']:
migrate_teams(simulation=True)
# creating divisions from division configuration file
if setup_steps in ['d', 'a']:
create_divisions(simulation=True)
# migrating players from json file to database
if setup_steps in ['p', 'a']:
migrate_players(simulation=True)
# retrieving player season statistics for all players in database
if setup_steps in ['ps', 'a']:
create_player_seasons(simulation=False)
# retrieving individual player data for all players in database
if setup_steps in ['pd', 'a']:
create_player_data(simulation=False)
+
+ if setup_steps in ['c']:
+ create_player_contracts(simulation=False)
|
ed5dcd72b661878913be224d641c5595c73ef049
|
tests/test_auditory.py
|
tests/test_auditory.py
|
from __future__ import division, print_function
import pytest
import numpy as np
from pambox import auditory as aud
import scipy.io as sio
from numpy.testing import assert_allclose
def test_lowpass_filtering_of_envelope():
mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat",
squeeze_me=True)
envelope = mat['unfiltered_env']
target = mat['lp_filtered_env']
filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.)
assert_allclose(filtered_envelope, target, atol=1e-7)
|
from __future__ import division, print_function
import pytest
import numpy as np
from pambox import auditory as aud
import scipy.io as sio
from numpy.testing import assert_allclose
def test_lowpass_filtering_of_envelope():
mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat",
squeeze_me=True)
envelope = mat['unfiltered_env']
target = mat['lp_filtered_env']
filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.)
assert_allclose(filtered_envelope, target, atol=1e-7)
def test_erb():
bw = aud.erbbw(1000)
assert_allclose(bw, 132.63, rtol=1e-4)
|
Test of the erb calculation
|
Test of the erb calculation
|
Python
|
bsd-3-clause
|
achabotl/pambox
|
from __future__ import division, print_function
import pytest
import numpy as np
from pambox import auditory as aud
import scipy.io as sio
from numpy.testing import assert_allclose
def test_lowpass_filtering_of_envelope():
mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat",
squeeze_me=True)
envelope = mat['unfiltered_env']
target = mat['lp_filtered_env']
filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.)
assert_allclose(filtered_envelope, target, atol=1e-7)
+ def test_erb():
+ bw = aud.erbbw(1000)
+ assert_allclose(bw, 132.63, rtol=1e-4)
|
Test of the erb calculation
|
## Code Before:
from __future__ import division, print_function
import pytest
import numpy as np
from pambox import auditory as aud
import scipy.io as sio
from numpy.testing import assert_allclose
def test_lowpass_filtering_of_envelope():
mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat",
squeeze_me=True)
envelope = mat['unfiltered_env']
target = mat['lp_filtered_env']
filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.)
assert_allclose(filtered_envelope, target, atol=1e-7)
## Instruction:
Test of the erb calculation
## Code After:
from __future__ import division, print_function
import pytest
import numpy as np
from pambox import auditory as aud
import scipy.io as sio
from numpy.testing import assert_allclose
def test_lowpass_filtering_of_envelope():
mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat",
squeeze_me=True)
envelope = mat['unfiltered_env']
target = mat['lp_filtered_env']
filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.)
assert_allclose(filtered_envelope, target, atol=1e-7)
def test_erb():
bw = aud.erbbw(1000)
assert_allclose(bw, 132.63, rtol=1e-4)
|
from __future__ import division, print_function
import pytest
import numpy as np
from pambox import auditory as aud
import scipy.io as sio
from numpy.testing import assert_allclose
def test_lowpass_filtering_of_envelope():
mat = sio.loadmat("./test_files/test_hilbert_env_and_lp_filtering_v1.mat",
squeeze_me=True)
envelope = mat['unfiltered_env']
target = mat['lp_filtered_env']
filtered_envelope = aud.lowpass_env_filtering(envelope, 150., 1, 22050.)
assert_allclose(filtered_envelope, target, atol=1e-7)
+ def test_erb():
+ bw = aud.erbbw(1000)
+ assert_allclose(bw, 132.63, rtol=1e-4)
|
64042be2b6febf64d601adaa6f85a542ae9b876d
|
sunpy/instr/iris/iris.py
|
sunpy/instr/iris/iris.py
|
import sunpy.io
import sunpy.time
import sunpy.map
__all__ = ['SJI_to_cube']
def SJI_to_cube(filename, start=0, stop=None):
"""
Read a SJI file and return a MapCube
..warning::
This function is a very early beta and is not stable. Further work is
on going to improve SunPy IRIS support.
Parameters
----------
filename: string
File to read
start:
Temporal axis index to create MapCube from
stop:
Temporal index to stop MapCube at
Returns
-------
iris_cube: sunpy.map.MapCube
A map cube of the SJI sequence
"""
hdus = sunpy.io.read_file(filename)
#Get the time delta
time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS'])
splits = time_range.split(hdus[0][0].shape[0])
if not stop:
stop = len(splits)
headers = [hdus[0][1]]*(stop-start)
datas = hdus[0][0][start:stop]
#Make the cube:
iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)
#Set the date/time
for i,m in enumerate(iris_cube):
m.meta['DATE-OBS'] = splits[i].center().isoformat()
return iris_cube
|
import sunpy.io
import sunpy.time
import sunpy.map
__all__ = ['SJI_to_cube']
def SJI_to_cube(filename, start=0, stop=None, hdu=0):
"""
Read a SJI file and return a MapCube
..warning::
This function is a very early beta and is not stable. Further work is
on going to improve SunPy IRIS support.
Parameters
----------
filename: string
File to read
start:
Temporal axis index to create MapCube from
stop:
Temporal index to stop MapCube at
hdu:
Choose hdu index
Returns
-------
iris_cube: sunpy.map.MapCube
A map cube of the SJI sequence
"""
hdus = sunpy.io.read_file(filename)
#Get the time delta
time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS'])
splits = time_range.split(hdus[hdu][0].shape[0])
if not stop:
stop = len(splits)
headers = [hdus[hdu][1]]*(stop-start)
datas = hdus[hdu][0][start:stop]
#Make the cube:
iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)
#Set the date/time
for i,m in enumerate(iris_cube):
m.meta['DATE-OBS'] = splits[i].center().isoformat()
return iris_cube
|
Change hdu[0] to hdu for optional indexing
|
Change hdu[0] to hdu for optional indexing
|
Python
|
bsd-2-clause
|
Alex-Ian-Hamilton/sunpy,dpshelio/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy
|
import sunpy.io
import sunpy.time
import sunpy.map
__all__ = ['SJI_to_cube']
- def SJI_to_cube(filename, start=0, stop=None):
+ def SJI_to_cube(filename, start=0, stop=None, hdu=0):
"""
Read a SJI file and return a MapCube
..warning::
This function is a very early beta and is not stable. Further work is
on going to improve SunPy IRIS support.
Parameters
----------
filename: string
File to read
start:
Temporal axis index to create MapCube from
stop:
Temporal index to stop MapCube at
+ hdu:
+ Choose hdu index
+
Returns
-------
iris_cube: sunpy.map.MapCube
A map cube of the SJI sequence
"""
hdus = sunpy.io.read_file(filename)
#Get the time delta
- time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS'])
+ time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS'])
- splits = time_range.split(hdus[0][0].shape[0])
+ splits = time_range.split(hdus[hdu][0].shape[0])
if not stop:
stop = len(splits)
- headers = [hdus[0][1]]*(stop-start)
+ headers = [hdus[hdu][1]]*(stop-start)
- datas = hdus[0][0][start:stop]
+ datas = hdus[hdu][0][start:stop]
#Make the cube:
iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)
#Set the date/time
for i,m in enumerate(iris_cube):
m.meta['DATE-OBS'] = splits[i].center().isoformat()
return iris_cube
+
|
Change hdu[0] to hdu for optional indexing
|
## Code Before:
import sunpy.io
import sunpy.time
import sunpy.map
__all__ = ['SJI_to_cube']
def SJI_to_cube(filename, start=0, stop=None):
"""
Read a SJI file and return a MapCube
..warning::
This function is a very early beta and is not stable. Further work is
on going to improve SunPy IRIS support.
Parameters
----------
filename: string
File to read
start:
Temporal axis index to create MapCube from
stop:
Temporal index to stop MapCube at
Returns
-------
iris_cube: sunpy.map.MapCube
A map cube of the SJI sequence
"""
hdus = sunpy.io.read_file(filename)
#Get the time delta
time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS'])
splits = time_range.split(hdus[0][0].shape[0])
if not stop:
stop = len(splits)
headers = [hdus[0][1]]*(stop-start)
datas = hdus[0][0][start:stop]
#Make the cube:
iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)
#Set the date/time
for i,m in enumerate(iris_cube):
m.meta['DATE-OBS'] = splits[i].center().isoformat()
return iris_cube
## Instruction:
Change hdu[0] to hdu for optional indexing
## Code After:
import sunpy.io
import sunpy.time
import sunpy.map
__all__ = ['SJI_to_cube']
def SJI_to_cube(filename, start=0, stop=None, hdu=0):
"""
Read a SJI file and return a MapCube
..warning::
This function is a very early beta and is not stable. Further work is
on going to improve SunPy IRIS support.
Parameters
----------
filename: string
File to read
start:
Temporal axis index to create MapCube from
stop:
Temporal index to stop MapCube at
hdu:
Choose hdu index
Returns
-------
iris_cube: sunpy.map.MapCube
A map cube of the SJI sequence
"""
hdus = sunpy.io.read_file(filename)
#Get the time delta
time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS'])
splits = time_range.split(hdus[hdu][0].shape[0])
if not stop:
stop = len(splits)
headers = [hdus[hdu][1]]*(stop-start)
datas = hdus[hdu][0][start:stop]
#Make the cube:
iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)
#Set the date/time
for i,m in enumerate(iris_cube):
m.meta['DATE-OBS'] = splits[i].center().isoformat()
return iris_cube
|
import sunpy.io
import sunpy.time
import sunpy.map
__all__ = ['SJI_to_cube']
- def SJI_to_cube(filename, start=0, stop=None):
+ def SJI_to_cube(filename, start=0, stop=None, hdu=0):
? +++++++
"""
Read a SJI file and return a MapCube
..warning::
This function is a very early beta and is not stable. Further work is
on going to improve SunPy IRIS support.
Parameters
----------
filename: string
File to read
start:
Temporal axis index to create MapCube from
stop:
Temporal index to stop MapCube at
+ hdu:
+ Choose hdu index
+
Returns
-------
iris_cube: sunpy.map.MapCube
A map cube of the SJI sequence
"""
hdus = sunpy.io.read_file(filename)
#Get the time delta
- time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS'])
? ^ ^
+ time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS'])
? ^^^ ^^^
- splits = time_range.split(hdus[0][0].shape[0])
? ^
+ splits = time_range.split(hdus[hdu][0].shape[0])
? ^^^
if not stop:
stop = len(splits)
- headers = [hdus[0][1]]*(stop-start)
? ^
+ headers = [hdus[hdu][1]]*(stop-start)
? ^^^
- datas = hdus[0][0][start:stop]
? ^
+ datas = hdus[hdu][0][start:stop]
? ^^^
#Make the cube:
iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)
#Set the date/time
for i,m in enumerate(iris_cube):
m.meta['DATE-OBS'] = splits[i].center().isoformat()
return iris_cube
|
43f67067c470386b6b24080642cc845ec1655f58
|
utils/networking.py
|
utils/networking.py
|
import fcntl
import socket
import struct
from contextlib import contextmanager
@contextmanager
def use_interface(ifname):
"""
:type ifname: str
"""
ip = _ip_address_for_interface(ifname.encode('ascii'))
original_socket = socket.socket
def rebound_socket(*args, **kwargs):
sock = original_socket(*args, **kwargs)
sock.bind((ip, 0))
return sock
socket.socket = rebound_socket
yield
socket.socket = original_socket
def _ip_address_for_interface(ifname):
"""
:type ifname: bytes
:rtype: str
"""
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
return socket.inet_ntoa(fcntl.ioctl(
sock.fileno(),
0x8915, # SIOCGIFADDR
struct.pack('256s', ifname[:15])
)[20:24])
|
import fcntl
import socket
import struct
from contextlib import contextmanager
@contextmanager
def use_interface(ifname):
"""
:type ifname: str
"""
ip = _ip_address_for_interface(ifname)
original_socket = socket.socket
def rebound_socket(*args, **kwargs):
sock = original_socket(*args, **kwargs)
sock.bind((ip, 0))
return sock
socket.socket = rebound_socket
yield
socket.socket = original_socket
def _ip_address_for_interface(ifname):
"""
:type ifname: str
:rtype: str
"""
ifname = ifname.encode('ascii')
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
return socket.inet_ntoa(fcntl.ioctl(
sock.fileno(),
0x8915, # SIOCGIFADDR
struct.pack('256s', ifname[:15])
)[20:24])
|
Make _ip_address_for_interface easier to use
|
Make _ip_address_for_interface easier to use
|
Python
|
apache-2.0
|
OPWEN/opwen-webapp,ascoderu/opwen-webapp,ascoderu/opwen-webapp,OPWEN/opwen-webapp,OPWEN/opwen-webapp,ascoderu/opwen-cloudserver,ascoderu/opwen-cloudserver,ascoderu/opwen-webapp
|
import fcntl
import socket
import struct
from contextlib import contextmanager
@contextmanager
def use_interface(ifname):
"""
:type ifname: str
"""
- ip = _ip_address_for_interface(ifname.encode('ascii'))
+ ip = _ip_address_for_interface(ifname)
original_socket = socket.socket
def rebound_socket(*args, **kwargs):
sock = original_socket(*args, **kwargs)
sock.bind((ip, 0))
return sock
socket.socket = rebound_socket
yield
socket.socket = original_socket
def _ip_address_for_interface(ifname):
"""
- :type ifname: bytes
+ :type ifname: str
:rtype: str
"""
+ ifname = ifname.encode('ascii')
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
return socket.inet_ntoa(fcntl.ioctl(
sock.fileno(),
0x8915, # SIOCGIFADDR
struct.pack('256s', ifname[:15])
)[20:24])
|
Make _ip_address_for_interface easier to use
|
## Code Before:
import fcntl
import socket
import struct
from contextlib import contextmanager
@contextmanager
def use_interface(ifname):
"""
:type ifname: str
"""
ip = _ip_address_for_interface(ifname.encode('ascii'))
original_socket = socket.socket
def rebound_socket(*args, **kwargs):
sock = original_socket(*args, **kwargs)
sock.bind((ip, 0))
return sock
socket.socket = rebound_socket
yield
socket.socket = original_socket
def _ip_address_for_interface(ifname):
"""
:type ifname: bytes
:rtype: str
"""
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
return socket.inet_ntoa(fcntl.ioctl(
sock.fileno(),
0x8915, # SIOCGIFADDR
struct.pack('256s', ifname[:15])
)[20:24])
## Instruction:
Make _ip_address_for_interface easier to use
## Code After:
import fcntl
import socket
import struct
from contextlib import contextmanager
@contextmanager
def use_interface(ifname):
"""
:type ifname: str
"""
ip = _ip_address_for_interface(ifname)
original_socket = socket.socket
def rebound_socket(*args, **kwargs):
sock = original_socket(*args, **kwargs)
sock.bind((ip, 0))
return sock
socket.socket = rebound_socket
yield
socket.socket = original_socket
def _ip_address_for_interface(ifname):
"""
:type ifname: str
:rtype: str
"""
ifname = ifname.encode('ascii')
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
return socket.inet_ntoa(fcntl.ioctl(
sock.fileno(),
0x8915, # SIOCGIFADDR
struct.pack('256s', ifname[:15])
)[20:24])
|
import fcntl
import socket
import struct
from contextlib import contextmanager
@contextmanager
def use_interface(ifname):
"""
:type ifname: str
"""
- ip = _ip_address_for_interface(ifname.encode('ascii'))
? --------------- -
+ ip = _ip_address_for_interface(ifname)
original_socket = socket.socket
def rebound_socket(*args, **kwargs):
sock = original_socket(*args, **kwargs)
sock.bind((ip, 0))
return sock
socket.socket = rebound_socket
yield
socket.socket = original_socket
def _ip_address_for_interface(ifname):
"""
- :type ifname: bytes
? ^^ ^^
+ :type ifname: str
? ^ ^
:rtype: str
"""
+ ifname = ifname.encode('ascii')
sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
return socket.inet_ntoa(fcntl.ioctl(
sock.fileno(),
0x8915, # SIOCGIFADDR
struct.pack('256s', ifname[:15])
)[20:24])
|
ded80de3c276b57cd36d94ab393937289f772a25
|
django_prometheus/db/backends/postgresql/base.py
|
django_prometheus/db/backends/postgresql/base.py
|
import django
import psycopg2.extensions
from django_prometheus.db.common import DatabaseWrapperMixin, \
ExportingCursorWrapper
if django.VERSION >= (1, 9):
from django.db.backends.postgresql import base
else:
from django.db.backends.postgresql_psycopg2 import base
class DatabaseFeatures(base.DatabaseFeatures):
"""Our database has the exact same features as the base one."""
pass
class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper):
def get_connection_params(self):
conn_params = super(DatabaseWrapper, self).get_connection_params()
conn_params['cursor_factory'] = ExportingCursorWrapper(
psycopg2.extensions.cursor,
self.alias,
self.vendor,
)
return conn_params
def create_cursor(self, name=None):
# cursor_factory is a kwarg to connect() so restore create_cursor()'s
# default behavior
return base.DatabaseWrapper.create_cursor(self, name=name)
|
import django
import psycopg2.extensions
from django_prometheus.db.common import DatabaseWrapperMixin, \
ExportingCursorWrapper
if django.VERSION >= (1, 9):
from django.db.backends.postgresql import base
else:
from django.db.backends.postgresql_psycopg2 import base
class DatabaseFeatures(base.DatabaseFeatures):
"""Our database has the exact same features as the base one."""
pass
class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper):
def get_connection_params(self):
conn_params = super(DatabaseWrapper, self).get_connection_params()
conn_params['cursor_factory'] = ExportingCursorWrapper(
psycopg2.extensions.cursor,
self.alias,
self.vendor,
)
return conn_params
def create_cursor(self, name=None):
# cursor_factory is a kwarg to connect() so restore create_cursor()'s
# default behavior
if django.VERSION >= (1, 11, 0):
return base.DatabaseWrapper.create_cursor(self, name=name)
else:
return base.DatabaseWrapper.create_cursor(self)
|
Fix backwards compatibility for postgresql backend on Django 1.10 and earlier
|
Fix backwards compatibility for postgresql backend on Django 1.10 and earlier
|
Python
|
apache-2.0
|
korfuri/django-prometheus,obytes/django-prometheus,korfuri/django-prometheus,obytes/django-prometheus
|
import django
import psycopg2.extensions
from django_prometheus.db.common import DatabaseWrapperMixin, \
ExportingCursorWrapper
if django.VERSION >= (1, 9):
from django.db.backends.postgresql import base
else:
from django.db.backends.postgresql_psycopg2 import base
class DatabaseFeatures(base.DatabaseFeatures):
"""Our database has the exact same features as the base one."""
pass
class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper):
def get_connection_params(self):
conn_params = super(DatabaseWrapper, self).get_connection_params()
conn_params['cursor_factory'] = ExportingCursorWrapper(
psycopg2.extensions.cursor,
self.alias,
self.vendor,
)
return conn_params
def create_cursor(self, name=None):
# cursor_factory is a kwarg to connect() so restore create_cursor()'s
# default behavior
+ if django.VERSION >= (1, 11, 0):
- return base.DatabaseWrapper.create_cursor(self, name=name)
+ return base.DatabaseWrapper.create_cursor(self, name=name)
+ else:
+ return base.DatabaseWrapper.create_cursor(self)
|
Fix backwards compatibility for postgresql backend on Django 1.10 and earlier
|
## Code Before:
import django
import psycopg2.extensions
from django_prometheus.db.common import DatabaseWrapperMixin, \
ExportingCursorWrapper
if django.VERSION >= (1, 9):
from django.db.backends.postgresql import base
else:
from django.db.backends.postgresql_psycopg2 import base
class DatabaseFeatures(base.DatabaseFeatures):
"""Our database has the exact same features as the base one."""
pass
class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper):
def get_connection_params(self):
conn_params = super(DatabaseWrapper, self).get_connection_params()
conn_params['cursor_factory'] = ExportingCursorWrapper(
psycopg2.extensions.cursor,
self.alias,
self.vendor,
)
return conn_params
def create_cursor(self, name=None):
# cursor_factory is a kwarg to connect() so restore create_cursor()'s
# default behavior
return base.DatabaseWrapper.create_cursor(self, name=name)
## Instruction:
Fix backwards compatibility for postgresql backend on Django 1.10 and earlier
## Code After:
import django
import psycopg2.extensions
from django_prometheus.db.common import DatabaseWrapperMixin, \
ExportingCursorWrapper
if django.VERSION >= (1, 9):
from django.db.backends.postgresql import base
else:
from django.db.backends.postgresql_psycopg2 import base
class DatabaseFeatures(base.DatabaseFeatures):
"""Our database has the exact same features as the base one."""
pass
class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper):
def get_connection_params(self):
conn_params = super(DatabaseWrapper, self).get_connection_params()
conn_params['cursor_factory'] = ExportingCursorWrapper(
psycopg2.extensions.cursor,
self.alias,
self.vendor,
)
return conn_params
def create_cursor(self, name=None):
# cursor_factory is a kwarg to connect() so restore create_cursor()'s
# default behavior
if django.VERSION >= (1, 11, 0):
return base.DatabaseWrapper.create_cursor(self, name=name)
else:
return base.DatabaseWrapper.create_cursor(self)
|
import django
import psycopg2.extensions
from django_prometheus.db.common import DatabaseWrapperMixin, \
ExportingCursorWrapper
if django.VERSION >= (1, 9):
from django.db.backends.postgresql import base
else:
from django.db.backends.postgresql_psycopg2 import base
class DatabaseFeatures(base.DatabaseFeatures):
"""Our database has the exact same features as the base one."""
pass
class DatabaseWrapper(DatabaseWrapperMixin, base.DatabaseWrapper):
def get_connection_params(self):
conn_params = super(DatabaseWrapper, self).get_connection_params()
conn_params['cursor_factory'] = ExportingCursorWrapper(
psycopg2.extensions.cursor,
self.alias,
self.vendor,
)
return conn_params
def create_cursor(self, name=None):
# cursor_factory is a kwarg to connect() so restore create_cursor()'s
# default behavior
+ if django.VERSION >= (1, 11, 0):
- return base.DatabaseWrapper.create_cursor(self, name=name)
+ return base.DatabaseWrapper.create_cursor(self, name=name)
? ++++
+ else:
+ return base.DatabaseWrapper.create_cursor(self)
|
20e8ef6bd68100a70b9d50013630ff71d8b7ec94
|
changes/artifacts/__init__.py
|
changes/artifacts/__init__.py
|
from __future__ import absolute_import, print_function
from .manager import Manager
from .coverage import CoverageHandler
from .xunit import XunitHandler
manager = Manager()
manager.register(CoverageHandler, ['coverage.xml'])
manager.register(XunitHandler, ['xunit.xml', 'junit.xml'])
|
from __future__ import absolute_import, print_function
from .manager import Manager
from .coverage import CoverageHandler
from .xunit import XunitHandler
manager = Manager()
manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml'])
manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml'])
|
Support wildcard matches on coverage/junit results
|
Support wildcard matches on coverage/junit results
|
Python
|
apache-2.0
|
dropbox/changes,bowlofstew/changes,wfxiang08/changes,dropbox/changes,wfxiang08/changes,bowlofstew/changes,bowlofstew/changes,dropbox/changes,bowlofstew/changes,wfxiang08/changes,wfxiang08/changes,dropbox/changes
|
from __future__ import absolute_import, print_function
from .manager import Manager
from .coverage import CoverageHandler
from .xunit import XunitHandler
manager = Manager()
- manager.register(CoverageHandler, ['coverage.xml'])
+ manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml'])
- manager.register(XunitHandler, ['xunit.xml', 'junit.xml'])
+ manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml'])
|
Support wildcard matches on coverage/junit results
|
## Code Before:
from __future__ import absolute_import, print_function
from .manager import Manager
from .coverage import CoverageHandler
from .xunit import XunitHandler
manager = Manager()
manager.register(CoverageHandler, ['coverage.xml'])
manager.register(XunitHandler, ['xunit.xml', 'junit.xml'])
## Instruction:
Support wildcard matches on coverage/junit results
## Code After:
from __future__ import absolute_import, print_function
from .manager import Manager
from .coverage import CoverageHandler
from .xunit import XunitHandler
manager = Manager()
manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml'])
manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml'])
|
from __future__ import absolute_import, print_function
from .manager import Manager
from .coverage import CoverageHandler
from .xunit import XunitHandler
manager = Manager()
- manager.register(CoverageHandler, ['coverage.xml'])
+ manager.register(CoverageHandler, ['coverage.xml', '*.coverage.xml'])
? ++++++++++++++++++
- manager.register(XunitHandler, ['xunit.xml', 'junit.xml'])
+ manager.register(XunitHandler, ['xunit.xml', 'junit.xml', '*.xunit.xml', '*.junit.xml'])
? ++++++++++++++++++++++++++++++
|
799d6738bd189fa202f45c10e7b5361f71f14c57
|
bin/request_domain.py
|
bin/request_domain.py
|
import requests, sys, json
otp = sys.argv[1]
domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain'
domain_txt_path = '/opt/cloudfleet/data/config/domain.txt'
print('retrieving domain for blimp: ' + domain_req_url)
r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp})
if r.status_code == 200:
print('Success: %s' % r.text)
result_dict = r.json()
if "domain" in result_dict:
with open(domain_txt_path, 'w') as domain_txt_file:
domain_txt_file.write(result_dict['domain'])
quit(0)
else:
quit(1)
else:
print('Error: %s' % r.text)
quit(1)
|
import requests, sys, json
otp = sys.argv[1]
domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain'
domain_txt_path = '/opt/cloudfleet/data/config/domain.txt'
print('retrieving domain for blimp: ' + domain_req_url)
r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp})
if r.status_code == 200:
print('Success: %s' % r.text)
result_dict = r.json()
if "domain" in result_dict:
with open(domain_txt_path, 'w') as domain_txt_file:
domain_txt_file.write(result_dict['domain'])
quit(0)
else:
quit(1)
else:
print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text)
quit(1)
|
Clarify error if otp is wrong
|
Clarify error if otp is wrong
|
Python
|
agpl-3.0
|
cloudfleet/blimp-engineroom,cloudfleet/blimp-engineroom
|
import requests, sys, json
otp = sys.argv[1]
domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain'
domain_txt_path = '/opt/cloudfleet/data/config/domain.txt'
print('retrieving domain for blimp: ' + domain_req_url)
r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp})
if r.status_code == 200:
print('Success: %s' % r.text)
result_dict = r.json()
if "domain" in result_dict:
with open(domain_txt_path, 'w') as domain_txt_file:
domain_txt_file.write(result_dict['domain'])
quit(0)
else:
quit(1)
else:
- print('Error: %s' % r.text)
+ print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text)
quit(1)
|
Clarify error if otp is wrong
|
## Code Before:
import requests, sys, json
otp = sys.argv[1]
domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain'
domain_txt_path = '/opt/cloudfleet/data/config/domain.txt'
print('retrieving domain for blimp: ' + domain_req_url)
r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp})
if r.status_code == 200:
print('Success: %s' % r.text)
result_dict = r.json()
if "domain" in result_dict:
with open(domain_txt_path, 'w') as domain_txt_file:
domain_txt_file.write(result_dict['domain'])
quit(0)
else:
quit(1)
else:
print('Error: %s' % r.text)
quit(1)
## Instruction:
Clarify error if otp is wrong
## Code After:
import requests, sys, json
otp = sys.argv[1]
domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain'
domain_txt_path = '/opt/cloudfleet/data/config/domain.txt'
print('retrieving domain for blimp: ' + domain_req_url)
r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp})
if r.status_code == 200:
print('Success: %s' % r.text)
result_dict = r.json()
if "domain" in result_dict:
with open(domain_txt_path, 'w') as domain_txt_file:
domain_txt_file.write(result_dict['domain'])
quit(0)
else:
quit(1)
else:
print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text)
quit(1)
|
import requests, sys, json
otp = sys.argv[1]
domain_req_url = 'https://spire.cloudfleet.io/api/v1/blimp/domain'
domain_txt_path = '/opt/cloudfleet/data/config/domain.txt'
print('retrieving domain for blimp: ' + domain_req_url)
r = requests.get(domain_req_url, headers={'X-AUTH-OTP': otp})
if r.status_code == 200:
print('Success: %s' % r.text)
result_dict = r.json()
if "domain" in result_dict:
with open(domain_txt_path, 'w') as domain_txt_file:
domain_txt_file.write(result_dict['domain'])
quit(0)
else:
quit(1)
else:
- print('Error: %s' % r.text)
+ print('Error: %s. Is the OTP correct? Double-check on Spire!' % r.text)
quit(1)
|
34ac848cc19477f032a78a4ccbc782d2694d1969
|
bluebottle/votes/models.py
|
bluebottle/votes/models.py
|
from django.db import models
from django.conf import settings
from django.utils.translation import ugettext as _
from django_extensions.db.fields import CreationDateTimeField
class Vote(models.Model):
"""
Mixin for generating an invoice reference.
"""
created = CreationDateTimeField(_('created'))
project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL)
voter = models.ForeignKey(settings.AUTH_USER_MODEL,
verbose_name=_('voter'), blank=True, null=True)
def __unicode__(self):
return "{0} -> {1}".format(self.voter, self.project)
class Meta:
unique_together = (('project', 'voter'), )
|
from django.db import models
from django.conf import settings
from django.utils.translation import ugettext as _
from django_extensions.db.fields import CreationDateTimeField
class Vote(models.Model):
"""
Mixin for generating an invoice reference.
"""
created = CreationDateTimeField(_('created'))
project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL)
voter = models.ForeignKey(settings.AUTH_USER_MODEL,
verbose_name=_('voter'), blank=True, null=True)
def __unicode__(self):
return "{0} -> {1}".format(self.voter, self.project)
class Meta:
unique_together = (('project', 'voter'), )
ordering = ('-created', )
|
Sort votes by created desc
|
Sort votes by created desc
BB-4430 #resolve
|
Python
|
bsd-3-clause
|
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,jfterpstra/bluebottle,onepercentclub/bluebottle
|
from django.db import models
from django.conf import settings
from django.utils.translation import ugettext as _
from django_extensions.db.fields import CreationDateTimeField
class Vote(models.Model):
"""
Mixin for generating an invoice reference.
"""
created = CreationDateTimeField(_('created'))
project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL)
voter = models.ForeignKey(settings.AUTH_USER_MODEL,
verbose_name=_('voter'), blank=True, null=True)
def __unicode__(self):
return "{0} -> {1}".format(self.voter, self.project)
class Meta:
unique_together = (('project', 'voter'), )
+ ordering = ('-created', )
|
Sort votes by created desc
|
## Code Before:
from django.db import models
from django.conf import settings
from django.utils.translation import ugettext as _
from django_extensions.db.fields import CreationDateTimeField
class Vote(models.Model):
"""
Mixin for generating an invoice reference.
"""
created = CreationDateTimeField(_('created'))
project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL)
voter = models.ForeignKey(settings.AUTH_USER_MODEL,
verbose_name=_('voter'), blank=True, null=True)
def __unicode__(self):
return "{0} -> {1}".format(self.voter, self.project)
class Meta:
unique_together = (('project', 'voter'), )
## Instruction:
Sort votes by created desc
## Code After:
from django.db import models
from django.conf import settings
from django.utils.translation import ugettext as _
from django_extensions.db.fields import CreationDateTimeField
class Vote(models.Model):
"""
Mixin for generating an invoice reference.
"""
created = CreationDateTimeField(_('created'))
project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL)
voter = models.ForeignKey(settings.AUTH_USER_MODEL,
verbose_name=_('voter'), blank=True, null=True)
def __unicode__(self):
return "{0} -> {1}".format(self.voter, self.project)
class Meta:
unique_together = (('project', 'voter'), )
ordering = ('-created', )
|
from django.db import models
from django.conf import settings
from django.utils.translation import ugettext as _
from django_extensions.db.fields import CreationDateTimeField
class Vote(models.Model):
"""
Mixin for generating an invoice reference.
"""
created = CreationDateTimeField(_('created'))
project = models.ForeignKey(settings.PROJECTS_PROJECT_MODEL)
voter = models.ForeignKey(settings.AUTH_USER_MODEL,
verbose_name=_('voter'), blank=True, null=True)
def __unicode__(self):
return "{0} -> {1}".format(self.voter, self.project)
class Meta:
unique_together = (('project', 'voter'), )
+ ordering = ('-created', )
|
5978eedb3147bc0f124335d9e408d6c4895de3a7
|
__init__.py
|
__init__.py
|
import os
import sys
import marshal
if sys.hexversion < 0x03030000:
raise ImportError('python >= 3.3 required')
if sys.implementation.cache_tag is None:
raise ImportError('python implementation does not use bytecode')
PY_TAG = sys.implementation.cache_tag
PY_VERSION = sys.hexversion
BUNDLE_DIR = os.path.join(__path__[0], 'bundle')
BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle')
if not os.path.exists(BUNDLE_FILE):
raise ImportError('python implementation {!r} not supported'.format(PY_TAG))
with open(BUNDLE_FILE, 'rb') as _fd:
for _c in marshal.load(_fd):
eval(_c)
del _c
del _fd
|
import os
import sys
import marshal
if not hasattr(sys, 'implementation'):
raise ImportError('python >= 3.3 required')
if sys.implementation.cache_tag is None:
raise ImportError('python implementation does not use bytecode')
PY_TAG = sys.implementation.cache_tag
PY_VERSION = sys.hexversion
BUNDLE_DIR = os.path.join(__path__[0], 'bundle')
BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle')
if not os.path.exists(BUNDLE_FILE):
raise ImportError('python implementation {!r} not supported'.format(PY_TAG))
with open(BUNDLE_FILE, 'rb') as _fd:
for _c in marshal.load(_fd):
eval(_c)
del _c
del _fd
|
Use a different way of ensuring 3.3+.
|
Use a different way of ensuring 3.3+.
|
Python
|
mit
|
pyos/dg
|
import os
import sys
import marshal
- if sys.hexversion < 0x03030000:
+ if not hasattr(sys, 'implementation'):
raise ImportError('python >= 3.3 required')
if sys.implementation.cache_tag is None:
raise ImportError('python implementation does not use bytecode')
PY_TAG = sys.implementation.cache_tag
PY_VERSION = sys.hexversion
BUNDLE_DIR = os.path.join(__path__[0], 'bundle')
BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle')
if not os.path.exists(BUNDLE_FILE):
raise ImportError('python implementation {!r} not supported'.format(PY_TAG))
with open(BUNDLE_FILE, 'rb') as _fd:
for _c in marshal.load(_fd):
eval(_c)
del _c
del _fd
|
Use a different way of ensuring 3.3+.
|
## Code Before:
import os
import sys
import marshal
if sys.hexversion < 0x03030000:
raise ImportError('python >= 3.3 required')
if sys.implementation.cache_tag is None:
raise ImportError('python implementation does not use bytecode')
PY_TAG = sys.implementation.cache_tag
PY_VERSION = sys.hexversion
BUNDLE_DIR = os.path.join(__path__[0], 'bundle')
BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle')
if not os.path.exists(BUNDLE_FILE):
raise ImportError('python implementation {!r} not supported'.format(PY_TAG))
with open(BUNDLE_FILE, 'rb') as _fd:
for _c in marshal.load(_fd):
eval(_c)
del _c
del _fd
## Instruction:
Use a different way of ensuring 3.3+.
## Code After:
import os
import sys
import marshal
if not hasattr(sys, 'implementation'):
raise ImportError('python >= 3.3 required')
if sys.implementation.cache_tag is None:
raise ImportError('python implementation does not use bytecode')
PY_TAG = sys.implementation.cache_tag
PY_VERSION = sys.hexversion
BUNDLE_DIR = os.path.join(__path__[0], 'bundle')
BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle')
if not os.path.exists(BUNDLE_FILE):
raise ImportError('python implementation {!r} not supported'.format(PY_TAG))
with open(BUNDLE_FILE, 'rb') as _fd:
for _c in marshal.load(_fd):
eval(_c)
del _c
del _fd
|
import os
import sys
import marshal
- if sys.hexversion < 0x03030000:
+ if not hasattr(sys, 'implementation'):
raise ImportError('python >= 3.3 required')
if sys.implementation.cache_tag is None:
raise ImportError('python implementation does not use bytecode')
PY_TAG = sys.implementation.cache_tag
PY_VERSION = sys.hexversion
BUNDLE_DIR = os.path.join(__path__[0], 'bundle')
BUNDLE_FILE = os.path.join(BUNDLE_DIR, PY_TAG + '.dgbundle')
if not os.path.exists(BUNDLE_FILE):
raise ImportError('python implementation {!r} not supported'.format(PY_TAG))
with open(BUNDLE_FILE, 'rb') as _fd:
for _c in marshal.load(_fd):
eval(_c)
del _c
del _fd
|
fa2fb3387912474eff2b6c2a14d6304fcf5cd1f8
|
erasmus/cogs/bible/testing_server_preferences_group.py
|
erasmus/cogs/bible/testing_server_preferences_group.py
|
from __future__ import annotations
from typing import TYPE_CHECKING
from botus_receptus.app_commands import test_guilds_only
from discord import app_commands
from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup
if TYPE_CHECKING:
from ...erasmus import Erasmus
from ...l10n import GroupLocalizer
from .types import ParentCog
@app_commands.default_permissions(administrator=True)
@app_commands.guild_only()
@test_guilds_only
class TestingServerPreferencesGroup(
app_commands.Group, name='test-server-prefs', description='Testing group'
):
bot: Erasmus
localizer: GroupLocalizer
daily_bread = DailyBreadPreferencesGroup()
def initialize_from_parent(self, parent: ParentCog, /) -> None:
self.bot = parent.bot
self.localizer = parent.localizer.for_group(self)
self.daily_bread.initialize_from_parent(self)
|
from __future__ import annotations
from typing import TYPE_CHECKING
from botus_receptus.app_commands import test_guilds_only
from discord import app_commands
from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup
if TYPE_CHECKING:
from ...erasmus import Erasmus
from ...l10n import GroupLocalizer
from .types import ParentCog
@app_commands.default_permissions(administrator=True)
@app_commands.guild_only()
@test_guilds_only
class TestingServerPreferencesGroup(
app_commands.Group, name='test-server-prefs', description='Testing group'
):
bot: Erasmus
localizer: GroupLocalizer
daily_bread = DailyBreadPreferencesGroup()
def initialize_from_parent(self, parent: ParentCog, /) -> None:
self.bot = parent.bot
self.localizer = parent.localizer.for_group('serverprefs')
self.daily_bread.initialize_from_parent(self)
|
Use serverprefs localizer for TestingServerPreferencesGroup
|
Use serverprefs localizer for TestingServerPreferencesGroup
|
Python
|
bsd-3-clause
|
bryanforbes/Erasmus
|
from __future__ import annotations
from typing import TYPE_CHECKING
from botus_receptus.app_commands import test_guilds_only
from discord import app_commands
from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup
if TYPE_CHECKING:
from ...erasmus import Erasmus
from ...l10n import GroupLocalizer
from .types import ParentCog
@app_commands.default_permissions(administrator=True)
@app_commands.guild_only()
@test_guilds_only
class TestingServerPreferencesGroup(
app_commands.Group, name='test-server-prefs', description='Testing group'
):
bot: Erasmus
localizer: GroupLocalizer
daily_bread = DailyBreadPreferencesGroup()
def initialize_from_parent(self, parent: ParentCog, /) -> None:
self.bot = parent.bot
- self.localizer = parent.localizer.for_group(self)
+ self.localizer = parent.localizer.for_group('serverprefs')
self.daily_bread.initialize_from_parent(self)
|
Use serverprefs localizer for TestingServerPreferencesGroup
|
## Code Before:
from __future__ import annotations
from typing import TYPE_CHECKING
from botus_receptus.app_commands import test_guilds_only
from discord import app_commands
from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup
if TYPE_CHECKING:
from ...erasmus import Erasmus
from ...l10n import GroupLocalizer
from .types import ParentCog
@app_commands.default_permissions(administrator=True)
@app_commands.guild_only()
@test_guilds_only
class TestingServerPreferencesGroup(
app_commands.Group, name='test-server-prefs', description='Testing group'
):
bot: Erasmus
localizer: GroupLocalizer
daily_bread = DailyBreadPreferencesGroup()
def initialize_from_parent(self, parent: ParentCog, /) -> None:
self.bot = parent.bot
self.localizer = parent.localizer.for_group(self)
self.daily_bread.initialize_from_parent(self)
## Instruction:
Use serverprefs localizer for TestingServerPreferencesGroup
## Code After:
from __future__ import annotations
from typing import TYPE_CHECKING
from botus_receptus.app_commands import test_guilds_only
from discord import app_commands
from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup
if TYPE_CHECKING:
from ...erasmus import Erasmus
from ...l10n import GroupLocalizer
from .types import ParentCog
@app_commands.default_permissions(administrator=True)
@app_commands.guild_only()
@test_guilds_only
class TestingServerPreferencesGroup(
app_commands.Group, name='test-server-prefs', description='Testing group'
):
bot: Erasmus
localizer: GroupLocalizer
daily_bread = DailyBreadPreferencesGroup()
def initialize_from_parent(self, parent: ParentCog, /) -> None:
self.bot = parent.bot
self.localizer = parent.localizer.for_group('serverprefs')
self.daily_bread.initialize_from_parent(self)
|
from __future__ import annotations
from typing import TYPE_CHECKING
from botus_receptus.app_commands import test_guilds_only
from discord import app_commands
from .daily_bread.daily_bread_preferences_group import DailyBreadPreferencesGroup
if TYPE_CHECKING:
from ...erasmus import Erasmus
from ...l10n import GroupLocalizer
from .types import ParentCog
@app_commands.default_permissions(administrator=True)
@app_commands.guild_only()
@test_guilds_only
class TestingServerPreferencesGroup(
app_commands.Group, name='test-server-prefs', description='Testing group'
):
bot: Erasmus
localizer: GroupLocalizer
daily_bread = DailyBreadPreferencesGroup()
def initialize_from_parent(self, parent: ParentCog, /) -> None:
self.bot = parent.bot
- self.localizer = parent.localizer.for_group(self)
? ^
+ self.localizer = parent.localizer.for_group('serverprefs')
? + ^^^^^^^ ++
self.daily_bread.initialize_from_parent(self)
|
8c26cb08dd08b7e34352e51b06ecb9129ac201a1
|
stagecraft/libs/schemas/schemas.py
|
stagecraft/libs/schemas/schemas.py
|
from django.conf import settings
from json import loads as json_loads
from os import path
def get_schema():
schema_root = path.join(
settings.BASE_DIR,
'stagecraft/apps/datasets/schemas/timestamp.json'
)
with open(schema_root) as f:
json_f = json_loads(f.read())
return json_f
|
from django.conf import settings
from json import loads as json_loads
from os import path
def get_schema():
schema_root = path.join(
settings.BASE_DIR,
'stagecraft/apps/datasets/schemas/timestamp.json'
)
with open(schema_root) as f:
schema = json_loads(f.read())
return schema
|
Make the schema return object a bit more obvious and descriptive
|
Make the schema return object a bit more obvious and descriptive
|
Python
|
mit
|
alphagov/stagecraft,alphagov/stagecraft,alphagov/stagecraft,alphagov/stagecraft
|
from django.conf import settings
from json import loads as json_loads
from os import path
def get_schema():
schema_root = path.join(
settings.BASE_DIR,
'stagecraft/apps/datasets/schemas/timestamp.json'
)
with open(schema_root) as f:
- json_f = json_loads(f.read())
+ schema = json_loads(f.read())
- return json_f
+ return schema
|
Make the schema return object a bit more obvious and descriptive
|
## Code Before:
from django.conf import settings
from json import loads as json_loads
from os import path
def get_schema():
schema_root = path.join(
settings.BASE_DIR,
'stagecraft/apps/datasets/schemas/timestamp.json'
)
with open(schema_root) as f:
json_f = json_loads(f.read())
return json_f
## Instruction:
Make the schema return object a bit more obvious and descriptive
## Code After:
from django.conf import settings
from json import loads as json_loads
from os import path
def get_schema():
schema_root = path.join(
settings.BASE_DIR,
'stagecraft/apps/datasets/schemas/timestamp.json'
)
with open(schema_root) as f:
schema = json_loads(f.read())
return schema
|
from django.conf import settings
from json import loads as json_loads
from os import path
def get_schema():
schema_root = path.join(
settings.BASE_DIR,
'stagecraft/apps/datasets/schemas/timestamp.json'
)
with open(schema_root) as f:
- json_f = json_loads(f.read())
? - ^^^^
+ schema = json_loads(f.read())
? ^^^^^
- return json_f
+ return schema
|
1de4a0edd0f3c43b53e3a91c10d23155889791c6
|
tca/chat/tests.py
|
tca/chat/tests.py
|
from django.test import TestCase
# Create your tests here.
|
from django.test import TestCase
from django.core.urlresolvers import reverse
from urllib import urlencode
import json
class ViewTestCaseMixin(object):
"""A mixin providing some convenience methods for testing views.
Expects that a ``view_name`` property exists on the class which
mixes it in.
"""
def get_view_url(self, *args, **kwargs):
return reverse(self.view_name, args=args, kwargs=kwargs)
def build_url(self, base_url, query_dict=None):
url_template = "{base_url}?{query_string}"
if query_dict is None:
return base_url
return url_template.format(
base_url=base_url,
query_string=urlencode(query_dict)
)
def get(self, parameters=None, *args, **kwargs):
"""
Sends a GET request to the view-under-test and returns the response
:param parameters: The query string parameters of the GET request
"""
base_url = self.get_view_url(*args, **kwargs)
return self.client.get(self.build_url(base_url, parameters))
def post(self, body=None, content_type='application/json', *args, **kwargs):
"""
Sends a POST request to the view-under-test and returns the response
:param body: The content to be included in the body of the request
"""
base_url = self.get_view_url(*args, **kwargs)
if body is None:
body = ''
return self.client.post(
self.build_url(base_url),
body,
content_type=content_type)
def post_json(self, json_payload, *args, **kwargs):
"""
Sends a POST request to the view-under-test and returns the response.
The body of the POST request is formed by serializing the
``json_payload`` object to JSON.
"""
payload = json.dumps(json_payload)
return self.post(
body=payload,
content_type='application/json',
*args, **kwargs)
|
Add a helper mixin for view test cases
|
Add a helper mixin for view test cases
The mixin defines some helper methods which are useful when testing
views (REST endpoints).
|
Python
|
bsd-3-clause
|
mlalic/TumCampusAppBackend,mlalic/TumCampusAppBackend
|
from django.test import TestCase
- # Create your tests here.
+ from django.core.urlresolvers import reverse
+ from urllib import urlencode
+ import json
+
+
+ class ViewTestCaseMixin(object):
+ """A mixin providing some convenience methods for testing views.
+
+ Expects that a ``view_name`` property exists on the class which
+ mixes it in.
+ """
+
+ def get_view_url(self, *args, **kwargs):
+ return reverse(self.view_name, args=args, kwargs=kwargs)
+
+ def build_url(self, base_url, query_dict=None):
+ url_template = "{base_url}?{query_string}"
+
+ if query_dict is None:
+ return base_url
+
+ return url_template.format(
+ base_url=base_url,
+ query_string=urlencode(query_dict)
+ )
+
+ def get(self, parameters=None, *args, **kwargs):
+ """
+ Sends a GET request to the view-under-test and returns the response
+
+ :param parameters: The query string parameters of the GET request
+ """
+ base_url = self.get_view_url(*args, **kwargs)
+
+ return self.client.get(self.build_url(base_url, parameters))
+
+ def post(self, body=None, content_type='application/json', *args, **kwargs):
+ """
+ Sends a POST request to the view-under-test and returns the response
+
+ :param body: The content to be included in the body of the request
+ """
+ base_url = self.get_view_url(*args, **kwargs)
+
+ if body is None:
+ body = ''
+
+ return self.client.post(
+ self.build_url(base_url),
+ body,
+ content_type=content_type)
+
+ def post_json(self, json_payload, *args, **kwargs):
+ """
+ Sends a POST request to the view-under-test and returns the response.
+ The body of the POST request is formed by serializing the
+ ``json_payload`` object to JSON.
+ """
+ payload = json.dumps(json_payload)
+
+ return self.post(
+ body=payload,
+ content_type='application/json',
+ *args, **kwargs)
+
|
Add a helper mixin for view test cases
|
## Code Before:
from django.test import TestCase
# Create your tests here.
## Instruction:
Add a helper mixin for view test cases
## Code After:
from django.test import TestCase
from django.core.urlresolvers import reverse
from urllib import urlencode
import json
class ViewTestCaseMixin(object):
"""A mixin providing some convenience methods for testing views.
Expects that a ``view_name`` property exists on the class which
mixes it in.
"""
def get_view_url(self, *args, **kwargs):
return reverse(self.view_name, args=args, kwargs=kwargs)
def build_url(self, base_url, query_dict=None):
url_template = "{base_url}?{query_string}"
if query_dict is None:
return base_url
return url_template.format(
base_url=base_url,
query_string=urlencode(query_dict)
)
def get(self, parameters=None, *args, **kwargs):
"""
Sends a GET request to the view-under-test and returns the response
:param parameters: The query string parameters of the GET request
"""
base_url = self.get_view_url(*args, **kwargs)
return self.client.get(self.build_url(base_url, parameters))
def post(self, body=None, content_type='application/json', *args, **kwargs):
"""
Sends a POST request to the view-under-test and returns the response
:param body: The content to be included in the body of the request
"""
base_url = self.get_view_url(*args, **kwargs)
if body is None:
body = ''
return self.client.post(
self.build_url(base_url),
body,
content_type=content_type)
def post_json(self, json_payload, *args, **kwargs):
"""
Sends a POST request to the view-under-test and returns the response.
The body of the POST request is formed by serializing the
``json_payload`` object to JSON.
"""
payload = json.dumps(json_payload)
return self.post(
body=payload,
content_type='application/json',
*args, **kwargs)
|
from django.test import TestCase
- # Create your tests here.
+ from django.core.urlresolvers import reverse
+ from urllib import urlencode
+
+ import json
+
+
+ class ViewTestCaseMixin(object):
+ """A mixin providing some convenience methods for testing views.
+
+ Expects that a ``view_name`` property exists on the class which
+ mixes it in.
+ """
+
+ def get_view_url(self, *args, **kwargs):
+ return reverse(self.view_name, args=args, kwargs=kwargs)
+
+ def build_url(self, base_url, query_dict=None):
+ url_template = "{base_url}?{query_string}"
+
+ if query_dict is None:
+ return base_url
+
+ return url_template.format(
+ base_url=base_url,
+ query_string=urlencode(query_dict)
+ )
+
+ def get(self, parameters=None, *args, **kwargs):
+ """
+ Sends a GET request to the view-under-test and returns the response
+
+ :param parameters: The query string parameters of the GET request
+ """
+ base_url = self.get_view_url(*args, **kwargs)
+
+ return self.client.get(self.build_url(base_url, parameters))
+
+ def post(self, body=None, content_type='application/json', *args, **kwargs):
+ """
+ Sends a POST request to the view-under-test and returns the response
+
+ :param body: The content to be included in the body of the request
+ """
+ base_url = self.get_view_url(*args, **kwargs)
+
+ if body is None:
+ body = ''
+
+ return self.client.post(
+ self.build_url(base_url),
+ body,
+ content_type=content_type)
+
+ def post_json(self, json_payload, *args, **kwargs):
+ """
+ Sends a POST request to the view-under-test and returns the response.
+ The body of the POST request is formed by serializing the
+ ``json_payload`` object to JSON.
+ """
+ payload = json.dumps(json_payload)
+
+ return self.post(
+ body=payload,
+ content_type='application/json',
+ *args, **kwargs)
|
fce1b1bdb5a39bbe57b750cd453a9697b8447d6b
|
chat.py
|
chat.py
|
import re
from redis import Redis
import json
from datetime import datetime
def is_valid_chatroom(chatroom):
return re.match('[A-Za-z_\\d]+$', chatroom) is not None
def get_redis():
return Redis()
def get_conversation(chatroom):
if chatroom is None or len(chatroom) == 0:
return None
# if chatroom doesn't exist create it!
storage = get_redis()
return [
json.loads(m)
for m in storage.lrange('notifexample:' + chatroom, 0, -1)
]
def send_message(chatroom, user_id, name, message):
if '<script>' in message:
message += '-- Not this time DefConFags'
storage = get_redis()
now = datetime.now()
created_on = now.strftime('%Y-%m-%d %H:%M:%S')
storage.rpush(
'notifexample:' + chatroom,
json.dumps({
'author': name,
'userID': user_id,
'message': message,
'createdOn': created_on
})
)
|
import re
from redis import Redis
import json
from datetime import datetime
def is_valid_chatroom(chatroom):
return re.match('[A-Za-z_\\d]+$', chatroom) is not None
def get_redis():
return Redis()
def get_conversation(chatroom):
if chatroom is None or len(chatroom) == 0:
return None
storage = get_redis()
return [
json.loads(m)
for m in storage.lrange('notifexample:' + chatroom, 0, -1)
]
def send_message(chatroom, user_id, name, message):
if '<script>' in message:
message += '-- Not this time DefConFags'
storage = get_redis()
now = datetime.now()
created_on = now.strftime('%Y-%m-%d %H:%M:%S')
# if chatroom doesn't exist create it!
storage.rpush(
'notifexample:' + chatroom,
json.dumps({
'author': name,
'userID': user_id,
'message': message,
'createdOn': created_on
})
)
|
Correct position of comment :)
|
Correct position of comment :)
|
Python
|
bsd-3-clause
|
arturosevilla/notification-server-example,arturosevilla/notification-server-example
|
import re
from redis import Redis
import json
from datetime import datetime
def is_valid_chatroom(chatroom):
return re.match('[A-Za-z_\\d]+$', chatroom) is not None
def get_redis():
return Redis()
def get_conversation(chatroom):
if chatroom is None or len(chatroom) == 0:
return None
- # if chatroom doesn't exist create it!
storage = get_redis()
return [
json.loads(m)
for m in storage.lrange('notifexample:' + chatroom, 0, -1)
]
def send_message(chatroom, user_id, name, message):
if '<script>' in message:
message += '-- Not this time DefConFags'
storage = get_redis()
now = datetime.now()
created_on = now.strftime('%Y-%m-%d %H:%M:%S')
+ # if chatroom doesn't exist create it!
storage.rpush(
'notifexample:' + chatroom,
json.dumps({
'author': name,
'userID': user_id,
'message': message,
'createdOn': created_on
})
)
|
Correct position of comment :)
|
## Code Before:
import re
from redis import Redis
import json
from datetime import datetime
def is_valid_chatroom(chatroom):
return re.match('[A-Za-z_\\d]+$', chatroom) is not None
def get_redis():
return Redis()
def get_conversation(chatroom):
if chatroom is None or len(chatroom) == 0:
return None
# if chatroom doesn't exist create it!
storage = get_redis()
return [
json.loads(m)
for m in storage.lrange('notifexample:' + chatroom, 0, -1)
]
def send_message(chatroom, user_id, name, message):
if '<script>' in message:
message += '-- Not this time DefConFags'
storage = get_redis()
now = datetime.now()
created_on = now.strftime('%Y-%m-%d %H:%M:%S')
storage.rpush(
'notifexample:' + chatroom,
json.dumps({
'author': name,
'userID': user_id,
'message': message,
'createdOn': created_on
})
)
## Instruction:
Correct position of comment :)
## Code After:
import re
from redis import Redis
import json
from datetime import datetime
def is_valid_chatroom(chatroom):
return re.match('[A-Za-z_\\d]+$', chatroom) is not None
def get_redis():
return Redis()
def get_conversation(chatroom):
if chatroom is None or len(chatroom) == 0:
return None
storage = get_redis()
return [
json.loads(m)
for m in storage.lrange('notifexample:' + chatroom, 0, -1)
]
def send_message(chatroom, user_id, name, message):
if '<script>' in message:
message += '-- Not this time DefConFags'
storage = get_redis()
now = datetime.now()
created_on = now.strftime('%Y-%m-%d %H:%M:%S')
# if chatroom doesn't exist create it!
storage.rpush(
'notifexample:' + chatroom,
json.dumps({
'author': name,
'userID': user_id,
'message': message,
'createdOn': created_on
})
)
|
import re
from redis import Redis
import json
from datetime import datetime
def is_valid_chatroom(chatroom):
return re.match('[A-Za-z_\\d]+$', chatroom) is not None
def get_redis():
return Redis()
def get_conversation(chatroom):
if chatroom is None or len(chatroom) == 0:
return None
- # if chatroom doesn't exist create it!
storage = get_redis()
return [
json.loads(m)
for m in storage.lrange('notifexample:' + chatroom, 0, -1)
]
def send_message(chatroom, user_id, name, message):
if '<script>' in message:
message += '-- Not this time DefConFags'
storage = get_redis()
now = datetime.now()
created_on = now.strftime('%Y-%m-%d %H:%M:%S')
+ # if chatroom doesn't exist create it!
storage.rpush(
'notifexample:' + chatroom,
json.dumps({
'author': name,
'userID': user_id,
'message': message,
'createdOn': created_on
})
)
|
175c72d97d073a64714cebef05bd37f0221f94fa
|
test_octave_kernel.py
|
test_octave_kernel.py
|
"""Example use of jupyter_kernel_test, with tests for IPython."""
import sys
import unittest
import jupyter_kernel_test as jkt
class OctaveKernelTests(jkt.KernelTests):
kernel_name = "octave"
language_name = "octave"
code_hello_world = "disp('hello, world')"
code_display_data = [
{'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'},
{'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'}
] if sys.platform == 'darwin' else []
completion_samples = [
{
'text': 'one',
'matches': {'ones', 'onenormest'},
},
]
code_page_something = "ones?"
if __name__ == '__main__':
unittest.main()
|
"""Example use of jupyter_kernel_test, with tests for IPython."""
import sys
import unittest
import jupyter_kernel_test as jkt
class OctaveKernelTests(jkt.KernelTests):
kernel_name = "octave"
language_name = "octave"
code_hello_world = "disp('hello, world')"
code_display_data = [
{'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'},
{'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'}
] if sys.platform == 'darwin' else []
completion_samples = [
{
'text': 'acos',
'matches': {'acos', 'acosd', 'acosh'},
},
]
code_page_something = "ones?"
if __name__ == '__main__':
unittest.main()
|
Fix tests with Octave 5.
|
Fix tests with Octave 5.
|
Python
|
bsd-3-clause
|
Calysto/octave_kernel,Calysto/octave_kernel
|
"""Example use of jupyter_kernel_test, with tests for IPython."""
import sys
import unittest
import jupyter_kernel_test as jkt
class OctaveKernelTests(jkt.KernelTests):
kernel_name = "octave"
language_name = "octave"
code_hello_world = "disp('hello, world')"
code_display_data = [
{'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'},
{'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'}
] if sys.platform == 'darwin' else []
completion_samples = [
{
- 'text': 'one',
+ 'text': 'acos',
- 'matches': {'ones', 'onenormest'},
+ 'matches': {'acos', 'acosd', 'acosh'},
},
]
code_page_something = "ones?"
if __name__ == '__main__':
unittest.main()
|
Fix tests with Octave 5.
|
## Code Before:
"""Example use of jupyter_kernel_test, with tests for IPython."""
import sys
import unittest
import jupyter_kernel_test as jkt
class OctaveKernelTests(jkt.KernelTests):
kernel_name = "octave"
language_name = "octave"
code_hello_world = "disp('hello, world')"
code_display_data = [
{'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'},
{'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'}
] if sys.platform == 'darwin' else []
completion_samples = [
{
'text': 'one',
'matches': {'ones', 'onenormest'},
},
]
code_page_something = "ones?"
if __name__ == '__main__':
unittest.main()
## Instruction:
Fix tests with Octave 5.
## Code After:
"""Example use of jupyter_kernel_test, with tests for IPython."""
import sys
import unittest
import jupyter_kernel_test as jkt
class OctaveKernelTests(jkt.KernelTests):
kernel_name = "octave"
language_name = "octave"
code_hello_world = "disp('hello, world')"
code_display_data = [
{'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'},
{'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'}
] if sys.platform == 'darwin' else []
completion_samples = [
{
'text': 'acos',
'matches': {'acos', 'acosd', 'acosh'},
},
]
code_page_something = "ones?"
if __name__ == '__main__':
unittest.main()
|
"""Example use of jupyter_kernel_test, with tests for IPython."""
import sys
import unittest
import jupyter_kernel_test as jkt
class OctaveKernelTests(jkt.KernelTests):
kernel_name = "octave"
language_name = "octave"
code_hello_world = "disp('hello, world')"
code_display_data = [
{'code': '%plot -f png\nplot([1,2,3])', 'mime': 'image/png'},
{'code': '%plot -f svg\nplot([1,2,3])', 'mime': 'image/svg+xml'}
] if sys.platform == 'darwin' else []
completion_samples = [
{
- 'text': 'one',
? ^^
+ 'text': 'acos',
? ++ ^
- 'matches': {'ones', 'onenormest'},
? -- ^^^ --- ^
+ 'matches': {'acos', 'acosd', 'acosh'},
? ++ ++ ^^^^^^^^ ^
},
]
code_page_something = "ones?"
if __name__ == '__main__':
unittest.main()
|
4d1ab55f2bbe8041421002a91dc4f58783913591
|
services/search_indexes.py
|
services/search_indexes.py
|
from aldryn_search.utils import get_index_base
from .models import Service
class ServiceIndex(get_index_base()):
haystack_use_for_indexing = True
index_title = True
def get_title(self, obj):
# XXX what about language? concatenate all available languages?
return obj.name_en
def get_index_queryset(self, language):
# XXX exclude objects with blank name for the selected language, not simply for EN
return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='')
def get_model(self):
return Service
def get_search_data(self, service, language, request):
# XXX return data for the selected language, not simply for EN
return ' '.join((
service.provider.name_en,
service.name_en,
service.area_of_service.name_en,
service.description_en,
service.additional_info_en,
service.type.name_en,
))
|
from aldryn_search.utils import get_index_base
from .models import Service
class ServiceIndex(get_index_base()):
haystack_use_for_indexing = True
index_title = True
def get_title(self, obj):
return obj.name
def get_index_queryset(self, language):
# For this language's index, don't include services with no name
# provided in this language.
return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{
'name_%s' % language: ''
})
def get_model(self):
return Service
def get_search_data(self, service, language, request):
description = getattr(service, 'description_%s' % language, '')
additional_info = getattr(service, 'additional_info_%s' % language, '')
return ' '.join((
service.provider.name,
service.name,
service.area_of_service.name,
description,
additional_info,
service.type.name,
))
|
Implement language-specific aspects of indexing
|
Implement language-specific aspects of indexing
|
Python
|
bsd-3-clause
|
theirc/ServiceInfo,theirc/ServiceInfo,theirc/ServiceInfo,theirc/ServiceInfo
|
from aldryn_search.utils import get_index_base
from .models import Service
class ServiceIndex(get_index_base()):
haystack_use_for_indexing = True
index_title = True
def get_title(self, obj):
- # XXX what about language? concatenate all available languages?
- return obj.name_en
+ return obj.name
def get_index_queryset(self, language):
- # XXX exclude objects with blank name for the selected language, not simply for EN
+ # For this language's index, don't include services with no name
+ # provided in this language.
- return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='')
+ return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{
+ 'name_%s' % language: ''
+ })
def get_model(self):
return Service
def get_search_data(self, service, language, request):
- # XXX return data for the selected language, not simply for EN
+ description = getattr(service, 'description_%s' % language, '')
+ additional_info = getattr(service, 'additional_info_%s' % language, '')
return ' '.join((
- service.provider.name_en,
+ service.provider.name,
- service.name_en,
+ service.name,
- service.area_of_service.name_en,
+ service.area_of_service.name,
- service.description_en,
+ description,
- service.additional_info_en,
+ additional_info,
- service.type.name_en,
+ service.type.name,
))
|
Implement language-specific aspects of indexing
|
## Code Before:
from aldryn_search.utils import get_index_base
from .models import Service
class ServiceIndex(get_index_base()):
haystack_use_for_indexing = True
index_title = True
def get_title(self, obj):
# XXX what about language? concatenate all available languages?
return obj.name_en
def get_index_queryset(self, language):
# XXX exclude objects with blank name for the selected language, not simply for EN
return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='')
def get_model(self):
return Service
def get_search_data(self, service, language, request):
# XXX return data for the selected language, not simply for EN
return ' '.join((
service.provider.name_en,
service.name_en,
service.area_of_service.name_en,
service.description_en,
service.additional_info_en,
service.type.name_en,
))
## Instruction:
Implement language-specific aspects of indexing
## Code After:
from aldryn_search.utils import get_index_base
from .models import Service
class ServiceIndex(get_index_base()):
haystack_use_for_indexing = True
index_title = True
def get_title(self, obj):
return obj.name
def get_index_queryset(self, language):
# For this language's index, don't include services with no name
# provided in this language.
return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{
'name_%s' % language: ''
})
def get_model(self):
return Service
def get_search_data(self, service, language, request):
description = getattr(service, 'description_%s' % language, '')
additional_info = getattr(service, 'additional_info_%s' % language, '')
return ' '.join((
service.provider.name,
service.name,
service.area_of_service.name,
description,
additional_info,
service.type.name,
))
|
from aldryn_search.utils import get_index_base
from .models import Service
class ServiceIndex(get_index_base()):
haystack_use_for_indexing = True
index_title = True
def get_title(self, obj):
- # XXX what about language? concatenate all available languages?
- return obj.name_en
? ---
+ return obj.name
def get_index_queryset(self, language):
- # XXX exclude objects with blank name for the selected language, not simply for EN
+ # For this language's index, don't include services with no name
+ # provided in this language.
- return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(name_en='')
? ^^^^^^^^^^^
+ return Service.objects.filter(status=Service.STATUS_CURRENT).exclude(**{
? ^^^
+ 'name_%s' % language: ''
+ })
def get_model(self):
return Service
def get_search_data(self, service, language, request):
- # XXX return data for the selected language, not simply for EN
+ description = getattr(service, 'description_%s' % language, '')
+ additional_info = getattr(service, 'additional_info_%s' % language, '')
return ' '.join((
- service.provider.name_en,
? ---
+ service.provider.name,
- service.name_en,
? ---
+ service.name,
- service.area_of_service.name_en,
? ---
+ service.area_of_service.name,
- service.description_en,
? -------- ---
+ description,
- service.additional_info_en,
? -------- ---
+ additional_info,
- service.type.name_en,
? ---
+ service.type.name,
))
|
609864faf36b9a82db9fd63d28b5a0da7a22c4f5
|
eforge/__init__.py
|
eforge/__init__.py
|
from eforge.menu import ItemOrder
EFORGE_PLUGIN = {
'name': 'EForge Core',
'credit': 'Copyright © 2010 Element43 and contributors',
'provides': {
'mnu': [('project-page', ItemOrder(000, 'Summary'))],
},
}
VERSION = (0, 5, 0, 'beta 1')
def get_version():
return '%d.%d.%d %s' % VERSION
|
from eforge.menu import ItemOrder
EFORGE_PLUGIN = {
'name': 'EForge Core',
'credit': 'Copyright © 2010 Element43 and contributors',
'provides': {
'mnu': [('project-page', ItemOrder(000, 'Summary'))],
},
}
VERSION = (0, 5, 99, '(git master)')
def get_version():
return '%d.%d.%d %s' % VERSION
|
Change master version information to 0.5.99 (git master)
|
Change master version information to 0.5.99 (git master)
Todo: We should probably add the smarts to EForge to grab the git
revision for master, at least if Dulwich is installed :-)
|
Python
|
isc
|
oshepherd/eforge,oshepherd/eforge,oshepherd/eforge
|
from eforge.menu import ItemOrder
EFORGE_PLUGIN = {
'name': 'EForge Core',
'credit': 'Copyright © 2010 Element43 and contributors',
'provides': {
'mnu': [('project-page', ItemOrder(000, 'Summary'))],
},
}
- VERSION = (0, 5, 0, 'beta 1')
+ VERSION = (0, 5, 99, '(git master)')
def get_version():
return '%d.%d.%d %s' % VERSION
|
Change master version information to 0.5.99 (git master)
|
## Code Before:
from eforge.menu import ItemOrder
EFORGE_PLUGIN = {
'name': 'EForge Core',
'credit': 'Copyright © 2010 Element43 and contributors',
'provides': {
'mnu': [('project-page', ItemOrder(000, 'Summary'))],
},
}
VERSION = (0, 5, 0, 'beta 1')
def get_version():
return '%d.%d.%d %s' % VERSION
## Instruction:
Change master version information to 0.5.99 (git master)
## Code After:
from eforge.menu import ItemOrder
EFORGE_PLUGIN = {
'name': 'EForge Core',
'credit': 'Copyright © 2010 Element43 and contributors',
'provides': {
'mnu': [('project-page', ItemOrder(000, 'Summary'))],
},
}
VERSION = (0, 5, 99, '(git master)')
def get_version():
return '%d.%d.%d %s' % VERSION
|
from eforge.menu import ItemOrder
EFORGE_PLUGIN = {
'name': 'EForge Core',
'credit': 'Copyright © 2010 Element43 and contributors',
'provides': {
'mnu': [('project-page', ItemOrder(000, 'Summary'))],
},
}
- VERSION = (0, 5, 0, 'beta 1')
+ VERSION = (0, 5, 99, '(git master)')
def get_version():
return '%d.%d.%d %s' % VERSION
|
959b5fd80a2eeb4ddb56dea07edd16c1aeabc4ff
|
userprofile/admin.py
|
userprofile/admin.py
|
from django.contrib import admin
from .models import Profile, Skill, DutyTime, Group
admin.site.register(Profile)
admin.site.register(Skill)
admin.site.register(DutyTime)
admin.site.register(Group)
|
from django.contrib import admin
from .models import Profile, Skill, DutyTime, Group
class ProfileAdmin(admin.ModelAdmin):
list_filter = (
('tos_accepted', admin.BooleanFieldListFilter),
)
admin.site.register(Profile, ProfileAdmin)
admin.site.register(Skill)
admin.site.register(DutyTime)
admin.site.register(Group)
|
Add filtering option to see profiles that have not accepted new tos
|
Add filtering option to see profiles that have not accepted new tos
|
Python
|
mit
|
hackerspace-ntnu/website,hackerspace-ntnu/website,hackerspace-ntnu/website
|
from django.contrib import admin
from .models import Profile, Skill, DutyTime, Group
+
+ class ProfileAdmin(admin.ModelAdmin):
+ list_filter = (
+ ('tos_accepted', admin.BooleanFieldListFilter),
+ )
+
+
- admin.site.register(Profile)
+ admin.site.register(Profile, ProfileAdmin)
admin.site.register(Skill)
admin.site.register(DutyTime)
admin.site.register(Group)
|
Add filtering option to see profiles that have not accepted new tos
|
## Code Before:
from django.contrib import admin
from .models import Profile, Skill, DutyTime, Group
admin.site.register(Profile)
admin.site.register(Skill)
admin.site.register(DutyTime)
admin.site.register(Group)
## Instruction:
Add filtering option to see profiles that have not accepted new tos
## Code After:
from django.contrib import admin
from .models import Profile, Skill, DutyTime, Group
class ProfileAdmin(admin.ModelAdmin):
list_filter = (
('tos_accepted', admin.BooleanFieldListFilter),
)
admin.site.register(Profile, ProfileAdmin)
admin.site.register(Skill)
admin.site.register(DutyTime)
admin.site.register(Group)
|
from django.contrib import admin
from .models import Profile, Skill, DutyTime, Group
+
+ class ProfileAdmin(admin.ModelAdmin):
+ list_filter = (
+ ('tos_accepted', admin.BooleanFieldListFilter),
+ )
+
+
- admin.site.register(Profile)
+ admin.site.register(Profile, ProfileAdmin)
? ++++++++++++++
admin.site.register(Skill)
admin.site.register(DutyTime)
admin.site.register(Group)
|
6fbe58692005e5c8b7a9c4f4e98984ae86d347a2
|
pinax/messages/context_processors.py
|
pinax/messages/context_processors.py
|
from .models import Thread
def user_messages(request):
c = {}
if request.user.is_authenticated():
c["inbox_count"] = Thread.inbox(request.user).count()
return c
|
from .models import Thread
def user_messages(request):
c = {}
if request.user.is_authenticated():
c["inbox_threads"] = Thread.inbox(request.user)
c["unread_threads"] = Thread.unread(request.user)
return c
|
Return querysets in context processor to be more useful
|
Return querysets in context processor to be more useful
|
Python
|
mit
|
eldarion/user_messages,pinax/pinax-messages,pinax/pinax-messages,arthur-wsw/pinax-messages,eldarion/user_messages,arthur-wsw/pinax-messages
|
from .models import Thread
def user_messages(request):
c = {}
if request.user.is_authenticated():
- c["inbox_count"] = Thread.inbox(request.user).count()
+ c["inbox_threads"] = Thread.inbox(request.user)
+ c["unread_threads"] = Thread.unread(request.user)
return c
|
Return querysets in context processor to be more useful
|
## Code Before:
from .models import Thread
def user_messages(request):
c = {}
if request.user.is_authenticated():
c["inbox_count"] = Thread.inbox(request.user).count()
return c
## Instruction:
Return querysets in context processor to be more useful
## Code After:
from .models import Thread
def user_messages(request):
c = {}
if request.user.is_authenticated():
c["inbox_threads"] = Thread.inbox(request.user)
c["unread_threads"] = Thread.unread(request.user)
return c
|
from .models import Thread
def user_messages(request):
c = {}
if request.user.is_authenticated():
- c["inbox_count"] = Thread.inbox(request.user).count()
? ---- --------
+ c["inbox_threads"] = Thread.inbox(request.user)
? ++++++
+ c["unread_threads"] = Thread.unread(request.user)
return c
|
aeb68225cc9c999b51b1733bffaf684280044c97
|
salt/utils/yamldumper.py
|
salt/utils/yamldumper.py
|
'''
salt.utils.yamldumper
~~~~~~~~~~~~~~~~~~~~~
'''
from __future__ import absolute_import
try:
from yaml import CDumper as Dumper
from yaml import CSafeDumper as SafeDumper
except ImportError:
from yaml import Dumper
from yaml import SafeDumper
from salt.utils.odict import OrderedDict
class OrderedDumper(Dumper):
'''
A YAML dumper that represents python OrderedDict as simple YAML map.
'''
class SafeOrderedDumper(SafeDumper):
'''
A YAML safe dumper that represents python OrderedDict as simple YAML map.
'''
def represent_ordereddict(dumper, data):
return dumper.represent_dict(data.items())
OrderedDumper.add_representer(OrderedDict, represent_ordereddict)
SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
|
'''
salt.utils.yamldumper
~~~~~~~~~~~~~~~~~~~~~
'''
# pylint: disable=W0232
# class has no __init__ method
from __future__ import absolute_import
try:
from yaml import CDumper as Dumper
from yaml import CSafeDumper as SafeDumper
except ImportError:
from yaml import Dumper
from yaml import SafeDumper
from salt.utils.odict import OrderedDict
class OrderedDumper(Dumper):
'''
A YAML dumper that represents python OrderedDict as simple YAML map.
'''
class SafeOrderedDumper(SafeDumper):
'''
A YAML safe dumper that represents python OrderedDict as simple YAML map.
'''
def represent_ordereddict(dumper, data):
return dumper.represent_dict(data.items())
OrderedDumper.add_representer(OrderedDict, represent_ordereddict)
SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
|
Disable W0232, no `__init__` method.
|
Disable W0232, no `__init__` method.
|
Python
|
apache-2.0
|
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
|
'''
salt.utils.yamldumper
~~~~~~~~~~~~~~~~~~~~~
'''
+ # pylint: disable=W0232
+ # class has no __init__ method
from __future__ import absolute_import
try:
from yaml import CDumper as Dumper
from yaml import CSafeDumper as SafeDumper
except ImportError:
from yaml import Dumper
from yaml import SafeDumper
from salt.utils.odict import OrderedDict
class OrderedDumper(Dumper):
'''
A YAML dumper that represents python OrderedDict as simple YAML map.
'''
class SafeOrderedDumper(SafeDumper):
'''
A YAML safe dumper that represents python OrderedDict as simple YAML map.
'''
def represent_ordereddict(dumper, data):
return dumper.represent_dict(data.items())
OrderedDumper.add_representer(OrderedDict, represent_ordereddict)
SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
|
Disable W0232, no `__init__` method.
|
## Code Before:
'''
salt.utils.yamldumper
~~~~~~~~~~~~~~~~~~~~~
'''
from __future__ import absolute_import
try:
from yaml import CDumper as Dumper
from yaml import CSafeDumper as SafeDumper
except ImportError:
from yaml import Dumper
from yaml import SafeDumper
from salt.utils.odict import OrderedDict
class OrderedDumper(Dumper):
'''
A YAML dumper that represents python OrderedDict as simple YAML map.
'''
class SafeOrderedDumper(SafeDumper):
'''
A YAML safe dumper that represents python OrderedDict as simple YAML map.
'''
def represent_ordereddict(dumper, data):
return dumper.represent_dict(data.items())
OrderedDumper.add_representer(OrderedDict, represent_ordereddict)
SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
## Instruction:
Disable W0232, no `__init__` method.
## Code After:
'''
salt.utils.yamldumper
~~~~~~~~~~~~~~~~~~~~~
'''
# pylint: disable=W0232
# class has no __init__ method
from __future__ import absolute_import
try:
from yaml import CDumper as Dumper
from yaml import CSafeDumper as SafeDumper
except ImportError:
from yaml import Dumper
from yaml import SafeDumper
from salt.utils.odict import OrderedDict
class OrderedDumper(Dumper):
'''
A YAML dumper that represents python OrderedDict as simple YAML map.
'''
class SafeOrderedDumper(SafeDumper):
'''
A YAML safe dumper that represents python OrderedDict as simple YAML map.
'''
def represent_ordereddict(dumper, data):
return dumper.represent_dict(data.items())
OrderedDumper.add_representer(OrderedDict, represent_ordereddict)
SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
|
'''
salt.utils.yamldumper
~~~~~~~~~~~~~~~~~~~~~
'''
+ # pylint: disable=W0232
+ # class has no __init__ method
from __future__ import absolute_import
try:
from yaml import CDumper as Dumper
from yaml import CSafeDumper as SafeDumper
except ImportError:
from yaml import Dumper
from yaml import SafeDumper
from salt.utils.odict import OrderedDict
class OrderedDumper(Dumper):
'''
A YAML dumper that represents python OrderedDict as simple YAML map.
'''
class SafeOrderedDumper(SafeDumper):
'''
A YAML safe dumper that represents python OrderedDict as simple YAML map.
'''
def represent_ordereddict(dumper, data):
return dumper.represent_dict(data.items())
OrderedDumper.add_representer(OrderedDict, represent_ordereddict)
SafeOrderedDumper.add_representer(OrderedDict, represent_ordereddict)
|
a7867806a6bd3abfd6bf2bcac6c490965be000e2
|
tests/test_completeness.py
|
tests/test_completeness.py
|
import unittest as unittest
from syntax import Syntax
from jscodegen import CodeGenerator
def add_cases(generator):
def class_decorator(cls):
"""Add tests to `cls` generated by `generator()`."""
for f, token in generator():
test = lambda self, i=token, f=f: f(self, i)
test.__name__ = "test %s" % token.name
setattr(cls, test.__name__, test)
return cls
return class_decorator
def _test_tokens():
def t(self, to):
c = CodeGenerator({})
func_name = to.name.lower()
try:
getattr(c, func_name)
self.assertTrue(True, func_name)
except AttributeError:
self.fail("Not implemented: %s" % func_name)
for token in Syntax:
yield t, token
class TestCase(unittest.TestCase):
pass
TestCase = add_cases(_test_tokens)(TestCase)
if __name__=="__main__":
unittest.main()
|
import unittest as unittest
from jscodegen.syntax import Syntax
from jscodegen import CodeGenerator
def add_cases(generator):
def class_decorator(cls):
"""Add tests to `cls` generated by `generator()`."""
for f, token in generator():
test = lambda self, i=token, f=f: f(self, i)
test.__name__ = "test %s" % token.name
setattr(cls, test.__name__, test)
return cls
return class_decorator
def _test_tokens():
def t(self, to):
c = CodeGenerator({})
func_name = to.name.lower()
try:
getattr(c, func_name)
self.assertTrue(True, func_name)
except AttributeError:
self.fail("Not implemented: %s" % func_name)
for token in Syntax:
yield t, token
class TestCase(unittest.TestCase):
pass
TestCase = add_cases(_test_tokens)(TestCase)
if __name__=="__main__":
unittest.main()
|
Fix an issue in the tests
|
Fix an issue in the tests
|
Python
|
mit
|
ksons/jscodegen.py
|
import unittest as unittest
- from syntax import Syntax
+ from jscodegen.syntax import Syntax
from jscodegen import CodeGenerator
def add_cases(generator):
def class_decorator(cls):
"""Add tests to `cls` generated by `generator()`."""
for f, token in generator():
test = lambda self, i=token, f=f: f(self, i)
test.__name__ = "test %s" % token.name
setattr(cls, test.__name__, test)
return cls
return class_decorator
def _test_tokens():
def t(self, to):
c = CodeGenerator({})
func_name = to.name.lower()
try:
getattr(c, func_name)
self.assertTrue(True, func_name)
except AttributeError:
self.fail("Not implemented: %s" % func_name)
for token in Syntax:
yield t, token
class TestCase(unittest.TestCase):
pass
TestCase = add_cases(_test_tokens)(TestCase)
if __name__=="__main__":
unittest.main()
+
|
Fix an issue in the tests
|
## Code Before:
import unittest as unittest
from syntax import Syntax
from jscodegen import CodeGenerator
def add_cases(generator):
def class_decorator(cls):
"""Add tests to `cls` generated by `generator()`."""
for f, token in generator():
test = lambda self, i=token, f=f: f(self, i)
test.__name__ = "test %s" % token.name
setattr(cls, test.__name__, test)
return cls
return class_decorator
def _test_tokens():
def t(self, to):
c = CodeGenerator({})
func_name = to.name.lower()
try:
getattr(c, func_name)
self.assertTrue(True, func_name)
except AttributeError:
self.fail("Not implemented: %s" % func_name)
for token in Syntax:
yield t, token
class TestCase(unittest.TestCase):
pass
TestCase = add_cases(_test_tokens)(TestCase)
if __name__=="__main__":
unittest.main()
## Instruction:
Fix an issue in the tests
## Code After:
import unittest as unittest
from jscodegen.syntax import Syntax
from jscodegen import CodeGenerator
def add_cases(generator):
def class_decorator(cls):
"""Add tests to `cls` generated by `generator()`."""
for f, token in generator():
test = lambda self, i=token, f=f: f(self, i)
test.__name__ = "test %s" % token.name
setattr(cls, test.__name__, test)
return cls
return class_decorator
def _test_tokens():
def t(self, to):
c = CodeGenerator({})
func_name = to.name.lower()
try:
getattr(c, func_name)
self.assertTrue(True, func_name)
except AttributeError:
self.fail("Not implemented: %s" % func_name)
for token in Syntax:
yield t, token
class TestCase(unittest.TestCase):
pass
TestCase = add_cases(_test_tokens)(TestCase)
if __name__=="__main__":
unittest.main()
|
import unittest as unittest
- from syntax import Syntax
+ from jscodegen.syntax import Syntax
? ++++++++++
from jscodegen import CodeGenerator
def add_cases(generator):
def class_decorator(cls):
"""Add tests to `cls` generated by `generator()`."""
for f, token in generator():
test = lambda self, i=token, f=f: f(self, i)
test.__name__ = "test %s" % token.name
setattr(cls, test.__name__, test)
return cls
return class_decorator
def _test_tokens():
def t(self, to):
c = CodeGenerator({})
func_name = to.name.lower()
try:
getattr(c, func_name)
self.assertTrue(True, func_name)
except AttributeError:
self.fail("Not implemented: %s" % func_name)
for token in Syntax:
yield t, token
class TestCase(unittest.TestCase):
pass
TestCase = add_cases(_test_tokens)(TestCase)
if __name__=="__main__":
unittest.main()
|
7bd82f6feb1a34dd7b855cfe2f421232229e19db
|
pages/search_indexes.py
|
pages/search_indexes.py
|
"""Django haystack `SearchIndex` module."""
from pages.models import Page
from haystack.indexes import SearchIndex, CharField, DateTimeField
from haystack import site
class PageIndex(SearchIndex):
"""Search index for pages content."""
text = CharField(document=True, use_template=True)
publication_date = DateTimeField(model_attr='publication_date')
def get_queryset(self):
"""Used when the entire index for model is updated."""
return Page.objects.published()
site.register(Page, PageIndex)
|
"""Django haystack `SearchIndex` module."""
from pages.models import Page
from haystack.indexes import SearchIndex, CharField, DateTimeField
from haystack import site
class PageIndex(SearchIndex):
"""Search index for pages content."""
text = CharField(document=True, use_template=True)
title = CharField(model_attr='title')
publication_date = DateTimeField(model_attr='publication_date')
def get_queryset(self):
"""Used when the entire index for model is updated."""
return Page.objects.published()
site.register(Page, PageIndex)
|
Add a title attribute to the SearchIndex for pages.
|
Add a title attribute to the SearchIndex for pages.
This is useful when displaying a list of search results because we
can display the title of the result without hitting the database to
actually pull the page.
|
Python
|
bsd-3-clause
|
remik/django-page-cms,pombredanne/django-page-cms-1,oliciv/django-page-cms,remik/django-page-cms,remik/django-page-cms,batiste/django-page-cms,oliciv/django-page-cms,batiste/django-page-cms,akaihola/django-page-cms,remik/django-page-cms,akaihola/django-page-cms,pombredanne/django-page-cms-1,oliciv/django-page-cms,batiste/django-page-cms,akaihola/django-page-cms,pombredanne/django-page-cms-1
|
"""Django haystack `SearchIndex` module."""
from pages.models import Page
from haystack.indexes import SearchIndex, CharField, DateTimeField
from haystack import site
class PageIndex(SearchIndex):
"""Search index for pages content."""
text = CharField(document=True, use_template=True)
+ title = CharField(model_attr='title')
publication_date = DateTimeField(model_attr='publication_date')
def get_queryset(self):
"""Used when the entire index for model is updated."""
return Page.objects.published()
site.register(Page, PageIndex)
|
Add a title attribute to the SearchIndex for pages.
|
## Code Before:
"""Django haystack `SearchIndex` module."""
from pages.models import Page
from haystack.indexes import SearchIndex, CharField, DateTimeField
from haystack import site
class PageIndex(SearchIndex):
"""Search index for pages content."""
text = CharField(document=True, use_template=True)
publication_date = DateTimeField(model_attr='publication_date')
def get_queryset(self):
"""Used when the entire index for model is updated."""
return Page.objects.published()
site.register(Page, PageIndex)
## Instruction:
Add a title attribute to the SearchIndex for pages.
## Code After:
"""Django haystack `SearchIndex` module."""
from pages.models import Page
from haystack.indexes import SearchIndex, CharField, DateTimeField
from haystack import site
class PageIndex(SearchIndex):
"""Search index for pages content."""
text = CharField(document=True, use_template=True)
title = CharField(model_attr='title')
publication_date = DateTimeField(model_attr='publication_date')
def get_queryset(self):
"""Used when the entire index for model is updated."""
return Page.objects.published()
site.register(Page, PageIndex)
|
"""Django haystack `SearchIndex` module."""
from pages.models import Page
from haystack.indexes import SearchIndex, CharField, DateTimeField
from haystack import site
class PageIndex(SearchIndex):
"""Search index for pages content."""
text = CharField(document=True, use_template=True)
+ title = CharField(model_attr='title')
publication_date = DateTimeField(model_attr='publication_date')
def get_queryset(self):
"""Used when the entire index for model is updated."""
return Page.objects.published()
site.register(Page, PageIndex)
|
b87073e7c7d4387b6608142de7fd6216a1d093b9
|
setup.py
|
setup.py
|
from distutils.core import setup
setup(name = "bitey",
description="Bitcode Import Tool",
long_description = """
Bitey allows LLVM bitcode to be directly imported into Python as
an high performance extension module without the need for writing wrappers.
""",
license="""BSD""",
version = "0.0",
author = "David Beazley",
author_email = "[email protected]",
maintainer = "David Beazley",
maintainer_email = "[email protected]",
url = "https://github.com/dabeaz/bitey/",
packages = ['bitey'],
classifiers = [
'Programming Language :: Python :: 2',
]
)
|
from distutils.core import setup
setup(name = "bitey",
description="Bitcode Import Tool",
long_description = """
Bitey allows LLVM bitcode to be directly imported into Python as
an high performance extension module without the need for writing wrappers.
""",
license="""BSD""",
version = "0.0",
author = "David Beazley",
author_email = "[email protected]",
maintainer = "David Beazley",
maintainer_email = "[email protected]",
url = "https://github.com/dabeaz/bitey/",
packages = ['bitey'],
classifiers = [
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
]
)
|
Add trove classifier for Python 3
|
Add trove classifier for Python 3
|
Python
|
bsd-3-clause
|
dabeaz/bitey,dabeaz/bitey
|
from distutils.core import setup
setup(name = "bitey",
description="Bitcode Import Tool",
long_description = """
Bitey allows LLVM bitcode to be directly imported into Python as
an high performance extension module without the need for writing wrappers.
""",
license="""BSD""",
version = "0.0",
author = "David Beazley",
author_email = "[email protected]",
maintainer = "David Beazley",
maintainer_email = "[email protected]",
url = "https://github.com/dabeaz/bitey/",
packages = ['bitey'],
classifiers = [
'Programming Language :: Python :: 2',
+ 'Programming Language :: Python :: 3',
]
)
|
Add trove classifier for Python 3
|
## Code Before:
from distutils.core import setup
setup(name = "bitey",
description="Bitcode Import Tool",
long_description = """
Bitey allows LLVM bitcode to be directly imported into Python as
an high performance extension module without the need for writing wrappers.
""",
license="""BSD""",
version = "0.0",
author = "David Beazley",
author_email = "[email protected]",
maintainer = "David Beazley",
maintainer_email = "[email protected]",
url = "https://github.com/dabeaz/bitey/",
packages = ['bitey'],
classifiers = [
'Programming Language :: Python :: 2',
]
)
## Instruction:
Add trove classifier for Python 3
## Code After:
from distutils.core import setup
setup(name = "bitey",
description="Bitcode Import Tool",
long_description = """
Bitey allows LLVM bitcode to be directly imported into Python as
an high performance extension module without the need for writing wrappers.
""",
license="""BSD""",
version = "0.0",
author = "David Beazley",
author_email = "[email protected]",
maintainer = "David Beazley",
maintainer_email = "[email protected]",
url = "https://github.com/dabeaz/bitey/",
packages = ['bitey'],
classifiers = [
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
]
)
|
from distutils.core import setup
setup(name = "bitey",
description="Bitcode Import Tool",
long_description = """
Bitey allows LLVM bitcode to be directly imported into Python as
an high performance extension module without the need for writing wrappers.
""",
license="""BSD""",
version = "0.0",
author = "David Beazley",
author_email = "[email protected]",
maintainer = "David Beazley",
maintainer_email = "[email protected]",
url = "https://github.com/dabeaz/bitey/",
packages = ['bitey'],
classifiers = [
'Programming Language :: Python :: 2',
+ 'Programming Language :: Python :: 3',
]
)
|
e5fd0b527877f5fab1d1a2e76ce32062a4a8d697
|
bika/lims/browser/batch/samples.py
|
bika/lims/browser/batch/samples.py
|
from bika.lims.browser.sample import SamplesView as _SV
from bika.lims.permissions import *
from Products.CMFCore.utils import getToolByName
from zope.interface import implements
from Products.CMFPlone.utils import safe_unicode
import plone
class SamplesView(_SV):
def __init__(self, context, request):
super(SamplesView, self).__init__(context, request)
self.view_url = self.context.absolute_url() + "/samples"
if 'path' in self.contentFilter:
del(self.contentFilter['path'])
def contentsMethod(self, contentFilter):
tool = getToolByName(self.context, self.catalog)
state = [x for x in self.review_states if x['id'] == self.review_state][0]
for k, v in state['contentFilter'].items():
self.contentFilter[k] = v
tool_samples = tool(contentFilter)
samples = {}
for sample in (p.getObject() for p in tool_samples):
for ar in sample.getAnalysisRequests():
if ar.getBatch().UID() == self.context.UID():
samples[sample.getId()] = sample
return samples.values()
|
from bika.lims.browser.sample import SamplesView as _SV
from bika.lims.permissions import *
from Products.CMFCore.utils import getToolByName
from zope.interface import implements
from Products.CMFPlone.utils import safe_unicode
import plone
class SamplesView(_SV):
def __init__(self, context, request):
super(SamplesView, self).__init__(context, request)
self.view_url = self.context.absolute_url() + "/samples"
if 'path' in self.contentFilter:
del(self.contentFilter['path'])
def contentsMethod(self, contentFilter):
tool = getToolByName(self.context, self.catalog)
state = [x for x in self.review_states if x['id'] == self.review_state][0]
for k, v in state['contentFilter'].items():
self.contentFilter[k] = v
tool_samples = tool(contentFilter)
samples = {}
for sample in (p.getObject() for p in tool_samples):
for ar in sample.getAnalysisRequests():
batch = ar.getBatch()
if batch and ar.getBatch().UID() == self.context.UID():
samples[sample.getId()] = sample
return samples.values()
|
Fix exception - batch is not required field of AR
|
Fix exception - batch is not required field of AR
|
Python
|
agpl-3.0
|
DeBortoliWines/Bika-LIMS,anneline/Bika-LIMS,labsanmartin/Bika-LIMS,rockfruit/bika.lims,anneline/Bika-LIMS,labsanmartin/Bika-LIMS,veroc/Bika-LIMS,anneline/Bika-LIMS,DeBortoliWines/Bika-LIMS,veroc/Bika-LIMS,veroc/Bika-LIMS,labsanmartin/Bika-LIMS,DeBortoliWines/Bika-LIMS,rockfruit/bika.lims
|
from bika.lims.browser.sample import SamplesView as _SV
from bika.lims.permissions import *
from Products.CMFCore.utils import getToolByName
from zope.interface import implements
from Products.CMFPlone.utils import safe_unicode
import plone
class SamplesView(_SV):
def __init__(self, context, request):
super(SamplesView, self).__init__(context, request)
self.view_url = self.context.absolute_url() + "/samples"
if 'path' in self.contentFilter:
del(self.contentFilter['path'])
def contentsMethod(self, contentFilter):
tool = getToolByName(self.context, self.catalog)
state = [x for x in self.review_states if x['id'] == self.review_state][0]
for k, v in state['contentFilter'].items():
self.contentFilter[k] = v
tool_samples = tool(contentFilter)
samples = {}
for sample in (p.getObject() for p in tool_samples):
for ar in sample.getAnalysisRequests():
+ batch = ar.getBatch()
- if ar.getBatch().UID() == self.context.UID():
+ if batch and ar.getBatch().UID() == self.context.UID():
- samples[sample.getId()] = sample
+ samples[sample.getId()] = sample
return samples.values()
|
Fix exception - batch is not required field of AR
|
## Code Before:
from bika.lims.browser.sample import SamplesView as _SV
from bika.lims.permissions import *
from Products.CMFCore.utils import getToolByName
from zope.interface import implements
from Products.CMFPlone.utils import safe_unicode
import plone
class SamplesView(_SV):
def __init__(self, context, request):
super(SamplesView, self).__init__(context, request)
self.view_url = self.context.absolute_url() + "/samples"
if 'path' in self.contentFilter:
del(self.contentFilter['path'])
def contentsMethod(self, contentFilter):
tool = getToolByName(self.context, self.catalog)
state = [x for x in self.review_states if x['id'] == self.review_state][0]
for k, v in state['contentFilter'].items():
self.contentFilter[k] = v
tool_samples = tool(contentFilter)
samples = {}
for sample in (p.getObject() for p in tool_samples):
for ar in sample.getAnalysisRequests():
if ar.getBatch().UID() == self.context.UID():
samples[sample.getId()] = sample
return samples.values()
## Instruction:
Fix exception - batch is not required field of AR
## Code After:
from bika.lims.browser.sample import SamplesView as _SV
from bika.lims.permissions import *
from Products.CMFCore.utils import getToolByName
from zope.interface import implements
from Products.CMFPlone.utils import safe_unicode
import plone
class SamplesView(_SV):
def __init__(self, context, request):
super(SamplesView, self).__init__(context, request)
self.view_url = self.context.absolute_url() + "/samples"
if 'path' in self.contentFilter:
del(self.contentFilter['path'])
def contentsMethod(self, contentFilter):
tool = getToolByName(self.context, self.catalog)
state = [x for x in self.review_states if x['id'] == self.review_state][0]
for k, v in state['contentFilter'].items():
self.contentFilter[k] = v
tool_samples = tool(contentFilter)
samples = {}
for sample in (p.getObject() for p in tool_samples):
for ar in sample.getAnalysisRequests():
batch = ar.getBatch()
if batch and ar.getBatch().UID() == self.context.UID():
samples[sample.getId()] = sample
return samples.values()
|
from bika.lims.browser.sample import SamplesView as _SV
from bika.lims.permissions import *
from Products.CMFCore.utils import getToolByName
from zope.interface import implements
from Products.CMFPlone.utils import safe_unicode
import plone
class SamplesView(_SV):
def __init__(self, context, request):
super(SamplesView, self).__init__(context, request)
self.view_url = self.context.absolute_url() + "/samples"
if 'path' in self.contentFilter:
del(self.contentFilter['path'])
def contentsMethod(self, contentFilter):
tool = getToolByName(self.context, self.catalog)
state = [x for x in self.review_states if x['id'] == self.review_state][0]
for k, v in state['contentFilter'].items():
self.contentFilter[k] = v
tool_samples = tool(contentFilter)
samples = {}
for sample in (p.getObject() for p in tool_samples):
for ar in sample.getAnalysisRequests():
+ batch = ar.getBatch()
- if ar.getBatch().UID() == self.context.UID():
+ if batch and ar.getBatch().UID() == self.context.UID():
? ++++++++++
- samples[sample.getId()] = sample
+ samples[sample.getId()] = sample
? ++++
return samples.values()
|
0241e253c68ca6862a3da26d29a649f65c27ae36
|
demos/chatroom/experiment.py
|
demos/chatroom/experiment.py
|
"""Coordination chatroom game."""
import dallinger as dlgr
from dallinger.config import get_config
try:
unicode = unicode
except NameError: # Python 3
unicode = str
config = get_config()
def extra_settings():
config.register('network', unicode)
config.register('n', int)
class CoordinationChatroom(dlgr.experiments.Experiment):
"""Define the structure of the experiment."""
def __init__(self, session):
"""Initialize the experiment."""
super(CoordinationChatroom, self).__init__(session)
self.experiment_repeats = 1
self.num_participants = config.get('n')
self.initial_recruitment_size = self.num_participants
self.quorum = self.num_participants
self.config = config
if not self.config.ready:
self.config.load_config()
self.setup()
def create_network(self):
"""Create a new network by reading the configuration file."""
class_ = getattr(
dlgr.networks,
self.config.get('network')
)
return class_(max_size=self.num_participants)
def info_post_request(self, node, info):
"""Run when a request to create an info is complete."""
for agent in node.neighbors():
node.transmit(what=info, to_whom=agent)
def create_node(self, participant, network):
"""Create a node for a participant."""
return dlgr.nodes.Agent(network=network, participant=participant)
|
"""Coordination chatroom game."""
import dallinger as dlgr
from dallinger.compat import unicode
from dallinger.config import get_config
config = get_config()
def extra_settings():
config.register('network', unicode)
config.register('n', int)
class CoordinationChatroom(dlgr.experiments.Experiment):
"""Define the structure of the experiment."""
def __init__(self, session):
"""Initialize the experiment."""
super(CoordinationChatroom, self).__init__(session)
self.experiment_repeats = 1
self.num_participants = config.get('n')
self.initial_recruitment_size = self.num_participants
self.quorum = self.num_participants
self.config = config
if not self.config.ready:
self.config.load_config()
self.setup()
def create_network(self):
"""Create a new network by reading the configuration file."""
class_ = getattr(
dlgr.networks,
self.config.get('network')
)
return class_(max_size=self.num_participants)
def info_post_request(self, node, info):
"""Run when a request to create an info is complete."""
for agent in node.neighbors():
node.transmit(what=info, to_whom=agent)
def create_node(self, participant, network):
"""Create a node for a participant."""
return dlgr.nodes.Agent(network=network, participant=participant)
|
Use compat for unicode import
|
Use compat for unicode import
|
Python
|
mit
|
Dallinger/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger
|
"""Coordination chatroom game."""
import dallinger as dlgr
+ from dallinger.compat import unicode
from dallinger.config import get_config
- try:
- unicode = unicode
- except NameError: # Python 3
- unicode = str
config = get_config()
def extra_settings():
config.register('network', unicode)
config.register('n', int)
class CoordinationChatroom(dlgr.experiments.Experiment):
"""Define the structure of the experiment."""
def __init__(self, session):
"""Initialize the experiment."""
super(CoordinationChatroom, self).__init__(session)
self.experiment_repeats = 1
self.num_participants = config.get('n')
self.initial_recruitment_size = self.num_participants
self.quorum = self.num_participants
self.config = config
if not self.config.ready:
self.config.load_config()
self.setup()
def create_network(self):
"""Create a new network by reading the configuration file."""
class_ = getattr(
dlgr.networks,
self.config.get('network')
)
return class_(max_size=self.num_participants)
def info_post_request(self, node, info):
"""Run when a request to create an info is complete."""
for agent in node.neighbors():
node.transmit(what=info, to_whom=agent)
def create_node(self, participant, network):
"""Create a node for a participant."""
return dlgr.nodes.Agent(network=network, participant=participant)
|
Use compat for unicode import
|
## Code Before:
"""Coordination chatroom game."""
import dallinger as dlgr
from dallinger.config import get_config
try:
unicode = unicode
except NameError: # Python 3
unicode = str
config = get_config()
def extra_settings():
config.register('network', unicode)
config.register('n', int)
class CoordinationChatroom(dlgr.experiments.Experiment):
"""Define the structure of the experiment."""
def __init__(self, session):
"""Initialize the experiment."""
super(CoordinationChatroom, self).__init__(session)
self.experiment_repeats = 1
self.num_participants = config.get('n')
self.initial_recruitment_size = self.num_participants
self.quorum = self.num_participants
self.config = config
if not self.config.ready:
self.config.load_config()
self.setup()
def create_network(self):
"""Create a new network by reading the configuration file."""
class_ = getattr(
dlgr.networks,
self.config.get('network')
)
return class_(max_size=self.num_participants)
def info_post_request(self, node, info):
"""Run when a request to create an info is complete."""
for agent in node.neighbors():
node.transmit(what=info, to_whom=agent)
def create_node(self, participant, network):
"""Create a node for a participant."""
return dlgr.nodes.Agent(network=network, participant=participant)
## Instruction:
Use compat for unicode import
## Code After:
"""Coordination chatroom game."""
import dallinger as dlgr
from dallinger.compat import unicode
from dallinger.config import get_config
config = get_config()
def extra_settings():
config.register('network', unicode)
config.register('n', int)
class CoordinationChatroom(dlgr.experiments.Experiment):
"""Define the structure of the experiment."""
def __init__(self, session):
"""Initialize the experiment."""
super(CoordinationChatroom, self).__init__(session)
self.experiment_repeats = 1
self.num_participants = config.get('n')
self.initial_recruitment_size = self.num_participants
self.quorum = self.num_participants
self.config = config
if not self.config.ready:
self.config.load_config()
self.setup()
def create_network(self):
"""Create a new network by reading the configuration file."""
class_ = getattr(
dlgr.networks,
self.config.get('network')
)
return class_(max_size=self.num_participants)
def info_post_request(self, node, info):
"""Run when a request to create an info is complete."""
for agent in node.neighbors():
node.transmit(what=info, to_whom=agent)
def create_node(self, participant, network):
"""Create a node for a participant."""
return dlgr.nodes.Agent(network=network, participant=participant)
|
"""Coordination chatroom game."""
import dallinger as dlgr
+ from dallinger.compat import unicode
from dallinger.config import get_config
- try:
- unicode = unicode
- except NameError: # Python 3
- unicode = str
config = get_config()
def extra_settings():
config.register('network', unicode)
config.register('n', int)
class CoordinationChatroom(dlgr.experiments.Experiment):
"""Define the structure of the experiment."""
def __init__(self, session):
"""Initialize the experiment."""
super(CoordinationChatroom, self).__init__(session)
self.experiment_repeats = 1
self.num_participants = config.get('n')
self.initial_recruitment_size = self.num_participants
self.quorum = self.num_participants
self.config = config
if not self.config.ready:
self.config.load_config()
self.setup()
def create_network(self):
"""Create a new network by reading the configuration file."""
class_ = getattr(
dlgr.networks,
self.config.get('network')
)
return class_(max_size=self.num_participants)
def info_post_request(self, node, info):
"""Run when a request to create an info is complete."""
for agent in node.neighbors():
node.transmit(what=info, to_whom=agent)
def create_node(self, participant, network):
"""Create a node for a participant."""
return dlgr.nodes.Agent(network=network, participant=participant)
|
727ec507284776f3eec91b644cd5bb112bdb0af1
|
july/people/forms.py
|
july/people/forms.py
|
from django import forms
class EditUserForm(forms.Form):
about_me = forms.CharField(widget=forms.Textarea, required=False)
url = forms.CharField(max_length=255, required=False)
facebook_url = forms.CharField(max_length=255, required=False)
email = forms.EmailField(max_length=255)
def __init__(self, *args, **kwargs):
user = kwargs.pop('user', None)
super(EditUserForm, self).__init__(*args, **kwargs)
if user:
self.fields['about_me'].initial=user.about_me
self.fields['url'].initial=user.url
self.fields['facebook_url'].initial=user.facebook_url
self.fields['email'].initial=user.email
|
from django import forms
class EditUserForm(forms.Form):
about_me = forms.CharField(widget=forms.Textarea, required=False)
url = forms.CharField(max_length=255, required=False)
facebook_url = forms.CharField(max_length=255, required=False)
email = forms.EmailField(max_length=255)
def __init__(self, *args, **kwargs):
user = kwargs.pop('user', None)
super(EditUserForm, self).__init__(*args, **kwargs)
if user:
self.fields['about_me'].initial=getattr(user, 'about_me', None)
self.fields['url'].initial=getattr(user, 'url', None)
self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None)
self.fields['email'].initial=user.email
|
Use getattr for expando props
|
Use getattr for expando props
|
Python
|
mit
|
julython/julython.org,julython/julython.org,ChimeraCoder/GOctober,julython/julython.org,ChimeraCoder/GOctober,julython/julython.org,ChimeraCoder/GOctober
|
from django import forms
class EditUserForm(forms.Form):
about_me = forms.CharField(widget=forms.Textarea, required=False)
url = forms.CharField(max_length=255, required=False)
facebook_url = forms.CharField(max_length=255, required=False)
email = forms.EmailField(max_length=255)
def __init__(self, *args, **kwargs):
user = kwargs.pop('user', None)
super(EditUserForm, self).__init__(*args, **kwargs)
if user:
- self.fields['about_me'].initial=user.about_me
+ self.fields['about_me'].initial=getattr(user, 'about_me', None)
- self.fields['url'].initial=user.url
+ self.fields['url'].initial=getattr(user, 'url', None)
- self.fields['facebook_url'].initial=user.facebook_url
+ self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None)
self.fields['email'].initial=user.email
|
Use getattr for expando props
|
## Code Before:
from django import forms
class EditUserForm(forms.Form):
about_me = forms.CharField(widget=forms.Textarea, required=False)
url = forms.CharField(max_length=255, required=False)
facebook_url = forms.CharField(max_length=255, required=False)
email = forms.EmailField(max_length=255)
def __init__(self, *args, **kwargs):
user = kwargs.pop('user', None)
super(EditUserForm, self).__init__(*args, **kwargs)
if user:
self.fields['about_me'].initial=user.about_me
self.fields['url'].initial=user.url
self.fields['facebook_url'].initial=user.facebook_url
self.fields['email'].initial=user.email
## Instruction:
Use getattr for expando props
## Code After:
from django import forms
class EditUserForm(forms.Form):
about_me = forms.CharField(widget=forms.Textarea, required=False)
url = forms.CharField(max_length=255, required=False)
facebook_url = forms.CharField(max_length=255, required=False)
email = forms.EmailField(max_length=255)
def __init__(self, *args, **kwargs):
user = kwargs.pop('user', None)
super(EditUserForm, self).__init__(*args, **kwargs)
if user:
self.fields['about_me'].initial=getattr(user, 'about_me', None)
self.fields['url'].initial=getattr(user, 'url', None)
self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None)
self.fields['email'].initial=user.email
|
from django import forms
class EditUserForm(forms.Form):
about_me = forms.CharField(widget=forms.Textarea, required=False)
url = forms.CharField(max_length=255, required=False)
facebook_url = forms.CharField(max_length=255, required=False)
email = forms.EmailField(max_length=255)
def __init__(self, *args, **kwargs):
user = kwargs.pop('user', None)
super(EditUserForm, self).__init__(*args, **kwargs)
if user:
- self.fields['about_me'].initial=user.about_me
? ^
+ self.fields['about_me'].initial=getattr(user, 'about_me', None)
? ++++++++ ^^^ ++++++++
- self.fields['url'].initial=user.url
? ^
+ self.fields['url'].initial=getattr(user, 'url', None)
? ++++++++ ^^^ ++++++++
- self.fields['facebook_url'].initial=user.facebook_url
? ^
+ self.fields['facebook_url'].initial=getattr(user, 'facebook_url', None)
? ++++++++ ^^^ ++++++++
self.fields['email'].initial=user.email
|
41fe44e99361d9006a8b196e9b886ffdb3e8e460
|
functional_tests/test_evexml.py
|
functional_tests/test_evexml.py
|
from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from django.test import tag
from django.shortcuts import reverse
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
MAX_WAIT = 10
@tag('functional')
class SubmissionTest(StaticLiveServerTestCase):
"""Tests for users who are submitting xml api key.
"""
@classmethod
def setUpClass(cls):
super(SubmissionTest, cls).setUpClass()
cls.browser = webdriver.Chrome()
cls.browser.maximize_window()
cls.browser.implicitly_wait(MAX_WAIT)
super(SubmissionTest, cls).setUpClass()
@classmethod
def tearDownClass(cls):
cls.browser.refresh()
cls.browser.quit()
super(SubmissionTest, cls).tearDownClass()
def tearDown(self):
self.browser.refresh()
def test_user_can_see_apikey_form(self):
"""A user should be able to see the form for submitting api keys.
"""
# They browse to the eve api keys page.
url = self.live_server_url + reverse('eveapi_submit')
self.browser.get(self.live_server_url)
# They see input boxes for keyID and vCode.
keyid_input = self.browser.find_element_by_name('keyID')
vcode_input = self.browser.find_element_by_name('vCode')
|
from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from django.test import tag
from django.shortcuts import reverse
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
MAX_WAIT = 10
@tag('functional')
class SubmissionTest(StaticLiveServerTestCase):
"""Tests for users who are submitting xml api key.
"""
@classmethod
def setUpClass(cls):
super(SubmissionTest, cls).setUpClass()
cls.browser = webdriver.Chrome()
cls.browser.maximize_window()
cls.browser.implicitly_wait(MAX_WAIT)
super(SubmissionTest, cls).setUpClass()
@classmethod
def tearDownClass(cls):
cls.browser.refresh()
cls.browser.quit()
super(SubmissionTest, cls).tearDownClass()
def tearDown(self):
self.browser.refresh()
def test_user_can_see_apikey_form(self):
"""A user should be able to see the form for submitting api keys.
"""
# They browse to the eve api keys page.
url = self.live_server_url + reverse('eveapi_add')
self.browser.get(url)
# They see input boxes for keyID and vCode.
keyid_input = self.browser.find_element_by_name('keyID')
vcode_input = self.browser.find_element_by_name('vCode')
|
Make test get correct url
|
Make test get correct url
|
Python
|
mit
|
randomic/aniauth-tdd,randomic/aniauth-tdd
|
from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from django.test import tag
from django.shortcuts import reverse
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
MAX_WAIT = 10
@tag('functional')
class SubmissionTest(StaticLiveServerTestCase):
"""Tests for users who are submitting xml api key.
"""
@classmethod
def setUpClass(cls):
super(SubmissionTest, cls).setUpClass()
cls.browser = webdriver.Chrome()
cls.browser.maximize_window()
cls.browser.implicitly_wait(MAX_WAIT)
super(SubmissionTest, cls).setUpClass()
@classmethod
def tearDownClass(cls):
cls.browser.refresh()
cls.browser.quit()
super(SubmissionTest, cls).tearDownClass()
def tearDown(self):
self.browser.refresh()
def test_user_can_see_apikey_form(self):
"""A user should be able to see the form for submitting api keys.
"""
# They browse to the eve api keys page.
- url = self.live_server_url + reverse('eveapi_submit')
+ url = self.live_server_url + reverse('eveapi_add')
- self.browser.get(self.live_server_url)
+ self.browser.get(url)
# They see input boxes for keyID and vCode.
keyid_input = self.browser.find_element_by_name('keyID')
vcode_input = self.browser.find_element_by_name('vCode')
|
Make test get correct url
|
## Code Before:
from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from django.test import tag
from django.shortcuts import reverse
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
MAX_WAIT = 10
@tag('functional')
class SubmissionTest(StaticLiveServerTestCase):
"""Tests for users who are submitting xml api key.
"""
@classmethod
def setUpClass(cls):
super(SubmissionTest, cls).setUpClass()
cls.browser = webdriver.Chrome()
cls.browser.maximize_window()
cls.browser.implicitly_wait(MAX_WAIT)
super(SubmissionTest, cls).setUpClass()
@classmethod
def tearDownClass(cls):
cls.browser.refresh()
cls.browser.quit()
super(SubmissionTest, cls).tearDownClass()
def tearDown(self):
self.browser.refresh()
def test_user_can_see_apikey_form(self):
"""A user should be able to see the form for submitting api keys.
"""
# They browse to the eve api keys page.
url = self.live_server_url + reverse('eveapi_submit')
self.browser.get(self.live_server_url)
# They see input boxes for keyID and vCode.
keyid_input = self.browser.find_element_by_name('keyID')
vcode_input = self.browser.find_element_by_name('vCode')
## Instruction:
Make test get correct url
## Code After:
from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from django.test import tag
from django.shortcuts import reverse
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
MAX_WAIT = 10
@tag('functional')
class SubmissionTest(StaticLiveServerTestCase):
"""Tests for users who are submitting xml api key.
"""
@classmethod
def setUpClass(cls):
super(SubmissionTest, cls).setUpClass()
cls.browser = webdriver.Chrome()
cls.browser.maximize_window()
cls.browser.implicitly_wait(MAX_WAIT)
super(SubmissionTest, cls).setUpClass()
@classmethod
def tearDownClass(cls):
cls.browser.refresh()
cls.browser.quit()
super(SubmissionTest, cls).tearDownClass()
def tearDown(self):
self.browser.refresh()
def test_user_can_see_apikey_form(self):
"""A user should be able to see the form for submitting api keys.
"""
# They browse to the eve api keys page.
url = self.live_server_url + reverse('eveapi_add')
self.browser.get(url)
# They see input boxes for keyID and vCode.
keyid_input = self.browser.find_element_by_name('keyID')
vcode_input = self.browser.find_element_by_name('vCode')
|
from django.contrib.staticfiles.testing import StaticLiveServerTestCase
from django.test import tag
from django.shortcuts import reverse
from selenium import webdriver
from selenium.webdriver.common.keys import Keys
MAX_WAIT = 10
@tag('functional')
class SubmissionTest(StaticLiveServerTestCase):
"""Tests for users who are submitting xml api key.
"""
@classmethod
def setUpClass(cls):
super(SubmissionTest, cls).setUpClass()
cls.browser = webdriver.Chrome()
cls.browser.maximize_window()
cls.browser.implicitly_wait(MAX_WAIT)
super(SubmissionTest, cls).setUpClass()
@classmethod
def tearDownClass(cls):
cls.browser.refresh()
cls.browser.quit()
super(SubmissionTest, cls).tearDownClass()
def tearDown(self):
self.browser.refresh()
def test_user_can_see_apikey_form(self):
"""A user should be able to see the form for submitting api keys.
"""
# They browse to the eve api keys page.
- url = self.live_server_url + reverse('eveapi_submit')
? ^^^^^^
+ url = self.live_server_url + reverse('eveapi_add')
? ^^^
- self.browser.get(self.live_server_url)
? -----------------
+ self.browser.get(url)
# They see input boxes for keyID and vCode.
keyid_input = self.browser.find_element_by_name('keyID')
vcode_input = self.browser.find_element_by_name('vCode')
|
596613c964311104098e64eeb349216bc7cd0023
|
saleor/demo/views.py
|
saleor/demo/views.py
|
from django.conf import settings
from django.shortcuts import render
from ..graphql.views import API_PATH, GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
ctx = {
"query": EXAMPLE_QUERY,
"api_url": request.build_absolute_uri(str(API_PATH)),
}
return render(request, "graphql/playground.html", ctx)
|
from django.conf import settings
from django.shortcuts import render
from ..graphql.views import GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
pwa_origin = settings.PWA_ORIGINS[0]
ctx = {
"query": EXAMPLE_QUERY,
"api_url": f"https://{pwa_origin}/graphql/",
}
return render(request, "graphql/playground.html", ctx)
|
Fix playground CSP for demo if deployed under proxied domain
|
Fix playground CSP for demo if deployed under proxied domain
|
Python
|
bsd-3-clause
|
mociepka/saleor,mociepka/saleor,mociepka/saleor
|
from django.conf import settings
from django.shortcuts import render
- from ..graphql.views import API_PATH, GraphQLView
+ from ..graphql.views import GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
+ pwa_origin = settings.PWA_ORIGINS[0]
ctx = {
"query": EXAMPLE_QUERY,
- "api_url": request.build_absolute_uri(str(API_PATH)),
+ "api_url": f"https://{pwa_origin}/graphql/",
}
return render(request, "graphql/playground.html", ctx)
|
Fix playground CSP for demo if deployed under proxied domain
|
## Code Before:
from django.conf import settings
from django.shortcuts import render
from ..graphql.views import API_PATH, GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
ctx = {
"query": EXAMPLE_QUERY,
"api_url": request.build_absolute_uri(str(API_PATH)),
}
return render(request, "graphql/playground.html", ctx)
## Instruction:
Fix playground CSP for demo if deployed under proxied domain
## Code After:
from django.conf import settings
from django.shortcuts import render
from ..graphql.views import GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
pwa_origin = settings.PWA_ORIGINS[0]
ctx = {
"query": EXAMPLE_QUERY,
"api_url": f"https://{pwa_origin}/graphql/",
}
return render(request, "graphql/playground.html", ctx)
|
from django.conf import settings
from django.shortcuts import render
- from ..graphql.views import API_PATH, GraphQLView
? ----------
+ from ..graphql.views import GraphQLView
EXAMPLE_QUERY = """# Welcome to Saleor GraphQL API!
#
# Type queries into this side of the screen, and you will see
# intelligent typeaheads aware of the current GraphQL type schema
# and live syntax and validation errors highlighted within the text.
#
# Here is an example query to fetch a list of products:
#
{
products(first: 5, channel: "%(channel_slug)s") {
edges {
node {
id
name
description
}
}
}
}
""" % {
"channel_slug": settings.DEFAULT_CHANNEL_SLUG
}
class DemoGraphQLView(GraphQLView):
def render_playground(self, request):
+ pwa_origin = settings.PWA_ORIGINS[0]
ctx = {
"query": EXAMPLE_QUERY,
- "api_url": request.build_absolute_uri(str(API_PATH)),
+ "api_url": f"https://{pwa_origin}/graphql/",
}
return render(request, "graphql/playground.html", ctx)
|
d08e8144b90d3fe89fd449d31bdb655d62f3a749
|
serfclient/connection.py
|
serfclient/connection.py
|
import socket
import sys
class SerfConnectionError(Exception):
pass
class SerfConnection(object):
"""
Manages RPC communication to and from a Serf agent.
"""
def __init__(self, host='localhost', port=7373):
self.host, self.port = host, port
self._socket = None
def __repr__(self):
return "%(class)s<host=%(host)s,port=%(port)s>" % {
'class': self.__class__.__name__,
'host': self.host,
'port': self.port,
}
def handshake(self):
"""
Sets up the connection with the Serf agent and does the
initial handshake.
"""
if self._socket:
return
try:
self._socket = socket.create_connection((self.host, self.port))
return True
except socket.error:
e = sys.exc_info()[1]
raise SerfConnectionError(self._error_message(e))
def _error_message(self, exception):
return "Error %s connecting %s:%s. %s." % \
(exception.args[0], self.host, self.port, exception.args[1])
|
import socket
import sys
class SerfConnectionError(Exception):
pass
class SerfConnection(object):
"""
Manages RPC communication to and from a Serf agent.
"""
def __init__(self, host='localhost', port=7373):
self.host, self.port = host, port
self._socket = None
def __repr__(self):
return "%(class)s<host=%(host)s,port=%(port)s>" % {
'class': self.__class__.__name__,
'host': self.host,
'port': self.port,
}
def handshake(self):
"""
Sets up the connection with the Serf agent and does the
initial handshake.
"""
if self._socket:
return True
else:
self._socket = self._connect()
return True
def _connect(self):
try:
return socket.create_connection((self.host, self.port), 3)
except socket.error:
e = sys.exc_info()[1]
raise SerfConnectionError(self._error_message(e))
def _error_message(self, exception):
return "Error %s connecting %s:%s. %s." % \
(exception.args[0], self.host, self.port, exception.args[1])
|
Move all 'connect' logic into a private method
|
Move all 'connect' logic into a private method
|
Python
|
mit
|
charleswhchan/serfclient-py,KushalP/serfclient-py
|
import socket
import sys
class SerfConnectionError(Exception):
pass
class SerfConnection(object):
"""
Manages RPC communication to and from a Serf agent.
"""
def __init__(self, host='localhost', port=7373):
self.host, self.port = host, port
self._socket = None
def __repr__(self):
return "%(class)s<host=%(host)s,port=%(port)s>" % {
'class': self.__class__.__name__,
'host': self.host,
'port': self.port,
}
def handshake(self):
"""
Sets up the connection with the Serf agent and does the
initial handshake.
"""
if self._socket:
- return
+ return True
+ else:
+ self._socket = self._connect()
+ return True
+
+ def _connect(self):
try:
- self._socket = socket.create_connection((self.host, self.port))
+ return socket.create_connection((self.host, self.port), 3)
- return True
except socket.error:
e = sys.exc_info()[1]
raise SerfConnectionError(self._error_message(e))
def _error_message(self, exception):
return "Error %s connecting %s:%s. %s." % \
(exception.args[0], self.host, self.port, exception.args[1])
|
Move all 'connect' logic into a private method
|
## Code Before:
import socket
import sys
class SerfConnectionError(Exception):
pass
class SerfConnection(object):
"""
Manages RPC communication to and from a Serf agent.
"""
def __init__(self, host='localhost', port=7373):
self.host, self.port = host, port
self._socket = None
def __repr__(self):
return "%(class)s<host=%(host)s,port=%(port)s>" % {
'class': self.__class__.__name__,
'host': self.host,
'port': self.port,
}
def handshake(self):
"""
Sets up the connection with the Serf agent and does the
initial handshake.
"""
if self._socket:
return
try:
self._socket = socket.create_connection((self.host, self.port))
return True
except socket.error:
e = sys.exc_info()[1]
raise SerfConnectionError(self._error_message(e))
def _error_message(self, exception):
return "Error %s connecting %s:%s. %s." % \
(exception.args[0], self.host, self.port, exception.args[1])
## Instruction:
Move all 'connect' logic into a private method
## Code After:
import socket
import sys
class SerfConnectionError(Exception):
pass
class SerfConnection(object):
"""
Manages RPC communication to and from a Serf agent.
"""
def __init__(self, host='localhost', port=7373):
self.host, self.port = host, port
self._socket = None
def __repr__(self):
return "%(class)s<host=%(host)s,port=%(port)s>" % {
'class': self.__class__.__name__,
'host': self.host,
'port': self.port,
}
def handshake(self):
"""
Sets up the connection with the Serf agent and does the
initial handshake.
"""
if self._socket:
return True
else:
self._socket = self._connect()
return True
def _connect(self):
try:
return socket.create_connection((self.host, self.port), 3)
except socket.error:
e = sys.exc_info()[1]
raise SerfConnectionError(self._error_message(e))
def _error_message(self, exception):
return "Error %s connecting %s:%s. %s." % \
(exception.args[0], self.host, self.port, exception.args[1])
|
import socket
import sys
class SerfConnectionError(Exception):
pass
class SerfConnection(object):
"""
Manages RPC communication to and from a Serf agent.
"""
def __init__(self, host='localhost', port=7373):
self.host, self.port = host, port
self._socket = None
def __repr__(self):
return "%(class)s<host=%(host)s,port=%(port)s>" % {
'class': self.__class__.__name__,
'host': self.host,
'port': self.port,
}
def handshake(self):
"""
Sets up the connection with the Serf agent and does the
initial handshake.
"""
if self._socket:
- return
+ return True
? +++++
+ else:
+ self._socket = self._connect()
+ return True
+
+ def _connect(self):
try:
- self._socket = socket.create_connection((self.host, self.port))
? ^^^^^^^^^^ ^^
+ return socket.create_connection((self.host, self.port), 3)
? ^ ^^^ +++
- return True
except socket.error:
e = sys.exc_info()[1]
raise SerfConnectionError(self._error_message(e))
def _error_message(self, exception):
return "Error %s connecting %s:%s. %s." % \
(exception.args[0], self.host, self.port, exception.args[1])
|
85769162560d83a58ccc92f818559ddd3dce2a09
|
pages/index.py
|
pages/index.py
|
import web
from modules.base import renderer
from modules.login import loginInstance
from modules.courses import Course
#Index page
class IndexPage:
#Simply display the page
def GET(self):
if loginInstance.isLoggedIn():
userInput = web.input();
if "logoff" in userInput:
loginInstance.disconnect();
return renderer.index(False)
else:
courses = Course.GetAllCoursesIds()
return renderer.main(courses)
else:
return renderer.index(False)
#Try to log in
def POST(self):
userInput = web.input();
if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password):
return renderer.main()
else:
return renderer.index(True)
|
import web
from modules.base import renderer
from modules.login import loginInstance
from modules.courses import Course
#Index page
class IndexPage:
#Simply display the page
def GET(self):
if loginInstance.isLoggedIn():
userInput = web.input();
if "logoff" in userInput:
loginInstance.disconnect();
return renderer.index(False)
else:
return renderer.main(Course.GetAllCoursesIds())
else:
return renderer.index(False)
#Try to log in
def POST(self):
userInput = web.input();
if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password):
return renderer.main(Course.GetAllCoursesIds())
else:
return renderer.index(True)
|
Fix another bug in the authentication
|
Fix another bug in the authentication
|
Python
|
agpl-3.0
|
layus/INGInious,GuillaumeDerval/INGInious,GuillaumeDerval/INGInious,layus/INGInious,layus/INGInious,GuillaumeDerval/INGInious,GuillaumeDerval/INGInious,layus/INGInious
|
import web
from modules.base import renderer
from modules.login import loginInstance
from modules.courses import Course
#Index page
class IndexPage:
#Simply display the page
def GET(self):
if loginInstance.isLoggedIn():
userInput = web.input();
if "logoff" in userInput:
loginInstance.disconnect();
return renderer.index(False)
else:
- courses = Course.GetAllCoursesIds()
- return renderer.main(courses)
+ return renderer.main(Course.GetAllCoursesIds())
else:
return renderer.index(False)
#Try to log in
def POST(self):
userInput = web.input();
if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password):
- return renderer.main()
+ return renderer.main(Course.GetAllCoursesIds())
else:
return renderer.index(True)
|
Fix another bug in the authentication
|
## Code Before:
import web
from modules.base import renderer
from modules.login import loginInstance
from modules.courses import Course
#Index page
class IndexPage:
#Simply display the page
def GET(self):
if loginInstance.isLoggedIn():
userInput = web.input();
if "logoff" in userInput:
loginInstance.disconnect();
return renderer.index(False)
else:
courses = Course.GetAllCoursesIds()
return renderer.main(courses)
else:
return renderer.index(False)
#Try to log in
def POST(self):
userInput = web.input();
if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password):
return renderer.main()
else:
return renderer.index(True)
## Instruction:
Fix another bug in the authentication
## Code After:
import web
from modules.base import renderer
from modules.login import loginInstance
from modules.courses import Course
#Index page
class IndexPage:
#Simply display the page
def GET(self):
if loginInstance.isLoggedIn():
userInput = web.input();
if "logoff" in userInput:
loginInstance.disconnect();
return renderer.index(False)
else:
return renderer.main(Course.GetAllCoursesIds())
else:
return renderer.index(False)
#Try to log in
def POST(self):
userInput = web.input();
if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password):
return renderer.main(Course.GetAllCoursesIds())
else:
return renderer.index(True)
|
import web
from modules.base import renderer
from modules.login import loginInstance
from modules.courses import Course
#Index page
class IndexPage:
#Simply display the page
def GET(self):
if loginInstance.isLoggedIn():
userInput = web.input();
if "logoff" in userInput:
loginInstance.disconnect();
return renderer.index(False)
else:
- courses = Course.GetAllCoursesIds()
- return renderer.main(courses)
? ^
+ return renderer.main(Course.GetAllCoursesIds())
? ^^^^^^^^^^^^^^ ++++ +
else:
return renderer.index(False)
#Try to log in
def POST(self):
userInput = web.input();
if "login" in userInput and "password" in userInput and loginInstance.connect(userInput.login,userInput.password):
- return renderer.main()
+ return renderer.main(Course.GetAllCoursesIds())
else:
return renderer.index(True)
|
a4e9198194e08b99e11802dd260bd5c203179211
|
Mollie/API/Object/Customer.py
|
Mollie/API/Object/Customer.py
|
from .Base import Base
class Customer(Base):
@property
def id(self):
return self.getProperty('id')
@property
def name(self):
return self.getProperty('name')
@property
def email(self):
return self.getProperty('email')
@property
def locale(self):
return self.getProperty('locale')
@property
def metadata(self):
if 'metadata' not in self:
return None
return self.getProperty('metadata')
@property
def mode(self):
return self.getProperty('mode')
@property
def resource(self):
return self.getProperty('resource')
@property
def createdAt(self):
return self.getProperty('createdAt')
|
from .Base import Base
class Customer(Base):
@property
def id(self):
return self.getProperty('id')
@property
def name(self):
return self.getProperty('name')
@property
def email(self):
return self.getProperty('email')
@property
def locale(self):
return self.getProperty('locale')
@property
def metadata(self):
return self.getProperty('metadata')
@property
def mode(self):
return self.getProperty('mode')
@property
def resource(self):
return self.getProperty('resource')
@property
def createdAt(self):
return self.getProperty('createdAt')
|
Remove forgotten if in customer.py
|
Remove forgotten if in customer.py
|
Python
|
bsd-2-clause
|
mollie/mollie-api-python
|
from .Base import Base
class Customer(Base):
@property
def id(self):
return self.getProperty('id')
@property
def name(self):
return self.getProperty('name')
@property
def email(self):
return self.getProperty('email')
@property
def locale(self):
return self.getProperty('locale')
@property
def metadata(self):
- if 'metadata' not in self:
- return None
return self.getProperty('metadata')
@property
def mode(self):
return self.getProperty('mode')
@property
def resource(self):
return self.getProperty('resource')
@property
def createdAt(self):
return self.getProperty('createdAt')
|
Remove forgotten if in customer.py
|
## Code Before:
from .Base import Base
class Customer(Base):
@property
def id(self):
return self.getProperty('id')
@property
def name(self):
return self.getProperty('name')
@property
def email(self):
return self.getProperty('email')
@property
def locale(self):
return self.getProperty('locale')
@property
def metadata(self):
if 'metadata' not in self:
return None
return self.getProperty('metadata')
@property
def mode(self):
return self.getProperty('mode')
@property
def resource(self):
return self.getProperty('resource')
@property
def createdAt(self):
return self.getProperty('createdAt')
## Instruction:
Remove forgotten if in customer.py
## Code After:
from .Base import Base
class Customer(Base):
@property
def id(self):
return self.getProperty('id')
@property
def name(self):
return self.getProperty('name')
@property
def email(self):
return self.getProperty('email')
@property
def locale(self):
return self.getProperty('locale')
@property
def metadata(self):
return self.getProperty('metadata')
@property
def mode(self):
return self.getProperty('mode')
@property
def resource(self):
return self.getProperty('resource')
@property
def createdAt(self):
return self.getProperty('createdAt')
|
from .Base import Base
class Customer(Base):
@property
def id(self):
return self.getProperty('id')
@property
def name(self):
return self.getProperty('name')
@property
def email(self):
return self.getProperty('email')
@property
def locale(self):
return self.getProperty('locale')
@property
def metadata(self):
- if 'metadata' not in self:
- return None
return self.getProperty('metadata')
@property
def mode(self):
return self.getProperty('mode')
@property
def resource(self):
return self.getProperty('resource')
@property
def createdAt(self):
return self.getProperty('createdAt')
|
c1e1c9d63d5334140aa71c025a90e9500b299307
|
functional_tests.py
|
functional_tests.py
|
from selenium import webdriver
browser = webdriver.Firefox()
browser.get('http://localhost:8000')
assert 'Django' in browser.title
|
from selenium import webdriver
import unittest
class NewVisitorTest(unittest.TestCase):
def setUp(self):
self.browser = webdriver.Firefox()
def tearDown(self):
self.browser.quit()
def test_can_start_a_list_and_retrieve_it_later(self):
# Rey has heard about a cool new online to-do app. He goes
# to check out its homepage
self.browser.get('http://localhost:8000')
# He notices the page title and header mention to-do lists
self.assertIn('To-Do', self.browser.title)
self.fail('Finish the test!')
# He is invited to enter a to-do item straight away
# He types "Make coffee" into a text box (Rey can't work when he is sleepy)
# There is still a text box inviting him to add another item.
# He enters "Go to the toilet" (Rey is very methodical)
# The page updates again, and now shows both items on his list
# Rey wonders whether the site will remember his list. Then he sees that
# the site has generated a unique URL for him
# He visits that URL - his to-do list is still there
# Satisfied, he goes back to sleep
if __name__ == '__main__':
unittest.main(warnings='ignore')
|
Update tests to be human-readable
|
Update tests to be human-readable
|
Python
|
apache-2.0
|
rocity/the-testing-goat,rocity/the-testing-goat
|
from selenium import webdriver
+ import unittest
+ class NewVisitorTest(unittest.TestCase):
+ def setUp(self):
- browser = webdriver.Firefox()
+ self.browser = webdriver.Firefox()
- browser.get('http://localhost:8000')
- assert 'Django' in browser.title
+ def tearDown(self):
+ self.browser.quit()
+
+ def test_can_start_a_list_and_retrieve_it_later(self):
+
+ # Rey has heard about a cool new online to-do app. He goes
+ # to check out its homepage
+ self.browser.get('http://localhost:8000')
+
+ # He notices the page title and header mention to-do lists
+ self.assertIn('To-Do', self.browser.title)
+ self.fail('Finish the test!')
+
+ # He is invited to enter a to-do item straight away
+
+ # He types "Make coffee" into a text box (Rey can't work when he is sleepy)
+
+ # There is still a text box inviting him to add another item.
+ # He enters "Go to the toilet" (Rey is very methodical)
+
+ # The page updates again, and now shows both items on his list
+
+ # Rey wonders whether the site will remember his list. Then he sees that
+ # the site has generated a unique URL for him
+
+ # He visits that URL - his to-do list is still there
+
+ # Satisfied, he goes back to sleep
+
+ if __name__ == '__main__':
+ unittest.main(warnings='ignore')
|
Update tests to be human-readable
|
## Code Before:
from selenium import webdriver
browser = webdriver.Firefox()
browser.get('http://localhost:8000')
assert 'Django' in browser.title
## Instruction:
Update tests to be human-readable
## Code After:
from selenium import webdriver
import unittest
class NewVisitorTest(unittest.TestCase):
def setUp(self):
self.browser = webdriver.Firefox()
def tearDown(self):
self.browser.quit()
def test_can_start_a_list_and_retrieve_it_later(self):
# Rey has heard about a cool new online to-do app. He goes
# to check out its homepage
self.browser.get('http://localhost:8000')
# He notices the page title and header mention to-do lists
self.assertIn('To-Do', self.browser.title)
self.fail('Finish the test!')
# He is invited to enter a to-do item straight away
# He types "Make coffee" into a text box (Rey can't work when he is sleepy)
# There is still a text box inviting him to add another item.
# He enters "Go to the toilet" (Rey is very methodical)
# The page updates again, and now shows both items on his list
# Rey wonders whether the site will remember his list. Then he sees that
# the site has generated a unique URL for him
# He visits that URL - his to-do list is still there
# Satisfied, he goes back to sleep
if __name__ == '__main__':
unittest.main(warnings='ignore')
|
from selenium import webdriver
+ import unittest
+ class NewVisitorTest(unittest.TestCase):
+ def setUp(self):
- browser = webdriver.Firefox()
+ self.browser = webdriver.Firefox()
? +++++++++++++
- browser.get('http://localhost:8000')
- assert 'Django' in browser.title
+ def tearDown(self):
+ self.browser.quit()
+
+ def test_can_start_a_list_and_retrieve_it_later(self):
+
+ # Rey has heard about a cool new online to-do app. He goes
+ # to check out its homepage
+ self.browser.get('http://localhost:8000')
+
+ # He notices the page title and header mention to-do lists
+ self.assertIn('To-Do', self.browser.title)
+ self.fail('Finish the test!')
+
+ # He is invited to enter a to-do item straight away
+
+ # He types "Make coffee" into a text box (Rey can't work when he is sleepy)
+
+ # There is still a text box inviting him to add another item.
+ # He enters "Go to the toilet" (Rey is very methodical)
+
+ # The page updates again, and now shows both items on his list
+
+ # Rey wonders whether the site will remember his list. Then he sees that
+ # the site has generated a unique URL for him
+
+ # He visits that URL - his to-do list is still there
+
+ # Satisfied, he goes back to sleep
+
+ if __name__ == '__main__':
+ unittest.main(warnings='ignore')
|
14bd2c0732b5871ac43991a237a8f12a334e982d
|
sirius/LI_V00/__init__.py
|
sirius/LI_V00/__init__.py
|
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = accelerator.create_accelerator
# -- default accelerator values for LI_V00 --
energy = _lattice._energy
single_bunch_charge = _lattice._single_bunch_charge
multi_bunch_charge = _lattice._multi_bunch_charge
pulse_duration_interval = _lattice._pulse_duration_interval
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'LI_V00'
family_data = _lattice._family_data
emittance = _lattice._emittance
|
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = accelerator.create_accelerator
# -- default accelerator values for LI_V00 --
energy = _lattice._energy
single_bunch_charge = _lattice._single_bunch_charge
multi_bunch_charge = _lattice._multi_bunch_charge
pulse_duration_interval = _lattice._pulse_duration_interval
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'LI_V00'
family_data = _lattice._family_data
emittance = _lattice._emittance
global_coupling = 1.0 # "round" beam
|
Add parameters of initial beam distribution at LI
|
Add parameters of initial beam distribution at LI
|
Python
|
mit
|
lnls-fac/sirius
|
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = accelerator.create_accelerator
# -- default accelerator values for LI_V00 --
energy = _lattice._energy
single_bunch_charge = _lattice._single_bunch_charge
multi_bunch_charge = _lattice._multi_bunch_charge
pulse_duration_interval = _lattice._pulse_duration_interval
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'LI_V00'
family_data = _lattice._family_data
- emittance = _lattice._emittance
+ emittance = _lattice._emittance
+ global_coupling = 1.0 # "round" beam
|
Add parameters of initial beam distribution at LI
|
## Code Before:
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = accelerator.create_accelerator
# -- default accelerator values for LI_V00 --
energy = _lattice._energy
single_bunch_charge = _lattice._single_bunch_charge
multi_bunch_charge = _lattice._multi_bunch_charge
pulse_duration_interval = _lattice._pulse_duration_interval
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'LI_V00'
family_data = _lattice._family_data
emittance = _lattice._emittance
## Instruction:
Add parameters of initial beam distribution at LI
## Code After:
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = accelerator.create_accelerator
# -- default accelerator values for LI_V00 --
energy = _lattice._energy
single_bunch_charge = _lattice._single_bunch_charge
multi_bunch_charge = _lattice._multi_bunch_charge
pulse_duration_interval = _lattice._pulse_duration_interval
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'LI_V00'
family_data = _lattice._family_data
emittance = _lattice._emittance
global_coupling = 1.0 # "round" beam
|
from . import lattice as _lattice
from . import accelerator as _accelerator
from . import record_names
create_accelerator = accelerator.create_accelerator
# -- default accelerator values for LI_V00 --
energy = _lattice._energy
single_bunch_charge = _lattice._single_bunch_charge
multi_bunch_charge = _lattice._multi_bunch_charge
pulse_duration_interval = _lattice._pulse_duration_interval
default_optics_mode = _lattice._default_optics_mode.label
lattice_version = 'LI_V00'
family_data = _lattice._family_data
- emittance = _lattice._emittance
+ emittance = _lattice._emittance
? +++++++++++
+ global_coupling = 1.0 # "round" beam
|
5edddcc85b0e21bb576b71db63d082c8ace5cf70
|
examples/boilerplates/samples/google_test.py
|
examples/boilerplates/samples/google_test.py
|
'''
Google.com testing example
'''
from seleniumbase import BaseCase
from google_objects import HomePage, ResultsPage
class GoogleTests(BaseCase):
def test_google_dot_com(self):
self.open('http://www.google.com')
self.assert_element(HomePage.search_button)
self.assert_element(HomePage.feeling_lucky_button)
self.update_text(HomePage.search_box, "github\n")
self.assert_text("github.com", ResultsPage.search_results)
self.assert_element(ResultsPage.google_logo)
self.click_link_text("Images")
self.assert_element('img[alt="Image result for github"]')
|
'''
Google.com testing example
'''
from seleniumbase import BaseCase
from google_objects import HomePage, ResultsPage
class GoogleTests(BaseCase):
def test_google_dot_com(self):
self.open('http://www.google.com')
self.assert_element(HomePage.search_button)
self.assert_element(HomePage.feeling_lucky_button)
self.update_text(HomePage.search_box, "github\n")
self.assert_text("github.com", ResultsPage.search_results)
self.click_link_text("Images")
self.assert_element('img[alt="Image result for github"]')
|
Update Google boilerplate test. (Logo frequently changes)
|
Update Google boilerplate test. (Logo frequently changes)
|
Python
|
mit
|
seleniumbase/SeleniumBase,mdmintz/seleniumspot,mdmintz/SeleniumBase,mdmintz/seleniumspot,mdmintz/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase
|
'''
Google.com testing example
'''
from seleniumbase import BaseCase
from google_objects import HomePage, ResultsPage
class GoogleTests(BaseCase):
def test_google_dot_com(self):
self.open('http://www.google.com')
self.assert_element(HomePage.search_button)
self.assert_element(HomePage.feeling_lucky_button)
self.update_text(HomePage.search_box, "github\n")
self.assert_text("github.com", ResultsPage.search_results)
- self.assert_element(ResultsPage.google_logo)
self.click_link_text("Images")
self.assert_element('img[alt="Image result for github"]')
|
Update Google boilerplate test. (Logo frequently changes)
|
## Code Before:
'''
Google.com testing example
'''
from seleniumbase import BaseCase
from google_objects import HomePage, ResultsPage
class GoogleTests(BaseCase):
def test_google_dot_com(self):
self.open('http://www.google.com')
self.assert_element(HomePage.search_button)
self.assert_element(HomePage.feeling_lucky_button)
self.update_text(HomePage.search_box, "github\n")
self.assert_text("github.com", ResultsPage.search_results)
self.assert_element(ResultsPage.google_logo)
self.click_link_text("Images")
self.assert_element('img[alt="Image result for github"]')
## Instruction:
Update Google boilerplate test. (Logo frequently changes)
## Code After:
'''
Google.com testing example
'''
from seleniumbase import BaseCase
from google_objects import HomePage, ResultsPage
class GoogleTests(BaseCase):
def test_google_dot_com(self):
self.open('http://www.google.com')
self.assert_element(HomePage.search_button)
self.assert_element(HomePage.feeling_lucky_button)
self.update_text(HomePage.search_box, "github\n")
self.assert_text("github.com", ResultsPage.search_results)
self.click_link_text("Images")
self.assert_element('img[alt="Image result for github"]')
|
'''
Google.com testing example
'''
from seleniumbase import BaseCase
from google_objects import HomePage, ResultsPage
class GoogleTests(BaseCase):
def test_google_dot_com(self):
self.open('http://www.google.com')
self.assert_element(HomePage.search_button)
self.assert_element(HomePage.feeling_lucky_button)
self.update_text(HomePage.search_box, "github\n")
self.assert_text("github.com", ResultsPage.search_results)
- self.assert_element(ResultsPage.google_logo)
self.click_link_text("Images")
self.assert_element('img[alt="Image result for github"]')
|
3a7428723c66010dec1d246beb63be371428d3fe
|
qipipe/staging/staging_helpers.py
|
qipipe/staging/staging_helpers.py
|
"""Pipeline utility functions."""
import re
_SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})'
"""The subject/session/series regexp pattern."""
def match_session_hierarchy(path):
"""
Matches the subject, session and series names from the given input path.
@param path: the path to match
@return: the matching (subject, session, series) tuple, or None if no match
"""
match = re.search(_SSS_REGEX, path)
if match:
return match.groups()
|
"""Pipeline utility functions."""
import re
from .staging_error import StagingError
_SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})'
"""The subject/session/series regexp pattern."""
def match_series_hierarchy(path):
"""
Matches the subject, session and series names from the given input path.
@param path: the path to match
@return: the matching (subject, session, series) tuple, or None if no match
"""
match = re.search(_SSS_REGEX, path)
if match:
return match.groups()
else:
raise StagingError("The path %s does match the subject/session/series pattern" % path)
|
Raise error if no match.
|
Raise error if no match.
|
Python
|
bsd-2-clause
|
ohsu-qin/qipipe
|
"""Pipeline utility functions."""
import re
+ from .staging_error import StagingError
_SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})'
"""The subject/session/series regexp pattern."""
- def match_session_hierarchy(path):
+ def match_series_hierarchy(path):
"""
Matches the subject, session and series names from the given input path.
@param path: the path to match
@return: the matching (subject, session, series) tuple, or None if no match
"""
match = re.search(_SSS_REGEX, path)
if match:
return match.groups()
+ else:
+ raise StagingError("The path %s does match the subject/session/series pattern" % path)
|
Raise error if no match.
|
## Code Before:
"""Pipeline utility functions."""
import re
_SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})'
"""The subject/session/series regexp pattern."""
def match_session_hierarchy(path):
"""
Matches the subject, session and series names from the given input path.
@param path: the path to match
@return: the matching (subject, session, series) tuple, or None if no match
"""
match = re.search(_SSS_REGEX, path)
if match:
return match.groups()
## Instruction:
Raise error if no match.
## Code After:
"""Pipeline utility functions."""
import re
from .staging_error import StagingError
_SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})'
"""The subject/session/series regexp pattern."""
def match_series_hierarchy(path):
"""
Matches the subject, session and series names from the given input path.
@param path: the path to match
@return: the matching (subject, session, series) tuple, or None if no match
"""
match = re.search(_SSS_REGEX, path)
if match:
return match.groups()
else:
raise StagingError("The path %s does match the subject/session/series pattern" % path)
|
"""Pipeline utility functions."""
import re
+ from .staging_error import StagingError
_SSS_REGEX = '(\w+\d{2})/(session\d{2})/(series\d{3})'
"""The subject/session/series regexp pattern."""
- def match_session_hierarchy(path):
? ----
+ def match_series_hierarchy(path):
? +++
"""
Matches the subject, session and series names from the given input path.
@param path: the path to match
@return: the matching (subject, session, series) tuple, or None if no match
"""
match = re.search(_SSS_REGEX, path)
if match:
return match.groups()
+ else:
+ raise StagingError("The path %s does match the subject/session/series pattern" % path)
|
49f332149ae8a9a3b5faf82bc20b46dfaeb0a3ad
|
indra/sources/ctd/api.py
|
indra/sources/ctd/api.py
|
import pandas
from .processor import CTDChemicalDiseaseProcessor, \
CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor
base_url = 'http://ctdbase.org/reports/'
urls = {
'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz',
'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz',
'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz',
}
processors = {
'chemical_gene': CTDChemicalGeneProcessor,
'chemical_disease': CTDChemicalDiseaseProcessor,
'gene_disease': CTDGeneDiseaseProcessor,
}
def process_from_web(subset):
if subset not in urls:
raise ValueError('%s is not a valid CTD subset.')
df = pandas.read_csv(urls[subset], sep='\t', comment='#',
header=None)
return process_dataframe(df)
def process_tsv(fname, subset):
df = pandas.read_csv(fname, sep='\t', comment='#', header=None)
return process_dataframe(df, subset)
def process_dataframe(df, subset):
if subset not in processors:
raise ValueError('%s is not a valid CTD subset.')
cp = processors[subset](df)
cp.extract_statements()
return cp
|
import pandas
from .processor import CTDChemicalDiseaseProcessor, \
CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor
base_url = 'http://ctdbase.org/reports/'
urls = {
'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz',
'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz',
'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz',
}
processors = {
'chemical_gene': CTDChemicalGeneProcessor,
'chemical_disease': CTDChemicalDiseaseProcessor,
'gene_disease': CTDGeneDiseaseProcessor,
}
def process_from_web(subset, url=None):
if subset not in urls:
raise ValueError('%s is not a valid CTD subset.' % subset)
url = url if url else urls[subset]
return _process_url_or_file(url, subset)
def process_tsv(fname, subset):
return _process_url_or_file(fname, subset)
def _process_url_or_file(path, subset):
df = pandas.read_csv(path, sep='\t', comment='#',
header=None, dtype=str, keep_default_na=False)
return process_dataframe(df, subset)
def process_dataframe(df, subset):
if subset not in processors:
raise ValueError('%s is not a valid CTD subset.' % subset)
cp = processors[subset](df)
cp.extract_statements()
return cp
|
Refactor API to have single pandas load
|
Refactor API to have single pandas load
|
Python
|
bsd-2-clause
|
sorgerlab/indra,bgyori/indra,johnbachman/indra,bgyori/indra,sorgerlab/belpy,sorgerlab/belpy,johnbachman/belpy,bgyori/indra,johnbachman/indra,sorgerlab/belpy,johnbachman/belpy,johnbachman/indra,johnbachman/belpy,sorgerlab/indra,sorgerlab/indra
|
import pandas
from .processor import CTDChemicalDiseaseProcessor, \
CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor
base_url = 'http://ctdbase.org/reports/'
urls = {
'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz',
'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz',
'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz',
}
processors = {
'chemical_gene': CTDChemicalGeneProcessor,
'chemical_disease': CTDChemicalDiseaseProcessor,
'gene_disease': CTDGeneDiseaseProcessor,
}
- def process_from_web(subset):
+ def process_from_web(subset, url=None):
if subset not in urls:
- raise ValueError('%s is not a valid CTD subset.')
+ raise ValueError('%s is not a valid CTD subset.' % subset)
+ url = url if url else urls[subset]
+ return _process_url_or_file(url, subset)
- df = pandas.read_csv(urls[subset], sep='\t', comment='#',
- header=None)
- return process_dataframe(df)
def process_tsv(fname, subset):
+ return _process_url_or_file(fname, subset)
+
+
+ def _process_url_or_file(path, subset):
- df = pandas.read_csv(fname, sep='\t', comment='#', header=None)
+ df = pandas.read_csv(path, sep='\t', comment='#',
+ header=None, dtype=str, keep_default_na=False)
return process_dataframe(df, subset)
def process_dataframe(df, subset):
if subset not in processors:
- raise ValueError('%s is not a valid CTD subset.')
+ raise ValueError('%s is not a valid CTD subset.' % subset)
cp = processors[subset](df)
cp.extract_statements()
return cp
|
Refactor API to have single pandas load
|
## Code Before:
import pandas
from .processor import CTDChemicalDiseaseProcessor, \
CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor
base_url = 'http://ctdbase.org/reports/'
urls = {
'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz',
'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz',
'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz',
}
processors = {
'chemical_gene': CTDChemicalGeneProcessor,
'chemical_disease': CTDChemicalDiseaseProcessor,
'gene_disease': CTDGeneDiseaseProcessor,
}
def process_from_web(subset):
if subset not in urls:
raise ValueError('%s is not a valid CTD subset.')
df = pandas.read_csv(urls[subset], sep='\t', comment='#',
header=None)
return process_dataframe(df)
def process_tsv(fname, subset):
df = pandas.read_csv(fname, sep='\t', comment='#', header=None)
return process_dataframe(df, subset)
def process_dataframe(df, subset):
if subset not in processors:
raise ValueError('%s is not a valid CTD subset.')
cp = processors[subset](df)
cp.extract_statements()
return cp
## Instruction:
Refactor API to have single pandas load
## Code After:
import pandas
from .processor import CTDChemicalDiseaseProcessor, \
CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor
base_url = 'http://ctdbase.org/reports/'
urls = {
'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz',
'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz',
'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz',
}
processors = {
'chemical_gene': CTDChemicalGeneProcessor,
'chemical_disease': CTDChemicalDiseaseProcessor,
'gene_disease': CTDGeneDiseaseProcessor,
}
def process_from_web(subset, url=None):
if subset not in urls:
raise ValueError('%s is not a valid CTD subset.' % subset)
url = url if url else urls[subset]
return _process_url_or_file(url, subset)
def process_tsv(fname, subset):
return _process_url_or_file(fname, subset)
def _process_url_or_file(path, subset):
df = pandas.read_csv(path, sep='\t', comment='#',
header=None, dtype=str, keep_default_na=False)
return process_dataframe(df, subset)
def process_dataframe(df, subset):
if subset not in processors:
raise ValueError('%s is not a valid CTD subset.' % subset)
cp = processors[subset](df)
cp.extract_statements()
return cp
|
import pandas
from .processor import CTDChemicalDiseaseProcessor, \
CTDGeneDiseaseProcessor, CTDChemicalGeneProcessor
base_url = 'http://ctdbase.org/reports/'
urls = {
'chemical_gene': base_url + 'CTD_chem_gene_ixns.tsv.gz',
'chemical_disease': base_url + 'CTD_chemicals_diseases.tsv.gz',
'gene_disease': base_url + 'CTD_genes_diseases.tsv.gz',
}
processors = {
'chemical_gene': CTDChemicalGeneProcessor,
'chemical_disease': CTDChemicalDiseaseProcessor,
'gene_disease': CTDGeneDiseaseProcessor,
}
- def process_from_web(subset):
+ def process_from_web(subset, url=None):
? ++++++++++
if subset not in urls:
- raise ValueError('%s is not a valid CTD subset.')
+ raise ValueError('%s is not a valid CTD subset.' % subset)
? +++++++++
+ url = url if url else urls[subset]
+ return _process_url_or_file(url, subset)
- df = pandas.read_csv(urls[subset], sep='\t', comment='#',
- header=None)
- return process_dataframe(df)
def process_tsv(fname, subset):
+ return _process_url_or_file(fname, subset)
+
+
+ def _process_url_or_file(path, subset):
- df = pandas.read_csv(fname, sep='\t', comment='#', header=None)
? ^^ ^^ -------------
+ df = pandas.read_csv(path, sep='\t', comment='#',
? ^ ^^
+ header=None, dtype=str, keep_default_na=False)
return process_dataframe(df, subset)
def process_dataframe(df, subset):
if subset not in processors:
- raise ValueError('%s is not a valid CTD subset.')
+ raise ValueError('%s is not a valid CTD subset.' % subset)
? +++++++++
cp = processors[subset](df)
cp.extract_statements()
return cp
|
48132de52d573f7f650ab693c1ad0b6007ebfaef
|
cybox/test/common/vocab_test.py
|
cybox/test/common/vocab_test.py
|
import unittest
from cybox.common.vocabs import VocabString
import cybox.test
from cybox.utils import normalize_to_xml
class TestVocabString(unittest.TestCase):
def test_plain(self):
a = VocabString("test_value")
self.assertTrue(a.is_plain())
def test_round_trip(self):
attr_dict = {
'value': "test_value",
'vocab_name': "test_a",
'vocab_reference': "test_b",
'condition': "test_d",
'apply_condition': "test_0",
'bit_mask': "test_1",
'pattern_type': "test_e",
'regex_syntax': "test_f",
'has_changed': "test_j",
'trend': "test_k",
}
attr_obj = VocabString.object_from_dict(attr_dict)
attr_dict2 = VocabString.dict_from_object(attr_obj)
cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type'])
if __name__ == "__main__":
unittest.main()
|
import unittest
from cybox.common.vocabs import VocabString
import cybox.test
from cybox.utils import normalize_to_xml
class TestVocabString(unittest.TestCase):
def test_plain(self):
a = VocabString("test_value")
self.assertTrue(a.is_plain())
def test_round_trip(self):
vocab_dict = {
'value': "test_value",
'vocab_name': "test_a",
'vocab_reference': "test_b",
'condition': "test_d",
# Leave out apply_condition since value is not a list.
'bit_mask': "test_1",
'pattern_type': "test_e",
'regex_syntax': "test_f",
'has_changed': "test_j",
'trend': "test_k",
}
vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
def test_round_trip_list(self):
vocab_dict = {
'value': ['Value1', 'Value2', 'Value3'],
'condition': "Equals",
'apply_condition': "ALL",
}
vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
if __name__ == "__main__":
unittest.main()
|
Clean up controlled vocab tests
|
Clean up controlled vocab tests
|
Python
|
bsd-3-clause
|
CybOXProject/python-cybox
|
import unittest
from cybox.common.vocabs import VocabString
import cybox.test
from cybox.utils import normalize_to_xml
class TestVocabString(unittest.TestCase):
def test_plain(self):
a = VocabString("test_value")
self.assertTrue(a.is_plain())
def test_round_trip(self):
- attr_dict = {
+ vocab_dict = {
'value': "test_value",
'vocab_name': "test_a",
'vocab_reference': "test_b",
'condition': "test_d",
- 'apply_condition': "test_0",
+ # Leave out apply_condition since value is not a list.
'bit_mask': "test_1",
'pattern_type': "test_e",
'regex_syntax': "test_f",
'has_changed': "test_j",
'trend': "test_k",
- }
+ }
+ vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
- attr_obj = VocabString.object_from_dict(attr_dict)
- attr_dict2 = VocabString.dict_from_object(attr_obj)
- cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type'])
+ cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
+
+ def test_round_trip_list(self):
+ vocab_dict = {
+ 'value': ['Value1', 'Value2', 'Value3'],
+ 'condition': "Equals",
+ 'apply_condition': "ALL",
+ }
+
+ vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
+ cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
if __name__ == "__main__":
unittest.main()
|
Clean up controlled vocab tests
|
## Code Before:
import unittest
from cybox.common.vocabs import VocabString
import cybox.test
from cybox.utils import normalize_to_xml
class TestVocabString(unittest.TestCase):
def test_plain(self):
a = VocabString("test_value")
self.assertTrue(a.is_plain())
def test_round_trip(self):
attr_dict = {
'value': "test_value",
'vocab_name': "test_a",
'vocab_reference': "test_b",
'condition': "test_d",
'apply_condition': "test_0",
'bit_mask': "test_1",
'pattern_type': "test_e",
'regex_syntax': "test_f",
'has_changed': "test_j",
'trend': "test_k",
}
attr_obj = VocabString.object_from_dict(attr_dict)
attr_dict2 = VocabString.dict_from_object(attr_obj)
cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type'])
if __name__ == "__main__":
unittest.main()
## Instruction:
Clean up controlled vocab tests
## Code After:
import unittest
from cybox.common.vocabs import VocabString
import cybox.test
from cybox.utils import normalize_to_xml
class TestVocabString(unittest.TestCase):
def test_plain(self):
a = VocabString("test_value")
self.assertTrue(a.is_plain())
def test_round_trip(self):
vocab_dict = {
'value': "test_value",
'vocab_name': "test_a",
'vocab_reference': "test_b",
'condition': "test_d",
# Leave out apply_condition since value is not a list.
'bit_mask': "test_1",
'pattern_type': "test_e",
'regex_syntax': "test_f",
'has_changed': "test_j",
'trend': "test_k",
}
vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
def test_round_trip_list(self):
vocab_dict = {
'value': ['Value1', 'Value2', 'Value3'],
'condition': "Equals",
'apply_condition': "ALL",
}
vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
if __name__ == "__main__":
unittest.main()
|
import unittest
from cybox.common.vocabs import VocabString
import cybox.test
from cybox.utils import normalize_to_xml
class TestVocabString(unittest.TestCase):
def test_plain(self):
a = VocabString("test_value")
self.assertTrue(a.is_plain())
def test_round_trip(self):
- attr_dict = {
? ^^^
+ vocab_dict = {
? +++ ^
'value': "test_value",
'vocab_name': "test_a",
'vocab_reference': "test_b",
'condition': "test_d",
- 'apply_condition': "test_0",
+ # Leave out apply_condition since value is not a list.
'bit_mask': "test_1",
'pattern_type': "test_e",
'regex_syntax': "test_f",
'has_changed': "test_j",
'trend': "test_k",
- }
+ }
? +
+ vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
- attr_obj = VocabString.object_from_dict(attr_dict)
- attr_dict2 = VocabString.dict_from_object(attr_obj)
- cybox.test.assert_equal_ignore(attr_dict, attr_dict2, ['xsi:type'])
? ^^^ ^^^
+ cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
? +++ ^ +++ ^
+
+ def test_round_trip_list(self):
+ vocab_dict = {
+ 'value': ['Value1', 'Value2', 'Value3'],
+ 'condition': "Equals",
+ 'apply_condition': "ALL",
+ }
+
+ vocab_dict2 = cybox.test.round_trip_dict(VocabString, vocab_dict)
+ cybox.test.assert_equal_ignore(vocab_dict, vocab_dict2, ['xsi:type'])
if __name__ == "__main__":
unittest.main()
|
5c41066e9c93c417253cbde325a18079c1c69d1a
|
scipy/sparse/linalg/isolve/__init__.py
|
scipy/sparse/linalg/isolve/__init__.py
|
"Iterative Solvers for Sparse Linear Systems"
#from info import __doc__
from iterative import *
from minres import minres
from lgmres import lgmres
from lsqr import lsqr
__all__ = filter(lambda s:not s.startswith('_'),dir())
from numpy.testing import Tester
test = Tester().test
bench = Tester().bench
|
"Iterative Solvers for Sparse Linear Systems"
#from info import __doc__
from iterative import *
from minres import minres
from lgmres import lgmres
from lsqr import lsqr
from lsmr import lsmr
__all__ = filter(lambda s:not s.startswith('_'),dir())
from numpy.testing import Tester
test = Tester().test
bench = Tester().bench
|
Add lsmr to isolve module.
|
ENH: Add lsmr to isolve module.
|
Python
|
bsd-3-clause
|
bkendzior/scipy,mikebenfield/scipy,anielsen001/scipy,nonhermitian/scipy,aeklant/scipy,apbard/scipy,jjhelmus/scipy,giorgiop/scipy,sonnyhu/scipy,befelix/scipy,jor-/scipy,gfyoung/scipy,piyush0609/scipy,pbrod/scipy,Shaswat27/scipy,pschella/scipy,chatcannon/scipy,Newman101/scipy,ilayn/scipy,Newman101/scipy,mtrbean/scipy,Shaswat27/scipy,newemailjdm/scipy,behzadnouri/scipy,woodscn/scipy,sonnyhu/scipy,jonycgn/scipy,behzadnouri/scipy,lukauskas/scipy,rgommers/scipy,efiring/scipy,newemailjdm/scipy,Eric89GXL/scipy,WillieMaddox/scipy,andim/scipy,ogrisel/scipy,ogrisel/scipy,FRidh/scipy,anielsen001/scipy,pnedunuri/scipy,mortonjt/scipy,gdooper/scipy,haudren/scipy,gertingold/scipy,lhilt/scipy,Srisai85/scipy,zerothi/scipy,pizzathief/scipy,futurulus/scipy,rgommers/scipy,mortonjt/scipy,piyush0609/scipy,piyush0609/scipy,piyush0609/scipy,Eric89GXL/scipy,matthewalbani/scipy,ilayn/scipy,fredrikw/scipy,pyramania/scipy,Kamp9/scipy,Gillu13/scipy,jonycgn/scipy,chatcannon/scipy,nvoron23/scipy,fernand/scipy,witcxc/scipy,Newman101/scipy,efiring/scipy,Newman101/scipy,argriffing/scipy,sargas/scipy,gertingold/scipy,raoulbq/scipy,minhlongdo/scipy,ChanderG/scipy,jjhelmus/scipy,maniteja123/scipy,felipebetancur/scipy,Gillu13/scipy,anntzer/scipy,maciejkula/scipy,jakevdp/scipy,Kamp9/scipy,trankmichael/scipy,woodscn/scipy,jjhelmus/scipy,zerothi/scipy,matthewalbani/scipy,josephcslater/scipy,apbard/scipy,mtrbean/scipy,richardotis/scipy,mortonjt/scipy,mortada/scipy,teoliphant/scipy,jonycgn/scipy,ilayn/scipy,ndchorley/scipy,Srisai85/scipy,zerothi/scipy,sonnyhu/scipy,anntzer/scipy,vigna/scipy,gef756/scipy,cpaulik/scipy,WarrenWeckesser/scipy,ndchorley/scipy,pnedunuri/scipy,jamestwebber/scipy,anielsen001/scipy,FRidh/scipy,vanpact/scipy,Newman101/scipy,pbrod/scipy,raoulbq/scipy,vigna/scipy,jakevdp/scipy,jsilter/scipy,sauliusl/scipy,e-q/scipy,aarchiba/scipy,piyush0609/scipy,cpaulik/scipy,gdooper/scipy,andim/scipy,teoliphant/scipy,hainm/scipy,behzadnouri/scipy,Stefan-Endres/scipy,petebachant/scipy,Shaswat27/scipy,gef756/scipy,jakevdp/scipy,kalvdans/scipy,ales-erjavec/scipy,kleskjr/scipy,mdhaber/scipy,mgaitan/scipy,perimosocordiae/scipy,pizzathief/scipy,aeklant/scipy,niknow/scipy,dch312/scipy,WarrenWeckesser/scipy,jjhelmus/scipy,mhogg/scipy,sargas/scipy,dominicelse/scipy,sriki18/scipy,jor-/scipy,petebachant/scipy,woodscn/scipy,vanpact/scipy,nonhermitian/scipy,futurulus/scipy,dominicelse/scipy,scipy/scipy,mdhaber/scipy,kalvdans/scipy,jsilter/scipy,ortylp/scipy,perimosocordiae/scipy,gfyoung/scipy,fernand/scipy,anntzer/scipy,aman-iitj/scipy,sauliusl/scipy,mikebenfield/scipy,matthew-brett/scipy,mgaitan/scipy,lukauskas/scipy,rmcgibbo/scipy,haudren/scipy,surhudm/scipy,sriki18/scipy,jonycgn/scipy,person142/scipy,befelix/scipy,jor-/scipy,efiring/scipy,vberaudi/scipy,pyramania/scipy,josephcslater/scipy,grlee77/scipy,zerothi/scipy,jsilter/scipy,larsmans/scipy,efiring/scipy,ales-erjavec/scipy,nvoron23/scipy,felipebetancur/scipy,jseabold/scipy,dch312/scipy,Gillu13/scipy,mortada/scipy,gef756/scipy,trankmichael/scipy,ortylp/scipy,Gillu13/scipy,ogrisel/scipy,nmayorov/scipy,e-q/scipy,witcxc/scipy,jseabold/scipy,vhaasteren/scipy,maniteja123/scipy,vigna/scipy,gfyoung/scipy,lhilt/scipy,giorgiop/scipy,grlee77/scipy,haudren/scipy,Eric89GXL/scipy,jamestwebber/scipy,raoulbq/scipy,zxsted/scipy,andim/scipy,mortada/scipy,befelix/scipy,richardotis/scipy,tylerjereddy/scipy,larsmans/scipy,jonycgn/scipy,niknow/scipy,jakevdp/scipy,nonhermitian/scipy,aman-iitj/scipy,teoliphant/scipy,lhilt/scipy,mortonjt/scipy,vanpact/scipy,felipebetancur/scipy,person142/scipy,woodscn/scipy,sriki18/scipy,Srisai85/scipy,FRidh/scipy,mhogg/scipy,behzadnouri/scipy,njwilson23/scipy,mingwpy/scipy,ilayn/scipy,ChanderG/scipy,Kamp9/scipy,minhlongdo/scipy,tylerjereddy/scipy,kleskjr/scipy,jseabold/scipy,aarchiba/scipy,josephcslater/scipy,andyfaff/scipy,dch312/scipy,mingwpy/scipy,raoulbq/scipy,Dapid/scipy,endolith/scipy,scipy/scipy,Srisai85/scipy,zxsted/scipy,fernand/scipy,pbrod/scipy,Stefan-Endres/scipy,rmcgibbo/scipy,zaxliu/scipy,rgommers/scipy,pschella/scipy,Eric89GXL/scipy,bkendzior/scipy,endolith/scipy,hainm/scipy,jonycgn/scipy,witcxc/scipy,zxsted/scipy,vberaudi/scipy,perimosocordiae/scipy,juliantaylor/scipy,grlee77/scipy,sriki18/scipy,Shaswat27/scipy,sonnyhu/scipy,bkendzior/scipy,nmayorov/scipy,Dapid/scipy,ales-erjavec/scipy,WillieMaddox/scipy,arokem/scipy,kleskjr/scipy,newemailjdm/scipy,mgaitan/scipy,aman-iitj/scipy,lukauskas/scipy,maciejkula/scipy,Dapid/scipy,niknow/scipy,ortylp/scipy,giorgiop/scipy,newemailjdm/scipy,WarrenWeckesser/scipy,lukauskas/scipy,mortada/scipy,jamestwebber/scipy,WillieMaddox/scipy,juliantaylor/scipy,juliantaylor/scipy,newemailjdm/scipy,nvoron23/scipy,FRidh/scipy,ilayn/scipy,Dapid/scipy,e-q/scipy,niknow/scipy,WarrenWeckesser/scipy,ndchorley/scipy,mgaitan/scipy,pizzathief/scipy,zaxliu/scipy,richardotis/scipy,pizzathief/scipy,gfyoung/scipy,matthew-brett/scipy,pbrod/scipy,nmayorov/scipy,mikebenfield/scipy,sauliusl/scipy,aarchiba/scipy,gertingold/scipy,zxsted/scipy,sriki18/scipy,zerothi/scipy,jamestwebber/scipy,surhudm/scipy,giorgiop/scipy,matthew-brett/scipy,scipy/scipy,lukauskas/scipy,pschella/scipy,felipebetancur/scipy,perimosocordiae/scipy,maniteja123/scipy,haudren/scipy,futurulus/scipy,ogrisel/scipy,surhudm/scipy,woodscn/scipy,jor-/scipy,niknow/scipy,perimosocordiae/scipy,mhogg/scipy,rmcgibbo/scipy,ales-erjavec/scipy,grlee77/scipy,mdhaber/scipy,fredrikw/scipy,pschella/scipy,Gillu13/scipy,vhaasteren/scipy,mtrbean/scipy,aeklant/scipy,dominicelse/scipy,giorgiop/scipy,vhaasteren/scipy,ales-erjavec/scipy,nvoron23/scipy,kalvdans/scipy,jakevdp/scipy,dominicelse/scipy,vberaudi/scipy,pizzathief/scipy,ChanderG/scipy,matthewalbani/scipy,fernand/scipy,gef756/scipy,ChanderG/scipy,FRidh/scipy,maniteja123/scipy,grlee77/scipy,nonhermitian/scipy,hainm/scipy,felipebetancur/scipy,zaxliu/scipy,vanpact/scipy,Newman101/scipy,cpaulik/scipy,ndchorley/scipy,nmayorov/scipy,gdooper/scipy,trankmichael/scipy,FRidh/scipy,richardotis/scipy,Gillu13/scipy,anntzer/scipy,lhilt/scipy,vigna/scipy,dch312/scipy,newemailjdm/scipy,maciejkula/scipy,efiring/scipy,tylerjereddy/scipy,gdooper/scipy,pyramania/scipy,mtrbean/scipy,gertingold/scipy,nonhermitian/scipy,apbard/scipy,bkendzior/scipy,giorgiop/scipy,Kamp9/scipy,piyush0609/scipy,WarrenWeckesser/scipy,fernand/scipy,ndchorley/scipy,Shaswat27/scipy,cpaulik/scipy,jor-/scipy,matthewalbani/scipy,pbrod/scipy,dch312/scipy,Kamp9/scipy,fredrikw/scipy,Shaswat27/scipy,cpaulik/scipy,larsmans/scipy,ortylp/scipy,mortonjt/scipy,njwilson23/scipy,ChanderG/scipy,mingwpy/scipy,hainm/scipy,scipy/scipy,vhaasteren/scipy,Dapid/scipy,haudren/scipy,Stefan-Endres/scipy,zxsted/scipy,sriki18/scipy,sauliusl/scipy,endolith/scipy,mingwpy/scipy,aeklant/scipy,argriffing/scipy,pnedunuri/scipy,futurulus/scipy,matthew-brett/scipy,fredrikw/scipy,juliantaylor/scipy,futurulus/scipy,teoliphant/scipy,scipy/scipy,petebachant/scipy,Eric89GXL/scipy,mortada/scipy,argriffing/scipy,behzadnouri/scipy,kalvdans/scipy,nmayorov/scipy,mtrbean/scipy,Stefan-Endres/scipy,befelix/scipy,hainm/scipy,surhudm/scipy,sargas/scipy,anielsen001/scipy,rmcgibbo/scipy,petebachant/scipy,mikebenfield/scipy,raoulbq/scipy,vhaasteren/scipy,person142/scipy,efiring/scipy,richardotis/scipy,jsilter/scipy,zaxliu/scipy,kleskjr/scipy,gertingold/scipy,petebachant/scipy,aarchiba/scipy,ales-erjavec/scipy,Srisai85/scipy,maciejkula/scipy,juliantaylor/scipy,arokem/scipy,zerothi/scipy,pyramania/scipy,endolith/scipy,e-q/scipy,rgommers/scipy,chatcannon/scipy,aman-iitj/scipy,nvoron23/scipy,pyramania/scipy,trankmichael/scipy,mortonjt/scipy,Stefan-Endres/scipy,chatcannon/scipy,trankmichael/scipy,apbard/scipy,ortylp/scipy,anntzer/scipy,jamestwebber/scipy,njwilson23/scipy,pnedunuri/scipy,argriffing/scipy,mingwpy/scipy,andim/scipy,Srisai85/scipy,zaxliu/scipy,person142/scipy,mhogg/scipy,aman-iitj/scipy,WillieMaddox/scipy,maniteja123/scipy,endolith/scipy,bkendzior/scipy,larsmans/scipy,kleskjr/scipy,sauliusl/scipy,Kamp9/scipy,cpaulik/scipy,richardotis/scipy,jseabold/scipy,dominicelse/scipy,surhudm/scipy,tylerjereddy/scipy,mgaitan/scipy,befelix/scipy,e-q/scipy,Dapid/scipy,ChanderG/scipy,perimosocordiae/scipy,witcxc/scipy,argriffing/scipy,Eric89GXL/scipy,rgommers/scipy,apbard/scipy,andyfaff/scipy,hainm/scipy,vhaasteren/scipy,woodscn/scipy,fernand/scipy,mgaitan/scipy,argriffing/scipy,tylerjereddy/scipy,andyfaff/scipy,arokem/scipy,behzadnouri/scipy,minhlongdo/scipy,felipebetancur/scipy,fredrikw/scipy,rmcgibbo/scipy,ndchorley/scipy,andyfaff/scipy,Stefan-Endres/scipy,vanpact/scipy,gef756/scipy,chatcannon/scipy,anielsen001/scipy,sauliusl/scipy,gfyoung/scipy,raoulbq/scipy,aman-iitj/scipy,mortada/scipy,zxsted/scipy,zaxliu/scipy,kalvdans/scipy,pnedunuri/scipy,minhlongdo/scipy,arokem/scipy,aeklant/scipy,sonnyhu/scipy,fredrikw/scipy,mingwpy/scipy,witcxc/scipy,njwilson23/scipy,minhlongdo/scipy,matthew-brett/scipy,larsmans/scipy,arokem/scipy,vberaudi/scipy,pnedunuri/scipy,andyfaff/scipy,mdhaber/scipy,gdooper/scipy,vigna/scipy,anntzer/scipy,andim/scipy,vanpact/scipy,pschella/scipy,matthewalbani/scipy,sargas/scipy,mdhaber/scipy,sonnyhu/scipy,njwilson23/scipy,endolith/scipy,maciejkula/scipy,jseabold/scipy,ortylp/scipy,sargas/scipy,teoliphant/scipy,njwilson23/scipy,josephcslater/scipy,gef756/scipy,andim/scipy,scipy/scipy,andyfaff/scipy,mikebenfield/scipy,WarrenWeckesser/scipy,pbrod/scipy,jsilter/scipy,niknow/scipy,jjhelmus/scipy,vberaudi/scipy,trankmichael/scipy,haudren/scipy,mhogg/scipy,lukauskas/scipy,nvoron23/scipy,mdhaber/scipy,mtrbean/scipy,larsmans/scipy,ogrisel/scipy,ilayn/scipy,mhogg/scipy,person142/scipy,rmcgibbo/scipy,lhilt/scipy,josephcslater/scipy,maniteja123/scipy,petebachant/scipy,anielsen001/scipy,vberaudi/scipy,jseabold/scipy,WillieMaddox/scipy,aarchiba/scipy,WillieMaddox/scipy,surhudm/scipy,chatcannon/scipy,minhlongdo/scipy,kleskjr/scipy,futurulus/scipy
|
"Iterative Solvers for Sparse Linear Systems"
#from info import __doc__
from iterative import *
from minres import minres
from lgmres import lgmres
from lsqr import lsqr
+ from lsmr import lsmr
__all__ = filter(lambda s:not s.startswith('_'),dir())
from numpy.testing import Tester
test = Tester().test
bench = Tester().bench
|
Add lsmr to isolve module.
|
## Code Before:
"Iterative Solvers for Sparse Linear Systems"
#from info import __doc__
from iterative import *
from minres import minres
from lgmres import lgmres
from lsqr import lsqr
__all__ = filter(lambda s:not s.startswith('_'),dir())
from numpy.testing import Tester
test = Tester().test
bench = Tester().bench
## Instruction:
Add lsmr to isolve module.
## Code After:
"Iterative Solvers for Sparse Linear Systems"
#from info import __doc__
from iterative import *
from minres import minres
from lgmres import lgmres
from lsqr import lsqr
from lsmr import lsmr
__all__ = filter(lambda s:not s.startswith('_'),dir())
from numpy.testing import Tester
test = Tester().test
bench = Tester().bench
|
"Iterative Solvers for Sparse Linear Systems"
#from info import __doc__
from iterative import *
from minres import minres
from lgmres import lgmres
from lsqr import lsqr
+ from lsmr import lsmr
__all__ = filter(lambda s:not s.startswith('_'),dir())
from numpy.testing import Tester
test = Tester().test
bench = Tester().bench
|
0f2950fcb44efc9b629242743574af503e8230d4
|
tip/algorithms/sorting/mergesort.py
|
tip/algorithms/sorting/mergesort.py
|
def merge(a, b):
if len(a) * len(b) == 0:
return a + b
v = (a[0] < b[0] and a or b).pop(0)
return [v] + merge(a, b)
def mergesort(list):
if len(list) < 2:
return list
m = len(list) / 2
return merge(mergesort(list[:m]), mergesort(list[m:]))
|
def merge(a, b):
if len(a) * len(b) == 0:
return a + b
v = (a[0] < b[0] and a or b).pop(0)
return [v] + merge(a, b)
def mergesort(list):
if len(list) < 2:
return list
m = len(list) / 2
return merge(mergesort(list[:int(m)]), mergesort(list[int(m):]))
|
Fix slices for Python 3
|
Fix slices for Python 3
|
Python
|
unlicense
|
davidgasquez/tip
|
def merge(a, b):
if len(a) * len(b) == 0:
return a + b
v = (a[0] < b[0] and a or b).pop(0)
return [v] + merge(a, b)
def mergesort(list):
if len(list) < 2:
return list
m = len(list) / 2
- return merge(mergesort(list[:m]), mergesort(list[m:]))
+ return merge(mergesort(list[:int(m)]), mergesort(list[int(m):]))
|
Fix slices for Python 3
|
## Code Before:
def merge(a, b):
if len(a) * len(b) == 0:
return a + b
v = (a[0] < b[0] and a or b).pop(0)
return [v] + merge(a, b)
def mergesort(list):
if len(list) < 2:
return list
m = len(list) / 2
return merge(mergesort(list[:m]), mergesort(list[m:]))
## Instruction:
Fix slices for Python 3
## Code After:
def merge(a, b):
if len(a) * len(b) == 0:
return a + b
v = (a[0] < b[0] and a or b).pop(0)
return [v] + merge(a, b)
def mergesort(list):
if len(list) < 2:
return list
m = len(list) / 2
return merge(mergesort(list[:int(m)]), mergesort(list[int(m):]))
|
def merge(a, b):
if len(a) * len(b) == 0:
return a + b
v = (a[0] < b[0] and a or b).pop(0)
return [v] + merge(a, b)
def mergesort(list):
if len(list) < 2:
return list
m = len(list) / 2
- return merge(mergesort(list[:m]), mergesort(list[m:]))
+ return merge(mergesort(list[:int(m)]), mergesort(list[int(m):]))
? ++++ + ++++ +
|
945aba9548b92f57fc25f9996bfa9c3811e64deb
|
server/resources.py
|
server/resources.py
|
from flask import request
from flask_restful import Resource, Api, abort, reqparse
from .models import db, Comment, Lecture
api = Api()
class CommentListResource(Resource):
def get(self, lecture_id):
db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not db_lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
db_comments = Comment.query.filter(Comment.lecture_id == lecture_id)
comments = [
{'id': c.id, 'content': c.content}
for c in db_comments
]
return {
'comments': comments
}
def post(self, lecture_id):
db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all()
if not db_lectures:
abort(404, message="Lecture {} does not exist".format(lecture_id))
parser = reqparse.RequestParser()
parser.add_argument('data', help='Text content of comment')
args = parser.parse_args()
if not args.data:
abort(400, message="Comment has no data parameter")
content = args.data
lecture = db_lectures[0]
comment = Comment(content, lecture)
db.session.add(comment)
db.session.commit()
return {
'id': comment.id
}
api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
|
from flask import request
from flask_restful import Resource, Api, abort, reqparse
from .models import db, Comment, Lecture
api = Api()
class CommentListResource(Resource):
def get(self, lecture_id):
db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not db_lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
db_comments = Comment.query.filter(Comment.lecture_id == lecture_id)
comments = [
{'id': c.id, 'content': c.content}
for c in db_comments
]
return {
'comments': comments
}
def post(self, lecture_id):
lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
parser = reqparse.RequestParser()
parser.add_argument('data', help='Text content of comment')
args = parser.parse_args()
if not args.data:
abort(400, message="Comment has no data parameter")
content = args.data
comment = Comment(content, lecture)
db.session.add(comment)
db.session.commit()
return {
'id': comment.id
}
api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
|
Change single Lecture query to use first() in stead of all()
|
Change single Lecture query to use first() in stead of all()
|
Python
|
mit
|
MACSIFS/IFS,MACSIFS/IFS,MACSIFS/IFS,MACSIFS/IFS
|
from flask import request
from flask_restful import Resource, Api, abort, reqparse
from .models import db, Comment, Lecture
api = Api()
class CommentListResource(Resource):
def get(self, lecture_id):
db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not db_lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
db_comments = Comment.query.filter(Comment.lecture_id == lecture_id)
comments = [
{'id': c.id, 'content': c.content}
for c in db_comments
]
return {
'comments': comments
}
def post(self, lecture_id):
- db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all()
+ lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
- if not db_lectures:
+ if not lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
parser = reqparse.RequestParser()
parser.add_argument('data', help='Text content of comment')
args = parser.parse_args()
if not args.data:
abort(400, message="Comment has no data parameter")
content = args.data
- lecture = db_lectures[0]
comment = Comment(content, lecture)
db.session.add(comment)
db.session.commit()
return {
'id': comment.id
}
api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
|
Change single Lecture query to use first() in stead of all()
|
## Code Before:
from flask import request
from flask_restful import Resource, Api, abort, reqparse
from .models import db, Comment, Lecture
api = Api()
class CommentListResource(Resource):
def get(self, lecture_id):
db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not db_lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
db_comments = Comment.query.filter(Comment.lecture_id == lecture_id)
comments = [
{'id': c.id, 'content': c.content}
for c in db_comments
]
return {
'comments': comments
}
def post(self, lecture_id):
db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all()
if not db_lectures:
abort(404, message="Lecture {} does not exist".format(lecture_id))
parser = reqparse.RequestParser()
parser.add_argument('data', help='Text content of comment')
args = parser.parse_args()
if not args.data:
abort(400, message="Comment has no data parameter")
content = args.data
lecture = db_lectures[0]
comment = Comment(content, lecture)
db.session.add(comment)
db.session.commit()
return {
'id': comment.id
}
api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
## Instruction:
Change single Lecture query to use first() in stead of all()
## Code After:
from flask import request
from flask_restful import Resource, Api, abort, reqparse
from .models import db, Comment, Lecture
api = Api()
class CommentListResource(Resource):
def get(self, lecture_id):
db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not db_lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
db_comments = Comment.query.filter(Comment.lecture_id == lecture_id)
comments = [
{'id': c.id, 'content': c.content}
for c in db_comments
]
return {
'comments': comments
}
def post(self, lecture_id):
lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
parser = reqparse.RequestParser()
parser.add_argument('data', help='Text content of comment')
args = parser.parse_args()
if not args.data:
abort(400, message="Comment has no data parameter")
content = args.data
comment = Comment(content, lecture)
db.session.add(comment)
db.session.commit()
return {
'id': comment.id
}
api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
|
from flask import request
from flask_restful import Resource, Api, abort, reqparse
from .models import db, Comment, Lecture
api = Api()
class CommentListResource(Resource):
def get(self, lecture_id):
db_lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
if not db_lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
db_comments = Comment.query.filter(Comment.lecture_id == lecture_id)
comments = [
{'id': c.id, 'content': c.content}
for c in db_comments
]
return {
'comments': comments
}
def post(self, lecture_id):
- db_lectures = Lecture.query.filter(Lecture.id == lecture_id).all()
? --- - ^^^
+ lecture = Lecture.query.filter(Lecture.id == lecture_id).first()
? ^^^^^
- if not db_lectures:
? --- -
+ if not lecture:
abort(404, message="Lecture {} does not exist".format(lecture_id))
parser = reqparse.RequestParser()
parser.add_argument('data', help='Text content of comment')
args = parser.parse_args()
if not args.data:
abort(400, message="Comment has no data parameter")
content = args.data
- lecture = db_lectures[0]
comment = Comment(content, lecture)
db.session.add(comment)
db.session.commit()
return {
'id': comment.id
}
api.add_resource(CommentListResource, '/api/0/lectures/<lecture_id>/comments')
|
58e2059c37d7464e7ab7a1681ea8f465e9378940
|
ukpostcode/__init__.py
|
ukpostcode/__init__.py
|
from __future__ import unicode_literals
__version__ = '0.1.0'
def formater(postcode):
'''
Format a UK Post Code to the official form. Expects a valid postcode.
'''
postcode = postcode.upper().replace(' ', '').strip()
inward = postcode[-3:]
outward = postcode[:-3]
return '%s %s' % (outward, inward)
|
from __future__ import unicode_literals
__version__ = '0.1.0'
def formater(postcode):
'''
Format a UK Post Code to the official form. Expects a valid postcode.
'''
postcode = postcode.upper().replace(' ', '').strip()
inward = postcode[-3:]
outward = postcode[:-3]
return '%s %s' % (outward, inward)
def validate(postcode):
'''
Returns True if the postcode is valid. False otherwise.
'''
raise NotImplementedError()
|
Prepare to code the validator
|
Prepare to code the validator
|
Python
|
apache-2.0
|
alanjds/pyukpostcode
|
from __future__ import unicode_literals
__version__ = '0.1.0'
def formater(postcode):
'''
Format a UK Post Code to the official form. Expects a valid postcode.
'''
postcode = postcode.upper().replace(' ', '').strip()
inward = postcode[-3:]
outward = postcode[:-3]
return '%s %s' % (outward, inward)
+
+ def validate(postcode):
+ '''
+ Returns True if the postcode is valid. False otherwise.
+ '''
+ raise NotImplementedError()
+
|
Prepare to code the validator
|
## Code Before:
from __future__ import unicode_literals
__version__ = '0.1.0'
def formater(postcode):
'''
Format a UK Post Code to the official form. Expects a valid postcode.
'''
postcode = postcode.upper().replace(' ', '').strip()
inward = postcode[-3:]
outward = postcode[:-3]
return '%s %s' % (outward, inward)
## Instruction:
Prepare to code the validator
## Code After:
from __future__ import unicode_literals
__version__ = '0.1.0'
def formater(postcode):
'''
Format a UK Post Code to the official form. Expects a valid postcode.
'''
postcode = postcode.upper().replace(' ', '').strip()
inward = postcode[-3:]
outward = postcode[:-3]
return '%s %s' % (outward, inward)
def validate(postcode):
'''
Returns True if the postcode is valid. False otherwise.
'''
raise NotImplementedError()
|
from __future__ import unicode_literals
__version__ = '0.1.0'
def formater(postcode):
'''
Format a UK Post Code to the official form. Expects a valid postcode.
'''
postcode = postcode.upper().replace(' ', '').strip()
inward = postcode[-3:]
outward = postcode[:-3]
return '%s %s' % (outward, inward)
+
+
+ def validate(postcode):
+ '''
+ Returns True if the postcode is valid. False otherwise.
+ '''
+ raise NotImplementedError()
|
2c02816c05f3863ef76b3a412ac5bad9eecfafdd
|
testrepository/tests/test_setup.py
|
testrepository/tests/test_setup.py
|
"""Tests for setup.py."""
import doctest
import os
import subprocess
import sys
from testtools import (
TestCase,
)
from testtools.matchers import (
DocTestMatches,
)
class TestCanSetup(TestCase):
def test_bdist(self):
# Single smoke test to make sure we can build a package.
path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py')
proc = subprocess.Popen([sys.executable, path, 'bdist'],
stdin=subprocess.PIPE, stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
output, _ = proc.communicate()
self.assertEqual(0, proc.returncode)
self.assertThat(output,
DocTestMatches("...running bdist...", doctest.ELLIPSIS))
|
"""Tests for setup.py."""
import doctest
import os
import subprocess
import sys
from testtools import (
TestCase,
)
from testtools.matchers import (
DocTestMatches,
)
class TestCanSetup(TestCase):
def test_bdist(self):
# Single smoke test to make sure we can build a package.
path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py')
proc = subprocess.Popen([sys.executable, path, 'bdist'],
stdin=subprocess.PIPE, stdout=subprocess.PIPE,
stderr=subprocess.STDOUT, universal_newlines=True)
output, _ = proc.communicate()
self.assertEqual(0, proc.returncode)
self.assertThat(output, DocTestMatches("""...
running install_scripts
...
adding '...testr'
...""", doctest.ELLIPSIS))
|
Make setup.py smoke test more specific again as requested in review
|
Make setup.py smoke test more specific again as requested in review
|
Python
|
apache-2.0
|
masayukig/stestr,masayukig/stestr,mtreinish/stestr,mtreinish/stestr
|
"""Tests for setup.py."""
import doctest
import os
import subprocess
import sys
from testtools import (
TestCase,
)
from testtools.matchers import (
DocTestMatches,
)
class TestCanSetup(TestCase):
def test_bdist(self):
# Single smoke test to make sure we can build a package.
path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py')
proc = subprocess.Popen([sys.executable, path, 'bdist'],
stdin=subprocess.PIPE, stdout=subprocess.PIPE,
- stderr=subprocess.STDOUT)
+ stderr=subprocess.STDOUT, universal_newlines=True)
output, _ = proc.communicate()
self.assertEqual(0, proc.returncode)
- self.assertThat(output,
- DocTestMatches("...running bdist...", doctest.ELLIPSIS))
+ self.assertThat(output, DocTestMatches("""...
+ running install_scripts
+ ...
+ adding '...testr'
+ ...""", doctest.ELLIPSIS))
|
Make setup.py smoke test more specific again as requested in review
|
## Code Before:
"""Tests for setup.py."""
import doctest
import os
import subprocess
import sys
from testtools import (
TestCase,
)
from testtools.matchers import (
DocTestMatches,
)
class TestCanSetup(TestCase):
def test_bdist(self):
# Single smoke test to make sure we can build a package.
path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py')
proc = subprocess.Popen([sys.executable, path, 'bdist'],
stdin=subprocess.PIPE, stdout=subprocess.PIPE,
stderr=subprocess.STDOUT)
output, _ = proc.communicate()
self.assertEqual(0, proc.returncode)
self.assertThat(output,
DocTestMatches("...running bdist...", doctest.ELLIPSIS))
## Instruction:
Make setup.py smoke test more specific again as requested in review
## Code After:
"""Tests for setup.py."""
import doctest
import os
import subprocess
import sys
from testtools import (
TestCase,
)
from testtools.matchers import (
DocTestMatches,
)
class TestCanSetup(TestCase):
def test_bdist(self):
# Single smoke test to make sure we can build a package.
path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py')
proc = subprocess.Popen([sys.executable, path, 'bdist'],
stdin=subprocess.PIPE, stdout=subprocess.PIPE,
stderr=subprocess.STDOUT, universal_newlines=True)
output, _ = proc.communicate()
self.assertEqual(0, proc.returncode)
self.assertThat(output, DocTestMatches("""...
running install_scripts
...
adding '...testr'
...""", doctest.ELLIPSIS))
|
"""Tests for setup.py."""
import doctest
import os
import subprocess
import sys
from testtools import (
TestCase,
)
from testtools.matchers import (
DocTestMatches,
)
class TestCanSetup(TestCase):
def test_bdist(self):
# Single smoke test to make sure we can build a package.
path = os.path.join(os.path.dirname(__file__), '..', '..', 'setup.py')
proc = subprocess.Popen([sys.executable, path, 'bdist'],
stdin=subprocess.PIPE, stdout=subprocess.PIPE,
- stderr=subprocess.STDOUT)
+ stderr=subprocess.STDOUT, universal_newlines=True)
output, _ = proc.communicate()
self.assertEqual(0, proc.returncode)
- self.assertThat(output,
- DocTestMatches("...running bdist...", doctest.ELLIPSIS))
+ self.assertThat(output, DocTestMatches("""...
+ running install_scripts
+ ...
+ adding '...testr'
+ ...""", doctest.ELLIPSIS))
|
87707340ac82f852937dae546380b5d5327f5bc7
|
txlege84/core/views.py
|
txlege84/core/views.py
|
from django.views.generic import ListView
from bills.mixins import AllSubjectsMixin
from core.mixins import ConveneTimeMixin
from legislators.mixins import AllLegislatorsMixin, ChambersMixin
from explainers.models import Explainer
from topics.models import Topic
class LandingView(AllSubjectsMixin, AllLegislatorsMixin,
ChambersMixin, ConveneTimeMixin, ListView):
model = Topic
template_name = 'landing.html'
def get_context_data(self, **kwargs):
context = super(LandingView, self).get_context_data(**kwargs)
context['explainer_list'] = Explainer.objects.all().published()
return context
|
from django.views.generic import ListView
from core.mixins import ConveneTimeMixin
from explainers.models import Explainer
from topics.models import Topic
class LandingView(ConveneTimeMixin, ListView):
model = Topic
template_name = 'landing.html'
def get_context_data(self, **kwargs):
context = super(LandingView, self).get_context_data(**kwargs)
context['explainer_list'] = Explainer.objects.all().published()
return context
|
Remove unneeded mixins from LandingView
|
Remove unneeded mixins from LandingView
|
Python
|
mit
|
texastribune/txlege84,texastribune/txlege84,texastribune/txlege84,texastribune/txlege84
|
from django.views.generic import ListView
- from bills.mixins import AllSubjectsMixin
from core.mixins import ConveneTimeMixin
- from legislators.mixins import AllLegislatorsMixin, ChambersMixin
from explainers.models import Explainer
from topics.models import Topic
+ class LandingView(ConveneTimeMixin, ListView):
- class LandingView(AllSubjectsMixin, AllLegislatorsMixin,
- ChambersMixin, ConveneTimeMixin, ListView):
model = Topic
template_name = 'landing.html'
def get_context_data(self, **kwargs):
context = super(LandingView, self).get_context_data(**kwargs)
context['explainer_list'] = Explainer.objects.all().published()
return context
|
Remove unneeded mixins from LandingView
|
## Code Before:
from django.views.generic import ListView
from bills.mixins import AllSubjectsMixin
from core.mixins import ConveneTimeMixin
from legislators.mixins import AllLegislatorsMixin, ChambersMixin
from explainers.models import Explainer
from topics.models import Topic
class LandingView(AllSubjectsMixin, AllLegislatorsMixin,
ChambersMixin, ConveneTimeMixin, ListView):
model = Topic
template_name = 'landing.html'
def get_context_data(self, **kwargs):
context = super(LandingView, self).get_context_data(**kwargs)
context['explainer_list'] = Explainer.objects.all().published()
return context
## Instruction:
Remove unneeded mixins from LandingView
## Code After:
from django.views.generic import ListView
from core.mixins import ConveneTimeMixin
from explainers.models import Explainer
from topics.models import Topic
class LandingView(ConveneTimeMixin, ListView):
model = Topic
template_name = 'landing.html'
def get_context_data(self, **kwargs):
context = super(LandingView, self).get_context_data(**kwargs)
context['explainer_list'] = Explainer.objects.all().published()
return context
|
from django.views.generic import ListView
- from bills.mixins import AllSubjectsMixin
from core.mixins import ConveneTimeMixin
- from legislators.mixins import AllLegislatorsMixin, ChambersMixin
from explainers.models import Explainer
from topics.models import Topic
+ class LandingView(ConveneTimeMixin, ListView):
- class LandingView(AllSubjectsMixin, AllLegislatorsMixin,
- ChambersMixin, ConveneTimeMixin, ListView):
model = Topic
template_name = 'landing.html'
def get_context_data(self, **kwargs):
context = super(LandingView, self).get_context_data(**kwargs)
context['explainer_list'] = Explainer.objects.all().published()
return context
|
0e835c6381374c5b00b7387057d056d679f635c4
|
zproject/legacy_urls.py
|
zproject/legacy_urls.py
|
from django.conf.urls import url
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists),
]
|
from django.urls import path
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
path('json/subscriptions/exists', zerver.views.streams.json_stream_exists),
]
|
Migrate legacy urls to use modern django pattern.
|
urls: Migrate legacy urls to use modern django pattern.
|
Python
|
apache-2.0
|
shubhamdhama/zulip,punchagan/zulip,kou/zulip,showell/zulip,hackerkid/zulip,timabbott/zulip,eeshangarg/zulip,kou/zulip,andersk/zulip,zulip/zulip,synicalsyntax/zulip,zulip/zulip,andersk/zulip,shubhamdhama/zulip,showell/zulip,kou/zulip,hackerkid/zulip,shubhamdhama/zulip,andersk/zulip,eeshangarg/zulip,brainwane/zulip,shubhamdhama/zulip,hackerkid/zulip,punchagan/zulip,showell/zulip,brainwane/zulip,rht/zulip,rht/zulip,shubhamdhama/zulip,hackerkid/zulip,eeshangarg/zulip,hackerkid/zulip,brainwane/zulip,brainwane/zulip,punchagan/zulip,punchagan/zulip,punchagan/zulip,synicalsyntax/zulip,timabbott/zulip,zulip/zulip,rht/zulip,eeshangarg/zulip,brainwane/zulip,punchagan/zulip,timabbott/zulip,timabbott/zulip,synicalsyntax/zulip,synicalsyntax/zulip,hackerkid/zulip,synicalsyntax/zulip,showell/zulip,timabbott/zulip,rht/zulip,kou/zulip,rht/zulip,synicalsyntax/zulip,timabbott/zulip,rht/zulip,eeshangarg/zulip,zulip/zulip,zulip/zulip,eeshangarg/zulip,zulip/zulip,punchagan/zulip,hackerkid/zulip,kou/zulip,andersk/zulip,rht/zulip,andersk/zulip,kou/zulip,shubhamdhama/zulip,andersk/zulip,showell/zulip,timabbott/zulip,synicalsyntax/zulip,eeshangarg/zulip,showell/zulip,kou/zulip,brainwane/zulip,brainwane/zulip,andersk/zulip,zulip/zulip,shubhamdhama/zulip,showell/zulip
|
- from django.conf.urls import url
+ from django.urls import path
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
- url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists),
+ path('json/subscriptions/exists', zerver.views.streams.json_stream_exists),
]
|
Migrate legacy urls to use modern django pattern.
|
## Code Before:
from django.conf.urls import url
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists),
]
## Instruction:
Migrate legacy urls to use modern django pattern.
## Code After:
from django.urls import path
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
path('json/subscriptions/exists', zerver.views.streams.json_stream_exists),
]
|
- from django.conf.urls import url
? ----- ^^^
+ from django.urls import path
? ^^^^
import zerver.views
import zerver.views.streams
import zerver.views.auth
import zerver.views.tutorial
import zerver.views.report
# Future endpoints should add to urls.py, which includes these legacy urls
legacy_urls = [
# These are json format views used by the web client. They require a logged in browser.
# We should remove this endpoint and all code related to it.
# It returns a 404 if the stream doesn't exist, which is confusing
# for devs, and I don't think we need to go to the server
# any more to find out about subscriptions, since they are already
# pushed to us via the event system.
- url(r'^json/subscriptions/exists$', zerver.views.streams.json_stream_exists),
? ^^^ - - -
+ path('json/subscriptions/exists', zerver.views.streams.json_stream_exists),
? ^^^^
]
|
65c5474936dca27023e45c1644fa2a9492e9a420
|
tests/convergence_tests/run_convergence_tests_lspr.py
|
tests/convergence_tests/run_convergence_tests_lspr.py
|
import os
import time
import subprocess
import datetime
from check_for_meshes import check_mesh
# tests to run
tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py']
# specify CUDA device to use
CUDA_DEVICE = '0'
ENV = os.environ.copy()
ENV['CUDA_DEVICE'] = CUDA_DEVICE
mesh_file = ''
folder_name = 'lspr_convergence_test_meshes'
rename_folder = 'geometry_lspr'
size = '~3MB'
check_mesh(mesh_file, folder_name, rename_folder, size)
tic = time.time()
for test in tests:
subprocess.call(['python', '{}'.format(test)])
toc = time.time()
print("Total runtime for convergence tests: ")
print(str(datetime.timedelta(seconds=(toc - tic))))
|
import os
import time
import subprocess
import datetime
from check_for_meshes import check_mesh
# tests to run
tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py']
# specify CUDA device to use
CUDA_DEVICE = '0'
ENV = os.environ.copy()
ENV['CUDA_DEVICE'] = CUDA_DEVICE
mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip'
folder_name = 'lspr_convergence_test_meshes'
rename_folder = 'geometry_lspr'
size = '~3MB'
check_mesh(mesh_file, folder_name, rename_folder, size)
tic = time.time()
for test in tests:
subprocess.call(['python', '{}'.format(test)])
toc = time.time()
print("Total runtime for convergence tests: ")
print(str(datetime.timedelta(seconds=(toc - tic))))
|
Add path to convergence test lspr zip file
|
Add path to convergence test lspr zip file
|
Python
|
bsd-3-clause
|
barbagroup/pygbe,barbagroup/pygbe,barbagroup/pygbe
|
import os
import time
import subprocess
import datetime
from check_for_meshes import check_mesh
# tests to run
tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py']
# specify CUDA device to use
CUDA_DEVICE = '0'
ENV = os.environ.copy()
ENV['CUDA_DEVICE'] = CUDA_DEVICE
- mesh_file = ''
+ mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip'
folder_name = 'lspr_convergence_test_meshes'
rename_folder = 'geometry_lspr'
size = '~3MB'
check_mesh(mesh_file, folder_name, rename_folder, size)
tic = time.time()
for test in tests:
subprocess.call(['python', '{}'.format(test)])
toc = time.time()
print("Total runtime for convergence tests: ")
print(str(datetime.timedelta(seconds=(toc - tic))))
|
Add path to convergence test lspr zip file
|
## Code Before:
import os
import time
import subprocess
import datetime
from check_for_meshes import check_mesh
# tests to run
tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py']
# specify CUDA device to use
CUDA_DEVICE = '0'
ENV = os.environ.copy()
ENV['CUDA_DEVICE'] = CUDA_DEVICE
mesh_file = ''
folder_name = 'lspr_convergence_test_meshes'
rename_folder = 'geometry_lspr'
size = '~3MB'
check_mesh(mesh_file, folder_name, rename_folder, size)
tic = time.time()
for test in tests:
subprocess.call(['python', '{}'.format(test)])
toc = time.time()
print("Total runtime for convergence tests: ")
print(str(datetime.timedelta(seconds=(toc - tic))))
## Instruction:
Add path to convergence test lspr zip file
## Code After:
import os
import time
import subprocess
import datetime
from check_for_meshes import check_mesh
# tests to run
tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py']
# specify CUDA device to use
CUDA_DEVICE = '0'
ENV = os.environ.copy()
ENV['CUDA_DEVICE'] = CUDA_DEVICE
mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip'
folder_name = 'lspr_convergence_test_meshes'
rename_folder = 'geometry_lspr'
size = '~3MB'
check_mesh(mesh_file, folder_name, rename_folder, size)
tic = time.time()
for test in tests:
subprocess.call(['python', '{}'.format(test)])
toc = time.time()
print("Total runtime for convergence tests: ")
print(str(datetime.timedelta(seconds=(toc - tic))))
|
import os
import time
import subprocess
import datetime
from check_for_meshes import check_mesh
# tests to run
tests = ['sphere_lspr.py', 'sphere_multiple_lspr.py']
# specify CUDA device to use
CUDA_DEVICE = '0'
ENV = os.environ.copy()
ENV['CUDA_DEVICE'] = CUDA_DEVICE
- mesh_file = ''
+ mesh_file = 'https://zenodo.org/record/580786/files/pygbe-lspr_convergence_test_meshes.zip'
folder_name = 'lspr_convergence_test_meshes'
rename_folder = 'geometry_lspr'
size = '~3MB'
check_mesh(mesh_file, folder_name, rename_folder, size)
tic = time.time()
for test in tests:
subprocess.call(['python', '{}'.format(test)])
toc = time.time()
print("Total runtime for convergence tests: ")
print(str(datetime.timedelta(seconds=(toc - tic))))
|
bad2fea8a3a8e7a7d1da9ee83ec48657824eaa07
|
tests/test_filesize.py
|
tests/test_filesize.py
|
from rich import filesize
def test_traditional():
assert filesize.decimal(0) == "0 bytes"
assert filesize.decimal(1) == "1 byte"
assert filesize.decimal(2) == "2 bytes"
assert filesize.decimal(1000) == "1.0 kB"
assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB"
def test_pick_unit_and_suffix():
units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"]
assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes")
assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
|
from rich import filesize
def test_traditional():
assert filesize.decimal(0) == "0 bytes"
assert filesize.decimal(1) == "1 byte"
assert filesize.decimal(2) == "2 bytes"
assert filesize.decimal(1000) == "1.0 kB"
assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB"
assert filesize.decimal(0, precision=2) == "0 bytes"
assert filesize.decimal(1111, precision=0) == "1 kB"
assert filesize.decimal(1111, precision=1) == "1.1 kB"
assert filesize.decimal(1111, precision=2) == "1.11 kB"
assert filesize.decimal(1111, separator="") == "1.1kB"
def test_pick_unit_and_suffix():
units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"]
assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes")
assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
|
Add some tests for new decimal() params
|
Add some tests for new decimal() params
|
Python
|
mit
|
willmcgugan/rich
|
from rich import filesize
def test_traditional():
assert filesize.decimal(0) == "0 bytes"
assert filesize.decimal(1) == "1 byte"
assert filesize.decimal(2) == "2 bytes"
assert filesize.decimal(1000) == "1.0 kB"
assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB"
+ assert filesize.decimal(0, precision=2) == "0 bytes"
+ assert filesize.decimal(1111, precision=0) == "1 kB"
+ assert filesize.decimal(1111, precision=1) == "1.1 kB"
+ assert filesize.decimal(1111, precision=2) == "1.11 kB"
+ assert filesize.decimal(1111, separator="") == "1.1kB"
def test_pick_unit_and_suffix():
units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"]
assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes")
assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
|
Add some tests for new decimal() params
|
## Code Before:
from rich import filesize
def test_traditional():
assert filesize.decimal(0) == "0 bytes"
assert filesize.decimal(1) == "1 byte"
assert filesize.decimal(2) == "2 bytes"
assert filesize.decimal(1000) == "1.0 kB"
assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB"
def test_pick_unit_and_suffix():
units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"]
assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes")
assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
## Instruction:
Add some tests for new decimal() params
## Code After:
from rich import filesize
def test_traditional():
assert filesize.decimal(0) == "0 bytes"
assert filesize.decimal(1) == "1 byte"
assert filesize.decimal(2) == "2 bytes"
assert filesize.decimal(1000) == "1.0 kB"
assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB"
assert filesize.decimal(0, precision=2) == "0 bytes"
assert filesize.decimal(1111, precision=0) == "1 kB"
assert filesize.decimal(1111, precision=1) == "1.1 kB"
assert filesize.decimal(1111, precision=2) == "1.11 kB"
assert filesize.decimal(1111, separator="") == "1.1kB"
def test_pick_unit_and_suffix():
units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"]
assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes")
assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
|
from rich import filesize
def test_traditional():
assert filesize.decimal(0) == "0 bytes"
assert filesize.decimal(1) == "1 byte"
assert filesize.decimal(2) == "2 bytes"
assert filesize.decimal(1000) == "1.0 kB"
assert filesize.decimal(1.5 * 1000 * 1000) == "1.5 MB"
+ assert filesize.decimal(0, precision=2) == "0 bytes"
+ assert filesize.decimal(1111, precision=0) == "1 kB"
+ assert filesize.decimal(1111, precision=1) == "1.1 kB"
+ assert filesize.decimal(1111, precision=2) == "1.11 kB"
+ assert filesize.decimal(1111, separator="") == "1.1kB"
def test_pick_unit_and_suffix():
units = ["bytes", "KB", "MB", "GB", "TB", "PB", "EB", "ZB", "YB"]
assert filesize.pick_unit_and_suffix(50, units, 1024) == (1, "bytes")
assert filesize.pick_unit_and_suffix(2048, units, 1024) == (1024, "KB")
|
3e20365624f02b70d8332ba7ff7da23961337f86
|
quickstart/python/understand/example-3/create_joke_samples.6.x.py
|
quickstart/python/understand/example-3/create_joke_samples.6.x.py
|
from twilio.rest import Client
# Your Account Sid and Auth Token from twilio.com/console
account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX'
auth_token = 'your_auth_token'
client = Client(account_sid, auth_token)
phrases = [
'Tell me a joke',
'Tell me a joke',
'I\'d like to hear a joke',
'Do you know any good jokes?',
'Joke',
'Tell joke',
'Tell me something funny',
'Make me laugh',
'I want to hear a joke',
'Can I hear a joke?',
'I like jokes',
'I\'d like to hear a punny joke'
]
# Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list
for phrase in phrases:
sample = client.preview.understand \
.assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.tasks('tell-a-joke') \
.samples \
.create(language='en-us', tagged_text=phrase)
print(sample.sid)
|
from twilio.rest import Client
# Your Account Sid and Auth Token from twilio.com/console
account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX'
auth_token = 'your_auth_token'
client = Client(account_sid, auth_token)
phrases = [
'Tell me a joke',
'Tell me a joke',
'I\'d like to hear a joke',
'Do you know any good jokes?',
'Joke',
'Tell joke',
'Tell me something funny',
'Make me laugh',
'I want to hear a joke',
'Can I hear a joke?',
'I like jokes',
'I\'d like to hear a punny joke'
]
# Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list
# Replace 'UDXXX...' with the SID for the task you just created.
for phrase in phrases:
sample = client.preview.understand \
.assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.samples \
.create(language='en-us', tagged_text=phrase)
print(sample.sid)
|
Update samples creation for intent rename
|
Update samples creation for intent rename
Update intent --> task, code comment
|
Python
|
mit
|
TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets
|
from twilio.rest import Client
# Your Account Sid and Auth Token from twilio.com/console
account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX'
auth_token = 'your_auth_token'
client = Client(account_sid, auth_token)
phrases = [
'Tell me a joke',
'Tell me a joke',
'I\'d like to hear a joke',
'Do you know any good jokes?',
'Joke',
'Tell joke',
'Tell me something funny',
'Make me laugh',
'I want to hear a joke',
'Can I hear a joke?',
'I like jokes',
'I\'d like to hear a punny joke'
]
# Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list
+ # Replace 'UDXXX...' with the SID for the task you just created.
for phrase in phrases:
sample = client.preview.understand \
.assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
- .tasks('tell-a-joke') \
+ .tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.samples \
.create(language='en-us', tagged_text=phrase)
print(sample.sid)
|
Update samples creation for intent rename
|
## Code Before:
from twilio.rest import Client
# Your Account Sid and Auth Token from twilio.com/console
account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX'
auth_token = 'your_auth_token'
client = Client(account_sid, auth_token)
phrases = [
'Tell me a joke',
'Tell me a joke',
'I\'d like to hear a joke',
'Do you know any good jokes?',
'Joke',
'Tell joke',
'Tell me something funny',
'Make me laugh',
'I want to hear a joke',
'Can I hear a joke?',
'I like jokes',
'I\'d like to hear a punny joke'
]
# Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list
for phrase in phrases:
sample = client.preview.understand \
.assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.tasks('tell-a-joke') \
.samples \
.create(language='en-us', tagged_text=phrase)
print(sample.sid)
## Instruction:
Update samples creation for intent rename
## Code After:
from twilio.rest import Client
# Your Account Sid and Auth Token from twilio.com/console
account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX'
auth_token = 'your_auth_token'
client = Client(account_sid, auth_token)
phrases = [
'Tell me a joke',
'Tell me a joke',
'I\'d like to hear a joke',
'Do you know any good jokes?',
'Joke',
'Tell joke',
'Tell me something funny',
'Make me laugh',
'I want to hear a joke',
'Can I hear a joke?',
'I like jokes',
'I\'d like to hear a punny joke'
]
# Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list
# Replace 'UDXXX...' with the SID for the task you just created.
for phrase in phrases:
sample = client.preview.understand \
.assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.samples \
.create(language='en-us', tagged_text=phrase)
print(sample.sid)
|
from twilio.rest import Client
# Your Account Sid and Auth Token from twilio.com/console
account_sid = 'ACXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX'
auth_token = 'your_auth_token'
client = Client(account_sid, auth_token)
phrases = [
'Tell me a joke',
'Tell me a joke',
'I\'d like to hear a joke',
'Do you know any good jokes?',
'Joke',
'Tell joke',
'Tell me something funny',
'Make me laugh',
'I want to hear a joke',
'Can I hear a joke?',
'I like jokes',
'I\'d like to hear a punny joke'
]
# Replace 'UAXXX...' with your Assistant's unique SID https://www.twilio.com/console/autopilot/list
+ # Replace 'UDXXX...' with the SID for the task you just created.
for phrase in phrases:
sample = client.preview.understand \
.assistants('UAXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
- .tasks('tell-a-joke') \
+ .tasks('UDXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX') \
.samples \
.create(language='en-us', tagged_text=phrase)
print(sample.sid)
|
d33a624fa6aedb93ae43ba1d2c0f6a76d90ff4a6
|
foldermd5sums.py
|
foldermd5sums.py
|
import json
import os
import sys
import hashlib
def get_md5sums(directory):
md5sums = []
for filename in os.listdir(directory):
md5 = hashlib.md5()
with open(os.path.join(directory, filename), 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
import json
import os
import sys
import hashlib
def get_relative_filepaths(base_directory):
""" Return a list of file paths without the base_directory prefix"""
file_list = []
for root, subFolders, files in os.walk('Data'):
relative_path="/".join(root.split('/')[1:])
for file in files:
file_list.append(os.path.join(relative_path,file))
return file_list
def get_md5sums(base_directory):
md5sums = []
for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
full_filepath = os.path.join(base_directory, filename)
with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
Allow directory of files to be indexed
|
ENH: Allow directory of files to be indexed
In the Data directory, there may be sub-directories of files that need to
be kept separate, but all of them need to be indexed.
|
Python
|
apache-2.0
|
zivy/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,zivy/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks,zivy/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks
|
import json
import os
import sys
import hashlib
+ def get_relative_filepaths(base_directory):
+ """ Return a list of file paths without the base_directory prefix"""
+ file_list = []
+ for root, subFolders, files in os.walk('Data'):
+ relative_path="/".join(root.split('/')[1:])
+ for file in files:
+ file_list.append(os.path.join(relative_path,file))
+ return file_list
+
+
- def get_md5sums(directory):
+ def get_md5sums(base_directory):
md5sums = []
- for filename in os.listdir(directory):
+ for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
- with open(os.path.join(directory, filename), 'rb') as fp:
+ full_filepath = os.path.join(base_directory, filename)
+ with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
-
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
Allow directory of files to be indexed
|
## Code Before:
import json
import os
import sys
import hashlib
def get_md5sums(directory):
md5sums = []
for filename in os.listdir(directory):
md5 = hashlib.md5()
with open(os.path.join(directory, filename), 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
## Instruction:
Allow directory of files to be indexed
## Code After:
import json
import os
import sys
import hashlib
def get_relative_filepaths(base_directory):
""" Return a list of file paths without the base_directory prefix"""
file_list = []
for root, subFolders, files in os.walk('Data'):
relative_path="/".join(root.split('/')[1:])
for file in files:
file_list.append(os.path.join(relative_path,file))
return file_list
def get_md5sums(base_directory):
md5sums = []
for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
full_filepath = os.path.join(base_directory, filename)
with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
import json
import os
import sys
import hashlib
+ def get_relative_filepaths(base_directory):
+ """ Return a list of file paths without the base_directory prefix"""
+ file_list = []
+ for root, subFolders, files in os.walk('Data'):
+ relative_path="/".join(root.split('/')[1:])
+ for file in files:
+ file_list.append(os.path.join(relative_path,file))
+ return file_list
+
+
- def get_md5sums(directory):
+ def get_md5sums(base_directory):
? +++++
md5sums = []
- for filename in os.listdir(directory):
+ for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
- with open(os.path.join(directory, filename), 'rb') as fp:
+ full_filepath = os.path.join(base_directory, filename)
+ with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
-
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
c6d7f2b1214e86f09431ab1d8e5c312f7a87081d
|
pttrack/views.py
|
pttrack/views.py
|
from django.shortcuts import render
from django.http import HttpResponse, HttpResponseRedirect
# Create your views here.from django.http import HttpResponse
def index(request):
return HttpResponse("Hello, world. You're at the BIG TABLE.")
def clindate(request, clindate):
(year, month, day) = clindate.split("-")
return HttpResponse("Clinic date %s" % year+" "+month+" "+day)
def patient(request, pt_uuid):
return HttpResponse("You're looking at patient %s" % pt_uuid)
def intake(request):
from . import forms as myforms
# if this is a POST request we need to process the form data
if request.method == 'POST':
# create a form instance and populate it with data from the request:
form = myforms.PatientForm(request.POST)
# check whether it's valid:
if form.is_valid():
# process the data in form.cleaned_data as required
# ...
# redirect to a new URL:
return HttpResponseRedirect('/')
# if a GET (or any other method) we'll create a blank form
else:
form = myforms.PatientForm()
return render(request, 'pttrack/intake.html', {'form': form})
|
from django.shortcuts import render
from django.http import HttpResponse, HttpResponseRedirect
from django.core.urlresolvers import reverse
from . import models as mymodels
# Create your views here.from django.http import HttpResponse
def index(request):
return HttpResponse("Hello, world. You're at the BIG TABLE.")
def clindate(request, clindate):
(year, month, day) = clindate.split("-")
return HttpResponse("Clinic date %s" % year+" "+month+" "+day)
def patient(request, pt_uuid):
return HttpResponse("You're looking at patient %s" % pt_uuid)
def intake(request):
from . import forms as myforms
# if this is a POST request we need to process the form data
if request.method == 'POST':
# create a form instance and populate it with data from the request:
form = myforms.PatientForm(request.POST)
# check whether it's valid:
if form.is_valid():
p = mymodels.Patient(**form.cleaned_data)
p.save()
# redirect to a new URL:
return HttpResponseRedirect(reverse(patient, args=(p.id,)))
# if a GET (or any other method) we'll create a blank form
else:
form = myforms.PatientForm()
return render(request, 'pttrack/intake.html', {'form': form})
|
Set up redirect at db saves for new patients.
|
Set up redirect at db saves for new patients.
|
Python
|
mit
|
SaturdayNeighborhoodHealthClinic/clintools,SaturdayNeighborhoodHealthClinic/clintools,SaturdayNeighborhoodHealthClinic/clintools
|
from django.shortcuts import render
from django.http import HttpResponse, HttpResponseRedirect
+ from django.core.urlresolvers import reverse
+
+ from . import models as mymodels
# Create your views here.from django.http import HttpResponse
def index(request):
return HttpResponse("Hello, world. You're at the BIG TABLE.")
def clindate(request, clindate):
(year, month, day) = clindate.split("-")
return HttpResponse("Clinic date %s" % year+" "+month+" "+day)
def patient(request, pt_uuid):
return HttpResponse("You're looking at patient %s" % pt_uuid)
def intake(request):
from . import forms as myforms
# if this is a POST request we need to process the form data
if request.method == 'POST':
# create a form instance and populate it with data from the request:
form = myforms.PatientForm(request.POST)
# check whether it's valid:
if form.is_valid():
- # process the data in form.cleaned_data as required
- # ...
+ p = mymodels.Patient(**form.cleaned_data)
+ p.save()
+
# redirect to a new URL:
- return HttpResponseRedirect('/')
+ return HttpResponseRedirect(reverse(patient, args=(p.id,)))
# if a GET (or any other method) we'll create a blank form
else:
form = myforms.PatientForm()
return render(request, 'pttrack/intake.html', {'form': form})
|
Set up redirect at db saves for new patients.
|
## Code Before:
from django.shortcuts import render
from django.http import HttpResponse, HttpResponseRedirect
# Create your views here.from django.http import HttpResponse
def index(request):
return HttpResponse("Hello, world. You're at the BIG TABLE.")
def clindate(request, clindate):
(year, month, day) = clindate.split("-")
return HttpResponse("Clinic date %s" % year+" "+month+" "+day)
def patient(request, pt_uuid):
return HttpResponse("You're looking at patient %s" % pt_uuid)
def intake(request):
from . import forms as myforms
# if this is a POST request we need to process the form data
if request.method == 'POST':
# create a form instance and populate it with data from the request:
form = myforms.PatientForm(request.POST)
# check whether it's valid:
if form.is_valid():
# process the data in form.cleaned_data as required
# ...
# redirect to a new URL:
return HttpResponseRedirect('/')
# if a GET (or any other method) we'll create a blank form
else:
form = myforms.PatientForm()
return render(request, 'pttrack/intake.html', {'form': form})
## Instruction:
Set up redirect at db saves for new patients.
## Code After:
from django.shortcuts import render
from django.http import HttpResponse, HttpResponseRedirect
from django.core.urlresolvers import reverse
from . import models as mymodels
# Create your views here.from django.http import HttpResponse
def index(request):
return HttpResponse("Hello, world. You're at the BIG TABLE.")
def clindate(request, clindate):
(year, month, day) = clindate.split("-")
return HttpResponse("Clinic date %s" % year+" "+month+" "+day)
def patient(request, pt_uuid):
return HttpResponse("You're looking at patient %s" % pt_uuid)
def intake(request):
from . import forms as myforms
# if this is a POST request we need to process the form data
if request.method == 'POST':
# create a form instance and populate it with data from the request:
form = myforms.PatientForm(request.POST)
# check whether it's valid:
if form.is_valid():
p = mymodels.Patient(**form.cleaned_data)
p.save()
# redirect to a new URL:
return HttpResponseRedirect(reverse(patient, args=(p.id,)))
# if a GET (or any other method) we'll create a blank form
else:
form = myforms.PatientForm()
return render(request, 'pttrack/intake.html', {'form': form})
|
from django.shortcuts import render
from django.http import HttpResponse, HttpResponseRedirect
+ from django.core.urlresolvers import reverse
+
+ from . import models as mymodels
# Create your views here.from django.http import HttpResponse
def index(request):
return HttpResponse("Hello, world. You're at the BIG TABLE.")
def clindate(request, clindate):
(year, month, day) = clindate.split("-")
return HttpResponse("Clinic date %s" % year+" "+month+" "+day)
def patient(request, pt_uuid):
return HttpResponse("You're looking at patient %s" % pt_uuid)
def intake(request):
from . import forms as myforms
# if this is a POST request we need to process the form data
if request.method == 'POST':
# create a form instance and populate it with data from the request:
form = myforms.PatientForm(request.POST)
# check whether it's valid:
if form.is_valid():
- # process the data in form.cleaned_data as required
- # ...
+ p = mymodels.Patient(**form.cleaned_data)
+ p.save()
+
# redirect to a new URL:
- return HttpResponseRedirect('/')
+ return HttpResponseRedirect(reverse(patient, args=(p.id,)))
# if a GET (or any other method) we'll create a blank form
else:
form = myforms.PatientForm()
return render(request, 'pttrack/intake.html', {'form': form})
|
0654d962918327e5143fb9250ad344de26e284eb
|
electrumx_server.py
|
electrumx_server.py
|
'''Script to kick off the server.'''
import logging
import traceback
from server.env import Env
from server.controller import Controller
def main():
'''Set up logging and run the server.'''
logging.basicConfig(level=logging.INFO,
format='%(asctime)s %(levelname)-9s %(message)-100s '
'%(name)s [%(filename)s:%(lineno)d]')
logging.info('ElectrumX server starting')
try:
controller = Controller(Env())
controller.run()
except Exception:
traceback.print_exc()
logging.critical('ElectrumX server terminated abnormally')
else:
logging.info('ElectrumX server terminated normally')
if __name__ == '__main__':
main()
|
'''Script to kick off the server.'''
import logging
import traceback
from server.env import Env
from server.controller import Controller
def main():
'''Set up logging and run the server.'''
logging.basicConfig(level=logging.INFO,
format='%(asctime)s %(levelname)-7s %(message)-100s '
'[%(filename)s:%(lineno)d]')
logging.info('ElectrumX server starting')
try:
controller = Controller(Env())
controller.run()
except Exception:
traceback.print_exc()
logging.critical('ElectrumX server terminated abnormally')
else:
logging.info('ElectrumX server terminated normally')
if __name__ == '__main__':
main()
|
Remove logger name from logs
|
Remove logger name from logs
|
Python
|
mit
|
thelazier/electrumx,shsmith/electrumx,shsmith/electrumx,erasmospunk/electrumx,erasmospunk/electrumx,thelazier/electrumx
|
'''Script to kick off the server.'''
import logging
import traceback
from server.env import Env
from server.controller import Controller
def main():
'''Set up logging and run the server.'''
logging.basicConfig(level=logging.INFO,
- format='%(asctime)s %(levelname)-9s %(message)-100s '
+ format='%(asctime)s %(levelname)-7s %(message)-100s '
- '%(name)s [%(filename)s:%(lineno)d]')
+ '[%(filename)s:%(lineno)d]')
logging.info('ElectrumX server starting')
try:
controller = Controller(Env())
controller.run()
except Exception:
traceback.print_exc()
logging.critical('ElectrumX server terminated abnormally')
else:
logging.info('ElectrumX server terminated normally')
if __name__ == '__main__':
main()
|
Remove logger name from logs
|
## Code Before:
'''Script to kick off the server.'''
import logging
import traceback
from server.env import Env
from server.controller import Controller
def main():
'''Set up logging and run the server.'''
logging.basicConfig(level=logging.INFO,
format='%(asctime)s %(levelname)-9s %(message)-100s '
'%(name)s [%(filename)s:%(lineno)d]')
logging.info('ElectrumX server starting')
try:
controller = Controller(Env())
controller.run()
except Exception:
traceback.print_exc()
logging.critical('ElectrumX server terminated abnormally')
else:
logging.info('ElectrumX server terminated normally')
if __name__ == '__main__':
main()
## Instruction:
Remove logger name from logs
## Code After:
'''Script to kick off the server.'''
import logging
import traceback
from server.env import Env
from server.controller import Controller
def main():
'''Set up logging and run the server.'''
logging.basicConfig(level=logging.INFO,
format='%(asctime)s %(levelname)-7s %(message)-100s '
'[%(filename)s:%(lineno)d]')
logging.info('ElectrumX server starting')
try:
controller = Controller(Env())
controller.run()
except Exception:
traceback.print_exc()
logging.critical('ElectrumX server terminated abnormally')
else:
logging.info('ElectrumX server terminated normally')
if __name__ == '__main__':
main()
|
'''Script to kick off the server.'''
import logging
import traceback
from server.env import Env
from server.controller import Controller
def main():
'''Set up logging and run the server.'''
logging.basicConfig(level=logging.INFO,
- format='%(asctime)s %(levelname)-9s %(message)-100s '
? ^
+ format='%(asctime)s %(levelname)-7s %(message)-100s '
? ^
- '%(name)s [%(filename)s:%(lineno)d]')
? ---------
+ '[%(filename)s:%(lineno)d]')
logging.info('ElectrumX server starting')
try:
controller = Controller(Env())
controller.run()
except Exception:
traceback.print_exc()
logging.critical('ElectrumX server terminated abnormally')
else:
logging.info('ElectrumX server terminated normally')
if __name__ == '__main__':
main()
|
a3f5e1338cc84c60b867fc04175253f7ab460912
|
relay_api/api/backend.py
|
relay_api/api/backend.py
|
import json
from relay_api.core.relay import relay
from relay_api.conf.config import relays
def init_relays():
for r in relays:
relays[r]["object"] = relay(relays[r]["gpio"])
relays[r]["state"] = relays[r]["object"].get_state()
def get_all_relays():
relays_dict = __get_relay_dict()
return json.dumps(relays_dict)
def get_relay(relay_name):
if relay_name not in relays:
return None
relay_dict = __get_relay_dict(relay_name)
return json.dumps(relay_dict)
def __get_relay_dict(relay_name=None):
if relay_name:
relay_dict = dict.copy(relays["relay_name"])
del(relay_dict["object"])
return relay_dict
relays_dict = dict.copy(relays)
for r in relays_dict:
del(relays_dict[r]["object"])
return relays_dict
|
import json
from relay_api.core.relay import relay
from relay_api.conf.config import relays
def init_relays():
for r in relays:
relays[r]["object"] = relay(relays[r]["gpio"])
relays[r]["state"] = relays[r]["object"].get_state()
def get_all_relays():
relays_dict = __get_relay_dict()
return json.dumps(relays_dict, indent=4)
def get_relay(relay_name):
if relay_name not in relays:
return None
relay_dict = __get_relay_dict(relay_name)
return json.dumps(relay_dict, indent=4)
def __get_relay_dict(relay_name=None):
if relay_name:
relay_dict = dict.copy(relays["relay_name"])
del(relay_dict["object"])
return relay_dict
relays_dict = dict.copy(relays)
for r in relays_dict:
del(relays_dict[r]["object"])
return relays_dict
|
Add indent in json to improve debugging
|
Add indent in json to improve debugging
|
Python
|
mit
|
pahumadad/raspi-relay-api
|
import json
from relay_api.core.relay import relay
from relay_api.conf.config import relays
def init_relays():
for r in relays:
relays[r]["object"] = relay(relays[r]["gpio"])
relays[r]["state"] = relays[r]["object"].get_state()
def get_all_relays():
relays_dict = __get_relay_dict()
- return json.dumps(relays_dict)
+ return json.dumps(relays_dict, indent=4)
def get_relay(relay_name):
if relay_name not in relays:
return None
relay_dict = __get_relay_dict(relay_name)
- return json.dumps(relay_dict)
+ return json.dumps(relay_dict, indent=4)
def __get_relay_dict(relay_name=None):
if relay_name:
relay_dict = dict.copy(relays["relay_name"])
del(relay_dict["object"])
return relay_dict
relays_dict = dict.copy(relays)
for r in relays_dict:
del(relays_dict[r]["object"])
return relays_dict
|
Add indent in json to improve debugging
|
## Code Before:
import json
from relay_api.core.relay import relay
from relay_api.conf.config import relays
def init_relays():
for r in relays:
relays[r]["object"] = relay(relays[r]["gpio"])
relays[r]["state"] = relays[r]["object"].get_state()
def get_all_relays():
relays_dict = __get_relay_dict()
return json.dumps(relays_dict)
def get_relay(relay_name):
if relay_name not in relays:
return None
relay_dict = __get_relay_dict(relay_name)
return json.dumps(relay_dict)
def __get_relay_dict(relay_name=None):
if relay_name:
relay_dict = dict.copy(relays["relay_name"])
del(relay_dict["object"])
return relay_dict
relays_dict = dict.copy(relays)
for r in relays_dict:
del(relays_dict[r]["object"])
return relays_dict
## Instruction:
Add indent in json to improve debugging
## Code After:
import json
from relay_api.core.relay import relay
from relay_api.conf.config import relays
def init_relays():
for r in relays:
relays[r]["object"] = relay(relays[r]["gpio"])
relays[r]["state"] = relays[r]["object"].get_state()
def get_all_relays():
relays_dict = __get_relay_dict()
return json.dumps(relays_dict, indent=4)
def get_relay(relay_name):
if relay_name not in relays:
return None
relay_dict = __get_relay_dict(relay_name)
return json.dumps(relay_dict, indent=4)
def __get_relay_dict(relay_name=None):
if relay_name:
relay_dict = dict.copy(relays["relay_name"])
del(relay_dict["object"])
return relay_dict
relays_dict = dict.copy(relays)
for r in relays_dict:
del(relays_dict[r]["object"])
return relays_dict
|
import json
from relay_api.core.relay import relay
from relay_api.conf.config import relays
def init_relays():
for r in relays:
relays[r]["object"] = relay(relays[r]["gpio"])
relays[r]["state"] = relays[r]["object"].get_state()
def get_all_relays():
relays_dict = __get_relay_dict()
- return json.dumps(relays_dict)
+ return json.dumps(relays_dict, indent=4)
? ++++++++++
def get_relay(relay_name):
if relay_name not in relays:
return None
relay_dict = __get_relay_dict(relay_name)
- return json.dumps(relay_dict)
+ return json.dumps(relay_dict, indent=4)
? ++++++++++
def __get_relay_dict(relay_name=None):
if relay_name:
relay_dict = dict.copy(relays["relay_name"])
del(relay_dict["object"])
return relay_dict
relays_dict = dict.copy(relays)
for r in relays_dict:
del(relays_dict[r]["object"])
return relays_dict
|
00ddeefdcdacb811f5e665a91139e165d7217f84
|
week1/poc_2048_merge_template.py
|
week1/poc_2048_merge_template.py
|
def merge(line):
"""
Function that merges a single row or column in 2048.
"""
l = len(line)
s1 = [0]*l
j = 0
for i in range(l):
if l[i] != 0:
s1[j] = l[i]
return []
a = [2,0,2,4]
print merge(a)
|
def merge(line):
"""
Function that merges a single row or column in 2048.
"""
l = len(line)
s1 = [0]*l
j = 0
for i in range(l):
if line[i] != 0:
s1[j] = line[i]
j += 1
return s1
a = [2,0,2,4]
print (merge(a))
|
Modify the correct merge 1 fct
|
Modify the correct merge 1 fct
|
Python
|
mit
|
Crescent-Saturn/Principles-of-Computing
|
def merge(line):
"""
Function that merges a single row or column in 2048.
"""
l = len(line)
s1 = [0]*l
j = 0
for i in range(l):
- if l[i] != 0:
+ if line[i] != 0:
- s1[j] = l[i]
-
+ s1[j] = line[i]
+ j += 1
- return []
+ return s1
a = [2,0,2,4]
- print merge(a)
+ print (merge(a))
|
Modify the correct merge 1 fct
|
## Code Before:
def merge(line):
"""
Function that merges a single row or column in 2048.
"""
l = len(line)
s1 = [0]*l
j = 0
for i in range(l):
if l[i] != 0:
s1[j] = l[i]
return []
a = [2,0,2,4]
print merge(a)
## Instruction:
Modify the correct merge 1 fct
## Code After:
def merge(line):
"""
Function that merges a single row or column in 2048.
"""
l = len(line)
s1 = [0]*l
j = 0
for i in range(l):
if line[i] != 0:
s1[j] = line[i]
j += 1
return s1
a = [2,0,2,4]
print (merge(a))
|
def merge(line):
"""
Function that merges a single row or column in 2048.
"""
l = len(line)
s1 = [0]*l
j = 0
for i in range(l):
- if l[i] != 0:
? ^
+ if line[i] != 0:
? ^^^^ +++
- s1[j] = l[i]
-
+ s1[j] = line[i]
+ j += 1
- return []
? ^^
+ return s1
? ^^
a = [2,0,2,4]
- print merge(a)
+ print (merge(a))
? + +
|
a0903bb9fd988662269e9f2ef7e38acd877a63d5
|
src/nodeconductor_saltstack/saltstack/handlers.py
|
src/nodeconductor_saltstack/saltstack/handlers.py
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
Add more details to event logs for property CRUD
|
Add more details to event logs for property CRUD
|
Python
|
mit
|
opennode/nodeconductor-saltstack
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
- '%s {property_name} has been created.' % instance.get_type_display_name(),
+ '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
- '%s {property_name} has been updated.' % instance.get_type_display_name(),
+ '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
- '%s {property_name} has been deleted.' % instance.get_type_display_name(),
+ '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
Add more details to event logs for property CRUD
|
## Code Before:
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
## Instruction:
Add more details to event logs for property CRUD
## Code After:
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
'%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
'%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
'%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
from __future__ import unicode_literals
import logging
from .log import event_logger
logger = logging.getLogger(__name__)
def log_saltstack_property_created(sender, instance, created=False, **kwargs):
if created:
event_logger.saltstack_property.info(
- '%s {property_name} has been created.' % instance.get_type_display_name(),
+ '%s {property_name} has been created in {tenant_full_name}.' % instance.get_type_display_name(),
? ++++++++++++++++++++++
event_type='saltstack_property_creation_succeeded',
event_context={
'property': instance,
})
else:
event_logger.saltstack_property.info(
- '%s {property_name} has been updated.' % instance.get_type_display_name(),
+ '%s {property_name} has been updated in {tenant_full_name}.' % instance.get_type_display_name(),
? ++++++++++++++++++++++
event_type='saltstack_property_update_succeeded',
event_context={
'property': instance,
})
def log_saltstack_property_deleted(sender, instance, **kwargs):
event_logger.saltstack_property.info(
- '%s {property_name} has been deleted.' % instance.get_type_display_name(),
+ '%s {property_name} has been deleted from {tenant_full_name}.' % instance.get_type_display_name(),
? ++++++++++++++++++++++++
event_type='saltstack_property_deletion_succeeded',
event_context={
'property': instance,
})
|
3222fab1b026250d9aee863d068137b03c13a05b
|
tests/test_check_dependencies.py
|
tests/test_check_dependencies.py
|
from check_dependencies import CheckDependencies
def test_default():
CheckDependencies(None)
def test_hydrotrend():
CheckDependencies("hydrotrend")
|
from check_dependencies import CheckDependencies
def test_default():
CheckDependencies(None)
def test_hydrotrend():
CheckDependencies("hydrotrend")
def test_cem():
CheckDependencies("cem")
|
Add dependency check test for CEM
|
Add dependency check test for CEM
|
Python
|
mit
|
csdms/rpm_models
|
from check_dependencies import CheckDependencies
def test_default():
CheckDependencies(None)
def test_hydrotrend():
CheckDependencies("hydrotrend")
+ def test_cem():
+ CheckDependencies("cem")
+
|
Add dependency check test for CEM
|
## Code Before:
from check_dependencies import CheckDependencies
def test_default():
CheckDependencies(None)
def test_hydrotrend():
CheckDependencies("hydrotrend")
## Instruction:
Add dependency check test for CEM
## Code After:
from check_dependencies import CheckDependencies
def test_default():
CheckDependencies(None)
def test_hydrotrend():
CheckDependencies("hydrotrend")
def test_cem():
CheckDependencies("cem")
|
from check_dependencies import CheckDependencies
def test_default():
CheckDependencies(None)
def test_hydrotrend():
CheckDependencies("hydrotrend")
+
+ def test_cem():
+ CheckDependencies("cem")
|
1599d4ed14fb3d7c7e551c9f6ce3f86d9df17cbd
|
mammoth/writers/html.py
|
mammoth/writers/html.py
|
from __future__ import unicode_literals
from .abc import Writer
import cgi
class HtmlWriter(Writer):
def __init__(self):
self._fragments = []
def text(self, text):
self._fragments.append(_escape_html(text))
def start(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1}>".format(name, attribute_string))
def end(self, name):
self._fragments.append("</{0}>".format(name))
def self_closing(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1} />".format(name, attribute_string))
def append(self, html):
self._fragments.append(html)
def as_string(self):
return "".join(self._fragments)
def _escape_html(text):
return cgi.escape(text, quote=True)
def _generate_attribute_string(attributes):
if attributes is None:
return ""
else:
return "".join(
' {0}="{1}"'.format(key, _escape_html(attributes[key]))
for key in sorted(attributes)
)
|
from __future__ import unicode_literals
from xml.sax.saxutils import escape
from .abc import Writer
class HtmlWriter(Writer):
def __init__(self):
self._fragments = []
def text(self, text):
self._fragments.append(_escape_html(text))
def start(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1}>".format(name, attribute_string))
def end(self, name):
self._fragments.append("</{0}>".format(name))
def self_closing(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1} />".format(name, attribute_string))
def append(self, html):
self._fragments.append(html)
def as_string(self):
return "".join(self._fragments)
def _escape_html(text):
return escape(text, {'"': """})
def _generate_attribute_string(attributes):
if attributes is None:
return ""
else:
return "".join(
' {0}="{1}"'.format(key, _escape_html(attributes[key]))
for key in sorted(attributes)
)
|
Use xml.sax.saxutils.escape instead of deprecated cgi.escape
|
Use xml.sax.saxutils.escape instead of deprecated cgi.escape
```
/usr/local/lib/python3.6/dist-packages/mammoth/writers/html.py:34: DeprecationWarning: cgi.escape is deprecated, use html.escape instead
return cgi.escape(text, quote=True)
```
|
Python
|
bsd-2-clause
|
mwilliamson/python-mammoth
|
from __future__ import unicode_literals
+ from xml.sax.saxutils import escape
from .abc import Writer
-
- import cgi
class HtmlWriter(Writer):
def __init__(self):
self._fragments = []
def text(self, text):
self._fragments.append(_escape_html(text))
def start(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1}>".format(name, attribute_string))
def end(self, name):
self._fragments.append("</{0}>".format(name))
def self_closing(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1} />".format(name, attribute_string))
def append(self, html):
self._fragments.append(html)
def as_string(self):
return "".join(self._fragments)
def _escape_html(text):
- return cgi.escape(text, quote=True)
+ return escape(text, {'"': """})
def _generate_attribute_string(attributes):
if attributes is None:
return ""
else:
return "".join(
' {0}="{1}"'.format(key, _escape_html(attributes[key]))
for key in sorted(attributes)
)
|
Use xml.sax.saxutils.escape instead of deprecated cgi.escape
|
## Code Before:
from __future__ import unicode_literals
from .abc import Writer
import cgi
class HtmlWriter(Writer):
def __init__(self):
self._fragments = []
def text(self, text):
self._fragments.append(_escape_html(text))
def start(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1}>".format(name, attribute_string))
def end(self, name):
self._fragments.append("</{0}>".format(name))
def self_closing(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1} />".format(name, attribute_string))
def append(self, html):
self._fragments.append(html)
def as_string(self):
return "".join(self._fragments)
def _escape_html(text):
return cgi.escape(text, quote=True)
def _generate_attribute_string(attributes):
if attributes is None:
return ""
else:
return "".join(
' {0}="{1}"'.format(key, _escape_html(attributes[key]))
for key in sorted(attributes)
)
## Instruction:
Use xml.sax.saxutils.escape instead of deprecated cgi.escape
## Code After:
from __future__ import unicode_literals
from xml.sax.saxutils import escape
from .abc import Writer
class HtmlWriter(Writer):
def __init__(self):
self._fragments = []
def text(self, text):
self._fragments.append(_escape_html(text))
def start(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1}>".format(name, attribute_string))
def end(self, name):
self._fragments.append("</{0}>".format(name))
def self_closing(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1} />".format(name, attribute_string))
def append(self, html):
self._fragments.append(html)
def as_string(self):
return "".join(self._fragments)
def _escape_html(text):
return escape(text, {'"': """})
def _generate_attribute_string(attributes):
if attributes is None:
return ""
else:
return "".join(
' {0}="{1}"'.format(key, _escape_html(attributes[key]))
for key in sorted(attributes)
)
|
from __future__ import unicode_literals
+ from xml.sax.saxutils import escape
from .abc import Writer
-
- import cgi
class HtmlWriter(Writer):
def __init__(self):
self._fragments = []
def text(self, text):
self._fragments.append(_escape_html(text))
def start(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1}>".format(name, attribute_string))
def end(self, name):
self._fragments.append("</{0}>".format(name))
def self_closing(self, name, attributes=None):
attribute_string = _generate_attribute_string(attributes)
self._fragments.append("<{0}{1} />".format(name, attribute_string))
def append(self, html):
self._fragments.append(html)
def as_string(self):
return "".join(self._fragments)
def _escape_html(text):
- return cgi.escape(text, quote=True)
+ return escape(text, {'"': """})
def _generate_attribute_string(attributes):
if attributes is None:
return ""
else:
return "".join(
' {0}="{1}"'.format(key, _escape_html(attributes[key]))
for key in sorted(attributes)
)
|
d88429d072f79c38d65ccaf3519495905f12f03f
|
calaccess_website/management/commands/updatedownloadswebsite.py
|
calaccess_website/management/commands/updatedownloadswebsite.py
|
import logging
from django.core.management import call_command
from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand
logger = logging.getLogger(__name__)
class Command(updatecommand):
"""
Update to the latest CAL-ACCESS snapshot and bake static website pages.
"""
help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages'
def add_arguments(self, parser):
"""
Adds custom arguments specific to this command.
"""
super(Command, self).add_arguments(parser)
parser.add_argument(
"--publish",
action="store_true",
dest="publish",
default=False,
help="Publish baked content"
)
def handle(self, *args, **options):
"""
Make it happen.
"""
super(Command, self).handle(*args, **options)
self.header('Creating latest file links')
call_command('createlatestlinks')
self.header('Baking downloads-website content')
call_command('build')
if options['publish']:
self.header('Publishing baked content to S3 bucket.')
call_command('publish')
self.success("Done!")
|
import logging
from django.core.management import call_command
from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand
logger = logging.getLogger(__name__)
class Command(updatecommand):
"""
Update to the latest CAL-ACCESS snapshot and bake static website pages.
"""
help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages'
def add_arguments(self, parser):
"""
Adds custom arguments specific to this command.
"""
super(Command, self).add_arguments(parser)
parser.add_argument(
"--publish",
action="store_true",
dest="publish",
default=False,
help="Publish baked content"
)
def handle(self, *args, **options):
"""
Make it happen.
"""
super(Command, self).handle(*args, **options)
call_command('processcalaccessdata')
self.header('Creating latest file links')
call_command('createlatestlinks')
self.header('Baking downloads-website content')
call_command('build')
if options['publish']:
self.header('Publishing baked content to S3 bucket.')
call_command('publish')
self.success("Done!")
|
Add processing cmd to update routine
|
Add processing cmd to update routine
|
Python
|
mit
|
california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website,california-civic-data-coalition/django-calaccess-downloads-website
|
import logging
from django.core.management import call_command
from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand
logger = logging.getLogger(__name__)
class Command(updatecommand):
"""
Update to the latest CAL-ACCESS snapshot and bake static website pages.
"""
help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages'
def add_arguments(self, parser):
"""
Adds custom arguments specific to this command.
"""
super(Command, self).add_arguments(parser)
parser.add_argument(
"--publish",
action="store_true",
dest="publish",
default=False,
help="Publish baked content"
)
def handle(self, *args, **options):
"""
Make it happen.
"""
super(Command, self).handle(*args, **options)
+ call_command('processcalaccessdata')
self.header('Creating latest file links')
call_command('createlatestlinks')
self.header('Baking downloads-website content')
call_command('build')
if options['publish']:
self.header('Publishing baked content to S3 bucket.')
call_command('publish')
self.success("Done!")
|
Add processing cmd to update routine
|
## Code Before:
import logging
from django.core.management import call_command
from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand
logger = logging.getLogger(__name__)
class Command(updatecommand):
"""
Update to the latest CAL-ACCESS snapshot and bake static website pages.
"""
help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages'
def add_arguments(self, parser):
"""
Adds custom arguments specific to this command.
"""
super(Command, self).add_arguments(parser)
parser.add_argument(
"--publish",
action="store_true",
dest="publish",
default=False,
help="Publish baked content"
)
def handle(self, *args, **options):
"""
Make it happen.
"""
super(Command, self).handle(*args, **options)
self.header('Creating latest file links')
call_command('createlatestlinks')
self.header('Baking downloads-website content')
call_command('build')
if options['publish']:
self.header('Publishing baked content to S3 bucket.')
call_command('publish')
self.success("Done!")
## Instruction:
Add processing cmd to update routine
## Code After:
import logging
from django.core.management import call_command
from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand
logger = logging.getLogger(__name__)
class Command(updatecommand):
"""
Update to the latest CAL-ACCESS snapshot and bake static website pages.
"""
help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages'
def add_arguments(self, parser):
"""
Adds custom arguments specific to this command.
"""
super(Command, self).add_arguments(parser)
parser.add_argument(
"--publish",
action="store_true",
dest="publish",
default=False,
help="Publish baked content"
)
def handle(self, *args, **options):
"""
Make it happen.
"""
super(Command, self).handle(*args, **options)
call_command('processcalaccessdata')
self.header('Creating latest file links')
call_command('createlatestlinks')
self.header('Baking downloads-website content')
call_command('build')
if options['publish']:
self.header('Publishing baked content to S3 bucket.')
call_command('publish')
self.success("Done!")
|
import logging
from django.core.management import call_command
from calaccess_raw.management.commands.updatecalaccessrawdata import Command as updatecommand
logger = logging.getLogger(__name__)
class Command(updatecommand):
"""
Update to the latest CAL-ACCESS snapshot and bake static website pages.
"""
help = 'Update to the latest CAL-ACCESS snapshot and bake static website pages'
def add_arguments(self, parser):
"""
Adds custom arguments specific to this command.
"""
super(Command, self).add_arguments(parser)
parser.add_argument(
"--publish",
action="store_true",
dest="publish",
default=False,
help="Publish baked content"
)
def handle(self, *args, **options):
"""
Make it happen.
"""
super(Command, self).handle(*args, **options)
+ call_command('processcalaccessdata')
self.header('Creating latest file links')
call_command('createlatestlinks')
self.header('Baking downloads-website content')
call_command('build')
if options['publish']:
self.header('Publishing baked content to S3 bucket.')
call_command('publish')
self.success("Done!")
|
7a00ff49799afc50da74a748d07c52fef57ebc84
|
setup.py
|
setup.py
|
import tungsten
from distutils.core import setup
setup(
name='Tungsten',
version=tungsten.__version__,
author='Seena Burns',
packages={'tungsten': 'tungsten'},
license=open('LICENSE.txt').read(),
description='Wolfram Alpha API built for Python.',
long_description=open('README.md').read(),
install_requires=[
"requests",
],
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7'
),
)
|
import tungsten
from distutils.core import setup
setup(
name='Tungsten',
version=tungsten.__version__,
author='Seena Burns',
author_email='[email protected]',
url='https://github.com/seenaburns/Tungsten',
packages={'tungsten': 'tungsten'},
license=open('LICENSE.txt').read(),
description='Wolfram Alpha API built for Python.',
long_description=open('README.md').read(),
install_requires=[
"requests",
],
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7'
),
)
|
Add url / author email for PyPI regs
|
Add url / author email for PyPI regs
|
Python
|
bsd-3-clause
|
seenaburns/Tungsten
|
import tungsten
from distutils.core import setup
setup(
name='Tungsten',
version=tungsten.__version__,
author='Seena Burns',
+ author_email='[email protected]',
+ url='https://github.com/seenaburns/Tungsten',
packages={'tungsten': 'tungsten'},
license=open('LICENSE.txt').read(),
description='Wolfram Alpha API built for Python.',
long_description=open('README.md').read(),
install_requires=[
"requests",
],
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7'
),
)
|
Add url / author email for PyPI regs
|
## Code Before:
import tungsten
from distutils.core import setup
setup(
name='Tungsten',
version=tungsten.__version__,
author='Seena Burns',
packages={'tungsten': 'tungsten'},
license=open('LICENSE.txt').read(),
description='Wolfram Alpha API built for Python.',
long_description=open('README.md').read(),
install_requires=[
"requests",
],
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7'
),
)
## Instruction:
Add url / author email for PyPI regs
## Code After:
import tungsten
from distutils.core import setup
setup(
name='Tungsten',
version=tungsten.__version__,
author='Seena Burns',
author_email='[email protected]',
url='https://github.com/seenaburns/Tungsten',
packages={'tungsten': 'tungsten'},
license=open('LICENSE.txt').read(),
description='Wolfram Alpha API built for Python.',
long_description=open('README.md').read(),
install_requires=[
"requests",
],
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7'
),
)
|
import tungsten
from distutils.core import setup
setup(
name='Tungsten',
version=tungsten.__version__,
author='Seena Burns',
+ author_email='[email protected]',
+ url='https://github.com/seenaburns/Tungsten',
packages={'tungsten': 'tungsten'},
license=open('LICENSE.txt').read(),
description='Wolfram Alpha API built for Python.',
long_description=open('README.md').read(),
install_requires=[
"requests",
],
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7'
),
)
|
446d36cbbf79083b9d41ea5b152c5a845560eb4b
|
whats_fresh/whats_fresh_api/tests/views/test_stories.py
|
whats_fresh/whats_fresh_api/tests/views/test_stories.py
|
from django.test import TestCase
from django.test.client import Client
from django.core.urlresolvers import reverse
from whats_fresh_api.models import *
from django.contrib.gis.db import models
import json
class StoriesTestCase(TestCase):
fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json']
def setUp(self):
self.expected_json = """
{
"error": {
"error_status": false,
"error_name": null,
"error_text": null,
"error_level": null
},
"story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..."
}"""
def test_url_endpoint(self):
url = reverse('story-details', kwargs={'id': '1'})
self.assertEqual(url, '/stories/1')
def test_json_equals(self):
c = Client()
response = c.get(reverse('story-details', kwargs={'id': '1'})).content
parsed_answer = json.loads(response)
expected_answer = json.loads(self.expected_json)
self.assertTrue(parsed_answer == expected_answer)
|
from django.test import TestCase
from django.test.client import Client
from django.core.urlresolvers import reverse
from whats_fresh_api.models import *
from django.contrib.gis.db import models
import json
class StoriesTestCase(TestCase):
fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json']
def setUp(self):
self.expected_json = """"
{
<<<<<<< HEAD
"error": {
"error_status": false,
"error_name": null,
"error_text": null,
"error_level": null
},
"story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..."
}"""
def test_url_endpoint(self):
url = reverse('story-details', kwargs={'id': '1'})
self.assertEqual(url, '/stories/1')
def test_json_equals(self):
c = Client()
response = c.get(reverse('story-details', kwargs={'id': '1'})).content
parsed_answer = json.loads(response)
expected_answer = json.loads(self.expected_json)
self.assertTrue(parsed_answer == expected_answer)
|
Add error field to expected JSON
|
Add error field to expected JSON
|
Python
|
apache-2.0
|
iCHAIT/whats-fresh-api,iCHAIT/whats-fresh-api,osu-cass/whats-fresh-api,iCHAIT/whats-fresh-api,iCHAIT/whats-fresh-api,osu-cass/whats-fresh-api,osu-cass/whats-fresh-api,osu-cass/whats-fresh-api
|
from django.test import TestCase
from django.test.client import Client
from django.core.urlresolvers import reverse
from whats_fresh_api.models import *
from django.contrib.gis.db import models
import json
class StoriesTestCase(TestCase):
fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json']
def setUp(self):
- self.expected_json = """
+ self.expected_json = """"
{
+ <<<<<<< HEAD
"error": {
"error_status": false,
"error_name": null,
"error_text": null,
"error_level": null
},
"story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..."
}"""
+
def test_url_endpoint(self):
url = reverse('story-details', kwargs={'id': '1'})
self.assertEqual(url, '/stories/1')
def test_json_equals(self):
c = Client()
response = c.get(reverse('story-details', kwargs={'id': '1'})).content
parsed_answer = json.loads(response)
expected_answer = json.loads(self.expected_json)
self.assertTrue(parsed_answer == expected_answer)
|
Add error field to expected JSON
|
## Code Before:
from django.test import TestCase
from django.test.client import Client
from django.core.urlresolvers import reverse
from whats_fresh_api.models import *
from django.contrib.gis.db import models
import json
class StoriesTestCase(TestCase):
fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json']
def setUp(self):
self.expected_json = """
{
"error": {
"error_status": false,
"error_name": null,
"error_text": null,
"error_level": null
},
"story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..."
}"""
def test_url_endpoint(self):
url = reverse('story-details', kwargs={'id': '1'})
self.assertEqual(url, '/stories/1')
def test_json_equals(self):
c = Client()
response = c.get(reverse('story-details', kwargs={'id': '1'})).content
parsed_answer = json.loads(response)
expected_answer = json.loads(self.expected_json)
self.assertTrue(parsed_answer == expected_answer)
## Instruction:
Add error field to expected JSON
## Code After:
from django.test import TestCase
from django.test.client import Client
from django.core.urlresolvers import reverse
from whats_fresh_api.models import *
from django.contrib.gis.db import models
import json
class StoriesTestCase(TestCase):
fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json']
def setUp(self):
self.expected_json = """"
{
<<<<<<< HEAD
"error": {
"error_status": false,
"error_name": null,
"error_text": null,
"error_level": null
},
"story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..."
}"""
def test_url_endpoint(self):
url = reverse('story-details', kwargs={'id': '1'})
self.assertEqual(url, '/stories/1')
def test_json_equals(self):
c = Client()
response = c.get(reverse('story-details', kwargs={'id': '1'})).content
parsed_answer = json.loads(response)
expected_answer = json.loads(self.expected_json)
self.assertTrue(parsed_answer == expected_answer)
|
from django.test import TestCase
from django.test.client import Client
from django.core.urlresolvers import reverse
from whats_fresh_api.models import *
from django.contrib.gis.db import models
import json
class StoriesTestCase(TestCase):
fixtures = ['whats_fresh_api/tests/testdata/test_fixtures.json']
def setUp(self):
- self.expected_json = """
+ self.expected_json = """"
? +
{
+ <<<<<<< HEAD
"error": {
"error_status": false,
"error_name": null,
"error_text": null,
"error_level": null
},
"story": "These are the voyages of the Starfish Enterblub; her five year mission -- to seek out new fish and new fishilizations..."
}"""
+
def test_url_endpoint(self):
url = reverse('story-details', kwargs={'id': '1'})
self.assertEqual(url, '/stories/1')
def test_json_equals(self):
c = Client()
response = c.get(reverse('story-details', kwargs={'id': '1'})).content
parsed_answer = json.loads(response)
expected_answer = json.loads(self.expected_json)
self.assertTrue(parsed_answer == expected_answer)
|
fba4fdf426b0a29ca06deb67587c2bd804adb017
|
tbgxmlutils/xmlutils.py
|
tbgxmlutils/xmlutils.py
|
from xml.dom import minidom
import xml.etree.ElementTree as ET
import xmltodict
def add(k, parent=None, txt=None, attrs=None):
if parent is None:
handle = ET.Element(k)
else:
handle = ET.SubElement(parent, k)
if txt: handle.text = unicode(txt)
try:
for k, v in attrs.iteritems(): handle.attrib[k] = v
except AttributeError:
pass
return handle
def etree2xml(e, encoding='UTF-8'):
return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e)
def pretty(xml=None, fn=None):
if fn is not None:
xml = minidom.parse(fn)
elif not isinstance(xml, minidom.Document):
xml = minidom.parseString(xml)
return xml.toprettyxml(indent=' ')
def xml_fn_to_json(fn):
fh = open(fn, 'r')
json = xmltodict.parse(fh.read())
return json
|
from xml.dom import minidom
import lxml.etree as ET
import xmltodict
def add(k, parent=None, txt=None, attrs=None):
if parent is None:
handle = ET.Element(k)
else:
handle = ET.SubElement(parent, k)
if txt: handle.text = unicode(txt)
try:
for k, v in attrs.iteritems(): handle.attrib[k] = v
except AttributeError:
pass
return handle
def etree2xml(e, encoding='UTF-8'):
return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e)
def pretty(xml=None, fn=None):
if fn is not None:
xml = minidom.parse(fn)
elif not isinstance(xml, minidom.Document):
xml = minidom.parseString(xml)
return xml.toprettyxml(indent=' ')
def xml_fn_to_json(fn):
fh = open(fn, 'r')
json = xmltodict.parse(fh.read())
return json
|
Use lxml instead of elementtree.
|
Use lxml instead of elementtree.
|
Python
|
mit
|
Schwarzschild/TBGXMLUtils
|
from xml.dom import minidom
- import xml.etree.ElementTree as ET
+ import lxml.etree as ET
import xmltodict
def add(k, parent=None, txt=None, attrs=None):
if parent is None:
handle = ET.Element(k)
else:
handle = ET.SubElement(parent, k)
if txt: handle.text = unicode(txt)
try:
for k, v in attrs.iteritems(): handle.attrib[k] = v
except AttributeError:
pass
return handle
def etree2xml(e, encoding='UTF-8'):
return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e)
def pretty(xml=None, fn=None):
if fn is not None:
xml = minidom.parse(fn)
elif not isinstance(xml, minidom.Document):
xml = minidom.parseString(xml)
return xml.toprettyxml(indent=' ')
def xml_fn_to_json(fn):
fh = open(fn, 'r')
json = xmltodict.parse(fh.read())
return json
|
Use lxml instead of elementtree.
|
## Code Before:
from xml.dom import minidom
import xml.etree.ElementTree as ET
import xmltodict
def add(k, parent=None, txt=None, attrs=None):
if parent is None:
handle = ET.Element(k)
else:
handle = ET.SubElement(parent, k)
if txt: handle.text = unicode(txt)
try:
for k, v in attrs.iteritems(): handle.attrib[k] = v
except AttributeError:
pass
return handle
def etree2xml(e, encoding='UTF-8'):
return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e)
def pretty(xml=None, fn=None):
if fn is not None:
xml = minidom.parse(fn)
elif not isinstance(xml, minidom.Document):
xml = minidom.parseString(xml)
return xml.toprettyxml(indent=' ')
def xml_fn_to_json(fn):
fh = open(fn, 'r')
json = xmltodict.parse(fh.read())
return json
## Instruction:
Use lxml instead of elementtree.
## Code After:
from xml.dom import minidom
import lxml.etree as ET
import xmltodict
def add(k, parent=None, txt=None, attrs=None):
if parent is None:
handle = ET.Element(k)
else:
handle = ET.SubElement(parent, k)
if txt: handle.text = unicode(txt)
try:
for k, v in attrs.iteritems(): handle.attrib[k] = v
except AttributeError:
pass
return handle
def etree2xml(e, encoding='UTF-8'):
return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e)
def pretty(xml=None, fn=None):
if fn is not None:
xml = minidom.parse(fn)
elif not isinstance(xml, minidom.Document):
xml = minidom.parseString(xml)
return xml.toprettyxml(indent=' ')
def xml_fn_to_json(fn):
fh = open(fn, 'r')
json = xmltodict.parse(fh.read())
return json
|
from xml.dom import minidom
- import xml.etree.ElementTree as ET
? ------------
+ import lxml.etree as ET
? +
import xmltodict
def add(k, parent=None, txt=None, attrs=None):
if parent is None:
handle = ET.Element(k)
else:
handle = ET.SubElement(parent, k)
if txt: handle.text = unicode(txt)
try:
for k, v in attrs.iteritems(): handle.attrib[k] = v
except AttributeError:
pass
return handle
def etree2xml(e, encoding='UTF-8'):
return ET.tostring(e, encoding=encoding) if encoding else ET.tostring(e)
def pretty(xml=None, fn=None):
if fn is not None:
xml = minidom.parse(fn)
elif not isinstance(xml, minidom.Document):
xml = minidom.parseString(xml)
return xml.toprettyxml(indent=' ')
def xml_fn_to_json(fn):
fh = open(fn, 'r')
json = xmltodict.parse(fh.read())
return json
|
ab5aac0c9b0e075901c4cd8dd5d134e79f0e0110
|
brasileirao/spiders/results_spider.py
|
brasileirao/spiders/results_spider.py
|
import scrapy
import scrapy.selector
from brasileirao.items import BrasileiraoItem
import hashlib
class ResultsSpider(scrapy.Spider):
name = "results"
start_urls = [
'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/',
]
def parse(self, response):
actual_round = 0
for rodada in response.css('.rodadas .confrontos li'):
actual_round += 1
for game in rodada.css(".confronto"):
home_team = game.css(".partida .time1")
away_team = game.css(".partida .time2")
item = BrasileiraoItem()
item['rodada'] = actual_round
item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8')
item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8')
item['home_score'] = home_team.css(".gols::text").extract_first()
item['away_score'] = away_team.css(".gols::text").extract_first()
item['date'] = game.css(".info-partida time::attr(datetime)").extract_first()
id = item['home_team'] + item['away_team']
item['id'] = hashlib.md5(id).hexdigest()
yield item
|
import scrapy
import scrapy.selector
from brasileirao.items import BrasileiraoItem
import hashlib
class ResultsSpider(scrapy.Spider):
name = "results"
start_urls = [
'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/',
]
def parse(self, response):
actual_round = 0
for rodada in response.css('.rodadas .confrontos li'):
actual_round += 1
for game in rodada.css(".confronto"):
home_team = game.css(".partida .time1")
away_team = game.css(".partida .time2")
item = BrasileiraoItem()
item['rodada'] = actual_round
item['home_team'] = home_team.css("abbr::attr(title)").extract_first()
item['away_team'] = away_team.css("abbr::attr(title)").extract_first()
item['home_score'] = home_team.css(".gols::text").extract_first()
item['away_score'] = away_team.css(".gols::text").extract_first()
item['date'] = game.css(".info-partida time::attr(datetime)").extract_first()
id = item['home_team'] + item['away_team']
item['id'] = hashlib.md5(id).hexdigest()
yield item
|
Set utf-8 as default encoding.
|
Set utf-8 as default encoding.
|
Python
|
mit
|
pghilardi/live-football-client
|
+
import scrapy
import scrapy.selector
from brasileirao.items import BrasileiraoItem
import hashlib
class ResultsSpider(scrapy.Spider):
name = "results"
start_urls = [
'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/',
]
def parse(self, response):
actual_round = 0
for rodada in response.css('.rodadas .confrontos li'):
actual_round += 1
for game in rodada.css(".confronto"):
home_team = game.css(".partida .time1")
away_team = game.css(".partida .time2")
item = BrasileiraoItem()
item['rodada'] = actual_round
- item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8')
+ item['home_team'] = home_team.css("abbr::attr(title)").extract_first()
- item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8')
+ item['away_team'] = away_team.css("abbr::attr(title)").extract_first()
item['home_score'] = home_team.css(".gols::text").extract_first()
item['away_score'] = away_team.css(".gols::text").extract_first()
item['date'] = game.css(".info-partida time::attr(datetime)").extract_first()
id = item['home_team'] + item['away_team']
item['id'] = hashlib.md5(id).hexdigest()
yield item
|
Set utf-8 as default encoding.
|
## Code Before:
import scrapy
import scrapy.selector
from brasileirao.items import BrasileiraoItem
import hashlib
class ResultsSpider(scrapy.Spider):
name = "results"
start_urls = [
'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/',
]
def parse(self, response):
actual_round = 0
for rodada in response.css('.rodadas .confrontos li'):
actual_round += 1
for game in rodada.css(".confronto"):
home_team = game.css(".partida .time1")
away_team = game.css(".partida .time2")
item = BrasileiraoItem()
item['rodada'] = actual_round
item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8')
item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8')
item['home_score'] = home_team.css(".gols::text").extract_first()
item['away_score'] = away_team.css(".gols::text").extract_first()
item['date'] = game.css(".info-partida time::attr(datetime)").extract_first()
id = item['home_team'] + item['away_team']
item['id'] = hashlib.md5(id).hexdigest()
yield item
## Instruction:
Set utf-8 as default encoding.
## Code After:
import scrapy
import scrapy.selector
from brasileirao.items import BrasileiraoItem
import hashlib
class ResultsSpider(scrapy.Spider):
name = "results"
start_urls = [
'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/',
]
def parse(self, response):
actual_round = 0
for rodada in response.css('.rodadas .confrontos li'):
actual_round += 1
for game in rodada.css(".confronto"):
home_team = game.css(".partida .time1")
away_team = game.css(".partida .time2")
item = BrasileiraoItem()
item['rodada'] = actual_round
item['home_team'] = home_team.css("abbr::attr(title)").extract_first()
item['away_team'] = away_team.css("abbr::attr(title)").extract_first()
item['home_score'] = home_team.css(".gols::text").extract_first()
item['away_score'] = away_team.css(".gols::text").extract_first()
item['date'] = game.css(".info-partida time::attr(datetime)").extract_first()
id = item['home_team'] + item['away_team']
item['id'] = hashlib.md5(id).hexdigest()
yield item
|
+
import scrapy
import scrapy.selector
from brasileirao.items import BrasileiraoItem
import hashlib
class ResultsSpider(scrapy.Spider):
name = "results"
start_urls = [
'https://esporte.uol.com.br/futebol/campeonatos/brasileirao/jogos/',
]
def parse(self, response):
actual_round = 0
for rodada in response.css('.rodadas .confrontos li'):
actual_round += 1
for game in rodada.css(".confronto"):
home_team = game.css(".partida .time1")
away_team = game.css(".partida .time2")
item = BrasileiraoItem()
item['rodada'] = actual_round
- item['home_team'] = home_team.css("abbr::attr(title)").extract_first().encode('utf8')
? ---------------
+ item['home_team'] = home_team.css("abbr::attr(title)").extract_first()
- item['away_team'] = away_team.css("abbr::attr(title)").extract_first().encode('utf8')
? ---------------
+ item['away_team'] = away_team.css("abbr::attr(title)").extract_first()
item['home_score'] = home_team.css(".gols::text").extract_first()
item['away_score'] = away_team.css(".gols::text").extract_first()
item['date'] = game.css(".info-partida time::attr(datetime)").extract_first()
id = item['home_team'] + item['away_team']
item['id'] = hashlib.md5(id).hexdigest()
yield item
|
55e0c877dbe1a073534c9cf445ffe58715160b8e
|
metadata/RomsLite/hooks/post-stage.py
|
metadata/RomsLite/hooks/post-stage.py
|
import os
import shutil
from wmt.utils.hook import find_simulation_input_file
def execute(env):
"""Perform post-stage tasks for running a component.
Parameters
----------
env : dict
A dict of component parameter values from WMT.
"""
for name in ('river_forcing_file', 'waves_forcing_file'):
src = find_simulation_input_file(env[name])
shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
|
import os
import shutil
from wmt.utils.hook import find_simulation_input_file
_DEFAULT_FILES = {
'river_forcing_file': 'river.nc',
'waves_forcing_file': 'waves.nc',
}
def execute(env):
"""Perform post-stage tasks for running a component.
Parameters
----------
env : dict
A dict of component parameter values from WMT.
"""
for name in _DEFAULT_FILES:
if env[name] != _DEFAULT_FILES[name]:
try:
os.remove(os.path.join(os.curdir, 'Forcing', env[name]))
except OSError:
pass
src = find_simulation_input_file(env[name])
shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
|
Remove default forcing files if not being used.
|
Remove default forcing files if not being used.
|
Python
|
mit
|
csdms/wmt-metadata
|
import os
import shutil
from wmt.utils.hook import find_simulation_input_file
+
+ _DEFAULT_FILES = {
+ 'river_forcing_file': 'river.nc',
+ 'waves_forcing_file': 'waves.nc',
+ }
def execute(env):
"""Perform post-stage tasks for running a component.
Parameters
----------
env : dict
A dict of component parameter values from WMT.
"""
- for name in ('river_forcing_file', 'waves_forcing_file'):
+ for name in _DEFAULT_FILES:
+ if env[name] != _DEFAULT_FILES[name]:
+ try:
+ os.remove(os.path.join(os.curdir, 'Forcing', env[name]))
+ except OSError:
+ pass
+
src = find_simulation_input_file(env[name])
shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
|
Remove default forcing files if not being used.
|
## Code Before:
import os
import shutil
from wmt.utils.hook import find_simulation_input_file
def execute(env):
"""Perform post-stage tasks for running a component.
Parameters
----------
env : dict
A dict of component parameter values from WMT.
"""
for name in ('river_forcing_file', 'waves_forcing_file'):
src = find_simulation_input_file(env[name])
shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
## Instruction:
Remove default forcing files if not being used.
## Code After:
import os
import shutil
from wmt.utils.hook import find_simulation_input_file
_DEFAULT_FILES = {
'river_forcing_file': 'river.nc',
'waves_forcing_file': 'waves.nc',
}
def execute(env):
"""Perform post-stage tasks for running a component.
Parameters
----------
env : dict
A dict of component parameter values from WMT.
"""
for name in _DEFAULT_FILES:
if env[name] != _DEFAULT_FILES[name]:
try:
os.remove(os.path.join(os.curdir, 'Forcing', env[name]))
except OSError:
pass
src = find_simulation_input_file(env[name])
shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
|
import os
import shutil
from wmt.utils.hook import find_simulation_input_file
+
+ _DEFAULT_FILES = {
+ 'river_forcing_file': 'river.nc',
+ 'waves_forcing_file': 'waves.nc',
+ }
def execute(env):
"""Perform post-stage tasks for running a component.
Parameters
----------
env : dict
A dict of component parameter values from WMT.
"""
- for name in ('river_forcing_file', 'waves_forcing_file'):
+ for name in _DEFAULT_FILES:
+ if env[name] != _DEFAULT_FILES[name]:
+ try:
+ os.remove(os.path.join(os.curdir, 'Forcing', env[name]))
+ except OSError:
+ pass
+
src = find_simulation_input_file(env[name])
shutil.copy(src, os.path.join(os.curdir, 'Forcing'))
|
db19dfa17261c3d04de0202b2809ba8abb70326b
|
tests/unit/test_moxstubout.py
|
tests/unit/test_moxstubout.py
|
from oslotest import base
from oslotest import moxstubout
class TestMoxStubout(base.BaseTestCase):
def _stubable(self):
pass
def test_basic_stubout(self):
f = self.useFixture(moxstubout.MoxStubout())
before = TestMoxStubout._stubable
f.mox.StubOutWithMock(TestMoxStubout, '_stubable')
after = TestMoxStubout._stubable
self.assertNotEqual(before, after)
f.cleanUp()
after2 = TestMoxStubout._stubable
self.assertEqual(before, after2)
|
from oslotest import base
from oslotest import moxstubout
class TestMoxStubout(base.BaseTestCase):
def _stubable(self):
pass
def test_basic_stubout(self):
f = self.useFixture(moxstubout.MoxStubout())
before = TestMoxStubout._stubable
f.mox.StubOutWithMock(TestMoxStubout, '_stubable')
after = TestMoxStubout._stubable
self.assertNotEqual(before, after)
f.cleanUp()
after2 = TestMoxStubout._stubable
self.assertEqual(before, after2)
f._clear_cleanups()
|
Fix build break with Fixtures 1.3
|
Fix build break with Fixtures 1.3
Our explicit call to cleanUp messes things up in latest
fixture, so we need to call _clear_cleanups to stop
the test from breaking
Change-Id: I8ce2309a94736b47fb347f37ab4027857e19c8a8
|
Python
|
apache-2.0
|
openstack/oslotest,openstack/oslotest
|
from oslotest import base
from oslotest import moxstubout
class TestMoxStubout(base.BaseTestCase):
def _stubable(self):
pass
def test_basic_stubout(self):
f = self.useFixture(moxstubout.MoxStubout())
before = TestMoxStubout._stubable
f.mox.StubOutWithMock(TestMoxStubout, '_stubable')
after = TestMoxStubout._stubable
self.assertNotEqual(before, after)
f.cleanUp()
after2 = TestMoxStubout._stubable
self.assertEqual(before, after2)
+ f._clear_cleanups()
|
Fix build break with Fixtures 1.3
|
## Code Before:
from oslotest import base
from oslotest import moxstubout
class TestMoxStubout(base.BaseTestCase):
def _stubable(self):
pass
def test_basic_stubout(self):
f = self.useFixture(moxstubout.MoxStubout())
before = TestMoxStubout._stubable
f.mox.StubOutWithMock(TestMoxStubout, '_stubable')
after = TestMoxStubout._stubable
self.assertNotEqual(before, after)
f.cleanUp()
after2 = TestMoxStubout._stubable
self.assertEqual(before, after2)
## Instruction:
Fix build break with Fixtures 1.3
## Code After:
from oslotest import base
from oslotest import moxstubout
class TestMoxStubout(base.BaseTestCase):
def _stubable(self):
pass
def test_basic_stubout(self):
f = self.useFixture(moxstubout.MoxStubout())
before = TestMoxStubout._stubable
f.mox.StubOutWithMock(TestMoxStubout, '_stubable')
after = TestMoxStubout._stubable
self.assertNotEqual(before, after)
f.cleanUp()
after2 = TestMoxStubout._stubable
self.assertEqual(before, after2)
f._clear_cleanups()
|
from oslotest import base
from oslotest import moxstubout
class TestMoxStubout(base.BaseTestCase):
def _stubable(self):
pass
def test_basic_stubout(self):
f = self.useFixture(moxstubout.MoxStubout())
before = TestMoxStubout._stubable
f.mox.StubOutWithMock(TestMoxStubout, '_stubable')
after = TestMoxStubout._stubable
self.assertNotEqual(before, after)
f.cleanUp()
after2 = TestMoxStubout._stubable
self.assertEqual(before, after2)
+ f._clear_cleanups()
|
1f958dc4439fbe435b1d0381d15860708f1f9745
|
constance/__init__.py
|
constance/__init__.py
|
from .base import Config
__version__ = '1.0a1'
try:
from django.apps import AppConfig # noqa
except ImportError:
config = Config()
else:
default_app_config = 'constance.apps.ConstanceConfig'
|
from .base import Config
from django.utils.functional import SimpleLazyObject
__version__ = '1.0a1'
try:
from django.apps import AppConfig # noqa
except ImportError:
config = SimpleLazyObject(Config)
else:
default_app_config = 'constance.apps.ConstanceConfig'
|
Make the config object lazy for old Djangos.
|
Make the config object lazy for old Djangos.
This should prevent import time side effects from instantiating the config object directly there.
|
Python
|
bsd-3-clause
|
gmflanagan/waterboy,vinnyrose/django-constance,jerzyk/django-constance,metalpriest/django-constance,thorgate/django-constance,jmerdich/django-constance,jazzband/django-constance,jonzlin95/django-constance,winzard/django-constance,metalpriest/django-constance,jazzband/django-constance,django-leonardo/django-constance,pombredanne/django-constance,APSL/django-constance,michaelkuty/django-constance,dmugtasimov/django-constance,thorgate/django-constance,Andrey86/django-constance,askabelin/django-constance,pombredanne/django-constance,jezdez/django-constance,Andrey86/django-constance,michaelkuty/django-constance,APSL/django-constance,jezdez/django-constance,jerzyk/django-constance,askabelin/django-constance,django-leonardo/django-constance,jmerdich/django-constance,jonzlin95/django-constance,jazzband/django-constance,dmugtasimov/django-constance,winzard/django-constance,vinnyrose/django-constance
|
from .base import Config
+ from django.utils.functional import SimpleLazyObject
__version__ = '1.0a1'
+
try:
from django.apps import AppConfig # noqa
except ImportError:
- config = Config()
+ config = SimpleLazyObject(Config)
else:
default_app_config = 'constance.apps.ConstanceConfig'
|
Make the config object lazy for old Djangos.
|
## Code Before:
from .base import Config
__version__ = '1.0a1'
try:
from django.apps import AppConfig # noqa
except ImportError:
config = Config()
else:
default_app_config = 'constance.apps.ConstanceConfig'
## Instruction:
Make the config object lazy for old Djangos.
## Code After:
from .base import Config
from django.utils.functional import SimpleLazyObject
__version__ = '1.0a1'
try:
from django.apps import AppConfig # noqa
except ImportError:
config = SimpleLazyObject(Config)
else:
default_app_config = 'constance.apps.ConstanceConfig'
|
from .base import Config
+ from django.utils.functional import SimpleLazyObject
__version__ = '1.0a1'
+
try:
from django.apps import AppConfig # noqa
except ImportError:
- config = Config()
+ config = SimpleLazyObject(Config)
else:
default_app_config = 'constance.apps.ConstanceConfig'
|
ca74738e9241230fd0cc843aa9b76f67494d02eb
|
python/intermediate/create_inter_python_data.py
|
python/intermediate/create_inter_python_data.py
|
"""Create the data for the Software Carpentry Intermediate Python lectures"""
import numpy as np
import pandas as pd
np.random.seed(26)
years = np.arange(1960, 2011)
temps = np.random.uniform(70, 90, len(years))
rainfalls = np.random.uniform(100, 300, len(years))
noise = 2 * np.random.randn(len(years))
mosquitos = 0.5 * temps + 0.7 * rainfalls + noise
data = zip(years, temps, rainfalls, mosquitos)
df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f')
df_short = df[-10:]
df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f')
|
"""Create the data for the Software Carpentry Intermediate Python lectures"""
import numpy as np
import pandas as pd
np.random.seed(26)
datasets = {'A1': [0, 0.5, 0.7, 10],
'A2': [0, 0.5, 0.7, 50],
'A3': [0, 0.5, 0.3, 50],
'B1': [3, 0.7, 0.2, 50],
'B2': [3, 0.7, 0.7, 50]}
def make_data(intercept, tempslope, rainfallslope, numyears):
years = np.arange(2010 - numyears, 2011)
temps = np.random.uniform(70, 90, len(years))
rainfalls = np.random.uniform(100, 300, len(years))
noise = 2 * np.random.randn(len(years))
mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise
return zip(years, temps, rainfalls, mosquitos)
def export_data(data, filename):
df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
df.to_csv(filename, index=False, float_format='%.0f')
for site in datasets:
data = make_data(*datasets[site])
if site == 'A1':
#create a shorter dataset for first example
data = data[-10:]
export_data(data, '%s_mosquito_data.csv' % site)
|
Allow creation of multiple example data files for Inter Python
|
Allow creation of multiple example data files for Inter Python
Generalizes the script for creating data files to allow for the
easy generation of larger numbers of data files.
|
Python
|
bsd-2-clause
|
selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest
|
"""Create the data for the Software Carpentry Intermediate Python lectures"""
import numpy as np
import pandas as pd
np.random.seed(26)
- years = np.arange(1960, 2011)
- temps = np.random.uniform(70, 90, len(years))
- rainfalls = np.random.uniform(100, 300, len(years))
- noise = 2 * np.random.randn(len(years))
- mosquitos = 0.5 * temps + 0.7 * rainfalls + noise
+ datasets = {'A1': [0, 0.5, 0.7, 10],
+ 'A2': [0, 0.5, 0.7, 50],
+ 'A3': [0, 0.5, 0.3, 50],
+ 'B1': [3, 0.7, 0.2, 50],
+ 'B2': [3, 0.7, 0.7, 50]}
+ def make_data(intercept, tempslope, rainfallslope, numyears):
+ years = np.arange(2010 - numyears, 2011)
+ temps = np.random.uniform(70, 90, len(years))
+ rainfalls = np.random.uniform(100, 300, len(years))
+ noise = 2 * np.random.randn(len(years))
+ mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise
- data = zip(years, temps, rainfalls, mosquitos)
+ return zip(years, temps, rainfalls, mosquitos)
- df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
- df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f')
- df_short = df[-10:]
- df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f')
+ def export_data(data, filename):
+ df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
+ df.to_csv(filename, index=False, float_format='%.0f')
+
+ for site in datasets:
+ data = make_data(*datasets[site])
+ if site == 'A1':
+ #create a shorter dataset for first example
+ data = data[-10:]
+ export_data(data, '%s_mosquito_data.csv' % site)
+
|
Allow creation of multiple example data files for Inter Python
|
## Code Before:
"""Create the data for the Software Carpentry Intermediate Python lectures"""
import numpy as np
import pandas as pd
np.random.seed(26)
years = np.arange(1960, 2011)
temps = np.random.uniform(70, 90, len(years))
rainfalls = np.random.uniform(100, 300, len(years))
noise = 2 * np.random.randn(len(years))
mosquitos = 0.5 * temps + 0.7 * rainfalls + noise
data = zip(years, temps, rainfalls, mosquitos)
df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f')
df_short = df[-10:]
df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f')
## Instruction:
Allow creation of multiple example data files for Inter Python
## Code After:
"""Create the data for the Software Carpentry Intermediate Python lectures"""
import numpy as np
import pandas as pd
np.random.seed(26)
datasets = {'A1': [0, 0.5, 0.7, 10],
'A2': [0, 0.5, 0.7, 50],
'A3': [0, 0.5, 0.3, 50],
'B1': [3, 0.7, 0.2, 50],
'B2': [3, 0.7, 0.7, 50]}
def make_data(intercept, tempslope, rainfallslope, numyears):
years = np.arange(2010 - numyears, 2011)
temps = np.random.uniform(70, 90, len(years))
rainfalls = np.random.uniform(100, 300, len(years))
noise = 2 * np.random.randn(len(years))
mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise
return zip(years, temps, rainfalls, mosquitos)
def export_data(data, filename):
df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
df.to_csv(filename, index=False, float_format='%.0f')
for site in datasets:
data = make_data(*datasets[site])
if site == 'A1':
#create a shorter dataset for first example
data = data[-10:]
export_data(data, '%s_mosquito_data.csv' % site)
|
"""Create the data for the Software Carpentry Intermediate Python lectures"""
import numpy as np
import pandas as pd
np.random.seed(26)
- years = np.arange(1960, 2011)
- temps = np.random.uniform(70, 90, len(years))
- rainfalls = np.random.uniform(100, 300, len(years))
- noise = 2 * np.random.randn(len(years))
- mosquitos = 0.5 * temps + 0.7 * rainfalls + noise
+ datasets = {'A1': [0, 0.5, 0.7, 10],
+ 'A2': [0, 0.5, 0.7, 50],
+ 'A3': [0, 0.5, 0.3, 50],
+ 'B1': [3, 0.7, 0.2, 50],
+ 'B2': [3, 0.7, 0.7, 50]}
+ def make_data(intercept, tempslope, rainfallslope, numyears):
+ years = np.arange(2010 - numyears, 2011)
+ temps = np.random.uniform(70, 90, len(years))
+ rainfalls = np.random.uniform(100, 300, len(years))
+ noise = 2 * np.random.randn(len(years))
+ mosquitos = intercept + tempslope * temps + rainfallslope * rainfalls + noise
- data = zip(years, temps, rainfalls, mosquitos)
? ^^ ^^^
+ return zip(years, temps, rainfalls, mosquitos)
? ^^^^^^ ^^^
+
+ def export_data(data, filename):
- df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
+ df = pd.DataFrame(data, columns=['year', 'temperature', 'rainfall','mosquitos'])
? ++++
- df.to_csv('mosquito_data_A2.csv', index=False, float_format='%.0f')
- df_short = df[-10:]
- df_short.to_csv('mosquito_data_A1.csv', index=False, float_format='%.0f')
+ df.to_csv(filename, index=False, float_format='%.0f')
+
+ for site in datasets:
+ data = make_data(*datasets[site])
+ if site == 'A1':
+ #create a shorter dataset for first example
+ data = data[-10:]
+ export_data(data, '%s_mosquito_data.csv' % site)
|
e3a95d00444fb981d7aaf6d3beffca8796a8891f
|
mycroft/frontends/tts/mimic_tts.py
|
mycroft/frontends/tts/mimic_tts.py
|
from subprocess import call
from mycroft.frontends.tts.tts_plugin import TtsPlugin
class MimicTts(TtsPlugin):
def read(self, text):
call(['mimic', '-t', text, '-voice', self.config['voice']])
|
from subprocess import call
from os.path import isdir
from mycroft.frontends.tts.tts_plugin import TtsPlugin
from mycroft.util.git_repo import GitRepo
class MimicTts(TtsPlugin):
def __init__(self, rt):
super().__init__(rt)
if not isdir(self.rt.paths.mimic_exe):
self.download_mimic()
def download_mimic(self):
repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master')
repo.try_pull()
repo.run_inside('./dependencies.sh --prefix="/usr/local"')
repo.run_inside('./autogen.sh')
repo.run_inside('./configure.sh --prefix="/usr/local"')
repo.run_inside('make -j2')
def read(self, text):
call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']])
|
Add download and compile step to mimic
|
Add download and compile step to mimic
|
Python
|
apache-2.0
|
MatthewScholefield/mycroft-simple,MatthewScholefield/mycroft-simple
|
from subprocess import call
+ from os.path import isdir
+
from mycroft.frontends.tts.tts_plugin import TtsPlugin
+ from mycroft.util.git_repo import GitRepo
class MimicTts(TtsPlugin):
+ def __init__(self, rt):
+ super().__init__(rt)
+
+ if not isdir(self.rt.paths.mimic_exe):
+ self.download_mimic()
+
+ def download_mimic(self):
+ repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master')
+ repo.try_pull()
+ repo.run_inside('./dependencies.sh --prefix="/usr/local"')
+ repo.run_inside('./autogen.sh')
+ repo.run_inside('./configure.sh --prefix="/usr/local"')
+ repo.run_inside('make -j2')
+
def read(self, text):
- call(['mimic', '-t', text, '-voice', self.config['voice']])
+ call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']])
|
Add download and compile step to mimic
|
## Code Before:
from subprocess import call
from mycroft.frontends.tts.tts_plugin import TtsPlugin
class MimicTts(TtsPlugin):
def read(self, text):
call(['mimic', '-t', text, '-voice', self.config['voice']])
## Instruction:
Add download and compile step to mimic
## Code After:
from subprocess import call
from os.path import isdir
from mycroft.frontends.tts.tts_plugin import TtsPlugin
from mycroft.util.git_repo import GitRepo
class MimicTts(TtsPlugin):
def __init__(self, rt):
super().__init__(rt)
if not isdir(self.rt.paths.mimic_exe):
self.download_mimic()
def download_mimic(self):
repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master')
repo.try_pull()
repo.run_inside('./dependencies.sh --prefix="/usr/local"')
repo.run_inside('./autogen.sh')
repo.run_inside('./configure.sh --prefix="/usr/local"')
repo.run_inside('make -j2')
def read(self, text):
call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']])
|
from subprocess import call
+ from os.path import isdir
+
from mycroft.frontends.tts.tts_plugin import TtsPlugin
+ from mycroft.util.git_repo import GitRepo
class MimicTts(TtsPlugin):
+ def __init__(self, rt):
+ super().__init__(rt)
+
+ if not isdir(self.rt.paths.mimic_exe):
+ self.download_mimic()
+
+ def download_mimic(self):
+ repo = GitRepo(self.rt.paths.mimic, self.config['url'], 'master')
+ repo.try_pull()
+ repo.run_inside('./dependencies.sh --prefix="/usr/local"')
+ repo.run_inside('./autogen.sh')
+ repo.run_inside('./configure.sh --prefix="/usr/local"')
+ repo.run_inside('make -j2')
+
def read(self, text):
- call(['mimic', '-t', text, '-voice', self.config['voice']])
? ^ ^
+ call([self.rt.paths.mimic_exe, '-t', text, '-voice', self.config['voice']])
? ^^^^^^^^^^^^^^ ^^^^
|
721f837cbfa0de8804def607908a9744b0d099a8
|
asl/vendor/__init__.py
|
asl/vendor/__init__.py
|
import sys
import os
_vendor_initialized = False
def append_paths(path, vendor_modules):
new_path = []
for v in vendor_modules:
new_path.append(path + os.sep + v)
sys.path = new_path + sys.path
def do_init():
global _vendor_initialized
if _vendor_initialized:
return
_vendor_initialized = True
path = "."
if os.path.exists('./asl/asl/vendor'):
path = os.path.abspath('./asl/asl/vendor')
else:
for p in sys.path:
if os.path.exists(p + '/vendor/'):
path = os.path.abspath(p + '/vendor/')
break
if os.path.exists(p + '/asl/vendor/'):
path = os.path.abspath(p + '/asl/vendor/')
break
vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson']
append_paths(path, vendor_modules)
do_init()
|
import sys
import os
_vendor_initialized = False
def append_paths(path, vendor_modules):
new_path = []
for v in vendor_modules:
new_path.append(path + os.sep + v)
sys.path = new_path + sys.path
def do_init():
global _vendor_initialized
if _vendor_initialized:
return
_vendor_initialized = True
path = os.path.dirname(__file__)
vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson']
append_paths(path, vendor_modules)
do_init()
|
Fix of vendor directory search.
|
Fix of vendor directory search.
|
Python
|
mit
|
AtteqCom/zsl,AtteqCom/zsl
|
import sys
import os
_vendor_initialized = False
def append_paths(path, vendor_modules):
new_path = []
for v in vendor_modules:
new_path.append(path + os.sep + v)
sys.path = new_path + sys.path
def do_init():
global _vendor_initialized
if _vendor_initialized:
return
_vendor_initialized = True
+ path = os.path.dirname(__file__)
- path = "."
- if os.path.exists('./asl/asl/vendor'):
- path = os.path.abspath('./asl/asl/vendor')
- else:
- for p in sys.path:
- if os.path.exists(p + '/vendor/'):
- path = os.path.abspath(p + '/vendor/')
- break
- if os.path.exists(p + '/asl/vendor/'):
- path = os.path.abspath(p + '/asl/vendor/')
- break
-
vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson']
append_paths(path, vendor_modules)
do_init()
|
Fix of vendor directory search.
|
## Code Before:
import sys
import os
_vendor_initialized = False
def append_paths(path, vendor_modules):
new_path = []
for v in vendor_modules:
new_path.append(path + os.sep + v)
sys.path = new_path + sys.path
def do_init():
global _vendor_initialized
if _vendor_initialized:
return
_vendor_initialized = True
path = "."
if os.path.exists('./asl/asl/vendor'):
path = os.path.abspath('./asl/asl/vendor')
else:
for p in sys.path:
if os.path.exists(p + '/vendor/'):
path = os.path.abspath(p + '/vendor/')
break
if os.path.exists(p + '/asl/vendor/'):
path = os.path.abspath(p + '/asl/vendor/')
break
vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson']
append_paths(path, vendor_modules)
do_init()
## Instruction:
Fix of vendor directory search.
## Code After:
import sys
import os
_vendor_initialized = False
def append_paths(path, vendor_modules):
new_path = []
for v in vendor_modules:
new_path.append(path + os.sep + v)
sys.path = new_path + sys.path
def do_init():
global _vendor_initialized
if _vendor_initialized:
return
_vendor_initialized = True
path = os.path.dirname(__file__)
vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson']
append_paths(path, vendor_modules)
do_init()
|
import sys
import os
_vendor_initialized = False
def append_paths(path, vendor_modules):
new_path = []
for v in vendor_modules:
new_path.append(path + os.sep + v)
sys.path = new_path + sys.path
def do_init():
global _vendor_initialized
if _vendor_initialized:
return
_vendor_initialized = True
+ path = os.path.dirname(__file__)
- path = "."
- if os.path.exists('./asl/asl/vendor'):
- path = os.path.abspath('./asl/asl/vendor')
- else:
- for p in sys.path:
- if os.path.exists(p + '/vendor/'):
- path = os.path.abspath(p + '/vendor/')
- break
- if os.path.exists(p + '/asl/vendor/'):
- path = os.path.abspath(p + '/asl/vendor/')
- break
-
vendor_modules = ['injector', 'flask_injector', 'redis-py', 'sqlalchemy/sqlalchemy-0_9_1/lib', 'bpython/bpython', 'sphinxapi', 'simplejson']
append_paths(path, vendor_modules)
do_init()
|
bd70ef56d95958b8f105bdff31b675d66c40bca8
|
serfnode/handler/supervisor.py
|
serfnode/handler/supervisor.py
|
import os
import subprocess
import docker_utils
import jinja2
env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs'))
def supervisor_install(block, **kwargs):
"""Update supervisor with `block` config.
- `block` is the name to a .conf template file (in directory
`/programs`)
- `kwargs` are the key/values to use in the template
"""
conf_filename = '{}.conf'.format(kwargs['target'])
template = env.get_template(block)
kwargs.update({
'DOCKER': docker_utils.DOCKER,
'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET,
'DOCKER_RUN': docker_utils.DOCKER_RUN})
conf = template.render(kwargs)
with open(os.path.join(
'/etc/supervisor/conf.d', conf_filename), 'w') as f:
f.write(conf)
def supervisor_exec(*args):
return subprocess.check_output(
['supervisorctl'] + list(args))
def supervisor_update():
supervisor_exec('reread')
supervisor_exec('update')
def start(block, **kwargs):
supervisor_install(block, **kwargs)
supervisor_update()
supervisor_exec('start', '{}:*'.format(kwargs['target']))
def stop(block):
supervisor_exec('stop', '{}:*'.format(block))
|
import os
import subprocess
import docker_utils
import docker
import jinja2
env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs'))
def supervisor_install(block, **kwargs):
"""Update supervisor with `block` config.
- `block` is the name to a .conf template file (in directory
`/programs`)
- `kwargs` are the key/values to use in the template
"""
conf_filename = '{}.conf'.format(kwargs['target'])
template = env.get_template(block)
kwargs.update({
'DOCKER': docker_utils.DOCKER,
'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET,
'DOCKER_RUN': docker_utils.DOCKER_RUN})
conf = template.render(kwargs)
with open(os.path.join(
'/etc/supervisor/conf.d', conf_filename), 'w') as f:
f.write(conf)
def supervisor_exec(*args):
return subprocess.check_output(
['supervisorctl'] + list(args))
def supervisor_update():
supervisor_exec('reread')
supervisor_exec('update')
def start(block, **kwargs):
supervisor_install(block, **kwargs)
supervisor_update()
supervisor_exec('start', '{}:*'.format(kwargs['target']))
def start_docker(target, name, cmdline):
start('app.conf', target=target,
ARGS='--cidfile=/app --name={} {}'.format(name, cmdline),
NAME=name)
def stop(block):
supervisor_exec('stop', '{}:*'.format(block))
|
Add convenience function to start docker
|
Add convenience function to start docker
Mainly to be used from supervisor.
|
Python
|
mit
|
waltermoreira/serfnode,waltermoreira/serfnode,waltermoreira/serfnode
|
import os
import subprocess
import docker_utils
+ import docker
import jinja2
env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs'))
def supervisor_install(block, **kwargs):
"""Update supervisor with `block` config.
- `block` is the name to a .conf template file (in directory
`/programs`)
- `kwargs` are the key/values to use in the template
"""
conf_filename = '{}.conf'.format(kwargs['target'])
template = env.get_template(block)
kwargs.update({
'DOCKER': docker_utils.DOCKER,
'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET,
'DOCKER_RUN': docker_utils.DOCKER_RUN})
conf = template.render(kwargs)
with open(os.path.join(
'/etc/supervisor/conf.d', conf_filename), 'w') as f:
f.write(conf)
def supervisor_exec(*args):
return subprocess.check_output(
['supervisorctl'] + list(args))
def supervisor_update():
supervisor_exec('reread')
supervisor_exec('update')
def start(block, **kwargs):
supervisor_install(block, **kwargs)
supervisor_update()
supervisor_exec('start', '{}:*'.format(kwargs['target']))
+ def start_docker(target, name, cmdline):
+ start('app.conf', target=target,
+ ARGS='--cidfile=/app --name={} {}'.format(name, cmdline),
+ NAME=name)
+
+
def stop(block):
supervisor_exec('stop', '{}:*'.format(block))
+
|
Add convenience function to start docker
|
## Code Before:
import os
import subprocess
import docker_utils
import jinja2
env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs'))
def supervisor_install(block, **kwargs):
"""Update supervisor with `block` config.
- `block` is the name to a .conf template file (in directory
`/programs`)
- `kwargs` are the key/values to use in the template
"""
conf_filename = '{}.conf'.format(kwargs['target'])
template = env.get_template(block)
kwargs.update({
'DOCKER': docker_utils.DOCKER,
'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET,
'DOCKER_RUN': docker_utils.DOCKER_RUN})
conf = template.render(kwargs)
with open(os.path.join(
'/etc/supervisor/conf.d', conf_filename), 'w') as f:
f.write(conf)
def supervisor_exec(*args):
return subprocess.check_output(
['supervisorctl'] + list(args))
def supervisor_update():
supervisor_exec('reread')
supervisor_exec('update')
def start(block, **kwargs):
supervisor_install(block, **kwargs)
supervisor_update()
supervisor_exec('start', '{}:*'.format(kwargs['target']))
def stop(block):
supervisor_exec('stop', '{}:*'.format(block))
## Instruction:
Add convenience function to start docker
## Code After:
import os
import subprocess
import docker_utils
import docker
import jinja2
env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs'))
def supervisor_install(block, **kwargs):
"""Update supervisor with `block` config.
- `block` is the name to a .conf template file (in directory
`/programs`)
- `kwargs` are the key/values to use in the template
"""
conf_filename = '{}.conf'.format(kwargs['target'])
template = env.get_template(block)
kwargs.update({
'DOCKER': docker_utils.DOCKER,
'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET,
'DOCKER_RUN': docker_utils.DOCKER_RUN})
conf = template.render(kwargs)
with open(os.path.join(
'/etc/supervisor/conf.d', conf_filename), 'w') as f:
f.write(conf)
def supervisor_exec(*args):
return subprocess.check_output(
['supervisorctl'] + list(args))
def supervisor_update():
supervisor_exec('reread')
supervisor_exec('update')
def start(block, **kwargs):
supervisor_install(block, **kwargs)
supervisor_update()
supervisor_exec('start', '{}:*'.format(kwargs['target']))
def start_docker(target, name, cmdline):
start('app.conf', target=target,
ARGS='--cidfile=/app --name={} {}'.format(name, cmdline),
NAME=name)
def stop(block):
supervisor_exec('stop', '{}:*'.format(block))
|
import os
import subprocess
import docker_utils
+ import docker
import jinja2
env = jinja2.Environment(loader=jinja2.FileSystemLoader('/programs'))
def supervisor_install(block, **kwargs):
"""Update supervisor with `block` config.
- `block` is the name to a .conf template file (in directory
`/programs`)
- `kwargs` are the key/values to use in the template
"""
conf_filename = '{}.conf'.format(kwargs['target'])
template = env.get_template(block)
kwargs.update({
'DOCKER': docker_utils.DOCKER,
'DOCKER_SOCKET': docker_utils.DOCKER_SOCKET,
'DOCKER_RUN': docker_utils.DOCKER_RUN})
conf = template.render(kwargs)
with open(os.path.join(
'/etc/supervisor/conf.d', conf_filename), 'w') as f:
f.write(conf)
def supervisor_exec(*args):
return subprocess.check_output(
['supervisorctl'] + list(args))
def supervisor_update():
supervisor_exec('reread')
supervisor_exec('update')
def start(block, **kwargs):
supervisor_install(block, **kwargs)
supervisor_update()
supervisor_exec('start', '{}:*'.format(kwargs['target']))
+ def start_docker(target, name, cmdline):
+ start('app.conf', target=target,
+ ARGS='--cidfile=/app --name={} {}'.format(name, cmdline),
+ NAME=name)
+
+
def stop(block):
supervisor_exec('stop', '{}:*'.format(block))
+
|
fce10cb35be29ba265f2ed189198703c718ad479
|
quantecon/__init__.py
|
quantecon/__init__.py
|
from . import models as models
from .compute_fp import compute_fixed_point
from .discrete_rv import DiscreteRV
from .ecdf import ECDF
from .estspec import smooth, periodogram, ar_periodogram
from .graph_tools import DiGraph
from .gridtools import cartesian, mlinspace
from .gth_solve import gth_solve
from .kalman import Kalman
from .lae import LAE
from .arma import ARMA
from .lqcontrol import LQ
from .lqnash import nnash
from .lss import LinearStateSpace
from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati
from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path
from .quadsums import var_quadratic_sum, m_quadratic_sum
from .markov import random_markov_chain, random_stochastic_matrix
from .rank_nullspace import rank_est, nullspace
from .robustlq import RBLQ
from .tauchen import approx_markov
from . import quad as quad
from .util import searchsorted, random_probvec, random_sample_without_replacement
#-Module Imports-#
import util.random as random
#Add Version Attribute
from .version import version as __version__
|
from . import models as models
from .compute_fp import compute_fixed_point
from .discrete_rv import DiscreteRV
from .ecdf import ECDF
from .estspec import smooth, periodogram, ar_periodogram
from .graph_tools import DiGraph
from .gridtools import cartesian, mlinspace
from .gth_solve import gth_solve
from .kalman import Kalman
from .lae import LAE
from .arma import ARMA
from .lqcontrol import LQ
from .lqnash import nnash
from .lss import LinearStateSpace
from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati
from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path
from .quadsums import var_quadratic_sum, m_quadratic_sum
from .markov import random_markov_chain, random_stochastic_matrix
from .rank_nullspace import rank_est, nullspace
from .robustlq import RBLQ
from .tauchen import approx_markov
from . import quad as quad
from .util import searchsorted, random_probvec, random_sample_without_replacement
#-Module Imports-#
from .util import random
#Add Version Attribute
from .version import version as __version__
|
Fix for python 3 relative import statement
|
Fix for python 3 relative import statement
|
Python
|
bsd-3-clause
|
andybrnr/QuantEcon.py,oyamad/QuantEcon.py,agutieda/QuantEcon.py,jviada/QuantEcon.py,QuantEcon/QuantEcon.py,QuantEcon/QuantEcon.py,gxxjjj/QuantEcon.py,agutieda/QuantEcon.py,gxxjjj/QuantEcon.py,oyamad/QuantEcon.py,jviada/QuantEcon.py,andybrnr/QuantEcon.py
|
from . import models as models
from .compute_fp import compute_fixed_point
from .discrete_rv import DiscreteRV
from .ecdf import ECDF
from .estspec import smooth, periodogram, ar_periodogram
from .graph_tools import DiGraph
from .gridtools import cartesian, mlinspace
from .gth_solve import gth_solve
from .kalman import Kalman
from .lae import LAE
from .arma import ARMA
from .lqcontrol import LQ
from .lqnash import nnash
from .lss import LinearStateSpace
from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati
from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path
from .quadsums import var_quadratic_sum, m_quadratic_sum
from .markov import random_markov_chain, random_stochastic_matrix
from .rank_nullspace import rank_est, nullspace
from .robustlq import RBLQ
from .tauchen import approx_markov
from . import quad as quad
from .util import searchsorted, random_probvec, random_sample_without_replacement
#-Module Imports-#
- import util.random as random
+ from .util import random
#Add Version Attribute
from .version import version as __version__
|
Fix for python 3 relative import statement
|
## Code Before:
from . import models as models
from .compute_fp import compute_fixed_point
from .discrete_rv import DiscreteRV
from .ecdf import ECDF
from .estspec import smooth, periodogram, ar_periodogram
from .graph_tools import DiGraph
from .gridtools import cartesian, mlinspace
from .gth_solve import gth_solve
from .kalman import Kalman
from .lae import LAE
from .arma import ARMA
from .lqcontrol import LQ
from .lqnash import nnash
from .lss import LinearStateSpace
from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati
from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path
from .quadsums import var_quadratic_sum, m_quadratic_sum
from .markov import random_markov_chain, random_stochastic_matrix
from .rank_nullspace import rank_est, nullspace
from .robustlq import RBLQ
from .tauchen import approx_markov
from . import quad as quad
from .util import searchsorted, random_probvec, random_sample_without_replacement
#-Module Imports-#
import util.random as random
#Add Version Attribute
from .version import version as __version__
## Instruction:
Fix for python 3 relative import statement
## Code After:
from . import models as models
from .compute_fp import compute_fixed_point
from .discrete_rv import DiscreteRV
from .ecdf import ECDF
from .estspec import smooth, periodogram, ar_periodogram
from .graph_tools import DiGraph
from .gridtools import cartesian, mlinspace
from .gth_solve import gth_solve
from .kalman import Kalman
from .lae import LAE
from .arma import ARMA
from .lqcontrol import LQ
from .lqnash import nnash
from .lss import LinearStateSpace
from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati
from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path
from .quadsums import var_quadratic_sum, m_quadratic_sum
from .markov import random_markov_chain, random_stochastic_matrix
from .rank_nullspace import rank_est, nullspace
from .robustlq import RBLQ
from .tauchen import approx_markov
from . import quad as quad
from .util import searchsorted, random_probvec, random_sample_without_replacement
#-Module Imports-#
from .util import random
#Add Version Attribute
from .version import version as __version__
|
from . import models as models
from .compute_fp import compute_fixed_point
from .discrete_rv import DiscreteRV
from .ecdf import ECDF
from .estspec import smooth, periodogram, ar_periodogram
from .graph_tools import DiGraph
from .gridtools import cartesian, mlinspace
from .gth_solve import gth_solve
from .kalman import Kalman
from .lae import LAE
from .arma import ARMA
from .lqcontrol import LQ
from .lqnash import nnash
from .lss import LinearStateSpace
from .matrix_eqn import solve_discrete_lyapunov, solve_discrete_riccati
from .mc_tools import MarkovChain, mc_compute_stationary, mc_sample_path
from .quadsums import var_quadratic_sum, m_quadratic_sum
from .markov import random_markov_chain, random_stochastic_matrix
from .rank_nullspace import rank_est, nullspace
from .robustlq import RBLQ
from .tauchen import approx_markov
from . import quad as quad
from .util import searchsorted, random_probvec, random_sample_without_replacement
#-Module Imports-#
- import util.random as random
+ from .util import random
#Add Version Attribute
from .version import version as __version__
|
898028dea2e04d52c32854752bda34d331c7696f
|
ynr/apps/candidatebot/management/commands/candidatebot_import_email_from_csv.py
|
ynr/apps/candidatebot/management/commands/candidatebot_import_email_from_csv.py
|
from __future__ import unicode_literals
import csv
from django.core.management.base import BaseCommand
from candidatebot.helpers import CandidateBot
from popolo.models import Person
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument(
'filename',
help='Path to the file with the email addresses'
)
parser.add_argument(
'--source',
help='Source of the data. The source CSV column takes precedence'
)
def handle(self, **options):
with open(options['filename'], 'r') as fh:
reader = csv.DictReader(fh)
for row in reader:
source = row.get('source', options.get('source'))
if not row['democlub_id']:
continue
if not source:
raise ValueError("A source is required")
try:
bot = CandidateBot(row['democlub_id'])
bot.add_email(row['email'])
bot.save(source)
# print(person)
except Person.DoesNotExist:
print("Person ID {} not found".format(
row['democlub_id']))
# print(row)
|
from __future__ import unicode_literals
import csv
from django.core.management.base import BaseCommand
from candidatebot.helpers import CandidateBot
from popolo.models import Person
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument(
'filename',
help='Path to the file with the email addresses'
)
parser.add_argument(
'--source',
help='Source of the data. The source CSV column takes precedence'
)
def handle(self, **options):
with open(options['filename'], 'r') as fh:
reader = csv.DictReader(fh)
for row in reader:
source = row.get('source', options.get('source'))
if not row['democlub_id']:
continue
if not source:
raise ValueError("A source is required")
try:
bot = CandidateBot(row['democlub_id'])
try:
bot.add_email(row['email'])
bot.save(source)
except ValueError:
#Email exists, move on
pass
except Person.DoesNotExist:
print("Person ID {} not found".format(
row['democlub_id']))
# print(row)
|
Move on if email exists
|
Move on if email exists
|
Python
|
agpl-3.0
|
DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative
|
from __future__ import unicode_literals
import csv
from django.core.management.base import BaseCommand
from candidatebot.helpers import CandidateBot
from popolo.models import Person
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument(
'filename',
help='Path to the file with the email addresses'
)
parser.add_argument(
'--source',
help='Source of the data. The source CSV column takes precedence'
)
def handle(self, **options):
with open(options['filename'], 'r') as fh:
reader = csv.DictReader(fh)
for row in reader:
source = row.get('source', options.get('source'))
if not row['democlub_id']:
continue
if not source:
raise ValueError("A source is required")
try:
bot = CandidateBot(row['democlub_id'])
+ try:
- bot.add_email(row['email'])
+ bot.add_email(row['email'])
- bot.save(source)
+ bot.save(source)
- # print(person)
+ except ValueError:
+ #Email exists, move on
+ pass
except Person.DoesNotExist:
print("Person ID {} not found".format(
row['democlub_id']))
# print(row)
|
Move on if email exists
|
## Code Before:
from __future__ import unicode_literals
import csv
from django.core.management.base import BaseCommand
from candidatebot.helpers import CandidateBot
from popolo.models import Person
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument(
'filename',
help='Path to the file with the email addresses'
)
parser.add_argument(
'--source',
help='Source of the data. The source CSV column takes precedence'
)
def handle(self, **options):
with open(options['filename'], 'r') as fh:
reader = csv.DictReader(fh)
for row in reader:
source = row.get('source', options.get('source'))
if not row['democlub_id']:
continue
if not source:
raise ValueError("A source is required")
try:
bot = CandidateBot(row['democlub_id'])
bot.add_email(row['email'])
bot.save(source)
# print(person)
except Person.DoesNotExist:
print("Person ID {} not found".format(
row['democlub_id']))
# print(row)
## Instruction:
Move on if email exists
## Code After:
from __future__ import unicode_literals
import csv
from django.core.management.base import BaseCommand
from candidatebot.helpers import CandidateBot
from popolo.models import Person
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument(
'filename',
help='Path to the file with the email addresses'
)
parser.add_argument(
'--source',
help='Source of the data. The source CSV column takes precedence'
)
def handle(self, **options):
with open(options['filename'], 'r') as fh:
reader = csv.DictReader(fh)
for row in reader:
source = row.get('source', options.get('source'))
if not row['democlub_id']:
continue
if not source:
raise ValueError("A source is required")
try:
bot = CandidateBot(row['democlub_id'])
try:
bot.add_email(row['email'])
bot.save(source)
except ValueError:
#Email exists, move on
pass
except Person.DoesNotExist:
print("Person ID {} not found".format(
row['democlub_id']))
# print(row)
|
from __future__ import unicode_literals
import csv
from django.core.management.base import BaseCommand
from candidatebot.helpers import CandidateBot
from popolo.models import Person
class Command(BaseCommand):
def add_arguments(self, parser):
parser.add_argument(
'filename',
help='Path to the file with the email addresses'
)
parser.add_argument(
'--source',
help='Source of the data. The source CSV column takes precedence'
)
def handle(self, **options):
with open(options['filename'], 'r') as fh:
reader = csv.DictReader(fh)
for row in reader:
source = row.get('source', options.get('source'))
if not row['democlub_id']:
continue
if not source:
raise ValueError("A source is required")
try:
bot = CandidateBot(row['democlub_id'])
+ try:
- bot.add_email(row['email'])
+ bot.add_email(row['email'])
? ++++
- bot.save(source)
+ bot.save(source)
? ++++
- # print(person)
+ except ValueError:
+ #Email exists, move on
+ pass
except Person.DoesNotExist:
print("Person ID {} not found".format(
row['democlub_id']))
# print(row)
|
6c2354a1e56477eb983b0adbcc2d15223c158184
|
foodsaving/subscriptions/consumers.py
|
foodsaving/subscriptions/consumers.py
|
from channels.auth import channel_session_user_from_http, channel_session_user
from django.utils import timezone
from foodsaving.subscriptions.models import ChannelSubscription
@channel_session_user_from_http
def ws_connect(message):
"""The user has connected! Register their channel subscription."""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel)
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_message(message):
"""They sent us a websocket message! We just update the ChannelSubscription lastseen time.."""
user = message.user
if not user.is_anonymous():
reply_channel = message.reply_channel.name
ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now())
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_disconnect(message):
"""The user has disconnected so we remove all their ChannelSubscriptions"""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
message.reply_channel.send({"accept": True})
|
from channels.auth import channel_session_user_from_http, channel_session_user
from django.utils import timezone
from foodsaving.subscriptions.models import ChannelSubscription
@channel_session_user_from_http
def ws_connect(message):
"""The user has connected! Register their channel subscription."""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel)
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_message(message):
"""They sent us a websocket message! We just update the ChannelSubscription lastseen time.."""
user = message.user
if not user.is_anonymous():
reply_channel = message.reply_channel.name
ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now())
@channel_session_user
def ws_disconnect(message):
"""The user has disconnected so we remove all their ChannelSubscriptions"""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
|
Remove redundant ws accept replies
|
Remove redundant ws accept replies
It's only relevent on connection
|
Python
|
agpl-3.0
|
yunity/yunity-core,yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/foodsaving-backend
|
from channels.auth import channel_session_user_from_http, channel_session_user
from django.utils import timezone
from foodsaving.subscriptions.models import ChannelSubscription
@channel_session_user_from_http
def ws_connect(message):
"""The user has connected! Register their channel subscription."""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel)
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_message(message):
"""They sent us a websocket message! We just update the ChannelSubscription lastseen time.."""
user = message.user
if not user.is_anonymous():
reply_channel = message.reply_channel.name
ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now())
- message.reply_channel.send({"accept": True})
@channel_session_user
def ws_disconnect(message):
"""The user has disconnected so we remove all their ChannelSubscriptions"""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
- message.reply_channel.send({"accept": True})
|
Remove redundant ws accept replies
|
## Code Before:
from channels.auth import channel_session_user_from_http, channel_session_user
from django.utils import timezone
from foodsaving.subscriptions.models import ChannelSubscription
@channel_session_user_from_http
def ws_connect(message):
"""The user has connected! Register their channel subscription."""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel)
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_message(message):
"""They sent us a websocket message! We just update the ChannelSubscription lastseen time.."""
user = message.user
if not user.is_anonymous():
reply_channel = message.reply_channel.name
ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now())
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_disconnect(message):
"""The user has disconnected so we remove all their ChannelSubscriptions"""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
message.reply_channel.send({"accept": True})
## Instruction:
Remove redundant ws accept replies
## Code After:
from channels.auth import channel_session_user_from_http, channel_session_user
from django.utils import timezone
from foodsaving.subscriptions.models import ChannelSubscription
@channel_session_user_from_http
def ws_connect(message):
"""The user has connected! Register their channel subscription."""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel)
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_message(message):
"""They sent us a websocket message! We just update the ChannelSubscription lastseen time.."""
user = message.user
if not user.is_anonymous():
reply_channel = message.reply_channel.name
ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now())
@channel_session_user
def ws_disconnect(message):
"""The user has disconnected so we remove all their ChannelSubscriptions"""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
|
from channels.auth import channel_session_user_from_http, channel_session_user
from django.utils import timezone
from foodsaving.subscriptions.models import ChannelSubscription
@channel_session_user_from_http
def ws_connect(message):
"""The user has connected! Register their channel subscription."""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.create(user=user, reply_channel=message.reply_channel)
message.reply_channel.send({"accept": True})
@channel_session_user
def ws_message(message):
"""They sent us a websocket message! We just update the ChannelSubscription lastseen time.."""
user = message.user
if not user.is_anonymous():
reply_channel = message.reply_channel.name
ChannelSubscription.objects.filter(user=user, reply_channel=reply_channel).update(lastseen_at=timezone.now())
- message.reply_channel.send({"accept": True})
@channel_session_user
def ws_disconnect(message):
"""The user has disconnected so we remove all their ChannelSubscriptions"""
user = message.user
if not user.is_anonymous():
ChannelSubscription.objects.filter(user=user, reply_channel=message.reply_channel).delete()
- message.reply_channel.send({"accept": True})
|
f890663daa329e3f22d0f619ed6acf9365308c7c
|
apps/ignite/views.py
|
apps/ignite/views.py
|
from django.shortcuts import get_object_or_404
import jingo
from challenges.models import Submission, Category
from projects.models import Project
def splash(request, project, slug, template_name='challenges/show.html'):
"""Show an individual project challenge."""
project = get_object_or_404(Project, slug=project)
challenge = get_object_or_404(project.challenge_set, slug=slug)
entries = Submission.objects.filter(
phase__challenge=challenge
).exclude(
is_draft=True
).extra(
order_by="?"
)
return jingo.render(request, 'ignite/splash.html', {
'challenge': challenge,
'project': project,
'phases': list(enumerate(challenge.phases.all(), start=1)),
'entries': entries[:10],
'categories': Category.objects.get_active_categories(),
})
|
from django.shortcuts import get_object_or_404
import jingo
from challenges.models import Submission, Category
from projects.models import Project
def splash(request, project, slug, template_name='challenges/show.html'):
"""Show an individual project challenge."""
project = get_object_or_404(Project, slug=project)
challenge = get_object_or_404(project.challenge_set, slug=slug)
entries = (Submission.objects.visible()
.filter(phase__challenge=challenge)
.order_by("?"))
return jingo.render(request, 'ignite/splash.html', {
'challenge': challenge,
'project': project,
'phases': list(enumerate(challenge.phases.all(), start=1)),
'entries': entries[:10],
'categories': Category.objects.get_active_categories(),
})
|
Update splash view to use visible() method.
|
Update splash view to use visible() method.
|
Python
|
bsd-3-clause
|
mozilla/mozilla-ignite,mozilla/mozilla-ignite,mozilla/mozilla-ignite,mozilla/mozilla-ignite
|
from django.shortcuts import get_object_or_404
import jingo
from challenges.models import Submission, Category
from projects.models import Project
def splash(request, project, slug, template_name='challenges/show.html'):
"""Show an individual project challenge."""
project = get_object_or_404(Project, slug=project)
challenge = get_object_or_404(project.challenge_set, slug=slug)
- entries = Submission.objects.filter(
+ entries = (Submission.objects.visible()
+ .filter(phase__challenge=challenge)
+ .order_by("?"))
- phase__challenge=challenge
- ).exclude(
- is_draft=True
- ).extra(
- order_by="?"
- )
+
return jingo.render(request, 'ignite/splash.html', {
'challenge': challenge,
'project': project,
'phases': list(enumerate(challenge.phases.all(), start=1)),
'entries': entries[:10],
'categories': Category.objects.get_active_categories(),
})
|
Update splash view to use visible() method.
|
## Code Before:
from django.shortcuts import get_object_or_404
import jingo
from challenges.models import Submission, Category
from projects.models import Project
def splash(request, project, slug, template_name='challenges/show.html'):
"""Show an individual project challenge."""
project = get_object_or_404(Project, slug=project)
challenge = get_object_or_404(project.challenge_set, slug=slug)
entries = Submission.objects.filter(
phase__challenge=challenge
).exclude(
is_draft=True
).extra(
order_by="?"
)
return jingo.render(request, 'ignite/splash.html', {
'challenge': challenge,
'project': project,
'phases': list(enumerate(challenge.phases.all(), start=1)),
'entries': entries[:10],
'categories': Category.objects.get_active_categories(),
})
## Instruction:
Update splash view to use visible() method.
## Code After:
from django.shortcuts import get_object_or_404
import jingo
from challenges.models import Submission, Category
from projects.models import Project
def splash(request, project, slug, template_name='challenges/show.html'):
"""Show an individual project challenge."""
project = get_object_or_404(Project, slug=project)
challenge = get_object_or_404(project.challenge_set, slug=slug)
entries = (Submission.objects.visible()
.filter(phase__challenge=challenge)
.order_by("?"))
return jingo.render(request, 'ignite/splash.html', {
'challenge': challenge,
'project': project,
'phases': list(enumerate(challenge.phases.all(), start=1)),
'entries': entries[:10],
'categories': Category.objects.get_active_categories(),
})
|
from django.shortcuts import get_object_or_404
import jingo
from challenges.models import Submission, Category
from projects.models import Project
def splash(request, project, slug, template_name='challenges/show.html'):
"""Show an individual project challenge."""
project = get_object_or_404(Project, slug=project)
challenge = get_object_or_404(project.challenge_set, slug=slug)
- entries = Submission.objects.filter(
? ^ - -
+ entries = (Submission.objects.visible()
? + ^ +++ +
+ .filter(phase__challenge=challenge)
+ .order_by("?"))
- phase__challenge=challenge
- ).exclude(
- is_draft=True
- ).extra(
- order_by="?"
- )
? -
+
return jingo.render(request, 'ignite/splash.html', {
'challenge': challenge,
'project': project,
'phases': list(enumerate(challenge.phases.all(), start=1)),
'entries': entries[:10],
'categories': Category.objects.get_active_categories(),
})
|
15013c51f602786265b59c1d4a7e894eae090d90
|
tests/test_normalize.py
|
tests/test_normalize.py
|
from hypothesis import assume, given
from utils import isclose, vectors
@given(v=vectors())
def test_normalize_length(v):
"""v.normalize().length == 1 and v == v.length * v.normalize()"""
assume(v)
assert isclose(v.normalize().length, 1)
assert v.isclose(v.length * v.normalize())
|
from hypothesis import assume, given
from utils import isclose, vectors
@given(v=vectors())
def test_normalize_length(v):
"""v.normalize().length == 1 and v == v.length * v.normalize()"""
assume(v)
assert isclose(v.normalize().length, 1)
assert v.isclose(v.length * v.normalize())
@given(v=vectors())
def test_normalize_angle(v):
"""Normalization preserves direction."""
assume(v)
assert angle_isclose(v.normalize().angle(v), 0)
|
Test that direction is preserved
|
tests/normalize: Test that direction is preserved
|
Python
|
artistic-2.0
|
ppb/ppb-vector,ppb/ppb-vector
|
from hypothesis import assume, given
from utils import isclose, vectors
@given(v=vectors())
def test_normalize_length(v):
"""v.normalize().length == 1 and v == v.length * v.normalize()"""
assume(v)
assert isclose(v.normalize().length, 1)
assert v.isclose(v.length * v.normalize())
+
+ @given(v=vectors())
+ def test_normalize_angle(v):
+ """Normalization preserves direction."""
+ assume(v)
+ assert angle_isclose(v.normalize().angle(v), 0)
+
|
Test that direction is preserved
|
## Code Before:
from hypothesis import assume, given
from utils import isclose, vectors
@given(v=vectors())
def test_normalize_length(v):
"""v.normalize().length == 1 and v == v.length * v.normalize()"""
assume(v)
assert isclose(v.normalize().length, 1)
assert v.isclose(v.length * v.normalize())
## Instruction:
Test that direction is preserved
## Code After:
from hypothesis import assume, given
from utils import isclose, vectors
@given(v=vectors())
def test_normalize_length(v):
"""v.normalize().length == 1 and v == v.length * v.normalize()"""
assume(v)
assert isclose(v.normalize().length, 1)
assert v.isclose(v.length * v.normalize())
@given(v=vectors())
def test_normalize_angle(v):
"""Normalization preserves direction."""
assume(v)
assert angle_isclose(v.normalize().angle(v), 0)
|
from hypothesis import assume, given
from utils import isclose, vectors
@given(v=vectors())
def test_normalize_length(v):
"""v.normalize().length == 1 and v == v.length * v.normalize()"""
assume(v)
assert isclose(v.normalize().length, 1)
assert v.isclose(v.length * v.normalize())
+
+
+ @given(v=vectors())
+ def test_normalize_angle(v):
+ """Normalization preserves direction."""
+ assume(v)
+ assert angle_isclose(v.normalize().angle(v), 0)
|
c1fbc761e10e06effa49ede1f8dbc04189999bd5
|
niftynet/layer/post_processing.py
|
niftynet/layer/post_processing.py
|
from __future__ import absolute_import, print_function
import tensorflow as tf
from niftynet.utilities.util_common import look_up_operations
from niftynet.layer.base_layer import Layer
SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"}
class PostProcessingLayer(Layer):
"""
This layer operation converts the raw network outputs into final inference
results.
"""
def __init__(self, func='', num_classes=0, name='post_processing'):
super(PostProcessingLayer, self).__init__(name=name)
self.func = look_up_operations(func.upper(), SUPPORTED_OPS)
self.num_classes = num_classes
def num_output_channels(self):
assert self._op._variables_created
if self.func == "SOFTMAX":
return self.num_classes
else:
return 1
def layer_op(self, inputs):
if self.func == "SOFTMAX":
output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32)
elif self.func == "ARGMAX":
output_tensor = tf.to_int64(tf.argmax(inputs, -1))
output_tensor = tf.expand_dims(output_tensor, axis=-1)
elif self.func == "IDENTITY":
output_tensor = tf.cast(inputs, tf.float32)
return output_tensor
|
from __future__ import absolute_import, print_function
import tensorflow as tf
from niftynet.utilities.util_common import look_up_operations
from niftynet.layer.base_layer import Layer
SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"}
class PostProcessingLayer(Layer):
"""
This layer operation converts the raw network outputs into final inference
results.
"""
def __init__(self, func='', num_classes=0, name='post_processing'):
super(PostProcessingLayer, self).__init__(name=name)
self.func = look_up_operations(func.upper(), SUPPORTED_OPS)
self.num_classes = num_classes
def num_output_channels(self):
assert self._op._variables_created
if self.func == "SOFTMAX":
return self.num_classes
else:
return 1
def layer_op(self, inputs):
if self.func == "SOFTMAX":
output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32)
elif self.func == "ARGMAX":
output_tensor = tf.argmax(inputs, -1,output_type=tf.int32)
output_tensor = tf.expand_dims(output_tensor, axis=-1)
elif self.func == "IDENTITY":
output_tensor = tf.cast(inputs, tf.float32)
return output_tensor
|
Change label output to int32 for compatibility with some viewers
|
Change label output to int32 for compatibility with some viewers
|
Python
|
apache-2.0
|
NifTK/NiftyNet,NifTK/NiftyNet,NifTK/NiftyNet,NifTK/NiftyNet
|
from __future__ import absolute_import, print_function
import tensorflow as tf
from niftynet.utilities.util_common import look_up_operations
from niftynet.layer.base_layer import Layer
SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"}
class PostProcessingLayer(Layer):
"""
This layer operation converts the raw network outputs into final inference
results.
"""
def __init__(self, func='', num_classes=0, name='post_processing'):
super(PostProcessingLayer, self).__init__(name=name)
self.func = look_up_operations(func.upper(), SUPPORTED_OPS)
self.num_classes = num_classes
def num_output_channels(self):
assert self._op._variables_created
if self.func == "SOFTMAX":
return self.num_classes
else:
return 1
def layer_op(self, inputs):
if self.func == "SOFTMAX":
output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32)
elif self.func == "ARGMAX":
- output_tensor = tf.to_int64(tf.argmax(inputs, -1))
+ output_tensor = tf.argmax(inputs, -1,output_type=tf.int32)
output_tensor = tf.expand_dims(output_tensor, axis=-1)
elif self.func == "IDENTITY":
output_tensor = tf.cast(inputs, tf.float32)
return output_tensor
|
Change label output to int32 for compatibility with some viewers
|
## Code Before:
from __future__ import absolute_import, print_function
import tensorflow as tf
from niftynet.utilities.util_common import look_up_operations
from niftynet.layer.base_layer import Layer
SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"}
class PostProcessingLayer(Layer):
"""
This layer operation converts the raw network outputs into final inference
results.
"""
def __init__(self, func='', num_classes=0, name='post_processing'):
super(PostProcessingLayer, self).__init__(name=name)
self.func = look_up_operations(func.upper(), SUPPORTED_OPS)
self.num_classes = num_classes
def num_output_channels(self):
assert self._op._variables_created
if self.func == "SOFTMAX":
return self.num_classes
else:
return 1
def layer_op(self, inputs):
if self.func == "SOFTMAX":
output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32)
elif self.func == "ARGMAX":
output_tensor = tf.to_int64(tf.argmax(inputs, -1))
output_tensor = tf.expand_dims(output_tensor, axis=-1)
elif self.func == "IDENTITY":
output_tensor = tf.cast(inputs, tf.float32)
return output_tensor
## Instruction:
Change label output to int32 for compatibility with some viewers
## Code After:
from __future__ import absolute_import, print_function
import tensorflow as tf
from niftynet.utilities.util_common import look_up_operations
from niftynet.layer.base_layer import Layer
SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"}
class PostProcessingLayer(Layer):
"""
This layer operation converts the raw network outputs into final inference
results.
"""
def __init__(self, func='', num_classes=0, name='post_processing'):
super(PostProcessingLayer, self).__init__(name=name)
self.func = look_up_operations(func.upper(), SUPPORTED_OPS)
self.num_classes = num_classes
def num_output_channels(self):
assert self._op._variables_created
if self.func == "SOFTMAX":
return self.num_classes
else:
return 1
def layer_op(self, inputs):
if self.func == "SOFTMAX":
output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32)
elif self.func == "ARGMAX":
output_tensor = tf.argmax(inputs, -1,output_type=tf.int32)
output_tensor = tf.expand_dims(output_tensor, axis=-1)
elif self.func == "IDENTITY":
output_tensor = tf.cast(inputs, tf.float32)
return output_tensor
|
from __future__ import absolute_import, print_function
import tensorflow as tf
from niftynet.utilities.util_common import look_up_operations
from niftynet.layer.base_layer import Layer
SUPPORTED_OPS = {"SOFTMAX", "ARGMAX", "IDENTITY"}
class PostProcessingLayer(Layer):
"""
This layer operation converts the raw network outputs into final inference
results.
"""
def __init__(self, func='', num_classes=0, name='post_processing'):
super(PostProcessingLayer, self).__init__(name=name)
self.func = look_up_operations(func.upper(), SUPPORTED_OPS)
self.num_classes = num_classes
def num_output_channels(self):
assert self._op._variables_created
if self.func == "SOFTMAX":
return self.num_classes
else:
return 1
def layer_op(self, inputs):
if self.func == "SOFTMAX":
output_tensor = tf.cast(tf.nn.softmax(inputs), tf.float32)
elif self.func == "ARGMAX":
- output_tensor = tf.to_int64(tf.argmax(inputs, -1))
+ output_tensor = tf.argmax(inputs, -1,output_type=tf.int32)
output_tensor = tf.expand_dims(output_tensor, axis=-1)
elif self.func == "IDENTITY":
output_tensor = tf.cast(inputs, tf.float32)
return output_tensor
|
4d1b96792f73777adaa0a79341901ca82f57839b
|
use/functional.py
|
use/functional.py
|
def pipe(*functions):
def closure(x):
for fn in functions:
if not out:
out = fn(x)
else:
out = fn(out)
return out
return closure
|
import collections
import functools
def pipe(*functions):
def closure(x):
for fn in functions:
if not out:
out = fn(x)
else:
out = fn(out)
return out
return closure
class memoize(object):
'''Decorator. Caches a function's return value each time it is called.
If called later with the same arguments, the cached value is returned
(not reevaluated).
'''
def __init__(self, func):
self.func = func
self.cache = {}
def __call__(self, *args):
if not isinstance(args, collections.Hashable):
# uncacheable. a list, for instance.
# better to not cache than blow up.
return self.func(*args)
if args in self.cache:
return self.cache[args]
else:
value = self.func(*args)
self.cache[args] = value
return value
def __repr__(self):
'''Return the function's docstring.'''
return self.func.__doc__
def __get__(self, obj, objtype):
'''Support instance methods.'''
return functools.partial(self.__call__, obj)
|
Add a simple memoize function
|
Add a simple memoize function
|
Python
|
mit
|
log0ymxm/corgi
|
+ import collections
+ import functools
+
+
def pipe(*functions):
def closure(x):
for fn in functions:
if not out:
out = fn(x)
else:
out = fn(out)
return out
return closure
+
+ class memoize(object):
+ '''Decorator. Caches a function's return value each time it is called.
+ If called later with the same arguments, the cached value is returned
+ (not reevaluated).
+ '''
+
+ def __init__(self, func):
+ self.func = func
+ self.cache = {}
+
+ def __call__(self, *args):
+ if not isinstance(args, collections.Hashable):
+ # uncacheable. a list, for instance.
+ # better to not cache than blow up.
+ return self.func(*args)
+ if args in self.cache:
+ return self.cache[args]
+ else:
+ value = self.func(*args)
+ self.cache[args] = value
+ return value
+
+ def __repr__(self):
+ '''Return the function's docstring.'''
+ return self.func.__doc__
+
+ def __get__(self, obj, objtype):
+ '''Support instance methods.'''
+ return functools.partial(self.__call__, obj)
+
|
Add a simple memoize function
|
## Code Before:
def pipe(*functions):
def closure(x):
for fn in functions:
if not out:
out = fn(x)
else:
out = fn(out)
return out
return closure
## Instruction:
Add a simple memoize function
## Code After:
import collections
import functools
def pipe(*functions):
def closure(x):
for fn in functions:
if not out:
out = fn(x)
else:
out = fn(out)
return out
return closure
class memoize(object):
'''Decorator. Caches a function's return value each time it is called.
If called later with the same arguments, the cached value is returned
(not reevaluated).
'''
def __init__(self, func):
self.func = func
self.cache = {}
def __call__(self, *args):
if not isinstance(args, collections.Hashable):
# uncacheable. a list, for instance.
# better to not cache than blow up.
return self.func(*args)
if args in self.cache:
return self.cache[args]
else:
value = self.func(*args)
self.cache[args] = value
return value
def __repr__(self):
'''Return the function's docstring.'''
return self.func.__doc__
def __get__(self, obj, objtype):
'''Support instance methods.'''
return functools.partial(self.__call__, obj)
|
+ import collections
+ import functools
+
+
def pipe(*functions):
def closure(x):
for fn in functions:
if not out:
out = fn(x)
else:
out = fn(out)
return out
return closure
+
+
+ class memoize(object):
+ '''Decorator. Caches a function's return value each time it is called.
+ If called later with the same arguments, the cached value is returned
+ (not reevaluated).
+ '''
+
+ def __init__(self, func):
+ self.func = func
+ self.cache = {}
+
+ def __call__(self, *args):
+ if not isinstance(args, collections.Hashable):
+ # uncacheable. a list, for instance.
+ # better to not cache than blow up.
+ return self.func(*args)
+ if args in self.cache:
+ return self.cache[args]
+ else:
+ value = self.func(*args)
+ self.cache[args] = value
+ return value
+
+ def __repr__(self):
+ '''Return the function's docstring.'''
+ return self.func.__doc__
+
+ def __get__(self, obj, objtype):
+ '''Support instance methods.'''
+ return functools.partial(self.__call__, obj)
|
ddd45afa0708682bb11d606e03e38aed111d7b9c
|
fireplace/cards/game/all.py
|
fireplace/cards/game/all.py
|
from ..utils import *
# The Coin
class GAME_005:
play = ManaThisTurn(CONTROLLER, 1)
|
from ..utils import *
# The Coin
class GAME_005:
play = ManaThisTurn(CONTROLLER, 1)
# Big Banana
class TB_006:
play = Buff(TARGET, "TB_006e")
# Deviate Banana
class TB_007:
play = Buff(TARGET, "TB_007e")
# Rotten Banana
class TB_008:
play = Hit(TARGET, 1)
|
Implement Big Banana, Deviate Banana, Rotten Banana
|
Implement Big Banana, Deviate Banana, Rotten Banana
|
Python
|
agpl-3.0
|
liujimj/fireplace,Ragowit/fireplace,butozerca/fireplace,butozerca/fireplace,smallnamespace/fireplace,amw2104/fireplace,smallnamespace/fireplace,beheh/fireplace,NightKev/fireplace,Meerkov/fireplace,Meerkov/fireplace,liujimj/fireplace,oftc-ftw/fireplace,Ragowit/fireplace,amw2104/fireplace,jleclanche/fireplace,oftc-ftw/fireplace
|
from ..utils import *
# The Coin
class GAME_005:
play = ManaThisTurn(CONTROLLER, 1)
+
+ # Big Banana
+ class TB_006:
+ play = Buff(TARGET, "TB_006e")
+
+
+ # Deviate Banana
+ class TB_007:
+ play = Buff(TARGET, "TB_007e")
+
+
+ # Rotten Banana
+ class TB_008:
+ play = Hit(TARGET, 1)
+
|
Implement Big Banana, Deviate Banana, Rotten Banana
|
## Code Before:
from ..utils import *
# The Coin
class GAME_005:
play = ManaThisTurn(CONTROLLER, 1)
## Instruction:
Implement Big Banana, Deviate Banana, Rotten Banana
## Code After:
from ..utils import *
# The Coin
class GAME_005:
play = ManaThisTurn(CONTROLLER, 1)
# Big Banana
class TB_006:
play = Buff(TARGET, "TB_006e")
# Deviate Banana
class TB_007:
play = Buff(TARGET, "TB_007e")
# Rotten Banana
class TB_008:
play = Hit(TARGET, 1)
|
from ..utils import *
# The Coin
class GAME_005:
play = ManaThisTurn(CONTROLLER, 1)
+
+
+ # Big Banana
+ class TB_006:
+ play = Buff(TARGET, "TB_006e")
+
+
+ # Deviate Banana
+ class TB_007:
+ play = Buff(TARGET, "TB_007e")
+
+
+ # Rotten Banana
+ class TB_008:
+ play = Hit(TARGET, 1)
|
0b41bdf6897bb070fc3d90aa5d90228e744dee60
|
sunpy/util/map_manager.py
|
sunpy/util/map_manager.py
|
import weakref
import sunpy
class MapManager(weakref.WeakSet):
"""Weak referenced set of maps created using functions decorated by manage_maps."""
def __repr__(self):
return str(self.data)
def manage_maps(fn):
"""Maps returned by functions decorated with manage_maps (eg. sunpy.make_map)
will be registered in the sunpy.map_manager list."""
def fn_manage_maps(*args, **kwargs):
ret = fn(*args, **kwargs)
sunpy.map_manager.add(ret)
return ret
return fn_manage_maps
|
import weakref
import sunpy
class MapManager(weakref.WeakSet):
"""Weak referenced set of maps created using functions decorated by manage_maps."""
pass
def manage_maps(fn):
"""Maps returned by functions decorated with manage_maps (eg. sunpy.make_map)
will be registered in the sunpy.map_manager list."""
def fn_manage_maps(*args, **kwargs):
ret = fn(*args, **kwargs)
sunpy.map_manager.add(ret)
return ret
return fn_manage_maps
|
Remove manager repr (user should not need to view contents)
|
Remove manager repr (user should not need to view contents)
|
Python
|
bsd-2-clause
|
mjm159/sunpy,dpshelio/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy,dpshelio/sunpy,Alex-Ian-Hamilton/sunpy,Alex-Ian-Hamilton/sunpy,mjm159/sunpy
|
import weakref
import sunpy
class MapManager(weakref.WeakSet):
"""Weak referenced set of maps created using functions decorated by manage_maps."""
+ pass
-
- def __repr__(self):
- return str(self.data)
def manage_maps(fn):
"""Maps returned by functions decorated with manage_maps (eg. sunpy.make_map)
will be registered in the sunpy.map_manager list."""
def fn_manage_maps(*args, **kwargs):
ret = fn(*args, **kwargs)
sunpy.map_manager.add(ret)
return ret
return fn_manage_maps
|
Remove manager repr (user should not need to view contents)
|
## Code Before:
import weakref
import sunpy
class MapManager(weakref.WeakSet):
"""Weak referenced set of maps created using functions decorated by manage_maps."""
def __repr__(self):
return str(self.data)
def manage_maps(fn):
"""Maps returned by functions decorated with manage_maps (eg. sunpy.make_map)
will be registered in the sunpy.map_manager list."""
def fn_manage_maps(*args, **kwargs):
ret = fn(*args, **kwargs)
sunpy.map_manager.add(ret)
return ret
return fn_manage_maps
## Instruction:
Remove manager repr (user should not need to view contents)
## Code After:
import weakref
import sunpy
class MapManager(weakref.WeakSet):
"""Weak referenced set of maps created using functions decorated by manage_maps."""
pass
def manage_maps(fn):
"""Maps returned by functions decorated with manage_maps (eg. sunpy.make_map)
will be registered in the sunpy.map_manager list."""
def fn_manage_maps(*args, **kwargs):
ret = fn(*args, **kwargs)
sunpy.map_manager.add(ret)
return ret
return fn_manage_maps
|
import weakref
import sunpy
class MapManager(weakref.WeakSet):
"""Weak referenced set of maps created using functions decorated by manage_maps."""
+ pass
-
- def __repr__(self):
- return str(self.data)
def manage_maps(fn):
"""Maps returned by functions decorated with manage_maps (eg. sunpy.make_map)
will be registered in the sunpy.map_manager list."""
def fn_manage_maps(*args, **kwargs):
ret = fn(*args, **kwargs)
sunpy.map_manager.add(ret)
return ret
return fn_manage_maps
|
b3413818bf651c13cef047132813fb26a185cd33
|
indra/tests/test_reading_files.py
|
indra/tests/test_reading_files.py
|
from os import path
from indra.tools.reading.read_files import read_files, get_readers
from nose.plugins.attrib import attr
@attr('slow', 'nonpublic')
def test_read_files():
"Test that the system can read files."
# Create the test files.
example_files = []
# Get txt content
abstract_txt = ("This is a paper that contains the phrase: MEK "
"phosphorylates ERK.")
with open('test_abstract.txt', 'w') as f:
f.write(abstract_txt)
example_files.append('test_abstract.txt')
# Get nxml content
pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)),
'pmc_cont_example.nxml')
if path.exists(pmc_test_fpath):
example_files.append(pmc_test_fpath)
assert len(example_files), "No content available to test."
# Now read them.
readers = get_readers()
outputs = read_files(example_files, readers)
N_out = len(outputs)
N_exp = 2*len(example_files)
assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
|
from os import path
from indra.tools.reading.read_files import read_files, get_reader_classes
from nose.plugins.attrib import attr
from indra.tools.reading.readers import EmptyReader
@attr('slow', 'nonpublic', 'notravis')
def test_read_files():
"Test that the system can read files."
# Create the test files.
example_files = []
# Get txt content
abstract_txt = ("This is a paper that contains the phrase: MEK "
"phosphorylates ERK.")
with open('test_abstract.txt', 'w') as f:
f.write(abstract_txt)
example_files.append('test_abstract.txt')
# Get nxml content
pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)),
'pmc_cont_example.nxml')
if path.exists(pmc_test_fpath):
example_files.append(pmc_test_fpath)
assert len(example_files), "No content available to test."
# Now read them.
reader_classes = get_reader_classes()
readers = []
for rc in reader_classes:
readers.append(rc())
outputs = read_files(example_files, readers)
N_out = len(outputs)
proper_readers = [r for r in readers if not isinstance(r, EmptyReader)]
N_exp = len(proper_readers)*len(example_files)
assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
|
Fix the reading files test.
|
Fix the reading files test.
|
Python
|
bsd-2-clause
|
johnbachman/belpy,pvtodorov/indra,sorgerlab/indra,bgyori/indra,pvtodorov/indra,sorgerlab/belpy,johnbachman/indra,sorgerlab/indra,sorgerlab/belpy,sorgerlab/belpy,bgyori/indra,pvtodorov/indra,johnbachman/belpy,johnbachman/indra,johnbachman/indra,bgyori/indra,sorgerlab/indra,pvtodorov/indra,johnbachman/belpy
|
from os import path
- from indra.tools.reading.read_files import read_files, get_readers
+ from indra.tools.reading.read_files import read_files, get_reader_classes
from nose.plugins.attrib import attr
+ from indra.tools.reading.readers import EmptyReader
+
- @attr('slow', 'nonpublic')
+ @attr('slow', 'nonpublic', 'notravis')
def test_read_files():
"Test that the system can read files."
# Create the test files.
example_files = []
# Get txt content
abstract_txt = ("This is a paper that contains the phrase: MEK "
"phosphorylates ERK.")
with open('test_abstract.txt', 'w') as f:
f.write(abstract_txt)
example_files.append('test_abstract.txt')
# Get nxml content
pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)),
'pmc_cont_example.nxml')
if path.exists(pmc_test_fpath):
example_files.append(pmc_test_fpath)
assert len(example_files), "No content available to test."
# Now read them.
- readers = get_readers()
+ reader_classes = get_reader_classes()
+ readers = []
+ for rc in reader_classes:
+ readers.append(rc())
outputs = read_files(example_files, readers)
N_out = len(outputs)
+ proper_readers = [r for r in readers if not isinstance(r, EmptyReader)]
- N_exp = 2*len(example_files)
+ N_exp = len(proper_readers)*len(example_files)
assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
|
Fix the reading files test.
|
## Code Before:
from os import path
from indra.tools.reading.read_files import read_files, get_readers
from nose.plugins.attrib import attr
@attr('slow', 'nonpublic')
def test_read_files():
"Test that the system can read files."
# Create the test files.
example_files = []
# Get txt content
abstract_txt = ("This is a paper that contains the phrase: MEK "
"phosphorylates ERK.")
with open('test_abstract.txt', 'w') as f:
f.write(abstract_txt)
example_files.append('test_abstract.txt')
# Get nxml content
pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)),
'pmc_cont_example.nxml')
if path.exists(pmc_test_fpath):
example_files.append(pmc_test_fpath)
assert len(example_files), "No content available to test."
# Now read them.
readers = get_readers()
outputs = read_files(example_files, readers)
N_out = len(outputs)
N_exp = 2*len(example_files)
assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
## Instruction:
Fix the reading files test.
## Code After:
from os import path
from indra.tools.reading.read_files import read_files, get_reader_classes
from nose.plugins.attrib import attr
from indra.tools.reading.readers import EmptyReader
@attr('slow', 'nonpublic', 'notravis')
def test_read_files():
"Test that the system can read files."
# Create the test files.
example_files = []
# Get txt content
abstract_txt = ("This is a paper that contains the phrase: MEK "
"phosphorylates ERK.")
with open('test_abstract.txt', 'w') as f:
f.write(abstract_txt)
example_files.append('test_abstract.txt')
# Get nxml content
pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)),
'pmc_cont_example.nxml')
if path.exists(pmc_test_fpath):
example_files.append(pmc_test_fpath)
assert len(example_files), "No content available to test."
# Now read them.
reader_classes = get_reader_classes()
readers = []
for rc in reader_classes:
readers.append(rc())
outputs = read_files(example_files, readers)
N_out = len(outputs)
proper_readers = [r for r in readers if not isinstance(r, EmptyReader)]
N_exp = len(proper_readers)*len(example_files)
assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
|
from os import path
- from indra.tools.reading.read_files import read_files, get_readers
+ from indra.tools.reading.read_files import read_files, get_reader_classes
? ++++ +++
from nose.plugins.attrib import attr
+ from indra.tools.reading.readers import EmptyReader
+
- @attr('slow', 'nonpublic')
+ @attr('slow', 'nonpublic', 'notravis')
? ++++++++++++
def test_read_files():
"Test that the system can read files."
# Create the test files.
example_files = []
# Get txt content
abstract_txt = ("This is a paper that contains the phrase: MEK "
"phosphorylates ERK.")
with open('test_abstract.txt', 'w') as f:
f.write(abstract_txt)
example_files.append('test_abstract.txt')
# Get nxml content
pmc_test_fpath = path.join(path.dirname(path.abspath(__file__)),
'pmc_cont_example.nxml')
if path.exists(pmc_test_fpath):
example_files.append(pmc_test_fpath)
assert len(example_files), "No content available to test."
# Now read them.
- readers = get_readers()
+ reader_classes = get_reader_classes()
? +++++++ +++++++
+ readers = []
+ for rc in reader_classes:
+ readers.append(rc())
outputs = read_files(example_files, readers)
N_out = len(outputs)
+ proper_readers = [r for r in readers if not isinstance(r, EmptyReader)]
- N_exp = 2*len(example_files)
? ^
+ N_exp = len(proper_readers)*len(example_files)
? ^^^^^^^^^^^^^^^^^^^
assert N_out == N_exp, "Expected %d outputs, got %d." % (N_exp, N_out)
|
9fe573614e2f3ca9a6e738afb7f1af84b541092c
|
invertedindex.py
|
invertedindex.py
|
class InvertedIndex:
def __init__(self):
self.index = dict()
def add_mail(self, mail):
for key in ["simple_terms_body", "complexe_terms_body"]:
for terms in mail[key]:
if terms in self.index.keys():
self.index[terms].append((mail["name"], mail[key][terms]))
else:
self.index[terms] = list()
self.index[terms].append((mail["name"], mail[key][terms]))
|
class InvertedIndex:
def __init__(self):
self.index = dict()
def add_mail(self, mail):
for key in ["simple_terms_body", "complexe_terms_body"]:
for terms in mail[key]:
if terms in self.index.keys():
self.index[terms].append((mail["name"], mail[key][terms]))
else:
self.index[terms] = list()
self.index[terms].append((mail["name"], mail[key][terms]))
def terms(self):
for terms in self.index.keys():
yield terms
def get_terms(self):
return self.index.keys()
def file_counter(self, terms):
for val in self.index[terms]:
yield val
def get_file_counter(self, terms):
return self.index.values()
def file(self, terms):
for val in file_counter(terms):
yield val[0]
def counter(self, terms):
for val in file_counter(terms):
yield val[1]
|
Add some access function to inverted index
|
Add some access function to inverted index
|
Python
|
mit
|
Nedgang/adt_project
|
class InvertedIndex:
def __init__(self):
self.index = dict()
def add_mail(self, mail):
for key in ["simple_terms_body", "complexe_terms_body"]:
for terms in mail[key]:
if terms in self.index.keys():
self.index[terms].append((mail["name"], mail[key][terms]))
else:
self.index[terms] = list()
self.index[terms].append((mail["name"], mail[key][terms]))
+ def terms(self):
+ for terms in self.index.keys():
+ yield terms
+ def get_terms(self):
+ return self.index.keys()
+
+ def file_counter(self, terms):
+ for val in self.index[terms]:
+ yield val
+
+ def get_file_counter(self, terms):
+ return self.index.values()
+
+ def file(self, terms):
+ for val in file_counter(terms):
+ yield val[0]
+
+ def counter(self, terms):
+ for val in file_counter(terms):
+ yield val[1]
+
|
Add some access function to inverted index
|
## Code Before:
class InvertedIndex:
def __init__(self):
self.index = dict()
def add_mail(self, mail):
for key in ["simple_terms_body", "complexe_terms_body"]:
for terms in mail[key]:
if terms in self.index.keys():
self.index[terms].append((mail["name"], mail[key][terms]))
else:
self.index[terms] = list()
self.index[terms].append((mail["name"], mail[key][terms]))
## Instruction:
Add some access function to inverted index
## Code After:
class InvertedIndex:
def __init__(self):
self.index = dict()
def add_mail(self, mail):
for key in ["simple_terms_body", "complexe_terms_body"]:
for terms in mail[key]:
if terms in self.index.keys():
self.index[terms].append((mail["name"], mail[key][terms]))
else:
self.index[terms] = list()
self.index[terms].append((mail["name"], mail[key][terms]))
def terms(self):
for terms in self.index.keys():
yield terms
def get_terms(self):
return self.index.keys()
def file_counter(self, terms):
for val in self.index[terms]:
yield val
def get_file_counter(self, terms):
return self.index.values()
def file(self, terms):
for val in file_counter(terms):
yield val[0]
def counter(self, terms):
for val in file_counter(terms):
yield val[1]
|
class InvertedIndex:
def __init__(self):
self.index = dict()
def add_mail(self, mail):
for key in ["simple_terms_body", "complexe_terms_body"]:
for terms in mail[key]:
if terms in self.index.keys():
self.index[terms].append((mail["name"], mail[key][terms]))
else:
self.index[terms] = list()
self.index[terms].append((mail["name"], mail[key][terms]))
+ def terms(self):
+ for terms in self.index.keys():
+ yield terms
+
+ def get_terms(self):
+ return self.index.keys()
+
+ def file_counter(self, terms):
+ for val in self.index[terms]:
+ yield val
+
+ def get_file_counter(self, terms):
+ return self.index.values()
+
+ def file(self, terms):
+ for val in file_counter(terms):
+ yield val[0]
+
+ def counter(self, terms):
+ for val in file_counter(terms):
+ yield val[1]
|
debe3a250a04986583589b1192cb6111b8b6c228
|
pydelhiconf/uix/screens/screenabout.py
|
pydelhiconf/uix/screens/screenabout.py
|
from kivy.uix.screenmanager import Screen
from kivy.lang import Builder
from kivy.factory import Factory
from functools import partial
import webbrowser
class ScreenAbout(Screen):
Builder.load_string('''
<ScreenAbout>
spacing: dp(9)
name: 'ScreenAbout'
ScrollView
id: scroll
ScrollGrid
AsyncImage
id: imgbt
allow_stretch: True
size_hint_y: None
height: dp(200)
BackLabel
id: comm_desc
''')
def on_pre_enter(self):
self.ids.scroll.opacity = 0
def on_enter(self, onsuccess=False):
from network import get_data
about = get_data('about', onsuccess=onsuccess)
if not about:
return
about = about.get('0.0.1')[0]
imbt = self.ids.imgbt
imbt.source = about['logo']
imbt.on_released = partial(webbrowser.open, about['website'])
self.ids.comm_desc.text = about['about']
Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
|
from kivy.uix.screenmanager import Screen
from kivy.lang import Builder
from kivy.factory import Factory
from functools import partial
import webbrowser
class ScreenAbout(Screen):
Builder.load_string('''
<ScreenAbout>
spacing: dp(9)
name: 'ScreenAbout'
ScrollView
id: scroll
ScrollGrid
AsyncImage
id: imgbt
allow_stretch: True
size_hint_y: None
height: dp(200)
BackLabel
id: comm_desc
FloatLayout
size_hint_y: None
height: dp(45)
ActiveButton
id: but
text: "Visit our website"
size_hint: None, None
width: dp(200)
center_x: comm_desc.center_x
top: comm_desc.y - dp(10)
''')
def on_pre_enter(self):
self.ids.scroll.opacity = 0
def on_enter(self, onsuccess=False):
from network import get_data
about = get_data('about', onsuccess=onsuccess)
if not about:
return
about = about.get('0.0.1')[0]
imbt = self.ids.imgbt
imbt.source = about['logo']
self.ids.but.on_released = partial(webbrowser.open, about['website'])
self.ids.comm_desc.text = about['about']
Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
|
Add button that links to website
|
Add button that links to website
|
Python
|
agpl-3.0
|
pydelhi/pydelhi_mobile,shivan1b/pydelhi_mobile,samukasmk/pythonbrasil_mobile,akshayaurora/PyDelhiMobile
|
from kivy.uix.screenmanager import Screen
from kivy.lang import Builder
from kivy.factory import Factory
from functools import partial
import webbrowser
class ScreenAbout(Screen):
Builder.load_string('''
<ScreenAbout>
spacing: dp(9)
name: 'ScreenAbout'
ScrollView
id: scroll
ScrollGrid
AsyncImage
id: imgbt
allow_stretch: True
size_hint_y: None
height: dp(200)
BackLabel
id: comm_desc
+ FloatLayout
+ size_hint_y: None
+ height: dp(45)
+ ActiveButton
+ id: but
+ text: "Visit our website"
+ size_hint: None, None
+ width: dp(200)
+ center_x: comm_desc.center_x
+ top: comm_desc.y - dp(10)
''')
def on_pre_enter(self):
self.ids.scroll.opacity = 0
def on_enter(self, onsuccess=False):
from network import get_data
about = get_data('about', onsuccess=onsuccess)
if not about:
return
about = about.get('0.0.1')[0]
imbt = self.ids.imgbt
imbt.source = about['logo']
- imbt.on_released = partial(webbrowser.open, about['website'])
+ self.ids.but.on_released = partial(webbrowser.open, about['website'])
self.ids.comm_desc.text = about['about']
Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
|
Add button that links to website
|
## Code Before:
from kivy.uix.screenmanager import Screen
from kivy.lang import Builder
from kivy.factory import Factory
from functools import partial
import webbrowser
class ScreenAbout(Screen):
Builder.load_string('''
<ScreenAbout>
spacing: dp(9)
name: 'ScreenAbout'
ScrollView
id: scroll
ScrollGrid
AsyncImage
id: imgbt
allow_stretch: True
size_hint_y: None
height: dp(200)
BackLabel
id: comm_desc
''')
def on_pre_enter(self):
self.ids.scroll.opacity = 0
def on_enter(self, onsuccess=False):
from network import get_data
about = get_data('about', onsuccess=onsuccess)
if not about:
return
about = about.get('0.0.1')[0]
imbt = self.ids.imgbt
imbt.source = about['logo']
imbt.on_released = partial(webbrowser.open, about['website'])
self.ids.comm_desc.text = about['about']
Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
## Instruction:
Add button that links to website
## Code After:
from kivy.uix.screenmanager import Screen
from kivy.lang import Builder
from kivy.factory import Factory
from functools import partial
import webbrowser
class ScreenAbout(Screen):
Builder.load_string('''
<ScreenAbout>
spacing: dp(9)
name: 'ScreenAbout'
ScrollView
id: scroll
ScrollGrid
AsyncImage
id: imgbt
allow_stretch: True
size_hint_y: None
height: dp(200)
BackLabel
id: comm_desc
FloatLayout
size_hint_y: None
height: dp(45)
ActiveButton
id: but
text: "Visit our website"
size_hint: None, None
width: dp(200)
center_x: comm_desc.center_x
top: comm_desc.y - dp(10)
''')
def on_pre_enter(self):
self.ids.scroll.opacity = 0
def on_enter(self, onsuccess=False):
from network import get_data
about = get_data('about', onsuccess=onsuccess)
if not about:
return
about = about.get('0.0.1')[0]
imbt = self.ids.imgbt
imbt.source = about['logo']
self.ids.but.on_released = partial(webbrowser.open, about['website'])
self.ids.comm_desc.text = about['about']
Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
|
from kivy.uix.screenmanager import Screen
from kivy.lang import Builder
from kivy.factory import Factory
from functools import partial
import webbrowser
class ScreenAbout(Screen):
Builder.load_string('''
<ScreenAbout>
spacing: dp(9)
name: 'ScreenAbout'
ScrollView
id: scroll
ScrollGrid
AsyncImage
id: imgbt
allow_stretch: True
size_hint_y: None
height: dp(200)
BackLabel
id: comm_desc
+ FloatLayout
+ size_hint_y: None
+ height: dp(45)
+ ActiveButton
+ id: but
+ text: "Visit our website"
+ size_hint: None, None
+ width: dp(200)
+ center_x: comm_desc.center_x
+ top: comm_desc.y - dp(10)
''')
def on_pre_enter(self):
self.ids.scroll.opacity = 0
def on_enter(self, onsuccess=False):
from network import get_data
about = get_data('about', onsuccess=onsuccess)
if not about:
return
about = about.get('0.0.1')[0]
imbt = self.ids.imgbt
imbt.source = about['logo']
- imbt.on_released = partial(webbrowser.open, about['website'])
? ^
+ self.ids.but.on_released = partial(webbrowser.open, about['website'])
? +++++ ^^^ +
self.ids.comm_desc.text = about['about']
Factory.Animation(opacity=1, d=.5).start(self.ids.scroll)
|
460a2430fbd8832f3fada1a74b754d71a27ac282
|
mockingjay/matcher.py
|
mockingjay/matcher.py
|
import abc
import re
class Matcher(object):
__metaclass__ = abc.ABCMeta
@abc.abstractmethod
def assert_request_matched(self, request):
"""
Assert that the request matched the spec in this matcher object.
"""
class HeaderMatcher(Matcher):
"""
Matcher for the request's header.
:param key: the name of the header
:param value: the value of the header
"""
def __init__(self, key, value):
self.key = key
self.value = value
def assert_request_matched(self, request):
assert request.headers.get(self.key) == self.value
class BodyMatcher(Matcher):
"""
Matcher for the request body.
:param body: can either be a string or a :class:`_sre.SRE_Pattern`: object
"""
def __init__(self, body):
self.body = body
def assert_request_matched(self, request):
if isinstance(self.body, re._pattern_type):
assert self.body.search(request.body) is not None
else:
assert request.body == self.body
|
import abc
import re
class StringOrPattern(object):
"""
A decorator object that wraps a string or a regex pattern so that it can
be compared against another string either literally or using the pattern.
"""
def __init__(self, subject):
self.subject = subject
def __eq__(self, other_str):
if isinstance(self.subject, re._pattern_type):
return self.subject.search(other_str) is not None
else:
return self.subject == other_str
def __hash__(self):
return self.subject.__hash__()
class Matcher(object):
__metaclass__ = abc.ABCMeta
@abc.abstractmethod
def assert_request_matched(self, request):
"""
Assert that the request matched the spec in this matcher object.
"""
class HeaderMatcher(Matcher):
"""
Matcher for the request's header.
:param key: the name of the header
:param value: the value of the header
"""
def __init__(self, key, value):
self.key = key
self.value = StringOrPattern(value)
def assert_request_matched(self, request):
assert request.headers.get(self.key) == self.value
class BodyMatcher(Matcher):
"""
Matcher for the request body.
:param body: can either be a string or a :class:`_sre.SRE_Pattern`: object
"""
def __init__(self, body):
self.body = StringOrPattern(body)
def assert_request_matched(self, request):
assert request.body == self.body
|
Allow all values to be compared with either literally or with a pattern
|
Allow all values to be compared with either literally or with a pattern
|
Python
|
bsd-3-clause
|
kevinjqiu/mockingjay
|
import abc
import re
+
+
+ class StringOrPattern(object):
+ """
+ A decorator object that wraps a string or a regex pattern so that it can
+ be compared against another string either literally or using the pattern.
+ """
+ def __init__(self, subject):
+ self.subject = subject
+
+ def __eq__(self, other_str):
+ if isinstance(self.subject, re._pattern_type):
+ return self.subject.search(other_str) is not None
+ else:
+ return self.subject == other_str
+
+ def __hash__(self):
+ return self.subject.__hash__()
class Matcher(object):
__metaclass__ = abc.ABCMeta
@abc.abstractmethod
def assert_request_matched(self, request):
"""
Assert that the request matched the spec in this matcher object.
"""
class HeaderMatcher(Matcher):
"""
Matcher for the request's header.
:param key: the name of the header
:param value: the value of the header
"""
def __init__(self, key, value):
self.key = key
- self.value = value
+ self.value = StringOrPattern(value)
def assert_request_matched(self, request):
assert request.headers.get(self.key) == self.value
class BodyMatcher(Matcher):
"""
Matcher for the request body.
:param body: can either be a string or a :class:`_sre.SRE_Pattern`: object
"""
def __init__(self, body):
- self.body = body
+ self.body = StringOrPattern(body)
def assert_request_matched(self, request):
- if isinstance(self.body, re._pattern_type):
- assert self.body.search(request.body) is not None
- else:
- assert request.body == self.body
+ assert request.body == self.body
|
Allow all values to be compared with either literally or with a pattern
|
## Code Before:
import abc
import re
class Matcher(object):
__metaclass__ = abc.ABCMeta
@abc.abstractmethod
def assert_request_matched(self, request):
"""
Assert that the request matched the spec in this matcher object.
"""
class HeaderMatcher(Matcher):
"""
Matcher for the request's header.
:param key: the name of the header
:param value: the value of the header
"""
def __init__(self, key, value):
self.key = key
self.value = value
def assert_request_matched(self, request):
assert request.headers.get(self.key) == self.value
class BodyMatcher(Matcher):
"""
Matcher for the request body.
:param body: can either be a string or a :class:`_sre.SRE_Pattern`: object
"""
def __init__(self, body):
self.body = body
def assert_request_matched(self, request):
if isinstance(self.body, re._pattern_type):
assert self.body.search(request.body) is not None
else:
assert request.body == self.body
## Instruction:
Allow all values to be compared with either literally or with a pattern
## Code After:
import abc
import re
class StringOrPattern(object):
"""
A decorator object that wraps a string or a regex pattern so that it can
be compared against another string either literally or using the pattern.
"""
def __init__(self, subject):
self.subject = subject
def __eq__(self, other_str):
if isinstance(self.subject, re._pattern_type):
return self.subject.search(other_str) is not None
else:
return self.subject == other_str
def __hash__(self):
return self.subject.__hash__()
class Matcher(object):
__metaclass__ = abc.ABCMeta
@abc.abstractmethod
def assert_request_matched(self, request):
"""
Assert that the request matched the spec in this matcher object.
"""
class HeaderMatcher(Matcher):
"""
Matcher for the request's header.
:param key: the name of the header
:param value: the value of the header
"""
def __init__(self, key, value):
self.key = key
self.value = StringOrPattern(value)
def assert_request_matched(self, request):
assert request.headers.get(self.key) == self.value
class BodyMatcher(Matcher):
"""
Matcher for the request body.
:param body: can either be a string or a :class:`_sre.SRE_Pattern`: object
"""
def __init__(self, body):
self.body = StringOrPattern(body)
def assert_request_matched(self, request):
assert request.body == self.body
|
import abc
import re
+
+
+ class StringOrPattern(object):
+ """
+ A decorator object that wraps a string or a regex pattern so that it can
+ be compared against another string either literally or using the pattern.
+ """
+ def __init__(self, subject):
+ self.subject = subject
+
+ def __eq__(self, other_str):
+ if isinstance(self.subject, re._pattern_type):
+ return self.subject.search(other_str) is not None
+ else:
+ return self.subject == other_str
+
+ def __hash__(self):
+ return self.subject.__hash__()
class Matcher(object):
__metaclass__ = abc.ABCMeta
@abc.abstractmethod
def assert_request_matched(self, request):
"""
Assert that the request matched the spec in this matcher object.
"""
class HeaderMatcher(Matcher):
"""
Matcher for the request's header.
:param key: the name of the header
:param value: the value of the header
"""
def __init__(self, key, value):
self.key = key
- self.value = value
+ self.value = StringOrPattern(value)
? ++++++++++++++++ +
def assert_request_matched(self, request):
assert request.headers.get(self.key) == self.value
class BodyMatcher(Matcher):
"""
Matcher for the request body.
:param body: can either be a string or a :class:`_sre.SRE_Pattern`: object
"""
def __init__(self, body):
- self.body = body
+ self.body = StringOrPattern(body)
def assert_request_matched(self, request):
- if isinstance(self.body, re._pattern_type):
- assert self.body.search(request.body) is not None
- else:
- assert request.body == self.body
? ----
+ assert request.body == self.body
|
95186f684328d5b84611f405d47d474c53cad619
|
cat.py
|
cat.py
|
import io
import aiohttp
from discord.ext import commands
import yaml
class Cat:
def __init__(self, bot):
self.bot = bot
with open('config.yaml') as file:
data = yaml.load(file)
self.key = data.get('cat_key', '')
self.url = 'http://thecatapi.com/api/images/get'
self.params = {'api_key': self.key,
'type': 'png,jpg'}
@commands.command()
async def cat(self, ctx):
s = self.bot.session
async with ctx.typing(), s.get(self.url, params=self.params) as resp:
image = io.BytesIO(await resp.content.read())
ext = resp.headers['Content-Type'].partition('/')[2]
await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}')
def setup(bot):
bot.add_cog(Cat(bot))
|
import io
import aiohttp
import discord
from discord.ext import commands
from lxml import etree
import yaml
class Cat:
def __init__(self, bot):
self.bot = bot
with open('config.yaml') as file:
data = yaml.load(file)
self.key = data.get('cat_key', '')
self.url = 'http://thecatapi.com/api/images/get'
self.params = {'api_key': self.key,
'type': 'png,jpg',
'format': 'xml',
}
@commands.command()
async def cat(self, ctx):
session = self.bot.session
async with ctx.typing():
async with session.get(self.url, params=self.params) as resp:
root = etree.fromstring(await resp.text())
url = root.find('.//url').text
embed = discord.Embed()
embed.set_image(url=url)
await ctx.send(embed=embed)
def setup(bot):
bot.add_cog(Cat(bot))
|
Send image in embed because aiohttp doesn't know how to parse links
|
Send image in embed because aiohttp doesn't know how to parse links
|
Python
|
mit
|
BeatButton/beattie,BeatButton/beattie-bot
|
import io
import aiohttp
+ import discord
from discord.ext import commands
+ from lxml import etree
import yaml
class Cat:
def __init__(self, bot):
self.bot = bot
with open('config.yaml') as file:
data = yaml.load(file)
self.key = data.get('cat_key', '')
self.url = 'http://thecatapi.com/api/images/get'
self.params = {'api_key': self.key,
- 'type': 'png,jpg'}
+ 'type': 'png,jpg',
+ 'format': 'xml',
+ }
@commands.command()
async def cat(self, ctx):
- s = self.bot.session
+ session = self.bot.session
+ async with ctx.typing():
- async with ctx.typing(), s.get(self.url, params=self.params) as resp:
+ async with session.get(self.url, params=self.params) as resp:
- image = io.BytesIO(await resp.content.read())
- ext = resp.headers['Content-Type'].partition('/')[2]
- await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}')
+ root = etree.fromstring(await resp.text())
+ url = root.find('.//url').text
+ embed = discord.Embed()
+ embed.set_image(url=url)
+ await ctx.send(embed=embed)
+
def setup(bot):
bot.add_cog(Cat(bot))
|
Send image in embed because aiohttp doesn't know how to parse links
|
## Code Before:
import io
import aiohttp
from discord.ext import commands
import yaml
class Cat:
def __init__(self, bot):
self.bot = bot
with open('config.yaml') as file:
data = yaml.load(file)
self.key = data.get('cat_key', '')
self.url = 'http://thecatapi.com/api/images/get'
self.params = {'api_key': self.key,
'type': 'png,jpg'}
@commands.command()
async def cat(self, ctx):
s = self.bot.session
async with ctx.typing(), s.get(self.url, params=self.params) as resp:
image = io.BytesIO(await resp.content.read())
ext = resp.headers['Content-Type'].partition('/')[2]
await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}')
def setup(bot):
bot.add_cog(Cat(bot))
## Instruction:
Send image in embed because aiohttp doesn't know how to parse links
## Code After:
import io
import aiohttp
import discord
from discord.ext import commands
from lxml import etree
import yaml
class Cat:
def __init__(self, bot):
self.bot = bot
with open('config.yaml') as file:
data = yaml.load(file)
self.key = data.get('cat_key', '')
self.url = 'http://thecatapi.com/api/images/get'
self.params = {'api_key': self.key,
'type': 'png,jpg',
'format': 'xml',
}
@commands.command()
async def cat(self, ctx):
session = self.bot.session
async with ctx.typing():
async with session.get(self.url, params=self.params) as resp:
root = etree.fromstring(await resp.text())
url = root.find('.//url').text
embed = discord.Embed()
embed.set_image(url=url)
await ctx.send(embed=embed)
def setup(bot):
bot.add_cog(Cat(bot))
|
import io
import aiohttp
+ import discord
from discord.ext import commands
+ from lxml import etree
import yaml
class Cat:
def __init__(self, bot):
self.bot = bot
with open('config.yaml') as file:
data = yaml.load(file)
self.key = data.get('cat_key', '')
self.url = 'http://thecatapi.com/api/images/get'
self.params = {'api_key': self.key,
- 'type': 'png,jpg'}
? ^
+ 'type': 'png,jpg',
? ^
+ 'format': 'xml',
+ }
@commands.command()
async def cat(self, ctx):
- s = self.bot.session
+ session = self.bot.session
? ++++++
+ async with ctx.typing():
- async with ctx.typing(), s.get(self.url, params=self.params) as resp:
? ^^^^^^^ ------
+ async with session.get(self.url, params=self.params) as resp:
? ++++ ^^^^ +
- image = io.BytesIO(await resp.content.read())
- ext = resp.headers['Content-Type'].partition('/')[2]
- await ctx.send(file=image, filename=f'{ctx.message.id}.{ext}')
+ root = etree.fromstring(await resp.text())
+ url = root.find('.//url').text
+ embed = discord.Embed()
+ embed.set_image(url=url)
+ await ctx.send(embed=embed)
+
def setup(bot):
bot.add_cog(Cat(bot))
|
0db43d894bfb419a7f4b538f755af47fc0b653cb
|
tests/unit/test_sharpspring.py
|
tests/unit/test_sharpspring.py
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post")
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "[email protected]",
"companyName": "Test Company",
}
sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class MockResponse:
def __init__(self, json_data, status_code):
self.json_data = json_data
self.status_code = status_code
def raise_for_status(self):
pass
def json(self):
return self.json_data
def mocked_requests_post_success(*args, **kwargs):
return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "[email protected]",
"companyName": "Test Company",
}
result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
self.assertTrue(result["result"]["creates"][0])
|
Add mock response to sharpspring test
|
Add mock response to sharpspring test
|
Python
|
apache-2.0
|
Code4SA/pmg-cms-2,Code4SA/pmg-cms-2,Code4SA/pmg-cms-2
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
+ class MockResponse:
+ def __init__(self, json_data, status_code):
+ self.json_data = json_data
+ self.status_code = status_code
+
+ def raise_for_status(self):
+ pass
+
+ def json(self):
+ return self.json_data
+
+
+ def mocked_requests_post_success(*args, **kwargs):
+
+ return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
+
+
class TestSharpspring(PMGTestCase):
- @patch("pmg.sharpspring.requests.post")
+ @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "[email protected]",
"companyName": "Test Company",
}
- sharpspring.call("createLeads", {"objects": [details]})
+ result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
+ self.assertTrue(result["result"]["creates"][0])
|
Add mock response to sharpspring test
|
## Code Before:
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post")
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "[email protected]",
"companyName": "Test Company",
}
sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
## Instruction:
Add mock response to sharpspring test
## Code After:
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class MockResponse:
def __init__(self, json_data, status_code):
self.json_data = json_data
self.status_code = status_code
def raise_for_status(self):
pass
def json(self):
return self.json_data
def mocked_requests_post_success(*args, **kwargs):
return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "[email protected]",
"companyName": "Test Company",
}
result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
self.assertTrue(result["result"]["creates"][0])
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
+ class MockResponse:
+ def __init__(self, json_data, status_code):
+ self.json_data = json_data
+ self.status_code = status_code
+
+ def raise_for_status(self):
+ pass
+
+ def json(self):
+ return self.json_data
+
+
+ def mocked_requests_post_success(*args, **kwargs):
+
+ return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
+
+
class TestSharpspring(PMGTestCase):
- @patch("pmg.sharpspring.requests.post")
+ @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "[email protected]",
"companyName": "Test Company",
}
- sharpspring.call("createLeads", {"objects": [details]})
+ result = sharpspring.call("createLeads", {"objects": [details]})
? +++++++++
post_mock.assert_called()
+ self.assertTrue(result["result"]["creates"][0])
|
3055fa16010a1b855142c2e5b866d76daee17c8f
|
markdown_gen/test/attributes_test.py
|
markdown_gen/test/attributes_test.py
|
import unittest
import markdown_gen.MardownGen as md
class AttributesTests(unittest.TestCase):
def test_italic(self):
expected = "*italic text*"
self.assertEqual(expected, md.gen_italic("italic text"))
expected = "_italic text alternative_"
self.assertEqual(expected, md.gen_italic("italic text alternative", True))
def test_bold(self):
expected = "**bold text**"
self.assertEqual(expected, md.gen_bold("bold text"))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_monspace(self):
expected = "`monospace`"
self.assertEqual(expected, md.gen_monospace("monospace"))
def test_strikethrough(self):
expected = "~~strikethrough~~"
self.assertEqual(expected, md.gen_strikethrough("strikethrough"))
if __name__ == '__main__':
unittest.main()
|
import unittest
import markdown_gen.MardownGen as md
class AttributesTests(unittest.TestCase):
def test_italic(self):
expected = "*italic text*"
self.assertEqual(expected, md.gen_italic("italic text"))
expected = "_italic text alternative_"
self.assertEqual(expected, md.gen_italic("italic text alternative", True))
def test_bold(self):
expected = "**bold text**"
self.assertEqual(expected, md.gen_bold("bold text"))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_bold_and_italic(self):
expected = "***bold and italic text***"
self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text")))
self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text")))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_monspace(self):
expected = "`monospace`"
self.assertEqual(expected, md.gen_monospace("monospace"))
def test_strikethrough(self):
expected = "~~strikethrough~~"
self.assertEqual(expected, md.gen_strikethrough("strikethrough"))
if __name__ == '__main__':
unittest.main()
|
Add test for bold and italic text
|
Add test for bold and italic text
|
Python
|
epl-1.0
|
LukasWoodtli/PyMarkdownGen
|
import unittest
import markdown_gen.MardownGen as md
class AttributesTests(unittest.TestCase):
def test_italic(self):
expected = "*italic text*"
self.assertEqual(expected, md.gen_italic("italic text"))
expected = "_italic text alternative_"
self.assertEqual(expected, md.gen_italic("italic text alternative", True))
def test_bold(self):
expected = "**bold text**"
self.assertEqual(expected, md.gen_bold("bold text"))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
+
+ def test_bold_and_italic(self):
+ expected = "***bold and italic text***"
+ self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text")))
+ self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text")))
+
+ expected = "__bold text alternative__"
+ self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_monspace(self):
expected = "`monospace`"
self.assertEqual(expected, md.gen_monospace("monospace"))
def test_strikethrough(self):
expected = "~~strikethrough~~"
self.assertEqual(expected, md.gen_strikethrough("strikethrough"))
if __name__ == '__main__':
unittest.main()
|
Add test for bold and italic text
|
## Code Before:
import unittest
import markdown_gen.MardownGen as md
class AttributesTests(unittest.TestCase):
def test_italic(self):
expected = "*italic text*"
self.assertEqual(expected, md.gen_italic("italic text"))
expected = "_italic text alternative_"
self.assertEqual(expected, md.gen_italic("italic text alternative", True))
def test_bold(self):
expected = "**bold text**"
self.assertEqual(expected, md.gen_bold("bold text"))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_monspace(self):
expected = "`monospace`"
self.assertEqual(expected, md.gen_monospace("monospace"))
def test_strikethrough(self):
expected = "~~strikethrough~~"
self.assertEqual(expected, md.gen_strikethrough("strikethrough"))
if __name__ == '__main__':
unittest.main()
## Instruction:
Add test for bold and italic text
## Code After:
import unittest
import markdown_gen.MardownGen as md
class AttributesTests(unittest.TestCase):
def test_italic(self):
expected = "*italic text*"
self.assertEqual(expected, md.gen_italic("italic text"))
expected = "_italic text alternative_"
self.assertEqual(expected, md.gen_italic("italic text alternative", True))
def test_bold(self):
expected = "**bold text**"
self.assertEqual(expected, md.gen_bold("bold text"))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_bold_and_italic(self):
expected = "***bold and italic text***"
self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text")))
self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text")))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_monspace(self):
expected = "`monospace`"
self.assertEqual(expected, md.gen_monospace("monospace"))
def test_strikethrough(self):
expected = "~~strikethrough~~"
self.assertEqual(expected, md.gen_strikethrough("strikethrough"))
if __name__ == '__main__':
unittest.main()
|
import unittest
import markdown_gen.MardownGen as md
class AttributesTests(unittest.TestCase):
def test_italic(self):
expected = "*italic text*"
self.assertEqual(expected, md.gen_italic("italic text"))
expected = "_italic text alternative_"
self.assertEqual(expected, md.gen_italic("italic text alternative", True))
def test_bold(self):
expected = "**bold text**"
self.assertEqual(expected, md.gen_bold("bold text"))
expected = "__bold text alternative__"
self.assertEqual(expected, md.gen_bold("bold text alternative", True))
+
+ def test_bold_and_italic(self):
+ expected = "***bold and italic text***"
+ self.assertEqual(expected, md.gen_italic(md.gen_bold("bold text")))
+ self.assertEqual(expected, md.gen_bold(md.gen_italic("bold text")))
+
+ expected = "__bold text alternative__"
+ self.assertEqual(expected, md.gen_bold("bold text alternative", True))
def test_monspace(self):
expected = "`monospace`"
self.assertEqual(expected, md.gen_monospace("monospace"))
def test_strikethrough(self):
expected = "~~strikethrough~~"
self.assertEqual(expected, md.gen_strikethrough("strikethrough"))
if __name__ == '__main__':
unittest.main()
|
0b82a5c10a9e728f6f5424429a70fd2951c9b5c5
|
pythonmisp/__init__.py
|
pythonmisp/__init__.py
|
from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute
|
from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError
|
Add MispTransportError in package import
|
Add MispTransportError in package import
|
Python
|
apache-2.0
|
nbareil/python-misp
|
- from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute
+ from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError
|
Add MispTransportError in package import
|
## Code Before:
from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute
## Instruction:
Add MispTransportError in package import
## Code After:
from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError
|
- from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute
+ from .misp import MispTag, MispEvent, MispServer, MispAttribute, MispShadowAttribute, MispTransportError
? ++++++++++++++++++++
|
110c362e3e8436700707c2306d115b3b2476a79d
|
core/models.py
|
core/models.py
|
from os import makedirs
from os.path import join, exists
from urllib import urlretrieve
from django.conf import settings
from social_auth.signals import socialauth_registered
def create_profile(sender, user, response, details, **kwargs):
try:
# twitter
photo_url = response["profile_image_url"]
photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1))
except KeyError:
# facebook
photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"]
path = join(settings.MEDIA_ROOT, "photos")
if not exists(path):
makedirs(path)
urlretrieve(photo_url, join(path, str(user.id)))
socialauth_registered.connect(create_profile, sender=None)
|
from os import makedirs
from os.path import join, exists
from urllib import urlretrieve
from django.conf import settings
from django.contrib.auth.models import User
from django.db import models
from django.db.models.signals import post_save
from django.dispatch import receiver
from social_auth.signals import socialauth_registered
class Account(models.Model):
"""
A user's account balance.
"""
user = models.OneToOneField(User)
balance = models.IntegerField(default=5000)
@receiver(post_save, sender=User)
def user_saved(sender, **kwargs):
"""
Create an initial account balance for new users.
"""
Account.objects.get_or_create(user=kwargs["instance"])
@receiver(socialauth_registered, sender=None)
def avatar(sender, user, response, details, **kwargs):
"""
Download the user's Twitter or Facebook avatar once they've
authenticated via either service.
"""
try:
# twitter
photo_url = response["profile_image_url"]
photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1))
except KeyError:
# facebook
photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"]
path = join(settings.MEDIA_ROOT, "photos")
if not exists(path):
makedirs(path)
urlretrieve(photo_url, join(path, str(user.id)))
|
Add initial account balance for users.
|
Add initial account balance for users.
|
Python
|
bsd-2-clause
|
stephenmcd/gamblor,stephenmcd/gamblor
|
from os import makedirs
from os.path import join, exists
from urllib import urlretrieve
from django.conf import settings
+ from django.contrib.auth.models import User
+ from django.db import models
+ from django.db.models.signals import post_save
+ from django.dispatch import receiver
from social_auth.signals import socialauth_registered
+ class Account(models.Model):
+ """
+ A user's account balance.
+ """
+ user = models.OneToOneField(User)
+ balance = models.IntegerField(default=5000)
+
+
+ @receiver(post_save, sender=User)
+ def user_saved(sender, **kwargs):
+ """
+ Create an initial account balance for new users.
+ """
+ Account.objects.get_or_create(user=kwargs["instance"])
+
+
+ @receiver(socialauth_registered, sender=None)
- def create_profile(sender, user, response, details, **kwargs):
+ def avatar(sender, user, response, details, **kwargs):
+ """
+ Download the user's Twitter or Facebook avatar once they've
+ authenticated via either service.
+ """
try:
# twitter
photo_url = response["profile_image_url"]
photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1))
except KeyError:
# facebook
photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"]
path = join(settings.MEDIA_ROOT, "photos")
if not exists(path):
makedirs(path)
urlretrieve(photo_url, join(path, str(user.id)))
- socialauth_registered.connect(create_profile, sender=None)
-
|
Add initial account balance for users.
|
## Code Before:
from os import makedirs
from os.path import join, exists
from urllib import urlretrieve
from django.conf import settings
from social_auth.signals import socialauth_registered
def create_profile(sender, user, response, details, **kwargs):
try:
# twitter
photo_url = response["profile_image_url"]
photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1))
except KeyError:
# facebook
photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"]
path = join(settings.MEDIA_ROOT, "photos")
if not exists(path):
makedirs(path)
urlretrieve(photo_url, join(path, str(user.id)))
socialauth_registered.connect(create_profile, sender=None)
## Instruction:
Add initial account balance for users.
## Code After:
from os import makedirs
from os.path import join, exists
from urllib import urlretrieve
from django.conf import settings
from django.contrib.auth.models import User
from django.db import models
from django.db.models.signals import post_save
from django.dispatch import receiver
from social_auth.signals import socialauth_registered
class Account(models.Model):
"""
A user's account balance.
"""
user = models.OneToOneField(User)
balance = models.IntegerField(default=5000)
@receiver(post_save, sender=User)
def user_saved(sender, **kwargs):
"""
Create an initial account balance for new users.
"""
Account.objects.get_or_create(user=kwargs["instance"])
@receiver(socialauth_registered, sender=None)
def avatar(sender, user, response, details, **kwargs):
"""
Download the user's Twitter or Facebook avatar once they've
authenticated via either service.
"""
try:
# twitter
photo_url = response["profile_image_url"]
photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1))
except KeyError:
# facebook
photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"]
path = join(settings.MEDIA_ROOT, "photos")
if not exists(path):
makedirs(path)
urlretrieve(photo_url, join(path, str(user.id)))
|
from os import makedirs
from os.path import join, exists
from urllib import urlretrieve
from django.conf import settings
+ from django.contrib.auth.models import User
+ from django.db import models
+ from django.db.models.signals import post_save
+ from django.dispatch import receiver
from social_auth.signals import socialauth_registered
+ class Account(models.Model):
+ """
+ A user's account balance.
+ """
+ user = models.OneToOneField(User)
+ balance = models.IntegerField(default=5000)
+
+
+ @receiver(post_save, sender=User)
+ def user_saved(sender, **kwargs):
+ """
+ Create an initial account balance for new users.
+ """
+ Account.objects.get_or_create(user=kwargs["instance"])
+
+
+ @receiver(socialauth_registered, sender=None)
- def create_profile(sender, user, response, details, **kwargs):
? ^^^ ^^^ -----
+ def avatar(sender, user, response, details, **kwargs):
? ^^ ^
+ """
+ Download the user's Twitter or Facebook avatar once they've
+ authenticated via either service.
+ """
try:
# twitter
photo_url = response["profile_image_url"]
photo_url = "_reasonably_small".join(photo_url.rsplit("_normal", 1))
except KeyError:
# facebook
photo_url = "http://graph.facebook.com/%s/picture?type=large" % response["id"]
path = join(settings.MEDIA_ROOT, "photos")
if not exists(path):
makedirs(path)
urlretrieve(photo_url, join(path, str(user.id)))
-
- socialauth_registered.connect(create_profile, sender=None)
|
7198133cf9d24f3d29d300366951b7eac8b2547f
|
alburnum/maas/viscera/users.py
|
alburnum/maas/viscera/users.py
|
"""Objects for users."""
__all__ = [
"User",
"Users",
]
from . import (
check,
Object,
ObjectField,
ObjectSet,
ObjectType,
)
class UsersType(ObjectType):
"""Metaclass for `Users`."""
def __iter__(cls):
return map(cls._object, cls._handler.read())
def create(cls, username, email, password, *, is_superuser=False):
data = cls._handler.create(
username=username, email=email, password=password,
is_superuser='1' if is_superuser else '0')
return cls._object(data)
class Users(ObjectSet, metaclass=UsersType):
"""The set of users."""
@classmethod
def read(cls):
return cls(cls)
class User(Object):
"""A user."""
username = ObjectField.Checked(
"username", check(str), check(str))
email = ObjectField.Checked(
"email", check(str), check(str))
is_admin = ObjectField.Checked(
"is_superuser", check(bool), check(bool))
|
"""Objects for users."""
__all__ = [
"User",
"Users",
]
from . import (
check,
Object,
ObjectField,
ObjectSet,
ObjectType,
)
class UsersType(ObjectType):
"""Metaclass for `Users`."""
def __iter__(cls):
return map(cls._object, cls._handler.read())
def create(cls, username, password, *, email=None, is_superuser=False):
if email is None:
email = "%[email protected]" % username
data = cls._handler.create(
username=username, email=email, password=password,
is_superuser='1' if is_superuser else '0')
return cls._object(data)
class Users(ObjectSet, metaclass=UsersType):
"""The set of users."""
@classmethod
def read(cls):
return cls(cls)
class User(Object):
"""A user."""
username = ObjectField.Checked(
"username", check(str), check(str))
email = ObjectField.Checked(
"email", check(str), check(str))
is_superuser = ObjectField.Checked(
"is_superuser", check(bool), check(bool))
|
Change to is_superuser and make email optional.
|
Change to is_superuser and make email optional.
|
Python
|
agpl-3.0
|
blakerouse/python-libmaas,alburnum/alburnum-maas-client
|
"""Objects for users."""
__all__ = [
"User",
"Users",
]
from . import (
check,
Object,
ObjectField,
ObjectSet,
ObjectType,
)
class UsersType(ObjectType):
"""Metaclass for `Users`."""
def __iter__(cls):
return map(cls._object, cls._handler.read())
- def create(cls, username, email, password, *, is_superuser=False):
+ def create(cls, username, password, *, email=None, is_superuser=False):
+ if email is None:
+ email = "%[email protected]" % username
data = cls._handler.create(
username=username, email=email, password=password,
is_superuser='1' if is_superuser else '0')
return cls._object(data)
class Users(ObjectSet, metaclass=UsersType):
"""The set of users."""
@classmethod
def read(cls):
return cls(cls)
class User(Object):
"""A user."""
username = ObjectField.Checked(
"username", check(str), check(str))
email = ObjectField.Checked(
"email", check(str), check(str))
- is_admin = ObjectField.Checked(
+ is_superuser = ObjectField.Checked(
"is_superuser", check(bool), check(bool))
|
Change to is_superuser and make email optional.
|
## Code Before:
"""Objects for users."""
__all__ = [
"User",
"Users",
]
from . import (
check,
Object,
ObjectField,
ObjectSet,
ObjectType,
)
class UsersType(ObjectType):
"""Metaclass for `Users`."""
def __iter__(cls):
return map(cls._object, cls._handler.read())
def create(cls, username, email, password, *, is_superuser=False):
data = cls._handler.create(
username=username, email=email, password=password,
is_superuser='1' if is_superuser else '0')
return cls._object(data)
class Users(ObjectSet, metaclass=UsersType):
"""The set of users."""
@classmethod
def read(cls):
return cls(cls)
class User(Object):
"""A user."""
username = ObjectField.Checked(
"username", check(str), check(str))
email = ObjectField.Checked(
"email", check(str), check(str))
is_admin = ObjectField.Checked(
"is_superuser", check(bool), check(bool))
## Instruction:
Change to is_superuser and make email optional.
## Code After:
"""Objects for users."""
__all__ = [
"User",
"Users",
]
from . import (
check,
Object,
ObjectField,
ObjectSet,
ObjectType,
)
class UsersType(ObjectType):
"""Metaclass for `Users`."""
def __iter__(cls):
return map(cls._object, cls._handler.read())
def create(cls, username, password, *, email=None, is_superuser=False):
if email is None:
email = "%[email protected]" % username
data = cls._handler.create(
username=username, email=email, password=password,
is_superuser='1' if is_superuser else '0')
return cls._object(data)
class Users(ObjectSet, metaclass=UsersType):
"""The set of users."""
@classmethod
def read(cls):
return cls(cls)
class User(Object):
"""A user."""
username = ObjectField.Checked(
"username", check(str), check(str))
email = ObjectField.Checked(
"email", check(str), check(str))
is_superuser = ObjectField.Checked(
"is_superuser", check(bool), check(bool))
|
"""Objects for users."""
__all__ = [
"User",
"Users",
]
from . import (
check,
Object,
ObjectField,
ObjectSet,
ObjectType,
)
class UsersType(ObjectType):
"""Metaclass for `Users`."""
def __iter__(cls):
return map(cls._object, cls._handler.read())
- def create(cls, username, email, password, *, is_superuser=False):
? -------
+ def create(cls, username, password, *, email=None, is_superuser=False):
? ++++++++++++
+ if email is None:
+ email = "%[email protected]" % username
data = cls._handler.create(
username=username, email=email, password=password,
is_superuser='1' if is_superuser else '0')
return cls._object(data)
class Users(ObjectSet, metaclass=UsersType):
"""The set of users."""
@classmethod
def read(cls):
return cls(cls)
class User(Object):
"""A user."""
username = ObjectField.Checked(
"username", check(str), check(str))
email = ObjectField.Checked(
"email", check(str), check(str))
- is_admin = ObjectField.Checked(
? ^^^^^
+ is_superuser = ObjectField.Checked(
? ^^^^^^^^^
"is_superuser", check(bool), check(bool))
|
cc92b1770acdc5a34eb32c596c0b2ece6bf32b0f
|
qiprofile_rest/server/settings.py
|
qiprofile_rest/server/settings.py
|
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
MONGO_DBNAME = 'qiprofile'
else:
MONGO_DBNAME = 'qiprofile_test'
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
|
"""This ``settings`` file specifies the Eve configuration."""
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
MONGO_DBNAME = 'qiprofile'
else:
MONGO_DBNAME = 'qiprofile_test'
# The MongoDB host default is localhost, but can be reset
# by the MONGO_HOST environment variable.
host = os.getenv('MONGO_HOST')
if host:
MONGO_HOST = host
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
|
Allow MONGO_HOST env var override.
|
Allow MONGO_HOST env var override.
|
Python
|
bsd-2-clause
|
ohsu-qin/qirest,ohsu-qin/qiprofile-rest
|
+ """This ``settings`` file specifies the Eve configuration."""
+
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
- MONGO_DBNAME = 'qiprofile'
+ MONGO_DBNAME = 'qiprofile'
else:
- MONGO_DBNAME = 'qiprofile_test'
+ MONGO_DBNAME = 'qiprofile_test'
+
+ # The MongoDB host default is localhost, but can be reset
+ # by the MONGO_HOST environment variable.
+ host = os.getenv('MONGO_HOST')
+ if host:
+ MONGO_HOST = host
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
|
Allow MONGO_HOST env var override.
|
## Code Before:
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
MONGO_DBNAME = 'qiprofile'
else:
MONGO_DBNAME = 'qiprofile_test'
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
## Instruction:
Allow MONGO_HOST env var override.
## Code After:
"""This ``settings`` file specifies the Eve configuration."""
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
MONGO_DBNAME = 'qiprofile'
else:
MONGO_DBNAME = 'qiprofile_test'
# The MongoDB host default is localhost, but can be reset
# by the MONGO_HOST environment variable.
host = os.getenv('MONGO_HOST')
if host:
MONGO_HOST = host
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
|
+ """This ``settings`` file specifies the Eve configuration."""
+
import os
# The run environment default is production.
# Modify this by setting the NODE_ENV environment variable.
env = os.getenv('NODE_ENV') or 'production'
# The MongoDB database.
if env == 'production':
- MONGO_DBNAME = 'qiprofile'
+ MONGO_DBNAME = 'qiprofile'
? ++
else:
- MONGO_DBNAME = 'qiprofile_test'
+ MONGO_DBNAME = 'qiprofile_test'
? ++
+
+ # The MongoDB host default is localhost, but can be reset
+ # by the MONGO_HOST environment variable.
+ host = os.getenv('MONGO_HOST')
+ if host:
+ MONGO_HOST = host
# Even though the domain is defined by the Eve MongoEngine
# adapter, a DOMAIN setting is required by Eve. This setting
# is only used to avoid an Eve complaint about a missing domain.
DOMAIN = {'eve-mongoengine': {}}
|
a28a29af31b1ea604ed97544b2d84a39c9ba3e7b
|
automation/src/rabird/automation/selenium/webelement.py
|
automation/src/rabird/automation/selenium/webelement.py
|
'''
@date 2014-11-16
@author Hong-she Liang <[email protected]>
'''
# Import the global selenium unit, not our selenium .
global_selenium = __import__('selenium')
import types
import time
def set_attribute(self, name, value):
value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'"
value = value.replace("\n", r"\n")
value = value.replace("\r", r"\r")
value = value.replace("\t", r"\t")
script = "arguments[0].setAttribute('%s', '%s')" % (name, value)
self._parent.execute_script(script, self)
def force_focus(self):
global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform()
def force_click(self):
self._parent.execute_script("arguments[0].click();", self);
|
'''
@date 2014-11-16
@author Hong-she Liang <[email protected]>
'''
# Import the global selenium unit, not our selenium .
global_selenium = __import__('selenium')
import types
import time
def set_attribute(self, name, value):
value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'"
value = value.replace("\n", r"\n")
value = value.replace("\r", r"\r")
value = value.replace("\t", r"\t")
script = "arguments[0].setAttribute('%s', '%s');" % (name, value)
self._parent.execute_script(script, self)
def force_focus(self):
self._parent.execute_script("arguments[0].focus();", self);
def force_click(self):
self._parent.execute_script("arguments[0].click();", self);
|
Use javascript to force focus on an element, because the action chains seems take no effect!
|
Use javascript to force focus on an element, because the action chains seems take no effect!
|
Python
|
apache-2.0
|
starofrainnight/rabird.core,starofrainnight/rabird.auto
|
'''
@date 2014-11-16
@author Hong-she Liang <[email protected]>
'''
# Import the global selenium unit, not our selenium .
global_selenium = __import__('selenium')
import types
import time
def set_attribute(self, name, value):
value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'"
value = value.replace("\n", r"\n")
value = value.replace("\r", r"\r")
value = value.replace("\t", r"\t")
- script = "arguments[0].setAttribute('%s', '%s')" % (name, value)
+ script = "arguments[0].setAttribute('%s', '%s');" % (name, value)
self._parent.execute_script(script, self)
def force_focus(self):
- global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform()
+ self._parent.execute_script("arguments[0].focus();", self);
def force_click(self):
self._parent.execute_script("arguments[0].click();", self);
|
Use javascript to force focus on an element, because the action chains seems take no effect!
|
## Code Before:
'''
@date 2014-11-16
@author Hong-she Liang <[email protected]>
'''
# Import the global selenium unit, not our selenium .
global_selenium = __import__('selenium')
import types
import time
def set_attribute(self, name, value):
value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'"
value = value.replace("\n", r"\n")
value = value.replace("\r", r"\r")
value = value.replace("\t", r"\t")
script = "arguments[0].setAttribute('%s', '%s')" % (name, value)
self._parent.execute_script(script, self)
def force_focus(self):
global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform()
def force_click(self):
self._parent.execute_script("arguments[0].click();", self);
## Instruction:
Use javascript to force focus on an element, because the action chains seems take no effect!
## Code After:
'''
@date 2014-11-16
@author Hong-she Liang <[email protected]>
'''
# Import the global selenium unit, not our selenium .
global_selenium = __import__('selenium')
import types
import time
def set_attribute(self, name, value):
value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'"
value = value.replace("\n", r"\n")
value = value.replace("\r", r"\r")
value = value.replace("\t", r"\t")
script = "arguments[0].setAttribute('%s', '%s');" % (name, value)
self._parent.execute_script(script, self)
def force_focus(self):
self._parent.execute_script("arguments[0].focus();", self);
def force_click(self):
self._parent.execute_script("arguments[0].click();", self);
|
'''
@date 2014-11-16
@author Hong-she Liang <[email protected]>
'''
# Import the global selenium unit, not our selenium .
global_selenium = __import__('selenium')
import types
import time
def set_attribute(self, name, value):
value = value.replace(r"'", r"\'") # Replace all r"'" with r"\'"
value = value.replace("\n", r"\n")
value = value.replace("\r", r"\r")
value = value.replace("\t", r"\t")
- script = "arguments[0].setAttribute('%s', '%s')" % (name, value)
+ script = "arguments[0].setAttribute('%s', '%s');" % (name, value)
? +
self._parent.execute_script(script, self)
def force_focus(self):
- global_selenium.webdriver.ActionChains(self._parent).move_to_element(self).perform()
+ self._parent.execute_script("arguments[0].focus();", self);
def force_click(self):
self._parent.execute_script("arguments[0].click();", self);
|
53a2d8781e3e5d8e5879d4ef7c62752483323cf9
|
bfg9000/shell/__init__.py
|
bfg9000/shell/__init__.py
|
import os
import subprocess
from ..platform_name import platform_name
if platform_name() == 'windows':
from .windows import *
else:
from .posix import *
class shell_list(list):
"""A special subclass of list used to mark that this command line uses
special shell characters."""
pass
def execute(args, shell=False, env=None, quiet=False):
if quiet:
devnull = open(os.devnull, 'wb')
try:
return subprocess.check_output(
args, universal_newlines=True, shell=shell, env=env
)
except:
if quiet:
devnull.close()
raise
|
import os
import subprocess
from ..platform_name import platform_name
if platform_name() == 'windows':
from .windows import *
else:
from .posix import *
class shell_list(list):
"""A special subclass of list used to mark that this command line uses
special shell characters."""
pass
def execute(args, shell=False, env=None, quiet=False):
stderr = None
if quiet:
stderr = open(os.devnull, 'wb')
try:
return subprocess.check_output(
args, universal_newlines=True, shell=shell, env=env, stderr=stderr
)
except:
if quiet:
stderr.close()
raise
|
Fix "quiet" mode for shell.execute()
|
Fix "quiet" mode for shell.execute()
|
Python
|
bsd-3-clause
|
jimporter/bfg9000,jimporter/bfg9000,jimporter/bfg9000,jimporter/bfg9000
|
import os
import subprocess
from ..platform_name import platform_name
if platform_name() == 'windows':
from .windows import *
else:
from .posix import *
class shell_list(list):
"""A special subclass of list used to mark that this command line uses
special shell characters."""
pass
def execute(args, shell=False, env=None, quiet=False):
+ stderr = None
if quiet:
- devnull = open(os.devnull, 'wb')
+ stderr = open(os.devnull, 'wb')
try:
return subprocess.check_output(
- args, universal_newlines=True, shell=shell, env=env
+ args, universal_newlines=True, shell=shell, env=env, stderr=stderr
)
except:
if quiet:
- devnull.close()
+ stderr.close()
raise
|
Fix "quiet" mode for shell.execute()
|
## Code Before:
import os
import subprocess
from ..platform_name import platform_name
if platform_name() == 'windows':
from .windows import *
else:
from .posix import *
class shell_list(list):
"""A special subclass of list used to mark that this command line uses
special shell characters."""
pass
def execute(args, shell=False, env=None, quiet=False):
if quiet:
devnull = open(os.devnull, 'wb')
try:
return subprocess.check_output(
args, universal_newlines=True, shell=shell, env=env
)
except:
if quiet:
devnull.close()
raise
## Instruction:
Fix "quiet" mode for shell.execute()
## Code After:
import os
import subprocess
from ..platform_name import platform_name
if platform_name() == 'windows':
from .windows import *
else:
from .posix import *
class shell_list(list):
"""A special subclass of list used to mark that this command line uses
special shell characters."""
pass
def execute(args, shell=False, env=None, quiet=False):
stderr = None
if quiet:
stderr = open(os.devnull, 'wb')
try:
return subprocess.check_output(
args, universal_newlines=True, shell=shell, env=env, stderr=stderr
)
except:
if quiet:
stderr.close()
raise
|
import os
import subprocess
from ..platform_name import platform_name
if platform_name() == 'windows':
from .windows import *
else:
from .posix import *
class shell_list(list):
"""A special subclass of list used to mark that this command line uses
special shell characters."""
pass
def execute(args, shell=False, env=None, quiet=False):
+ stderr = None
if quiet:
- devnull = open(os.devnull, 'wb')
? ^^^^^
+ stderr = open(os.devnull, 'wb')
? ++ ^^
try:
return subprocess.check_output(
- args, universal_newlines=True, shell=shell, env=env
+ args, universal_newlines=True, shell=shell, env=env, stderr=stderr
? +++++++++++++++
)
except:
if quiet:
- devnull.close()
? ^^^^^
+ stderr.close()
? ++ ^^
raise
|
70c98a42326471d3ed615def61954905673c5972
|
typhon/nonlte/__init__.py
|
typhon/nonlte/__init__.py
|
from .version import __version__
try:
__ATRASU_SETUP__
except:
__ATRASU_SETUP__ = False
if not __ATRASU_SETUP__:
from . import spectra
from . import setup_atmosphere
from . import const
from . import nonltecalc
from . import mathmatics
from . import rtc
|
try:
__ATRASU_SETUP__
except:
__ATRASU_SETUP__ = False
if not __ATRASU_SETUP__:
from . import spectra
from . import setup_atmosphere
from . import const
from . import nonltecalc
from . import mathmatics
from . import rtc
|
Remove import of removed version module.
|
Remove import of removed version module.
|
Python
|
mit
|
atmtools/typhon,atmtools/typhon
|
-
- from .version import __version__
try:
__ATRASU_SETUP__
except:
__ATRASU_SETUP__ = False
if not __ATRASU_SETUP__:
from . import spectra
from . import setup_atmosphere
from . import const
from . import nonltecalc
from . import mathmatics
from . import rtc
|
Remove import of removed version module.
|
## Code Before:
from .version import __version__
try:
__ATRASU_SETUP__
except:
__ATRASU_SETUP__ = False
if not __ATRASU_SETUP__:
from . import spectra
from . import setup_atmosphere
from . import const
from . import nonltecalc
from . import mathmatics
from . import rtc
## Instruction:
Remove import of removed version module.
## Code After:
try:
__ATRASU_SETUP__
except:
__ATRASU_SETUP__ = False
if not __ATRASU_SETUP__:
from . import spectra
from . import setup_atmosphere
from . import const
from . import nonltecalc
from . import mathmatics
from . import rtc
|
-
- from .version import __version__
try:
__ATRASU_SETUP__
except:
__ATRASU_SETUP__ = False
if not __ATRASU_SETUP__:
from . import spectra
from . import setup_atmosphere
from . import const
from . import nonltecalc
from . import mathmatics
from . import rtc
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.