commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 10
2.94k
| new_contents
stringlengths 21
3.18k
| subject
stringlengths 16
444
| message
stringlengths 17
2.63k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43k
| ndiff
stringlengths 52
3.32k
| instruction
stringlengths 16
444
| content
stringlengths 133
4.32k
| fuzzy_diff
stringlengths 16
3.18k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4ff1eb00f8e212d280ac858feb4efcc795d97d80
|
tests/test_models.py
|
tests/test_models.py
|
import pytest
from suddendev.models import GameController
def test_create_game(session):
pass
|
import pytest
from suddendev.models import GameSetup
def test_create_game(session):
game_setup = GameSetup('ASDF')
assert game_setup.player_count == 1
|
Fix broken import in model tests.
|
[NG] Fix broken import in model tests.
|
Python
|
mit
|
SuddenDevs/SuddenDev,SuddenDevs/SuddenDev,SuddenDevs/SuddenDev,SuddenDevs/SuddenDev
|
import pytest
- from suddendev.models import GameController
+ from suddendev.models import GameSetup
def test_create_game(session):
- pass
+ game_setup = GameSetup('ASDF')
+ assert game_setup.player_count == 1
|
Fix broken import in model tests.
|
## Code Before:
import pytest
from suddendev.models import GameController
def test_create_game(session):
pass
## Instruction:
Fix broken import in model tests.
## Code After:
import pytest
from suddendev.models import GameSetup
def test_create_game(session):
game_setup = GameSetup('ASDF')
assert game_setup.player_count == 1
|
...
import pytest
from suddendev.models import GameSetup
...
def test_create_game(session):
game_setup = GameSetup('ASDF')
assert game_setup.player_count == 1
...
|
f696e320d66f375779692ec073f15d3d6d466059
|
edx_data_research/parsing/parse_sql.py
|
edx_data_research/parsing/parse_sql.py
|
import subprocess
from edx_data_research.parsing.parse import Parse
class SQL(Parse):
def __init__(self, args):
super(SQL, self).__init__(args)
self._collections = args.collection
self.sql_file = args.sql_file
def migrate(self):
subprocess.check_call(['mongoimport', '-d', self.db_name, '-c',
self._collections, '--type', 'tsv', '--file',
self.sql_file, '--headerline'])
|
import subprocess
from edx_data_research.parsing.parse import Parse
class SQL(Parse):
def __init__(self, args):
super(SQL, self).__init__(args)
self._collections = args.collection
self.sql_file = args.sql_file
def migrate(self):
subprocess.check_call(['mongoimport', '-d', self.db_name, '-c',
self._collections, '--type', 'tsv', '--file',
self.sql_file, '--headerline', '--drop'])
|
Update mongo import of files to drop existing collection first
|
Update mongo import of files to drop existing collection first
|
Python
|
mit
|
McGillX/edx_data_research,McGillX/edx_data_research,McGillX/edx_data_research
|
import subprocess
from edx_data_research.parsing.parse import Parse
class SQL(Parse):
def __init__(self, args):
super(SQL, self).__init__(args)
self._collections = args.collection
self.sql_file = args.sql_file
def migrate(self):
subprocess.check_call(['mongoimport', '-d', self.db_name, '-c',
self._collections, '--type', 'tsv', '--file',
- self.sql_file, '--headerline'])
+ self.sql_file, '--headerline', '--drop'])
|
Update mongo import of files to drop existing collection first
|
## Code Before:
import subprocess
from edx_data_research.parsing.parse import Parse
class SQL(Parse):
def __init__(self, args):
super(SQL, self).__init__(args)
self._collections = args.collection
self.sql_file = args.sql_file
def migrate(self):
subprocess.check_call(['mongoimport', '-d', self.db_name, '-c',
self._collections, '--type', 'tsv', '--file',
self.sql_file, '--headerline'])
## Instruction:
Update mongo import of files to drop existing collection first
## Code After:
import subprocess
from edx_data_research.parsing.parse import Parse
class SQL(Parse):
def __init__(self, args):
super(SQL, self).__init__(args)
self._collections = args.collection
self.sql_file = args.sql_file
def migrate(self):
subprocess.check_call(['mongoimport', '-d', self.db_name, '-c',
self._collections, '--type', 'tsv', '--file',
self.sql_file, '--headerline', '--drop'])
|
...
self._collections, '--type', 'tsv', '--file',
self.sql_file, '--headerline', '--drop'])
...
|
12a9ef54d82d9508852e5596dbb9df321986e067
|
tests/test_heroku.py
|
tests/test_heroku.py
|
"""Tests for the Wallace API."""
import subprocess
import re
import requests
class TestHeroku(object):
"""The Heroku test class."""
def test_sandbox(self):
"""Launch the experiment on Heroku."""
sandbox_output = subprocess.check_output(
"cd examples/bartlett1932; wallace sandbox --verbose", shell=True)
id = re.search(
'Running as experiment (.*)...', sandbox_output).group(1)
r = requests.get("http://{}.herokuapp.com/summary".format(id))
assert r.json()['status'] == []
subprocess.call(
"heroku apps:destroy --app {} --confirm {}".format(id),
shell=True)
|
"""Tests for the Wallace API."""
import subprocess
import re
import os
import requests
class TestHeroku(object):
"""The Heroku test class."""
sandbox_output = subprocess.check_output(
"cd examples/bartlett1932; wallace sandbox --verbose", shell=True)
os.environ['app_id'] = re.search(
'Running as experiment (.*)...', sandbox_output).group(1)
@classmethod
def teardown_class(cls):
"""Remove the app from Heroku."""
app_id = os.environ['app_id']
subprocess.call(
"heroku apps:destroy --app {} --confirm {}".format(app_id, app_id),
shell=True)
def test_summary(self):
"""Launch the experiment on Heroku."""
app_id = os.environ['app_id']
r = requests.get("http://{}.herokuapp.com/summary".format(app_id))
assert r.json()['status'] == []
|
Refactor Heroku tests to have shared setup
|
Refactor Heroku tests to have shared setup
|
Python
|
mit
|
Dallinger/Dallinger,berkeley-cocosci/Wallace,jcpeterson/Dallinger,jcpeterson/Dallinger,suchow/Wallace,berkeley-cocosci/Wallace,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,berkeley-cocosci/Wallace,Dallinger/Dallinger,jcpeterson/Dallinger,suchow/Wallace,suchow/Wallace,jcpeterson/Dallinger,Dallinger/Dallinger
|
"""Tests for the Wallace API."""
import subprocess
import re
+ import os
import requests
class TestHeroku(object):
"""The Heroku test class."""
+ sandbox_output = subprocess.check_output(
+ "cd examples/bartlett1932; wallace sandbox --verbose", shell=True)
+
+ os.environ['app_id'] = re.search(
+ 'Running as experiment (.*)...', sandbox_output).group(1)
+
+ @classmethod
+ def teardown_class(cls):
+ """Remove the app from Heroku."""
+ app_id = os.environ['app_id']
+ subprocess.call(
+ "heroku apps:destroy --app {} --confirm {}".format(app_id, app_id),
+ shell=True)
+
- def test_sandbox(self):
+ def test_summary(self):
"""Launch the experiment on Heroku."""
+ app_id = os.environ['app_id']
- sandbox_output = subprocess.check_output(
- "cd examples/bartlett1932; wallace sandbox --verbose", shell=True)
-
- id = re.search(
- 'Running as experiment (.*)...', sandbox_output).group(1)
-
- r = requests.get("http://{}.herokuapp.com/summary".format(id))
+ r = requests.get("http://{}.herokuapp.com/summary".format(app_id))
-
assert r.json()['status'] == []
- subprocess.call(
- "heroku apps:destroy --app {} --confirm {}".format(id),
- shell=True)
-
|
Refactor Heroku tests to have shared setup
|
## Code Before:
"""Tests for the Wallace API."""
import subprocess
import re
import requests
class TestHeroku(object):
"""The Heroku test class."""
def test_sandbox(self):
"""Launch the experiment on Heroku."""
sandbox_output = subprocess.check_output(
"cd examples/bartlett1932; wallace sandbox --verbose", shell=True)
id = re.search(
'Running as experiment (.*)...', sandbox_output).group(1)
r = requests.get("http://{}.herokuapp.com/summary".format(id))
assert r.json()['status'] == []
subprocess.call(
"heroku apps:destroy --app {} --confirm {}".format(id),
shell=True)
## Instruction:
Refactor Heroku tests to have shared setup
## Code After:
"""Tests for the Wallace API."""
import subprocess
import re
import os
import requests
class TestHeroku(object):
"""The Heroku test class."""
sandbox_output = subprocess.check_output(
"cd examples/bartlett1932; wallace sandbox --verbose", shell=True)
os.environ['app_id'] = re.search(
'Running as experiment (.*)...', sandbox_output).group(1)
@classmethod
def teardown_class(cls):
"""Remove the app from Heroku."""
app_id = os.environ['app_id']
subprocess.call(
"heroku apps:destroy --app {} --confirm {}".format(app_id, app_id),
shell=True)
def test_summary(self):
"""Launch the experiment on Heroku."""
app_id = os.environ['app_id']
r = requests.get("http://{}.herokuapp.com/summary".format(app_id))
assert r.json()['status'] == []
|
...
import re
import os
import requests
...
sandbox_output = subprocess.check_output(
"cd examples/bartlett1932; wallace sandbox --verbose", shell=True)
os.environ['app_id'] = re.search(
'Running as experiment (.*)...', sandbox_output).group(1)
@classmethod
def teardown_class(cls):
"""Remove the app from Heroku."""
app_id = os.environ['app_id']
subprocess.call(
"heroku apps:destroy --app {} --confirm {}".format(app_id, app_id),
shell=True)
def test_summary(self):
"""Launch the experiment on Heroku."""
app_id = os.environ['app_id']
r = requests.get("http://{}.herokuapp.com/summary".format(app_id))
assert r.json()['status'] == []
...
|
a20fc95d3a4dc194ef4f6d227976ff5bba229aaa
|
feincms/__init__.py
|
feincms/__init__.py
|
VERSION = (1, 4, 2)
__version__ = '.'.join(map(str, VERSION))
class LazySettings(object):
def _load_settings(self):
from feincms import default_settings
from django.conf import settings as django_settings
for key in dir(default_settings):
if not key.startswith(('FEINCMS_', '_HACK_')):
continue
setattr(self, key, getattr(django_settings, key,
getattr(default_settings, key)))
def __getattr__(self, attr):
self._load_settings()
del self.__class__.__getattr__
return self.__dict__[attr]
settings = LazySettings()
COMPLETELY_LOADED = False
def ensure_completely_loaded():
"""
This method ensures all models are completely loaded
FeinCMS requires Django to be completely initialized before proceeding,
because of the extension mechanism and the dynamically created content
types.
For more informations, have a look at issue #23 on github:
http://github.com/feincms/feincms/issues#issue/23
"""
global COMPLETELY_LOADED
if COMPLETELY_LOADED:
return True
from django.core.management.validation import get_validation_errors
from StringIO import StringIO
get_validation_errors(StringIO(), None)
COMPLETELY_LOADED = True
return True
|
VERSION = (1, 4, 2)
__version__ = '.'.join(map(str, VERSION))
class LazySettings(object):
def _load_settings(self):
from feincms import default_settings
from django.conf import settings as django_settings
for key in dir(default_settings):
if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')):
continue
setattr(self, key, getattr(django_settings, key,
getattr(default_settings, key)))
def __getattr__(self, attr):
self._load_settings()
del self.__class__.__getattr__
return self.__dict__[attr]
settings = LazySettings()
COMPLETELY_LOADED = False
def ensure_completely_loaded():
"""
This method ensures all models are completely loaded
FeinCMS requires Django to be completely initialized before proceeding,
because of the extension mechanism and the dynamically created content
types.
For more informations, have a look at issue #23 on github:
http://github.com/feincms/feincms/issues#issue/23
"""
global COMPLETELY_LOADED
if COMPLETELY_LOADED:
return True
from django.core.management.validation import get_validation_errors
from StringIO import StringIO
get_validation_errors(StringIO(), None)
COMPLETELY_LOADED = True
return True
|
Fix a Python 2.4 incompatibility that snuck in
|
Fix a Python 2.4 incompatibility that snuck in
Fixes github issue #214
|
Python
|
bsd-3-clause
|
matthiask/django-content-editor,nickburlett/feincms,matthiask/django-content-editor,matthiask/feincms2-content,michaelkuty/feincms,matthiask/feincms2-content,pjdelport/feincms,mjl/feincms,joshuajonah/feincms,matthiask/django-content-editor,feincms/feincms,nickburlett/feincms,nickburlett/feincms,mjl/feincms,joshuajonah/feincms,joshuajonah/feincms,pjdelport/feincms,joshuajonah/feincms,michaelkuty/feincms,pjdelport/feincms,feincms/feincms,michaelkuty/feincms,michaelkuty/feincms,matthiask/feincms2-content,mjl/feincms,nickburlett/feincms,feincms/feincms,matthiask/django-content-editor
|
VERSION = (1, 4, 2)
__version__ = '.'.join(map(str, VERSION))
class LazySettings(object):
def _load_settings(self):
from feincms import default_settings
from django.conf import settings as django_settings
for key in dir(default_settings):
- if not key.startswith(('FEINCMS_', '_HACK_')):
+ if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')):
continue
setattr(self, key, getattr(django_settings, key,
getattr(default_settings, key)))
def __getattr__(self, attr):
self._load_settings()
del self.__class__.__getattr__
return self.__dict__[attr]
settings = LazySettings()
COMPLETELY_LOADED = False
def ensure_completely_loaded():
"""
This method ensures all models are completely loaded
FeinCMS requires Django to be completely initialized before proceeding,
because of the extension mechanism and the dynamically created content
types.
For more informations, have a look at issue #23 on github:
http://github.com/feincms/feincms/issues#issue/23
"""
global COMPLETELY_LOADED
if COMPLETELY_LOADED:
return True
from django.core.management.validation import get_validation_errors
from StringIO import StringIO
get_validation_errors(StringIO(), None)
COMPLETELY_LOADED = True
return True
|
Fix a Python 2.4 incompatibility that snuck in
|
## Code Before:
VERSION = (1, 4, 2)
__version__ = '.'.join(map(str, VERSION))
class LazySettings(object):
def _load_settings(self):
from feincms import default_settings
from django.conf import settings as django_settings
for key in dir(default_settings):
if not key.startswith(('FEINCMS_', '_HACK_')):
continue
setattr(self, key, getattr(django_settings, key,
getattr(default_settings, key)))
def __getattr__(self, attr):
self._load_settings()
del self.__class__.__getattr__
return self.__dict__[attr]
settings = LazySettings()
COMPLETELY_LOADED = False
def ensure_completely_loaded():
"""
This method ensures all models are completely loaded
FeinCMS requires Django to be completely initialized before proceeding,
because of the extension mechanism and the dynamically created content
types.
For more informations, have a look at issue #23 on github:
http://github.com/feincms/feincms/issues#issue/23
"""
global COMPLETELY_LOADED
if COMPLETELY_LOADED:
return True
from django.core.management.validation import get_validation_errors
from StringIO import StringIO
get_validation_errors(StringIO(), None)
COMPLETELY_LOADED = True
return True
## Instruction:
Fix a Python 2.4 incompatibility that snuck in
## Code After:
VERSION = (1, 4, 2)
__version__ = '.'.join(map(str, VERSION))
class LazySettings(object):
def _load_settings(self):
from feincms import default_settings
from django.conf import settings as django_settings
for key in dir(default_settings):
if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')):
continue
setattr(self, key, getattr(django_settings, key,
getattr(default_settings, key)))
def __getattr__(self, attr):
self._load_settings()
del self.__class__.__getattr__
return self.__dict__[attr]
settings = LazySettings()
COMPLETELY_LOADED = False
def ensure_completely_loaded():
"""
This method ensures all models are completely loaded
FeinCMS requires Django to be completely initialized before proceeding,
because of the extension mechanism and the dynamically created content
types.
For more informations, have a look at issue #23 on github:
http://github.com/feincms/feincms/issues#issue/23
"""
global COMPLETELY_LOADED
if COMPLETELY_LOADED:
return True
from django.core.management.validation import get_validation_errors
from StringIO import StringIO
get_validation_errors(StringIO(), None)
COMPLETELY_LOADED = True
return True
|
...
for key in dir(default_settings):
if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')):
continue
...
|
d0ca3952a34a74f0167b76bbedfa3cf8875a399c
|
var/spack/repos/builtin/packages/py-scikit-learn/package.py
|
var/spack/repos/builtin/packages/py-scikit-learn/package.py
|
from spack import *
class PyScikitLearn(Package):
""""""
homepage = "https://pypi.python.org/pypi/scikit-learn"
url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz"
version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d')
version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
|
from spack import *
class PyScikitLearn(Package):
""""""
homepage = "https://pypi.python.org/pypi/scikit-learn"
url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz"
version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d')
version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e')
version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
|
Add version 0.17.1 of scikit-learn.
|
Add version 0.17.1 of scikit-learn.
|
Python
|
lgpl-2.1
|
matthiasdiener/spack,mfherbst/spack,EmreAtes/spack,TheTimmy/spack,iulian787/spack,iulian787/spack,iulian787/spack,mfherbst/spack,tmerrick1/spack,mfherbst/spack,LLNL/spack,LLNL/spack,tmerrick1/spack,skosukhin/spack,TheTimmy/spack,skosukhin/spack,TheTimmy/spack,skosukhin/spack,matthiasdiener/spack,mfherbst/spack,mfherbst/spack,lgarren/spack,LLNL/spack,matthiasdiener/spack,tmerrick1/spack,krafczyk/spack,krafczyk/spack,TheTimmy/spack,EmreAtes/spack,LLNL/spack,krafczyk/spack,iulian787/spack,lgarren/spack,tmerrick1/spack,EmreAtes/spack,EmreAtes/spack,krafczyk/spack,lgarren/spack,matthiasdiener/spack,lgarren/spack,iulian787/spack,LLNL/spack,lgarren/spack,skosukhin/spack,krafczyk/spack,skosukhin/spack,matthiasdiener/spack,TheTimmy/spack,EmreAtes/spack,tmerrick1/spack
|
from spack import *
class PyScikitLearn(Package):
""""""
homepage = "https://pypi.python.org/pypi/scikit-learn"
url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz"
version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d')
version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e')
+ version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
|
Add version 0.17.1 of scikit-learn.
|
## Code Before:
from spack import *
class PyScikitLearn(Package):
""""""
homepage = "https://pypi.python.org/pypi/scikit-learn"
url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz"
version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d')
version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
## Instruction:
Add version 0.17.1 of scikit-learn.
## Code After:
from spack import *
class PyScikitLearn(Package):
""""""
homepage = "https://pypi.python.org/pypi/scikit-learn"
url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz"
version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d')
version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e')
version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc')
extends('python')
def install(self, spec, prefix):
python('setup.py', 'install', '--prefix=%s' % prefix)
|
...
version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e')
version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc')
...
|
77541e5b3956d9e6b130810211fcae10de29eb85
|
tests/integration/base.py
|
tests/integration/base.py
|
import righteous
from ConfigParser import SafeConfigParser
from ..compat import unittest
class RighteousIntegrationTestCase(unittest.TestCase):
def setUp(self):
config = SafeConfigParser()
config.read('righteous.config')
if not config.has_section('auth'):
raise Exception('Please create a righteous.config file with '
'appropriate credentials')
self.auth = dict(
(key, config.get('auth', key))
for key in config.options('auth'))
self.server = dict(
(key, config.get('server-defaults', key))
for key in config.options('server-defaults'))
righteous.init(
self.auth['username'], self.auth['password'],
self.auth['account_id'], **self.server)
self.config = config
self.username = self.auth['username']
def test_login(self):
self.assertTrue(righteous.login())
assert False
|
import righteous
from ConfigParser import SafeConfigParser
from ..compat import unittest
class RighteousIntegrationTestCase(unittest.TestCase):
def setUp(self):
config = SafeConfigParser()
config.read('righteous.config')
if not config.has_section('auth'):
raise Exception('Please create a righteous.config file with '
'appropriate credentials')
self.auth = dict(
(key, config.get('auth', key))
for key in config.options('auth'))
self.server = dict(
(key, config.get('server-defaults', key))
for key in config.options('server-defaults'))
righteous.init(
self.auth['username'], self.auth['password'],
self.auth['account_id'], **self.server)
self.config = config
self.username = self.auth['username']
def test_login(self):
self.assertTrue(righteous.login())
|
Make the integration test fail, so we can see the request / response
|
Make the integration test fail, so we can see the request / response
|
Python
|
unlicense
|
michaeljoseph/righteous,michaeljoseph/righteous
|
import righteous
from ConfigParser import SafeConfigParser
from ..compat import unittest
class RighteousIntegrationTestCase(unittest.TestCase):
def setUp(self):
config = SafeConfigParser()
config.read('righteous.config')
if not config.has_section('auth'):
raise Exception('Please create a righteous.config file with '
'appropriate credentials')
self.auth = dict(
(key, config.get('auth', key))
for key in config.options('auth'))
self.server = dict(
(key, config.get('server-defaults', key))
for key in config.options('server-defaults'))
righteous.init(
self.auth['username'], self.auth['password'],
self.auth['account_id'], **self.server)
self.config = config
self.username = self.auth['username']
def test_login(self):
self.assertTrue(righteous.login())
- assert False
|
Make the integration test fail, so we can see the request / response
|
## Code Before:
import righteous
from ConfigParser import SafeConfigParser
from ..compat import unittest
class RighteousIntegrationTestCase(unittest.TestCase):
def setUp(self):
config = SafeConfigParser()
config.read('righteous.config')
if not config.has_section('auth'):
raise Exception('Please create a righteous.config file with '
'appropriate credentials')
self.auth = dict(
(key, config.get('auth', key))
for key in config.options('auth'))
self.server = dict(
(key, config.get('server-defaults', key))
for key in config.options('server-defaults'))
righteous.init(
self.auth['username'], self.auth['password'],
self.auth['account_id'], **self.server)
self.config = config
self.username = self.auth['username']
def test_login(self):
self.assertTrue(righteous.login())
assert False
## Instruction:
Make the integration test fail, so we can see the request / response
## Code After:
import righteous
from ConfigParser import SafeConfigParser
from ..compat import unittest
class RighteousIntegrationTestCase(unittest.TestCase):
def setUp(self):
config = SafeConfigParser()
config.read('righteous.config')
if not config.has_section('auth'):
raise Exception('Please create a righteous.config file with '
'appropriate credentials')
self.auth = dict(
(key, config.get('auth', key))
for key in config.options('auth'))
self.server = dict(
(key, config.get('server-defaults', key))
for key in config.options('server-defaults'))
righteous.init(
self.auth['username'], self.auth['password'],
self.auth['account_id'], **self.server)
self.config = config
self.username = self.auth['username']
def test_login(self):
self.assertTrue(righteous.login())
|
// ... existing code ...
self.assertTrue(righteous.login())
// ... rest of the code ...
|
cb08d25f49b8b4c5177c8afdd9a69330992ee854
|
tests/replay/test_replay.py
|
tests/replay/test_replay.py
|
import pytest
from cookiecutter import replay, main, exceptions
def test_get_replay_file_name():
"""Make sure that replay.get_file_name generates a valid json file path."""
assert replay.get_file_name('foo', 'bar') == 'foo/bar.json'
@pytest.fixture(params=[
{'no_input': True},
{'extra_context': {}},
{'no_input': True, 'extra_context': {}},
])
def invalid_kwargs(request):
return request.param
def test_raise_on_invalid_mode(invalid_kwargs):
with pytest.raises(exceptions.InvalidModeException):
main.cookiecutter('foo', replay=True, **invalid_kwargs)
|
import pytest
from cookiecutter import replay, main, exceptions
def test_get_replay_file_name():
"""Make sure that replay.get_file_name generates a valid json file path."""
assert replay.get_file_name('foo', 'bar') == 'foo/bar.json'
@pytest.fixture(params=[
{'no_input': True},
{'extra_context': {}},
{'no_input': True, 'extra_context': {}},
])
def invalid_kwargs(request):
return request.param
def test_raise_on_invalid_mode(invalid_kwargs):
with pytest.raises(exceptions.InvalidModeException):
main.cookiecutter('foo', replay=True, **invalid_kwargs)
def test_main_does_not_invoke_dump_but_load(mocker):
mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
mock_replay_dump = mocker.patch('cookiecutter.main.dump')
mock_replay_load = mocker.patch('cookiecutter.main.load')
main.cookiecutter('foobar', replay=True)
assert not mock_prompt.called
assert not mock_gen_context.called
assert not mock_replay_dump.called
assert mock_replay_load.called
assert mock_gen_files.called
def test_main_does_not_invoke_load_but_dump(mocker):
mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
mock_replay_dump = mocker.patch('cookiecutter.main.dump')
mock_replay_load = mocker.patch('cookiecutter.main.load')
main.cookiecutter('foobar', replay=False)
assert mock_prompt.called
assert mock_gen_context.called
assert mock_replay_dump.called
assert not mock_replay_load.called
assert mock_gen_files.called
|
Add tests for a correct behaviour in cookiecutter.main for replay
|
Add tests for a correct behaviour in cookiecutter.main for replay
|
Python
|
bsd-3-clause
|
christabor/cookiecutter,luzfcb/cookiecutter,hackebrot/cookiecutter,cguardia/cookiecutter,pjbull/cookiecutter,dajose/cookiecutter,michaeljoseph/cookiecutter,moi65/cookiecutter,terryjbates/cookiecutter,takeflight/cookiecutter,terryjbates/cookiecutter,luzfcb/cookiecutter,agconti/cookiecutter,cguardia/cookiecutter,christabor/cookiecutter,audreyr/cookiecutter,stevepiercy/cookiecutter,willingc/cookiecutter,venumech/cookiecutter,stevepiercy/cookiecutter,takeflight/cookiecutter,pjbull/cookiecutter,benthomasson/cookiecutter,agconti/cookiecutter,benthomasson/cookiecutter,Springerle/cookiecutter,ramiroluz/cookiecutter,audreyr/cookiecutter,moi65/cookiecutter,dajose/cookiecutter,hackebrot/cookiecutter,michaeljoseph/cookiecutter,Springerle/cookiecutter,ramiroluz/cookiecutter,venumech/cookiecutter,willingc/cookiecutter
|
import pytest
from cookiecutter import replay, main, exceptions
def test_get_replay_file_name():
"""Make sure that replay.get_file_name generates a valid json file path."""
assert replay.get_file_name('foo', 'bar') == 'foo/bar.json'
@pytest.fixture(params=[
{'no_input': True},
{'extra_context': {}},
{'no_input': True, 'extra_context': {}},
])
def invalid_kwargs(request):
return request.param
def test_raise_on_invalid_mode(invalid_kwargs):
with pytest.raises(exceptions.InvalidModeException):
main.cookiecutter('foo', replay=True, **invalid_kwargs)
+
+ def test_main_does_not_invoke_dump_but_load(mocker):
+ mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
+ mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
+ mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
+ mock_replay_dump = mocker.patch('cookiecutter.main.dump')
+ mock_replay_load = mocker.patch('cookiecutter.main.load')
+
+ main.cookiecutter('foobar', replay=True)
+
+ assert not mock_prompt.called
+ assert not mock_gen_context.called
+ assert not mock_replay_dump.called
+ assert mock_replay_load.called
+ assert mock_gen_files.called
+
+
+ def test_main_does_not_invoke_load_but_dump(mocker):
+ mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
+ mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
+ mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
+ mock_replay_dump = mocker.patch('cookiecutter.main.dump')
+ mock_replay_load = mocker.patch('cookiecutter.main.load')
+
+ main.cookiecutter('foobar', replay=False)
+
+ assert mock_prompt.called
+ assert mock_gen_context.called
+ assert mock_replay_dump.called
+ assert not mock_replay_load.called
+ assert mock_gen_files.called
+
|
Add tests for a correct behaviour in cookiecutter.main for replay
|
## Code Before:
import pytest
from cookiecutter import replay, main, exceptions
def test_get_replay_file_name():
"""Make sure that replay.get_file_name generates a valid json file path."""
assert replay.get_file_name('foo', 'bar') == 'foo/bar.json'
@pytest.fixture(params=[
{'no_input': True},
{'extra_context': {}},
{'no_input': True, 'extra_context': {}},
])
def invalid_kwargs(request):
return request.param
def test_raise_on_invalid_mode(invalid_kwargs):
with pytest.raises(exceptions.InvalidModeException):
main.cookiecutter('foo', replay=True, **invalid_kwargs)
## Instruction:
Add tests for a correct behaviour in cookiecutter.main for replay
## Code After:
import pytest
from cookiecutter import replay, main, exceptions
def test_get_replay_file_name():
"""Make sure that replay.get_file_name generates a valid json file path."""
assert replay.get_file_name('foo', 'bar') == 'foo/bar.json'
@pytest.fixture(params=[
{'no_input': True},
{'extra_context': {}},
{'no_input': True, 'extra_context': {}},
])
def invalid_kwargs(request):
return request.param
def test_raise_on_invalid_mode(invalid_kwargs):
with pytest.raises(exceptions.InvalidModeException):
main.cookiecutter('foo', replay=True, **invalid_kwargs)
def test_main_does_not_invoke_dump_but_load(mocker):
mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
mock_replay_dump = mocker.patch('cookiecutter.main.dump')
mock_replay_load = mocker.patch('cookiecutter.main.load')
main.cookiecutter('foobar', replay=True)
assert not mock_prompt.called
assert not mock_gen_context.called
assert not mock_replay_dump.called
assert mock_replay_load.called
assert mock_gen_files.called
def test_main_does_not_invoke_load_but_dump(mocker):
mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
mock_replay_dump = mocker.patch('cookiecutter.main.dump')
mock_replay_load = mocker.patch('cookiecutter.main.load')
main.cookiecutter('foobar', replay=False)
assert mock_prompt.called
assert mock_gen_context.called
assert mock_replay_dump.called
assert not mock_replay_load.called
assert mock_gen_files.called
|
# ... existing code ...
main.cookiecutter('foo', replay=True, **invalid_kwargs)
def test_main_does_not_invoke_dump_but_load(mocker):
mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
mock_replay_dump = mocker.patch('cookiecutter.main.dump')
mock_replay_load = mocker.patch('cookiecutter.main.load')
main.cookiecutter('foobar', replay=True)
assert not mock_prompt.called
assert not mock_gen_context.called
assert not mock_replay_dump.called
assert mock_replay_load.called
assert mock_gen_files.called
def test_main_does_not_invoke_load_but_dump(mocker):
mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config')
mock_gen_context = mocker.patch('cookiecutter.main.generate_context')
mock_gen_files = mocker.patch('cookiecutter.main.generate_files')
mock_replay_dump = mocker.patch('cookiecutter.main.dump')
mock_replay_load = mocker.patch('cookiecutter.main.load')
main.cookiecutter('foobar', replay=False)
assert mock_prompt.called
assert mock_gen_context.called
assert mock_replay_dump.called
assert not mock_replay_load.called
assert mock_gen_files.called
# ... rest of the code ...
|
dfaa49b31e8abd10456761110d0cadc1b7c7640d
|
zaqar/transport/wsgi/app.py
|
zaqar/transport/wsgi/app.py
|
from oslo_config import cfg
from zaqar import bootstrap
# Use the global CONF instance
conf = cfg.CONF
conf(project='zaqar', prog='zaqar-queues', args=[])
boot = bootstrap.Bootstrap(conf)
conf.drivers.transport = 'wsgi'
app = boot.transport.app
|
from oslo_config import cfg
from oslo_log import log
from zaqar import bootstrap
# Use the global CONF instance
conf = cfg.CONF
log.register_options(conf)
conf(project='zaqar', prog='zaqar-queues', args=[])
log.setup(conf, 'zaqar')
boot = bootstrap.Bootstrap(conf)
conf.drivers.transport = 'wsgi'
app = boot.transport.app
|
Make the log work when deploy Zaqar with uwsgi
|
Make the log work when deploy Zaqar with uwsgi
The zaqar-wsgi runs under uwsgi by devstack can't
print any WARNING, DEBUG, ERROR or INFO log now.
This path add the log initialization for uwsgi boot.
Change-Id: Ifcd6be908442275d2acbde2562e593b2ca87b277
Cloese-bug: #1645492
|
Python
|
apache-2.0
|
openstack/zaqar,openstack/zaqar,openstack/zaqar,openstack/zaqar
|
from oslo_config import cfg
+ from oslo_log import log
from zaqar import bootstrap
# Use the global CONF instance
conf = cfg.CONF
+ log.register_options(conf)
conf(project='zaqar', prog='zaqar-queues', args=[])
+ log.setup(conf, 'zaqar')
boot = bootstrap.Bootstrap(conf)
conf.drivers.transport = 'wsgi'
app = boot.transport.app
|
Make the log work when deploy Zaqar with uwsgi
|
## Code Before:
from oslo_config import cfg
from zaqar import bootstrap
# Use the global CONF instance
conf = cfg.CONF
conf(project='zaqar', prog='zaqar-queues', args=[])
boot = bootstrap.Bootstrap(conf)
conf.drivers.transport = 'wsgi'
app = boot.transport.app
## Instruction:
Make the log work when deploy Zaqar with uwsgi
## Code After:
from oslo_config import cfg
from oslo_log import log
from zaqar import bootstrap
# Use the global CONF instance
conf = cfg.CONF
log.register_options(conf)
conf(project='zaqar', prog='zaqar-queues', args=[])
log.setup(conf, 'zaqar')
boot = bootstrap.Bootstrap(conf)
conf.drivers.transport = 'wsgi'
app = boot.transport.app
|
...
from oslo_config import cfg
from oslo_log import log
...
conf = cfg.CONF
log.register_options(conf)
conf(project='zaqar', prog='zaqar-queues', args=[])
log.setup(conf, 'zaqar')
...
|
4636c9394138534fc39cc5bdac373b97919ffd01
|
server/info/services.py
|
server/info/services.py
|
"""info services."""
from info.models import Article, News, Column
def get_column_object(uid):
"""Get column object."""
try:
obj = Column.objects.get(uid=uid)
except Column.DoesNotExist:
obj = None
return obj
def get_articles_by_column(uid):
"""Get_articles_by_column."""
queryset = Article.objects.filter(column__uid=uid).order_by('id')
return queryset
def get_columns_queryset():
"""Get_columns_queryset."""
queryset = Column.objects.all().order_by('-id')
return queryset
def get_article_queryset():
"""Get article queryset."""
queryset = Article.objects.all().order_by('-id')
return queryset
def get_article_object(uid):
"""Get article object."""
return Article.objects.get(uid=uid)
def get_news_queryset():
"""Get news queryset."""
return News.objects.all().order_by('-id')
|
"""info services."""
from info.models import Article, News, Column
def get_column_object(uid):
"""Get column object."""
try:
obj = Column.objects.get(uid=uid)
except Column.DoesNotExist:
obj = None
return obj
def get_articles_by_column(uid):
"""Get_articles_by_column."""
queryset = Article.objects.filter(
column__uid=uid
).order_by('id')
return queryset
def get_columns_queryset():
"""Get_columns_queryset."""
queryset = Column.objects.all().only('uid', 'name').order_by('-id')
return queryset
def get_article_queryset():
"""Get article queryset."""
queryset = Article.objects.all().order_by('-id')
return queryset
def get_article_object(uid):
"""Get article object."""
return Article.objects.get(uid=uid)
def get_news_queryset():
"""Get news queryset."""
return News.objects.all().order_by('-id')
|
Modify django orm filter, add only
|
Modify django orm filter, add only
|
Python
|
mit
|
istommao/codingcatweb,istommao/codingcatweb,istommao/codingcatweb
|
"""info services."""
from info.models import Article, News, Column
def get_column_object(uid):
"""Get column object."""
try:
obj = Column.objects.get(uid=uid)
except Column.DoesNotExist:
obj = None
return obj
def get_articles_by_column(uid):
"""Get_articles_by_column."""
- queryset = Article.objects.filter(column__uid=uid).order_by('id')
+ queryset = Article.objects.filter(
+ column__uid=uid
+ ).order_by('id')
return queryset
def get_columns_queryset():
"""Get_columns_queryset."""
- queryset = Column.objects.all().order_by('-id')
+ queryset = Column.objects.all().only('uid', 'name').order_by('-id')
return queryset
def get_article_queryset():
"""Get article queryset."""
queryset = Article.objects.all().order_by('-id')
return queryset
def get_article_object(uid):
"""Get article object."""
return Article.objects.get(uid=uid)
def get_news_queryset():
"""Get news queryset."""
return News.objects.all().order_by('-id')
|
Modify django orm filter, add only
|
## Code Before:
"""info services."""
from info.models import Article, News, Column
def get_column_object(uid):
"""Get column object."""
try:
obj = Column.objects.get(uid=uid)
except Column.DoesNotExist:
obj = None
return obj
def get_articles_by_column(uid):
"""Get_articles_by_column."""
queryset = Article.objects.filter(column__uid=uid).order_by('id')
return queryset
def get_columns_queryset():
"""Get_columns_queryset."""
queryset = Column.objects.all().order_by('-id')
return queryset
def get_article_queryset():
"""Get article queryset."""
queryset = Article.objects.all().order_by('-id')
return queryset
def get_article_object(uid):
"""Get article object."""
return Article.objects.get(uid=uid)
def get_news_queryset():
"""Get news queryset."""
return News.objects.all().order_by('-id')
## Instruction:
Modify django orm filter, add only
## Code After:
"""info services."""
from info.models import Article, News, Column
def get_column_object(uid):
"""Get column object."""
try:
obj = Column.objects.get(uid=uid)
except Column.DoesNotExist:
obj = None
return obj
def get_articles_by_column(uid):
"""Get_articles_by_column."""
queryset = Article.objects.filter(
column__uid=uid
).order_by('id')
return queryset
def get_columns_queryset():
"""Get_columns_queryset."""
queryset = Column.objects.all().only('uid', 'name').order_by('-id')
return queryset
def get_article_queryset():
"""Get article queryset."""
queryset = Article.objects.all().order_by('-id')
return queryset
def get_article_object(uid):
"""Get article object."""
return Article.objects.get(uid=uid)
def get_news_queryset():
"""Get news queryset."""
return News.objects.all().order_by('-id')
|
...
"""Get_articles_by_column."""
queryset = Article.objects.filter(
column__uid=uid
).order_by('id')
...
"""Get_columns_queryset."""
queryset = Column.objects.all().only('uid', 'name').order_by('-id')
...
|
10dea74d7f7946e9bab8c99b489793708845183c
|
fireplace/cards/wog/hunter.py
|
fireplace/cards/wog/hunter.py
|
from ..utils import *
##
# Minions
class OG_179:
"Fiery Bat"
deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1)
class OG_292:
"Forlorn Stalker"
play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e")
OG_292e = buff(+1, +1)
|
from ..utils import *
##
# Minions
class OG_179:
"Fiery Bat"
deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1)
class OG_292:
"Forlorn Stalker"
play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e")
OG_292e = buff(+1, +1)
##
# Spells
class OG_045:
"Infest"
play = Buff(FRIENDLY_MINIONS, "OG_045a")
class OG_045a:
"Nerubian Spores"
deathrattle = Give(CONTROLLER, RandomBeast())
tags = {GameTag.DEATHRATTLE: True}
class OG_061:
"On the Hunt"
play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t")
class OG_211:
"Call of the Wild"
play = (
Summon(CONTROLLER, "NEW1_034"),
Summon(CONTROLLER, "NEW1_033"),
Summon(CONTROLLER, "NEW1_032")
)
|
Implement Infest, On the Hunt, Call of the Wild
|
Implement Infest, On the Hunt, Call of the Wild
|
Python
|
agpl-3.0
|
jleclanche/fireplace,beheh/fireplace,NightKev/fireplace
|
from ..utils import *
##
# Minions
class OG_179:
"Fiery Bat"
deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1)
class OG_292:
"Forlorn Stalker"
play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e")
OG_292e = buff(+1, +1)
+
+ ##
+ # Spells
+
+ class OG_045:
+ "Infest"
+ play = Buff(FRIENDLY_MINIONS, "OG_045a")
+
+ class OG_045a:
+ "Nerubian Spores"
+ deathrattle = Give(CONTROLLER, RandomBeast())
+ tags = {GameTag.DEATHRATTLE: True}
+
+
+ class OG_061:
+ "On the Hunt"
+ play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t")
+
+
+ class OG_211:
+ "Call of the Wild"
+ play = (
+ Summon(CONTROLLER, "NEW1_034"),
+ Summon(CONTROLLER, "NEW1_033"),
+ Summon(CONTROLLER, "NEW1_032")
+ )
+
|
Implement Infest, On the Hunt, Call of the Wild
|
## Code Before:
from ..utils import *
##
# Minions
class OG_179:
"Fiery Bat"
deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1)
class OG_292:
"Forlorn Stalker"
play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e")
OG_292e = buff(+1, +1)
## Instruction:
Implement Infest, On the Hunt, Call of the Wild
## Code After:
from ..utils import *
##
# Minions
class OG_179:
"Fiery Bat"
deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1)
class OG_292:
"Forlorn Stalker"
play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e")
OG_292e = buff(+1, +1)
##
# Spells
class OG_045:
"Infest"
play = Buff(FRIENDLY_MINIONS, "OG_045a")
class OG_045a:
"Nerubian Spores"
deathrattle = Give(CONTROLLER, RandomBeast())
tags = {GameTag.DEATHRATTLE: True}
class OG_061:
"On the Hunt"
play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t")
class OG_211:
"Call of the Wild"
play = (
Summon(CONTROLLER, "NEW1_034"),
Summon(CONTROLLER, "NEW1_033"),
Summon(CONTROLLER, "NEW1_032")
)
|
// ... existing code ...
OG_292e = buff(+1, +1)
##
# Spells
class OG_045:
"Infest"
play = Buff(FRIENDLY_MINIONS, "OG_045a")
class OG_045a:
"Nerubian Spores"
deathrattle = Give(CONTROLLER, RandomBeast())
tags = {GameTag.DEATHRATTLE: True}
class OG_061:
"On the Hunt"
play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t")
class OG_211:
"Call of the Wild"
play = (
Summon(CONTROLLER, "NEW1_034"),
Summon(CONTROLLER, "NEW1_033"),
Summon(CONTROLLER, "NEW1_032")
)
// ... rest of the code ...
|
a70bb058bd93831b755079f5fee495088b620c6d
|
taiga/locale/api.py
|
taiga/locale/api.py
|
from django.conf import settings
from taiga.base import response
from taiga.base.api.viewsets import ReadOnlyListViewSet
from . import permissions
class LocalesViewSet(ReadOnlyListViewSet):
permission_classes = (permissions.LocalesPermission,)
def list(self, request, *args, **kwargs):
locales = [{"code": c, "name": n} for c, n in settings.LANGUAGES]
return response.Ok(locales)
|
from django.conf import settings
from taiga.base import response
from taiga.base.api.viewsets import ReadOnlyListViewSet
from . import permissions
class LocalesViewSet(ReadOnlyListViewSet):
permission_classes = (permissions.LocalesPermission,)
def list(self, request, *args, **kwargs):
locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES]
return response.Ok(locales)
|
Add bidi (right-to-left layout) attr to locale resource
|
Add bidi (right-to-left layout) attr to locale resource
|
Python
|
agpl-3.0
|
crr0004/taiga-back,CoolCloud/taiga-back,Rademade/taiga-back,seanchen/taiga-back,Tigerwhit4/taiga-back,seanchen/taiga-back,astronaut1712/taiga-back,dycodedev/taiga-back,dycodedev/taiga-back,WALR/taiga-back,forging2012/taiga-back,xdevelsistemas/taiga-back-community,xdevelsistemas/taiga-back-community,rajiteh/taiga-back,Tigerwhit4/taiga-back,taigaio/taiga-back,EvgeneOskin/taiga-back,CMLL/taiga-back,EvgeneOskin/taiga-back,Rademade/taiga-back,EvgeneOskin/taiga-back,seanchen/taiga-back,rajiteh/taiga-back,CMLL/taiga-back,gauravjns/taiga-back,obimod/taiga-back,CMLL/taiga-back,astagi/taiga-back,dayatz/taiga-back,astronaut1712/taiga-back,CMLL/taiga-back,astagi/taiga-back,joshisa/taiga-back,coopsource/taiga-back,dayatz/taiga-back,gauravjns/taiga-back,forging2012/taiga-back,joshisa/taiga-back,gam-phon/taiga-back,Rademade/taiga-back,Tigerwhit4/taiga-back,CoolCloud/taiga-back,crr0004/taiga-back,coopsource/taiga-back,obimod/taiga-back,dycodedev/taiga-back,gam-phon/taiga-back,WALR/taiga-back,gam-phon/taiga-back,gam-phon/taiga-back,taigaio/taiga-back,xdevelsistemas/taiga-back-community,CoolCloud/taiga-back,jeffdwyatt/taiga-back,taigaio/taiga-back,coopsource/taiga-back,forging2012/taiga-back,WALR/taiga-back,gauravjns/taiga-back,joshisa/taiga-back,crr0004/taiga-back,jeffdwyatt/taiga-back,forging2012/taiga-back,CoolCloud/taiga-back,seanchen/taiga-back,EvgeneOskin/taiga-back,joshisa/taiga-back,obimod/taiga-back,Rademade/taiga-back,WALR/taiga-back,Rademade/taiga-back,jeffdwyatt/taiga-back,dycodedev/taiga-back,bdang2012/taiga-back-casting,obimod/taiga-back,astronaut1712/taiga-back,astagi/taiga-back,dayatz/taiga-back,rajiteh/taiga-back,astagi/taiga-back,bdang2012/taiga-back-casting,crr0004/taiga-back,gauravjns/taiga-back,coopsource/taiga-back,bdang2012/taiga-back-casting,astronaut1712/taiga-back,jeffdwyatt/taiga-back,Tigerwhit4/taiga-back,bdang2012/taiga-back-casting,rajiteh/taiga-back
|
from django.conf import settings
from taiga.base import response
from taiga.base.api.viewsets import ReadOnlyListViewSet
from . import permissions
class LocalesViewSet(ReadOnlyListViewSet):
permission_classes = (permissions.LocalesPermission,)
def list(self, request, *args, **kwargs):
- locales = [{"code": c, "name": n} for c, n in settings.LANGUAGES]
+ locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES]
return response.Ok(locales)
|
Add bidi (right-to-left layout) attr to locale resource
|
## Code Before:
from django.conf import settings
from taiga.base import response
from taiga.base.api.viewsets import ReadOnlyListViewSet
from . import permissions
class LocalesViewSet(ReadOnlyListViewSet):
permission_classes = (permissions.LocalesPermission,)
def list(self, request, *args, **kwargs):
locales = [{"code": c, "name": n} for c, n in settings.LANGUAGES]
return response.Ok(locales)
## Instruction:
Add bidi (right-to-left layout) attr to locale resource
## Code After:
from django.conf import settings
from taiga.base import response
from taiga.base.api.viewsets import ReadOnlyListViewSet
from . import permissions
class LocalesViewSet(ReadOnlyListViewSet):
permission_classes = (permissions.LocalesPermission,)
def list(self, request, *args, **kwargs):
locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES]
return response.Ok(locales)
|
...
def list(self, request, *args, **kwargs):
locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES]
return response.Ok(locales)
...
|
8803f6058255237dff39549426ca6a513a25193c
|
website_product_supplier/__openerp__.py
|
website_product_supplier/__openerp__.py
|
{
'name': "Website Product Supplier",
'category': 'Website',
'version': '8.0.1.0.0',
'depends': [
'website_sale',
'website_portal_purchase',
],
'data': [
'security/ir.model.access.csv',
'views/product_supplier_view.xml',
'views/website_portal.xml',
'views/assets.xml',
'security/website_product_supplier_security.xml',
],
'qweb': ['static/src/xml/*.xml'],
'author': 'Antiun Ingeniería S.L., '
'Incaser Informatica S.L., '
'Odoo Community Association (OCA)',
'website': 'http://www.antiun.com',
'license': 'AGPL-3',
'installable': True,
}
|
{
'name': "Website Product Supplier",
'category': 'Website',
'version': '8.0.1.0.0',
'depends': [
'website_sale',
'website_portal_purchase',
],
'data': [
'security/ir.model.access.csv',
'views/product_supplier_view.xml',
'views/website_portal.xml',
'views/assets.xml',
'security/website_product_supplier_security.xml',
],
'images': [],
'qweb': ['static/src/xml/*.xml'],
'author': 'Antiun Ingeniería S.L., '
'Incaser Informatica S.L., '
'Odoo Community Association (OCA)',
'website': 'http://www.antiun.com',
'license': 'AGPL-3',
'installable': True,
}
|
Add images key in manifest file
|
[FIX] website_product_supplier: Add images key in manifest file
|
Python
|
agpl-3.0
|
nuobit/website,open-synergy/website,gfcapalbo/website,LasLabs/website,acsone/website,nuobit/website,LasLabs/website,Yajo/website,LasLabs/website,gfcapalbo/website,kaerdsar/website,Yajo/website,nuobit/website,nuobit/website,Yajo/website,gfcapalbo/website,acsone/website,kaerdsar/website,LasLabs/website,open-synergy/website,acsone/website,acsone/website,gfcapalbo/website,open-synergy/website,brain-tec/website,brain-tec/website,open-synergy/website,brain-tec/website,kaerdsar/website,brain-tec/website,Yajo/website
|
{
'name': "Website Product Supplier",
'category': 'Website',
'version': '8.0.1.0.0',
'depends': [
'website_sale',
'website_portal_purchase',
],
'data': [
'security/ir.model.access.csv',
'views/product_supplier_view.xml',
'views/website_portal.xml',
'views/assets.xml',
'security/website_product_supplier_security.xml',
],
+ 'images': [],
'qweb': ['static/src/xml/*.xml'],
'author': 'Antiun Ingeniería S.L., '
'Incaser Informatica S.L., '
'Odoo Community Association (OCA)',
'website': 'http://www.antiun.com',
'license': 'AGPL-3',
'installable': True,
}
|
Add images key in manifest file
|
## Code Before:
{
'name': "Website Product Supplier",
'category': 'Website',
'version': '8.0.1.0.0',
'depends': [
'website_sale',
'website_portal_purchase',
],
'data': [
'security/ir.model.access.csv',
'views/product_supplier_view.xml',
'views/website_portal.xml',
'views/assets.xml',
'security/website_product_supplier_security.xml',
],
'qweb': ['static/src/xml/*.xml'],
'author': 'Antiun Ingeniería S.L., '
'Incaser Informatica S.L., '
'Odoo Community Association (OCA)',
'website': 'http://www.antiun.com',
'license': 'AGPL-3',
'installable': True,
}
## Instruction:
Add images key in manifest file
## Code After:
{
'name': "Website Product Supplier",
'category': 'Website',
'version': '8.0.1.0.0',
'depends': [
'website_sale',
'website_portal_purchase',
],
'data': [
'security/ir.model.access.csv',
'views/product_supplier_view.xml',
'views/website_portal.xml',
'views/assets.xml',
'security/website_product_supplier_security.xml',
],
'images': [],
'qweb': ['static/src/xml/*.xml'],
'author': 'Antiun Ingeniería S.L., '
'Incaser Informatica S.L., '
'Odoo Community Association (OCA)',
'website': 'http://www.antiun.com',
'license': 'AGPL-3',
'installable': True,
}
|
...
],
'images': [],
'qweb': ['static/src/xml/*.xml'],
...
|
fb02617b29cab97a70a1a11b0d3b7b62b834aa3b
|
server.py
|
server.py
|
from flask import Flask
from flask import request
import flask
import hashlib
import json
import gzip
app = Flask(__name__)
stored_files = {}
@app.route('/profile/<type>', methods=['GET'])
def get_dummy_files(type):
if type == 'lawyer':
pass
elif type == 'doctor:':
pass
elif type == 'female':
pass
elif type == 'male':
pass
else:
return "No files here\n"
return "Sent files\n"
@app.route('/<int:id>', methods=['GET'])
def get_file(id):
if id in stored_files:
return stored_files[id]
else:
return "No such file\n"
@app.route('/', methods=['POST'])
def upload_file():
data = json.loads(request.data)
uploaded_file = data['uploaded_file']
salt = data['salt']
id = hashlib.sha256(uploaded_file.encode()).hexdigest()
stored_files[id] = (uploaded_file, salt)
return "File stored\n"
if __name__ == "__main__":
app.run()
|
from flask import Flask
from flask import request
import flask
import hashlib
import json
import gzip
app = Flask(__name__)
stored_files = {}
@app.route('/profile/<type>', methods=['GET'])
def get_dummy_files(type):
if type == 'lawyer':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'doctor:':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'female':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'male':
gzip_address = './zipfiles/doc.tar.gz'
else:
return "No files here\n"
gzip_file = open(gzip_address).read()
return bytearray(gzip_file)
@app.route('/<int:id>', methods=['GET'])
def get_file(id):
if id in stored_files:
return stored_files[id]
else:
return "No such file\n"
@app.route('/', methods=['POST'])
def upload_file():
data = json.loads(request.data)
uploaded_file = data['uploaded_file']
salt = data['salt']
id = hashlib.sha256(uploaded_file.encode()).hexdigest()
stored_files[id] = (uploaded_file, salt)
return "File stored\n"
if __name__ == "__main__":
app.run()
|
Structure for sending dummy files
|
Structure for sending dummy files
|
Python
|
mit
|
rotemh/soteria
|
from flask import Flask
from flask import request
import flask
import hashlib
import json
import gzip
app = Flask(__name__)
stored_files = {}
@app.route('/profile/<type>', methods=['GET'])
def get_dummy_files(type):
if type == 'lawyer':
- pass
+ gzip_address = './zipfiles/doc.tar.gz'
elif type == 'doctor:':
- pass
+ gzip_address = './zipfiles/doc.tar.gz'
elif type == 'female':
- pass
+ gzip_address = './zipfiles/doc.tar.gz'
elif type == 'male':
- pass
+ gzip_address = './zipfiles/doc.tar.gz'
else:
return "No files here\n"
- return "Sent files\n"
+ gzip_file = open(gzip_address).read()
+ return bytearray(gzip_file)
@app.route('/<int:id>', methods=['GET'])
def get_file(id):
if id in stored_files:
return stored_files[id]
else:
return "No such file\n"
@app.route('/', methods=['POST'])
def upload_file():
data = json.loads(request.data)
uploaded_file = data['uploaded_file']
salt = data['salt']
id = hashlib.sha256(uploaded_file.encode()).hexdigest()
stored_files[id] = (uploaded_file, salt)
return "File stored\n"
if __name__ == "__main__":
app.run()
|
Structure for sending dummy files
|
## Code Before:
from flask import Flask
from flask import request
import flask
import hashlib
import json
import gzip
app = Flask(__name__)
stored_files = {}
@app.route('/profile/<type>', methods=['GET'])
def get_dummy_files(type):
if type == 'lawyer':
pass
elif type == 'doctor:':
pass
elif type == 'female':
pass
elif type == 'male':
pass
else:
return "No files here\n"
return "Sent files\n"
@app.route('/<int:id>', methods=['GET'])
def get_file(id):
if id in stored_files:
return stored_files[id]
else:
return "No such file\n"
@app.route('/', methods=['POST'])
def upload_file():
data = json.loads(request.data)
uploaded_file = data['uploaded_file']
salt = data['salt']
id = hashlib.sha256(uploaded_file.encode()).hexdigest()
stored_files[id] = (uploaded_file, salt)
return "File stored\n"
if __name__ == "__main__":
app.run()
## Instruction:
Structure for sending dummy files
## Code After:
from flask import Flask
from flask import request
import flask
import hashlib
import json
import gzip
app = Flask(__name__)
stored_files = {}
@app.route('/profile/<type>', methods=['GET'])
def get_dummy_files(type):
if type == 'lawyer':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'doctor:':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'female':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'male':
gzip_address = './zipfiles/doc.tar.gz'
else:
return "No files here\n"
gzip_file = open(gzip_address).read()
return bytearray(gzip_file)
@app.route('/<int:id>', methods=['GET'])
def get_file(id):
if id in stored_files:
return stored_files[id]
else:
return "No such file\n"
@app.route('/', methods=['POST'])
def upload_file():
data = json.loads(request.data)
uploaded_file = data['uploaded_file']
salt = data['salt']
id = hashlib.sha256(uploaded_file.encode()).hexdigest()
stored_files[id] = (uploaded_file, salt)
return "File stored\n"
if __name__ == "__main__":
app.run()
|
# ... existing code ...
if type == 'lawyer':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'doctor:':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'female':
gzip_address = './zipfiles/doc.tar.gz'
elif type == 'male':
gzip_address = './zipfiles/doc.tar.gz'
else:
# ... modified code ...
gzip_file = open(gzip_address).read()
return bytearray(gzip_file)
# ... rest of the code ...
|
c017d8fa711724fc7acb7e90b85f208be074d1ec
|
drupdates/plugins/repolist/__init__.py
|
drupdates/plugins/repolist/__init__.py
|
from drupdates.utils import *
from drupdates.repos import *
'''
Note: you need an ssh key set up with Stash to make this script work
'''
class repolist(repoTool):
def __init__(self):
currentDir = os.path.dirname(os.path.realpath(__file__))
self.localsettings = Settings(currentDir)
def gitRepos(self):
#Get list of Stash repos in the Rain Project.
repoDict = self.localsettings.get('repoDict')
if not repoDict:
return {}
else:
return repoDict
|
from drupdates.utils import *
from drupdates.repos import *
'''
Note: you need an ssh key set up with Stash to make this script work
'''
class repolist(repoTool):
def __init__(self):
currentDir = os.path.dirname(os.path.realpath(__file__))
self.localsettings = Settings(currentDir)
def gitRepos(self):
#Get list of Stash repos in the Rain Project.
repoDict = self.localsettings.get('repoDict')
if (not repoDict) or (type(repoDict) is not dict):
return {}
else:
return repoDict
|
Add extra check to repo list to verify a dictionary is returned
|
Add extra check to repo list to verify a dictionary is returned
|
Python
|
mit
|
jalama/drupdates
|
from drupdates.utils import *
from drupdates.repos import *
'''
Note: you need an ssh key set up with Stash to make this script work
'''
class repolist(repoTool):
def __init__(self):
currentDir = os.path.dirname(os.path.realpath(__file__))
self.localsettings = Settings(currentDir)
def gitRepos(self):
#Get list of Stash repos in the Rain Project.
repoDict = self.localsettings.get('repoDict')
- if not repoDict:
+ if (not repoDict) or (type(repoDict) is not dict):
return {}
else:
return repoDict
|
Add extra check to repo list to verify a dictionary is returned
|
## Code Before:
from drupdates.utils import *
from drupdates.repos import *
'''
Note: you need an ssh key set up with Stash to make this script work
'''
class repolist(repoTool):
def __init__(self):
currentDir = os.path.dirname(os.path.realpath(__file__))
self.localsettings = Settings(currentDir)
def gitRepos(self):
#Get list of Stash repos in the Rain Project.
repoDict = self.localsettings.get('repoDict')
if not repoDict:
return {}
else:
return repoDict
## Instruction:
Add extra check to repo list to verify a dictionary is returned
## Code After:
from drupdates.utils import *
from drupdates.repos import *
'''
Note: you need an ssh key set up with Stash to make this script work
'''
class repolist(repoTool):
def __init__(self):
currentDir = os.path.dirname(os.path.realpath(__file__))
self.localsettings = Settings(currentDir)
def gitRepos(self):
#Get list of Stash repos in the Rain Project.
repoDict = self.localsettings.get('repoDict')
if (not repoDict) or (type(repoDict) is not dict):
return {}
else:
return repoDict
|
...
repoDict = self.localsettings.get('repoDict')
if (not repoDict) or (type(repoDict) is not dict):
return {}
...
|
df6b13a70241b616f49d4dcc25073084c371f5b1
|
share/models/creative/base.py
|
share/models/creative/base.py
|
from django.db import models
from share.models.base import ShareObject
from share.models.people import Person
from share.models.base import TypedShareObjectMeta
from share.models.creative.meta import Venue, Institution, Funder, Award, Tag
from share.models.fields import ShareForeignKey, ShareManyToManyField
class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta):
title = models.TextField()
description = models.TextField()
contributors = ShareManyToManyField(Person, through='Contributor')
institutions = ShareManyToManyField(Institution, through='ThroughInstitutions')
venues = ShareManyToManyField(Venue, through='ThroughVenues')
funders = ShareManyToManyField(Funder, through='ThroughFunders')
awards = ShareManyToManyField(Award, through='ThroughAwards')
subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True)
# Note: Null allows inserting of None but returns it as an empty string
tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags')
created = models.DateTimeField(null=True)
published = models.DateTimeField(null=True)
free_to_read_type = models.URLField(blank=True)
free_to_read_date = models.DateTimeField(null=True)
rights = models.TextField()
language = models.TextField()
class CreativeWork(AbstractCreativeWork):
pass
|
from django.db import models
from share.models.base import ShareObject
from share.models.people import Person
from share.models.base import TypedShareObjectMeta
from share.models.creative.meta import Venue, Institution, Funder, Award, Tag
from share.models.fields import ShareForeignKey, ShareManyToManyField
class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta):
title = models.TextField()
description = models.TextField()
contributors = ShareManyToManyField(Person, through='Contributor')
institutions = ShareManyToManyField(Institution, through='ThroughInstitutions')
venues = ShareManyToManyField(Venue, through='ThroughVenues')
funders = ShareManyToManyField(Funder, through='ThroughFunders')
awards = ShareManyToManyField(Award, through='ThroughAwards')
subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True)
# Note: Null allows inserting of None but returns it as an empty string
tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags')
created = models.DateTimeField(null=True)
published = models.DateTimeField(null=True)
free_to_read_type = models.URLField(blank=True)
free_to_read_date = models.DateTimeField(null=True)
rights = models.TextField(blank=True, null=True)
language = models.TextField(blank=True, null=True)
class CreativeWork(AbstractCreativeWork):
pass
|
Swap out license with rights
|
Swap out license with rights
|
Python
|
apache-2.0
|
CenterForOpenScience/SHARE,aaxelb/SHARE,zamattiac/SHARE,zamattiac/SHARE,CenterForOpenScience/SHARE,CenterForOpenScience/SHARE,aaxelb/SHARE,zamattiac/SHARE,laurenbarker/SHARE,aaxelb/SHARE,laurenbarker/SHARE,laurenbarker/SHARE
|
from django.db import models
from share.models.base import ShareObject
from share.models.people import Person
from share.models.base import TypedShareObjectMeta
from share.models.creative.meta import Venue, Institution, Funder, Award, Tag
from share.models.fields import ShareForeignKey, ShareManyToManyField
class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta):
title = models.TextField()
description = models.TextField()
contributors = ShareManyToManyField(Person, through='Contributor')
institutions = ShareManyToManyField(Institution, through='ThroughInstitutions')
venues = ShareManyToManyField(Venue, through='ThroughVenues')
funders = ShareManyToManyField(Funder, through='ThroughFunders')
awards = ShareManyToManyField(Award, through='ThroughAwards')
subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True)
# Note: Null allows inserting of None but returns it as an empty string
tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags')
created = models.DateTimeField(null=True)
published = models.DateTimeField(null=True)
free_to_read_type = models.URLField(blank=True)
free_to_read_date = models.DateTimeField(null=True)
- rights = models.TextField()
+
+ rights = models.TextField(blank=True, null=True)
- language = models.TextField()
+ language = models.TextField(blank=True, null=True)
class CreativeWork(AbstractCreativeWork):
pass
|
Swap out license with rights
|
## Code Before:
from django.db import models
from share.models.base import ShareObject
from share.models.people import Person
from share.models.base import TypedShareObjectMeta
from share.models.creative.meta import Venue, Institution, Funder, Award, Tag
from share.models.fields import ShareForeignKey, ShareManyToManyField
class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta):
title = models.TextField()
description = models.TextField()
contributors = ShareManyToManyField(Person, through='Contributor')
institutions = ShareManyToManyField(Institution, through='ThroughInstitutions')
venues = ShareManyToManyField(Venue, through='ThroughVenues')
funders = ShareManyToManyField(Funder, through='ThroughFunders')
awards = ShareManyToManyField(Award, through='ThroughAwards')
subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True)
# Note: Null allows inserting of None but returns it as an empty string
tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags')
created = models.DateTimeField(null=True)
published = models.DateTimeField(null=True)
free_to_read_type = models.URLField(blank=True)
free_to_read_date = models.DateTimeField(null=True)
rights = models.TextField()
language = models.TextField()
class CreativeWork(AbstractCreativeWork):
pass
## Instruction:
Swap out license with rights
## Code After:
from django.db import models
from share.models.base import ShareObject
from share.models.people import Person
from share.models.base import TypedShareObjectMeta
from share.models.creative.meta import Venue, Institution, Funder, Award, Tag
from share.models.fields import ShareForeignKey, ShareManyToManyField
class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta):
title = models.TextField()
description = models.TextField()
contributors = ShareManyToManyField(Person, through='Contributor')
institutions = ShareManyToManyField(Institution, through='ThroughInstitutions')
venues = ShareManyToManyField(Venue, through='ThroughVenues')
funders = ShareManyToManyField(Funder, through='ThroughFunders')
awards = ShareManyToManyField(Award, through='ThroughAwards')
subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True)
# Note: Null allows inserting of None but returns it as an empty string
tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags')
created = models.DateTimeField(null=True)
published = models.DateTimeField(null=True)
free_to_read_type = models.URLField(blank=True)
free_to_read_date = models.DateTimeField(null=True)
rights = models.TextField(blank=True, null=True)
language = models.TextField(blank=True, null=True)
class CreativeWork(AbstractCreativeWork):
pass
|
...
free_to_read_date = models.DateTimeField(null=True)
rights = models.TextField(blank=True, null=True)
language = models.TextField(blank=True, null=True)
...
|
c2598058722531662aab8831640fc367689d2a43
|
tests/utils/test_process_word_vectors.py
|
tests/utils/test_process_word_vectors.py
|
import inspect
import os
import pytest
import numpy as np
from subprocess import call
from utils.preprocess_text_word_vectors import txtvec2npy
def test_text_word2vec2npy():
# check whether files are present in folder
vectors_name = 'wiki.fiu_vro.vec'
path = os.path.dirname(inspect.getfile(inspect.currentframe()))
if not os.path.exists(path + '/' + vectors_name):
call(["wget https://s3-us-west-1.amazonaws.com/fasttext-vectors/" + vectors_name + " -O " +
path + "/" + vectors_name],
shell=True)
txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4])
vectors = np.load('./' + vectors_name[:-4] + '.npy').item()
assert len(list(vectors)) == 8769
assert vectors['kihlkunnan'].shape[0] == 300
if __name__ == '__main__':
pytest.main([__file__])
|
import inspect
import os
import pytest
import numpy as np
from subprocess import call
from utils.preprocess_text_word_vectors import txtvec2npy
def test_text_word2vec2npy():
# check whether files are present in folder
vectors_name = 'wiki.fiu_vro.vec'
path = os.path.dirname(inspect.getfile(inspect.currentframe()))
if not os.path.exists(path + '/' + vectors_name):
call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " +
path + "/" + vectors_name],
shell=True)
txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4])
vectors = np.load('./' + vectors_name[:-4] + '.npy').item()
assert len(list(vectors)) == 8769
assert vectors['kihlkunnan'].shape[0] == 300
if __name__ == '__main__':
pytest.main([__file__])
|
Update Fasttext pretrained vectors location
|
Update Fasttext pretrained vectors location
|
Python
|
mit
|
lvapeab/nmt-keras,lvapeab/nmt-keras
|
import inspect
import os
import pytest
import numpy as np
from subprocess import call
from utils.preprocess_text_word_vectors import txtvec2npy
def test_text_word2vec2npy():
# check whether files are present in folder
vectors_name = 'wiki.fiu_vro.vec'
path = os.path.dirname(inspect.getfile(inspect.currentframe()))
if not os.path.exists(path + '/' + vectors_name):
- call(["wget https://s3-us-west-1.amazonaws.com/fasttext-vectors/" + vectors_name + " -O " +
+ call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " +
path + "/" + vectors_name],
shell=True)
txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4])
vectors = np.load('./' + vectors_name[:-4] + '.npy').item()
assert len(list(vectors)) == 8769
assert vectors['kihlkunnan'].shape[0] == 300
if __name__ == '__main__':
pytest.main([__file__])
|
Update Fasttext pretrained vectors location
|
## Code Before:
import inspect
import os
import pytest
import numpy as np
from subprocess import call
from utils.preprocess_text_word_vectors import txtvec2npy
def test_text_word2vec2npy():
# check whether files are present in folder
vectors_name = 'wiki.fiu_vro.vec'
path = os.path.dirname(inspect.getfile(inspect.currentframe()))
if not os.path.exists(path + '/' + vectors_name):
call(["wget https://s3-us-west-1.amazonaws.com/fasttext-vectors/" + vectors_name + " -O " +
path + "/" + vectors_name],
shell=True)
txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4])
vectors = np.load('./' + vectors_name[:-4] + '.npy').item()
assert len(list(vectors)) == 8769
assert vectors['kihlkunnan'].shape[0] == 300
if __name__ == '__main__':
pytest.main([__file__])
## Instruction:
Update Fasttext pretrained vectors location
## Code After:
import inspect
import os
import pytest
import numpy as np
from subprocess import call
from utils.preprocess_text_word_vectors import txtvec2npy
def test_text_word2vec2npy():
# check whether files are present in folder
vectors_name = 'wiki.fiu_vro.vec'
path = os.path.dirname(inspect.getfile(inspect.currentframe()))
if not os.path.exists(path + '/' + vectors_name):
call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " +
path + "/" + vectors_name],
shell=True)
txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4])
vectors = np.load('./' + vectors_name[:-4] + '.npy').item()
assert len(list(vectors)) == 8769
assert vectors['kihlkunnan'].shape[0] == 300
if __name__ == '__main__':
pytest.main([__file__])
|
# ... existing code ...
if not os.path.exists(path + '/' + vectors_name):
call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " +
path + "/" + vectors_name],
# ... rest of the code ...
|
bbf3d68b9566a826f404aa1ab3da198d765dca58
|
contacts/rules.py
|
contacts/rules.py
|
ALLOWED_FIELDS = [
'name',
'first_name',
'last_name',
'phone_number',
'photo',
'email',
'twitter'
]
|
ALLOWED_FIELDS = [
'name',
'phone_number',
'first_name',
'last_name',
'phone_number',
'photo',
'email',
'twitter'
]
|
Add 'phone_number' field to ALLOWED_FIELDS.
|
Add 'phone_number' field to ALLOWED_FIELDS.
|
Python
|
mit
|
heimann/contacts
|
ALLOWED_FIELDS = [
'name',
+ 'phone_number',
'first_name',
'last_name',
'phone_number',
'photo',
'email',
'twitter'
]
|
Add 'phone_number' field to ALLOWED_FIELDS.
|
## Code Before:
ALLOWED_FIELDS = [
'name',
'first_name',
'last_name',
'phone_number',
'photo',
'email',
'twitter'
]
## Instruction:
Add 'phone_number' field to ALLOWED_FIELDS.
## Code After:
ALLOWED_FIELDS = [
'name',
'phone_number',
'first_name',
'last_name',
'phone_number',
'photo',
'email',
'twitter'
]
|
# ... existing code ...
'name',
'phone_number',
'first_name',
# ... rest of the code ...
|
8164d048b47299377b4db7d9fc0198e24b07bdb3
|
engine/geometry.py
|
engine/geometry.py
|
from math import cos, sin, pi, hypot
def rotate(polygon, angle):
rotated_points = []
cos_result = cos(angle)
sin_result = sin(angle)
for point in polygon:
x = point[0] * cos_result - point[1] * sin_result
y = point[0] * sin_result + point[1] * cos_result
rotated_points.append((x, y))
return rotated_points
def move(point, direction, amount):
return [point[0] + amount * cos(direction),
point[1] + amount * sin(direction)]
def distance(point1, point2):
return hypot(point1[0] - point2[0], point1[1] - point2[1])
|
from math import cos, sin, pi, hypot
def rotate(polygon, angle):
rotated_points = []
cos_result = cos(angle)
sin_result = sin(angle)
for point in polygon:
x = point[0] * cos_result - point[1] * sin_result
y = point[0] * sin_result + point[1] * cos_result
rotated_points.append((x, y))
return rotated_points
def move(point, direction, amount):
return [int(point[0] + amount * cos(direction)),
int(point[1] + amount * sin(direction))]
def distance(point1, point2):
return hypot(point1[0] - point2[0], point1[1] - point2[1])
|
Fix move to return only int, draw functions cannot handle floats as coordinates
|
Fix move to return only int, draw functions cannot handle floats as coordinates
|
Python
|
apache-2.0
|
PGHM/spacebattle
|
from math import cos, sin, pi, hypot
def rotate(polygon, angle):
rotated_points = []
cos_result = cos(angle)
sin_result = sin(angle)
for point in polygon:
x = point[0] * cos_result - point[1] * sin_result
y = point[0] * sin_result + point[1] * cos_result
rotated_points.append((x, y))
return rotated_points
def move(point, direction, amount):
- return [point[0] + amount * cos(direction),
+ return [int(point[0] + amount * cos(direction)),
- point[1] + amount * sin(direction)]
+ int(point[1] + amount * sin(direction))]
def distance(point1, point2):
return hypot(point1[0] - point2[0], point1[1] - point2[1])
|
Fix move to return only int, draw functions cannot handle floats as coordinates
|
## Code Before:
from math import cos, sin, pi, hypot
def rotate(polygon, angle):
rotated_points = []
cos_result = cos(angle)
sin_result = sin(angle)
for point in polygon:
x = point[0] * cos_result - point[1] * sin_result
y = point[0] * sin_result + point[1] * cos_result
rotated_points.append((x, y))
return rotated_points
def move(point, direction, amount):
return [point[0] + amount * cos(direction),
point[1] + amount * sin(direction)]
def distance(point1, point2):
return hypot(point1[0] - point2[0], point1[1] - point2[1])
## Instruction:
Fix move to return only int, draw functions cannot handle floats as coordinates
## Code After:
from math import cos, sin, pi, hypot
def rotate(polygon, angle):
rotated_points = []
cos_result = cos(angle)
sin_result = sin(angle)
for point in polygon:
x = point[0] * cos_result - point[1] * sin_result
y = point[0] * sin_result + point[1] * cos_result
rotated_points.append((x, y))
return rotated_points
def move(point, direction, amount):
return [int(point[0] + amount * cos(direction)),
int(point[1] + amount * sin(direction))]
def distance(point1, point2):
return hypot(point1[0] - point2[0], point1[1] - point2[1])
|
...
def move(point, direction, amount):
return [int(point[0] + amount * cos(direction)),
int(point[1] + amount * sin(direction))]
...
|
ddc44c6673cff4121eaaa47d8d075d63b82a85fe
|
runreport.py
|
runreport.py
|
import os
import json
import saulify.sitespec as sitespec
SPEC_DIRECTORY = "sitespecs"
if __name__ == "__main__":
for fname in os.listdir(SPEC_DIRECTORY):
fpath = os.path.join(SPEC_DIRECTORY, fname)
test_cases = sitespec.load_testcases(fpath)
for test_case in test_cases:
result = test_case.run()
print(json.dumps(result))
|
import os
import json
import argparse
import saulify.sitespec as sitespec
SPEC_DIRECTORY = "sitespecs"
parser = argparse.ArgumentParser()
parser.add_argument("-p", "--pretty", help="Pretty print test results",
action="store_true")
args = parser.parse_args()
def test_passed(report):
""" Whether all components of a scraper test succeeded """
if report["status"] != "OK":
return False
for result in report["result"].values():
if result["missing"]:
return False
return True
def print_report(report):
""" Converts test report dictionary to a human-readable format """
if report["status"] == "OK":
result = "PASS" if test_passed(report) else "FAIL"
else:
result = "EXCEPTION"
print("{0} : {1}".format(result, report["url"]))
if report["status"] == "EXCEPTION":
print(report["message"])
elif test_passed(report):
r = report["result"]
stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r])
print("Found " + stats)
else:
for category, result in report["result"].items():
if result["missing"]:
count = len(result["missing"])
print("Missing {0} {1}:".format(count, category))
for item in result["missing"]:
print(item)
if __name__ == "__main__":
for fname in os.listdir(SPEC_DIRECTORY):
fpath = os.path.join(SPEC_DIRECTORY, fname)
test_cases = sitespec.load_testcases(fpath)
for test_case in test_cases:
report = test_case.run()
if args.pretty:
print_report(report)
print("\n")
else:
print(json.dumps(report))
|
Add optional pretty printing to test runner
|
Add optional pretty printing to test runner
|
Python
|
agpl-3.0
|
asm-products/saulify-web,asm-products/saulify-web,asm-products/saulify-web
|
import os
import json
+ import argparse
import saulify.sitespec as sitespec
SPEC_DIRECTORY = "sitespecs"
+
+
+ parser = argparse.ArgumentParser()
+ parser.add_argument("-p", "--pretty", help="Pretty print test results",
+ action="store_true")
+ args = parser.parse_args()
+
+
+ def test_passed(report):
+ """ Whether all components of a scraper test succeeded """
+
+ if report["status"] != "OK":
+ return False
+
+ for result in report["result"].values():
+ if result["missing"]:
+ return False
+
+ return True
+
+
+ def print_report(report):
+ """ Converts test report dictionary to a human-readable format """
+
+ if report["status"] == "OK":
+ result = "PASS" if test_passed(report) else "FAIL"
+ else:
+ result = "EXCEPTION"
+
+ print("{0} : {1}".format(result, report["url"]))
+
+ if report["status"] == "EXCEPTION":
+ print(report["message"])
+
+ elif test_passed(report):
+ r = report["result"]
+ stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r])
+ print("Found " + stats)
+
+ else:
+ for category, result in report["result"].items():
+ if result["missing"]:
+ count = len(result["missing"])
+ print("Missing {0} {1}:".format(count, category))
+ for item in result["missing"]:
+ print(item)
if __name__ == "__main__":
for fname in os.listdir(SPEC_DIRECTORY):
fpath = os.path.join(SPEC_DIRECTORY, fname)
test_cases = sitespec.load_testcases(fpath)
for test_case in test_cases:
- result = test_case.run()
+ report = test_case.run()
+ if args.pretty:
+ print_report(report)
+ print("\n")
+ else:
- print(json.dumps(result))
+ print(json.dumps(report))
|
Add optional pretty printing to test runner
|
## Code Before:
import os
import json
import saulify.sitespec as sitespec
SPEC_DIRECTORY = "sitespecs"
if __name__ == "__main__":
for fname in os.listdir(SPEC_DIRECTORY):
fpath = os.path.join(SPEC_DIRECTORY, fname)
test_cases = sitespec.load_testcases(fpath)
for test_case in test_cases:
result = test_case.run()
print(json.dumps(result))
## Instruction:
Add optional pretty printing to test runner
## Code After:
import os
import json
import argparse
import saulify.sitespec as sitespec
SPEC_DIRECTORY = "sitespecs"
parser = argparse.ArgumentParser()
parser.add_argument("-p", "--pretty", help="Pretty print test results",
action="store_true")
args = parser.parse_args()
def test_passed(report):
""" Whether all components of a scraper test succeeded """
if report["status"] != "OK":
return False
for result in report["result"].values():
if result["missing"]:
return False
return True
def print_report(report):
""" Converts test report dictionary to a human-readable format """
if report["status"] == "OK":
result = "PASS" if test_passed(report) else "FAIL"
else:
result = "EXCEPTION"
print("{0} : {1}".format(result, report["url"]))
if report["status"] == "EXCEPTION":
print(report["message"])
elif test_passed(report):
r = report["result"]
stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r])
print("Found " + stats)
else:
for category, result in report["result"].items():
if result["missing"]:
count = len(result["missing"])
print("Missing {0} {1}:".format(count, category))
for item in result["missing"]:
print(item)
if __name__ == "__main__":
for fname in os.listdir(SPEC_DIRECTORY):
fpath = os.path.join(SPEC_DIRECTORY, fname)
test_cases = sitespec.load_testcases(fpath)
for test_case in test_cases:
report = test_case.run()
if args.pretty:
print_report(report)
print("\n")
else:
print(json.dumps(report))
|
# ... existing code ...
import json
import argparse
# ... modified code ...
SPEC_DIRECTORY = "sitespecs"
parser = argparse.ArgumentParser()
parser.add_argument("-p", "--pretty", help="Pretty print test results",
action="store_true")
args = parser.parse_args()
def test_passed(report):
""" Whether all components of a scraper test succeeded """
if report["status"] != "OK":
return False
for result in report["result"].values():
if result["missing"]:
return False
return True
def print_report(report):
""" Converts test report dictionary to a human-readable format """
if report["status"] == "OK":
result = "PASS" if test_passed(report) else "FAIL"
else:
result = "EXCEPTION"
print("{0} : {1}".format(result, report["url"]))
if report["status"] == "EXCEPTION":
print(report["message"])
elif test_passed(report):
r = report["result"]
stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r])
print("Found " + stats)
else:
for category, result in report["result"].items():
if result["missing"]:
count = len(result["missing"])
print("Missing {0} {1}:".format(count, category))
for item in result["missing"]:
print(item)
...
for test_case in test_cases:
report = test_case.run()
if args.pretty:
print_report(report)
print("\n")
else:
print(json.dumps(report))
# ... rest of the code ...
|
a35d6f59d214741f554dde1363d2eac7addb04cb
|
crypto_enigma/__init__.py
|
crypto_enigma/__init__.py
|
"""An Enigma machine simulator with rich textual display functionality."""
from ._version import __version__, __author__
#__all__ = ['machine', 'components']
from .components import *
from .machine import *
|
from ._version import __version__, __author__
#__all__ = ['machine', 'components']
from .components import *
from .machine import *
|
Add limitations to package documentation
|
Add limitations to package documentation
|
Python
|
bsd-3-clause
|
orome/crypto-enigma-py
|
-
- """An Enigma machine simulator with rich textual display functionality."""
from ._version import __version__, __author__
#__all__ = ['machine', 'components']
from .components import *
from .machine import *
|
Add limitations to package documentation
|
## Code Before:
"""An Enigma machine simulator with rich textual display functionality."""
from ._version import __version__, __author__
#__all__ = ['machine', 'components']
from .components import *
from .machine import *
## Instruction:
Add limitations to package documentation
## Code After:
from ._version import __version__, __author__
#__all__ = ['machine', 'components']
from .components import *
from .machine import *
|
// ... existing code ...
// ... rest of the code ...
|
ba1186c47e5f3466faeea9f2d5bf96948d5f7183
|
confuzzle.py
|
confuzzle.py
|
import sys
import argparse
import yaml
from jinja2 import Template
def render(template_string, context_dict):
template = Template(template_string)
return template.render(**context_dict)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used")
parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read")
parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used")
args = parser.parse_args()
context_dict = yaml.load(args.config.read())
template_string = args.template.read()
rendered = render(template_string, context_dict)
args.out.write(rendered)
if __name__ == "__main__":
main()
|
import sys
import argparse
import yaml
import jinja2
def render(template_string, context_dict, strict=False):
template = jinja2.Template(template_string)
if strict:
template.environment.undefined = jinja2.StrictUndefined
return template.render(**context_dict)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used")
parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read")
parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used")
parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables")
args = parser.parse_args()
context_dict = yaml.load(args.config.read())
template_string = args.template.read()
rendered = render(template_string, context_dict, args.strict)
args.out.write(rendered)
if __name__ == "__main__":
main()
|
Add --strict flag to raise exception on undefined variables
|
Add --strict flag to raise exception on undefined variables
|
Python
|
unlicense
|
j4mie/confuzzle
|
import sys
import argparse
import yaml
- from jinja2 import Template
+ import jinja2
- def render(template_string, context_dict):
+ def render(template_string, context_dict, strict=False):
- template = Template(template_string)
+ template = jinja2.Template(template_string)
+ if strict:
+ template.environment.undefined = jinja2.StrictUndefined
return template.render(**context_dict)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used")
parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read")
parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used")
+ parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables")
args = parser.parse_args()
context_dict = yaml.load(args.config.read())
template_string = args.template.read()
- rendered = render(template_string, context_dict)
+ rendered = render(template_string, context_dict, args.strict)
args.out.write(rendered)
if __name__ == "__main__":
main()
|
Add --strict flag to raise exception on undefined variables
|
## Code Before:
import sys
import argparse
import yaml
from jinja2 import Template
def render(template_string, context_dict):
template = Template(template_string)
return template.render(**context_dict)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used")
parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read")
parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used")
args = parser.parse_args()
context_dict = yaml.load(args.config.read())
template_string = args.template.read()
rendered = render(template_string, context_dict)
args.out.write(rendered)
if __name__ == "__main__":
main()
## Instruction:
Add --strict flag to raise exception on undefined variables
## Code After:
import sys
import argparse
import yaml
import jinja2
def render(template_string, context_dict, strict=False):
template = jinja2.Template(template_string)
if strict:
template.environment.undefined = jinja2.StrictUndefined
return template.render(**context_dict)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used")
parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read")
parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used")
parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables")
args = parser.parse_args()
context_dict = yaml.load(args.config.read())
template_string = args.template.read()
rendered = render(template_string, context_dict, args.strict)
args.out.write(rendered)
if __name__ == "__main__":
main()
|
# ... existing code ...
import yaml
import jinja2
# ... modified code ...
def render(template_string, context_dict, strict=False):
template = jinja2.Template(template_string)
if strict:
template.environment.undefined = jinja2.StrictUndefined
return template.render(**context_dict)
...
parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used")
parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables")
...
rendered = render(template_string, context_dict, args.strict)
args.out.write(rendered)
# ... rest of the code ...
|
e890ac9ef00193beac77b757c62911553cebf656
|
test.py
|
test.py
|
import urllib
urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', '/home/pi/img/img.jpg')
|
import urllib
urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg')
|
Change save path to local path
|
Change save path to local path
|
Python
|
mit
|
adampiskorski/lpr_poc
|
import urllib
- urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', '/home/pi/img/img.jpg')
+ urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg')
|
Change save path to local path
|
## Code Before:
import urllib
urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', '/home/pi/img/img.jpg')
## Instruction:
Change save path to local path
## Code After:
import urllib
urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg')
|
// ... existing code ...
import urllib
urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg')
// ... rest of the code ...
|
1cc6ec9f328d3ce045a4a1a50138b11c0b23cc3a
|
pyfr/ctypesutil.py
|
pyfr/ctypesutil.py
|
import ctypes
import ctypes.util
import os
import sys
def find_libc():
if sys.platform == 'win32':
return ctypes.util.find_msvcrt()
else:
return ctypes.util.find_library('c')
def load_library(name):
lname = platform_libname(name)
sdirs = platform_libdirs()
# First attempt to utilise the system search path
try:
return ctypes.CDLL(lname)
# Otherwise, if this fails then run our own search
except OSError:
for sd in sdirs:
try:
return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname)))
except OSError:
pass
else:
raise OSError('Unable to load {0}'.format(name))
def platform_libname(name):
if sys.platform == 'darwin':
return 'lib{0}.dylib'.format(name)
elif sys.platform == 'win32':
return '{0}.dll'.format(name)
else:
return 'lib{0}.so'.format(name)
def platform_libdirs():
path = os.environ.get('PYFR_LIBRARY_PATH', '')
dirs = [d for d in path.split(':') if d]
# On Mac OS X append the default path used by MacPorts
if sys.platform == 'darwin':
return dirs + ['/opt/local/lib']
# Otherwise just return
else:
return dirs
|
import ctypes
import ctypes.util
import os
import sys
def find_libc():
if sys.platform == 'win32':
return ctypes.util.find_msvcrt()
else:
return ctypes.util.find_library('c')
def load_library(name):
# If an explicit override has been given then use it
lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper()))
if lpath:
return ctypes.CDLL(lpath)
# Otherwise synthesise the library name and start searching
lname = platform_libname(name)
# Start with system search path
try:
return ctypes.CDLL(lname)
# ..and if this fails then run our own search
except OSError:
for sd in platform_libdirs():
try:
return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname)))
except OSError:
pass
else:
raise OSError('Unable to load {0}'.format(name))
def platform_libname(name):
if sys.platform == 'darwin':
return 'lib{0}.dylib'.format(name)
elif sys.platform == 'win32':
return '{0}.dll'.format(name)
else:
return 'lib{0}.so'.format(name)
def platform_libdirs():
path = os.environ.get('PYFR_LIBRARY_PATH', '')
dirs = [d for d in path.split(':') if d]
# On Mac OS X append the default path used by MacPorts
if sys.platform == 'darwin':
return dirs + ['/opt/local/lib']
# Otherwise just return
else:
return dirs
|
Enable library paths to be explicitly specified.
|
Enable library paths to be explicitly specified.
All shared libraries loaded through the load_library function
can bow be specified explicitly through a suitable environmental
variable
PYFR_<LIB>_LIBRARY_PATH=/path/to/lib.here
where <LIB> corresponds to the name of the library, e.g. METIS.
|
Python
|
bsd-3-clause
|
BrianVermeire/PyFR
|
import ctypes
import ctypes.util
import os
import sys
def find_libc():
if sys.platform == 'win32':
return ctypes.util.find_msvcrt()
else:
return ctypes.util.find_library('c')
def load_library(name):
+ # If an explicit override has been given then use it
+ lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper()))
+ if lpath:
+ return ctypes.CDLL(lpath)
+
+ # Otherwise synthesise the library name and start searching
lname = platform_libname(name)
- sdirs = platform_libdirs()
- # First attempt to utilise the system search path
+ # Start with system search path
try:
return ctypes.CDLL(lname)
- # Otherwise, if this fails then run our own search
+ # ..and if this fails then run our own search
except OSError:
- for sd in sdirs:
+ for sd in platform_libdirs():
try:
return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname)))
except OSError:
pass
else:
raise OSError('Unable to load {0}'.format(name))
def platform_libname(name):
if sys.platform == 'darwin':
return 'lib{0}.dylib'.format(name)
elif sys.platform == 'win32':
return '{0}.dll'.format(name)
else:
return 'lib{0}.so'.format(name)
def platform_libdirs():
path = os.environ.get('PYFR_LIBRARY_PATH', '')
dirs = [d for d in path.split(':') if d]
# On Mac OS X append the default path used by MacPorts
if sys.platform == 'darwin':
return dirs + ['/opt/local/lib']
# Otherwise just return
else:
return dirs
|
Enable library paths to be explicitly specified.
|
## Code Before:
import ctypes
import ctypes.util
import os
import sys
def find_libc():
if sys.platform == 'win32':
return ctypes.util.find_msvcrt()
else:
return ctypes.util.find_library('c')
def load_library(name):
lname = platform_libname(name)
sdirs = platform_libdirs()
# First attempt to utilise the system search path
try:
return ctypes.CDLL(lname)
# Otherwise, if this fails then run our own search
except OSError:
for sd in sdirs:
try:
return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname)))
except OSError:
pass
else:
raise OSError('Unable to load {0}'.format(name))
def platform_libname(name):
if sys.platform == 'darwin':
return 'lib{0}.dylib'.format(name)
elif sys.platform == 'win32':
return '{0}.dll'.format(name)
else:
return 'lib{0}.so'.format(name)
def platform_libdirs():
path = os.environ.get('PYFR_LIBRARY_PATH', '')
dirs = [d for d in path.split(':') if d]
# On Mac OS X append the default path used by MacPorts
if sys.platform == 'darwin':
return dirs + ['/opt/local/lib']
# Otherwise just return
else:
return dirs
## Instruction:
Enable library paths to be explicitly specified.
## Code After:
import ctypes
import ctypes.util
import os
import sys
def find_libc():
if sys.platform == 'win32':
return ctypes.util.find_msvcrt()
else:
return ctypes.util.find_library('c')
def load_library(name):
# If an explicit override has been given then use it
lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper()))
if lpath:
return ctypes.CDLL(lpath)
# Otherwise synthesise the library name and start searching
lname = platform_libname(name)
# Start with system search path
try:
return ctypes.CDLL(lname)
# ..and if this fails then run our own search
except OSError:
for sd in platform_libdirs():
try:
return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname)))
except OSError:
pass
else:
raise OSError('Unable to load {0}'.format(name))
def platform_libname(name):
if sys.platform == 'darwin':
return 'lib{0}.dylib'.format(name)
elif sys.platform == 'win32':
return '{0}.dll'.format(name)
else:
return 'lib{0}.so'.format(name)
def platform_libdirs():
path = os.environ.get('PYFR_LIBRARY_PATH', '')
dirs = [d for d in path.split(':') if d]
# On Mac OS X append the default path used by MacPorts
if sys.platform == 'darwin':
return dirs + ['/opt/local/lib']
# Otherwise just return
else:
return dirs
|
# ... existing code ...
def load_library(name):
# If an explicit override has been given then use it
lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper()))
if lpath:
return ctypes.CDLL(lpath)
# Otherwise synthesise the library name and start searching
lname = platform_libname(name)
# Start with system search path
try:
# ... modified code ...
return ctypes.CDLL(lname)
# ..and if this fails then run our own search
except OSError:
for sd in platform_libdirs():
try:
# ... rest of the code ...
|
86a992dc15482087773f1591752a667a6014ba5d
|
docker/settings/celery.py
|
docker/settings/celery.py
|
from .docker_compose import DockerBaseSettings
class CeleryDevSettings(DockerBaseSettings):
pass
CeleryDevSettings.load_settings(__name__)
|
from .docker_compose import DockerBaseSettings
class CeleryDevSettings(DockerBaseSettings):
# Since we can't properly set CORS on Azurite container
# (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561)
# trying to fetch ``objects.inv`` from celery container fails because the
# URL is like http://docs.dev.readthedocs.io/... and it should be
# http://storage:10000/... This setting fixes that.
# Once we can use CORS, we should define this setting in the
# ``docker_compose.py`` file instead.
AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000'
CeleryDevSettings.load_settings(__name__)
|
Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME
|
Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME
We can't access docs.dev.readthedocs.io from celery container because
that domain points to 127.0.0.1 and we don't have the storage in that
IP. So, we need to override the AZURE_MEDIA_STORAGE_HOSTNAME in the
celery container to point to the storage.
We should do this directly in `docker_compose.py` settings file, but
since we can't configure CORS in Azurite we can't do it yet.
|
Python
|
mit
|
rtfd/readthedocs.org,rtfd/readthedocs.org,rtfd/readthedocs.org,rtfd/readthedocs.org
|
from .docker_compose import DockerBaseSettings
class CeleryDevSettings(DockerBaseSettings):
- pass
+ # Since we can't properly set CORS on Azurite container
+ # (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561)
+ # trying to fetch ``objects.inv`` from celery container fails because the
+ # URL is like http://docs.dev.readthedocs.io/... and it should be
+ # http://storage:10000/... This setting fixes that.
+ # Once we can use CORS, we should define this setting in the
+ # ``docker_compose.py`` file instead.
+ AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000'
+
CeleryDevSettings.load_settings(__name__)
|
Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME
|
## Code Before:
from .docker_compose import DockerBaseSettings
class CeleryDevSettings(DockerBaseSettings):
pass
CeleryDevSettings.load_settings(__name__)
## Instruction:
Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME
## Code After:
from .docker_compose import DockerBaseSettings
class CeleryDevSettings(DockerBaseSettings):
# Since we can't properly set CORS on Azurite container
# (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561)
# trying to fetch ``objects.inv`` from celery container fails because the
# URL is like http://docs.dev.readthedocs.io/... and it should be
# http://storage:10000/... This setting fixes that.
# Once we can use CORS, we should define this setting in the
# ``docker_compose.py`` file instead.
AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000'
CeleryDevSettings.load_settings(__name__)
|
# ... existing code ...
class CeleryDevSettings(DockerBaseSettings):
# Since we can't properly set CORS on Azurite container
# (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561)
# trying to fetch ``objects.inv`` from celery container fails because the
# URL is like http://docs.dev.readthedocs.io/... and it should be
# http://storage:10000/... This setting fixes that.
# Once we can use CORS, we should define this setting in the
# ``docker_compose.py`` file instead.
AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000'
# ... rest of the code ...
|
1666f883e3f6a497971b484c9ba875df2f6693a2
|
test/testall.py
|
test/testall.py
|
import os
import re
import sys
from _common import unittest
pkgpath = os.path.dirname(__file__) or '.'
sys.path.append(pkgpath)
os.chdir(pkgpath)
def suite():
s = unittest.TestSuite()
# Get the suite() of every module in this directory beginning with
# "test_".
for fname in os.listdir(pkgpath):
match = re.match(r'(test_\S+)\.py$', fname)
if match:
modname = match.group(1)
s.addTest(__import__(modname).suite())
return s
if __name__ == '__main__':
unittest.main(defaultTest='suite')
|
import os
import re
import sys
from _common import unittest
pkgpath = os.path.dirname(__file__) or '.'
sys.path.append(pkgpath)
os.chdir(pkgpath)
# Make sure we use local version of beetsplug and not system namespaced version
# for tests
try:
del sys.modules["beetsplug"]
except KeyError:
pass
def suite():
s = unittest.TestSuite()
# Get the suite() of every module in this directory beginning with
# "test_".
for fname in os.listdir(pkgpath):
match = re.match(r'(test_\S+)\.py$', fname)
if match:
modname = match.group(1)
s.addTest(__import__(modname).suite())
return s
if __name__ == '__main__':
unittest.main(defaultTest='suite')
|
Fix python namespaces for test runs
|
Fix python namespaces for test runs
We need to make sure we don't use namespaced versions that are already installed
on the system but rather use local version from current sources
|
Python
|
mit
|
SusannaMaria/beets,mathstuf/beets,mathstuf/beets,YetAnotherNerd/beets,lengtche/beets,LordSputnik/beets,shamangeorge/beets,ibmibmibm/beets,m-urban/beets,krig/beets,lightwang1/beets,shamangeorge/beets,MyTunesFreeMusic/privacy-policy,jcoady9/beets,SusannaMaria/beets,beetbox/beets,Andypsamp/CODfinalJUNIT,Andypsamp/CODfinalJUNIT,jcoady9/beets,pkess/beets,PierreRust/beets,tima/beets,mried/beets,pkess/beets,Freso/beets,bj-yinyan/beets,beetbox/beets,dfc/beets,YetAnotherNerd/beets,tima/beets,ruippeixotog/beets,diego-plan9/beets,drm00/beets,ruippeixotog/beets,marcuskrahl/beets,kareemallen/beets,arabenjamin/beets,drm00/beets,parapente/beets,Dishwishy/beets,madmouser1/beets,imsparsh/beets,Freso/beets,mathstuf/beets,andremiller/beets,LordSputnik/beets,moodboom/beets,YetAnotherNerd/beets,mosesfistos1/beetbox,multikatt/beets,jackwilsdon/beets,jmwatte/beets,jayme-github/beets,asteven/beets,xsteadfastx/beets,m-urban/beets,bj-yinyan/beets,YetAnotherNerd/beets,LordSputnik/beets,Dishwishy/beets,Kraymer/beets,mosesfistos1/beetbox,ruippeixotog/beets,jcoady9/beets,randybias/beets,untitaker/beets,PierreRust/beets,beetbox/beets,mried/beets,artemutin/beets,shanemikel/beets,Freso/beets,Andypsamp/CODfinalJUNIT,lightwang1/beets,shanemikel/beets,kelvinhammond/beets,mried/beets,gabrielaraujof/beets,ttsda/beets,randybias/beets,krig/beets,sadatay/beets,sampsyo/beets,parapente/beets,kareemallen/beets,ttsda/beets,swt30/beets,PierreRust/beets,imsparsh/beets,sampsyo/beets,madmouser1/beets,gabrielaraujof/beets,pkess/beets,jackwilsdon/beets,m-urban/beets,arabenjamin/beets,drm00/beets,Andypsamp/CODfinalJUNIT,beetbox/beets,sadatay/beets,ibmibmibm/beets,untitaker/beets,moodboom/beets,SusannaMaria/beets,sampsyo/beets,tima/beets,ttsda/beets,jmwatte/beets,kelvinhammond/beets,drm00/beets,jayme-github/beets,xsteadfastx/beets,kareemallen/beets,jackwilsdon/beets,MyTunesFreeMusic/privacy-policy,xsteadfastx/beets,jbaiter/beets,m-urban/beets,parapente/beets,lengtche/beets,randybias/beets,mosesfistos1/beetbox,tima/beets,andremiller/beets,PierreRust/beets,kelvinhammond/beets,artemutin/beets,marcuskrahl/beets,diego-plan9/beets,xsteadfastx/beets,gabrielaraujof/beets,arabenjamin/beets,Wen777/beets,imsparsh/beets,swt30/beets,arabenjamin/beets,diego-plan9/beets,asteven/beets,MyTunesFreeMusic/privacy-policy,lengtche/beets,dfc/beets,sampsyo/beets,bj-yinyan/beets,Kraymer/beets,moodboom/beets,shanemikel/beets,swt30/beets,madmouser1/beets,asteven/beets,Freso/beets,ttsda/beets,Kraymer/beets,randybias/beets,Andypsamp/CODjunit,parapente/beets,Wen777/beets,jcoady9/beets,swt30/beets,multikatt/beets,bj-yinyan/beets,kareemallen/beets,ruippeixotog/beets,Andypsamp/CODjunit,shamangeorge/beets,lengtche/beets,MyTunesFreeMusic/privacy-policy,lightwang1/beets,lightwang1/beets,LordSputnik/beets,artemutin/beets,Wen777/beets,untitaker/beets,multikatt/beets,Andypsamp/CODfinalJUNIT,marcuskrahl/beets,shamangeorge/beets,andremiller/beets,mried/beets,jackwilsdon/beets,dfc/beets,gabrielaraujof/beets,mosesfistos1/beetbox,SusannaMaria/beets,marcuskrahl/beets,asteven/beets,Andypsamp/CODjunit,moodboom/beets,madmouser1/beets,ibmibmibm/beets,dfc/beets,artemutin/beets,diego-plan9/beets,sadatay/beets,Andypsamp/CODjunit,sadatay/beets,Dishwishy/beets,ibmibmibm/beets,Dishwishy/beets,mathstuf/beets,Kraymer/beets,pkess/beets,imsparsh/beets,Andypsamp/CODjunit,krig/beets,jbaiter/beets,jmwatte/beets,multikatt/beets,jmwatte/beets,shanemikel/beets,kelvinhammond/beets,untitaker/beets
|
import os
import re
import sys
from _common import unittest
pkgpath = os.path.dirname(__file__) or '.'
sys.path.append(pkgpath)
os.chdir(pkgpath)
+
+ # Make sure we use local version of beetsplug and not system namespaced version
+ # for tests
+ try:
+ del sys.modules["beetsplug"]
+ except KeyError:
+ pass
def suite():
s = unittest.TestSuite()
# Get the suite() of every module in this directory beginning with
# "test_".
for fname in os.listdir(pkgpath):
match = re.match(r'(test_\S+)\.py$', fname)
if match:
modname = match.group(1)
s.addTest(__import__(modname).suite())
return s
if __name__ == '__main__':
unittest.main(defaultTest='suite')
|
Fix python namespaces for test runs
|
## Code Before:
import os
import re
import sys
from _common import unittest
pkgpath = os.path.dirname(__file__) or '.'
sys.path.append(pkgpath)
os.chdir(pkgpath)
def suite():
s = unittest.TestSuite()
# Get the suite() of every module in this directory beginning with
# "test_".
for fname in os.listdir(pkgpath):
match = re.match(r'(test_\S+)\.py$', fname)
if match:
modname = match.group(1)
s.addTest(__import__(modname).suite())
return s
if __name__ == '__main__':
unittest.main(defaultTest='suite')
## Instruction:
Fix python namespaces for test runs
## Code After:
import os
import re
import sys
from _common import unittest
pkgpath = os.path.dirname(__file__) or '.'
sys.path.append(pkgpath)
os.chdir(pkgpath)
# Make sure we use local version of beetsplug and not system namespaced version
# for tests
try:
del sys.modules["beetsplug"]
except KeyError:
pass
def suite():
s = unittest.TestSuite()
# Get the suite() of every module in this directory beginning with
# "test_".
for fname in os.listdir(pkgpath):
match = re.match(r'(test_\S+)\.py$', fname)
if match:
modname = match.group(1)
s.addTest(__import__(modname).suite())
return s
if __name__ == '__main__':
unittest.main(defaultTest='suite')
|
// ... existing code ...
os.chdir(pkgpath)
# Make sure we use local version of beetsplug and not system namespaced version
# for tests
try:
del sys.modules["beetsplug"]
except KeyError:
pass
// ... rest of the code ...
|
37fa40a9b5260f8090adaa8c15d3767c0867574f
|
python/fusion_engine_client/messages/__init__.py
|
python/fusion_engine_client/messages/__init__.py
|
from .core import *
from . import ros
message_type_to_class = {
# Navigation solution messages.
PoseMessage.MESSAGE_TYPE: PoseMessage,
PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage,
GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage,
GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage,
# Sensor measurement messages.
IMUMeasurement.MESSAGE_TYPE: IMUMeasurement,
# ROS messages.
ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage,
ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage,
ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage,
# Command and control messages.
CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage,
MessageRequest.MESSAGE_TYPE: MessageRequest,
ResetRequest.MESSAGE_TYPE: ResetRequest,
VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage,
EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage,
}
|
from .core import *
from . import ros
message_type_to_class = {
# Navigation solution messages.
PoseMessage.MESSAGE_TYPE: PoseMessage,
PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage,
GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage,
GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage,
# Sensor measurement messages.
IMUMeasurement.MESSAGE_TYPE: IMUMeasurement,
# ROS messages.
ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage,
ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage,
ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage,
# Command and control messages.
CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage,
MessageRequest.MESSAGE_TYPE: MessageRequest,
ResetRequest.MESSAGE_TYPE: ResetRequest,
VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage,
EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage,
}
messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')]
|
Create a list of messages that contain system time.
|
Create a list of messages that contain system time.
|
Python
|
mit
|
PointOneNav/fusion-engine-client,PointOneNav/fusion-engine-client,PointOneNav/fusion-engine-client
|
from .core import *
from . import ros
message_type_to_class = {
# Navigation solution messages.
PoseMessage.MESSAGE_TYPE: PoseMessage,
PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage,
GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage,
GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage,
# Sensor measurement messages.
IMUMeasurement.MESSAGE_TYPE: IMUMeasurement,
# ROS messages.
ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage,
ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage,
ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage,
# Command and control messages.
CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage,
MessageRequest.MESSAGE_TYPE: MessageRequest,
ResetRequest.MESSAGE_TYPE: ResetRequest,
VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage,
EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage,
}
+ messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')]
+
|
Create a list of messages that contain system time.
|
## Code Before:
from .core import *
from . import ros
message_type_to_class = {
# Navigation solution messages.
PoseMessage.MESSAGE_TYPE: PoseMessage,
PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage,
GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage,
GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage,
# Sensor measurement messages.
IMUMeasurement.MESSAGE_TYPE: IMUMeasurement,
# ROS messages.
ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage,
ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage,
ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage,
# Command and control messages.
CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage,
MessageRequest.MESSAGE_TYPE: MessageRequest,
ResetRequest.MESSAGE_TYPE: ResetRequest,
VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage,
EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage,
}
## Instruction:
Create a list of messages that contain system time.
## Code After:
from .core import *
from . import ros
message_type_to_class = {
# Navigation solution messages.
PoseMessage.MESSAGE_TYPE: PoseMessage,
PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage,
GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage,
GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage,
# Sensor measurement messages.
IMUMeasurement.MESSAGE_TYPE: IMUMeasurement,
# ROS messages.
ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage,
ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage,
ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage,
# Command and control messages.
CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage,
MessageRequest.MESSAGE_TYPE: MessageRequest,
ResetRequest.MESSAGE_TYPE: ResetRequest,
VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage,
EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage,
}
messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')]
|
...
}
messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')]
...
|
c354d130cb542c2a5d57e519ce49175daa597e9c
|
froide/accesstoken/apps.py
|
froide/accesstoken/apps.py
|
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class AccessTokenConfig(AppConfig):
name = 'froide.accesstoken'
verbose_name = _('Secret Access Token')
def ready(self):
from froide.account import account_canceled
account_canceled.connect(cancel_user)
def cancel_user(sender, user=None, **kwargs):
from .models import AccessToken
if user is None:
return
AccessToken.objects.filter(user=user).delete()
|
import json
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class AccessTokenConfig(AppConfig):
name = 'froide.accesstoken'
verbose_name = _('Secret Access Token')
def ready(self):
from froide.account import account_canceled
from froide.account.export import registry
account_canceled.connect(cancel_user)
registry.register(export_user_data)
def cancel_user(sender, user=None, **kwargs):
from .models import AccessToken
if user is None:
return
AccessToken.objects.filter(user=user).delete()
def export_user_data(user):
from .models import AccessToken
access_tokens = (
AccessToken.objects.filter(user=user)
)
if access_tokens:
yield ('access_tokens.json', json.dumps([
{
'purpose': a.purpose,
'timestamp': a.timestamp.isoformat(),
}
for a in access_tokens]).encode('utf-8')
)
|
Add user data export for accesstokens
|
Add user data export for accesstokens
|
Python
|
mit
|
fin/froide,fin/froide,fin/froide,fin/froide,stefanw/froide,stefanw/froide,stefanw/froide,stefanw/froide,stefanw/froide
|
+ import json
+
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class AccessTokenConfig(AppConfig):
name = 'froide.accesstoken'
verbose_name = _('Secret Access Token')
def ready(self):
from froide.account import account_canceled
+ from froide.account.export import registry
account_canceled.connect(cancel_user)
+ registry.register(export_user_data)
def cancel_user(sender, user=None, **kwargs):
from .models import AccessToken
if user is None:
return
AccessToken.objects.filter(user=user).delete()
+
+ def export_user_data(user):
+ from .models import AccessToken
+
+ access_tokens = (
+ AccessToken.objects.filter(user=user)
+ )
+ if access_tokens:
+ yield ('access_tokens.json', json.dumps([
+ {
+ 'purpose': a.purpose,
+ 'timestamp': a.timestamp.isoformat(),
+ }
+ for a in access_tokens]).encode('utf-8')
+ )
+
|
Add user data export for accesstokens
|
## Code Before:
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class AccessTokenConfig(AppConfig):
name = 'froide.accesstoken'
verbose_name = _('Secret Access Token')
def ready(self):
from froide.account import account_canceled
account_canceled.connect(cancel_user)
def cancel_user(sender, user=None, **kwargs):
from .models import AccessToken
if user is None:
return
AccessToken.objects.filter(user=user).delete()
## Instruction:
Add user data export for accesstokens
## Code After:
import json
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class AccessTokenConfig(AppConfig):
name = 'froide.accesstoken'
verbose_name = _('Secret Access Token')
def ready(self):
from froide.account import account_canceled
from froide.account.export import registry
account_canceled.connect(cancel_user)
registry.register(export_user_data)
def cancel_user(sender, user=None, **kwargs):
from .models import AccessToken
if user is None:
return
AccessToken.objects.filter(user=user).delete()
def export_user_data(user):
from .models import AccessToken
access_tokens = (
AccessToken.objects.filter(user=user)
)
if access_tokens:
yield ('access_tokens.json', json.dumps([
{
'purpose': a.purpose,
'timestamp': a.timestamp.isoformat(),
}
for a in access_tokens]).encode('utf-8')
)
|
// ... existing code ...
import json
from django.apps import AppConfig
// ... modified code ...
from froide.account import account_canceled
from froide.account.export import registry
...
account_canceled.connect(cancel_user)
registry.register(export_user_data)
...
AccessToken.objects.filter(user=user).delete()
def export_user_data(user):
from .models import AccessToken
access_tokens = (
AccessToken.objects.filter(user=user)
)
if access_tokens:
yield ('access_tokens.json', json.dumps([
{
'purpose': a.purpose,
'timestamp': a.timestamp.isoformat(),
}
for a in access_tokens]).encode('utf-8')
)
// ... rest of the code ...
|
6160da958f4b8ecb1553c7bcca0b32bc1a5a1649
|
tests/conftest.py
|
tests/conftest.py
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
shutil.rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
shutil.rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
|
Make sure .git test directory is removed on Windows
|
Make sure .git test directory is removed on Windows
|
Python
|
bsd-3-clause
|
scopatz/rever,ergs/rever
|
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
+ import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
- shutil.rmtree(repo)
+ rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
- shutil.rmtree(repo)
+ rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
+
+ def rmtree(dirname):
+ """Remove a directory, even if it has read-only files (Windows).
+ Git creates read-only files that must be removed on teardown. See
+ https://stackoverflow.com/questions/2656322 for more info.
+
+ Parameters
+ ----------
+ dirname : str
+ Directory to be removed
+ """
+ try:
+ shutil.rmtree(dirname)
+ except PermissionError:
+ if sys.platform == 'win32':
+ subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
+ else:
+ raise
+
|
Make sure .git test directory is removed on Windows
|
## Code Before:
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
shutil.rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
shutil.rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
## Instruction:
Make sure .git test directory is removed on Windows
## Code After:
import os
import shutil
import tempfile
import builtins
import subprocess
import pytest
import sys
from rever import environ
@pytest.fixture
def gitrepo(request):
"""A test fixutre that creates and destroys a git repo in a temporary
directory.
This will yield the path to the repo.
"""
cwd = os.getcwd()
name = request.node.name
repo = os.path.join(tempfile.gettempdir(), name)
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
os.chdir(repo)
with open('README', 'w') as f:
f.write('testing ' + name)
subprocess.run(['git', 'add', '.'])
subprocess.run(['git', 'commit', '-am', 'Initial readme'])
with environ.context():
yield repo
os.chdir(cwd)
rmtree(repo)
@pytest.fixture
def gitecho(request):
aliases = builtins.aliases
aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n'
yield None
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
|
...
import pytest
import sys
...
if os.path.exists(repo):
rmtree(repo)
subprocess.run(['git', 'init', repo])
...
os.chdir(cwd)
rmtree(repo)
...
del aliases['git']
def rmtree(dirname):
"""Remove a directory, even if it has read-only files (Windows).
Git creates read-only files that must be removed on teardown. See
https://stackoverflow.com/questions/2656322 for more info.
Parameters
----------
dirname : str
Directory to be removed
"""
try:
shutil.rmtree(dirname)
except PermissionError:
if sys.platform == 'win32':
subprocess.check_call(['del', '/F/S/Q', dirname], shell=True)
else:
raise
...
|
6e6c60613180bb3d7e2d019129e57d1a2c33286d
|
backend/backend/models.py
|
backend/backend/models.py
|
from django.db import models
class Animal(models.Model):
MALE = 'male'
FEMALE = 'female'
GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female'))
father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father")
mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother")
name = models.CharField(max_length = 100)
dob = models.IntegerField()
gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE)
active = models.BooleanField()
own = models.BooleanField()
class Meta:
unique_together = ("name", "dob")
|
from django.db import models
from django.core.validators import MaxValueValidator, MaxLengthValidator
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
from datetime import datetime
def current_year():
return datetime.now().year
class Animal(models.Model):
MALE = 'male'
FEMALE = 'female'
GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female'))
father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father")
mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother")
name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)])
dob = models.IntegerField(validators = [MaxValueValidator(current_year())])
gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE)
active = models.BooleanField()
own = models.BooleanField()
class Meta:
unique_together = ("name", "dob")
|
Add length validator to name. Add dob validator can't be higher than current year.
|
Add length validator to name.
Add dob validator can't be higher than current year.
|
Python
|
apache-2.0
|
mmlado/animal_pairing,mmlado/animal_pairing
|
from django.db import models
+ from django.core.validators import MaxValueValidator, MaxLengthValidator
+ from django.core.exceptions import ValidationError
+ from django.utils.translation import gettext_lazy as _
+ from datetime import datetime
+
+ def current_year():
+ return datetime.now().year
class Animal(models.Model):
MALE = 'male'
FEMALE = 'female'
GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female'))
father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father")
mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother")
- name = models.CharField(max_length = 100)
- dob = models.IntegerField()
+ name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)])
+ dob = models.IntegerField(validators = [MaxValueValidator(current_year())])
gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE)
active = models.BooleanField()
own = models.BooleanField()
class Meta:
unique_together = ("name", "dob")
|
Add length validator to name. Add dob validator can't be higher than current year.
|
## Code Before:
from django.db import models
class Animal(models.Model):
MALE = 'male'
FEMALE = 'female'
GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female'))
father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father")
mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother")
name = models.CharField(max_length = 100)
dob = models.IntegerField()
gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE)
active = models.BooleanField()
own = models.BooleanField()
class Meta:
unique_together = ("name", "dob")
## Instruction:
Add length validator to name. Add dob validator can't be higher than current year.
## Code After:
from django.db import models
from django.core.validators import MaxValueValidator, MaxLengthValidator
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
from datetime import datetime
def current_year():
return datetime.now().year
class Animal(models.Model):
MALE = 'male'
FEMALE = 'female'
GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female'))
father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father")
mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother")
name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)])
dob = models.IntegerField(validators = [MaxValueValidator(current_year())])
gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE)
active = models.BooleanField()
own = models.BooleanField()
class Meta:
unique_together = ("name", "dob")
|
...
from django.db import models
from django.core.validators import MaxValueValidator, MaxLengthValidator
from django.core.exceptions import ValidationError
from django.utils.translation import gettext_lazy as _
from datetime import datetime
def current_year():
return datetime.now().year
...
mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother")
name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)])
dob = models.IntegerField(validators = [MaxValueValidator(current_year())])
gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE)
...
|
90a724313902e3d95f1a37d9102af1544c9bc61d
|
segments/set_term_title.py
|
segments/set_term_title.py
|
def add_term_title_segment():
term = os.getenv('TERM')
if not (('xterm' in term) or ('rxvt' in term)):
return
if powerline.args.shell == 'bash':
set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]'
elif powerline.args.shell == 'zsh':
set_title = '\\e]0;%n@%m: %~\\a'
else:
import socket
set_title = '\\e]0;%s@%s: %s\\a' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD'))
powerline.append(set_title, None, None, '')
add_term_title_segment()
|
def add_term_title_segment():
term = os.getenv('TERM')
if not (('xterm' in term) or ('rxvt' in term)):
return
if powerline.args.shell == 'bash':
set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]'
elif powerline.args.shell == 'zsh':
set_title = '\033]0;%n@%m: %~\007'
else:
import socket
set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD'))
powerline.append(set_title, None, None, '')
add_term_title_segment()
|
Fix use of escape characters in "set terminal title" segment.
|
Fix use of escape characters in "set terminal title" segment.
Escape characters were incorrect for non-BASH shells.
|
Python
|
mit
|
nicholascapo/powerline-shell,b-ryan/powerline-shell,junix/powerline-shell,wrgoldstein/powerline-shell,rbanffy/powerline-shell,b-ryan/powerline-shell,mart-e/powerline-shell,blieque/powerline-shell,paulhybryant/powerline-shell,tswsl1989/powerline-shell,torbjornvatn/powerline-shell,MartinWetterwald/powerline-shell,iKrishneel/powerline-shell,fellipecastro/powerline-shell,ceholden/powerline-shell,banga/powerline-shell,banga/powerline-shell,handsomecheung/powerline-shell,saghul/shline,strycore/powerline-shell,bitIO/powerline-shell,intfrr/powerline-shell,yc2prime/powerline-shell,mcdope/powerline-shell,milkbikis/powerline-shell,paulhybryant/powerline-shell,JulianVolodia/powerline-shell,dtrip/powerline-shell,paol/powerline-shell,Menci/powerline-shell,LeonardoGentile/powerline-shell
|
def add_term_title_segment():
term = os.getenv('TERM')
if not (('xterm' in term) or ('rxvt' in term)):
return
if powerline.args.shell == 'bash':
set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]'
elif powerline.args.shell == 'zsh':
- set_title = '\\e]0;%n@%m: %~\\a'
+ set_title = '\033]0;%n@%m: %~\007'
else:
import socket
- set_title = '\\e]0;%s@%s: %s\\a' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD'))
+ set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD'))
powerline.append(set_title, None, None, '')
add_term_title_segment()
|
Fix use of escape characters in "set terminal title" segment.
|
## Code Before:
def add_term_title_segment():
term = os.getenv('TERM')
if not (('xterm' in term) or ('rxvt' in term)):
return
if powerline.args.shell == 'bash':
set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]'
elif powerline.args.shell == 'zsh':
set_title = '\\e]0;%n@%m: %~\\a'
else:
import socket
set_title = '\\e]0;%s@%s: %s\\a' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD'))
powerline.append(set_title, None, None, '')
add_term_title_segment()
## Instruction:
Fix use of escape characters in "set terminal title" segment.
## Code After:
def add_term_title_segment():
term = os.getenv('TERM')
if not (('xterm' in term) or ('rxvt' in term)):
return
if powerline.args.shell == 'bash':
set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]'
elif powerline.args.shell == 'zsh':
set_title = '\033]0;%n@%m: %~\007'
else:
import socket
set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD'))
powerline.append(set_title, None, None, '')
add_term_title_segment()
|
# ... existing code ...
elif powerline.args.shell == 'zsh':
set_title = '\033]0;%n@%m: %~\007'
else:
# ... modified code ...
import socket
set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD'))
# ... rest of the code ...
|
a8f125236308cbfc9bb2eb5b225a0ac92a3a95e4
|
ANN.py
|
ANN.py
|
from random import random
class Neuron:
def __init__(self, parents=[]):
self.parents = parents
self.weights = [random() for parent in parents]
def get_output(self):
return sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1
output = property(get_output)
class NeuronNetwork:
neurons = []
def __init__(self, inputs, outputs, rows, columns):
self.neurons = []
for row in xrange(rows + 2):
self.neurons.append([])
if row == 0:
for input_ in xrange(inputs):
self.neurons[row].append(Neuron(parents=[]))
elif row == rows + 1:
for output in xrange(outputs):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
else:
for column in xrange(columns):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
|
from random import random
class Neuron:
output = None
def __init__(self, parents=[]):
self.parents = parents
self.weights = [random() for parent in parents]
def calculate(self):
self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1
class NeuronNetwork:
neurons = []
def __init__(self, inputs, outputs, rows, columns):
self.neurons = []
for row in xrange(rows + 2):
self.neurons.append([])
if row == 0:
for input_ in xrange(inputs):
self.neurons[row].append(Neuron(parents=[]))
elif row == rows + 1:
for output in xrange(outputs):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
else:
for column in xrange(columns):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
|
Store output instead of calculating it each time
|
Store output instead of calculating it each time
|
Python
|
mit
|
tysonzero/py-ann
|
from random import random
class Neuron:
+ output = None
+
def __init__(self, parents=[]):
self.parents = parents
self.weights = [random() for parent in parents]
- def get_output(self):
+ def calculate(self):
- return sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1
+ self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1
-
- output = property(get_output)
class NeuronNetwork:
neurons = []
def __init__(self, inputs, outputs, rows, columns):
self.neurons = []
for row in xrange(rows + 2):
self.neurons.append([])
if row == 0:
for input_ in xrange(inputs):
self.neurons[row].append(Neuron(parents=[]))
elif row == rows + 1:
for output in xrange(outputs):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
else:
for column in xrange(columns):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
|
Store output instead of calculating it each time
|
## Code Before:
from random import random
class Neuron:
def __init__(self, parents=[]):
self.parents = parents
self.weights = [random() for parent in parents]
def get_output(self):
return sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1
output = property(get_output)
class NeuronNetwork:
neurons = []
def __init__(self, inputs, outputs, rows, columns):
self.neurons = []
for row in xrange(rows + 2):
self.neurons.append([])
if row == 0:
for input_ in xrange(inputs):
self.neurons[row].append(Neuron(parents=[]))
elif row == rows + 1:
for output in xrange(outputs):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
else:
for column in xrange(columns):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
## Instruction:
Store output instead of calculating it each time
## Code After:
from random import random
class Neuron:
output = None
def __init__(self, parents=[]):
self.parents = parents
self.weights = [random() for parent in parents]
def calculate(self):
self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1
class NeuronNetwork:
neurons = []
def __init__(self, inputs, outputs, rows, columns):
self.neurons = []
for row in xrange(rows + 2):
self.neurons.append([])
if row == 0:
for input_ in xrange(inputs):
self.neurons[row].append(Neuron(parents=[]))
elif row == rows + 1:
for output in xrange(outputs):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
else:
for column in xrange(columns):
self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
|
...
class Neuron:
output = None
def __init__(self, parents=[]):
...
def calculate(self):
self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1
...
|
222a87ef324f66baf8113020b41d336c459ab847
|
stdnum/fi/__init__.py
|
stdnum/fi/__init__.py
|
"""Collection of Finnish numbers."""
# provide vat as an alias
from stdnum.fi import alv as vat
from stdnum.fi import ytunnus as businessid
|
"""Collection of Finnish numbers."""
# provide vat as an alias
from stdnum.fi import alv as vat
from stdnum.fi import ytunnus as businessid
from stdnum.fi import hetu as personalid
|
Add alias to hetu in for finnish personal id code
|
Add alias to hetu in for finnish personal id code
|
Python
|
lgpl-2.1
|
holvi/python-stdnum,holvi/python-stdnum,arthurdejong/python-stdnum,arthurdejong/python-stdnum,holvi/python-stdnum,arthurdejong/python-stdnum
|
"""Collection of Finnish numbers."""
# provide vat as an alias
from stdnum.fi import alv as vat
from stdnum.fi import ytunnus as businessid
+ from stdnum.fi import hetu as personalid
|
Add alias to hetu in for finnish personal id code
|
## Code Before:
"""Collection of Finnish numbers."""
# provide vat as an alias
from stdnum.fi import alv as vat
from stdnum.fi import ytunnus as businessid
## Instruction:
Add alias to hetu in for finnish personal id code
## Code After:
"""Collection of Finnish numbers."""
# provide vat as an alias
from stdnum.fi import alv as vat
from stdnum.fi import ytunnus as businessid
from stdnum.fi import hetu as personalid
|
# ... existing code ...
from stdnum.fi import ytunnus as businessid
from stdnum.fi import hetu as personalid
# ... rest of the code ...
|
2f2cef54a98e2328a638d9bbdfd2e0312606d906
|
plugins/GCodeWriter/__init__.py
|
plugins/GCodeWriter/__init__.py
|
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File")
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
Add mime types to GCodeWriter plugin
|
Add mime types to GCodeWriter plugin
|
Python
|
agpl-3.0
|
Curahelper/Cura,senttech/Cura,fieldOfView/Cura,hmflash/Cura,lo0ol/Ultimaker-Cura,Curahelper/Cura,markwal/Cura,ad1217/Cura,senttech/Cura,ad1217/Cura,lo0ol/Ultimaker-Cura,totalretribution/Cura,hmflash/Cura,bq/Ultimaker-Cura,ynotstartups/Wanhao,fieldOfView/Cura,totalretribution/Cura,fxtentacle/Cura,fxtentacle/Cura,ynotstartups/Wanhao,markwal/Cura,bq/Ultimaker-Cura
|
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
- "description": catalog.i18nc("GCode Writer File Description", "GCode File")
+ "description": catalog.i18nc("GCode Writer File Description", "GCode File"),
+ "mime_types": [
+ "text/x-gcode"
+ ]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
Add mime types to GCodeWriter plugin
|
## Code Before:
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File")
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
## Instruction:
Add mime types to GCodeWriter plugin
## Code After:
from . import GCodeWriter
from UM.i18n import i18nCatalog
catalog = i18nCatalog("cura")
def getMetaData():
return {
"type": "mesh_writer",
"plugin": {
"name": "GCode Writer",
"author": "Ultimaker",
"version": "1.0",
"description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file")
},
"mesh_writer": {
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
}
def register(app):
return { "mesh_writer": GCodeWriter.GCodeWriter() }
|
# ... existing code ...
"extension": "gcode",
"description": catalog.i18nc("GCode Writer File Description", "GCode File"),
"mime_types": [
"text/x-gcode"
]
}
# ... rest of the code ...
|
8f8b313a1b5118b6528e5152252128e075de0401
|
tests/test_terrain.py
|
tests/test_terrain.py
|
import unittest
from randterrainpy import *
class TerrainTesterPy(unittest.TestCase):
def setUp(self):
pass
|
import unittest
from randterrainpy import *
class TerrainTesterPy(unittest.TestCase):
def setUp(self):
self.ter1 = Terrain(1, 1)
self.ter2 = Terrain(2, 4)
self.ter3 = Terrain(1, 1)
def test_getitem(self):
self.assertEqual(self.ter1[0, 0], 0)
self.assertEqual(self.ter2[1, 2], 0)
def test_eq(self):
self.assertEqual(self.ter1, self.ter3)
self.assertNotEqual(self.ter1, self.ter2)
def test_setitem(self):
self.ter1[0, 0] = 1
self.assertEqual(self.ter1[0, 0], 1)
self.ter2[1, 2] = 0.5
self.assertEqual(self.ter2[1, 2], 0.5)
def test_add(self):
self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2)
self.assertEqual(self.ter1+self.ter3, self.ter1)
if __name__ == "__main__":
unittest.main()
|
Add tests for indexing, equality and addition for Terrain
|
Add tests for indexing, equality and addition for Terrain
|
Python
|
mit
|
jackromo/RandTerrainPy
|
import unittest
from randterrainpy import *
class TerrainTesterPy(unittest.TestCase):
def setUp(self):
- pass
+ self.ter1 = Terrain(1, 1)
+ self.ter2 = Terrain(2, 4)
+ self.ter3 = Terrain(1, 1)
+ def test_getitem(self):
+ self.assertEqual(self.ter1[0, 0], 0)
+ self.assertEqual(self.ter2[1, 2], 0)
+
+ def test_eq(self):
+ self.assertEqual(self.ter1, self.ter3)
+ self.assertNotEqual(self.ter1, self.ter2)
+
+ def test_setitem(self):
+ self.ter1[0, 0] = 1
+ self.assertEqual(self.ter1[0, 0], 1)
+ self.ter2[1, 2] = 0.5
+ self.assertEqual(self.ter2[1, 2], 0.5)
+
+ def test_add(self):
+ self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2)
+ self.assertEqual(self.ter1+self.ter3, self.ter1)
+
+
+ if __name__ == "__main__":
+ unittest.main()
+
|
Add tests for indexing, equality and addition for Terrain
|
## Code Before:
import unittest
from randterrainpy import *
class TerrainTesterPy(unittest.TestCase):
def setUp(self):
pass
## Instruction:
Add tests for indexing, equality and addition for Terrain
## Code After:
import unittest
from randterrainpy import *
class TerrainTesterPy(unittest.TestCase):
def setUp(self):
self.ter1 = Terrain(1, 1)
self.ter2 = Terrain(2, 4)
self.ter3 = Terrain(1, 1)
def test_getitem(self):
self.assertEqual(self.ter1[0, 0], 0)
self.assertEqual(self.ter2[1, 2], 0)
def test_eq(self):
self.assertEqual(self.ter1, self.ter3)
self.assertNotEqual(self.ter1, self.ter2)
def test_setitem(self):
self.ter1[0, 0] = 1
self.assertEqual(self.ter1[0, 0], 1)
self.ter2[1, 2] = 0.5
self.assertEqual(self.ter2[1, 2], 0.5)
def test_add(self):
self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2)
self.assertEqual(self.ter1+self.ter3, self.ter1)
if __name__ == "__main__":
unittest.main()
|
# ... existing code ...
def setUp(self):
self.ter1 = Terrain(1, 1)
self.ter2 = Terrain(2, 4)
self.ter3 = Terrain(1, 1)
def test_getitem(self):
self.assertEqual(self.ter1[0, 0], 0)
self.assertEqual(self.ter2[1, 2], 0)
def test_eq(self):
self.assertEqual(self.ter1, self.ter3)
self.assertNotEqual(self.ter1, self.ter2)
def test_setitem(self):
self.ter1[0, 0] = 1
self.assertEqual(self.ter1[0, 0], 1)
self.ter2[1, 2] = 0.5
self.assertEqual(self.ter2[1, 2], 0.5)
def test_add(self):
self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2)
self.assertEqual(self.ter1+self.ter3, self.ter1)
if __name__ == "__main__":
unittest.main()
# ... rest of the code ...
|
f849961e75dc956d669813fddb5b13627b224e1e
|
pyang/plugins/name.py
|
pyang/plugins/name.py
|
import optparse
from pyang import plugin
def pyang_plugin_init():
plugin.register_plugin(NamePlugin())
class NamePlugin(plugin.PyangPlugin):
def add_output_format(self, fmts):
self.multiple_modules = True
fmts['name'] = self
def add_opts(self, optparser):
optlist = [
optparse.make_option("--name-print-revision",
dest="print_revision",
action="store_true",
help="Print the name and revision in name@revision format"),
]
g = optparser.add_option_group("Name output specific options")
g.add_options(optlist)
def setup_fmt(self, ctx):
ctx.implicit_errors = False
def emit(self, ctx, modules, fd):
emit_name(ctx, modules, fd)
def emit_name(ctx, modules, fd):
for module in modules:
bstr = ""
rstr = ""
if ctx.opts.print_revision:
r = module.search_one('revision')
if r is not None:
rstr = '@%s' % r.arg
b = module.search_one('belongs-to')
if b is not None:
bstr = " (belongs-to %s)" % b.arg
fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
|
import optparse
from pyang import plugin
def pyang_plugin_init():
plugin.register_plugin(NamePlugin())
class NamePlugin(plugin.PyangPlugin):
def add_output_format(self, fmts):
self.multiple_modules = True
fmts['name'] = self
def add_opts(self, optparser):
optlist = [
optparse.make_option("--name-print-revision",
dest="print_revision",
action="store_true",
help="Print the name and revision in name@revision format"),
]
g = optparser.add_option_group("Name output specific options")
g.add_options(optlist)
def setup_fmt(self, ctx):
ctx.implicit_errors = False
def emit(self, ctx, modules, fd):
emit_name(ctx, modules, fd)
def emit_name(ctx, modules, fd):
for module in modules:
bstr = ""
rstr = ""
if ctx.opts.print_revision:
rs = module.i_latest_revision
if rs is None:
r = module.search_one('revision')
if r is not None:
rs = r.arg
if rs is not None:
rstr = '@%s' % rs
b = module.search_one('belongs-to')
if b is not None:
bstr = " (belongs-to %s)" % b.arg
fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
|
Use i_latest_revision to ensure we get the latest revision.
|
Use i_latest_revision to ensure we get the latest revision.
|
Python
|
isc
|
mbj4668/pyang,mbj4668/pyang
|
import optparse
from pyang import plugin
def pyang_plugin_init():
plugin.register_plugin(NamePlugin())
class NamePlugin(plugin.PyangPlugin):
def add_output_format(self, fmts):
self.multiple_modules = True
fmts['name'] = self
def add_opts(self, optparser):
optlist = [
optparse.make_option("--name-print-revision",
dest="print_revision",
action="store_true",
help="Print the name and revision in name@revision format"),
]
g = optparser.add_option_group("Name output specific options")
g.add_options(optlist)
def setup_fmt(self, ctx):
ctx.implicit_errors = False
def emit(self, ctx, modules, fd):
emit_name(ctx, modules, fd)
def emit_name(ctx, modules, fd):
for module in modules:
bstr = ""
rstr = ""
if ctx.opts.print_revision:
+ rs = module.i_latest_revision
+ if rs is None:
- r = module.search_one('revision')
+ r = module.search_one('revision')
+ if r is not None:
+ rs = r.arg
- if r is not None:
+ if rs is not None:
- rstr = '@%s' % r.arg
+ rstr = '@%s' % rs
b = module.search_one('belongs-to')
if b is not None:
bstr = " (belongs-to %s)" % b.arg
fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
|
Use i_latest_revision to ensure we get the latest revision.
|
## Code Before:
import optparse
from pyang import plugin
def pyang_plugin_init():
plugin.register_plugin(NamePlugin())
class NamePlugin(plugin.PyangPlugin):
def add_output_format(self, fmts):
self.multiple_modules = True
fmts['name'] = self
def add_opts(self, optparser):
optlist = [
optparse.make_option("--name-print-revision",
dest="print_revision",
action="store_true",
help="Print the name and revision in name@revision format"),
]
g = optparser.add_option_group("Name output specific options")
g.add_options(optlist)
def setup_fmt(self, ctx):
ctx.implicit_errors = False
def emit(self, ctx, modules, fd):
emit_name(ctx, modules, fd)
def emit_name(ctx, modules, fd):
for module in modules:
bstr = ""
rstr = ""
if ctx.opts.print_revision:
r = module.search_one('revision')
if r is not None:
rstr = '@%s' % r.arg
b = module.search_one('belongs-to')
if b is not None:
bstr = " (belongs-to %s)" % b.arg
fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
## Instruction:
Use i_latest_revision to ensure we get the latest revision.
## Code After:
import optparse
from pyang import plugin
def pyang_plugin_init():
plugin.register_plugin(NamePlugin())
class NamePlugin(plugin.PyangPlugin):
def add_output_format(self, fmts):
self.multiple_modules = True
fmts['name'] = self
def add_opts(self, optparser):
optlist = [
optparse.make_option("--name-print-revision",
dest="print_revision",
action="store_true",
help="Print the name and revision in name@revision format"),
]
g = optparser.add_option_group("Name output specific options")
g.add_options(optlist)
def setup_fmt(self, ctx):
ctx.implicit_errors = False
def emit(self, ctx, modules, fd):
emit_name(ctx, modules, fd)
def emit_name(ctx, modules, fd):
for module in modules:
bstr = ""
rstr = ""
if ctx.opts.print_revision:
rs = module.i_latest_revision
if rs is None:
r = module.search_one('revision')
if r is not None:
rs = r.arg
if rs is not None:
rstr = '@%s' % rs
b = module.search_one('belongs-to')
if b is not None:
bstr = " (belongs-to %s)" % b.arg
fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
|
// ... existing code ...
if ctx.opts.print_revision:
rs = module.i_latest_revision
if rs is None:
r = module.search_one('revision')
if r is not None:
rs = r.arg
if rs is not None:
rstr = '@%s' % rs
b = module.search_one('belongs-to')
// ... rest of the code ...
|
695b3f628b56cd1dbe050f07692559ff3685290c
|
templatefinder/__init__.py
|
templatefinder/__init__.py
|
from __future__ import absolute_import
from .utils import *
VERSION = (0, 5,)
|
from __future__ import absolute_import
# this is required for setup.py to work
try:
from .utils import *
except ImportError:
pass
VERSION = (0, 5, 1,)
|
Fix setup.py for installs without Django
|
Fix setup.py for installs without Django
|
Python
|
bsd-2-clause
|
TyMaszWeb/django-template-finder
|
from __future__ import absolute_import
+ # this is required for setup.py to work
+ try:
- from .utils import *
+ from .utils import *
+ except ImportError:
+ pass
- VERSION = (0, 5,)
+ VERSION = (0, 5, 1,)
|
Fix setup.py for installs without Django
|
## Code Before:
from __future__ import absolute_import
from .utils import *
VERSION = (0, 5,)
## Instruction:
Fix setup.py for installs without Django
## Code After:
from __future__ import absolute_import
# this is required for setup.py to work
try:
from .utils import *
except ImportError:
pass
VERSION = (0, 5, 1,)
|
// ... existing code ...
# this is required for setup.py to work
try:
from .utils import *
except ImportError:
pass
// ... modified code ...
VERSION = (0, 5, 1,)
// ... rest of the code ...
|
4de03c57bf4f4995eb8c8859e0a40b7c5fc9942b
|
desktop/libs/libzookeeper/src/libzookeeper/models.py
|
desktop/libs/libzookeeper/src/libzookeeper/models.py
|
from kazoo.client import KazooClient
from libzookeeper.conf import PRINCIPAL_NAME
def get_children_data(ensemble, namespace, read_only=True):
zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=PRINCIPAL_NAME.get())
zk.start()
children_data = []
children = zk.get_children(namespace)
for node in children:
data, stat = zk.get("%s/%s" % (namespace, node))
children_data.append(data)
zk.stop()
return children_data
|
from kazoo.client import KazooClient
from hadoop import cluster
from desktop.lib.exceptions_renderable import PopupException
from libzookeeper.conf import PRINCIPAL_NAME
def get_children_data(ensemble, namespace, read_only=True):
hdfs = cluster.get_hdfs()
if hdfs is None:
raise PopupException(_('No [hdfs] configured in hue.ini.'))
if hdfs.security_enabled:
sasl_server_principal = PRINCIPAL_NAME.get()
else:
sasl_server_principal = None
zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal)
zk.start()
children_data = []
children = zk.get_children(namespace)
for node in children:
data, stat = zk.get("%s/%s" % (namespace, node))
children_data.append(data)
zk.stop()
return children_data
|
Enable Kerberos automatically based on HDFS security
|
[libzookeeper] Enable Kerberos automatically based on HDFS security
We don't need another property that way and Kerberos is a all or nothing
setup.
Even if HDFS is not used in Hue, the default hue.ini has security set
to false.
|
Python
|
apache-2.0
|
pratikmallya/hue,jjmleiro/hue,lumig242/Hue-Integration-with-CDAP,cloudera/hue,pratikmallya/hue,xiangel/hue,Peddle/hue,x303597316/hue,cloudera/hue,rahul67/hue,kawamon/hue,yongshengwang/hue,MobinRanjbar/hue,x303597316/hue,xq262144/hue,jayceyxc/hue,mapr/hue,yongshengwang/hue,pratikmallya/hue,sanjeevtripurari/hue,lumig242/Hue-Integration-with-CDAP,jayceyxc/hue,cloudera/hue,jounex/hue,Peddle/hue,ChenJunor/hue,jayceyxc/hue,kawamon/hue,Peddle/hue,vmax-feihu/hue,MobinRanjbar/hue,vmax-feihu/hue,ahmed-mahran/hue,ahmed-mahran/hue,kawamon/hue,kawamon/hue,GitHublong/hue,hdinsight/hue,Peddle/hue,kawamon/hue,xiangel/hue,yoer/hue,pratikmallya/hue,yongshengwang/hue,kawamon/hue,cloudera/hue,jounex/hue,cloudera/hue,rahul67/hue,javachengwc/hue,azureplus/hue,kawamon/hue,vmax-feihu/hue,fangxingli/hue,hdinsight/hue,jounex/hue,cloudera/hue,sanjeevtripurari/hue,rahul67/hue,lumig242/Hue-Integration-with-CDAP,jjmleiro/hue,kawamon/hue,cloudera/hue,javachengwc/hue,cloudera/hue,yongshengwang/hue,jjmleiro/hue,mapr/hue,yoer/hue,yongshengwang/hue,todaychi/hue,kawamon/hue,GitHublong/hue,xq262144/hue,jounex/hue,kawamon/hue,ChenJunor/hue,cloudera/hue,x303597316/hue,jayceyxc/hue,todaychi/hue,jounex/hue,GitHublong/hue,hdinsight/hue,yoer/hue,hdinsight/hue,fangxingli/hue,kawamon/hue,lumig242/Hue-Integration-with-CDAP,x303597316/hue,ahmed-mahran/hue,azureplus/hue,xiangel/hue,todaychi/hue,jjmleiro/hue,jounex/hue,cloudera/hue,fangxingli/hue,vmax-feihu/hue,yongshengwang/hue,hdinsight/hue,kawamon/hue,ChenJunor/hue,todaychi/hue,lumig242/Hue-Integration-with-CDAP,rahul67/hue,javachengwc/hue,todaychi/hue,todaychi/hue,rahul67/hue,kawamon/hue,GitHublong/hue,yoer/hue,fangxingli/hue,javachengwc/hue,todaychi/hue,cloudera/hue,cloudera/hue,ChenJunor/hue,jayceyxc/hue,jayceyxc/hue,Peddle/hue,sanjeevtripurari/hue,lumig242/Hue-Integration-with-CDAP,xq262144/hue,pratikmallya/hue,GitHublong/hue,x303597316/hue,jayceyxc/hue,ahmed-mahran/hue,MobinRanjbar/hue,MobinRanjbar/hue,vmax-feihu/hue,fangxingli/hue,mapr/hue,xiangel/hue,kawamon/hue,lumig242/Hue-Integration-with-CDAP,mapr/hue,mapr/hue,fangxingli/hue,ChenJunor/hue,jayceyxc/hue,jjmleiro/hue,GitHublong/hue,todaychi/hue,pratikmallya/hue,cloudera/hue,ahmed-mahran/hue,cloudera/hue,xq262144/hue,rahul67/hue,jayceyxc/hue,azureplus/hue,jjmleiro/hue,MobinRanjbar/hue,javachengwc/hue,azureplus/hue,mapr/hue,jounex/hue,xiangel/hue,vmax-feihu/hue,hdinsight/hue,sanjeevtripurari/hue,ahmed-mahran/hue,yoer/hue,ahmed-mahran/hue,pratikmallya/hue,hdinsight/hue,cloudera/hue,ChenJunor/hue,xiangel/hue,kawamon/hue,xq262144/hue,yongshengwang/hue,jounex/hue,hdinsight/hue,MobinRanjbar/hue,rahul67/hue,Peddle/hue,kawamon/hue,xq262144/hue,azureplus/hue,lumig242/Hue-Integration-with-CDAP,vmax-feihu/hue,Peddle/hue,xiangel/hue,sanjeevtripurari/hue,sanjeevtripurari/hue,x303597316/hue,MobinRanjbar/hue,javachengwc/hue,xq262144/hue,fangxingli/hue,x303597316/hue,mapr/hue,xq262144/hue,pratikmallya/hue,sanjeevtripurari/hue,javachengwc/hue,sanjeevtripurari/hue,yoer/hue,rahul67/hue,kawamon/hue,azureplus/hue,jjmleiro/hue,Peddle/hue,ChenJunor/hue,MobinRanjbar/hue,cloudera/hue,yongshengwang/hue,azureplus/hue,yoer/hue,ChenJunor/hue,cloudera/hue,vmax-feihu/hue,cloudera/hue,azureplus/hue,yoer/hue,jjmleiro/hue,javachengwc/hue,GitHublong/hue,Peddle/hue,ahmed-mahran/hue,todaychi/hue,x303597316/hue,jjmleiro/hue,xiangel/hue,lumig242/Hue-Integration-with-CDAP,kawamon/hue,fangxingli/hue,GitHublong/hue,xq262144/hue
|
from kazoo.client import KazooClient
+
+ from hadoop import cluster
+ from desktop.lib.exceptions_renderable import PopupException
+
from libzookeeper.conf import PRINCIPAL_NAME
def get_children_data(ensemble, namespace, read_only=True):
+ hdfs = cluster.get_hdfs()
+ if hdfs is None:
+ raise PopupException(_('No [hdfs] configured in hue.ini.'))
+
+ if hdfs.security_enabled:
+ sasl_server_principal = PRINCIPAL_NAME.get()
+ else:
+ sasl_server_principal = None
+
- zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=PRINCIPAL_NAME.get())
+ zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal)
zk.start()
children_data = []
children = zk.get_children(namespace)
for node in children:
data, stat = zk.get("%s/%s" % (namespace, node))
children_data.append(data)
zk.stop()
-
+
return children_data
|
Enable Kerberos automatically based on HDFS security
|
## Code Before:
from kazoo.client import KazooClient
from libzookeeper.conf import PRINCIPAL_NAME
def get_children_data(ensemble, namespace, read_only=True):
zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=PRINCIPAL_NAME.get())
zk.start()
children_data = []
children = zk.get_children(namespace)
for node in children:
data, stat = zk.get("%s/%s" % (namespace, node))
children_data.append(data)
zk.stop()
return children_data
## Instruction:
Enable Kerberos automatically based on HDFS security
## Code After:
from kazoo.client import KazooClient
from hadoop import cluster
from desktop.lib.exceptions_renderable import PopupException
from libzookeeper.conf import PRINCIPAL_NAME
def get_children_data(ensemble, namespace, read_only=True):
hdfs = cluster.get_hdfs()
if hdfs is None:
raise PopupException(_('No [hdfs] configured in hue.ini.'))
if hdfs.security_enabled:
sasl_server_principal = PRINCIPAL_NAME.get()
else:
sasl_server_principal = None
zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal)
zk.start()
children_data = []
children = zk.get_children(namespace)
for node in children:
data, stat = zk.get("%s/%s" % (namespace, node))
children_data.append(data)
zk.stop()
return children_data
|
// ... existing code ...
from kazoo.client import KazooClient
from hadoop import cluster
from desktop.lib.exceptions_renderable import PopupException
from libzookeeper.conf import PRINCIPAL_NAME
// ... modified code ...
def get_children_data(ensemble, namespace, read_only=True):
hdfs = cluster.get_hdfs()
if hdfs is None:
raise PopupException(_('No [hdfs] configured in hue.ini.'))
if hdfs.security_enabled:
sasl_server_principal = PRINCIPAL_NAME.get()
else:
sasl_server_principal = None
zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal)
...
zk.stop()
return children_data
// ... rest of the code ...
|
382b8df7a25732ee8384c02d776472a93c18a0ea
|
vcspull/__about__.py
|
vcspull/__about__.py
|
__title__ = 'vcspull'
__package_name__ = 'vcspull'
__description__ = 'synchronize your repos'
__version__ = '1.2.0'
__author__ = 'Tony Narlock'
__github__ = 'https://github.com/vcs-python/vcspull'
__pypi__ = 'https://pypi.org/project/vcspull/'
__email__ = '[email protected]'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013-2018 Tony Narlock'
|
__title__ = 'vcspull'
__package_name__ = 'vcspull'
__description__ = 'synchronize your repos'
__version__ = '1.2.0'
__author__ = 'Tony Narlock'
__github__ = 'https://github.com/vcs-python/vcspull'
__docs__ = 'https://vcspull.git-pull.com'
__tracker__ = 'https://github.com/vcs-python/vcspull/issues'
__pypi__ = 'https://pypi.org/project/vcspull/'
__email__ = '[email protected]'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013-Tony Narlock'
|
Add docs / tracker to metadata
|
Add docs / tracker to metadata
|
Python
|
mit
|
tony/vcspull,tony/vcspull
|
__title__ = 'vcspull'
__package_name__ = 'vcspull'
__description__ = 'synchronize your repos'
__version__ = '1.2.0'
__author__ = 'Tony Narlock'
__github__ = 'https://github.com/vcs-python/vcspull'
+ __docs__ = 'https://vcspull.git-pull.com'
+ __tracker__ = 'https://github.com/vcs-python/vcspull/issues'
__pypi__ = 'https://pypi.org/project/vcspull/'
__email__ = '[email protected]'
__license__ = 'MIT'
- __copyright__ = 'Copyright 2013-2018 Tony Narlock'
+ __copyright__ = 'Copyright 2013-Tony Narlock'
|
Add docs / tracker to metadata
|
## Code Before:
__title__ = 'vcspull'
__package_name__ = 'vcspull'
__description__ = 'synchronize your repos'
__version__ = '1.2.0'
__author__ = 'Tony Narlock'
__github__ = 'https://github.com/vcs-python/vcspull'
__pypi__ = 'https://pypi.org/project/vcspull/'
__email__ = '[email protected]'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013-2018 Tony Narlock'
## Instruction:
Add docs / tracker to metadata
## Code After:
__title__ = 'vcspull'
__package_name__ = 'vcspull'
__description__ = 'synchronize your repos'
__version__ = '1.2.0'
__author__ = 'Tony Narlock'
__github__ = 'https://github.com/vcs-python/vcspull'
__docs__ = 'https://vcspull.git-pull.com'
__tracker__ = 'https://github.com/vcs-python/vcspull/issues'
__pypi__ = 'https://pypi.org/project/vcspull/'
__email__ = '[email protected]'
__license__ = 'MIT'
__copyright__ = 'Copyright 2013-Tony Narlock'
|
# ... existing code ...
__github__ = 'https://github.com/vcs-python/vcspull'
__docs__ = 'https://vcspull.git-pull.com'
__tracker__ = 'https://github.com/vcs-python/vcspull/issues'
__pypi__ = 'https://pypi.org/project/vcspull/'
# ... modified code ...
__license__ = 'MIT'
__copyright__ = 'Copyright 2013-Tony Narlock'
# ... rest of the code ...
|
27ab3ad3d1ce869baec85264b840da49ff43f82f
|
scripts/sync_exceeded_traffic_limits.py
|
scripts/sync_exceeded_traffic_limits.py
|
import os
from flask import _request_ctx_stack, g, request
from sqlalchemy import create_engine
from sqlalchemy.orm import scoped_session, sessionmaker
from pycroft.model import session
from pycroft.model.session import set_scoped_session
from scripts.schema import AlembicHelper, SchemaStrategist
from pycroft.lib import traffic
def main():
try:
connection_string = os.environ['PYCROFT_DB_URI']
except KeyError:
raise RuntimeError("Environment variable PYCROFT_DB_URI must be "
"set to an SQLAlchemy connection string.")
engine = create_engine(connection_string)
connection = engine.connect()
state = AlembicHelper(connection)
strategy = SchemaStrategist(state).determine_schema_strategy()
strategy()
engine = create_engine(connection_string)
set_scoped_session(scoped_session(sessionmaker(bind=engine),
scopefunc=lambda: _request_ctx_stack.top))
print("Starting synchronization of exceeded traffic limits.")
traffic.sync_exceeded_traffic_limits()
session.session.commit()
print("Finished synchronization.")
if __name__ == "__main__":
main()
|
import os
from flask import _request_ctx_stack, g, request
from sqlalchemy import create_engine
from sqlalchemy.orm import scoped_session, sessionmaker
from pycroft.model import session
from pycroft.model.session import set_scoped_session
from scripts.schema import AlembicHelper, SchemaStrategist
from pycroft.lib import traffic
def main():
try:
connection_string = os.environ['PYCROFT_DB_URI']
except KeyError:
raise RuntimeError("Environment variable PYCROFT_DB_URI must be "
"set to an SQLAlchemy connection string.")
engine = create_engine(connection_string)
connection = engine.connect()
state = AlembicHelper(connection)
strategist = SchemaStrategist(state)
is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version
if not is_up_to_date:
print("Schema is not up to date!")
return
set_scoped_session(scoped_session(sessionmaker(bind=engine),
scopefunc=lambda: _request_ctx_stack.top))
print("Starting synchronization of exceeded traffic limits.")
traffic.sync_exceeded_traffic_limits()
session.session.commit()
print("Finished synchronization.")
if __name__ == "__main__":
main()
|
Add schema version check to sync script
|
Add schema version check to sync script
|
Python
|
apache-2.0
|
agdsn/pycroft,agdsn/pycroft,agdsn/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,lukasjuhrich/pycroft,lukasjuhrich/pycroft,lukasjuhrich/pycroft,agdsn/pycroft
|
import os
from flask import _request_ctx_stack, g, request
from sqlalchemy import create_engine
from sqlalchemy.orm import scoped_session, sessionmaker
from pycroft.model import session
from pycroft.model.session import set_scoped_session
from scripts.schema import AlembicHelper, SchemaStrategist
from pycroft.lib import traffic
def main():
try:
connection_string = os.environ['PYCROFT_DB_URI']
except KeyError:
raise RuntimeError("Environment variable PYCROFT_DB_URI must be "
"set to an SQLAlchemy connection string.")
engine = create_engine(connection_string)
connection = engine.connect()
state = AlembicHelper(connection)
- strategy = SchemaStrategist(state).determine_schema_strategy()
- strategy()
+ strategist = SchemaStrategist(state)
+ is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version
+ if not is_up_to_date:
+ print("Schema is not up to date!")
+ return
- engine = create_engine(connection_string)
set_scoped_session(scoped_session(sessionmaker(bind=engine),
scopefunc=lambda: _request_ctx_stack.top))
print("Starting synchronization of exceeded traffic limits.")
traffic.sync_exceeded_traffic_limits()
session.session.commit()
print("Finished synchronization.")
if __name__ == "__main__":
main()
|
Add schema version check to sync script
|
## Code Before:
import os
from flask import _request_ctx_stack, g, request
from sqlalchemy import create_engine
from sqlalchemy.orm import scoped_session, sessionmaker
from pycroft.model import session
from pycroft.model.session import set_scoped_session
from scripts.schema import AlembicHelper, SchemaStrategist
from pycroft.lib import traffic
def main():
try:
connection_string = os.environ['PYCROFT_DB_URI']
except KeyError:
raise RuntimeError("Environment variable PYCROFT_DB_URI must be "
"set to an SQLAlchemy connection string.")
engine = create_engine(connection_string)
connection = engine.connect()
state = AlembicHelper(connection)
strategy = SchemaStrategist(state).determine_schema_strategy()
strategy()
engine = create_engine(connection_string)
set_scoped_session(scoped_session(sessionmaker(bind=engine),
scopefunc=lambda: _request_ctx_stack.top))
print("Starting synchronization of exceeded traffic limits.")
traffic.sync_exceeded_traffic_limits()
session.session.commit()
print("Finished synchronization.")
if __name__ == "__main__":
main()
## Instruction:
Add schema version check to sync script
## Code After:
import os
from flask import _request_ctx_stack, g, request
from sqlalchemy import create_engine
from sqlalchemy.orm import scoped_session, sessionmaker
from pycroft.model import session
from pycroft.model.session import set_scoped_session
from scripts.schema import AlembicHelper, SchemaStrategist
from pycroft.lib import traffic
def main():
try:
connection_string = os.environ['PYCROFT_DB_URI']
except KeyError:
raise RuntimeError("Environment variable PYCROFT_DB_URI must be "
"set to an SQLAlchemy connection string.")
engine = create_engine(connection_string)
connection = engine.connect()
state = AlembicHelper(connection)
strategist = SchemaStrategist(state)
is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version
if not is_up_to_date:
print("Schema is not up to date!")
return
set_scoped_session(scoped_session(sessionmaker(bind=engine),
scopefunc=lambda: _request_ctx_stack.top))
print("Starting synchronization of exceeded traffic limits.")
traffic.sync_exceeded_traffic_limits()
session.session.commit()
print("Finished synchronization.")
if __name__ == "__main__":
main()
|
// ... existing code ...
state = AlembicHelper(connection)
strategist = SchemaStrategist(state)
is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version
if not is_up_to_date:
print("Schema is not up to date!")
return
set_scoped_session(scoped_session(sessionmaker(bind=engine),
// ... rest of the code ...
|
4583c9949143e58bf400fc86e27d634aa382f605
|
tests/test_expanded.py
|
tests/test_expanded.py
|
from mycli.packages.expanded import expanded_table
def test_expanded_table_renders():
input = [("hello", 123), ("world", 456)]
expected = """-[ RECORD 0 ]
name | hello
age | 123
-[ RECORD 1 ]
name | world
age | 456
"""
assert expected == expanded_table(input, ["name", "age"])
|
from mycli.packages.expanded import expanded_table
def test_expanded_table_renders():
input = [("hello", 123), ("world", 456)]
expected = """***************************[ 1. row ]***************************
name | hello
age | 123
***************************[ 2. row ]***************************
name | world
age | 456
"""
assert expected == expanded_table(input, ["name", "age"])
|
Update expanded tests to match mysql style.
|
Update expanded tests to match mysql style.
|
Python
|
bsd-3-clause
|
oguzy/mycli,chenpingzhao/mycli,ZuoGuocai/mycli,evook/mycli,jinstrive/mycli,j-bennet/mycli,danieljwest/mycli,suzukaze/mycli,thanatoskira/mycli,chenpingzhao/mycli,j-bennet/mycli,brewneaux/mycli,webwlsong/mycli,MnO2/rediscli,brewneaux/mycli,shoma/mycli,mdsrosa/mycli,oguzy/mycli,danieljwest/mycli,jinstrive/mycli,thanatoskira/mycli,ZuoGuocai/mycli,webwlsong/mycli,evook/mycli,suzukaze/mycli,D-e-e-m-o/mycli,MnO2/rediscli,shoma/mycli,martijnengler/mycli,mdsrosa/mycli,martijnengler/mycli,D-e-e-m-o/mycli
|
from mycli.packages.expanded import expanded_table
def test_expanded_table_renders():
input = [("hello", 123), ("world", 456)]
- expected = """-[ RECORD 0 ]
+ expected = """***************************[ 1. row ]***************************
name | hello
age | 123
- -[ RECORD 1 ]
+ ***************************[ 2. row ]***************************
name | world
age | 456
"""
assert expected == expanded_table(input, ["name", "age"])
|
Update expanded tests to match mysql style.
|
## Code Before:
from mycli.packages.expanded import expanded_table
def test_expanded_table_renders():
input = [("hello", 123), ("world", 456)]
expected = """-[ RECORD 0 ]
name | hello
age | 123
-[ RECORD 1 ]
name | world
age | 456
"""
assert expected == expanded_table(input, ["name", "age"])
## Instruction:
Update expanded tests to match mysql style.
## Code After:
from mycli.packages.expanded import expanded_table
def test_expanded_table_renders():
input = [("hello", 123), ("world", 456)]
expected = """***************************[ 1. row ]***************************
name | hello
age | 123
***************************[ 2. row ]***************************
name | world
age | 456
"""
assert expected == expanded_table(input, ["name", "age"])
|
// ... existing code ...
expected = """***************************[ 1. row ]***************************
name | hello
// ... modified code ...
age | 123
***************************[ 2. row ]***************************
name | world
// ... rest of the code ...
|
6a379b806dd1992ad3dd2b728878ed35e8d0ea3c
|
cdf/utils.py
|
cdf/utils.py
|
def get_major_dot_minor_version(version):
"""
Convert full VERSION Django tuple to
a dotted string containing MAJOR.MINOR.
For example, (1, 9, 3, 'final', 0) will result in '1.9'
"""
return '.'.join(version.split('.')[:2])
|
def get_major_dot_minor_version(version):
"""
Convert full VERSION Django tuple to
a dotted string containing MAJOR.MINOR.
For example, (1, 9, 3, 'final', 0) will result in '1.9'
"""
return '.'.join([str(v) for v in version[:2]])
|
Fix getting major.minor django version
|
Fix getting major.minor django version
|
Python
|
mit
|
ana-balica/classy-django-forms,ana-balica/classy-django-forms,ana-balica/classy-django-forms
|
def get_major_dot_minor_version(version):
"""
Convert full VERSION Django tuple to
a dotted string containing MAJOR.MINOR.
For example, (1, 9, 3, 'final', 0) will result in '1.9'
"""
- return '.'.join(version.split('.')[:2])
+ return '.'.join([str(v) for v in version[:2]])
|
Fix getting major.minor django version
|
## Code Before:
def get_major_dot_minor_version(version):
"""
Convert full VERSION Django tuple to
a dotted string containing MAJOR.MINOR.
For example, (1, 9, 3, 'final', 0) will result in '1.9'
"""
return '.'.join(version.split('.')[:2])
## Instruction:
Fix getting major.minor django version
## Code After:
def get_major_dot_minor_version(version):
"""
Convert full VERSION Django tuple to
a dotted string containing MAJOR.MINOR.
For example, (1, 9, 3, 'final', 0) will result in '1.9'
"""
return '.'.join([str(v) for v in version[:2]])
|
...
"""
return '.'.join([str(v) for v in version[:2]])
...
|
07058595e43290524d28b53b5919fb76f16c618b
|
test/test_validators.py
|
test/test_validators.py
|
from unittest import TestCase
from win_unc import validators as V
class TestIsValidDriveLetter(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_drive_letter('A'))
self.assertTrue(V.is_valid_drive_letter('Z'))
self.assertTrue(V.is_valid_drive_letter('a'))
self.assertTrue(V.is_valid_drive_letter('z'))
def test_invalid(self):
self.assertFalse(V.is_valid_drive_letter(''))
self.assertFalse(V.is_valid_drive_letter(':'))
self.assertFalse(V.is_valid_drive_letter('aa'))
self.assertFalse(V.is_valid_drive_letter('a:'))
|
from unittest import TestCase
from win_unc import validators as V
class TestIsValidDriveLetter(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_drive_letter('A'))
self.assertTrue(V.is_valid_drive_letter('Z'))
self.assertTrue(V.is_valid_drive_letter('a'))
self.assertTrue(V.is_valid_drive_letter('z'))
def test_invalid(self):
self.assertFalse(V.is_valid_drive_letter(''))
self.assertFalse(V.is_valid_drive_letter(':'))
self.assertFalse(V.is_valid_drive_letter('aa'))
self.assertFalse(V.is_valid_drive_letter('a:'))
class TestIsValidUncPath(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_unc_path(r'\\a'))
self.assertTrue(V.is_valid_unc_path(r'\\a\b\c'))
self.assertTrue(V.is_valid_unc_path(r'\\ABC\\'))
def test_invalid(self):
self.assertFalse(V.is_valid_unc_path(''))
self.assertFalse(V.is_valid_unc_path(r'\\'))
self.assertFalse(V.is_valid_unc_path(r'\\\a'))
self.assertFalse(V.is_valid_unc_path(r'C:\path'))
self.assertFalse(V.is_valid_unc_path(r'\\<a>'))
|
Add tests for UNC path validator
|
Add tests for UNC path validator
|
Python
|
mit
|
CovenantEyes/py_win_unc,nithinphilips/py_win_unc
|
from unittest import TestCase
from win_unc import validators as V
class TestIsValidDriveLetter(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_drive_letter('A'))
self.assertTrue(V.is_valid_drive_letter('Z'))
self.assertTrue(V.is_valid_drive_letter('a'))
self.assertTrue(V.is_valid_drive_letter('z'))
def test_invalid(self):
self.assertFalse(V.is_valid_drive_letter(''))
self.assertFalse(V.is_valid_drive_letter(':'))
self.assertFalse(V.is_valid_drive_letter('aa'))
self.assertFalse(V.is_valid_drive_letter('a:'))
+
+ class TestIsValidUncPath(TestCase):
+ def test_valid(self):
+ self.assertTrue(V.is_valid_unc_path(r'\\a'))
+ self.assertTrue(V.is_valid_unc_path(r'\\a\b\c'))
+ self.assertTrue(V.is_valid_unc_path(r'\\ABC\\'))
+
+ def test_invalid(self):
+ self.assertFalse(V.is_valid_unc_path(''))
+ self.assertFalse(V.is_valid_unc_path(r'\\'))
+ self.assertFalse(V.is_valid_unc_path(r'\\\a'))
+ self.assertFalse(V.is_valid_unc_path(r'C:\path'))
+ self.assertFalse(V.is_valid_unc_path(r'\\<a>'))
+
|
Add tests for UNC path validator
|
## Code Before:
from unittest import TestCase
from win_unc import validators as V
class TestIsValidDriveLetter(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_drive_letter('A'))
self.assertTrue(V.is_valid_drive_letter('Z'))
self.assertTrue(V.is_valid_drive_letter('a'))
self.assertTrue(V.is_valid_drive_letter('z'))
def test_invalid(self):
self.assertFalse(V.is_valid_drive_letter(''))
self.assertFalse(V.is_valid_drive_letter(':'))
self.assertFalse(V.is_valid_drive_letter('aa'))
self.assertFalse(V.is_valid_drive_letter('a:'))
## Instruction:
Add tests for UNC path validator
## Code After:
from unittest import TestCase
from win_unc import validators as V
class TestIsValidDriveLetter(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_drive_letter('A'))
self.assertTrue(V.is_valid_drive_letter('Z'))
self.assertTrue(V.is_valid_drive_letter('a'))
self.assertTrue(V.is_valid_drive_letter('z'))
def test_invalid(self):
self.assertFalse(V.is_valid_drive_letter(''))
self.assertFalse(V.is_valid_drive_letter(':'))
self.assertFalse(V.is_valid_drive_letter('aa'))
self.assertFalse(V.is_valid_drive_letter('a:'))
class TestIsValidUncPath(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_unc_path(r'\\a'))
self.assertTrue(V.is_valid_unc_path(r'\\a\b\c'))
self.assertTrue(V.is_valid_unc_path(r'\\ABC\\'))
def test_invalid(self):
self.assertFalse(V.is_valid_unc_path(''))
self.assertFalse(V.is_valid_unc_path(r'\\'))
self.assertFalse(V.is_valid_unc_path(r'\\\a'))
self.assertFalse(V.is_valid_unc_path(r'C:\path'))
self.assertFalse(V.is_valid_unc_path(r'\\<a>'))
|
// ... existing code ...
self.assertFalse(V.is_valid_drive_letter('a:'))
class TestIsValidUncPath(TestCase):
def test_valid(self):
self.assertTrue(V.is_valid_unc_path(r'\\a'))
self.assertTrue(V.is_valid_unc_path(r'\\a\b\c'))
self.assertTrue(V.is_valid_unc_path(r'\\ABC\\'))
def test_invalid(self):
self.assertFalse(V.is_valid_unc_path(''))
self.assertFalse(V.is_valid_unc_path(r'\\'))
self.assertFalse(V.is_valid_unc_path(r'\\\a'))
self.assertFalse(V.is_valid_unc_path(r'C:\path'))
self.assertFalse(V.is_valid_unc_path(r'\\<a>'))
// ... rest of the code ...
|
6f557ed73372aa5823393a53b079bf4cec7511b8
|
docker/ssladapter/ssladapter.py
|
docker/ssladapter/ssladapter.py
|
from distutils.version import StrictVersion
from requests.adapters import HTTPAdapter
try:
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
import urllib3
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that uses an arbitrary SSL version.'''
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
urllib_ver = urllib3.__version__
if urllib3 and StrictVersion(urllib_ver) <= StrictVersion('1.5'):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block)
else:
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
|
from distutils.version import StrictVersion
from requests.adapters import HTTPAdapter
try:
import requests.packages.urllib3 as urllib3
except ImportError:
import urllib3
PoolManager = urllib3.poolmanager.PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that uses an arbitrary SSL version.'''
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
urllib_ver = urllib3.__version__.split('-')[0]
if urllib3 and urllib_ver != 'dev' and \
StrictVersion(urllib_ver) <= StrictVersion('1.5'):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block)
else:
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
|
Fix some urllib3 import issues
|
Fix some urllib3 import issues
|
Python
|
apache-2.0
|
runcom/docker-py,kaiyou/docker-py,TomasTomecek/docker-py,vpetersson/docker-py,dimaspivak/docker-py,paulbellamy/docker-py,ClusterHQ/docker-py,terminalmage/docker-py,ticosax/docker-py,kaiyou/docker-py,dimaspivak/docker-py,auready/docker-py,dnephin/docker-py,docker/docker-py,docker/docker-py,mrfuxi/docker-py,bboreham/docker-py,rhatdan/docker-py,mangalaman93/docker-py,hibooboo2/docker-py,rancher/docker-py,tbeadle/docker-py,vitalyisaev2/docker-py,sourcelair/docker-py,uian/docker-py,delfick/docker-py,olsaki/docker-py,youhong316/docker-py,funkyfuture/docker-py,ColinHuang/docker-py,ssanderson/docker-py,youhong316/docker-py,MohamedAshiqrh/docker-py,auready/docker-py,v-boyko/docker-py,clarete/docker-py,bfirsh/docker-py,aiden0z/docker-py,tangkun75/docker-py,mnowster/docker-py,Faylixe/docker-py,gamechanger/docker-py,shakamunyi/docker-py,jpopelka/docker-py,mikedougherty/docker-py,zwqzhangweiqiang/docker-py,mark-adams/docker-py,funkyfuture/docker-py,shishir-a412ed/docker-py,minzhang28/docker-py,shakamunyi/docker-py,ColinHuang/docker-py,tshauck/docker-py,vpetersson/docker-py,Faylixe/docker-py,jamesmarva/docker-py,wehkamp/docker-py,wallrj/docker-py,vdemeester/docker-py,minzhang28/docker-py,leolujuyi/docker-py,mohitsoni/docker-py,Melraidin/docker-py,bfirsh/docker-py,aiden0z/docker-py,vdemeester/docker-py,Ye-Yong-Chi/docker-py,erikced/docker-py,PierreF/docker-py,kpavel/docker-py,schu/docker-py,tristan0x/docker-py,rastaman/docker-py,stevenewey/docker-py,rschmidtz/docker-py,dlorenc/docker-py,jhowardmsft/docker-py,wlan0/docker-py,uggla/docker-py
|
from distutils.version import StrictVersion
from requests.adapters import HTTPAdapter
try:
- from requests.packages.urllib3.poolmanager import PoolManager
+ import requests.packages.urllib3 as urllib3
except ImportError:
import urllib3
- from urllib3.poolmanager import PoolManager
+
+
+ PoolManager = urllib3.poolmanager.PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that uses an arbitrary SSL version.'''
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
- urllib_ver = urllib3.__version__
+ urllib_ver = urllib3.__version__.split('-')[0]
+ if urllib3 and urllib_ver != 'dev' and \
- if urllib3 and StrictVersion(urllib_ver) <= StrictVersion('1.5'):
+ StrictVersion(urllib_ver) <= StrictVersion('1.5'):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block)
else:
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
|
Fix some urllib3 import issues
|
## Code Before:
from distutils.version import StrictVersion
from requests.adapters import HTTPAdapter
try:
from requests.packages.urllib3.poolmanager import PoolManager
except ImportError:
import urllib3
from urllib3.poolmanager import PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that uses an arbitrary SSL version.'''
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
urllib_ver = urllib3.__version__
if urllib3 and StrictVersion(urllib_ver) <= StrictVersion('1.5'):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block)
else:
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
## Instruction:
Fix some urllib3 import issues
## Code After:
from distutils.version import StrictVersion
from requests.adapters import HTTPAdapter
try:
import requests.packages.urllib3 as urllib3
except ImportError:
import urllib3
PoolManager = urllib3.poolmanager.PoolManager
class SSLAdapter(HTTPAdapter):
'''An HTTPS Transport Adapter that uses an arbitrary SSL version.'''
def __init__(self, ssl_version=None, **kwargs):
self.ssl_version = ssl_version
super(SSLAdapter, self).__init__(**kwargs)
def init_poolmanager(self, connections, maxsize, block=False):
urllib_ver = urllib3.__version__.split('-')[0]
if urllib3 and urllib_ver != 'dev' and \
StrictVersion(urllib_ver) <= StrictVersion('1.5'):
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block)
else:
self.poolmanager = PoolManager(num_pools=connections,
maxsize=maxsize,
block=block,
ssl_version=self.ssl_version)
|
...
try:
import requests.packages.urllib3 as urllib3
except ImportError:
...
import urllib3
PoolManager = urllib3.poolmanager.PoolManager
...
def init_poolmanager(self, connections, maxsize, block=False):
urllib_ver = urllib3.__version__.split('-')[0]
if urllib3 and urllib_ver != 'dev' and \
StrictVersion(urllib_ver) <= StrictVersion('1.5'):
self.poolmanager = PoolManager(num_pools=connections,
...
|
a4a37a783efcfd1cbb21acc29077c8096a0a0198
|
spacy/lang/pl/__init__.py
|
spacy/lang/pl/__init__.py
|
from __future__ import unicode_literals
from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
from .stop_words import STOP_WORDS
from ..tokenizer_exceptions import BASE_EXCEPTIONS
from ...language import Language
from ...attrs import LANG
from ...util import update_exc
class Polish(Language):
lang = 'pl'
class Defaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'pl'
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS)
stop_words = set(STOP_WORDS)
__all__ = ['Polish']
|
from __future__ import unicode_literals
from .stop_words import STOP_WORDS
from ..tokenizer_exceptions import BASE_EXCEPTIONS
from ...language import Language
from ...attrs import LANG
from ...util import update_exc
class Polish(Language):
lang = 'pl'
class Defaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'pl'
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS)
stop_words = set(STOP_WORDS)
__all__ = ['Polish']
|
Remove import from non-existing module
|
Remove import from non-existing module
|
Python
|
mit
|
honnibal/spaCy,spacy-io/spaCy,explosion/spaCy,recognai/spaCy,spacy-io/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,spacy-io/spaCy,honnibal/spaCy,honnibal/spaCy,recognai/spaCy,aikramer2/spaCy,honnibal/spaCy,explosion/spaCy,aikramer2/spaCy,explosion/spaCy,aikramer2/spaCy,spacy-io/spaCy,recognai/spaCy,spacy-io/spaCy,aikramer2/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy
|
from __future__ import unicode_literals
- from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
from .stop_words import STOP_WORDS
from ..tokenizer_exceptions import BASE_EXCEPTIONS
from ...language import Language
from ...attrs import LANG
from ...util import update_exc
class Polish(Language):
lang = 'pl'
class Defaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'pl'
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS)
stop_words = set(STOP_WORDS)
__all__ = ['Polish']
|
Remove import from non-existing module
|
## Code Before:
from __future__ import unicode_literals
from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS
from .stop_words import STOP_WORDS
from ..tokenizer_exceptions import BASE_EXCEPTIONS
from ...language import Language
from ...attrs import LANG
from ...util import update_exc
class Polish(Language):
lang = 'pl'
class Defaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'pl'
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS)
stop_words = set(STOP_WORDS)
__all__ = ['Polish']
## Instruction:
Remove import from non-existing module
## Code After:
from __future__ import unicode_literals
from .stop_words import STOP_WORDS
from ..tokenizer_exceptions import BASE_EXCEPTIONS
from ...language import Language
from ...attrs import LANG
from ...util import update_exc
class Polish(Language):
lang = 'pl'
class Defaults(Language.Defaults):
lex_attr_getters = dict(Language.Defaults.lex_attr_getters)
lex_attr_getters[LANG] = lambda text: 'pl'
tokenizer_exceptions = update_exc(BASE_EXCEPTIONS)
stop_words = set(STOP_WORDS)
__all__ = ['Polish']
|
...
from .stop_words import STOP_WORDS
...
|
4a6f76857a626dd756675a4fe1dd3660cf63d8b7
|
alg_fibonacci.py
|
alg_fibonacci.py
|
from __future__ import print_function
import time
def fibonacci(n):
"""Get nth number of Fibonacci series by recursion."""
if n == 0:
return 0
elif n == 1 or n == 2:
return 1
else:
return fibonacci(n - 1) + fibonacci(n - 2)
def main():
n = 13
print('{}th number of Fibonacci series: {}'
.format(n, fibonacci(n)))
if __name__ == '__main__':
main()
|
from __future__ import print_function
def fibonacci(n):
"""Get nth number of Fibonacci series by recursion."""
if n == 0:
return 0
elif n == 1 or n == 2:
return 1
else:
return fibonacci(n - 1) + fibonacci(n - 2)
def main():
import time
n = 13
print('{}th number of Fibonacci series: {}'
.format(n, fibonacci(n)))
if __name__ == '__main__':
main()
|
Move module time to main()
|
Move module time to main()
|
Python
|
bsd-2-clause
|
bowen0701/algorithms_data_structures
|
from __future__ import print_function
- import time
def fibonacci(n):
"""Get nth number of Fibonacci series by recursion."""
if n == 0:
return 0
elif n == 1 or n == 2:
return 1
else:
return fibonacci(n - 1) + fibonacci(n - 2)
def main():
+ import time
n = 13
print('{}th number of Fibonacci series: {}'
.format(n, fibonacci(n)))
if __name__ == '__main__':
main()
|
Move module time to main()
|
## Code Before:
from __future__ import print_function
import time
def fibonacci(n):
"""Get nth number of Fibonacci series by recursion."""
if n == 0:
return 0
elif n == 1 or n == 2:
return 1
else:
return fibonacci(n - 1) + fibonacci(n - 2)
def main():
n = 13
print('{}th number of Fibonacci series: {}'
.format(n, fibonacci(n)))
if __name__ == '__main__':
main()
## Instruction:
Move module time to main()
## Code After:
from __future__ import print_function
def fibonacci(n):
"""Get nth number of Fibonacci series by recursion."""
if n == 0:
return 0
elif n == 1 or n == 2:
return 1
else:
return fibonacci(n - 1) + fibonacci(n - 2)
def main():
import time
n = 13
print('{}th number of Fibonacci series: {}'
.format(n, fibonacci(n)))
if __name__ == '__main__':
main()
|
// ... existing code ...
from __future__ import print_function
// ... modified code ...
def main():
import time
n = 13
// ... rest of the code ...
|
c713273fe145418113d750579f8b135dc513c3b8
|
config.py
|
config.py
|
import os
if os.environ.get('DATABASE_URL') is None:
SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db'
else:
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
|
import os
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
|
Delete default case for SQLALCHEMY_DATABASE_URI
|
Delete default case for SQLALCHEMY_DATABASE_URI
if user doesn't set it, he coud have some problems with SQLite
|
Python
|
mit
|
Stark-Mountain/meetup-facebook-bot,Stark-Mountain/meetup-facebook-bot
|
import os
- if os.environ.get('DATABASE_URL') is None:
- SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db'
- else:
- SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
+ SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
-
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
|
Delete default case for SQLALCHEMY_DATABASE_URI
|
## Code Before:
import os
if os.environ.get('DATABASE_URL') is None:
SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db'
else:
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
## Instruction:
Delete default case for SQLALCHEMY_DATABASE_URI
## Code After:
import os
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
|
// ... existing code ...
SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']
SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
// ... rest of the code ...
|
8fc274021a8c0813f3fc3568d1d7984112952b9c
|
pytilemap/qtsupport.py
|
pytilemap/qtsupport.py
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
Use Cache location instead of temp folder
|
Use Cache location instead of temp folder
|
Python
|
mit
|
allebacco/PyTileMap
|
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
- return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
+ return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
- return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
+ return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
Use Cache location instead of temp folder
|
## Code Before:
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.TempLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
## Instruction:
Use Cache location instead of temp folder
## Code After:
import sys
import sip
import qtpy
__all__ = [
'getQVariantValue',
'wheelAngleDelta',
]
try:
if qtpy.PYQT5:
QVARIANT_API = 2
else:
QVARIANT_API = sip.getapi('QVariant')
except ValueError:
QVARIANT_API = 1
if QVARIANT_API == 1:
def getQVariantValue(variant):
return variant.toPyObject()
else:
def getQVariantValue(variant):
return variant
if qtpy.PYQT5:
def wheelAngleDelta(wheelEvent):
return wheelEvent.angleDelta().y()
else:
def wheelAngleDelta(wheelEvent):
return wheelEvent.delta()
if qtpy.PYQT5:
from qtpy.QtCore import QStandardPaths
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
else:
from qtpy.QtGui import QDesktopServices
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
|
# ... existing code ...
def getTemporaryFolder():
return QStandardPaths.writableLocation(QStandardPaths.CacheLocation)
# ... modified code ...
def getTemporaryFolder():
return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
# ... rest of the code ...
|
9d5abdaefa483574cdd81da8d8d4e63ef68f5ab8
|
crossfolium/__init__.py
|
crossfolium/__init__.py
|
import crossfolium.marker_function as marker_function
from crossfolium.crossfolium import (
Crossfilter,
PieFilter,
RowBarFilter,
BarFilter,
TableFilter,
CountFilter,
ResetFilter,
GeoChoroplethFilter,
)
from .map import (
FeatureGroupFilter,
HeatmapFilter,
)
__version__ = "0.0.0"
__all__ = [
'__version__',
'marker_function',
'Crossfilter',
'PieFilter',
'RowBarFilter',
'BarFilter',
'FeatureGroupFilter',
'TableFilter',
'CountFilter',
'ResetFilter',
'HeatmapFilter',
'GeoChoroplethFilter',
]
|
from __future__ import absolute_import
from crossfolium import marker_function
from crossfolium.crossfolium import (
Crossfilter,
PieFilter,
RowBarFilter,
BarFilter,
TableFilter,
CountFilter,
ResetFilter,
GeoChoroplethFilter,
)
from crossfolium.map import (
FeatureGroupFilter,
HeatmapFilter,
)
__version__ = "0.0.0"
__all__ = [
'__version__',
'marker_function',
'Crossfilter',
'PieFilter',
'RowBarFilter',
'BarFilter',
'FeatureGroupFilter',
'TableFilter',
'CountFilter',
'ResetFilter',
'HeatmapFilter',
'GeoChoroplethFilter',
]
|
Handle absolute import for py27
|
Handle absolute import for py27
|
Python
|
mit
|
BibMartin/crossfolium,BibMartin/crossfolium
|
+ from __future__ import absolute_import
- import crossfolium.marker_function as marker_function
+ from crossfolium import marker_function
from crossfolium.crossfolium import (
Crossfilter,
PieFilter,
RowBarFilter,
BarFilter,
TableFilter,
CountFilter,
ResetFilter,
GeoChoroplethFilter,
)
- from .map import (
+ from crossfolium.map import (
FeatureGroupFilter,
HeatmapFilter,
)
__version__ = "0.0.0"
__all__ = [
'__version__',
'marker_function',
'Crossfilter',
'PieFilter',
'RowBarFilter',
'BarFilter',
'FeatureGroupFilter',
'TableFilter',
'CountFilter',
'ResetFilter',
'HeatmapFilter',
'GeoChoroplethFilter',
]
|
Handle absolute import for py27
|
## Code Before:
import crossfolium.marker_function as marker_function
from crossfolium.crossfolium import (
Crossfilter,
PieFilter,
RowBarFilter,
BarFilter,
TableFilter,
CountFilter,
ResetFilter,
GeoChoroplethFilter,
)
from .map import (
FeatureGroupFilter,
HeatmapFilter,
)
__version__ = "0.0.0"
__all__ = [
'__version__',
'marker_function',
'Crossfilter',
'PieFilter',
'RowBarFilter',
'BarFilter',
'FeatureGroupFilter',
'TableFilter',
'CountFilter',
'ResetFilter',
'HeatmapFilter',
'GeoChoroplethFilter',
]
## Instruction:
Handle absolute import for py27
## Code After:
from __future__ import absolute_import
from crossfolium import marker_function
from crossfolium.crossfolium import (
Crossfilter,
PieFilter,
RowBarFilter,
BarFilter,
TableFilter,
CountFilter,
ResetFilter,
GeoChoroplethFilter,
)
from crossfolium.map import (
FeatureGroupFilter,
HeatmapFilter,
)
__version__ = "0.0.0"
__all__ = [
'__version__',
'marker_function',
'Crossfilter',
'PieFilter',
'RowBarFilter',
'BarFilter',
'FeatureGroupFilter',
'TableFilter',
'CountFilter',
'ResetFilter',
'HeatmapFilter',
'GeoChoroplethFilter',
]
|
...
from __future__ import absolute_import
from crossfolium import marker_function
...
from crossfolium.map import (
FeatureGroupFilter,
...
|
6aa8f148b3b3975363d5d4a763f5abb45ea6cbd8
|
databin/parsers/__init__.py
|
databin/parsers/__init__.py
|
from databin.parsers.util import ParseException
from databin.parsers.simple import parse_csv, parse_tsv
from databin.parsers.psql import parse_psql
PARSERS = [
('Comma-Separated Values', 'csv', parse_csv),
('Tab-Separated Values', 'tsv', parse_tsv),
('Excel copy & paste', 'excel', parse_tsv),
('psql Shell', 'psql', parse_psql),
]
def parse(format, data):
for name, key, func in PARSERS:
if key == format:
return func(data)
raise ParseException()
def get_parsers():
for name, key, func in PARSERS:
yield (key, name)
|
from databin.parsers.util import ParseException
from databin.parsers.simple import parse_csv, parse_tsv
from databin.parsers.psql import parse_psql
PARSERS = [
('Excel copy & paste', 'excel', parse_tsv),
('Comma-Separated Values', 'csv', parse_csv),
('Tab-Separated Values', 'tsv', parse_tsv),
('psql Shell', 'psql', parse_psql),
]
def parse(format, data):
for name, key, func in PARSERS:
if key == format:
return func(data)
raise ParseException()
def get_parsers():
for name, key, func in PARSERS:
yield (key, name)
|
Make excel format the default
|
Make excel format the default
|
Python
|
mit
|
LeTristanB/Pastable,pudo/databin,LeTristanB/Pastable
|
from databin.parsers.util import ParseException
from databin.parsers.simple import parse_csv, parse_tsv
from databin.parsers.psql import parse_psql
PARSERS = [
+ ('Excel copy & paste', 'excel', parse_tsv),
('Comma-Separated Values', 'csv', parse_csv),
('Tab-Separated Values', 'tsv', parse_tsv),
- ('Excel copy & paste', 'excel', parse_tsv),
('psql Shell', 'psql', parse_psql),
]
def parse(format, data):
for name, key, func in PARSERS:
if key == format:
return func(data)
raise ParseException()
def get_parsers():
for name, key, func in PARSERS:
yield (key, name)
|
Make excel format the default
|
## Code Before:
from databin.parsers.util import ParseException
from databin.parsers.simple import parse_csv, parse_tsv
from databin.parsers.psql import parse_psql
PARSERS = [
('Comma-Separated Values', 'csv', parse_csv),
('Tab-Separated Values', 'tsv', parse_tsv),
('Excel copy & paste', 'excel', parse_tsv),
('psql Shell', 'psql', parse_psql),
]
def parse(format, data):
for name, key, func in PARSERS:
if key == format:
return func(data)
raise ParseException()
def get_parsers():
for name, key, func in PARSERS:
yield (key, name)
## Instruction:
Make excel format the default
## Code After:
from databin.parsers.util import ParseException
from databin.parsers.simple import parse_csv, parse_tsv
from databin.parsers.psql import parse_psql
PARSERS = [
('Excel copy & paste', 'excel', parse_tsv),
('Comma-Separated Values', 'csv', parse_csv),
('Tab-Separated Values', 'tsv', parse_tsv),
('psql Shell', 'psql', parse_psql),
]
def parse(format, data):
for name, key, func in PARSERS:
if key == format:
return func(data)
raise ParseException()
def get_parsers():
for name, key, func in PARSERS:
yield (key, name)
|
# ... existing code ...
PARSERS = [
('Excel copy & paste', 'excel', parse_tsv),
('Comma-Separated Values', 'csv', parse_csv),
# ... modified code ...
('Tab-Separated Values', 'tsv', parse_tsv),
('psql Shell', 'psql', parse_psql),
# ... rest of the code ...
|
fc41951d1e395c3cdc8994b4c025e9776d67d4e0
|
http.py
|
http.py
|
from django.http import HttpResponse
class HttpResponseCreated(HttpResponse):
status_code = 201
class HttpResponseNoContent(HttpResponse):
status_code = 204
class HttpResponseNotAllowed(HttpResponse):
status_code = 405
def __init__(self, allow_headers):
"""
RFC2616: The response MUST include an Allow header containing
a list of valid methods for the requested resource.
"""
super(HttpResponseNotAllowed, self).__init__()
try:
iter(allow_headers)
except TypeError:
self['Allow'] = allow_headers
else:
self['Allow'] = ", ".join(allow_headers)
class HttpResponseNotAcceptable(HttpResponse):
status_code = 406
class HttpResponseConflict(HttpResponse):
status_code = 409
class HttpResponseNotImplemented(HttpResponse):
status_code = 501
|
from django.http import HttpResponse
class HttpResponseCreated(HttpResponse):
status_code = 201
class HttpResponseNoContent(HttpResponse):
status_code = 204
class HttpResponseNotAllowed(HttpResponse):
status_code = 405
def __init__(self, allow_headers):
"""
RFC2616: The response MUST include an Allow header containing
a list of valid methods for the requested resource.
"""
super(HttpResponseNotAllowed, self).__init__()
try:
iter(allow_headers)
except TypeError:
self['Allow'] = allow_headers
else:
self['Allow'] = ", ".join(allow_headers)
class HttpResponseNotAcceptable(HttpResponse):
status_code = 406
class HttpResponseConflict(HttpResponse):
status_code = 409
class HttpResponseUnsupportedMediaType(HttpResponse):
status_code = 415
class HttpResponseNotImplemented(HttpResponse):
status_code = 501
|
Add Response class for unsupported media
|
Add Response class for unsupported media
|
Python
|
mit
|
danrex/django-riv,danrex/django-riv
|
from django.http import HttpResponse
class HttpResponseCreated(HttpResponse):
status_code = 201
class HttpResponseNoContent(HttpResponse):
status_code = 204
class HttpResponseNotAllowed(HttpResponse):
status_code = 405
def __init__(self, allow_headers):
"""
RFC2616: The response MUST include an Allow header containing
a list of valid methods for the requested resource.
"""
super(HttpResponseNotAllowed, self).__init__()
try:
iter(allow_headers)
except TypeError:
self['Allow'] = allow_headers
else:
self['Allow'] = ", ".join(allow_headers)
class HttpResponseNotAcceptable(HttpResponse):
status_code = 406
class HttpResponseConflict(HttpResponse):
status_code = 409
+ class HttpResponseUnsupportedMediaType(HttpResponse):
+ status_code = 415
+
class HttpResponseNotImplemented(HttpResponse):
status_code = 501
|
Add Response class for unsupported media
|
## Code Before:
from django.http import HttpResponse
class HttpResponseCreated(HttpResponse):
status_code = 201
class HttpResponseNoContent(HttpResponse):
status_code = 204
class HttpResponseNotAllowed(HttpResponse):
status_code = 405
def __init__(self, allow_headers):
"""
RFC2616: The response MUST include an Allow header containing
a list of valid methods for the requested resource.
"""
super(HttpResponseNotAllowed, self).__init__()
try:
iter(allow_headers)
except TypeError:
self['Allow'] = allow_headers
else:
self['Allow'] = ", ".join(allow_headers)
class HttpResponseNotAcceptable(HttpResponse):
status_code = 406
class HttpResponseConflict(HttpResponse):
status_code = 409
class HttpResponseNotImplemented(HttpResponse):
status_code = 501
## Instruction:
Add Response class for unsupported media
## Code After:
from django.http import HttpResponse
class HttpResponseCreated(HttpResponse):
status_code = 201
class HttpResponseNoContent(HttpResponse):
status_code = 204
class HttpResponseNotAllowed(HttpResponse):
status_code = 405
def __init__(self, allow_headers):
"""
RFC2616: The response MUST include an Allow header containing
a list of valid methods for the requested resource.
"""
super(HttpResponseNotAllowed, self).__init__()
try:
iter(allow_headers)
except TypeError:
self['Allow'] = allow_headers
else:
self['Allow'] = ", ".join(allow_headers)
class HttpResponseNotAcceptable(HttpResponse):
status_code = 406
class HttpResponseConflict(HttpResponse):
status_code = 409
class HttpResponseUnsupportedMediaType(HttpResponse):
status_code = 415
class HttpResponseNotImplemented(HttpResponse):
status_code = 501
|
// ... existing code ...
class HttpResponseUnsupportedMediaType(HttpResponse):
status_code = 415
class HttpResponseNotImplemented(HttpResponse):
// ... rest of the code ...
|
a32e61e9cdf2eababb568659766688a731b121cb
|
warlock/__init__.py
|
warlock/__init__.py
|
"""Public-facing Warlock API"""
from warlock.core import model_factory # NOQA
from warlock.exceptions import InvalidOperation # NOQA
|
"""Public-facing Warlock API"""
from warlock.core import model_factory # noqa: F401
from warlock.exceptions import InvalidOperation # noqa: F401
|
Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook
|
Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook
|
Python
|
apache-2.0
|
bcwaldon/warlock
|
"""Public-facing Warlock API"""
- from warlock.core import model_factory # NOQA
+ from warlock.core import model_factory # noqa: F401
- from warlock.exceptions import InvalidOperation # NOQA
+ from warlock.exceptions import InvalidOperation # noqa: F401
|
Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook
|
## Code Before:
"""Public-facing Warlock API"""
from warlock.core import model_factory # NOQA
from warlock.exceptions import InvalidOperation # NOQA
## Instruction:
Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook
## Code After:
"""Public-facing Warlock API"""
from warlock.core import model_factory # noqa: F401
from warlock.exceptions import InvalidOperation # noqa: F401
|
...
from warlock.core import model_factory # noqa: F401
from warlock.exceptions import InvalidOperation # noqa: F401
...
|
61bbd4e8fc0712fe56614481173eb86d409eb8d7
|
tests/test_linked_list.py
|
tests/test_linked_list.py
|
from unittest import TestCase
from pystructures.linked_lists import LinkedList, Node
class TestNode(TestCase):
def test_value(self):
""" A simple test to check the Node's value """
node = Node(10)
self.assertEqual(10, node.value)
def test_improper_node(self):
""" A test to check if an invalid data type is set as a node's next"""
node = Node(10)
with self.assertRaises(ValueError):
node.next = "Hello"
class TestLinkedList(TestCase):
def test_insert(self):
""" A simple test to check if insertion works as expected in a singly linked list """
l = LinkedList()
results = [l.insert(val) for val in xrange(10, 100, 10)]
self.assertEqual(len(set(results)), 1)
self.assertTrue(results[0], msg="Testing for successful insertion...")
self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
|
from builtins import range
from unittest import TestCase
from pystructures.linked_lists import LinkedList, Node
class TestNode(TestCase):
def test_value(self):
""" A simple test to check the Node's value """
node = Node(10)
self.assertEqual(10, node.value)
def test_improper_node(self):
""" A test to check if an invalid data type is set as a node's next"""
node = Node(10)
with self.assertRaises(ValueError):
node.next = "Hello"
class TestLinkedList(TestCase):
def test_insert(self):
""" A simple test to check if insertion works as expected in a singly linked list """
l = LinkedList()
results = [l.insert(val) for val in range(10, 100, 10)]
self.assertEqual(len(set(results)), 1)
self.assertTrue(results[0], msg="Testing for successful insertion...")
self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
|
Fix range issue with travis
|
Fix range issue with travis
|
Python
|
mit
|
apranav19/pystructures
|
+ from builtins import range
from unittest import TestCase
from pystructures.linked_lists import LinkedList, Node
class TestNode(TestCase):
def test_value(self):
""" A simple test to check the Node's value """
node = Node(10)
self.assertEqual(10, node.value)
def test_improper_node(self):
""" A test to check if an invalid data type is set as a node's next"""
node = Node(10)
with self.assertRaises(ValueError):
node.next = "Hello"
class TestLinkedList(TestCase):
def test_insert(self):
""" A simple test to check if insertion works as expected in a singly linked list """
l = LinkedList()
- results = [l.insert(val) for val in xrange(10, 100, 10)]
+ results = [l.insert(val) for val in range(10, 100, 10)]
self.assertEqual(len(set(results)), 1)
self.assertTrue(results[0], msg="Testing for successful insertion...")
self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
|
Fix range issue with travis
|
## Code Before:
from unittest import TestCase
from pystructures.linked_lists import LinkedList, Node
class TestNode(TestCase):
def test_value(self):
""" A simple test to check the Node's value """
node = Node(10)
self.assertEqual(10, node.value)
def test_improper_node(self):
""" A test to check if an invalid data type is set as a node's next"""
node = Node(10)
with self.assertRaises(ValueError):
node.next = "Hello"
class TestLinkedList(TestCase):
def test_insert(self):
""" A simple test to check if insertion works as expected in a singly linked list """
l = LinkedList()
results = [l.insert(val) for val in xrange(10, 100, 10)]
self.assertEqual(len(set(results)), 1)
self.assertTrue(results[0], msg="Testing for successful insertion...")
self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
## Instruction:
Fix range issue with travis
## Code After:
from builtins import range
from unittest import TestCase
from pystructures.linked_lists import LinkedList, Node
class TestNode(TestCase):
def test_value(self):
""" A simple test to check the Node's value """
node = Node(10)
self.assertEqual(10, node.value)
def test_improper_node(self):
""" A test to check if an invalid data type is set as a node's next"""
node = Node(10)
with self.assertRaises(ValueError):
node.next = "Hello"
class TestLinkedList(TestCase):
def test_insert(self):
""" A simple test to check if insertion works as expected in a singly linked list """
l = LinkedList()
results = [l.insert(val) for val in range(10, 100, 10)]
self.assertEqual(len(set(results)), 1)
self.assertTrue(results[0], msg="Testing for successful insertion...")
self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
|
...
from builtins import range
from unittest import TestCase
...
l = LinkedList()
results = [l.insert(val) for val in range(10, 100, 10)]
self.assertEqual(len(set(results)), 1)
...
|
9851430922f9c14583c9eb17062629f6ea99c258
|
turbustat/tests/test_vcs.py
|
turbustat/tests/test_vcs.py
|
'''
Test functions for VCS
'''
from unittest import TestCase
import numpy as np
import numpy.testing as npt
from ..statistics import VCS, VCS_Distance
from ._testing_data import \
dataset1, dataset2, computed_data, computed_distances
class testVCS(TestCase):
def setUp(self):
self.dataset1 = dataset1
self.dataset2 = dataset2
def test_VCS_method(self):
self.tester = VCS(dataset1["cube"]).run()
npt.assert_allclose(self.tester.ps1D, computed_data['vcs_val'])
def test_VCS_distance(self):
self.tester_dist = \
VCS_Distance(dataset1["cube"], dataset2["cube"])
self.tester_dist = self.tester_dist.distance_metric()
npt.assert_almost_equal(self.tester_dist.distance,
computed_distances['vcs_distance'])
|
'''
Test functions for VCS
'''
import pytest
import numpy as np
import numpy.testing as npt
from ..statistics import VCS, VCS_Distance
from ._testing_data import \
dataset1, dataset2, computed_data, computed_distances
def test_VCS_method():
tester = VCS(dataset1["cube"]).run()
npt.assert_allclose(tester.ps1D, computed_data['vcs_val'])
npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val'])
def test_VCS_distance():
tester_dist = \
VCS_Distance(dataset1["cube"], dataset2["cube"])
tester_dist = tester_dist.distance_metric()
npt.assert_almost_equal(tester_dist.distance,
computed_distances['vcs_distance'])
# Add tests for: VCS changing the spectral width, pixel and spectral units,
|
Reformat VCS tests; need updated unit test values!
|
Reformat VCS tests; need updated unit test values!
|
Python
|
mit
|
e-koch/TurbuStat,Astroua/TurbuStat
|
'''
Test functions for VCS
'''
- from unittest import TestCase
+ import pytest
import numpy as np
import numpy.testing as npt
from ..statistics import VCS, VCS_Distance
from ._testing_data import \
dataset1, dataset2, computed_data, computed_distances
- class testVCS(TestCase):
+ def test_VCS_method():
+ tester = VCS(dataset1["cube"]).run()
+ npt.assert_allclose(tester.ps1D, computed_data['vcs_val'])
- def setUp(self):
- self.dataset1 = dataset1
- self.dataset2 = dataset2
+ npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val'])
- def test_VCS_method(self):
- self.tester = VCS(dataset1["cube"]).run()
- npt.assert_allclose(self.tester.ps1D, computed_data['vcs_val'])
- def test_VCS_distance(self):
+ def test_VCS_distance():
- self.tester_dist = \
+ tester_dist = \
- VCS_Distance(dataset1["cube"], dataset2["cube"])
+ VCS_Distance(dataset1["cube"], dataset2["cube"])
- self.tester_dist = self.tester_dist.distance_metric()
+ tester_dist = tester_dist.distance_metric()
- npt.assert_almost_equal(self.tester_dist.distance,
+ npt.assert_almost_equal(tester_dist.distance,
- computed_distances['vcs_distance'])
+ computed_distances['vcs_distance'])
+ # Add tests for: VCS changing the spectral width, pixel and spectral units,
+
|
Reformat VCS tests; need updated unit test values!
|
## Code Before:
'''
Test functions for VCS
'''
from unittest import TestCase
import numpy as np
import numpy.testing as npt
from ..statistics import VCS, VCS_Distance
from ._testing_data import \
dataset1, dataset2, computed_data, computed_distances
class testVCS(TestCase):
def setUp(self):
self.dataset1 = dataset1
self.dataset2 = dataset2
def test_VCS_method(self):
self.tester = VCS(dataset1["cube"]).run()
npt.assert_allclose(self.tester.ps1D, computed_data['vcs_val'])
def test_VCS_distance(self):
self.tester_dist = \
VCS_Distance(dataset1["cube"], dataset2["cube"])
self.tester_dist = self.tester_dist.distance_metric()
npt.assert_almost_equal(self.tester_dist.distance,
computed_distances['vcs_distance'])
## Instruction:
Reformat VCS tests; need updated unit test values!
## Code After:
'''
Test functions for VCS
'''
import pytest
import numpy as np
import numpy.testing as npt
from ..statistics import VCS, VCS_Distance
from ._testing_data import \
dataset1, dataset2, computed_data, computed_distances
def test_VCS_method():
tester = VCS(dataset1["cube"]).run()
npt.assert_allclose(tester.ps1D, computed_data['vcs_val'])
npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val'])
def test_VCS_distance():
tester_dist = \
VCS_Distance(dataset1["cube"], dataset2["cube"])
tester_dist = tester_dist.distance_metric()
npt.assert_almost_equal(tester_dist.distance,
computed_distances['vcs_distance'])
# Add tests for: VCS changing the spectral width, pixel and spectral units,
|
# ... existing code ...
import pytest
# ... modified code ...
def test_VCS_method():
tester = VCS(dataset1["cube"]).run()
npt.assert_allclose(tester.ps1D, computed_data['vcs_val'])
npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val'])
def test_VCS_distance():
tester_dist = \
VCS_Distance(dataset1["cube"], dataset2["cube"])
tester_dist = tester_dist.distance_metric()
npt.assert_almost_equal(tester_dist.distance,
computed_distances['vcs_distance'])
# Add tests for: VCS changing the spectral width, pixel and spectral units,
# ... rest of the code ...
|
f48cb4fd946c8fa4b6157b8e1ea9ad8b385bc67a
|
src/hades/bin/generate_config.py
|
src/hades/bin/generate_config.py
|
import os
import sys
from hades import constants
from hades.common.cli import ArgumentParser, parser as common_parser
from hades.config.generate import ConfigGenerator
from hades.config.loader import load_config
def main():
parser = ArgumentParser(parents=[common_parser])
parser.add_argument(dest='source', metavar='SOURCE',
help="Template file name or template directory name")
parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?',
help="Destination file or directory (default is stdout"
"for files; required for directories)")
args = parser.parse_args()
config = load_config(args.config)
template_dir = constants.templatedir
generator = ConfigGenerator(template_dir, config)
source_path = os.path.join(template_dir, args.source)
if os.path.isdir(source_path):
generator.from_directory(args.source, args.destination)
elif os.path.isfile(source_path):
if args.destination is None:
generator.from_file(args.source, sys.stdout)
else:
with open(args.destination, 'w', encoding='utf-8') as f:
generator.from_file(args.source, f)
else:
print("No such file or directory {} in {}".format(args.source,
template_dir),
file=sys.stderr)
return os.EX_NOINPUT
if __name__ == '__main__':
sys.exit(main())
|
import logging
import os
import sys
from hades import constants
from hades.common.cli import ArgumentParser, parser as common_parser
from hades.config.generate import ConfigGenerator
from hades.config.loader import load_config
logger = logging.getLogger()
def main():
parser = ArgumentParser(parents=[common_parser])
parser.add_argument(dest='source', metavar='SOURCE',
help="Template file name or template directory name")
parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?',
help="Destination file or directory (default is stdout"
"for files; required for directories)")
args = parser.parse_args()
config = load_config(args.config)
template_dir = constants.templatedir
generator = ConfigGenerator(template_dir, config)
source_path = os.path.join(template_dir, args.source)
if os.path.isdir(source_path):
generator.from_directory(args.source, args.destination)
elif os.path.isfile(source_path):
if args.destination is None:
generator.from_file(args.source, sys.stdout)
else:
with open(args.destination, 'w', encoding='utf-8') as f:
generator.from_file(args.source, f)
else:
logger.critical("No such file or directory {} in {}"
.format(args.source, template_dir))
return os.EX_NOINPUT
if __name__ == '__main__':
sys.exit(main())
|
Use logger for hades-generate-config error messages
|
Use logger for hades-generate-config error messages
|
Python
|
mit
|
agdsn/hades,agdsn/hades,agdsn/hades,agdsn/hades,agdsn/hades
|
+ import logging
import os
import sys
from hades import constants
from hades.common.cli import ArgumentParser, parser as common_parser
from hades.config.generate import ConfigGenerator
from hades.config.loader import load_config
+
+ logger = logging.getLogger()
def main():
parser = ArgumentParser(parents=[common_parser])
parser.add_argument(dest='source', metavar='SOURCE',
help="Template file name or template directory name")
parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?',
help="Destination file or directory (default is stdout"
"for files; required for directories)")
args = parser.parse_args()
config = load_config(args.config)
template_dir = constants.templatedir
generator = ConfigGenerator(template_dir, config)
source_path = os.path.join(template_dir, args.source)
if os.path.isdir(source_path):
generator.from_directory(args.source, args.destination)
elif os.path.isfile(source_path):
if args.destination is None:
generator.from_file(args.source, sys.stdout)
else:
with open(args.destination, 'w', encoding='utf-8') as f:
generator.from_file(args.source, f)
else:
+ logger.critical("No such file or directory {} in {}"
+ .format(args.source, template_dir))
- print("No such file or directory {} in {}".format(args.source,
- template_dir),
- file=sys.stderr)
return os.EX_NOINPUT
if __name__ == '__main__':
sys.exit(main())
|
Use logger for hades-generate-config error messages
|
## Code Before:
import os
import sys
from hades import constants
from hades.common.cli import ArgumentParser, parser as common_parser
from hades.config.generate import ConfigGenerator
from hades.config.loader import load_config
def main():
parser = ArgumentParser(parents=[common_parser])
parser.add_argument(dest='source', metavar='SOURCE',
help="Template file name or template directory name")
parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?',
help="Destination file or directory (default is stdout"
"for files; required for directories)")
args = parser.parse_args()
config = load_config(args.config)
template_dir = constants.templatedir
generator = ConfigGenerator(template_dir, config)
source_path = os.path.join(template_dir, args.source)
if os.path.isdir(source_path):
generator.from_directory(args.source, args.destination)
elif os.path.isfile(source_path):
if args.destination is None:
generator.from_file(args.source, sys.stdout)
else:
with open(args.destination, 'w', encoding='utf-8') as f:
generator.from_file(args.source, f)
else:
print("No such file or directory {} in {}".format(args.source,
template_dir),
file=sys.stderr)
return os.EX_NOINPUT
if __name__ == '__main__':
sys.exit(main())
## Instruction:
Use logger for hades-generate-config error messages
## Code After:
import logging
import os
import sys
from hades import constants
from hades.common.cli import ArgumentParser, parser as common_parser
from hades.config.generate import ConfigGenerator
from hades.config.loader import load_config
logger = logging.getLogger()
def main():
parser = ArgumentParser(parents=[common_parser])
parser.add_argument(dest='source', metavar='SOURCE',
help="Template file name or template directory name")
parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?',
help="Destination file or directory (default is stdout"
"for files; required for directories)")
args = parser.parse_args()
config = load_config(args.config)
template_dir = constants.templatedir
generator = ConfigGenerator(template_dir, config)
source_path = os.path.join(template_dir, args.source)
if os.path.isdir(source_path):
generator.from_directory(args.source, args.destination)
elif os.path.isfile(source_path):
if args.destination is None:
generator.from_file(args.source, sys.stdout)
else:
with open(args.destination, 'w', encoding='utf-8') as f:
generator.from_file(args.source, f)
else:
logger.critical("No such file or directory {} in {}"
.format(args.source, template_dir))
return os.EX_NOINPUT
if __name__ == '__main__':
sys.exit(main())
|
# ... existing code ...
import logging
import os
# ... modified code ...
from hades.config.loader import load_config
logger = logging.getLogger()
...
else:
logger.critical("No such file or directory {} in {}"
.format(args.source, template_dir))
return os.EX_NOINPUT
# ... rest of the code ...
|
59d44ba76a9b2f98375fa2f893dabc0376de6f82
|
localeurl/models.py
|
localeurl/models.py
|
from django.conf import settings
from django.core import urlresolvers
from django.utils import translation
from localeurl import utils
def reverse(*args, **kwargs):
reverse_kwargs = kwargs.get('kwargs') or {}
locale = utils.supported_language(reverse_kwargs.pop(
'locale', translation.get_language()))
url = django_reverse(*args, **kwargs)
_, path = utils.strip_script_prefix(url)
return utils.locale_url(path, locale)
django_reverse = None
def patch_reverse():
"""
Monkey-patches the urlresolvers.reverse function. Will not patch twice.
"""
global django_reverse
if urlresolvers.reverse is not reverse:
django_reverse = urlresolvers.reverse
urlresolvers.reverse = reverse
if settings.USE_I18N:
patch_reverse()
|
from django.conf import settings
from django.core import urlresolvers
from django.utils import translation
from django.contrib.auth import views as auth_views
from localeurl import utils
def reverse(*args, **kwargs):
reverse_kwargs = kwargs.get('kwargs') or {}
locale = utils.supported_language(reverse_kwargs.pop(
'locale', translation.get_language()))
url = django_reverse(*args, **kwargs)
_, path = utils.strip_script_prefix(url)
return utils.locale_url(path, locale)
django_reverse = None
def patch_reverse():
"""
Monkey-patches the urlresolvers.reverse function. Will not patch twice.
"""
global django_reverse
if urlresolvers.reverse is not reverse:
django_reverse = urlresolvers.reverse
urlresolvers.reverse = reverse
def redirect_to_login(next, login_url, *args, **kwargs):
if not login_url:
login_url = settings.LOGIN_URL
login_url = utils.locale_url(login_url, translation.get_language())
return django_redirect_to_login(next, login_url, *args, **kwargs)
django_redirect_to_login = None
def patch_redirect_to_login():
"""
Monkey-patches the redirect_to_login function. Will not patch twice.
"""
global django_redirect_to_login
if auth_views.redirect_to_login is not redirect_to_login:
django_redirect_to_login = auth_views.redirect_to_login
auth_views.redirect_to_login = redirect_to_login
if settings.USE_I18N:
patch_reverse()
patch_redirect_to_login()
|
Patch redirect_to_login to maintain locale
|
Patch redirect_to_login to maintain locale
Signed-off-by: Simon Luijk <[email protected]>
|
Python
|
mit
|
simonluijk/django-localeurl
|
from django.conf import settings
from django.core import urlresolvers
from django.utils import translation
+ from django.contrib.auth import views as auth_views
from localeurl import utils
+
def reverse(*args, **kwargs):
reverse_kwargs = kwargs.get('kwargs') or {}
locale = utils.supported_language(reverse_kwargs.pop(
'locale', translation.get_language()))
url = django_reverse(*args, **kwargs)
_, path = utils.strip_script_prefix(url)
return utils.locale_url(path, locale)
django_reverse = None
def patch_reverse():
"""
Monkey-patches the urlresolvers.reverse function. Will not patch twice.
"""
global django_reverse
if urlresolvers.reverse is not reverse:
django_reverse = urlresolvers.reverse
urlresolvers.reverse = reverse
+
+ def redirect_to_login(next, login_url, *args, **kwargs):
+ if not login_url:
+ login_url = settings.LOGIN_URL
+ login_url = utils.locale_url(login_url, translation.get_language())
+ return django_redirect_to_login(next, login_url, *args, **kwargs)
+
+ django_redirect_to_login = None
+
+ def patch_redirect_to_login():
+ """
+ Monkey-patches the redirect_to_login function. Will not patch twice.
+ """
+ global django_redirect_to_login
+ if auth_views.redirect_to_login is not redirect_to_login:
+ django_redirect_to_login = auth_views.redirect_to_login
+ auth_views.redirect_to_login = redirect_to_login
+
+
if settings.USE_I18N:
patch_reverse()
+ patch_redirect_to_login()
|
Patch redirect_to_login to maintain locale
|
## Code Before:
from django.conf import settings
from django.core import urlresolvers
from django.utils import translation
from localeurl import utils
def reverse(*args, **kwargs):
reverse_kwargs = kwargs.get('kwargs') or {}
locale = utils.supported_language(reverse_kwargs.pop(
'locale', translation.get_language()))
url = django_reverse(*args, **kwargs)
_, path = utils.strip_script_prefix(url)
return utils.locale_url(path, locale)
django_reverse = None
def patch_reverse():
"""
Monkey-patches the urlresolvers.reverse function. Will not patch twice.
"""
global django_reverse
if urlresolvers.reverse is not reverse:
django_reverse = urlresolvers.reverse
urlresolvers.reverse = reverse
if settings.USE_I18N:
patch_reverse()
## Instruction:
Patch redirect_to_login to maintain locale
## Code After:
from django.conf import settings
from django.core import urlresolvers
from django.utils import translation
from django.contrib.auth import views as auth_views
from localeurl import utils
def reverse(*args, **kwargs):
reverse_kwargs = kwargs.get('kwargs') or {}
locale = utils.supported_language(reverse_kwargs.pop(
'locale', translation.get_language()))
url = django_reverse(*args, **kwargs)
_, path = utils.strip_script_prefix(url)
return utils.locale_url(path, locale)
django_reverse = None
def patch_reverse():
"""
Monkey-patches the urlresolvers.reverse function. Will not patch twice.
"""
global django_reverse
if urlresolvers.reverse is not reverse:
django_reverse = urlresolvers.reverse
urlresolvers.reverse = reverse
def redirect_to_login(next, login_url, *args, **kwargs):
if not login_url:
login_url = settings.LOGIN_URL
login_url = utils.locale_url(login_url, translation.get_language())
return django_redirect_to_login(next, login_url, *args, **kwargs)
django_redirect_to_login = None
def patch_redirect_to_login():
"""
Monkey-patches the redirect_to_login function. Will not patch twice.
"""
global django_redirect_to_login
if auth_views.redirect_to_login is not redirect_to_login:
django_redirect_to_login = auth_views.redirect_to_login
auth_views.redirect_to_login = redirect_to_login
if settings.USE_I18N:
patch_reverse()
patch_redirect_to_login()
|
...
from django.utils import translation
from django.contrib.auth import views as auth_views
from localeurl import utils
...
def redirect_to_login(next, login_url, *args, **kwargs):
if not login_url:
login_url = settings.LOGIN_URL
login_url = utils.locale_url(login_url, translation.get_language())
return django_redirect_to_login(next, login_url, *args, **kwargs)
django_redirect_to_login = None
def patch_redirect_to_login():
"""
Monkey-patches the redirect_to_login function. Will not patch twice.
"""
global django_redirect_to_login
if auth_views.redirect_to_login is not redirect_to_login:
django_redirect_to_login = auth_views.redirect_to_login
auth_views.redirect_to_login = redirect_to_login
if settings.USE_I18N:
...
patch_reverse()
patch_redirect_to_login()
...
|
9e73de0014b3f88b9e94ead11a878c6bc3819782
|
selenium_testcase/tests/test_navigation.py
|
selenium_testcase/tests/test_navigation.py
|
from __future__ import absolute_import
from ..testcases import SeleniumLiveTestCase
class NavigationTestCase(SeleniumLiveTestCase):
test_templates = [
(r'^nav_1/$', 'nav_1.html'),
(r'^nav_1/nav_2/$', 'nav_2.html')
]
def test_get_page(self):
""" Test that you can traverse the page tree. """
self.get_page("/nav_1/")
self.should_see("This is nav 1.")
self.get_page("/nav_1/nav_2/")
self.should_see("This is nav 2.")
|
from __future__ import absolute_import
from ..testcases import SeleniumLiveTestCase
class NavigationTestCase(SeleniumLiveTestCase):
test_templates = [
(r'^nav_1/$', 'nav_1.html'),
(r'^nav_1/nav_2/$', 'nav_2.html')
]
def test_get_page(self):
""" Test that you can traverse the page tree. """
self.get_page("/nav_1/")
self.has_title("Navigation 1")
self.title_contains("1")
self.should_see("This is nav 1.")
self.get_page("/nav_1/nav_2/")
self.should_see("This is nav 2.")
def test_get_bad_page(self):
""" Test that /bogus/ is not found. """
self.get_page("/bogus/")
self.should_see("Not Found")
self.should_see(
"The requested URL /bogus/ was not found on this server.")
def test_missing_content_with_retry(self):
""" Test retry for missing content, LONG RETRIES! """
self.get_page("/nav_1/")
self.should_not_see("This is nav 2.")
self.url_should_not_contain("nav_2")
self.has_not_title("Navigation 2")
self.title_does_not_contain("2")
self.assertRaises(
AssertionError, self.click_button, "not_there_dude")
self.not_at_page("/nav_2/")
|
Test missing content and failed navigation tests.
|
Test missing content and failed navigation tests.
This commit adds unit tests outside of the happy path where
a url does not exist or the test is looking for conten that
doesn't exist on the page. Since testing for missing informaion
requires timeouts to be sure, some of these tests take several
seconds to execute.
|
Python
|
bsd-3-clause
|
nimbis/django-selenium-testcase,nimbis/django-selenium-testcase
|
from __future__ import absolute_import
from ..testcases import SeleniumLiveTestCase
class NavigationTestCase(SeleniumLiveTestCase):
test_templates = [
(r'^nav_1/$', 'nav_1.html'),
(r'^nav_1/nav_2/$', 'nav_2.html')
]
def test_get_page(self):
""" Test that you can traverse the page tree. """
self.get_page("/nav_1/")
+ self.has_title("Navigation 1")
+ self.title_contains("1")
self.should_see("This is nav 1.")
self.get_page("/nav_1/nav_2/")
self.should_see("This is nav 2.")
+ def test_get_bad_page(self):
+ """ Test that /bogus/ is not found. """
+ self.get_page("/bogus/")
+ self.should_see("Not Found")
+ self.should_see(
+ "The requested URL /bogus/ was not found on this server.")
+
+ def test_missing_content_with_retry(self):
+ """ Test retry for missing content, LONG RETRIES! """
+ self.get_page("/nav_1/")
+ self.should_not_see("This is nav 2.")
+ self.url_should_not_contain("nav_2")
+ self.has_not_title("Navigation 2")
+ self.title_does_not_contain("2")
+ self.assertRaises(
+ AssertionError, self.click_button, "not_there_dude")
+ self.not_at_page("/nav_2/")
+
|
Test missing content and failed navigation tests.
|
## Code Before:
from __future__ import absolute_import
from ..testcases import SeleniumLiveTestCase
class NavigationTestCase(SeleniumLiveTestCase):
test_templates = [
(r'^nav_1/$', 'nav_1.html'),
(r'^nav_1/nav_2/$', 'nav_2.html')
]
def test_get_page(self):
""" Test that you can traverse the page tree. """
self.get_page("/nav_1/")
self.should_see("This is nav 1.")
self.get_page("/nav_1/nav_2/")
self.should_see("This is nav 2.")
## Instruction:
Test missing content and failed navigation tests.
## Code After:
from __future__ import absolute_import
from ..testcases import SeleniumLiveTestCase
class NavigationTestCase(SeleniumLiveTestCase):
test_templates = [
(r'^nav_1/$', 'nav_1.html'),
(r'^nav_1/nav_2/$', 'nav_2.html')
]
def test_get_page(self):
""" Test that you can traverse the page tree. """
self.get_page("/nav_1/")
self.has_title("Navigation 1")
self.title_contains("1")
self.should_see("This is nav 1.")
self.get_page("/nav_1/nav_2/")
self.should_see("This is nav 2.")
def test_get_bad_page(self):
""" Test that /bogus/ is not found. """
self.get_page("/bogus/")
self.should_see("Not Found")
self.should_see(
"The requested URL /bogus/ was not found on this server.")
def test_missing_content_with_retry(self):
""" Test retry for missing content, LONG RETRIES! """
self.get_page("/nav_1/")
self.should_not_see("This is nav 2.")
self.url_should_not_contain("nav_2")
self.has_not_title("Navigation 2")
self.title_does_not_contain("2")
self.assertRaises(
AssertionError, self.click_button, "not_there_dude")
self.not_at_page("/nav_2/")
|
...
self.get_page("/nav_1/")
self.has_title("Navigation 1")
self.title_contains("1")
self.should_see("This is nav 1.")
...
self.should_see("This is nav 2.")
def test_get_bad_page(self):
""" Test that /bogus/ is not found. """
self.get_page("/bogus/")
self.should_see("Not Found")
self.should_see(
"The requested URL /bogus/ was not found on this server.")
def test_missing_content_with_retry(self):
""" Test retry for missing content, LONG RETRIES! """
self.get_page("/nav_1/")
self.should_not_see("This is nav 2.")
self.url_should_not_contain("nav_2")
self.has_not_title("Navigation 2")
self.title_does_not_contain("2")
self.assertRaises(
AssertionError, self.click_button, "not_there_dude")
self.not_at_page("/nav_2/")
...
|
8d0b9da511d55191609ffbd88a8b11afd6ff0367
|
remedy/radremedy.py
|
remedy/radremedy.py
|
from flask import Flask, url_for, request, abort
from flask.ext.script import Manager
from flask.ext.migrate import Migrate, MigrateCommand
from rad.models import db, Resource
def create_app(config, models=()):
from remedyblueprint import remedy, url_for_other_page
app = Flask(__name__)
app.config.from_object(config)
app.register_blueprint(remedy)
# searching configurations
app.jinja_env.trim_blocks = True
# Register the paging helper method with Jinja2
app.jinja_env.globals['url_for_other_page'] = url_for_other_page
db.init_app(app)
Migrate(app, db, directory=app.config['MIGRATIONS_DIR'])
manager = Manager(app)
manager.add_command('db', MigrateCommand)
# turning API off for now
# from api_manager import init_api_manager
# api_manager = init_api_manager(app, db)
# map(lambda m: api_manager.create_api(m), models)
return app, manager
if __name__ == '__main__':
app, manager = create_app('config.BaseConfig', (Resource, ))
with app.app_context():
manager.run()
|
from flask import Flask, url_for, request, abort
from flask.ext.script import Manager
from flask.ext.migrate import Migrate, MigrateCommand
from flask.ext.login import current_user
from rad.models import db, Resource
def create_app(config, models=()):
app = Flask(__name__)
app.config.from_object(config)
from remedyblueprint import remedy, url_for_other_page
app.register_blueprint(remedy)
from auth.user_auth import auth, login_manager
app.register_blueprint(auth)
login_manager.init_app(app)
# searching configurations
app.jinja_env.trim_blocks = True
# Register the paging helper method with Jinja2
app.jinja_env.globals['url_for_other_page'] = url_for_other_page
app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous()
db.init_app(app)
Migrate(app, db, directory=app.config['MIGRATIONS_DIR'])
manager = Manager(app)
manager.add_command('db', MigrateCommand)
# turning API off for now
# from api_manager import init_api_manager
# api_manager = init_api_manager(app, db)
# map(lambda m: api_manager.create_api(m), models)
return app, manager
if __name__ == '__main__':
application, manager = create_app('config.BaseConfig', (Resource, ))
with application.app_context():
manager.run()
|
Move around imports and not shadow app
|
Move around imports and not shadow app
|
Python
|
mpl-2.0
|
radioprotector/radremedy,radioprotector/radremedy,radioprotector/radremedy,AllieDeford/radremedy,AllieDeford/radremedy,radremedy/radremedy,radremedy/radremedy,radremedy/radremedy,AllieDeford/radremedy,radioprotector/radremedy,radremedy/radremedy
|
from flask import Flask, url_for, request, abort
from flask.ext.script import Manager
from flask.ext.migrate import Migrate, MigrateCommand
+ from flask.ext.login import current_user
from rad.models import db, Resource
def create_app(config, models=()):
- from remedyblueprint import remedy, url_for_other_page
-
app = Flask(__name__)
app.config.from_object(config)
+ from remedyblueprint import remedy, url_for_other_page
app.register_blueprint(remedy)
+
+ from auth.user_auth import auth, login_manager
+ app.register_blueprint(auth)
+ login_manager.init_app(app)
# searching configurations
app.jinja_env.trim_blocks = True
# Register the paging helper method with Jinja2
app.jinja_env.globals['url_for_other_page'] = url_for_other_page
+ app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous()
db.init_app(app)
Migrate(app, db, directory=app.config['MIGRATIONS_DIR'])
manager = Manager(app)
manager.add_command('db', MigrateCommand)
# turning API off for now
# from api_manager import init_api_manager
# api_manager = init_api_manager(app, db)
# map(lambda m: api_manager.create_api(m), models)
return app, manager
if __name__ == '__main__':
- app, manager = create_app('config.BaseConfig', (Resource, ))
+ application, manager = create_app('config.BaseConfig', (Resource, ))
- with app.app_context():
+ with application.app_context():
manager.run()
|
Move around imports and not shadow app
|
## Code Before:
from flask import Flask, url_for, request, abort
from flask.ext.script import Manager
from flask.ext.migrate import Migrate, MigrateCommand
from rad.models import db, Resource
def create_app(config, models=()):
from remedyblueprint import remedy, url_for_other_page
app = Flask(__name__)
app.config.from_object(config)
app.register_blueprint(remedy)
# searching configurations
app.jinja_env.trim_blocks = True
# Register the paging helper method with Jinja2
app.jinja_env.globals['url_for_other_page'] = url_for_other_page
db.init_app(app)
Migrate(app, db, directory=app.config['MIGRATIONS_DIR'])
manager = Manager(app)
manager.add_command('db', MigrateCommand)
# turning API off for now
# from api_manager import init_api_manager
# api_manager = init_api_manager(app, db)
# map(lambda m: api_manager.create_api(m), models)
return app, manager
if __name__ == '__main__':
app, manager = create_app('config.BaseConfig', (Resource, ))
with app.app_context():
manager.run()
## Instruction:
Move around imports and not shadow app
## Code After:
from flask import Flask, url_for, request, abort
from flask.ext.script import Manager
from flask.ext.migrate import Migrate, MigrateCommand
from flask.ext.login import current_user
from rad.models import db, Resource
def create_app(config, models=()):
app = Flask(__name__)
app.config.from_object(config)
from remedyblueprint import remedy, url_for_other_page
app.register_blueprint(remedy)
from auth.user_auth import auth, login_manager
app.register_blueprint(auth)
login_manager.init_app(app)
# searching configurations
app.jinja_env.trim_blocks = True
# Register the paging helper method with Jinja2
app.jinja_env.globals['url_for_other_page'] = url_for_other_page
app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous()
db.init_app(app)
Migrate(app, db, directory=app.config['MIGRATIONS_DIR'])
manager = Manager(app)
manager.add_command('db', MigrateCommand)
# turning API off for now
# from api_manager import init_api_manager
# api_manager = init_api_manager(app, db)
# map(lambda m: api_manager.create_api(m), models)
return app, manager
if __name__ == '__main__':
application, manager = create_app('config.BaseConfig', (Resource, ))
with application.app_context():
manager.run()
|
...
from flask.ext.migrate import Migrate, MigrateCommand
from flask.ext.login import current_user
from rad.models import db, Resource
...
app = Flask(__name__)
...
from remedyblueprint import remedy, url_for_other_page
app.register_blueprint(remedy)
from auth.user_auth import auth, login_manager
app.register_blueprint(auth)
login_manager.init_app(app)
...
app.jinja_env.globals['url_for_other_page'] = url_for_other_page
app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous()
...
if __name__ == '__main__':
application, manager = create_app('config.BaseConfig', (Resource, ))
with application.app_context():
manager.run()
...
|
ccf60e9e79b8b2db8cbf7918caf23314e8790134
|
lib/reporter.py
|
lib/reporter.py
|
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s>" % name)
print("\t<status=\"%s\" />" % status)
if status != "SKIP":
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
Fix the XML format produced
|
Fix the XML format produced
|
Python
|
apache-2.0
|
CESNET/secant,CESNET/secant
|
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
- print("<%s>" % name)
- print("\t<status=\"%s\" />" % status)
+ print("<%s status=\"%s\">" % (name, status))
- if status != "SKIP":
- print("\t<outcome>")
+ print("\t<outcome>")
- for line in sys.stdin:
+ for line in sys.stdin:
- # Escaping, ... !
+ # Escaping, ... !
- print(line.rstrip(os.linesep))
+ print(line.rstrip(os.linesep))
- print("\t</outcome>")
+ print("\t</outcome>")
print("</%s>" % name)
|
Fix the XML format produced
|
## Code Before:
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s>" % name)
print("\t<status=\"%s\" />" % status)
if status != "SKIP":
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
## Instruction:
Fix the XML format produced
## Code After:
import sys
import os
name = sys.argv[1]
status = sys.stdin.readline()
status = status.rstrip(os.linesep)
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
print("</%s>" % name)
|
...
print("<%s status=\"%s\">" % (name, status))
print("\t<outcome>")
for line in sys.stdin:
# Escaping, ... !
print(line.rstrip(os.linesep))
print("\t</outcome>")
...
|
2e361627ca94b3a3b1cdd9583d22ca8ff81a0591
|
rpn/util.py
|
rpn/util.py
|
from functools import wraps
import subprocess
_SELECTIONS = {
'+': 'clipboard',
'*': 'primary',
}
def _store_selection(data, selection):
with subprocess.Popen(['xclip',
'-selection', selection],
stdin=subprocess.PIPE) as xclip:
xclip.stdin.write(str(data).encode())
def _load_selection(selection):
with subprocess.Popen(['xclip',
'-selection', selection,
'-o'], stdout=PIPE) as xclip:
return xclip.stdout.read().decode()
class RPNError(Exception):
pass
def wrap_user_errors(fmt):
'''
Ugly hack decorator that converts exceptions to warnings.
Passes through RPNErrors.
'''
def decorator(f):
@wraps(f)
def wrapper(*args, **kwargs):
try:
return f(*args, **kwargs)
except RPNError:
raise
except Exception as e:
raise RPNError(fmt.format(*args, **kwargs), e)
return wrapper
return decorator
|
from functools import wraps
import subprocess
_SELECTIONS = {
'+': 'clipboard',
'*': 'primary',
}
def _store_selection(data, selection):
with subprocess.Popen(['xclip',
'-selection', selection],
stdin=subprocess.PIPE) as xclip:
xclip.stdin.write(str(data).encode())
def _load_selection(selection):
with subprocess.Popen(['xclip',
'-selection', selection,
'-o'], stdout=subprocess.PIPE) as xclip:
return xclip.stdout.read().decode()
class RPNError(Exception):
pass
def wrap_user_errors(fmt):
'''
Ugly hack decorator that converts exceptions to warnings.
Passes through RPNErrors.
'''
def decorator(f):
@wraps(f)
def wrapper(*args, **kwargs):
try:
return f(*args, **kwargs)
except RPNError:
raise
except Exception as e:
raise RPNError(fmt.format(*args, **kwargs), e)
return wrapper
return decorator
|
Fix typo resulting in NameError
|
Fix typo resulting in NameError
|
Python
|
isc
|
pilona/RPN,pilona/RPN
|
from functools import wraps
import subprocess
_SELECTIONS = {
'+': 'clipboard',
'*': 'primary',
}
def _store_selection(data, selection):
with subprocess.Popen(['xclip',
'-selection', selection],
stdin=subprocess.PIPE) as xclip:
xclip.stdin.write(str(data).encode())
def _load_selection(selection):
with subprocess.Popen(['xclip',
'-selection', selection,
- '-o'], stdout=PIPE) as xclip:
+ '-o'], stdout=subprocess.PIPE) as xclip:
return xclip.stdout.read().decode()
class RPNError(Exception):
pass
def wrap_user_errors(fmt):
'''
Ugly hack decorator that converts exceptions to warnings.
Passes through RPNErrors.
'''
def decorator(f):
@wraps(f)
def wrapper(*args, **kwargs):
try:
return f(*args, **kwargs)
except RPNError:
raise
except Exception as e:
raise RPNError(fmt.format(*args, **kwargs), e)
return wrapper
return decorator
|
Fix typo resulting in NameError
|
## Code Before:
from functools import wraps
import subprocess
_SELECTIONS = {
'+': 'clipboard',
'*': 'primary',
}
def _store_selection(data, selection):
with subprocess.Popen(['xclip',
'-selection', selection],
stdin=subprocess.PIPE) as xclip:
xclip.stdin.write(str(data).encode())
def _load_selection(selection):
with subprocess.Popen(['xclip',
'-selection', selection,
'-o'], stdout=PIPE) as xclip:
return xclip.stdout.read().decode()
class RPNError(Exception):
pass
def wrap_user_errors(fmt):
'''
Ugly hack decorator that converts exceptions to warnings.
Passes through RPNErrors.
'''
def decorator(f):
@wraps(f)
def wrapper(*args, **kwargs):
try:
return f(*args, **kwargs)
except RPNError:
raise
except Exception as e:
raise RPNError(fmt.format(*args, **kwargs), e)
return wrapper
return decorator
## Instruction:
Fix typo resulting in NameError
## Code After:
from functools import wraps
import subprocess
_SELECTIONS = {
'+': 'clipboard',
'*': 'primary',
}
def _store_selection(data, selection):
with subprocess.Popen(['xclip',
'-selection', selection],
stdin=subprocess.PIPE) as xclip:
xclip.stdin.write(str(data).encode())
def _load_selection(selection):
with subprocess.Popen(['xclip',
'-selection', selection,
'-o'], stdout=subprocess.PIPE) as xclip:
return xclip.stdout.read().decode()
class RPNError(Exception):
pass
def wrap_user_errors(fmt):
'''
Ugly hack decorator that converts exceptions to warnings.
Passes through RPNErrors.
'''
def decorator(f):
@wraps(f)
def wrapper(*args, **kwargs):
try:
return f(*args, **kwargs)
except RPNError:
raise
except Exception as e:
raise RPNError(fmt.format(*args, **kwargs), e)
return wrapper
return decorator
|
# ... existing code ...
'-selection', selection,
'-o'], stdout=subprocess.PIPE) as xclip:
return xclip.stdout.read().decode()
# ... rest of the code ...
|
b4687eb7fda33323cad8d42f9819a3ee223d3822
|
web/config/local_settings.py
|
web/config/local_settings.py
|
import os
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
if os.getenv("MEMCACHE_HOSTS"):
CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
import os
import json, requests
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"):
rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME")
r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json()
r = map(lambda x: x["primary_ip"] + ":80", r)
CLUSTER_SERVERS = [str(x) for x in r]
if os.getenv("MEMCACHE_HOSTS"):
CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
Add graphite cluster discovery support using rancher
|
Add graphite cluster discovery support using rancher
|
Python
|
apache-2.0
|
Banno/graphite-setup,Banno/graphite-setup,Banno/graphite-setup
|
import os
+ import json, requests
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
+ elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"):
+ rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME")
+ r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json()
+ r = map(lambda x: x["primary_ip"] + ":80", r)
+ CLUSTER_SERVERS = [str(x) for x in r]
if os.getenv("MEMCACHE_HOSTS"):
CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
Add graphite cluster discovery support using rancher
|
## Code Before:
import os
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
if os.getenv("MEMCACHE_HOSTS"):
CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
## Instruction:
Add graphite cluster discovery support using rancher
## Code After:
import os
import json, requests
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"):
rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME")
r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json()
r = map(lambda x: x["primary_ip"] + ":80", r)
CLUSTER_SERVERS = [str(x) for x in r]
if os.getenv("MEMCACHE_HOSTS"):
CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
...
import os
import json, requests
from datetime import datetime
...
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"):
rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME")
r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json()
r = map(lambda x: x["primary_ip"] + ":80", r)
CLUSTER_SERVERS = [str(x) for x in r]
...
|
3af22fd5583ee110f731b9e1ebecba67ebee2bd4
|
sendwithus/exceptions.py
|
sendwithus/exceptions.py
|
class SendwithusError(Exception):
"""Base class for Sendwithus API errors"""
class AuthenticationError(SendwithusError):
"""API Authentication Failed"""
class APIError(SendwithusError):
"""4xx - Invalid Request (Client error)"""
class ServerError(SendwithusError):
"""5xx - Failed Request (Server error)"""
|
class SendwithusError(Exception):
"""Base class for Sendwithus API errors"""
def __init__(self, content=None):
self.content = content
class AuthenticationError(SendwithusError):
"""API Authentication Failed"""
class APIError(SendwithusError):
"""4xx - Invalid Request (Client error)"""
class ServerError(SendwithusError):
"""5xx - Failed Request (Server error)"""
|
Add a constructor to SendwithusError that stores content
|
Add a constructor to SendwithusError that stores content
|
Python
|
apache-2.0
|
sendwithus/sendwithus_python
|
class SendwithusError(Exception):
"""Base class for Sendwithus API errors"""
+
+ def __init__(self, content=None):
+ self.content = content
class AuthenticationError(SendwithusError):
"""API Authentication Failed"""
class APIError(SendwithusError):
"""4xx - Invalid Request (Client error)"""
class ServerError(SendwithusError):
"""5xx - Failed Request (Server error)"""
|
Add a constructor to SendwithusError that stores content
|
## Code Before:
class SendwithusError(Exception):
"""Base class for Sendwithus API errors"""
class AuthenticationError(SendwithusError):
"""API Authentication Failed"""
class APIError(SendwithusError):
"""4xx - Invalid Request (Client error)"""
class ServerError(SendwithusError):
"""5xx - Failed Request (Server error)"""
## Instruction:
Add a constructor to SendwithusError that stores content
## Code After:
class SendwithusError(Exception):
"""Base class for Sendwithus API errors"""
def __init__(self, content=None):
self.content = content
class AuthenticationError(SendwithusError):
"""API Authentication Failed"""
class APIError(SendwithusError):
"""4xx - Invalid Request (Client error)"""
class ServerError(SendwithusError):
"""5xx - Failed Request (Server error)"""
|
...
"""Base class for Sendwithus API errors"""
def __init__(self, content=None):
self.content = content
...
|
62451e8c5b3d93409fa4bcc7ec29827be6253e88
|
website/registries/utils.py
|
website/registries/utils.py
|
REG_CAMPAIGNS = {
'prereg': 'Prereg Challenge',
'registered_report': 'Registered Report Protocol Preregistration',
}
def get_campaign_schema(campaign):
from osf.models import RegistrationSchema
if campaign not in REG_CAMPAIGNS:
raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys())))
schema_name = REG_CAMPAIGNS[campaign]
return RegistrationSchema.objects.get(name=schema_name, schema_version=2)
def drafts_for_user(user, campaign=None):
from osf.models import DraftRegistration, Node
from guardian.shortcuts import get_objects_for_user
if not user or user.is_anonymous:
return None
node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True)
if campaign:
drafts = DraftRegistration.objects.filter(
registration_schema=get_campaign_schema(campaign),
approval=None,
registered_node=None,
deleted__isnull=True,
branched_from__in=list(node_qs),
initiator=user
)
else:
drafts = DraftRegistration.objects.filter(
approval=None,
registered_node=None,
deleted__isnull=True,
branched_from__in=list(node_qs),
initiator=user
)
return drafts
|
REG_CAMPAIGNS = {
'prereg': 'Prereg Challenge',
'registered_report': 'Registered Report Protocol Preregistration',
}
def get_campaign_schema(campaign):
from osf.models import RegistrationSchema
if campaign not in REG_CAMPAIGNS:
raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys())))
schema_name = REG_CAMPAIGNS[campaign]
return RegistrationSchema.objects.get(name=schema_name, schema_version=2)
def drafts_for_user(user, campaign=None):
from osf.models import DraftRegistration, Node
from guardian.shortcuts import get_objects_for_user
if not user or user.is_anonymous:
return None
node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True)
drafts = DraftRegistration.objects.filter(
approval=None,
registered_node=None,
deleted__isnull=True,
branched_from__in=node_qs,
initiator=user
)
if campaign:
drafts = drafts.filter(
registration_schema=get_campaign_schema(campaign),
)
return drafts
|
Speed up draft registrations query.
|
Speed up draft registrations query.
|
Python
|
apache-2.0
|
baylee-d/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,adlius/osf.io,mattclark/osf.io,felliott/osf.io,saradbowman/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,felliott/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,baylee-d/osf.io,mattclark/osf.io,brianjgeiger/osf.io,mattclark/osf.io,Johnetordoff/osf.io,brianjgeiger/osf.io,brianjgeiger/osf.io,baylee-d/osf.io,adlius/osf.io,cslzchen/osf.io,aaxelb/osf.io,aaxelb/osf.io,felliott/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,mfraezz/osf.io,mfraezz/osf.io,adlius/osf.io,CenterForOpenScience/osf.io,aaxelb/osf.io,aaxelb/osf.io,mfraezz/osf.io,felliott/osf.io,saradbowman/osf.io,CenterForOpenScience/osf.io,adlius/osf.io
|
REG_CAMPAIGNS = {
'prereg': 'Prereg Challenge',
'registered_report': 'Registered Report Protocol Preregistration',
}
def get_campaign_schema(campaign):
from osf.models import RegistrationSchema
if campaign not in REG_CAMPAIGNS:
raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys())))
schema_name = REG_CAMPAIGNS[campaign]
return RegistrationSchema.objects.get(name=schema_name, schema_version=2)
def drafts_for_user(user, campaign=None):
from osf.models import DraftRegistration, Node
from guardian.shortcuts import get_objects_for_user
if not user or user.is_anonymous:
return None
- node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True)
+ node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True)
+
+ drafts = DraftRegistration.objects.filter(
+ approval=None,
+ registered_node=None,
+ deleted__isnull=True,
+ branched_from__in=node_qs,
+ initiator=user
+ )
if campaign:
- drafts = DraftRegistration.objects.filter(
+ drafts = drafts.filter(
registration_schema=get_campaign_schema(campaign),
- approval=None,
- registered_node=None,
- deleted__isnull=True,
- branched_from__in=list(node_qs),
- initiator=user
)
+
- else:
- drafts = DraftRegistration.objects.filter(
- approval=None,
- registered_node=None,
- deleted__isnull=True,
- branched_from__in=list(node_qs),
- initiator=user
- )
return drafts
|
Speed up draft registrations query.
|
## Code Before:
REG_CAMPAIGNS = {
'prereg': 'Prereg Challenge',
'registered_report': 'Registered Report Protocol Preregistration',
}
def get_campaign_schema(campaign):
from osf.models import RegistrationSchema
if campaign not in REG_CAMPAIGNS:
raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys())))
schema_name = REG_CAMPAIGNS[campaign]
return RegistrationSchema.objects.get(name=schema_name, schema_version=2)
def drafts_for_user(user, campaign=None):
from osf.models import DraftRegistration, Node
from guardian.shortcuts import get_objects_for_user
if not user or user.is_anonymous:
return None
node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True)
if campaign:
drafts = DraftRegistration.objects.filter(
registration_schema=get_campaign_schema(campaign),
approval=None,
registered_node=None,
deleted__isnull=True,
branched_from__in=list(node_qs),
initiator=user
)
else:
drafts = DraftRegistration.objects.filter(
approval=None,
registered_node=None,
deleted__isnull=True,
branched_from__in=list(node_qs),
initiator=user
)
return drafts
## Instruction:
Speed up draft registrations query.
## Code After:
REG_CAMPAIGNS = {
'prereg': 'Prereg Challenge',
'registered_report': 'Registered Report Protocol Preregistration',
}
def get_campaign_schema(campaign):
from osf.models import RegistrationSchema
if campaign not in REG_CAMPAIGNS:
raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys())))
schema_name = REG_CAMPAIGNS[campaign]
return RegistrationSchema.objects.get(name=schema_name, schema_version=2)
def drafts_for_user(user, campaign=None):
from osf.models import DraftRegistration, Node
from guardian.shortcuts import get_objects_for_user
if not user or user.is_anonymous:
return None
node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True)
drafts = DraftRegistration.objects.filter(
approval=None,
registered_node=None,
deleted__isnull=True,
branched_from__in=node_qs,
initiator=user
)
if campaign:
drafts = drafts.filter(
registration_schema=get_campaign_schema(campaign),
)
return drafts
|
...
node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True)
drafts = DraftRegistration.objects.filter(
approval=None,
registered_node=None,
deleted__isnull=True,
branched_from__in=node_qs,
initiator=user
)
...
if campaign:
drafts = drafts.filter(
registration_schema=get_campaign_schema(campaign),
)
return drafts
...
|
8a9f707960c3b39488c9bbee6ce7f22c6fbfc853
|
web/config/local_settings.py
|
web/config/local_settings.py
|
import os
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
if os.getenv("MEMCACHE_HOSTS"):
CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
import os
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
if os.getenv("MEMCACHE_HOSTS"):
MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
Fix memcache hosts setting from env
|
Fix memcache hosts setting from env
Before this fix if one had set OS env vars for both CLUSTER_SERVERS and
MEMCACHE_HOSTS the value of later would override the former and the
graphite web application fails to show any metrics.
|
Python
|
apache-2.0
|
Banno/graphite-setup,Banno/graphite-setup,Banno/graphite-setup
|
import os
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
if os.getenv("MEMCACHE_HOSTS"):
- CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
+ MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
Fix memcache hosts setting from env
|
## Code Before:
import os
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
if os.getenv("MEMCACHE_HOSTS"):
CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
## Instruction:
Fix memcache hosts setting from env
## Code After:
import os
from datetime import datetime
LOG_DIR = '/var/log/graphite'
if os.getenv("CARBONLINK_HOSTS"):
CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',')
if os.getenv("CLUSTER_SERVERS"):
CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',')
if os.getenv("MEMCACHE_HOSTS"):
MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',')
if os.getenv("WHISPER_DIR"):
WHISPER_DIR = os.getenv("WHISPER_DIR")
SECRET_KEY = str(datetime.now())
|
# ... existing code ...
if os.getenv("MEMCACHE_HOSTS"):
MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',')
# ... rest of the code ...
|
26ffa0cdd1389e2a364531cd20e9f37ee1565cce
|
base/view_utils.py
|
base/view_utils.py
|
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger
# standard library
def paginate(request, objects, page_size=25):
paginator = Paginator(objects, page_size)
page = request.GET.get('p')
try:
paginated_objects = paginator.page(page)
except PageNotAnInteger:
# If page is not an integer, deliver first page.
paginated_objects = paginator.page(1)
except EmptyPage:
# If page is out of range (e.g. 9999), deliver last page of results.
paginated_objects = paginator.page(paginator.num_pages)
return paginated_objects
def clean_query_string(request):
clean_query_set = request.GET.copy()
clean_query_set = dict(
(k, v) for k, v in request.GET.items() if not k.startswith('o')
)
try:
del clean_query_set['p']
except:
pass
mstring = []
for key in clean_query_set.keys():
valuelist = request.GET.getlist(key)
mstring.extend(['%s=%s' % (key, val) for val in valuelist])
return '&'.join(mstring)
|
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger
# standard library
def paginate(request, objects, page_size=25):
paginator = Paginator(objects, page_size)
page = request.GET.get('p')
try:
paginated_objects = paginator.page(page)
except PageNotAnInteger:
# If page is not an integer, deliver first page.
paginated_objects = paginator.page(1)
except EmptyPage:
# If page is out of range (e.g. 9999), deliver last page of results.
paginated_objects = paginator.page(paginator.num_pages)
return paginated_objects
def clean_query_string(request):
clean_query_set = request.GET.copy()
clean_query_set = dict(
(k, v) for k, v in request.GET.items() if k != 'o'
)
try:
del clean_query_set['p']
except:
pass
mstring = []
for key in clean_query_set.keys():
valuelist = request.GET.getlist(key)
mstring.extend(['%s=%s' % (key, val) for val in valuelist])
return '&'.join(mstring)
|
Use 'o' as the order by parameter in clean_query_string
|
Use 'o' as the order by parameter in clean_query_string
|
Python
|
mit
|
magnet-cl/django-project-template-py3,Angoreher/xcero,Angoreher/xcero,magnet-cl/django-project-template-py3,magnet-cl/django-project-template-py3,magnet-cl/django-project-template-py3,Angoreher/xcero,Angoreher/xcero
|
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger
# standard library
def paginate(request, objects, page_size=25):
paginator = Paginator(objects, page_size)
page = request.GET.get('p')
try:
paginated_objects = paginator.page(page)
except PageNotAnInteger:
# If page is not an integer, deliver first page.
paginated_objects = paginator.page(1)
except EmptyPage:
# If page is out of range (e.g. 9999), deliver last page of results.
paginated_objects = paginator.page(paginator.num_pages)
return paginated_objects
def clean_query_string(request):
clean_query_set = request.GET.copy()
clean_query_set = dict(
- (k, v) for k, v in request.GET.items() if not k.startswith('o')
+ (k, v) for k, v in request.GET.items() if k != 'o'
)
try:
del clean_query_set['p']
except:
pass
mstring = []
for key in clean_query_set.keys():
valuelist = request.GET.getlist(key)
mstring.extend(['%s=%s' % (key, val) for val in valuelist])
return '&'.join(mstring)
|
Use 'o' as the order by parameter in clean_query_string
|
## Code Before:
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger
# standard library
def paginate(request, objects, page_size=25):
paginator = Paginator(objects, page_size)
page = request.GET.get('p')
try:
paginated_objects = paginator.page(page)
except PageNotAnInteger:
# If page is not an integer, deliver first page.
paginated_objects = paginator.page(1)
except EmptyPage:
# If page is out of range (e.g. 9999), deliver last page of results.
paginated_objects = paginator.page(paginator.num_pages)
return paginated_objects
def clean_query_string(request):
clean_query_set = request.GET.copy()
clean_query_set = dict(
(k, v) for k, v in request.GET.items() if not k.startswith('o')
)
try:
del clean_query_set['p']
except:
pass
mstring = []
for key in clean_query_set.keys():
valuelist = request.GET.getlist(key)
mstring.extend(['%s=%s' % (key, val) for val in valuelist])
return '&'.join(mstring)
## Instruction:
Use 'o' as the order by parameter in clean_query_string
## Code After:
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger
# standard library
def paginate(request, objects, page_size=25):
paginator = Paginator(objects, page_size)
page = request.GET.get('p')
try:
paginated_objects = paginator.page(page)
except PageNotAnInteger:
# If page is not an integer, deliver first page.
paginated_objects = paginator.page(1)
except EmptyPage:
# If page is out of range (e.g. 9999), deliver last page of results.
paginated_objects = paginator.page(paginator.num_pages)
return paginated_objects
def clean_query_string(request):
clean_query_set = request.GET.copy()
clean_query_set = dict(
(k, v) for k, v in request.GET.items() if k != 'o'
)
try:
del clean_query_set['p']
except:
pass
mstring = []
for key in clean_query_set.keys():
valuelist = request.GET.getlist(key)
mstring.extend(['%s=%s' % (key, val) for val in valuelist])
return '&'.join(mstring)
|
...
clean_query_set = dict(
(k, v) for k, v in request.GET.items() if k != 'o'
)
...
|
3af265ab0740378267a3c3e9cc85bb21468bf2e0
|
engine/cli.py
|
engine/cli.py
|
from engine.event import *
from engine.action import *
from engine.code import *
from engine.player import *
from engine.round import *
from engine.team import *
def processInput():
userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n")
if 'f' in userText:
jailCode = input("enter jail code: ")
Action.fleePlayerWithCode(jailCode)
Stats.printPlayersDetailed()
elif 's' in userText:
mobile = input("enter mobile: ")
code = input("enter code: ")
Action.handleSms(mobile, code)
Stats.printPlayersDetailed()
elif 'a' in userText:
name = input("enter name: ")
mobile = input("enter mobile: ")
#email = input("enter email: ")
Action.addPlayer(name, mobile, "")
Stats.printPlayersDetailed()
elif 'w' in userText:
hash = input("enter player hash: ")
code = input("enter code: ")
Action.handleWeb(hash, code)
Stats.printPlayersDetailed()
elif 't' in userText:
name = input("enter player name: ")
team = input("enter team name: ")
Action.addPlayerToTeam(name, team)
Stats.printPlayersDetailed()
elif 'p' in userText:
Stats.printStats()
elif 'c' in userText:
name = input("enter name: ")
message = input("enter text: ")
playerId = Player._getIdByName(name)
Action.sayToMyTeam(playerId, message)
|
from engine.action import Action, Stats
from engine.player import Player
def processInput():
userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n")
if 'f' in userText:
jailCode = input("enter jail code: ")
Action.fleePlayerWithCode(jailCode)
Stats.printPlayersDetailed()
elif 's' in userText:
mobile = input("enter mobile: ")
code = input("enter code: ")
Action.handleSms(mobile, code)
Stats.printPlayersDetailed()
elif 'a' in userText:
name = input("enter name: ")
mobile = input("enter mobile: ")
#email = input("enter email: ")
Action.addPlayer(name, mobile, "")
Stats.printPlayersDetailed()
elif 'w' in userText:
hash = input("enter player hash: ")
code = input("enter code: ")
Action.handleWeb(hash, code)
Stats.printPlayersDetailed()
elif 't' in userText:
name = input("enter player name: ")
team = input("enter team name: ")
Action.addPlayerToTeam(name, team)
Stats.printPlayersDetailed()
elif 'p' in userText:
Stats.printStats()
elif 'c' in userText:
name = input("enter name: ")
message = input("enter text: ")
playerId = Player._getIdByName(name)
Action.sayToMyTeam(playerId, message)
|
Remove a few unnecessary imports
|
Remove a few unnecessary imports
|
Python
|
bsd-2-clause
|
mahfiaz/spotter_irl,mahfiaz/spotter_irl,mahfiaz/spotter_irl
|
- from engine.event import *
- from engine.action import *
+ from engine.action import Action, Stats
- from engine.code import *
- from engine.player import *
+ from engine.player import Player
- from engine.round import *
- from engine.team import *
-
def processInput():
userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n")
if 'f' in userText:
jailCode = input("enter jail code: ")
Action.fleePlayerWithCode(jailCode)
Stats.printPlayersDetailed()
elif 's' in userText:
mobile = input("enter mobile: ")
code = input("enter code: ")
Action.handleSms(mobile, code)
Stats.printPlayersDetailed()
elif 'a' in userText:
name = input("enter name: ")
mobile = input("enter mobile: ")
#email = input("enter email: ")
Action.addPlayer(name, mobile, "")
Stats.printPlayersDetailed()
elif 'w' in userText:
hash = input("enter player hash: ")
code = input("enter code: ")
Action.handleWeb(hash, code)
Stats.printPlayersDetailed()
elif 't' in userText:
name = input("enter player name: ")
team = input("enter team name: ")
Action.addPlayerToTeam(name, team)
Stats.printPlayersDetailed()
elif 'p' in userText:
Stats.printStats()
elif 'c' in userText:
name = input("enter name: ")
message = input("enter text: ")
playerId = Player._getIdByName(name)
Action.sayToMyTeam(playerId, message)
-
|
Remove a few unnecessary imports
|
## Code Before:
from engine.event import *
from engine.action import *
from engine.code import *
from engine.player import *
from engine.round import *
from engine.team import *
def processInput():
userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n")
if 'f' in userText:
jailCode = input("enter jail code: ")
Action.fleePlayerWithCode(jailCode)
Stats.printPlayersDetailed()
elif 's' in userText:
mobile = input("enter mobile: ")
code = input("enter code: ")
Action.handleSms(mobile, code)
Stats.printPlayersDetailed()
elif 'a' in userText:
name = input("enter name: ")
mobile = input("enter mobile: ")
#email = input("enter email: ")
Action.addPlayer(name, mobile, "")
Stats.printPlayersDetailed()
elif 'w' in userText:
hash = input("enter player hash: ")
code = input("enter code: ")
Action.handleWeb(hash, code)
Stats.printPlayersDetailed()
elif 't' in userText:
name = input("enter player name: ")
team = input("enter team name: ")
Action.addPlayerToTeam(name, team)
Stats.printPlayersDetailed()
elif 'p' in userText:
Stats.printStats()
elif 'c' in userText:
name = input("enter name: ")
message = input("enter text: ")
playerId = Player._getIdByName(name)
Action.sayToMyTeam(playerId, message)
## Instruction:
Remove a few unnecessary imports
## Code After:
from engine.action import Action, Stats
from engine.player import Player
def processInput():
userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n")
if 'f' in userText:
jailCode = input("enter jail code: ")
Action.fleePlayerWithCode(jailCode)
Stats.printPlayersDetailed()
elif 's' in userText:
mobile = input("enter mobile: ")
code = input("enter code: ")
Action.handleSms(mobile, code)
Stats.printPlayersDetailed()
elif 'a' in userText:
name = input("enter name: ")
mobile = input("enter mobile: ")
#email = input("enter email: ")
Action.addPlayer(name, mobile, "")
Stats.printPlayersDetailed()
elif 'w' in userText:
hash = input("enter player hash: ")
code = input("enter code: ")
Action.handleWeb(hash, code)
Stats.printPlayersDetailed()
elif 't' in userText:
name = input("enter player name: ")
team = input("enter team name: ")
Action.addPlayerToTeam(name, team)
Stats.printPlayersDetailed()
elif 'p' in userText:
Stats.printStats()
elif 'c' in userText:
name = input("enter name: ")
message = input("enter text: ")
playerId = Player._getIdByName(name)
Action.sayToMyTeam(playerId, message)
|
// ... existing code ...
from engine.action import Action, Stats
from engine.player import Player
// ... modified code ...
Action.sayToMyTeam(playerId, message)
// ... rest of the code ...
|
59a08fff34f095f601ced76cd7b2e27665824146
|
setup.py
|
setup.py
|
from distutils.core import setup
setup(name='webracer',
version='0.2.0',
description='Comprehensive web application testing library',
author='Oleg Pudeyev',
author_email='[email protected]',
url='http://github.com/p/webracer',
packages=['webracer', 'webracer.utils'],
data_files=['LICENSE', 'README.rst'],
)
|
from distutils.core import setup
import os.path
PACKAGE = "webracer"
setup(name=PACKAGE,
version='0.2.0',
description='Comprehensive web application testing library',
author='Oleg Pudeyev',
author_email='[email protected]',
url='http://github.com/p/webracer',
packages=['webracer', 'webracer.utils'],
data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))],
)
|
Put license and readme into share/doc/webracer rather than installation root
|
Put license and readme into share/doc/webracer rather than installation root
|
Python
|
bsd-2-clause
|
p/webracer
|
from distutils.core import setup
+ import os.path
- setup(name='webracer',
+ PACKAGE = "webracer"
+
+ setup(name=PACKAGE,
version='0.2.0',
description='Comprehensive web application testing library',
author='Oleg Pudeyev',
author_email='[email protected]',
url='http://github.com/p/webracer',
packages=['webracer', 'webracer.utils'],
- data_files=['LICENSE', 'README.rst'],
+ data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))],
)
|
Put license and readme into share/doc/webracer rather than installation root
|
## Code Before:
from distutils.core import setup
setup(name='webracer',
version='0.2.0',
description='Comprehensive web application testing library',
author='Oleg Pudeyev',
author_email='[email protected]',
url='http://github.com/p/webracer',
packages=['webracer', 'webracer.utils'],
data_files=['LICENSE', 'README.rst'],
)
## Instruction:
Put license and readme into share/doc/webracer rather than installation root
## Code After:
from distutils.core import setup
import os.path
PACKAGE = "webracer"
setup(name=PACKAGE,
version='0.2.0',
description='Comprehensive web application testing library',
author='Oleg Pudeyev',
author_email='[email protected]',
url='http://github.com/p/webracer',
packages=['webracer', 'webracer.utils'],
data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))],
)
|
// ... existing code ...
from distutils.core import setup
import os.path
PACKAGE = "webracer"
setup(name=PACKAGE,
version='0.2.0',
// ... modified code ...
packages=['webracer', 'webracer.utils'],
data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))],
)
// ... rest of the code ...
|
c65ed9ec976c440b46dedc514daf883bba940282
|
myElsClient.py
|
myElsClient.py
|
import requests
class myElsClient:
"""A class that implements a Python interface to api.elsevier.com"""
# local variables
__base_url = "http://api.elsevier.com/"
def __init__(self, apiKey):
"""Instantiates a client with a given API Key."""
self.apiKey = apiKey
def getBaseURL(self):
return self.__base_url
|
import requests
class myElsClient:
"""A class that implements a Python interface to api.elsevier.com"""
# local variables
__base_url = "http://api.elsevier.com/"
# constructors
def __init__(self, apiKey):
"""Instantiates a client with a given API Key."""
self.apiKey = apiKey
# configuration functions
"""Sets an institutional token for customer authentication"""
def setInstToken(self, instToken):
self.instToken = instToken
# utility access functions
def getBaseURL(self):
"""Returns the base URL currently configured for Elsevier's APIs"""
return self.__base_url
|
Add ability to set insttoken
|
Add ability to set insttoken
|
Python
|
bsd-3-clause
|
ElsevierDev/elsapy
|
import requests
class myElsClient:
"""A class that implements a Python interface to api.elsevier.com"""
# local variables
__base_url = "http://api.elsevier.com/"
-
+
+ # constructors
def __init__(self, apiKey):
"""Instantiates a client with a given API Key."""
self.apiKey = apiKey
+ # configuration functions
+ """Sets an institutional token for customer authentication"""
+ def setInstToken(self, instToken):
+ self.instToken = instToken
+
+ # utility access functions
def getBaseURL(self):
+ """Returns the base URL currently configured for Elsevier's APIs"""
return self.__base_url
|
Add ability to set insttoken
|
## Code Before:
import requests
class myElsClient:
"""A class that implements a Python interface to api.elsevier.com"""
# local variables
__base_url = "http://api.elsevier.com/"
def __init__(self, apiKey):
"""Instantiates a client with a given API Key."""
self.apiKey = apiKey
def getBaseURL(self):
return self.__base_url
## Instruction:
Add ability to set insttoken
## Code After:
import requests
class myElsClient:
"""A class that implements a Python interface to api.elsevier.com"""
# local variables
__base_url = "http://api.elsevier.com/"
# constructors
def __init__(self, apiKey):
"""Instantiates a client with a given API Key."""
self.apiKey = apiKey
# configuration functions
"""Sets an institutional token for customer authentication"""
def setInstToken(self, instToken):
self.instToken = instToken
# utility access functions
def getBaseURL(self):
"""Returns the base URL currently configured for Elsevier's APIs"""
return self.__base_url
|
# ... existing code ...
__base_url = "http://api.elsevier.com/"
# constructors
def __init__(self, apiKey):
# ... modified code ...
# configuration functions
"""Sets an institutional token for customer authentication"""
def setInstToken(self, instToken):
self.instToken = instToken
# utility access functions
def getBaseURL(self):
"""Returns the base URL currently configured for Elsevier's APIs"""
return self.__base_url
# ... rest of the code ...
|
97894a171d1831838da28b757aabb352bc5ecfd9
|
patches/sitecustomize.py
|
patches/sitecustomize.py
|
import torch
import os
_HOST_FILE = "/etc/hosts"
kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN")
if kaggle_proxy_token:
from google.auth import credentials
from google.cloud import bigquery
from google.cloud.bigquery._http import Connection
host_entry = os.getenv("KAGGLE_DATA_PROXY_HOST_ENTRY")
if host_entry:
host_entry = host_entry.strip()
with open(_HOST_FILE, "r") as host_file:
for line in host_file:
if line.strip() == host_entry:
break
else:
with open(_HOST_FILE, "a") as host_file_append:
host_file_append.write("\n" # Just in case it wasn't newline terminated.
+ host_entry
+ "\n")
Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL")
Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token
bq_client = bigquery.Client
bigquery.Client = lambda *args, **kwargs: bq_client(
*args,
credentials=credentials.AnonymousCredentials(),
project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"),
**kwargs)
credentials.AnonymousCredentials.refresh = lambda *args: None
|
import torch
import os
kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN")
if kaggle_proxy_token:
from google.auth import credentials
from google.cloud import bigquery
from google.cloud.bigquery._http import Connection
Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL")
Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token
bq_client = bigquery.Client
bigquery.Client = lambda *args, **kwargs: bq_client(
*args,
credentials=credentials.AnonymousCredentials(),
project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"),
**kwargs)
credentials.AnonymousCredentials.refresh = lambda *args: None
|
Revert "Support adding an /etc/host entry for the data proxy, if asked."
|
Revert "Support adding an /etc/host entry for the data proxy, if asked."
This reverts commit 062f975d92c5795feb530e3ea1914d3c7dd3a96b.
There is no more need for this support in the docker image. It is fully
externally handled through docker run's `--add-host` feature.
|
Python
|
apache-2.0
|
Kaggle/docker-python,Kaggle/docker-python
|
import torch
import os
-
- _HOST_FILE = "/etc/hosts"
kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN")
if kaggle_proxy_token:
from google.auth import credentials
from google.cloud import bigquery
from google.cloud.bigquery._http import Connection
-
- host_entry = os.getenv("KAGGLE_DATA_PROXY_HOST_ENTRY")
- if host_entry:
- host_entry = host_entry.strip()
- with open(_HOST_FILE, "r") as host_file:
- for line in host_file:
- if line.strip() == host_entry:
- break
- else:
- with open(_HOST_FILE, "a") as host_file_append:
- host_file_append.write("\n" # Just in case it wasn't newline terminated.
- + host_entry
- + "\n")
Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL")
Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token
bq_client = bigquery.Client
bigquery.Client = lambda *args, **kwargs: bq_client(
*args,
credentials=credentials.AnonymousCredentials(),
project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"),
**kwargs)
credentials.AnonymousCredentials.refresh = lambda *args: None
|
Revert "Support adding an /etc/host entry for the data proxy, if asked."
|
## Code Before:
import torch
import os
_HOST_FILE = "/etc/hosts"
kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN")
if kaggle_proxy_token:
from google.auth import credentials
from google.cloud import bigquery
from google.cloud.bigquery._http import Connection
host_entry = os.getenv("KAGGLE_DATA_PROXY_HOST_ENTRY")
if host_entry:
host_entry = host_entry.strip()
with open(_HOST_FILE, "r") as host_file:
for line in host_file:
if line.strip() == host_entry:
break
else:
with open(_HOST_FILE, "a") as host_file_append:
host_file_append.write("\n" # Just in case it wasn't newline terminated.
+ host_entry
+ "\n")
Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL")
Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token
bq_client = bigquery.Client
bigquery.Client = lambda *args, **kwargs: bq_client(
*args,
credentials=credentials.AnonymousCredentials(),
project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"),
**kwargs)
credentials.AnonymousCredentials.refresh = lambda *args: None
## Instruction:
Revert "Support adding an /etc/host entry for the data proxy, if asked."
## Code After:
import torch
import os
kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN")
if kaggle_proxy_token:
from google.auth import credentials
from google.cloud import bigquery
from google.cloud.bigquery._http import Connection
Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL")
Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token
bq_client = bigquery.Client
bigquery.Client = lambda *args, **kwargs: bq_client(
*args,
credentials=credentials.AnonymousCredentials(),
project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"),
**kwargs)
credentials.AnonymousCredentials.refresh = lambda *args: None
|
// ... existing code ...
import os
// ... modified code ...
from google.cloud.bigquery._http import Connection
// ... rest of the code ...
|
2b43ab4eb41e305c5bdadf5c338e134e5569249d
|
tests/conftest.py
|
tests/conftest.py
|
import pytest
import os
import tarfile
BASEDIR = os.path.dirname(__file__)
@pytest.fixture(autouse=False)
def set_up(tmpdir):
# print BASEDIR
tmpdir.chdir()
tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz"))
tar.extractall()
tar.close()
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
|
import pytest
import os
import tarfile
BASEDIR = os.path.dirname(__file__)
@pytest.fixture(autouse=False)
def set_up(tmpdir):
# print BASEDIR
tmpdir.chdir()
tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz"))
tar.extractall()
tar.close()
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
def session_teardown():
"""Tear down testing session"""
print('Tearing down testing session')
os.chdir(BASEDIR) # so html coverage report lands in correct directory
@pytest.fixture(scope='session', autouse=True)
def session_setup(request):
"""Set up testing session"""
print('Setting up testing session')
request.addfinalizer(session_teardown)
|
Add session setup and teardown fixtures.
|
Add session setup and teardown fixtures.
|
Python
|
mit
|
bilderbuchi/ofStateManager
|
import pytest
import os
import tarfile
BASEDIR = os.path.dirname(__file__)
@pytest.fixture(autouse=False)
def set_up(tmpdir):
# print BASEDIR
tmpdir.chdir()
tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz"))
tar.extractall()
tar.close()
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
+
+ def session_teardown():
+ """Tear down testing session"""
+ print('Tearing down testing session')
+ os.chdir(BASEDIR) # so html coverage report lands in correct directory
+
+
+ @pytest.fixture(scope='session', autouse=True)
+ def session_setup(request):
+ """Set up testing session"""
+ print('Setting up testing session')
+
+ request.addfinalizer(session_teardown)
+
|
Add session setup and teardown fixtures.
|
## Code Before:
import pytest
import os
import tarfile
BASEDIR = os.path.dirname(__file__)
@pytest.fixture(autouse=False)
def set_up(tmpdir):
# print BASEDIR
tmpdir.chdir()
tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz"))
tar.extractall()
tar.close()
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
## Instruction:
Add session setup and teardown fixtures.
## Code After:
import pytest
import os
import tarfile
BASEDIR = os.path.dirname(__file__)
@pytest.fixture(autouse=False)
def set_up(tmpdir):
# print BASEDIR
tmpdir.chdir()
tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz"))
tar.extractall()
tar.close()
os.chdir('MockRepos')
print('In directory ' + os.getcwd())
# does not need teardown, since tmpdir directories get autodeleted
def session_teardown():
"""Tear down testing session"""
print('Tearing down testing session')
os.chdir(BASEDIR) # so html coverage report lands in correct directory
@pytest.fixture(scope='session', autouse=True)
def session_setup(request):
"""Set up testing session"""
print('Setting up testing session')
request.addfinalizer(session_teardown)
|
...
# does not need teardown, since tmpdir directories get autodeleted
def session_teardown():
"""Tear down testing session"""
print('Tearing down testing session')
os.chdir(BASEDIR) # so html coverage report lands in correct directory
@pytest.fixture(scope='session', autouse=True)
def session_setup(request):
"""Set up testing session"""
print('Setting up testing session')
request.addfinalizer(session_teardown)
...
|
efaa172668b8961734fa8a10650dc3191b4a7348
|
website/project/metadata/authorizers/__init__.py
|
website/project/metadata/authorizers/__init__.py
|
import json
import os
import logging
logger = logging.getLogger(__name__)
HERE = os.path.dirname(os.path.realpath(__file__))
groups = json.load(open('{0}/defaults.json'.format(HERE)))
try:
fp = open('{0}/local.json'.format(HERE))
except IOError:
logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.')
for group, members in json.load(fp).iteritems():
if group not in groups:
groups[group] = members
else:
groups[group] = set(groups[group]) | set(members)
def members_for(group):
global_members = set(groups['global'])
return global_members | set(groups.get(group, []))
|
import json
import os
import logging
logger = logging.getLogger(__name__)
HERE = os.path.dirname(os.path.realpath(__file__))
groups = json.load(open('{0}/defaults.json'.format(HERE)))
fp = None
try:
fp = open('{0}/local.json'.format(HERE))
except IOError:
logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.')
if fp:
for group, members in json.load(fp).iteritems():
if group not in groups:
groups[group] = members
else:
groups[group] = set(groups[group]) | set(members)
def members_for(group):
global_members = set(groups['global'])
return global_members | set(groups.get(group, []))
|
Allow local.json to be missing
|
Allow local.json to be missing
|
Python
|
apache-2.0
|
kch8qx/osf.io,acshi/osf.io,binoculars/osf.io,abought/osf.io,mluo613/osf.io,cslzchen/osf.io,chrisseto/osf.io,ticklemepierce/osf.io,cslzchen/osf.io,brianjgeiger/osf.io,kwierman/osf.io,brandonPurvis/osf.io,icereval/osf.io,TomBaxter/osf.io,doublebits/osf.io,mluke93/osf.io,wearpants/osf.io,alexschiller/osf.io,billyhunt/osf.io,danielneis/osf.io,rdhyee/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,Ghalko/osf.io,mattclark/osf.io,rdhyee/osf.io,kwierman/osf.io,cwisecarver/osf.io,mluke93/osf.io,Johnetordoff/osf.io,GageGaskins/osf.io,abought/osf.io,RomanZWang/osf.io,acshi/osf.io,alexschiller/osf.io,cslzchen/osf.io,GageGaskins/osf.io,SSJohns/osf.io,KAsante95/osf.io,hmoco/osf.io,saradbowman/osf.io,adlius/osf.io,zamattiac/osf.io,binoculars/osf.io,monikagrabowska/osf.io,CenterForOpenScience/osf.io,DanielSBrown/osf.io,emetsger/osf.io,SSJohns/osf.io,brianjgeiger/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io,samanehsan/osf.io,KAsante95/osf.io,emetsger/osf.io,TomBaxter/osf.io,monikagrabowska/osf.io,danielneis/osf.io,samanehsan/osf.io,mfraezz/osf.io,chennan47/osf.io,kch8qx/osf.io,GageGaskins/osf.io,kwierman/osf.io,asanfilippo7/osf.io,caneruguz/osf.io,acshi/osf.io,Nesiehr/osf.io,caseyrollins/osf.io,GageGaskins/osf.io,doublebits/osf.io,sloria/osf.io,adlius/osf.io,mluke93/osf.io,samanehsan/osf.io,samchrisinger/osf.io,jnayak1/osf.io,jnayak1/osf.io,samchrisinger/osf.io,TomHeatwole/osf.io,binoculars/osf.io,samchrisinger/osf.io,billyhunt/osf.io,crcresearch/osf.io,Johnetordoff/osf.io,doublebits/osf.io,RomanZWang/osf.io,KAsante95/osf.io,icereval/osf.io,amyshi188/osf.io,aaxelb/osf.io,leb2dg/osf.io,zamattiac/osf.io,monikagrabowska/osf.io,zamattiac/osf.io,chrisseto/osf.io,alexschiller/osf.io,TomHeatwole/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io,alexschiller/osf.io,mluo613/osf.io,TomBaxter/osf.io,hmoco/osf.io,chrisseto/osf.io,billyhunt/osf.io,cwisecarver/osf.io,mluo613/osf.io,emetsger/osf.io,Nesiehr/osf.io,Ghalko/osf.io,HalcyonChimera/osf.io,adlius/osf.io,KAsante95/osf.io,felliott/osf.io,aaxelb/osf.io,caneruguz/osf.io,ticklemepierce/osf.io,leb2dg/osf.io,felliott/osf.io,caseyrollins/osf.io,danielneis/osf.io,cslzchen/osf.io,chennan47/osf.io,Nesiehr/osf.io,asanfilippo7/osf.io,crcresearch/osf.io,rdhyee/osf.io,baylee-d/osf.io,alexschiller/osf.io,samanehsan/osf.io,jnayak1/osf.io,danielneis/osf.io,amyshi188/osf.io,DanielSBrown/osf.io,HalcyonChimera/osf.io,wearpants/osf.io,mluo613/osf.io,baylee-d/osf.io,crcresearch/osf.io,mluo613/osf.io,aaxelb/osf.io,laurenrevere/osf.io,erinspace/osf.io,kch8qx/osf.io,chrisseto/osf.io,chennan47/osf.io,doublebits/osf.io,monikagrabowska/osf.io,pattisdr/osf.io,laurenrevere/osf.io,caseyrollins/osf.io,icereval/osf.io,abought/osf.io,emetsger/osf.io,TomHeatwole/osf.io,brianjgeiger/osf.io,RomanZWang/osf.io,brianjgeiger/osf.io,erinspace/osf.io,caneruguz/osf.io,amyshi188/osf.io,RomanZWang/osf.io,asanfilippo7/osf.io,acshi/osf.io,hmoco/osf.io,acshi/osf.io,CenterForOpenScience/osf.io,wearpants/osf.io,zamattiac/osf.io,billyhunt/osf.io,kwierman/osf.io,amyshi188/osf.io,rdhyee/osf.io,Johnetordoff/osf.io,erinspace/osf.io,zachjanicki/osf.io,kch8qx/osf.io,doublebits/osf.io,mfraezz/osf.io,GageGaskins/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,Johnetordoff/osf.io,laurenrevere/osf.io,samchrisinger/osf.io,mfraezz/osf.io,hmoco/osf.io,RomanZWang/osf.io,wearpants/osf.io,brandonPurvis/osf.io,sloria/osf.io,CenterForOpenScience/osf.io,KAsante95/osf.io,mattclark/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,pattisdr/osf.io,jnayak1/osf.io,mluke93/osf.io,asanfilippo7/osf.io,adlius/osf.io,cwisecarver/osf.io,mfraezz/osf.io,Nesiehr/osf.io,caneruguz/osf.io,zachjanicki/osf.io,SSJohns/osf.io,felliott/osf.io,TomHeatwole/osf.io,mattclark/osf.io,Ghalko/osf.io,cwisecarver/osf.io,ticklemepierce/osf.io,abought/osf.io,ticklemepierce/osf.io,sloria/osf.io,billyhunt/osf.io,felliott/osf.io,SSJohns/osf.io,Ghalko/osf.io,HalcyonChimera/osf.io,pattisdr/osf.io,baylee-d/osf.io,aaxelb/osf.io,saradbowman/osf.io,monikagrabowska/osf.io,kch8qx/osf.io,HalcyonChimera/osf.io
|
import json
import os
import logging
logger = logging.getLogger(__name__)
HERE = os.path.dirname(os.path.realpath(__file__))
groups = json.load(open('{0}/defaults.json'.format(HERE)))
+ fp = None
try:
fp = open('{0}/local.json'.format(HERE))
except IOError:
logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.')
+ if fp:
- for group, members in json.load(fp).iteritems():
+ for group, members in json.load(fp).iteritems():
- if group not in groups:
+ if group not in groups:
- groups[group] = members
+ groups[group] = members
- else:
+ else:
- groups[group] = set(groups[group]) | set(members)
+ groups[group] = set(groups[group]) | set(members)
def members_for(group):
global_members = set(groups['global'])
return global_members | set(groups.get(group, []))
|
Allow local.json to be missing
|
## Code Before:
import json
import os
import logging
logger = logging.getLogger(__name__)
HERE = os.path.dirname(os.path.realpath(__file__))
groups = json.load(open('{0}/defaults.json'.format(HERE)))
try:
fp = open('{0}/local.json'.format(HERE))
except IOError:
logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.')
for group, members in json.load(fp).iteritems():
if group not in groups:
groups[group] = members
else:
groups[group] = set(groups[group]) | set(members)
def members_for(group):
global_members = set(groups['global'])
return global_members | set(groups.get(group, []))
## Instruction:
Allow local.json to be missing
## Code After:
import json
import os
import logging
logger = logging.getLogger(__name__)
HERE = os.path.dirname(os.path.realpath(__file__))
groups = json.load(open('{0}/defaults.json'.format(HERE)))
fp = None
try:
fp = open('{0}/local.json'.format(HERE))
except IOError:
logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.')
if fp:
for group, members in json.load(fp).iteritems():
if group not in groups:
groups[group] = members
else:
groups[group] = set(groups[group]) | set(members)
def members_for(group):
global_members = set(groups['global'])
return global_members | set(groups.get(group, []))
|
...
groups = json.load(open('{0}/defaults.json'.format(HERE)))
fp = None
try:
...
logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.')
if fp:
for group, members in json.load(fp).iteritems():
if group not in groups:
groups[group] = members
else:
groups[group] = set(groups[group]) | set(members)
...
|
eff3195097e9599b87f5cec9bbae744b91ae16cf
|
buses/utils.py
|
buses/utils.py
|
import re
def minify(template_source):
template_source = re.sub(r'(\n *)+', '\n', template_source)
template_source = re.sub(r'({%.+%})\n+', r'\1', template_source)
return template_source
|
import re
from haystack.utils import default_get_identifier
def minify(template_source):
template_source = re.sub(r'(\n *)+', '\n', template_source)
template_source = re.sub(r'({%.+%})\n+', r'\1', template_source)
return template_source
def get_identifier(obj_or_string):
if isinstance(obj_or_string, basestring):
return obj_or_string
return default_get_identifier(obj_or_string)
|
Add custom Hastack get_identifier function
|
Add custom Hastack get_identifier function
|
Python
|
mpl-2.0
|
jclgoodwin/bustimes.org.uk,jclgoodwin/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,jclgoodwin/bustimes.org.uk,jclgoodwin/bustimes.org.uk
|
import re
+ from haystack.utils import default_get_identifier
def minify(template_source):
template_source = re.sub(r'(\n *)+', '\n', template_source)
template_source = re.sub(r'({%.+%})\n+', r'\1', template_source)
return template_source
+ def get_identifier(obj_or_string):
+ if isinstance(obj_or_string, basestring):
+ return obj_or_string
+ return default_get_identifier(obj_or_string)
+
|
Add custom Hastack get_identifier function
|
## Code Before:
import re
def minify(template_source):
template_source = re.sub(r'(\n *)+', '\n', template_source)
template_source = re.sub(r'({%.+%})\n+', r'\1', template_source)
return template_source
## Instruction:
Add custom Hastack get_identifier function
## Code After:
import re
from haystack.utils import default_get_identifier
def minify(template_source):
template_source = re.sub(r'(\n *)+', '\n', template_source)
template_source = re.sub(r'({%.+%})\n+', r'\1', template_source)
return template_source
def get_identifier(obj_or_string):
if isinstance(obj_or_string, basestring):
return obj_or_string
return default_get_identifier(obj_or_string)
|
# ... existing code ...
import re
from haystack.utils import default_get_identifier
# ... modified code ...
return template_source
def get_identifier(obj_or_string):
if isinstance(obj_or_string, basestring):
return obj_or_string
return default_get_identifier(obj_or_string)
# ... rest of the code ...
|
afc94c1a1ebf14dbb393234233055915132a9fb8
|
django_ethereum_events/apps.py
|
django_ethereum_events/apps.py
|
from django.apps import AppConfig
from django.conf import settings
class EthereumEventsConfig(AppConfig):
name = 'django_ethereum_events'
def ready(self):
super(EthereumEventsConfig, self).ready()
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS, force=True)
|
from django.apps import AppConfig
from django.conf import settings
class EthereumEventsConfig(AppConfig):
name = 'django_ethereum_events'
|
Fix for the previous commit (Celery app removal)
|
Fix for the previous commit (Celery app removal)
Haven't paid enough attention and missed what ready method does.
Removed the code. Libraries shouldn't do this - it's main project responsibility.
|
Python
|
mit
|
artemistomaras/django-ethereum-events,artemistomaras/django-ethereum-events
|
from django.apps import AppConfig
from django.conf import settings
class EthereumEventsConfig(AppConfig):
name = 'django_ethereum_events'
- def ready(self):
- super(EthereumEventsConfig, self).ready()
- app.config_from_object('django.conf:settings')
- app.autodiscover_tasks(lambda: settings.INSTALLED_APPS, force=True)
-
|
Fix for the previous commit (Celery app removal)
|
## Code Before:
from django.apps import AppConfig
from django.conf import settings
class EthereumEventsConfig(AppConfig):
name = 'django_ethereum_events'
def ready(self):
super(EthereumEventsConfig, self).ready()
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS, force=True)
## Instruction:
Fix for the previous commit (Celery app removal)
## Code After:
from django.apps import AppConfig
from django.conf import settings
class EthereumEventsConfig(AppConfig):
name = 'django_ethereum_events'
|
# ... existing code ...
name = 'django_ethereum_events'
# ... rest of the code ...
|
e30d433153d9ad2f1d931f7f48b0ebbe9ba6763c
|
modules/new_module/new_module.py
|
modules/new_module/new_module.py
|
from models import custom_modules
from . import handlers
def register_module():
"""Registers this module in the registry."""
global_urls = [
('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url
]
course_urls = [
('/new-course-url', handlers.NewURLHandler)
] # Course URLs go on mycourse.appspot.com/course-name/url
global custom_module
custom_module = custom_modules.Module(
'New module title (has to be unique)',
'Implements some functionality',
global_urls, course_urls)
return custom_module
|
import logging
from models import custom_modules
from . import handlers
def register_module():
"""Registers this module in the registry."""
def on_module_enabled():
logging.info('Module new_module.py was just enabled')
def on_module_disabled():
logging.info('Module new_module.py was just dissabled')
global_urls = [
('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url
]
course_urls = [
('/new-course-url', handlers.NewURLHandler)
] # Course URLs go on mycourse.appspot.com/course-name/url
global custom_module
custom_module = custom_modules.Module(
'New module title (has to be unique)',
'Implements some functionality',
global_urls, course_urls,
notify_module_disabled=on_module_disabled,
notify_module_enabled=on_module_enabled)
return custom_module
|
Add enable and dissable hooks
|
Add enable and dissable hooks
|
Python
|
apache-2.0
|
UniMOOC/gcb-new-module,UniMOOC/gcb-new-module,UniMOOC/gcb-new-module,UniMOOC/gcb-new-module
|
+
+ import logging
from models import custom_modules
from . import handlers
def register_module():
"""Registers this module in the registry."""
+
+ def on_module_enabled():
+ logging.info('Module new_module.py was just enabled')
+
+ def on_module_disabled():
+ logging.info('Module new_module.py was just dissabled')
global_urls = [
('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url
]
course_urls = [
('/new-course-url', handlers.NewURLHandler)
] # Course URLs go on mycourse.appspot.com/course-name/url
global custom_module
custom_module = custom_modules.Module(
'New module title (has to be unique)',
'Implements some functionality',
- global_urls, course_urls)
+ global_urls, course_urls,
+ notify_module_disabled=on_module_disabled,
+ notify_module_enabled=on_module_enabled)
return custom_module
|
Add enable and dissable hooks
|
## Code Before:
from models import custom_modules
from . import handlers
def register_module():
"""Registers this module in the registry."""
global_urls = [
('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url
]
course_urls = [
('/new-course-url', handlers.NewURLHandler)
] # Course URLs go on mycourse.appspot.com/course-name/url
global custom_module
custom_module = custom_modules.Module(
'New module title (has to be unique)',
'Implements some functionality',
global_urls, course_urls)
return custom_module
## Instruction:
Add enable and dissable hooks
## Code After:
import logging
from models import custom_modules
from . import handlers
def register_module():
"""Registers this module in the registry."""
def on_module_enabled():
logging.info('Module new_module.py was just enabled')
def on_module_disabled():
logging.info('Module new_module.py was just dissabled')
global_urls = [
('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url
]
course_urls = [
('/new-course-url', handlers.NewURLHandler)
] # Course URLs go on mycourse.appspot.com/course-name/url
global custom_module
custom_module = custom_modules.Module(
'New module title (has to be unique)',
'Implements some functionality',
global_urls, course_urls,
notify_module_disabled=on_module_disabled,
notify_module_enabled=on_module_enabled)
return custom_module
|
# ... existing code ...
import logging
# ... modified code ...
"""Registers this module in the registry."""
def on_module_enabled():
logging.info('Module new_module.py was just enabled')
def on_module_disabled():
logging.info('Module new_module.py was just dissabled')
...
'Implements some functionality',
global_urls, course_urls,
notify_module_disabled=on_module_disabled,
notify_module_enabled=on_module_enabled)
return custom_module
# ... rest of the code ...
|
83a16ba4485f3e483adc20352cb0cef7c02f8ef2
|
tests/test_config_schema.py
|
tests/test_config_schema.py
|
from __future__ import unicode_literals, division, absolute_import
import jsonschema
from flexget import config_schema
from flexget import plugin
from tests import FlexGetBase
class TestSchemaValidator(FlexGetBase):
def test_plugin_schemas_are_valid(self):
for p in plugin.plugins.values():
if p.schema is None:
continue
try:
config_schema.SchemaValidator.check_schema(p.schema)
except jsonschema.SchemaError as e:
assert False, 'plugin `%s` has an invalid schema. %s %s' % (
p.name, '/'.join(str(p) for p in e.path), e.message)
def test_resolves_local_refs(self):
schema = {'$ref': '/schema/plugin/accept_all'}
v = config_schema.SchemaValidator(schema)
# accept_all schema should be for type boolean
assert v.is_valid(True)
assert not v.is_valid(14)
def test_custom_format_checker(self):
schema = {'type': 'string', 'format': 'quality'}
v = config_schema.SchemaValidator(schema)
assert v.is_valid('720p')
assert not v.is_valid('aoeu')
|
from __future__ import unicode_literals, division, absolute_import
import jsonschema
from flexget import config_schema
from tests import FlexGetBase
class TestSchemaValidator(FlexGetBase):
def test_registered_schemas_are_valid(self):
for path in config_schema.schema_paths:
schema = config_schema.resolve_ref(path)
try:
config_schema.SchemaValidator.check_schema(schema)
except jsonschema.SchemaError as e:
assert False, 'plugin `%s` has an invalid schema. %s %s' % (
path, '/'.join(str(p) for p in e.path), e.message)
def test_resolves_local_refs(self):
schema = {'$ref': '/schema/plugin/accept_all'}
v = config_schema.SchemaValidator(schema)
# accept_all schema should be for type boolean
assert v.is_valid(True)
assert not v.is_valid(14)
def test_custom_format_checker(self):
schema = {'type': 'string', 'format': 'quality'}
v = config_schema.SchemaValidator(schema)
assert v.is_valid('720p')
assert not v.is_valid('aoeu')
|
Convert unit test to test all registered schemas instead of plugins directly.
|
Convert unit test to test all registered schemas instead of plugins directly.
|
Python
|
mit
|
vfrc2/Flexget,poulpito/Flexget,jacobmetrick/Flexget,ibrahimkarahan/Flexget,Danfocus/Flexget,oxc/Flexget,ibrahimkarahan/Flexget,asm0dey/Flexget,qvazzler/Flexget,dsemi/Flexget,crawln45/Flexget,thalamus/Flexget,tsnoam/Flexget,patsissons/Flexget,tsnoam/Flexget,vfrc2/Flexget,Danfocus/Flexget,drwyrm/Flexget,v17al/Flexget,grrr2/Flexget,lildadou/Flexget,OmgOhnoes/Flexget,voriux/Flexget,crawln45/Flexget,ianstalk/Flexget,Danfocus/Flexget,spencerjanssen/Flexget,JorisDeRieck/Flexget,ratoaq2/Flexget,tobinjt/Flexget,thalamus/Flexget,ratoaq2/Flexget,gazpachoking/Flexget,qk4l/Flexget,grrr2/Flexget,lildadou/Flexget,tarzasai/Flexget,lildadou/Flexget,OmgOhnoes/Flexget,ZefQ/Flexget,antivirtel/Flexget,vfrc2/Flexget,Flexget/Flexget,qvazzler/Flexget,tobinjt/Flexget,patsissons/Flexget,xfouloux/Flexget,JorisDeRieck/Flexget,camon/Flexget,jawilson/Flexget,qk4l/Flexget,offbyone/Flexget,tarzasai/Flexget,poulpito/Flexget,sean797/Flexget,camon/Flexget,jacobmetrick/Flexget,Pretagonist/Flexget,tobinjt/Flexget,thalamus/Flexget,tsnoam/Flexget,tarzasai/Flexget,ratoaq2/Flexget,Pretagonist/Flexget,ianstalk/Flexget,ibrahimkarahan/Flexget,ZefQ/Flexget,gazpachoking/Flexget,grrr2/Flexget,Flexget/Flexget,malkavi/Flexget,Flexget/Flexget,spencerjanssen/Flexget,tvcsantos/Flexget,X-dark/Flexget,Flexget/Flexget,LynxyssCZ/Flexget,v17al/Flexget,asm0dey/Flexget,dsemi/Flexget,drwyrm/Flexget,tobinjt/Flexget,offbyone/Flexget,JorisDeRieck/Flexget,oxc/Flexget,cvium/Flexget,qk4l/Flexget,antivirtel/Flexget,LynxyssCZ/Flexget,malkavi/Flexget,cvium/Flexget,tvcsantos/Flexget,ianstalk/Flexget,poulpito/Flexget,drwyrm/Flexget,malkavi/Flexget,X-dark/Flexget,patsissons/Flexget,antivirtel/Flexget,LynxyssCZ/Flexget,offbyone/Flexget,JorisDeRieck/Flexget,jacobmetrick/Flexget,crawln45/Flexget,cvium/Flexget,jawilson/Flexget,jawilson/Flexget,oxc/Flexget,Danfocus/Flexget,xfouloux/Flexget,asm0dey/Flexget,crawln45/Flexget,spencerjanssen/Flexget,dsemi/Flexget,X-dark/Flexget,malkavi/Flexget,v17al/Flexget,ZefQ/Flexget,Pretagonist/Flexget,OmgOhnoes/Flexget,xfouloux/Flexget,jawilson/Flexget,LynxyssCZ/Flexget,sean797/Flexget,qvazzler/Flexget,voriux/Flexget,sean797/Flexget
|
from __future__ import unicode_literals, division, absolute_import
import jsonschema
from flexget import config_schema
- from flexget import plugin
from tests import FlexGetBase
class TestSchemaValidator(FlexGetBase):
- def test_plugin_schemas_are_valid(self):
+ def test_registered_schemas_are_valid(self):
+ for path in config_schema.schema_paths:
+ schema = config_schema.resolve_ref(path)
- for p in plugin.plugins.values():
- if p.schema is None:
- continue
try:
- config_schema.SchemaValidator.check_schema(p.schema)
+ config_schema.SchemaValidator.check_schema(schema)
except jsonschema.SchemaError as e:
assert False, 'plugin `%s` has an invalid schema. %s %s' % (
- p.name, '/'.join(str(p) for p in e.path), e.message)
+ path, '/'.join(str(p) for p in e.path), e.message)
def test_resolves_local_refs(self):
schema = {'$ref': '/schema/plugin/accept_all'}
v = config_schema.SchemaValidator(schema)
# accept_all schema should be for type boolean
assert v.is_valid(True)
assert not v.is_valid(14)
def test_custom_format_checker(self):
schema = {'type': 'string', 'format': 'quality'}
v = config_schema.SchemaValidator(schema)
assert v.is_valid('720p')
assert not v.is_valid('aoeu')
|
Convert unit test to test all registered schemas instead of plugins directly.
|
## Code Before:
from __future__ import unicode_literals, division, absolute_import
import jsonschema
from flexget import config_schema
from flexget import plugin
from tests import FlexGetBase
class TestSchemaValidator(FlexGetBase):
def test_plugin_schemas_are_valid(self):
for p in plugin.plugins.values():
if p.schema is None:
continue
try:
config_schema.SchemaValidator.check_schema(p.schema)
except jsonschema.SchemaError as e:
assert False, 'plugin `%s` has an invalid schema. %s %s' % (
p.name, '/'.join(str(p) for p in e.path), e.message)
def test_resolves_local_refs(self):
schema = {'$ref': '/schema/plugin/accept_all'}
v = config_schema.SchemaValidator(schema)
# accept_all schema should be for type boolean
assert v.is_valid(True)
assert not v.is_valid(14)
def test_custom_format_checker(self):
schema = {'type': 'string', 'format': 'quality'}
v = config_schema.SchemaValidator(schema)
assert v.is_valid('720p')
assert not v.is_valid('aoeu')
## Instruction:
Convert unit test to test all registered schemas instead of plugins directly.
## Code After:
from __future__ import unicode_literals, division, absolute_import
import jsonschema
from flexget import config_schema
from tests import FlexGetBase
class TestSchemaValidator(FlexGetBase):
def test_registered_schemas_are_valid(self):
for path in config_schema.schema_paths:
schema = config_schema.resolve_ref(path)
try:
config_schema.SchemaValidator.check_schema(schema)
except jsonschema.SchemaError as e:
assert False, 'plugin `%s` has an invalid schema. %s %s' % (
path, '/'.join(str(p) for p in e.path), e.message)
def test_resolves_local_refs(self):
schema = {'$ref': '/schema/plugin/accept_all'}
v = config_schema.SchemaValidator(schema)
# accept_all schema should be for type boolean
assert v.is_valid(True)
assert not v.is_valid(14)
def test_custom_format_checker(self):
schema = {'type': 'string', 'format': 'quality'}
v = config_schema.SchemaValidator(schema)
assert v.is_valid('720p')
assert not v.is_valid('aoeu')
|
# ... existing code ...
from flexget import config_schema
from tests import FlexGetBase
# ... modified code ...
class TestSchemaValidator(FlexGetBase):
def test_registered_schemas_are_valid(self):
for path in config_schema.schema_paths:
schema = config_schema.resolve_ref(path)
try:
config_schema.SchemaValidator.check_schema(schema)
except jsonschema.SchemaError as e:
...
assert False, 'plugin `%s` has an invalid schema. %s %s' % (
path, '/'.join(str(p) for p in e.path), e.message)
# ... rest of the code ...
|
4daefdb0a4def961572fc22d0fe01a394b11fad9
|
tests/test_httpclient.py
|
tests/test_httpclient.py
|
try:
import unittest2 as unittest
except ImportError:
import unittest
import sys
sys.path.append('..')
from pyrabbit import http
class TestHTTPClient(unittest.TestCase):
"""
Except for the init test, these are largely functional tests that
require a RabbitMQ management API to be available on localhost:55672
"""
def setUp(self):
self.c = http.HTTPClient('localhost:55672', 'guest', 'guest')
def test_client_init(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest')
self.assertIsInstance(c, http.HTTPClient)
def test_client_init_sets_default_timeout(self):
self.assertEqual(self.c.client.timeout, 1)
def test_client_init_with_timeout(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5)
self.assertEqual(c.client.timeout, 5)
|
try:
import unittest2 as unittest
except ImportError:
import unittest
import sys
sys.path.append('..')
from pyrabbit import http
class TestHTTPClient(unittest.TestCase):
"""
Except for the init test, these are largely functional tests that
require a RabbitMQ management API to be available on localhost:55672
"""
def setUp(self):
self.c = http.HTTPClient('localhost:55672', 'guest', 'guest')
def test_client_init(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest')
self.assertIsInstance(c, http.HTTPClient)
def test_client_init_sets_credentials(self):
domain = ''
expected_credentials = [(domain, 'guest', 'guest')]
self.assertEqual(
self.c.client.credentials.credentials, expected_credentials)
def test_client_init_sets_default_timeout(self):
self.assertEqual(self.c.client.timeout, 1)
def test_client_init_with_timeout(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5)
self.assertEqual(c.client.timeout, 5)
|
Test creation of HTTP credentials
|
tests.http: Test creation of HTTP credentials
|
Python
|
bsd-3-clause
|
ranjithlav/pyrabbit,bkjones/pyrabbit,NeCTAR-RC/pyrabbit,chaos95/pyrabbit,switchtower/pyrabbit
|
try:
import unittest2 as unittest
except ImportError:
import unittest
import sys
sys.path.append('..')
from pyrabbit import http
class TestHTTPClient(unittest.TestCase):
"""
Except for the init test, these are largely functional tests that
require a RabbitMQ management API to be available on localhost:55672
"""
def setUp(self):
self.c = http.HTTPClient('localhost:55672', 'guest', 'guest')
def test_client_init(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest')
self.assertIsInstance(c, http.HTTPClient)
+ def test_client_init_sets_credentials(self):
+ domain = ''
+ expected_credentials = [(domain, 'guest', 'guest')]
+ self.assertEqual(
+ self.c.client.credentials.credentials, expected_credentials)
+
def test_client_init_sets_default_timeout(self):
self.assertEqual(self.c.client.timeout, 1)
def test_client_init_with_timeout(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5)
self.assertEqual(c.client.timeout, 5)
|
Test creation of HTTP credentials
|
## Code Before:
try:
import unittest2 as unittest
except ImportError:
import unittest
import sys
sys.path.append('..')
from pyrabbit import http
class TestHTTPClient(unittest.TestCase):
"""
Except for the init test, these are largely functional tests that
require a RabbitMQ management API to be available on localhost:55672
"""
def setUp(self):
self.c = http.HTTPClient('localhost:55672', 'guest', 'guest')
def test_client_init(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest')
self.assertIsInstance(c, http.HTTPClient)
def test_client_init_sets_default_timeout(self):
self.assertEqual(self.c.client.timeout, 1)
def test_client_init_with_timeout(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5)
self.assertEqual(c.client.timeout, 5)
## Instruction:
Test creation of HTTP credentials
## Code After:
try:
import unittest2 as unittest
except ImportError:
import unittest
import sys
sys.path.append('..')
from pyrabbit import http
class TestHTTPClient(unittest.TestCase):
"""
Except for the init test, these are largely functional tests that
require a RabbitMQ management API to be available on localhost:55672
"""
def setUp(self):
self.c = http.HTTPClient('localhost:55672', 'guest', 'guest')
def test_client_init(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest')
self.assertIsInstance(c, http.HTTPClient)
def test_client_init_sets_credentials(self):
domain = ''
expected_credentials = [(domain, 'guest', 'guest')]
self.assertEqual(
self.c.client.credentials.credentials, expected_credentials)
def test_client_init_sets_default_timeout(self):
self.assertEqual(self.c.client.timeout, 1)
def test_client_init_with_timeout(self):
c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5)
self.assertEqual(c.client.timeout, 5)
|
// ... existing code ...
def test_client_init_sets_credentials(self):
domain = ''
expected_credentials = [(domain, 'guest', 'guest')]
self.assertEqual(
self.c.client.credentials.credentials, expected_credentials)
def test_client_init_sets_default_timeout(self):
// ... rest of the code ...
|
18bd0bcc0d892aef4ea9babfc6ec2af6e40cea62
|
manager/urls.py
|
manager/urls.py
|
from django.conf.urls import url
from manager import views
urlpatterns = [
url(r'^$', views.package_list, name='package_list'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail')
]
|
from django.conf.urls import url
from manager import views
urlpatterns = [
url(r'^$', views.package_list, name='package_list'),
url(r'^packages/$', views.package_list, name='package_list'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail')
]
|
Add alternative package list url
|
Add alternative package list url
|
Python
|
mit
|
colajam93/aurpackager,colajam93/aurpackager,colajam93/aurpackager,colajam93/aurpackager
|
from django.conf.urls import url
from manager import views
urlpatterns = [
url(r'^$', views.package_list, name='package_list'),
+ url(r'^packages/$', views.package_list, name='package_list'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail')
]
|
Add alternative package list url
|
## Code Before:
from django.conf.urls import url
from manager import views
urlpatterns = [
url(r'^$', views.package_list, name='package_list'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail')
]
## Instruction:
Add alternative package list url
## Code After:
from django.conf.urls import url
from manager import views
urlpatterns = [
url(r'^$', views.package_list, name='package_list'),
url(r'^packages/$', views.package_list, name='package_list'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail')
]
|
...
url(r'^$', views.package_list, name='package_list'),
url(r'^packages/$', views.package_list, name='package_list'),
url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'),
...
|
a03eb91088943a4b3ed0ae5fc87b104562a4a645
|
location_field/urls.py
|
location_field/urls.py
|
try:
from django.conf.urls import patterns # Django>=1.6
except ImportError:
from django.conf.urls.defaults import patterns # Django<1.6
import os
app_dir = os.path.dirname(__file__)
urlpatterns = patterns(
'',
(r'^media/(.*)$', 'django.views.static.serve', {
'document_root': '%s/media' % app_dir}),
)
|
from django.conf.urls import patterns
import os
app_dir = os.path.dirname(__file__)
urlpatterns = patterns(
'',
(r'^media/(.*)$', 'django.views.static.serve', {
'document_root': '%s/media' % app_dir}),
)
|
Drop support for Django 1.6
|
Drop support for Django 1.6
|
Python
|
mit
|
Mixser/django-location-field,recklessromeo/django-location-field,Mixser/django-location-field,voodmania/django-location-field,recklessromeo/django-location-field,undernewmanagement/django-location-field,voodmania/django-location-field,caioariede/django-location-field,caioariede/django-location-field,undernewmanagement/django-location-field,Mixser/django-location-field,undernewmanagement/django-location-field,caioariede/django-location-field,recklessromeo/django-location-field,voodmania/django-location-field
|
- try:
- from django.conf.urls import patterns # Django>=1.6
+ from django.conf.urls import patterns
- except ImportError:
- from django.conf.urls.defaults import patterns # Django<1.6
import os
app_dir = os.path.dirname(__file__)
urlpatterns = patterns(
'',
(r'^media/(.*)$', 'django.views.static.serve', {
'document_root': '%s/media' % app_dir}),
)
|
Drop support for Django 1.6
|
## Code Before:
try:
from django.conf.urls import patterns # Django>=1.6
except ImportError:
from django.conf.urls.defaults import patterns # Django<1.6
import os
app_dir = os.path.dirname(__file__)
urlpatterns = patterns(
'',
(r'^media/(.*)$', 'django.views.static.serve', {
'document_root': '%s/media' % app_dir}),
)
## Instruction:
Drop support for Django 1.6
## Code After:
from django.conf.urls import patterns
import os
app_dir = os.path.dirname(__file__)
urlpatterns = patterns(
'',
(r'^media/(.*)$', 'django.views.static.serve', {
'document_root': '%s/media' % app_dir}),
)
|
# ... existing code ...
from django.conf.urls import patterns
# ... rest of the code ...
|
ecac9283bc831a6879f21e80e1b98818683ff6a4
|
atlas/prodtask/management/commands/pthealthcheck.py
|
atlas/prodtask/management/commands/pthealthcheck.py
|
from django.core.management.base import BaseCommand, CommandError
import time
from django_celery_beat.models import PeriodicTask
from django.utils import timezone
from datetime import timedelta
from atlas.prodtask.views import send_alarm_message
class Command(BaseCommand):
args = 'None'
help = 'Check celery beat health'
def handle(self, *args, **options):
if not args:
try:
try:
last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0]
except Exception as e:
send_alarm_message('Alarm: the celery beat health check problem',
f'Celery beat health check problem {e}')
raise e
if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3):
send_alarm_message('Alarm: the celery beat is stuck',
f'Celery beat last updated {last_executed_task.last_run_at}')
except Exception as e:
raise CommandError('Some problem during alarm mail sending check: %s'%e)
|
from django.core.management.base import BaseCommand, CommandError
import time
from django_celery_beat.models import PeriodicTask
from django.utils import timezone
from datetime import timedelta
from atlas.prodtask.views import send_alarm_message
class Command(BaseCommand):
args = 'None'
help = 'Check celery beat health'
def handle(self, *args, **options):
if not args:
self.stdout.write(f'Start celery beat health check {timezone.now()}')
try:
try:
last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0]
except Exception as e:
send_alarm_message('Alarm: the celery beat health check problem',
f'Celery beat health check problem {e}')
raise e
if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3):
send_alarm_message('Alarm: the celery beat is stuck',
f'Celery beat last updated {last_executed_task.last_run_at}')
except Exception as e:
raise CommandError('Some problem during alarm mail sending check: %s'%e)
|
Add logging for health check
|
Add logging for health check
|
Python
|
apache-2.0
|
PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas
|
from django.core.management.base import BaseCommand, CommandError
import time
from django_celery_beat.models import PeriodicTask
from django.utils import timezone
from datetime import timedelta
from atlas.prodtask.views import send_alarm_message
class Command(BaseCommand):
args = 'None'
help = 'Check celery beat health'
def handle(self, *args, **options):
if not args:
+ self.stdout.write(f'Start celery beat health check {timezone.now()}')
+
try:
try:
last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0]
except Exception as e:
send_alarm_message('Alarm: the celery beat health check problem',
f'Celery beat health check problem {e}')
raise e
if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3):
send_alarm_message('Alarm: the celery beat is stuck',
f'Celery beat last updated {last_executed_task.last_run_at}')
except Exception as e:
raise CommandError('Some problem during alarm mail sending check: %s'%e)
|
Add logging for health check
|
## Code Before:
from django.core.management.base import BaseCommand, CommandError
import time
from django_celery_beat.models import PeriodicTask
from django.utils import timezone
from datetime import timedelta
from atlas.prodtask.views import send_alarm_message
class Command(BaseCommand):
args = 'None'
help = 'Check celery beat health'
def handle(self, *args, **options):
if not args:
try:
try:
last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0]
except Exception as e:
send_alarm_message('Alarm: the celery beat health check problem',
f'Celery beat health check problem {e}')
raise e
if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3):
send_alarm_message('Alarm: the celery beat is stuck',
f'Celery beat last updated {last_executed_task.last_run_at}')
except Exception as e:
raise CommandError('Some problem during alarm mail sending check: %s'%e)
## Instruction:
Add logging for health check
## Code After:
from django.core.management.base import BaseCommand, CommandError
import time
from django_celery_beat.models import PeriodicTask
from django.utils import timezone
from datetime import timedelta
from atlas.prodtask.views import send_alarm_message
class Command(BaseCommand):
args = 'None'
help = 'Check celery beat health'
def handle(self, *args, **options):
if not args:
self.stdout.write(f'Start celery beat health check {timezone.now()}')
try:
try:
last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0]
except Exception as e:
send_alarm_message('Alarm: the celery beat health check problem',
f'Celery beat health check problem {e}')
raise e
if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3):
send_alarm_message('Alarm: the celery beat is stuck',
f'Celery beat last updated {last_executed_task.last_run_at}')
except Exception as e:
raise CommandError('Some problem during alarm mail sending check: %s'%e)
|
# ... existing code ...
if not args:
self.stdout.write(f'Start celery beat health check {timezone.now()}')
try:
# ... rest of the code ...
|
442b083e9d1618569aa96a653ed2c0e4dfc27e59
|
saleor/search/forms.py
|
saleor/search/forms.py
|
from django import forms
from .backends import get_search_backend
class SearchForm(forms.Form):
q = forms.CharField(label='Query', required=True)
def search(self, model_or_queryset):
backend = get_search_backend('default')
query = self.cleaned_data['q']
results = backend.search(query, model_or_queryset=model_or_queryset)
return results
|
from django import forms
from django.utils.translation import pgettext
from .backends import get_search_backend
class SearchForm(forms.Form):
q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True)
def search(self, model_or_queryset):
backend = get_search_backend('default')
query = self.cleaned_data['q']
results = backend.search(query, model_or_queryset=model_or_queryset)
return results
|
Add contextual marker for search app
|
Add contextual marker for search app
|
Python
|
bsd-3-clause
|
jreigel/saleor,car3oon/saleor,KenMutemi/saleor,HyperManTT/ECommerceSaleor,itbabu/saleor,UITools/saleor,HyperManTT/ECommerceSaleor,UITools/saleor,maferelo/saleor,tfroehlich82/saleor,mociepka/saleor,mociepka/saleor,UITools/saleor,maferelo/saleor,UITools/saleor,tfroehlich82/saleor,UITools/saleor,jreigel/saleor,itbabu/saleor,car3oon/saleor,KenMutemi/saleor,jreigel/saleor,car3oon/saleor,HyperManTT/ECommerceSaleor,KenMutemi/saleor,tfroehlich82/saleor,itbabu/saleor,maferelo/saleor,mociepka/saleor
|
from django import forms
+ from django.utils.translation import pgettext
from .backends import get_search_backend
class SearchForm(forms.Form):
- q = forms.CharField(label='Query', required=True)
+ q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True)
def search(self, model_or_queryset):
backend = get_search_backend('default')
query = self.cleaned_data['q']
results = backend.search(query, model_or_queryset=model_or_queryset)
return results
|
Add contextual marker for search app
|
## Code Before:
from django import forms
from .backends import get_search_backend
class SearchForm(forms.Form):
q = forms.CharField(label='Query', required=True)
def search(self, model_or_queryset):
backend = get_search_backend('default')
query = self.cleaned_data['q']
results = backend.search(query, model_or_queryset=model_or_queryset)
return results
## Instruction:
Add contextual marker for search app
## Code After:
from django import forms
from django.utils.translation import pgettext
from .backends import get_search_backend
class SearchForm(forms.Form):
q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True)
def search(self, model_or_queryset):
backend = get_search_backend('default')
query = self.cleaned_data['q']
results = backend.search(query, model_or_queryset=model_or_queryset)
return results
|
# ... existing code ...
from django import forms
from django.utils.translation import pgettext
# ... modified code ...
class SearchForm(forms.Form):
q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True)
# ... rest of the code ...
|
73e15928a8427eb5a6e4a886660b9493e50cd699
|
currencies/models.py
|
currencies/models.py
|
from django.db import models
from django.utils.translation import gettext_lazy as _
class Currency(models.Model):
code = models.CharField(_('code'), max_length=3)
name = models.CharField(_('name'), max_length=35)
symbol = models.CharField(_('symbol'), max_length=1, blank=True)
factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4,
help_text=_('Specifies the difference of the currency to default one.'))
is_active = models.BooleanField(_('active'), default=True,
help_text=_('The currency will be available.'))
is_default = models.BooleanField(_('default'), default=False,
help_text=_('Make this the default currency.'))
class Meta:
verbose_name = _('currency')
verbose_name_plural = _('currencies')
def __unicode__(self):
return self.code
def save(self, **kwargs):
# Make sure the default currency is unique
if self.is_default:
Currency.objects.filter(is_default=True).update(is_default=False)
super(Currency, self).save(**kwargs)
|
from django.db import models
from django.utils.translation import gettext_lazy as _
class Currency(models.Model):
code = models.CharField(_('code'), max_length=3)
name = models.CharField(_('name'), max_length=35)
symbol = models.CharField(_('symbol'), max_length=1, blank=True)
factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4,
help_text=_('Specifies the difference of the currency to default one.'))
is_active = models.BooleanField(_('active'), default=True,
help_text=_('The currency will be available.'))
is_base = models.BooleanField(_('base'), default=False,
help_text=_('Make this the base currency against which rates are calculated.'))
is_default = models.BooleanField(_('default'), default=False,
help_text=_('Make this the default user currency.'))
class Meta:
verbose_name = _('currency')
verbose_name_plural = _('currencies')
def __unicode__(self):
return self.code
def save(self, **kwargs):
# Make sure the base and default currencies are unique
if self.is_base:
Currency.objects.filter(is_base=True).update(is_base=False)
if self.is_default:
Currency.objects.filter(is_default=True).update(is_default=False)
super(Currency, self).save(**kwargs)
|
Add a Currency.is_base field (currently unused)
|
Add a Currency.is_base field (currently unused)
|
Python
|
bsd-3-clause
|
pathakamit88/django-currencies,panosl/django-currencies,ydaniv/django-currencies,mysociety/django-currencies,panosl/django-currencies,barseghyanartur/django-currencies,bashu/django-simple-currencies,pathakamit88/django-currencies,ydaniv/django-currencies,marcosalcazar/django-currencies,jmp0xf/django-currencies,racitup/django-currencies,mysociety/django-currencies,marcosalcazar/django-currencies,bashu/django-simple-currencies,racitup/django-currencies
|
from django.db import models
from django.utils.translation import gettext_lazy as _
class Currency(models.Model):
code = models.CharField(_('code'), max_length=3)
name = models.CharField(_('name'), max_length=35)
symbol = models.CharField(_('symbol'), max_length=1, blank=True)
factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4,
help_text=_('Specifies the difference of the currency to default one.'))
is_active = models.BooleanField(_('active'), default=True,
help_text=_('The currency will be available.'))
+ is_base = models.BooleanField(_('base'), default=False,
+ help_text=_('Make this the base currency against which rates are calculated.'))
is_default = models.BooleanField(_('default'), default=False,
- help_text=_('Make this the default currency.'))
+ help_text=_('Make this the default user currency.'))
class Meta:
verbose_name = _('currency')
verbose_name_plural = _('currencies')
def __unicode__(self):
return self.code
def save(self, **kwargs):
- # Make sure the default currency is unique
+ # Make sure the base and default currencies are unique
+ if self.is_base:
+ Currency.objects.filter(is_base=True).update(is_base=False)
if self.is_default:
Currency.objects.filter(is_default=True).update(is_default=False)
super(Currency, self).save(**kwargs)
|
Add a Currency.is_base field (currently unused)
|
## Code Before:
from django.db import models
from django.utils.translation import gettext_lazy as _
class Currency(models.Model):
code = models.CharField(_('code'), max_length=3)
name = models.CharField(_('name'), max_length=35)
symbol = models.CharField(_('symbol'), max_length=1, blank=True)
factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4,
help_text=_('Specifies the difference of the currency to default one.'))
is_active = models.BooleanField(_('active'), default=True,
help_text=_('The currency will be available.'))
is_default = models.BooleanField(_('default'), default=False,
help_text=_('Make this the default currency.'))
class Meta:
verbose_name = _('currency')
verbose_name_plural = _('currencies')
def __unicode__(self):
return self.code
def save(self, **kwargs):
# Make sure the default currency is unique
if self.is_default:
Currency.objects.filter(is_default=True).update(is_default=False)
super(Currency, self).save(**kwargs)
## Instruction:
Add a Currency.is_base field (currently unused)
## Code After:
from django.db import models
from django.utils.translation import gettext_lazy as _
class Currency(models.Model):
code = models.CharField(_('code'), max_length=3)
name = models.CharField(_('name'), max_length=35)
symbol = models.CharField(_('symbol'), max_length=1, blank=True)
factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4,
help_text=_('Specifies the difference of the currency to default one.'))
is_active = models.BooleanField(_('active'), default=True,
help_text=_('The currency will be available.'))
is_base = models.BooleanField(_('base'), default=False,
help_text=_('Make this the base currency against which rates are calculated.'))
is_default = models.BooleanField(_('default'), default=False,
help_text=_('Make this the default user currency.'))
class Meta:
verbose_name = _('currency')
verbose_name_plural = _('currencies')
def __unicode__(self):
return self.code
def save(self, **kwargs):
# Make sure the base and default currencies are unique
if self.is_base:
Currency.objects.filter(is_base=True).update(is_base=False)
if self.is_default:
Currency.objects.filter(is_default=True).update(is_default=False)
super(Currency, self).save(**kwargs)
|
// ... existing code ...
help_text=_('The currency will be available.'))
is_base = models.BooleanField(_('base'), default=False,
help_text=_('Make this the base currency against which rates are calculated.'))
is_default = models.BooleanField(_('default'), default=False,
help_text=_('Make this the default user currency.'))
// ... modified code ...
def save(self, **kwargs):
# Make sure the base and default currencies are unique
if self.is_base:
Currency.objects.filter(is_base=True).update(is_base=False)
if self.is_default:
// ... rest of the code ...
|
220e0008924878f774f570cc0122c563f2c17465
|
recipes/migrations/0010_auto_20150919_1228.py
|
recipes/migrations/0010_auto_20150919_1228.py
|
from __future__ import unicode_literals
from django.db import models, migrations
def change_to_usage(apps, schema_editor):
Recipe = apps.get_model("recipes", "Recipe")
Ingredient = apps.get_model("recipes", "Ingredient")
IngredientUsage = apps.get_model("recipes", "IngredientUsage")
for recipe in Recipe.objects.all():
for ingredient in recipe.ingredient_set.all():
u = IngredientUsage(recipe=r, ingredient=i, quantity=1)
u.save()
class Migration(migrations.Migration):
dependencies = [
('recipes', '0009_auto_20150919_1226'),
]
operations = [
migrations.RunPython(change_to_usage),
]
|
from __future__ import unicode_literals
from django.db import models, migrations
def change_to_usage(apps, schema_editor):
Recipe = apps.get_model("recipes", "Recipe")
Ingredient = apps.get_model("recipes", "Ingredient")
IngredientUsage = apps.get_model("recipes", "IngredientUsage")
for recipe in Recipe.objects.all():
for ingredient in recipe.ingredient_set.all():
u = IngredientUsage(recipe=recipe, ingredient=ingredient,
quantity=1)
u.save()
class Migration(migrations.Migration):
dependencies = [
('recipes', '0009_auto_20150919_1226'),
]
operations = [
migrations.RunPython(change_to_usage),
]
|
Make the data migration actually work
|
Make the data migration actually work
|
Python
|
agpl-3.0
|
XeryusTC/rotd,XeryusTC/rotd,XeryusTC/rotd
|
from __future__ import unicode_literals
from django.db import models, migrations
def change_to_usage(apps, schema_editor):
Recipe = apps.get_model("recipes", "Recipe")
Ingredient = apps.get_model("recipes", "Ingredient")
IngredientUsage = apps.get_model("recipes", "IngredientUsage")
for recipe in Recipe.objects.all():
for ingredient in recipe.ingredient_set.all():
- u = IngredientUsage(recipe=r, ingredient=i, quantity=1)
+ u = IngredientUsage(recipe=recipe, ingredient=ingredient,
+ quantity=1)
u.save()
class Migration(migrations.Migration):
dependencies = [
('recipes', '0009_auto_20150919_1226'),
]
operations = [
migrations.RunPython(change_to_usage),
]
|
Make the data migration actually work
|
## Code Before:
from __future__ import unicode_literals
from django.db import models, migrations
def change_to_usage(apps, schema_editor):
Recipe = apps.get_model("recipes", "Recipe")
Ingredient = apps.get_model("recipes", "Ingredient")
IngredientUsage = apps.get_model("recipes", "IngredientUsage")
for recipe in Recipe.objects.all():
for ingredient in recipe.ingredient_set.all():
u = IngredientUsage(recipe=r, ingredient=i, quantity=1)
u.save()
class Migration(migrations.Migration):
dependencies = [
('recipes', '0009_auto_20150919_1226'),
]
operations = [
migrations.RunPython(change_to_usage),
]
## Instruction:
Make the data migration actually work
## Code After:
from __future__ import unicode_literals
from django.db import models, migrations
def change_to_usage(apps, schema_editor):
Recipe = apps.get_model("recipes", "Recipe")
Ingredient = apps.get_model("recipes", "Ingredient")
IngredientUsage = apps.get_model("recipes", "IngredientUsage")
for recipe in Recipe.objects.all():
for ingredient in recipe.ingredient_set.all():
u = IngredientUsage(recipe=recipe, ingredient=ingredient,
quantity=1)
u.save()
class Migration(migrations.Migration):
dependencies = [
('recipes', '0009_auto_20150919_1226'),
]
operations = [
migrations.RunPython(change_to_usage),
]
|
// ... existing code ...
for ingredient in recipe.ingredient_set.all():
u = IngredientUsage(recipe=recipe, ingredient=ingredient,
quantity=1)
u.save()
// ... rest of the code ...
|
8fe8717b4e2afe6329d2dd25210371df3eab2b4f
|
test/test_stdlib.py
|
test/test_stdlib.py
|
import pep543.stdlib
from .backend_tests import SimpleNegotiation
class TestSimpleNegotiationStdlib(SimpleNegotiation):
BACKEND = pep543.stdlib.STDLIB_BACKEND
|
import pep543
import pep543.stdlib
import pytest
from .backend_tests import SimpleNegotiation
CONTEXTS = (
pep543.stdlib.STDLIB_BACKEND.client_context,
pep543.stdlib.STDLIB_BACKEND.server_context
)
def assert_wrap_fails(context, exception):
"""
A convenient helper that calls wrap_buffers with the appropriate number of
arugments and asserts that it raises the appropriate error.
"""
if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context):
with pytest.raises(exception):
context.wrap_buffers(server_hostname=None)
else:
with pytest.raises(exception):
context.wrap_buffers()
class TestSimpleNegotiationStdlib(SimpleNegotiation):
BACKEND = pep543.stdlib.STDLIB_BACKEND
class TestStdlibErrorHandling(object):
"""
Validate that the stdlib backend can do sensible error handling in specific
situations that it cannot handle.
"""
@pytest.mark.parametrize(
'lowest,highest', (
(object(), None), (None, object()), (object(), object())
)
)
@pytest.mark.parametrize('context', CONTEXTS)
def test_bad_values_for_versions_client(self, lowest, highest, context):
"""
Using TLSConfiguration objects with a bad value for their minimum
version raises a TLSError with Client contexts.
"""
config = pep543.TLSConfiguration(
validate_certificates=False,
lowest_supported_version=lowest,
highest_supported_version=highest
)
ctx = context(config)
assert_wrap_fails(ctx, pep543.TLSError)
|
Test that we reject bad TLS versions
|
Test that we reject bad TLS versions
|
Python
|
mit
|
python-hyper/pep543
|
+ import pep543
import pep543.stdlib
+ import pytest
+
from .backend_tests import SimpleNegotiation
+
+
+ CONTEXTS = (
+ pep543.stdlib.STDLIB_BACKEND.client_context,
+ pep543.stdlib.STDLIB_BACKEND.server_context
+ )
+
+
+ def assert_wrap_fails(context, exception):
+ """
+ A convenient helper that calls wrap_buffers with the appropriate number of
+ arugments and asserts that it raises the appropriate error.
+ """
+ if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context):
+ with pytest.raises(exception):
+ context.wrap_buffers(server_hostname=None)
+ else:
+ with pytest.raises(exception):
+ context.wrap_buffers()
class TestSimpleNegotiationStdlib(SimpleNegotiation):
BACKEND = pep543.stdlib.STDLIB_BACKEND
+
+ class TestStdlibErrorHandling(object):
+ """
+ Validate that the stdlib backend can do sensible error handling in specific
+ situations that it cannot handle.
+ """
+ @pytest.mark.parametrize(
+ 'lowest,highest', (
+ (object(), None), (None, object()), (object(), object())
+ )
+ )
+ @pytest.mark.parametrize('context', CONTEXTS)
+ def test_bad_values_for_versions_client(self, lowest, highest, context):
+ """
+ Using TLSConfiguration objects with a bad value for their minimum
+ version raises a TLSError with Client contexts.
+ """
+ config = pep543.TLSConfiguration(
+ validate_certificates=False,
+ lowest_supported_version=lowest,
+ highest_supported_version=highest
+ )
+ ctx = context(config)
+ assert_wrap_fails(ctx, pep543.TLSError)
+
|
Test that we reject bad TLS versions
|
## Code Before:
import pep543.stdlib
from .backend_tests import SimpleNegotiation
class TestSimpleNegotiationStdlib(SimpleNegotiation):
BACKEND = pep543.stdlib.STDLIB_BACKEND
## Instruction:
Test that we reject bad TLS versions
## Code After:
import pep543
import pep543.stdlib
import pytest
from .backend_tests import SimpleNegotiation
CONTEXTS = (
pep543.stdlib.STDLIB_BACKEND.client_context,
pep543.stdlib.STDLIB_BACKEND.server_context
)
def assert_wrap_fails(context, exception):
"""
A convenient helper that calls wrap_buffers with the appropriate number of
arugments and asserts that it raises the appropriate error.
"""
if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context):
with pytest.raises(exception):
context.wrap_buffers(server_hostname=None)
else:
with pytest.raises(exception):
context.wrap_buffers()
class TestSimpleNegotiationStdlib(SimpleNegotiation):
BACKEND = pep543.stdlib.STDLIB_BACKEND
class TestStdlibErrorHandling(object):
"""
Validate that the stdlib backend can do sensible error handling in specific
situations that it cannot handle.
"""
@pytest.mark.parametrize(
'lowest,highest', (
(object(), None), (None, object()), (object(), object())
)
)
@pytest.mark.parametrize('context', CONTEXTS)
def test_bad_values_for_versions_client(self, lowest, highest, context):
"""
Using TLSConfiguration objects with a bad value for their minimum
version raises a TLSError with Client contexts.
"""
config = pep543.TLSConfiguration(
validate_certificates=False,
lowest_supported_version=lowest,
highest_supported_version=highest
)
ctx = context(config)
assert_wrap_fails(ctx, pep543.TLSError)
|
// ... existing code ...
import pep543
import pep543.stdlib
import pytest
// ... modified code ...
CONTEXTS = (
pep543.stdlib.STDLIB_BACKEND.client_context,
pep543.stdlib.STDLIB_BACKEND.server_context
)
def assert_wrap_fails(context, exception):
"""
A convenient helper that calls wrap_buffers with the appropriate number of
arugments and asserts that it raises the appropriate error.
"""
if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context):
with pytest.raises(exception):
context.wrap_buffers(server_hostname=None)
else:
with pytest.raises(exception):
context.wrap_buffers()
class TestSimpleNegotiationStdlib(SimpleNegotiation):
...
BACKEND = pep543.stdlib.STDLIB_BACKEND
class TestStdlibErrorHandling(object):
"""
Validate that the stdlib backend can do sensible error handling in specific
situations that it cannot handle.
"""
@pytest.mark.parametrize(
'lowest,highest', (
(object(), None), (None, object()), (object(), object())
)
)
@pytest.mark.parametrize('context', CONTEXTS)
def test_bad_values_for_versions_client(self, lowest, highest, context):
"""
Using TLSConfiguration objects with a bad value for their minimum
version raises a TLSError with Client contexts.
"""
config = pep543.TLSConfiguration(
validate_certificates=False,
lowest_supported_version=lowest,
highest_supported_version=highest
)
ctx = context(config)
assert_wrap_fails(ctx, pep543.TLSError)
// ... rest of the code ...
|
332452cf7ccd6d3ee583be9a6aac27b14771263f
|
source/services/omdb_service.py
|
source/services/omdb_service.py
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class OmdbService:
__API_URL = 'http://www.omdbapi.com/?'
def __init__(self, movie_id):
self.id = movie_id
def get_rt_rating(self):
payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'}
response = requests.post(self.__API_URL, params=payload)
movie_info = response.json()
ratings = []
ratings.append(movie_info['tomatoMeter'])
ratings.append(movie_info['tomatoUserMeter'])
return RTRating(ratings)
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class OmdbService:
__API_URL = 'http://www.omdbapi.com/?'
def __init__(self, movie_id):
self.id = movie_id
def get_rt_rating(self):
payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'}
response = requests.post(self.__API_URL, params=payload)
movie_info = response.json()
scores = []
scores.append(movie_info['tomatoMeter'])
scores.append(movie_info['tomatoUserMeter'])
rt_rating = RTRating(scores)
rt_rating.link = movie_info['tomatoURL']
return rt_rating
|
Add url to RTRating object
|
Add url to RTRating object
|
Python
|
mit
|
jeremyrea/caterblu,jeremyrea/caterblu,jeremyrea/caterblu,jeremyrea/caterblu
|
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class OmdbService:
__API_URL = 'http://www.omdbapi.com/?'
def __init__(self, movie_id):
self.id = movie_id
def get_rt_rating(self):
payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'}
response = requests.post(self.__API_URL, params=payload)
movie_info = response.json()
- ratings = []
+ scores = []
- ratings.append(movie_info['tomatoMeter'])
+ scores.append(movie_info['tomatoMeter'])
- ratings.append(movie_info['tomatoUserMeter'])
+ scores.append(movie_info['tomatoUserMeter'])
- return RTRating(ratings)
+ rt_rating = RTRating(scores)
+ rt_rating.link = movie_info['tomatoURL']
+ return rt_rating
+
|
Add url to RTRating object
|
## Code Before:
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class OmdbService:
__API_URL = 'http://www.omdbapi.com/?'
def __init__(self, movie_id):
self.id = movie_id
def get_rt_rating(self):
payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'}
response = requests.post(self.__API_URL, params=payload)
movie_info = response.json()
ratings = []
ratings.append(movie_info['tomatoMeter'])
ratings.append(movie_info['tomatoUserMeter'])
return RTRating(ratings)
## Instruction:
Add url to RTRating object
## Code After:
import requests
from bs4 import BeautifulSoup
from source.models.rt_rating import RTRating
class OmdbService:
__API_URL = 'http://www.omdbapi.com/?'
def __init__(self, movie_id):
self.id = movie_id
def get_rt_rating(self):
payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'}
response = requests.post(self.__API_URL, params=payload)
movie_info = response.json()
scores = []
scores.append(movie_info['tomatoMeter'])
scores.append(movie_info['tomatoUserMeter'])
rt_rating = RTRating(scores)
rt_rating.link = movie_info['tomatoURL']
return rt_rating
|
# ... existing code ...
scores = []
scores.append(movie_info['tomatoMeter'])
scores.append(movie_info['tomatoUserMeter'])
rt_rating = RTRating(scores)
rt_rating.link = movie_info['tomatoURL']
return rt_rating
# ... rest of the code ...
|
e120aafad13138abb5d98bbad12c0d8bdc532b30
|
lglass/web/application.py
|
lglass/web/application.py
|
import bottle
from lglass.web.helpers import render_template, with_config
app = bottle.Bottle()
def index_handler():
return render_template("index.html")
@with_config
def robots_txt_handler(config):
if config["robots.txt"] is not None:
return bottle.static_file(config["robots.txt"])
bottle.abort(404, "File not found")
app.route("/", "GET", index_handler)
app.route("/robots.txt", "GET", robots_txt_handler)
import lglass.web.registry
app.route("/obj", "GET", lglass.web.registry.show_object_types)
app.route("/obj/<type>", "GET", lglass.web.registry.show_objects)
app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object)
app.route("/whois/<query>", "GET", lglass.web.registry.whois_query)
app.route("/whois", "POST", lglass.web.registry.whois_query)
app.route("/flush", "POST", lglass.web.registry.flush_cache)
|
import bottle
from lglass.web.helpers import render_template, with_config
app = bottle.Bottle()
def index_handler():
return render_template("index.html")
@with_config
def robots_txt_handler(config):
if config["robots.txt"] is not None:
return open(config["robots.txt"])
bottle.abort(404, "File not found")
app.route("/", "GET", index_handler)
app.route("/robots.txt", "GET", robots_txt_handler)
import lglass.web.registry
app.route("/obj", "GET", lglass.web.registry.show_object_types)
app.route("/obj/<type>", "GET", lglass.web.registry.show_objects)
app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object)
app.route("/whois/<query>", "GET", lglass.web.registry.whois_query)
app.route("/whois", "POST", lglass.web.registry.whois_query)
app.route("/flush", "POST", lglass.web.registry.flush_cache)
|
Replace static_file by open call
|
Replace static_file by open call
|
Python
|
mit
|
fritz0705/lglass
|
import bottle
from lglass.web.helpers import render_template, with_config
app = bottle.Bottle()
def index_handler():
return render_template("index.html")
@with_config
def robots_txt_handler(config):
if config["robots.txt"] is not None:
- return bottle.static_file(config["robots.txt"])
+ return open(config["robots.txt"])
bottle.abort(404, "File not found")
app.route("/", "GET", index_handler)
app.route("/robots.txt", "GET", robots_txt_handler)
import lglass.web.registry
app.route("/obj", "GET", lglass.web.registry.show_object_types)
app.route("/obj/<type>", "GET", lglass.web.registry.show_objects)
app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object)
app.route("/whois/<query>", "GET", lglass.web.registry.whois_query)
app.route("/whois", "POST", lglass.web.registry.whois_query)
app.route("/flush", "POST", lglass.web.registry.flush_cache)
|
Replace static_file by open call
|
## Code Before:
import bottle
from lglass.web.helpers import render_template, with_config
app = bottle.Bottle()
def index_handler():
return render_template("index.html")
@with_config
def robots_txt_handler(config):
if config["robots.txt"] is not None:
return bottle.static_file(config["robots.txt"])
bottle.abort(404, "File not found")
app.route("/", "GET", index_handler)
app.route("/robots.txt", "GET", robots_txt_handler)
import lglass.web.registry
app.route("/obj", "GET", lglass.web.registry.show_object_types)
app.route("/obj/<type>", "GET", lglass.web.registry.show_objects)
app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object)
app.route("/whois/<query>", "GET", lglass.web.registry.whois_query)
app.route("/whois", "POST", lglass.web.registry.whois_query)
app.route("/flush", "POST", lglass.web.registry.flush_cache)
## Instruction:
Replace static_file by open call
## Code After:
import bottle
from lglass.web.helpers import render_template, with_config
app = bottle.Bottle()
def index_handler():
return render_template("index.html")
@with_config
def robots_txt_handler(config):
if config["robots.txt"] is not None:
return open(config["robots.txt"])
bottle.abort(404, "File not found")
app.route("/", "GET", index_handler)
app.route("/robots.txt", "GET", robots_txt_handler)
import lglass.web.registry
app.route("/obj", "GET", lglass.web.registry.show_object_types)
app.route("/obj/<type>", "GET", lglass.web.registry.show_objects)
app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object)
app.route("/whois/<query>", "GET", lglass.web.registry.whois_query)
app.route("/whois", "POST", lglass.web.registry.whois_query)
app.route("/flush", "POST", lglass.web.registry.flush_cache)
|
...
if config["robots.txt"] is not None:
return open(config["robots.txt"])
bottle.abort(404, "File not found")
...
|
8e0f2271b19504886728ccf5d060778c027c79ca
|
ide/views.py
|
ide/views.py
|
import json
from werkzeug.routing import BaseConverter
from flask import render_template, request, abort
import requests
from ide import app
from ide.projects import get_all_projects, Project
MCLABAAS_URL = 'http://localhost:4242'
@app.route('/')
def index():
return render_template('index.html', projects=get_all_projects())
@app.route('/parse', methods=['POST'])
def parse():
return requests.post(MCLABAAS_URL + '/ast', data=request.data).text
class ProjectConverter(BaseConverter):
def to_python(self, value):
return Project(value)
def to_url(self, value):
return BaseConverter.to_url(value.name)
app.url_map.converters['project'] = ProjectConverter
@app.route('/project/<project:project>/')
def project(project):
if not project.exists():
abort(404)
return render_template('project.html')
@app.route('/project/<project:project>/tree', methods=['GET'])
def tree(project):
return json.dumps(project.tree())
@app.route('/project/<project:project>/read', methods=['GET'])
def read(project):
return project.read_file(request.args['path'])
@app.route('/project/<project:project>/write', methods=['POST'])
def write(project):
project.write_file(request.form['path'], request.form['contents'])
return json.dumps({'status': 'OK'})
|
import json
from werkzeug.routing import BaseConverter
from flask import render_template, request, abort
import requests
from ide import app
from ide.projects import get_all_projects, Project
MCLABAAS_URL = 'http://localhost:4242'
@app.route('/')
def index():
return render_template('index.html', projects=get_all_projects())
@app.route('/parse', methods=['POST'])
def parse():
return requests.post(MCLABAAS_URL + '/ast', data=request.data).text
class ProjectConverter(BaseConverter):
def to_python(self, value):
project = Project(value)
if not project.exists():
abort(404)
return project
def to_url(self, value):
return BaseConverter.to_url(value.name)
app.url_map.converters['project'] = ProjectConverter
@app.route('/project/<project:project>/')
def project(project):
return render_template('project.html')
@app.route('/project/<project:project>/tree', methods=['GET'])
def tree(project):
return json.dumps(project.tree())
@app.route('/project/<project:project>/read', methods=['GET'])
def read(project):
return project.read_file(request.args['path'])
@app.route('/project/<project:project>/write', methods=['POST'])
def write(project):
project.write_file(request.form['path'], request.form['contents'])
return json.dumps({'status': 'OK'})
|
Check if project exists inside ProjectConverter.
|
Check if project exists inside ProjectConverter.
|
Python
|
apache-2.0
|
Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide
|
import json
from werkzeug.routing import BaseConverter
from flask import render_template, request, abort
import requests
from ide import app
from ide.projects import get_all_projects, Project
MCLABAAS_URL = 'http://localhost:4242'
@app.route('/')
def index():
return render_template('index.html', projects=get_all_projects())
@app.route('/parse', methods=['POST'])
def parse():
return requests.post(MCLABAAS_URL + '/ast', data=request.data).text
class ProjectConverter(BaseConverter):
def to_python(self, value):
- return Project(value)
+ project = Project(value)
+ if not project.exists():
+ abort(404)
+ return project
def to_url(self, value):
return BaseConverter.to_url(value.name)
app.url_map.converters['project'] = ProjectConverter
@app.route('/project/<project:project>/')
def project(project):
- if not project.exists():
- abort(404)
return render_template('project.html')
@app.route('/project/<project:project>/tree', methods=['GET'])
def tree(project):
return json.dumps(project.tree())
@app.route('/project/<project:project>/read', methods=['GET'])
def read(project):
return project.read_file(request.args['path'])
@app.route('/project/<project:project>/write', methods=['POST'])
def write(project):
project.write_file(request.form['path'], request.form['contents'])
return json.dumps({'status': 'OK'})
|
Check if project exists inside ProjectConverter.
|
## Code Before:
import json
from werkzeug.routing import BaseConverter
from flask import render_template, request, abort
import requests
from ide import app
from ide.projects import get_all_projects, Project
MCLABAAS_URL = 'http://localhost:4242'
@app.route('/')
def index():
return render_template('index.html', projects=get_all_projects())
@app.route('/parse', methods=['POST'])
def parse():
return requests.post(MCLABAAS_URL + '/ast', data=request.data).text
class ProjectConverter(BaseConverter):
def to_python(self, value):
return Project(value)
def to_url(self, value):
return BaseConverter.to_url(value.name)
app.url_map.converters['project'] = ProjectConverter
@app.route('/project/<project:project>/')
def project(project):
if not project.exists():
abort(404)
return render_template('project.html')
@app.route('/project/<project:project>/tree', methods=['GET'])
def tree(project):
return json.dumps(project.tree())
@app.route('/project/<project:project>/read', methods=['GET'])
def read(project):
return project.read_file(request.args['path'])
@app.route('/project/<project:project>/write', methods=['POST'])
def write(project):
project.write_file(request.form['path'], request.form['contents'])
return json.dumps({'status': 'OK'})
## Instruction:
Check if project exists inside ProjectConverter.
## Code After:
import json
from werkzeug.routing import BaseConverter
from flask import render_template, request, abort
import requests
from ide import app
from ide.projects import get_all_projects, Project
MCLABAAS_URL = 'http://localhost:4242'
@app.route('/')
def index():
return render_template('index.html', projects=get_all_projects())
@app.route('/parse', methods=['POST'])
def parse():
return requests.post(MCLABAAS_URL + '/ast', data=request.data).text
class ProjectConverter(BaseConverter):
def to_python(self, value):
project = Project(value)
if not project.exists():
abort(404)
return project
def to_url(self, value):
return BaseConverter.to_url(value.name)
app.url_map.converters['project'] = ProjectConverter
@app.route('/project/<project:project>/')
def project(project):
return render_template('project.html')
@app.route('/project/<project:project>/tree', methods=['GET'])
def tree(project):
return json.dumps(project.tree())
@app.route('/project/<project:project>/read', methods=['GET'])
def read(project):
return project.read_file(request.args['path'])
@app.route('/project/<project:project>/write', methods=['POST'])
def write(project):
project.write_file(request.form['path'], request.form['contents'])
return json.dumps({'status': 'OK'})
|
...
def to_python(self, value):
project = Project(value)
if not project.exists():
abort(404)
return project
...
def project(project):
return render_template('project.html')
...
|
5cf0e2e9d68d2e0fca3780608f33d5d8cdaef8a9
|
admin/metrics/views.py
|
admin/metrics/views.py
|
from django.views.generic import TemplateView
from django.contrib.auth.mixins import PermissionRequiredMixin
from admin.base.settings import KEEN_CREDENTIALS
from admin.base.utils import OSFAdmin
class MetricsView(OSFAdmin, TemplateView, PermissionRequiredMixin):
template_name = 'metrics/osf_metrics.html'
permission_required = 'admin.view_metrics'
def get_context_data(self, **kwargs):
kwargs.update(KEEN_CREDENTIALS.copy())
return super(MetricsView, self).get_context_data(**kwargs)
|
from django.views.generic import TemplateView
from django.contrib.auth.mixins import PermissionRequiredMixin
from admin.base.settings import KEEN_CREDENTIALS
class MetricsView(TemplateView, PermissionRequiredMixin):
template_name = 'metrics/osf_metrics.html'
permission_required = 'admin.view_metrics'
def get_context_data(self, **kwargs):
kwargs.update(KEEN_CREDENTIALS.copy())
return super(MetricsView, self).get_context_data(**kwargs)
|
Remove one more reference to old group permissions
|
Remove one more reference to old group permissions
|
Python
|
apache-2.0
|
erinspace/osf.io,adlius/osf.io,hmoco/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,cslzchen/osf.io,saradbowman/osf.io,icereval/osf.io,icereval/osf.io,leb2dg/osf.io,icereval/osf.io,sloria/osf.io,caseyrollins/osf.io,sloria/osf.io,Nesiehr/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,felliott/osf.io,monikagrabowska/osf.io,brianjgeiger/osf.io,mattclark/osf.io,cwisecarver/osf.io,mattclark/osf.io,monikagrabowska/osf.io,cwisecarver/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,laurenrevere/osf.io,chennan47/osf.io,acshi/osf.io,Nesiehr/osf.io,laurenrevere/osf.io,erinspace/osf.io,acshi/osf.io,erinspace/osf.io,mfraezz/osf.io,mfraezz/osf.io,Johnetordoff/osf.io,aaxelb/osf.io,aaxelb/osf.io,hmoco/osf.io,binoculars/osf.io,brianjgeiger/osf.io,pattisdr/osf.io,binoculars/osf.io,mattclark/osf.io,hmoco/osf.io,chennan47/osf.io,chrisseto/osf.io,laurenrevere/osf.io,CenterForOpenScience/osf.io,chennan47/osf.io,caneruguz/osf.io,hmoco/osf.io,monikagrabowska/osf.io,adlius/osf.io,monikagrabowska/osf.io,TomBaxter/osf.io,aaxelb/osf.io,baylee-d/osf.io,mfraezz/osf.io,saradbowman/osf.io,chrisseto/osf.io,sloria/osf.io,caseyrollins/osf.io,caneruguz/osf.io,Nesiehr/osf.io,crcresearch/osf.io,leb2dg/osf.io,HalcyonChimera/osf.io,pattisdr/osf.io,crcresearch/osf.io,acshi/osf.io,acshi/osf.io,brianjgeiger/osf.io,adlius/osf.io,felliott/osf.io,leb2dg/osf.io,Johnetordoff/osf.io,cwisecarver/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,leb2dg/osf.io,CenterForOpenScience/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,felliott/osf.io,aaxelb/osf.io,acshi/osf.io,binoculars/osf.io,baylee-d/osf.io,chrisseto/osf.io,caneruguz/osf.io,CenterForOpenScience/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,pattisdr/osf.io,felliott/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,adlius/osf.io,caneruguz/osf.io,baylee-d/osf.io,caseyrollins/osf.io,cwisecarver/osf.io,crcresearch/osf.io,chrisseto/osf.io,Nesiehr/osf.io
|
from django.views.generic import TemplateView
from django.contrib.auth.mixins import PermissionRequiredMixin
from admin.base.settings import KEEN_CREDENTIALS
- from admin.base.utils import OSFAdmin
-
-
- class MetricsView(OSFAdmin, TemplateView, PermissionRequiredMixin):
+ class MetricsView(TemplateView, PermissionRequiredMixin):
template_name = 'metrics/osf_metrics.html'
permission_required = 'admin.view_metrics'
def get_context_data(self, **kwargs):
kwargs.update(KEEN_CREDENTIALS.copy())
return super(MetricsView, self).get_context_data(**kwargs)
|
Remove one more reference to old group permissions
|
## Code Before:
from django.views.generic import TemplateView
from django.contrib.auth.mixins import PermissionRequiredMixin
from admin.base.settings import KEEN_CREDENTIALS
from admin.base.utils import OSFAdmin
class MetricsView(OSFAdmin, TemplateView, PermissionRequiredMixin):
template_name = 'metrics/osf_metrics.html'
permission_required = 'admin.view_metrics'
def get_context_data(self, **kwargs):
kwargs.update(KEEN_CREDENTIALS.copy())
return super(MetricsView, self).get_context_data(**kwargs)
## Instruction:
Remove one more reference to old group permissions
## Code After:
from django.views.generic import TemplateView
from django.contrib.auth.mixins import PermissionRequiredMixin
from admin.base.settings import KEEN_CREDENTIALS
class MetricsView(TemplateView, PermissionRequiredMixin):
template_name = 'metrics/osf_metrics.html'
permission_required = 'admin.view_metrics'
def get_context_data(self, **kwargs):
kwargs.update(KEEN_CREDENTIALS.copy())
return super(MetricsView, self).get_context_data(**kwargs)
|
# ... existing code ...
class MetricsView(TemplateView, PermissionRequiredMixin):
template_name = 'metrics/osf_metrics.html'
# ... rest of the code ...
|
1cf82c6efa0550c5a0ba7160f82f77db6e3358ec
|
panoptes/test/test_mount.py
|
panoptes/test/test_mount.py
|
from panoptes.mount.ioptron import iOptronMount
class TestOptronMount:
mount = None
def setup(self):
print ("TestMount:setup() before each test method")
def teardown(self):
print ("TestMount:teardown() after each test method")
@classmethod
def setup_class(cls):
print ("setup_class() before any methods in this class")
_Mounts = []
for name in os.listdir(os.path.dirname(__file__)):
if not name.startswith('_') and name.endswith('.py'):
name = '.' + os.path.splitext(name)[0]
try:
module = importlib.import_module(name,'panoptes')
_Mounts.append(module)
except ImportError as err:
self.logger.warn('Failed to load mount plugin: {}'.format(err))
@classmethod
def teardown_class(cls):
print ("teardown_class() after any methods in this class")
def test_is_connected_false(self):
pass
def test_connect(self):
pass
def test_is_connected_true(self):
pass
def test_is_slewing(self):
pass
def test_check_coordinates(self):
pass
def test_sync_coordinates(self):
pass
def test_slew_to_coordinates(self):
pass
def test_slew_to_park(self):
pass
def test_echo(self):
pass
|
import os
import importlib
import warnings
class TestOptronMount:
mount = None
def setup(self):
print ("TestMount:setup() before each test method")
def teardown(self):
print ("TestMount:teardown() after each test method")
@classmethod
def setup_class(cls):
mount_dir = os.path.dirname(__file__) + '/../mount/'
print ("setup_class() before any methods in this class")
_Mounts = []
for name in os.listdir(os.path.dirname(mount_dir)):
if not name.startswith('_') and name.endswith('.py'):
name = '.' + os.path.splitext(name)[0]
try:
module = importlib.import_module(name,'panoptes.mount')
_Mounts.append(module)
except ImportError as err:
warnings.warn('Failed to load mount plugin: {}'.format(err))
@classmethod
def teardown_class(cls):
print ("teardown_class() after any methods in this class")
def test_is_connected_false(self):
pass
def test_connect(self):
pass
def test_is_connected_true(self):
pass
def test_is_slewing(self):
pass
def test_check_coordinates(self):
pass
def test_sync_coordinates(self):
pass
def test_slew_to_coordinates(self):
pass
def test_slew_to_park(self):
pass
def test_echo(self):
pass
|
Test file loops over all the mounts
|
Test file loops over all the mounts
|
Python
|
mit
|
AstroHuntsman/POCS,AstroHuntsman/POCS,panoptes/POCS,fmin2958/POCS,joshwalawender/POCS,AstroHuntsman/POCS,panoptes/POCS,fmin2958/POCS,Guokr1991/POCS,panoptes/POCS,Guokr1991/POCS,joshwalawender/POCS,Guokr1991/POCS,Guokr1991/POCS,panoptes/POCS,AstroHuntsman/POCS,joshwalawender/POCS,fmin2958/POCS
|
- from panoptes.mount.ioptron import iOptronMount
+ import os
+ import importlib
+ import warnings
class TestOptronMount:
mount = None
def setup(self):
print ("TestMount:setup() before each test method")
def teardown(self):
print ("TestMount:teardown() after each test method")
@classmethod
def setup_class(cls):
+ mount_dir = os.path.dirname(__file__) + '/../mount/'
print ("setup_class() before any methods in this class")
_Mounts = []
- for name in os.listdir(os.path.dirname(__file__)):
+ for name in os.listdir(os.path.dirname(mount_dir)):
if not name.startswith('_') and name.endswith('.py'):
name = '.' + os.path.splitext(name)[0]
try:
- module = importlib.import_module(name,'panoptes')
+ module = importlib.import_module(name,'panoptes.mount')
_Mounts.append(module)
except ImportError as err:
- self.logger.warn('Failed to load mount plugin: {}'.format(err))
+ warnings.warn('Failed to load mount plugin: {}'.format(err))
@classmethod
def teardown_class(cls):
print ("teardown_class() after any methods in this class")
def test_is_connected_false(self):
pass
def test_connect(self):
pass
def test_is_connected_true(self):
pass
def test_is_slewing(self):
pass
def test_check_coordinates(self):
pass
def test_sync_coordinates(self):
pass
def test_slew_to_coordinates(self):
pass
def test_slew_to_park(self):
pass
def test_echo(self):
pass
|
Test file loops over all the mounts
|
## Code Before:
from panoptes.mount.ioptron import iOptronMount
class TestOptronMount:
mount = None
def setup(self):
print ("TestMount:setup() before each test method")
def teardown(self):
print ("TestMount:teardown() after each test method")
@classmethod
def setup_class(cls):
print ("setup_class() before any methods in this class")
_Mounts = []
for name in os.listdir(os.path.dirname(__file__)):
if not name.startswith('_') and name.endswith('.py'):
name = '.' + os.path.splitext(name)[0]
try:
module = importlib.import_module(name,'panoptes')
_Mounts.append(module)
except ImportError as err:
self.logger.warn('Failed to load mount plugin: {}'.format(err))
@classmethod
def teardown_class(cls):
print ("teardown_class() after any methods in this class")
def test_is_connected_false(self):
pass
def test_connect(self):
pass
def test_is_connected_true(self):
pass
def test_is_slewing(self):
pass
def test_check_coordinates(self):
pass
def test_sync_coordinates(self):
pass
def test_slew_to_coordinates(self):
pass
def test_slew_to_park(self):
pass
def test_echo(self):
pass
## Instruction:
Test file loops over all the mounts
## Code After:
import os
import importlib
import warnings
class TestOptronMount:
mount = None
def setup(self):
print ("TestMount:setup() before each test method")
def teardown(self):
print ("TestMount:teardown() after each test method")
@classmethod
def setup_class(cls):
mount_dir = os.path.dirname(__file__) + '/../mount/'
print ("setup_class() before any methods in this class")
_Mounts = []
for name in os.listdir(os.path.dirname(mount_dir)):
if not name.startswith('_') and name.endswith('.py'):
name = '.' + os.path.splitext(name)[0]
try:
module = importlib.import_module(name,'panoptes.mount')
_Mounts.append(module)
except ImportError as err:
warnings.warn('Failed to load mount plugin: {}'.format(err))
@classmethod
def teardown_class(cls):
print ("teardown_class() after any methods in this class")
def test_is_connected_false(self):
pass
def test_connect(self):
pass
def test_is_connected_true(self):
pass
def test_is_slewing(self):
pass
def test_check_coordinates(self):
pass
def test_sync_coordinates(self):
pass
def test_slew_to_coordinates(self):
pass
def test_slew_to_park(self):
pass
def test_echo(self):
pass
|
...
import os
import importlib
import warnings
...
def setup_class(cls):
mount_dir = os.path.dirname(__file__) + '/../mount/'
print ("setup_class() before any methods in this class")
...
_Mounts = []
for name in os.listdir(os.path.dirname(mount_dir)):
if not name.startswith('_') and name.endswith('.py'):
...
try:
module = importlib.import_module(name,'panoptes.mount')
_Mounts.append(module)
...
except ImportError as err:
warnings.warn('Failed to load mount plugin: {}'.format(err))
...
|
8898f23a429112cd80e6a2c8321b0de44aeaee7e
|
blanc_basic_pages/forms.py
|
blanc_basic_pages/forms.py
|
from django import forms
from django.conf import settings
from mptt.forms import MPTTAdminForm
from .models import Page
TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', (
('', 'Default'),
))
class PageAdminForm(MPTTAdminForm):
class Meta:
model = Page
exclude = ()
def __init__(self, *args, **kwargs):
super(PageAdminForm, self).__init__(*args, **kwargs)
# The list of templates is defined in settings, however as we can't have dynamic choices in
# models due to migrations - we change the form choices instead.
self.fields['template_name'] = forms.ChoiceField(choices=TEMPLATE_CHOICES, required=False)
|
from django import forms
from django.conf import settings
from mptt.forms import MPTTAdminForm
from .models import Page
TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', (
('', 'Default'),
))
class PageAdminForm(MPTTAdminForm):
class Meta:
model = Page
exclude = ()
widgets = {
# The list of templates is defined in settings, however as we can't have dynamic
# choices in models due to migrations - we change the form choices instead.
'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES),
}
|
Use custom widget for template choices instead
|
Use custom widget for template choices instead
A bit more Djangonic than tweaking self.fields
|
Python
|
bsd-3-clause
|
blancltd/blanc-basic-pages
|
from django import forms
from django.conf import settings
from mptt.forms import MPTTAdminForm
from .models import Page
TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', (
('', 'Default'),
))
class PageAdminForm(MPTTAdminForm):
class Meta:
model = Page
exclude = ()
+ widgets = {
+ # The list of templates is defined in settings, however as we can't have dynamic
+ # choices in models due to migrations - we change the form choices instead.
+ 'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES),
+ }
- def __init__(self, *args, **kwargs):
- super(PageAdminForm, self).__init__(*args, **kwargs)
-
- # The list of templates is defined in settings, however as we can't have dynamic choices in
- # models due to migrations - we change the form choices instead.
- self.fields['template_name'] = forms.ChoiceField(choices=TEMPLATE_CHOICES, required=False)
-
|
Use custom widget for template choices instead
|
## Code Before:
from django import forms
from django.conf import settings
from mptt.forms import MPTTAdminForm
from .models import Page
TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', (
('', 'Default'),
))
class PageAdminForm(MPTTAdminForm):
class Meta:
model = Page
exclude = ()
def __init__(self, *args, **kwargs):
super(PageAdminForm, self).__init__(*args, **kwargs)
# The list of templates is defined in settings, however as we can't have dynamic choices in
# models due to migrations - we change the form choices instead.
self.fields['template_name'] = forms.ChoiceField(choices=TEMPLATE_CHOICES, required=False)
## Instruction:
Use custom widget for template choices instead
## Code After:
from django import forms
from django.conf import settings
from mptt.forms import MPTTAdminForm
from .models import Page
TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', (
('', 'Default'),
))
class PageAdminForm(MPTTAdminForm):
class Meta:
model = Page
exclude = ()
widgets = {
# The list of templates is defined in settings, however as we can't have dynamic
# choices in models due to migrations - we change the form choices instead.
'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES),
}
|
# ... existing code ...
exclude = ()
widgets = {
# The list of templates is defined in settings, however as we can't have dynamic
# choices in models due to migrations - we change the form choices instead.
'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES),
}
# ... rest of the code ...
|
b63b22678a005baa6195854b65cc1828061febba
|
vx/mode.py
|
vx/mode.py
|
import vx
import os.path
def mode_from_filename(file):
root, ext = os.path.splitext(file)
ext = ext if ext else root
mode = None
if ext == '.c':
return c_mode
class mode:
def __init__(self, window):
self.breaks = ('_', ' ', '\n', '\t')
self.keywords = ()
class python_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#')
self.keywords = ('return', 'for', 'while', 'break', 'continue', 'def')
class c_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#')
self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
|
import vx
import os.path
def mode_from_filename(file):
root, ext = os.path.splitext(file)
ext = ext if ext else root
mode = None
if ext == '.c':
return c_mode
elif ext == '.py':
return python_mode
class mode:
def __init__(self, window):
self.breaks = ('_', ' ', '\n', '\t')
self.keywords = ()
class python_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#')
self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import')
class c_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#')
self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
|
Add .py extension handling and more python keywords
|
Add .py extension handling and more python keywords
|
Python
|
mit
|
philipdexter/vx,philipdexter/vx
|
import vx
import os.path
def mode_from_filename(file):
root, ext = os.path.splitext(file)
ext = ext if ext else root
mode = None
if ext == '.c':
return c_mode
+ elif ext == '.py':
+ return python_mode
class mode:
def __init__(self, window):
self.breaks = ('_', ' ', '\n', '\t')
self.keywords = ()
class python_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#')
- self.keywords = ('return', 'for', 'while', 'break', 'continue', 'def')
+ self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import')
class c_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#')
self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
|
Add .py extension handling and more python keywords
|
## Code Before:
import vx
import os.path
def mode_from_filename(file):
root, ext = os.path.splitext(file)
ext = ext if ext else root
mode = None
if ext == '.c':
return c_mode
class mode:
def __init__(self, window):
self.breaks = ('_', ' ', '\n', '\t')
self.keywords = ()
class python_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#')
self.keywords = ('return', 'for', 'while', 'break', 'continue', 'def')
class c_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#')
self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
## Instruction:
Add .py extension handling and more python keywords
## Code After:
import vx
import os.path
def mode_from_filename(file):
root, ext = os.path.splitext(file)
ext = ext if ext else root
mode = None
if ext == '.c':
return c_mode
elif ext == '.py':
return python_mode
class mode:
def __init__(self, window):
self.breaks = ('_', ' ', '\n', '\t')
self.keywords = ()
class python_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#')
self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import')
class c_mode(mode):
def __init__(self, window):
super().__init__(window)
self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#')
self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
|
// ... existing code ...
return c_mode
elif ext == '.py':
return python_mode
// ... modified code ...
self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import')
// ... rest of the code ...
|
605443886582d13c2b45b19fad86854bf4e8ddbd
|
backend/catalogue/serializers.py
|
backend/catalogue/serializers.py
|
from rest_framework import serializers
from .models import Release, Track, Comment
class CommentSerializer(serializers.ModelSerializer):
class Meta:
model = Comment
fields = ('id', 'comment')
class TrackSerializer(serializers.ModelSerializer):
cdid = serializers.StringRelatedField(
read_only=True
)
class Meta:
model = Track
fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'cdid')
class ReleaseSerializer(serializers.ModelSerializer):
tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks')
comments = serializers.HyperlinkedIdentityField(view_name='release-comments')
class Meta:
model = Release
fields = ('id', 'arrivaldate', 'artist', 'title', 'year', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
|
from rest_framework import serializers
from .models import Release, Track, Comment
class CommentSerializer(serializers.ModelSerializer):
class Meta:
model = Comment
fields = ('id', 'comment')
class TrackSerializer(serializers.ModelSerializer):
class Meta:
model = Track
fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release')
class ReleaseSerializer(serializers.ModelSerializer):
tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks')
comments = serializers.HyperlinkedIdentityField(view_name='release-comments')
class Meta:
model = Release
fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
|
Add more fields to Release serializer.
|
Add more fields to Release serializer.
|
Python
|
mit
|
ThreeDRadio/playlists,ThreeDRadio/playlists,ThreeDRadio/playlists
|
from rest_framework import serializers
from .models import Release, Track, Comment
class CommentSerializer(serializers.ModelSerializer):
class Meta:
model = Comment
fields = ('id', 'comment')
class TrackSerializer(serializers.ModelSerializer):
- cdid = serializers.StringRelatedField(
- read_only=True
- )
class Meta:
model = Track
- fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'cdid')
+ fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release')
class ReleaseSerializer(serializers.ModelSerializer):
tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks')
comments = serializers.HyperlinkedIdentityField(view_name='release-comments')
class Meta:
model = Release
- fields = ('id', 'arrivaldate', 'artist', 'title', 'year', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
+ fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
|
Add more fields to Release serializer.
|
## Code Before:
from rest_framework import serializers
from .models import Release, Track, Comment
class CommentSerializer(serializers.ModelSerializer):
class Meta:
model = Comment
fields = ('id', 'comment')
class TrackSerializer(serializers.ModelSerializer):
cdid = serializers.StringRelatedField(
read_only=True
)
class Meta:
model = Track
fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'cdid')
class ReleaseSerializer(serializers.ModelSerializer):
tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks')
comments = serializers.HyperlinkedIdentityField(view_name='release-comments')
class Meta:
model = Release
fields = ('id', 'arrivaldate', 'artist', 'title', 'year', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
## Instruction:
Add more fields to Release serializer.
## Code After:
from rest_framework import serializers
from .models import Release, Track, Comment
class CommentSerializer(serializers.ModelSerializer):
class Meta:
model = Comment
fields = ('id', 'comment')
class TrackSerializer(serializers.ModelSerializer):
class Meta:
model = Track
fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release')
class ReleaseSerializer(serializers.ModelSerializer):
tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks')
comments = serializers.HyperlinkedIdentityField(view_name='release-comments')
class Meta:
model = Release
fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
|
// ... existing code ...
class TrackSerializer(serializers.ModelSerializer):
// ... modified code ...
model = Track
fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release')
...
model = Release
fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
// ... rest of the code ...
|
c1e1c9d63d5334140aa71c025a90e9500b299307
|
functional_tests.py
|
functional_tests.py
|
from selenium import webdriver
browser = webdriver.Firefox()
browser.get('http://localhost:8000')
assert 'Django' in browser.title
|
from selenium import webdriver
import unittest
class NewVisitorTest(unittest.TestCase):
def setUp(self):
self.browser = webdriver.Firefox()
def tearDown(self):
self.browser.quit()
def test_can_start_a_list_and_retrieve_it_later(self):
# Rey has heard about a cool new online to-do app. He goes
# to check out its homepage
self.browser.get('http://localhost:8000')
# He notices the page title and header mention to-do lists
self.assertIn('To-Do', self.browser.title)
self.fail('Finish the test!')
# He is invited to enter a to-do item straight away
# He types "Make coffee" into a text box (Rey can't work when he is sleepy)
# There is still a text box inviting him to add another item.
# He enters "Go to the toilet" (Rey is very methodical)
# The page updates again, and now shows both items on his list
# Rey wonders whether the site will remember his list. Then he sees that
# the site has generated a unique URL for him
# He visits that URL - his to-do list is still there
# Satisfied, he goes back to sleep
if __name__ == '__main__':
unittest.main(warnings='ignore')
|
Update tests to be human-readable
|
Update tests to be human-readable
|
Python
|
apache-2.0
|
rocity/the-testing-goat,rocity/the-testing-goat
|
from selenium import webdriver
+ import unittest
+ class NewVisitorTest(unittest.TestCase):
+ def setUp(self):
- browser = webdriver.Firefox()
+ self.browser = webdriver.Firefox()
- browser.get('http://localhost:8000')
- assert 'Django' in browser.title
+ def tearDown(self):
+ self.browser.quit()
+
+ def test_can_start_a_list_and_retrieve_it_later(self):
+
+ # Rey has heard about a cool new online to-do app. He goes
+ # to check out its homepage
+ self.browser.get('http://localhost:8000')
+
+ # He notices the page title and header mention to-do lists
+ self.assertIn('To-Do', self.browser.title)
+ self.fail('Finish the test!')
+
+ # He is invited to enter a to-do item straight away
+
+ # He types "Make coffee" into a text box (Rey can't work when he is sleepy)
+
+ # There is still a text box inviting him to add another item.
+ # He enters "Go to the toilet" (Rey is very methodical)
+
+ # The page updates again, and now shows both items on his list
+
+ # Rey wonders whether the site will remember his list. Then he sees that
+ # the site has generated a unique URL for him
+
+ # He visits that URL - his to-do list is still there
+
+ # Satisfied, he goes back to sleep
+
+ if __name__ == '__main__':
+ unittest.main(warnings='ignore')
|
Update tests to be human-readable
|
## Code Before:
from selenium import webdriver
browser = webdriver.Firefox()
browser.get('http://localhost:8000')
assert 'Django' in browser.title
## Instruction:
Update tests to be human-readable
## Code After:
from selenium import webdriver
import unittest
class NewVisitorTest(unittest.TestCase):
def setUp(self):
self.browser = webdriver.Firefox()
def tearDown(self):
self.browser.quit()
def test_can_start_a_list_and_retrieve_it_later(self):
# Rey has heard about a cool new online to-do app. He goes
# to check out its homepage
self.browser.get('http://localhost:8000')
# He notices the page title and header mention to-do lists
self.assertIn('To-Do', self.browser.title)
self.fail('Finish the test!')
# He is invited to enter a to-do item straight away
# He types "Make coffee" into a text box (Rey can't work when he is sleepy)
# There is still a text box inviting him to add another item.
# He enters "Go to the toilet" (Rey is very methodical)
# The page updates again, and now shows both items on his list
# Rey wonders whether the site will remember his list. Then he sees that
# the site has generated a unique URL for him
# He visits that URL - his to-do list is still there
# Satisfied, he goes back to sleep
if __name__ == '__main__':
unittest.main(warnings='ignore')
|
# ... existing code ...
from selenium import webdriver
import unittest
class NewVisitorTest(unittest.TestCase):
def setUp(self):
self.browser = webdriver.Firefox()
def tearDown(self):
self.browser.quit()
def test_can_start_a_list_and_retrieve_it_later(self):
# Rey has heard about a cool new online to-do app. He goes
# to check out its homepage
self.browser.get('http://localhost:8000')
# He notices the page title and header mention to-do lists
self.assertIn('To-Do', self.browser.title)
self.fail('Finish the test!')
# He is invited to enter a to-do item straight away
# He types "Make coffee" into a text box (Rey can't work when he is sleepy)
# There is still a text box inviting him to add another item.
# He enters "Go to the toilet" (Rey is very methodical)
# The page updates again, and now shows both items on his list
# Rey wonders whether the site will remember his list. Then he sees that
# the site has generated a unique URL for him
# He visits that URL - his to-do list is still there
# Satisfied, he goes back to sleep
if __name__ == '__main__':
unittest.main(warnings='ignore')
# ... rest of the code ...
|
24da0f84e1a844b1f53e5afafc34cfcc915a9a67
|
corehq/apps/userreports/tests/test_report_rendering.py
|
corehq/apps/userreports/tests/test_report_rendering.py
|
from __future__ import absolute_import
from __future__ import unicode_literals
from django.test import SimpleTestCase
from corehq.apps.userreports.reports.view import ConfigurableReportView
class VeryFakeReportView(ConfigurableReportView):
# note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing
def __init__(self, data):
self._data = data
@property
def export_table(self):
return self._data
class ReportRenderingTest(SimpleTestCase):
def test_email_response_unicode(self):
report = VeryFakeReportView(data=[
['hello', 'हिन्दी']
])
# this used to fail: https://manage.dimagi.com/default.asp?263803
report.email_response
|
from __future__ import absolute_import
from __future__ import unicode_literals
from django.test import SimpleTestCase
from corehq.apps.userreports.reports.view import ConfigurableReportView
from corehq.apps.userreports.reports.util import ReportExport
class VeryFakeReportExport(ReportExport):
def __init__(self, data):
self._data = data
def get_table(self):
return self._data
class VeryFakeReportView(ConfigurableReportView):
# note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing
def __init__(self, data):
self._data = data
@property
def report_export(self):
return VeryFakeReportExport(self._data)
class ReportRenderingTest(SimpleTestCase):
def test_email_response_unicode(self):
report = VeryFakeReportView(data=[
['hello', 'हिन्दी']
])
# this used to fail: https://manage.dimagi.com/default.asp?263803
report.email_response
|
Update report_rendering test to use ReportExport
|
Update report_rendering test to use ReportExport
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
from __future__ import absolute_import
from __future__ import unicode_literals
+
from django.test import SimpleTestCase
+
from corehq.apps.userreports.reports.view import ConfigurableReportView
+ from corehq.apps.userreports.reports.util import ReportExport
+
+
+ class VeryFakeReportExport(ReportExport):
+ def __init__(self, data):
+ self._data = data
+
+ def get_table(self):
+ return self._data
class VeryFakeReportView(ConfigurableReportView):
# note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing
def __init__(self, data):
self._data = data
@property
- def export_table(self):
+ def report_export(self):
- return self._data
+ return VeryFakeReportExport(self._data)
class ReportRenderingTest(SimpleTestCase):
def test_email_response_unicode(self):
report = VeryFakeReportView(data=[
['hello', 'हिन्दी']
])
# this used to fail: https://manage.dimagi.com/default.asp?263803
report.email_response
|
Update report_rendering test to use ReportExport
|
## Code Before:
from __future__ import absolute_import
from __future__ import unicode_literals
from django.test import SimpleTestCase
from corehq.apps.userreports.reports.view import ConfigurableReportView
class VeryFakeReportView(ConfigurableReportView):
# note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing
def __init__(self, data):
self._data = data
@property
def export_table(self):
return self._data
class ReportRenderingTest(SimpleTestCase):
def test_email_response_unicode(self):
report = VeryFakeReportView(data=[
['hello', 'हिन्दी']
])
# this used to fail: https://manage.dimagi.com/default.asp?263803
report.email_response
## Instruction:
Update report_rendering test to use ReportExport
## Code After:
from __future__ import absolute_import
from __future__ import unicode_literals
from django.test import SimpleTestCase
from corehq.apps.userreports.reports.view import ConfigurableReportView
from corehq.apps.userreports.reports.util import ReportExport
class VeryFakeReportExport(ReportExport):
def __init__(self, data):
self._data = data
def get_table(self):
return self._data
class VeryFakeReportView(ConfigurableReportView):
# note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing
def __init__(self, data):
self._data = data
@property
def report_export(self):
return VeryFakeReportExport(self._data)
class ReportRenderingTest(SimpleTestCase):
def test_email_response_unicode(self):
report = VeryFakeReportView(data=[
['hello', 'हिन्दी']
])
# this used to fail: https://manage.dimagi.com/default.asp?263803
report.email_response
|
# ... existing code ...
from __future__ import unicode_literals
from django.test import SimpleTestCase
from corehq.apps.userreports.reports.view import ConfigurableReportView
from corehq.apps.userreports.reports.util import ReportExport
class VeryFakeReportExport(ReportExport):
def __init__(self, data):
self._data = data
def get_table(self):
return self._data
# ... modified code ...
@property
def report_export(self):
return VeryFakeReportExport(self._data)
# ... rest of the code ...
|
6f0a35372d625f923b9093194540cf0b0e9f054d
|
platformio_api/__init__.py
|
platformio_api/__init__.py
|
import json
import logging.config
import os
from time import tzset
VERSION = (0, 3, 0)
__version__ = ".".join([str(s) for s in VERSION])
__title__ = "platformio-api"
__description__ = ("An API for PlatformIO")
__url__ = "https://github.com/ivankravets/platformio-api"
__author__ = "Ivan Kravets"
__email__ = "[email protected]"
__license__ = "MIT License"
__copyright__ = "Copyright (C) 2014-2015 Ivan Kravets"
config = dict(
SQLALCHEMY_DATABASE_URI=None,
GITHUB_LOGIN=None,
GITHUB_PASSWORD=None,
DL_PIO_DIR=None,
DL_PIO_URL=None,
MAX_DLFILE_SIZE=1024*1024*10,
LOGGING=dict(version=1)
)
assert "PIOAPI_CONFIG_PATH" in os.environ
with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f:
config.update(json.load(f))
# configure logging for packages
logging.basicConfig()
logging.config.dictConfig(config['LOGGING'])
# setup time zone to UTC globally
os.environ['TZ'] = "+00:00"
tzset()
|
import json
import logging.config
import os
from time import tzset
VERSION = (0, 3, 0)
__version__ = ".".join([str(s) for s in VERSION])
__title__ = "platformio-api"
__description__ = ("An API for PlatformIO")
__url__ = "https://github.com/ivankravets/platformio-api"
__author__ = "Ivan Kravets"
__email__ = "[email protected]"
__license__ = "MIT License"
__copyright__ = "Copyright (C) 2014-2015 Ivan Kravets"
config = dict(
SQLALCHEMY_DATABASE_URI=None,
GITHUB_LOGIN=None,
GITHUB_PASSWORD=None,
DL_PIO_DIR=None,
DL_PIO_URL=None,
MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb
LOGGING=dict(version=1)
)
assert "PIOAPI_CONFIG_PATH" in os.environ
with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f:
config.update(json.load(f))
# configure logging for packages
logging.basicConfig()
logging.config.dictConfig(config['LOGGING'])
# setup time zone to UTC globally
os.environ['TZ'] = "+00:00"
tzset()
|
Increase repo size to 20Mb
|
Increase repo size to 20Mb
|
Python
|
apache-2.0
|
orgkhnargh/platformio-api,platformio/platformio-api
|
import json
import logging.config
import os
from time import tzset
VERSION = (0, 3, 0)
__version__ = ".".join([str(s) for s in VERSION])
__title__ = "platformio-api"
__description__ = ("An API for PlatformIO")
__url__ = "https://github.com/ivankravets/platformio-api"
__author__ = "Ivan Kravets"
__email__ = "[email protected]"
__license__ = "MIT License"
__copyright__ = "Copyright (C) 2014-2015 Ivan Kravets"
config = dict(
SQLALCHEMY_DATABASE_URI=None,
GITHUB_LOGIN=None,
GITHUB_PASSWORD=None,
DL_PIO_DIR=None,
DL_PIO_URL=None,
- MAX_DLFILE_SIZE=1024*1024*10,
+ MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb
LOGGING=dict(version=1)
)
assert "PIOAPI_CONFIG_PATH" in os.environ
with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f:
config.update(json.load(f))
# configure logging for packages
logging.basicConfig()
logging.config.dictConfig(config['LOGGING'])
# setup time zone to UTC globally
os.environ['TZ'] = "+00:00"
tzset()
|
Increase repo size to 20Mb
|
## Code Before:
import json
import logging.config
import os
from time import tzset
VERSION = (0, 3, 0)
__version__ = ".".join([str(s) for s in VERSION])
__title__ = "platformio-api"
__description__ = ("An API for PlatformIO")
__url__ = "https://github.com/ivankravets/platformio-api"
__author__ = "Ivan Kravets"
__email__ = "[email protected]"
__license__ = "MIT License"
__copyright__ = "Copyright (C) 2014-2015 Ivan Kravets"
config = dict(
SQLALCHEMY_DATABASE_URI=None,
GITHUB_LOGIN=None,
GITHUB_PASSWORD=None,
DL_PIO_DIR=None,
DL_PIO_URL=None,
MAX_DLFILE_SIZE=1024*1024*10,
LOGGING=dict(version=1)
)
assert "PIOAPI_CONFIG_PATH" in os.environ
with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f:
config.update(json.load(f))
# configure logging for packages
logging.basicConfig()
logging.config.dictConfig(config['LOGGING'])
# setup time zone to UTC globally
os.environ['TZ'] = "+00:00"
tzset()
## Instruction:
Increase repo size to 20Mb
## Code After:
import json
import logging.config
import os
from time import tzset
VERSION = (0, 3, 0)
__version__ = ".".join([str(s) for s in VERSION])
__title__ = "platformio-api"
__description__ = ("An API for PlatformIO")
__url__ = "https://github.com/ivankravets/platformio-api"
__author__ = "Ivan Kravets"
__email__ = "[email protected]"
__license__ = "MIT License"
__copyright__ = "Copyright (C) 2014-2015 Ivan Kravets"
config = dict(
SQLALCHEMY_DATABASE_URI=None,
GITHUB_LOGIN=None,
GITHUB_PASSWORD=None,
DL_PIO_DIR=None,
DL_PIO_URL=None,
MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb
LOGGING=dict(version=1)
)
assert "PIOAPI_CONFIG_PATH" in os.environ
with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f:
config.update(json.load(f))
# configure logging for packages
logging.basicConfig()
logging.config.dictConfig(config['LOGGING'])
# setup time zone to UTC globally
os.environ['TZ'] = "+00:00"
tzset()
|
...
DL_PIO_URL=None,
MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb
LOGGING=dict(version=1)
...
|
4a827bfff24758677e9c1d9d3b186fc14f23e0bb
|
lib/oeqa/runtime/cases/parselogs_rpi.py
|
lib/oeqa/runtime/cases/parselogs_rpi.py
|
from oeqa.runtime.cases.parselogs import *
rpi_errors = [
'bcmgenet fd580000.genet: failed to get enet-eee clock',
'bcmgenet fd580000.genet: failed to get enet-wol clock',
'bcmgenet fd580000.genet: failed to get enet clock',
'bcmgenet fd580000.ethernet: failed to get enet-eee clock',
'bcmgenet fd580000.ethernet: failed to get enet-wol clock',
'bcmgenet fd580000.ethernet: failed to get enet clock',
]
ignore_errors['raspberrypi4'] = rpi_errors + common_errors
ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors
ignore_errors['raspberrypi3'] = rpi_errors + common_errors
ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors
class ParseLogsTestRpi(ParseLogsTest):
pass
|
from oeqa.runtime.cases.parselogs import *
rpi_errors = [
]
ignore_errors['raspberrypi4'] = rpi_errors + common_errors
ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors
ignore_errors['raspberrypi3'] = rpi_errors + common_errors
ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors
class ParseLogsTestRpi(ParseLogsTest):
pass
|
Update the error regexps to 5.10 kernel
|
parselogs: Update the error regexps to 5.10 kernel
The old messages are no longer necessary
Signed-off-by: Khem Raj <[email protected]>
|
Python
|
mit
|
agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi
|
from oeqa.runtime.cases.parselogs import *
rpi_errors = [
- 'bcmgenet fd580000.genet: failed to get enet-eee clock',
- 'bcmgenet fd580000.genet: failed to get enet-wol clock',
- 'bcmgenet fd580000.genet: failed to get enet clock',
- 'bcmgenet fd580000.ethernet: failed to get enet-eee clock',
- 'bcmgenet fd580000.ethernet: failed to get enet-wol clock',
- 'bcmgenet fd580000.ethernet: failed to get enet clock',
]
ignore_errors['raspberrypi4'] = rpi_errors + common_errors
ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors
ignore_errors['raspberrypi3'] = rpi_errors + common_errors
ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors
class ParseLogsTestRpi(ParseLogsTest):
pass
|
Update the error regexps to 5.10 kernel
|
## Code Before:
from oeqa.runtime.cases.parselogs import *
rpi_errors = [
'bcmgenet fd580000.genet: failed to get enet-eee clock',
'bcmgenet fd580000.genet: failed to get enet-wol clock',
'bcmgenet fd580000.genet: failed to get enet clock',
'bcmgenet fd580000.ethernet: failed to get enet-eee clock',
'bcmgenet fd580000.ethernet: failed to get enet-wol clock',
'bcmgenet fd580000.ethernet: failed to get enet clock',
]
ignore_errors['raspberrypi4'] = rpi_errors + common_errors
ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors
ignore_errors['raspberrypi3'] = rpi_errors + common_errors
ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors
class ParseLogsTestRpi(ParseLogsTest):
pass
## Instruction:
Update the error regexps to 5.10 kernel
## Code After:
from oeqa.runtime.cases.parselogs import *
rpi_errors = [
]
ignore_errors['raspberrypi4'] = rpi_errors + common_errors
ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors
ignore_errors['raspberrypi3'] = rpi_errors + common_errors
ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors
class ParseLogsTestRpi(ParseLogsTest):
pass
|
# ... existing code ...
rpi_errors = [
]
# ... rest of the code ...
|
62ba442ac447dbb4482dd15f70075d224d0e5a0e
|
scripts/test_conda_build_log.py
|
scripts/test_conda_build_log.py
|
import pytest
import log_parser
import os
@pytest.fixture
def parsed_log():
logname = os.path.join(os.path.split(os.path.abspath(__file__))[0],
'build.log')
gen = list(log_parser.read_log_from_script(logname))
parsed = {built_name: log_parser.parse_conda_build(lines)
for name, built_name, lines in gen}
return parsed
def test_parse_conda_build(parsed_log):
# make sure that we have at least one thing that was parsed
assert len(parsed_log) >= 1
def test_parse_init(parsed_log):
# make sure we are getting the build command out of every single entry
for pkg_name, parsed in parsed_log.items():
parsed_init = log_parser.parse_init(parsed['init'])
assert 'build_command' in parsed_init
def test_parse_build(parsed_log):
# make sure we are getting either an error or the build string out of the
# build section
for pkg_name, parsed in parsed_log.items():
if 'build' not in parsed:
# not all packages will successfully build
continue
# if there is a build section, then parse it
parsed_build = log_parser.parse_build(parsed['build'])
if parsed_build['built_name'] == 'failed':
assert parsed_build['error'] != []
else:
assert parsed_build['error'] == []
|
import pytest
import log_parser
import os
@pytest.fixture
def parsed_log():
logname = os.path.join(os.path.split(os.path.abspath(__file__))[0],
'build.log')
gen = list(log_parser.read_log_from_script(logname))
parsed = {built_name: log_parser.parse_conda_build(lines)
for name, built_name, lines in gen}
return parsed
def test_parse_conda_build(parsed_log):
# make sure that we have at least one thing that was parsed
assert len(parsed_log) >= 1
def test_parse_init(parsed_log):
# make sure we are getting the build command out of every single entry
for pkg_name, parsed in parsed_log.items():
parsed_init = log_parser.parse_init(parsed['init'])
assert 'build_command' in parsed_init
assert 'err' in parsed_init
def test_parse_build(parsed_log):
# make sure we are getting either an error or the build string out of the
# build section
for pkg_name, parsed in parsed_log.items():
if 'build' not in parsed:
# not all packages will successfully build
continue
# if there is a build section, then parse it
parsed_build = log_parser.parse_build(parsed['build'])
if parsed_build['built_name'] == 'failed':
assert parsed_build['error'] != []
else:
assert parsed_build['error'] == []
|
Make sure there is an error field
|
TST: Make sure there is an error field
|
Python
|
bsd-3-clause
|
NSLS-II/lightsource2-recipes,NSLS-II/auto-build-tagged-recipes,NSLS-II/lightsource2-recipes,NSLS-II/auto-build-tagged-recipes,NSLS-II/lightsource2-recipes,NSLS-II/lightsource2-recipes
|
import pytest
import log_parser
import os
@pytest.fixture
def parsed_log():
logname = os.path.join(os.path.split(os.path.abspath(__file__))[0],
'build.log')
gen = list(log_parser.read_log_from_script(logname))
parsed = {built_name: log_parser.parse_conda_build(lines)
for name, built_name, lines in gen}
return parsed
def test_parse_conda_build(parsed_log):
# make sure that we have at least one thing that was parsed
assert len(parsed_log) >= 1
def test_parse_init(parsed_log):
# make sure we are getting the build command out of every single entry
for pkg_name, parsed in parsed_log.items():
parsed_init = log_parser.parse_init(parsed['init'])
assert 'build_command' in parsed_init
+ assert 'err' in parsed_init
def test_parse_build(parsed_log):
# make sure we are getting either an error or the build string out of the
# build section
for pkg_name, parsed in parsed_log.items():
if 'build' not in parsed:
# not all packages will successfully build
continue
# if there is a build section, then parse it
parsed_build = log_parser.parse_build(parsed['build'])
if parsed_build['built_name'] == 'failed':
assert parsed_build['error'] != []
else:
assert parsed_build['error'] == []
|
Make sure there is an error field
|
## Code Before:
import pytest
import log_parser
import os
@pytest.fixture
def parsed_log():
logname = os.path.join(os.path.split(os.path.abspath(__file__))[0],
'build.log')
gen = list(log_parser.read_log_from_script(logname))
parsed = {built_name: log_parser.parse_conda_build(lines)
for name, built_name, lines in gen}
return parsed
def test_parse_conda_build(parsed_log):
# make sure that we have at least one thing that was parsed
assert len(parsed_log) >= 1
def test_parse_init(parsed_log):
# make sure we are getting the build command out of every single entry
for pkg_name, parsed in parsed_log.items():
parsed_init = log_parser.parse_init(parsed['init'])
assert 'build_command' in parsed_init
def test_parse_build(parsed_log):
# make sure we are getting either an error or the build string out of the
# build section
for pkg_name, parsed in parsed_log.items():
if 'build' not in parsed:
# not all packages will successfully build
continue
# if there is a build section, then parse it
parsed_build = log_parser.parse_build(parsed['build'])
if parsed_build['built_name'] == 'failed':
assert parsed_build['error'] != []
else:
assert parsed_build['error'] == []
## Instruction:
Make sure there is an error field
## Code After:
import pytest
import log_parser
import os
@pytest.fixture
def parsed_log():
logname = os.path.join(os.path.split(os.path.abspath(__file__))[0],
'build.log')
gen = list(log_parser.read_log_from_script(logname))
parsed = {built_name: log_parser.parse_conda_build(lines)
for name, built_name, lines in gen}
return parsed
def test_parse_conda_build(parsed_log):
# make sure that we have at least one thing that was parsed
assert len(parsed_log) >= 1
def test_parse_init(parsed_log):
# make sure we are getting the build command out of every single entry
for pkg_name, parsed in parsed_log.items():
parsed_init = log_parser.parse_init(parsed['init'])
assert 'build_command' in parsed_init
assert 'err' in parsed_init
def test_parse_build(parsed_log):
# make sure we are getting either an error or the build string out of the
# build section
for pkg_name, parsed in parsed_log.items():
if 'build' not in parsed:
# not all packages will successfully build
continue
# if there is a build section, then parse it
parsed_build = log_parser.parse_build(parsed['build'])
if parsed_build['built_name'] == 'failed':
assert parsed_build['error'] != []
else:
assert parsed_build['error'] == []
|
// ... existing code ...
assert 'build_command' in parsed_init
assert 'err' in parsed_init
// ... rest of the code ...
|
e5ac63b4615b4166d7e7866c9f169e4c9f86f46c
|
setup.py
|
setup.py
|
from distutils.core import setup
setup(
name='django-emailuser',
version='1.0',
description='simple User model identified by email address',
packages=['emailuser'],
author='Mark Paschal',
author_email='[email protected]',
url='https://github.com/duncaningram/django-emailuser',
classifiers=[
'Framework :: Django',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
],
)
|
from distutils.core import setup
setup(
name='django-emailuser',
version='1.0',
description='simple User model identified by email address',
packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'],
author='Mark Paschal',
author_email='[email protected]',
url='https://github.com/duncaningram/django-emailuser',
classifiers=[
'Framework :: Django',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
],
)
|
Install the management command too when installing as a distribution
|
Install the management command too when installing as a distribution
|
Python
|
mit
|
markpasc/django-emailuser,duncaningram/django-emailuser
|
from distutils.core import setup
setup(
name='django-emailuser',
version='1.0',
description='simple User model identified by email address',
- packages=['emailuser'],
+ packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'],
author='Mark Paschal',
author_email='[email protected]',
url='https://github.com/duncaningram/django-emailuser',
classifiers=[
'Framework :: Django',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
],
)
|
Install the management command too when installing as a distribution
|
## Code Before:
from distutils.core import setup
setup(
name='django-emailuser',
version='1.0',
description='simple User model identified by email address',
packages=['emailuser'],
author='Mark Paschal',
author_email='[email protected]',
url='https://github.com/duncaningram/django-emailuser',
classifiers=[
'Framework :: Django',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
],
)
## Instruction:
Install the management command too when installing as a distribution
## Code After:
from distutils.core import setup
setup(
name='django-emailuser',
version='1.0',
description='simple User model identified by email address',
packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'],
author='Mark Paschal',
author_email='[email protected]',
url='https://github.com/duncaningram/django-emailuser',
classifiers=[
'Framework :: Django',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
],
)
|
# ... existing code ...
description='simple User model identified by email address',
packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'],
# ... rest of the code ...
|
839d884d3dca3e799a235b1d2d69acf998f520f9
|
barsystem_base/management/commands/import_people.py
|
barsystem_base/management/commands/import_people.py
|
from django.core.management.base import BaseCommand, CommandError
from barsystem_base.models import Person
class Command(BaseCommand):
args = '<filename>'
help = 'Import list of people'
csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',')
def handle(self, *args, **kwargs):
if len(args) == 0:
raise CommandError('Please supply filename')
with open(args[0], 'r') as f:
columns = None
for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']:
# print(line)
# take header
if columns is None:
columns = line
continue
values = dict(zip(columns, line))
values['active'] = values['type'] != 'hidden'
try:
p = Person.objects.get(id=values['id'])
except Person.DoesNotExist:
p = Person()
for key, val in values.items():
if hasattr(p, key):
setattr(p, key, val)
print(p)
p.save()
print('Done')
|
from django.core.management.base import BaseCommand, CommandError
from barsystem_base.models import Person, Token
class Command(BaseCommand):
args = '<filename>'
help = 'Import list of people'
csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',')
def handle(self, *args, **kwargs):
if len(args) == 0:
raise CommandError('Please supply filename')
with open(args[0], 'r') as f:
columns = None
for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']:
# print(line)
# take header
if columns is None:
columns = line
continue
values = dict(zip(columns, line))
values['active'] = values['type'] != 'hidden'
try:
p = Person.objects.get(id=values['id'])
except Person.DoesNotExist:
p = Person()
for key, val in values.items():
if hasattr(p, key):
setattr(p, key, val)
print(p)
p.save()
t = Token()
t.type = 'ibutton'
t.value = values['token']
t.person = p
t.save()
print('Done')
|
Add ibutton when importing old people
|
Add ibutton when importing old people
|
Python
|
mit
|
TkkrLab/barsystem,TkkrLab/barsystem,TkkrLab/barsystem
|
from django.core.management.base import BaseCommand, CommandError
- from barsystem_base.models import Person
+ from barsystem_base.models import Person, Token
class Command(BaseCommand):
args = '<filename>'
help = 'Import list of people'
csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',')
def handle(self, *args, **kwargs):
if len(args) == 0:
raise CommandError('Please supply filename')
with open(args[0], 'r') as f:
columns = None
for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']:
# print(line)
# take header
if columns is None:
columns = line
continue
values = dict(zip(columns, line))
values['active'] = values['type'] != 'hidden'
try:
p = Person.objects.get(id=values['id'])
except Person.DoesNotExist:
p = Person()
for key, val in values.items():
if hasattr(p, key):
setattr(p, key, val)
print(p)
p.save()
+ t = Token()
+ t.type = 'ibutton'
+ t.value = values['token']
+ t.person = p
+ t.save()
print('Done')
|
Add ibutton when importing old people
|
## Code Before:
from django.core.management.base import BaseCommand, CommandError
from barsystem_base.models import Person
class Command(BaseCommand):
args = '<filename>'
help = 'Import list of people'
csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',')
def handle(self, *args, **kwargs):
if len(args) == 0:
raise CommandError('Please supply filename')
with open(args[0], 'r') as f:
columns = None
for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']:
# print(line)
# take header
if columns is None:
columns = line
continue
values = dict(zip(columns, line))
values['active'] = values['type'] != 'hidden'
try:
p = Person.objects.get(id=values['id'])
except Person.DoesNotExist:
p = Person()
for key, val in values.items():
if hasattr(p, key):
setattr(p, key, val)
print(p)
p.save()
print('Done')
## Instruction:
Add ibutton when importing old people
## Code After:
from django.core.management.base import BaseCommand, CommandError
from barsystem_base.models import Person, Token
class Command(BaseCommand):
args = '<filename>'
help = 'Import list of people'
csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',')
def handle(self, *args, **kwargs):
if len(args) == 0:
raise CommandError('Please supply filename')
with open(args[0], 'r') as f:
columns = None
for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']:
# print(line)
# take header
if columns is None:
columns = line
continue
values = dict(zip(columns, line))
values['active'] = values['type'] != 'hidden'
try:
p = Person.objects.get(id=values['id'])
except Person.DoesNotExist:
p = Person()
for key, val in values.items():
if hasattr(p, key):
setattr(p, key, val)
print(p)
p.save()
t = Token()
t.type = 'ibutton'
t.value = values['token']
t.person = p
t.save()
print('Done')
|
...
from django.core.management.base import BaseCommand, CommandError
from barsystem_base.models import Person, Token
...
p.save()
t = Token()
t.type = 'ibutton'
t.value = values['token']
t.person = p
t.save()
print('Done')
...
|
f764b52558cb02b8e31b9695a724e4c4e80872dd
|
iscc_bench/readers/__init__.py
|
iscc_bench/readers/__init__.py
|
from iscc_bench.readers.bxbooks import bxbooks
from iscc_bench.readers.dnbrdf import dnbrdf
from iscc_bench.readers.harvard import harvard
from iscc_bench.readers.openlibrary import openlibrary
from iscc_bench.readers.libgen import libgen
ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen)
|
from iscc_bench.readers.bxbooks import bxbooks
from iscc_bench.readers.dnbrdf import dnbrdf
from iscc_bench.readers.harvard import harvard
from iscc_bench.readers.openlibrary import openlibrary
from iscc_bench.readers.libgen import libgen
from iscc_bench.readers.caltech101 import caltech_101
from iscc_bench.readers.caltech256 import caltech_256
ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen)
ALL_IMAGE_READERS = (caltech_101, caltech_256)
|
Add image readers to package scope
|
Add image readers to package scope
|
Python
|
bsd-2-clause
|
coblo/isccbench
|
from iscc_bench.readers.bxbooks import bxbooks
from iscc_bench.readers.dnbrdf import dnbrdf
from iscc_bench.readers.harvard import harvard
from iscc_bench.readers.openlibrary import openlibrary
from iscc_bench.readers.libgen import libgen
-
+ from iscc_bench.readers.caltech101 import caltech_101
+ from iscc_bench.readers.caltech256 import caltech_256
ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen)
+ ALL_IMAGE_READERS = (caltech_101, caltech_256)
|
Add image readers to package scope
|
## Code Before:
from iscc_bench.readers.bxbooks import bxbooks
from iscc_bench.readers.dnbrdf import dnbrdf
from iscc_bench.readers.harvard import harvard
from iscc_bench.readers.openlibrary import openlibrary
from iscc_bench.readers.libgen import libgen
ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen)
## Instruction:
Add image readers to package scope
## Code After:
from iscc_bench.readers.bxbooks import bxbooks
from iscc_bench.readers.dnbrdf import dnbrdf
from iscc_bench.readers.harvard import harvard
from iscc_bench.readers.openlibrary import openlibrary
from iscc_bench.readers.libgen import libgen
from iscc_bench.readers.caltech101 import caltech_101
from iscc_bench.readers.caltech256 import caltech_256
ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen)
ALL_IMAGE_READERS = (caltech_101, caltech_256)
|
# ... existing code ...
from iscc_bench.readers.libgen import libgen
from iscc_bench.readers.caltech101 import caltech_101
from iscc_bench.readers.caltech256 import caltech_256
# ... modified code ...
ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen)
ALL_IMAGE_READERS = (caltech_101, caltech_256)
# ... rest of the code ...
|
0c3529bd264d5512e31d828c65676baff6edefa6
|
pinax/waitinglist/templatetags/pinax_waitinglist_tags.py
|
pinax/waitinglist/templatetags/pinax_waitinglist_tags.py
|
from django import template
from ..forms import WaitingListEntryForm
register = template.Library()
@register.assignment_tag
def waitinglist_entry_form():
"""
Get a (new) form object to post a new comment.
Syntax::
{% waitinglist_entry_form as [varname] %}
"""
return WaitingListEntryForm()
|
from django import template
from ..forms import WaitingListEntryForm
register = template.Library()
@register.simple_tag(takes_context=True)
def waitinglist_entry_form(context):
"""
Get a (new) form object to post a new comment.
Syntax::
{% waitinglist_entry_form as [varname] %}
"""
initial = {}
if "request" in context:
initial.update({
"referrer": context["request"].META.get("HTTP_REFERER", ""),
"campaign": context["request"].GET.get("wlc", "")
})
return WaitingListEntryForm(initial=initial)
|
Update template tag to also take context
|
Update template tag to also take context
|
Python
|
mit
|
pinax/pinax-waitinglist,pinax/pinax-waitinglist
|
from django import template
from ..forms import WaitingListEntryForm
register = template.Library()
- @register.assignment_tag
+ @register.simple_tag(takes_context=True)
- def waitinglist_entry_form():
+ def waitinglist_entry_form(context):
"""
Get a (new) form object to post a new comment.
Syntax::
{% waitinglist_entry_form as [varname] %}
"""
+ initial = {}
+ if "request" in context:
+ initial.update({
+ "referrer": context["request"].META.get("HTTP_REFERER", ""),
+ "campaign": context["request"].GET.get("wlc", "")
+ })
- return WaitingListEntryForm()
+ return WaitingListEntryForm(initial=initial)
|
Update template tag to also take context
|
## Code Before:
from django import template
from ..forms import WaitingListEntryForm
register = template.Library()
@register.assignment_tag
def waitinglist_entry_form():
"""
Get a (new) form object to post a new comment.
Syntax::
{% waitinglist_entry_form as [varname] %}
"""
return WaitingListEntryForm()
## Instruction:
Update template tag to also take context
## Code After:
from django import template
from ..forms import WaitingListEntryForm
register = template.Library()
@register.simple_tag(takes_context=True)
def waitinglist_entry_form(context):
"""
Get a (new) form object to post a new comment.
Syntax::
{% waitinglist_entry_form as [varname] %}
"""
initial = {}
if "request" in context:
initial.update({
"referrer": context["request"].META.get("HTTP_REFERER", ""),
"campaign": context["request"].GET.get("wlc", "")
})
return WaitingListEntryForm(initial=initial)
|
// ... existing code ...
@register.simple_tag(takes_context=True)
def waitinglist_entry_form(context):
"""
// ... modified code ...
"""
initial = {}
if "request" in context:
initial.update({
"referrer": context["request"].META.get("HTTP_REFERER", ""),
"campaign": context["request"].GET.get("wlc", "")
})
return WaitingListEntryForm(initial=initial)
// ... rest of the code ...
|
c30347f34967ee7634676e0e4e27164910f9e52b
|
regparser/tree/xml_parser/note_processor.py
|
regparser/tree/xml_parser/note_processor.py
|
from regparser.tree.depth import markers as mtypes, optional_rules
from regparser.tree.struct import Node
from regparser.tree.xml_parser import (
paragraph_processor, simple_hierarchy_processor)
class IgnoreNotesHeader(paragraph_processor.BaseMatcher):
"""We don't want to include "Note:" and "Notes:" headers"""
def matches(self, xml):
return xml.tag == 'HD' and xml.text.lower().startswith('note')
def derive_nodes(self, xml, processor=None):
return []
class NoteProcessor(paragraph_processor.ParagraphProcessor):
MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(),
IgnoreNotesHeader(),
paragraph_processor.IgnoreTagMatcher('PRTPAGE')]
def additional_constraints(self):
return [optional_rules.limit_paragraph_types(
mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)]
class NoteMatcher(paragraph_processor.BaseMatcher):
"""Processes the contents of NOTE and NOTES tags using a NoteProcessor"""
def matches(self, xml):
return xml.tag in ('NOTE', 'NOTES')
def derive_nodes(self, xml, processor=None):
processor = NoteProcessor()
node = Node(label=[mtypes.MARKERLESS], source_xml=xml,
node_type=Node.NOTE)
return [processor.process(xml, node)]
|
import re
from regparser.tree.depth import markers as mtypes, optional_rules
from regparser.tree.struct import Node
from regparser.tree.xml_parser import (
paragraph_processor, simple_hierarchy_processor)
class IgnoreNotesHeader(paragraph_processor.BaseMatcher):
"""We don't want to include "Note:" and "Notes:" headers"""
REGEX = re.compile('notes?:?\s*$', re.IGNORECASE)
def matches(self, xml):
return xml.tag == 'HD' and self.REGEX.match(xml.text or '')
def derive_nodes(self, xml, processor=None):
return []
class NoteProcessor(paragraph_processor.ParagraphProcessor):
MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(),
IgnoreNotesHeader(),
paragraph_processor.IgnoreTagMatcher('PRTPAGE')]
def additional_constraints(self):
return [optional_rules.limit_paragraph_types(
mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)]
class NoteMatcher(paragraph_processor.BaseMatcher):
"""Processes the contents of NOTE and NOTES tags using a NoteProcessor"""
def matches(self, xml):
return xml.tag in ('NOTE', 'NOTES')
def derive_nodes(self, xml, processor=None):
processor = NoteProcessor()
node = Node(label=[mtypes.MARKERLESS], source_xml=xml,
node_type=Node.NOTE)
return [processor.process(xml, node)]
|
Use regex rather than string match for Note:
|
Use regex rather than string match for Note:
Per suggestion from @tadhg-ohiggins
|
Python
|
cc0-1.0
|
eregs/regulations-parser,tadhg-ohiggins/regulations-parser,tadhg-ohiggins/regulations-parser,cmc333333/regulations-parser,cmc333333/regulations-parser,eregs/regulations-parser
|
+ import re
+
from regparser.tree.depth import markers as mtypes, optional_rules
from regparser.tree.struct import Node
from regparser.tree.xml_parser import (
paragraph_processor, simple_hierarchy_processor)
class IgnoreNotesHeader(paragraph_processor.BaseMatcher):
"""We don't want to include "Note:" and "Notes:" headers"""
+ REGEX = re.compile('notes?:?\s*$', re.IGNORECASE)
+
def matches(self, xml):
- return xml.tag == 'HD' and xml.text.lower().startswith('note')
+ return xml.tag == 'HD' and self.REGEX.match(xml.text or '')
def derive_nodes(self, xml, processor=None):
return []
class NoteProcessor(paragraph_processor.ParagraphProcessor):
MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(),
IgnoreNotesHeader(),
paragraph_processor.IgnoreTagMatcher('PRTPAGE')]
def additional_constraints(self):
return [optional_rules.limit_paragraph_types(
mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)]
class NoteMatcher(paragraph_processor.BaseMatcher):
"""Processes the contents of NOTE and NOTES tags using a NoteProcessor"""
def matches(self, xml):
return xml.tag in ('NOTE', 'NOTES')
def derive_nodes(self, xml, processor=None):
processor = NoteProcessor()
node = Node(label=[mtypes.MARKERLESS], source_xml=xml,
node_type=Node.NOTE)
return [processor.process(xml, node)]
|
Use regex rather than string match for Note:
|
## Code Before:
from regparser.tree.depth import markers as mtypes, optional_rules
from regparser.tree.struct import Node
from regparser.tree.xml_parser import (
paragraph_processor, simple_hierarchy_processor)
class IgnoreNotesHeader(paragraph_processor.BaseMatcher):
"""We don't want to include "Note:" and "Notes:" headers"""
def matches(self, xml):
return xml.tag == 'HD' and xml.text.lower().startswith('note')
def derive_nodes(self, xml, processor=None):
return []
class NoteProcessor(paragraph_processor.ParagraphProcessor):
MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(),
IgnoreNotesHeader(),
paragraph_processor.IgnoreTagMatcher('PRTPAGE')]
def additional_constraints(self):
return [optional_rules.limit_paragraph_types(
mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)]
class NoteMatcher(paragraph_processor.BaseMatcher):
"""Processes the contents of NOTE and NOTES tags using a NoteProcessor"""
def matches(self, xml):
return xml.tag in ('NOTE', 'NOTES')
def derive_nodes(self, xml, processor=None):
processor = NoteProcessor()
node = Node(label=[mtypes.MARKERLESS], source_xml=xml,
node_type=Node.NOTE)
return [processor.process(xml, node)]
## Instruction:
Use regex rather than string match for Note:
## Code After:
import re
from regparser.tree.depth import markers as mtypes, optional_rules
from regparser.tree.struct import Node
from regparser.tree.xml_parser import (
paragraph_processor, simple_hierarchy_processor)
class IgnoreNotesHeader(paragraph_processor.BaseMatcher):
"""We don't want to include "Note:" and "Notes:" headers"""
REGEX = re.compile('notes?:?\s*$', re.IGNORECASE)
def matches(self, xml):
return xml.tag == 'HD' and self.REGEX.match(xml.text or '')
def derive_nodes(self, xml, processor=None):
return []
class NoteProcessor(paragraph_processor.ParagraphProcessor):
MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(),
IgnoreNotesHeader(),
paragraph_processor.IgnoreTagMatcher('PRTPAGE')]
def additional_constraints(self):
return [optional_rules.limit_paragraph_types(
mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)]
class NoteMatcher(paragraph_processor.BaseMatcher):
"""Processes the contents of NOTE and NOTES tags using a NoteProcessor"""
def matches(self, xml):
return xml.tag in ('NOTE', 'NOTES')
def derive_nodes(self, xml, processor=None):
processor = NoteProcessor()
node = Node(label=[mtypes.MARKERLESS], source_xml=xml,
node_type=Node.NOTE)
return [processor.process(xml, node)]
|
// ... existing code ...
import re
from regparser.tree.depth import markers as mtypes, optional_rules
// ... modified code ...
"""We don't want to include "Note:" and "Notes:" headers"""
REGEX = re.compile('notes?:?\s*$', re.IGNORECASE)
def matches(self, xml):
return xml.tag == 'HD' and self.REGEX.match(xml.text or '')
// ... rest of the code ...
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.