commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
3.18k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43k
ndiff
stringlengths
52
3.32k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
fuzzy_diff
stringlengths
16
3.18k
4ff1eb00f8e212d280ac858feb4efcc795d97d80
tests/test_models.py
tests/test_models.py
import pytest from suddendev.models import GameController def test_create_game(session): pass
import pytest from suddendev.models import GameSetup def test_create_game(session): game_setup = GameSetup('ASDF') assert game_setup.player_count == 1
Fix broken import in model tests.
[NG] Fix broken import in model tests.
Python
mit
SuddenDevs/SuddenDev,SuddenDevs/SuddenDev,SuddenDevs/SuddenDev,SuddenDevs/SuddenDev
import pytest - from suddendev.models import GameController + from suddendev.models import GameSetup def test_create_game(session): - pass + game_setup = GameSetup('ASDF') + assert game_setup.player_count == 1
Fix broken import in model tests.
## Code Before: import pytest from suddendev.models import GameController def test_create_game(session): pass ## Instruction: Fix broken import in model tests. ## Code After: import pytest from suddendev.models import GameSetup def test_create_game(session): game_setup = GameSetup('ASDF') assert game_setup.player_count == 1
... import pytest from suddendev.models import GameSetup ... def test_create_game(session): game_setup = GameSetup('ASDF') assert game_setup.player_count == 1 ...
f696e320d66f375779692ec073f15d3d6d466059
edx_data_research/parsing/parse_sql.py
edx_data_research/parsing/parse_sql.py
import subprocess from edx_data_research.parsing.parse import Parse class SQL(Parse): def __init__(self, args): super(SQL, self).__init__(args) self._collections = args.collection self.sql_file = args.sql_file def migrate(self): subprocess.check_call(['mongoimport', '-d', self.db_name, '-c', self._collections, '--type', 'tsv', '--file', self.sql_file, '--headerline'])
import subprocess from edx_data_research.parsing.parse import Parse class SQL(Parse): def __init__(self, args): super(SQL, self).__init__(args) self._collections = args.collection self.sql_file = args.sql_file def migrate(self): subprocess.check_call(['mongoimport', '-d', self.db_name, '-c', self._collections, '--type', 'tsv', '--file', self.sql_file, '--headerline', '--drop'])
Update mongo import of files to drop existing collection first
Update mongo import of files to drop existing collection first
Python
mit
McGillX/edx_data_research,McGillX/edx_data_research,McGillX/edx_data_research
import subprocess from edx_data_research.parsing.parse import Parse class SQL(Parse): def __init__(self, args): super(SQL, self).__init__(args) self._collections = args.collection self.sql_file = args.sql_file def migrate(self): subprocess.check_call(['mongoimport', '-d', self.db_name, '-c', self._collections, '--type', 'tsv', '--file', - self.sql_file, '--headerline']) + self.sql_file, '--headerline', '--drop'])
Update mongo import of files to drop existing collection first
## Code Before: import subprocess from edx_data_research.parsing.parse import Parse class SQL(Parse): def __init__(self, args): super(SQL, self).__init__(args) self._collections = args.collection self.sql_file = args.sql_file def migrate(self): subprocess.check_call(['mongoimport', '-d', self.db_name, '-c', self._collections, '--type', 'tsv', '--file', self.sql_file, '--headerline']) ## Instruction: Update mongo import of files to drop existing collection first ## Code After: import subprocess from edx_data_research.parsing.parse import Parse class SQL(Parse): def __init__(self, args): super(SQL, self).__init__(args) self._collections = args.collection self.sql_file = args.sql_file def migrate(self): subprocess.check_call(['mongoimport', '-d', self.db_name, '-c', self._collections, '--type', 'tsv', '--file', self.sql_file, '--headerline', '--drop'])
... self._collections, '--type', 'tsv', '--file', self.sql_file, '--headerline', '--drop']) ...
12a9ef54d82d9508852e5596dbb9df321986e067
tests/test_heroku.py
tests/test_heroku.py
"""Tests for the Wallace API.""" import subprocess import re import requests class TestHeroku(object): """The Heroku test class.""" def test_sandbox(self): """Launch the experiment on Heroku.""" sandbox_output = subprocess.check_output( "cd examples/bartlett1932; wallace sandbox --verbose", shell=True) id = re.search( 'Running as experiment (.*)...', sandbox_output).group(1) r = requests.get("http://{}.herokuapp.com/summary".format(id)) assert r.json()['status'] == [] subprocess.call( "heroku apps:destroy --app {} --confirm {}".format(id), shell=True)
"""Tests for the Wallace API.""" import subprocess import re import os import requests class TestHeroku(object): """The Heroku test class.""" sandbox_output = subprocess.check_output( "cd examples/bartlett1932; wallace sandbox --verbose", shell=True) os.environ['app_id'] = re.search( 'Running as experiment (.*)...', sandbox_output).group(1) @classmethod def teardown_class(cls): """Remove the app from Heroku.""" app_id = os.environ['app_id'] subprocess.call( "heroku apps:destroy --app {} --confirm {}".format(app_id, app_id), shell=True) def test_summary(self): """Launch the experiment on Heroku.""" app_id = os.environ['app_id'] r = requests.get("http://{}.herokuapp.com/summary".format(app_id)) assert r.json()['status'] == []
Refactor Heroku tests to have shared setup
Refactor Heroku tests to have shared setup
Python
mit
Dallinger/Dallinger,berkeley-cocosci/Wallace,jcpeterson/Dallinger,jcpeterson/Dallinger,suchow/Wallace,berkeley-cocosci/Wallace,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,berkeley-cocosci/Wallace,Dallinger/Dallinger,jcpeterson/Dallinger,suchow/Wallace,suchow/Wallace,jcpeterson/Dallinger,Dallinger/Dallinger
"""Tests for the Wallace API.""" import subprocess import re + import os import requests class TestHeroku(object): """The Heroku test class.""" + sandbox_output = subprocess.check_output( + "cd examples/bartlett1932; wallace sandbox --verbose", shell=True) + + os.environ['app_id'] = re.search( + 'Running as experiment (.*)...', sandbox_output).group(1) + + @classmethod + def teardown_class(cls): + """Remove the app from Heroku.""" + app_id = os.environ['app_id'] + subprocess.call( + "heroku apps:destroy --app {} --confirm {}".format(app_id, app_id), + shell=True) + - def test_sandbox(self): + def test_summary(self): """Launch the experiment on Heroku.""" + app_id = os.environ['app_id'] - sandbox_output = subprocess.check_output( - "cd examples/bartlett1932; wallace sandbox --verbose", shell=True) - - id = re.search( - 'Running as experiment (.*)...', sandbox_output).group(1) - - r = requests.get("http://{}.herokuapp.com/summary".format(id)) + r = requests.get("http://{}.herokuapp.com/summary".format(app_id)) - assert r.json()['status'] == [] - subprocess.call( - "heroku apps:destroy --app {} --confirm {}".format(id), - shell=True) -
Refactor Heroku tests to have shared setup
## Code Before: """Tests for the Wallace API.""" import subprocess import re import requests class TestHeroku(object): """The Heroku test class.""" def test_sandbox(self): """Launch the experiment on Heroku.""" sandbox_output = subprocess.check_output( "cd examples/bartlett1932; wallace sandbox --verbose", shell=True) id = re.search( 'Running as experiment (.*)...', sandbox_output).group(1) r = requests.get("http://{}.herokuapp.com/summary".format(id)) assert r.json()['status'] == [] subprocess.call( "heroku apps:destroy --app {} --confirm {}".format(id), shell=True) ## Instruction: Refactor Heroku tests to have shared setup ## Code After: """Tests for the Wallace API.""" import subprocess import re import os import requests class TestHeroku(object): """The Heroku test class.""" sandbox_output = subprocess.check_output( "cd examples/bartlett1932; wallace sandbox --verbose", shell=True) os.environ['app_id'] = re.search( 'Running as experiment (.*)...', sandbox_output).group(1) @classmethod def teardown_class(cls): """Remove the app from Heroku.""" app_id = os.environ['app_id'] subprocess.call( "heroku apps:destroy --app {} --confirm {}".format(app_id, app_id), shell=True) def test_summary(self): """Launch the experiment on Heroku.""" app_id = os.environ['app_id'] r = requests.get("http://{}.herokuapp.com/summary".format(app_id)) assert r.json()['status'] == []
... import re import os import requests ... sandbox_output = subprocess.check_output( "cd examples/bartlett1932; wallace sandbox --verbose", shell=True) os.environ['app_id'] = re.search( 'Running as experiment (.*)...', sandbox_output).group(1) @classmethod def teardown_class(cls): """Remove the app from Heroku.""" app_id = os.environ['app_id'] subprocess.call( "heroku apps:destroy --app {} --confirm {}".format(app_id, app_id), shell=True) def test_summary(self): """Launch the experiment on Heroku.""" app_id = os.environ['app_id'] r = requests.get("http://{}.herokuapp.com/summary".format(app_id)) assert r.json()['status'] == [] ...
a20fc95d3a4dc194ef4f6d227976ff5bba229aaa
feincms/__init__.py
feincms/__init__.py
VERSION = (1, 4, 2) __version__ = '.'.join(map(str, VERSION)) class LazySettings(object): def _load_settings(self): from feincms import default_settings from django.conf import settings as django_settings for key in dir(default_settings): if not key.startswith(('FEINCMS_', '_HACK_')): continue setattr(self, key, getattr(django_settings, key, getattr(default_settings, key))) def __getattr__(self, attr): self._load_settings() del self.__class__.__getattr__ return self.__dict__[attr] settings = LazySettings() COMPLETELY_LOADED = False def ensure_completely_loaded(): """ This method ensures all models are completely loaded FeinCMS requires Django to be completely initialized before proceeding, because of the extension mechanism and the dynamically created content types. For more informations, have a look at issue #23 on github: http://github.com/feincms/feincms/issues#issue/23 """ global COMPLETELY_LOADED if COMPLETELY_LOADED: return True from django.core.management.validation import get_validation_errors from StringIO import StringIO get_validation_errors(StringIO(), None) COMPLETELY_LOADED = True return True
VERSION = (1, 4, 2) __version__ = '.'.join(map(str, VERSION)) class LazySettings(object): def _load_settings(self): from feincms import default_settings from django.conf import settings as django_settings for key in dir(default_settings): if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')): continue setattr(self, key, getattr(django_settings, key, getattr(default_settings, key))) def __getattr__(self, attr): self._load_settings() del self.__class__.__getattr__ return self.__dict__[attr] settings = LazySettings() COMPLETELY_LOADED = False def ensure_completely_loaded(): """ This method ensures all models are completely loaded FeinCMS requires Django to be completely initialized before proceeding, because of the extension mechanism and the dynamically created content types. For more informations, have a look at issue #23 on github: http://github.com/feincms/feincms/issues#issue/23 """ global COMPLETELY_LOADED if COMPLETELY_LOADED: return True from django.core.management.validation import get_validation_errors from StringIO import StringIO get_validation_errors(StringIO(), None) COMPLETELY_LOADED = True return True
Fix a Python 2.4 incompatibility that snuck in
Fix a Python 2.4 incompatibility that snuck in Fixes github issue #214
Python
bsd-3-clause
matthiask/django-content-editor,nickburlett/feincms,matthiask/django-content-editor,matthiask/feincms2-content,michaelkuty/feincms,matthiask/feincms2-content,pjdelport/feincms,mjl/feincms,joshuajonah/feincms,matthiask/django-content-editor,feincms/feincms,nickburlett/feincms,nickburlett/feincms,mjl/feincms,joshuajonah/feincms,joshuajonah/feincms,pjdelport/feincms,joshuajonah/feincms,michaelkuty/feincms,pjdelport/feincms,feincms/feincms,michaelkuty/feincms,michaelkuty/feincms,matthiask/feincms2-content,mjl/feincms,nickburlett/feincms,feincms/feincms,matthiask/django-content-editor
VERSION = (1, 4, 2) __version__ = '.'.join(map(str, VERSION)) class LazySettings(object): def _load_settings(self): from feincms import default_settings from django.conf import settings as django_settings for key in dir(default_settings): - if not key.startswith(('FEINCMS_', '_HACK_')): + if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')): continue setattr(self, key, getattr(django_settings, key, getattr(default_settings, key))) def __getattr__(self, attr): self._load_settings() del self.__class__.__getattr__ return self.__dict__[attr] settings = LazySettings() COMPLETELY_LOADED = False def ensure_completely_loaded(): """ This method ensures all models are completely loaded FeinCMS requires Django to be completely initialized before proceeding, because of the extension mechanism and the dynamically created content types. For more informations, have a look at issue #23 on github: http://github.com/feincms/feincms/issues#issue/23 """ global COMPLETELY_LOADED if COMPLETELY_LOADED: return True from django.core.management.validation import get_validation_errors from StringIO import StringIO get_validation_errors(StringIO(), None) COMPLETELY_LOADED = True return True
Fix a Python 2.4 incompatibility that snuck in
## Code Before: VERSION = (1, 4, 2) __version__ = '.'.join(map(str, VERSION)) class LazySettings(object): def _load_settings(self): from feincms import default_settings from django.conf import settings as django_settings for key in dir(default_settings): if not key.startswith(('FEINCMS_', '_HACK_')): continue setattr(self, key, getattr(django_settings, key, getattr(default_settings, key))) def __getattr__(self, attr): self._load_settings() del self.__class__.__getattr__ return self.__dict__[attr] settings = LazySettings() COMPLETELY_LOADED = False def ensure_completely_loaded(): """ This method ensures all models are completely loaded FeinCMS requires Django to be completely initialized before proceeding, because of the extension mechanism and the dynamically created content types. For more informations, have a look at issue #23 on github: http://github.com/feincms/feincms/issues#issue/23 """ global COMPLETELY_LOADED if COMPLETELY_LOADED: return True from django.core.management.validation import get_validation_errors from StringIO import StringIO get_validation_errors(StringIO(), None) COMPLETELY_LOADED = True return True ## Instruction: Fix a Python 2.4 incompatibility that snuck in ## Code After: VERSION = (1, 4, 2) __version__ = '.'.join(map(str, VERSION)) class LazySettings(object): def _load_settings(self): from feincms import default_settings from django.conf import settings as django_settings for key in dir(default_settings): if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')): continue setattr(self, key, getattr(django_settings, key, getattr(default_settings, key))) def __getattr__(self, attr): self._load_settings() del self.__class__.__getattr__ return self.__dict__[attr] settings = LazySettings() COMPLETELY_LOADED = False def ensure_completely_loaded(): """ This method ensures all models are completely loaded FeinCMS requires Django to be completely initialized before proceeding, because of the extension mechanism and the dynamically created content types. For more informations, have a look at issue #23 on github: http://github.com/feincms/feincms/issues#issue/23 """ global COMPLETELY_LOADED if COMPLETELY_LOADED: return True from django.core.management.validation import get_validation_errors from StringIO import StringIO get_validation_errors(StringIO(), None) COMPLETELY_LOADED = True return True
... for key in dir(default_settings): if not (key.startswith('FEINCMS_') or key.startswith('_HACK_')): continue ...
d0ca3952a34a74f0167b76bbedfa3cf8875a399c
var/spack/repos/builtin/packages/py-scikit-learn/package.py
var/spack/repos/builtin/packages/py-scikit-learn/package.py
from spack import * class PyScikitLearn(Package): """""" homepage = "https://pypi.python.org/pypi/scikit-learn" url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz" version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d') version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e') extends('python') def install(self, spec, prefix): python('setup.py', 'install', '--prefix=%s' % prefix)
from spack import * class PyScikitLearn(Package): """""" homepage = "https://pypi.python.org/pypi/scikit-learn" url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz" version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d') version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e') version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc') extends('python') def install(self, spec, prefix): python('setup.py', 'install', '--prefix=%s' % prefix)
Add version 0.17.1 of scikit-learn.
Add version 0.17.1 of scikit-learn.
Python
lgpl-2.1
matthiasdiener/spack,mfherbst/spack,EmreAtes/spack,TheTimmy/spack,iulian787/spack,iulian787/spack,iulian787/spack,mfherbst/spack,tmerrick1/spack,mfherbst/spack,LLNL/spack,LLNL/spack,tmerrick1/spack,skosukhin/spack,TheTimmy/spack,skosukhin/spack,TheTimmy/spack,skosukhin/spack,matthiasdiener/spack,mfherbst/spack,mfherbst/spack,lgarren/spack,LLNL/spack,matthiasdiener/spack,tmerrick1/spack,krafczyk/spack,krafczyk/spack,TheTimmy/spack,EmreAtes/spack,LLNL/spack,krafczyk/spack,iulian787/spack,lgarren/spack,tmerrick1/spack,EmreAtes/spack,EmreAtes/spack,krafczyk/spack,lgarren/spack,matthiasdiener/spack,lgarren/spack,iulian787/spack,LLNL/spack,lgarren/spack,skosukhin/spack,krafczyk/spack,skosukhin/spack,matthiasdiener/spack,TheTimmy/spack,EmreAtes/spack,tmerrick1/spack
from spack import * class PyScikitLearn(Package): """""" homepage = "https://pypi.python.org/pypi/scikit-learn" url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz" version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d') version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e') + version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc') extends('python') def install(self, spec, prefix): python('setup.py', 'install', '--prefix=%s' % prefix)
Add version 0.17.1 of scikit-learn.
## Code Before: from spack import * class PyScikitLearn(Package): """""" homepage = "https://pypi.python.org/pypi/scikit-learn" url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz" version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d') version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e') extends('python') def install(self, spec, prefix): python('setup.py', 'install', '--prefix=%s' % prefix) ## Instruction: Add version 0.17.1 of scikit-learn. ## Code After: from spack import * class PyScikitLearn(Package): """""" homepage = "https://pypi.python.org/pypi/scikit-learn" url = "https://pypi.python.org/packages/source/s/scikit-learn/scikit-learn-0.15.2.tar.gz" version('0.15.2', 'd9822ad0238e17b382a3c756ea94fe0d') version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e') version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc') extends('python') def install(self, spec, prefix): python('setup.py', 'install', '--prefix=%s' % prefix)
... version('0.16.1', '363ddda501e3b6b61726aa40b8dbdb7e') version('0.17.1', 'a2f8b877e6d99b1ed737144f5a478dfc') ...
77541e5b3956d9e6b130810211fcae10de29eb85
tests/integration/base.py
tests/integration/base.py
import righteous from ConfigParser import SafeConfigParser from ..compat import unittest class RighteousIntegrationTestCase(unittest.TestCase): def setUp(self): config = SafeConfigParser() config.read('righteous.config') if not config.has_section('auth'): raise Exception('Please create a righteous.config file with ' 'appropriate credentials') self.auth = dict( (key, config.get('auth', key)) for key in config.options('auth')) self.server = dict( (key, config.get('server-defaults', key)) for key in config.options('server-defaults')) righteous.init( self.auth['username'], self.auth['password'], self.auth['account_id'], **self.server) self.config = config self.username = self.auth['username'] def test_login(self): self.assertTrue(righteous.login()) assert False
import righteous from ConfigParser import SafeConfigParser from ..compat import unittest class RighteousIntegrationTestCase(unittest.TestCase): def setUp(self): config = SafeConfigParser() config.read('righteous.config') if not config.has_section('auth'): raise Exception('Please create a righteous.config file with ' 'appropriate credentials') self.auth = dict( (key, config.get('auth', key)) for key in config.options('auth')) self.server = dict( (key, config.get('server-defaults', key)) for key in config.options('server-defaults')) righteous.init( self.auth['username'], self.auth['password'], self.auth['account_id'], **self.server) self.config = config self.username = self.auth['username'] def test_login(self): self.assertTrue(righteous.login())
Make the integration test fail, so we can see the request / response
Make the integration test fail, so we can see the request / response
Python
unlicense
michaeljoseph/righteous,michaeljoseph/righteous
import righteous from ConfigParser import SafeConfigParser from ..compat import unittest class RighteousIntegrationTestCase(unittest.TestCase): def setUp(self): config = SafeConfigParser() config.read('righteous.config') if not config.has_section('auth'): raise Exception('Please create a righteous.config file with ' 'appropriate credentials') self.auth = dict( (key, config.get('auth', key)) for key in config.options('auth')) self.server = dict( (key, config.get('server-defaults', key)) for key in config.options('server-defaults')) righteous.init( self.auth['username'], self.auth['password'], self.auth['account_id'], **self.server) self.config = config self.username = self.auth['username'] def test_login(self): self.assertTrue(righteous.login()) - assert False
Make the integration test fail, so we can see the request / response
## Code Before: import righteous from ConfigParser import SafeConfigParser from ..compat import unittest class RighteousIntegrationTestCase(unittest.TestCase): def setUp(self): config = SafeConfigParser() config.read('righteous.config') if not config.has_section('auth'): raise Exception('Please create a righteous.config file with ' 'appropriate credentials') self.auth = dict( (key, config.get('auth', key)) for key in config.options('auth')) self.server = dict( (key, config.get('server-defaults', key)) for key in config.options('server-defaults')) righteous.init( self.auth['username'], self.auth['password'], self.auth['account_id'], **self.server) self.config = config self.username = self.auth['username'] def test_login(self): self.assertTrue(righteous.login()) assert False ## Instruction: Make the integration test fail, so we can see the request / response ## Code After: import righteous from ConfigParser import SafeConfigParser from ..compat import unittest class RighteousIntegrationTestCase(unittest.TestCase): def setUp(self): config = SafeConfigParser() config.read('righteous.config') if not config.has_section('auth'): raise Exception('Please create a righteous.config file with ' 'appropriate credentials') self.auth = dict( (key, config.get('auth', key)) for key in config.options('auth')) self.server = dict( (key, config.get('server-defaults', key)) for key in config.options('server-defaults')) righteous.init( self.auth['username'], self.auth['password'], self.auth['account_id'], **self.server) self.config = config self.username = self.auth['username'] def test_login(self): self.assertTrue(righteous.login())
// ... existing code ... self.assertTrue(righteous.login()) // ... rest of the code ...
cb08d25f49b8b4c5177c8afdd9a69330992ee854
tests/replay/test_replay.py
tests/replay/test_replay.py
import pytest from cookiecutter import replay, main, exceptions def test_get_replay_file_name(): """Make sure that replay.get_file_name generates a valid json file path.""" assert replay.get_file_name('foo', 'bar') == 'foo/bar.json' @pytest.fixture(params=[ {'no_input': True}, {'extra_context': {}}, {'no_input': True, 'extra_context': {}}, ]) def invalid_kwargs(request): return request.param def test_raise_on_invalid_mode(invalid_kwargs): with pytest.raises(exceptions.InvalidModeException): main.cookiecutter('foo', replay=True, **invalid_kwargs)
import pytest from cookiecutter import replay, main, exceptions def test_get_replay_file_name(): """Make sure that replay.get_file_name generates a valid json file path.""" assert replay.get_file_name('foo', 'bar') == 'foo/bar.json' @pytest.fixture(params=[ {'no_input': True}, {'extra_context': {}}, {'no_input': True, 'extra_context': {}}, ]) def invalid_kwargs(request): return request.param def test_raise_on_invalid_mode(invalid_kwargs): with pytest.raises(exceptions.InvalidModeException): main.cookiecutter('foo', replay=True, **invalid_kwargs) def test_main_does_not_invoke_dump_but_load(mocker): mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') mock_gen_context = mocker.patch('cookiecutter.main.generate_context') mock_gen_files = mocker.patch('cookiecutter.main.generate_files') mock_replay_dump = mocker.patch('cookiecutter.main.dump') mock_replay_load = mocker.patch('cookiecutter.main.load') main.cookiecutter('foobar', replay=True) assert not mock_prompt.called assert not mock_gen_context.called assert not mock_replay_dump.called assert mock_replay_load.called assert mock_gen_files.called def test_main_does_not_invoke_load_but_dump(mocker): mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') mock_gen_context = mocker.patch('cookiecutter.main.generate_context') mock_gen_files = mocker.patch('cookiecutter.main.generate_files') mock_replay_dump = mocker.patch('cookiecutter.main.dump') mock_replay_load = mocker.patch('cookiecutter.main.load') main.cookiecutter('foobar', replay=False) assert mock_prompt.called assert mock_gen_context.called assert mock_replay_dump.called assert not mock_replay_load.called assert mock_gen_files.called
Add tests for a correct behaviour in cookiecutter.main for replay
Add tests for a correct behaviour in cookiecutter.main for replay
Python
bsd-3-clause
christabor/cookiecutter,luzfcb/cookiecutter,hackebrot/cookiecutter,cguardia/cookiecutter,pjbull/cookiecutter,dajose/cookiecutter,michaeljoseph/cookiecutter,moi65/cookiecutter,terryjbates/cookiecutter,takeflight/cookiecutter,terryjbates/cookiecutter,luzfcb/cookiecutter,agconti/cookiecutter,cguardia/cookiecutter,christabor/cookiecutter,audreyr/cookiecutter,stevepiercy/cookiecutter,willingc/cookiecutter,venumech/cookiecutter,stevepiercy/cookiecutter,takeflight/cookiecutter,pjbull/cookiecutter,benthomasson/cookiecutter,agconti/cookiecutter,benthomasson/cookiecutter,Springerle/cookiecutter,ramiroluz/cookiecutter,audreyr/cookiecutter,moi65/cookiecutter,dajose/cookiecutter,hackebrot/cookiecutter,michaeljoseph/cookiecutter,Springerle/cookiecutter,ramiroluz/cookiecutter,venumech/cookiecutter,willingc/cookiecutter
import pytest from cookiecutter import replay, main, exceptions def test_get_replay_file_name(): """Make sure that replay.get_file_name generates a valid json file path.""" assert replay.get_file_name('foo', 'bar') == 'foo/bar.json' @pytest.fixture(params=[ {'no_input': True}, {'extra_context': {}}, {'no_input': True, 'extra_context': {}}, ]) def invalid_kwargs(request): return request.param def test_raise_on_invalid_mode(invalid_kwargs): with pytest.raises(exceptions.InvalidModeException): main.cookiecutter('foo', replay=True, **invalid_kwargs) + + def test_main_does_not_invoke_dump_but_load(mocker): + mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') + mock_gen_context = mocker.patch('cookiecutter.main.generate_context') + mock_gen_files = mocker.patch('cookiecutter.main.generate_files') + mock_replay_dump = mocker.patch('cookiecutter.main.dump') + mock_replay_load = mocker.patch('cookiecutter.main.load') + + main.cookiecutter('foobar', replay=True) + + assert not mock_prompt.called + assert not mock_gen_context.called + assert not mock_replay_dump.called + assert mock_replay_load.called + assert mock_gen_files.called + + + def test_main_does_not_invoke_load_but_dump(mocker): + mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') + mock_gen_context = mocker.patch('cookiecutter.main.generate_context') + mock_gen_files = mocker.patch('cookiecutter.main.generate_files') + mock_replay_dump = mocker.patch('cookiecutter.main.dump') + mock_replay_load = mocker.patch('cookiecutter.main.load') + + main.cookiecutter('foobar', replay=False) + + assert mock_prompt.called + assert mock_gen_context.called + assert mock_replay_dump.called + assert not mock_replay_load.called + assert mock_gen_files.called +
Add tests for a correct behaviour in cookiecutter.main for replay
## Code Before: import pytest from cookiecutter import replay, main, exceptions def test_get_replay_file_name(): """Make sure that replay.get_file_name generates a valid json file path.""" assert replay.get_file_name('foo', 'bar') == 'foo/bar.json' @pytest.fixture(params=[ {'no_input': True}, {'extra_context': {}}, {'no_input': True, 'extra_context': {}}, ]) def invalid_kwargs(request): return request.param def test_raise_on_invalid_mode(invalid_kwargs): with pytest.raises(exceptions.InvalidModeException): main.cookiecutter('foo', replay=True, **invalid_kwargs) ## Instruction: Add tests for a correct behaviour in cookiecutter.main for replay ## Code After: import pytest from cookiecutter import replay, main, exceptions def test_get_replay_file_name(): """Make sure that replay.get_file_name generates a valid json file path.""" assert replay.get_file_name('foo', 'bar') == 'foo/bar.json' @pytest.fixture(params=[ {'no_input': True}, {'extra_context': {}}, {'no_input': True, 'extra_context': {}}, ]) def invalid_kwargs(request): return request.param def test_raise_on_invalid_mode(invalid_kwargs): with pytest.raises(exceptions.InvalidModeException): main.cookiecutter('foo', replay=True, **invalid_kwargs) def test_main_does_not_invoke_dump_but_load(mocker): mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') mock_gen_context = mocker.patch('cookiecutter.main.generate_context') mock_gen_files = mocker.patch('cookiecutter.main.generate_files') mock_replay_dump = mocker.patch('cookiecutter.main.dump') mock_replay_load = mocker.patch('cookiecutter.main.load') main.cookiecutter('foobar', replay=True) assert not mock_prompt.called assert not mock_gen_context.called assert not mock_replay_dump.called assert mock_replay_load.called assert mock_gen_files.called def test_main_does_not_invoke_load_but_dump(mocker): mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') mock_gen_context = mocker.patch('cookiecutter.main.generate_context') mock_gen_files = mocker.patch('cookiecutter.main.generate_files') mock_replay_dump = mocker.patch('cookiecutter.main.dump') mock_replay_load = mocker.patch('cookiecutter.main.load') main.cookiecutter('foobar', replay=False) assert mock_prompt.called assert mock_gen_context.called assert mock_replay_dump.called assert not mock_replay_load.called assert mock_gen_files.called
# ... existing code ... main.cookiecutter('foo', replay=True, **invalid_kwargs) def test_main_does_not_invoke_dump_but_load(mocker): mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') mock_gen_context = mocker.patch('cookiecutter.main.generate_context') mock_gen_files = mocker.patch('cookiecutter.main.generate_files') mock_replay_dump = mocker.patch('cookiecutter.main.dump') mock_replay_load = mocker.patch('cookiecutter.main.load') main.cookiecutter('foobar', replay=True) assert not mock_prompt.called assert not mock_gen_context.called assert not mock_replay_dump.called assert mock_replay_load.called assert mock_gen_files.called def test_main_does_not_invoke_load_but_dump(mocker): mock_prompt = mocker.patch('cookiecutter.main.prompt_for_config') mock_gen_context = mocker.patch('cookiecutter.main.generate_context') mock_gen_files = mocker.patch('cookiecutter.main.generate_files') mock_replay_dump = mocker.patch('cookiecutter.main.dump') mock_replay_load = mocker.patch('cookiecutter.main.load') main.cookiecutter('foobar', replay=False) assert mock_prompt.called assert mock_gen_context.called assert mock_replay_dump.called assert not mock_replay_load.called assert mock_gen_files.called # ... rest of the code ...
dfaa49b31e8abd10456761110d0cadc1b7c7640d
zaqar/transport/wsgi/app.py
zaqar/transport/wsgi/app.py
from oslo_config import cfg from zaqar import bootstrap # Use the global CONF instance conf = cfg.CONF conf(project='zaqar', prog='zaqar-queues', args=[]) boot = bootstrap.Bootstrap(conf) conf.drivers.transport = 'wsgi' app = boot.transport.app
from oslo_config import cfg from oslo_log import log from zaqar import bootstrap # Use the global CONF instance conf = cfg.CONF log.register_options(conf) conf(project='zaqar', prog='zaqar-queues', args=[]) log.setup(conf, 'zaqar') boot = bootstrap.Bootstrap(conf) conf.drivers.transport = 'wsgi' app = boot.transport.app
Make the log work when deploy Zaqar with uwsgi
Make the log work when deploy Zaqar with uwsgi The zaqar-wsgi runs under uwsgi by devstack can't print any WARNING, DEBUG, ERROR or INFO log now. This path add the log initialization for uwsgi boot. Change-Id: Ifcd6be908442275d2acbde2562e593b2ca87b277 Cloese-bug: #1645492
Python
apache-2.0
openstack/zaqar,openstack/zaqar,openstack/zaqar,openstack/zaqar
from oslo_config import cfg + from oslo_log import log from zaqar import bootstrap # Use the global CONF instance conf = cfg.CONF + log.register_options(conf) conf(project='zaqar', prog='zaqar-queues', args=[]) + log.setup(conf, 'zaqar') boot = bootstrap.Bootstrap(conf) conf.drivers.transport = 'wsgi' app = boot.transport.app
Make the log work when deploy Zaqar with uwsgi
## Code Before: from oslo_config import cfg from zaqar import bootstrap # Use the global CONF instance conf = cfg.CONF conf(project='zaqar', prog='zaqar-queues', args=[]) boot = bootstrap.Bootstrap(conf) conf.drivers.transport = 'wsgi' app = boot.transport.app ## Instruction: Make the log work when deploy Zaqar with uwsgi ## Code After: from oslo_config import cfg from oslo_log import log from zaqar import bootstrap # Use the global CONF instance conf = cfg.CONF log.register_options(conf) conf(project='zaqar', prog='zaqar-queues', args=[]) log.setup(conf, 'zaqar') boot = bootstrap.Bootstrap(conf) conf.drivers.transport = 'wsgi' app = boot.transport.app
... from oslo_config import cfg from oslo_log import log ... conf = cfg.CONF log.register_options(conf) conf(project='zaqar', prog='zaqar-queues', args=[]) log.setup(conf, 'zaqar') ...
4636c9394138534fc39cc5bdac373b97919ffd01
server/info/services.py
server/info/services.py
"""info services.""" from info.models import Article, News, Column def get_column_object(uid): """Get column object.""" try: obj = Column.objects.get(uid=uid) except Column.DoesNotExist: obj = None return obj def get_articles_by_column(uid): """Get_articles_by_column.""" queryset = Article.objects.filter(column__uid=uid).order_by('id') return queryset def get_columns_queryset(): """Get_columns_queryset.""" queryset = Column.objects.all().order_by('-id') return queryset def get_article_queryset(): """Get article queryset.""" queryset = Article.objects.all().order_by('-id') return queryset def get_article_object(uid): """Get article object.""" return Article.objects.get(uid=uid) def get_news_queryset(): """Get news queryset.""" return News.objects.all().order_by('-id')
"""info services.""" from info.models import Article, News, Column def get_column_object(uid): """Get column object.""" try: obj = Column.objects.get(uid=uid) except Column.DoesNotExist: obj = None return obj def get_articles_by_column(uid): """Get_articles_by_column.""" queryset = Article.objects.filter( column__uid=uid ).order_by('id') return queryset def get_columns_queryset(): """Get_columns_queryset.""" queryset = Column.objects.all().only('uid', 'name').order_by('-id') return queryset def get_article_queryset(): """Get article queryset.""" queryset = Article.objects.all().order_by('-id') return queryset def get_article_object(uid): """Get article object.""" return Article.objects.get(uid=uid) def get_news_queryset(): """Get news queryset.""" return News.objects.all().order_by('-id')
Modify django orm filter, add only
Modify django orm filter, add only
Python
mit
istommao/codingcatweb,istommao/codingcatweb,istommao/codingcatweb
"""info services.""" from info.models import Article, News, Column def get_column_object(uid): """Get column object.""" try: obj = Column.objects.get(uid=uid) except Column.DoesNotExist: obj = None return obj def get_articles_by_column(uid): """Get_articles_by_column.""" - queryset = Article.objects.filter(column__uid=uid).order_by('id') + queryset = Article.objects.filter( + column__uid=uid + ).order_by('id') return queryset def get_columns_queryset(): """Get_columns_queryset.""" - queryset = Column.objects.all().order_by('-id') + queryset = Column.objects.all().only('uid', 'name').order_by('-id') return queryset def get_article_queryset(): """Get article queryset.""" queryset = Article.objects.all().order_by('-id') return queryset def get_article_object(uid): """Get article object.""" return Article.objects.get(uid=uid) def get_news_queryset(): """Get news queryset.""" return News.objects.all().order_by('-id')
Modify django orm filter, add only
## Code Before: """info services.""" from info.models import Article, News, Column def get_column_object(uid): """Get column object.""" try: obj = Column.objects.get(uid=uid) except Column.DoesNotExist: obj = None return obj def get_articles_by_column(uid): """Get_articles_by_column.""" queryset = Article.objects.filter(column__uid=uid).order_by('id') return queryset def get_columns_queryset(): """Get_columns_queryset.""" queryset = Column.objects.all().order_by('-id') return queryset def get_article_queryset(): """Get article queryset.""" queryset = Article.objects.all().order_by('-id') return queryset def get_article_object(uid): """Get article object.""" return Article.objects.get(uid=uid) def get_news_queryset(): """Get news queryset.""" return News.objects.all().order_by('-id') ## Instruction: Modify django orm filter, add only ## Code After: """info services.""" from info.models import Article, News, Column def get_column_object(uid): """Get column object.""" try: obj = Column.objects.get(uid=uid) except Column.DoesNotExist: obj = None return obj def get_articles_by_column(uid): """Get_articles_by_column.""" queryset = Article.objects.filter( column__uid=uid ).order_by('id') return queryset def get_columns_queryset(): """Get_columns_queryset.""" queryset = Column.objects.all().only('uid', 'name').order_by('-id') return queryset def get_article_queryset(): """Get article queryset.""" queryset = Article.objects.all().order_by('-id') return queryset def get_article_object(uid): """Get article object.""" return Article.objects.get(uid=uid) def get_news_queryset(): """Get news queryset.""" return News.objects.all().order_by('-id')
... """Get_articles_by_column.""" queryset = Article.objects.filter( column__uid=uid ).order_by('id') ... """Get_columns_queryset.""" queryset = Column.objects.all().only('uid', 'name').order_by('-id') ...
10dea74d7f7946e9bab8c99b489793708845183c
fireplace/cards/wog/hunter.py
fireplace/cards/wog/hunter.py
from ..utils import * ## # Minions class OG_179: "Fiery Bat" deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1) class OG_292: "Forlorn Stalker" play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e") OG_292e = buff(+1, +1)
from ..utils import * ## # Minions class OG_179: "Fiery Bat" deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1) class OG_292: "Forlorn Stalker" play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e") OG_292e = buff(+1, +1) ## # Spells class OG_045: "Infest" play = Buff(FRIENDLY_MINIONS, "OG_045a") class OG_045a: "Nerubian Spores" deathrattle = Give(CONTROLLER, RandomBeast()) tags = {GameTag.DEATHRATTLE: True} class OG_061: "On the Hunt" play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t") class OG_211: "Call of the Wild" play = ( Summon(CONTROLLER, "NEW1_034"), Summon(CONTROLLER, "NEW1_033"), Summon(CONTROLLER, "NEW1_032") )
Implement Infest, On the Hunt, Call of the Wild
Implement Infest, On the Hunt, Call of the Wild
Python
agpl-3.0
jleclanche/fireplace,beheh/fireplace,NightKev/fireplace
from ..utils import * ## # Minions class OG_179: "Fiery Bat" deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1) class OG_292: "Forlorn Stalker" play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e") OG_292e = buff(+1, +1) + + ## + # Spells + + class OG_045: + "Infest" + play = Buff(FRIENDLY_MINIONS, "OG_045a") + + class OG_045a: + "Nerubian Spores" + deathrattle = Give(CONTROLLER, RandomBeast()) + tags = {GameTag.DEATHRATTLE: True} + + + class OG_061: + "On the Hunt" + play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t") + + + class OG_211: + "Call of the Wild" + play = ( + Summon(CONTROLLER, "NEW1_034"), + Summon(CONTROLLER, "NEW1_033"), + Summon(CONTROLLER, "NEW1_032") + ) +
Implement Infest, On the Hunt, Call of the Wild
## Code Before: from ..utils import * ## # Minions class OG_179: "Fiery Bat" deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1) class OG_292: "Forlorn Stalker" play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e") OG_292e = buff(+1, +1) ## Instruction: Implement Infest, On the Hunt, Call of the Wild ## Code After: from ..utils import * ## # Minions class OG_179: "Fiery Bat" deathrattle = Hit(RANDOM_ENEMY_CHARACTER, 1) class OG_292: "Forlorn Stalker" play = Buff(FRIENDLY_HAND + MINION + DEATHRATTLE, "OG_292e") OG_292e = buff(+1, +1) ## # Spells class OG_045: "Infest" play = Buff(FRIENDLY_MINIONS, "OG_045a") class OG_045a: "Nerubian Spores" deathrattle = Give(CONTROLLER, RandomBeast()) tags = {GameTag.DEATHRATTLE: True} class OG_061: "On the Hunt" play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t") class OG_211: "Call of the Wild" play = ( Summon(CONTROLLER, "NEW1_034"), Summon(CONTROLLER, "NEW1_033"), Summon(CONTROLLER, "NEW1_032") )
// ... existing code ... OG_292e = buff(+1, +1) ## # Spells class OG_045: "Infest" play = Buff(FRIENDLY_MINIONS, "OG_045a") class OG_045a: "Nerubian Spores" deathrattle = Give(CONTROLLER, RandomBeast()) tags = {GameTag.DEATHRATTLE: True} class OG_061: "On the Hunt" play = Hit(TARGET, 1), Summon(CONTROLLER, "OG_061t") class OG_211: "Call of the Wild" play = ( Summon(CONTROLLER, "NEW1_034"), Summon(CONTROLLER, "NEW1_033"), Summon(CONTROLLER, "NEW1_032") ) // ... rest of the code ...
a70bb058bd93831b755079f5fee495088b620c6d
taiga/locale/api.py
taiga/locale/api.py
from django.conf import settings from taiga.base import response from taiga.base.api.viewsets import ReadOnlyListViewSet from . import permissions class LocalesViewSet(ReadOnlyListViewSet): permission_classes = (permissions.LocalesPermission,) def list(self, request, *args, **kwargs): locales = [{"code": c, "name": n} for c, n in settings.LANGUAGES] return response.Ok(locales)
from django.conf import settings from taiga.base import response from taiga.base.api.viewsets import ReadOnlyListViewSet from . import permissions class LocalesViewSet(ReadOnlyListViewSet): permission_classes = (permissions.LocalesPermission,) def list(self, request, *args, **kwargs): locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES] return response.Ok(locales)
Add bidi (right-to-left layout) attr to locale resource
Add bidi (right-to-left layout) attr to locale resource
Python
agpl-3.0
crr0004/taiga-back,CoolCloud/taiga-back,Rademade/taiga-back,seanchen/taiga-back,Tigerwhit4/taiga-back,seanchen/taiga-back,astronaut1712/taiga-back,dycodedev/taiga-back,dycodedev/taiga-back,WALR/taiga-back,forging2012/taiga-back,xdevelsistemas/taiga-back-community,xdevelsistemas/taiga-back-community,rajiteh/taiga-back,Tigerwhit4/taiga-back,taigaio/taiga-back,EvgeneOskin/taiga-back,CMLL/taiga-back,EvgeneOskin/taiga-back,Rademade/taiga-back,EvgeneOskin/taiga-back,seanchen/taiga-back,rajiteh/taiga-back,CMLL/taiga-back,gauravjns/taiga-back,obimod/taiga-back,CMLL/taiga-back,astagi/taiga-back,dayatz/taiga-back,astronaut1712/taiga-back,CMLL/taiga-back,astagi/taiga-back,joshisa/taiga-back,coopsource/taiga-back,dayatz/taiga-back,gauravjns/taiga-back,forging2012/taiga-back,joshisa/taiga-back,gam-phon/taiga-back,Rademade/taiga-back,Tigerwhit4/taiga-back,CoolCloud/taiga-back,crr0004/taiga-back,coopsource/taiga-back,obimod/taiga-back,dycodedev/taiga-back,gam-phon/taiga-back,WALR/taiga-back,gam-phon/taiga-back,gam-phon/taiga-back,taigaio/taiga-back,xdevelsistemas/taiga-back-community,CoolCloud/taiga-back,jeffdwyatt/taiga-back,taigaio/taiga-back,coopsource/taiga-back,forging2012/taiga-back,WALR/taiga-back,gauravjns/taiga-back,joshisa/taiga-back,crr0004/taiga-back,jeffdwyatt/taiga-back,forging2012/taiga-back,CoolCloud/taiga-back,seanchen/taiga-back,EvgeneOskin/taiga-back,joshisa/taiga-back,obimod/taiga-back,Rademade/taiga-back,WALR/taiga-back,Rademade/taiga-back,jeffdwyatt/taiga-back,dycodedev/taiga-back,bdang2012/taiga-back-casting,obimod/taiga-back,astronaut1712/taiga-back,astagi/taiga-back,dayatz/taiga-back,rajiteh/taiga-back,astagi/taiga-back,bdang2012/taiga-back-casting,crr0004/taiga-back,gauravjns/taiga-back,coopsource/taiga-back,bdang2012/taiga-back-casting,astronaut1712/taiga-back,jeffdwyatt/taiga-back,Tigerwhit4/taiga-back,bdang2012/taiga-back-casting,rajiteh/taiga-back
from django.conf import settings from taiga.base import response from taiga.base.api.viewsets import ReadOnlyListViewSet from . import permissions class LocalesViewSet(ReadOnlyListViewSet): permission_classes = (permissions.LocalesPermission,) def list(self, request, *args, **kwargs): - locales = [{"code": c, "name": n} for c, n in settings.LANGUAGES] + locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES] return response.Ok(locales)
Add bidi (right-to-left layout) attr to locale resource
## Code Before: from django.conf import settings from taiga.base import response from taiga.base.api.viewsets import ReadOnlyListViewSet from . import permissions class LocalesViewSet(ReadOnlyListViewSet): permission_classes = (permissions.LocalesPermission,) def list(self, request, *args, **kwargs): locales = [{"code": c, "name": n} for c, n in settings.LANGUAGES] return response.Ok(locales) ## Instruction: Add bidi (right-to-left layout) attr to locale resource ## Code After: from django.conf import settings from taiga.base import response from taiga.base.api.viewsets import ReadOnlyListViewSet from . import permissions class LocalesViewSet(ReadOnlyListViewSet): permission_classes = (permissions.LocalesPermission,) def list(self, request, *args, **kwargs): locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES] return response.Ok(locales)
... def list(self, request, *args, **kwargs): locales = [{"code": c, "name": n, "bidi": c in settings.LANGUAGES_BIDI} for c, n in settings.LANGUAGES] return response.Ok(locales) ...
8803f6058255237dff39549426ca6a513a25193c
website_product_supplier/__openerp__.py
website_product_supplier/__openerp__.py
{ 'name': "Website Product Supplier", 'category': 'Website', 'version': '8.0.1.0.0', 'depends': [ 'website_sale', 'website_portal_purchase', ], 'data': [ 'security/ir.model.access.csv', 'views/product_supplier_view.xml', 'views/website_portal.xml', 'views/assets.xml', 'security/website_product_supplier_security.xml', ], 'qweb': ['static/src/xml/*.xml'], 'author': 'Antiun Ingeniería S.L., ' 'Incaser Informatica S.L., ' 'Odoo Community Association (OCA)', 'website': 'http://www.antiun.com', 'license': 'AGPL-3', 'installable': True, }
{ 'name': "Website Product Supplier", 'category': 'Website', 'version': '8.0.1.0.0', 'depends': [ 'website_sale', 'website_portal_purchase', ], 'data': [ 'security/ir.model.access.csv', 'views/product_supplier_view.xml', 'views/website_portal.xml', 'views/assets.xml', 'security/website_product_supplier_security.xml', ], 'images': [], 'qweb': ['static/src/xml/*.xml'], 'author': 'Antiun Ingeniería S.L., ' 'Incaser Informatica S.L., ' 'Odoo Community Association (OCA)', 'website': 'http://www.antiun.com', 'license': 'AGPL-3', 'installable': True, }
Add images key in manifest file
[FIX] website_product_supplier: Add images key in manifest file
Python
agpl-3.0
nuobit/website,open-synergy/website,gfcapalbo/website,LasLabs/website,acsone/website,nuobit/website,LasLabs/website,Yajo/website,LasLabs/website,gfcapalbo/website,kaerdsar/website,Yajo/website,nuobit/website,nuobit/website,Yajo/website,gfcapalbo/website,acsone/website,kaerdsar/website,LasLabs/website,open-synergy/website,acsone/website,acsone/website,gfcapalbo/website,open-synergy/website,brain-tec/website,brain-tec/website,open-synergy/website,brain-tec/website,kaerdsar/website,brain-tec/website,Yajo/website
{ 'name': "Website Product Supplier", 'category': 'Website', 'version': '8.0.1.0.0', 'depends': [ 'website_sale', 'website_portal_purchase', ], 'data': [ 'security/ir.model.access.csv', 'views/product_supplier_view.xml', 'views/website_portal.xml', 'views/assets.xml', 'security/website_product_supplier_security.xml', ], + 'images': [], 'qweb': ['static/src/xml/*.xml'], 'author': 'Antiun Ingeniería S.L., ' 'Incaser Informatica S.L., ' 'Odoo Community Association (OCA)', 'website': 'http://www.antiun.com', 'license': 'AGPL-3', 'installable': True, }
Add images key in manifest file
## Code Before: { 'name': "Website Product Supplier", 'category': 'Website', 'version': '8.0.1.0.0', 'depends': [ 'website_sale', 'website_portal_purchase', ], 'data': [ 'security/ir.model.access.csv', 'views/product_supplier_view.xml', 'views/website_portal.xml', 'views/assets.xml', 'security/website_product_supplier_security.xml', ], 'qweb': ['static/src/xml/*.xml'], 'author': 'Antiun Ingeniería S.L., ' 'Incaser Informatica S.L., ' 'Odoo Community Association (OCA)', 'website': 'http://www.antiun.com', 'license': 'AGPL-3', 'installable': True, } ## Instruction: Add images key in manifest file ## Code After: { 'name': "Website Product Supplier", 'category': 'Website', 'version': '8.0.1.0.0', 'depends': [ 'website_sale', 'website_portal_purchase', ], 'data': [ 'security/ir.model.access.csv', 'views/product_supplier_view.xml', 'views/website_portal.xml', 'views/assets.xml', 'security/website_product_supplier_security.xml', ], 'images': [], 'qweb': ['static/src/xml/*.xml'], 'author': 'Antiun Ingeniería S.L., ' 'Incaser Informatica S.L., ' 'Odoo Community Association (OCA)', 'website': 'http://www.antiun.com', 'license': 'AGPL-3', 'installable': True, }
... ], 'images': [], 'qweb': ['static/src/xml/*.xml'], ...
fb02617b29cab97a70a1a11b0d3b7b62b834aa3b
server.py
server.py
from flask import Flask from flask import request import flask import hashlib import json import gzip app = Flask(__name__) stored_files = {} @app.route('/profile/<type>', methods=['GET']) def get_dummy_files(type): if type == 'lawyer': pass elif type == 'doctor:': pass elif type == 'female': pass elif type == 'male': pass else: return "No files here\n" return "Sent files\n" @app.route('/<int:id>', methods=['GET']) def get_file(id): if id in stored_files: return stored_files[id] else: return "No such file\n" @app.route('/', methods=['POST']) def upload_file(): data = json.loads(request.data) uploaded_file = data['uploaded_file'] salt = data['salt'] id = hashlib.sha256(uploaded_file.encode()).hexdigest() stored_files[id] = (uploaded_file, salt) return "File stored\n" if __name__ == "__main__": app.run()
from flask import Flask from flask import request import flask import hashlib import json import gzip app = Flask(__name__) stored_files = {} @app.route('/profile/<type>', methods=['GET']) def get_dummy_files(type): if type == 'lawyer': gzip_address = './zipfiles/doc.tar.gz' elif type == 'doctor:': gzip_address = './zipfiles/doc.tar.gz' elif type == 'female': gzip_address = './zipfiles/doc.tar.gz' elif type == 'male': gzip_address = './zipfiles/doc.tar.gz' else: return "No files here\n" gzip_file = open(gzip_address).read() return bytearray(gzip_file) @app.route('/<int:id>', methods=['GET']) def get_file(id): if id in stored_files: return stored_files[id] else: return "No such file\n" @app.route('/', methods=['POST']) def upload_file(): data = json.loads(request.data) uploaded_file = data['uploaded_file'] salt = data['salt'] id = hashlib.sha256(uploaded_file.encode()).hexdigest() stored_files[id] = (uploaded_file, salt) return "File stored\n" if __name__ == "__main__": app.run()
Structure for sending dummy files
Structure for sending dummy files
Python
mit
rotemh/soteria
from flask import Flask from flask import request import flask import hashlib import json import gzip app = Flask(__name__) stored_files = {} @app.route('/profile/<type>', methods=['GET']) def get_dummy_files(type): if type == 'lawyer': - pass + gzip_address = './zipfiles/doc.tar.gz' elif type == 'doctor:': - pass + gzip_address = './zipfiles/doc.tar.gz' elif type == 'female': - pass + gzip_address = './zipfiles/doc.tar.gz' elif type == 'male': - pass + gzip_address = './zipfiles/doc.tar.gz' else: return "No files here\n" - return "Sent files\n" + gzip_file = open(gzip_address).read() + return bytearray(gzip_file) @app.route('/<int:id>', methods=['GET']) def get_file(id): if id in stored_files: return stored_files[id] else: return "No such file\n" @app.route('/', methods=['POST']) def upload_file(): data = json.loads(request.data) uploaded_file = data['uploaded_file'] salt = data['salt'] id = hashlib.sha256(uploaded_file.encode()).hexdigest() stored_files[id] = (uploaded_file, salt) return "File stored\n" if __name__ == "__main__": app.run()
Structure for sending dummy files
## Code Before: from flask import Flask from flask import request import flask import hashlib import json import gzip app = Flask(__name__) stored_files = {} @app.route('/profile/<type>', methods=['GET']) def get_dummy_files(type): if type == 'lawyer': pass elif type == 'doctor:': pass elif type == 'female': pass elif type == 'male': pass else: return "No files here\n" return "Sent files\n" @app.route('/<int:id>', methods=['GET']) def get_file(id): if id in stored_files: return stored_files[id] else: return "No such file\n" @app.route('/', methods=['POST']) def upload_file(): data = json.loads(request.data) uploaded_file = data['uploaded_file'] salt = data['salt'] id = hashlib.sha256(uploaded_file.encode()).hexdigest() stored_files[id] = (uploaded_file, salt) return "File stored\n" if __name__ == "__main__": app.run() ## Instruction: Structure for sending dummy files ## Code After: from flask import Flask from flask import request import flask import hashlib import json import gzip app = Flask(__name__) stored_files = {} @app.route('/profile/<type>', methods=['GET']) def get_dummy_files(type): if type == 'lawyer': gzip_address = './zipfiles/doc.tar.gz' elif type == 'doctor:': gzip_address = './zipfiles/doc.tar.gz' elif type == 'female': gzip_address = './zipfiles/doc.tar.gz' elif type == 'male': gzip_address = './zipfiles/doc.tar.gz' else: return "No files here\n" gzip_file = open(gzip_address).read() return bytearray(gzip_file) @app.route('/<int:id>', methods=['GET']) def get_file(id): if id in stored_files: return stored_files[id] else: return "No such file\n" @app.route('/', methods=['POST']) def upload_file(): data = json.loads(request.data) uploaded_file = data['uploaded_file'] salt = data['salt'] id = hashlib.sha256(uploaded_file.encode()).hexdigest() stored_files[id] = (uploaded_file, salt) return "File stored\n" if __name__ == "__main__": app.run()
# ... existing code ... if type == 'lawyer': gzip_address = './zipfiles/doc.tar.gz' elif type == 'doctor:': gzip_address = './zipfiles/doc.tar.gz' elif type == 'female': gzip_address = './zipfiles/doc.tar.gz' elif type == 'male': gzip_address = './zipfiles/doc.tar.gz' else: # ... modified code ... gzip_file = open(gzip_address).read() return bytearray(gzip_file) # ... rest of the code ...
c017d8fa711724fc7acb7e90b85f208be074d1ec
drupdates/plugins/repolist/__init__.py
drupdates/plugins/repolist/__init__.py
from drupdates.utils import * from drupdates.repos import * ''' Note: you need an ssh key set up with Stash to make this script work ''' class repolist(repoTool): def __init__(self): currentDir = os.path.dirname(os.path.realpath(__file__)) self.localsettings = Settings(currentDir) def gitRepos(self): #Get list of Stash repos in the Rain Project. repoDict = self.localsettings.get('repoDict') if not repoDict: return {} else: return repoDict
from drupdates.utils import * from drupdates.repos import * ''' Note: you need an ssh key set up with Stash to make this script work ''' class repolist(repoTool): def __init__(self): currentDir = os.path.dirname(os.path.realpath(__file__)) self.localsettings = Settings(currentDir) def gitRepos(self): #Get list of Stash repos in the Rain Project. repoDict = self.localsettings.get('repoDict') if (not repoDict) or (type(repoDict) is not dict): return {} else: return repoDict
Add extra check to repo list to verify a dictionary is returned
Add extra check to repo list to verify a dictionary is returned
Python
mit
jalama/drupdates
from drupdates.utils import * from drupdates.repos import * ''' Note: you need an ssh key set up with Stash to make this script work ''' class repolist(repoTool): def __init__(self): currentDir = os.path.dirname(os.path.realpath(__file__)) self.localsettings = Settings(currentDir) def gitRepos(self): #Get list of Stash repos in the Rain Project. repoDict = self.localsettings.get('repoDict') - if not repoDict: + if (not repoDict) or (type(repoDict) is not dict): return {} else: return repoDict
Add extra check to repo list to verify a dictionary is returned
## Code Before: from drupdates.utils import * from drupdates.repos import * ''' Note: you need an ssh key set up with Stash to make this script work ''' class repolist(repoTool): def __init__(self): currentDir = os.path.dirname(os.path.realpath(__file__)) self.localsettings = Settings(currentDir) def gitRepos(self): #Get list of Stash repos in the Rain Project. repoDict = self.localsettings.get('repoDict') if not repoDict: return {} else: return repoDict ## Instruction: Add extra check to repo list to verify a dictionary is returned ## Code After: from drupdates.utils import * from drupdates.repos import * ''' Note: you need an ssh key set up with Stash to make this script work ''' class repolist(repoTool): def __init__(self): currentDir = os.path.dirname(os.path.realpath(__file__)) self.localsettings = Settings(currentDir) def gitRepos(self): #Get list of Stash repos in the Rain Project. repoDict = self.localsettings.get('repoDict') if (not repoDict) or (type(repoDict) is not dict): return {} else: return repoDict
... repoDict = self.localsettings.get('repoDict') if (not repoDict) or (type(repoDict) is not dict): return {} ...
df6b13a70241b616f49d4dcc25073084c371f5b1
share/models/creative/base.py
share/models/creative/base.py
from django.db import models from share.models.base import ShareObject from share.models.people import Person from share.models.base import TypedShareObjectMeta from share.models.creative.meta import Venue, Institution, Funder, Award, Tag from share.models.fields import ShareForeignKey, ShareManyToManyField class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta): title = models.TextField() description = models.TextField() contributors = ShareManyToManyField(Person, through='Contributor') institutions = ShareManyToManyField(Institution, through='ThroughInstitutions') venues = ShareManyToManyField(Venue, through='ThroughVenues') funders = ShareManyToManyField(Funder, through='ThroughFunders') awards = ShareManyToManyField(Award, through='ThroughAwards') subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True) # Note: Null allows inserting of None but returns it as an empty string tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags') created = models.DateTimeField(null=True) published = models.DateTimeField(null=True) free_to_read_type = models.URLField(blank=True) free_to_read_date = models.DateTimeField(null=True) rights = models.TextField() language = models.TextField() class CreativeWork(AbstractCreativeWork): pass
from django.db import models from share.models.base import ShareObject from share.models.people import Person from share.models.base import TypedShareObjectMeta from share.models.creative.meta import Venue, Institution, Funder, Award, Tag from share.models.fields import ShareForeignKey, ShareManyToManyField class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta): title = models.TextField() description = models.TextField() contributors = ShareManyToManyField(Person, through='Contributor') institutions = ShareManyToManyField(Institution, through='ThroughInstitutions') venues = ShareManyToManyField(Venue, through='ThroughVenues') funders = ShareManyToManyField(Funder, through='ThroughFunders') awards = ShareManyToManyField(Award, through='ThroughAwards') subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True) # Note: Null allows inserting of None but returns it as an empty string tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags') created = models.DateTimeField(null=True) published = models.DateTimeField(null=True) free_to_read_type = models.URLField(blank=True) free_to_read_date = models.DateTimeField(null=True) rights = models.TextField(blank=True, null=True) language = models.TextField(blank=True, null=True) class CreativeWork(AbstractCreativeWork): pass
Swap out license with rights
Swap out license with rights
Python
apache-2.0
CenterForOpenScience/SHARE,aaxelb/SHARE,zamattiac/SHARE,zamattiac/SHARE,CenterForOpenScience/SHARE,CenterForOpenScience/SHARE,aaxelb/SHARE,zamattiac/SHARE,laurenbarker/SHARE,aaxelb/SHARE,laurenbarker/SHARE,laurenbarker/SHARE
from django.db import models from share.models.base import ShareObject from share.models.people import Person from share.models.base import TypedShareObjectMeta from share.models.creative.meta import Venue, Institution, Funder, Award, Tag from share.models.fields import ShareForeignKey, ShareManyToManyField class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta): title = models.TextField() description = models.TextField() contributors = ShareManyToManyField(Person, through='Contributor') institutions = ShareManyToManyField(Institution, through='ThroughInstitutions') venues = ShareManyToManyField(Venue, through='ThroughVenues') funders = ShareManyToManyField(Funder, through='ThroughFunders') awards = ShareManyToManyField(Award, through='ThroughAwards') subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True) # Note: Null allows inserting of None but returns it as an empty string tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags') created = models.DateTimeField(null=True) published = models.DateTimeField(null=True) free_to_read_type = models.URLField(blank=True) free_to_read_date = models.DateTimeField(null=True) - rights = models.TextField() + + rights = models.TextField(blank=True, null=True) - language = models.TextField() + language = models.TextField(blank=True, null=True) class CreativeWork(AbstractCreativeWork): pass
Swap out license with rights
## Code Before: from django.db import models from share.models.base import ShareObject from share.models.people import Person from share.models.base import TypedShareObjectMeta from share.models.creative.meta import Venue, Institution, Funder, Award, Tag from share.models.fields import ShareForeignKey, ShareManyToManyField class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta): title = models.TextField() description = models.TextField() contributors = ShareManyToManyField(Person, through='Contributor') institutions = ShareManyToManyField(Institution, through='ThroughInstitutions') venues = ShareManyToManyField(Venue, through='ThroughVenues') funders = ShareManyToManyField(Funder, through='ThroughFunders') awards = ShareManyToManyField(Award, through='ThroughAwards') subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True) # Note: Null allows inserting of None but returns it as an empty string tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags') created = models.DateTimeField(null=True) published = models.DateTimeField(null=True) free_to_read_type = models.URLField(blank=True) free_to_read_date = models.DateTimeField(null=True) rights = models.TextField() language = models.TextField() class CreativeWork(AbstractCreativeWork): pass ## Instruction: Swap out license with rights ## Code After: from django.db import models from share.models.base import ShareObject from share.models.people import Person from share.models.base import TypedShareObjectMeta from share.models.creative.meta import Venue, Institution, Funder, Award, Tag from share.models.fields import ShareForeignKey, ShareManyToManyField class AbstractCreativeWork(ShareObject, metaclass=TypedShareObjectMeta): title = models.TextField() description = models.TextField() contributors = ShareManyToManyField(Person, through='Contributor') institutions = ShareManyToManyField(Institution, through='ThroughInstitutions') venues = ShareManyToManyField(Venue, through='ThroughVenues') funders = ShareManyToManyField(Funder, through='ThroughFunders') awards = ShareManyToManyField(Award, through='ThroughAwards') subject = ShareForeignKey(Tag, related_name='subjected_%(class)s', null=True) # Note: Null allows inserting of None but returns it as an empty string tags = ShareManyToManyField(Tag, related_name='tagged_%(class)s', through='ThroughTags') created = models.DateTimeField(null=True) published = models.DateTimeField(null=True) free_to_read_type = models.URLField(blank=True) free_to_read_date = models.DateTimeField(null=True) rights = models.TextField(blank=True, null=True) language = models.TextField(blank=True, null=True) class CreativeWork(AbstractCreativeWork): pass
... free_to_read_date = models.DateTimeField(null=True) rights = models.TextField(blank=True, null=True) language = models.TextField(blank=True, null=True) ...
c2598058722531662aab8831640fc367689d2a43
tests/utils/test_process_word_vectors.py
tests/utils/test_process_word_vectors.py
import inspect import os import pytest import numpy as np from subprocess import call from utils.preprocess_text_word_vectors import txtvec2npy def test_text_word2vec2npy(): # check whether files are present in folder vectors_name = 'wiki.fiu_vro.vec' path = os.path.dirname(inspect.getfile(inspect.currentframe())) if not os.path.exists(path + '/' + vectors_name): call(["wget https://s3-us-west-1.amazonaws.com/fasttext-vectors/" + vectors_name + " -O " + path + "/" + vectors_name], shell=True) txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4]) vectors = np.load('./' + vectors_name[:-4] + '.npy').item() assert len(list(vectors)) == 8769 assert vectors['kihlkunnan'].shape[0] == 300 if __name__ == '__main__': pytest.main([__file__])
import inspect import os import pytest import numpy as np from subprocess import call from utils.preprocess_text_word_vectors import txtvec2npy def test_text_word2vec2npy(): # check whether files are present in folder vectors_name = 'wiki.fiu_vro.vec' path = os.path.dirname(inspect.getfile(inspect.currentframe())) if not os.path.exists(path + '/' + vectors_name): call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " + path + "/" + vectors_name], shell=True) txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4]) vectors = np.load('./' + vectors_name[:-4] + '.npy').item() assert len(list(vectors)) == 8769 assert vectors['kihlkunnan'].shape[0] == 300 if __name__ == '__main__': pytest.main([__file__])
Update Fasttext pretrained vectors location
Update Fasttext pretrained vectors location
Python
mit
lvapeab/nmt-keras,lvapeab/nmt-keras
import inspect import os import pytest import numpy as np from subprocess import call from utils.preprocess_text_word_vectors import txtvec2npy def test_text_word2vec2npy(): # check whether files are present in folder vectors_name = 'wiki.fiu_vro.vec' path = os.path.dirname(inspect.getfile(inspect.currentframe())) if not os.path.exists(path + '/' + vectors_name): - call(["wget https://s3-us-west-1.amazonaws.com/fasttext-vectors/" + vectors_name + " -O " + + call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " + path + "/" + vectors_name], shell=True) txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4]) vectors = np.load('./' + vectors_name[:-4] + '.npy').item() assert len(list(vectors)) == 8769 assert vectors['kihlkunnan'].shape[0] == 300 if __name__ == '__main__': pytest.main([__file__])
Update Fasttext pretrained vectors location
## Code Before: import inspect import os import pytest import numpy as np from subprocess import call from utils.preprocess_text_word_vectors import txtvec2npy def test_text_word2vec2npy(): # check whether files are present in folder vectors_name = 'wiki.fiu_vro.vec' path = os.path.dirname(inspect.getfile(inspect.currentframe())) if not os.path.exists(path + '/' + vectors_name): call(["wget https://s3-us-west-1.amazonaws.com/fasttext-vectors/" + vectors_name + " -O " + path + "/" + vectors_name], shell=True) txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4]) vectors = np.load('./' + vectors_name[:-4] + '.npy').item() assert len(list(vectors)) == 8769 assert vectors['kihlkunnan'].shape[0] == 300 if __name__ == '__main__': pytest.main([__file__]) ## Instruction: Update Fasttext pretrained vectors location ## Code After: import inspect import os import pytest import numpy as np from subprocess import call from utils.preprocess_text_word_vectors import txtvec2npy def test_text_word2vec2npy(): # check whether files are present in folder vectors_name = 'wiki.fiu_vro.vec' path = os.path.dirname(inspect.getfile(inspect.currentframe())) if not os.path.exists(path + '/' + vectors_name): call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " + path + "/" + vectors_name], shell=True) txtvec2npy(path + '/' + vectors_name, './', vectors_name[:-4]) vectors = np.load('./' + vectors_name[:-4] + '.npy').item() assert len(list(vectors)) == 8769 assert vectors['kihlkunnan'].shape[0] == 300 if __name__ == '__main__': pytest.main([__file__])
# ... existing code ... if not os.path.exists(path + '/' + vectors_name): call(["wget https://dl.fbaipublicfiles.com/fasttext/vectors-wiki/" + vectors_name + " -O " + path + "/" + vectors_name], # ... rest of the code ...
bbf3d68b9566a826f404aa1ab3da198d765dca58
contacts/rules.py
contacts/rules.py
ALLOWED_FIELDS = [ 'name', 'first_name', 'last_name', 'phone_number', 'photo', 'email', 'twitter' ]
ALLOWED_FIELDS = [ 'name', 'phone_number', 'first_name', 'last_name', 'phone_number', 'photo', 'email', 'twitter' ]
Add 'phone_number' field to ALLOWED_FIELDS.
Add 'phone_number' field to ALLOWED_FIELDS.
Python
mit
heimann/contacts
ALLOWED_FIELDS = [ 'name', + 'phone_number', 'first_name', 'last_name', 'phone_number', 'photo', 'email', 'twitter' ]
Add 'phone_number' field to ALLOWED_FIELDS.
## Code Before: ALLOWED_FIELDS = [ 'name', 'first_name', 'last_name', 'phone_number', 'photo', 'email', 'twitter' ] ## Instruction: Add 'phone_number' field to ALLOWED_FIELDS. ## Code After: ALLOWED_FIELDS = [ 'name', 'phone_number', 'first_name', 'last_name', 'phone_number', 'photo', 'email', 'twitter' ]
# ... existing code ... 'name', 'phone_number', 'first_name', # ... rest of the code ...
8164d048b47299377b4db7d9fc0198e24b07bdb3
engine/geometry.py
engine/geometry.py
from math import cos, sin, pi, hypot def rotate(polygon, angle): rotated_points = [] cos_result = cos(angle) sin_result = sin(angle) for point in polygon: x = point[0] * cos_result - point[1] * sin_result y = point[0] * sin_result + point[1] * cos_result rotated_points.append((x, y)) return rotated_points def move(point, direction, amount): return [point[0] + amount * cos(direction), point[1] + amount * sin(direction)] def distance(point1, point2): return hypot(point1[0] - point2[0], point1[1] - point2[1])
from math import cos, sin, pi, hypot def rotate(polygon, angle): rotated_points = [] cos_result = cos(angle) sin_result = sin(angle) for point in polygon: x = point[0] * cos_result - point[1] * sin_result y = point[0] * sin_result + point[1] * cos_result rotated_points.append((x, y)) return rotated_points def move(point, direction, amount): return [int(point[0] + amount * cos(direction)), int(point[1] + amount * sin(direction))] def distance(point1, point2): return hypot(point1[0] - point2[0], point1[1] - point2[1])
Fix move to return only int, draw functions cannot handle floats as coordinates
Fix move to return only int, draw functions cannot handle floats as coordinates
Python
apache-2.0
PGHM/spacebattle
from math import cos, sin, pi, hypot def rotate(polygon, angle): rotated_points = [] cos_result = cos(angle) sin_result = sin(angle) for point in polygon: x = point[0] * cos_result - point[1] * sin_result y = point[0] * sin_result + point[1] * cos_result rotated_points.append((x, y)) return rotated_points def move(point, direction, amount): - return [point[0] + amount * cos(direction), + return [int(point[0] + amount * cos(direction)), - point[1] + amount * sin(direction)] + int(point[1] + amount * sin(direction))] def distance(point1, point2): return hypot(point1[0] - point2[0], point1[1] - point2[1])
Fix move to return only int, draw functions cannot handle floats as coordinates
## Code Before: from math import cos, sin, pi, hypot def rotate(polygon, angle): rotated_points = [] cos_result = cos(angle) sin_result = sin(angle) for point in polygon: x = point[0] * cos_result - point[1] * sin_result y = point[0] * sin_result + point[1] * cos_result rotated_points.append((x, y)) return rotated_points def move(point, direction, amount): return [point[0] + amount * cos(direction), point[1] + amount * sin(direction)] def distance(point1, point2): return hypot(point1[0] - point2[0], point1[1] - point2[1]) ## Instruction: Fix move to return only int, draw functions cannot handle floats as coordinates ## Code After: from math import cos, sin, pi, hypot def rotate(polygon, angle): rotated_points = [] cos_result = cos(angle) sin_result = sin(angle) for point in polygon: x = point[0] * cos_result - point[1] * sin_result y = point[0] * sin_result + point[1] * cos_result rotated_points.append((x, y)) return rotated_points def move(point, direction, amount): return [int(point[0] + amount * cos(direction)), int(point[1] + amount * sin(direction))] def distance(point1, point2): return hypot(point1[0] - point2[0], point1[1] - point2[1])
... def move(point, direction, amount): return [int(point[0] + amount * cos(direction)), int(point[1] + amount * sin(direction))] ...
ddc44c6673cff4121eaaa47d8d075d63b82a85fe
runreport.py
runreport.py
import os import json import saulify.sitespec as sitespec SPEC_DIRECTORY = "sitespecs" if __name__ == "__main__": for fname in os.listdir(SPEC_DIRECTORY): fpath = os.path.join(SPEC_DIRECTORY, fname) test_cases = sitespec.load_testcases(fpath) for test_case in test_cases: result = test_case.run() print(json.dumps(result))
import os import json import argparse import saulify.sitespec as sitespec SPEC_DIRECTORY = "sitespecs" parser = argparse.ArgumentParser() parser.add_argument("-p", "--pretty", help="Pretty print test results", action="store_true") args = parser.parse_args() def test_passed(report): """ Whether all components of a scraper test succeeded """ if report["status"] != "OK": return False for result in report["result"].values(): if result["missing"]: return False return True def print_report(report): """ Converts test report dictionary to a human-readable format """ if report["status"] == "OK": result = "PASS" if test_passed(report) else "FAIL" else: result = "EXCEPTION" print("{0} : {1}".format(result, report["url"])) if report["status"] == "EXCEPTION": print(report["message"]) elif test_passed(report): r = report["result"] stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r]) print("Found " + stats) else: for category, result in report["result"].items(): if result["missing"]: count = len(result["missing"]) print("Missing {0} {1}:".format(count, category)) for item in result["missing"]: print(item) if __name__ == "__main__": for fname in os.listdir(SPEC_DIRECTORY): fpath = os.path.join(SPEC_DIRECTORY, fname) test_cases = sitespec.load_testcases(fpath) for test_case in test_cases: report = test_case.run() if args.pretty: print_report(report) print("\n") else: print(json.dumps(report))
Add optional pretty printing to test runner
Add optional pretty printing to test runner
Python
agpl-3.0
asm-products/saulify-web,asm-products/saulify-web,asm-products/saulify-web
import os import json + import argparse import saulify.sitespec as sitespec SPEC_DIRECTORY = "sitespecs" + + + parser = argparse.ArgumentParser() + parser.add_argument("-p", "--pretty", help="Pretty print test results", + action="store_true") + args = parser.parse_args() + + + def test_passed(report): + """ Whether all components of a scraper test succeeded """ + + if report["status"] != "OK": + return False + + for result in report["result"].values(): + if result["missing"]: + return False + + return True + + + def print_report(report): + """ Converts test report dictionary to a human-readable format """ + + if report["status"] == "OK": + result = "PASS" if test_passed(report) else "FAIL" + else: + result = "EXCEPTION" + + print("{0} : {1}".format(result, report["url"])) + + if report["status"] == "EXCEPTION": + print(report["message"]) + + elif test_passed(report): + r = report["result"] + stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r]) + print("Found " + stats) + + else: + for category, result in report["result"].items(): + if result["missing"]: + count = len(result["missing"]) + print("Missing {0} {1}:".format(count, category)) + for item in result["missing"]: + print(item) if __name__ == "__main__": for fname in os.listdir(SPEC_DIRECTORY): fpath = os.path.join(SPEC_DIRECTORY, fname) test_cases = sitespec.load_testcases(fpath) for test_case in test_cases: - result = test_case.run() + report = test_case.run() + if args.pretty: + print_report(report) + print("\n") + else: - print(json.dumps(result)) + print(json.dumps(report))
Add optional pretty printing to test runner
## Code Before: import os import json import saulify.sitespec as sitespec SPEC_DIRECTORY = "sitespecs" if __name__ == "__main__": for fname in os.listdir(SPEC_DIRECTORY): fpath = os.path.join(SPEC_DIRECTORY, fname) test_cases = sitespec.load_testcases(fpath) for test_case in test_cases: result = test_case.run() print(json.dumps(result)) ## Instruction: Add optional pretty printing to test runner ## Code After: import os import json import argparse import saulify.sitespec as sitespec SPEC_DIRECTORY = "sitespecs" parser = argparse.ArgumentParser() parser.add_argument("-p", "--pretty", help="Pretty print test results", action="store_true") args = parser.parse_args() def test_passed(report): """ Whether all components of a scraper test succeeded """ if report["status"] != "OK": return False for result in report["result"].values(): if result["missing"]: return False return True def print_report(report): """ Converts test report dictionary to a human-readable format """ if report["status"] == "OK": result = "PASS" if test_passed(report) else "FAIL" else: result = "EXCEPTION" print("{0} : {1}".format(result, report["url"])) if report["status"] == "EXCEPTION": print(report["message"]) elif test_passed(report): r = report["result"] stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r]) print("Found " + stats) else: for category, result in report["result"].items(): if result["missing"]: count = len(result["missing"]) print("Missing {0} {1}:".format(count, category)) for item in result["missing"]: print(item) if __name__ == "__main__": for fname in os.listdir(SPEC_DIRECTORY): fpath = os.path.join(SPEC_DIRECTORY, fname) test_cases = sitespec.load_testcases(fpath) for test_case in test_cases: report = test_case.run() if args.pretty: print_report(report) print("\n") else: print(json.dumps(report))
# ... existing code ... import json import argparse # ... modified code ... SPEC_DIRECTORY = "sitespecs" parser = argparse.ArgumentParser() parser.add_argument("-p", "--pretty", help="Pretty print test results", action="store_true") args = parser.parse_args() def test_passed(report): """ Whether all components of a scraper test succeeded """ if report["status"] != "OK": return False for result in report["result"].values(): if result["missing"]: return False return True def print_report(report): """ Converts test report dictionary to a human-readable format """ if report["status"] == "OK": result = "PASS" if test_passed(report) else "FAIL" else: result = "EXCEPTION" print("{0} : {1}".format(result, report["url"])) if report["status"] == "EXCEPTION": print(report["message"]) elif test_passed(report): r = report["result"] stats = ", ".join(["{0} {1}".format(len(r[c]["found"]), c) for c in r]) print("Found " + stats) else: for category, result in report["result"].items(): if result["missing"]: count = len(result["missing"]) print("Missing {0} {1}:".format(count, category)) for item in result["missing"]: print(item) ... for test_case in test_cases: report = test_case.run() if args.pretty: print_report(report) print("\n") else: print(json.dumps(report)) # ... rest of the code ...
a35d6f59d214741f554dde1363d2eac7addb04cb
crypto_enigma/__init__.py
crypto_enigma/__init__.py
"""An Enigma machine simulator with rich textual display functionality.""" from ._version import __version__, __author__ #__all__ = ['machine', 'components'] from .components import * from .machine import *
from ._version import __version__, __author__ #__all__ = ['machine', 'components'] from .components import * from .machine import *
Add limitations to package documentation
Add limitations to package documentation
Python
bsd-3-clause
orome/crypto-enigma-py
- - """An Enigma machine simulator with rich textual display functionality.""" from ._version import __version__, __author__ #__all__ = ['machine', 'components'] from .components import * from .machine import *
Add limitations to package documentation
## Code Before: """An Enigma machine simulator with rich textual display functionality.""" from ._version import __version__, __author__ #__all__ = ['machine', 'components'] from .components import * from .machine import * ## Instruction: Add limitations to package documentation ## Code After: from ._version import __version__, __author__ #__all__ = ['machine', 'components'] from .components import * from .machine import *
// ... existing code ... // ... rest of the code ...
ba1186c47e5f3466faeea9f2d5bf96948d5f7183
confuzzle.py
confuzzle.py
import sys import argparse import yaml from jinja2 import Template def render(template_string, context_dict): template = Template(template_string) return template.render(**context_dict) def main(): parser = argparse.ArgumentParser() parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used") parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read") parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used") args = parser.parse_args() context_dict = yaml.load(args.config.read()) template_string = args.template.read() rendered = render(template_string, context_dict) args.out.write(rendered) if __name__ == "__main__": main()
import sys import argparse import yaml import jinja2 def render(template_string, context_dict, strict=False): template = jinja2.Template(template_string) if strict: template.environment.undefined = jinja2.StrictUndefined return template.render(**context_dict) def main(): parser = argparse.ArgumentParser() parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used") parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read") parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used") parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables") args = parser.parse_args() context_dict = yaml.load(args.config.read()) template_string = args.template.read() rendered = render(template_string, context_dict, args.strict) args.out.write(rendered) if __name__ == "__main__": main()
Add --strict flag to raise exception on undefined variables
Add --strict flag to raise exception on undefined variables
Python
unlicense
j4mie/confuzzle
import sys import argparse import yaml - from jinja2 import Template + import jinja2 - def render(template_string, context_dict): + def render(template_string, context_dict, strict=False): - template = Template(template_string) + template = jinja2.Template(template_string) + if strict: + template.environment.undefined = jinja2.StrictUndefined return template.render(**context_dict) def main(): parser = argparse.ArgumentParser() parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used") parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read") parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used") + parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables") args = parser.parse_args() context_dict = yaml.load(args.config.read()) template_string = args.template.read() - rendered = render(template_string, context_dict) + rendered = render(template_string, context_dict, args.strict) args.out.write(rendered) if __name__ == "__main__": main()
Add --strict flag to raise exception on undefined variables
## Code Before: import sys import argparse import yaml from jinja2 import Template def render(template_string, context_dict): template = Template(template_string) return template.render(**context_dict) def main(): parser = argparse.ArgumentParser() parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used") parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read") parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used") args = parser.parse_args() context_dict = yaml.load(args.config.read()) template_string = args.template.read() rendered = render(template_string, context_dict) args.out.write(rendered) if __name__ == "__main__": main() ## Instruction: Add --strict flag to raise exception on undefined variables ## Code After: import sys import argparse import yaml import jinja2 def render(template_string, context_dict, strict=False): template = jinja2.Template(template_string) if strict: template.environment.undefined = jinja2.StrictUndefined return template.render(**context_dict) def main(): parser = argparse.ArgumentParser() parser.add_argument('template', nargs='?', type=argparse.FileType('r'), default=sys.stdin, help="Config file template. If not supplied, stdin is used") parser.add_argument('config', type=argparse.FileType('r'), help="YAML data file to read") parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used") parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables") args = parser.parse_args() context_dict = yaml.load(args.config.read()) template_string = args.template.read() rendered = render(template_string, context_dict, args.strict) args.out.write(rendered) if __name__ == "__main__": main()
# ... existing code ... import yaml import jinja2 # ... modified code ... def render(template_string, context_dict, strict=False): template = jinja2.Template(template_string) if strict: template.environment.undefined = jinja2.StrictUndefined return template.render(**context_dict) ... parser.add_argument('--out', '-o', dest='out', type=argparse.FileType('w'), default=sys.stdout, help="Output file to write. If not supplied, stdout is used") parser.add_argument('--strict', dest='strict', action='store_true', default=False, help="Raise an exception on undefined variables") ... rendered = render(template_string, context_dict, args.strict) args.out.write(rendered) # ... rest of the code ...
e890ac9ef00193beac77b757c62911553cebf656
test.py
test.py
import urllib urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', '/home/pi/img/img.jpg')
import urllib urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg')
Change save path to local path
Change save path to local path
Python
mit
adampiskorski/lpr_poc
import urllib - urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', '/home/pi/img/img.jpg') + urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg')
Change save path to local path
## Code Before: import urllib urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', '/home/pi/img/img.jpg') ## Instruction: Change save path to local path ## Code After: import urllib urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg')
// ... existing code ... import urllib urllib.urlretrieve('http://192.168.0.13:8080/photoaf.jpg', 'img.jpg') // ... rest of the code ...
1cc6ec9f328d3ce045a4a1a50138b11c0b23cc3a
pyfr/ctypesutil.py
pyfr/ctypesutil.py
import ctypes import ctypes.util import os import sys def find_libc(): if sys.platform == 'win32': return ctypes.util.find_msvcrt() else: return ctypes.util.find_library('c') def load_library(name): lname = platform_libname(name) sdirs = platform_libdirs() # First attempt to utilise the system search path try: return ctypes.CDLL(lname) # Otherwise, if this fails then run our own search except OSError: for sd in sdirs: try: return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname))) except OSError: pass else: raise OSError('Unable to load {0}'.format(name)) def platform_libname(name): if sys.platform == 'darwin': return 'lib{0}.dylib'.format(name) elif sys.platform == 'win32': return '{0}.dll'.format(name) else: return 'lib{0}.so'.format(name) def platform_libdirs(): path = os.environ.get('PYFR_LIBRARY_PATH', '') dirs = [d for d in path.split(':') if d] # On Mac OS X append the default path used by MacPorts if sys.platform == 'darwin': return dirs + ['/opt/local/lib'] # Otherwise just return else: return dirs
import ctypes import ctypes.util import os import sys def find_libc(): if sys.platform == 'win32': return ctypes.util.find_msvcrt() else: return ctypes.util.find_library('c') def load_library(name): # If an explicit override has been given then use it lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper())) if lpath: return ctypes.CDLL(lpath) # Otherwise synthesise the library name and start searching lname = platform_libname(name) # Start with system search path try: return ctypes.CDLL(lname) # ..and if this fails then run our own search except OSError: for sd in platform_libdirs(): try: return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname))) except OSError: pass else: raise OSError('Unable to load {0}'.format(name)) def platform_libname(name): if sys.platform == 'darwin': return 'lib{0}.dylib'.format(name) elif sys.platform == 'win32': return '{0}.dll'.format(name) else: return 'lib{0}.so'.format(name) def platform_libdirs(): path = os.environ.get('PYFR_LIBRARY_PATH', '') dirs = [d for d in path.split(':') if d] # On Mac OS X append the default path used by MacPorts if sys.platform == 'darwin': return dirs + ['/opt/local/lib'] # Otherwise just return else: return dirs
Enable library paths to be explicitly specified.
Enable library paths to be explicitly specified. All shared libraries loaded through the load_library function can bow be specified explicitly through a suitable environmental variable PYFR_<LIB>_LIBRARY_PATH=/path/to/lib.here where <LIB> corresponds to the name of the library, e.g. METIS.
Python
bsd-3-clause
BrianVermeire/PyFR
import ctypes import ctypes.util import os import sys def find_libc(): if sys.platform == 'win32': return ctypes.util.find_msvcrt() else: return ctypes.util.find_library('c') def load_library(name): + # If an explicit override has been given then use it + lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper())) + if lpath: + return ctypes.CDLL(lpath) + + # Otherwise synthesise the library name and start searching lname = platform_libname(name) - sdirs = platform_libdirs() - # First attempt to utilise the system search path + # Start with system search path try: return ctypes.CDLL(lname) - # Otherwise, if this fails then run our own search + # ..and if this fails then run our own search except OSError: - for sd in sdirs: + for sd in platform_libdirs(): try: return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname))) except OSError: pass else: raise OSError('Unable to load {0}'.format(name)) def platform_libname(name): if sys.platform == 'darwin': return 'lib{0}.dylib'.format(name) elif sys.platform == 'win32': return '{0}.dll'.format(name) else: return 'lib{0}.so'.format(name) def platform_libdirs(): path = os.environ.get('PYFR_LIBRARY_PATH', '') dirs = [d for d in path.split(':') if d] # On Mac OS X append the default path used by MacPorts if sys.platform == 'darwin': return dirs + ['/opt/local/lib'] # Otherwise just return else: return dirs
Enable library paths to be explicitly specified.
## Code Before: import ctypes import ctypes.util import os import sys def find_libc(): if sys.platform == 'win32': return ctypes.util.find_msvcrt() else: return ctypes.util.find_library('c') def load_library(name): lname = platform_libname(name) sdirs = platform_libdirs() # First attempt to utilise the system search path try: return ctypes.CDLL(lname) # Otherwise, if this fails then run our own search except OSError: for sd in sdirs: try: return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname))) except OSError: pass else: raise OSError('Unable to load {0}'.format(name)) def platform_libname(name): if sys.platform == 'darwin': return 'lib{0}.dylib'.format(name) elif sys.platform == 'win32': return '{0}.dll'.format(name) else: return 'lib{0}.so'.format(name) def platform_libdirs(): path = os.environ.get('PYFR_LIBRARY_PATH', '') dirs = [d for d in path.split(':') if d] # On Mac OS X append the default path used by MacPorts if sys.platform == 'darwin': return dirs + ['/opt/local/lib'] # Otherwise just return else: return dirs ## Instruction: Enable library paths to be explicitly specified. ## Code After: import ctypes import ctypes.util import os import sys def find_libc(): if sys.platform == 'win32': return ctypes.util.find_msvcrt() else: return ctypes.util.find_library('c') def load_library(name): # If an explicit override has been given then use it lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper())) if lpath: return ctypes.CDLL(lpath) # Otherwise synthesise the library name and start searching lname = platform_libname(name) # Start with system search path try: return ctypes.CDLL(lname) # ..and if this fails then run our own search except OSError: for sd in platform_libdirs(): try: return ctypes.CDLL(os.path.abspath(os.path.join(sd, lname))) except OSError: pass else: raise OSError('Unable to load {0}'.format(name)) def platform_libname(name): if sys.platform == 'darwin': return 'lib{0}.dylib'.format(name) elif sys.platform == 'win32': return '{0}.dll'.format(name) else: return 'lib{0}.so'.format(name) def platform_libdirs(): path = os.environ.get('PYFR_LIBRARY_PATH', '') dirs = [d for d in path.split(':') if d] # On Mac OS X append the default path used by MacPorts if sys.platform == 'darwin': return dirs + ['/opt/local/lib'] # Otherwise just return else: return dirs
# ... existing code ... def load_library(name): # If an explicit override has been given then use it lpath = os.environ.get('PYFR_{0}_LIBRARY_PATH'.format(name.upper())) if lpath: return ctypes.CDLL(lpath) # Otherwise synthesise the library name and start searching lname = platform_libname(name) # Start with system search path try: # ... modified code ... return ctypes.CDLL(lname) # ..and if this fails then run our own search except OSError: for sd in platform_libdirs(): try: # ... rest of the code ...
86a992dc15482087773f1591752a667a6014ba5d
docker/settings/celery.py
docker/settings/celery.py
from .docker_compose import DockerBaseSettings class CeleryDevSettings(DockerBaseSettings): pass CeleryDevSettings.load_settings(__name__)
from .docker_compose import DockerBaseSettings class CeleryDevSettings(DockerBaseSettings): # Since we can't properly set CORS on Azurite container # (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561) # trying to fetch ``objects.inv`` from celery container fails because the # URL is like http://docs.dev.readthedocs.io/... and it should be # http://storage:10000/... This setting fixes that. # Once we can use CORS, we should define this setting in the # ``docker_compose.py`` file instead. AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000' CeleryDevSettings.load_settings(__name__)
Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME
Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME We can't access docs.dev.readthedocs.io from celery container because that domain points to 127.0.0.1 and we don't have the storage in that IP. So, we need to override the AZURE_MEDIA_STORAGE_HOSTNAME in the celery container to point to the storage. We should do this directly in `docker_compose.py` settings file, but since we can't configure CORS in Azurite we can't do it yet.
Python
mit
rtfd/readthedocs.org,rtfd/readthedocs.org,rtfd/readthedocs.org,rtfd/readthedocs.org
from .docker_compose import DockerBaseSettings class CeleryDevSettings(DockerBaseSettings): - pass + # Since we can't properly set CORS on Azurite container + # (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561) + # trying to fetch ``objects.inv`` from celery container fails because the + # URL is like http://docs.dev.readthedocs.io/... and it should be + # http://storage:10000/... This setting fixes that. + # Once we can use CORS, we should define this setting in the + # ``docker_compose.py`` file instead. + AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000' + CeleryDevSettings.load_settings(__name__)
Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME
## Code Before: from .docker_compose import DockerBaseSettings class CeleryDevSettings(DockerBaseSettings): pass CeleryDevSettings.load_settings(__name__) ## Instruction: Use proper domain for AZURE_MEDIA_STORAGE_HOSTNAME ## Code After: from .docker_compose import DockerBaseSettings class CeleryDevSettings(DockerBaseSettings): # Since we can't properly set CORS on Azurite container # (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561) # trying to fetch ``objects.inv`` from celery container fails because the # URL is like http://docs.dev.readthedocs.io/... and it should be # http://storage:10000/... This setting fixes that. # Once we can use CORS, we should define this setting in the # ``docker_compose.py`` file instead. AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000' CeleryDevSettings.load_settings(__name__)
# ... existing code ... class CeleryDevSettings(DockerBaseSettings): # Since we can't properly set CORS on Azurite container # (see https://github.com/Azure/Azurite/issues/55#issuecomment-503380561) # trying to fetch ``objects.inv`` from celery container fails because the # URL is like http://docs.dev.readthedocs.io/... and it should be # http://storage:10000/... This setting fixes that. # Once we can use CORS, we should define this setting in the # ``docker_compose.py`` file instead. AZURE_MEDIA_STORAGE_HOSTNAME = 'storage:10000' # ... rest of the code ...
1666f883e3f6a497971b484c9ba875df2f6693a2
test/testall.py
test/testall.py
import os import re import sys from _common import unittest pkgpath = os.path.dirname(__file__) or '.' sys.path.append(pkgpath) os.chdir(pkgpath) def suite(): s = unittest.TestSuite() # Get the suite() of every module in this directory beginning with # "test_". for fname in os.listdir(pkgpath): match = re.match(r'(test_\S+)\.py$', fname) if match: modname = match.group(1) s.addTest(__import__(modname).suite()) return s if __name__ == '__main__': unittest.main(defaultTest='suite')
import os import re import sys from _common import unittest pkgpath = os.path.dirname(__file__) or '.' sys.path.append(pkgpath) os.chdir(pkgpath) # Make sure we use local version of beetsplug and not system namespaced version # for tests try: del sys.modules["beetsplug"] except KeyError: pass def suite(): s = unittest.TestSuite() # Get the suite() of every module in this directory beginning with # "test_". for fname in os.listdir(pkgpath): match = re.match(r'(test_\S+)\.py$', fname) if match: modname = match.group(1) s.addTest(__import__(modname).suite()) return s if __name__ == '__main__': unittest.main(defaultTest='suite')
Fix python namespaces for test runs
Fix python namespaces for test runs We need to make sure we don't use namespaced versions that are already installed on the system but rather use local version from current sources
Python
mit
SusannaMaria/beets,mathstuf/beets,mathstuf/beets,YetAnotherNerd/beets,lengtche/beets,LordSputnik/beets,shamangeorge/beets,ibmibmibm/beets,m-urban/beets,krig/beets,lightwang1/beets,shamangeorge/beets,MyTunesFreeMusic/privacy-policy,jcoady9/beets,SusannaMaria/beets,beetbox/beets,Andypsamp/CODfinalJUNIT,Andypsamp/CODfinalJUNIT,jcoady9/beets,pkess/beets,PierreRust/beets,tima/beets,mried/beets,pkess/beets,Freso/beets,bj-yinyan/beets,beetbox/beets,dfc/beets,YetAnotherNerd/beets,tima/beets,ruippeixotog/beets,diego-plan9/beets,drm00/beets,ruippeixotog/beets,marcuskrahl/beets,kareemallen/beets,arabenjamin/beets,drm00/beets,parapente/beets,Dishwishy/beets,madmouser1/beets,imsparsh/beets,Freso/beets,mathstuf/beets,andremiller/beets,LordSputnik/beets,moodboom/beets,YetAnotherNerd/beets,mosesfistos1/beetbox,multikatt/beets,jackwilsdon/beets,jmwatte/beets,jayme-github/beets,asteven/beets,xsteadfastx/beets,m-urban/beets,bj-yinyan/beets,YetAnotherNerd/beets,LordSputnik/beets,Dishwishy/beets,Kraymer/beets,mosesfistos1/beetbox,ruippeixotog/beets,jcoady9/beets,randybias/beets,untitaker/beets,PierreRust/beets,beetbox/beets,mried/beets,artemutin/beets,shanemikel/beets,Freso/beets,Andypsamp/CODfinalJUNIT,lightwang1/beets,shanemikel/beets,kelvinhammond/beets,mried/beets,gabrielaraujof/beets,ttsda/beets,randybias/beets,krig/beets,sadatay/beets,sampsyo/beets,parapente/beets,kareemallen/beets,ttsda/beets,swt30/beets,PierreRust/beets,imsparsh/beets,sampsyo/beets,madmouser1/beets,gabrielaraujof/beets,pkess/beets,jackwilsdon/beets,m-urban/beets,arabenjamin/beets,drm00/beets,Andypsamp/CODfinalJUNIT,beetbox/beets,sadatay/beets,ibmibmibm/beets,untitaker/beets,moodboom/beets,SusannaMaria/beets,sampsyo/beets,tima/beets,ttsda/beets,jmwatte/beets,kelvinhammond/beets,drm00/beets,jayme-github/beets,xsteadfastx/beets,kareemallen/beets,jackwilsdon/beets,MyTunesFreeMusic/privacy-policy,xsteadfastx/beets,jbaiter/beets,m-urban/beets,parapente/beets,lengtche/beets,randybias/beets,mosesfistos1/beetbox,tima/beets,andremiller/beets,PierreRust/beets,kelvinhammond/beets,artemutin/beets,marcuskrahl/beets,diego-plan9/beets,xsteadfastx/beets,gabrielaraujof/beets,arabenjamin/beets,Wen777/beets,imsparsh/beets,swt30/beets,arabenjamin/beets,diego-plan9/beets,asteven/beets,MyTunesFreeMusic/privacy-policy,lengtche/beets,dfc/beets,sampsyo/beets,bj-yinyan/beets,Kraymer/beets,moodboom/beets,shanemikel/beets,swt30/beets,madmouser1/beets,asteven/beets,Freso/beets,ttsda/beets,Kraymer/beets,randybias/beets,Andypsamp/CODjunit,parapente/beets,Wen777/beets,jcoady9/beets,swt30/beets,multikatt/beets,bj-yinyan/beets,kareemallen/beets,ruippeixotog/beets,Andypsamp/CODjunit,shamangeorge/beets,lengtche/beets,MyTunesFreeMusic/privacy-policy,lightwang1/beets,lightwang1/beets,LordSputnik/beets,artemutin/beets,Wen777/beets,untitaker/beets,multikatt/beets,Andypsamp/CODfinalJUNIT,marcuskrahl/beets,shamangeorge/beets,andremiller/beets,mried/beets,jackwilsdon/beets,dfc/beets,gabrielaraujof/beets,mosesfistos1/beetbox,SusannaMaria/beets,marcuskrahl/beets,asteven/beets,Andypsamp/CODjunit,moodboom/beets,madmouser1/beets,ibmibmibm/beets,dfc/beets,artemutin/beets,diego-plan9/beets,sadatay/beets,Andypsamp/CODjunit,sadatay/beets,Dishwishy/beets,ibmibmibm/beets,Dishwishy/beets,mathstuf/beets,Kraymer/beets,pkess/beets,imsparsh/beets,Andypsamp/CODjunit,krig/beets,jbaiter/beets,jmwatte/beets,multikatt/beets,jmwatte/beets,shanemikel/beets,kelvinhammond/beets,untitaker/beets
import os import re import sys from _common import unittest pkgpath = os.path.dirname(__file__) or '.' sys.path.append(pkgpath) os.chdir(pkgpath) + + # Make sure we use local version of beetsplug and not system namespaced version + # for tests + try: + del sys.modules["beetsplug"] + except KeyError: + pass def suite(): s = unittest.TestSuite() # Get the suite() of every module in this directory beginning with # "test_". for fname in os.listdir(pkgpath): match = re.match(r'(test_\S+)\.py$', fname) if match: modname = match.group(1) s.addTest(__import__(modname).suite()) return s if __name__ == '__main__': unittest.main(defaultTest='suite')
Fix python namespaces for test runs
## Code Before: import os import re import sys from _common import unittest pkgpath = os.path.dirname(__file__) or '.' sys.path.append(pkgpath) os.chdir(pkgpath) def suite(): s = unittest.TestSuite() # Get the suite() of every module in this directory beginning with # "test_". for fname in os.listdir(pkgpath): match = re.match(r'(test_\S+)\.py$', fname) if match: modname = match.group(1) s.addTest(__import__(modname).suite()) return s if __name__ == '__main__': unittest.main(defaultTest='suite') ## Instruction: Fix python namespaces for test runs ## Code After: import os import re import sys from _common import unittest pkgpath = os.path.dirname(__file__) or '.' sys.path.append(pkgpath) os.chdir(pkgpath) # Make sure we use local version of beetsplug and not system namespaced version # for tests try: del sys.modules["beetsplug"] except KeyError: pass def suite(): s = unittest.TestSuite() # Get the suite() of every module in this directory beginning with # "test_". for fname in os.listdir(pkgpath): match = re.match(r'(test_\S+)\.py$', fname) if match: modname = match.group(1) s.addTest(__import__(modname).suite()) return s if __name__ == '__main__': unittest.main(defaultTest='suite')
// ... existing code ... os.chdir(pkgpath) # Make sure we use local version of beetsplug and not system namespaced version # for tests try: del sys.modules["beetsplug"] except KeyError: pass // ... rest of the code ...
37fa40a9b5260f8090adaa8c15d3767c0867574f
python/fusion_engine_client/messages/__init__.py
python/fusion_engine_client/messages/__init__.py
from .core import * from . import ros message_type_to_class = { # Navigation solution messages. PoseMessage.MESSAGE_TYPE: PoseMessage, PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage, GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage, GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage, # Sensor measurement messages. IMUMeasurement.MESSAGE_TYPE: IMUMeasurement, # ROS messages. ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage, ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage, ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage, # Command and control messages. CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage, MessageRequest.MESSAGE_TYPE: MessageRequest, ResetRequest.MESSAGE_TYPE: ResetRequest, VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage, EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage, }
from .core import * from . import ros message_type_to_class = { # Navigation solution messages. PoseMessage.MESSAGE_TYPE: PoseMessage, PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage, GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage, GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage, # Sensor measurement messages. IMUMeasurement.MESSAGE_TYPE: IMUMeasurement, # ROS messages. ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage, ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage, ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage, # Command and control messages. CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage, MessageRequest.MESSAGE_TYPE: MessageRequest, ResetRequest.MESSAGE_TYPE: ResetRequest, VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage, EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage, } messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')]
Create a list of messages that contain system time.
Create a list of messages that contain system time.
Python
mit
PointOneNav/fusion-engine-client,PointOneNav/fusion-engine-client,PointOneNav/fusion-engine-client
from .core import * from . import ros message_type_to_class = { # Navigation solution messages. PoseMessage.MESSAGE_TYPE: PoseMessage, PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage, GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage, GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage, # Sensor measurement messages. IMUMeasurement.MESSAGE_TYPE: IMUMeasurement, # ROS messages. ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage, ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage, ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage, # Command and control messages. CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage, MessageRequest.MESSAGE_TYPE: MessageRequest, ResetRequest.MESSAGE_TYPE: ResetRequest, VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage, EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage, } + messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')] +
Create a list of messages that contain system time.
## Code Before: from .core import * from . import ros message_type_to_class = { # Navigation solution messages. PoseMessage.MESSAGE_TYPE: PoseMessage, PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage, GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage, GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage, # Sensor measurement messages. IMUMeasurement.MESSAGE_TYPE: IMUMeasurement, # ROS messages. ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage, ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage, ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage, # Command and control messages. CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage, MessageRequest.MESSAGE_TYPE: MessageRequest, ResetRequest.MESSAGE_TYPE: ResetRequest, VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage, EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage, } ## Instruction: Create a list of messages that contain system time. ## Code After: from .core import * from . import ros message_type_to_class = { # Navigation solution messages. PoseMessage.MESSAGE_TYPE: PoseMessage, PoseAuxMessage.MESSAGE_TYPE: PoseAuxMessage, GNSSInfoMessage.MESSAGE_TYPE: GNSSInfoMessage, GNSSSatelliteMessage.MESSAGE_TYPE: GNSSSatelliteMessage, # Sensor measurement messages. IMUMeasurement.MESSAGE_TYPE: IMUMeasurement, # ROS messages. ros.PoseMessage.MESSAGE_TYPE: ros.PoseMessage, ros.GPSFixMessage.MESSAGE_TYPE: ros.GPSFixMessage, ros.IMUMessage.MESSAGE_TYPE: ros.IMUMessage, # Command and control messages. CommandResponseMessage.MESSAGE_TYPE: CommandResponseMessage, MessageRequest.MESSAGE_TYPE: MessageRequest, ResetRequest.MESSAGE_TYPE: ResetRequest, VersionInfoMessage.MESSAGE_TYPE: VersionInfoMessage, EventNotificationMessage.MESSAGE_TYPE: EventNotificationMessage, } messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')]
... } messages_with_system_time = [t for t, c in message_type_to_class.items() if hasattr(c(), 'system_time_ns')] ...
c354d130cb542c2a5d57e519ce49175daa597e9c
froide/accesstoken/apps.py
froide/accesstoken/apps.py
from django.apps import AppConfig from django.utils.translation import ugettext_lazy as _ class AccessTokenConfig(AppConfig): name = 'froide.accesstoken' verbose_name = _('Secret Access Token') def ready(self): from froide.account import account_canceled account_canceled.connect(cancel_user) def cancel_user(sender, user=None, **kwargs): from .models import AccessToken if user is None: return AccessToken.objects.filter(user=user).delete()
import json from django.apps import AppConfig from django.utils.translation import ugettext_lazy as _ class AccessTokenConfig(AppConfig): name = 'froide.accesstoken' verbose_name = _('Secret Access Token') def ready(self): from froide.account import account_canceled from froide.account.export import registry account_canceled.connect(cancel_user) registry.register(export_user_data) def cancel_user(sender, user=None, **kwargs): from .models import AccessToken if user is None: return AccessToken.objects.filter(user=user).delete() def export_user_data(user): from .models import AccessToken access_tokens = ( AccessToken.objects.filter(user=user) ) if access_tokens: yield ('access_tokens.json', json.dumps([ { 'purpose': a.purpose, 'timestamp': a.timestamp.isoformat(), } for a in access_tokens]).encode('utf-8') )
Add user data export for accesstokens
Add user data export for accesstokens
Python
mit
fin/froide,fin/froide,fin/froide,fin/froide,stefanw/froide,stefanw/froide,stefanw/froide,stefanw/froide,stefanw/froide
+ import json + from django.apps import AppConfig from django.utils.translation import ugettext_lazy as _ class AccessTokenConfig(AppConfig): name = 'froide.accesstoken' verbose_name = _('Secret Access Token') def ready(self): from froide.account import account_canceled + from froide.account.export import registry account_canceled.connect(cancel_user) + registry.register(export_user_data) def cancel_user(sender, user=None, **kwargs): from .models import AccessToken if user is None: return AccessToken.objects.filter(user=user).delete() + + def export_user_data(user): + from .models import AccessToken + + access_tokens = ( + AccessToken.objects.filter(user=user) + ) + if access_tokens: + yield ('access_tokens.json', json.dumps([ + { + 'purpose': a.purpose, + 'timestamp': a.timestamp.isoformat(), + } + for a in access_tokens]).encode('utf-8') + ) +
Add user data export for accesstokens
## Code Before: from django.apps import AppConfig from django.utils.translation import ugettext_lazy as _ class AccessTokenConfig(AppConfig): name = 'froide.accesstoken' verbose_name = _('Secret Access Token') def ready(self): from froide.account import account_canceled account_canceled.connect(cancel_user) def cancel_user(sender, user=None, **kwargs): from .models import AccessToken if user is None: return AccessToken.objects.filter(user=user).delete() ## Instruction: Add user data export for accesstokens ## Code After: import json from django.apps import AppConfig from django.utils.translation import ugettext_lazy as _ class AccessTokenConfig(AppConfig): name = 'froide.accesstoken' verbose_name = _('Secret Access Token') def ready(self): from froide.account import account_canceled from froide.account.export import registry account_canceled.connect(cancel_user) registry.register(export_user_data) def cancel_user(sender, user=None, **kwargs): from .models import AccessToken if user is None: return AccessToken.objects.filter(user=user).delete() def export_user_data(user): from .models import AccessToken access_tokens = ( AccessToken.objects.filter(user=user) ) if access_tokens: yield ('access_tokens.json', json.dumps([ { 'purpose': a.purpose, 'timestamp': a.timestamp.isoformat(), } for a in access_tokens]).encode('utf-8') )
// ... existing code ... import json from django.apps import AppConfig // ... modified code ... from froide.account import account_canceled from froide.account.export import registry ... account_canceled.connect(cancel_user) registry.register(export_user_data) ... AccessToken.objects.filter(user=user).delete() def export_user_data(user): from .models import AccessToken access_tokens = ( AccessToken.objects.filter(user=user) ) if access_tokens: yield ('access_tokens.json', json.dumps([ { 'purpose': a.purpose, 'timestamp': a.timestamp.isoformat(), } for a in access_tokens]).encode('utf-8') ) // ... rest of the code ...
6160da958f4b8ecb1553c7bcca0b32bc1a5a1649
tests/conftest.py
tests/conftest.py
import os import shutil import tempfile import builtins import subprocess import pytest from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): shutil.rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) shutil.rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git']
import os import shutil import tempfile import builtins import subprocess import pytest import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise
Make sure .git test directory is removed on Windows
Make sure .git test directory is removed on Windows
Python
bsd-3-clause
scopatz/rever,ergs/rever
import os import shutil import tempfile import builtins import subprocess import pytest + import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): - shutil.rmtree(repo) + rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) - shutil.rmtree(repo) + rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] + + def rmtree(dirname): + """Remove a directory, even if it has read-only files (Windows). + Git creates read-only files that must be removed on teardown. See + https://stackoverflow.com/questions/2656322 for more info. + + Parameters + ---------- + dirname : str + Directory to be removed + """ + try: + shutil.rmtree(dirname) + except PermissionError: + if sys.platform == 'win32': + subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) + else: + raise +
Make sure .git test directory is removed on Windows
## Code Before: import os import shutil import tempfile import builtins import subprocess import pytest from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): shutil.rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) shutil.rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] ## Instruction: Make sure .git test directory is removed on Windows ## Code After: import os import shutil import tempfile import builtins import subprocess import pytest import sys from rever import environ @pytest.fixture def gitrepo(request): """A test fixutre that creates and destroys a git repo in a temporary directory. This will yield the path to the repo. """ cwd = os.getcwd() name = request.node.name repo = os.path.join(tempfile.gettempdir(), name) if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) os.chdir(repo) with open('README', 'w') as f: f.write('testing ' + name) subprocess.run(['git', 'add', '.']) subprocess.run(['git', 'commit', '-am', 'Initial readme']) with environ.context(): yield repo os.chdir(cwd) rmtree(repo) @pytest.fixture def gitecho(request): aliases = builtins.aliases aliases['git'] = lambda args: 'Would have run: ' + ' '.join(args) + '\n' yield None del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise
... import pytest import sys ... if os.path.exists(repo): rmtree(repo) subprocess.run(['git', 'init', repo]) ... os.chdir(cwd) rmtree(repo) ... del aliases['git'] def rmtree(dirname): """Remove a directory, even if it has read-only files (Windows). Git creates read-only files that must be removed on teardown. See https://stackoverflow.com/questions/2656322 for more info. Parameters ---------- dirname : str Directory to be removed """ try: shutil.rmtree(dirname) except PermissionError: if sys.platform == 'win32': subprocess.check_call(['del', '/F/S/Q', dirname], shell=True) else: raise ...
6e6c60613180bb3d7e2d019129e57d1a2c33286d
backend/backend/models.py
backend/backend/models.py
from django.db import models class Animal(models.Model): MALE = 'male' FEMALE = 'female' GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female')) father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father") mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother") name = models.CharField(max_length = 100) dob = models.IntegerField() gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE) active = models.BooleanField() own = models.BooleanField() class Meta: unique_together = ("name", "dob")
from django.db import models from django.core.validators import MaxValueValidator, MaxLengthValidator from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ from datetime import datetime def current_year(): return datetime.now().year class Animal(models.Model): MALE = 'male' FEMALE = 'female' GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female')) father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father") mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother") name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)]) dob = models.IntegerField(validators = [MaxValueValidator(current_year())]) gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE) active = models.BooleanField() own = models.BooleanField() class Meta: unique_together = ("name", "dob")
Add length validator to name. Add dob validator can't be higher than current year.
Add length validator to name. Add dob validator can't be higher than current year.
Python
apache-2.0
mmlado/animal_pairing,mmlado/animal_pairing
from django.db import models + from django.core.validators import MaxValueValidator, MaxLengthValidator + from django.core.exceptions import ValidationError + from django.utils.translation import gettext_lazy as _ + from datetime import datetime + + def current_year(): + return datetime.now().year class Animal(models.Model): MALE = 'male' FEMALE = 'female' GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female')) father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father") mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother") - name = models.CharField(max_length = 100) - dob = models.IntegerField() + name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)]) + dob = models.IntegerField(validators = [MaxValueValidator(current_year())]) gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE) active = models.BooleanField() own = models.BooleanField() class Meta: unique_together = ("name", "dob")
Add length validator to name. Add dob validator can't be higher than current year.
## Code Before: from django.db import models class Animal(models.Model): MALE = 'male' FEMALE = 'female' GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female')) father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father") mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother") name = models.CharField(max_length = 100) dob = models.IntegerField() gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE) active = models.BooleanField() own = models.BooleanField() class Meta: unique_together = ("name", "dob") ## Instruction: Add length validator to name. Add dob validator can't be higher than current year. ## Code After: from django.db import models from django.core.validators import MaxValueValidator, MaxLengthValidator from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ from datetime import datetime def current_year(): return datetime.now().year class Animal(models.Model): MALE = 'male' FEMALE = 'female' GENDER_CHOICES = ((MALE, 'Male'), (FEMALE, 'Female')) father = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_father") mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother") name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)]) dob = models.IntegerField(validators = [MaxValueValidator(current_year())]) gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE) active = models.BooleanField() own = models.BooleanField() class Meta: unique_together = ("name", "dob")
... from django.db import models from django.core.validators import MaxValueValidator, MaxLengthValidator from django.core.exceptions import ValidationError from django.utils.translation import gettext_lazy as _ from datetime import datetime def current_year(): return datetime.now().year ... mother = models.ForeignKey("self", null = True, on_delete = models.SET_NULL, related_name = "child_mother") name = models.CharField(max_length = 100, validators = [MaxLengthValidator(100)]) dob = models.IntegerField(validators = [MaxValueValidator(current_year())]) gender = models.CharField(max_length = 6, choices = GENDER_CHOICES, default = FEMALE) ...
90a724313902e3d95f1a37d9102af1544c9bc61d
segments/set_term_title.py
segments/set_term_title.py
def add_term_title_segment(): term = os.getenv('TERM') if not (('xterm' in term) or ('rxvt' in term)): return if powerline.args.shell == 'bash': set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]' elif powerline.args.shell == 'zsh': set_title = '\\e]0;%n@%m: %~\\a' else: import socket set_title = '\\e]0;%s@%s: %s\\a' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD')) powerline.append(set_title, None, None, '') add_term_title_segment()
def add_term_title_segment(): term = os.getenv('TERM') if not (('xterm' in term) or ('rxvt' in term)): return if powerline.args.shell == 'bash': set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]' elif powerline.args.shell == 'zsh': set_title = '\033]0;%n@%m: %~\007' else: import socket set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD')) powerline.append(set_title, None, None, '') add_term_title_segment()
Fix use of escape characters in "set terminal title" segment.
Fix use of escape characters in "set terminal title" segment. Escape characters were incorrect for non-BASH shells.
Python
mit
nicholascapo/powerline-shell,b-ryan/powerline-shell,junix/powerline-shell,wrgoldstein/powerline-shell,rbanffy/powerline-shell,b-ryan/powerline-shell,mart-e/powerline-shell,blieque/powerline-shell,paulhybryant/powerline-shell,tswsl1989/powerline-shell,torbjornvatn/powerline-shell,MartinWetterwald/powerline-shell,iKrishneel/powerline-shell,fellipecastro/powerline-shell,ceholden/powerline-shell,banga/powerline-shell,banga/powerline-shell,handsomecheung/powerline-shell,saghul/shline,strycore/powerline-shell,bitIO/powerline-shell,intfrr/powerline-shell,yc2prime/powerline-shell,mcdope/powerline-shell,milkbikis/powerline-shell,paulhybryant/powerline-shell,JulianVolodia/powerline-shell,dtrip/powerline-shell,paol/powerline-shell,Menci/powerline-shell,LeonardoGentile/powerline-shell
def add_term_title_segment(): term = os.getenv('TERM') if not (('xterm' in term) or ('rxvt' in term)): return if powerline.args.shell == 'bash': set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]' elif powerline.args.shell == 'zsh': - set_title = '\\e]0;%n@%m: %~\\a' + set_title = '\033]0;%n@%m: %~\007' else: import socket - set_title = '\\e]0;%s@%s: %s\\a' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD')) + set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD')) powerline.append(set_title, None, None, '') add_term_title_segment()
Fix use of escape characters in "set terminal title" segment.
## Code Before: def add_term_title_segment(): term = os.getenv('TERM') if not (('xterm' in term) or ('rxvt' in term)): return if powerline.args.shell == 'bash': set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]' elif powerline.args.shell == 'zsh': set_title = '\\e]0;%n@%m: %~\\a' else: import socket set_title = '\\e]0;%s@%s: %s\\a' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD')) powerline.append(set_title, None, None, '') add_term_title_segment() ## Instruction: Fix use of escape characters in "set terminal title" segment. ## Code After: def add_term_title_segment(): term = os.getenv('TERM') if not (('xterm' in term) or ('rxvt' in term)): return if powerline.args.shell == 'bash': set_title = '\\[\\e]0;\\u@\\h: \\w\\a\\]' elif powerline.args.shell == 'zsh': set_title = '\033]0;%n@%m: %~\007' else: import socket set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD')) powerline.append(set_title, None, None, '') add_term_title_segment()
# ... existing code ... elif powerline.args.shell == 'zsh': set_title = '\033]0;%n@%m: %~\007' else: # ... modified code ... import socket set_title = '\033]0;%s@%s: %s\007' % (os.getenv('USER'), socket.gethostname().split('.')[0], powerline.cwd or os.getenv('PWD')) # ... rest of the code ...
a8f125236308cbfc9bb2eb5b225a0ac92a3a95e4
ANN.py
ANN.py
from random import random class Neuron: def __init__(self, parents=[]): self.parents = parents self.weights = [random() for parent in parents] def get_output(self): return sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1 output = property(get_output) class NeuronNetwork: neurons = [] def __init__(self, inputs, outputs, rows, columns): self.neurons = [] for row in xrange(rows + 2): self.neurons.append([]) if row == 0: for input_ in xrange(inputs): self.neurons[row].append(Neuron(parents=[])) elif row == rows + 1: for output in xrange(outputs): self.neurons[row].append(Neuron(parents=self.neurons[row - 1])) else: for column in xrange(columns): self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
from random import random class Neuron: output = None def __init__(self, parents=[]): self.parents = parents self.weights = [random() for parent in parents] def calculate(self): self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1 class NeuronNetwork: neurons = [] def __init__(self, inputs, outputs, rows, columns): self.neurons = [] for row in xrange(rows + 2): self.neurons.append([]) if row == 0: for input_ in xrange(inputs): self.neurons[row].append(Neuron(parents=[])) elif row == rows + 1: for output in xrange(outputs): self.neurons[row].append(Neuron(parents=self.neurons[row - 1])) else: for column in xrange(columns): self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
Store output instead of calculating it each time
Store output instead of calculating it each time
Python
mit
tysonzero/py-ann
from random import random class Neuron: + output = None + def __init__(self, parents=[]): self.parents = parents self.weights = [random() for parent in parents] - def get_output(self): + def calculate(self): - return sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1 + self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1 - - output = property(get_output) class NeuronNetwork: neurons = [] def __init__(self, inputs, outputs, rows, columns): self.neurons = [] for row in xrange(rows + 2): self.neurons.append([]) if row == 0: for input_ in xrange(inputs): self.neurons[row].append(Neuron(parents=[])) elif row == rows + 1: for output in xrange(outputs): self.neurons[row].append(Neuron(parents=self.neurons[row - 1])) else: for column in xrange(columns): self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
Store output instead of calculating it each time
## Code Before: from random import random class Neuron: def __init__(self, parents=[]): self.parents = parents self.weights = [random() for parent in parents] def get_output(self): return sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1 output = property(get_output) class NeuronNetwork: neurons = [] def __init__(self, inputs, outputs, rows, columns): self.neurons = [] for row in xrange(rows + 2): self.neurons.append([]) if row == 0: for input_ in xrange(inputs): self.neurons[row].append(Neuron(parents=[])) elif row == rows + 1: for output in xrange(outputs): self.neurons[row].append(Neuron(parents=self.neurons[row - 1])) else: for column in xrange(columns): self.neurons[row].append(Neuron(parents=self.neurons[row - 1])) ## Instruction: Store output instead of calculating it each time ## Code After: from random import random class Neuron: output = None def __init__(self, parents=[]): self.parents = parents self.weights = [random() for parent in parents] def calculate(self): self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1 class NeuronNetwork: neurons = [] def __init__(self, inputs, outputs, rows, columns): self.neurons = [] for row in xrange(rows + 2): self.neurons.append([]) if row == 0: for input_ in xrange(inputs): self.neurons[row].append(Neuron(parents=[])) elif row == rows + 1: for output in xrange(outputs): self.neurons[row].append(Neuron(parents=self.neurons[row - 1])) else: for column in xrange(columns): self.neurons[row].append(Neuron(parents=self.neurons[row - 1]))
... class Neuron: output = None def __init__(self, parents=[]): ... def calculate(self): self.output = sum([parent.output * self.weights[i] for i, parent in enumerate(self.parents)]) >= 1 ...
222a87ef324f66baf8113020b41d336c459ab847
stdnum/fi/__init__.py
stdnum/fi/__init__.py
"""Collection of Finnish numbers.""" # provide vat as an alias from stdnum.fi import alv as vat from stdnum.fi import ytunnus as businessid
"""Collection of Finnish numbers.""" # provide vat as an alias from stdnum.fi import alv as vat from stdnum.fi import ytunnus as businessid from stdnum.fi import hetu as personalid
Add alias to hetu in for finnish personal id code
Add alias to hetu in for finnish personal id code
Python
lgpl-2.1
holvi/python-stdnum,holvi/python-stdnum,arthurdejong/python-stdnum,arthurdejong/python-stdnum,holvi/python-stdnum,arthurdejong/python-stdnum
"""Collection of Finnish numbers.""" # provide vat as an alias from stdnum.fi import alv as vat from stdnum.fi import ytunnus as businessid + from stdnum.fi import hetu as personalid
Add alias to hetu in for finnish personal id code
## Code Before: """Collection of Finnish numbers.""" # provide vat as an alias from stdnum.fi import alv as vat from stdnum.fi import ytunnus as businessid ## Instruction: Add alias to hetu in for finnish personal id code ## Code After: """Collection of Finnish numbers.""" # provide vat as an alias from stdnum.fi import alv as vat from stdnum.fi import ytunnus as businessid from stdnum.fi import hetu as personalid
# ... existing code ... from stdnum.fi import ytunnus as businessid from stdnum.fi import hetu as personalid # ... rest of the code ...
2f2cef54a98e2328a638d9bbdfd2e0312606d906
plugins/GCodeWriter/__init__.py
plugins/GCodeWriter/__init__.py
from . import GCodeWriter from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "type": "mesh_writer", "plugin": { "name": "GCode Writer", "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file") }, "mesh_writer": { "extension": "gcode", "description": catalog.i18nc("GCode Writer File Description", "GCode File") } } def register(app): return { "mesh_writer": GCodeWriter.GCodeWriter() }
from . import GCodeWriter from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "type": "mesh_writer", "plugin": { "name": "GCode Writer", "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file") }, "mesh_writer": { "extension": "gcode", "description": catalog.i18nc("GCode Writer File Description", "GCode File"), "mime_types": [ "text/x-gcode" ] } } def register(app): return { "mesh_writer": GCodeWriter.GCodeWriter() }
Add mime types to GCodeWriter plugin
Add mime types to GCodeWriter plugin
Python
agpl-3.0
Curahelper/Cura,senttech/Cura,fieldOfView/Cura,hmflash/Cura,lo0ol/Ultimaker-Cura,Curahelper/Cura,markwal/Cura,ad1217/Cura,senttech/Cura,ad1217/Cura,lo0ol/Ultimaker-Cura,totalretribution/Cura,hmflash/Cura,bq/Ultimaker-Cura,ynotstartups/Wanhao,fieldOfView/Cura,totalretribution/Cura,fxtentacle/Cura,fxtentacle/Cura,ynotstartups/Wanhao,markwal/Cura,bq/Ultimaker-Cura
from . import GCodeWriter from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "type": "mesh_writer", "plugin": { "name": "GCode Writer", "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file") }, "mesh_writer": { "extension": "gcode", - "description": catalog.i18nc("GCode Writer File Description", "GCode File") + "description": catalog.i18nc("GCode Writer File Description", "GCode File"), + "mime_types": [ + "text/x-gcode" + ] } } def register(app): return { "mesh_writer": GCodeWriter.GCodeWriter() }
Add mime types to GCodeWriter plugin
## Code Before: from . import GCodeWriter from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "type": "mesh_writer", "plugin": { "name": "GCode Writer", "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file") }, "mesh_writer": { "extension": "gcode", "description": catalog.i18nc("GCode Writer File Description", "GCode File") } } def register(app): return { "mesh_writer": GCodeWriter.GCodeWriter() } ## Instruction: Add mime types to GCodeWriter plugin ## Code After: from . import GCodeWriter from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "type": "mesh_writer", "plugin": { "name": "GCode Writer", "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("GCode Writer Plugin Description", "Writes GCode to a file") }, "mesh_writer": { "extension": "gcode", "description": catalog.i18nc("GCode Writer File Description", "GCode File"), "mime_types": [ "text/x-gcode" ] } } def register(app): return { "mesh_writer": GCodeWriter.GCodeWriter() }
# ... existing code ... "extension": "gcode", "description": catalog.i18nc("GCode Writer File Description", "GCode File"), "mime_types": [ "text/x-gcode" ] } # ... rest of the code ...
8f8b313a1b5118b6528e5152252128e075de0401
tests/test_terrain.py
tests/test_terrain.py
import unittest from randterrainpy import * class TerrainTesterPy(unittest.TestCase): def setUp(self): pass
import unittest from randterrainpy import * class TerrainTesterPy(unittest.TestCase): def setUp(self): self.ter1 = Terrain(1, 1) self.ter2 = Terrain(2, 4) self.ter3 = Terrain(1, 1) def test_getitem(self): self.assertEqual(self.ter1[0, 0], 0) self.assertEqual(self.ter2[1, 2], 0) def test_eq(self): self.assertEqual(self.ter1, self.ter3) self.assertNotEqual(self.ter1, self.ter2) def test_setitem(self): self.ter1[0, 0] = 1 self.assertEqual(self.ter1[0, 0], 1) self.ter2[1, 2] = 0.5 self.assertEqual(self.ter2[1, 2], 0.5) def test_add(self): self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2) self.assertEqual(self.ter1+self.ter3, self.ter1) if __name__ == "__main__": unittest.main()
Add tests for indexing, equality and addition for Terrain
Add tests for indexing, equality and addition for Terrain
Python
mit
jackromo/RandTerrainPy
import unittest from randterrainpy import * class TerrainTesterPy(unittest.TestCase): def setUp(self): - pass + self.ter1 = Terrain(1, 1) + self.ter2 = Terrain(2, 4) + self.ter3 = Terrain(1, 1) + def test_getitem(self): + self.assertEqual(self.ter1[0, 0], 0) + self.assertEqual(self.ter2[1, 2], 0) + + def test_eq(self): + self.assertEqual(self.ter1, self.ter3) + self.assertNotEqual(self.ter1, self.ter2) + + def test_setitem(self): + self.ter1[0, 0] = 1 + self.assertEqual(self.ter1[0, 0], 1) + self.ter2[1, 2] = 0.5 + self.assertEqual(self.ter2[1, 2], 0.5) + + def test_add(self): + self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2) + self.assertEqual(self.ter1+self.ter3, self.ter1) + + + if __name__ == "__main__": + unittest.main() +
Add tests for indexing, equality and addition for Terrain
## Code Before: import unittest from randterrainpy import * class TerrainTesterPy(unittest.TestCase): def setUp(self): pass ## Instruction: Add tests for indexing, equality and addition for Terrain ## Code After: import unittest from randterrainpy import * class TerrainTesterPy(unittest.TestCase): def setUp(self): self.ter1 = Terrain(1, 1) self.ter2 = Terrain(2, 4) self.ter3 = Terrain(1, 1) def test_getitem(self): self.assertEqual(self.ter1[0, 0], 0) self.assertEqual(self.ter2[1, 2], 0) def test_eq(self): self.assertEqual(self.ter1, self.ter3) self.assertNotEqual(self.ter1, self.ter2) def test_setitem(self): self.ter1[0, 0] = 1 self.assertEqual(self.ter1[0, 0], 1) self.ter2[1, 2] = 0.5 self.assertEqual(self.ter2[1, 2], 0.5) def test_add(self): self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2) self.assertEqual(self.ter1+self.ter3, self.ter1) if __name__ == "__main__": unittest.main()
# ... existing code ... def setUp(self): self.ter1 = Terrain(1, 1) self.ter2 = Terrain(2, 4) self.ter3 = Terrain(1, 1) def test_getitem(self): self.assertEqual(self.ter1[0, 0], 0) self.assertEqual(self.ter2[1, 2], 0) def test_eq(self): self.assertEqual(self.ter1, self.ter3) self.assertNotEqual(self.ter1, self.ter2) def test_setitem(self): self.ter1[0, 0] = 1 self.assertEqual(self.ter1[0, 0], 1) self.ter2[1, 2] = 0.5 self.assertEqual(self.ter2[1, 2], 0.5) def test_add(self): self.assertRaises(InvalidDimensionsError, self.ter1.__add__, self.ter2) self.assertEqual(self.ter1+self.ter3, self.ter1) if __name__ == "__main__": unittest.main() # ... rest of the code ...
f849961e75dc956d669813fddb5b13627b224e1e
pyang/plugins/name.py
pyang/plugins/name.py
import optparse from pyang import plugin def pyang_plugin_init(): plugin.register_plugin(NamePlugin()) class NamePlugin(plugin.PyangPlugin): def add_output_format(self, fmts): self.multiple_modules = True fmts['name'] = self def add_opts(self, optparser): optlist = [ optparse.make_option("--name-print-revision", dest="print_revision", action="store_true", help="Print the name and revision in name@revision format"), ] g = optparser.add_option_group("Name output specific options") g.add_options(optlist) def setup_fmt(self, ctx): ctx.implicit_errors = False def emit(self, ctx, modules, fd): emit_name(ctx, modules, fd) def emit_name(ctx, modules, fd): for module in modules: bstr = "" rstr = "" if ctx.opts.print_revision: r = module.search_one('revision') if r is not None: rstr = '@%s' % r.arg b = module.search_one('belongs-to') if b is not None: bstr = " (belongs-to %s)" % b.arg fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
import optparse from pyang import plugin def pyang_plugin_init(): plugin.register_plugin(NamePlugin()) class NamePlugin(plugin.PyangPlugin): def add_output_format(self, fmts): self.multiple_modules = True fmts['name'] = self def add_opts(self, optparser): optlist = [ optparse.make_option("--name-print-revision", dest="print_revision", action="store_true", help="Print the name and revision in name@revision format"), ] g = optparser.add_option_group("Name output specific options") g.add_options(optlist) def setup_fmt(self, ctx): ctx.implicit_errors = False def emit(self, ctx, modules, fd): emit_name(ctx, modules, fd) def emit_name(ctx, modules, fd): for module in modules: bstr = "" rstr = "" if ctx.opts.print_revision: rs = module.i_latest_revision if rs is None: r = module.search_one('revision') if r is not None: rs = r.arg if rs is not None: rstr = '@%s' % rs b = module.search_one('belongs-to') if b is not None: bstr = " (belongs-to %s)" % b.arg fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
Use i_latest_revision to ensure we get the latest revision.
Use i_latest_revision to ensure we get the latest revision.
Python
isc
mbj4668/pyang,mbj4668/pyang
import optparse from pyang import plugin def pyang_plugin_init(): plugin.register_plugin(NamePlugin()) class NamePlugin(plugin.PyangPlugin): def add_output_format(self, fmts): self.multiple_modules = True fmts['name'] = self def add_opts(self, optparser): optlist = [ optparse.make_option("--name-print-revision", dest="print_revision", action="store_true", help="Print the name and revision in name@revision format"), ] g = optparser.add_option_group("Name output specific options") g.add_options(optlist) def setup_fmt(self, ctx): ctx.implicit_errors = False def emit(self, ctx, modules, fd): emit_name(ctx, modules, fd) def emit_name(ctx, modules, fd): for module in modules: bstr = "" rstr = "" if ctx.opts.print_revision: + rs = module.i_latest_revision + if rs is None: - r = module.search_one('revision') + r = module.search_one('revision') + if r is not None: + rs = r.arg - if r is not None: + if rs is not None: - rstr = '@%s' % r.arg + rstr = '@%s' % rs b = module.search_one('belongs-to') if b is not None: bstr = " (belongs-to %s)" % b.arg fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
Use i_latest_revision to ensure we get the latest revision.
## Code Before: import optparse from pyang import plugin def pyang_plugin_init(): plugin.register_plugin(NamePlugin()) class NamePlugin(plugin.PyangPlugin): def add_output_format(self, fmts): self.multiple_modules = True fmts['name'] = self def add_opts(self, optparser): optlist = [ optparse.make_option("--name-print-revision", dest="print_revision", action="store_true", help="Print the name and revision in name@revision format"), ] g = optparser.add_option_group("Name output specific options") g.add_options(optlist) def setup_fmt(self, ctx): ctx.implicit_errors = False def emit(self, ctx, modules, fd): emit_name(ctx, modules, fd) def emit_name(ctx, modules, fd): for module in modules: bstr = "" rstr = "" if ctx.opts.print_revision: r = module.search_one('revision') if r is not None: rstr = '@%s' % r.arg b = module.search_one('belongs-to') if b is not None: bstr = " (belongs-to %s)" % b.arg fd.write("%s%s%s\n" % (module.arg, rstr, bstr)) ## Instruction: Use i_latest_revision to ensure we get the latest revision. ## Code After: import optparse from pyang import plugin def pyang_plugin_init(): plugin.register_plugin(NamePlugin()) class NamePlugin(plugin.PyangPlugin): def add_output_format(self, fmts): self.multiple_modules = True fmts['name'] = self def add_opts(self, optparser): optlist = [ optparse.make_option("--name-print-revision", dest="print_revision", action="store_true", help="Print the name and revision in name@revision format"), ] g = optparser.add_option_group("Name output specific options") g.add_options(optlist) def setup_fmt(self, ctx): ctx.implicit_errors = False def emit(self, ctx, modules, fd): emit_name(ctx, modules, fd) def emit_name(ctx, modules, fd): for module in modules: bstr = "" rstr = "" if ctx.opts.print_revision: rs = module.i_latest_revision if rs is None: r = module.search_one('revision') if r is not None: rs = r.arg if rs is not None: rstr = '@%s' % rs b = module.search_one('belongs-to') if b is not None: bstr = " (belongs-to %s)" % b.arg fd.write("%s%s%s\n" % (module.arg, rstr, bstr))
// ... existing code ... if ctx.opts.print_revision: rs = module.i_latest_revision if rs is None: r = module.search_one('revision') if r is not None: rs = r.arg if rs is not None: rstr = '@%s' % rs b = module.search_one('belongs-to') // ... rest of the code ...
695b3f628b56cd1dbe050f07692559ff3685290c
templatefinder/__init__.py
templatefinder/__init__.py
from __future__ import absolute_import from .utils import * VERSION = (0, 5,)
from __future__ import absolute_import # this is required for setup.py to work try: from .utils import * except ImportError: pass VERSION = (0, 5, 1,)
Fix setup.py for installs without Django
Fix setup.py for installs without Django
Python
bsd-2-clause
TyMaszWeb/django-template-finder
from __future__ import absolute_import + # this is required for setup.py to work + try: - from .utils import * + from .utils import * + except ImportError: + pass - VERSION = (0, 5,) + VERSION = (0, 5, 1,)
Fix setup.py for installs without Django
## Code Before: from __future__ import absolute_import from .utils import * VERSION = (0, 5,) ## Instruction: Fix setup.py for installs without Django ## Code After: from __future__ import absolute_import # this is required for setup.py to work try: from .utils import * except ImportError: pass VERSION = (0, 5, 1,)
// ... existing code ... # this is required for setup.py to work try: from .utils import * except ImportError: pass // ... modified code ... VERSION = (0, 5, 1,) // ... rest of the code ...
4de03c57bf4f4995eb8c8859e0a40b7c5fc9942b
desktop/libs/libzookeeper/src/libzookeeper/models.py
desktop/libs/libzookeeper/src/libzookeeper/models.py
from kazoo.client import KazooClient from libzookeeper.conf import PRINCIPAL_NAME def get_children_data(ensemble, namespace, read_only=True): zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=PRINCIPAL_NAME.get()) zk.start() children_data = [] children = zk.get_children(namespace) for node in children: data, stat = zk.get("%s/%s" % (namespace, node)) children_data.append(data) zk.stop() return children_data
from kazoo.client import KazooClient from hadoop import cluster from desktop.lib.exceptions_renderable import PopupException from libzookeeper.conf import PRINCIPAL_NAME def get_children_data(ensemble, namespace, read_only=True): hdfs = cluster.get_hdfs() if hdfs is None: raise PopupException(_('No [hdfs] configured in hue.ini.')) if hdfs.security_enabled: sasl_server_principal = PRINCIPAL_NAME.get() else: sasl_server_principal = None zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal) zk.start() children_data = [] children = zk.get_children(namespace) for node in children: data, stat = zk.get("%s/%s" % (namespace, node)) children_data.append(data) zk.stop() return children_data
Enable Kerberos automatically based on HDFS security
[libzookeeper] Enable Kerberos automatically based on HDFS security We don't need another property that way and Kerberos is a all or nothing setup. Even if HDFS is not used in Hue, the default hue.ini has security set to false.
Python
apache-2.0
pratikmallya/hue,jjmleiro/hue,lumig242/Hue-Integration-with-CDAP,cloudera/hue,pratikmallya/hue,xiangel/hue,Peddle/hue,x303597316/hue,cloudera/hue,rahul67/hue,kawamon/hue,yongshengwang/hue,MobinRanjbar/hue,x303597316/hue,xq262144/hue,jayceyxc/hue,mapr/hue,yongshengwang/hue,pratikmallya/hue,sanjeevtripurari/hue,lumig242/Hue-Integration-with-CDAP,jayceyxc/hue,cloudera/hue,jounex/hue,Peddle/hue,ChenJunor/hue,jayceyxc/hue,kawamon/hue,Peddle/hue,vmax-feihu/hue,MobinRanjbar/hue,vmax-feihu/hue,ahmed-mahran/hue,ahmed-mahran/hue,kawamon/hue,kawamon/hue,GitHublong/hue,hdinsight/hue,Peddle/hue,kawamon/hue,xiangel/hue,yoer/hue,pratikmallya/hue,yongshengwang/hue,kawamon/hue,cloudera/hue,jounex/hue,cloudera/hue,rahul67/hue,javachengwc/hue,azureplus/hue,kawamon/hue,vmax-feihu/hue,fangxingli/hue,hdinsight/hue,jounex/hue,cloudera/hue,sanjeevtripurari/hue,rahul67/hue,lumig242/Hue-Integration-with-CDAP,jjmleiro/hue,kawamon/hue,cloudera/hue,javachengwc/hue,cloudera/hue,yongshengwang/hue,jjmleiro/hue,mapr/hue,yoer/hue,yongshengwang/hue,todaychi/hue,kawamon/hue,GitHublong/hue,xq262144/hue,jounex/hue,kawamon/hue,ChenJunor/hue,cloudera/hue,x303597316/hue,jayceyxc/hue,todaychi/hue,jounex/hue,GitHublong/hue,hdinsight/hue,yoer/hue,hdinsight/hue,fangxingli/hue,kawamon/hue,lumig242/Hue-Integration-with-CDAP,x303597316/hue,ahmed-mahran/hue,azureplus/hue,xiangel/hue,todaychi/hue,jjmleiro/hue,jounex/hue,cloudera/hue,fangxingli/hue,vmax-feihu/hue,yongshengwang/hue,hdinsight/hue,kawamon/hue,ChenJunor/hue,todaychi/hue,lumig242/Hue-Integration-with-CDAP,rahul67/hue,javachengwc/hue,todaychi/hue,todaychi/hue,rahul67/hue,kawamon/hue,GitHublong/hue,yoer/hue,fangxingli/hue,javachengwc/hue,todaychi/hue,cloudera/hue,cloudera/hue,ChenJunor/hue,jayceyxc/hue,jayceyxc/hue,Peddle/hue,sanjeevtripurari/hue,lumig242/Hue-Integration-with-CDAP,xq262144/hue,pratikmallya/hue,GitHublong/hue,x303597316/hue,jayceyxc/hue,ahmed-mahran/hue,MobinRanjbar/hue,MobinRanjbar/hue,vmax-feihu/hue,fangxingli/hue,mapr/hue,xiangel/hue,kawamon/hue,lumig242/Hue-Integration-with-CDAP,mapr/hue,mapr/hue,fangxingli/hue,ChenJunor/hue,jayceyxc/hue,jjmleiro/hue,GitHublong/hue,todaychi/hue,pratikmallya/hue,cloudera/hue,ahmed-mahran/hue,cloudera/hue,xq262144/hue,rahul67/hue,jayceyxc/hue,azureplus/hue,jjmleiro/hue,MobinRanjbar/hue,javachengwc/hue,azureplus/hue,mapr/hue,jounex/hue,xiangel/hue,vmax-feihu/hue,hdinsight/hue,sanjeevtripurari/hue,ahmed-mahran/hue,yoer/hue,ahmed-mahran/hue,pratikmallya/hue,hdinsight/hue,cloudera/hue,ChenJunor/hue,xiangel/hue,kawamon/hue,xq262144/hue,yongshengwang/hue,jounex/hue,hdinsight/hue,MobinRanjbar/hue,rahul67/hue,Peddle/hue,kawamon/hue,xq262144/hue,azureplus/hue,lumig242/Hue-Integration-with-CDAP,vmax-feihu/hue,Peddle/hue,xiangel/hue,sanjeevtripurari/hue,sanjeevtripurari/hue,x303597316/hue,MobinRanjbar/hue,javachengwc/hue,xq262144/hue,fangxingli/hue,x303597316/hue,mapr/hue,xq262144/hue,pratikmallya/hue,sanjeevtripurari/hue,javachengwc/hue,sanjeevtripurari/hue,yoer/hue,rahul67/hue,kawamon/hue,azureplus/hue,jjmleiro/hue,Peddle/hue,ChenJunor/hue,MobinRanjbar/hue,cloudera/hue,yongshengwang/hue,azureplus/hue,yoer/hue,ChenJunor/hue,cloudera/hue,vmax-feihu/hue,cloudera/hue,azureplus/hue,yoer/hue,jjmleiro/hue,javachengwc/hue,GitHublong/hue,Peddle/hue,ahmed-mahran/hue,todaychi/hue,x303597316/hue,jjmleiro/hue,xiangel/hue,lumig242/Hue-Integration-with-CDAP,kawamon/hue,fangxingli/hue,GitHublong/hue,xq262144/hue
from kazoo.client import KazooClient + + from hadoop import cluster + from desktop.lib.exceptions_renderable import PopupException + from libzookeeper.conf import PRINCIPAL_NAME def get_children_data(ensemble, namespace, read_only=True): + hdfs = cluster.get_hdfs() + if hdfs is None: + raise PopupException(_('No [hdfs] configured in hue.ini.')) + + if hdfs.security_enabled: + sasl_server_principal = PRINCIPAL_NAME.get() + else: + sasl_server_principal = None + - zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=PRINCIPAL_NAME.get()) + zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal) zk.start() children_data = [] children = zk.get_children(namespace) for node in children: data, stat = zk.get("%s/%s" % (namespace, node)) children_data.append(data) zk.stop() - + return children_data
Enable Kerberos automatically based on HDFS security
## Code Before: from kazoo.client import KazooClient from libzookeeper.conf import PRINCIPAL_NAME def get_children_data(ensemble, namespace, read_only=True): zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=PRINCIPAL_NAME.get()) zk.start() children_data = [] children = zk.get_children(namespace) for node in children: data, stat = zk.get("%s/%s" % (namespace, node)) children_data.append(data) zk.stop() return children_data ## Instruction: Enable Kerberos automatically based on HDFS security ## Code After: from kazoo.client import KazooClient from hadoop import cluster from desktop.lib.exceptions_renderable import PopupException from libzookeeper.conf import PRINCIPAL_NAME def get_children_data(ensemble, namespace, read_only=True): hdfs = cluster.get_hdfs() if hdfs is None: raise PopupException(_('No [hdfs] configured in hue.ini.')) if hdfs.security_enabled: sasl_server_principal = PRINCIPAL_NAME.get() else: sasl_server_principal = None zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal) zk.start() children_data = [] children = zk.get_children(namespace) for node in children: data, stat = zk.get("%s/%s" % (namespace, node)) children_data.append(data) zk.stop() return children_data
// ... existing code ... from kazoo.client import KazooClient from hadoop import cluster from desktop.lib.exceptions_renderable import PopupException from libzookeeper.conf import PRINCIPAL_NAME // ... modified code ... def get_children_data(ensemble, namespace, read_only=True): hdfs = cluster.get_hdfs() if hdfs is None: raise PopupException(_('No [hdfs] configured in hue.ini.')) if hdfs.security_enabled: sasl_server_principal = PRINCIPAL_NAME.get() else: sasl_server_principal = None zk = KazooClient(hosts=ensemble, read_only=read_only, sasl_server_principal=sasl_server_principal) ... zk.stop() return children_data // ... rest of the code ...
382b8df7a25732ee8384c02d776472a93c18a0ea
vcspull/__about__.py
vcspull/__about__.py
__title__ = 'vcspull' __package_name__ = 'vcspull' __description__ = 'synchronize your repos' __version__ = '1.2.0' __author__ = 'Tony Narlock' __github__ = 'https://github.com/vcs-python/vcspull' __pypi__ = 'https://pypi.org/project/vcspull/' __email__ = '[email protected]' __license__ = 'MIT' __copyright__ = 'Copyright 2013-2018 Tony Narlock'
__title__ = 'vcspull' __package_name__ = 'vcspull' __description__ = 'synchronize your repos' __version__ = '1.2.0' __author__ = 'Tony Narlock' __github__ = 'https://github.com/vcs-python/vcspull' __docs__ = 'https://vcspull.git-pull.com' __tracker__ = 'https://github.com/vcs-python/vcspull/issues' __pypi__ = 'https://pypi.org/project/vcspull/' __email__ = '[email protected]' __license__ = 'MIT' __copyright__ = 'Copyright 2013-Tony Narlock'
Add docs / tracker to metadata
Add docs / tracker to metadata
Python
mit
tony/vcspull,tony/vcspull
__title__ = 'vcspull' __package_name__ = 'vcspull' __description__ = 'synchronize your repos' __version__ = '1.2.0' __author__ = 'Tony Narlock' __github__ = 'https://github.com/vcs-python/vcspull' + __docs__ = 'https://vcspull.git-pull.com' + __tracker__ = 'https://github.com/vcs-python/vcspull/issues' __pypi__ = 'https://pypi.org/project/vcspull/' __email__ = '[email protected]' __license__ = 'MIT' - __copyright__ = 'Copyright 2013-2018 Tony Narlock' + __copyright__ = 'Copyright 2013-Tony Narlock'
Add docs / tracker to metadata
## Code Before: __title__ = 'vcspull' __package_name__ = 'vcspull' __description__ = 'synchronize your repos' __version__ = '1.2.0' __author__ = 'Tony Narlock' __github__ = 'https://github.com/vcs-python/vcspull' __pypi__ = 'https://pypi.org/project/vcspull/' __email__ = '[email protected]' __license__ = 'MIT' __copyright__ = 'Copyright 2013-2018 Tony Narlock' ## Instruction: Add docs / tracker to metadata ## Code After: __title__ = 'vcspull' __package_name__ = 'vcspull' __description__ = 'synchronize your repos' __version__ = '1.2.0' __author__ = 'Tony Narlock' __github__ = 'https://github.com/vcs-python/vcspull' __docs__ = 'https://vcspull.git-pull.com' __tracker__ = 'https://github.com/vcs-python/vcspull/issues' __pypi__ = 'https://pypi.org/project/vcspull/' __email__ = '[email protected]' __license__ = 'MIT' __copyright__ = 'Copyright 2013-Tony Narlock'
# ... existing code ... __github__ = 'https://github.com/vcs-python/vcspull' __docs__ = 'https://vcspull.git-pull.com' __tracker__ = 'https://github.com/vcs-python/vcspull/issues' __pypi__ = 'https://pypi.org/project/vcspull/' # ... modified code ... __license__ = 'MIT' __copyright__ = 'Copyright 2013-Tony Narlock' # ... rest of the code ...
27ab3ad3d1ce869baec85264b840da49ff43f82f
scripts/sync_exceeded_traffic_limits.py
scripts/sync_exceeded_traffic_limits.py
import os from flask import _request_ctx_stack, g, request from sqlalchemy import create_engine from sqlalchemy.orm import scoped_session, sessionmaker from pycroft.model import session from pycroft.model.session import set_scoped_session from scripts.schema import AlembicHelper, SchemaStrategist from pycroft.lib import traffic def main(): try: connection_string = os.environ['PYCROFT_DB_URI'] except KeyError: raise RuntimeError("Environment variable PYCROFT_DB_URI must be " "set to an SQLAlchemy connection string.") engine = create_engine(connection_string) connection = engine.connect() state = AlembicHelper(connection) strategy = SchemaStrategist(state).determine_schema_strategy() strategy() engine = create_engine(connection_string) set_scoped_session(scoped_session(sessionmaker(bind=engine), scopefunc=lambda: _request_ctx_stack.top)) print("Starting synchronization of exceeded traffic limits.") traffic.sync_exceeded_traffic_limits() session.session.commit() print("Finished synchronization.") if __name__ == "__main__": main()
import os from flask import _request_ctx_stack, g, request from sqlalchemy import create_engine from sqlalchemy.orm import scoped_session, sessionmaker from pycroft.model import session from pycroft.model.session import set_scoped_session from scripts.schema import AlembicHelper, SchemaStrategist from pycroft.lib import traffic def main(): try: connection_string = os.environ['PYCROFT_DB_URI'] except KeyError: raise RuntimeError("Environment variable PYCROFT_DB_URI must be " "set to an SQLAlchemy connection string.") engine = create_engine(connection_string) connection = engine.connect() state = AlembicHelper(connection) strategist = SchemaStrategist(state) is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version if not is_up_to_date: print("Schema is not up to date!") return set_scoped_session(scoped_session(sessionmaker(bind=engine), scopefunc=lambda: _request_ctx_stack.top)) print("Starting synchronization of exceeded traffic limits.") traffic.sync_exceeded_traffic_limits() session.session.commit() print("Finished synchronization.") if __name__ == "__main__": main()
Add schema version check to sync script
Add schema version check to sync script
Python
apache-2.0
agdsn/pycroft,agdsn/pycroft,agdsn/pycroft,lukasjuhrich/pycroft,agdsn/pycroft,lukasjuhrich/pycroft,lukasjuhrich/pycroft,lukasjuhrich/pycroft,agdsn/pycroft
import os from flask import _request_ctx_stack, g, request from sqlalchemy import create_engine from sqlalchemy.orm import scoped_session, sessionmaker from pycroft.model import session from pycroft.model.session import set_scoped_session from scripts.schema import AlembicHelper, SchemaStrategist from pycroft.lib import traffic def main(): try: connection_string = os.environ['PYCROFT_DB_URI'] except KeyError: raise RuntimeError("Environment variable PYCROFT_DB_URI must be " "set to an SQLAlchemy connection string.") engine = create_engine(connection_string) connection = engine.connect() state = AlembicHelper(connection) - strategy = SchemaStrategist(state).determine_schema_strategy() - strategy() + strategist = SchemaStrategist(state) + is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version + if not is_up_to_date: + print("Schema is not up to date!") + return - engine = create_engine(connection_string) set_scoped_session(scoped_session(sessionmaker(bind=engine), scopefunc=lambda: _request_ctx_stack.top)) print("Starting synchronization of exceeded traffic limits.") traffic.sync_exceeded_traffic_limits() session.session.commit() print("Finished synchronization.") if __name__ == "__main__": main()
Add schema version check to sync script
## Code Before: import os from flask import _request_ctx_stack, g, request from sqlalchemy import create_engine from sqlalchemy.orm import scoped_session, sessionmaker from pycroft.model import session from pycroft.model.session import set_scoped_session from scripts.schema import AlembicHelper, SchemaStrategist from pycroft.lib import traffic def main(): try: connection_string = os.environ['PYCROFT_DB_URI'] except KeyError: raise RuntimeError("Environment variable PYCROFT_DB_URI must be " "set to an SQLAlchemy connection string.") engine = create_engine(connection_string) connection = engine.connect() state = AlembicHelper(connection) strategy = SchemaStrategist(state).determine_schema_strategy() strategy() engine = create_engine(connection_string) set_scoped_session(scoped_session(sessionmaker(bind=engine), scopefunc=lambda: _request_ctx_stack.top)) print("Starting synchronization of exceeded traffic limits.") traffic.sync_exceeded_traffic_limits() session.session.commit() print("Finished synchronization.") if __name__ == "__main__": main() ## Instruction: Add schema version check to sync script ## Code After: import os from flask import _request_ctx_stack, g, request from sqlalchemy import create_engine from sqlalchemy.orm import scoped_session, sessionmaker from pycroft.model import session from pycroft.model.session import set_scoped_session from scripts.schema import AlembicHelper, SchemaStrategist from pycroft.lib import traffic def main(): try: connection_string = os.environ['PYCROFT_DB_URI'] except KeyError: raise RuntimeError("Environment variable PYCROFT_DB_URI must be " "set to an SQLAlchemy connection string.") engine = create_engine(connection_string) connection = engine.connect() state = AlembicHelper(connection) strategist = SchemaStrategist(state) is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version if not is_up_to_date: print("Schema is not up to date!") return set_scoped_session(scoped_session(sessionmaker(bind=engine), scopefunc=lambda: _request_ctx_stack.top)) print("Starting synchronization of exceeded traffic limits.") traffic.sync_exceeded_traffic_limits() session.session.commit() print("Finished synchronization.") if __name__ == "__main__": main()
// ... existing code ... state = AlembicHelper(connection) strategist = SchemaStrategist(state) is_up_to_date = strategist.helper.running_version == strategist.helper.desired_version if not is_up_to_date: print("Schema is not up to date!") return set_scoped_session(scoped_session(sessionmaker(bind=engine), // ... rest of the code ...
4583c9949143e58bf400fc86e27d634aa382f605
tests/test_expanded.py
tests/test_expanded.py
from mycli.packages.expanded import expanded_table def test_expanded_table_renders(): input = [("hello", 123), ("world", 456)] expected = """-[ RECORD 0 ] name | hello age | 123 -[ RECORD 1 ] name | world age | 456 """ assert expected == expanded_table(input, ["name", "age"])
from mycli.packages.expanded import expanded_table def test_expanded_table_renders(): input = [("hello", 123), ("world", 456)] expected = """***************************[ 1. row ]*************************** name | hello age | 123 ***************************[ 2. row ]*************************** name | world age | 456 """ assert expected == expanded_table(input, ["name", "age"])
Update expanded tests to match mysql style.
Update expanded tests to match mysql style.
Python
bsd-3-clause
oguzy/mycli,chenpingzhao/mycli,ZuoGuocai/mycli,evook/mycli,jinstrive/mycli,j-bennet/mycli,danieljwest/mycli,suzukaze/mycli,thanatoskira/mycli,chenpingzhao/mycli,j-bennet/mycli,brewneaux/mycli,webwlsong/mycli,MnO2/rediscli,brewneaux/mycli,shoma/mycli,mdsrosa/mycli,oguzy/mycli,danieljwest/mycli,jinstrive/mycli,thanatoskira/mycli,ZuoGuocai/mycli,webwlsong/mycli,evook/mycli,suzukaze/mycli,D-e-e-m-o/mycli,MnO2/rediscli,shoma/mycli,martijnengler/mycli,mdsrosa/mycli,martijnengler/mycli,D-e-e-m-o/mycli
from mycli.packages.expanded import expanded_table def test_expanded_table_renders(): input = [("hello", 123), ("world", 456)] - expected = """-[ RECORD 0 ] + expected = """***************************[ 1. row ]*************************** name | hello age | 123 - -[ RECORD 1 ] + ***************************[ 2. row ]*************************** name | world age | 456 """ assert expected == expanded_table(input, ["name", "age"])
Update expanded tests to match mysql style.
## Code Before: from mycli.packages.expanded import expanded_table def test_expanded_table_renders(): input = [("hello", 123), ("world", 456)] expected = """-[ RECORD 0 ] name | hello age | 123 -[ RECORD 1 ] name | world age | 456 """ assert expected == expanded_table(input, ["name", "age"]) ## Instruction: Update expanded tests to match mysql style. ## Code After: from mycli.packages.expanded import expanded_table def test_expanded_table_renders(): input = [("hello", 123), ("world", 456)] expected = """***************************[ 1. row ]*************************** name | hello age | 123 ***************************[ 2. row ]*************************** name | world age | 456 """ assert expected == expanded_table(input, ["name", "age"])
// ... existing code ... expected = """***************************[ 1. row ]*************************** name | hello // ... modified code ... age | 123 ***************************[ 2. row ]*************************** name | world // ... rest of the code ...
6a379b806dd1992ad3dd2b728878ed35e8d0ea3c
cdf/utils.py
cdf/utils.py
def get_major_dot_minor_version(version): """ Convert full VERSION Django tuple to a dotted string containing MAJOR.MINOR. For example, (1, 9, 3, 'final', 0) will result in '1.9' """ return '.'.join(version.split('.')[:2])
def get_major_dot_minor_version(version): """ Convert full VERSION Django tuple to a dotted string containing MAJOR.MINOR. For example, (1, 9, 3, 'final', 0) will result in '1.9' """ return '.'.join([str(v) for v in version[:2]])
Fix getting major.minor django version
Fix getting major.minor django version
Python
mit
ana-balica/classy-django-forms,ana-balica/classy-django-forms,ana-balica/classy-django-forms
def get_major_dot_minor_version(version): """ Convert full VERSION Django tuple to a dotted string containing MAJOR.MINOR. For example, (1, 9, 3, 'final', 0) will result in '1.9' """ - return '.'.join(version.split('.')[:2]) + return '.'.join([str(v) for v in version[:2]])
Fix getting major.minor django version
## Code Before: def get_major_dot_minor_version(version): """ Convert full VERSION Django tuple to a dotted string containing MAJOR.MINOR. For example, (1, 9, 3, 'final', 0) will result in '1.9' """ return '.'.join(version.split('.')[:2]) ## Instruction: Fix getting major.minor django version ## Code After: def get_major_dot_minor_version(version): """ Convert full VERSION Django tuple to a dotted string containing MAJOR.MINOR. For example, (1, 9, 3, 'final', 0) will result in '1.9' """ return '.'.join([str(v) for v in version[:2]])
... """ return '.'.join([str(v) for v in version[:2]]) ...
07058595e43290524d28b53b5919fb76f16c618b
test/test_validators.py
test/test_validators.py
from unittest import TestCase from win_unc import validators as V class TestIsValidDriveLetter(TestCase): def test_valid(self): self.assertTrue(V.is_valid_drive_letter('A')) self.assertTrue(V.is_valid_drive_letter('Z')) self.assertTrue(V.is_valid_drive_letter('a')) self.assertTrue(V.is_valid_drive_letter('z')) def test_invalid(self): self.assertFalse(V.is_valid_drive_letter('')) self.assertFalse(V.is_valid_drive_letter(':')) self.assertFalse(V.is_valid_drive_letter('aa')) self.assertFalse(V.is_valid_drive_letter('a:'))
from unittest import TestCase from win_unc import validators as V class TestIsValidDriveLetter(TestCase): def test_valid(self): self.assertTrue(V.is_valid_drive_letter('A')) self.assertTrue(V.is_valid_drive_letter('Z')) self.assertTrue(V.is_valid_drive_letter('a')) self.assertTrue(V.is_valid_drive_letter('z')) def test_invalid(self): self.assertFalse(V.is_valid_drive_letter('')) self.assertFalse(V.is_valid_drive_letter(':')) self.assertFalse(V.is_valid_drive_letter('aa')) self.assertFalse(V.is_valid_drive_letter('a:')) class TestIsValidUncPath(TestCase): def test_valid(self): self.assertTrue(V.is_valid_unc_path(r'\\a')) self.assertTrue(V.is_valid_unc_path(r'\\a\b\c')) self.assertTrue(V.is_valid_unc_path(r'\\ABC\\')) def test_invalid(self): self.assertFalse(V.is_valid_unc_path('')) self.assertFalse(V.is_valid_unc_path(r'\\')) self.assertFalse(V.is_valid_unc_path(r'\\\a')) self.assertFalse(V.is_valid_unc_path(r'C:\path')) self.assertFalse(V.is_valid_unc_path(r'\\<a>'))
Add tests for UNC path validator
Add tests for UNC path validator
Python
mit
CovenantEyes/py_win_unc,nithinphilips/py_win_unc
from unittest import TestCase from win_unc import validators as V class TestIsValidDriveLetter(TestCase): def test_valid(self): self.assertTrue(V.is_valid_drive_letter('A')) self.assertTrue(V.is_valid_drive_letter('Z')) self.assertTrue(V.is_valid_drive_letter('a')) self.assertTrue(V.is_valid_drive_letter('z')) def test_invalid(self): self.assertFalse(V.is_valid_drive_letter('')) self.assertFalse(V.is_valid_drive_letter(':')) self.assertFalse(V.is_valid_drive_letter('aa')) self.assertFalse(V.is_valid_drive_letter('a:')) + + class TestIsValidUncPath(TestCase): + def test_valid(self): + self.assertTrue(V.is_valid_unc_path(r'\\a')) + self.assertTrue(V.is_valid_unc_path(r'\\a\b\c')) + self.assertTrue(V.is_valid_unc_path(r'\\ABC\\')) + + def test_invalid(self): + self.assertFalse(V.is_valid_unc_path('')) + self.assertFalse(V.is_valid_unc_path(r'\\')) + self.assertFalse(V.is_valid_unc_path(r'\\\a')) + self.assertFalse(V.is_valid_unc_path(r'C:\path')) + self.assertFalse(V.is_valid_unc_path(r'\\<a>')) +
Add tests for UNC path validator
## Code Before: from unittest import TestCase from win_unc import validators as V class TestIsValidDriveLetter(TestCase): def test_valid(self): self.assertTrue(V.is_valid_drive_letter('A')) self.assertTrue(V.is_valid_drive_letter('Z')) self.assertTrue(V.is_valid_drive_letter('a')) self.assertTrue(V.is_valid_drive_letter('z')) def test_invalid(self): self.assertFalse(V.is_valid_drive_letter('')) self.assertFalse(V.is_valid_drive_letter(':')) self.assertFalse(V.is_valid_drive_letter('aa')) self.assertFalse(V.is_valid_drive_letter('a:')) ## Instruction: Add tests for UNC path validator ## Code After: from unittest import TestCase from win_unc import validators as V class TestIsValidDriveLetter(TestCase): def test_valid(self): self.assertTrue(V.is_valid_drive_letter('A')) self.assertTrue(V.is_valid_drive_letter('Z')) self.assertTrue(V.is_valid_drive_letter('a')) self.assertTrue(V.is_valid_drive_letter('z')) def test_invalid(self): self.assertFalse(V.is_valid_drive_letter('')) self.assertFalse(V.is_valid_drive_letter(':')) self.assertFalse(V.is_valid_drive_letter('aa')) self.assertFalse(V.is_valid_drive_letter('a:')) class TestIsValidUncPath(TestCase): def test_valid(self): self.assertTrue(V.is_valid_unc_path(r'\\a')) self.assertTrue(V.is_valid_unc_path(r'\\a\b\c')) self.assertTrue(V.is_valid_unc_path(r'\\ABC\\')) def test_invalid(self): self.assertFalse(V.is_valid_unc_path('')) self.assertFalse(V.is_valid_unc_path(r'\\')) self.assertFalse(V.is_valid_unc_path(r'\\\a')) self.assertFalse(V.is_valid_unc_path(r'C:\path')) self.assertFalse(V.is_valid_unc_path(r'\\<a>'))
// ... existing code ... self.assertFalse(V.is_valid_drive_letter('a:')) class TestIsValidUncPath(TestCase): def test_valid(self): self.assertTrue(V.is_valid_unc_path(r'\\a')) self.assertTrue(V.is_valid_unc_path(r'\\a\b\c')) self.assertTrue(V.is_valid_unc_path(r'\\ABC\\')) def test_invalid(self): self.assertFalse(V.is_valid_unc_path('')) self.assertFalse(V.is_valid_unc_path(r'\\')) self.assertFalse(V.is_valid_unc_path(r'\\\a')) self.assertFalse(V.is_valid_unc_path(r'C:\path')) self.assertFalse(V.is_valid_unc_path(r'\\<a>')) // ... rest of the code ...
6f557ed73372aa5823393a53b079bf4cec7511b8
docker/ssladapter/ssladapter.py
docker/ssladapter/ssladapter.py
from distutils.version import StrictVersion from requests.adapters import HTTPAdapter try: from requests.packages.urllib3.poolmanager import PoolManager except ImportError: import urllib3 from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that uses an arbitrary SSL version.''' def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): urllib_ver = urllib3.__version__ if urllib3 and StrictVersion(urllib_ver) <= StrictVersion('1.5'): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block) else: self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
from distutils.version import StrictVersion from requests.adapters import HTTPAdapter try: import requests.packages.urllib3 as urllib3 except ImportError: import urllib3 PoolManager = urllib3.poolmanager.PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that uses an arbitrary SSL version.''' def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): urllib_ver = urllib3.__version__.split('-')[0] if urllib3 and urllib_ver != 'dev' and \ StrictVersion(urllib_ver) <= StrictVersion('1.5'): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block) else: self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
Fix some urllib3 import issues
Fix some urllib3 import issues
Python
apache-2.0
runcom/docker-py,kaiyou/docker-py,TomasTomecek/docker-py,vpetersson/docker-py,dimaspivak/docker-py,paulbellamy/docker-py,ClusterHQ/docker-py,terminalmage/docker-py,ticosax/docker-py,kaiyou/docker-py,dimaspivak/docker-py,auready/docker-py,dnephin/docker-py,docker/docker-py,docker/docker-py,mrfuxi/docker-py,bboreham/docker-py,rhatdan/docker-py,mangalaman93/docker-py,hibooboo2/docker-py,rancher/docker-py,tbeadle/docker-py,vitalyisaev2/docker-py,sourcelair/docker-py,uian/docker-py,delfick/docker-py,olsaki/docker-py,youhong316/docker-py,funkyfuture/docker-py,ColinHuang/docker-py,ssanderson/docker-py,youhong316/docker-py,MohamedAshiqrh/docker-py,auready/docker-py,v-boyko/docker-py,clarete/docker-py,bfirsh/docker-py,aiden0z/docker-py,tangkun75/docker-py,mnowster/docker-py,Faylixe/docker-py,gamechanger/docker-py,shakamunyi/docker-py,jpopelka/docker-py,mikedougherty/docker-py,zwqzhangweiqiang/docker-py,mark-adams/docker-py,funkyfuture/docker-py,shishir-a412ed/docker-py,minzhang28/docker-py,shakamunyi/docker-py,ColinHuang/docker-py,tshauck/docker-py,vpetersson/docker-py,Faylixe/docker-py,jamesmarva/docker-py,wehkamp/docker-py,wallrj/docker-py,vdemeester/docker-py,minzhang28/docker-py,leolujuyi/docker-py,mohitsoni/docker-py,Melraidin/docker-py,bfirsh/docker-py,aiden0z/docker-py,vdemeester/docker-py,Ye-Yong-Chi/docker-py,erikced/docker-py,PierreF/docker-py,kpavel/docker-py,schu/docker-py,tristan0x/docker-py,rastaman/docker-py,stevenewey/docker-py,rschmidtz/docker-py,dlorenc/docker-py,jhowardmsft/docker-py,wlan0/docker-py,uggla/docker-py
from distutils.version import StrictVersion from requests.adapters import HTTPAdapter try: - from requests.packages.urllib3.poolmanager import PoolManager + import requests.packages.urllib3 as urllib3 except ImportError: import urllib3 - from urllib3.poolmanager import PoolManager + + + PoolManager = urllib3.poolmanager.PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that uses an arbitrary SSL version.''' def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): - urllib_ver = urllib3.__version__ + urllib_ver = urllib3.__version__.split('-')[0] + if urllib3 and urllib_ver != 'dev' and \ - if urllib3 and StrictVersion(urllib_ver) <= StrictVersion('1.5'): + StrictVersion(urllib_ver) <= StrictVersion('1.5'): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block) else: self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
Fix some urllib3 import issues
## Code Before: from distutils.version import StrictVersion from requests.adapters import HTTPAdapter try: from requests.packages.urllib3.poolmanager import PoolManager except ImportError: import urllib3 from urllib3.poolmanager import PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that uses an arbitrary SSL version.''' def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): urllib_ver = urllib3.__version__ if urllib3 and StrictVersion(urllib_ver) <= StrictVersion('1.5'): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block) else: self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version) ## Instruction: Fix some urllib3 import issues ## Code After: from distutils.version import StrictVersion from requests.adapters import HTTPAdapter try: import requests.packages.urllib3 as urllib3 except ImportError: import urllib3 PoolManager = urllib3.poolmanager.PoolManager class SSLAdapter(HTTPAdapter): '''An HTTPS Transport Adapter that uses an arbitrary SSL version.''' def __init__(self, ssl_version=None, **kwargs): self.ssl_version = ssl_version super(SSLAdapter, self).__init__(**kwargs) def init_poolmanager(self, connections, maxsize, block=False): urllib_ver = urllib3.__version__.split('-')[0] if urllib3 and urllib_ver != 'dev' and \ StrictVersion(urllib_ver) <= StrictVersion('1.5'): self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block) else: self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize, block=block, ssl_version=self.ssl_version)
... try: import requests.packages.urllib3 as urllib3 except ImportError: ... import urllib3 PoolManager = urllib3.poolmanager.PoolManager ... def init_poolmanager(self, connections, maxsize, block=False): urllib_ver = urllib3.__version__.split('-')[0] if urllib3 and urllib_ver != 'dev' and \ StrictVersion(urllib_ver) <= StrictVersion('1.5'): self.poolmanager = PoolManager(num_pools=connections, ...
a4a37a783efcfd1cbb21acc29077c8096a0a0198
spacy/lang/pl/__init__.py
spacy/lang/pl/__init__.py
from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .stop_words import STOP_WORDS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class Polish(Language): lang = 'pl' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'pl' tokenizer_exceptions = update_exc(BASE_EXCEPTIONS) stop_words = set(STOP_WORDS) __all__ = ['Polish']
from __future__ import unicode_literals from .stop_words import STOP_WORDS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class Polish(Language): lang = 'pl' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'pl' tokenizer_exceptions = update_exc(BASE_EXCEPTIONS) stop_words = set(STOP_WORDS) __all__ = ['Polish']
Remove import from non-existing module
Remove import from non-existing module
Python
mit
honnibal/spaCy,spacy-io/spaCy,explosion/spaCy,recognai/spaCy,spacy-io/spaCy,explosion/spaCy,spacy-io/spaCy,explosion/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,spacy-io/spaCy,honnibal/spaCy,honnibal/spaCy,recognai/spaCy,aikramer2/spaCy,honnibal/spaCy,explosion/spaCy,aikramer2/spaCy,explosion/spaCy,aikramer2/spaCy,spacy-io/spaCy,recognai/spaCy,spacy-io/spaCy,aikramer2/spaCy,aikramer2/spaCy,recognai/spaCy,recognai/spaCy
from __future__ import unicode_literals - from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .stop_words import STOP_WORDS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class Polish(Language): lang = 'pl' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'pl' tokenizer_exceptions = update_exc(BASE_EXCEPTIONS) stop_words = set(STOP_WORDS) __all__ = ['Polish']
Remove import from non-existing module
## Code Before: from __future__ import unicode_literals from .tokenizer_exceptions import TOKENIZER_EXCEPTIONS from .stop_words import STOP_WORDS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class Polish(Language): lang = 'pl' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'pl' tokenizer_exceptions = update_exc(BASE_EXCEPTIONS) stop_words = set(STOP_WORDS) __all__ = ['Polish'] ## Instruction: Remove import from non-existing module ## Code After: from __future__ import unicode_literals from .stop_words import STOP_WORDS from ..tokenizer_exceptions import BASE_EXCEPTIONS from ...language import Language from ...attrs import LANG from ...util import update_exc class Polish(Language): lang = 'pl' class Defaults(Language.Defaults): lex_attr_getters = dict(Language.Defaults.lex_attr_getters) lex_attr_getters[LANG] = lambda text: 'pl' tokenizer_exceptions = update_exc(BASE_EXCEPTIONS) stop_words = set(STOP_WORDS) __all__ = ['Polish']
... from .stop_words import STOP_WORDS ...
4a6f76857a626dd756675a4fe1dd3660cf63d8b7
alg_fibonacci.py
alg_fibonacci.py
from __future__ import print_function import time def fibonacci(n): """Get nth number of Fibonacci series by recursion.""" if n == 0: return 0 elif n == 1 or n == 2: return 1 else: return fibonacci(n - 1) + fibonacci(n - 2) def main(): n = 13 print('{}th number of Fibonacci series: {}' .format(n, fibonacci(n))) if __name__ == '__main__': main()
from __future__ import print_function def fibonacci(n): """Get nth number of Fibonacci series by recursion.""" if n == 0: return 0 elif n == 1 or n == 2: return 1 else: return fibonacci(n - 1) + fibonacci(n - 2) def main(): import time n = 13 print('{}th number of Fibonacci series: {}' .format(n, fibonacci(n))) if __name__ == '__main__': main()
Move module time to main()
Move module time to main()
Python
bsd-2-clause
bowen0701/algorithms_data_structures
from __future__ import print_function - import time def fibonacci(n): """Get nth number of Fibonacci series by recursion.""" if n == 0: return 0 elif n == 1 or n == 2: return 1 else: return fibonacci(n - 1) + fibonacci(n - 2) def main(): + import time n = 13 print('{}th number of Fibonacci series: {}' .format(n, fibonacci(n))) if __name__ == '__main__': main()
Move module time to main()
## Code Before: from __future__ import print_function import time def fibonacci(n): """Get nth number of Fibonacci series by recursion.""" if n == 0: return 0 elif n == 1 or n == 2: return 1 else: return fibonacci(n - 1) + fibonacci(n - 2) def main(): n = 13 print('{}th number of Fibonacci series: {}' .format(n, fibonacci(n))) if __name__ == '__main__': main() ## Instruction: Move module time to main() ## Code After: from __future__ import print_function def fibonacci(n): """Get nth number of Fibonacci series by recursion.""" if n == 0: return 0 elif n == 1 or n == 2: return 1 else: return fibonacci(n - 1) + fibonacci(n - 2) def main(): import time n = 13 print('{}th number of Fibonacci series: {}' .format(n, fibonacci(n))) if __name__ == '__main__': main()
// ... existing code ... from __future__ import print_function // ... modified code ... def main(): import time n = 13 // ... rest of the code ...
c713273fe145418113d750579f8b135dc513c3b8
config.py
config.py
import os if os.environ.get('DATABASE_URL') is None: SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db' else: SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL'] SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
import os SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL'] SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
Delete default case for SQLALCHEMY_DATABASE_URI
Delete default case for SQLALCHEMY_DATABASE_URI if user doesn't set it, he coud have some problems with SQLite
Python
mit
Stark-Mountain/meetup-facebook-bot,Stark-Mountain/meetup-facebook-bot
import os - if os.environ.get('DATABASE_URL') is None: - SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db' - else: - SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL'] + SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL'] - SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
Delete default case for SQLALCHEMY_DATABASE_URI
## Code Before: import os if os.environ.get('DATABASE_URL') is None: SQLALCHEMY_DATABASE_URI = 'sqlite:///meetup.db' else: SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL'] SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning ## Instruction: Delete default case for SQLALCHEMY_DATABASE_URI ## Code After: import os SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL'] SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning
// ... existing code ... SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL'] SQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning // ... rest of the code ...
8fc274021a8c0813f3fc3568d1d7984112952b9c
pytilemap/qtsupport.py
pytilemap/qtsupport.py
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.TempLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.TempLocation)
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
Use Cache location instead of temp folder
Use Cache location instead of temp folder
Python
mit
allebacco/PyTileMap
import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): - return QStandardPaths.writableLocation(QStandardPaths.TempLocation) + return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): - return QDesktopServices.storageLocation(QDesktopServices.TempLocation) + return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
Use Cache location instead of temp folder
## Code Before: import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.TempLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.TempLocation) ## Instruction: Use Cache location instead of temp folder ## Code After: import sys import sip import qtpy __all__ = [ 'getQVariantValue', 'wheelAngleDelta', ] try: if qtpy.PYQT5: QVARIANT_API = 2 else: QVARIANT_API = sip.getapi('QVariant') except ValueError: QVARIANT_API = 1 if QVARIANT_API == 1: def getQVariantValue(variant): return variant.toPyObject() else: def getQVariantValue(variant): return variant if qtpy.PYQT5: def wheelAngleDelta(wheelEvent): return wheelEvent.angleDelta().y() else: def wheelAngleDelta(wheelEvent): return wheelEvent.delta() if qtpy.PYQT5: from qtpy.QtCore import QStandardPaths def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) else: from qtpy.QtGui import QDesktopServices def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation)
# ... existing code ... def getTemporaryFolder(): return QStandardPaths.writableLocation(QStandardPaths.CacheLocation) # ... modified code ... def getTemporaryFolder(): return QDesktopServices.storageLocation(QDesktopServices.CacheLocation) # ... rest of the code ...
9d5abdaefa483574cdd81da8d8d4e63ef68f5ab8
crossfolium/__init__.py
crossfolium/__init__.py
import crossfolium.marker_function as marker_function from crossfolium.crossfolium import ( Crossfilter, PieFilter, RowBarFilter, BarFilter, TableFilter, CountFilter, ResetFilter, GeoChoroplethFilter, ) from .map import ( FeatureGroupFilter, HeatmapFilter, ) __version__ = "0.0.0" __all__ = [ '__version__', 'marker_function', 'Crossfilter', 'PieFilter', 'RowBarFilter', 'BarFilter', 'FeatureGroupFilter', 'TableFilter', 'CountFilter', 'ResetFilter', 'HeatmapFilter', 'GeoChoroplethFilter', ]
from __future__ import absolute_import from crossfolium import marker_function from crossfolium.crossfolium import ( Crossfilter, PieFilter, RowBarFilter, BarFilter, TableFilter, CountFilter, ResetFilter, GeoChoroplethFilter, ) from crossfolium.map import ( FeatureGroupFilter, HeatmapFilter, ) __version__ = "0.0.0" __all__ = [ '__version__', 'marker_function', 'Crossfilter', 'PieFilter', 'RowBarFilter', 'BarFilter', 'FeatureGroupFilter', 'TableFilter', 'CountFilter', 'ResetFilter', 'HeatmapFilter', 'GeoChoroplethFilter', ]
Handle absolute import for py27
Handle absolute import for py27
Python
mit
BibMartin/crossfolium,BibMartin/crossfolium
+ from __future__ import absolute_import - import crossfolium.marker_function as marker_function + from crossfolium import marker_function from crossfolium.crossfolium import ( Crossfilter, PieFilter, RowBarFilter, BarFilter, TableFilter, CountFilter, ResetFilter, GeoChoroplethFilter, ) - from .map import ( + from crossfolium.map import ( FeatureGroupFilter, HeatmapFilter, ) __version__ = "0.0.0" __all__ = [ '__version__', 'marker_function', 'Crossfilter', 'PieFilter', 'RowBarFilter', 'BarFilter', 'FeatureGroupFilter', 'TableFilter', 'CountFilter', 'ResetFilter', 'HeatmapFilter', 'GeoChoroplethFilter', ]
Handle absolute import for py27
## Code Before: import crossfolium.marker_function as marker_function from crossfolium.crossfolium import ( Crossfilter, PieFilter, RowBarFilter, BarFilter, TableFilter, CountFilter, ResetFilter, GeoChoroplethFilter, ) from .map import ( FeatureGroupFilter, HeatmapFilter, ) __version__ = "0.0.0" __all__ = [ '__version__', 'marker_function', 'Crossfilter', 'PieFilter', 'RowBarFilter', 'BarFilter', 'FeatureGroupFilter', 'TableFilter', 'CountFilter', 'ResetFilter', 'HeatmapFilter', 'GeoChoroplethFilter', ] ## Instruction: Handle absolute import for py27 ## Code After: from __future__ import absolute_import from crossfolium import marker_function from crossfolium.crossfolium import ( Crossfilter, PieFilter, RowBarFilter, BarFilter, TableFilter, CountFilter, ResetFilter, GeoChoroplethFilter, ) from crossfolium.map import ( FeatureGroupFilter, HeatmapFilter, ) __version__ = "0.0.0" __all__ = [ '__version__', 'marker_function', 'Crossfilter', 'PieFilter', 'RowBarFilter', 'BarFilter', 'FeatureGroupFilter', 'TableFilter', 'CountFilter', 'ResetFilter', 'HeatmapFilter', 'GeoChoroplethFilter', ]
... from __future__ import absolute_import from crossfolium import marker_function ... from crossfolium.map import ( FeatureGroupFilter, ...
6aa8f148b3b3975363d5d4a763f5abb45ea6cbd8
databin/parsers/__init__.py
databin/parsers/__init__.py
from databin.parsers.util import ParseException from databin.parsers.simple import parse_csv, parse_tsv from databin.parsers.psql import parse_psql PARSERS = [ ('Comma-Separated Values', 'csv', parse_csv), ('Tab-Separated Values', 'tsv', parse_tsv), ('Excel copy & paste', 'excel', parse_tsv), ('psql Shell', 'psql', parse_psql), ] def parse(format, data): for name, key, func in PARSERS: if key == format: return func(data) raise ParseException() def get_parsers(): for name, key, func in PARSERS: yield (key, name)
from databin.parsers.util import ParseException from databin.parsers.simple import parse_csv, parse_tsv from databin.parsers.psql import parse_psql PARSERS = [ ('Excel copy & paste', 'excel', parse_tsv), ('Comma-Separated Values', 'csv', parse_csv), ('Tab-Separated Values', 'tsv', parse_tsv), ('psql Shell', 'psql', parse_psql), ] def parse(format, data): for name, key, func in PARSERS: if key == format: return func(data) raise ParseException() def get_parsers(): for name, key, func in PARSERS: yield (key, name)
Make excel format the default
Make excel format the default
Python
mit
LeTristanB/Pastable,pudo/databin,LeTristanB/Pastable
from databin.parsers.util import ParseException from databin.parsers.simple import parse_csv, parse_tsv from databin.parsers.psql import parse_psql PARSERS = [ + ('Excel copy & paste', 'excel', parse_tsv), ('Comma-Separated Values', 'csv', parse_csv), ('Tab-Separated Values', 'tsv', parse_tsv), - ('Excel copy & paste', 'excel', parse_tsv), ('psql Shell', 'psql', parse_psql), ] def parse(format, data): for name, key, func in PARSERS: if key == format: return func(data) raise ParseException() def get_parsers(): for name, key, func in PARSERS: yield (key, name)
Make excel format the default
## Code Before: from databin.parsers.util import ParseException from databin.parsers.simple import parse_csv, parse_tsv from databin.parsers.psql import parse_psql PARSERS = [ ('Comma-Separated Values', 'csv', parse_csv), ('Tab-Separated Values', 'tsv', parse_tsv), ('Excel copy & paste', 'excel', parse_tsv), ('psql Shell', 'psql', parse_psql), ] def parse(format, data): for name, key, func in PARSERS: if key == format: return func(data) raise ParseException() def get_parsers(): for name, key, func in PARSERS: yield (key, name) ## Instruction: Make excel format the default ## Code After: from databin.parsers.util import ParseException from databin.parsers.simple import parse_csv, parse_tsv from databin.parsers.psql import parse_psql PARSERS = [ ('Excel copy & paste', 'excel', parse_tsv), ('Comma-Separated Values', 'csv', parse_csv), ('Tab-Separated Values', 'tsv', parse_tsv), ('psql Shell', 'psql', parse_psql), ] def parse(format, data): for name, key, func in PARSERS: if key == format: return func(data) raise ParseException() def get_parsers(): for name, key, func in PARSERS: yield (key, name)
# ... existing code ... PARSERS = [ ('Excel copy & paste', 'excel', parse_tsv), ('Comma-Separated Values', 'csv', parse_csv), # ... modified code ... ('Tab-Separated Values', 'tsv', parse_tsv), ('psql Shell', 'psql', parse_psql), # ... rest of the code ...
fc41951d1e395c3cdc8994b4c025e9776d67d4e0
http.py
http.py
from django.http import HttpResponse class HttpResponseCreated(HttpResponse): status_code = 201 class HttpResponseNoContent(HttpResponse): status_code = 204 class HttpResponseNotAllowed(HttpResponse): status_code = 405 def __init__(self, allow_headers): """ RFC2616: The response MUST include an Allow header containing a list of valid methods for the requested resource. """ super(HttpResponseNotAllowed, self).__init__() try: iter(allow_headers) except TypeError: self['Allow'] = allow_headers else: self['Allow'] = ", ".join(allow_headers) class HttpResponseNotAcceptable(HttpResponse): status_code = 406 class HttpResponseConflict(HttpResponse): status_code = 409 class HttpResponseNotImplemented(HttpResponse): status_code = 501
from django.http import HttpResponse class HttpResponseCreated(HttpResponse): status_code = 201 class HttpResponseNoContent(HttpResponse): status_code = 204 class HttpResponseNotAllowed(HttpResponse): status_code = 405 def __init__(self, allow_headers): """ RFC2616: The response MUST include an Allow header containing a list of valid methods for the requested resource. """ super(HttpResponseNotAllowed, self).__init__() try: iter(allow_headers) except TypeError: self['Allow'] = allow_headers else: self['Allow'] = ", ".join(allow_headers) class HttpResponseNotAcceptable(HttpResponse): status_code = 406 class HttpResponseConflict(HttpResponse): status_code = 409 class HttpResponseUnsupportedMediaType(HttpResponse): status_code = 415 class HttpResponseNotImplemented(HttpResponse): status_code = 501
Add Response class for unsupported media
Add Response class for unsupported media
Python
mit
danrex/django-riv,danrex/django-riv
from django.http import HttpResponse class HttpResponseCreated(HttpResponse): status_code = 201 class HttpResponseNoContent(HttpResponse): status_code = 204 class HttpResponseNotAllowed(HttpResponse): status_code = 405 def __init__(self, allow_headers): """ RFC2616: The response MUST include an Allow header containing a list of valid methods for the requested resource. """ super(HttpResponseNotAllowed, self).__init__() try: iter(allow_headers) except TypeError: self['Allow'] = allow_headers else: self['Allow'] = ", ".join(allow_headers) class HttpResponseNotAcceptable(HttpResponse): status_code = 406 class HttpResponseConflict(HttpResponse): status_code = 409 + class HttpResponseUnsupportedMediaType(HttpResponse): + status_code = 415 + class HttpResponseNotImplemented(HttpResponse): status_code = 501
Add Response class for unsupported media
## Code Before: from django.http import HttpResponse class HttpResponseCreated(HttpResponse): status_code = 201 class HttpResponseNoContent(HttpResponse): status_code = 204 class HttpResponseNotAllowed(HttpResponse): status_code = 405 def __init__(self, allow_headers): """ RFC2616: The response MUST include an Allow header containing a list of valid methods for the requested resource. """ super(HttpResponseNotAllowed, self).__init__() try: iter(allow_headers) except TypeError: self['Allow'] = allow_headers else: self['Allow'] = ", ".join(allow_headers) class HttpResponseNotAcceptable(HttpResponse): status_code = 406 class HttpResponseConflict(HttpResponse): status_code = 409 class HttpResponseNotImplemented(HttpResponse): status_code = 501 ## Instruction: Add Response class for unsupported media ## Code After: from django.http import HttpResponse class HttpResponseCreated(HttpResponse): status_code = 201 class HttpResponseNoContent(HttpResponse): status_code = 204 class HttpResponseNotAllowed(HttpResponse): status_code = 405 def __init__(self, allow_headers): """ RFC2616: The response MUST include an Allow header containing a list of valid methods for the requested resource. """ super(HttpResponseNotAllowed, self).__init__() try: iter(allow_headers) except TypeError: self['Allow'] = allow_headers else: self['Allow'] = ", ".join(allow_headers) class HttpResponseNotAcceptable(HttpResponse): status_code = 406 class HttpResponseConflict(HttpResponse): status_code = 409 class HttpResponseUnsupportedMediaType(HttpResponse): status_code = 415 class HttpResponseNotImplemented(HttpResponse): status_code = 501
// ... existing code ... class HttpResponseUnsupportedMediaType(HttpResponse): status_code = 415 class HttpResponseNotImplemented(HttpResponse): // ... rest of the code ...
a32e61e9cdf2eababb568659766688a731b121cb
warlock/__init__.py
warlock/__init__.py
"""Public-facing Warlock API""" from warlock.core import model_factory # NOQA from warlock.exceptions import InvalidOperation # NOQA
"""Public-facing Warlock API""" from warlock.core import model_factory # noqa: F401 from warlock.exceptions import InvalidOperation # noqa: F401
Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook
Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook
Python
apache-2.0
bcwaldon/warlock
"""Public-facing Warlock API""" - from warlock.core import model_factory # NOQA + from warlock.core import model_factory # noqa: F401 - from warlock.exceptions import InvalidOperation # NOQA + from warlock.exceptions import InvalidOperation # noqa: F401
Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook
## Code Before: """Public-facing Warlock API""" from warlock.core import model_factory # NOQA from warlock.exceptions import InvalidOperation # NOQA ## Instruction: Apply 'no blanket NOQA statements' fixes enforced by pre-commit hook ## Code After: """Public-facing Warlock API""" from warlock.core import model_factory # noqa: F401 from warlock.exceptions import InvalidOperation # noqa: F401
... from warlock.core import model_factory # noqa: F401 from warlock.exceptions import InvalidOperation # noqa: F401 ...
61bbd4e8fc0712fe56614481173eb86d409eb8d7
tests/test_linked_list.py
tests/test_linked_list.py
from unittest import TestCase from pystructures.linked_lists import LinkedList, Node class TestNode(TestCase): def test_value(self): """ A simple test to check the Node's value """ node = Node(10) self.assertEqual(10, node.value) def test_improper_node(self): """ A test to check if an invalid data type is set as a node's next""" node = Node(10) with self.assertRaises(ValueError): node.next = "Hello" class TestLinkedList(TestCase): def test_insert(self): """ A simple test to check if insertion works as expected in a singly linked list """ l = LinkedList() results = [l.insert(val) for val in xrange(10, 100, 10)] self.assertEqual(len(set(results)), 1) self.assertTrue(results[0], msg="Testing for successful insertion...") self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
from builtins import range from unittest import TestCase from pystructures.linked_lists import LinkedList, Node class TestNode(TestCase): def test_value(self): """ A simple test to check the Node's value """ node = Node(10) self.assertEqual(10, node.value) def test_improper_node(self): """ A test to check if an invalid data type is set as a node's next""" node = Node(10) with self.assertRaises(ValueError): node.next = "Hello" class TestLinkedList(TestCase): def test_insert(self): """ A simple test to check if insertion works as expected in a singly linked list """ l = LinkedList() results = [l.insert(val) for val in range(10, 100, 10)] self.assertEqual(len(set(results)), 1) self.assertTrue(results[0], msg="Testing for successful insertion...") self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
Fix range issue with travis
Fix range issue with travis
Python
mit
apranav19/pystructures
+ from builtins import range from unittest import TestCase from pystructures.linked_lists import LinkedList, Node class TestNode(TestCase): def test_value(self): """ A simple test to check the Node's value """ node = Node(10) self.assertEqual(10, node.value) def test_improper_node(self): """ A test to check if an invalid data type is set as a node's next""" node = Node(10) with self.assertRaises(ValueError): node.next = "Hello" class TestLinkedList(TestCase): def test_insert(self): """ A simple test to check if insertion works as expected in a singly linked list """ l = LinkedList() - results = [l.insert(val) for val in xrange(10, 100, 10)] + results = [l.insert(val) for val in range(10, 100, 10)] self.assertEqual(len(set(results)), 1) self.assertTrue(results[0], msg="Testing for successful insertion...") self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
Fix range issue with travis
## Code Before: from unittest import TestCase from pystructures.linked_lists import LinkedList, Node class TestNode(TestCase): def test_value(self): """ A simple test to check the Node's value """ node = Node(10) self.assertEqual(10, node.value) def test_improper_node(self): """ A test to check if an invalid data type is set as a node's next""" node = Node(10) with self.assertRaises(ValueError): node.next = "Hello" class TestLinkedList(TestCase): def test_insert(self): """ A simple test to check if insertion works as expected in a singly linked list """ l = LinkedList() results = [l.insert(val) for val in xrange(10, 100, 10)] self.assertEqual(len(set(results)), 1) self.assertTrue(results[0], msg="Testing for successful insertion...") self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...") ## Instruction: Fix range issue with travis ## Code After: from builtins import range from unittest import TestCase from pystructures.linked_lists import LinkedList, Node class TestNode(TestCase): def test_value(self): """ A simple test to check the Node's value """ node = Node(10) self.assertEqual(10, node.value) def test_improper_node(self): """ A test to check if an invalid data type is set as a node's next""" node = Node(10) with self.assertRaises(ValueError): node.next = "Hello" class TestLinkedList(TestCase): def test_insert(self): """ A simple test to check if insertion works as expected in a singly linked list """ l = LinkedList() results = [l.insert(val) for val in range(10, 100, 10)] self.assertEqual(len(set(results)), 1) self.assertTrue(results[0], msg="Testing for successful insertion...") self.assertEqual(len(results), l.size, msg="Testing if # of results equal list size...")
... from builtins import range from unittest import TestCase ... l = LinkedList() results = [l.insert(val) for val in range(10, 100, 10)] self.assertEqual(len(set(results)), 1) ...
9851430922f9c14583c9eb17062629f6ea99c258
turbustat/tests/test_vcs.py
turbustat/tests/test_vcs.py
''' Test functions for VCS ''' from unittest import TestCase import numpy as np import numpy.testing as npt from ..statistics import VCS, VCS_Distance from ._testing_data import \ dataset1, dataset2, computed_data, computed_distances class testVCS(TestCase): def setUp(self): self.dataset1 = dataset1 self.dataset2 = dataset2 def test_VCS_method(self): self.tester = VCS(dataset1["cube"]).run() npt.assert_allclose(self.tester.ps1D, computed_data['vcs_val']) def test_VCS_distance(self): self.tester_dist = \ VCS_Distance(dataset1["cube"], dataset2["cube"]) self.tester_dist = self.tester_dist.distance_metric() npt.assert_almost_equal(self.tester_dist.distance, computed_distances['vcs_distance'])
''' Test functions for VCS ''' import pytest import numpy as np import numpy.testing as npt from ..statistics import VCS, VCS_Distance from ._testing_data import \ dataset1, dataset2, computed_data, computed_distances def test_VCS_method(): tester = VCS(dataset1["cube"]).run() npt.assert_allclose(tester.ps1D, computed_data['vcs_val']) npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val']) def test_VCS_distance(): tester_dist = \ VCS_Distance(dataset1["cube"], dataset2["cube"]) tester_dist = tester_dist.distance_metric() npt.assert_almost_equal(tester_dist.distance, computed_distances['vcs_distance']) # Add tests for: VCS changing the spectral width, pixel and spectral units,
Reformat VCS tests; need updated unit test values!
Reformat VCS tests; need updated unit test values!
Python
mit
e-koch/TurbuStat,Astroua/TurbuStat
''' Test functions for VCS ''' - from unittest import TestCase + import pytest import numpy as np import numpy.testing as npt from ..statistics import VCS, VCS_Distance from ._testing_data import \ dataset1, dataset2, computed_data, computed_distances - class testVCS(TestCase): + def test_VCS_method(): + tester = VCS(dataset1["cube"]).run() + npt.assert_allclose(tester.ps1D, computed_data['vcs_val']) - def setUp(self): - self.dataset1 = dataset1 - self.dataset2 = dataset2 + npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val']) - def test_VCS_method(self): - self.tester = VCS(dataset1["cube"]).run() - npt.assert_allclose(self.tester.ps1D, computed_data['vcs_val']) - def test_VCS_distance(self): + def test_VCS_distance(): - self.tester_dist = \ + tester_dist = \ - VCS_Distance(dataset1["cube"], dataset2["cube"]) + VCS_Distance(dataset1["cube"], dataset2["cube"]) - self.tester_dist = self.tester_dist.distance_metric() + tester_dist = tester_dist.distance_metric() - npt.assert_almost_equal(self.tester_dist.distance, + npt.assert_almost_equal(tester_dist.distance, - computed_distances['vcs_distance']) + computed_distances['vcs_distance']) + # Add tests for: VCS changing the spectral width, pixel and spectral units, +
Reformat VCS tests; need updated unit test values!
## Code Before: ''' Test functions for VCS ''' from unittest import TestCase import numpy as np import numpy.testing as npt from ..statistics import VCS, VCS_Distance from ._testing_data import \ dataset1, dataset2, computed_data, computed_distances class testVCS(TestCase): def setUp(self): self.dataset1 = dataset1 self.dataset2 = dataset2 def test_VCS_method(self): self.tester = VCS(dataset1["cube"]).run() npt.assert_allclose(self.tester.ps1D, computed_data['vcs_val']) def test_VCS_distance(self): self.tester_dist = \ VCS_Distance(dataset1["cube"], dataset2["cube"]) self.tester_dist = self.tester_dist.distance_metric() npt.assert_almost_equal(self.tester_dist.distance, computed_distances['vcs_distance']) ## Instruction: Reformat VCS tests; need updated unit test values! ## Code After: ''' Test functions for VCS ''' import pytest import numpy as np import numpy.testing as npt from ..statistics import VCS, VCS_Distance from ._testing_data import \ dataset1, dataset2, computed_data, computed_distances def test_VCS_method(): tester = VCS(dataset1["cube"]).run() npt.assert_allclose(tester.ps1D, computed_data['vcs_val']) npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val']) def test_VCS_distance(): tester_dist = \ VCS_Distance(dataset1["cube"], dataset2["cube"]) tester_dist = tester_dist.distance_metric() npt.assert_almost_equal(tester_dist.distance, computed_distances['vcs_distance']) # Add tests for: VCS changing the spectral width, pixel and spectral units,
# ... existing code ... import pytest # ... modified code ... def test_VCS_method(): tester = VCS(dataset1["cube"]).run() npt.assert_allclose(tester.ps1D, computed_data['vcs_val']) npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val']) def test_VCS_distance(): tester_dist = \ VCS_Distance(dataset1["cube"], dataset2["cube"]) tester_dist = tester_dist.distance_metric() npt.assert_almost_equal(tester_dist.distance, computed_distances['vcs_distance']) # Add tests for: VCS changing the spectral width, pixel and spectral units, # ... rest of the code ...
f48cb4fd946c8fa4b6157b8e1ea9ad8b385bc67a
src/hades/bin/generate_config.py
src/hades/bin/generate_config.py
import os import sys from hades import constants from hades.common.cli import ArgumentParser, parser as common_parser from hades.config.generate import ConfigGenerator from hades.config.loader import load_config def main(): parser = ArgumentParser(parents=[common_parser]) parser.add_argument(dest='source', metavar='SOURCE', help="Template file name or template directory name") parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?', help="Destination file or directory (default is stdout" "for files; required for directories)") args = parser.parse_args() config = load_config(args.config) template_dir = constants.templatedir generator = ConfigGenerator(template_dir, config) source_path = os.path.join(template_dir, args.source) if os.path.isdir(source_path): generator.from_directory(args.source, args.destination) elif os.path.isfile(source_path): if args.destination is None: generator.from_file(args.source, sys.stdout) else: with open(args.destination, 'w', encoding='utf-8') as f: generator.from_file(args.source, f) else: print("No such file or directory {} in {}".format(args.source, template_dir), file=sys.stderr) return os.EX_NOINPUT if __name__ == '__main__': sys.exit(main())
import logging import os import sys from hades import constants from hades.common.cli import ArgumentParser, parser as common_parser from hades.config.generate import ConfigGenerator from hades.config.loader import load_config logger = logging.getLogger() def main(): parser = ArgumentParser(parents=[common_parser]) parser.add_argument(dest='source', metavar='SOURCE', help="Template file name or template directory name") parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?', help="Destination file or directory (default is stdout" "for files; required for directories)") args = parser.parse_args() config = load_config(args.config) template_dir = constants.templatedir generator = ConfigGenerator(template_dir, config) source_path = os.path.join(template_dir, args.source) if os.path.isdir(source_path): generator.from_directory(args.source, args.destination) elif os.path.isfile(source_path): if args.destination is None: generator.from_file(args.source, sys.stdout) else: with open(args.destination, 'w', encoding='utf-8') as f: generator.from_file(args.source, f) else: logger.critical("No such file or directory {} in {}" .format(args.source, template_dir)) return os.EX_NOINPUT if __name__ == '__main__': sys.exit(main())
Use logger for hades-generate-config error messages
Use logger for hades-generate-config error messages
Python
mit
agdsn/hades,agdsn/hades,agdsn/hades,agdsn/hades,agdsn/hades
+ import logging import os import sys from hades import constants from hades.common.cli import ArgumentParser, parser as common_parser from hades.config.generate import ConfigGenerator from hades.config.loader import load_config + + logger = logging.getLogger() def main(): parser = ArgumentParser(parents=[common_parser]) parser.add_argument(dest='source', metavar='SOURCE', help="Template file name or template directory name") parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?', help="Destination file or directory (default is stdout" "for files; required for directories)") args = parser.parse_args() config = load_config(args.config) template_dir = constants.templatedir generator = ConfigGenerator(template_dir, config) source_path = os.path.join(template_dir, args.source) if os.path.isdir(source_path): generator.from_directory(args.source, args.destination) elif os.path.isfile(source_path): if args.destination is None: generator.from_file(args.source, sys.stdout) else: with open(args.destination, 'w', encoding='utf-8') as f: generator.from_file(args.source, f) else: + logger.critical("No such file or directory {} in {}" + .format(args.source, template_dir)) - print("No such file or directory {} in {}".format(args.source, - template_dir), - file=sys.stderr) return os.EX_NOINPUT if __name__ == '__main__': sys.exit(main())
Use logger for hades-generate-config error messages
## Code Before: import os import sys from hades import constants from hades.common.cli import ArgumentParser, parser as common_parser from hades.config.generate import ConfigGenerator from hades.config.loader import load_config def main(): parser = ArgumentParser(parents=[common_parser]) parser.add_argument(dest='source', metavar='SOURCE', help="Template file name or template directory name") parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?', help="Destination file or directory (default is stdout" "for files; required for directories)") args = parser.parse_args() config = load_config(args.config) template_dir = constants.templatedir generator = ConfigGenerator(template_dir, config) source_path = os.path.join(template_dir, args.source) if os.path.isdir(source_path): generator.from_directory(args.source, args.destination) elif os.path.isfile(source_path): if args.destination is None: generator.from_file(args.source, sys.stdout) else: with open(args.destination, 'w', encoding='utf-8') as f: generator.from_file(args.source, f) else: print("No such file or directory {} in {}".format(args.source, template_dir), file=sys.stderr) return os.EX_NOINPUT if __name__ == '__main__': sys.exit(main()) ## Instruction: Use logger for hades-generate-config error messages ## Code After: import logging import os import sys from hades import constants from hades.common.cli import ArgumentParser, parser as common_parser from hades.config.generate import ConfigGenerator from hades.config.loader import load_config logger = logging.getLogger() def main(): parser = ArgumentParser(parents=[common_parser]) parser.add_argument(dest='source', metavar='SOURCE', help="Template file name or template directory name") parser.add_argument(dest='destination', metavar='DESTINATION', nargs='?', help="Destination file or directory (default is stdout" "for files; required for directories)") args = parser.parse_args() config = load_config(args.config) template_dir = constants.templatedir generator = ConfigGenerator(template_dir, config) source_path = os.path.join(template_dir, args.source) if os.path.isdir(source_path): generator.from_directory(args.source, args.destination) elif os.path.isfile(source_path): if args.destination is None: generator.from_file(args.source, sys.stdout) else: with open(args.destination, 'w', encoding='utf-8') as f: generator.from_file(args.source, f) else: logger.critical("No such file or directory {} in {}" .format(args.source, template_dir)) return os.EX_NOINPUT if __name__ == '__main__': sys.exit(main())
# ... existing code ... import logging import os # ... modified code ... from hades.config.loader import load_config logger = logging.getLogger() ... else: logger.critical("No such file or directory {} in {}" .format(args.source, template_dir)) return os.EX_NOINPUT # ... rest of the code ...
59d44ba76a9b2f98375fa2f893dabc0376de6f82
localeurl/models.py
localeurl/models.py
from django.conf import settings from django.core import urlresolvers from django.utils import translation from localeurl import utils def reverse(*args, **kwargs): reverse_kwargs = kwargs.get('kwargs') or {} locale = utils.supported_language(reverse_kwargs.pop( 'locale', translation.get_language())) url = django_reverse(*args, **kwargs) _, path = utils.strip_script_prefix(url) return utils.locale_url(path, locale) django_reverse = None def patch_reverse(): """ Monkey-patches the urlresolvers.reverse function. Will not patch twice. """ global django_reverse if urlresolvers.reverse is not reverse: django_reverse = urlresolvers.reverse urlresolvers.reverse = reverse if settings.USE_I18N: patch_reverse()
from django.conf import settings from django.core import urlresolvers from django.utils import translation from django.contrib.auth import views as auth_views from localeurl import utils def reverse(*args, **kwargs): reverse_kwargs = kwargs.get('kwargs') or {} locale = utils.supported_language(reverse_kwargs.pop( 'locale', translation.get_language())) url = django_reverse(*args, **kwargs) _, path = utils.strip_script_prefix(url) return utils.locale_url(path, locale) django_reverse = None def patch_reverse(): """ Monkey-patches the urlresolvers.reverse function. Will not patch twice. """ global django_reverse if urlresolvers.reverse is not reverse: django_reverse = urlresolvers.reverse urlresolvers.reverse = reverse def redirect_to_login(next, login_url, *args, **kwargs): if not login_url: login_url = settings.LOGIN_URL login_url = utils.locale_url(login_url, translation.get_language()) return django_redirect_to_login(next, login_url, *args, **kwargs) django_redirect_to_login = None def patch_redirect_to_login(): """ Monkey-patches the redirect_to_login function. Will not patch twice. """ global django_redirect_to_login if auth_views.redirect_to_login is not redirect_to_login: django_redirect_to_login = auth_views.redirect_to_login auth_views.redirect_to_login = redirect_to_login if settings.USE_I18N: patch_reverse() patch_redirect_to_login()
Patch redirect_to_login to maintain locale
Patch redirect_to_login to maintain locale Signed-off-by: Simon Luijk <[email protected]>
Python
mit
simonluijk/django-localeurl
from django.conf import settings from django.core import urlresolvers from django.utils import translation + from django.contrib.auth import views as auth_views from localeurl import utils + def reverse(*args, **kwargs): reverse_kwargs = kwargs.get('kwargs') or {} locale = utils.supported_language(reverse_kwargs.pop( 'locale', translation.get_language())) url = django_reverse(*args, **kwargs) _, path = utils.strip_script_prefix(url) return utils.locale_url(path, locale) django_reverse = None def patch_reverse(): """ Monkey-patches the urlresolvers.reverse function. Will not patch twice. """ global django_reverse if urlresolvers.reverse is not reverse: django_reverse = urlresolvers.reverse urlresolvers.reverse = reverse + + def redirect_to_login(next, login_url, *args, **kwargs): + if not login_url: + login_url = settings.LOGIN_URL + login_url = utils.locale_url(login_url, translation.get_language()) + return django_redirect_to_login(next, login_url, *args, **kwargs) + + django_redirect_to_login = None + + def patch_redirect_to_login(): + """ + Monkey-patches the redirect_to_login function. Will not patch twice. + """ + global django_redirect_to_login + if auth_views.redirect_to_login is not redirect_to_login: + django_redirect_to_login = auth_views.redirect_to_login + auth_views.redirect_to_login = redirect_to_login + + if settings.USE_I18N: patch_reverse() + patch_redirect_to_login()
Patch redirect_to_login to maintain locale
## Code Before: from django.conf import settings from django.core import urlresolvers from django.utils import translation from localeurl import utils def reverse(*args, **kwargs): reverse_kwargs = kwargs.get('kwargs') or {} locale = utils.supported_language(reverse_kwargs.pop( 'locale', translation.get_language())) url = django_reverse(*args, **kwargs) _, path = utils.strip_script_prefix(url) return utils.locale_url(path, locale) django_reverse = None def patch_reverse(): """ Monkey-patches the urlresolvers.reverse function. Will not patch twice. """ global django_reverse if urlresolvers.reverse is not reverse: django_reverse = urlresolvers.reverse urlresolvers.reverse = reverse if settings.USE_I18N: patch_reverse() ## Instruction: Patch redirect_to_login to maintain locale ## Code After: from django.conf import settings from django.core import urlresolvers from django.utils import translation from django.contrib.auth import views as auth_views from localeurl import utils def reverse(*args, **kwargs): reverse_kwargs = kwargs.get('kwargs') or {} locale = utils.supported_language(reverse_kwargs.pop( 'locale', translation.get_language())) url = django_reverse(*args, **kwargs) _, path = utils.strip_script_prefix(url) return utils.locale_url(path, locale) django_reverse = None def patch_reverse(): """ Monkey-patches the urlresolvers.reverse function. Will not patch twice. """ global django_reverse if urlresolvers.reverse is not reverse: django_reverse = urlresolvers.reverse urlresolvers.reverse = reverse def redirect_to_login(next, login_url, *args, **kwargs): if not login_url: login_url = settings.LOGIN_URL login_url = utils.locale_url(login_url, translation.get_language()) return django_redirect_to_login(next, login_url, *args, **kwargs) django_redirect_to_login = None def patch_redirect_to_login(): """ Monkey-patches the redirect_to_login function. Will not patch twice. """ global django_redirect_to_login if auth_views.redirect_to_login is not redirect_to_login: django_redirect_to_login = auth_views.redirect_to_login auth_views.redirect_to_login = redirect_to_login if settings.USE_I18N: patch_reverse() patch_redirect_to_login()
... from django.utils import translation from django.contrib.auth import views as auth_views from localeurl import utils ... def redirect_to_login(next, login_url, *args, **kwargs): if not login_url: login_url = settings.LOGIN_URL login_url = utils.locale_url(login_url, translation.get_language()) return django_redirect_to_login(next, login_url, *args, **kwargs) django_redirect_to_login = None def patch_redirect_to_login(): """ Monkey-patches the redirect_to_login function. Will not patch twice. """ global django_redirect_to_login if auth_views.redirect_to_login is not redirect_to_login: django_redirect_to_login = auth_views.redirect_to_login auth_views.redirect_to_login = redirect_to_login if settings.USE_I18N: ... patch_reverse() patch_redirect_to_login() ...
9e73de0014b3f88b9e94ead11a878c6bc3819782
selenium_testcase/tests/test_navigation.py
selenium_testcase/tests/test_navigation.py
from __future__ import absolute_import from ..testcases import SeleniumLiveTestCase class NavigationTestCase(SeleniumLiveTestCase): test_templates = [ (r'^nav_1/$', 'nav_1.html'), (r'^nav_1/nav_2/$', 'nav_2.html') ] def test_get_page(self): """ Test that you can traverse the page tree. """ self.get_page("/nav_1/") self.should_see("This is nav 1.") self.get_page("/nav_1/nav_2/") self.should_see("This is nav 2.")
from __future__ import absolute_import from ..testcases import SeleniumLiveTestCase class NavigationTestCase(SeleniumLiveTestCase): test_templates = [ (r'^nav_1/$', 'nav_1.html'), (r'^nav_1/nav_2/$', 'nav_2.html') ] def test_get_page(self): """ Test that you can traverse the page tree. """ self.get_page("/nav_1/") self.has_title("Navigation 1") self.title_contains("1") self.should_see("This is nav 1.") self.get_page("/nav_1/nav_2/") self.should_see("This is nav 2.") def test_get_bad_page(self): """ Test that /bogus/ is not found. """ self.get_page("/bogus/") self.should_see("Not Found") self.should_see( "The requested URL /bogus/ was not found on this server.") def test_missing_content_with_retry(self): """ Test retry for missing content, LONG RETRIES! """ self.get_page("/nav_1/") self.should_not_see("This is nav 2.") self.url_should_not_contain("nav_2") self.has_not_title("Navigation 2") self.title_does_not_contain("2") self.assertRaises( AssertionError, self.click_button, "not_there_dude") self.not_at_page("/nav_2/")
Test missing content and failed navigation tests.
Test missing content and failed navigation tests. This commit adds unit tests outside of the happy path where a url does not exist or the test is looking for conten that doesn't exist on the page. Since testing for missing informaion requires timeouts to be sure, some of these tests take several seconds to execute.
Python
bsd-3-clause
nimbis/django-selenium-testcase,nimbis/django-selenium-testcase
from __future__ import absolute_import from ..testcases import SeleniumLiveTestCase class NavigationTestCase(SeleniumLiveTestCase): test_templates = [ (r'^nav_1/$', 'nav_1.html'), (r'^nav_1/nav_2/$', 'nav_2.html') ] def test_get_page(self): """ Test that you can traverse the page tree. """ self.get_page("/nav_1/") + self.has_title("Navigation 1") + self.title_contains("1") self.should_see("This is nav 1.") self.get_page("/nav_1/nav_2/") self.should_see("This is nav 2.") + def test_get_bad_page(self): + """ Test that /bogus/ is not found. """ + self.get_page("/bogus/") + self.should_see("Not Found") + self.should_see( + "The requested URL /bogus/ was not found on this server.") + + def test_missing_content_with_retry(self): + """ Test retry for missing content, LONG RETRIES! """ + self.get_page("/nav_1/") + self.should_not_see("This is nav 2.") + self.url_should_not_contain("nav_2") + self.has_not_title("Navigation 2") + self.title_does_not_contain("2") + self.assertRaises( + AssertionError, self.click_button, "not_there_dude") + self.not_at_page("/nav_2/") +
Test missing content and failed navigation tests.
## Code Before: from __future__ import absolute_import from ..testcases import SeleniumLiveTestCase class NavigationTestCase(SeleniumLiveTestCase): test_templates = [ (r'^nav_1/$', 'nav_1.html'), (r'^nav_1/nav_2/$', 'nav_2.html') ] def test_get_page(self): """ Test that you can traverse the page tree. """ self.get_page("/nav_1/") self.should_see("This is nav 1.") self.get_page("/nav_1/nav_2/") self.should_see("This is nav 2.") ## Instruction: Test missing content and failed navigation tests. ## Code After: from __future__ import absolute_import from ..testcases import SeleniumLiveTestCase class NavigationTestCase(SeleniumLiveTestCase): test_templates = [ (r'^nav_1/$', 'nav_1.html'), (r'^nav_1/nav_2/$', 'nav_2.html') ] def test_get_page(self): """ Test that you can traverse the page tree. """ self.get_page("/nav_1/") self.has_title("Navigation 1") self.title_contains("1") self.should_see("This is nav 1.") self.get_page("/nav_1/nav_2/") self.should_see("This is nav 2.") def test_get_bad_page(self): """ Test that /bogus/ is not found. """ self.get_page("/bogus/") self.should_see("Not Found") self.should_see( "The requested URL /bogus/ was not found on this server.") def test_missing_content_with_retry(self): """ Test retry for missing content, LONG RETRIES! """ self.get_page("/nav_1/") self.should_not_see("This is nav 2.") self.url_should_not_contain("nav_2") self.has_not_title("Navigation 2") self.title_does_not_contain("2") self.assertRaises( AssertionError, self.click_button, "not_there_dude") self.not_at_page("/nav_2/")
... self.get_page("/nav_1/") self.has_title("Navigation 1") self.title_contains("1") self.should_see("This is nav 1.") ... self.should_see("This is nav 2.") def test_get_bad_page(self): """ Test that /bogus/ is not found. """ self.get_page("/bogus/") self.should_see("Not Found") self.should_see( "The requested URL /bogus/ was not found on this server.") def test_missing_content_with_retry(self): """ Test retry for missing content, LONG RETRIES! """ self.get_page("/nav_1/") self.should_not_see("This is nav 2.") self.url_should_not_contain("nav_2") self.has_not_title("Navigation 2") self.title_does_not_contain("2") self.assertRaises( AssertionError, self.click_button, "not_there_dude") self.not_at_page("/nav_2/") ...
8d0b9da511d55191609ffbd88a8b11afd6ff0367
remedy/radremedy.py
remedy/radremedy.py
from flask import Flask, url_for, request, abort from flask.ext.script import Manager from flask.ext.migrate import Migrate, MigrateCommand from rad.models import db, Resource def create_app(config, models=()): from remedyblueprint import remedy, url_for_other_page app = Flask(__name__) app.config.from_object(config) app.register_blueprint(remedy) # searching configurations app.jinja_env.trim_blocks = True # Register the paging helper method with Jinja2 app.jinja_env.globals['url_for_other_page'] = url_for_other_page db.init_app(app) Migrate(app, db, directory=app.config['MIGRATIONS_DIR']) manager = Manager(app) manager.add_command('db', MigrateCommand) # turning API off for now # from api_manager import init_api_manager # api_manager = init_api_manager(app, db) # map(lambda m: api_manager.create_api(m), models) return app, manager if __name__ == '__main__': app, manager = create_app('config.BaseConfig', (Resource, )) with app.app_context(): manager.run()
from flask import Flask, url_for, request, abort from flask.ext.script import Manager from flask.ext.migrate import Migrate, MigrateCommand from flask.ext.login import current_user from rad.models import db, Resource def create_app(config, models=()): app = Flask(__name__) app.config.from_object(config) from remedyblueprint import remedy, url_for_other_page app.register_blueprint(remedy) from auth.user_auth import auth, login_manager app.register_blueprint(auth) login_manager.init_app(app) # searching configurations app.jinja_env.trim_blocks = True # Register the paging helper method with Jinja2 app.jinja_env.globals['url_for_other_page'] = url_for_other_page app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous() db.init_app(app) Migrate(app, db, directory=app.config['MIGRATIONS_DIR']) manager = Manager(app) manager.add_command('db', MigrateCommand) # turning API off for now # from api_manager import init_api_manager # api_manager = init_api_manager(app, db) # map(lambda m: api_manager.create_api(m), models) return app, manager if __name__ == '__main__': application, manager = create_app('config.BaseConfig', (Resource, )) with application.app_context(): manager.run()
Move around imports and not shadow app
Move around imports and not shadow app
Python
mpl-2.0
radioprotector/radremedy,radioprotector/radremedy,radioprotector/radremedy,AllieDeford/radremedy,AllieDeford/radremedy,radremedy/radremedy,radremedy/radremedy,radremedy/radremedy,AllieDeford/radremedy,radioprotector/radremedy,radremedy/radremedy
from flask import Flask, url_for, request, abort from flask.ext.script import Manager from flask.ext.migrate import Migrate, MigrateCommand + from flask.ext.login import current_user from rad.models import db, Resource def create_app(config, models=()): - from remedyblueprint import remedy, url_for_other_page - app = Flask(__name__) app.config.from_object(config) + from remedyblueprint import remedy, url_for_other_page app.register_blueprint(remedy) + + from auth.user_auth import auth, login_manager + app.register_blueprint(auth) + login_manager.init_app(app) # searching configurations app.jinja_env.trim_blocks = True # Register the paging helper method with Jinja2 app.jinja_env.globals['url_for_other_page'] = url_for_other_page + app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous() db.init_app(app) Migrate(app, db, directory=app.config['MIGRATIONS_DIR']) manager = Manager(app) manager.add_command('db', MigrateCommand) # turning API off for now # from api_manager import init_api_manager # api_manager = init_api_manager(app, db) # map(lambda m: api_manager.create_api(m), models) return app, manager if __name__ == '__main__': - app, manager = create_app('config.BaseConfig', (Resource, )) + application, manager = create_app('config.BaseConfig', (Resource, )) - with app.app_context(): + with application.app_context(): manager.run()
Move around imports and not shadow app
## Code Before: from flask import Flask, url_for, request, abort from flask.ext.script import Manager from flask.ext.migrate import Migrate, MigrateCommand from rad.models import db, Resource def create_app(config, models=()): from remedyblueprint import remedy, url_for_other_page app = Flask(__name__) app.config.from_object(config) app.register_blueprint(remedy) # searching configurations app.jinja_env.trim_blocks = True # Register the paging helper method with Jinja2 app.jinja_env.globals['url_for_other_page'] = url_for_other_page db.init_app(app) Migrate(app, db, directory=app.config['MIGRATIONS_DIR']) manager = Manager(app) manager.add_command('db', MigrateCommand) # turning API off for now # from api_manager import init_api_manager # api_manager = init_api_manager(app, db) # map(lambda m: api_manager.create_api(m), models) return app, manager if __name__ == '__main__': app, manager = create_app('config.BaseConfig', (Resource, )) with app.app_context(): manager.run() ## Instruction: Move around imports and not shadow app ## Code After: from flask import Flask, url_for, request, abort from flask.ext.script import Manager from flask.ext.migrate import Migrate, MigrateCommand from flask.ext.login import current_user from rad.models import db, Resource def create_app(config, models=()): app = Flask(__name__) app.config.from_object(config) from remedyblueprint import remedy, url_for_other_page app.register_blueprint(remedy) from auth.user_auth import auth, login_manager app.register_blueprint(auth) login_manager.init_app(app) # searching configurations app.jinja_env.trim_blocks = True # Register the paging helper method with Jinja2 app.jinja_env.globals['url_for_other_page'] = url_for_other_page app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous() db.init_app(app) Migrate(app, db, directory=app.config['MIGRATIONS_DIR']) manager = Manager(app) manager.add_command('db', MigrateCommand) # turning API off for now # from api_manager import init_api_manager # api_manager = init_api_manager(app, db) # map(lambda m: api_manager.create_api(m), models) return app, manager if __name__ == '__main__': application, manager = create_app('config.BaseConfig', (Resource, )) with application.app_context(): manager.run()
... from flask.ext.migrate import Migrate, MigrateCommand from flask.ext.login import current_user from rad.models import db, Resource ... app = Flask(__name__) ... from remedyblueprint import remedy, url_for_other_page app.register_blueprint(remedy) from auth.user_auth import auth, login_manager app.register_blueprint(auth) login_manager.init_app(app) ... app.jinja_env.globals['url_for_other_page'] = url_for_other_page app.jinja_env.globals['logged_in'] = lambda : not current_user.is_anonymous() ... if __name__ == '__main__': application, manager = create_app('config.BaseConfig', (Resource, )) with application.app_context(): manager.run() ...
ccf60e9e79b8b2db8cbf7918caf23314e8790134
lib/reporter.py
lib/reporter.py
import sys import os name = sys.argv[1] status = sys.stdin.readline() status = status.rstrip(os.linesep) print("<%s>" % name) print("\t<status=\"%s\" />" % status) if status != "SKIP": print("\t<outcome>") for line in sys.stdin: # Escaping, ... ! print(line.rstrip(os.linesep)) print("\t</outcome>") print("</%s>" % name)
import sys import os name = sys.argv[1] status = sys.stdin.readline() status = status.rstrip(os.linesep) print("<%s status=\"%s\">" % (name, status)) print("\t<outcome>") for line in sys.stdin: # Escaping, ... ! print(line.rstrip(os.linesep)) print("\t</outcome>") print("</%s>" % name)
Fix the XML format produced
Fix the XML format produced
Python
apache-2.0
CESNET/secant,CESNET/secant
import sys import os name = sys.argv[1] status = sys.stdin.readline() status = status.rstrip(os.linesep) - print("<%s>" % name) - print("\t<status=\"%s\" />" % status) + print("<%s status=\"%s\">" % (name, status)) - if status != "SKIP": - print("\t<outcome>") + print("\t<outcome>") - for line in sys.stdin: + for line in sys.stdin: - # Escaping, ... ! + # Escaping, ... ! - print(line.rstrip(os.linesep)) + print(line.rstrip(os.linesep)) - print("\t</outcome>") + print("\t</outcome>") print("</%s>" % name)
Fix the XML format produced
## Code Before: import sys import os name = sys.argv[1] status = sys.stdin.readline() status = status.rstrip(os.linesep) print("<%s>" % name) print("\t<status=\"%s\" />" % status) if status != "SKIP": print("\t<outcome>") for line in sys.stdin: # Escaping, ... ! print(line.rstrip(os.linesep)) print("\t</outcome>") print("</%s>" % name) ## Instruction: Fix the XML format produced ## Code After: import sys import os name = sys.argv[1] status = sys.stdin.readline() status = status.rstrip(os.linesep) print("<%s status=\"%s\">" % (name, status)) print("\t<outcome>") for line in sys.stdin: # Escaping, ... ! print(line.rstrip(os.linesep)) print("\t</outcome>") print("</%s>" % name)
... print("<%s status=\"%s\">" % (name, status)) print("\t<outcome>") for line in sys.stdin: # Escaping, ... ! print(line.rstrip(os.linesep)) print("\t</outcome>") ...
2e361627ca94b3a3b1cdd9583d22ca8ff81a0591
rpn/util.py
rpn/util.py
from functools import wraps import subprocess _SELECTIONS = { '+': 'clipboard', '*': 'primary', } def _store_selection(data, selection): with subprocess.Popen(['xclip', '-selection', selection], stdin=subprocess.PIPE) as xclip: xclip.stdin.write(str(data).encode()) def _load_selection(selection): with subprocess.Popen(['xclip', '-selection', selection, '-o'], stdout=PIPE) as xclip: return xclip.stdout.read().decode() class RPNError(Exception): pass def wrap_user_errors(fmt): ''' Ugly hack decorator that converts exceptions to warnings. Passes through RPNErrors. ''' def decorator(f): @wraps(f) def wrapper(*args, **kwargs): try: return f(*args, **kwargs) except RPNError: raise except Exception as e: raise RPNError(fmt.format(*args, **kwargs), e) return wrapper return decorator
from functools import wraps import subprocess _SELECTIONS = { '+': 'clipboard', '*': 'primary', } def _store_selection(data, selection): with subprocess.Popen(['xclip', '-selection', selection], stdin=subprocess.PIPE) as xclip: xclip.stdin.write(str(data).encode()) def _load_selection(selection): with subprocess.Popen(['xclip', '-selection', selection, '-o'], stdout=subprocess.PIPE) as xclip: return xclip.stdout.read().decode() class RPNError(Exception): pass def wrap_user_errors(fmt): ''' Ugly hack decorator that converts exceptions to warnings. Passes through RPNErrors. ''' def decorator(f): @wraps(f) def wrapper(*args, **kwargs): try: return f(*args, **kwargs) except RPNError: raise except Exception as e: raise RPNError(fmt.format(*args, **kwargs), e) return wrapper return decorator
Fix typo resulting in NameError
Fix typo resulting in NameError
Python
isc
pilona/RPN,pilona/RPN
from functools import wraps import subprocess _SELECTIONS = { '+': 'clipboard', '*': 'primary', } def _store_selection(data, selection): with subprocess.Popen(['xclip', '-selection', selection], stdin=subprocess.PIPE) as xclip: xclip.stdin.write(str(data).encode()) def _load_selection(selection): with subprocess.Popen(['xclip', '-selection', selection, - '-o'], stdout=PIPE) as xclip: + '-o'], stdout=subprocess.PIPE) as xclip: return xclip.stdout.read().decode() class RPNError(Exception): pass def wrap_user_errors(fmt): ''' Ugly hack decorator that converts exceptions to warnings. Passes through RPNErrors. ''' def decorator(f): @wraps(f) def wrapper(*args, **kwargs): try: return f(*args, **kwargs) except RPNError: raise except Exception as e: raise RPNError(fmt.format(*args, **kwargs), e) return wrapper return decorator
Fix typo resulting in NameError
## Code Before: from functools import wraps import subprocess _SELECTIONS = { '+': 'clipboard', '*': 'primary', } def _store_selection(data, selection): with subprocess.Popen(['xclip', '-selection', selection], stdin=subprocess.PIPE) as xclip: xclip.stdin.write(str(data).encode()) def _load_selection(selection): with subprocess.Popen(['xclip', '-selection', selection, '-o'], stdout=PIPE) as xclip: return xclip.stdout.read().decode() class RPNError(Exception): pass def wrap_user_errors(fmt): ''' Ugly hack decorator that converts exceptions to warnings. Passes through RPNErrors. ''' def decorator(f): @wraps(f) def wrapper(*args, **kwargs): try: return f(*args, **kwargs) except RPNError: raise except Exception as e: raise RPNError(fmt.format(*args, **kwargs), e) return wrapper return decorator ## Instruction: Fix typo resulting in NameError ## Code After: from functools import wraps import subprocess _SELECTIONS = { '+': 'clipboard', '*': 'primary', } def _store_selection(data, selection): with subprocess.Popen(['xclip', '-selection', selection], stdin=subprocess.PIPE) as xclip: xclip.stdin.write(str(data).encode()) def _load_selection(selection): with subprocess.Popen(['xclip', '-selection', selection, '-o'], stdout=subprocess.PIPE) as xclip: return xclip.stdout.read().decode() class RPNError(Exception): pass def wrap_user_errors(fmt): ''' Ugly hack decorator that converts exceptions to warnings. Passes through RPNErrors. ''' def decorator(f): @wraps(f) def wrapper(*args, **kwargs): try: return f(*args, **kwargs) except RPNError: raise except Exception as e: raise RPNError(fmt.format(*args, **kwargs), e) return wrapper return decorator
# ... existing code ... '-selection', selection, '-o'], stdout=subprocess.PIPE) as xclip: return xclip.stdout.read().decode() # ... rest of the code ...
b4687eb7fda33323cad8d42f9819a3ee223d3822
web/config/local_settings.py
web/config/local_settings.py
import os from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') if os.getenv("MEMCACHE_HOSTS"): CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
import os import json, requests from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"): rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME") r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json() r = map(lambda x: x["primary_ip"] + ":80", r) CLUSTER_SERVERS = [str(x) for x in r] if os.getenv("MEMCACHE_HOSTS"): CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
Add graphite cluster discovery support using rancher
Add graphite cluster discovery support using rancher
Python
apache-2.0
Banno/graphite-setup,Banno/graphite-setup,Banno/graphite-setup
import os + import json, requests from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') + elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"): + rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME") + r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json() + r = map(lambda x: x["primary_ip"] + ":80", r) + CLUSTER_SERVERS = [str(x) for x in r] if os.getenv("MEMCACHE_HOSTS"): CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
Add graphite cluster discovery support using rancher
## Code Before: import os from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') if os.getenv("MEMCACHE_HOSTS"): CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now()) ## Instruction: Add graphite cluster discovery support using rancher ## Code After: import os import json, requests from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"): rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME") r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json() r = map(lambda x: x["primary_ip"] + ":80", r) CLUSTER_SERVERS = [str(x) for x in r] if os.getenv("MEMCACHE_HOSTS"): CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
... import os import json, requests from datetime import datetime ... CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') elif os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME"): rancher_carbonlink_service_url = "http://rancher-metadata/2015-12-19/services/%s/containers" % os.getenv("RANCHER_GRAPHITE_CLUSTER_SERVICE_NAME") r = requests.get(rancher_carbonlink_service_url, headers={"Accept": "application/json"}).json() r = map(lambda x: x["primary_ip"] + ":80", r) CLUSTER_SERVERS = [str(x) for x in r] ...
3af22fd5583ee110f731b9e1ebecba67ebee2bd4
sendwithus/exceptions.py
sendwithus/exceptions.py
class SendwithusError(Exception): """Base class for Sendwithus API errors""" class AuthenticationError(SendwithusError): """API Authentication Failed""" class APIError(SendwithusError): """4xx - Invalid Request (Client error)""" class ServerError(SendwithusError): """5xx - Failed Request (Server error)"""
class SendwithusError(Exception): """Base class for Sendwithus API errors""" def __init__(self, content=None): self.content = content class AuthenticationError(SendwithusError): """API Authentication Failed""" class APIError(SendwithusError): """4xx - Invalid Request (Client error)""" class ServerError(SendwithusError): """5xx - Failed Request (Server error)"""
Add a constructor to SendwithusError that stores content
Add a constructor to SendwithusError that stores content
Python
apache-2.0
sendwithus/sendwithus_python
class SendwithusError(Exception): """Base class for Sendwithus API errors""" + + def __init__(self, content=None): + self.content = content class AuthenticationError(SendwithusError): """API Authentication Failed""" class APIError(SendwithusError): """4xx - Invalid Request (Client error)""" class ServerError(SendwithusError): """5xx - Failed Request (Server error)"""
Add a constructor to SendwithusError that stores content
## Code Before: class SendwithusError(Exception): """Base class for Sendwithus API errors""" class AuthenticationError(SendwithusError): """API Authentication Failed""" class APIError(SendwithusError): """4xx - Invalid Request (Client error)""" class ServerError(SendwithusError): """5xx - Failed Request (Server error)""" ## Instruction: Add a constructor to SendwithusError that stores content ## Code After: class SendwithusError(Exception): """Base class for Sendwithus API errors""" def __init__(self, content=None): self.content = content class AuthenticationError(SendwithusError): """API Authentication Failed""" class APIError(SendwithusError): """4xx - Invalid Request (Client error)""" class ServerError(SendwithusError): """5xx - Failed Request (Server error)"""
... """Base class for Sendwithus API errors""" def __init__(self, content=None): self.content = content ...
62451e8c5b3d93409fa4bcc7ec29827be6253e88
website/registries/utils.py
website/registries/utils.py
REG_CAMPAIGNS = { 'prereg': 'Prereg Challenge', 'registered_report': 'Registered Report Protocol Preregistration', } def get_campaign_schema(campaign): from osf.models import RegistrationSchema if campaign not in REG_CAMPAIGNS: raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys()))) schema_name = REG_CAMPAIGNS[campaign] return RegistrationSchema.objects.get(name=schema_name, schema_version=2) def drafts_for_user(user, campaign=None): from osf.models import DraftRegistration, Node from guardian.shortcuts import get_objects_for_user if not user or user.is_anonymous: return None node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True) if campaign: drafts = DraftRegistration.objects.filter( registration_schema=get_campaign_schema(campaign), approval=None, registered_node=None, deleted__isnull=True, branched_from__in=list(node_qs), initiator=user ) else: drafts = DraftRegistration.objects.filter( approval=None, registered_node=None, deleted__isnull=True, branched_from__in=list(node_qs), initiator=user ) return drafts
REG_CAMPAIGNS = { 'prereg': 'Prereg Challenge', 'registered_report': 'Registered Report Protocol Preregistration', } def get_campaign_schema(campaign): from osf.models import RegistrationSchema if campaign not in REG_CAMPAIGNS: raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys()))) schema_name = REG_CAMPAIGNS[campaign] return RegistrationSchema.objects.get(name=schema_name, schema_version=2) def drafts_for_user(user, campaign=None): from osf.models import DraftRegistration, Node from guardian.shortcuts import get_objects_for_user if not user or user.is_anonymous: return None node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True) drafts = DraftRegistration.objects.filter( approval=None, registered_node=None, deleted__isnull=True, branched_from__in=node_qs, initiator=user ) if campaign: drafts = drafts.filter( registration_schema=get_campaign_schema(campaign), ) return drafts
Speed up draft registrations query.
Speed up draft registrations query.
Python
apache-2.0
baylee-d/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,adlius/osf.io,mattclark/osf.io,felliott/osf.io,saradbowman/osf.io,Johnetordoff/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,felliott/osf.io,CenterForOpenScience/osf.io,cslzchen/osf.io,baylee-d/osf.io,mattclark/osf.io,brianjgeiger/osf.io,mattclark/osf.io,Johnetordoff/osf.io,brianjgeiger/osf.io,brianjgeiger/osf.io,baylee-d/osf.io,adlius/osf.io,cslzchen/osf.io,aaxelb/osf.io,aaxelb/osf.io,felliott/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,mfraezz/osf.io,mfraezz/osf.io,adlius/osf.io,CenterForOpenScience/osf.io,aaxelb/osf.io,aaxelb/osf.io,mfraezz/osf.io,felliott/osf.io,saradbowman/osf.io,CenterForOpenScience/osf.io,adlius/osf.io
REG_CAMPAIGNS = { 'prereg': 'Prereg Challenge', 'registered_report': 'Registered Report Protocol Preregistration', } def get_campaign_schema(campaign): from osf.models import RegistrationSchema if campaign not in REG_CAMPAIGNS: raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys()))) schema_name = REG_CAMPAIGNS[campaign] return RegistrationSchema.objects.get(name=schema_name, schema_version=2) def drafts_for_user(user, campaign=None): from osf.models import DraftRegistration, Node from guardian.shortcuts import get_objects_for_user if not user or user.is_anonymous: return None - node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True) + node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True) + + drafts = DraftRegistration.objects.filter( + approval=None, + registered_node=None, + deleted__isnull=True, + branched_from__in=node_qs, + initiator=user + ) if campaign: - drafts = DraftRegistration.objects.filter( + drafts = drafts.filter( registration_schema=get_campaign_schema(campaign), - approval=None, - registered_node=None, - deleted__isnull=True, - branched_from__in=list(node_qs), - initiator=user ) + - else: - drafts = DraftRegistration.objects.filter( - approval=None, - registered_node=None, - deleted__isnull=True, - branched_from__in=list(node_qs), - initiator=user - ) return drafts
Speed up draft registrations query.
## Code Before: REG_CAMPAIGNS = { 'prereg': 'Prereg Challenge', 'registered_report': 'Registered Report Protocol Preregistration', } def get_campaign_schema(campaign): from osf.models import RegistrationSchema if campaign not in REG_CAMPAIGNS: raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys()))) schema_name = REG_CAMPAIGNS[campaign] return RegistrationSchema.objects.get(name=schema_name, schema_version=2) def drafts_for_user(user, campaign=None): from osf.models import DraftRegistration, Node from guardian.shortcuts import get_objects_for_user if not user or user.is_anonymous: return None node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True) if campaign: drafts = DraftRegistration.objects.filter( registration_schema=get_campaign_schema(campaign), approval=None, registered_node=None, deleted__isnull=True, branched_from__in=list(node_qs), initiator=user ) else: drafts = DraftRegistration.objects.filter( approval=None, registered_node=None, deleted__isnull=True, branched_from__in=list(node_qs), initiator=user ) return drafts ## Instruction: Speed up draft registrations query. ## Code After: REG_CAMPAIGNS = { 'prereg': 'Prereg Challenge', 'registered_report': 'Registered Report Protocol Preregistration', } def get_campaign_schema(campaign): from osf.models import RegistrationSchema if campaign not in REG_CAMPAIGNS: raise ValueError('campaign must be one of: {}'.format(', '.join(REG_CAMPAIGNS.keys()))) schema_name = REG_CAMPAIGNS[campaign] return RegistrationSchema.objects.get(name=schema_name, schema_version=2) def drafts_for_user(user, campaign=None): from osf.models import DraftRegistration, Node from guardian.shortcuts import get_objects_for_user if not user or user.is_anonymous: return None node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True) drafts = DraftRegistration.objects.filter( approval=None, registered_node=None, deleted__isnull=True, branched_from__in=node_qs, initiator=user ) if campaign: drafts = drafts.filter( registration_schema=get_campaign_schema(campaign), ) return drafts
... node_qs = get_objects_for_user(user, 'admin_node', Node, with_superuser=False).exclude(is_deleted=True).values_list('id', flat=True) drafts = DraftRegistration.objects.filter( approval=None, registered_node=None, deleted__isnull=True, branched_from__in=node_qs, initiator=user ) ... if campaign: drafts = drafts.filter( registration_schema=get_campaign_schema(campaign), ) return drafts ...
8a9f707960c3b39488c9bbee6ce7f22c6fbfc853
web/config/local_settings.py
web/config/local_settings.py
import os from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') if os.getenv("MEMCACHE_HOSTS"): CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
import os from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') if os.getenv("MEMCACHE_HOSTS"): MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
Fix memcache hosts setting from env
Fix memcache hosts setting from env Before this fix if one had set OS env vars for both CLUSTER_SERVERS and MEMCACHE_HOSTS the value of later would override the former and the graphite web application fails to show any metrics.
Python
apache-2.0
Banno/graphite-setup,Banno/graphite-setup,Banno/graphite-setup
import os from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') if os.getenv("MEMCACHE_HOSTS"): - CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') + MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
Fix memcache hosts setting from env
## Code Before: import os from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') if os.getenv("MEMCACHE_HOSTS"): CLUSTER_SERVERS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now()) ## Instruction: Fix memcache hosts setting from env ## Code After: import os from datetime import datetime LOG_DIR = '/var/log/graphite' if os.getenv("CARBONLINK_HOSTS"): CARBONLINK_HOSTS = os.getenv("CARBONLINK_HOSTS").split(',') if os.getenv("CLUSTER_SERVERS"): CLUSTER_SERVERS = os.getenv("CLUSTER_SERVERS").split(',') if os.getenv("MEMCACHE_HOSTS"): MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',') if os.getenv("WHISPER_DIR"): WHISPER_DIR = os.getenv("WHISPER_DIR") SECRET_KEY = str(datetime.now())
# ... existing code ... if os.getenv("MEMCACHE_HOSTS"): MEMCACHE_HOSTS = os.getenv("MEMCACHE_HOSTS").split(',') # ... rest of the code ...
26ffa0cdd1389e2a364531cd20e9f37ee1565cce
base/view_utils.py
base/view_utils.py
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger # standard library def paginate(request, objects, page_size=25): paginator = Paginator(objects, page_size) page = request.GET.get('p') try: paginated_objects = paginator.page(page) except PageNotAnInteger: # If page is not an integer, deliver first page. paginated_objects = paginator.page(1) except EmptyPage: # If page is out of range (e.g. 9999), deliver last page of results. paginated_objects = paginator.page(paginator.num_pages) return paginated_objects def clean_query_string(request): clean_query_set = request.GET.copy() clean_query_set = dict( (k, v) for k, v in request.GET.items() if not k.startswith('o') ) try: del clean_query_set['p'] except: pass mstring = [] for key in clean_query_set.keys(): valuelist = request.GET.getlist(key) mstring.extend(['%s=%s' % (key, val) for val in valuelist]) return '&'.join(mstring)
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger # standard library def paginate(request, objects, page_size=25): paginator = Paginator(objects, page_size) page = request.GET.get('p') try: paginated_objects = paginator.page(page) except PageNotAnInteger: # If page is not an integer, deliver first page. paginated_objects = paginator.page(1) except EmptyPage: # If page is out of range (e.g. 9999), deliver last page of results. paginated_objects = paginator.page(paginator.num_pages) return paginated_objects def clean_query_string(request): clean_query_set = request.GET.copy() clean_query_set = dict( (k, v) for k, v in request.GET.items() if k != 'o' ) try: del clean_query_set['p'] except: pass mstring = [] for key in clean_query_set.keys(): valuelist = request.GET.getlist(key) mstring.extend(['%s=%s' % (key, val) for val in valuelist]) return '&'.join(mstring)
Use 'o' as the order by parameter in clean_query_string
Use 'o' as the order by parameter in clean_query_string
Python
mit
magnet-cl/django-project-template-py3,Angoreher/xcero,Angoreher/xcero,magnet-cl/django-project-template-py3,magnet-cl/django-project-template-py3,magnet-cl/django-project-template-py3,Angoreher/xcero,Angoreher/xcero
from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger # standard library def paginate(request, objects, page_size=25): paginator = Paginator(objects, page_size) page = request.GET.get('p') try: paginated_objects = paginator.page(page) except PageNotAnInteger: # If page is not an integer, deliver first page. paginated_objects = paginator.page(1) except EmptyPage: # If page is out of range (e.g. 9999), deliver last page of results. paginated_objects = paginator.page(paginator.num_pages) return paginated_objects def clean_query_string(request): clean_query_set = request.GET.copy() clean_query_set = dict( - (k, v) for k, v in request.GET.items() if not k.startswith('o') + (k, v) for k, v in request.GET.items() if k != 'o' ) try: del clean_query_set['p'] except: pass mstring = [] for key in clean_query_set.keys(): valuelist = request.GET.getlist(key) mstring.extend(['%s=%s' % (key, val) for val in valuelist]) return '&'.join(mstring)
Use 'o' as the order by parameter in clean_query_string
## Code Before: from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger # standard library def paginate(request, objects, page_size=25): paginator = Paginator(objects, page_size) page = request.GET.get('p') try: paginated_objects = paginator.page(page) except PageNotAnInteger: # If page is not an integer, deliver first page. paginated_objects = paginator.page(1) except EmptyPage: # If page is out of range (e.g. 9999), deliver last page of results. paginated_objects = paginator.page(paginator.num_pages) return paginated_objects def clean_query_string(request): clean_query_set = request.GET.copy() clean_query_set = dict( (k, v) for k, v in request.GET.items() if not k.startswith('o') ) try: del clean_query_set['p'] except: pass mstring = [] for key in clean_query_set.keys(): valuelist = request.GET.getlist(key) mstring.extend(['%s=%s' % (key, val) for val in valuelist]) return '&'.join(mstring) ## Instruction: Use 'o' as the order by parameter in clean_query_string ## Code After: from django.core.paginator import Paginator, EmptyPage, PageNotAnInteger # standard library def paginate(request, objects, page_size=25): paginator = Paginator(objects, page_size) page = request.GET.get('p') try: paginated_objects = paginator.page(page) except PageNotAnInteger: # If page is not an integer, deliver first page. paginated_objects = paginator.page(1) except EmptyPage: # If page is out of range (e.g. 9999), deliver last page of results. paginated_objects = paginator.page(paginator.num_pages) return paginated_objects def clean_query_string(request): clean_query_set = request.GET.copy() clean_query_set = dict( (k, v) for k, v in request.GET.items() if k != 'o' ) try: del clean_query_set['p'] except: pass mstring = [] for key in clean_query_set.keys(): valuelist = request.GET.getlist(key) mstring.extend(['%s=%s' % (key, val) for val in valuelist]) return '&'.join(mstring)
... clean_query_set = dict( (k, v) for k, v in request.GET.items() if k != 'o' ) ...
3af265ab0740378267a3c3e9cc85bb21468bf2e0
engine/cli.py
engine/cli.py
from engine.event import * from engine.action import * from engine.code import * from engine.player import * from engine.round import * from engine.team import * def processInput(): userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n") if 'f' in userText: jailCode = input("enter jail code: ") Action.fleePlayerWithCode(jailCode) Stats.printPlayersDetailed() elif 's' in userText: mobile = input("enter mobile: ") code = input("enter code: ") Action.handleSms(mobile, code) Stats.printPlayersDetailed() elif 'a' in userText: name = input("enter name: ") mobile = input("enter mobile: ") #email = input("enter email: ") Action.addPlayer(name, mobile, "") Stats.printPlayersDetailed() elif 'w' in userText: hash = input("enter player hash: ") code = input("enter code: ") Action.handleWeb(hash, code) Stats.printPlayersDetailed() elif 't' in userText: name = input("enter player name: ") team = input("enter team name: ") Action.addPlayerToTeam(name, team) Stats.printPlayersDetailed() elif 'p' in userText: Stats.printStats() elif 'c' in userText: name = input("enter name: ") message = input("enter text: ") playerId = Player._getIdByName(name) Action.sayToMyTeam(playerId, message)
from engine.action import Action, Stats from engine.player import Player def processInput(): userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n") if 'f' in userText: jailCode = input("enter jail code: ") Action.fleePlayerWithCode(jailCode) Stats.printPlayersDetailed() elif 's' in userText: mobile = input("enter mobile: ") code = input("enter code: ") Action.handleSms(mobile, code) Stats.printPlayersDetailed() elif 'a' in userText: name = input("enter name: ") mobile = input("enter mobile: ") #email = input("enter email: ") Action.addPlayer(name, mobile, "") Stats.printPlayersDetailed() elif 'w' in userText: hash = input("enter player hash: ") code = input("enter code: ") Action.handleWeb(hash, code) Stats.printPlayersDetailed() elif 't' in userText: name = input("enter player name: ") team = input("enter team name: ") Action.addPlayerToTeam(name, team) Stats.printPlayersDetailed() elif 'p' in userText: Stats.printStats() elif 'c' in userText: name = input("enter name: ") message = input("enter text: ") playerId = Player._getIdByName(name) Action.sayToMyTeam(playerId, message)
Remove a few unnecessary imports
Remove a few unnecessary imports
Python
bsd-2-clause
mahfiaz/spotter_irl,mahfiaz/spotter_irl,mahfiaz/spotter_irl
- from engine.event import * - from engine.action import * + from engine.action import Action, Stats - from engine.code import * - from engine.player import * + from engine.player import Player - from engine.round import * - from engine.team import * - def processInput(): userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n") if 'f' in userText: jailCode = input("enter jail code: ") Action.fleePlayerWithCode(jailCode) Stats.printPlayersDetailed() elif 's' in userText: mobile = input("enter mobile: ") code = input("enter code: ") Action.handleSms(mobile, code) Stats.printPlayersDetailed() elif 'a' in userText: name = input("enter name: ") mobile = input("enter mobile: ") #email = input("enter email: ") Action.addPlayer(name, mobile, "") Stats.printPlayersDetailed() elif 'w' in userText: hash = input("enter player hash: ") code = input("enter code: ") Action.handleWeb(hash, code) Stats.printPlayersDetailed() elif 't' in userText: name = input("enter player name: ") team = input("enter team name: ") Action.addPlayerToTeam(name, team) Stats.printPlayersDetailed() elif 'p' in userText: Stats.printStats() elif 'c' in userText: name = input("enter name: ") message = input("enter text: ") playerId = Player._getIdByName(name) Action.sayToMyTeam(playerId, message) -
Remove a few unnecessary imports
## Code Before: from engine.event import * from engine.action import * from engine.code import * from engine.player import * from engine.round import * from engine.team import * def processInput(): userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n") if 'f' in userText: jailCode = input("enter jail code: ") Action.fleePlayerWithCode(jailCode) Stats.printPlayersDetailed() elif 's' in userText: mobile = input("enter mobile: ") code = input("enter code: ") Action.handleSms(mobile, code) Stats.printPlayersDetailed() elif 'a' in userText: name = input("enter name: ") mobile = input("enter mobile: ") #email = input("enter email: ") Action.addPlayer(name, mobile, "") Stats.printPlayersDetailed() elif 'w' in userText: hash = input("enter player hash: ") code = input("enter code: ") Action.handleWeb(hash, code) Stats.printPlayersDetailed() elif 't' in userText: name = input("enter player name: ") team = input("enter team name: ") Action.addPlayerToTeam(name, team) Stats.printPlayersDetailed() elif 'p' in userText: Stats.printStats() elif 'c' in userText: name = input("enter name: ") message = input("enter text: ") playerId = Player._getIdByName(name) Action.sayToMyTeam(playerId, message) ## Instruction: Remove a few unnecessary imports ## Code After: from engine.action import Action, Stats from engine.player import Player def processInput(): userText = input("Enter command [Add player] [Team player] [Spot] [Web spot] [Flee jail] [Print] [teamChat]: \n") if 'f' in userText: jailCode = input("enter jail code: ") Action.fleePlayerWithCode(jailCode) Stats.printPlayersDetailed() elif 's' in userText: mobile = input("enter mobile: ") code = input("enter code: ") Action.handleSms(mobile, code) Stats.printPlayersDetailed() elif 'a' in userText: name = input("enter name: ") mobile = input("enter mobile: ") #email = input("enter email: ") Action.addPlayer(name, mobile, "") Stats.printPlayersDetailed() elif 'w' in userText: hash = input("enter player hash: ") code = input("enter code: ") Action.handleWeb(hash, code) Stats.printPlayersDetailed() elif 't' in userText: name = input("enter player name: ") team = input("enter team name: ") Action.addPlayerToTeam(name, team) Stats.printPlayersDetailed() elif 'p' in userText: Stats.printStats() elif 'c' in userText: name = input("enter name: ") message = input("enter text: ") playerId = Player._getIdByName(name) Action.sayToMyTeam(playerId, message)
// ... existing code ... from engine.action import Action, Stats from engine.player import Player // ... modified code ... Action.sayToMyTeam(playerId, message) // ... rest of the code ...
59a08fff34f095f601ced76cd7b2e27665824146
setup.py
setup.py
from distutils.core import setup setup(name='webracer', version='0.2.0', description='Comprehensive web application testing library', author='Oleg Pudeyev', author_email='[email protected]', url='http://github.com/p/webracer', packages=['webracer', 'webracer.utils'], data_files=['LICENSE', 'README.rst'], )
from distutils.core import setup import os.path PACKAGE = "webracer" setup(name=PACKAGE, version='0.2.0', description='Comprehensive web application testing library', author='Oleg Pudeyev', author_email='[email protected]', url='http://github.com/p/webracer', packages=['webracer', 'webracer.utils'], data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))], )
Put license and readme into share/doc/webracer rather than installation root
Put license and readme into share/doc/webracer rather than installation root
Python
bsd-2-clause
p/webracer
from distutils.core import setup + import os.path - setup(name='webracer', + PACKAGE = "webracer" + + setup(name=PACKAGE, version='0.2.0', description='Comprehensive web application testing library', author='Oleg Pudeyev', author_email='[email protected]', url='http://github.com/p/webracer', packages=['webracer', 'webracer.utils'], - data_files=['LICENSE', 'README.rst'], + data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))], )
Put license and readme into share/doc/webracer rather than installation root
## Code Before: from distutils.core import setup setup(name='webracer', version='0.2.0', description='Comprehensive web application testing library', author='Oleg Pudeyev', author_email='[email protected]', url='http://github.com/p/webracer', packages=['webracer', 'webracer.utils'], data_files=['LICENSE', 'README.rst'], ) ## Instruction: Put license and readme into share/doc/webracer rather than installation root ## Code After: from distutils.core import setup import os.path PACKAGE = "webracer" setup(name=PACKAGE, version='0.2.0', description='Comprehensive web application testing library', author='Oleg Pudeyev', author_email='[email protected]', url='http://github.com/p/webracer', packages=['webracer', 'webracer.utils'], data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))], )
// ... existing code ... from distutils.core import setup import os.path PACKAGE = "webracer" setup(name=PACKAGE, version='0.2.0', // ... modified code ... packages=['webracer', 'webracer.utils'], data_files=[(os.path.join('share', 'doc', PACKAGE), ('LICENSE', 'README.rst'))], ) // ... rest of the code ...
c65ed9ec976c440b46dedc514daf883bba940282
myElsClient.py
myElsClient.py
import requests class myElsClient: """A class that implements a Python interface to api.elsevier.com""" # local variables __base_url = "http://api.elsevier.com/" def __init__(self, apiKey): """Instantiates a client with a given API Key.""" self.apiKey = apiKey def getBaseURL(self): return self.__base_url
import requests class myElsClient: """A class that implements a Python interface to api.elsevier.com""" # local variables __base_url = "http://api.elsevier.com/" # constructors def __init__(self, apiKey): """Instantiates a client with a given API Key.""" self.apiKey = apiKey # configuration functions """Sets an institutional token for customer authentication""" def setInstToken(self, instToken): self.instToken = instToken # utility access functions def getBaseURL(self): """Returns the base URL currently configured for Elsevier's APIs""" return self.__base_url
Add ability to set insttoken
Add ability to set insttoken
Python
bsd-3-clause
ElsevierDev/elsapy
import requests class myElsClient: """A class that implements a Python interface to api.elsevier.com""" # local variables __base_url = "http://api.elsevier.com/" - + + # constructors def __init__(self, apiKey): """Instantiates a client with a given API Key.""" self.apiKey = apiKey + # configuration functions + """Sets an institutional token for customer authentication""" + def setInstToken(self, instToken): + self.instToken = instToken + + # utility access functions def getBaseURL(self): + """Returns the base URL currently configured for Elsevier's APIs""" return self.__base_url
Add ability to set insttoken
## Code Before: import requests class myElsClient: """A class that implements a Python interface to api.elsevier.com""" # local variables __base_url = "http://api.elsevier.com/" def __init__(self, apiKey): """Instantiates a client with a given API Key.""" self.apiKey = apiKey def getBaseURL(self): return self.__base_url ## Instruction: Add ability to set insttoken ## Code After: import requests class myElsClient: """A class that implements a Python interface to api.elsevier.com""" # local variables __base_url = "http://api.elsevier.com/" # constructors def __init__(self, apiKey): """Instantiates a client with a given API Key.""" self.apiKey = apiKey # configuration functions """Sets an institutional token for customer authentication""" def setInstToken(self, instToken): self.instToken = instToken # utility access functions def getBaseURL(self): """Returns the base URL currently configured for Elsevier's APIs""" return self.__base_url
# ... existing code ... __base_url = "http://api.elsevier.com/" # constructors def __init__(self, apiKey): # ... modified code ... # configuration functions """Sets an institutional token for customer authentication""" def setInstToken(self, instToken): self.instToken = instToken # utility access functions def getBaseURL(self): """Returns the base URL currently configured for Elsevier's APIs""" return self.__base_url # ... rest of the code ...
97894a171d1831838da28b757aabb352bc5ecfd9
patches/sitecustomize.py
patches/sitecustomize.py
import torch import os _HOST_FILE = "/etc/hosts" kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN") if kaggle_proxy_token: from google.auth import credentials from google.cloud import bigquery from google.cloud.bigquery._http import Connection host_entry = os.getenv("KAGGLE_DATA_PROXY_HOST_ENTRY") if host_entry: host_entry = host_entry.strip() with open(_HOST_FILE, "r") as host_file: for line in host_file: if line.strip() == host_entry: break else: with open(_HOST_FILE, "a") as host_file_append: host_file_append.write("\n" # Just in case it wasn't newline terminated. + host_entry + "\n") Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL") Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token bq_client = bigquery.Client bigquery.Client = lambda *args, **kwargs: bq_client( *args, credentials=credentials.AnonymousCredentials(), project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"), **kwargs) credentials.AnonymousCredentials.refresh = lambda *args: None
import torch import os kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN") if kaggle_proxy_token: from google.auth import credentials from google.cloud import bigquery from google.cloud.bigquery._http import Connection Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL") Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token bq_client = bigquery.Client bigquery.Client = lambda *args, **kwargs: bq_client( *args, credentials=credentials.AnonymousCredentials(), project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"), **kwargs) credentials.AnonymousCredentials.refresh = lambda *args: None
Revert "Support adding an /etc/host entry for the data proxy, if asked."
Revert "Support adding an /etc/host entry for the data proxy, if asked." This reverts commit 062f975d92c5795feb530e3ea1914d3c7dd3a96b. There is no more need for this support in the docker image. It is fully externally handled through docker run's `--add-host` feature.
Python
apache-2.0
Kaggle/docker-python,Kaggle/docker-python
import torch import os - - _HOST_FILE = "/etc/hosts" kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN") if kaggle_proxy_token: from google.auth import credentials from google.cloud import bigquery from google.cloud.bigquery._http import Connection - - host_entry = os.getenv("KAGGLE_DATA_PROXY_HOST_ENTRY") - if host_entry: - host_entry = host_entry.strip() - with open(_HOST_FILE, "r") as host_file: - for line in host_file: - if line.strip() == host_entry: - break - else: - with open(_HOST_FILE, "a") as host_file_append: - host_file_append.write("\n" # Just in case it wasn't newline terminated. - + host_entry - + "\n") Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL") Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token bq_client = bigquery.Client bigquery.Client = lambda *args, **kwargs: bq_client( *args, credentials=credentials.AnonymousCredentials(), project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"), **kwargs) credentials.AnonymousCredentials.refresh = lambda *args: None
Revert "Support adding an /etc/host entry for the data proxy, if asked."
## Code Before: import torch import os _HOST_FILE = "/etc/hosts" kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN") if kaggle_proxy_token: from google.auth import credentials from google.cloud import bigquery from google.cloud.bigquery._http import Connection host_entry = os.getenv("KAGGLE_DATA_PROXY_HOST_ENTRY") if host_entry: host_entry = host_entry.strip() with open(_HOST_FILE, "r") as host_file: for line in host_file: if line.strip() == host_entry: break else: with open(_HOST_FILE, "a") as host_file_append: host_file_append.write("\n" # Just in case it wasn't newline terminated. + host_entry + "\n") Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL") Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token bq_client = bigquery.Client bigquery.Client = lambda *args, **kwargs: bq_client( *args, credentials=credentials.AnonymousCredentials(), project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"), **kwargs) credentials.AnonymousCredentials.refresh = lambda *args: None ## Instruction: Revert "Support adding an /etc/host entry for the data proxy, if asked." ## Code After: import torch import os kaggle_proxy_token = os.getenv("KAGGLE_DATA_PROXY_TOKEN") if kaggle_proxy_token: from google.auth import credentials from google.cloud import bigquery from google.cloud.bigquery._http import Connection Connection.API_BASE_URL = os.getenv("KAGGLE_DATA_PROXY_URL") Connection._EXTRA_HEADERS["X-KAGGLE-PROXY-DATA"] = kaggle_proxy_token bq_client = bigquery.Client bigquery.Client = lambda *args, **kwargs: bq_client( *args, credentials=credentials.AnonymousCredentials(), project=os.getenv("KAGGLE_DATA_PROXY_PROJECT"), **kwargs) credentials.AnonymousCredentials.refresh = lambda *args: None
// ... existing code ... import os // ... modified code ... from google.cloud.bigquery._http import Connection // ... rest of the code ...
2b43ab4eb41e305c5bdadf5c338e134e5569249d
tests/conftest.py
tests/conftest.py
import pytest import os import tarfile BASEDIR = os.path.dirname(__file__) @pytest.fixture(autouse=False) def set_up(tmpdir): # print BASEDIR tmpdir.chdir() tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz")) tar.extractall() tar.close() os.chdir('MockRepos') print('In directory ' + os.getcwd()) # does not need teardown, since tmpdir directories get autodeleted
import pytest import os import tarfile BASEDIR = os.path.dirname(__file__) @pytest.fixture(autouse=False) def set_up(tmpdir): # print BASEDIR tmpdir.chdir() tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz")) tar.extractall() tar.close() os.chdir('MockRepos') print('In directory ' + os.getcwd()) # does not need teardown, since tmpdir directories get autodeleted def session_teardown(): """Tear down testing session""" print('Tearing down testing session') os.chdir(BASEDIR) # so html coverage report lands in correct directory @pytest.fixture(scope='session', autouse=True) def session_setup(request): """Set up testing session""" print('Setting up testing session') request.addfinalizer(session_teardown)
Add session setup and teardown fixtures.
Add session setup and teardown fixtures.
Python
mit
bilderbuchi/ofStateManager
import pytest import os import tarfile BASEDIR = os.path.dirname(__file__) @pytest.fixture(autouse=False) def set_up(tmpdir): # print BASEDIR tmpdir.chdir() tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz")) tar.extractall() tar.close() os.chdir('MockRepos') print('In directory ' + os.getcwd()) # does not need teardown, since tmpdir directories get autodeleted + + def session_teardown(): + """Tear down testing session""" + print('Tearing down testing session') + os.chdir(BASEDIR) # so html coverage report lands in correct directory + + + @pytest.fixture(scope='session', autouse=True) + def session_setup(request): + """Set up testing session""" + print('Setting up testing session') + + request.addfinalizer(session_teardown) +
Add session setup and teardown fixtures.
## Code Before: import pytest import os import tarfile BASEDIR = os.path.dirname(__file__) @pytest.fixture(autouse=False) def set_up(tmpdir): # print BASEDIR tmpdir.chdir() tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz")) tar.extractall() tar.close() os.chdir('MockRepos') print('In directory ' + os.getcwd()) # does not need teardown, since tmpdir directories get autodeleted ## Instruction: Add session setup and teardown fixtures. ## Code After: import pytest import os import tarfile BASEDIR = os.path.dirname(__file__) @pytest.fixture(autouse=False) def set_up(tmpdir): # print BASEDIR tmpdir.chdir() tar = tarfile.open(os.path.join(BASEDIR, "MockRepos.tar.gz")) tar.extractall() tar.close() os.chdir('MockRepos') print('In directory ' + os.getcwd()) # does not need teardown, since tmpdir directories get autodeleted def session_teardown(): """Tear down testing session""" print('Tearing down testing session') os.chdir(BASEDIR) # so html coverage report lands in correct directory @pytest.fixture(scope='session', autouse=True) def session_setup(request): """Set up testing session""" print('Setting up testing session') request.addfinalizer(session_teardown)
... # does not need teardown, since tmpdir directories get autodeleted def session_teardown(): """Tear down testing session""" print('Tearing down testing session') os.chdir(BASEDIR) # so html coverage report lands in correct directory @pytest.fixture(scope='session', autouse=True) def session_setup(request): """Set up testing session""" print('Setting up testing session') request.addfinalizer(session_teardown) ...
efaa172668b8961734fa8a10650dc3191b4a7348
website/project/metadata/authorizers/__init__.py
website/project/metadata/authorizers/__init__.py
import json import os import logging logger = logging.getLogger(__name__) HERE = os.path.dirname(os.path.realpath(__file__)) groups = json.load(open('{0}/defaults.json'.format(HERE))) try: fp = open('{0}/local.json'.format(HERE)) except IOError: logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.') for group, members in json.load(fp).iteritems(): if group not in groups: groups[group] = members else: groups[group] = set(groups[group]) | set(members) def members_for(group): global_members = set(groups['global']) return global_members | set(groups.get(group, []))
import json import os import logging logger = logging.getLogger(__name__) HERE = os.path.dirname(os.path.realpath(__file__)) groups = json.load(open('{0}/defaults.json'.format(HERE))) fp = None try: fp = open('{0}/local.json'.format(HERE)) except IOError: logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.') if fp: for group, members in json.load(fp).iteritems(): if group not in groups: groups[group] = members else: groups[group] = set(groups[group]) | set(members) def members_for(group): global_members = set(groups['global']) return global_members | set(groups.get(group, []))
Allow local.json to be missing
Allow local.json to be missing
Python
apache-2.0
kch8qx/osf.io,acshi/osf.io,binoculars/osf.io,abought/osf.io,mluo613/osf.io,cslzchen/osf.io,chrisseto/osf.io,ticklemepierce/osf.io,cslzchen/osf.io,brianjgeiger/osf.io,kwierman/osf.io,brandonPurvis/osf.io,icereval/osf.io,TomBaxter/osf.io,doublebits/osf.io,mluke93/osf.io,wearpants/osf.io,alexschiller/osf.io,billyhunt/osf.io,danielneis/osf.io,rdhyee/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,Ghalko/osf.io,mattclark/osf.io,rdhyee/osf.io,kwierman/osf.io,cwisecarver/osf.io,mluke93/osf.io,Johnetordoff/osf.io,GageGaskins/osf.io,abought/osf.io,RomanZWang/osf.io,acshi/osf.io,alexschiller/osf.io,cslzchen/osf.io,GageGaskins/osf.io,SSJohns/osf.io,KAsante95/osf.io,hmoco/osf.io,saradbowman/osf.io,adlius/osf.io,zamattiac/osf.io,binoculars/osf.io,monikagrabowska/osf.io,CenterForOpenScience/osf.io,DanielSBrown/osf.io,emetsger/osf.io,SSJohns/osf.io,brianjgeiger/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io,samanehsan/osf.io,KAsante95/osf.io,emetsger/osf.io,TomBaxter/osf.io,monikagrabowska/osf.io,danielneis/osf.io,samanehsan/osf.io,mfraezz/osf.io,chennan47/osf.io,kch8qx/osf.io,GageGaskins/osf.io,kwierman/osf.io,asanfilippo7/osf.io,caneruguz/osf.io,acshi/osf.io,Nesiehr/osf.io,caseyrollins/osf.io,GageGaskins/osf.io,doublebits/osf.io,sloria/osf.io,adlius/osf.io,mluke93/osf.io,samanehsan/osf.io,samchrisinger/osf.io,jnayak1/osf.io,jnayak1/osf.io,samchrisinger/osf.io,TomHeatwole/osf.io,binoculars/osf.io,samchrisinger/osf.io,billyhunt/osf.io,crcresearch/osf.io,Johnetordoff/osf.io,doublebits/osf.io,RomanZWang/osf.io,KAsante95/osf.io,icereval/osf.io,amyshi188/osf.io,aaxelb/osf.io,leb2dg/osf.io,zamattiac/osf.io,monikagrabowska/osf.io,zamattiac/osf.io,chrisseto/osf.io,alexschiller/osf.io,TomHeatwole/osf.io,brandonPurvis/osf.io,zachjanicki/osf.io,alexschiller/osf.io,mluo613/osf.io,TomBaxter/osf.io,hmoco/osf.io,chrisseto/osf.io,billyhunt/osf.io,cwisecarver/osf.io,mluo613/osf.io,emetsger/osf.io,Nesiehr/osf.io,Ghalko/osf.io,HalcyonChimera/osf.io,adlius/osf.io,KAsante95/osf.io,felliott/osf.io,aaxelb/osf.io,caneruguz/osf.io,ticklemepierce/osf.io,leb2dg/osf.io,felliott/osf.io,caseyrollins/osf.io,danielneis/osf.io,cslzchen/osf.io,chennan47/osf.io,Nesiehr/osf.io,asanfilippo7/osf.io,crcresearch/osf.io,rdhyee/osf.io,baylee-d/osf.io,alexschiller/osf.io,samanehsan/osf.io,jnayak1/osf.io,danielneis/osf.io,amyshi188/osf.io,DanielSBrown/osf.io,HalcyonChimera/osf.io,wearpants/osf.io,mluo613/osf.io,baylee-d/osf.io,crcresearch/osf.io,mluo613/osf.io,aaxelb/osf.io,laurenrevere/osf.io,erinspace/osf.io,kch8qx/osf.io,chrisseto/osf.io,chennan47/osf.io,doublebits/osf.io,monikagrabowska/osf.io,pattisdr/osf.io,laurenrevere/osf.io,caseyrollins/osf.io,icereval/osf.io,abought/osf.io,emetsger/osf.io,TomHeatwole/osf.io,brianjgeiger/osf.io,RomanZWang/osf.io,brianjgeiger/osf.io,erinspace/osf.io,caneruguz/osf.io,amyshi188/osf.io,RomanZWang/osf.io,asanfilippo7/osf.io,acshi/osf.io,hmoco/osf.io,acshi/osf.io,CenterForOpenScience/osf.io,wearpants/osf.io,zamattiac/osf.io,billyhunt/osf.io,kwierman/osf.io,amyshi188/osf.io,rdhyee/osf.io,Johnetordoff/osf.io,erinspace/osf.io,zachjanicki/osf.io,kch8qx/osf.io,doublebits/osf.io,mfraezz/osf.io,GageGaskins/osf.io,brandonPurvis/osf.io,CenterForOpenScience/osf.io,Johnetordoff/osf.io,laurenrevere/osf.io,samchrisinger/osf.io,mfraezz/osf.io,hmoco/osf.io,RomanZWang/osf.io,wearpants/osf.io,brandonPurvis/osf.io,sloria/osf.io,CenterForOpenScience/osf.io,KAsante95/osf.io,mattclark/osf.io,leb2dg/osf.io,DanielSBrown/osf.io,pattisdr/osf.io,jnayak1/osf.io,mluke93/osf.io,asanfilippo7/osf.io,adlius/osf.io,cwisecarver/osf.io,mfraezz/osf.io,Nesiehr/osf.io,caneruguz/osf.io,zachjanicki/osf.io,SSJohns/osf.io,felliott/osf.io,TomHeatwole/osf.io,mattclark/osf.io,Ghalko/osf.io,cwisecarver/osf.io,ticklemepierce/osf.io,abought/osf.io,ticklemepierce/osf.io,sloria/osf.io,billyhunt/osf.io,felliott/osf.io,SSJohns/osf.io,Ghalko/osf.io,HalcyonChimera/osf.io,pattisdr/osf.io,baylee-d/osf.io,aaxelb/osf.io,saradbowman/osf.io,monikagrabowska/osf.io,kch8qx/osf.io,HalcyonChimera/osf.io
import json import os import logging logger = logging.getLogger(__name__) HERE = os.path.dirname(os.path.realpath(__file__)) groups = json.load(open('{0}/defaults.json'.format(HERE))) + fp = None try: fp = open('{0}/local.json'.format(HERE)) except IOError: logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.') + if fp: - for group, members in json.load(fp).iteritems(): + for group, members in json.load(fp).iteritems(): - if group not in groups: + if group not in groups: - groups[group] = members + groups[group] = members - else: + else: - groups[group] = set(groups[group]) | set(members) + groups[group] = set(groups[group]) | set(members) def members_for(group): global_members = set(groups['global']) return global_members | set(groups.get(group, []))
Allow local.json to be missing
## Code Before: import json import os import logging logger = logging.getLogger(__name__) HERE = os.path.dirname(os.path.realpath(__file__)) groups = json.load(open('{0}/defaults.json'.format(HERE))) try: fp = open('{0}/local.json'.format(HERE)) except IOError: logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.') for group, members in json.load(fp).iteritems(): if group not in groups: groups[group] = members else: groups[group] = set(groups[group]) | set(members) def members_for(group): global_members = set(groups['global']) return global_members | set(groups.get(group, [])) ## Instruction: Allow local.json to be missing ## Code After: import json import os import logging logger = logging.getLogger(__name__) HERE = os.path.dirname(os.path.realpath(__file__)) groups = json.load(open('{0}/defaults.json'.format(HERE))) fp = None try: fp = open('{0}/local.json'.format(HERE)) except IOError: logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.') if fp: for group, members in json.load(fp).iteritems(): if group not in groups: groups[group] = members else: groups[group] = set(groups[group]) | set(members) def members_for(group): global_members = set(groups['global']) return global_members | set(groups.get(group, []))
... groups = json.load(open('{0}/defaults.json'.format(HERE))) fp = None try: ... logger.info('No local.json found to populate lists of DraftRegistrationApproval authorizers.') if fp: for group, members in json.load(fp).iteritems(): if group not in groups: groups[group] = members else: groups[group] = set(groups[group]) | set(members) ...
eff3195097e9599b87f5cec9bbae744b91ae16cf
buses/utils.py
buses/utils.py
import re def minify(template_source): template_source = re.sub(r'(\n *)+', '\n', template_source) template_source = re.sub(r'({%.+%})\n+', r'\1', template_source) return template_source
import re from haystack.utils import default_get_identifier def minify(template_source): template_source = re.sub(r'(\n *)+', '\n', template_source) template_source = re.sub(r'({%.+%})\n+', r'\1', template_source) return template_source def get_identifier(obj_or_string): if isinstance(obj_or_string, basestring): return obj_or_string return default_get_identifier(obj_or_string)
Add custom Hastack get_identifier function
Add custom Hastack get_identifier function
Python
mpl-2.0
jclgoodwin/bustimes.org.uk,jclgoodwin/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,stev-0/bustimes.org.uk,jclgoodwin/bustimes.org.uk,jclgoodwin/bustimes.org.uk
import re + from haystack.utils import default_get_identifier def minify(template_source): template_source = re.sub(r'(\n *)+', '\n', template_source) template_source = re.sub(r'({%.+%})\n+', r'\1', template_source) return template_source + def get_identifier(obj_or_string): + if isinstance(obj_or_string, basestring): + return obj_or_string + return default_get_identifier(obj_or_string) +
Add custom Hastack get_identifier function
## Code Before: import re def minify(template_source): template_source = re.sub(r'(\n *)+', '\n', template_source) template_source = re.sub(r'({%.+%})\n+', r'\1', template_source) return template_source ## Instruction: Add custom Hastack get_identifier function ## Code After: import re from haystack.utils import default_get_identifier def minify(template_source): template_source = re.sub(r'(\n *)+', '\n', template_source) template_source = re.sub(r'({%.+%})\n+', r'\1', template_source) return template_source def get_identifier(obj_or_string): if isinstance(obj_or_string, basestring): return obj_or_string return default_get_identifier(obj_or_string)
# ... existing code ... import re from haystack.utils import default_get_identifier # ... modified code ... return template_source def get_identifier(obj_or_string): if isinstance(obj_or_string, basestring): return obj_or_string return default_get_identifier(obj_or_string) # ... rest of the code ...
afc94c1a1ebf14dbb393234233055915132a9fb8
django_ethereum_events/apps.py
django_ethereum_events/apps.py
from django.apps import AppConfig from django.conf import settings class EthereumEventsConfig(AppConfig): name = 'django_ethereum_events' def ready(self): super(EthereumEventsConfig, self).ready() app.config_from_object('django.conf:settings') app.autodiscover_tasks(lambda: settings.INSTALLED_APPS, force=True)
from django.apps import AppConfig from django.conf import settings class EthereumEventsConfig(AppConfig): name = 'django_ethereum_events'
Fix for the previous commit (Celery app removal)
Fix for the previous commit (Celery app removal) Haven't paid enough attention and missed what ready method does. Removed the code. Libraries shouldn't do this - it's main project responsibility.
Python
mit
artemistomaras/django-ethereum-events,artemistomaras/django-ethereum-events
from django.apps import AppConfig from django.conf import settings class EthereumEventsConfig(AppConfig): name = 'django_ethereum_events' - def ready(self): - super(EthereumEventsConfig, self).ready() - app.config_from_object('django.conf:settings') - app.autodiscover_tasks(lambda: settings.INSTALLED_APPS, force=True) -
Fix for the previous commit (Celery app removal)
## Code Before: from django.apps import AppConfig from django.conf import settings class EthereumEventsConfig(AppConfig): name = 'django_ethereum_events' def ready(self): super(EthereumEventsConfig, self).ready() app.config_from_object('django.conf:settings') app.autodiscover_tasks(lambda: settings.INSTALLED_APPS, force=True) ## Instruction: Fix for the previous commit (Celery app removal) ## Code After: from django.apps import AppConfig from django.conf import settings class EthereumEventsConfig(AppConfig): name = 'django_ethereum_events'
# ... existing code ... name = 'django_ethereum_events' # ... rest of the code ...
e30d433153d9ad2f1d931f7f48b0ebbe9ba6763c
modules/new_module/new_module.py
modules/new_module/new_module.py
from models import custom_modules from . import handlers def register_module(): """Registers this module in the registry.""" global_urls = [ ('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url ] course_urls = [ ('/new-course-url', handlers.NewURLHandler) ] # Course URLs go on mycourse.appspot.com/course-name/url global custom_module custom_module = custom_modules.Module( 'New module title (has to be unique)', 'Implements some functionality', global_urls, course_urls) return custom_module
import logging from models import custom_modules from . import handlers def register_module(): """Registers this module in the registry.""" def on_module_enabled(): logging.info('Module new_module.py was just enabled') def on_module_disabled(): logging.info('Module new_module.py was just dissabled') global_urls = [ ('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url ] course_urls = [ ('/new-course-url', handlers.NewURLHandler) ] # Course URLs go on mycourse.appspot.com/course-name/url global custom_module custom_module = custom_modules.Module( 'New module title (has to be unique)', 'Implements some functionality', global_urls, course_urls, notify_module_disabled=on_module_disabled, notify_module_enabled=on_module_enabled) return custom_module
Add enable and dissable hooks
Add enable and dissable hooks
Python
apache-2.0
UniMOOC/gcb-new-module,UniMOOC/gcb-new-module,UniMOOC/gcb-new-module,UniMOOC/gcb-new-module
+ + import logging from models import custom_modules from . import handlers def register_module(): """Registers this module in the registry.""" + + def on_module_enabled(): + logging.info('Module new_module.py was just enabled') + + def on_module_disabled(): + logging.info('Module new_module.py was just dissabled') global_urls = [ ('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url ] course_urls = [ ('/new-course-url', handlers.NewURLHandler) ] # Course URLs go on mycourse.appspot.com/course-name/url global custom_module custom_module = custom_modules.Module( 'New module title (has to be unique)', 'Implements some functionality', - global_urls, course_urls) + global_urls, course_urls, + notify_module_disabled=on_module_disabled, + notify_module_enabled=on_module_enabled) return custom_module
Add enable and dissable hooks
## Code Before: from models import custom_modules from . import handlers def register_module(): """Registers this module in the registry.""" global_urls = [ ('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url ] course_urls = [ ('/new-course-url', handlers.NewURLHandler) ] # Course URLs go on mycourse.appspot.com/course-name/url global custom_module custom_module = custom_modules.Module( 'New module title (has to be unique)', 'Implements some functionality', global_urls, course_urls) return custom_module ## Instruction: Add enable and dissable hooks ## Code After: import logging from models import custom_modules from . import handlers def register_module(): """Registers this module in the registry.""" def on_module_enabled(): logging.info('Module new_module.py was just enabled') def on_module_disabled(): logging.info('Module new_module.py was just dissabled') global_urls = [ ('/new-global-url', handlers.NewURLHandler) # Global URLs go on mycourse.appspot.com/url ] course_urls = [ ('/new-course-url', handlers.NewURLHandler) ] # Course URLs go on mycourse.appspot.com/course-name/url global custom_module custom_module = custom_modules.Module( 'New module title (has to be unique)', 'Implements some functionality', global_urls, course_urls, notify_module_disabled=on_module_disabled, notify_module_enabled=on_module_enabled) return custom_module
# ... existing code ... import logging # ... modified code ... """Registers this module in the registry.""" def on_module_enabled(): logging.info('Module new_module.py was just enabled') def on_module_disabled(): logging.info('Module new_module.py was just dissabled') ... 'Implements some functionality', global_urls, course_urls, notify_module_disabled=on_module_disabled, notify_module_enabled=on_module_enabled) return custom_module # ... rest of the code ...
83a16ba4485f3e483adc20352cb0cef7c02f8ef2
tests/test_config_schema.py
tests/test_config_schema.py
from __future__ import unicode_literals, division, absolute_import import jsonschema from flexget import config_schema from flexget import plugin from tests import FlexGetBase class TestSchemaValidator(FlexGetBase): def test_plugin_schemas_are_valid(self): for p in plugin.plugins.values(): if p.schema is None: continue try: config_schema.SchemaValidator.check_schema(p.schema) except jsonschema.SchemaError as e: assert False, 'plugin `%s` has an invalid schema. %s %s' % ( p.name, '/'.join(str(p) for p in e.path), e.message) def test_resolves_local_refs(self): schema = {'$ref': '/schema/plugin/accept_all'} v = config_schema.SchemaValidator(schema) # accept_all schema should be for type boolean assert v.is_valid(True) assert not v.is_valid(14) def test_custom_format_checker(self): schema = {'type': 'string', 'format': 'quality'} v = config_schema.SchemaValidator(schema) assert v.is_valid('720p') assert not v.is_valid('aoeu')
from __future__ import unicode_literals, division, absolute_import import jsonschema from flexget import config_schema from tests import FlexGetBase class TestSchemaValidator(FlexGetBase): def test_registered_schemas_are_valid(self): for path in config_schema.schema_paths: schema = config_schema.resolve_ref(path) try: config_schema.SchemaValidator.check_schema(schema) except jsonschema.SchemaError as e: assert False, 'plugin `%s` has an invalid schema. %s %s' % ( path, '/'.join(str(p) for p in e.path), e.message) def test_resolves_local_refs(self): schema = {'$ref': '/schema/plugin/accept_all'} v = config_schema.SchemaValidator(schema) # accept_all schema should be for type boolean assert v.is_valid(True) assert not v.is_valid(14) def test_custom_format_checker(self): schema = {'type': 'string', 'format': 'quality'} v = config_schema.SchemaValidator(schema) assert v.is_valid('720p') assert not v.is_valid('aoeu')
Convert unit test to test all registered schemas instead of plugins directly.
Convert unit test to test all registered schemas instead of plugins directly.
Python
mit
vfrc2/Flexget,poulpito/Flexget,jacobmetrick/Flexget,ibrahimkarahan/Flexget,Danfocus/Flexget,oxc/Flexget,ibrahimkarahan/Flexget,asm0dey/Flexget,qvazzler/Flexget,dsemi/Flexget,crawln45/Flexget,thalamus/Flexget,tsnoam/Flexget,patsissons/Flexget,tsnoam/Flexget,vfrc2/Flexget,Danfocus/Flexget,drwyrm/Flexget,v17al/Flexget,grrr2/Flexget,lildadou/Flexget,OmgOhnoes/Flexget,voriux/Flexget,crawln45/Flexget,ianstalk/Flexget,Danfocus/Flexget,spencerjanssen/Flexget,JorisDeRieck/Flexget,ratoaq2/Flexget,tobinjt/Flexget,thalamus/Flexget,ratoaq2/Flexget,gazpachoking/Flexget,qk4l/Flexget,grrr2/Flexget,lildadou/Flexget,tarzasai/Flexget,lildadou/Flexget,OmgOhnoes/Flexget,ZefQ/Flexget,antivirtel/Flexget,vfrc2/Flexget,Flexget/Flexget,qvazzler/Flexget,tobinjt/Flexget,patsissons/Flexget,xfouloux/Flexget,JorisDeRieck/Flexget,camon/Flexget,jawilson/Flexget,qk4l/Flexget,offbyone/Flexget,tarzasai/Flexget,poulpito/Flexget,sean797/Flexget,camon/Flexget,jacobmetrick/Flexget,Pretagonist/Flexget,tobinjt/Flexget,thalamus/Flexget,tsnoam/Flexget,tarzasai/Flexget,ratoaq2/Flexget,Pretagonist/Flexget,ianstalk/Flexget,ibrahimkarahan/Flexget,ZefQ/Flexget,gazpachoking/Flexget,grrr2/Flexget,Flexget/Flexget,malkavi/Flexget,Flexget/Flexget,spencerjanssen/Flexget,tvcsantos/Flexget,X-dark/Flexget,Flexget/Flexget,LynxyssCZ/Flexget,v17al/Flexget,asm0dey/Flexget,dsemi/Flexget,drwyrm/Flexget,tobinjt/Flexget,offbyone/Flexget,JorisDeRieck/Flexget,oxc/Flexget,cvium/Flexget,qk4l/Flexget,antivirtel/Flexget,LynxyssCZ/Flexget,malkavi/Flexget,cvium/Flexget,tvcsantos/Flexget,ianstalk/Flexget,poulpito/Flexget,drwyrm/Flexget,malkavi/Flexget,X-dark/Flexget,patsissons/Flexget,antivirtel/Flexget,LynxyssCZ/Flexget,offbyone/Flexget,JorisDeRieck/Flexget,jacobmetrick/Flexget,crawln45/Flexget,cvium/Flexget,jawilson/Flexget,jawilson/Flexget,oxc/Flexget,Danfocus/Flexget,xfouloux/Flexget,asm0dey/Flexget,crawln45/Flexget,spencerjanssen/Flexget,dsemi/Flexget,X-dark/Flexget,malkavi/Flexget,v17al/Flexget,ZefQ/Flexget,Pretagonist/Flexget,OmgOhnoes/Flexget,xfouloux/Flexget,jawilson/Flexget,LynxyssCZ/Flexget,sean797/Flexget,qvazzler/Flexget,voriux/Flexget,sean797/Flexget
from __future__ import unicode_literals, division, absolute_import import jsonschema from flexget import config_schema - from flexget import plugin from tests import FlexGetBase class TestSchemaValidator(FlexGetBase): - def test_plugin_schemas_are_valid(self): + def test_registered_schemas_are_valid(self): + for path in config_schema.schema_paths: + schema = config_schema.resolve_ref(path) - for p in plugin.plugins.values(): - if p.schema is None: - continue try: - config_schema.SchemaValidator.check_schema(p.schema) + config_schema.SchemaValidator.check_schema(schema) except jsonschema.SchemaError as e: assert False, 'plugin `%s` has an invalid schema. %s %s' % ( - p.name, '/'.join(str(p) for p in e.path), e.message) + path, '/'.join(str(p) for p in e.path), e.message) def test_resolves_local_refs(self): schema = {'$ref': '/schema/plugin/accept_all'} v = config_schema.SchemaValidator(schema) # accept_all schema should be for type boolean assert v.is_valid(True) assert not v.is_valid(14) def test_custom_format_checker(self): schema = {'type': 'string', 'format': 'quality'} v = config_schema.SchemaValidator(schema) assert v.is_valid('720p') assert not v.is_valid('aoeu')
Convert unit test to test all registered schemas instead of plugins directly.
## Code Before: from __future__ import unicode_literals, division, absolute_import import jsonschema from flexget import config_schema from flexget import plugin from tests import FlexGetBase class TestSchemaValidator(FlexGetBase): def test_plugin_schemas_are_valid(self): for p in plugin.plugins.values(): if p.schema is None: continue try: config_schema.SchemaValidator.check_schema(p.schema) except jsonschema.SchemaError as e: assert False, 'plugin `%s` has an invalid schema. %s %s' % ( p.name, '/'.join(str(p) for p in e.path), e.message) def test_resolves_local_refs(self): schema = {'$ref': '/schema/plugin/accept_all'} v = config_schema.SchemaValidator(schema) # accept_all schema should be for type boolean assert v.is_valid(True) assert not v.is_valid(14) def test_custom_format_checker(self): schema = {'type': 'string', 'format': 'quality'} v = config_schema.SchemaValidator(schema) assert v.is_valid('720p') assert not v.is_valid('aoeu') ## Instruction: Convert unit test to test all registered schemas instead of plugins directly. ## Code After: from __future__ import unicode_literals, division, absolute_import import jsonschema from flexget import config_schema from tests import FlexGetBase class TestSchemaValidator(FlexGetBase): def test_registered_schemas_are_valid(self): for path in config_schema.schema_paths: schema = config_schema.resolve_ref(path) try: config_schema.SchemaValidator.check_schema(schema) except jsonschema.SchemaError as e: assert False, 'plugin `%s` has an invalid schema. %s %s' % ( path, '/'.join(str(p) for p in e.path), e.message) def test_resolves_local_refs(self): schema = {'$ref': '/schema/plugin/accept_all'} v = config_schema.SchemaValidator(schema) # accept_all schema should be for type boolean assert v.is_valid(True) assert not v.is_valid(14) def test_custom_format_checker(self): schema = {'type': 'string', 'format': 'quality'} v = config_schema.SchemaValidator(schema) assert v.is_valid('720p') assert not v.is_valid('aoeu')
# ... existing code ... from flexget import config_schema from tests import FlexGetBase # ... modified code ... class TestSchemaValidator(FlexGetBase): def test_registered_schemas_are_valid(self): for path in config_schema.schema_paths: schema = config_schema.resolve_ref(path) try: config_schema.SchemaValidator.check_schema(schema) except jsonschema.SchemaError as e: ... assert False, 'plugin `%s` has an invalid schema. %s %s' % ( path, '/'.join(str(p) for p in e.path), e.message) # ... rest of the code ...
4daefdb0a4def961572fc22d0fe01a394b11fad9
tests/test_httpclient.py
tests/test_httpclient.py
try: import unittest2 as unittest except ImportError: import unittest import sys sys.path.append('..') from pyrabbit import http class TestHTTPClient(unittest.TestCase): """ Except for the init test, these are largely functional tests that require a RabbitMQ management API to be available on localhost:55672 """ def setUp(self): self.c = http.HTTPClient('localhost:55672', 'guest', 'guest') def test_client_init(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest') self.assertIsInstance(c, http.HTTPClient) def test_client_init_sets_default_timeout(self): self.assertEqual(self.c.client.timeout, 1) def test_client_init_with_timeout(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5) self.assertEqual(c.client.timeout, 5)
try: import unittest2 as unittest except ImportError: import unittest import sys sys.path.append('..') from pyrabbit import http class TestHTTPClient(unittest.TestCase): """ Except for the init test, these are largely functional tests that require a RabbitMQ management API to be available on localhost:55672 """ def setUp(self): self.c = http.HTTPClient('localhost:55672', 'guest', 'guest') def test_client_init(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest') self.assertIsInstance(c, http.HTTPClient) def test_client_init_sets_credentials(self): domain = '' expected_credentials = [(domain, 'guest', 'guest')] self.assertEqual( self.c.client.credentials.credentials, expected_credentials) def test_client_init_sets_default_timeout(self): self.assertEqual(self.c.client.timeout, 1) def test_client_init_with_timeout(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5) self.assertEqual(c.client.timeout, 5)
Test creation of HTTP credentials
tests.http: Test creation of HTTP credentials
Python
bsd-3-clause
ranjithlav/pyrabbit,bkjones/pyrabbit,NeCTAR-RC/pyrabbit,chaos95/pyrabbit,switchtower/pyrabbit
try: import unittest2 as unittest except ImportError: import unittest import sys sys.path.append('..') from pyrabbit import http class TestHTTPClient(unittest.TestCase): """ Except for the init test, these are largely functional tests that require a RabbitMQ management API to be available on localhost:55672 """ def setUp(self): self.c = http.HTTPClient('localhost:55672', 'guest', 'guest') def test_client_init(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest') self.assertIsInstance(c, http.HTTPClient) + def test_client_init_sets_credentials(self): + domain = '' + expected_credentials = [(domain, 'guest', 'guest')] + self.assertEqual( + self.c.client.credentials.credentials, expected_credentials) + def test_client_init_sets_default_timeout(self): self.assertEqual(self.c.client.timeout, 1) def test_client_init_with_timeout(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5) self.assertEqual(c.client.timeout, 5)
Test creation of HTTP credentials
## Code Before: try: import unittest2 as unittest except ImportError: import unittest import sys sys.path.append('..') from pyrabbit import http class TestHTTPClient(unittest.TestCase): """ Except for the init test, these are largely functional tests that require a RabbitMQ management API to be available on localhost:55672 """ def setUp(self): self.c = http.HTTPClient('localhost:55672', 'guest', 'guest') def test_client_init(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest') self.assertIsInstance(c, http.HTTPClient) def test_client_init_sets_default_timeout(self): self.assertEqual(self.c.client.timeout, 1) def test_client_init_with_timeout(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5) self.assertEqual(c.client.timeout, 5) ## Instruction: Test creation of HTTP credentials ## Code After: try: import unittest2 as unittest except ImportError: import unittest import sys sys.path.append('..') from pyrabbit import http class TestHTTPClient(unittest.TestCase): """ Except for the init test, these are largely functional tests that require a RabbitMQ management API to be available on localhost:55672 """ def setUp(self): self.c = http.HTTPClient('localhost:55672', 'guest', 'guest') def test_client_init(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest') self.assertIsInstance(c, http.HTTPClient) def test_client_init_sets_credentials(self): domain = '' expected_credentials = [(domain, 'guest', 'guest')] self.assertEqual( self.c.client.credentials.credentials, expected_credentials) def test_client_init_sets_default_timeout(self): self.assertEqual(self.c.client.timeout, 1) def test_client_init_with_timeout(self): c = http.HTTPClient('localhost:55672', 'guest', 'guest', 5) self.assertEqual(c.client.timeout, 5)
// ... existing code ... def test_client_init_sets_credentials(self): domain = '' expected_credentials = [(domain, 'guest', 'guest')] self.assertEqual( self.c.client.credentials.credentials, expected_credentials) def test_client_init_sets_default_timeout(self): // ... rest of the code ...
18bd0bcc0d892aef4ea9babfc6ec2af6e40cea62
manager/urls.py
manager/urls.py
from django.conf.urls import url from manager import views urlpatterns = [ url(r'^$', views.package_list, name='package_list'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail') ]
from django.conf.urls import url from manager import views urlpatterns = [ url(r'^$', views.package_list, name='package_list'), url(r'^packages/$', views.package_list, name='package_list'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail') ]
Add alternative package list url
Add alternative package list url
Python
mit
colajam93/aurpackager,colajam93/aurpackager,colajam93/aurpackager,colajam93/aurpackager
from django.conf.urls import url from manager import views urlpatterns = [ url(r'^$', views.package_list, name='package_list'), + url(r'^packages/$', views.package_list, name='package_list'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail') ]
Add alternative package list url
## Code Before: from django.conf.urls import url from manager import views urlpatterns = [ url(r'^$', views.package_list, name='package_list'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail') ] ## Instruction: Add alternative package list url ## Code After: from django.conf.urls import url from manager import views urlpatterns = [ url(r'^$', views.package_list, name='package_list'), url(r'^packages/$', views.package_list, name='package_list'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/build/$', views.package_build, name='package_build'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/(?P<build_number>\d+)/$', views.build_detail, name='build_detail') ]
... url(r'^$', views.package_list, name='package_list'), url(r'^packages/$', views.package_list, name='package_list'), url(r'^packages/(?P<package_name>[a-zA-Z0-9_+-]+)/$', views.package_detail, name='package_detail'), ...
a03eb91088943a4b3ed0ae5fc87b104562a4a645
location_field/urls.py
location_field/urls.py
try: from django.conf.urls import patterns # Django>=1.6 except ImportError: from django.conf.urls.defaults import patterns # Django<1.6 import os app_dir = os.path.dirname(__file__) urlpatterns = patterns( '', (r'^media/(.*)$', 'django.views.static.serve', { 'document_root': '%s/media' % app_dir}), )
from django.conf.urls import patterns import os app_dir = os.path.dirname(__file__) urlpatterns = patterns( '', (r'^media/(.*)$', 'django.views.static.serve', { 'document_root': '%s/media' % app_dir}), )
Drop support for Django 1.6
Drop support for Django 1.6
Python
mit
Mixser/django-location-field,recklessromeo/django-location-field,Mixser/django-location-field,voodmania/django-location-field,recklessromeo/django-location-field,undernewmanagement/django-location-field,voodmania/django-location-field,caioariede/django-location-field,caioariede/django-location-field,undernewmanagement/django-location-field,Mixser/django-location-field,undernewmanagement/django-location-field,caioariede/django-location-field,recklessromeo/django-location-field,voodmania/django-location-field
- try: - from django.conf.urls import patterns # Django>=1.6 + from django.conf.urls import patterns - except ImportError: - from django.conf.urls.defaults import patterns # Django<1.6 import os app_dir = os.path.dirname(__file__) urlpatterns = patterns( '', (r'^media/(.*)$', 'django.views.static.serve', { 'document_root': '%s/media' % app_dir}), )
Drop support for Django 1.6
## Code Before: try: from django.conf.urls import patterns # Django>=1.6 except ImportError: from django.conf.urls.defaults import patterns # Django<1.6 import os app_dir = os.path.dirname(__file__) urlpatterns = patterns( '', (r'^media/(.*)$', 'django.views.static.serve', { 'document_root': '%s/media' % app_dir}), ) ## Instruction: Drop support for Django 1.6 ## Code After: from django.conf.urls import patterns import os app_dir = os.path.dirname(__file__) urlpatterns = patterns( '', (r'^media/(.*)$', 'django.views.static.serve', { 'document_root': '%s/media' % app_dir}), )
# ... existing code ... from django.conf.urls import patterns # ... rest of the code ...
ecac9283bc831a6879f21e80e1b98818683ff6a4
atlas/prodtask/management/commands/pthealthcheck.py
atlas/prodtask/management/commands/pthealthcheck.py
from django.core.management.base import BaseCommand, CommandError import time from django_celery_beat.models import PeriodicTask from django.utils import timezone from datetime import timedelta from atlas.prodtask.views import send_alarm_message class Command(BaseCommand): args = 'None' help = 'Check celery beat health' def handle(self, *args, **options): if not args: try: try: last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0] except Exception as e: send_alarm_message('Alarm: the celery beat health check problem', f'Celery beat health check problem {e}') raise e if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3): send_alarm_message('Alarm: the celery beat is stuck', f'Celery beat last updated {last_executed_task.last_run_at}') except Exception as e: raise CommandError('Some problem during alarm mail sending check: %s'%e)
from django.core.management.base import BaseCommand, CommandError import time from django_celery_beat.models import PeriodicTask from django.utils import timezone from datetime import timedelta from atlas.prodtask.views import send_alarm_message class Command(BaseCommand): args = 'None' help = 'Check celery beat health' def handle(self, *args, **options): if not args: self.stdout.write(f'Start celery beat health check {timezone.now()}') try: try: last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0] except Exception as e: send_alarm_message('Alarm: the celery beat health check problem', f'Celery beat health check problem {e}') raise e if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3): send_alarm_message('Alarm: the celery beat is stuck', f'Celery beat last updated {last_executed_task.last_run_at}') except Exception as e: raise CommandError('Some problem during alarm mail sending check: %s'%e)
Add logging for health check
Add logging for health check
Python
apache-2.0
PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas,PanDAWMS/panda-bigmon-atlas
from django.core.management.base import BaseCommand, CommandError import time from django_celery_beat.models import PeriodicTask from django.utils import timezone from datetime import timedelta from atlas.prodtask.views import send_alarm_message class Command(BaseCommand): args = 'None' help = 'Check celery beat health' def handle(self, *args, **options): if not args: + self.stdout.write(f'Start celery beat health check {timezone.now()}') + try: try: last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0] except Exception as e: send_alarm_message('Alarm: the celery beat health check problem', f'Celery beat health check problem {e}') raise e if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3): send_alarm_message('Alarm: the celery beat is stuck', f'Celery beat last updated {last_executed_task.last_run_at}') except Exception as e: raise CommandError('Some problem during alarm mail sending check: %s'%e)
Add logging for health check
## Code Before: from django.core.management.base import BaseCommand, CommandError import time from django_celery_beat.models import PeriodicTask from django.utils import timezone from datetime import timedelta from atlas.prodtask.views import send_alarm_message class Command(BaseCommand): args = 'None' help = 'Check celery beat health' def handle(self, *args, **options): if not args: try: try: last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0] except Exception as e: send_alarm_message('Alarm: the celery beat health check problem', f'Celery beat health check problem {e}') raise e if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3): send_alarm_message('Alarm: the celery beat is stuck', f'Celery beat last updated {last_executed_task.last_run_at}') except Exception as e: raise CommandError('Some problem during alarm mail sending check: %s'%e) ## Instruction: Add logging for health check ## Code After: from django.core.management.base import BaseCommand, CommandError import time from django_celery_beat.models import PeriodicTask from django.utils import timezone from datetime import timedelta from atlas.prodtask.views import send_alarm_message class Command(BaseCommand): args = 'None' help = 'Check celery beat health' def handle(self, *args, **options): if not args: self.stdout.write(f'Start celery beat health check {timezone.now()}') try: try: last_executed_task = PeriodicTask.objects.all().order_by('-last_run_at')[0] except Exception as e: send_alarm_message('Alarm: the celery beat health check problem', f'Celery beat health check problem {e}') raise e if (timezone.now() - last_executed_task.last_run_at) > timedelta(hours=3): send_alarm_message('Alarm: the celery beat is stuck', f'Celery beat last updated {last_executed_task.last_run_at}') except Exception as e: raise CommandError('Some problem during alarm mail sending check: %s'%e)
# ... existing code ... if not args: self.stdout.write(f'Start celery beat health check {timezone.now()}') try: # ... rest of the code ...
442b083e9d1618569aa96a653ed2c0e4dfc27e59
saleor/search/forms.py
saleor/search/forms.py
from django import forms from .backends import get_search_backend class SearchForm(forms.Form): q = forms.CharField(label='Query', required=True) def search(self, model_or_queryset): backend = get_search_backend('default') query = self.cleaned_data['q'] results = backend.search(query, model_or_queryset=model_or_queryset) return results
from django import forms from django.utils.translation import pgettext from .backends import get_search_backend class SearchForm(forms.Form): q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True) def search(self, model_or_queryset): backend = get_search_backend('default') query = self.cleaned_data['q'] results = backend.search(query, model_or_queryset=model_or_queryset) return results
Add contextual marker for search app
Add contextual marker for search app
Python
bsd-3-clause
jreigel/saleor,car3oon/saleor,KenMutemi/saleor,HyperManTT/ECommerceSaleor,itbabu/saleor,UITools/saleor,HyperManTT/ECommerceSaleor,UITools/saleor,maferelo/saleor,tfroehlich82/saleor,mociepka/saleor,mociepka/saleor,UITools/saleor,maferelo/saleor,UITools/saleor,tfroehlich82/saleor,UITools/saleor,jreigel/saleor,itbabu/saleor,car3oon/saleor,KenMutemi/saleor,jreigel/saleor,car3oon/saleor,HyperManTT/ECommerceSaleor,KenMutemi/saleor,tfroehlich82/saleor,itbabu/saleor,maferelo/saleor,mociepka/saleor
from django import forms + from django.utils.translation import pgettext from .backends import get_search_backend class SearchForm(forms.Form): - q = forms.CharField(label='Query', required=True) + q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True) def search(self, model_or_queryset): backend = get_search_backend('default') query = self.cleaned_data['q'] results = backend.search(query, model_or_queryset=model_or_queryset) return results
Add contextual marker for search app
## Code Before: from django import forms from .backends import get_search_backend class SearchForm(forms.Form): q = forms.CharField(label='Query', required=True) def search(self, model_or_queryset): backend = get_search_backend('default') query = self.cleaned_data['q'] results = backend.search(query, model_or_queryset=model_or_queryset) return results ## Instruction: Add contextual marker for search app ## Code After: from django import forms from django.utils.translation import pgettext from .backends import get_search_backend class SearchForm(forms.Form): q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True) def search(self, model_or_queryset): backend = get_search_backend('default') query = self.cleaned_data['q'] results = backend.search(query, model_or_queryset=model_or_queryset) return results
# ... existing code ... from django import forms from django.utils.translation import pgettext # ... modified code ... class SearchForm(forms.Form): q = forms.CharField(label=pgettext('Search form label', 'Query'), required=True) # ... rest of the code ...
73e15928a8427eb5a6e4a886660b9493e50cd699
currencies/models.py
currencies/models.py
from django.db import models from django.utils.translation import gettext_lazy as _ class Currency(models.Model): code = models.CharField(_('code'), max_length=3) name = models.CharField(_('name'), max_length=35) symbol = models.CharField(_('symbol'), max_length=1, blank=True) factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4, help_text=_('Specifies the difference of the currency to default one.')) is_active = models.BooleanField(_('active'), default=True, help_text=_('The currency will be available.')) is_default = models.BooleanField(_('default'), default=False, help_text=_('Make this the default currency.')) class Meta: verbose_name = _('currency') verbose_name_plural = _('currencies') def __unicode__(self): return self.code def save(self, **kwargs): # Make sure the default currency is unique if self.is_default: Currency.objects.filter(is_default=True).update(is_default=False) super(Currency, self).save(**kwargs)
from django.db import models from django.utils.translation import gettext_lazy as _ class Currency(models.Model): code = models.CharField(_('code'), max_length=3) name = models.CharField(_('name'), max_length=35) symbol = models.CharField(_('symbol'), max_length=1, blank=True) factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4, help_text=_('Specifies the difference of the currency to default one.')) is_active = models.BooleanField(_('active'), default=True, help_text=_('The currency will be available.')) is_base = models.BooleanField(_('base'), default=False, help_text=_('Make this the base currency against which rates are calculated.')) is_default = models.BooleanField(_('default'), default=False, help_text=_('Make this the default user currency.')) class Meta: verbose_name = _('currency') verbose_name_plural = _('currencies') def __unicode__(self): return self.code def save(self, **kwargs): # Make sure the base and default currencies are unique if self.is_base: Currency.objects.filter(is_base=True).update(is_base=False) if self.is_default: Currency.objects.filter(is_default=True).update(is_default=False) super(Currency, self).save(**kwargs)
Add a Currency.is_base field (currently unused)
Add a Currency.is_base field (currently unused)
Python
bsd-3-clause
pathakamit88/django-currencies,panosl/django-currencies,ydaniv/django-currencies,mysociety/django-currencies,panosl/django-currencies,barseghyanartur/django-currencies,bashu/django-simple-currencies,pathakamit88/django-currencies,ydaniv/django-currencies,marcosalcazar/django-currencies,jmp0xf/django-currencies,racitup/django-currencies,mysociety/django-currencies,marcosalcazar/django-currencies,bashu/django-simple-currencies,racitup/django-currencies
from django.db import models from django.utils.translation import gettext_lazy as _ class Currency(models.Model): code = models.CharField(_('code'), max_length=3) name = models.CharField(_('name'), max_length=35) symbol = models.CharField(_('symbol'), max_length=1, blank=True) factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4, help_text=_('Specifies the difference of the currency to default one.')) is_active = models.BooleanField(_('active'), default=True, help_text=_('The currency will be available.')) + is_base = models.BooleanField(_('base'), default=False, + help_text=_('Make this the base currency against which rates are calculated.')) is_default = models.BooleanField(_('default'), default=False, - help_text=_('Make this the default currency.')) + help_text=_('Make this the default user currency.')) class Meta: verbose_name = _('currency') verbose_name_plural = _('currencies') def __unicode__(self): return self.code def save(self, **kwargs): - # Make sure the default currency is unique + # Make sure the base and default currencies are unique + if self.is_base: + Currency.objects.filter(is_base=True).update(is_base=False) if self.is_default: Currency.objects.filter(is_default=True).update(is_default=False) super(Currency, self).save(**kwargs)
Add a Currency.is_base field (currently unused)
## Code Before: from django.db import models from django.utils.translation import gettext_lazy as _ class Currency(models.Model): code = models.CharField(_('code'), max_length=3) name = models.CharField(_('name'), max_length=35) symbol = models.CharField(_('symbol'), max_length=1, blank=True) factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4, help_text=_('Specifies the difference of the currency to default one.')) is_active = models.BooleanField(_('active'), default=True, help_text=_('The currency will be available.')) is_default = models.BooleanField(_('default'), default=False, help_text=_('Make this the default currency.')) class Meta: verbose_name = _('currency') verbose_name_plural = _('currencies') def __unicode__(self): return self.code def save(self, **kwargs): # Make sure the default currency is unique if self.is_default: Currency.objects.filter(is_default=True).update(is_default=False) super(Currency, self).save(**kwargs) ## Instruction: Add a Currency.is_base field (currently unused) ## Code After: from django.db import models from django.utils.translation import gettext_lazy as _ class Currency(models.Model): code = models.CharField(_('code'), max_length=3) name = models.CharField(_('name'), max_length=35) symbol = models.CharField(_('symbol'), max_length=1, blank=True) factor = models.DecimalField(_('factor'), max_digits=10, decimal_places=4, help_text=_('Specifies the difference of the currency to default one.')) is_active = models.BooleanField(_('active'), default=True, help_text=_('The currency will be available.')) is_base = models.BooleanField(_('base'), default=False, help_text=_('Make this the base currency against which rates are calculated.')) is_default = models.BooleanField(_('default'), default=False, help_text=_('Make this the default user currency.')) class Meta: verbose_name = _('currency') verbose_name_plural = _('currencies') def __unicode__(self): return self.code def save(self, **kwargs): # Make sure the base and default currencies are unique if self.is_base: Currency.objects.filter(is_base=True).update(is_base=False) if self.is_default: Currency.objects.filter(is_default=True).update(is_default=False) super(Currency, self).save(**kwargs)
// ... existing code ... help_text=_('The currency will be available.')) is_base = models.BooleanField(_('base'), default=False, help_text=_('Make this the base currency against which rates are calculated.')) is_default = models.BooleanField(_('default'), default=False, help_text=_('Make this the default user currency.')) // ... modified code ... def save(self, **kwargs): # Make sure the base and default currencies are unique if self.is_base: Currency.objects.filter(is_base=True).update(is_base=False) if self.is_default: // ... rest of the code ...
220e0008924878f774f570cc0122c563f2c17465
recipes/migrations/0010_auto_20150919_1228.py
recipes/migrations/0010_auto_20150919_1228.py
from __future__ import unicode_literals from django.db import models, migrations def change_to_usage(apps, schema_editor): Recipe = apps.get_model("recipes", "Recipe") Ingredient = apps.get_model("recipes", "Ingredient") IngredientUsage = apps.get_model("recipes", "IngredientUsage") for recipe in Recipe.objects.all(): for ingredient in recipe.ingredient_set.all(): u = IngredientUsage(recipe=r, ingredient=i, quantity=1) u.save() class Migration(migrations.Migration): dependencies = [ ('recipes', '0009_auto_20150919_1226'), ] operations = [ migrations.RunPython(change_to_usage), ]
from __future__ import unicode_literals from django.db import models, migrations def change_to_usage(apps, schema_editor): Recipe = apps.get_model("recipes", "Recipe") Ingredient = apps.get_model("recipes", "Ingredient") IngredientUsage = apps.get_model("recipes", "IngredientUsage") for recipe in Recipe.objects.all(): for ingredient in recipe.ingredient_set.all(): u = IngredientUsage(recipe=recipe, ingredient=ingredient, quantity=1) u.save() class Migration(migrations.Migration): dependencies = [ ('recipes', '0009_auto_20150919_1226'), ] operations = [ migrations.RunPython(change_to_usage), ]
Make the data migration actually work
Make the data migration actually work
Python
agpl-3.0
XeryusTC/rotd,XeryusTC/rotd,XeryusTC/rotd
from __future__ import unicode_literals from django.db import models, migrations def change_to_usage(apps, schema_editor): Recipe = apps.get_model("recipes", "Recipe") Ingredient = apps.get_model("recipes", "Ingredient") IngredientUsage = apps.get_model("recipes", "IngredientUsage") for recipe in Recipe.objects.all(): for ingredient in recipe.ingredient_set.all(): - u = IngredientUsage(recipe=r, ingredient=i, quantity=1) + u = IngredientUsage(recipe=recipe, ingredient=ingredient, + quantity=1) u.save() class Migration(migrations.Migration): dependencies = [ ('recipes', '0009_auto_20150919_1226'), ] operations = [ migrations.RunPython(change_to_usage), ]
Make the data migration actually work
## Code Before: from __future__ import unicode_literals from django.db import models, migrations def change_to_usage(apps, schema_editor): Recipe = apps.get_model("recipes", "Recipe") Ingredient = apps.get_model("recipes", "Ingredient") IngredientUsage = apps.get_model("recipes", "IngredientUsage") for recipe in Recipe.objects.all(): for ingredient in recipe.ingredient_set.all(): u = IngredientUsage(recipe=r, ingredient=i, quantity=1) u.save() class Migration(migrations.Migration): dependencies = [ ('recipes', '0009_auto_20150919_1226'), ] operations = [ migrations.RunPython(change_to_usage), ] ## Instruction: Make the data migration actually work ## Code After: from __future__ import unicode_literals from django.db import models, migrations def change_to_usage(apps, schema_editor): Recipe = apps.get_model("recipes", "Recipe") Ingredient = apps.get_model("recipes", "Ingredient") IngredientUsage = apps.get_model("recipes", "IngredientUsage") for recipe in Recipe.objects.all(): for ingredient in recipe.ingredient_set.all(): u = IngredientUsage(recipe=recipe, ingredient=ingredient, quantity=1) u.save() class Migration(migrations.Migration): dependencies = [ ('recipes', '0009_auto_20150919_1226'), ] operations = [ migrations.RunPython(change_to_usage), ]
// ... existing code ... for ingredient in recipe.ingredient_set.all(): u = IngredientUsage(recipe=recipe, ingredient=ingredient, quantity=1) u.save() // ... rest of the code ...
8fe8717b4e2afe6329d2dd25210371df3eab2b4f
test/test_stdlib.py
test/test_stdlib.py
import pep543.stdlib from .backend_tests import SimpleNegotiation class TestSimpleNegotiationStdlib(SimpleNegotiation): BACKEND = pep543.stdlib.STDLIB_BACKEND
import pep543 import pep543.stdlib import pytest from .backend_tests import SimpleNegotiation CONTEXTS = ( pep543.stdlib.STDLIB_BACKEND.client_context, pep543.stdlib.STDLIB_BACKEND.server_context ) def assert_wrap_fails(context, exception): """ A convenient helper that calls wrap_buffers with the appropriate number of arugments and asserts that it raises the appropriate error. """ if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context): with pytest.raises(exception): context.wrap_buffers(server_hostname=None) else: with pytest.raises(exception): context.wrap_buffers() class TestSimpleNegotiationStdlib(SimpleNegotiation): BACKEND = pep543.stdlib.STDLIB_BACKEND class TestStdlibErrorHandling(object): """ Validate that the stdlib backend can do sensible error handling in specific situations that it cannot handle. """ @pytest.mark.parametrize( 'lowest,highest', ( (object(), None), (None, object()), (object(), object()) ) ) @pytest.mark.parametrize('context', CONTEXTS) def test_bad_values_for_versions_client(self, lowest, highest, context): """ Using TLSConfiguration objects with a bad value for their minimum version raises a TLSError with Client contexts. """ config = pep543.TLSConfiguration( validate_certificates=False, lowest_supported_version=lowest, highest_supported_version=highest ) ctx = context(config) assert_wrap_fails(ctx, pep543.TLSError)
Test that we reject bad TLS versions
Test that we reject bad TLS versions
Python
mit
python-hyper/pep543
+ import pep543 import pep543.stdlib + import pytest + from .backend_tests import SimpleNegotiation + + + CONTEXTS = ( + pep543.stdlib.STDLIB_BACKEND.client_context, + pep543.stdlib.STDLIB_BACKEND.server_context + ) + + + def assert_wrap_fails(context, exception): + """ + A convenient helper that calls wrap_buffers with the appropriate number of + arugments and asserts that it raises the appropriate error. + """ + if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context): + with pytest.raises(exception): + context.wrap_buffers(server_hostname=None) + else: + with pytest.raises(exception): + context.wrap_buffers() class TestSimpleNegotiationStdlib(SimpleNegotiation): BACKEND = pep543.stdlib.STDLIB_BACKEND + + class TestStdlibErrorHandling(object): + """ + Validate that the stdlib backend can do sensible error handling in specific + situations that it cannot handle. + """ + @pytest.mark.parametrize( + 'lowest,highest', ( + (object(), None), (None, object()), (object(), object()) + ) + ) + @pytest.mark.parametrize('context', CONTEXTS) + def test_bad_values_for_versions_client(self, lowest, highest, context): + """ + Using TLSConfiguration objects with a bad value for their minimum + version raises a TLSError with Client contexts. + """ + config = pep543.TLSConfiguration( + validate_certificates=False, + lowest_supported_version=lowest, + highest_supported_version=highest + ) + ctx = context(config) + assert_wrap_fails(ctx, pep543.TLSError) +
Test that we reject bad TLS versions
## Code Before: import pep543.stdlib from .backend_tests import SimpleNegotiation class TestSimpleNegotiationStdlib(SimpleNegotiation): BACKEND = pep543.stdlib.STDLIB_BACKEND ## Instruction: Test that we reject bad TLS versions ## Code After: import pep543 import pep543.stdlib import pytest from .backend_tests import SimpleNegotiation CONTEXTS = ( pep543.stdlib.STDLIB_BACKEND.client_context, pep543.stdlib.STDLIB_BACKEND.server_context ) def assert_wrap_fails(context, exception): """ A convenient helper that calls wrap_buffers with the appropriate number of arugments and asserts that it raises the appropriate error. """ if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context): with pytest.raises(exception): context.wrap_buffers(server_hostname=None) else: with pytest.raises(exception): context.wrap_buffers() class TestSimpleNegotiationStdlib(SimpleNegotiation): BACKEND = pep543.stdlib.STDLIB_BACKEND class TestStdlibErrorHandling(object): """ Validate that the stdlib backend can do sensible error handling in specific situations that it cannot handle. """ @pytest.mark.parametrize( 'lowest,highest', ( (object(), None), (None, object()), (object(), object()) ) ) @pytest.mark.parametrize('context', CONTEXTS) def test_bad_values_for_versions_client(self, lowest, highest, context): """ Using TLSConfiguration objects with a bad value for their minimum version raises a TLSError with Client contexts. """ config = pep543.TLSConfiguration( validate_certificates=False, lowest_supported_version=lowest, highest_supported_version=highest ) ctx = context(config) assert_wrap_fails(ctx, pep543.TLSError)
// ... existing code ... import pep543 import pep543.stdlib import pytest // ... modified code ... CONTEXTS = ( pep543.stdlib.STDLIB_BACKEND.client_context, pep543.stdlib.STDLIB_BACKEND.server_context ) def assert_wrap_fails(context, exception): """ A convenient helper that calls wrap_buffers with the appropriate number of arugments and asserts that it raises the appropriate error. """ if isinstance(context, pep543.stdlib.STDLIB_BACKEND.client_context): with pytest.raises(exception): context.wrap_buffers(server_hostname=None) else: with pytest.raises(exception): context.wrap_buffers() class TestSimpleNegotiationStdlib(SimpleNegotiation): ... BACKEND = pep543.stdlib.STDLIB_BACKEND class TestStdlibErrorHandling(object): """ Validate that the stdlib backend can do sensible error handling in specific situations that it cannot handle. """ @pytest.mark.parametrize( 'lowest,highest', ( (object(), None), (None, object()), (object(), object()) ) ) @pytest.mark.parametrize('context', CONTEXTS) def test_bad_values_for_versions_client(self, lowest, highest, context): """ Using TLSConfiguration objects with a bad value for their minimum version raises a TLSError with Client contexts. """ config = pep543.TLSConfiguration( validate_certificates=False, lowest_supported_version=lowest, highest_supported_version=highest ) ctx = context(config) assert_wrap_fails(ctx, pep543.TLSError) // ... rest of the code ...
332452cf7ccd6d3ee583be9a6aac27b14771263f
source/services/omdb_service.py
source/services/omdb_service.py
import requests from bs4 import BeautifulSoup from source.models.rt_rating import RTRating class OmdbService: __API_URL = 'http://www.omdbapi.com/?' def __init__(self, movie_id): self.id = movie_id def get_rt_rating(self): payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'} response = requests.post(self.__API_URL, params=payload) movie_info = response.json() ratings = [] ratings.append(movie_info['tomatoMeter']) ratings.append(movie_info['tomatoUserMeter']) return RTRating(ratings)
import requests from bs4 import BeautifulSoup from source.models.rt_rating import RTRating class OmdbService: __API_URL = 'http://www.omdbapi.com/?' def __init__(self, movie_id): self.id = movie_id def get_rt_rating(self): payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'} response = requests.post(self.__API_URL, params=payload) movie_info = response.json() scores = [] scores.append(movie_info['tomatoMeter']) scores.append(movie_info['tomatoUserMeter']) rt_rating = RTRating(scores) rt_rating.link = movie_info['tomatoURL'] return rt_rating
Add url to RTRating object
Add url to RTRating object
Python
mit
jeremyrea/caterblu,jeremyrea/caterblu,jeremyrea/caterblu,jeremyrea/caterblu
import requests from bs4 import BeautifulSoup from source.models.rt_rating import RTRating class OmdbService: __API_URL = 'http://www.omdbapi.com/?' def __init__(self, movie_id): self.id = movie_id def get_rt_rating(self): payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'} response = requests.post(self.__API_URL, params=payload) movie_info = response.json() - ratings = [] + scores = [] - ratings.append(movie_info['tomatoMeter']) + scores.append(movie_info['tomatoMeter']) - ratings.append(movie_info['tomatoUserMeter']) + scores.append(movie_info['tomatoUserMeter']) - return RTRating(ratings) + rt_rating = RTRating(scores) + rt_rating.link = movie_info['tomatoURL'] + return rt_rating +
Add url to RTRating object
## Code Before: import requests from bs4 import BeautifulSoup from source.models.rt_rating import RTRating class OmdbService: __API_URL = 'http://www.omdbapi.com/?' def __init__(self, movie_id): self.id = movie_id def get_rt_rating(self): payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'} response = requests.post(self.__API_URL, params=payload) movie_info = response.json() ratings = [] ratings.append(movie_info['tomatoMeter']) ratings.append(movie_info['tomatoUserMeter']) return RTRating(ratings) ## Instruction: Add url to RTRating object ## Code After: import requests from bs4 import BeautifulSoup from source.models.rt_rating import RTRating class OmdbService: __API_URL = 'http://www.omdbapi.com/?' def __init__(self, movie_id): self.id = movie_id def get_rt_rating(self): payload = {'i': self.id, 'plot': 'short', 'r': 'json', 'tomatoes': 'true'} response = requests.post(self.__API_URL, params=payload) movie_info = response.json() scores = [] scores.append(movie_info['tomatoMeter']) scores.append(movie_info['tomatoUserMeter']) rt_rating = RTRating(scores) rt_rating.link = movie_info['tomatoURL'] return rt_rating
# ... existing code ... scores = [] scores.append(movie_info['tomatoMeter']) scores.append(movie_info['tomatoUserMeter']) rt_rating = RTRating(scores) rt_rating.link = movie_info['tomatoURL'] return rt_rating # ... rest of the code ...
e120aafad13138abb5d98bbad12c0d8bdc532b30
lglass/web/application.py
lglass/web/application.py
import bottle from lglass.web.helpers import render_template, with_config app = bottle.Bottle() def index_handler(): return render_template("index.html") @with_config def robots_txt_handler(config): if config["robots.txt"] is not None: return bottle.static_file(config["robots.txt"]) bottle.abort(404, "File not found") app.route("/", "GET", index_handler) app.route("/robots.txt", "GET", robots_txt_handler) import lglass.web.registry app.route("/obj", "GET", lglass.web.registry.show_object_types) app.route("/obj/<type>", "GET", lglass.web.registry.show_objects) app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object) app.route("/whois/<query>", "GET", lglass.web.registry.whois_query) app.route("/whois", "POST", lglass.web.registry.whois_query) app.route("/flush", "POST", lglass.web.registry.flush_cache)
import bottle from lglass.web.helpers import render_template, with_config app = bottle.Bottle() def index_handler(): return render_template("index.html") @with_config def robots_txt_handler(config): if config["robots.txt"] is not None: return open(config["robots.txt"]) bottle.abort(404, "File not found") app.route("/", "GET", index_handler) app.route("/robots.txt", "GET", robots_txt_handler) import lglass.web.registry app.route("/obj", "GET", lglass.web.registry.show_object_types) app.route("/obj/<type>", "GET", lglass.web.registry.show_objects) app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object) app.route("/whois/<query>", "GET", lglass.web.registry.whois_query) app.route("/whois", "POST", lglass.web.registry.whois_query) app.route("/flush", "POST", lglass.web.registry.flush_cache)
Replace static_file by open call
Replace static_file by open call
Python
mit
fritz0705/lglass
import bottle from lglass.web.helpers import render_template, with_config app = bottle.Bottle() def index_handler(): return render_template("index.html") @with_config def robots_txt_handler(config): if config["robots.txt"] is not None: - return bottle.static_file(config["robots.txt"]) + return open(config["robots.txt"]) bottle.abort(404, "File not found") app.route("/", "GET", index_handler) app.route("/robots.txt", "GET", robots_txt_handler) import lglass.web.registry app.route("/obj", "GET", lglass.web.registry.show_object_types) app.route("/obj/<type>", "GET", lglass.web.registry.show_objects) app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object) app.route("/whois/<query>", "GET", lglass.web.registry.whois_query) app.route("/whois", "POST", lglass.web.registry.whois_query) app.route("/flush", "POST", lglass.web.registry.flush_cache)
Replace static_file by open call
## Code Before: import bottle from lglass.web.helpers import render_template, with_config app = bottle.Bottle() def index_handler(): return render_template("index.html") @with_config def robots_txt_handler(config): if config["robots.txt"] is not None: return bottle.static_file(config["robots.txt"]) bottle.abort(404, "File not found") app.route("/", "GET", index_handler) app.route("/robots.txt", "GET", robots_txt_handler) import lglass.web.registry app.route("/obj", "GET", lglass.web.registry.show_object_types) app.route("/obj/<type>", "GET", lglass.web.registry.show_objects) app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object) app.route("/whois/<query>", "GET", lglass.web.registry.whois_query) app.route("/whois", "POST", lglass.web.registry.whois_query) app.route("/flush", "POST", lglass.web.registry.flush_cache) ## Instruction: Replace static_file by open call ## Code After: import bottle from lglass.web.helpers import render_template, with_config app = bottle.Bottle() def index_handler(): return render_template("index.html") @with_config def robots_txt_handler(config): if config["robots.txt"] is not None: return open(config["robots.txt"]) bottle.abort(404, "File not found") app.route("/", "GET", index_handler) app.route("/robots.txt", "GET", robots_txt_handler) import lglass.web.registry app.route("/obj", "GET", lglass.web.registry.show_object_types) app.route("/obj/<type>", "GET", lglass.web.registry.show_objects) app.route("/obj/<type>/<primary_key>", "GET", lglass.web.registry.show_object) app.route("/whois/<query>", "GET", lglass.web.registry.whois_query) app.route("/whois", "POST", lglass.web.registry.whois_query) app.route("/flush", "POST", lglass.web.registry.flush_cache)
... if config["robots.txt"] is not None: return open(config["robots.txt"]) bottle.abort(404, "File not found") ...
8e0f2271b19504886728ccf5d060778c027c79ca
ide/views.py
ide/views.py
import json from werkzeug.routing import BaseConverter from flask import render_template, request, abort import requests from ide import app from ide.projects import get_all_projects, Project MCLABAAS_URL = 'http://localhost:4242' @app.route('/') def index(): return render_template('index.html', projects=get_all_projects()) @app.route('/parse', methods=['POST']) def parse(): return requests.post(MCLABAAS_URL + '/ast', data=request.data).text class ProjectConverter(BaseConverter): def to_python(self, value): return Project(value) def to_url(self, value): return BaseConverter.to_url(value.name) app.url_map.converters['project'] = ProjectConverter @app.route('/project/<project:project>/') def project(project): if not project.exists(): abort(404) return render_template('project.html') @app.route('/project/<project:project>/tree', methods=['GET']) def tree(project): return json.dumps(project.tree()) @app.route('/project/<project:project>/read', methods=['GET']) def read(project): return project.read_file(request.args['path']) @app.route('/project/<project:project>/write', methods=['POST']) def write(project): project.write_file(request.form['path'], request.form['contents']) return json.dumps({'status': 'OK'})
import json from werkzeug.routing import BaseConverter from flask import render_template, request, abort import requests from ide import app from ide.projects import get_all_projects, Project MCLABAAS_URL = 'http://localhost:4242' @app.route('/') def index(): return render_template('index.html', projects=get_all_projects()) @app.route('/parse', methods=['POST']) def parse(): return requests.post(MCLABAAS_URL + '/ast', data=request.data).text class ProjectConverter(BaseConverter): def to_python(self, value): project = Project(value) if not project.exists(): abort(404) return project def to_url(self, value): return BaseConverter.to_url(value.name) app.url_map.converters['project'] = ProjectConverter @app.route('/project/<project:project>/') def project(project): return render_template('project.html') @app.route('/project/<project:project>/tree', methods=['GET']) def tree(project): return json.dumps(project.tree()) @app.route('/project/<project:project>/read', methods=['GET']) def read(project): return project.read_file(request.args['path']) @app.route('/project/<project:project>/write', methods=['POST']) def write(project): project.write_file(request.form['path'], request.form['contents']) return json.dumps({'status': 'OK'})
Check if project exists inside ProjectConverter.
Check if project exists inside ProjectConverter.
Python
apache-2.0
Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide,Sable/mclab-ide
import json from werkzeug.routing import BaseConverter from flask import render_template, request, abort import requests from ide import app from ide.projects import get_all_projects, Project MCLABAAS_URL = 'http://localhost:4242' @app.route('/') def index(): return render_template('index.html', projects=get_all_projects()) @app.route('/parse', methods=['POST']) def parse(): return requests.post(MCLABAAS_URL + '/ast', data=request.data).text class ProjectConverter(BaseConverter): def to_python(self, value): - return Project(value) + project = Project(value) + if not project.exists(): + abort(404) + return project def to_url(self, value): return BaseConverter.to_url(value.name) app.url_map.converters['project'] = ProjectConverter @app.route('/project/<project:project>/') def project(project): - if not project.exists(): - abort(404) return render_template('project.html') @app.route('/project/<project:project>/tree', methods=['GET']) def tree(project): return json.dumps(project.tree()) @app.route('/project/<project:project>/read', methods=['GET']) def read(project): return project.read_file(request.args['path']) @app.route('/project/<project:project>/write', methods=['POST']) def write(project): project.write_file(request.form['path'], request.form['contents']) return json.dumps({'status': 'OK'})
Check if project exists inside ProjectConverter.
## Code Before: import json from werkzeug.routing import BaseConverter from flask import render_template, request, abort import requests from ide import app from ide.projects import get_all_projects, Project MCLABAAS_URL = 'http://localhost:4242' @app.route('/') def index(): return render_template('index.html', projects=get_all_projects()) @app.route('/parse', methods=['POST']) def parse(): return requests.post(MCLABAAS_URL + '/ast', data=request.data).text class ProjectConverter(BaseConverter): def to_python(self, value): return Project(value) def to_url(self, value): return BaseConverter.to_url(value.name) app.url_map.converters['project'] = ProjectConverter @app.route('/project/<project:project>/') def project(project): if not project.exists(): abort(404) return render_template('project.html') @app.route('/project/<project:project>/tree', methods=['GET']) def tree(project): return json.dumps(project.tree()) @app.route('/project/<project:project>/read', methods=['GET']) def read(project): return project.read_file(request.args['path']) @app.route('/project/<project:project>/write', methods=['POST']) def write(project): project.write_file(request.form['path'], request.form['contents']) return json.dumps({'status': 'OK'}) ## Instruction: Check if project exists inside ProjectConverter. ## Code After: import json from werkzeug.routing import BaseConverter from flask import render_template, request, abort import requests from ide import app from ide.projects import get_all_projects, Project MCLABAAS_URL = 'http://localhost:4242' @app.route('/') def index(): return render_template('index.html', projects=get_all_projects()) @app.route('/parse', methods=['POST']) def parse(): return requests.post(MCLABAAS_URL + '/ast', data=request.data).text class ProjectConverter(BaseConverter): def to_python(self, value): project = Project(value) if not project.exists(): abort(404) return project def to_url(self, value): return BaseConverter.to_url(value.name) app.url_map.converters['project'] = ProjectConverter @app.route('/project/<project:project>/') def project(project): return render_template('project.html') @app.route('/project/<project:project>/tree', methods=['GET']) def tree(project): return json.dumps(project.tree()) @app.route('/project/<project:project>/read', methods=['GET']) def read(project): return project.read_file(request.args['path']) @app.route('/project/<project:project>/write', methods=['POST']) def write(project): project.write_file(request.form['path'], request.form['contents']) return json.dumps({'status': 'OK'})
... def to_python(self, value): project = Project(value) if not project.exists(): abort(404) return project ... def project(project): return render_template('project.html') ...
5cf0e2e9d68d2e0fca3780608f33d5d8cdaef8a9
admin/metrics/views.py
admin/metrics/views.py
from django.views.generic import TemplateView from django.contrib.auth.mixins import PermissionRequiredMixin from admin.base.settings import KEEN_CREDENTIALS from admin.base.utils import OSFAdmin class MetricsView(OSFAdmin, TemplateView, PermissionRequiredMixin): template_name = 'metrics/osf_metrics.html' permission_required = 'admin.view_metrics' def get_context_data(self, **kwargs): kwargs.update(KEEN_CREDENTIALS.copy()) return super(MetricsView, self).get_context_data(**kwargs)
from django.views.generic import TemplateView from django.contrib.auth.mixins import PermissionRequiredMixin from admin.base.settings import KEEN_CREDENTIALS class MetricsView(TemplateView, PermissionRequiredMixin): template_name = 'metrics/osf_metrics.html' permission_required = 'admin.view_metrics' def get_context_data(self, **kwargs): kwargs.update(KEEN_CREDENTIALS.copy()) return super(MetricsView, self).get_context_data(**kwargs)
Remove one more reference to old group permissions
Remove one more reference to old group permissions
Python
apache-2.0
erinspace/osf.io,adlius/osf.io,hmoco/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,cslzchen/osf.io,saradbowman/osf.io,icereval/osf.io,icereval/osf.io,leb2dg/osf.io,icereval/osf.io,sloria/osf.io,caseyrollins/osf.io,sloria/osf.io,Nesiehr/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,felliott/osf.io,monikagrabowska/osf.io,brianjgeiger/osf.io,mattclark/osf.io,cwisecarver/osf.io,mattclark/osf.io,monikagrabowska/osf.io,cwisecarver/osf.io,mfraezz/osf.io,CenterForOpenScience/osf.io,laurenrevere/osf.io,chennan47/osf.io,acshi/osf.io,Nesiehr/osf.io,laurenrevere/osf.io,erinspace/osf.io,acshi/osf.io,erinspace/osf.io,mfraezz/osf.io,mfraezz/osf.io,Johnetordoff/osf.io,aaxelb/osf.io,aaxelb/osf.io,hmoco/osf.io,binoculars/osf.io,brianjgeiger/osf.io,pattisdr/osf.io,binoculars/osf.io,mattclark/osf.io,hmoco/osf.io,chennan47/osf.io,chrisseto/osf.io,laurenrevere/osf.io,CenterForOpenScience/osf.io,chennan47/osf.io,caneruguz/osf.io,hmoco/osf.io,monikagrabowska/osf.io,adlius/osf.io,monikagrabowska/osf.io,TomBaxter/osf.io,aaxelb/osf.io,baylee-d/osf.io,mfraezz/osf.io,saradbowman/osf.io,chrisseto/osf.io,sloria/osf.io,caseyrollins/osf.io,caneruguz/osf.io,Nesiehr/osf.io,crcresearch/osf.io,leb2dg/osf.io,HalcyonChimera/osf.io,pattisdr/osf.io,crcresearch/osf.io,acshi/osf.io,acshi/osf.io,brianjgeiger/osf.io,adlius/osf.io,felliott/osf.io,leb2dg/osf.io,Johnetordoff/osf.io,cwisecarver/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,leb2dg/osf.io,CenterForOpenScience/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,felliott/osf.io,aaxelb/osf.io,acshi/osf.io,binoculars/osf.io,baylee-d/osf.io,chrisseto/osf.io,caneruguz/osf.io,CenterForOpenScience/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,pattisdr/osf.io,felliott/osf.io,TomBaxter/osf.io,HalcyonChimera/osf.io,adlius/osf.io,caneruguz/osf.io,baylee-d/osf.io,caseyrollins/osf.io,cwisecarver/osf.io,crcresearch/osf.io,chrisseto/osf.io,Nesiehr/osf.io
from django.views.generic import TemplateView from django.contrib.auth.mixins import PermissionRequiredMixin from admin.base.settings import KEEN_CREDENTIALS - from admin.base.utils import OSFAdmin - - - class MetricsView(OSFAdmin, TemplateView, PermissionRequiredMixin): + class MetricsView(TemplateView, PermissionRequiredMixin): template_name = 'metrics/osf_metrics.html' permission_required = 'admin.view_metrics' def get_context_data(self, **kwargs): kwargs.update(KEEN_CREDENTIALS.copy()) return super(MetricsView, self).get_context_data(**kwargs)
Remove one more reference to old group permissions
## Code Before: from django.views.generic import TemplateView from django.contrib.auth.mixins import PermissionRequiredMixin from admin.base.settings import KEEN_CREDENTIALS from admin.base.utils import OSFAdmin class MetricsView(OSFAdmin, TemplateView, PermissionRequiredMixin): template_name = 'metrics/osf_metrics.html' permission_required = 'admin.view_metrics' def get_context_data(self, **kwargs): kwargs.update(KEEN_CREDENTIALS.copy()) return super(MetricsView, self).get_context_data(**kwargs) ## Instruction: Remove one more reference to old group permissions ## Code After: from django.views.generic import TemplateView from django.contrib.auth.mixins import PermissionRequiredMixin from admin.base.settings import KEEN_CREDENTIALS class MetricsView(TemplateView, PermissionRequiredMixin): template_name = 'metrics/osf_metrics.html' permission_required = 'admin.view_metrics' def get_context_data(self, **kwargs): kwargs.update(KEEN_CREDENTIALS.copy()) return super(MetricsView, self).get_context_data(**kwargs)
# ... existing code ... class MetricsView(TemplateView, PermissionRequiredMixin): template_name = 'metrics/osf_metrics.html' # ... rest of the code ...
1cf82c6efa0550c5a0ba7160f82f77db6e3358ec
panoptes/test/test_mount.py
panoptes/test/test_mount.py
from panoptes.mount.ioptron import iOptronMount class TestOptronMount: mount = None def setup(self): print ("TestMount:setup() before each test method") def teardown(self): print ("TestMount:teardown() after each test method") @classmethod def setup_class(cls): print ("setup_class() before any methods in this class") _Mounts = [] for name in os.listdir(os.path.dirname(__file__)): if not name.startswith('_') and name.endswith('.py'): name = '.' + os.path.splitext(name)[0] try: module = importlib.import_module(name,'panoptes') _Mounts.append(module) except ImportError as err: self.logger.warn('Failed to load mount plugin: {}'.format(err)) @classmethod def teardown_class(cls): print ("teardown_class() after any methods in this class") def test_is_connected_false(self): pass def test_connect(self): pass def test_is_connected_true(self): pass def test_is_slewing(self): pass def test_check_coordinates(self): pass def test_sync_coordinates(self): pass def test_slew_to_coordinates(self): pass def test_slew_to_park(self): pass def test_echo(self): pass
import os import importlib import warnings class TestOptronMount: mount = None def setup(self): print ("TestMount:setup() before each test method") def teardown(self): print ("TestMount:teardown() after each test method") @classmethod def setup_class(cls): mount_dir = os.path.dirname(__file__) + '/../mount/' print ("setup_class() before any methods in this class") _Mounts = [] for name in os.listdir(os.path.dirname(mount_dir)): if not name.startswith('_') and name.endswith('.py'): name = '.' + os.path.splitext(name)[0] try: module = importlib.import_module(name,'panoptes.mount') _Mounts.append(module) except ImportError as err: warnings.warn('Failed to load mount plugin: {}'.format(err)) @classmethod def teardown_class(cls): print ("teardown_class() after any methods in this class") def test_is_connected_false(self): pass def test_connect(self): pass def test_is_connected_true(self): pass def test_is_slewing(self): pass def test_check_coordinates(self): pass def test_sync_coordinates(self): pass def test_slew_to_coordinates(self): pass def test_slew_to_park(self): pass def test_echo(self): pass
Test file loops over all the mounts
Test file loops over all the mounts
Python
mit
AstroHuntsman/POCS,AstroHuntsman/POCS,panoptes/POCS,fmin2958/POCS,joshwalawender/POCS,AstroHuntsman/POCS,panoptes/POCS,fmin2958/POCS,Guokr1991/POCS,panoptes/POCS,Guokr1991/POCS,joshwalawender/POCS,Guokr1991/POCS,Guokr1991/POCS,panoptes/POCS,AstroHuntsman/POCS,joshwalawender/POCS,fmin2958/POCS
- from panoptes.mount.ioptron import iOptronMount + import os + import importlib + import warnings class TestOptronMount: mount = None def setup(self): print ("TestMount:setup() before each test method") def teardown(self): print ("TestMount:teardown() after each test method") @classmethod def setup_class(cls): + mount_dir = os.path.dirname(__file__) + '/../mount/' print ("setup_class() before any methods in this class") _Mounts = [] - for name in os.listdir(os.path.dirname(__file__)): + for name in os.listdir(os.path.dirname(mount_dir)): if not name.startswith('_') and name.endswith('.py'): name = '.' + os.path.splitext(name)[0] try: - module = importlib.import_module(name,'panoptes') + module = importlib.import_module(name,'panoptes.mount') _Mounts.append(module) except ImportError as err: - self.logger.warn('Failed to load mount plugin: {}'.format(err)) + warnings.warn('Failed to load mount plugin: {}'.format(err)) @classmethod def teardown_class(cls): print ("teardown_class() after any methods in this class") def test_is_connected_false(self): pass def test_connect(self): pass def test_is_connected_true(self): pass def test_is_slewing(self): pass def test_check_coordinates(self): pass def test_sync_coordinates(self): pass def test_slew_to_coordinates(self): pass def test_slew_to_park(self): pass def test_echo(self): pass
Test file loops over all the mounts
## Code Before: from panoptes.mount.ioptron import iOptronMount class TestOptronMount: mount = None def setup(self): print ("TestMount:setup() before each test method") def teardown(self): print ("TestMount:teardown() after each test method") @classmethod def setup_class(cls): print ("setup_class() before any methods in this class") _Mounts = [] for name in os.listdir(os.path.dirname(__file__)): if not name.startswith('_') and name.endswith('.py'): name = '.' + os.path.splitext(name)[0] try: module = importlib.import_module(name,'panoptes') _Mounts.append(module) except ImportError as err: self.logger.warn('Failed to load mount plugin: {}'.format(err)) @classmethod def teardown_class(cls): print ("teardown_class() after any methods in this class") def test_is_connected_false(self): pass def test_connect(self): pass def test_is_connected_true(self): pass def test_is_slewing(self): pass def test_check_coordinates(self): pass def test_sync_coordinates(self): pass def test_slew_to_coordinates(self): pass def test_slew_to_park(self): pass def test_echo(self): pass ## Instruction: Test file loops over all the mounts ## Code After: import os import importlib import warnings class TestOptronMount: mount = None def setup(self): print ("TestMount:setup() before each test method") def teardown(self): print ("TestMount:teardown() after each test method") @classmethod def setup_class(cls): mount_dir = os.path.dirname(__file__) + '/../mount/' print ("setup_class() before any methods in this class") _Mounts = [] for name in os.listdir(os.path.dirname(mount_dir)): if not name.startswith('_') and name.endswith('.py'): name = '.' + os.path.splitext(name)[0] try: module = importlib.import_module(name,'panoptes.mount') _Mounts.append(module) except ImportError as err: warnings.warn('Failed to load mount plugin: {}'.format(err)) @classmethod def teardown_class(cls): print ("teardown_class() after any methods in this class") def test_is_connected_false(self): pass def test_connect(self): pass def test_is_connected_true(self): pass def test_is_slewing(self): pass def test_check_coordinates(self): pass def test_sync_coordinates(self): pass def test_slew_to_coordinates(self): pass def test_slew_to_park(self): pass def test_echo(self): pass
... import os import importlib import warnings ... def setup_class(cls): mount_dir = os.path.dirname(__file__) + '/../mount/' print ("setup_class() before any methods in this class") ... _Mounts = [] for name in os.listdir(os.path.dirname(mount_dir)): if not name.startswith('_') and name.endswith('.py'): ... try: module = importlib.import_module(name,'panoptes.mount') _Mounts.append(module) ... except ImportError as err: warnings.warn('Failed to load mount plugin: {}'.format(err)) ...
8898f23a429112cd80e6a2c8321b0de44aeaee7e
blanc_basic_pages/forms.py
blanc_basic_pages/forms.py
from django import forms from django.conf import settings from mptt.forms import MPTTAdminForm from .models import Page TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', ( ('', 'Default'), )) class PageAdminForm(MPTTAdminForm): class Meta: model = Page exclude = () def __init__(self, *args, **kwargs): super(PageAdminForm, self).__init__(*args, **kwargs) # The list of templates is defined in settings, however as we can't have dynamic choices in # models due to migrations - we change the form choices instead. self.fields['template_name'] = forms.ChoiceField(choices=TEMPLATE_CHOICES, required=False)
from django import forms from django.conf import settings from mptt.forms import MPTTAdminForm from .models import Page TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', ( ('', 'Default'), )) class PageAdminForm(MPTTAdminForm): class Meta: model = Page exclude = () widgets = { # The list of templates is defined in settings, however as we can't have dynamic # choices in models due to migrations - we change the form choices instead. 'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES), }
Use custom widget for template choices instead
Use custom widget for template choices instead A bit more Djangonic than tweaking self.fields
Python
bsd-3-clause
blancltd/blanc-basic-pages
from django import forms from django.conf import settings from mptt.forms import MPTTAdminForm from .models import Page TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', ( ('', 'Default'), )) class PageAdminForm(MPTTAdminForm): class Meta: model = Page exclude = () + widgets = { + # The list of templates is defined in settings, however as we can't have dynamic + # choices in models due to migrations - we change the form choices instead. + 'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES), + } - def __init__(self, *args, **kwargs): - super(PageAdminForm, self).__init__(*args, **kwargs) - - # The list of templates is defined in settings, however as we can't have dynamic choices in - # models due to migrations - we change the form choices instead. - self.fields['template_name'] = forms.ChoiceField(choices=TEMPLATE_CHOICES, required=False) -
Use custom widget for template choices instead
## Code Before: from django import forms from django.conf import settings from mptt.forms import MPTTAdminForm from .models import Page TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', ( ('', 'Default'), )) class PageAdminForm(MPTTAdminForm): class Meta: model = Page exclude = () def __init__(self, *args, **kwargs): super(PageAdminForm, self).__init__(*args, **kwargs) # The list of templates is defined in settings, however as we can't have dynamic choices in # models due to migrations - we change the form choices instead. self.fields['template_name'] = forms.ChoiceField(choices=TEMPLATE_CHOICES, required=False) ## Instruction: Use custom widget for template choices instead ## Code After: from django import forms from django.conf import settings from mptt.forms import MPTTAdminForm from .models import Page TEMPLATE_CHOICES = getattr(settings, 'PAGE_TEMPLATES', ( ('', 'Default'), )) class PageAdminForm(MPTTAdminForm): class Meta: model = Page exclude = () widgets = { # The list of templates is defined in settings, however as we can't have dynamic # choices in models due to migrations - we change the form choices instead. 'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES), }
# ... existing code ... exclude = () widgets = { # The list of templates is defined in settings, however as we can't have dynamic # choices in models due to migrations - we change the form choices instead. 'template_name': forms.widgets.Select(choices=TEMPLATE_CHOICES), } # ... rest of the code ...
b63b22678a005baa6195854b65cc1828061febba
vx/mode.py
vx/mode.py
import vx import os.path def mode_from_filename(file): root, ext = os.path.splitext(file) ext = ext if ext else root mode = None if ext == '.c': return c_mode class mode: def __init__(self, window): self.breaks = ('_', ' ', '\n', '\t') self.keywords = () class python_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#') self.keywords = ('return', 'for', 'while', 'break', 'continue', 'def') class c_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#') self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
import vx import os.path def mode_from_filename(file): root, ext = os.path.splitext(file) ext = ext if ext else root mode = None if ext == '.c': return c_mode elif ext == '.py': return python_mode class mode: def __init__(self, window): self.breaks = ('_', ' ', '\n', '\t') self.keywords = () class python_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#') self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import') class c_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#') self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
Add .py extension handling and more python keywords
Add .py extension handling and more python keywords
Python
mit
philipdexter/vx,philipdexter/vx
import vx import os.path def mode_from_filename(file): root, ext = os.path.splitext(file) ext = ext if ext else root mode = None if ext == '.c': return c_mode + elif ext == '.py': + return python_mode class mode: def __init__(self, window): self.breaks = ('_', ' ', '\n', '\t') self.keywords = () class python_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#') - self.keywords = ('return', 'for', 'while', 'break', 'continue', 'def') + self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import') class c_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#') self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
Add .py extension handling and more python keywords
## Code Before: import vx import os.path def mode_from_filename(file): root, ext = os.path.splitext(file) ext = ext if ext else root mode = None if ext == '.c': return c_mode class mode: def __init__(self, window): self.breaks = ('_', ' ', '\n', '\t') self.keywords = () class python_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#') self.keywords = ('return', 'for', 'while', 'break', 'continue', 'def') class c_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#') self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"') ## Instruction: Add .py extension handling and more python keywords ## Code After: import vx import os.path def mode_from_filename(file): root, ext = os.path.splitext(file) ext = ext if ext else root mode = None if ext == '.c': return c_mode elif ext == '.py': return python_mode class mode: def __init__(self, window): self.breaks = ('_', ' ', '\n', '\t') self.keywords = () class python_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '{', '}', '.', ',', '#') self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import') class c_mode(mode): def __init__(self, window): super().__init__(window) self.breaks = ('_', ' ', '\n', '\t', '(', ')', '<', '>', '.', ',', '#') self.keywords = ('#include', '#define', 'if', 'else', 'return', 'goto', 'break', 'continue', r'"(?:[^"\\]|\\.)*"')
// ... existing code ... return c_mode elif ext == '.py': return python_mode // ... modified code ... self.keywords = ('class', 'return', 'for', 'while', 'break', 'continue', 'def', 'from', 'import') // ... rest of the code ...
605443886582d13c2b45b19fad86854bf4e8ddbd
backend/catalogue/serializers.py
backend/catalogue/serializers.py
from rest_framework import serializers from .models import Release, Track, Comment class CommentSerializer(serializers.ModelSerializer): class Meta: model = Comment fields = ('id', 'comment') class TrackSerializer(serializers.ModelSerializer): cdid = serializers.StringRelatedField( read_only=True ) class Meta: model = Track fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'cdid') class ReleaseSerializer(serializers.ModelSerializer): tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks') comments = serializers.HyperlinkedIdentityField(view_name='release-comments') class Meta: model = Release fields = ('id', 'arrivaldate', 'artist', 'title', 'year', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
from rest_framework import serializers from .models import Release, Track, Comment class CommentSerializer(serializers.ModelSerializer): class Meta: model = Comment fields = ('id', 'comment') class TrackSerializer(serializers.ModelSerializer): class Meta: model = Track fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release') class ReleaseSerializer(serializers.ModelSerializer): tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks') comments = serializers.HyperlinkedIdentityField(view_name='release-comments') class Meta: model = Release fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
Add more fields to Release serializer.
Add more fields to Release serializer.
Python
mit
ThreeDRadio/playlists,ThreeDRadio/playlists,ThreeDRadio/playlists
from rest_framework import serializers from .models import Release, Track, Comment class CommentSerializer(serializers.ModelSerializer): class Meta: model = Comment fields = ('id', 'comment') class TrackSerializer(serializers.ModelSerializer): - cdid = serializers.StringRelatedField( - read_only=True - ) class Meta: model = Track - fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'cdid') + fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release') class ReleaseSerializer(serializers.ModelSerializer): tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks') comments = serializers.HyperlinkedIdentityField(view_name='release-comments') class Meta: model = Release - fields = ('id', 'arrivaldate', 'artist', 'title', 'year', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments') + fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
Add more fields to Release serializer.
## Code Before: from rest_framework import serializers from .models import Release, Track, Comment class CommentSerializer(serializers.ModelSerializer): class Meta: model = Comment fields = ('id', 'comment') class TrackSerializer(serializers.ModelSerializer): cdid = serializers.StringRelatedField( read_only=True ) class Meta: model = Track fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'cdid') class ReleaseSerializer(serializers.ModelSerializer): tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks') comments = serializers.HyperlinkedIdentityField(view_name='release-comments') class Meta: model = Release fields = ('id', 'arrivaldate', 'artist', 'title', 'year', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments') ## Instruction: Add more fields to Release serializer. ## Code After: from rest_framework import serializers from .models import Release, Track, Comment class CommentSerializer(serializers.ModelSerializer): class Meta: model = Comment fields = ('id', 'comment') class TrackSerializer(serializers.ModelSerializer): class Meta: model = Track fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release') class ReleaseSerializer(serializers.ModelSerializer): tracks = serializers.HyperlinkedIdentityField(view_name='release-tracks') comments = serializers.HyperlinkedIdentityField(view_name='release-comments') class Meta: model = Release fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments')
// ... existing code ... class TrackSerializer(serializers.ModelSerializer): // ... modified code ... model = Track fields = ('trackid', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'release') ... model = Release fields = ('id', 'arrivaldate', 'artist', 'title', 'year','company','genre','format', 'local', 'cpa', 'compilation', 'female', 'tracks', 'comments') // ... rest of the code ...
c1e1c9d63d5334140aa71c025a90e9500b299307
functional_tests.py
functional_tests.py
from selenium import webdriver browser = webdriver.Firefox() browser.get('http://localhost:8000') assert 'Django' in browser.title
from selenium import webdriver import unittest class NewVisitorTest(unittest.TestCase): def setUp(self): self.browser = webdriver.Firefox() def tearDown(self): self.browser.quit() def test_can_start_a_list_and_retrieve_it_later(self): # Rey has heard about a cool new online to-do app. He goes # to check out its homepage self.browser.get('http://localhost:8000') # He notices the page title and header mention to-do lists self.assertIn('To-Do', self.browser.title) self.fail('Finish the test!') # He is invited to enter a to-do item straight away # He types "Make coffee" into a text box (Rey can't work when he is sleepy) # There is still a text box inviting him to add another item. # He enters "Go to the toilet" (Rey is very methodical) # The page updates again, and now shows both items on his list # Rey wonders whether the site will remember his list. Then he sees that # the site has generated a unique URL for him # He visits that URL - his to-do list is still there # Satisfied, he goes back to sleep if __name__ == '__main__': unittest.main(warnings='ignore')
Update tests to be human-readable
Update tests to be human-readable
Python
apache-2.0
rocity/the-testing-goat,rocity/the-testing-goat
from selenium import webdriver + import unittest + class NewVisitorTest(unittest.TestCase): + def setUp(self): - browser = webdriver.Firefox() + self.browser = webdriver.Firefox() - browser.get('http://localhost:8000') - assert 'Django' in browser.title + def tearDown(self): + self.browser.quit() + + def test_can_start_a_list_and_retrieve_it_later(self): + + # Rey has heard about a cool new online to-do app. He goes + # to check out its homepage + self.browser.get('http://localhost:8000') + + # He notices the page title and header mention to-do lists + self.assertIn('To-Do', self.browser.title) + self.fail('Finish the test!') + + # He is invited to enter a to-do item straight away + + # He types "Make coffee" into a text box (Rey can't work when he is sleepy) + + # There is still a text box inviting him to add another item. + # He enters "Go to the toilet" (Rey is very methodical) + + # The page updates again, and now shows both items on his list + + # Rey wonders whether the site will remember his list. Then he sees that + # the site has generated a unique URL for him + + # He visits that URL - his to-do list is still there + + # Satisfied, he goes back to sleep + + if __name__ == '__main__': + unittest.main(warnings='ignore')
Update tests to be human-readable
## Code Before: from selenium import webdriver browser = webdriver.Firefox() browser.get('http://localhost:8000') assert 'Django' in browser.title ## Instruction: Update tests to be human-readable ## Code After: from selenium import webdriver import unittest class NewVisitorTest(unittest.TestCase): def setUp(self): self.browser = webdriver.Firefox() def tearDown(self): self.browser.quit() def test_can_start_a_list_and_retrieve_it_later(self): # Rey has heard about a cool new online to-do app. He goes # to check out its homepage self.browser.get('http://localhost:8000') # He notices the page title and header mention to-do lists self.assertIn('To-Do', self.browser.title) self.fail('Finish the test!') # He is invited to enter a to-do item straight away # He types "Make coffee" into a text box (Rey can't work when he is sleepy) # There is still a text box inviting him to add another item. # He enters "Go to the toilet" (Rey is very methodical) # The page updates again, and now shows both items on his list # Rey wonders whether the site will remember his list. Then he sees that # the site has generated a unique URL for him # He visits that URL - his to-do list is still there # Satisfied, he goes back to sleep if __name__ == '__main__': unittest.main(warnings='ignore')
# ... existing code ... from selenium import webdriver import unittest class NewVisitorTest(unittest.TestCase): def setUp(self): self.browser = webdriver.Firefox() def tearDown(self): self.browser.quit() def test_can_start_a_list_and_retrieve_it_later(self): # Rey has heard about a cool new online to-do app. He goes # to check out its homepage self.browser.get('http://localhost:8000') # He notices the page title and header mention to-do lists self.assertIn('To-Do', self.browser.title) self.fail('Finish the test!') # He is invited to enter a to-do item straight away # He types "Make coffee" into a text box (Rey can't work when he is sleepy) # There is still a text box inviting him to add another item. # He enters "Go to the toilet" (Rey is very methodical) # The page updates again, and now shows both items on his list # Rey wonders whether the site will remember his list. Then he sees that # the site has generated a unique URL for him # He visits that URL - his to-do list is still there # Satisfied, he goes back to sleep if __name__ == '__main__': unittest.main(warnings='ignore') # ... rest of the code ...
24da0f84e1a844b1f53e5afafc34cfcc915a9a67
corehq/apps/userreports/tests/test_report_rendering.py
corehq/apps/userreports/tests/test_report_rendering.py
from __future__ import absolute_import from __future__ import unicode_literals from django.test import SimpleTestCase from corehq.apps.userreports.reports.view import ConfigurableReportView class VeryFakeReportView(ConfigurableReportView): # note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing def __init__(self, data): self._data = data @property def export_table(self): return self._data class ReportRenderingTest(SimpleTestCase): def test_email_response_unicode(self): report = VeryFakeReportView(data=[ ['hello', 'हिन्दी'] ]) # this used to fail: https://manage.dimagi.com/default.asp?263803 report.email_response
from __future__ import absolute_import from __future__ import unicode_literals from django.test import SimpleTestCase from corehq.apps.userreports.reports.view import ConfigurableReportView from corehq.apps.userreports.reports.util import ReportExport class VeryFakeReportExport(ReportExport): def __init__(self, data): self._data = data def get_table(self): return self._data class VeryFakeReportView(ConfigurableReportView): # note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing def __init__(self, data): self._data = data @property def report_export(self): return VeryFakeReportExport(self._data) class ReportRenderingTest(SimpleTestCase): def test_email_response_unicode(self): report = VeryFakeReportView(data=[ ['hello', 'हिन्दी'] ]) # this used to fail: https://manage.dimagi.com/default.asp?263803 report.email_response
Update report_rendering test to use ReportExport
Update report_rendering test to use ReportExport
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
from __future__ import absolute_import from __future__ import unicode_literals + from django.test import SimpleTestCase + from corehq.apps.userreports.reports.view import ConfigurableReportView + from corehq.apps.userreports.reports.util import ReportExport + + + class VeryFakeReportExport(ReportExport): + def __init__(self, data): + self._data = data + + def get_table(self): + return self._data class VeryFakeReportView(ConfigurableReportView): # note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing def __init__(self, data): self._data = data @property - def export_table(self): + def report_export(self): - return self._data + return VeryFakeReportExport(self._data) class ReportRenderingTest(SimpleTestCase): def test_email_response_unicode(self): report = VeryFakeReportView(data=[ ['hello', 'हिन्दी'] ]) # this used to fail: https://manage.dimagi.com/default.asp?263803 report.email_response
Update report_rendering test to use ReportExport
## Code Before: from __future__ import absolute_import from __future__ import unicode_literals from django.test import SimpleTestCase from corehq.apps.userreports.reports.view import ConfigurableReportView class VeryFakeReportView(ConfigurableReportView): # note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing def __init__(self, data): self._data = data @property def export_table(self): return self._data class ReportRenderingTest(SimpleTestCase): def test_email_response_unicode(self): report = VeryFakeReportView(data=[ ['hello', 'हिन्दी'] ]) # this used to fail: https://manage.dimagi.com/default.asp?263803 report.email_response ## Instruction: Update report_rendering test to use ReportExport ## Code After: from __future__ import absolute_import from __future__ import unicode_literals from django.test import SimpleTestCase from corehq.apps.userreports.reports.view import ConfigurableReportView from corehq.apps.userreports.reports.util import ReportExport class VeryFakeReportExport(ReportExport): def __init__(self, data): self._data = data def get_table(self): return self._data class VeryFakeReportView(ConfigurableReportView): # note: this is very coupled to what it tests below, but it beats bootstrapping a whole UCR thing def __init__(self, data): self._data = data @property def report_export(self): return VeryFakeReportExport(self._data) class ReportRenderingTest(SimpleTestCase): def test_email_response_unicode(self): report = VeryFakeReportView(data=[ ['hello', 'हिन्दी'] ]) # this used to fail: https://manage.dimagi.com/default.asp?263803 report.email_response
# ... existing code ... from __future__ import unicode_literals from django.test import SimpleTestCase from corehq.apps.userreports.reports.view import ConfigurableReportView from corehq.apps.userreports.reports.util import ReportExport class VeryFakeReportExport(ReportExport): def __init__(self, data): self._data = data def get_table(self): return self._data # ... modified code ... @property def report_export(self): return VeryFakeReportExport(self._data) # ... rest of the code ...
6f0a35372d625f923b9093194540cf0b0e9f054d
platformio_api/__init__.py
platformio_api/__init__.py
import json import logging.config import os from time import tzset VERSION = (0, 3, 0) __version__ = ".".join([str(s) for s in VERSION]) __title__ = "platformio-api" __description__ = ("An API for PlatformIO") __url__ = "https://github.com/ivankravets/platformio-api" __author__ = "Ivan Kravets" __email__ = "[email protected]" __license__ = "MIT License" __copyright__ = "Copyright (C) 2014-2015 Ivan Kravets" config = dict( SQLALCHEMY_DATABASE_URI=None, GITHUB_LOGIN=None, GITHUB_PASSWORD=None, DL_PIO_DIR=None, DL_PIO_URL=None, MAX_DLFILE_SIZE=1024*1024*10, LOGGING=dict(version=1) ) assert "PIOAPI_CONFIG_PATH" in os.environ with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f: config.update(json.load(f)) # configure logging for packages logging.basicConfig() logging.config.dictConfig(config['LOGGING']) # setup time zone to UTC globally os.environ['TZ'] = "+00:00" tzset()
import json import logging.config import os from time import tzset VERSION = (0, 3, 0) __version__ = ".".join([str(s) for s in VERSION]) __title__ = "platformio-api" __description__ = ("An API for PlatformIO") __url__ = "https://github.com/ivankravets/platformio-api" __author__ = "Ivan Kravets" __email__ = "[email protected]" __license__ = "MIT License" __copyright__ = "Copyright (C) 2014-2015 Ivan Kravets" config = dict( SQLALCHEMY_DATABASE_URI=None, GITHUB_LOGIN=None, GITHUB_PASSWORD=None, DL_PIO_DIR=None, DL_PIO_URL=None, MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb LOGGING=dict(version=1) ) assert "PIOAPI_CONFIG_PATH" in os.environ with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f: config.update(json.load(f)) # configure logging for packages logging.basicConfig() logging.config.dictConfig(config['LOGGING']) # setup time zone to UTC globally os.environ['TZ'] = "+00:00" tzset()
Increase repo size to 20Mb
Increase repo size to 20Mb
Python
apache-2.0
orgkhnargh/platformio-api,platformio/platformio-api
import json import logging.config import os from time import tzset VERSION = (0, 3, 0) __version__ = ".".join([str(s) for s in VERSION]) __title__ = "platformio-api" __description__ = ("An API for PlatformIO") __url__ = "https://github.com/ivankravets/platformio-api" __author__ = "Ivan Kravets" __email__ = "[email protected]" __license__ = "MIT License" __copyright__ = "Copyright (C) 2014-2015 Ivan Kravets" config = dict( SQLALCHEMY_DATABASE_URI=None, GITHUB_LOGIN=None, GITHUB_PASSWORD=None, DL_PIO_DIR=None, DL_PIO_URL=None, - MAX_DLFILE_SIZE=1024*1024*10, + MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb LOGGING=dict(version=1) ) assert "PIOAPI_CONFIG_PATH" in os.environ with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f: config.update(json.load(f)) # configure logging for packages logging.basicConfig() logging.config.dictConfig(config['LOGGING']) # setup time zone to UTC globally os.environ['TZ'] = "+00:00" tzset()
Increase repo size to 20Mb
## Code Before: import json import logging.config import os from time import tzset VERSION = (0, 3, 0) __version__ = ".".join([str(s) for s in VERSION]) __title__ = "platformio-api" __description__ = ("An API for PlatformIO") __url__ = "https://github.com/ivankravets/platformio-api" __author__ = "Ivan Kravets" __email__ = "[email protected]" __license__ = "MIT License" __copyright__ = "Copyright (C) 2014-2015 Ivan Kravets" config = dict( SQLALCHEMY_DATABASE_URI=None, GITHUB_LOGIN=None, GITHUB_PASSWORD=None, DL_PIO_DIR=None, DL_PIO_URL=None, MAX_DLFILE_SIZE=1024*1024*10, LOGGING=dict(version=1) ) assert "PIOAPI_CONFIG_PATH" in os.environ with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f: config.update(json.load(f)) # configure logging for packages logging.basicConfig() logging.config.dictConfig(config['LOGGING']) # setup time zone to UTC globally os.environ['TZ'] = "+00:00" tzset() ## Instruction: Increase repo size to 20Mb ## Code After: import json import logging.config import os from time import tzset VERSION = (0, 3, 0) __version__ = ".".join([str(s) for s in VERSION]) __title__ = "platformio-api" __description__ = ("An API for PlatformIO") __url__ = "https://github.com/ivankravets/platformio-api" __author__ = "Ivan Kravets" __email__ = "[email protected]" __license__ = "MIT License" __copyright__ = "Copyright (C) 2014-2015 Ivan Kravets" config = dict( SQLALCHEMY_DATABASE_URI=None, GITHUB_LOGIN=None, GITHUB_PASSWORD=None, DL_PIO_DIR=None, DL_PIO_URL=None, MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb LOGGING=dict(version=1) ) assert "PIOAPI_CONFIG_PATH" in os.environ with open(os.environ.get("PIOAPI_CONFIG_PATH")) as f: config.update(json.load(f)) # configure logging for packages logging.basicConfig() logging.config.dictConfig(config['LOGGING']) # setup time zone to UTC globally os.environ['TZ'] = "+00:00" tzset()
... DL_PIO_URL=None, MAX_DLFILE_SIZE=1024*1024*20, # 20 Mb LOGGING=dict(version=1) ...
4a827bfff24758677e9c1d9d3b186fc14f23e0bb
lib/oeqa/runtime/cases/parselogs_rpi.py
lib/oeqa/runtime/cases/parselogs_rpi.py
from oeqa.runtime.cases.parselogs import * rpi_errors = [ 'bcmgenet fd580000.genet: failed to get enet-eee clock', 'bcmgenet fd580000.genet: failed to get enet-wol clock', 'bcmgenet fd580000.genet: failed to get enet clock', 'bcmgenet fd580000.ethernet: failed to get enet-eee clock', 'bcmgenet fd580000.ethernet: failed to get enet-wol clock', 'bcmgenet fd580000.ethernet: failed to get enet clock', ] ignore_errors['raspberrypi4'] = rpi_errors + common_errors ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors ignore_errors['raspberrypi3'] = rpi_errors + common_errors ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors class ParseLogsTestRpi(ParseLogsTest): pass
from oeqa.runtime.cases.parselogs import * rpi_errors = [ ] ignore_errors['raspberrypi4'] = rpi_errors + common_errors ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors ignore_errors['raspberrypi3'] = rpi_errors + common_errors ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors class ParseLogsTestRpi(ParseLogsTest): pass
Update the error regexps to 5.10 kernel
parselogs: Update the error regexps to 5.10 kernel The old messages are no longer necessary Signed-off-by: Khem Raj <[email protected]>
Python
mit
agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi,agherzan/meta-raspberrypi
from oeqa.runtime.cases.parselogs import * rpi_errors = [ - 'bcmgenet fd580000.genet: failed to get enet-eee clock', - 'bcmgenet fd580000.genet: failed to get enet-wol clock', - 'bcmgenet fd580000.genet: failed to get enet clock', - 'bcmgenet fd580000.ethernet: failed to get enet-eee clock', - 'bcmgenet fd580000.ethernet: failed to get enet-wol clock', - 'bcmgenet fd580000.ethernet: failed to get enet clock', ] ignore_errors['raspberrypi4'] = rpi_errors + common_errors ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors ignore_errors['raspberrypi3'] = rpi_errors + common_errors ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors class ParseLogsTestRpi(ParseLogsTest): pass
Update the error regexps to 5.10 kernel
## Code Before: from oeqa.runtime.cases.parselogs import * rpi_errors = [ 'bcmgenet fd580000.genet: failed to get enet-eee clock', 'bcmgenet fd580000.genet: failed to get enet-wol clock', 'bcmgenet fd580000.genet: failed to get enet clock', 'bcmgenet fd580000.ethernet: failed to get enet-eee clock', 'bcmgenet fd580000.ethernet: failed to get enet-wol clock', 'bcmgenet fd580000.ethernet: failed to get enet clock', ] ignore_errors['raspberrypi4'] = rpi_errors + common_errors ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors ignore_errors['raspberrypi3'] = rpi_errors + common_errors ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors class ParseLogsTestRpi(ParseLogsTest): pass ## Instruction: Update the error regexps to 5.10 kernel ## Code After: from oeqa.runtime.cases.parselogs import * rpi_errors = [ ] ignore_errors['raspberrypi4'] = rpi_errors + common_errors ignore_errors['raspberrypi4-64'] = rpi_errors + common_errors ignore_errors['raspberrypi3'] = rpi_errors + common_errors ignore_errors['raspberrypi3-64'] = rpi_errors + common_errors class ParseLogsTestRpi(ParseLogsTest): pass
# ... existing code ... rpi_errors = [ ] # ... rest of the code ...
62ba442ac447dbb4482dd15f70075d224d0e5a0e
scripts/test_conda_build_log.py
scripts/test_conda_build_log.py
import pytest import log_parser import os @pytest.fixture def parsed_log(): logname = os.path.join(os.path.split(os.path.abspath(__file__))[0], 'build.log') gen = list(log_parser.read_log_from_script(logname)) parsed = {built_name: log_parser.parse_conda_build(lines) for name, built_name, lines in gen} return parsed def test_parse_conda_build(parsed_log): # make sure that we have at least one thing that was parsed assert len(parsed_log) >= 1 def test_parse_init(parsed_log): # make sure we are getting the build command out of every single entry for pkg_name, parsed in parsed_log.items(): parsed_init = log_parser.parse_init(parsed['init']) assert 'build_command' in parsed_init def test_parse_build(parsed_log): # make sure we are getting either an error or the build string out of the # build section for pkg_name, parsed in parsed_log.items(): if 'build' not in parsed: # not all packages will successfully build continue # if there is a build section, then parse it parsed_build = log_parser.parse_build(parsed['build']) if parsed_build['built_name'] == 'failed': assert parsed_build['error'] != [] else: assert parsed_build['error'] == []
import pytest import log_parser import os @pytest.fixture def parsed_log(): logname = os.path.join(os.path.split(os.path.abspath(__file__))[0], 'build.log') gen = list(log_parser.read_log_from_script(logname)) parsed = {built_name: log_parser.parse_conda_build(lines) for name, built_name, lines in gen} return parsed def test_parse_conda_build(parsed_log): # make sure that we have at least one thing that was parsed assert len(parsed_log) >= 1 def test_parse_init(parsed_log): # make sure we are getting the build command out of every single entry for pkg_name, parsed in parsed_log.items(): parsed_init = log_parser.parse_init(parsed['init']) assert 'build_command' in parsed_init assert 'err' in parsed_init def test_parse_build(parsed_log): # make sure we are getting either an error or the build string out of the # build section for pkg_name, parsed in parsed_log.items(): if 'build' not in parsed: # not all packages will successfully build continue # if there is a build section, then parse it parsed_build = log_parser.parse_build(parsed['build']) if parsed_build['built_name'] == 'failed': assert parsed_build['error'] != [] else: assert parsed_build['error'] == []
Make sure there is an error field
TST: Make sure there is an error field
Python
bsd-3-clause
NSLS-II/lightsource2-recipes,NSLS-II/auto-build-tagged-recipes,NSLS-II/lightsource2-recipes,NSLS-II/auto-build-tagged-recipes,NSLS-II/lightsource2-recipes,NSLS-II/lightsource2-recipes
import pytest import log_parser import os @pytest.fixture def parsed_log(): logname = os.path.join(os.path.split(os.path.abspath(__file__))[0], 'build.log') gen = list(log_parser.read_log_from_script(logname)) parsed = {built_name: log_parser.parse_conda_build(lines) for name, built_name, lines in gen} return parsed def test_parse_conda_build(parsed_log): # make sure that we have at least one thing that was parsed assert len(parsed_log) >= 1 def test_parse_init(parsed_log): # make sure we are getting the build command out of every single entry for pkg_name, parsed in parsed_log.items(): parsed_init = log_parser.parse_init(parsed['init']) assert 'build_command' in parsed_init + assert 'err' in parsed_init def test_parse_build(parsed_log): # make sure we are getting either an error or the build string out of the # build section for pkg_name, parsed in parsed_log.items(): if 'build' not in parsed: # not all packages will successfully build continue # if there is a build section, then parse it parsed_build = log_parser.parse_build(parsed['build']) if parsed_build['built_name'] == 'failed': assert parsed_build['error'] != [] else: assert parsed_build['error'] == []
Make sure there is an error field
## Code Before: import pytest import log_parser import os @pytest.fixture def parsed_log(): logname = os.path.join(os.path.split(os.path.abspath(__file__))[0], 'build.log') gen = list(log_parser.read_log_from_script(logname)) parsed = {built_name: log_parser.parse_conda_build(lines) for name, built_name, lines in gen} return parsed def test_parse_conda_build(parsed_log): # make sure that we have at least one thing that was parsed assert len(parsed_log) >= 1 def test_parse_init(parsed_log): # make sure we are getting the build command out of every single entry for pkg_name, parsed in parsed_log.items(): parsed_init = log_parser.parse_init(parsed['init']) assert 'build_command' in parsed_init def test_parse_build(parsed_log): # make sure we are getting either an error or the build string out of the # build section for pkg_name, parsed in parsed_log.items(): if 'build' not in parsed: # not all packages will successfully build continue # if there is a build section, then parse it parsed_build = log_parser.parse_build(parsed['build']) if parsed_build['built_name'] == 'failed': assert parsed_build['error'] != [] else: assert parsed_build['error'] == [] ## Instruction: Make sure there is an error field ## Code After: import pytest import log_parser import os @pytest.fixture def parsed_log(): logname = os.path.join(os.path.split(os.path.abspath(__file__))[0], 'build.log') gen = list(log_parser.read_log_from_script(logname)) parsed = {built_name: log_parser.parse_conda_build(lines) for name, built_name, lines in gen} return parsed def test_parse_conda_build(parsed_log): # make sure that we have at least one thing that was parsed assert len(parsed_log) >= 1 def test_parse_init(parsed_log): # make sure we are getting the build command out of every single entry for pkg_name, parsed in parsed_log.items(): parsed_init = log_parser.parse_init(parsed['init']) assert 'build_command' in parsed_init assert 'err' in parsed_init def test_parse_build(parsed_log): # make sure we are getting either an error or the build string out of the # build section for pkg_name, parsed in parsed_log.items(): if 'build' not in parsed: # not all packages will successfully build continue # if there is a build section, then parse it parsed_build = log_parser.parse_build(parsed['build']) if parsed_build['built_name'] == 'failed': assert parsed_build['error'] != [] else: assert parsed_build['error'] == []
// ... existing code ... assert 'build_command' in parsed_init assert 'err' in parsed_init // ... rest of the code ...
e5ac63b4615b4166d7e7866c9f169e4c9f86f46c
setup.py
setup.py
from distutils.core import setup setup( name='django-emailuser', version='1.0', description='simple User model identified by email address', packages=['emailuser'], author='Mark Paschal', author_email='[email protected]', url='https://github.com/duncaningram/django-emailuser', classifiers=[ 'Framework :: Django', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', ], )
from distutils.core import setup setup( name='django-emailuser', version='1.0', description='simple User model identified by email address', packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'], author='Mark Paschal', author_email='[email protected]', url='https://github.com/duncaningram/django-emailuser', classifiers=[ 'Framework :: Django', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', ], )
Install the management command too when installing as a distribution
Install the management command too when installing as a distribution
Python
mit
markpasc/django-emailuser,duncaningram/django-emailuser
from distutils.core import setup setup( name='django-emailuser', version='1.0', description='simple User model identified by email address', - packages=['emailuser'], + packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'], author='Mark Paschal', author_email='[email protected]', url='https://github.com/duncaningram/django-emailuser', classifiers=[ 'Framework :: Django', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', ], )
Install the management command too when installing as a distribution
## Code Before: from distutils.core import setup setup( name='django-emailuser', version='1.0', description='simple User model identified by email address', packages=['emailuser'], author='Mark Paschal', author_email='[email protected]', url='https://github.com/duncaningram/django-emailuser', classifiers=[ 'Framework :: Django', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', ], ) ## Instruction: Install the management command too when installing as a distribution ## Code After: from distutils.core import setup setup( name='django-emailuser', version='1.0', description='simple User model identified by email address', packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'], author='Mark Paschal', author_email='[email protected]', url='https://github.com/duncaningram/django-emailuser', classifiers=[ 'Framework :: Django', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 3', ], )
# ... existing code ... description='simple User model identified by email address', packages=['emailuser', 'emailuser.management', 'emailuser.management.commands'], # ... rest of the code ...
839d884d3dca3e799a235b1d2d69acf998f520f9
barsystem_base/management/commands/import_people.py
barsystem_base/management/commands/import_people.py
from django.core.management.base import BaseCommand, CommandError from barsystem_base.models import Person class Command(BaseCommand): args = '<filename>' help = 'Import list of people' csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',') def handle(self, *args, **kwargs): if len(args) == 0: raise CommandError('Please supply filename') with open(args[0], 'r') as f: columns = None for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']: # print(line) # take header if columns is None: columns = line continue values = dict(zip(columns, line)) values['active'] = values['type'] != 'hidden' try: p = Person.objects.get(id=values['id']) except Person.DoesNotExist: p = Person() for key, val in values.items(): if hasattr(p, key): setattr(p, key, val) print(p) p.save() print('Done')
from django.core.management.base import BaseCommand, CommandError from barsystem_base.models import Person, Token class Command(BaseCommand): args = '<filename>' help = 'Import list of people' csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',') def handle(self, *args, **kwargs): if len(args) == 0: raise CommandError('Please supply filename') with open(args[0], 'r') as f: columns = None for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']: # print(line) # take header if columns is None: columns = line continue values = dict(zip(columns, line)) values['active'] = values['type'] != 'hidden' try: p = Person.objects.get(id=values['id']) except Person.DoesNotExist: p = Person() for key, val in values.items(): if hasattr(p, key): setattr(p, key, val) print(p) p.save() t = Token() t.type = 'ibutton' t.value = values['token'] t.person = p t.save() print('Done')
Add ibutton when importing old people
Add ibutton when importing old people
Python
mit
TkkrLab/barsystem,TkkrLab/barsystem,TkkrLab/barsystem
from django.core.management.base import BaseCommand, CommandError - from barsystem_base.models import Person + from barsystem_base.models import Person, Token class Command(BaseCommand): args = '<filename>' help = 'Import list of people' csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',') def handle(self, *args, **kwargs): if len(args) == 0: raise CommandError('Please supply filename') with open(args[0], 'r') as f: columns = None for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']: # print(line) # take header if columns is None: columns = line continue values = dict(zip(columns, line)) values['active'] = values['type'] != 'hidden' try: p = Person.objects.get(id=values['id']) except Person.DoesNotExist: p = Person() for key, val in values.items(): if hasattr(p, key): setattr(p, key, val) print(p) p.save() + t = Token() + t.type = 'ibutton' + t.value = values['token'] + t.person = p + t.save() print('Done')
Add ibutton when importing old people
## Code Before: from django.core.management.base import BaseCommand, CommandError from barsystem_base.models import Person class Command(BaseCommand): args = '<filename>' help = 'Import list of people' csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',') def handle(self, *args, **kwargs): if len(args) == 0: raise CommandError('Please supply filename') with open(args[0], 'r') as f: columns = None for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']: # print(line) # take header if columns is None: columns = line continue values = dict(zip(columns, line)) values['active'] = values['type'] != 'hidden' try: p = Person.objects.get(id=values['id']) except Person.DoesNotExist: p = Person() for key, val in values.items(): if hasattr(p, key): setattr(p, key, val) print(p) p.save() print('Done') ## Instruction: Add ibutton when importing old people ## Code After: from django.core.management.base import BaseCommand, CommandError from barsystem_base.models import Person, Token class Command(BaseCommand): args = '<filename>' help = 'Import list of people' csv_columns = 'id,first_name,last_name,nick_name,amount,type,token'.split(',') def handle(self, *args, **kwargs): if len(args) == 0: raise CommandError('Please supply filename') with open(args[0], 'r') as f: columns = None for line in [line.strip().split(',') for line in f.readlines() if line[0] != '#']: # print(line) # take header if columns is None: columns = line continue values = dict(zip(columns, line)) values['active'] = values['type'] != 'hidden' try: p = Person.objects.get(id=values['id']) except Person.DoesNotExist: p = Person() for key, val in values.items(): if hasattr(p, key): setattr(p, key, val) print(p) p.save() t = Token() t.type = 'ibutton' t.value = values['token'] t.person = p t.save() print('Done')
... from django.core.management.base import BaseCommand, CommandError from barsystem_base.models import Person, Token ... p.save() t = Token() t.type = 'ibutton' t.value = values['token'] t.person = p t.save() print('Done') ...
f764b52558cb02b8e31b9695a724e4c4e80872dd
iscc_bench/readers/__init__.py
iscc_bench/readers/__init__.py
from iscc_bench.readers.bxbooks import bxbooks from iscc_bench.readers.dnbrdf import dnbrdf from iscc_bench.readers.harvard import harvard from iscc_bench.readers.openlibrary import openlibrary from iscc_bench.readers.libgen import libgen ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen)
from iscc_bench.readers.bxbooks import bxbooks from iscc_bench.readers.dnbrdf import dnbrdf from iscc_bench.readers.harvard import harvard from iscc_bench.readers.openlibrary import openlibrary from iscc_bench.readers.libgen import libgen from iscc_bench.readers.caltech101 import caltech_101 from iscc_bench.readers.caltech256 import caltech_256 ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen) ALL_IMAGE_READERS = (caltech_101, caltech_256)
Add image readers to package scope
Add image readers to package scope
Python
bsd-2-clause
coblo/isccbench
from iscc_bench.readers.bxbooks import bxbooks from iscc_bench.readers.dnbrdf import dnbrdf from iscc_bench.readers.harvard import harvard from iscc_bench.readers.openlibrary import openlibrary from iscc_bench.readers.libgen import libgen - + from iscc_bench.readers.caltech101 import caltech_101 + from iscc_bench.readers.caltech256 import caltech_256 ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen) + ALL_IMAGE_READERS = (caltech_101, caltech_256)
Add image readers to package scope
## Code Before: from iscc_bench.readers.bxbooks import bxbooks from iscc_bench.readers.dnbrdf import dnbrdf from iscc_bench.readers.harvard import harvard from iscc_bench.readers.openlibrary import openlibrary from iscc_bench.readers.libgen import libgen ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen) ## Instruction: Add image readers to package scope ## Code After: from iscc_bench.readers.bxbooks import bxbooks from iscc_bench.readers.dnbrdf import dnbrdf from iscc_bench.readers.harvard import harvard from iscc_bench.readers.openlibrary import openlibrary from iscc_bench.readers.libgen import libgen from iscc_bench.readers.caltech101 import caltech_101 from iscc_bench.readers.caltech256 import caltech_256 ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen) ALL_IMAGE_READERS = (caltech_101, caltech_256)
# ... existing code ... from iscc_bench.readers.libgen import libgen from iscc_bench.readers.caltech101 import caltech_101 from iscc_bench.readers.caltech256 import caltech_256 # ... modified code ... ALL_READERS = (bxbooks, dnbrdf, harvard, openlibrary, libgen) ALL_IMAGE_READERS = (caltech_101, caltech_256) # ... rest of the code ...
0c3529bd264d5512e31d828c65676baff6edefa6
pinax/waitinglist/templatetags/pinax_waitinglist_tags.py
pinax/waitinglist/templatetags/pinax_waitinglist_tags.py
from django import template from ..forms import WaitingListEntryForm register = template.Library() @register.assignment_tag def waitinglist_entry_form(): """ Get a (new) form object to post a new comment. Syntax:: {% waitinglist_entry_form as [varname] %} """ return WaitingListEntryForm()
from django import template from ..forms import WaitingListEntryForm register = template.Library() @register.simple_tag(takes_context=True) def waitinglist_entry_form(context): """ Get a (new) form object to post a new comment. Syntax:: {% waitinglist_entry_form as [varname] %} """ initial = {} if "request" in context: initial.update({ "referrer": context["request"].META.get("HTTP_REFERER", ""), "campaign": context["request"].GET.get("wlc", "") }) return WaitingListEntryForm(initial=initial)
Update template tag to also take context
Update template tag to also take context
Python
mit
pinax/pinax-waitinglist,pinax/pinax-waitinglist
from django import template from ..forms import WaitingListEntryForm register = template.Library() - @register.assignment_tag + @register.simple_tag(takes_context=True) - def waitinglist_entry_form(): + def waitinglist_entry_form(context): """ Get a (new) form object to post a new comment. Syntax:: {% waitinglist_entry_form as [varname] %} """ + initial = {} + if "request" in context: + initial.update({ + "referrer": context["request"].META.get("HTTP_REFERER", ""), + "campaign": context["request"].GET.get("wlc", "") + }) - return WaitingListEntryForm() + return WaitingListEntryForm(initial=initial)
Update template tag to also take context
## Code Before: from django import template from ..forms import WaitingListEntryForm register = template.Library() @register.assignment_tag def waitinglist_entry_form(): """ Get a (new) form object to post a new comment. Syntax:: {% waitinglist_entry_form as [varname] %} """ return WaitingListEntryForm() ## Instruction: Update template tag to also take context ## Code After: from django import template from ..forms import WaitingListEntryForm register = template.Library() @register.simple_tag(takes_context=True) def waitinglist_entry_form(context): """ Get a (new) form object to post a new comment. Syntax:: {% waitinglist_entry_form as [varname] %} """ initial = {} if "request" in context: initial.update({ "referrer": context["request"].META.get("HTTP_REFERER", ""), "campaign": context["request"].GET.get("wlc", "") }) return WaitingListEntryForm(initial=initial)
// ... existing code ... @register.simple_tag(takes_context=True) def waitinglist_entry_form(context): """ // ... modified code ... """ initial = {} if "request" in context: initial.update({ "referrer": context["request"].META.get("HTTP_REFERER", ""), "campaign": context["request"].GET.get("wlc", "") }) return WaitingListEntryForm(initial=initial) // ... rest of the code ...
c30347f34967ee7634676e0e4e27164910f9e52b
regparser/tree/xml_parser/note_processor.py
regparser/tree/xml_parser/note_processor.py
from regparser.tree.depth import markers as mtypes, optional_rules from regparser.tree.struct import Node from regparser.tree.xml_parser import ( paragraph_processor, simple_hierarchy_processor) class IgnoreNotesHeader(paragraph_processor.BaseMatcher): """We don't want to include "Note:" and "Notes:" headers""" def matches(self, xml): return xml.tag == 'HD' and xml.text.lower().startswith('note') def derive_nodes(self, xml, processor=None): return [] class NoteProcessor(paragraph_processor.ParagraphProcessor): MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(), IgnoreNotesHeader(), paragraph_processor.IgnoreTagMatcher('PRTPAGE')] def additional_constraints(self): return [optional_rules.limit_paragraph_types( mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)] class NoteMatcher(paragraph_processor.BaseMatcher): """Processes the contents of NOTE and NOTES tags using a NoteProcessor""" def matches(self, xml): return xml.tag in ('NOTE', 'NOTES') def derive_nodes(self, xml, processor=None): processor = NoteProcessor() node = Node(label=[mtypes.MARKERLESS], source_xml=xml, node_type=Node.NOTE) return [processor.process(xml, node)]
import re from regparser.tree.depth import markers as mtypes, optional_rules from regparser.tree.struct import Node from regparser.tree.xml_parser import ( paragraph_processor, simple_hierarchy_processor) class IgnoreNotesHeader(paragraph_processor.BaseMatcher): """We don't want to include "Note:" and "Notes:" headers""" REGEX = re.compile('notes?:?\s*$', re.IGNORECASE) def matches(self, xml): return xml.tag == 'HD' and self.REGEX.match(xml.text or '') def derive_nodes(self, xml, processor=None): return [] class NoteProcessor(paragraph_processor.ParagraphProcessor): MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(), IgnoreNotesHeader(), paragraph_processor.IgnoreTagMatcher('PRTPAGE')] def additional_constraints(self): return [optional_rules.limit_paragraph_types( mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)] class NoteMatcher(paragraph_processor.BaseMatcher): """Processes the contents of NOTE and NOTES tags using a NoteProcessor""" def matches(self, xml): return xml.tag in ('NOTE', 'NOTES') def derive_nodes(self, xml, processor=None): processor = NoteProcessor() node = Node(label=[mtypes.MARKERLESS], source_xml=xml, node_type=Node.NOTE) return [processor.process(xml, node)]
Use regex rather than string match for Note:
Use regex rather than string match for Note: Per suggestion from @tadhg-ohiggins
Python
cc0-1.0
eregs/regulations-parser,tadhg-ohiggins/regulations-parser,tadhg-ohiggins/regulations-parser,cmc333333/regulations-parser,cmc333333/regulations-parser,eregs/regulations-parser
+ import re + from regparser.tree.depth import markers as mtypes, optional_rules from regparser.tree.struct import Node from regparser.tree.xml_parser import ( paragraph_processor, simple_hierarchy_processor) class IgnoreNotesHeader(paragraph_processor.BaseMatcher): """We don't want to include "Note:" and "Notes:" headers""" + REGEX = re.compile('notes?:?\s*$', re.IGNORECASE) + def matches(self, xml): - return xml.tag == 'HD' and xml.text.lower().startswith('note') + return xml.tag == 'HD' and self.REGEX.match(xml.text or '') def derive_nodes(self, xml, processor=None): return [] class NoteProcessor(paragraph_processor.ParagraphProcessor): MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(), IgnoreNotesHeader(), paragraph_processor.IgnoreTagMatcher('PRTPAGE')] def additional_constraints(self): return [optional_rules.limit_paragraph_types( mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)] class NoteMatcher(paragraph_processor.BaseMatcher): """Processes the contents of NOTE and NOTES tags using a NoteProcessor""" def matches(self, xml): return xml.tag in ('NOTE', 'NOTES') def derive_nodes(self, xml, processor=None): processor = NoteProcessor() node = Node(label=[mtypes.MARKERLESS], source_xml=xml, node_type=Node.NOTE) return [processor.process(xml, node)]
Use regex rather than string match for Note:
## Code Before: from regparser.tree.depth import markers as mtypes, optional_rules from regparser.tree.struct import Node from regparser.tree.xml_parser import ( paragraph_processor, simple_hierarchy_processor) class IgnoreNotesHeader(paragraph_processor.BaseMatcher): """We don't want to include "Note:" and "Notes:" headers""" def matches(self, xml): return xml.tag == 'HD' and xml.text.lower().startswith('note') def derive_nodes(self, xml, processor=None): return [] class NoteProcessor(paragraph_processor.ParagraphProcessor): MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(), IgnoreNotesHeader(), paragraph_processor.IgnoreTagMatcher('PRTPAGE')] def additional_constraints(self): return [optional_rules.limit_paragraph_types( mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)] class NoteMatcher(paragraph_processor.BaseMatcher): """Processes the contents of NOTE and NOTES tags using a NoteProcessor""" def matches(self, xml): return xml.tag in ('NOTE', 'NOTES') def derive_nodes(self, xml, processor=None): processor = NoteProcessor() node = Node(label=[mtypes.MARKERLESS], source_xml=xml, node_type=Node.NOTE) return [processor.process(xml, node)] ## Instruction: Use regex rather than string match for Note: ## Code After: import re from regparser.tree.depth import markers as mtypes, optional_rules from regparser.tree.struct import Node from regparser.tree.xml_parser import ( paragraph_processor, simple_hierarchy_processor) class IgnoreNotesHeader(paragraph_processor.BaseMatcher): """We don't want to include "Note:" and "Notes:" headers""" REGEX = re.compile('notes?:?\s*$', re.IGNORECASE) def matches(self, xml): return xml.tag == 'HD' and self.REGEX.match(xml.text or '') def derive_nodes(self, xml, processor=None): return [] class NoteProcessor(paragraph_processor.ParagraphProcessor): MATCHERS = [simple_hierarchy_processor.DepthParagraphMatcher(), IgnoreNotesHeader(), paragraph_processor.IgnoreTagMatcher('PRTPAGE')] def additional_constraints(self): return [optional_rules.limit_paragraph_types( mtypes.lower, mtypes.ints, mtypes.roman, mtypes.markerless)] class NoteMatcher(paragraph_processor.BaseMatcher): """Processes the contents of NOTE and NOTES tags using a NoteProcessor""" def matches(self, xml): return xml.tag in ('NOTE', 'NOTES') def derive_nodes(self, xml, processor=None): processor = NoteProcessor() node = Node(label=[mtypes.MARKERLESS], source_xml=xml, node_type=Node.NOTE) return [processor.process(xml, node)]
// ... existing code ... import re from regparser.tree.depth import markers as mtypes, optional_rules // ... modified code ... """We don't want to include "Note:" and "Notes:" headers""" REGEX = re.compile('notes?:?\s*$', re.IGNORECASE) def matches(self, xml): return xml.tag == 'HD' and self.REGEX.match(xml.text or '') // ... rest of the code ...